Skip to main content

All Questions

0 votes
1 answer
809 views

ModuleNotFoundError: No module named 'huggingface_hub.inference._types'

I am running a RAG pipeline, with LlamaIndex and quantized LLama3-8B-Instruct. I just installed these libraries: !pip install --upgrade huggingface_hub !pip install --upgrade peft !pip install llama-...
Hoang Cuong Nguyen's user avatar
1 vote
1 answer
1k views

LlamaParse not able to parse documents inside directory

Whenever I try to use LlamaParse I get an error that states the file_input must be a file path string, file bytes, or buffer object. parser = LlamaParse(result_type="markdown") ...
verstandskies's user avatar
0 votes
0 answers
99 views

Maximum recursion depth exceeded when using ColBERT reranker for Llama 3

I am looking to deploy ColBERT reranker for my RAG pipeline, with a T4 GPU (the LLM that I am using is Meta-LLaMa-3-8B-Instruct, which has already been quantized to 4bit): import torch from ...
Hoang Cuong Nguyen's user avatar
0 votes
0 answers
137 views

GoogleGenerativeAIError: Error embedding content - Using LlamaIndex

I am trying to use GoogleGenerativeAIEmbeddings from llamaindex for creating the embeddings, but from past two days it is giving the error as: GoogleGenerativeAIError: Error embedding content: ...
chandu135's user avatar
0 votes
0 answers
116 views

Getting RateLimitError: Error code: 429 while evaluating query engine despite using OpenSource LLM Meta Llama2-70B-Chat-hf

I've been working on developing a Question-Answering Bot utilizing RAG (Retrieval-Augmented Generation) architecture. To achieve this, I've integrated LlamaIndex and the LLama2-70B-chat-hf model via ...
Muhammad Haseeb's user avatar
0 votes
1 answer
438 views

Is there any need to perform preprocessing while using LlamaParser & MarkdownElementNodeParser? how can I add additional steps for Preprocessing?

Improved response: I am developing a Retrieval Augmented Generation (RAG) based application that involves extracting data from a large PDF file containing over 200 pages. To accomplish this, I am ...
Muhammad Haseeb's user avatar
1 vote
0 answers
853 views

How to return the source document of retrieved nodes following llama index engine query?

My goal is rather straightforward: I want to return the document from which my nodes are derived for my RAG application. I want to do this so I can know which document is being most frequently ...
JP1990's user avatar
  • 53
0 votes
0 answers
578 views

Using llama index but avoiding the tiktoken API call

I want to use llama_index but when I import the package I get the following error ConnectionError: ('Connected aborted.', ConnectionResetError(10054, 'An existing connection was forcibly closed by the ...
Eisen's user avatar
  • 1,927