All Questions
8 questions
0
votes
1
answer
809
views
ModuleNotFoundError: No module named 'huggingface_hub.inference._types'
I am running a RAG pipeline, with LlamaIndex and quantized LLama3-8B-Instruct. I just installed these libraries:
!pip install --upgrade huggingface_hub
!pip install --upgrade peft
!pip install llama-...
1
vote
1
answer
1k
views
LlamaParse not able to parse documents inside directory
Whenever I try to use LlamaParse I get an error that states the file_input must be a file path string, file bytes, or buffer object.
parser = LlamaParse(result_type="markdown")
...
0
votes
0
answers
99
views
Maximum recursion depth exceeded when using ColBERT reranker for Llama 3
I am looking to deploy ColBERT reranker for my RAG pipeline, with a T4 GPU (the LLM that I am using is Meta-LLaMa-3-8B-Instruct, which has already been quantized to 4bit):
import torch
from ...
0
votes
0
answers
137
views
GoogleGenerativeAIError: Error embedding content - Using LlamaIndex
I am trying to use GoogleGenerativeAIEmbeddings from llamaindex for creating the embeddings, but from past two days it is giving the error as: GoogleGenerativeAIError: Error embedding content: ...
0
votes
0
answers
116
views
Getting RateLimitError: Error code: 429 while evaluating query engine despite using OpenSource LLM Meta Llama2-70B-Chat-hf
I've been working on developing a Question-Answering Bot utilizing RAG (Retrieval-Augmented Generation) architecture. To achieve this, I've integrated LlamaIndex and the LLama2-70B-chat-hf model via ...
0
votes
1
answer
438
views
Is there any need to perform preprocessing while using LlamaParser & MarkdownElementNodeParser? how can I add additional steps for Preprocessing?
Improved response:
I am developing a Retrieval Augmented Generation (RAG) based application that involves extracting data from a large PDF file containing over 200 pages. To accomplish this, I am ...
1
vote
0
answers
853
views
How to return the source document of retrieved nodes following llama index engine query?
My goal is rather straightforward: I want to return the document from which my nodes are derived for my RAG application. I want to do this so I can know which document is being most frequently ...
0
votes
0
answers
578
views
Using llama index but avoiding the tiktoken API call
I want to use llama_index but when I import the package I get the following error
ConnectionError: ('Connected aborted.', ConnectionResetError(10054, 'An existing connection was forcibly closed by the ...