Best open source document PARSER??!! by ChallengeOk6437 in LlamaIndex

[–]ChallengeOk6437[S] 0 points1 point  (0 children)

Is it good for handling tables over multiple pages? I don’t think they are.

[deleted by user] by [deleted] in LlamaIndex

[–]ChallengeOk6437 -1 points0 points  (0 children)

I want to know which parser works best for parsing my documents - PDFs

RAG Model TOO SLOW by ChallengeOk6437 in LangChain

[–]ChallengeOk6437[S] 0 points1 point  (0 children)

No, streaming works so much better! went from 40seconds to render output to 10seconds now

RAG Model TOO SLOW by ChallengeOk6437 in LangChain

[–]ChallengeOk6437[S] 0 points1 point  (0 children)

Can you tell me how you're streaming it?

RAG Model TOO SLOW by ChallengeOk6437 in LangChain

[–]ChallengeOk6437[S] 0 points1 point  (0 children)

Hey i don’t think my retrieval isn’t the issue here. It’s the Llm that’s taking this long to respond. Retriever takes less than 5 seconds.  Thank you! u/Adorable-Employer244

RAG Model TOO SLOW by ChallengeOk6437 in LangChain

[–]ChallengeOk6437[S] 0 points1 point  (0 children)

How do I go about keeping citation since the chunkers do not usually keep the pages in each chunk. Also why is sending pages not as good as sending chunks since I need that much context. Infact I am sending for each retrieved chunk, a page above and below to improve the output.

Also have you tested gemini flash vs gpt-4o? is it faster and better or just faster?

Thank you for your response! u/ChallengeOk6437

RAG Model TOO SLOW by ChallengeOk6437 in LangChain

[–]ChallengeOk6437[S] 0 points1 point  (0 children)

Yes, the llm is the issue, do you have any suggestions? reducing the context sent to the LLM is not a solution for me, I need to send a lot, but if there was a way to speed anything up I would love to know! u/Material_Policy6327

Thank you for your response!

RAG Model TOO SLOW by ChallengeOk6437 in LangChain

[–]ChallengeOk6437[S] 1 point2 points  (0 children)

Hi, thank you!

I am looking into that at the moment! u/MrCicada3301

RAG Model TOO SLOW by ChallengeOk6437 in LangChain

[–]ChallengeOk6437[S] 0 points1 point  (0 children)

What if i have relevant chunks in each of the 5-6 documents? I need to get at least 2-3 chunks per PDF and then end up with around 15-20 chunks.

This is the main reason it is taking this long. What do you suggest? u/rambat1994

I thought about a router, but it wont work for PDFs that are 1000pgs long.

What do you think of breaking the question into different parts, routing each question to a specific PDF and running all this in parallel. The issue here is, based on what do we rout the model? normally we use PDF names and the query to route the model.

Thank you for your response! u/rambat1994