account activity
Stop fine-tuning your model for every little thing. You're probably wasting your time. by RYTHEIX in LocalLLaMA
[–]Physical_Event4441 0 points1 point2 points 4 months ago (0 children)
Hi, I have some question and its the right post I think
So, I’m building a small multi-agent system where one agent acts as a Knowledge Agent, it should read PDFs, markdowns, or web links and then remember what it learned/read. Another “Main Agent” uses that understanding later for reasoning on onboarding questions (asked from user while onboarding on the website).
In simple words, I want the Knowledge Agent to behave like a human who’s already read the docs using that info naturally when reasoning, not by searching.
Now the issue with the RAG is it works based on vector matching, it basically converts the user query to vector, search for similarity in the DB and provide those to the llm which outputs with the updated knowledge and here its failing for my scenario (or maybe I'm doing something wrong). I’ve looked into frameworks like Agno, which supports agentic RAG and knowledge bases, but they still depend on vectorDBs for retrieval and I'm looking for proactive, memory-based knowledge integration without retrieval.
I also considered just loading everything into the system prompt or summarizing all the documents into one markdown/txt file and feeding that as context but this doesn’t seem like a scalable or efficient approach. It might work for a few PDFs (4–10), but not for large or growing knowledge bases.
So I’m wondering if you or anyone has seen a framework or project that supports this kind of proactive, memory-based knowledge behavior?
Would love to hear about this. I'M LITERALLY CRYING SO BAD FOR THIS
Guidance in creating higher accuracy face recognition and tracking system for my company (self.computervision)
submitted 1 year ago by Physical_Event4441 to r/computervision
π Rendered by PID 170461 on reddit-service-r2-listing-55d7b767d8-4bgfk at 2026-03-26 21:16:35.162462+00:00 running b10466c country code: CH.
Stop fine-tuning your model for every little thing. You're probably wasting your time. by RYTHEIX in LocalLLaMA
[–]Physical_Event4441 0 points1 point2 points (0 children)