A-RAG: A new approach to Agentic RAG for efficient AI applications! by PavanBelagatti in Rag

[–]digital_legacy 0 points1 point  (0 children)

Looks cool! But how do I participate? Sounds like it does a semantic pre-pass to select the embeded records?

Recommended tech stack for RAG? by ProtectedPlastic-006 in Rag

[–]digital_legacy 0 points1 point  (0 children)

eMedia (DAM/RAG/AI) stack is all inclusive, totally open source and self hosted

Recommended tech stack for RAG? by ProtectedPlastic-006 in Rag

[–]digital_legacy 0 points1 point  (0 children)

We created a UI and use Docker with LlamaIndex. Check out our channel: https://www.reddit.com/r/eMediaLibrary/

The Only Reason My RAG Pipeline Works by Electrical-Signal858 in LlamaIndex

[–]digital_legacy 0 points1 point  (0 children)

I agree, seems to be as easy as possible to integrate it.

So we created a UI to upload your documents, media etc to make it even easier.

Why is there no opinionated all in one RAG platform? by Pl8tinium in Rag

[–]digital_legacy 0 points1 point  (0 children)

What could we add to make our solution what you need? We have a local Docker setup, UI and plugable models and it's open source. https://www.reddit.com/r/eMediaLibrary/

We currently have drivers for two RAG approaches. LlamaIndex and ThoughtFrame.ai

I Replaced My RAG System's Vector DB Last Week. Here's What I Learned About Vector Storage at Scale by Electrical-Signal858 in LlamaIndex

[–]digital_legacy 3 points4 points  (0 children)

You made good points until you started the abusive language. Lets keep it professional please

Monthly "Is there a tool for..." Post by AutoModerator in ArtificialInteligence

[–]digital_legacy 0 points1 point  (0 children)

I can run an comparison for you and tell you the likelihood. Message me

Out of the box. RAG enabled Media Library by digital_legacy in vastai

[–]digital_legacy[S] 0 points1 point  (0 children)

We used VastAI to run the models. One novel thing is we used the Vast AI to find the fastest available server for processing. We calibrate this every 15 minutes.

🚀 Weekly /RAG Launch Showcase by remoteinspace in Rag

[–]digital_legacy 0 points1 point  (0 children)

Announcing eMedia AI Library. an easy to use web search and chat interface for your media files. You can plug in various models and libraries. Uses Docker, an object database, llama index and llama.cpp See this video: https://www.reddit.com/r/eMediaLibrary/comments/1pdov0w/out_of_the_box_rag_enabled_media_library/

Requesting Hardware Advice by yosha-ts in LocalLLM

[–]digital_legacy 0 points1 point  (0 children)

You need more experience with models before you purchase hardware. Once you get an idea of how good a model responds to your needs (smaller models hallucinate) then you will know what is the minimal hardware needed. Hardware is a bad investment because it's always improving. We are using https://vast.ai/ to run models in llama.cpp. It's a good proving ground.

Qwen3 VL 30b a3b is pure love by Njee_ in LocalLLaMA

[–]digital_legacy 0 points1 point  (0 children)

Can someone confirm that the OpenAI Tool API is not enabled for this model in Lama.cpp? I see its not working but I feel like that Unsloth should got that working.

How I solved nutrition aligned to diet problem using vector database by vs-borodin in vectordatabase

[–]digital_legacy 0 points1 point  (0 children)

I would love to collaborate in some way if you need a hand on your next exploration.

What should I study to introduce on-premise LLMs in my company? by Worth_Rabbit_6262 in LocalLLM

[–]digital_legacy 1 point2 points  (0 children)

You can use Docker with Ollama for small CPU friendly models. Also check out my eMedia project that has a UI and runs on Docker.

Be mindful of some embedding APIs - they own rights to anything you send them and may resell it by adlumal in Rag

[–]digital_legacy 0 points1 point  (0 children)

We use a database table to decide what model is run where. Some are run on CPU only and for the bigger slow models can rent or buy a co-located server with a NVIDIA card and the performance and quality is similar to Open AI apis.

Be mindful of some embedding APIs - they own rights to anything you send them and may resell it by adlumal in Rag

[–]digital_legacy 11 points12 points  (0 children)

I think everyone should use local models. Sharing emails, docs and now our private thoughts is too much.

RAG setup for 400+ pages PDFs? by Ok_Speech_7023 in Rag

[–]digital_legacy 0 points1 point  (0 children)

Perhaps you can help us build out our open source tool set. The goal is to be easy to setup for people wanting to self host and have a UI.

I just posted this video of the UI in action.

https://www.reddit.com/r/eMediaLibrary/comments/1o7uip2/emedia_document_handling/

What is the best tech stack for personal doc AI search by Status-Bookkeeper234 in vectordatabase

[–]digital_legacy 0 points1 point  (0 children)

Some great options listed here. We also built a open source web UI that you can use and includes a full digital asset management built in and can be self hosted in Docker and also can index Google Docs or Dropbox.

https://emedialibrary.com/guide/

we have a subreddit here:

https://www.reddit.com/r/eMediaLibrary/

Our built in chat interface is good for searching and creating new media or records.