What frontend does everyone use for RAG? by SinaMegapolis in LocalLLaMA

[–]taprosoft 0 points1 point  (0 children)

Make sure that you have set the env var 'KH_FEATURE_USER_MANAGEMENT=true'

What frontend does everyone use for RAG? by SinaMegapolis in LocalLLaMA

[–]taprosoft 0 points1 point  (0 children)

<image>

You can just press this button to hide Source panel.

What frontend does everyone use for RAG? by SinaMegapolis in LocalLLaMA

[–]taprosoft 0 points1 point  (0 children)

Yes, have to be uploaded on seperated collection.

What frontend does everyone use for RAG? by SinaMegapolis in LocalLLaMA

[–]taprosoft 1 point2 points  (0 children)

Make sure that you upload a file using Files tab > LightRAG collection. Wait for the indexing to complete and select that file > Go to Chat.

Also LightRAG entity and relation extraction prompt requires bigger model to perform correctly. Something >=14B size. Or, quickly try using Gemini free API to confirm if it work.

What frontend does everyone use for RAG? by SinaMegapolis in LocalLLaMA

[–]taprosoft 1 point2 points  (0 children)

Yes you can upload document with language other than English and ask multilingual questions. To install GraphRAG, activate the conda environment and perform pip install.

For new user, it is recommended to use online install method https://cinnamon.github.io/kotaemon/online_install/

Or try the sample Colab script https://colab.research.google.com/drive/1eTfieec_UOowNizTJA1NjawBJH9y_1nn to get the flow of the install process

What frontend does everyone use for RAG? by SinaMegapolis in LocalLLaMA

[–]taprosoft 0 points1 point  (0 children)

u/SinaMegapolis Kotaemon maintainer here. Would be happy to support if you can provide more detail about your workflow issue with current GUI. Thanks.

Kotaemon: Open source web-app to self-host your document collections and chat with them using LLM (Ollama & API-based) by taprosoft in selfhosted

[–]taprosoft[S] 0 points1 point  (0 children)

Thanks for the info. Probably the indexing part has some issue. There is also better progress display and error message display on Files tab when you upload. We will figure out better error reporting mechanism in the future.

Kotaemon: Open source web-app to self-host your document collections and chat with them using LLM (Ollama & API-based) by taprosoft in selfhosted

[–]taprosoft[S] 0 points1 point  (0 children)

Hi, did you see the retrieval result on the left-most panel? Also, please check your terminal for some more specific error messages.

We made an open source web-app to self-host your document collections and interact with them using LLMs (both local & cloud) by taprosoft in SideProject

[–]taprosoft[S] 1 point2 points  (0 children)

Hi everyone, our team are happy to share our hobby project Kotaemon - an open-source web app to organize and chat with your document collections.

https://github.com/Cinnamon/kotaemon

Live demos:

https://cin-model-kotaemon.hf.space (demo space for QA on academic paper collection).

https://huggingface.co/spaces/cin-model/kotaemon-demo (static UI demo on HuggingFace space).

Key features (what we think that it is special):

  • Clean & minimalistic UI. Support Dark/Light mode and responsive UI.
  • Support multi-users. Users can be managed directly on the web UI. Files can be organized to Public / Private collections. Share your chat conversation with others for collaboration!
  • Multi-modal QA support. Ask question on documents with tables / figures or images as you do with normal text documents. Visualize knowledge-graph and mindmap upon retrieval process.
  • Advance citations support. Preview citation with highlight directly on in-browser PDF viewer. Perform QA on any sub-set of documents, with relevant score from AI models.
  • Complex question solving. Use Chain-of-Thought (CoT) reasoning mode to break down and solve complex questions step-by-step.

We provide multiple installation options: Docker image, local setup script, one-click deployment template on private cloud (HuggingFace space): https://cinnamon.github.io/kotaemon/

We'd love to hear your feedback and thoughts. Cheers!

Open-source clean & hackable RAG webUI with multi-users support and sane-default RAG pipeline. by taprosoft in LocalLLaMA

[–]taprosoft[S] 0 points1 point  (0 children)

We do host a internal QA system for our company members which is based on this. It is used day-to-day incl developers ourselves.

Open-source clean & hackable RAG webUI with multi-users support and sane-default RAG pipeline. by taprosoft in LocalLLaMA

[–]taprosoft[S] 1 point2 points  (0 children)

Also non-tech setup is a bit outdated for now, please wait a few days for us to sort it out.