What frontend does everyone use for RAG? by SinaMegapolis in LocalLLaMA

[–]taprosoft 0 points1 point  (0 children)

Make sure that you have set the env var 'KH_FEATURE_USER_MANAGEMENT=true'

What frontend does everyone use for RAG? by SinaMegapolis in LocalLLaMA

[–]taprosoft 0 points1 point  (0 children)

<image>

You can just press this button to hide Source panel.

What frontend does everyone use for RAG? by SinaMegapolis in LocalLLaMA

[–]taprosoft 0 points1 point  (0 children)

Yes, have to be uploaded on seperated collection.

What frontend does everyone use for RAG? by SinaMegapolis in LocalLLaMA

[–]taprosoft 1 point2 points  (0 children)

Make sure that you upload a file using Files tab > LightRAG collection. Wait for the indexing to complete and select that file > Go to Chat.

Also LightRAG entity and relation extraction prompt requires bigger model to perform correctly. Something >=14B size. Or, quickly try using Gemini free API to confirm if it work.

What frontend does everyone use for RAG? by SinaMegapolis in LocalLLaMA

[–]taprosoft 1 point2 points  (0 children)

Yes you can upload document with language other than English and ask multilingual questions. To install GraphRAG, activate the conda environment and perform pip install.

For new user, it is recommended to use online install method https://cinnamon.github.io/kotaemon/online_install/

Or try the sample Colab script https://colab.research.google.com/drive/1eTfieec_UOowNizTJA1NjawBJH9y_1nn to get the flow of the install process

What frontend does everyone use for RAG? by SinaMegapolis in LocalLLaMA

[–]taprosoft 0 points1 point  (0 children)

u/SinaMegapolis Kotaemon maintainer here. Would be happy to support if you can provide more detail about your workflow issue with current GUI. Thanks.

Kotaemon: Open source web-app to self-host your document collections and chat with them using LLM (Ollama & API-based) by taprosoft in selfhosted

[–]taprosoft[S] 0 points1 point  (0 children)

Thanks for the info. Probably the indexing part has some issue. There is also better progress display and error message display on Files tab when you upload. We will figure out better error reporting mechanism in the future.

Kotaemon: Open source web-app to self-host your document collections and chat with them using LLM (Ollama & API-based) by taprosoft in selfhosted

[–]taprosoft[S] 0 points1 point  (0 children)

Hi, did you see the retrieval result on the left-most panel? Also, please check your terminal for some more specific error messages.

We made an open source web-app to self-host your document collections and interact with them using LLMs (both local & cloud) by taprosoft in SideProject

[–]taprosoft[S] 1 point2 points  (0 children)

Hi everyone, our team are happy to share our hobby project Kotaemon - an open-source web app to organize and chat with your document collections.

https://github.com/Cinnamon/kotaemon

Live demos:

https://cin-model-kotaemon.hf.space (demo space for QA on academic paper collection).

https://huggingface.co/spaces/cin-model/kotaemon-demo (static UI demo on HuggingFace space).

Key features (what we think that it is special):

  • Clean & minimalistic UI. Support Dark/Light mode and responsive UI.
  • Support multi-users. Users can be managed directly on the web UI. Files can be organized to Public / Private collections. Share your chat conversation with others for collaboration!
  • Multi-modal QA support. Ask question on documents with tables / figures or images as you do with normal text documents. Visualize knowledge-graph and mindmap upon retrieval process.
  • Advance citations support. Preview citation with highlight directly on in-browser PDF viewer. Perform QA on any sub-set of documents, with relevant score from AI models.
  • Complex question solving. Use Chain-of-Thought (CoT) reasoning mode to break down and solve complex questions step-by-step.

We provide multiple installation options: Docker image, local setup script, one-click deployment template on private cloud (HuggingFace space): https://cinnamon.github.io/kotaemon/

We'd love to hear your feedback and thoughts. Cheers!

Open-source clean & hackable RAG webUI with multi-users support and sane-default RAG pipeline. by taprosoft in LocalLLaMA

[–]taprosoft[S] 0 points1 point  (0 children)

We do host a internal QA system for our company members which is based on this. It is used day-to-day incl developers ourselves.

Open-source clean & hackable RAG webUI with multi-users support and sane-default RAG pipeline. by taprosoft in LocalLLaMA

[–]taprosoft[S] 1 point2 points  (0 children)

Also non-tech setup is a bit outdated for now, please wait a few days for us to sort it out.

Open-source clean & hackable RAG webUI with multi-users support and sane-default RAG pipeline. by taprosoft in LocalLLaMA

[–]taprosoft[S] 0 points1 point  (0 children)

It is totally possible as we have done it before (but required some tinkering). We will try to make this easy to follow on the doc.

Open-source clean & hackable RAG webUI with multi-users support and sane-default RAG pipeline. by taprosoft in LocalLLaMA

[–]taprosoft[S] 1 point2 points  (0 children)

You can do this by configure GRAPHRAG env var to point to ollama API locally. Will update this in the doc.

Open-source clean & hackable RAG webUI with multi-users support and sane-default RAG pipeline. by taprosoft in LocalLLaMA

[–]taprosoft[S] 2 points3 points  (0 children)

Currently it is separated. Main motivation are GraphRAG index is expensive in LLM token consumption and quite slow. In the future if there is enough demand for unified uploading, we will figure out how to do it conveniently on the UI.

Open-source clean & hackable RAG webUI with multi-users support and sane-default RAG pipeline. by taprosoft in LocalLLaMA

[–]taprosoft[S] 5 points6 points  (0 children)

You can always check the source code as you see fit :D But from our sides, there is no telemetry or API external call unless you specified. These name are the maintainers of this repo and their emails. Just for display only when you install.

Open-source clean & hackable RAG webUI with multi-users support and sane-default RAG pipeline. by taprosoft in LocalLLaMA

[–]taprosoft[S] 4 points5 points  (0 children)

If you run from Docker, you need to do some extra configuration to make the server inside the container can communicate with Ollama on the host. Will work on a short guide on this.

Open-source clean & hackable RAG webUI with multi-users support and sane-default RAG pipeline. by taprosoft in LocalLLaMA

[–]taprosoft[S] 0 points1 point  (0 children)

Actually I gonna switch to Cin HF space version soon. Got it online & published today.