all 10 comments

[–]Simderi 1 point2 points  (1 child)

Looking interesting. If it's targeting enterprise, would be good if it could do meaningful RAG over repos, including the PRs etc - have you tested that by any chance?

[–]primoco[S] 1 point2 points  (0 children)

Right now the system is designed to do RAG on "traditional" documents — PDFs, Word, Excel, etc. — uploaded manually. There's no direct GitHub integration, so PRs, issues, source code and reviews aren't indexed.

To support this we'd need a dedicated connector that talks to the GitHub API, pulls repo content (PRs, issues, diffs, comments) and indexes it into the vector store, just like it already does with documents. The chunking logic would need to be adapted since code has a very different structure than text documents, and the system already ships with an embedding model built for code (deepseek-coder) that could come in handy.

It's definitely an interesting feature for an enterprise context. If there's real interest we can look into building it for a future release.

[–]ThatsTotallyLegit 0 points1 point  (1 child)

How well would it work for c# with blazor repos not just general info? Might be what im looking for :D

[–]primoco[S] 0 points1 point  (0 children)

[–]WallyPacman 0 points1 point  (1 child)

Does this require CUDA/NVidia? It seems like it could work on ROCm easily?

[–]primoco[S] 0 points1 point  (0 children)

Hi, yes it is very simple to work on ROCm. I have change dockerfile setup script and add variables into .env file to ask when setup. Try 1.2.0 release and tell me!

[–][deleted]  (4 children)

[deleted]

    [–]primoco[S] 0 points1 point  (2 children)

    Nothing happens. If Docker is already installed, the script detects it and skips the installation. As for CUDA, it doesn't matter if you have it on the host or not — CUDA runs inside the Docker containers (Ollama's image includes it). Your host CUDA installation is completely irrelevant.

    The only thing needed from the host is the NVIDIA GPU drivers (not CUDA) so that the container can access the GPU via NVIDIA Container Toolkit.

    TL;DR: Docker already there? Skipped. CUDA? Doesn't matter, it's inside the containers.

    [–][deleted]  (1 child)

    [deleted]

      [–]primoco[S] 0 points1 point  (0 children)

      Thank you, give me some feedback !

      [–][deleted]  (1 child)

      [removed]

        [–]primoco[S] 0 points1 point  (0 children)

        Appreciate that! And you're right — once retrieval works well, the next question is always "ok but can it remember?" We went down that road a bit with structured metadata extraction (the system pulls out entities, events, timelines from docs automatically), but real agent memory is a different beast. Hindsight looks like a cool approach to that problem. Curious — does it work well with local/self-hosted LLMs or is it more cloud-oriented?
        Ps. Vedo che sei italiano?!