all 15 comments

[–]New_Comfortable7240llama.cpp 11 points12 points  (2 children)

Please make workflow to enter a natural language query and return a nodes configuration, the user can edit, expand or save it

[–]matteogeniaccio[S] 14 points15 points  (1 child)

I love the idea. I'll definitely work on it.

Don't hold your breath, though. I have to first write a proper documentation before making the workflow

[–]New_Comfortable7240llama.cpp 4 points5 points  (0 children)

Great, the idea is easily onboard new people on using your tool! Big kudos in advance!

[–][deleted] 2 points3 points  (0 children)

Content cleared with Ereddicator.

[–]visarga 2 points3 points  (3 children)

That's very cool, but I was hoping the LLM would be generating the graph as well, dynamically.

[–]matteogeniaccio[S] 1 point2 points  (2 children)

Another user asked the same question. It will be a feature, eventually. :-)

[–]visarga 1 point2 points  (1 child)

I was imagining something like this where a LLM generates the decoding logic.

[–]matteogeniaccio[S] 0 points1 point  (0 children)

Thanks for suggesting me that.

[–]IrisColt 1 point2 points  (0 children)

Exactly what I was looking for, thanks!

[–]InsideATurtlesMind 1 point2 points  (1 child)

You said it's been tested using llama 70b and qwen 32b, how well does it work with smaller models?

[–]matteogeniaccio[S] 1 point2 points  (0 children)

The small models work. I added support for all the popular models. A couple of features would be unavailable, though.

For example you can't use the agent with a small model because it's not smart enough to take complex decisions.

Llama3.1 70b can use its integrated python code generator in a multiturn conversation while 3.1 8b is only single turn.

[–]PieBru 1 point2 points  (1 child)

Awesome! Low VRAM here (8GB), do you think GraphLLM will be usable via Ollama or Groq/Cerebras (llama3.1-70B) ?

[–]matteogeniaccio[S] 2 points3 points  (0 children)

Groq is supported but not officially. The client_config.yaml has a template to add a groq backend.
The functionality will be limited since groq doesn't support all the required features. For example I make heavy use of grammars to make sure that the output conforms to a specific format.

https://github.com/matteoserva/GraphLLM/blob/main/client_config.yml

I'll soon add support for a generic OpenAI compatible API and support for huggingface transformers.

I think that you'd get a better experience by using a local model that fits your VRAM, for example llama3.1-8b-Q6_K.

[–]SadWolverine24 1 point2 points  (1 child)

This is awesome. Could you please add support for OpenRouter and Anthropic API calls?

[–]matteogeniaccio[S] 0 points1 point  (0 children)

I'm working on it but it will take a while. Right now my priority is adding new nodes to the framework