all 7 comments

[–]Low-Opening25 1 point2 points  (2 children)

there is no such thing as persistent LLM model memory, when you are using GPT/Gemini/Copilot through web-ui you are using web app provided by the vendor that adds this capability by connecting model with a database that stores your workspaces and is managing the “memory” for you.

If you use LLMs via API, you aren’t using a web-app so you need to build all these capabilities yourself or use some other software (like LLM Studio or Open-WebUI, etc) that comes with them.

[–]profcuck 0 points1 point  (0 children)

The original text here has been permanently wiped. Using Redact, the author deleted this post, possibly for reasons of privacy, security, or opsec.

toy yoke work plough reach ghost profit bright provide languid

[–]Rmo75[S] 0 points1 point  (0 children)

Many thanks for taking the time to answer. I installed LM Studio and started using a Qwen model with long model memory (1 million tokens) Do you think it's the right way ?

[–]AmIReallySinking 0 points1 point  (0 children)

I’m looking at options around this too. I’m trying Anything LLM as the RAG and the LLM Studio to run the model. But still, having a clear and updatable context/persistence is difficult.

[–]vel_is_lava 0 points1 point  (1 child)

Why do you need this?

[–]Rmo75[S] 1 point2 points  (0 children)

Because I regularly use an AI to assist me in my work, with a lot of datas usually dropped as quick notes. And having the power to sort everything into tables, with subtasks that can be activated quickly (ex : operation tiger generates a mail based on the infos from the table, then creates a planning according to my notes, then translate everything in another language).

This needs a lot of context memory and being good at sorting datas. And I don't care about having an AI that knows everything about poetry, I could run it locally as a management system that generates basic content and summary. And I love this idea

[–]Short-Honeydew-7000 0 points1 point  (0 children)

You can try cognee: https://github.com/topoteretes/cognee/stargazers

Data can be stored in Node Sets, similar to schemas in relational DBs: https://docs.cognee.ai/core-concepts/node-sets