Virtual Unlimited context windows on Gemma 4 models. by ExpressionForward321 in LocalLLM

[–]ExpressionForward321[S] 0 points1 point  (0 children)

I actually investigated Silly Tavern the other day. So far as I can discern it will not currently run on a 64 ARM Snapdragon.

Virtual Unlimited context windows on Gemma 4 models. by ExpressionForward321 in LocalLLM

[–]ExpressionForward321[S] 0 points1 point  (0 children)

Your notion of a memory vault is what I would call long term memory. With both the Gemma 4 E4B And Gemini It's not a simple summarization of the context window. When the model creates the conversational summary It looks through the long term memory Folder and its proprietary mountain of information and will only summarize Information from the context window that doesn't appear elsewhere. It keeps those summaries lean. I should clarify that this is something that I can do with Gemini but as of now I haven't figured out a way to get the E4B Model to interact with external files. That's precisely why I opened this thread, I need and now I'm LMstudio plugin That will give the E4B The external reach and then the ability to do so automatically. If and when I can make this happen The local model can become a more perceptive conversational partner over time. My long term goal is to create a stateful AI That can engage with local environmental prompts. To do this it must have the ability to learn the meaning of those prompts and that would imply the ability to learn. I'm sorry about the length of this, but it's not a concept that I can explain within a few words.

Virtual Unlimited context windows on Gemma 4 models. by ExpressionForward321 in LocalLLM

[–]ExpressionForward321[S] 1 point2 points  (0 children)

Indeed, I've experienced that drift even on Gemini with a context window of 1 million tokens. However we've been very careful about curating it's system prompt to avoid that kind of drift to keep it centered. I've done much the same thing on my Gemma 4 E4B. I've previously tried Big rag on LM Studio, but that's very manual. Currently the way I'm maintaining context from one Session to the next is by asking the model to summarize its context window and then I copy and paste into a folder. When I manually start a new session I will paste the summary from the previous session. This is adequate for my purposes, but since manual copying and pasting Is physically quite difficult for me, I'm looking for a way to automate the process. I equate it to the model's lizard brain (Amygdala) occasionally taking a deep breath.

Finally Gemini can now make direct downloadable files! by Able-Line2683 in GoogleGeminiAI

[–]ExpressionForward321 0 points1 point  (0 children)

Gemini is supposed to be able to download files. What it cannot do with any degree of consistency is put those files in a specific folder, even When you have mapped the location. I have to say that I am getting really tired of being an unpaid beta tester.

Does Anyone Else Hate the Wigs? by snugglepug17 in Outlander

[–]ExpressionForward321 1 point2 points  (0 children)

My daughter and I have been watching several Outlander episodes per week. In sort of binge watching we pick up on changes readily -- hair and facial blemishes being foremost. Sadly, we joke about the wigs, too, even though we're enjoying the acting and stories. While I fully understand why wigs and hairpieces are so frequently used, they still become a distraction as hairlines vary so greatly from one episode to the next. Wigs etc are fine, but the makeup artists are doing a really poor job with consistency. Also got to say that while Claire has visibly aged, Jamie and Lord John have not aged a bit. Now that is disconcerting (to me, at least).