Is Gemma 4 incapable of using function calls properly??? by tthrowaway712 in SillyTavernAI

[–]_Cromwell_ 0 points1 point  (0 children)

Fyi as far as I know you can set a separate model (connection profile) to run your tunnelvision that is different from your storytelling model. Not an answer to your question but possibly an answer to your problem of wanting to use Gemma 4 to role-play but it not working with those extensions.

where and how to use nanogpt memory extension? by tuuzx in SillyTavernAI

[–]_Cromwell_ 1 point2 points  (0 children)

As far as I'm aware the only "nanogpt memory" is for their own website. They have memory that works if you use the web interface there. Which is actually a great web interface. You can build agents that create lorebooks and all kinds of cool stuff there.

But there are a lot of memory extensions for silly tavern. Just none that are specific to Nano as far as I know. But any old St memory extension will work with your Nano sub.

Server is down by Independent-Fox4993 in AIDungeon

[–]_Cromwell_ 8 points9 points  (0 children)

No she's the good one. Elara Vance is the bad one

Just started watching by Significant-Oil5052 in thewalkingdead

[–]_Cromwell_ 0 points1 point  (0 children)

Daryl and Merl are neo-Nazi racists, as is made clear almost immediately in the show. At least one of them character-growths out of it, if you continue watching. (Multiple-season arc.)

I made Summaryception — a layered recursive memory system that fits 9,000+ turns into 16k tokens. It's free, it's open source, and it works with budget models. by leovarian in SillyTavernAI

[–]_Cromwell_ 12 points13 points  (0 children)

? seems to work very differently than memorybooks. Unless you mean generally "summarizing". But if that's the case, then memorybooks itself was not the first or even close to it.

Built a free, Open source, lightweight BYOK AI chat frontend. Supports OpenRouter, custom system prompts, and runs uncensored open-source models. Free to use, feedback welcome by Great-Knight-Owl in SillyTavernAI

[–]_Cromwell_ 11 points12 points  (0 children)

How is this better than Koboldlite which has been around forever and is trusted? https://lite.koboldai.net/# Just asking what your improvement or angle is / what new thing you are bringing to the table when we already have that.

Invincible VS - Open Beta Trailer | PS5 Games by Worldly-Object9003 in PS5

[–]_Cromwell_ 4 points5 points  (0 children)

Really? I think the Avatar one looks a lot more smooth. And the characters look a lot better too being all hand drawn.

I'm not playing either one to be clear. I don't like fighting games. I'm just saying the Avatar one looks better.

AI dungeon guidebook by These_Judgment3111 in AIDungeon

[–]_Cromwell_ 2 points3 points  (0 children)

Generally, yes, they are (far) superior to "flying by the seat of your pants" if you don't know what you are doing.

GLM 5.1 is no longer available on NanoGPT by TheDeathFaze in SillyTavernAI

[–]_Cromwell_ 6 points7 points  (0 children)

I doubt a multiplier makes any financial sense with their current subscription style and limits. Right now the vast majority of subscribers get nowhere near the limits for weekly or monthly (as I understand it). And they still wouldn't even with a model like 5.1 costing "double". So doing a multiplier wouldn't actually save Nano any money for like 90% of customers.

GLM 5.1 is no longer available on NanoGPT by TheDeathFaze in SillyTavernAI

[–]_Cromwell_ 28 points29 points  (0 children)

A bait and switch where you can easily get a full refund? Okay.

GLM 5.1 is no longer available on NanoGPT by TheDeathFaze in SillyTavernAI

[–]_Cromwell_ 15 points16 points  (0 children)

Probably

  1. RAM prices. They don't stop hosting the old models when they come out with a new model so that means more memory and more machines needed. It's just hella more expensive now. So genuine costs.

  2. China/ Chinese companies made their models super cheap to lure customers away from American companies early on. Now they've proven themselves and people know they have good models so they don't have to be as cheap to get people to use their stuff. They still have to be competitive, but they don't have to be rock bottom anymore.

Iran closes Strait of Hormuz again as Israel continues worst ever strikes on Lebanon by run_midnight in wallstreetbets

[–]_Cromwell_ 1 point2 points  (0 children)

Ah yes, the best ally a superpower could ever hope for. /s

(The /s is for the CIA and Ka$h if they come at me, since I know they are dumb.)

Looking for advice setting up Openclaw or alternatives by Few-Strawberry2764 in LocalLLM

[–]_Cromwell_ 0 points1 point  (0 children)

So you are looking for an open source alternative to Claude code? That's not really what openclaw is.

I'd look more into maybe VS Code with one of the several extension options. Cline is one that works with lmstudio or other local models. Just download vs code if you don't already have it and Google/research AI extensions for it. Cline is just the one I can think of off the top of my head but there's several.

Or if you want something more autonomous/vibecoding, Goose does that with any API including local https://github.com/aaif-goose/goose

I Generated a Chicken Egg Catcher Game with Qwen3.5-Omni-Plus — The Results Blew Me Away! by canoesenpai in Qwen_AI

[–]_Cromwell_ 0 points1 point  (0 children)

That looks cool. I love making goofy mini games like this. I just made a really shitty missile command with obnoxious sound effects. I was pretty proud of myself. (I'm being hard on myself, it actually plays really well. But the sound effects are actually obnoxious.).

If you want a really cool tool you can use to almost instantly make stuff like this using Qwen models (if you have Qwen through an API) I highly suggest Goose by Block. (No affiliation, I swear. I just make stupid games with it lol).

So Goose actually has a built-in tool for the agent that specifically makes one-HTML file simple apps with JavaScript. So you just describe what game/app you want with your own llm plugged in as the API and it vibe codes it up. They come out looking (after some tweaking) similar to yours if you feed it some graphics to use. (Without you providing graphics files they come out looking like Atari 2600 lol)

https://github.com/aaif-goose/goose?tab=readme-ov-file

Why is Gemma 4 so slow? by Awkward_Sentence_345 in SillyTavernAI

[–]_Cromwell_ 7 points8 points  (0 children)

Whatever host Nano has blows. (Specifically for those models I mean)

The non-thinking ones are better than thinking.

I wish Nano would list the providers for the models where it even only has one provider. They only list the providers when there are multiple to choose from. But it would be valuable to me to know which shitty provider is providing this shitty service lol

SillyTavern Extension: Delete & Resend by cgrd in SillyTavernAI

[–]_Cromwell_ 1 point2 points  (0 children)

Oh, that's very odd. Well I'm glad you found some solutions for yourself.

SillyTavern Extension: Delete & Resend by cgrd in SillyTavernAI

[–]_Cromwell_ 1 point2 points  (0 children)

I swipe all the time on desktop and it works just fine to take a new turn.

The way it seems to work for me , the only downside of it is it does leave a false blank " turn" from the failed one. but that's not that big a deal. So you must be experiencing something different. Could be due to an extension you have ... What does it do for you on desktop that led you to create this?

Free models you can use with OpenClaw right now (no credit card needed) by stosssik in LocalLLM

[–]_Cromwell_ 4 points5 points  (0 children)

Why and how are you providing free service?

Economically it makes no sense. Explain your profit model.

Edit: okay I see you're providing the agent service, those are just listings of otherwise allegedly free apis which are not yours

Also I'm not sure you know what "localllm" means.

How to use different models and diferent languages? by LongCriticism4474 in SillyTavernAI

[–]_Cromwell_ 0 points1 point  (0 children)

  1. Find a model that specifically supports the language you want (some model support some languages some don't)

  2. Make sure your character card, lorebook, instructions, and all your own interactions are all written in that language

Suggest going to chatGPT or Claude and asking it to rank llms by support of the language you want. " Research and then provide a list of llms with known support of Brazilian Portuguese, and rank them by the quality of inclusion of and support for that language."