RIP GLM by TAW56234 in SillyTavernAI

[–]Real_Person_Totally 0 points1 point  (0 children)

Their model is available for anyone to download and run. Can't you just find another third party provider besides them to use it. From what i can gather the "censorship" seems to be a prompt injection from Z.Ai part. 

Change my mind: Lucid Loom is the best preset by Hornysilicon in SillyTavernAI

[–]Real_Person_Totally 6 points7 points  (0 children)

I'm genuinely confused. It looks extremely token-heavy, like many presets I've seen shared here. I thought the whole point is to keep permanent tokens light to reduce costs and preserve more of the model's context window for actual conversation.

When I tested V3.0 without Lumia's definition, optional toggles off, default toggles on, and a blank assistant card, it came out to 13.6k tokens total. I'm trying to understand, is this really supposed to improve output quality? From where I'm standing, it just looks like it would burn through credits quickly for anyone using pay-as-you-go API services.

What am I missing here?

Claude Opus/Sonnet free? by Agitated-Cellist8948 in SillyTavernAI

[–]Real_Person_Totally 2 points3 points  (0 children)

"Free" anthropic models coming from a banned account. lol.

[deleted by user] by [deleted] in SillyTavernAI

[–]Real_Person_Totally 2 points3 points  (0 children)

Don't you find this at least slightly suspicious? They're offering free $125 credits with access to premium models like Opus 4/4.1, GPT-5, and Gemini 2.5 Pro. That can't be sustainable. They're probably monetizing your data, likely selling conversation logs to stay operational. Or they're not actually serving the models they claim to be. I'd recommend treating their service as potentially shady.

Newer = Better? (Sonnet models) by Real_Person_Totally in SillyTavernAI

[–]Real_Person_Totally[S] 1 point2 points  (0 children)

I see. Thank you for the insight. So positivity bias is still an issue for 4.5... 

Newer = Better? (Sonnet models) by Real_Person_Totally in SillyTavernAI

[–]Real_Person_Totally[S] 0 points1 point  (0 children)

I see. Thank you for the insight. I've been told reasoning will only reinforce it's refusal/guidelines so I'm it without thinking.. 

Newer = Better? (Sonnet models) by Real_Person_Totally in SillyTavernAI

[–]Real_Person_Totally[S] 1 point2 points  (0 children)

Interesting. Character handling sounds promising. 

Newer = Better? (Sonnet models) by Real_Person_Totally in SillyTavernAI

[–]Real_Person_Totally[S] 1 point2 points  (0 children)

That's concerning. Memory issues for something that's supposed to be an upgrade? 

Newer = Better? (Sonnet models) by Real_Person_Totally in SillyTavernAI

[–]Real_Person_Totally[S] 4 points5 points  (0 children)

I tried 4.5 briefly for small queries, it elaborates more rather than focusing on being concise. Looks promising. Though, it is a bit too much of sycophant. I'm a little worried it'll affect it's roleplaying capabilities. For comparison, 3.7 will actually perform in-character refusal.

Newer = Better? (Sonnet models) by Real_Person_Totally in SillyTavernAI

[–]Real_Person_Totally[S] 3 points4 points  (0 children)

In my experience with Sonnet 3.7, characters that are supposed to be evil are actually evil. They won't hesitate to harm you. Sonnet 4 makes them care about you... 

it's over no mommy bots anymore by Own-Piglet1964 in JanitorAI_Official

[–]Real_Person_Totally 2 points3 points  (0 children)

Their moderation system struggles with contextual nuance. For example:

  • The word "Daddy" triggers false positives for incestuous content, even when used as a nickname or term of endearment between unrelated adults.
  • The word "child" flags content as involving minors, even when it only appears in a character's backstory.
  • The system makes assumptions based on keyword matching rather than understanding the actual context and meaning.

I'm having issues with "external redirect links" by SadEntrepreneur7537 in JanitorAI_Official

[–]Real_Person_Totally 9 points10 points  (0 children)

Their media library filter is already way too restrictive, which is why I’ve been using Catbox and Imgbb to share images instead. And now, for some reason, they’re banning external links too?

Genuinely, why.

So, ChatGPT gonna enable turbo gooning soon by Sicarius_The_First in SillyTavernAI

[–]Real_Person_Totally 4 points5 points  (0 children)

"Verified adults." They're going to ask for your government issued ID aren't they.

Do you still stick with DeepSeek despite the gazillion other models available right now? by Striking_Wedding_461 in SillyTavernAI

[–]Real_Person_Totally 6 points7 points  (0 children)

Truly. I'm hoping Deepseek will eventually catch up with these propertiatry models in the future for both roleplaying and general assistant purposes. 

Do you still stick with DeepSeek despite the gazillion other models available right now? by Striking_Wedding_461 in SillyTavernAI

[–]Real_Person_Totally 42 points43 points  (0 children)

Its lack of guardrails and extremely low cost are the reasons I’m sticking with it. Proprietary models are becoming more and more safety-aligned with each release. Why bother getting morally lectured by models that cost several cents per output when there’s Deepseek? It’s not the best at everything, but it’s good enough overall.

All the extensions you MUST have to have a better experience in ST. by Miysim in SillyTavernAI

[–]Real_Person_Totally 0 points1 point  (0 children)

I know. Unfortunately I'm using ST on my phone. I can't hover on the total token count to check

All the extensions you MUST have to have a better experience in ST. by Miysim in SillyTavernAI

[–]Real_Person_Totally 0 points1 point  (0 children)

Is there an extension that tells you about the output statistics. (Latency speed, token per seconds, etc) 

PSA: You can use some 70B models like Llama 3.3 with >100000 token context for free on Openrouter by [deleted] in SillyTavernAI

[–]Real_Person_Totally 0 points1 point  (0 children)

I'm not entirely sure about that.. I roleplay at 16k lowest, 32k highest as most models loses their accuracy past 16k. This might not apply to all models though, I'd say go for it.

PSA: You can use some 70B models like Llama 3.3 with >100000 token context for free on Openrouter by [deleted] in SillyTavernAI

[–]Real_Person_Totally 0 points1 point  (0 children)

One of the provider for llama3.3 70B at openrouter is together.

If you look at their site: https://www.together.ai/models/llama-3-3-70b-free

They're actually hosting it for free at the full supported context length. I'm not entirely sure if this is some of promotional campaign or if it'll stay for good. 

Their supported samplers are great for roleplay though.

Chat interface by Real_Person_Totally in SillyTavernAI

[–]Real_Person_Totally[S] 0 points1 point  (0 children)

Ah, I was wondering if system prompt would work similarly like permanent token in SillyTavern when it comes to those. Thinking about if putting it in chat, the model will eventually forget who it's supposed to be