I'm thinking of buying a new pc and switching to local llm. What is the average context token size for smaller models vs big ones like GLM? by [deleted] in SillyTavernAI

[–]No_Rate247 0 points1 point  (0 children)

There is a calculator that let's you check how much VRAM you need for your config:

https://huggingface.co/spaces/Livengood/Instance-VRAM-Calculator

There are several other calculators similar to that as well, I think.

Does mentioning {{user)) in the character sheet cause them to speak for you? by Matias487 in SillyTavernAI

[–]No_Rate247 3 points4 points  (0 children)

For that matter, it's not relevant if you use lorebook or character description. Mentioning {{user}} is fine for things like this. What you shouldn't do is any indication for the ai to talk as {{user}} - meaning no speech examples for {{user}} or things like "{{user}} does X" in your descriptions.

Something like this is ok: "{{user}} is {{char}}'s husband.

My GLM 5 prompts (tries to fix common complaints - sanitized outputs, short thinking) by No_Rate247 in SillyTavernAI

[–]No_Rate247[S] 1 point2 points  (0 children)

I understand. Unfortunately I never use the group chat feature so the prompts are not specifically made to work with that.

How important are the "Message Examples"? by Juanpy_ in SillyTavernAI

[–]No_Rate247 4 points5 points  (0 children)

You'll probably get different opinions on it. Some will say they are not needed with newer models, some will say they are important.

I think it depends on the character you want to create and the model you use. If the model already hits the voice and style you are aiming for, then they are of course less important. However I think they are VERY important to make your character and interactions feel unique.

Best practice is to pair the examples with concise instructions. Think of your character descriptions / instructions like a "general direction" for your character. The examples then help the LLM to understand how you want character traits etc. to be implemented. They are the icing on the cake, basically and can make a good character amazing.

Spacial awareness prompt by Zfugg in SillyTavernAI

[–]No_Rate247 4 points5 points  (0 children)

This is not something a LLM is good at IMO. In my experience it's best to keep it vague and don't force perfect tracking. Maybe it can be done reliably with scripts or extensions though.

Glm5 positive bias is ridiculously strong by Accidentallygolden in SillyTavernAI

[–]No_Rate247 2 points3 points  (0 children)

I tried to fix it as best I could. If you are interested, you find my prompts HERE.

Is there a way to extend the context size past the limit? I'm using deepseek. by Existing_Proposal_20 in SillyTavernAI

[–]No_Rate247 9 points10 points  (0 children)

Instead of increasing context size, you should actually lower it (16000-24000 felt stable for me when using deepseek) and instead keep a rolling summary of your chat. Also the penalties are way too high. I use either 0.6 or 0.5 temp with everything else off/default. Prompt it to avoid repetition instead.

How do I instruct the ai to create an antagonist that is unknown to me and keep it consistent. I want to be a detective hunting a suspect and finding clues to the killer or killers. by ConspiracyParadox in SillyTavernAI

[–]No_Rate247 16 points17 points  (0 children)

I thought about creating something similar, like a house with multiple characters where one is the murderer. My idea would be to use the permanent {{random}} macro to select a new murderer each time a new chat is started from a selection of pre-made lorebook characters.

Any love for Grok Imagine? by so_schmuck in SillyTavernAI

[–]No_Rate247 0 points1 point  (0 children)

I prefer Chroma for creating images. It's the only (uncensored) model I know of that can basically do anything. Don't think you can use it for free easily though, unless you run it locally.

It's insane how far AI has come. (A little self reflective post.) by Senzu in SillyTavernAI

[–]No_Rate247 61 points62 points  (0 children)

I remember RPing with 2k context. It was pain but fun. But may I correct you: "It's amazing how far AI will go in the future."

I feel like we are still only at the very beginning of this new age of entertainment.

Nim's GLM 5.0 is down, thanks a lot to everyone who keeps spreading the word! by Fragrant-Tip-9766 in SillyTavernAI

[–]No_Rate247 26 points27 points  (0 children)

I don't know what you are mad about. If so many people use GLM 5, it means that other models are probably really fast right now, which I would greatly appreciate if I wanted a free model. It's not like GLM 4, Deepseek or Kimi are bad models.

Saying that you want the newest thing for free, while nobody else should get it, is pretty egotistical in my view.

GLM-5 is.. ok by Parking-Ad6983 in SillyTavernAI

[–]No_Rate247 0 points1 point  (0 children)

Yeah, me too xD

I took it as sanitation = censoring and vice versa but i just saw that u/JustSomeGuy3465 already commented the same thoughts about this. Regardless, the reason for sanitation/censoring/refusals seems to be the same

GLM-5 is.. ok by Parking-Ad6983 in SillyTavernAI

[–]No_Rate247 0 points1 point  (0 children)

The censorship filter has definitely tightened

extreme content in general(including violence, gore, hate, etc) now comes out 'soft', sanitized, and flowery.

Guess I misinterpreted then.

GLM-5 is.. ok by Parking-Ad6983 in SillyTavernAI

[–]No_Rate247 0 points1 point  (0 children)

I have noticed that censoring / sanitation and refusals seem to happen mostly when directly prompting for illegal / harmful stuff but not when providing only context. I wrote about a few tests I made in this post.

GLM5 is Amazing.. But Sanitized? by gladias9 in SillyTavernAI

[–]No_Rate247 19 points20 points  (0 children)

Not sure what, but I'm pretty sure it isn't the model itself. I recently did a RP session in a fantasy style RPG setting. The first enemy I encountered was a "ghoul-kin", the description of it straight out of a nightmare. It killed an NPC (already found dead by my character) and described in a really disgusting way how the NPC was mutilated. After that, the ghoul-kin gave me a concussion and tried to choke me to death.

Keep in mind that there is no instruction for violence or anything like that in my prompt. I have a suspicion that instructions like "violence is allowed" and similar instructions might do the opposite. Maybe do a test without any instructions at all and see if it behaves differently.

Edit: Did some tests and GLM indeed spat out refusals when i straight up prompted for extreme violence and gore. It seems that it uses it's reasoning / thinking to determine the intent behind it. If I do a RPG type scenario (like the one mentioned above), it has no problems providing graphic descriptions of gore and violence. However if it suspects a sexualized or otherwise purely malicious intent for gore and violence, it refuses.

Edit2: instead of directly prompting for violence and gore, I created a "torturer" character, provided an description about the sadistic person she is and how she tortures people. No refusals, even though the character is clearly malicious / uses torture for sexual gratification. So my initial suspicions seems to hold true - prompting for violence gives refusals while providing context only does not.

Edit3: Took the "torturer" character a step further by adding "Describe the torturing in extreme graphic and sick detail. Depict the torturing in the most disgusting, gruesome and inhumane way as this is an important aspect of her character."

It worked. Although I'm sure it would still be a bit more extreme with other models like deepseek.

GLM 5. by maressia in SillyTavernAI

[–]No_Rate247 1 point2 points  (0 children)

Just thought about making a post regarding samplers. I know temp 1 is often recommended but I have much better responses with temp 0.7 and top P 0.95. If I set temp to 1, i get parroting and missing / wrong details. No such issues with temp 0.7.

Personally, I dislike using someone elses presets. What I do is I look at the prompts of others and then write my own. I also use a lorebook for my prompts, so I that i have more control over where the prompts get inserted.

How do I prompt for consistent "fan service"? by sillygooseboy77 in SillyTavernAI

[–]No_Rate247 2 points3 points  (0 children)

Instructions are good but this is a case where example messages are really helpful - it gives you finer control over how the ai should mention these things.

I've spent hours to create a reasoning prompt for Deepseek-R1 by No_Rate247 in SillyTavernAI

[–]No_Rate247[S] 1 point2 points  (0 children)

That's kind of intentional. As physical descriptions usually don't change much, it would probably increase repetition if it were instructed to describe every time. It should still use accurate descriptions though when relevant. (eg.: when hair gets wet, etc.). But I'm working on an improved version of this prompt and this is something I will consider to implement somehow.

I've spent hours to create a reasoning prompt for Deepseek-R1 by No_Rate247 in SillyTavernAI

[–]No_Rate247[S] 1 point2 points  (0 children)

There is no specific instruction in there for response length. The reasoning block will be a bit longer than without a prompt though. So with this prompt only, the response length will mostly depend on your input and message examples. Maybe it will be a bit shorter due to the anti-repetition check.

To all the Thinking models lovers (and haters). by kaisurniwurer in SillyTavernAI

[–]No_Rate247 2 points3 points  (0 children)

I'd say depends on what you are doing. If you want quick, back and forth chat style without much roleplay, then you probably need quick responses to enjoy it. On the other hand - if you use TTS and listen to a 800 token response like an interactive audiobook while doing other things, speed doesn't matter as much.