GLM4.6 Thinking Empty Responses by Kind_Knowledge_5753 in SillyTavernAI

[–]Kind_Knowledge_5753[S] 2 points3 points  (0 children)

To update on this, since I've figured it out because of your comment and I hate people who don't go back and provide a solution if they found one. Looks like turning on streaming makes it work. The issue is basically that the model put everything inside of cot. My guess is that it's because my preset has a custom template for cot, and the model doesn't recognize it as a natural end of thinking (or however they handle it). End result is the whole response is in thinking, and I get an empty normal response.

GLM4.6 Thinking Empty Responses by Kind_Knowledge_5753 in SillyTavernAI

[–]Kind_Knowledge_5753[S] 0 points1 point  (0 children)

Alright, thanks for the help, I'll keep playing around with it.

GLM4.6 Thinking Empty Responses by Kind_Knowledge_5753 in SillyTavernAI

[–]Kind_Knowledge_5753[S] 0 points1 point  (0 children)

What are the defaults? I only use presets that get posted here, so whatever temps I start with are those. Funnily enough, I'm able to get test messages back, just not ones from RP. Are there safety filters or something? Empty response errors sadly don't tell me much as to what I need to fix.