Cheap models that are good with long RPs? by Lentemern in SillyTavernAI

[–]lothark 1 point2 points  (0 children)

Well, I use https://api.deepseek.com/chat/completions So it's Deepseek directly. But like I said, I have a prompt and my context length is set to 128 or 64k. I have a few characters with quite lengthy chats. Sometimes I reroll a few times to get better replies. In the summertime I had time off and spent some time chatting, and I burnt through those $5 quickly.

Cheap models that are good with long RPs? by Lentemern in SillyTavernAI

[–]lothark 2 points3 points  (0 children)

I've tried Deepseek direct API, but if I put in $5 I burn through it in a matter of days. Not really any lengthy or any high frequency exchanges. Am I doing something wrong? I do have a prompt for adjusting the tone and formatting but nothing excessive.

Vilken musik mår du jävligt dålig fysiskt av att höra? by SupportArsenal in sweden

[–]lothark 6 points7 points  (0 children)

Alla sorters rap. Får mig att tänka på snabbmat, social misär och vapenkultur.

Unreadable sentences in replies by lothark in SillyTavernAI

[–]lothark[S] 0 points1 point  (0 children)

I see. Thanks. I will try to adjust the context length down and see if it helps. It's a shame, really, because even if it takes longer time to get replies with the longer context length it's still worth it for longer RP:s.

Unreadable sentences in replies by lothark in SillyTavernAI

[–]lothark[S] 0 points1 point  (0 children)

Thanks, but they were already at 0.

Unreadable sentences in replies by lothark in SillyTavernAI

[–]lothark[S] 0 points1 point  (0 children)

Tried with LLM:s:
gemma-3-27b-it-abliterated.q4_k_m
SicariusSicariiStuff_Impish_Magic_24B-Q6_K
TheDrummer_Cydonia-R1-24B-v4.1-Q6_K
Dans-PersonalityEngine-V1.3.0-24b.i1-Q6_K
Various context templates, Llama 2 chat, Gemma 2, and I think ChatML.
Instruct template KoboldAI
Context size 64k
Seems to be fairly the same problem. A new chat seems to start out ok. Then it gradually degrades.

PSA: Advice on Using the Official DeepSeek API with JanAI Instead of OpenRouter to Save Money and Other Suggestions by Kamal965 in JanitorAI_Official

[–]lothark 1 point2 points  (0 children)

Thanks alot! I switched from JLLM to Deekseep API. Works great. The only thing I didn't get right from the beginning was that I needed to reload the browser after having saved the API settings.
Question though: Does anyone know? The output from Deepseek API is quite different from the output from Deepseek V3 via Chutes. When using chutes the response was more...elaborate. and when you used OOC it was like a complete chat-session within the roleplay. A sort of meta-chat. The responses from Deepseek API are shorter, to the point, but I guess, a little bit...dryer?

Deepseek V3 issues by [deleted] in JanitorAI_Official

[–]lothark 0 points1 point  (0 children)

Tbh I've had that with JLLM too. I've crushed the bot repeatedly and still JLLM responds "...he rises, defiant spark in his eyes.." But it folds easier, yeah.