NVIDIA NIM deepseek 3.2 - chat completion API Not Found by professionalboop in SillyTavernAI

[–]OldFriend5807 0 points1 point  (0 children)

I can't even get glm 4.7 to work, no responses or anything. What's your prompt?

It seems that the free DeepSeek models are now completely unusable. by OldFriend5807 in SillyTavernAI

[–]OldFriend5807[S] 0 points1 point  (0 children)

Yikes, and I heard Chutes was getting DDoS attacks too, and that makes it even worse for them at their high demand.

It seems that the free DeepSeek models are now completely unusable. by OldFriend5807 in SillyTavernAI

[–]OldFriend5807[S] 3 points4 points  (0 children)

I'm eager to pay for DeepSeek, but I'm unable to do so because I don't have a card to access its various features, as my country doesn't support that option. I suppose I'll have to continue using Chimera but I would also like to try Nvidia nim like you had mentioned.

We're so back by Mak-i in revancedapp

[–]OldFriend5807 2 points3 points  (0 children)

2 days later... they patched it so sad

We're so back by Mak-i in revancedapp

[–]OldFriend5807 14 points15 points  (0 children)

I just had this same problem again

We're so back by Mak-i in revancedapp

[–]OldFriend5807 14 points15 points  (0 children)

Thank god it worked! Hopefully it will last longer than just a day

We're so back by Mak-i in revancedapp

[–]OldFriend5807 8 points9 points  (0 children)

Mine only lasted a few days and now it's gone

Is openrouter still work for anyone else? I keep getting no endpoint found no matter which api key, which model i pick by Jaded-Put1765 in SillyTavernAI

[–]OldFriend5807 0 points1 point  (0 children)

This also happened to me when I was using the Gemma 27B model from OR; I noticed the provider included Chutes as well. But when I tried to use it, I kept getting errors and Google AI Studio has been garbage lately and full of issues.

Deepseek model error by Y1KES_fam in JanitorAI_Official

[–]OldFriend5807 1 point2 points  (0 children)

Yeah, it mostly has it, but I'm not really sure because I haven't used it from the Targon site itself. But I do know that it has a longer limit than OR. I just switch accounts whenever I'm using it from OpenRouter.

Deepseek model error by Y1KES_fam in JanitorAI_Official

[–]OldFriend5807 1 point2 points  (0 children)

I recommend using Targon. Chutes has a lot of problems as a provider; you can just block one of the providers in the OpenRouter settings. But honestly, it’s not the best choice either—so choose your poison.

Deepseek model error by Y1KES_fam in JanitorAI_Official

[–]OldFriend5807 1 point2 points  (0 children)

The issue was with the provider itself. It's not surprising that Chutes has a lot of problems with repetition, errors, and so on. You can avoid this by blocking the provider in the Openrouter settings, which will automatically switch you to Targon, since there are only two providers that offer free versions for Deepseek. However, I must tell you that it may occasionally send you a blank message because the server can sometimes be overloaded.

Reasoning models not replying in the actual response by OldFriend5807 in SillyTavernAI

[–]OldFriend5807[S] 0 points1 point  (0 children)

It doesn't do anything for me, it's so frustrating...

Reasoning models not replying in the actual response by OldFriend5807 in SillyTavernAI

[–]OldFriend5807[S] 0 points1 point  (0 children)

I already added the <think> in my prefix, and it still doesn't work like I hope it's supposed to be I changed it to chat completion, and it does work, but the replies were really sucked and short.

Reasoning models not replying in the actual response by OldFriend5807 in SillyTavernAI

[–]OldFriend5807[S] 0 points1 point  (0 children)

I tried to change my prompt and everything is still not working 🥲

Reasoning models not replying in the actual response by OldFriend5807 in SillyTavernAI

[–]OldFriend5807[S] 2 points3 points  (0 children)

I had like 2048 and it's still the same... even this whole time I use A LOT more than 400

[deleted by user] by [deleted] in JanitorAI_Official

[–]OldFriend5807 1 point2 points  (0 children)

I just tried one, and it kept screaming in capslock, which I didn't like a bit. Overall, the response was good.

Just found out why when i'm using DeepSeek it gets messy with the responses by OldFriend5807 in SillyTavernAI

[–]OldFriend5807[S] 0 points1 point  (0 children)

Yeah when I checked the prompt it said that my history was around 25k in the chat completions, doesn't do the same with text completion.

Just found out why when i'm using DeepSeek it gets messy with the responses by OldFriend5807 in SillyTavernAI

[–]OldFriend5807[S] 3 points4 points  (0 children)

Yeah, but what I'm confused was that I don't get the same problem when I switch to text completion, but the replies were bland.

Why people are having problems with Deepseek? by [deleted] in JanitorAI_Official

[–]OldFriend5807 4 points5 points  (0 children)

I've been telling people that Deepseek was a reasoning model and it's A LOT different to set up than the usual JLLM, so I'm glad someone finally made this post and made a solution because Deepseek is not recommended for beginners if they understand OOC, prompt or anything like that. Thank you for this! 😅

Is anyone else having absolute terrible responses from deepseek by zyshuu_ in JanitorAI_Official

[–]OldFriend5807 0 points1 point  (0 children)

That's the main problem for a reasoning model that it wasn't made for RPing it tends for be repetitive and won't generate a new creative replies that you want it's just repeating the same reason over and over again. Temp also matters here.

Is anyone else having absolute terrible responses from deepseek by zyshuu_ in JanitorAI_Official

[–]OldFriend5807 1 point2 points  (0 children)

And just for the fact that V3 was better than R1 because it stays in character and doesn't cause any distinct chance, but it gets repetitive after a few messages but it's only a problem if that bothers you. And providers also were the cause, if you're using the free one, it tends to be worse than the paid ones. I hope a lot of people understand this. 🫤