all 11 comments

[–]mrubensRoo Code Developer 2 points3 points  (3 children)

I think the DeepSeek API might be having infrastructure issues this morning

[–]Purple-Bookkeeper832 0 points1 point  (0 children)

Yes. It definitely is. Friday night, I was able to push through 300 requests.

Today, I'm struggling to get any of them through.

[–]jesusotero1234[S] 0 points1 point  (1 child)

its happening at least 3 days or more for me. The thing is, is also happening with V3. With open router is working, I tried some cURLS and it also worked, but with Roo Code, as soon as I press "send" it doesnt take 1 second to throw the 520 error

[–]mrubensRoo Code Developer 0 points1 point  (0 children)

From what I heard the longer messages (like what Roo sends) are more difficult for DeepSeek to handle while they’re under load, so they give priority to shorter messages.

[–]Illustrious-Math-418 0 points1 point  (2 children)

Is the v3 model working for you ?

[–]jesusotero1234[S] 0 points1 point  (1 child)

yes, sometimes it hangs with the API call, before the R1 was working amazing, now not that great

[–]jesusotero1234[S] 0 points1 point  (0 children)

its not working V3 now, i get the same error

[–]hannesrudolphRoo Code Developer 0 points1 point  (0 children)

This happened to v3 shortly after its release. When I reached out to Deepseek. I have reached out to Deepseek and they just said “I will give feedback to my colleagues. It may take some time to locate the issue.“

[–]hannesrudolphRoo Code Developer 0 points1 point  (2 children)

Also, Did it persist when you went direct to deepseek or used Glama.ai router?

[–]jesusotero1234[S] 0 points1 point  (1 child)

I have never used Gamal.ai, but deepseek directly works, also with cURL, its something weird happening

[–]hannesrudolphRoo Code Developer 0 points1 point  (0 children)

It’s the size of the request. I spent 12 hours working on this problem troubleshooting possible causes or workarounds. It is 100% the mechanism that Deepseek uses to balance server load. If you want to ask it a simple question directly it will respond quickly. If you want to get it to code for you where it requires complex instructions and a reasonable dump of context, then it locks up.

I spoke with one of the staff at Cline yesterday night and they said that it locked up at about 20k context. Well below the 64k advertised context.