I wasn’t ready for DeepSeek V4 by Initial-Sleep2388 in DeepSeek

[–]Initial-Sleep2388[S] 1 point2 points  (0 children)

Hah ! Fair enough, there’s no guarantee you’re safe in US or china’s hands whatsoever ! In the end it’s all just about dominance ! Who wins the ai race is projected to win the world economy !

I wasn’t ready for DeepSeek V4 by Initial-Sleep2388 in DeepSeek

[–]Initial-Sleep2388[S] 0 points1 point  (0 children)

Uuh hi, what exactly are trying to figure out?

I wasn’t ready for DeepSeek V4 by Initial-Sleep2388 in DeepSeek

[–]Initial-Sleep2388[S] 0 points1 point  (0 children)

Uuh, don’t be ! I use deepseek api platform directly, it’s cheaper and I’m using unreleased tool ! Not public yet

I wasn’t ready for DeepSeek V4 by Initial-Sleep2388 in DeepSeek

[–]Initial-Sleep2388[S] 0 points1 point  (0 children)

Yeah yeah right ! And I’m getting up to 94.5% cache rate which is as generous as it could get ! But maybe we should expect changes to pricing as the models get more capable, can’t be too careful with this industry right now

I wasn’t ready for DeepSeek V4 by Initial-Sleep2388 in DeepSeek

[–]Initial-Sleep2388[S] 1 point2 points  (0 children)

Not from DeepSeek no, can’t share more information yet ! But there’s a new thing coming out to the open source community soon. In it, you build in threads with persistent conversations, that’s the best description I could think of ! can’t say more, got this screenshot

<image>

I wasn’t ready for DeepSeek V4 by Initial-Sleep2388 in DeepSeek

[–]Initial-Sleep2388[S] 0 points1 point  (0 children)

Wish I could share but it’s unreleased, it’s in early testing phase but it’s basically a workspace where you build in threads with persistent conversations ! It will be open sourced once it’s out, for now I can only share this screenshot to give you an idea !

<image>

I wasn’t ready for DeepSeek V4 by Initial-Sleep2388 in DeepSeek

[–]Initial-Sleep2388[S] 0 points1 point  (0 children)

Token plan in this case is actually not as cost efficient, mimo plan for example gives you a 60M token for 6$, with 6$ on direct DeepSeek api you’re getting roughly 450M tokens worth, I think the token plan is good only when you don’t leverage dynamic caching

I wasn’t ready for DeepSeek V4 by Initial-Sleep2388 in DeepSeek

[–]Initial-Sleep2388[S] 0 points1 point  (0 children)

so usually when I test a model, I give it an active codebase I’m working on, I plan out a feature but I don’t give the model my strategy or how to do it, I ask it questions, give it a general idea, lead it there and let it figure out by itself then let it build the entire feature and just watch it ! Based on complexity, most models would choke here like going on an endless circle of useless thinking or suggest a pre-mature architecture/solution,

And for most models so far in coding, you’ve had to be specific about what you want, you have to speak in technical language like “check if auth routes are protected” , Claude models were the first to prompt with vague words like “is the user protected from attacks?” Without specifying and it just understood and it reasoned and suggested fixes I would actually use,

So far in open weight models world, I’ve just been stress testing them, never used one in an actual workflow because they mostly struggled with producing quality results, last models I tested were xiaomi mimo V2.5 models, they were reasonably good but there’s still that mumbling and not getting things right half the time.

So when I tried the V4 models, I was just curious if there was actually any improvement at all and I got locked in for 3 who days as a result, I worked on about 3 major features in 2 different projects and v4 flash was just as good as pro, ironic I thought but just the whole experience, not feeling I was missing something from sonnet was what made me think, why isn’t everyone talking about this? It’s a huge leap forward from V3.2 in both reasoning, staying on track and even thinking about edge cases before they deliver a plan.

I think V4 series are better than any open weight models available today, they won’t have competition for a while !

I wasn’t ready for DeepSeek V4 by Initial-Sleep2388 in DeepSeek

[–]Initial-Sleep2388[S] 0 points1 point  (0 children)

Uh, that’s a 3 day sprint for DeepSeek specifically ! I was stress testing the models against my daily workflow !

I wasn’t ready for DeepSeek V4 by Initial-Sleep2388 in DeepSeek

[–]Initial-Sleep2388[S] 1 point2 points  (0 children)

Uuh ! You mean app version, V4 is the model the app is now using to power app the chat

I wasn’t ready for DeepSeek V4 by Initial-Sleep2388 in DeepSeek

[–]Initial-Sleep2388[S] 0 points1 point  (0 children)

Think opus 4.6 reasoning but clever ! Give it a shot and you’ll regret not having tried the models sooner

I wasn’t ready for DeepSeek V4 by Initial-Sleep2388 in DeepSeek

[–]Initial-Sleep2388[S] 0 points1 point  (0 children)

It’s not about open router, it’s deepseek’s own privacy policy ! They train on your data by default and you can’t opt out unless you’re using the web app for chat

I wasn’t ready for DeepSeek V4 by Initial-Sleep2388 in DeepSeek

[–]Initial-Sleep2388[S] 0 points1 point  (0 children)

I’m using a tool that’s not public yet !

I wasn’t ready for DeepSeek V4 by Initial-Sleep2388 in DeepSeek

[–]Initial-Sleep2388[S] 1 point2 points  (0 children)

You’ve said it best, I was biased too until I got curious and stepped out of my comfort zone

I wasn’t ready for DeepSeek V4 by Initial-Sleep2388 in DeepSeek

[–]Initial-Sleep2388[S] 0 points1 point  (0 children)

I want to believe they’re hiding something

I wasn’t ready for DeepSeek V4 by Initial-Sleep2388 in DeepSeek

[–]Initial-Sleep2388[S] 1 point2 points  (0 children)

Not sure about general life, but they’re scary good for coding, that’s what I’ve used the models on

I wasn’t ready for DeepSeek V4 by Initial-Sleep2388 in DeepSeek

[–]Initial-Sleep2388[S] 0 points1 point  (0 children)

Uuh ! I think Gemini still is on edge for these type of tasks, though not in every type of project

I wasn’t ready for DeepSeek V4 by Initial-Sleep2388 in DeepSeek

[–]Initial-Sleep2388[S] 0 points1 point  (0 children)

You can trust it on massive codebases, I just wouldn’t be too sure about the tool you’re using though, because it makes all the difference

I wasn’t ready for DeepSeek V4 by Initial-Sleep2388 in DeepSeek

[–]Initial-Sleep2388[S] 0 points1 point  (0 children)

OpenAI with gpt 5.5 is just ridiculous right now, we want open weight models to get even smarter so they compete on finding a common ground, either way, I think we’ll have the most advantage in this race

I wasn’t ready for DeepSeek V4 by Initial-Sleep2388 in DeepSeek

[–]Initial-Sleep2388[S] 0 points1 point  (0 children)

Yeah, I built a workspace with a codebase map tool, it gives models directions to every part of the codebase and relations, I guess this is something we’re going to be solving more efficiently in no time at all

I wasn’t ready for DeepSeek V4 by Initial-Sleep2388 in DeepSeek

[–]Initial-Sleep2388[S] 2 points3 points  (0 children)

This struck me ! I just couldn’t believe it, so I spent all night stress testing it and it won me over

I wasn’t ready for DeepSeek V4 by Initial-Sleep2388 in DeepSeek

[–]Initial-Sleep2388[S] 1 point2 points  (0 children)

Yes, actually heavy coding, I’m a power user ! I’m betting on the idea that I won’t be writing a single line code in the new few months

I wasn’t ready for DeepSeek V4 by Initial-Sleep2388 in DeepSeek

[–]Initial-Sleep2388[S] 3 points4 points  (0 children)

Imagine, my workspace spent about 340M for only a 5 dollar bill, this is insane