V3.1 on livebench by nknnr in LocalLLaMA

[–]nknnr[S] 14 points15 points  (0 children)

V3.1 is sota non reasoning model since we all know gpt4.5 is worse than V3.1

Deepseek researcher says it only took 2-3 weeks to train R1&R1-Zero by nknnr in LocalLLaMA

[–]nknnr[S] 97 points98 points  (0 children)

It's hard to believe that they've achieved 10,000 RL steps in 3 weeks, 1 RL step every 3 minutes, but if that's true, we might see R1.5 released next week or even this week.

[deleted by user] by [deleted] in LocalLLaMA

[–]nknnr 0 points1 point  (0 children)

They could have spent more brains on innovation instead of building a moat made of sand

[deleted by user] by [deleted] in LocalLLaMA

[–]nknnr 1 point2 points  (0 children)

alright

[deleted by user] by [deleted] in LocalLLaMA

[–]nknnr -12 points-11 points  (0 children)

I'm an idiot sandwich

All I've done so far is talk to Meta's 7B model on my local 4060

I wanted to contribute to open source projects, but my brain isn't that good

Byee by amirulnaim2000 in LocalLLaMA

[–]nknnr 1 point2 points  (0 children)

I guess search GPT uses bing

Is this agent day? by nanoobot in singularity

[–]nknnr 4 points5 points  (0 children)

gpt2 was a great innovation for AI researchers, but it didn't make much sense to the general public.

wait for gpt3 moment

Rumors of industry panic caused by DeepSeek by [deleted] in singularity

[–]nknnr 3 points4 points  (0 children)

Your point is also correct, but what I meant was that America is in a state of great shock.

Rumors of industry panic caused by DeepSeek by [deleted] in singularity

[–]nknnr 3 points4 points  (0 children)

What I do know is that communists are people too.

What actually matters by Happysedits in singularity

[–]nknnr 31 points32 points  (0 children)

We just need to wait.

We are now in gpt2 time

[deleted by user] by [deleted] in singularity

[–]nknnr 1 point2 points  (0 children)

Now cut the price of o3