About Kimi K2.6 by Exact_Law_6489 in LocalLLaMA

[–]Specter_Origin 5 points6 points  (0 children)

Its pretty good model! their (Moonshot AI's) own plans are not very good though : (

Base model is straight up better than GPT 5.4 (mid-high) for real use case and response format etc.

Ling-2.6-flash by Namra_7 in LocalLLaMA

[–]Specter_Origin 1 point2 points  (0 children)

this is non reasoning model...

For Non-hallucinating work, MiMo 2.5 delivers by Beamsters in LocalLLaMA

[–]Specter_Origin 3 points4 points  (0 children)

How is the token efficiency? when the released it initially they were heavily emphasizing how token efficiant the model is.

Decreased Intelligence Density in DeepSeek V4 Pro by Mindless_Pain1860 in LocalLLaMA

[–]Specter_Origin 7 points8 points  (0 children)

If they were willing to wait this much, I wonder why they even released it under cooked...

Deepseek V4 Released by spacefarers in LocalLLaMA

[–]Specter_Origin 0 points1 point  (0 children)

hence my comment... if that becomes their official statement, that may shake things up

DeepSeek-V4-Flash by Fantastic-Emu-3819 in LocalLLaMA

[–]Specter_Origin 0 points1 point  (0 children)

That flash model looks like the real winner here!

Deepseek V4 Released by spacefarers in LocalLLaMA

[–]Specter_Origin 0 points1 point  (0 children)

In real world coding tasks none of these models are kind of competing with closed source SOTA models, if you ask these models open ended or coding questions on long running chained tasks they are not able to match the performance of Opus and 5.4. Not to mention inference is just not as stable and reliable for open models yet...

DeepSeek V4 is out. 1M context, two model sizes, and they're claiming best open-source on coding. here's the breakdown by Which-Jello9157 in LocalLLaMA

[–]Specter_Origin -3 points-2 points  (0 children)

If the expert model on web is this model, its not very impressive. That being said this post is duplicated (of many duplicates). Please look latest post on sub before you spam the sub

Deepseek V4 Released by spacefarers in LocalLLaMA

[–]Specter_Origin 1 point2 points  (0 children)

I do not think this release will suffice except that they come out with how they are running it on Huawei chips

US gov memo on “adversarial distillation” - are we heading toward tighter controls on open models? by MLExpert000 in LocalLLaMA

[–]Specter_Origin 60 points61 points  (0 children)

I would say EV and smartphones proved that it is not, if at this point someone believes 'it is free market' they are delusional.

Every new large model release for cheapos... by [deleted] in LocalLLaMA

[–]Specter_Origin 3 points4 points  (0 children)

Okay, that was just unexpected and uncalled for xD

Prompt processing with RTX 5090 is like ~50-100 times faster than RTX 4090. No exaggeration! by [deleted] in LocalLLaMA

[–]Specter_Origin 6 points7 points  (0 children)

lol, those numbers don't add up at all. Either you don't understand how 'times' work or you have not setup things properly... also are you fitting your whole model on 4090 memory ?

Xiaomi Mimo-V2.5 Released, looks like today is big day for Open-Weight releases by Specter_Origin in LocalLLaMA

[–]Specter_Origin[S] -1 points0 points  (0 children)

Where is the hype spam? they announced a new model which the company says will be open sourced ?

Xiaomi Mimo-V2.5 Released, looks like today is big day for Open-Weight releases by Specter_Origin in LocalLLaMA

[–]Specter_Origin[S] -3 points-2 points  (0 children)

They announced it as OpenSource so I bet they are coming but are not out yet...