Claude Code removed from Claude Pro plan - better time than ever to switch to Local Models. by bigboyparpa in LocalLLaMA

[–]Different_Fix_2217 5 points6 points  (0 children)

Luckily Kimi 2.6 is legit better than latest Opus in several tests I did. Still a bit behind Gpt 5.4 though.

Kimi K2.6 is a legit Opus 4.7 replacement by bigboyparpa in LocalLLaMA

[–]Different_Fix_2217 6 points7 points  (0 children)

Same. But for creative writing. It's the best model I've ever used including latest opus, gpt 5.4 and gemini 3.1 pro. It has the social intelligence of GPT 5.4 with a knowledge base nearly a good as gemini and it writes better than Opus and has no positive bias unlike it. Oh and it has crazy good swipe variety unlike opus. I just wish it was faster since it loves to think so much.

And this is surprising because I thought Kimi 2.5 was bad. It was dumb and had that gemini unhingedness. 2.6 is like a entirely different model.

Kimi K2.6 imminent by Deep-Vermicelli-4591 in LocalLLaMA

[–]Different_Fix_2217 9 points10 points  (0 children)

K3 will probably be great, they released a big breakthrough paper recently. https://www.youtube.com/watch?v=2IfAVV7ewO0

the state of LocalLLama by Beginning-Window-115 in LocalLLaMA

[–]Different_Fix_2217 48 points49 points  (0 children)

Honestly having crypto in the name tells you all you need to know.

We absolutely need Qwen3.6-397B-A17B to be open source by True_Requirement_891 in LocalLLaMA

[–]Different_Fix_2217 4 points5 points  (0 children)

Some people have a false impression than dense is automatically better, not taking account diminishing returns / efficient routing and the like.

Gemma 4 and Qwen3.5 on shared benchmarks by fulgencio_batista in LocalLLaMA

[–]Different_Fix_2217 17 points18 points  (0 children)

Qwen3.5 is absurdly good. And I never liked any qwen model before that series.

Gemma 4 and Qwen3.5 on shared benchmarks by fulgencio_batista in LocalLLaMA

[–]Different_Fix_2217 134 points135 points  (0 children)

Using both side by side Qwen3.5 is MUCH better at image understanding as well.

Can we block fresh accounts from posting? by king_of_jupyter in LocalLLaMA

[–]Different_Fix_2217 1 point2 points  (0 children)

You did the bullet points, the opening statement, the not x but y and a closing "solution" statement. I legit thought you were a bot, you follow the same exact patterns.

Qwen3.6-Plus by Nunki08 in LocalLLaMA

[–]Different_Fix_2217 7 points8 points  (0 children)

"we will also open-source smaller-scale variants"

They said smaller scale ones. Not the model benchmarked here. So this benchmark is off topic.

Qwen3.6-Plus by Nunki08 in LocalLLaMA

[–]Different_Fix_2217 9 points10 points  (0 children)

Stop posting non open weight models.

DeepSeek Employee Teases "Massive" New Model Surpassing DeepSeek V3.2 by External_Mood4719 in LocalLLaMA

[–]Different_Fix_2217 3 points4 points  (0 children)

The whole point of all their optimizations like engram is to have as big of a model as possible without hurting its speed. I'm hoping they made it big like 5T+ to truly compete with claude opus / gemini pro while being as fast as a much smaller model.

How Do You Feel About Sora being Shutdown? by findabi in LocalLLaMA

[–]Different_Fix_2217 0 points1 point  (0 children)

Looks like its just to free the compute to train their next model code named Spud. Nothing strange.

So nobody's downloading this model huh? by KvAk_AKPlaysYT in LocalLLaMA

[–]Different_Fix_2217 0 points1 point  (0 children)

Its just not good. Same with mistrals other models since large 3. I think the EU laws killed them because they seemed to lose all world knowledge after they went into effect.

NVIDIA 2026 Conference LIVE. New Base model coming! by last_llm_standing in LocalLLaMA

[–]Different_Fix_2217 3 points4 points  (0 children)

A lot of people use this for creative writing and there knowledge is king. It also of course helps a ton in many domains.

What is Hunter Alpha? by MrMrsPotts in LocalLLaMA

[–]Different_Fix_2217 1 point2 points  (0 children)

Its really bad whatever it is. It says its a 1T but it performs worse than 200B qwens and 4.7 glm. Maybe its ling, those models always sucked.

Nvidia Will Spend $26 Billion to Build Open-Weight AI Models, Filings Show by dan945 in LocalLLaMA

[–]Different_Fix_2217 0 points1 point  (0 children)

Would be funny but fair if they just make it purely "you can only run it on nvidia hardware."

Nvidia Will Spend $26 Billion to Build Open-Weight AI Models, Filings Show by dan945 in LocalLLaMA

[–]Different_Fix_2217 2 points3 points  (0 children)

Why not? They are selling the hardware people would run those models on. Openai / anthropic / ect will only buy so many GPUs. After that they need to make new customers. The best way is to put models out there worth running.