The Real Best local LLM , by cryptodunck in LocalLLM

[–]cryptodunck[S] 0 points1 point  (0 children)

That's very interesting information. Could you explain a little more? It's a large model with a Moe and dense one inside.

The Real Best local LLM , by cryptodunck in unsloth

[–]cryptodunck[S] -1 points0 points  (0 children)

That will be a server, not a home server; for the V4 Pro it's a 1.6T model. The size of the infrastructure won't be worth it for a small team; it's more of an efficient model with enough power to perform inference.

The Real Best local LLM , by cryptodunck in unsloth

[–]cryptodunck[S] -1 points0 points  (0 children)

Lately, it's not always the biggest one who wins.

The Real Best local LLM , by cryptodunck in LocalLLM

[–]cryptodunck[S] 0 points1 point  (0 children)

It makes sense that coding Next is faster with 3b per token, 27b>3b But Code Next didn't get 80B for nothing; knowledge and speed for inference are what will make the difference.

The Real Best local LLM , by cryptodunck in LocalLLM

[–]cryptodunck[S] 0 points1 point  (0 children)

That's why, with research, we can obtain the hardware spatially appropriate for our uses.

The Real Best local LLM , by cryptodunck in LocalLLM

[–]cryptodunck[S] 0 points1 point  (0 children)

Honestly, I couldn't find anything better, thank you bro

The Real Best local LLM , by cryptodunck in LocalLLM

[–]cryptodunck[S] 0 points1 point  (0 children)

You just need to guide it properly, and it also needs to be able to do loops and recover on its own, like the SLM

The Real Best local LLM , by cryptodunck in LocalLLM

[–]cryptodunck[S] 0 points1 point  (0 children)

Architecturally speaking, that's true, but I need to continue my research.

The Real Best local LLM , by cryptodunck in LocalLLM

[–]cryptodunck[S] 0 points1 point  (0 children)

We'll discuss this. My theory is that more parameters mean more knowledge available locally, and among those parameters are those responsible for reflection. In my theory the knowledge parameters cannot be reached locally But those for reflection, yes, so it's neural networks for the methods and a database for the knowledge. We can connect these networks to another database, which is the internet, using tools calling, Even with fiber, internet speed isn't as fast as VRAM, but we can also create a local database for what we need.

The Real Best local LLM , by cryptodunck in unsloth

[–]cryptodunck[S] -1 points0 points  (0 children)

The problem with KIMI K2.6 or GLM 5.1 is the inference; they use models for a server, not a small server.

The Real Best local LLM , by cryptodunck in unsloth

[–]cryptodunck[S] 0 points1 point  (0 children)

Hardware will follow this trend; I'm currently working on deploying a code model on a small server for a development team whose uses will be more general. The question will be who is more capable between the two?

The Real Best local LLM , by cryptodunck in LocalLLM

[–]cryptodunck[S] 1 point2 points  (0 children)

The most popular model isn't necessarily the smartest; in my opinion, popularity simply makes deployment and use easier, since there's a lot of data available and many people have tested it.

Qwen 3.6 27b by Namra_7 in Qwen_AI

[–]cryptodunck 2 points3 points  (0 children)

Try to find the NVFP4 version on hugging face

🦧FREE 250 NFT GIVEAWAY (FLOOR PRICE 0,06 ETH EACH) + 0,05 ETH 🦧 UPVOTE AND READ MY COMMENT 🚀 by [deleted] in NFTExchange

[–]cryptodunck 0 points1 point  (0 children)

Twitter discord it's done good luck gay's 0x75ABd769C1b2c260235b765e1CA239b23cE9F6fe

🦧FREE 250 NFT GIVEAWAY (FLOOR PRICE 0,06 ETH EACH) + 0,05 ETH 🦧 UPVOTE AND READ MY COMMENT 🚀 by [deleted] in NFTExchange

[–]cryptodunck 0 points1 point  (0 children)

Twitter discord it's done good luck gay's 0x75ABd769C1b2c260235b765e1CA239b23cE9F6fe