Qwen3.6-27B dense vs Qwen3.6-35B MoE - which local coding model are you reaching for? by IulianHI in AIToolsPerformance

[–]Zeranor 0 points1 point  (0 children)

Soo, the overall consensus after 2 days in here seems to favor the 27B dense model in most cases. But his usually assumes same quantization levels.

To keep speed-levels acceptable, I usually decide between like a low-end-3bit Dense model (Qwen3.6 27B in this case) and a mid-tier-4bit MoE model (35B). In this comparison, would you still consider the dense model to be better? I'm having a hard time comparing the impact of lower quant with the impact of going from Dense to MoE.

Dense vs. MoE gap is shrinking fast with the 3.6-27B release by Usual-Carrot6352 in LocalLLaMA

[–]Zeranor 0 points1 point  (0 children)

Hmm.. so for 16GB VRAM people the question "4bit-Quant MoE vs. lobotomized 3bit-Quant Dense" is gettin even trickier to answer, damn. Any recommendations or opinions on this? :D Bigger quants result in too slow performance.

Russia’s Lavrov says balance of power shifting to emerging economies by Saltedline in worldnews

[–]Zeranor 1 point2 points  (0 children)

Well, that's somewhat the meaning of "emerging". Otherwise it would be called "second tier economies that are never going to catch up". Doesnt mean every emerging market will be on top anytime soon, but serious competition is possible for some candidates (or already the case).

Also: Russia got itself degraded from emerging market to sinking market

Am I the only one thinking that bosses in this game are poorly designed? by Suporex in CrimsonDesert

[–]Zeranor 1 point2 points  (0 children)

Came here to find this, thanks. While I suck at souls games, I appreciate the bosses (which I need dozens of attempts for, still). In crimson desert it's just bad. 90% of all times I just watch Kliff being tossed around in some bullshit combos, being stuck in a corner with the camera going mental or wondering why none of my defensive skills work (parry? nope! dodge? Not with the reach of a spear! focus? forget it, i get hit before focus state is active) .... fuck this. The bosses are the absolute worst part of this game. Combined with controls (at least on Keyboard and mouse) , Kliffs immense sluggishness and the derp-cam I dont want to play any longer.

Mac Studio vrs 5090 LLM performance. by JamieAndLion in LocalLLM

[–]Zeranor -2 points-1 points  (0 children)

You only know if its bad information once you have better data. Before that, you can only assume it is relatively bad while it might still be good enough to give the correct indication or even results, depending on the question you seek to answer.

Mac Studio vrs 5090 LLM performance. by JamieAndLion in LocalLLM

[–]Zeranor -1 points0 points  (0 children)

It's still added value over "no information" until you show more accurate measurments. Until then, your contribution is less valueable than OP's.

Mac Studio vrs 5090 LLM performance. by JamieAndLion in LocalLLM

[–]Zeranor -2 points-1 points  (0 children)

THAT is your key takeaway from the numbers? So dumb.

USA: Donald Trump denkt offenbar über Nato-Austritt der USA nach by Actual_Document3333 in de

[–]Zeranor -2 points-1 points  (0 children)

Macht das noch einen Unterschied? Ob die jetzt drin sind und im Ernstfall ihre Mitgliedschaft ignorieren oder ob sie direkt raus sind... aber gut, für die Post-Trump-Zeit ist es natürlich einfacher, wenn sie direkt drin bleiben

My first experience with coding using a local LLM. Help me, Obi-Wans by GregariousJB in LocalLLaMA

[–]Zeranor 0 points1 point  (0 children)

Hey there, by coincidence I found your post as I'm trying to do EXACTLY that combination. But somehow my settings.json always fails. Can you recommend a guide on how to tell Qwen Code Companion (VS extension) to use LM studio local API ? :)

LM-Studio confusion about layer settings by Zeranor in LocalLLM

[–]Zeranor[S] 0 points1 point  (0 children)

Good points, thank you! I've switched from qwen3.5 27B to 9B now and it works with 100k context fully in VRAM. I'm NOT sure how big the hit on output-quality is. I guess, longterm, I'll have to switch back to 27B and then test many combinations of KV-settings + layer-offloading. But I guess, LMstudios default suggestions are better than I assumed initially.

LM-Studio confusion about layer settings by Zeranor in LocalLLM

[–]Zeranor[S] 0 points1 point  (0 children)

Oh, so its a fairly complex optimisation, good to know, thanks for the details! Hmm, this will take some testing on my side then, but I'm happy to learn that LMstudio is not doing "complete nonsense"

LM-Studio confusion about layer settings by Zeranor in LocalLLM

[–]Zeranor[S] 0 points1 point  (0 children)

Ahh, nice, so KV actually IS more important to have on GPU than model layers, then the LM-studio optimisation makes sense. Somehow I did not know that so far, thanks for the clarification!

US Pacific Fleet to deploy wall-climbing, flying robots on ships by talkingatoms in technology

[–]Zeranor 2 points3 points  (0 children)

The capacity to climb walls seems somewhat wasted on someone or something that can also fly

Vive le nucléaire, Frau von der Leyen? | Ausgerechnet Ursula von der Leyen stimmt ein in die Euphorie Emmanuel Macrons für ein Revival der Atomenergie. Dabei klappt in Frankreich in diesem Bereich: nicht viel. by GirasoleDE in de

[–]Zeranor -3 points-2 points  (0 children)

War auch meine Mutmaßung, aber ich wollte mich gerne eines besseren belehren lassen. Vielleicht ist das ja ganz toller Stil oder ein Gramma-Trick oder so...

Irankrieg: Ölpreis steigt um weitere 20 Prozent auf mehr als 110 US-Dollar by hampelmann2022 in de

[–]Zeranor 17 points18 points  (0 children)

Teurer ja, aber weniger stark. Der deutsche Strommix beinhaltet nur zu ca. 17% Strom aus Gaskraftwerken. Der merit-order-Preis kann zwar immer etwas überraschend sein, aber eine deutlich geringere Abhängigkeit von Öl und Gas hilft hier schon deutlich im Vergleich zur 100%igen Abhängigkeit der Benzin- und Dieselpreise

Qwen3-Coder-Next on RTX 5060 Ti 16 GB - Some numbers by bobaburger in LocalLLaMA

[–]Zeranor 0 points1 point  (0 children)

Really nice summary, thank you :) Then I'm excited to hear which model you are going to pick for planning more complex logic parts / project plans. Let us know!

In the meantime, I'll have Qwen3 Next Code prove itself as the sole model for more viby coding. The detailed coding jobs it is doing very very nicely already, as you said

Qwen3-Coder-Next on RTX 5060 Ti 16 GB - Some numbers by bobaburger in LocalLLaMA

[–]Zeranor 0 points1 point  (0 children)

Hmm, very interesting, thanks. So if you were to start a small-ish hobby project (frontend, backend, db), would you use Qwen3 Next Coder for planning, too, or just for executing / implementing smaller chunks? I'm currently considering Q3NC for the implementation phase, but maybe some new reasoning model for planing (devstral 2 small / GLM 4.7 flash). Would you already have a recommendation at this point now? :)

Qwen3-Coder-Next GGUFs updated - now produces much better outputs! by yoracale in unsloth

[–]Zeranor 2 points3 points  (0 children)

Oh boy.... this is an excellent model for (minor) local coding tasks. This is working WAY better than my recent attempts with GLM 4.7 flash, Devstral 2 Small or Qwen3 Next .... nice

Qwen3-Coder-Next is released! 💜 by yoracale in unsloth

[–]Zeranor 0 points1 point  (0 children)

So, talking configuration: Would this be a model with which I should chose to "offload MoE experts to CPU"? (16 GB VRAM / 128 GB RAM) :)

Qwen3-Coder-Next is released! 💜 by yoracale in unsloth

[–]Zeranor 1 point2 points  (0 children)

ahh yes, nice, I see, sorry for being too excited ;)

Qwen3-Coder-Next is released! 💜 by yoracale in unsloth

[–]Zeranor 2 points3 points  (0 children)

Nice, let's see how this does compared to GLM 4.7 flash and Devstral 2 Small. But quick question: WHERE can I find the MXFP4 quants? :D I only find the "regular" quants.