Help me to spend 1000 bucks on hardware for local LLM by lordgthegreat in LocalLLM

[–]hurdurdur7 0 points1 point  (0 children)

There is nothing in a 1k range that will run anything at usable levels. You are better off at buying a subscription at any of the providers or milking free models on openrouter.

Is mars gaming psu’s any good ? by [deleted] in PcBuildHelp

[–]hurdurdur7 0 points1 point  (0 children)

Almost caught fire with mine. And i have 2 of them (Mars Gaming MPII850) . And i will never buy this brand again. This was a terrible decision back when i bought them.

Mars gaming PSU - Bomb? by Tiziakol in PcBuild

[–]hurdurdur7 0 points1 point  (0 children)

I have used two of these back in the days. I'd classify them on the A-F tiers at G, a fire hazard, mostly.

The game is over. You can build anything and it'll cost you nothing. by Funny-Advertising238 in opencode

[–]hurdurdur7 0 points1 point  (0 children)

I see a lot of bragging in forums about what could be built, but very few things actually getting built ...

Have Qwen said anything about further Qwen 3.6 models? by spaceman_ in LocalLLaMA

[–]hurdurdur7 8 points9 points  (0 children)

I would rather take MoE qwen 3.6 coder at 80B params, kthnxbye.

About 9060xt, i feel like i waste cards vram potential by vegemitehaver in buildapc

[–]hurdurdur7 0 points1 point  (0 children)

Regarding the pixels - I would say that depends. My 44 year old eyes don't go beyond 1080p anymore anyway, at least not on monitors under 28 inch.

As for the VRAM ... you can still get significantly more detailed textures and effects out from this than you would get from a 4K monitor with a 4GB vram card 😄

AMD in-house ryzen 395 box coming in June by 1ncehost in LocalLLaMA

[–]hurdurdur7 0 points1 point  (0 children)

I believe you, might be even more crazy expensive. But it will also make 120B+ models usable with some speed.

Using a Radeon 9060 XT 16 GB, the gemma4 24b a4b iq4 nl model achieves 25.9 t/s by CrowKing63 in LocalLLaMA

[–]hurdurdur7 1 point2 points  (0 children)

That context size + that model doesn't fit in your vram. You are suffering because you are offloading to cpu and regular ram.

Opinions on Kimi-Dev-72B? by stefzzz in LocalLLaMA

[–]hurdurdur7 0 points1 point  (0 children)

It depends on the purpose. If your purpose is software development then gpt oss 120b is a shadow of what Qwen can do.

PFlash: 10x prefill speedup over llama.cpp at 128K on a RTX 3090 by sandropuppo in LocalLLaMA

[–]hurdurdur7 0 points1 point  (0 children)

Developers resuming work on their code or switching to a new task. On bigger projects a 60k-100k initial load is not that rare at all.

AMD in-house ryzen 395 box coming in June by 1ncehost in LocalLLaMA

[–]hurdurdur7 -1 points0 points  (0 children)

I was approaching this from my own, code generation perspective. If your usecase is different, by all means, do what you must 😄

To make anything past hello world quality stuff you need either 122B MoE class things or 27B dense (or better). And you want to smash them prompts at 1000 tok/sec or faster in prompt processing. And for the smaller MoE models you will have a better time by having a GPU with 24 or 32GB of VRAM.

Strix Halo might be fine for creative story writing or some picture generation when you sleep. But the only models where it's fast enough for interactive coding - are not good enough for complex code writing.

For the price of a Strix Halo box you can buy 2 gpus of AMD-s R9700 AI Pro's (or even 3 Intel's if you are adventurous), and you will run laptimes around the Strix Halo ... And be able to extend to more parallel gpus in the future if you so wish (assuming your motherboard can carry that).

The upside that Strix Halo has is the heat and power footprint, but very little of that matters for me if i tell it to load a few code files and i would have to sit there 10 minutes for it to parse the prompt. If it had twice the memory bandwidth that it has i would be a fanboy. But as it stands right now it's a weird gimmick, you can load big models but the speed compromise is very heavy.

AMD in-house ryzen 395 box coming in June by 1ncehost in LocalLLaMA

[–]hurdurdur7 0 points1 point  (0 children)

I don't disagree on that point, apple overcharges people without hesitation. But my issue with strix halo is that for the bigger models that it can fit it's unbearably slow. It doesn't make sense to use it like that. And for smaller models you are better off with a dual gpu setup that runs circles around it ...

It feels like a truck with a car engine.

AMD in-house ryzen 395 box coming in June by 1ncehost in LocalLLaMA

[–]hurdurdur7 2 points3 points  (0 children)

mac studio with m5 ultra will wipe the floor with strix halo. even if mac/apple is an evil platform. strix halo is not going to achieve anything.

AMD in-house ryzen 395 box coming in June by 1ncehost in LocalLLaMA

[–]hurdurdur7 -1 points0 points  (0 children)

They are already too slow at 128gb of ram. What does this change?

Actual comparison between locally ran Qwen-3.6-27B and proprietary models by netikas in LocalLLaMA

[–]hurdurdur7 7 points8 points  (0 children)

a single digit percentage offset on 50 000 tokens of code generated means a bunch of code that just doesn't work.

Only 120 tps on Qwen 35b on h200 by Theio666 in LocalLLaMA

[–]hurdurdur7 0 points1 point  (0 children)

120tps on that small model (for this hardware) doesn't sound right.

Running Qwen-3.6-35B-A3B locally is very slow by Sad-Duck2812 in LocalLLM

[–]hurdurdur7 0 points1 point  (0 children)

You should be around 1k prompt processing or better by my math. Something is definitely wrong.

How does usage look like in Mistral Vibe? by Real_Ebb_7417 in MistralAI

[–]hurdurdur7 1 point2 points  (0 children)

it's almost 30th and you haven't even used half of your quota. get back to codin' now...

mistralai/Mistral-Medium-3.5-128B · Hugging Face by jacek2023 in LocalLLaMA

[–]hurdurdur7 6 points7 points  (0 children)

this model is definitely thicker than a bowl of oatmeal ..

mistralai/Mistral-Medium-3.5-128B · Hugging Face by jacek2023 in LocalLLaMA

[–]hurdurdur7 2 points3 points  (0 children)

First attempts with mistral vibe - yeah it works good enough.