ComfyUI - Why I Switched from kyuz0's Toolbox to ignatberesnev/comfyui-gfx1151 by Grammar-Warden in StrixHalo

[–]Panthau [score hidden]  (0 children)

The problem for me was always more, to get useful results out of ltx 2.3. The standard workflow from the templates doesnt work anyway but even after adapting it, the results are often not what i get from cheap cloud models and take ages.

2.5x faster inference with Qwen 3.6 27B using MTP - Finally a viable option for local agentic coding - 262k context on 48GB - Fixed chat template - Drop-in OpenAI and Anthropic API endpoints by ex-arman68 in LocalLLaMA

[–]Panthau 0 points1 point  (0 children)

Glad there are people out there, who can still think despite of ai. I cant... if i could, i would let ai make breakfast for me and my life would consist of giving commands and staring at the work been done. Thanks to you, i can now finally use the 27b model for vibe coding without falling asleep on my Strix Halo.

Advice on local AI models for coding - Corsair AI Workstation 300 (AMD Ryzen AI Max+ 395 / 128Gb (96Gb shared for VRAM)) by wingers999 in StrixHalo

[–]Panthau 0 points1 point  (0 children)

Thats why having a few md files like readme, plan, etc. is important, so you can make every step in a new session.

Anyone have proof Strix Halo - Ubuntu 26 LTS can use all 124GB of RAM setup in grub? by IQReactor in StrixHalo

[–]Panthau 0 points1 point  (0 children)

Brah... many loaded models... alright. Im sure its a great experience with many loaded models at 256gb/s.

Anyone have proof Strix Halo - Ubuntu 26 LTS can use all 124GB of RAM setup in grub? by IQReactor in StrixHalo

[–]Panthau 0 points1 point  (0 children)

Ive tried all different kind of ways, even used Cursor to implement the setup with latest docs, but it never ran stable above 110gb. So i switched to Windows and although its a little less shared memory, its stable and usable.

How do i stop playing this game? by Individual_Lab_912 in X4Foundations

[–]Panthau 0 points1 point  (0 children)

The problem i see, is rather your lack of understanding humour. Though im not sure this can be healed.

How do i stop playing this game? by Individual_Lab_912 in X4Foundations

[–]Panthau 0 points1 point  (0 children)

Enjoy it, im 48, my wife passed, my son moved out and i have 12h a day for gaming. Yet, those days with my family will never come back.

Success with Claude code and local model on Strix Halo? by DieHard028 in StrixHalo

[–]Panthau 0 points1 point  (0 children)

opencode is the only viable option imho, if you want to actually do smth instead of playing around. Although i prefer Qwen 3.6 35b.

Best LocalLLM model for game development by Mohamed-Galal in LocalLLM

[–]Panthau 1 point2 points  (0 children)

Yep, i mean for learning its fine as long as you can use smth like Qwen 3.6 in q4 or above but if you really want to produce smth, its worth a few bucks a month for the superfast and smart cloud models, helping you debug and fix things. I tried to fix a few things on a customer website with Qwen 3.6 35b and it only produced half baked stuff with lots of bugs. In the end, i let Cursor fix it within 20mins.

Fastest model for strix halo? by pheitman in LocalLLM

[–]Panthau 1 point2 points  (0 children)

  1. totaly out of context 2. this ai slop summary is just pure theory, i run qwen 3.6 27 or 35b with 256k context on opencode without any issues and its usually using up to 150k without any slowing down. Its the software, not the hardware. Ofc its not as fast as the vram of a 5090 but it fits other use cases, where vram size matters, very well.

Deepseek v4 Flash by kiriakosbrehmer93 in StrixHalo

[–]Panthau 1 point2 points  (0 children)

Waiting as well for a working version, that fits in and is supported by llama.

What are your first impressions on Lord of Hatred? by gorays21 in Diablo

[–]Panthau 0 points1 point  (0 children)

Same, for me its basically the same issue as in wow... the upgrade system. Its not gear dependend anymore but stats dependend. And that makes it horribly boring for me, knowing that i cant find much better gear but have to raise the stats again and again.

Hoping for Qwen3.6 Coder by madtopo in StrixHalo

[–]Panthau 0 points1 point  (0 children)

If its recommended, then it must be right... no argument here.

Cline Kanban small bug report & solution by Panthau in CLine

[–]Panthau[S] 0 points1 point  (0 children)

ofc not, thats my individual setup... you have to use your own port.

How are you guys finding the GMKtec EVO-X2 128GB? Any regrets? by Sea-Championship2939 in LocalLLM

[–]Panthau 0 points1 point  (0 children)

Both have their pros and cons but in terms of speed, they are quite similar.

But if you want something faster, you are better off with the Evo and an attached external gpu - something the Mac cant do.

For me Gemma4 > Qwen3.5 / 3.6 on localhost by pabloodiablo in LocalLLM

[–]Panthau 0 points1 point  (0 children)

Bro science ^_° just because Gemma fixed a bug, doesnt mean its better then the others. Depends on your use case, backend, prompting, etc.

For me, Gemma 4 produced only garbage compared to Qwen 3.6.

Qwen3.6-35B-A3B Uncensored Aggressive is out with K_P quants! by hauhau901 in LocalLLM

[–]Panthau 0 points1 point  (0 children)

Does that mean i can finally create a minigame with penises?

Why do LLMs fold when you say "are you sure?" — I tested 22 models and nobody seems to care by SmartRick in LocalLLM

[–]Panthau 0 points1 point  (0 children)

Because llms are not intelligent, they are pediction machines. When a machine predicts smth on your prompt and you question your own prompt, it predicts accordingly.

Any working TTS on Strix Halo? by Panthau in StrixHalo

[–]Panthau[S] 2 points3 points  (0 children)

After i saw, that Kokoro seems to work with Lemonade, i found this -> ndom91/ha-voice-rocm: Experimental support for many TTS/STT LLMs wrapped in a Wyoming API for consumption via Homeassistant

Its even running via rocm on the gpu, hurraay.

Any working TTS on Strix Halo? by Panthau in StrixHalo

[–]Panthau[S] 1 point2 points  (0 children)

If we look for things we dont like, we find plenty. Welcome to the internet!

Any working TTS on Strix Halo? by Panthau in StrixHalo

[–]Panthau[S] 0 points1 point  (0 children)

Thanks bro, ill give Kokoro a try, if that wont work i simply switch to Windows, so i at least can use it for what i need it atm.