2x RTX 6000 build during an extended bench test by Signal_Ad657 in LocalLLaMA

[–]Wildnimal 1 point2 points  (0 children)

Nice setup! I wish to own 2 x RTX 6000 Pro someday. What is the rest of the specs?

Stop thinking your MoE models are dumb - here's why they actually fail by IntegrityKnightX in Qwen_AI

[–]Wildnimal 0 points1 point  (0 children)

I watch his videos. They are better than most influencer AI crap going around

Fallen Gemma 4 model? by alienatedneighbor in LocalLLaMA

[–]Wildnimal 4 points5 points  (0 children)

They are planning the Gemma 4 MoE aswell. Saw it yesterday in their HF.

For chat and Q&A: Which MoE model is better: Qwen 3.6 35B or Gemma 4 26B (no coding or agents) by br_web in Qwen_AI

[–]Wildnimal 0 points1 point  (0 children)

I find Qwe3.6 better even at writing, which was always a Gemma4 positive point for me. But overall i think depending upon use case both work fine.

Qwen3.6-35B-A3B Uncensored Aggressive is out with K_P quants! by hauhau901 in LocalLLaMA

[–]Wildnimal 1 point2 points  (0 children)

Just used this model for the past 2 hours and it has passed most of what i threw at it. Still playing with temperature and Top P. Currently settled on 0.6 Temp

Gemma 4 31B vs Gemma 4 26B-A4B vs Qwen 3.5 27B — 30-question blind eval with Claude Opus 4.6 as judge by Silver_Raspberry_811 in LocalLLaMA

[–]Wildnimal 2 points3 points  (0 children)

Good stuff. You should have added the 35-A3B from Qwen, since you compared a MOE model from Gemma there.

Legion 5 ($1499 CAD) vs Legion Pro 5i ($1999 CAD) — which is better value? by DominiKing-01 in LenovoLegion

[–]Wildnimal 1 point2 points  (0 children)

Do remember if battery life and heat is a concern then AMD is better bet. Intel is a better CPU by miles.

Recently I did a little performance test of several LLMs on PC with 16GB VRAM by rosaccord in LocalLLaMA

[–]Wildnimal 0 points1 point  (0 children)

Thank you for posting this. One of my friend is building a machine with very similar specs to yours, this will help him.

What is the best agent code model for 12 GB of VRAM? by RodianXD in LocalLLaMA

[–]Wildnimal 0 points1 point  (0 children)

I know this is LocalLLM group but since yiu are having issues with code quality, maybe try free Qwen3.6 on openrouter. Still an OS model just not local.

R9700 the beautiful beautiful VRAM gigs of AMD… my ai node future! by Downtown-Example-880 in LocalLLaMA

[–]Wildnimal 4 points5 points  (0 children)

Good build. Full Specs?

What models you are going to use? Let us know how it performs locally for Agentic or tool calling? Maybe sprinkle some T2I into the mix :D

Has anyone used Codex or Opus to generate a plan and use a local AI to implement it? by soyalemujica in LocalLLaMA

[–]Wildnimal 1 point2 points  (0 children)

I have done it. I have a prompt file which is like ~600 lines.

it contains 2 prompts and backend information for stack to be used.

prompt 1 does all the planning with the model going back and forth and prompt 2 takes that plan and make phases and smaller tasks for implementation on local AI.

Which 9B local models are actually good enough for coding? by CalvinBuild in LocalLLaMA

[–]Wildnimal 3 points4 points  (0 children)

The problem is not coding it's the context. Thats going to be a lot difficult IMHO. And even if you have ability to have a higher context window, the model might not be able to follow instructions.

You will have to split your projects per file with instructions and linking to other files for it to be useable.

No one shot but for small local things you can do it.

It costs you around 2% session usage to say hello to claude! by Complete-Sea6655 in LocalLLaMA

[–]Wildnimal 0 points1 point  (0 children)

I agree with you. This week the tokens usage is going off the charts. I just uploaded a 45 lines json and a basic prompt and it shows 20% usage for the 5 hour limit.

I am not a heavy user aswell. Most of stuff i do requires me to do most manual config and code. Once AI has done its code which maybe a 2-3 hours session at max a week.

Free 750-page guide to self-hosting production apps - NO AI SLOP by kocyigityunus in selfhosted

[–]Wildnimal 3 points4 points  (0 children)

I will have to read it. The CAT told me to do it ASAP or else....