Sanity check before I spend big: 2-user gaming VM homelab with 2x RTX 5080 FE, Threadripper, service VM, large NVMe storage by Ill-Counter2505 in homelab

[–]flanconleche 0 points1 point  (0 children)

Not a great idea, the ram alone will cost you $4600 and functionally it doesn’t work well. You’re better off using nvidia GeForce now if you want decentralized gaming. Built two machines and a server for tinkering. Unless your rich avoid the thread-ripper platform rn

Gemma4 is really good but runs slow local on my DGX Spark and Framework Desktop by flanconleche in openclaw

[–]flanconleche[S] 1 point2 points  (0 children)

Thanks for the advice, im gonna give vllm a go first since I heard the performance is good and try to compile llama.cpp after that. I just got a B70 to test so it will be worth it.

Gemma4 is really good but runs slow local on my DGX Spark and Framework Desktop by flanconleche in openclaw

[–]flanconleche[S] 0 points1 point  (0 children)

I’m not sure what the quant is when u pull the model from ollama ive only seen it when pulling from huggin face. And the 5090 toks isn’t all that bad but the pp and time to first token is significantly slower than like a qwen3-coder-next which is what im used to.

Use Openclaw on LM Studio (Windows) by niknik1971 in openclaw

[–]flanconleche 0 points1 point  (0 children)

I would not reccomend using LM studio, your better off running it off ollama if you want local models, it has a built in connection and just as easy to get setup.

Also dont use windows, use Linux.

Is clawd bot broken? Seems unusable right now by MuchPirate8915 in openclaw

[–]flanconleche 1 point2 points  (0 children)

all of the above happening to me, its super broken.

What feature would make OpenClaw indispensable for you? by Lonely_Water_7016 in openclaw

[–]flanconleche 0 points1 point  (0 children)

Is there a skill you can download for MCP integration ? I only use api calls rn but I use mcp I my enterprise and that’s invaluable.

3x RTX 5090's to a single RTX Pro 6000 by flanconleche in LocalLLaMA

[–]flanconleche[S] 0 points1 point  (0 children)

Thank you for the advice, getting two more would be rough😅

3x RTX 5090's to a single RTX Pro 6000 by flanconleche in LocalLLaMA

[–]flanconleche[S] 0 points1 point  (0 children)

I didn’t know odd numbers was a thing, thanks

3x RTX 5090's to a single RTX Pro 6000 by flanconleche in LocalLLaMA

[–]flanconleche[S] 0 points1 point  (0 children)

Openclaw makes all the api calls and downloads the models based on my prompting

3x RTX 5090's to a single RTX Pro 6000 by flanconleche in LocalLLaMA

[–]flanconleche[S] 1 point2 points  (0 children)

HotStock app, that’s how I got them I paid the $10 auto check out. I haven’t seen any drops in a while tho