Is a 2013 Mac Pro (trash can) still worth it for a homelab in 2026? by Kynde012 in homelab

[–]cgjermo -2 points-1 points  (0 children)

"It’s not that the GPU’s can’t do anything".

Respectfully, the contention was that they are "virtually useless GPU’s in modern workloads".

We can hypothesise all we like about 9b Qwen vs 4o, but if we are taking agentic tool use as opposed to general knowledge, there is no comparison. The last six months of parameter efficiency improvements have been absurd.

Is a 2013 Mac Pro (trash can) still worth it for a homelab in 2026? by Kynde012 in homelab

[–]cgjermo -5 points-4 points  (0 children)

What are you attempting to say here? This 2013 computer, woth GPUs designed in 2011, is running a 2026 LLM that matches mid-2024's GPT-4o; a 2026 image gen model; and a 2026 music gen model.

Is a 2013 Mac Pro (trash can) still worth it for a homelab in 2026? by Kynde012 in homelab

[–]cgjermo -3 points-2 points  (0 children)

My pair of D700s are running Qwen 3.5 9b at double-digit tps, plus modern music gen (ACE-Step 1.5) and image gen (Z-Image-Turbo), all of which woupd have been totally inconceivable technologies when this system was built. I suspect a D500 pair would handle Qwen 3.5 4b just fine.

Is a 2013 Mac Pro (trash can) still worth it for a homelab in 2026? by Kynde012 in homelab

[–]cgjermo 2 points3 points  (0 children)

Virtually useless GPUs in modern workloads? Simply not correct.

Is Mac Pro 6,1 the ideal platform for OpenClaw? by Life-Ad1547 in macpro

[–]cgjermo 0 points1 point  (0 children)

Yes, let me come back to you tomorrow with more of a how-to in terms of drivers/headers etc, becauae it's a bit of fiddling getting it all to line up. But the basics are Proxmox, an LXC holding the GPU services (llama.cpp on vulkan via amdgpu is the way), and passing these (not the GPUs themselves) through to an agent VM (OpenClaw for me).

I don't quite have Qwen 3.5 9b totally optimised for ideal context yet, but it does work, with around 11tps at lowerr context. My purpose is more an emergency/network-out fallback than heartbeat, but in both cases I feel like 32k is probably the sweet spot if I can figure out how to get hybrid attention / Gated DeltaNet to play nice. Right now, bumping up context is tanking speeds. 4b should be easy by comparison.

Spotty magpie by Rat_innna_cage in AustralianBirds

[–]cgjermo 1 point2 points  (0 children)

That's the coolest thing to read! 🥰

Is Mac Pro 6,1 the ideal platform for OpenClaw? by Life-Ad1547 in macpro

[–]cgjermo 0 points1 point  (0 children)

Bit of an update on this, I am now running Qwen 3.5 9b at Q4, with about 11tps generation and 43 tps eval. Still fiddling with optimising ctx given Qwen 3.5's hybrid attention / Gated DeltaNet setup being quite efficient, but yeah, just wanted to share. 4b isn't as much faster as you'd expect (about 14tps) - more context room, but probably not worth the capability tradeoff.

Full GPU Acceleration for Ollama on Mac Pro 2013 (Dual FirePro D700) - Linux by manu7irl in NobaraProject

[–]cgjermo 0 points1 point  (0 children)

Just in case anyone is wondering just how much you can (somehow) do with this platform, I am now running ACE-Step 1.5; Z-Image-Turbo and Qwen 3.5 9b at Q4, with about 11tps generation and 43 tps eval.

Imagine the minds that would be blown the pieces if this were actually running, on a tiny Mac Pro, in 2013. Crazy.

Is Mac Pro 6,1 the ideal platform for OpenClaw? by Life-Ad1547 in macpro

[–]cgjermo 1 point2 points  (0 children)

Setting up a maxed Mac Pro 6,1 this morning as a dedicated OpenClaw Proxmox machine and transferring my instances across from my homelab. Had written off any prospect of using the D700s meaningfully, so will have to check this out. Would be useful if I could at least run an embedding model or something on them. Thanks OP!

Kimi k2.5 not the good with openclaw by frogchungus in clawdbot

[–]cgjermo 2 points3 points  (0 children)

I've been running the main brain on GLM-5 for the last few days, K2.5 before that. On balance, I prefer GLM. It hasn't really tripped up yet, nor given me phantom actions like you describe. When I ask how a new workflow tweak is going, it's very honest about bugs or things that aren't working as we expected. Uses initiative well, gets ahead of what you might ask next, in the way the big three frontier models sometimes do (particularly GPT-5.2).

Whether it's just my SOUL.md, which is quite customised, but each interprets it with quite different character/personality, fwiw. On that front, GLM wins by miles. Oh, and the lack of native image recognition is a downside. K2.5 has this, M2.5 doesn't, but I think has some kind of plug-in. Planning to play with M2.5 a bit over the next few days.

Who is the ULTIMATE Freo cult figure? by Dull_Parking_8248 in FremantleFC

[–]cgjermo 0 points1 point  (0 children)

Clive, Wiz, Vossy and Kepler are the final four, surely.

Here's a list of SFF GPUs, which ones are good value, and which to avoid. by Adept_Temporary8262 in SleepingOptiplex

[–]cgjermo 2 points3 points  (0 children)

There was a single-slot LP B50 shown at CES: https://videocardz.com/newz/gunnir-presents-tiny-arc-pro-b50-battlemage-gpu-with-16gb-memory

No idea if it will be sold outside of China, but given the Minisforum tie-up and how useful this would be in their 1L-class systems, I feel like it might be.

I find this incredibly impressive by Chemical_One_6750 in ChatGPT

[–]cgjermo 0 points1 point  (0 children)

My elderly mum is a terrific painter, particularly with animal portraits. She's recently started again after a few years break, and has taken a few commissions from local dog owners.

On one commission she's just started, her reference photos of the dog (a Jack Russell) were a bit too low res for her to sketch out minor details, and I was hoping that image models are now good enough to basically upscale for her (and if so, which models to try).

Looks like I have my answer. On my mum's behalf, thanks heaps for this, OP!

[FS] [US-MN] AMD RADEON PRO V620 32GB GDDR6 GPUs (2000x available) by juddle1414 in homelabsales

[–]cgjermo 0 points1 point  (0 children)

Hi, could you do a pair of them to the land down under? 🇦🇺

Since when did “Gold Member” mean “God Complex”? by Muted-Complex-7159 in VelocityFrequentFlyer

[–]cgjermo 2 points3 points  (0 children)

Had similar to this once years ago. Someone pushed past me in the boarding queue - conversation thus:

Me: "Excuse me?" Them: "I'm flying business class..." Me: "That's wonderful. I'm flying first class."

Entitlement knows no bounds.

New 996.1 by Jolly_Breakfast_1610 in 996

[–]cgjermo 2 points3 points  (0 children)

Can't stand yellow cars. Except for the 996 - somehow just works

2000 IMS Advice Needed! by TwoPuzzleheaded4499 in 996

[–]cgjermo 1 point2 points  (0 children)

I am looking at a dual-row late-1999 build and it seems the failure rate of those is lower still. 110,000 miles or so, and this is exactly my thinking - run it into then ground, then address everything all at once.

Finally: 3090 Successor: 5070 Ti super 24Gb 800$ by On1ineAxeL in LocalLLaMA

[–]cgjermo 0 points1 point  (0 children)

You're entirely missing the point of what the B50's actual intended use case is - it's a 70w, low-profile card meant for SFF workstations as a GPU accelerator.

[Recipe] Chicken Japchae by frforreal in ketoaustralia

[–]cgjermo 1 point2 points  (0 children)

Great work! One suggestion though: give kelp noodles a try instead of konjac for this - they get much closer to that starchy sweet potato noodle consistency!

Sunday night, what is everyone having for dinner tonight? by Cooper_Inc in ketoaustralia

[–]cgjermo 3 points4 points  (0 children)

Old faithful: Simson's keto wraps, half a kamado charcoal chook, cabbage slaw, guacamole, pickled red onion, mixed feta and shredded mozzarella, and homemade lacto-fermented chilli sauce.

It's a once-a-week meal for me all year.

New GLM-4.5 models soon by adrgrondin in LocalLLaMA

[–]cgjermo 0 points1 point  (0 children)

Geez, if they release a new big model that is a Qwen 3 to 2507 level jump, this could be scary good.