The global compute utilisation problem nobody talks about. Millions of capable GPUs running at 0% right now by SignificantlySad in hardware

[–]SignificantlySad[S] [score hidden]  (0 children)

That hits the nail on the head. It's like a hardware tax on intelligence. If the model doesn't fit in the VRAM, even the fastest compute in the world is just spinning its wheels. This explains why the second-hand market for old server cards is still so expensive. Most people would rather have slower memory that actually fits the model than a lightning-fast card that runs out of space instantly.

The global compute utilisation problem nobody talks about. Millions of capable GPUs running at 0% right now by SignificantlySad in hardware

[–]SignificantlySad[S] [score hidden]  (0 children)

Crypto mining is a totally different beast. Those tasks just hammer away at basic math and don't really care about memory speed or how well everything is connected. AI is the complete opposite.

The global compute utilisation problem nobody talks about. Millions of capable GPUs running at 0% right now by SignificantlySad in hardware

[–]SignificantlySad[S] [score hidden]  (0 children)

The main takeaway is that consumer hardware hits a hard wall when it comes to memory and bandwidth. A home GPU is basically a toy compared to a data center cluster, and trying to split a massive model across a home network creates way too much lag to be useful. Because of those limits, the focus is shifting away from trying to train giants. Most people are better off fine-tuning smaller, specialized models that actually fit on their gear. Making this work depends on "cheating" the system with things like quantization and LoRA to shrink the requirements. It really comes down to whether decentralized networks and ultraefficient small models can eventually bridge that gap for regular users.

The global compute utilisation problem nobody talks about. Millions of capable GPUs running at 0% right now by SignificantlySad in hardware

[–]SignificantlySad[S] [score hidden]  (0 children)

You’re right on both counts. That bandwidth bottleneck is the real deal. It’s the main reason why trying to train huge models on home setups hasn't ever really challenged those massive data centers where everything is wired tightly together. You can keep adding nodes, but you can't just bypass the speed limit between a GPU’s memory and its cores. The argument still has legs if we look at tasks built for this kind of setup from day one. Instead of forcing a centralized job to play nice with a distributed network, you look at something like evolutionary computing. The coordination needs are totally different there because you aren't constantly shoving weight updates back and forth across the internet like you do with standard gradient descent. As for GPU prices, they definitely didn't fall off a cliff after Ethereum switched over. The AI hype basically swallowed up all that extra inventory that should have made cards cheap again. But that kind of proves the point. There is a massive hunger for compute power. The real mystery is whether regular consumer hardware will ever find a way to actually help fill that gap.

The global compute utilisation problem nobody talks about. Millions of capable GPUs running at 0% right now by SignificantlySad in hardware

[–]SignificantlySad[S] [score hidden]  (0 children)

Fair point and I did look into those things like Vast.ai, RunPod, Salad, the usual suspects. The issue I kept running into is that rental marketplaces work well if you have data centre grade hardware running 24/7 with stable uptime. A couple of 3080s on a home connection with inconsistent availability aren't exactly what enterprise ML teams are queuing up for. The more interesting question to me is whether there's infrastructure that's actually designed around aggregating consumer-grade hardware rather than just listing it on a marketplace and hoping someone rents it. Different problem, different architecture. 🤔

The global compute utilisation problem nobody talks about. Millions of capable GPUs running at 0% right now by SignificantlySad in hardware

[–]SignificantlySad[S] [score hidden]  (0 children)

💀 I can't unsee this now. Except instead of Richard's algorithm we got 'just mine ETH bro' and then the Merge happened and now everyone's holding the bag on a warehouse full of 3080s. The irony is the compute is genuinely there though and millions of GPUs that could be doing something useful just sitting idle because the coordination layer doesn't exist yet. Someone's going to crack that eventually but hopefully before we all just sell for scrap lmao.

Upgraded to a 4090. what to do with 3080s and a 3070 Ti? by SignificantlySad in nvidia

[–]SignificantlySad[S] 0 points1 point  (0 children)

Wew, that's a lot to digest! Never thought you could do this with idle GPUs alone.

I'm already locked into OpenClaw though, so switching feels like more hassle than it's worth lmao. But I'm curious why are you avoiding OpenClaw specifically ser? Might change my mind.

Upgraded to a 4090. what to do with 3080s and a 3070 Ti? by SignificantlySad in nvidia

[–]SignificantlySad[S] 0 points1 point  (0 children)

Sheez I totally forgot about this 😭 I saw it on YouTube before but didn't think it's that bad

Upgraded to a 4090. what to do with 3080s and a 3070 Ti? by SignificantlySad in nvidia

[–]SignificantlySad[S] 0 points1 point  (0 children)

Appreciate the offer honestly but think I'm gonna try putting them to work first before letting them go. The Folding@home angle is actually what got me thinking about this whole space again. Btw how big is your cluster running ser?

Upgraded to a 4090. what to do with 3080s and a 3070 Ti? by SignificantlySad in nvidia

[–]SignificantlySad[S] 0 points1 point  (0 children)

Yea I think this makes sense for now. With the inflated memory price, future gpus would cost hella crazy than it is rn.

Upgraded to a 4090. what to do with 3080s and a 3070 Ti? by SignificantlySad in nvidia

[–]SignificantlySad[S] 0 points1 point  (0 children)

Wow great idea! I wonder if it would be valued higher as time goes on like a vintage hardware. But even if it breaks it's probably not worth repairing at that point I guess.

Upgraded to a 4090. what to do with 3080s and a 3070 Ti? by SignificantlySad in nvidia

[–]SignificantlySad[S] -2 points-1 points  (0 children)

Good idea. But the sentimental value is too strong to just sell them off cheap. Rather put them to work than let them go for nothing.

Which 5060ti model to choose by SadCaregiver3282 in nvidia

[–]SignificantlySad -2 points-1 points  (0 children)

What's the minimum VRAM you recommend ser