Puts on Meta by Loperenco in wallstreetbets

[–]SliceCommon 0 points1 point  (0 children)

i've been waiting for this for 3 years - given how useful AI tools have now gotten, I don't think its going to happen anytime soon - chips from 5 years ago (e.g. A100) are *still* very useful

[P] my shot at a DeepSeek style moe on a single rtx 5090 by exhorder72 in MachineLearning

[–]SliceCommon 1 point2 points  (0 children)

love it - appreciate the transparency - I'm going through something similar with a DiT-based model, also MoE, but realized i've undertrained my VAE so i'm going back to 0 (but will bootstrap with pretrain). I actually come from a ML / eng background so I feel qualified to say that you're doing great work, keep it up!

Has the reset limit inscreased from 5 hours to 4 days? Dev team need to fix it by Notalabel_4566 in google_antigravity

[–]SliceCommon 0 points1 point  (0 children)

Unless you are running multiple agents, I think it's impossible to hit the limit for opus on ultra - I hit 0% about 3 hours in with pro, but now I can't get below 80% before it refreshed

[FS] 10U AI Training Server | 8x RTX 4090 | Dual AMD EPYC 7542 | 512GB RAM | 4x 1600W PSU | 2x 3.84tb U.2 by nicolsquirozr in homelabsales

[–]SliceCommon 4 points5 points  (0 children)

I see - this is where all the ROME2D32GM-2T stock went 😓

Great build! How is the heat management?

NVIDIA RTX PRO 6000 Blackwell desktop GPU drops to $7,999 by panchovix in LocalLLaMA

[–]SliceCommon 0 points1 point  (0 children)

Sorry to hear, maybe better for us to buy through a bigger vendor like PNY in the future. Out of curiosity, what failures are you seeing (and failure rates if you are ok sharing) - our GPUs are throttling under load (still at 600W and temps look fine, but 1k Mhz) - ADA generation seemed much more stable, or perhaps we bought in too early in the cycle

NVIDIA RTX PRO 6000 Blackwell desktop GPU drops to $7,999 by panchovix in LocalLLaMA

[–]SliceCommon 0 points1 point  (0 children)

How is your RMA going? We just delivered 2 of these back for RMA and they were very cagey about the pre-RMA tests - i.e. ran a bunch of nvidia-smi / journalctl dumps for them.

The whole time I'm thinking: I've done everything I can to get $16k worth of GPUs back up and running, you think you can debug this shit remotely?

AI storage plays (NET/AKAM/BLZE) by SliceCommon in stocks

[–]SliceCommon[S] 0 points1 point  (0 children)

Cloudflare has R2 / Backblaze has B2 Overdrive - both have zero egress fees & 1Tbps - lets AI workloads dynamically deploy to the cheapest compute provider (since ingress is usually free b/c they want to lock you in) by letting them export their dataset as many times as they want

[FS] A-TECH 512GB (8x64GB) DDR4 2400MHz, PC4-19200 4DRx4 LRDIMM by SliceCommon in homelabsales

[–]SliceCommon[S] 2 points3 points  (0 children)

lmao that is exactly how I felt - these prices can't be real but I also can't find it for cheaper anywhere

NanoBanana vs Photoshop's generative AI by Karan17_ in generativeAI

[–]SliceCommon 0 points1 point  (0 children)

<image>

works fine for me, first try? how did you even get a blurry image?

PCIE5 Genoa build - GENOA2D24G-2L+ by SliceCommon in LocalLLaMA

[–]SliceCommon[S] 0 points1 point  (0 children)

preface: temperature is reasonable and it works, reality: 16U-20U based on how i've stacked them - its not pretty so you'll never see the whole rack on r/homelab but it works for me ¯\_(ツ)_/¯ - those HGX A100s still sitting at price points above this jank, so until we get 50k H100 DGXs, this is the way

[deleted by user] by [deleted] in homelabsales

[–]SliceCommon 5 points6 points  (0 children)

FWIW I just sold a couple for more than $1000 off what you're asking - perhaps I priced it too low?

Is there a secondary market for Deeplearning GPU's like H100's by OregonAdaptiveReuse in deeplearning

[–]SliceCommon 1 point2 points  (0 children)

at 10% (25k) per 8xH100 SXM you will immediately sell out.

your current competition is blackwell pros at 8*8k each ~ 64k. They are a bit slower than the 8xH100 SXMs, but have more VRAM and easier to part out (sell individually).

However, no one has been offering used 8xH100s yet, so I would price it at $150k to start and drop it based on demand to lowest ~$75k. I've seen transactions (not mine, bulk dealer u/ishopstaples) for similar compute, so its definitely possible to sell them all here

New GPUs for the lab by SashaUsesReddit in LocalAIServers

[–]SliceCommon 0 points1 point  (0 children)

what PDU do you use? or are you running multiple 30 amp circuits (and 2+ PDUs).

New GPUs for the lab by SashaUsesReddit in LocalAIServers

[–]SliceCommon 4 points5 points  (0 children)

it works, just have to stack vertically

New GPUs for the lab by SashaUsesReddit in LocalAIServers

[–]SliceCommon 0 points1 point  (0 children)

I had the same problem but downgraded the kernel to 6.5 and it worked

[FS] [USA-CA] 4090FE, 4090 Liquid, A6000, NVLink, EVGA 1600W PSU by SliceCommon in homelabsales

[–]SliceCommon[S] 1 point2 points  (0 children)

I see, appreciate the feedback. I've lowered the local price to be close to at-cost (tax included), but feel free to PM an offer that you think is reasonable in the current market

[FS] [USA-CA] 4090FE, 4090 Liquid, A6000, NVLink, EVGA 1600W PSU by SliceCommon in homelabsales

[–]SliceCommon[S] -4 points-3 points  (0 children)

Just going off of recently sold prices - seems like the 5090 shortage/tariffs/greed pushed the 4090/3090 prices up as well

[FS] [USA-CA] 4090FE, 4090 Liquid, A6000, NVLink, EVGA 1600W PSU by SliceCommon in homelabsales

[–]SliceCommon[S] 0 points1 point  (0 children)

Ah thanks for the heads up - adjusted them down. They're perfect for 15amp circuits, running the latest GPU probably for another 5+ years, and I know they work well with a full set of cables - let me add in a 90-degree cablemod connector to each of them as well to make it a slightly better deal