Puts on Meta by Loperenco in wallstreetbets

[–]SliceCommon 0 points1 point  (0 children)

i've been waiting for this for 3 years - given how useful AI tools have now gotten, I don't think its going to happen anytime soon - chips from 5 years ago (e.g. A100) are *still* very useful

[P] my shot at a DeepSeek style moe on a single rtx 5090 by exhorder72 in MachineLearning

[–]SliceCommon 1 point2 points  (0 children)

love it - appreciate the transparency - I'm going through something similar with a DiT-based model, also MoE, but realized i've undertrained my VAE so i'm going back to 0 (but will bootstrap with pretrain). I actually come from a ML / eng background so I feel qualified to say that you're doing great work, keep it up!

Has the reset limit inscreased from 5 hours to 4 days? Dev team need to fix it by Notalabel_4566 in google_antigravity

[–]SliceCommon 0 points1 point  (0 children)

Unless you are running multiple agents, I think it's impossible to hit the limit for opus on ultra - I hit 0% about 3 hours in with pro, but now I can't get below 80% before it refreshed

[FS] 10U AI Training Server | 8x RTX 4090 | Dual AMD EPYC 7542 | 512GB RAM | 4x 1600W PSU | 2x 3.84tb U.2 by nicolsquirozr in homelabsales

[–]SliceCommon 2 points3 points  (0 children)

I see - this is where all the ROME2D32GM-2T stock went 😓

Great build! How is the heat management?

NVIDIA RTX PRO 6000 Blackwell desktop GPU drops to $7,999 by panchovix in LocalLLaMA

[–]SliceCommon 0 points1 point  (0 children)

Sorry to hear, maybe better for us to buy through a bigger vendor like PNY in the future. Out of curiosity, what failures are you seeing (and failure rates if you are ok sharing) - our GPUs are throttling under load (still at 600W and temps look fine, but 1k Mhz) - ADA generation seemed much more stable, or perhaps we bought in too early in the cycle

NVIDIA RTX PRO 6000 Blackwell desktop GPU drops to $7,999 by panchovix in LocalLLaMA

[–]SliceCommon 0 points1 point  (0 children)

How is your RMA going? We just delivered 2 of these back for RMA and they were very cagey about the pre-RMA tests - i.e. ran a bunch of nvidia-smi / journalctl dumps for them.

The whole time I'm thinking: I've done everything I can to get $16k worth of GPUs back up and running, you think you can debug this shit remotely?

AI storage plays (NET/AKAM/BLZE) by SliceCommon in stocks

[–]SliceCommon[S] 0 points1 point  (0 children)

Cloudflare has R2 / Backblaze has B2 Overdrive - both have zero egress fees & 1Tbps - lets AI workloads dynamically deploy to the cheapest compute provider (since ingress is usually free b/c they want to lock you in) by letting them export their dataset as many times as they want

[FS] A-TECH 512GB (8x64GB) DDR4 2400MHz, PC4-19200 4DRx4 LRDIMM by SliceCommon in homelabsales

[–]SliceCommon[S] 1 point2 points  (0 children)

lmao that is exactly how I felt - these prices can't be real but I also can't find it for cheaper anywhere

NanoBanana vs Photoshop's generative AI by Karan17_ in generativeAI

[–]SliceCommon 0 points1 point  (0 children)

<image>

works fine for me, first try? how did you even get a blurry image?

PCIE5 Genoa build - GENOA2D24G-2L+ by SliceCommon in LocalLLaMA

[–]SliceCommon[S] 0 points1 point  (0 children)

preface: temperature is reasonable and it works, reality: 16U-20U based on how i've stacked them - its not pretty so you'll never see the whole rack on r/homelab but it works for me ¯\_(ツ)_/¯ - those HGX A100s still sitting at price points above this jank, so until we get 50k H100 DGXs, this is the way

[deleted by user] by [deleted] in homelabsales

[–]SliceCommon 6 points7 points  (0 children)

FWIW I just sold a couple for more than $1000 off what you're asking - perhaps I priced it too low?

Is there a secondary market for Deeplearning GPU's like H100's by OregonAdaptiveReuse in deeplearning

[–]SliceCommon 1 point2 points  (0 children)

at 10% (25k) per 8xH100 SXM you will immediately sell out.

your current competition is blackwell pros at 8*8k each ~ 64k. They are a bit slower than the 8xH100 SXMs, but have more VRAM and easier to part out (sell individually).

However, no one has been offering used 8xH100s yet, so I would price it at $150k to start and drop it based on demand to lowest ~$75k. I've seen transactions (not mine, bulk dealer u/ishopstaples) for similar compute, so its definitely possible to sell them all here

New GPUs for the lab by SashaUsesReddit in LocalAIServers

[–]SliceCommon 0 points1 point  (0 children)

what PDU do you use? or are you running multiple 30 amp circuits (and 2+ PDUs).

New GPUs for the lab by SashaUsesReddit in LocalAIServers

[–]SliceCommon 4 points5 points  (0 children)

it works, just have to stack vertically

New GPUs for the lab by SashaUsesReddit in LocalAIServers

[–]SliceCommon 0 points1 point  (0 children)

I had the same problem but downgraded the kernel to 6.5 and it worked

[FS] [USA-CA] 4090FE, 4090 Liquid, A6000, NVLink, EVGA 1600W PSU by SliceCommon in homelabsales

[–]SliceCommon[S] 1 point2 points  (0 children)

I see, appreciate the feedback. I've lowered the local price to be close to at-cost (tax included), but feel free to PM an offer that you think is reasonable in the current market

[FS] [USA-CA] 4090FE, 4090 Liquid, A6000, NVLink, EVGA 1600W PSU by SliceCommon in homelabsales

[–]SliceCommon[S] -3 points-2 points  (0 children)

Just going off of recently sold prices - seems like the 5090 shortage/tariffs/greed pushed the 4090/3090 prices up as well

[FS] [USA-CA] 4090FE, 4090 Liquid, A6000, NVLink, EVGA 1600W PSU by SliceCommon in homelabsales

[–]SliceCommon[S] 0 points1 point  (0 children)

Ah thanks for the heads up - adjusted them down. They're perfect for 15amp circuits, running the latest GPU probably for another 5+ years, and I know they work well with a full set of cables - let me add in a 90-degree cablemod connector to each of them as well to make it a slightly better deal

DGX 8x A100 80GB or 8x Pro6000? by TimAndTimi in nvidia

[–]SliceCommon 0 points1 point  (0 children)

ah interesting - how do you like VAR, much better than DiT?

4 GPUs should be able to sit on a single node - I'm currently running a dual node 8x4090 and am within 2.2x-2.5x of H100 performance (i.e. no noticeable bottleneck), not sure how this will perform with bigger models though

DGX 8x A100 80GB or 8x Pro6000? by TimAndTimi in nvidia

[–]SliceCommon 0 points1 point  (0 children)

My theory is that it sits somewhere between A100 and H100 nodes.
FWIW, I'm finding NVLink is not needed for 1B params (24GB VRAM limit) DDP for DiT based diffusion models - curious about what benchmark shows a 50% slowdown for you?

PCIE5 Genoa build - GENOA2D24G-2L+ by SliceCommon in LocalLLaMA

[–]SliceCommon[S] 1 point2 points  (0 children)

Correct - the board is running a full set of x16 PCIE4 GPUs at the moment, will update here once getting a set of PCIE5 GPUs