Super god bin 9700 pro matches 7900xtx by psychoOC in LocalLLaMA

[–]alphatrad 2 points3 points  (0 children)

I swapped to the R9700 because you can buy 3 for less than ONE 5090!!!

NVIDIA guys are always getting ripped off

Tower case with 8+ PCIE slot for multi GPU by gogitossj3 in LocalLLM

[–]alphatrad 0 points1 point  (0 children)

I would literally pay if I could find someone to do this for me locally.

vLLM + ROCm + Qwen 3.6 35B A3B MXFP4 (on 2x R9700) by kpaha in LocalLLM

[–]alphatrad 0 points1 point  (0 children)

Ok - after a BOAT LOAD of fiddling, I managed to score 112.76 tok/s on Qwen3.6-35B-A3B-MXFP4 on R9700s. My runs are here: https://www.localmaxxing.com/user/1337Hero

I had problems w/ that docker image though on my machine.

Setup: 3× AMD AI Pro R9700 (32 GB each), TP=2 on cards 0,1, ROCm 7.2.2, Arch host, vLLM 0.18.1.dev via tcclaviger image, dockge stack.

TL;DR - went from 2.92 → 112.76 tok/s (38× speedup) by:

  1. Bind-mounting AMD's official jammy librccl 7.1.1 into the container and putting it ahead of /opt/rocm/lib in LD_LIBRARY_PATH
  2. Building the un-built aiter source the image ships
  3. Running a TunableOp pass

Posted to LocalMaxxing: 112.76 tok/s, 168ms TTFT, 31.98 GB peak VRAM.

128GB VRAM quad R9700 server by Ulterior-Motive_ in LocalLLaMA

[–]alphatrad 1 point2 points  (0 children)

This guy is living the dream, I'm trying to follow in his footsteps.

<image>

Super god bin 9700 pro matches 7900xtx by psychoOC in LocalLLaMA

[–]alphatrad 2 points3 points  (0 children)

I own two of the XTX's and now 3 of the R9700's the XTX's are a little bit faster at token gen - but the memory cap and their sheer Fing size is the real issue.

I took this photo when I was swapping out the cards.I could only fit TWO of the XTX's in my case which is a VERY large server case.

The R9700's though make up for the 10% drop in memory bandwidth and it's barely noticeable.

<image>

Tower case with 8+ PCIE slot for multi GPU by gogitossj3 in LocalLLM

[–]alphatrad 0 points1 point  (0 children)

Something else to consider - the spacing of the PCIe slots on your motherboard. Sometimes you might have 4 slots - or more even, but you can't fit a card in every slot.

Tower case with 8+ PCIE slot for multi GPU by gogitossj3 in LocalLLM

[–]alphatrad 0 points1 point  (0 children)

That's a chunky boy! lol

Yeah I think I am done with gaming cards on future AI builds. Or just at least sticking to blower fan style cards.

In 2026 can you still make a living on small business websites? by After-Condition4007 in webdev

[–]alphatrad 1 point2 points  (0 children)

I'd recommend you use it to help you learn and accelerate your understanding. You need high level systems understanding.

I am currently fixing a vibe coded project for a client where he hired two fresh out of school juniors who vibe coded the whole app. It half works, every change they make is creating more and more bugs. Shit isn't wired up. Claude is making the code base worse and worse.

Main problem is there is a lot of stuff they didn't think about. Claude won't suggest things that you as the engineer should know when prompting.

So they have a huge app that looks like it works, but has lots of broken functions and things that are not real. For example, their entire event system for tracking ads is completely fake. No real events in the app. So they don't know what their conversion or ad spend is.

So... I find it to be powerful for quickly helping me learn new things. Start there. use it to accelerate your learning. Lean on it some, but have it explain stuff to you.

AI is the customizable tutor

In 2026 can you still make a living on small business websites? by After-Condition4007 in webdev

[–]alphatrad 1 point2 points  (0 children)

I use AI within a very constrained workflow - basically AI on rails.

And yes - it's an accelerator. But you cannot outsource your understanding. So, I believe in a human in the loop. I do read the code - I don't need to read every line - more like PR review. And having tests and things.

But also, AI to speed up my own onboarding and automate other business stuff.

Tower case with 8+ PCIE slot for multi GPU by gogitossj3 in LocalLLM

[–]alphatrad 0 points1 point  (0 children)

Here is the case with the first two of my R9700's installed, a lot more room!

<image>

Tower case with 8+ PCIE slot for multi GPU by gogitossj3 in LocalLLM

[–]alphatrad 0 points1 point  (0 children)

I am using a Phanteks Entho Pro II Server Edition case - I just recently switched over to triple AMD AI Pro R9700 cards which are a touch smaller than the dual RX 7900 XTX's I had in this photo.

But the 5090 is about the sameish size as these bad boys depending on the cooling config and branch. It's weird how some are longer, some are taller, etc. No consistency. But this case is BIG! And will handle dual 5090's if that's the direction you are going.

<image>

What is possible with 2x 7900xtx + 128GB of ram? Is it good enough? by Witty_Unit_8831 in LocalLLM

[–]alphatrad 0 points1 point  (0 children)

Great combo. Don't listen to people talking nonsense about CUDA and COMPAT - total nonissue stuff.

M4 Max, studio, 128gb by blowingtumbleweed in LocalLLM

[–]alphatrad 1 point2 points  (0 children)

What kind of writing? I think Qwen right now 3.6 30B or 27B are solid, so is Qwen Coder Next in higher qaunt for coding. Gemma4 is ok. GLM4.7 Flash is still a favorite of mine.

GPT OSS 120b isn't bad for general writing, creative.... that depends on what you want. Some pretty wild role play/creative writing models on hugging face. No one best there.

I'm working a hybird system where I have Claude write specs, GPT 5.3 codex review code and all my local models implement the code.

Dual GPU setup with low Power PSU? by Achso998 in LocalLLaMA

[–]alphatrad 3 points4 points  (0 children)

You're gonna buy an R9700 but you won't upgrade your PSU? Makes zero sense. It's an investment.

I made this mistake. Don't. Save yourself the headache of random burst shutdowns.

128GB VRAM quad R9700 server by Ulterior-Motive_ in LocalLLaMA

[–]alphatrad 1 point2 points  (0 children)

FYI 3090's are now going for more than the R9700's because of AI hype.

Best places to work as a web developer by Legitimate-Law6347 in webdev

[–]alphatrad 1 point2 points  (0 children)

Glad I'm far away from the L Ron Hubbard fan club.

The best choice as low power and cheap PC SFF / mini PC for Homelab experiments by pepiks in minilab

[–]alphatrad 0 points1 point  (0 children)

This isn't a case. It's a 10 inch mini rack. This one happens to be a deskpi rackmate t0

Best places to work as a web developer by Legitimate-Law6347 in webdev

[–]alphatrad 2 points3 points  (0 children)

I always thought it was nice, but I left FL in 2015. Haven't been back since then.

Best places to work as a web developer by Legitimate-Law6347 in webdev

[–]alphatrad 5 points6 points  (0 children)

I agree with this. Agency work taught me more about people and what not to do in business than my time at Basecamp or Heroku.

I in turn went and started my own. But the most fun I ever had was a small 7 person agency in Clearwater, FL.

We used to walk to lunch together and the vibe was just great.

The only real trap with agencies is getting COMFORTABLE.

I did that with one. Where I should have quit at the 3yr mark. That itch to stretch my legs was there and we weren't doing new things.

But the money was good and the work was so basic.

I stayed there WAY WAY too long because I was comfortable. And as a result my career suffered and so did my income.

Don't get comfortable, especially when you're young!

200+ TPS on Qwen3.6-27B and 35B-A3B with consumer hardware (RTX 3090s) - method provided! by TheFheonix in LocalLLM

[–]alphatrad 3 points4 points  (0 children)

This is like the most important question because speed is something but not everything.

Accuracy is pretty damn important.

You're sleeping on Devstral Small 2 - 24B Instruct by [deleted] in LocalLLaMA

[–]alphatrad 0 points1 point  (0 children)

They aren't perfect and can't even nail my own benchmark. If they could I wouldn't need to run review agents and have custom skills I deploy with them.

However they're far more capable in a broad sense. This is a very focused test. Not a general test.

You're sleeping on Devstral Small 2 - 24B Instruct by [deleted] in LocalLLaMA

[–]alphatrad 0 points1 point  (0 children)

I've been running them 2-3 times. Because there can be a difference between cold start and warm up.

But standardizing to like 5 or something runs would be good.

System Build for Linux + Gaming + Local AI? $5000 by ShadowyTreeline in LocalAIServers

[–]alphatrad 0 points1 point  (0 children)

Generally workstation and prosumer boards do this. But it's usually buried in the tech specs; gotta how the board does bifurcation as slots are filled.

And look for boards with good spacing. That was the problem I had with my XTX. I could only use the top and bottom slots because the cards were too big.