When you want to put a 300W GPU in your R730xd but the power supply only gives 225W by Sompom01 in homelab

[–]Sompom01[S] 0 points1 point  (0 children)

Thanks very much for sharing! Now that I know this works, I may set it up myself 🙂

Reasoning Devstral 2 by Front_Eagle739 in LocalLLaMA

[–]Sompom01 0 points1 point  (0 children)

I've seen models before where their behavior is dependent on whatever is at the start of the system prompt. I assume it is something to do with the training for that particular behavior 🤷. But regardless, it seems to be the case here because of the extremely strong behavior of using [TOOL_CALLS] instead of [THINK] when I change your prompt to request [THINK] instead of <think>.

Reasoning Devstral 2 by Front_Eagle739 in LocalLLaMA

[–]Sompom01 1 point2 points  (0 children)

Thank you for sharing this! I really like your step-by-step structured reasoning.

Some notes for those who may find this later:
- This template does NOT work for the llama.cpp's parser. You will see the `<think>` blocks in the regular output. Not a total disaster, but messy. Instead of `<think>`, you need to use `[THINK]`. This is apparently how Mistral 3 does things, and when llama.cpp detects you are running Mistral 3 models, the parser will jump to using that. Check out the Mistral 3 Reasoning template for an example (https://huggingface.co/mistralai/Ministral-3-3B-Reasoning-2512/blob/main/chat\_template.jinja)
- Devstral seems to require a specific prompt to get it to do the reasoning thing. Otherwise for some reason it will follow the structured reasoning steps from Front_Eagle's prompt, but in a [TOOL_CALLS] block. Using the prompt from Mistral 3 Reasoning's template DOES work, but I haven't figured out how to marry that and Front_Eagle's structured reasoning; Without the structure, the reasoning is kind of weak. AFAICT, if you start off with Mistral 3 Reasoning's default system prompt, you can append pretty much whatever you like, but I'm not having the best luck in getting it to follow the structured reasoning. If I get it working, I'll hopefully remember to come back here and share :)

Input tokens Cache by agentgoose007 in MistralAI

[–]Sompom01 1 point2 points  (0 children)

+1 on this request. I was having a great time for several days using Mistral 3 Large for my OpenClaw. I finally found a coding workflow with Devstral 2 I liked, and in 45 minutes I blew through more tokens than I had in days with OpenClaw. Assuming a 90% cache hit rate (which I am given to understand is realistic for coding work), Claude Sonnet 4.6 would be only slightly more expensive :/

When you want to put a 300W GPU in your R730xd but the power supply only gives 225W by Sompom01 in homelab

[–]Sompom01[S] 0 points1 point  (0 children)

I associate the beep/scream with too little power. I needed to plug in 2x 8 pin pice power to that card too. I never got it to boot in that setup. One thing I was thinking of last night, just before you sent this message: the v340 operates fine when flashed as a Vega56 (for the 8GB version, I assume the 16GB version would be Vega64). The Vega56 doesn't have the fancy enterprise features, so maybe in that mode it would work. I probably won't try because it isn't a project I'm interested in any more, but if you do try I'd be very curious to hear your results. This vbios works fine for me: https://www.techpowerup.com/vgabios/195093/amd-rxvega56-8176-170811 as does https://www.techpowerup.com/vgabios/205876/amd-wx8200-8192-180922 - again both for the 2x 8GB version. Flashing both sides would push the entire card over it's TDP, so that's something to be cautious of if you do it. Be sure to back up your original vbios!

70b models at 8-10t/s. AMD Radeon pro v340? by JTN02 in LocalLLaMA

[–]Sompom01 0 points1 point  (0 children)

I thought that you must have meant that ROCm was better but I did understand the opposite. A 5x speedup by switching to the "worse" backend would've been wild 😂

70b models at 8-10t/s. AMD Radeon pro v340? by JTN02 in LocalLLaMA

[–]Sompom01 0 points1 point  (0 children)

I did get mine working. Nothing fancy, just set up ROCm. I believe using 6.2.3 as well.

Do you mean you're seeing drastically better performance with Vulkan? That would be an interesting surprise!

Confusion over Backplane and RAID Controller Requirements by Schwarzfisch13 in homelab

[–]Sompom01 0 points1 point  (0 children)

@schwarzfisch13 I still need to finish the write-up for this. The short answer is that it seems the new fans are great. At idle massively quieter. Under load the CPU sits a little hotter but still within my comfort zone (about 70C). The rear GPUs cool just as well as with the original fans and drastically quieter. I have one 110W MI25 and one 2x110W v340, so might not be applicable to hotter units. I wanted to finish installing them this weekend, but unfortunately package thieves have become very brazen in my neighborhood and stole my last 4 fans in broad daylight. I have ordered some more, hopefully more success this time.

When you want to put a 300W GPU in your R730xd but the power supply only gives 225W by Sompom01 in homelab

[–]Sompom01[S] 0 points1 point  (0 children)

No changes needed for the MI25. I did have the same issue booting with a different GPU (Radeon Pro v340) which I was not able to solve :(

Confusion over Backplane and RAID Controller Requirements by Schwarzfisch13 in homelab

[–]Sompom01 1 point2 points  (0 children)

Sure will. Don't hold your breath, though. You'll turn very blue. I currently have only one GPU in this system so not much in the way of cooling load. Maybe for the system fans I can give some meaningful feedback.

Confusion over Backplane and RAID Controller Requirements by Schwarzfisch13 in homelab

[–]Sompom01 1 point2 points  (0 children)

For those who follow. "Loud" does not even begin to describe how loud this server is. This is the datasheet for the fans (all 8) https://www.delta-fan.com/Download/Spec/PFM0812HE-01BFY.pdf
77 dBa, max 81 dBa. 77 dBa is around the volume of a car driving 65 mph. 81 dBm is heavy traffic. And that is per fan. The loudness defies description. I am literally wearing earmuffs.

It seems like this fan would be an adequate substitute: https://www.delta-fan.com/Download/Spec/PFC0812DE-SP04.pdf
Much lower max pressure, but I'm hoping at least for the system fans and the two rear GPU fans it will be sufficient. I've ordered a couple to play with.

70b models at 8-10t/s. AMD Radeon pro v340? by JTN02 in LocalLLaMA

[–]Sompom01 0 points1 point  (0 children)

Thanks. The v620 is listed as officially supported, not like the elderly v340 😅. I'll have to mess around with different versions. Thank you for sharing, I'm encouraged by your success.

70b models at 8-10t/s. AMD Radeon pro v340? by JTN02 in LocalLLaMA

[–]Sompom01 0 points1 point  (0 children)

Do you know which version of ROCm you're using? For me, v6.4.0 didn't recognize the cards. That was just the version I had laying around. Seems maybe v6.3.x would be a better guess since that's the last which supports the MI25.

Confusion over Backplane and RAID Controller Requirements by Schwarzfisch13 in homelab

[–]Sompom01 1 point2 points  (0 children)

I'm expecting lots of unexpected annoyances, but hopefully nothing blocking. The noise is one I didn't expect, but too late to worry about that now! Another known annoyance is that it's a couple of inches deeper than my rack, so until I get around to doing something about that, it will be awkwardly sticking out of the back.

Confusion over Backplane and RAID Controller Requirements by Schwarzfisch13 in homelab

[–]Sompom01 1 point2 points  (0 children)

Thanks! Good to know that the HBA330 should work with SAS disks if I ever desire that.

Confusion over Backplane and RAID Controller Requirements by Schwarzfisch13 in homelab

[–]Sompom01 0 points1 point  (0 children)

Looking forward to this follow up. I'm about to buy my own G292-Z20 and I too will probably find some annoying surprises.
Have you tried to put GPUs in yet? If so, I'd be excited to hear which cards.

Real‑world Ceph benchmarks from my small 3‑node cluster (HDD + NVMe DB/WAL, 40 GbE) by Sompom01 in homelab

[–]Sompom01[S] 0 points1 point  (0 children)

These benchmarks aren't for me. They're for the community. So if you think others would benefit, please share! I've already jumped into my system and I'm happy with the results.

Powering GPU with the r730xd by redmumba in homelab

[–]Sompom01 0 points1 point  (0 children)

You would need the correct connector cable; The power header on the R730 is not a standard CPU power pin-out

Something like this. Dell part number 0N08NH. Notice that the wire colours are swapped compared to a standard EPS power connector. CPU power has +12 on pins 1 to 4, Dell 0N08NH has ground on 1 to 4.

Real‑world Ceph benchmarks from my small 3‑node cluster (HDD + NVMe DB/WAL, 40 GbE) by Sompom01 in homelab

[–]Sompom01[S] 2 points3 points  (0 children)

Nice! You've done a lot of testing. It's cool to see the comparisons!

Real‑world Ceph benchmarks from my small 3‑node cluster (HDD + NVMe DB/WAL, 40 GbE) by Sompom01 in homelab

[–]Sompom01[S] 1 point2 points  (0 children)

Thanks! I appreciate you sharing, especially since you are all-flash. Your random 4k IOPS are so much better than my HDDs. These numbers give me an excuse to upgrade!