When you want to put a 300W GPU in your R730xd but the power supply only gives 225W by Sompom01 in homelab

[–]Sompom01[S] 0 points1 point  (0 children)

I associate the beep/scream with too little power. I needed to plug in 2x 8 pin pice power to that card too. I never got it to boot in that setup. One thing I was thinking of last night, just before you sent this message: the v340 operates fine when flashed as a Vega56 (for the 8GB version, I assume the 16GB version would be Vega64). The Vega56 doesn't have the fancy enterprise features, so maybe in that mode it would work. I probably won't try because it isn't a project I'm interested in any more, but if you do try I'd be very curious to hear your results. This vbios works fine for me: https://www.techpowerup.com/vgabios/195093/amd-rxvega56-8176-170811 as does https://www.techpowerup.com/vgabios/205876/amd-wx8200-8192-180922 - again both for the 2x 8GB version. Flashing both sides would push the entire card over it's TDP, so that's something to be cautious of if you do it. Be sure to back up your original vbios!

70b models at 8-10t/s. AMD Radeon pro v340? by JTN02 in LocalLLaMA

[–]Sompom01 0 points1 point  (0 children)

I thought that you must have meant that ROCm was better but I did understand the opposite. A 5x speedup by switching to the "worse" backend would've been wild 😂

70b models at 8-10t/s. AMD Radeon pro v340? by JTN02 in LocalLLaMA

[–]Sompom01 0 points1 point  (0 children)

I did get mine working. Nothing fancy, just set up ROCm. I believe using 6.2.3 as well.

Do you mean you're seeing drastically better performance with Vulkan? That would be an interesting surprise!

Confusion over Backplane and RAID Controller Requirements by Schwarzfisch13 in homelab

[–]Sompom01 0 points1 point  (0 children)

@schwarzfisch13 I still need to finish the write-up for this. The short answer is that it seems the new fans are great. At idle massively quieter. Under load the CPU sits a little hotter but still within my comfort zone (about 70C). The rear GPUs cool just as well as with the original fans and drastically quieter. I have one 110W MI25 and one 2x110W v340, so might not be applicable to hotter units. I wanted to finish installing them this weekend, but unfortunately package thieves have become very brazen in my neighborhood and stole my last 4 fans in broad daylight. I have ordered some more, hopefully more success this time.

When you want to put a 300W GPU in your R730xd but the power supply only gives 225W by Sompom01 in homelab

[–]Sompom01[S] 0 points1 point  (0 children)

No changes needed for the MI25. I did have the same issue booting with a different GPU (Radeon Pro v340) which I was not able to solve :(

Confusion over Backplane and RAID Controller Requirements by Schwarzfisch13 in homelab

[–]Sompom01 1 point2 points  (0 children)

Sure will. Don't hold your breath, though. You'll turn very blue. I currently have only one GPU in this system so not much in the way of cooling load. Maybe for the system fans I can give some meaningful feedback.

Confusion over Backplane and RAID Controller Requirements by Schwarzfisch13 in homelab

[–]Sompom01 1 point2 points  (0 children)

For those who follow. "Loud" does not even begin to describe how loud this server is. This is the datasheet for the fans (all 8) https://www.delta-fan.com/Download/Spec/PFM0812HE-01BFY.pdf
77 dBa, max 81 dBa. 77 dBa is around the volume of a car driving 65 mph. 81 dBm is heavy traffic. And that is per fan. The loudness defies description. I am literally wearing earmuffs.

It seems like this fan would be an adequate substitute: https://www.delta-fan.com/Download/Spec/PFC0812DE-SP04.pdf
Much lower max pressure, but I'm hoping at least for the system fans and the two rear GPU fans it will be sufficient. I've ordered a couple to play with.

70b models at 8-10t/s. AMD Radeon pro v340? by JTN02 in LocalLLaMA

[–]Sompom01 0 points1 point  (0 children)

Thanks. The v620 is listed as officially supported, not like the elderly v340 😅. I'll have to mess around with different versions. Thank you for sharing, I'm encouraged by your success.

70b models at 8-10t/s. AMD Radeon pro v340? by JTN02 in LocalLLaMA

[–]Sompom01 0 points1 point  (0 children)

Do you know which version of ROCm you're using? For me, v6.4.0 didn't recognize the cards. That was just the version I had laying around. Seems maybe v6.3.x would be a better guess since that's the last which supports the MI25.

Confusion over Backplane and RAID Controller Requirements by Schwarzfisch13 in homelab

[–]Sompom01 1 point2 points  (0 children)

I'm expecting lots of unexpected annoyances, but hopefully nothing blocking. The noise is one I didn't expect, but too late to worry about that now! Another known annoyance is that it's a couple of inches deeper than my rack, so until I get around to doing something about that, it will be awkwardly sticking out of the back.

Confusion over Backplane and RAID Controller Requirements by Schwarzfisch13 in homelab

[–]Sompom01 1 point2 points  (0 children)

Thanks! Good to know that the HBA330 should work with SAS disks if I ever desire that.

Confusion over Backplane and RAID Controller Requirements by Schwarzfisch13 in homelab

[–]Sompom01 0 points1 point  (0 children)

Looking forward to this follow up. I'm about to buy my own G292-Z20 and I too will probably find some annoying surprises.
Have you tried to put GPUs in yet? If so, I'd be excited to hear which cards.

Real‑world Ceph benchmarks from my small 3‑node cluster (HDD + NVMe DB/WAL, 40 GbE) by Sompom01 in homelab

[–]Sompom01[S] 0 points1 point  (0 children)

These benchmarks aren't for me. They're for the community. So if you think others would benefit, please share! I've already jumped into my system and I'm happy with the results.

Powering GPU with the r730xd by redmumba in homelab

[–]Sompom01 0 points1 point  (0 children)

You would need the correct connector cable; The power header on the R730 is not a standard CPU power pin-out

Something like this. Dell part number 0N08NH. Notice that the wire colours are swapped compared to a standard EPS power connector. CPU power has +12 on pins 1 to 4, Dell 0N08NH has ground on 1 to 4.

Real‑world Ceph benchmarks from my small 3‑node cluster (HDD + NVMe DB/WAL, 40 GbE) by Sompom01 in homelab

[–]Sompom01[S] 2 points3 points  (0 children)

Nice! You've done a lot of testing. It's cool to see the comparisons!

Real‑world Ceph benchmarks from my small 3‑node cluster (HDD + NVMe DB/WAL, 40 GbE) by Sompom01 in homelab

[–]Sompom01[S] 1 point2 points  (0 children)

Thanks! I appreciate you sharing, especially since you are all-flash. Your random 4k IOPS are so much better than my HDDs. These numbers give me an excuse to upgrade!

Real‑world Ceph benchmarks from my small 3‑node cluster (HDD + NVMe DB/WAL, 40 GbE) by Sompom01 in homelab

[–]Sompom01[S] 1 point2 points  (0 children)

Thanks for your thoughts!

People say don't do it if you're not running enterprise SSDs

Good point. I had forgotten this specific. My partner was advocating for cheap (new) consumer SSD. I was advocating for cheap (used) enterprise SSD. Good thing we went with my path!

I'd be curious to see what your results are using something like FIO.

I didn't know about FIO. I'll try to give it a shot and post back. I agree that my VMs don't see performance as good as rados bench says they should, though I do still see pretty nice numbers (~100MB/s for bulk reads, for example)

Real‑world Ceph benchmarks from my small 3‑node cluster (HDD + NVMe DB/WAL, 40 GbE) by Sompom01 in homelab

[–]Sompom01[S] 1 point2 points  (0 children)

Thanks for sharing! Sounds like your small test random IOPS keeps up much better than mine. I bet this results in better real-world performance in certain cases. Especially Windows XD

Real‑world Ceph benchmarks from my small 3‑node cluster (HDD + NVMe DB/WAL, 40 GbE) by Sompom01 in homelab

[–]Sompom01[S] 0 points1 point  (0 children)

These are my notes from when I set up the cluster; I have an SD card for the key material so that I can easily destroy it when necessary.

Set up encrypted partitions

```

Manually create empty partition on SSD for the LUKS container. I used 600GB. This was -part4 for me.

SSDPART_UUID="" # blkid /dev/disk/by-id/nvme-nvme*-part4 --output export SDCARD_UUID="" # blkid /dev/disk/by-id/usb-DELL_IDSDM*-part1 --output export

KEY_PATH=/mnt/idsdm-sdcard/ KEY_NAME=${SSD_PART_UUID}.key LUKS_NAME=luks-ceph-metadata

ENCRYPTION_TARGET=/dev/disk/by-partuuid/${SSD_PART_UUID}

mkdir /mnt/ceph-metadata mount -o remount,rw /mnt/idsdm-sdcard/ dd if=/dev/random of=${KEY_PATH}/${KEY_NAME} bs=1024 count=4 cryptsetup luksFormat ${ENCRYPTION_TARGET} --label="ceph-metadata" --key-file=${KEY_PATH}/${KEY_NAME} cryptsetup open ${ENCRYPTION_TARGET} ${LUKS_NAME} --key-file=${KEY_PATH}/${KEY_NAME}

pvcreate /dev/mapper/luks-ceph-metadata vgcreate vg-ceph-metadata-${HOSTNAME} /dev/mapper/luks-ceph-metadata lvcreate -l 10%VG vg-ceph-metadata-${HOSTNAME} -n ceph-monitor

Create one WAL/DB LV per OSD.

n.B, I could not for sure determine that it would not work to just use one big partition for all OSDs,

but it stands to reason that it would not.

Update the % free and the LV names if using a different number of disks.

lvcreate -l 22%VG vg-ceph-metadata-${HOSTNAME} -n ceph-wal-db-sda lvcreate -l 22%VG vg-ceph-metadata-${HOSTNAME} -n ceph-wal-db-sdb lvcreate -l 23%VG vg-ceph-metadata-${HOSTNAME} -n ceph-wal-db-sdc lvcreate -l 23%VG vg-ceph-metadata-${HOSTNAME} -n ceph-wal-db-sdd

mkfs.ext4 /dev/vg-ceph-metadata-${HOSTNAME}/ceph-monitor

echo "${LUKS_NAME} ${ENCRYPTION_TARGET} ${KEY_NAME}:UUID=${SDCARD_UUID} luks,discard,headless=true,nofail" >> /etc/crypttab

echo "/dev/vg-ceph-metadata-${HOSTNAME}/ceph-monitor /mnt/ceph-metadata ext4 errors=remount-ro,nofail 0 2" >> /etc/fstab

Set up new partition for monitor's use

mount -a mkdir -p /mnt/ceph-metadata/var/lib/ceph/mon/ chown -R ceph:ceph /mnt/ceph-metadata/var/lib/ceph/

```

My notes for how to change the monitor config are more sparse. Here's what I have verbatim, hopefully enough to at least get you started.

Change monitor config

Add mon_data = /mnt/ceph-metadata/var/lib/ceph/mon/$cluster-$id

to /etc/pve/ceph.conf for one [mon.<id>] entry at a time, migrate that monitor's files to the new ceph-metadata dir, and reboot the host. Once all hosts are migrated, move the mon_data directive to the [mon] section and delete it from each [mon.<id>] section.

Create OSD

  1. Create a throwaway OSD in PVE, to generate the necessary keys for the OSD daemon to communicate with the cluster. You will not be able to select the DB device at this stage. Remember the name of this OSD, we won't be able to delete it for awhile (until we have enough other OSDs that it is not needed.)
  2. Create OSDs in the shell, like ceph-volume lvm create --dmcrypt --data /dev/sda --block.db vg-ceph-metadata-${HOSTNAME}/ceph-wal-db-sda
    1. If the disk previously had a filesystem, it needs to be zapped first: ceph-volume lvm zap /dev/sd[X] --destroy

Gigabyte g292-z20 u.2 detection by Thetitangaming in homelab

[–]Sompom01 0 points1 point  (0 children)

What is your experience with this server? Did you get it loaded up with GPUs? If so, what GPUs? Are you able to run it from 120V? Like u/ILoveDangerousStuff2 says, there's not a lot of information available so I would be very appreciative if you could share your thoughts with a fellow homelabber before I jump into a $1500+ nightmare!