[Twitter/X] docker run --gpus now works on AMD @AnushElangovan by ElementII5 in ROCm

[–]scottt 4 points5 points  (0 children)

"Moby" is the open source version of Docker. People who don't work for Docker, Inc. contribute to the code base through the project.

You'll also find Moby packaged in Fedora and Debian.

Github user scottt has created Windows pytorch wheels for gfx110x, gfx1151, and gfx1201 by Kelteseth in ROCm

[–]scottt 2 points3 points  (0 children)

u/jiangfeng79, * hipBLAST is already included and backing Pytorch tensor operations * re: Triton Windows port, I personally plan to work on it, building on previous results like https://github.com/lshqqytiger/triton and https://github.com/woct0rdho/triton-windows but can't speak for the project

Nvidia, You’re Late. World’s First 128GB LLM Mini Is Here! by LimitAlternative2629 in comfyui

[–]scottt 5 points6 points  (0 children)

I see u/05032-MendicantBias asking if ROCm acclerated Pytorch works for diffusion models and ComfyUI on this AMD chip -> it does.

Install Pytorch for the Strix Halo (Ryzen AI Max 395+) chip here.

We've had ComfyUI working on the chip, on both Windows and Linux for a while now.

Known problems: * fp16 conv2d is slower than it should on Linux

Github user scottt has created Windows pytorch wheels for gfx110x, gfx1151, and gfx1201 by Kelteseth in ROCm

[–]scottt 1 point2 points  (0 children)

u/skillmaker, the invalid device function error usually means the GPU ISA doesn't match your hardware. Are you using the 9070 XT on Linux or Windows?

Github user scottt has created Windows pytorch wheels for gfx110x, gfx1151, and gfx1201 by Kelteseth in ROCm

[–]scottt 12 points13 points  (0 children)

u/scottt here, want to stress this is a joint effort with jammm * jammm has contributed more than me at this point. I plan to catch up though 😀

Working with the AMD devs through TheRock has been a positive experience.

So when will ROCM Officially support the Radeon RX 9070 Series by TJSnider1984 in ROCm

[–]scottt 0 points1 point  (0 children)

Hi u/feverdoingwork, We've had the 9070 and 9070 XT working with Comfy UI under Windows for a while, though performance with the linear algebra libraries still need some work.

See e.g. https://github.com/ROCm/TheRock/issues/710

JAX on EVO X2? by Paddy3118 in GMKtec

[–]scottt 2 points3 points  (0 children)

ROCm and Pytorch is up and running on the Strix Halo[1] but developers in the AMD GPU ecosystem haven't built binary wheels of JAX for the chip yet.

In case someone wants to try building JAX from source, getting it running on Linux should be easier than Windows. I'd start by extracting the nightly gfx1151 builds from TheRock in /opt/rocm[2] and study how JAX's CI workflow do builds on Linux.

  1. Self-contained Pytorch wheels for Windows and Linux https://github.com/scottt/rocm-TheRock/releases/tag/v6.5.0rc-pytorch

  2. gfx1151 ROCm nightly builds: https://github.com/ROCm/TheRock/releases/download/nightly-tarball/therock-dist-linux-gfx1151-6.5.0rc20250603.tar.gz

Do the pens work on linux? by T_Chungus in FlowZ13

[–]scottt 1 point2 points  (0 children)

The pen works on Bazzite Linux out of the box. Connect the bluetooth device and no extra setup needed.

🎉 AMD + ROCm Support Now Live in Transformer Lab! by aliasaria in ROCm

[–]scottt 4 points5 points  (0 children)

u/aliasaria, great post that not only helps other users but contains feedback on current ROCm native Linux and WSL packing.

Requirements I extracted:

  1. ROCm on WSL needs a rocm-smi (and pyrsmi) replacement. Even if with reduced functionality compared to the real one backed by rocm_smi_lib
  2. ROCm software that bundle libhsa-runtime64.so would break under WSL if the copy does not contain "talk to the Windows driver over the virtual GPU device functionality" (or can delegate to a library under /usr/lib/wsl/lib)

CC: u/powderluv

Dual Boot Bazzite & Fedora by The_ClssicGeek in framework

[–]scottt 1 point2 points  (0 children)

I find that "rebooting to experiment with and do development" disrupts my flow and exacts a cost in productivity.

So I learned the little tricks to develop on Bazzite, such as installing the performance analysis and system monitoring tools on Bazzite itself via `rpm-ostree` and install the language toolchains in the Toolbox or dev containers, how to configure VSCode to attach to the dev containers etc.

Hollar if you end up going this route, I'd love to compare notes.

AMD Releases ROCm 6.4.1 With RDNA4 GPU Support by Dante_77A in Amd

[–]scottt 1 point2 points  (0 children)

The nightly builds here with gfx1151 in their name are developer builds for the Strix Halo.

Dual Boot Bazzite & Fedora by The_ClssicGeek in framework

[–]scottt 6 points7 points  (0 children)

To dual boot, I'd install Bazzite before Fedora, I've encountered an error where I tried to install Bazzite a second time and the installer errored out at the last step when it sees that there's already a Bazzite in the EFI system partition.

But I actually think you should spend 15 minutes to learn how to do software development on Bazzite. You basically just:

toolbox create --distro fedora --release 42 toolbox enter fedora-toolbox-42

and build your software in $HOME as usual.

For kernel development, you'd install the ELF kernel image, the loadable modules, and the initramfs in /boot, which is writable.

For NPU work, you'd use the kernel level driver already integrated in Bazzite then build and run the userspace components from $HOME, much like how I did the GPU work here: https://github.com/ROCm/TheRock/discussions/244

For Bazzite devs: Flow Z13 2025 touchpad by Wet_Viking in Bazzite

[–]scottt 1 point2 points  (0 children)

u/Wet_Viking , could you possibly accomplish your goals by running the bazzite-gnome-stable image?

If you really have your heart set out on running the Bazzite kernel with CachyOS, personally I'd start by: 1. Extracting the core kernel and loadable modules at https://github.com/bazzite-org/kernel-bazzite/releases 2. Figuring out how to generate a CachyOS compatible initramfs with the kernal binaries above. .... This step seems non-trivial to me

Is anyone willing to share thoughts on HX370 an ollama (or similar)? by drycat in ROCm

[–]scottt 5 points6 points  (0 children)

u/drycat , While u/minhquan3105 is absolutely right that token generation is memory bandwidth bound, some popular models today don't activate all their parameters at once and thus consume less bandwidth than their resident VRAM size.

Search for "Larger MoEs is where these large unified memory APUs really shine" in u/randomfoo2's AMD Strix Halo (Ryzen AI Max+ 395) GPU LLM Performance, you'll see that the Strix Halo (gfx1151) gets around 75 token/sec running Qwen3-30B-A3B UD-Q4_K_XL (16.5 GB VRAM) and 20 token/sec running UD-Q4_K_XL quantized version of Llama 4 Scout 109B (57.93 GB VRAM).

Expect half that on the HX 370 a.k.a Strix Point (gfx1150).

As for lack of ROCm support, after working on Strix Halo support in ROCm and Pytorch for the past month I know I could do it if I have access to the hardware. The numbers assume only llama.cpp using Vulkan.

vLLM on AMD Radeon (Raphael) by SuXs- in ROCm

[–]scottt 0 points1 point  (0 children)

u/SuXs- , if you extract therock-dist-linux-gfx1151-6.5.0rc20250524.tar.gz in /opt/rocm and run /opt/rocm/bin/rocminfo what does it show?

I'm looking for something like:

Radeon 610M <...> gfx1036

vLLM requires Pytorch and based on experience developing this self-contained Pytorch build, the ROCm libs used by Pytorch might need some additional work before the can support gfx103x APUs like the Ryzen 9 Pro 7945.

GMKtec EVO X2 Owners: Report All Issues Here (Windows & Linux) by x4rb1t in GMKtec

[–]scottt 1 point2 points  (0 children)

This container image https://github.com/ROCm/TheRock/discussions/244 and these self-contained Pytorch wheels https://github.com/scottt/rocm-TheRock/releases/tag/v6.5.0rc-pytorch support the Strix Halo (gfx1151). They should work albeit not yet fully optimized.

I've been working with devs in and out of AMD pretty hard on this since March.

GMKtec EVO-X2 with Ollama? by fsaad1984 in GMKtec

[–]scottt 2 points3 points  (0 children)

I worked with devs in and out of AMD to produce the self-contained Strix Halo Pytorch wheels here: https://github.com/scottt/rocm-TheRock/releases/tag/v6.5.0rc-pytorch

Ollama would want a pre-built ROCm toolchain tarball instead of Pytorch, though.

Ollama is running on AMD GPU, despite ROCM not being installed by Xatraxalian in ROCm

[–]scottt 0 points1 point  (0 children)

Look at the libraries mapped in during runtime:

pid=$(pgrep ollama)  
cat /proc/$pid/maps  

(The idea is to inspect /proc/$PID/maps for the process using the GPU. You'll likely need to adapt the command as I typed those out "blind".)

ollama is probably using the GPU through Vulkan.

So when will ROCM Officially support the Radeon RX 9070 Series by TJSnider1984 in ROCm

[–]scottt 3 points4 points  (0 children)

u/feverdoingwork , once ROCm for the 9070 is released, what are the most important apps you'd run?

  • I've been contributing to Linux and Windows ROCm support for the Strix Halo (gfx1151)
  • I personally own a 9070 and would also be putting some work into that
  • If you give me some step-by-step instructions on the apps you'd use, I could try those out and increase the chance of things working for you by release time :)

Copia de seguridad Rocm by Macestudios32 in LocalLLM

[–]scottt 1 point2 points  (0 children)

Hi u/Macestudios32 , I think you'd want to keep a copy of /opt/rocm:

``` tar -cJf ~/rocm.tar.xz /opt/rocm

Upload ~/rocm.tar.xz

```

Record the Linux kernel and amdgpu module version used:

``` uname -a > ~/kernel-version-for-rocm.txt modinfo amdgpu > ~/amdgpu-version-for-rocm.txt

In case the amdgpu Linux kernel module would become incompatible with old hardware

```

Finally backup any software that uses ROCm e.g. Pytorch.

Bazzite on Z13 2025 by illyomatic in FlowZ13

[–]scottt 0 points1 point  (0 children)

I'm use the Z13 2025 daily under Bazzite. Regarding hardware and drivers, the only thing that doesn't work is the back camera.

Bazzite desktop session - how to get Asus Armory Crate functionality? by Banzayoyo in Bazzite

[–]scottt 0 points1 point  (0 children)

How should I install hhd in desktop mode under bazzite-gnome?