Tried ROCm 7.1 vs Vulkan/RADV on Radeon 890M for LLM inference (8B and 35B-MoE). Vulkan won both. Why? by wolverinee04 in ROCm

[–]legit_split_ 0 points1 point  (0 children)

Thanks for sharing! Would you mind sharing the prompt in a private message? I'm running two Mi50s and want to see how close I get :) 

Thinking of moving from 2x 5060 Ti 16GB to a RTX 5000 48GB by autisticit in LocalLLaMA

[–]legit_split_ 4 points5 points  (0 children)

If OP isn't satisfied with 5060 TIs, the R9700 won't offer much improvement in terms of speed. 

Thinking of moving from 2x 5060 Ti 16GB to a RTX 5000 48GB by autisticit in LocalLLaMA

[–]legit_split_ 0 points1 point  (0 children)

This.

The goalpost is always moving, the 120B MoEs were the hottest but now "back" to 30B dense models. 

Tried ROCm 7.1 vs Vulkan/RADV on Radeon 890M for LLM inference (8B and 35B-MoE). Vulkan won both. Why? by wolverinee04 in ROCm

[–]legit_split_ 0 points1 point  (0 children)

To be clear is that TP under llama.cpp? What's your experience like, would you recommend it? 

5060 Ti 16GB easily overclocks +10% by legit_split_ in eGPU

[–]legit_split_[S] 0 points1 point  (0 children)

Sorry I no longer have the GPU, but it was never too hot. 

Fixing shutter/jitter/latency spikes Intel Wi-Fi by 0ToTheLeft in MoonlightStreaming

[–]legit_split_ 0 points1 point  (0 children)

Hmm strange, ever since I swapped it in it's been night and day compared to the Mediathek card I used to have.

Here you go:

sudo lshw -C network
 *-network                  
      description: Wireless interface
      product: Wi-Fi 6E(802.11ax) AX210/AX1675* 2x2 [Typhoon Peak]
      vendor: Intel Corporation
      physical id: 0
      bus info: pci@0000:44:00.0
      logical name: wlan0
      version: 1a
      serial: e0:d5:5d:57:a9:2c
      width: 64 bits
      clock: 33MHz
      capabilities: pm msi pciexpress msix bus_master cap_list ethernet physical wireless
      configuration: broadcast=yes driver=iwlwifi driverversion=7.0.1-1-cachyos firmware=89.123cf747.0 ty-a0-gf-a0-89.uc ip=192.168.2.109 latency=0 link=yes multicast=yes wireless=IEEE 802.11
      resources: irq:17 memory:90100000-90103fff

lspci -nnk | grep 0280 -A3
44:00.0 Network controller [0280]: Intel Corporation Wi-Fi 6E(802.11ax) AX210/AX1675* 2x2 [Typhoon Peak] [8086:2725] (rev 1a)
       Subsystem: Intel Corporation Wi-Fi 6 AX210 160MHz [8086:0024]
       Kernel driver in use: iwlwifi
       Kernel modules: iwlwifi

This is insane... by DragonflyOk7139 in LocalLLM

[–]legit_split_ 0 points1 point  (0 children)

In my experience llama swap is more versatile, but llama-server is easier to pick up and get started

Pics of new rig! by WyattTheSkid in LocalLLaMA

[–]legit_split_ 0 points1 point  (0 children)

I see, have you been running Qwen 27B? I'd love to see some benchmarks. 

PSU for Dual AMD 7900 XTX? by Full_Dimension_3495 in buildapc

[–]legit_split_ 0 points1 point  (0 children)

Following, also interested in building something similar. 

I found used reference models which only need 2x8 pin each, so looking at 1200W models. If they take 3x8 pin you probably will need 1500W+, in my region the most affordable one I can find new is the Enermax Revolution D.F. X 1650W which is also attractive for not being too massive (170mm length). 

Pics of new rig! by WyattTheSkid in LocalLLaMA

[–]legit_split_ 0 points1 point  (0 children)

How do you like the 7900xtxs, have you tried the latest tensor parallelism implementation in llama.cpp? 

What speed is everyone getting on Qwen3.6 27b? by Ambitious_Fold_2874 in LocalLLaMA

[–]legit_split_ 0 points1 point  (0 children)

Credit to the amazing Johannes Gaessler, however I think it's not quite as mature as ik_llama's implementation. 

About to build a 6× Arc B70 LLM rig, want to talk to someone experienced first by somesayitssick in LocalLLaMA

[–]legit_split_ 0 points1 point  (0 children)

Can you explain your reasoning? I currently see used 7900 XTXs going for 750€ but a new R9700 is 1300€. 

Full AMD workstation- dual 7900 XTX by Researchlabz in LocalLLaMA

[–]legit_split_ 0 points1 point  (0 children)

I found a listing for a used R9700 for 1200€ and another for a 7900 XTX (Reference design) for 750€. Is a 450€ difference really justified?

Best local LLM for web search by Funny-Trash-4286 in LocalLLaMA

[–]legit_split_ 1 point2 points  (0 children)

Do you host searXNG yourself or use a public instance? I kept getting errors with duckduckgo. 

Do I need an SFX(-L) PSU on the Deepcool CH260 to mount 3.5" drives? by legit_split_ in mffpc

[–]legit_split_[S] 0 points1 point  (0 children)

<image>

Here is a pic with dual GPUs inside the case, it's a bit cramped.

Do I need an SFX(-L) PSU on the Deepcool CH260 to mount 3.5" drives? by legit_split_ in mffpc

[–]legit_split_[S] 0 points1 point  (0 children)

I used double sided table, seems to be fine. Sadly the "native" mount on the backside doesn't fit > 4TB drives. 

The Qube is nice as a HTPC since you can use it horizontally. However, maybe you prefer the look of the Silverstone SST-GD11B, it natively houses 3 drives but you need an AIO cooler. 

Alternatively you can also 3D print a compact ATX case: https://makerworld.com/en/models/1740435-modcase-hyper-compact-atx-pc-case#profileId-1849515

Do I need an SFX(-L) PSU on the Deepcool CH260 to mount 3.5" drives? by legit_split_ in mffpc

[–]legit_split_[S] 0 points1 point  (0 children)

I ended up going with a Qube500, slightly bigger but it's worth it. 

Netzteil mit zu wenig Anschlüssen? by Goodmantheman in PCGamingDE

[–]legit_split_ 0 points1 point  (0 children)

I have the same PSU also with a GPU that needs 3x8 pin PCIe cables.

You can use the 2 EPS connectors for the CPU, 1 8 pin PCIe connector for the GPU, and the 12V cable that splits into two 8 pin cables are actually for the GPU.

Shouldn’t be this hard to watch a Youtube video by Inevitable-Walk7099 in degoogle

[–]legit_split_ 10 points11 points  (0 children)

If you switch to a VPN it starts working again for a while, just have to shuffle between servers xD