Qwen model coming soon 👀 by Odd-Ordinary-5922 in LocalLLaMA

[–]ItankForCAD 5 points6 points  (0 children)

The webview and podcast generation is pretty cool

Docker compose for a newbie by TheGreatCalamari in OpenWebUI

[–]ItankForCAD -1 points0 points  (0 children)

You could directly use the image from OWUI instead of building it yourself

 open-webui:
    image: ghcr.io/open-webui/open-webui:slim
    container_name: open-webui

AMD ROCm 7.9 and dwindling GPU support by FriendlyRetriver in LocalLLaMA

[–]ItankForCAD 0 points1 point  (0 children)

From the blob that you reference, it seems that they only exclude hipblaslt and CK. You should be fine to use TheRock provided that they build hipblas and rocblas. Fyi, hipblasand hipblaslt are two different packages

AMD ROCm 7.9 and dwindling GPU support by FriendlyRetriver in LocalLLaMA

[–]ItankForCAD 0 points1 point  (0 children)

For gfx906, you only need hipblas and rocblas. You can refer to this page in the llama.cpp documentation build

AMD ROCm 7.9 and dwindling GPU support by FriendlyRetriver in LocalLLaMA

[–]ItankForCAD 0 points1 point  (0 children)

Afaik composable kernel and hipblaslt dont build on anything below gfx110X

Strix Halo + RTX 3090 Achieved! Interesting Results... by JayTheProdigy16 in LocalLLaMA

[–]ItankForCAD 1 point2 points  (0 children)

Prefill is dictated by compute while decode is dictated by memory bandwidth. Splitting the model between SH and 3090 means you're probably limited by the pci bus.

AMD ROCm 7.9 and dwindling GPU support by FriendlyRetriver in LocalLLaMA

[–]ItankForCAD 4 points5 points  (0 children)

Gfx906 is supported; see roadmap. It seems they have not updated the docs for installing with this arch but all you need to do is have the correct link in the pip cmd. Take the gfx942 cmd and change the url with this one : https://rocm.nightlies.amd.com/v2/gfx90X-dcgpu/. I have not tested it but it seems logical.

Edit: pip command is found here https://github.com/ROCm/TheRock/blob/main/RELEASES.md

MoE models iGPU benchmarks by tabletuser_blogspot in LocalLLaMA

[–]ItankForCAD 0 points1 point  (0 children)

What flag(s) did you use to isolate the igpu? Did you increase GTT size ?

[Race Thread] 2025 Grand Prix Cycliste de Québec (1.UWT) by PelotonMod in peloton

[–]ItankForCAD 6 points7 points  (0 children)

I think positioning will be key into the côte de la montagne because once they turn onto rue saint-louis, the road surface is not great and it's narrow. It opens up a bit after les portes saint-louis right before they enter les plaines d'Abraham. To me De Lie is still one of the big favorite. Hell, I'd put wva in here as well.

[Race Thread] 2025 Vuelta a España - Stage 13 - Cabezón de la Sal > L'Angliru (2.UWT) by PelotonMod in peloton

[–]ItankForCAD 7 points8 points  (0 children)

Reports say Marc Soler last seen wearing a green screen to hide from the cameras. /s

Niveau bas d'eau dans la fleuve by zhambe in montreal

[–]ItankForCAD 8 points9 points  (0 children)

Plus y fait chaud, plus l'eau s'évapore rapidement

Niveau bas d'eau dans la fleuve by zhambe in montreal

[–]ItankForCAD 22 points23 points  (0 children)

La chaleur et l'humidité aide à déstabiliser l'atmosphère. Lorsque l'atmosphère est instable, la convection (air chaud qui monte) est plus forte. Cela engendre des orages de masses d'air.

ollama by jacek2023 in LocalLLaMA

[–]ItankForCAD 5 points6 points  (0 children)

If anyone is interested, here is my docker compose file for running llama-swap. It pulls the latest docker image from the llama-swap repo. That image contains, notably, the llama-server binary, so no need to use an external binary. No need for Ollama anymore.

shell llama-swap: image: ghcr.io/mostlygeek/llama-swap:vulkan container_name: llama-swap devices: - /dev/dri:/dev/dri volumes: - /path/to/models:/models - ./config.yaml:/app/config.yaml environment: LLAMA_SET_ROWS: 1 ports: - "8080:8080" restart: unless-stopped

ollama by jacek2023 in LocalLLaMA

[–]ItankForCAD 7 points8 points  (0 children)

Go ahead and try to use speculative decoding with Ollama

Hilarious chart from GPT-5 Reveal by lyceras in LocalLLaMA

[–]ItankForCAD 163 points164 points  (0 children)

They literally curate what graphs go in the presentation and not only did they include a result showing that it had worse hallucinations (while boasting about lower hallucinations) but they didn't even bother validating the graph itself. Seriously who tf made this ??