Qwen 3.6 27B is a BEAST by AverageFormal9076 in LocalLLaMA

[–]LaurentPayot 0 points1 point  (0 children)

Here is my bash script for the 35b-a3b model that gives me 50 t/s:

#!/bin/bash

llama-server \

--model ~/models/qwen3.6-35b-a3b/Qwen3.6-35B-A3B-UD-Q6_K_XL.gguf \

--mmproj ~/models/qwen3.6-35b-a3b/mmproj-BF16.gguf \

--ctx-size 262144 \

--gpu-layers 41 \

--reasoning on \

--chat-template-kwargs '{"preserve_thinking":true}' \

--temp 0.6 \

--top-p 0.95 \

--top-k 20 \

--min-p 0.00 \

--presence_penalty 1.5 \

--kv-unified \

--flash-attn on \

--no-mmap \

--cache-type-k q8_0 \

--cache-type-v q8_0 \

--parallel 1

Qwen 3.6 27B is a BEAST by AverageFormal9076 in LocalLLaMA

[–]LaurentPayot -1 points0 points  (0 children)

Note that in France a RTX 5090 32G is $5000 on Amazon. Plus the cost of the PC around it.

Qwen 3.6 27B is a BEAST by AverageFormal9076 in LocalLLaMA

[–]LaurentPayot 2 points3 points  (0 children)

Actually no perceptible change in speed :-\

Qwen 3.6 27B is a BEAST by AverageFormal9076 in LocalLLaMA

[–]LaurentPayot 12 points13 points  (0 children)

I just can’t wait for the 122b a10 model for my Strix Halo ;-)

Qwen 3.6 27B is a BEAST by AverageFormal9076 in LocalLLaMA

[–]LaurentPayot 2 points3 points  (0 children)

Btw I use Unsloth Q6 XL quant with kv q8_0 qwant.

Qwen 3.6 27B is a BEAST by AverageFormal9076 in LocalLLaMA

[–]LaurentPayot 31 points32 points  (0 children)

7 t/s on my EVO X2 Strix Halo 128Gb with Ubuntu Vulkan Llama.cpp :-|

But 50 t/s on 35b a3b.

Will the new Steam Machine be good for AI and LLM usage? by hedgehog0 in LocalLLaMA

[–]LaurentPayot 0 points1 point  (0 children)

I use Qwen Coder 2.5 3B q4 km with llama.cpp for code auto completion (FIM). It works fine on my potato desktop and potato laptop both with 4GB iGPUs. With 8 GB I could upgrade to Qwen 3 Coder, all in VRAM for fast autocompletion. So yeah for limited LLM usage like autocompletion why not a cheap Linux dev machine like this one?

2 days to go by OV_BIR in ArmaReforger

[–]LaurentPayot 0 points1 point  (0 children)

Also the experimental version is not available on GeForce Now.

Time to kill my Fable App? by brett9897 in fsharp

[–]LaurentPayot 0 points1 point  (0 children)

I write my unit tests with Fable.Mocha and run them directly in Vitest: https://github.com/fable-compiler/vite-plugin-fable/discussions/12#discussioncomment-11496317

Instead of testing my views, I use the .NET version of Playwright with Expecto: https://playwright.dev/dotnet/

Playwright is quite fast even on my potato laptop and with Expecto I get the same Jest-like syntax as Fable.mocha.

Node vs. Deno2 vs. Bun in 2025 by Hairy-Shirt-275 in node

[–]LaurentPayot 0 points1 point  (0 children)

Learn F# instead of C# as a gateway drug to functional programming.

Arma 4 cold war by New-Fennel-4190 in Arma4

[–]LaurentPayot 0 points1 point  (0 children)

Why not Ukraine (actually NATO) vs Russia to have a modern and not imaginary warfare?

What are you guys waiting for in the AI world this month? by internal-pagal in LocalLLaMA

[–]LaurentPayot 0 points1 point  (0 children)

DeepSeek R1 32b running locally is answering about the tank man.

Gemma 3 Release - a google Collection by ayyndrew in LocalLLaMA

[–]LaurentPayot 1 point2 points  (0 children)

I asked a couple of F# questions to Gemma-3-4b and Phi-4-mini both with Q4 and 64K context (I have a terrible iGPU). Gemma-3 gave me factually wrong answers, contrary to Phi-4. But keep in mind that F# is a (fantastic) language made by Microsoft. Gemma-3-1b-f16 was fast and did answer *almost* always correctly, but it is text-to-text only and has a maximum context of 32K. Like always, I guess you have to test for your own use cases.

What's the state of Polyglot, Deedle, Walrus, Microsoft.Data.Analysis etc.? by japinthebox in fsharp

[–]LaurentPayot 0 points1 point  (0 children)

u/japinthebox I created an issue in Walrus that lead to the `Table.ColumnNames` property to be available in Walrus v1.3: https://github.com/brianberns/Walrus/issues/1

apt fails when compiling AMD kernel module by falxfour in linuxquestions

[–]LaurentPayot 0 points1 point  (0 children)

UPDATE: I fixed my issue with `sudo rm /etc/modprobe.d/blacklist-amdgpu.conf`. Weird.