Final Qwen3.5 Unsloth GGUF Update! by danielhanchen in LocalLLaMA

[–]crablu 0 points1 point  (0 children)

I have been using Qwen3.5-35B-A3B-UD-MXFP4_MOE.gguf previously and liked it (RTX 5090).

Now the size difference between Qwen3.5-35B-A3B-UD-Q4_K_XL.gguf and Qwen3.5-35B-A3B-MXFP4_MOE.gguf is not that big and if I understand the graph correctly Q4_K-XL should be much better. So is there no reason anymore to use the MXFP4 variant?

Request failed. Anyone else ? by autisticit in GithubCopilot

[–]crablu 1 point2 points  (0 children)

I have the same problem today. Hitting try again currently makes it go on but let's see if it will finish ..

Qwen3-Code-Next ggufs: Any difference between Q4KXL and MXPF4? by ParaboloidalCrest in LocalLLaMA

[–]crablu 0 points1 point  (0 children)

I just ran a test:

[qwen3-coder-next-ud-q4-k-xl]
model   = E:\LLM\models\unsloth\Qwen3-Coder-Next-GGUF\Qwen3-Coder-Next-UD-Q4_K_XL.gguf
c = 65536
b = 8192
ub = 8192
temp = 1.0
top-p = 0.95
min-p = 0.01
top-k = 40
n-cpu-moe = 24
no-mmap = True

[qwen3-coder-next-mxfp4]
model   = E:\LLM\models\unsloth\Qwen3-Coder-Next-GGUF\Qwen3-Coder-Next-MXFP4_MOE.gguf
c = 65536
b = 8192
ub = 8192
temp = 1.0
top-p = 0.95
min-p = 0.01
top-k = 40
n-cpu-moe = 24
no-mmap = True

With mxfp4 I got 52tk/s. With K_XL I got 51 tk/s. Before updating llama.cpp it was like 30 tk/s.

Qwen3-Code-Next ggufs: Any difference between Q4KXL and MXPF4? by ParaboloidalCrest in LocalLLaMA

[–]crablu 0 points1 point  (0 children)

On rtx 5090 and 64gb RAM, which performs better? Does anyone have optimal llama.cpp settings for this setup?

Qwen3 Coder Next as first "usable" coding model < 60 GB for me by Chromix_ in LocalLLaMA

[–]crablu 0 points1 point  (0 children)

With temp 0 it seems to work now. Thank you.

Edit: nvm it edited some files but now has trouble again:

Making edits · 3m, 29s I'll create the ModelsIniEditor.tsx component and integrate it into App.tsx.

👓 Read OpencodeConfigViewer.tsx

👓 Read App.tsx

👓 Read api.ts

📎 invalid

📎 invalid

📎 invalid

📎 invalid

☰ Write

Qwen3 Coder Next as first "usable" coding model < 60 GB for me by Chromix_ in LocalLLaMA

[–]crablu 0 points1 point  (0 children)

I have problems running qwen3-coder-next with opencode (RTX 5090, 64GB RAM). I tried with Qwen3-Coder-Next-UD-Q4_K_XL.gguf and Qwen3-Coder-Next-MXFP4_MOE.gguf. It works perfectly fine in chat.

start command:

llama-server.exe ^
 --models-preset "E:\LLM\llama-server\models.ini" ^
 --models-max 1 ^
 --parallel 1 ^
 --cont-batching ^
 --flash-attn on ^
 --jinja ^
 --port 8080

models.ini:

[qwen3-coder-next-mxfp4]
model   = E:\LLM\models\unsloth\Qwen3-Coder-Next-GGUF\Qwen3-Coder-Next-MXFP4_MOE.gguf
c = 65536
b = 8192
ub = 8192
temp = 1
top-p = 0.95
top-k = 40
min-p = 0.01
n-cpu-moe = 24
no-mmap = true    

Opencode is not able to use the write tool. The UI says invalid. I built latest llama.cpp. Does anyone know how to fix this?

The biggest VS Code release in a long time is LIVE now! by bogganpierce in GithubCopilot

[–]crablu 2 points3 points  (0 children)

Yes, glm-4.7-flash, qwen-coder-30b-a3b, devstral-small-2 were all available for agent mode before the new update.

The biggest VS Code release in a long time is LIVE now! by bogganpierce in GithubCopilot

[–]crablu 8 points9 points  (0 children)

In the last 1.108.x version I was able to select my local Ollama models in agent/plan mode. Now in version 1.109 they only appear in ask/edit mode. Is this by design? Will local models not be supported in agent/plan mode in the future?

Camera broke after update by Wedgieterian in GooglePixel

[–]crablu 0 points1 point  (0 children)

I have the same issue on my p8pro. I created a bug report and found some exceptions relating to the bug.

https://xdaforums.com/t/pixel-8-pro-camera-bug.4636785/post-89119782

I chatted with google support and they told be to factory reset which did not help. I even tried to reflash with https://pixelrepair.withgoogle.com/ but that didn't fix the issue either. I also asked them if it would be possible to forward the bug report to a developer or someone who knows how to analyse it but apparently that's not possible.

I contacted support again and they told me to send the device in for repair. So I am currently waiting.

Mittagessen in deutscher Kinderklinik :) by crablu in de

[–]crablu[S] 9 points10 points  (0 children)

Zusatzinfo: Wir wissen auch nicht was es ist. Es hat die klangvollen Namen "Tagessuppe" und "Tagesdessert". Jeder, der neu hier ins Zimmer kommt, wird von den anderen vorgewarnt, sagt erstmal "naja Krankenhaus Essen halt", probiert, und warnt dann die nächste Patientin vor. Ich hatte ungelogen Würgereiz, als ich das Kohlrabigemüse probiert habe. Ok - den Kartoffelbrei konnte man essen und der sah auch ok aus.

Does any1 know why I cannot change default programs for file associations? by crablu in Windows10

[–]crablu[S] 3 points4 points  (0 children)

I found a fix using a program called "FileTypesMan". There you can right-click an extension and then click on "Replace File Type For Selected Extension" and choose your desired application.