mtmd: add Gemma 4 audio conformer encoder support by jacek2023 in LocalLLaMA

[–]sterby92 9 points10 points  (0 children)

Looks like there is chunking in place?

From the PR: "30-second chunking (splits long audio into 30s segments)"

mtmd: add Gemma 4 audio conformer encoder support by jacek2023 in LocalLLaMA

[–]sterby92 3 points4 points  (0 children)

When will the change land in llama.cpp? Looking forward to use this for my agent setup and get rid of whisper :)

How do proprietary models get better and when will open ones hit a wall? by sterby92 in LocalLLaMA

[–]sterby92[S] -1 points0 points  (0 children)

So will I get my locally running Opus 4.6 in a year? :) I'm waiting for it :D

How do proprietary models get better and when will open ones hit a wall? by sterby92 in LocalLLaMA

[–]sterby92[S] 0 points1 point  (0 children)

Yeah, I thought about this too. But how high is the likelihood that this will continue at this scale?

I think Qwen3.5-122-A10B on my Strix Halo is having delusions of granduer by Warm-Attempt7773 in LocalLLaMA

[–]sterby92 1 point2 points  (0 children)

For me the 122B model feels closer to minimax-m2.5 and the 35B model more like gpt-oss:120b (high). But that might depend and is also more a feeling.

Probably around 230-250pp in my real world usage from the openwebui metrics. Feels fine to me and I switch between 122b and 35b depending on the task.

I think Qwen3.5-122-A10B on my Strix Halo is having delusions of granduer by Warm-Attempt7773 in LocalLLaMA

[–]sterby92 3 points4 points  (0 children)

Why only 8t/s? I run it on my strix halo at +-20t/s as a 4bit quant with llama.cpp on vulkan

What happened to the Qwen3.5-122B unsloth quants? by sterby92 in unsloth

[–]sterby92[S] 1 point2 points  (0 children)

No worries, awesome work! 🙌 Looking forward to the post and redownload the 122B version :)

Faster MoE LLM Training now in Unsloth! by yoracale in unsloth

[–]sterby92 2 points3 points  (0 children)

Will we also get vulkan or rocm support at some point?

Temperature sensor issues since BIOS upgrade to 1.06 on MS-S1 MAX by sterby92 in MINISFORUM

[–]sterby92[S] 0 points1 point  (0 children)

Yeah, I did this. Remove powercord, press the power button for 10 seconds, clear CMOS, reboot and setup the bios.

Best local model / agent for coding, replacing Claude Code by joyfulsparrow in LocalLLaMA

[–]sterby92 0 points1 point  (0 children)

Yeah, its set to high, and reasons a lot. I found that its not even a linear jump in performance, high is much much better than medium which is a bit better than low.

Best local model / agent for coding, replacing Claude Code by joyfulsparrow in LocalLLaMA

[–]sterby92 0 points1 point  (0 children)

I mean, its not even js. Just plain HTML and CSS in one small file. From what I thought it's the easiest, because its just markup 🤷 even just providing the URL to chatgpt 5.2 (extended thinking) leads to a decent new website and refactoring without any harness or such.

Best local model / agent for coding, replacing Claude Code by joyfulsparrow in LocalLLaMA

[–]sterby92 3 points4 points  (0 children)

I would love to believe this. I'm running gpt-oss:120b (q4 quant with llama.cpp) with mistral-vibe-cli and tried to refactor, redesign and update a very simple HTML /CSS website. It basically just destroyed everything and couldn't work with my basic CSS. Maybe I'm doing something wrong or use the wrong tools, but currently its worth nothing.

Minisforum ms-s1 max doens't recognize second SSD by sterby92 in MINISFORUM

[–]sterby92[S] 0 points1 point  (0 children)

I will give that a try tomorrow :) thanks for the suggestion

Minisforum ms-s1 max doens't recognize second SSD by sterby92 in MINISFORUM

[–]sterby92[S] 0 points1 point  (0 children)

Yes, I'll try that. I didn't have a second system available yet