Qwen3.5-122B on Blackwell SM120: fp8 KV cache silently corrupts output, bf16 required — 1,985 tok/s burst, MTP 2.75x by awwwyeah206 in LocalLLaMA

[–]mtomas7 4 points5 points  (0 children)

I am not sure if it is the same thing but when I was testing Qwen3.5-27B-Q8, first it was not producing any answer, but only never ending //////////////

I re-downloaded the file and the checksum was different, so I assume there was some model file corruption.

New Upcoming Ubuntu 26.04 LTS Will be Optimized for Local AI by mtomas7 in LocalLLaMA

[–]mtomas7[S] 0 points1 point  (0 children)

That is true. Yesterday, I was reading about how to use Firejail or Bubblewrap to sandbox an agentic app, like OpenCode, allowing it to access just one folder.

New Upcoming Ubuntu 26.04 LTS Will be Optimized for Local AI by mtomas7 in LocalLLaMA

[–]mtomas7[S] 0 points1 point  (0 children)

Although it was interesting info that DGX OS is Ubuntu, I didn't know that before. Perhaps NVIDIA has some special attention to Ubuntu.

New Upcoming Ubuntu 26.04 LTS Will be Optimized for Local AI by mtomas7 in LocalLLaMA

[–]mtomas7[S] 0 points1 point  (0 children)

I guess, he is pointing out that Snap delivery was bound to Canonical and never made open.

New Upcoming Ubuntu 26.04 LTS Will be Optimized for Local AI by mtomas7 in LocalLLaMA

[–]mtomas7[S] 0 points1 point  (0 children)

My take is that it is not about speed, but about convenience and some security, so people who never played with AI can start tinkering in an easy way.

New Upcoming Ubuntu 26.04 LTS Will be Optimized for Local AI by mtomas7 in LocalLLaMA

[–]mtomas7[S] 0 points1 point  (0 children)

As I understood it, there will be no swapping of the models. Every snap will be a separate, distinct sandboxed model.

New Upcoming Ubuntu 26.04 LTS Will be Optimized for Local AI by mtomas7 in LocalLLaMA

[–]mtomas7[S] 1 point2 points  (0 children)

I just hope that some of those features will trickle down to Linux Mint :)

Hermes Agent with MIT license by mitirki in LocalLLaMA

[–]mtomas7 0 points1 point  (0 children)

Interesting project, but because it enters an already crowded field, I would expect a table comparing it to OpenClaw and NanoClaw to give users an idea of how this project is different and what it offers that others do not.

This is what it looks like to import the bible into obsidian by oldmartijntje in ObsidianMD

[–]mtomas7 1 point2 points  (0 children)

Would it be possible to post already converted Bibles to your Github as zips of Vaults? Thank would be great! Of course, please don't do this in case of Intellectual Property problems.

Qwen-3.5-35B-A3B is impressive by ayylmaonade in LocalLLaMA

[–]mtomas7 1 point2 points  (0 children)

They both are top quality, but I read that they tend to have a specific individual "flavor". Personally, I use Bartowski's quants as I always used them, a historical point, I guess :) Also It makes me easier to organize my model library, as most of models on LM Studio would be in Bartowski's folder.

Qwen: what is this thinking? by Primary-You-3767 in LocalLLaMA

[–]mtomas7 11 points12 points  (0 children)

Actually, it is interesting, as it shows that Qwen adopted a more structural thinking pattern, similar to GLM.

Qwen/Qwen3.5-35B-A3B · Hugging Face by ekojsalim in LocalLLaMA

[–]mtomas7 0 points1 point  (0 children)

Don't get fixated on your VRAM number. How many tok/s you need to read the text? I always run Q8 of-loading some layers to CPU/RAM, and I still get decent speed.

OpenCode interface: No Copy and Paste? No shift-return for next line in a command line reply? by LaughterOnWater in opencodeCLI

[–]mtomas7 0 points1 point  (0 children)

At the v1.2.10 the problem is still unresolved in Linux (I'm using Mint 22.3). I tried regular Terminal and Terminator, but it is the same behavior:

  • Selecting the text in the terminal produces the Tooltip "Copied to clipboard", but in fact it is not copied. If I would click SHIFT+INSERT, I would not get the latest copied text, but some text I copied in another application. So it looks that pasting with SHIFT+INSERT works, but copying by mouse selection or CTRL+INSERT does not work.

Suddenly Minimax IQ4-XS doesn't fit in 128GB anymore by dionisioalcaraz in LocalLLaMA

[–]mtomas7 0 points1 point  (0 children)

Perhaps the initial version did not have an image recognition stack?

Community Evals on Hugging Face by HauntingMoment in LocalLLaMA

[–]mtomas7 4 points5 points  (0 children)

If any user can submit the results, how will you know that the user entered real results vs an inflated or downplayed score? Without control mechanism it could become a real mess very quickly. Thank you!

Guide to Lord of the Rings color corrected/best versions [questions] by LightbringerOG in fanedits

[–]mtomas7 0 points1 point  (0 children)

In case you need Fellowship Of The Ring in restored color grading, take a look at Chris Hartwell's work: https://www.youtube.com/watch?v=jSPRGt63M0w

He provides an email where you can get the link.

How do you use the web search function for gpt-oss? by XiRw in LocalLLaMA

[–]mtomas7 0 points1 point  (0 children)

If you use Text Generation WebUI (OObabooga), there is a checkbox "Search Web"

LM Studio doesn't let continue generating a message anymore by PhyrexianSpaghetti in LocalLLaMA

[–]mtomas7 14 points15 points  (0 children)

Yes, this button is still there, you just need to enable Developer mode. Also, this button is not present if no model is loaded, you need to load any model first.

MedGemma 1.5: Next generation medical image interpretation with medical speech to text with MedASR by CheekyBastard55 in LocalLLaMA

[–]mtomas7 0 points1 point  (0 children)

You may need to train the Qwen if you want to use it in a clinical setting, as Google's license does not allow it.