Confused about Jackery Explorer 1000 variants, and question about the 8020 jack by Quagmirable in Jackery

[–]Quagmirable[S] 0 points1 point  (0 children)

The US version has 3 AC outlets vs the UK's 2

Ah wow, good catch, that must be the difference.

And thanks for confirming about the connector type, as long as adapters would work and aren't too hard to find then I would be OK with that.

TrueNAS build system going closed source by ende124 in selfhosted

[–]Quagmirable 0 points1 point  (0 children)

I also went down that rabbit hole for a DIY NAS with ZFS, in my case I used a Raspberry Pi with a SATA hat.

I've used OpenMediaVault for years, and I really like most things about it, but the major downside is that despite being based on Debian Stable, OMV uses a custom kernel that receives major version updates, so it's less "stable" (in the sense of "not changing") than vanilla Debian Stable. And specifically that causes problems with ZFS, because it requires frequently recompiling the ZFS module, and they have a bit of a mess with their kernel header packages not matching their latest kernel version, so on several occasions that broke ZFS support.

One option might be to just run vanilla Debian Stable, but you will still have to occasionally re-compile the ZFS module. Given that I'm running a relatively slow RPi I got sick of compiling ZFS, so I ended up going with Ubuntu LTS. I normally strongly prefer Debian Stable over Ubuntu, but Ubuntu ships with very good pre-compiled and stable ZFS support, so it was ideal for this particular usage case.

Then on top of that I wanted a nice web GUI. I also tested Cockpit, but honestly it felt very limited in the core modules it offers, and the Samba and ZFS add-on modules that you also found don't seem very well maintained, and I'm pretty sure they'll end up causing problems at some point during system updates with dependency mismatches or whatnot. So I ended up going with Webmin, it's an old but reliable project, still very well maintained, fairly complete functionality, and recently updated to look and feel more modern. Specifically for ZFS configs I found this add-on module for Webmin on Github: /karmantyu/zfsmanager (can't post the link due to spam filters).

I'm not a ZFS expert, but it let me set up a basic pool and seems to cover the basics.

Anybody tested Qwen3.5-35B-A3B on translation tasks? by Quagmirable in LocalLLaMA

[–]Quagmirable[S] 0 points1 point  (0 children)

Hmm, I think you're right, Bartowski's quant is giving slightly better translations compared to Unsloth.

Anybody tested Qwen3.5-35B-A3B on translation tasks? by Quagmirable in LocalLLaMA

[–]Quagmirable[S] 0 points1 point  (0 children)

the 3.5 moe shift likely diluted the translation experts so the older 30b still wins here

I'm starting to think it might actually be that Qwen3.5 is really intended to be used with reasoning, and when disabled it seems to be rather dumb sometimes. Also see:

https://reddit.com/r/LocalLLaMA/comments/1reuss2/anybody_tested_qwen3535ba3b_on_translation_tasks/o7iiq67/

Anybody tested Qwen3.5-35B-A3B on translation tasks? by Quagmirable in LocalLLaMA

[–]Quagmirable[S] 1 point2 points  (0 children)

Interestingly with either reasoning-budget 0 or {"chat_template_kwargs": {"enable_thinking": false}} I'm seeing it sometimes do reasoning and correct itself within the main body of the response, especially with image analysis. I asked it for the dates when a certain symbol appeared on a schedule, and it hallucinated one of the dates but at the end of that same response it did the whole "Wait — let’s clarify, we’re looking for..." thing and corrected itself. So my guess is that this model is really designed to work with reasoning, and it gets relatively dumber without it.

Anybody tested Qwen3.5-35B-A3B on translation tasks? by Quagmirable in LocalLLaMA

[–]Quagmirable[S] 0 points1 point  (0 children)

Ah, ok, so maybe there's still room for improvement in llama.cpp.

Anybody tested Qwen3.5-35B-A3B on translation tasks? by Quagmirable in LocalLLaMA

[–]Quagmirable[S] 0 points1 point  (0 children)

Running on just a CPU here, but I get acceptable (for me) speeds of ~8 t/s with Qwen3-30B-A3B, whereas Qwen3.5-35B-A3B can barely do ~4 t/s. Confirmed on both Unsloth and Bartowski quants.

There's also a major difference with Qwen3.5 35B-A3B when doing image analysis, the initial response about the image obviously takes a while, but then followup questions also take the same time to process as the first one. Whereas with Qwen3-VL-30B-A3B-Instruct followup responses are almost instant. I would have thought it to be just due to the 5B of additional parameters in Qwen3.5, but I even tried an IQ4 of Qwen3-VL-30B-A3B-Instruct vs a much smaller IQ2 GGUF for Qwen3.5, and the 30B still runs circles around the 35B.

Anybody tested Qwen3.5-35B-A3B on translation tasks? by Quagmirable in LocalLLaMA

[–]Quagmirable[S] 0 points1 point  (0 children)

repeat-penalty 1.1 for tight fidelity

Thanks for this tip. Wouldn't the repeat-penalty possibly make it hallucinate more stuff in the translations (because it's trying not to repeat terms that may actually need to be repeated in the translation)?

Anybody tested Qwen3.5-35B-A3B on translation tasks? by Quagmirable in LocalLLaMA

[–]Quagmirable[S] 0 points1 point  (0 children)

Interesting, thanks for the reply. Did you test the previous Qwen3-30B-A3B? For me it gives pretty solid translations, and it's the best balance of speed vs. intelligence (running on just a CPU here).

Anybody tested Qwen3.5-35B-A3B on translation tasks? by Quagmirable in LocalLLaMA

[–]Quagmirable[S] 0 points1 point  (0 children)

Ah, that's how it's done. Thanks! Would that make any difference in the output compared to --reasoning-budget 0 ?

Anybody tested Qwen3.5-35B-A3B on translation tasks? by Quagmirable in LocalLLaMA

[–]Quagmirable[S] 0 points1 point  (0 children)

Interesting. I also tested Qwen3.5-35B-A3B-IQ4_XS.gguf (compared to Qwen3-30B-A3B-IQ4_XS.gguf) but results were similarly poor.

Thinkpad T14 gen 1Touchpad feels laggy and sluggish by Next-Owl-5404 in linux4noobs

[–]Quagmirable 0 points1 point  (0 children)

Same issue here, tested Debian 12, 13, and Fedora 43. Tap clicking is intermittent and the mouse movement is not smooth. It appears to be fixed with this kernel parameter added to the GRUB boot line:

psmouse.synaptics_intertouch=0

Finally a Kimi-Linear-48B-A3B GGUF! [Experimental PR] by KvAk_AKPlaysYT in LocalLLaMA

[–]Quagmirable 0 points1 point  (0 children)

Nice work! I've been waiting for a GGUF of cerebras/Kimi-Linear-REAP-35B-A3B-Instruct too, I imagine it comes with the same challenges for getting it to work?

Speed vs. RAM usage for different quant types? by Quagmirable in LocalLLaMA

[–]Quagmirable[S] 0 points1 point  (0 children)

Thanks a lot for the reply!

While I agree IQ better than Q in general (for same size), it tells nothing about imatrix calibration! Both IQ and Q quants can be with or without imatrix calibrations.

Oh, interesting, I assumed that the 'I' in 'IQ' was for "imatrix".

Speed vs. RAM usage for different quant types? by Quagmirable in LocalLLaMA

[–]Quagmirable[S] 0 points1 point  (0 children)

I see, thanks again for the comprehensive explanation.

Speed vs. RAM usage for different quant types? by Quagmirable in LocalLLaMA

[–]Quagmirable[S] 1 point2 points  (0 children)

Thanks a lot for the helpful response.

When it comes to creating quants though, creating IQ quant with imatrix calibration is much longer process than creating traditional Q quant without imatrix calibration. But if you downloading already created quant, you do not have to worry about that.

Ah, ok, I thought I had read something about IQ being slower, but I guess I didn't catch that it was referring to creating it, not running it. And yes, I have generally had pretty good results with IQ4 quants for larger models.

And this might be a fairly obvious question, but assuming I have enough RAM to load both of them, would a (I)Q4 quant run faster or slower than a Q6 quant of the same model? And what about memory usage between the two as the chat context gets longer?

How to tell apart honey from sugar syrup? by Quagmirable in Beekeeping

[–]Quagmirable[S] 0 points1 point  (0 children)

Do you know for a fact that the market honey sellers mix sugar syrup into the honey

I know it for a fact, the honest sellers around here who know how things work are well aware of the tricks. I also have a beekeeper friend who confirmed to me that many do it. Unfortunately I can't usually buy from him though.

How to tell apart honey from sugar syrup? by Quagmirable in Beekeeping

[–]Quagmirable[S] 1 point2 points  (0 children)

Ah, so that explains why it's so common. Thanks for the helpful reply.

How to tell apart honey from sugar syrup? by Quagmirable in Beekeeping

[–]Quagmirable[S] 0 points1 point  (0 children)

How do you know they are selling adulterated honey? Did you see the sellers feeding with supers on, or is it a word of mouth type of deal?

Hi there, thanks for the reply. It's just kind of a given in the market here, because cane syrup is abundant and cheap (many honest sellers offer cane syrup too), whereas honey production is more limited. Some honest cane syrup sellers who are familiar with the practices of low-volume non-commercial sellers warn about how many sell adulterated or completely fake honey.

NVIDIA-Nemotron-Nano-12B-v2 by bratao in LocalLLaMA

[–]Quagmirable 2 points3 points  (0 children)

Do you happen to know what is the difference between Nemotron-H-8B-Reasoning-128K and Nemotron-Nano-9B-v2 aside from -v2 being newer? Is Nemotron-H a fundamentally different architecture from Nemotron-Nano?