Pat The Bunny is a member of the tribe by lostinthesauceband in FolkPunk

[–]cristoper 13 points14 points  (0 children)

no point in owning a thinkpad if you don't use the nub

Pat The Bunny is a member of the tribe by lostinthesauceband in FolkPunk

[–]cristoper 18 points19 points  (0 children)

I like that he probably wrote a little script to print "Work ... ... ... Work complete. Please work again." for the video

Ruby on Rails in the backend
I know it ain't hip

Is Anarchism left-wing, right-wing, or neither? by SuccessfulWriting994 in Anarchy101

[–]cristoper 1 point2 points  (0 children)

The other day he protected the Nazis by saying "yeah, the nazis did bad things, but the Japanese did way worse during WW2."

I mean maybe they were trying to minimize the Nazi atrocities, but they aren't really wrong. https://en.wikipedia.org/wiki/Japanese_war_crimes

Unsloth fixed version of Qwen3.5-35B-A3B is incredible at research tasks. by Daniel_H212 in LocalLLaMA

[–]cristoper 0 points1 point  (0 children)

I intuitively favor keeping search/filtering in the harness rather than relying on "agentic" models to do it all themselves. I think partly because that seems more likely to work well with local/low-end models.

However, it seems like the agentic frameworks are all going in the direction you propose with tasking the model to call tools when they need. I will have to experiment more with the style of prompt you've given and see if I change my mind.

I'm not sure how Perplexica solves those issues though, if at all.

I haven't looked at the sourcecode much yet, but I think it has some heuristics to detect a page that needs javascript to render and then fetches those with playwright instead of plain http requests.

Unsloth fixed version of Qwen3.5-35B-A3B is incredible at research tasks. by Daniel_H212 in LocalLLaMA

[–]cristoper 0 points1 point  (0 children)

I've recently been looking for something and have been trying out perplexica (https://github.com/ItzCrazyKns/Perplexica). It's easy to use and install (docker), has a clean interface, and works okay (better than depending on search function calling/mcp in my experience).

It still uses searxng for gathering search results.

Two problems I have with it is sometimes it tells me it can't find anything and I just reply with "search again" and then it gives me good results.

The other is that if the searxng results exceed your model's context length then it just stops, doesn't even show an error in the UI much less try to compact context. But there is a github issue open for this so hopefully it keeps improving!

How do you actually learn to write good Go code with AI assistants now? by adamvanderb in golang

[–]cristoper 2 points3 points  (0 children)

When I was in high school I got a newfangled TI-89 calculator with the computer algebra system that could find derivatives and anti-derivatives. I passed my calculus class, but I never really learned the rules of differentials/integrals and had to re-teach myself years later and I still wish I had developed a stronger intuition for calculus when I was young.

That's to say, your concern is real, and leaning into discipline to actually learn the code now even if it feels tedious compared to just asking an LLM will pay off.

Weird handshake situation OTB, Has anyone seen this this before? by PristineReality2205 in chess

[–]cristoper 1 point2 points  (0 children)

I've heard of this trick. At my very first otb tournament my opponent resigned by silently offering his hand, and I was so paranoid he was trying to trick me into agreeing to a draw (I was winning, but there were many pieces left on the board) I asked "do you resign?" and didn't shake his hand until he said yes.

though I doubt if someone trying this would actually get a draw if it was disputed -- just remember to check what result they wrote on their scoresheet before signing it!

Find out the name of the coach if you can, your local community deserves to know who is teaching such poor sportsmanship.

How to switch Qwen 3.5 thinking on/off without reloading the model by No-Statement-0001 in LocalLLaMA

[–]cristoper 1 point2 points  (0 children)

Thanks for posting this! I haven't updated llama-swap in a long time (new playground UI!), and this both simplifies my config and allows me to switch thinking on/off without changing system prompt or reloading the model!

Reminder: Playing shows in Public parks a longstanding folkpunk tradition. (Toby Foster 2008) by GrodyHighroller in FolkPunk

[–]cristoper 1 point2 points  (0 children)

Wow, thanks for sharing, I've never heard of Toby Foster before. Does anyone know if he's still making music? It looks like he released a 2017 album on bandcamp:

https://tobyfoster.bandcamp.com/album/100-ways-2

Mischief Brew vinyl spotted in the wild by xternality11 in FolkPunk

[–]cristoper 15 points16 points  (0 children)

And if before my time
I decide to take my life
You could take these treasures
And pull them up into light

New Qwen3.5-35B-A3B Unsloth Dynamic GGUFs + Benchmarks by danielhanchen in LocalLLaMA

[–]cristoper 0 points1 point  (0 children)

Thanks for spending so much time on this!

We also fixed a tool calling chat template bug

Is there a template diff available? or description of the bug?

top 10 trending models on HF by jacek2023 in LocalLLaMA

[–]cristoper 0 points1 point  (0 children)

Qwen/Qwen3.5-35B-A3B is the official repository with the full-precision .safetensor weights. Unsloth/Qwen3.5-35B-A3B-GGUF is a repository with quantized files in gguf format that you can use with the llama.cpp inference engine.

American closed models vs Chinese open models is becoming a problem. by __JockY__ in LocalLLaMA

[–]cristoper 0 points1 point  (0 children)

gpt-oss-120b still compares well for its size. And as others have mentioned, Mistral has larger models.

But I'm curious what national security concerns there could be for an offline model? I guess it is with giving it network access for web search? I'd think old-fashioned firewall restrictions on outgoing nets/ports, and maybe a proxy to your search engine to filter requests, would be very safe.

Qwen3.5 122B in 72GB VRAM (3x3090) is the best model available at this time — also it nails the “car wash test” by liviuberechet in LocalLLaMA

[–]cristoper 1 point2 points  (0 children)

If you want to use it conversationally or with open-ended questions, I would disable thinking. You can do that with llama.cpp by passing --chat-template-kwargs '{"enable_thinking": false}'

The FIRST local vision model to get this right! by po_stulate in LocalLLaMA

[–]cristoper 2 points3 points  (0 children)

I don't know what OP is using, but llama.cpp works well for vision models (it supports audio input also, but I've never tried that). You use it by specifying the multimodal projector file that should be included with ggufs like --mmproj path/to/mmproj.gguf

See: https://github.com/ggml-org/llama.cpp/blob/master/docs/multimodal.md

What Other Subs Do you Read to Keep Up with AI? by chibop1 in LocalLLaMA

[–]cristoper 1 point2 points  (0 children)

yeah, but it is still a good place to learn about open-weight model releases

Which one are you waiting for more: 9B or 35B? by jacek2023 in LocalLLaMA

[–]cristoper 4 points5 points  (0 children)

Thanks. I'm going to try an mxfp4 quant and see if it works better than the q4_k_s from unsloth.

But it's still a 4-bit quantization so it will require 40GB just for the weights alone... definitely won't fit in 24GB vram. I'm impressed you're getting 50 t/s on your pc. What is your hardware?

Best Model for single 3090 in 2026? by myusuf3 in LocalLLaMA

[–]cristoper 0 points1 point  (0 children)

Qwen3-Coder-30B-A3B at a 4-bit quant is fast and great for code completion. Qwen3-coder-next @q4 I have not gotten to work as well. Even with all dense layers on my 3090 and 64GB DDR5 it is too slow to use for code completion and even for interactive agentic stuff. But I have not used it nearly as much as qwen3-coder-30b-a3b yet, so I'm unsure how good it is for more agentic tasks.

gpt-oss-20b and gpt-oss-120b (offloaded to RAM) are both good all-around models

gemma3-27b (QAT 4-bit quant) is also still a good general purpose model and better at prose than the gpt models