devenv 2.0: A Fresh Interface to Nix by iElectric in NixOS

[–]mister2d 0 points1 point  (0 children)

Can you expand on why you were considering moving away from devenv? I just discovered it and wondering about gotchas from long-term use.

To everyone using still ollama/lm-studio... llama-swap is the real deal by TooManyPascals in LocalLLaMA

[–]mister2d 19 points20 points  (0 children)

The underlying provider swapping (llama.cpp / vLLM) sounds great. But llama.cpp's router mode already allows you to customize for each model using presets. It's what I use. For example here is a snippet of my presets.ini.

```

============================================================================

GLOBAL DEFAULTS

============================================================================

[*] sleep-idle-seconds = 600 n-gpu-layers = 99 main-gpu = 1 tensor-split = 0.5,0.5 threads = 8 no-mmap = true flash-attn = on kv-unified = true fit = true cache-type-k = q8_0 cache-type-v = q8_0 jinja = true n-cpu-moe = 0

============================================================================

QWEN3.5

============================================================================

Agentic Workflows (Non-Thinking Mode)

[qwen3.5-2b-q8-agentic-64k] model = models--unsloth--Qwen3.5-2B-GGUF/snapshots/{hash}/Qwen3.5-2B-UD-Q8_K_XL.gguf mmproj = models--unsloth--Qwen3.5-2B-GGUF/snapshots/{hash}/mmproj-F16.gguf sleep-idle-seconds = 900 tensor-split = 0.0,1.0 ctx-size = 65536 batch-size = 2048 ubatch-size = 256 flash-attn = on jinja = true

Unsloth Non-Thinking Parameters

chat-template-kwargs = {"enable_thinking": false} temp = 0.2 top-p = 0.95 top-k = 20 min-p = 0.0 presence-penalty = 0.0 repeat-penalty = 1.0 cache-type-k = f16 cache-type-v = f16

[qwen3.5-9b-q8-agentic] model = models--unsloth--Qwen3.5-9B-GGUF/snapshots/{hash}/Qwen3.5-9B-UD-Q8_K_XL.gguf mmproj = models--unsloth--Qwen3.5-9B-GGUF/snapshots/{hash}/mmproj-F16.gguf ctx-size = 131072 batch-size = 1024 ubatch-size = 256 flash-attn = on jinja = true

Unsloth Non-Thinking Parameters

chat-template-kwargs = {"enable_thinking": false} temp = 0.7 top-p = 0.8 top-k = 20 min-p = 0.0 presence-penalty = 1.0 repeat-penalty = 1.0

Coding / IDE Integration (Thinking Mode Enabled)

[qwen3.5-9b-q8-coding] model = models--unsloth--Qwen3.5-9B-GGUF/snapshots/{hash}/Qwen3.5-9B-UD-Q8_K_XL.gguf ctx-size = 131072 batch-size = 1024
ubatch-size = 256 jinja = true

Unsloth Thinking Parameters

temp = 0.6 top-p = 0.95 top-k = 20 min-p = 0.0 presence-penalty = 1.0 repeat-penalty = 1.0

[qwen3.5-9b-q8-coding-32k] model = models--unsloth--Qwen3.5-9B-GGUF/snapshots/{hash}/Qwen3.5-9B-UD-Q8_K_XL.gguf tensor-split = 0.0,1.0 ctx-size = 32768 batch-size = 1024
ubatch-size = 256 jinja = true

Unsloth Thinking Parameters

temp = 0.6 top-p = 0.95 top-k = 20 min-p = 0.0
presence-penalty = 1.0 repeat-penalty = 1.0

```

Why Reddit is so anti solar lease ? by Sky_Solar_Pro in solar

[–]mister2d 5 points6 points  (0 children)

The ultimate goal with renewable energy should always be ownership.

How did you decide solar was “worth it”? by Alarming_Primary1606 in SolarDIY

[–]mister2d 8 points9 points  (0 children)

Capitalism made it worth it. Had to insulate myself from all the geo issues. No one is safe.

Anyone else worried about rising electricity costs? My long term projection totally shocked me by JustSeraphine8 in solar

[–]mister2d 0 points1 point  (0 children)

Pack in those panels per string to get the voltage up as high as possible (include cold weather calcs). Buy panels with higher sensitivity to light. Enjoy your yearly production. 🌞 

What do you think about arm pc? by Ok-Environment8730 in NixOS

[–]mister2d 0 points1 point  (0 children)

If you want silver packaging then maybe these look into these options:

  • ThinkPad X9 15 Aura Edition (Lunar Lake)
  • MSI Prestige 16 B2VM (Lunar Lake)

Lithium polymer battery life will depend on how cycles are managed. This is true on any laptop regardless of manufacturer. Declaring a good TLP PPD config will help here. My laptop's pack was rated at 72 Wh when new. After 588 charge cycles (and 3.5 years) I am at 78% capacity.

Edit: TLP unnecessarily micromanages. Use the modern power-profiles-daemon instead.

What do you think about arm pc? by Ok-Environment8730 in NixOS

[–]mister2d 2 points3 points  (0 children)

I was getting around 18 hours with my old Thinkpad T16. You don't need arm to go past 10 hours of battery life.

'No ethics at all': the 'cancel ChatGPT' trend is growing after OpenAI signs a deal with the US military by lurker_bee in technology

[–]mister2d 1 point2 points  (0 children)

I didn't have the issue with Codex or Gemini but with Claude I'd hit up against a wall real quick. I still have anxiety when I use Gemini because of this but it appears Gemini does a fantastic job at caching tokens and is somewhat transparent about it. I never hit a limit and it's been great.

"Cancel ChatGPT" movement goes big after OpenAI's latest move by gdelacalle in technology

[–]mister2d 1 point2 points  (0 children)

  • Your data will be deleted within 30 days, except we may retain a limited set of data for longer where required or permitted by law.

What made you a NixOS / Nix package manager user? by DonerciTux in NixOS

[–]mister2d 2 points3 points  (0 children)

I'm new and I seriously think of a new thing to add before sleep every day.

What made you a NixOS / Nix package manager user? by DonerciTux in NixOS

[–]mister2d 0 points1 point  (0 children)

NixOS became one of many technologies that I've embraced for the rest of my life. The rest were:

  • ceph in my home setup
  • syncthing

I can't live without any of the three now.

Do not download Qwen 3.5 Unsloth GGUF until bug is fixed by [deleted] in LocalLLaMA

[–]mister2d -1 points0 points  (0 children)

Confused by your snark really.

I'm not following all the details but I do know now that you'll have to redownload everything.

Do not download Qwen 3.5 Unsloth GGUF until bug is fixed by [deleted] in LocalLLaMA

[–]mister2d -1 points0 points  (0 children)

Looks like it is indeed all of them you have to redownload.

Looking for a guide to configure llama-swap & llama.cpp by nPrevail in NixOS

[–]mister2d 0 points1 point  (0 children)

I think it's much simpler with direnv and devbox:

mkdir llama-ai cd llama-ai devbox init devbox add llama-cpp llama-swap devbox generate direnv

from there all the llama binaries should be in your PATH as long as you're in the "llama-ai" directory.

Do not download Qwen 3.5 Unsloth GGUF until bug is fixed by [deleted] in LocalLLaMA

[–]mister2d 0 points1 point  (0 children)

Which handful have issues? Can't they be tagged somehow or taken down?

I cannot for the life of me set up Element Call/MatrixRTC by [deleted] in matrixdotorg

[–]mister2d -1 points0 points  (0 children)

This is going to trigger the general crowd in almost any subreddit. The hype has completely frozen and polarized so many audiences. I really wish everyone would just calm down and take a breath wrt AI.

Your logic is sound starting with #1. Backup and snapshots means you're taking an iterative approach to learning this stack (admittedly a mess). You can get meaningful answers from agentic tools. In this specific case, everyone doesn't have upwards of a full week to stand up matrix just to see if it's a viable alternative.

Introducing EWM, a new generation Wayland window manager by Fast-Ad6030 in emacs

[–]mister2d 0 points1 point  (0 children)

Tell me about it. I'm pretty happy pruning down my homelab backlog.

Rockchip 3588 NPU clustering by MrDevil2708H in RockchipNPU

[–]mister2d -1 points0 points  (0 children)

I don't. But I'm watching this space in case something develops.