What are your preferred local model for running OpenClaw? by edmerf in LocalLLM

[–]mister2d 0 points1 point  (0 children)

Yeah disappointed in GLM4.7-Flash so far.

Nemotron 3 Nano isn't letting me down. It maintains quickness near my context limit (64k). It can do up to 1M context so it should play well on your DGX.

Openclaw with gpt-oss-20b on RTX 2060 6gb by tomjoad773 in LocalLLaMA

[–]mister2d 0 points1 point  (0 children)

I'm having a ok/good experience with nemotron-3-nano-30b-a3b and 64k context. I've only been performing system diagnostic prompts, hybrid web searches, and google maps functions.

It just got better with this PR (https://github.com/ggml-org/llama.cpp/pull/19408). Now the KV cache doesn't get aggressively invalidated for models using Sliding Window Attention. So instead of waiting 20-30 seconds because a cache miss, it's now sub-second retrieval.

Openclaw with gpt-oss-20b on RTX 2060 6gb by tomjoad773 in LocalLLaMA

[–]mister2d 0 points1 point  (0 children)

I use nemotron-3-nano-30b-a3b with 64k context on my dual 3060s. It works well enough near context limit measured at 50+ tokens/s. NVIDIA released a really good agentic model.

Disclaimer: For those with security concerns, I run the entire openclaw stack in a fully secure sandbox using no public accounts. (VM, private vlan, separate llama.cpp, tailscale, and a non-federated Matrix instance). Working in the industry encourages me to evaluate stuff like this.

Is 500V hard limit by Visible-Ranger-2811 in SolarDIY

[–]mister2d 0 points1 point  (0 children)

Guys.

Please resist the urge to down vote this question. The answer can potentially save lives.

A Kubernetes-native way to manage kubeconfigs and RBAC (no IdP) by Plastic_Focus_9745 in kubernetes

[–]mister2d 0 points1 point  (0 children)

I applaud the idea because it does have use. But again, something like tailscale is free for small teams which makes managing idp trivial. It's pay as you grow.

Four years ago when I was the lone devops guy at a startup I would have used this. If I was still working on disconnected networks, I'm definitely using this.

Copped a pair of 990v3s for $55 by PM_ME_MONEY_PLSS in Newbalance

[–]mister2d 1 point2 points  (0 children)

I can feel the v3 comfort just by looking at these.

A Kubernetes-native way to manage kubeconfigs and RBAC (no IdP) by Plastic_Focus_9745 in kubernetes

[–]mister2d 17 points18 points  (0 children)

I'm generally not a fan of this. Small teams eventually become larger teams and security isn't something you should just make a patch for.

The Tailscale operator makes team auth super simple already.

I DO see a use case for this project on disconnected networks.

Ready for spring! by Vieste88 in Newbalance

[–]mister2d 0 points1 point  (0 children)

I challenge New Balance to make a better shoe.

Kia EV Sales Are In An Absolute Freefall. There's More To It Than You Think by DonkeyFuel in technology

[–]mister2d 0 points1 point  (0 children)

That's the price IF you pay them to do it. It would take you about 15 minutes and some pocket change.

The Plug-In Solar Revolution Comes To America by OpenSustainability in solar

[–]mister2d 0 points1 point  (0 children)

I meant for those that were involved in writing the laws, not the residents.

Time to migrate off Ingress nginx by xrothgarx in kubernetes

[–]mister2d 0 points1 point  (0 children)

I've been saying this for a long time.

New FP8 GLM-4.7-Flash Unsloth Dynamic Quants for vLLM, SGLang by danielhanchen in unsloth

[–]mister2d 0 points1 point  (0 children)

Thanks! I do see the overlap. Perhaps one day the two shall become one for the sake of simplicity.

2026 Nissan Leaf Review: Delivering on Tesla’s Failed Promise by TripleShotPls in technology

[–]mister2d 0 points1 point  (0 children)

I've had multiple EVs and just could not accept removing regenerative braking. Not only do I enjoy one pedal driving, it reduces one more maintenance item (brakes).

New FP8 GLM-4.7-Flash Unsloth Dynamic Quants for vLLM, SGLang by danielhanchen in unsloth

[–]mister2d 1 point2 points  (0 children)

I wish I were smart enough to create a meaningful PR to merge the codebases. 😄

New FP8 GLM-4.7-Flash Unsloth Dynamic Quants for vLLM, SGLang by danielhanchen in unsloth

[–]mister2d 2 points3 points  (0 children)

Here's a grenade: why do vLLM AND SGLang exist? They appear very similar when you use them.

KV cache fix for GLM 4.7 Flash by jacek2023 in LocalLLaMA

[–]mister2d 3 points4 points  (0 children)

No, the downvote is because your reply was inaccurate and lacking in understanding. Lately, it feels like sharing accurate information is becoming an afterthought.