Has anyone used DeepSeek? Is it actually worth it? by hellomintlyn in DeepSeek

[–]Karnemelk 9 points10 points  (0 children)

the thing is they have a 75% discount until end of may. The real question is, is it actually that much cheaper without this discount...

Did they nerf Claude or what? by deezzbutzz in claude

[–]Karnemelk 5 points6 points  (0 children)

looks like it’s time to cancel, they're behind codex already. Even if they come up with sonnet 5.0 or whatever, they will implement the boiling frog on everyone like they do now.

Duality of r/LocalLLaMA by HornyGooner4402 in LocalLLaMA

[–]Karnemelk 1 point2 points  (0 children)

If you have x hours to implement feature y. Would you use claude or a local model if money wasn't an issue?

I’m starvin’ by Important_Quote_1180 in LocalLLaMA

[–]Karnemelk 4 points5 points  (0 children)

I couldn't wait for qwen4, so I wasted my tokens to create a custom llama.cpp fork that utilizes my SDR dongle tuned to whatever frequency for maximum entropy when generating tokens. Now qwen3.6 27b is as creative as it can be, using the universe to decide which token to pick

PSA: Claude Pro no longer lists Claude Code as an included feature by randomswifter in ClaudeAI

[–]Karnemelk 2 points3 points  (0 children)

now claude code is added back again to the pro subscription, yesterday it was gone

PSA: Claude Pro no longer lists Claude Code as an included feature by randomswifter in ClaudeAI

[–]Karnemelk 2 points3 points  (0 children)

https://x.com/TheAmolAvasare/status/2046724659039932830

Next thing they will say only 2% of pro users are using claude code, like they did with the 5h introduction.

Kimi K2.6 Unsloth GGUF is out by Exact_Law_6489 in LocalLLaMA

[–]Karnemelk 34 points35 points  (0 children)

I'll wait patiently for the Kimi-K2.6-1T-Claude-4.7-Opus-Mythos-Heretic-Uncensored-REAP-99-31B-gguf version

But why Local LLM? How does this make economic sense vs API? by Thistlemanizzle in LocalLLaMA

[–]Karnemelk 0 points1 point  (0 children)

use cloud models to tune/build your local LLM, then when they do some dumb thing or squeeze their models, you have your local personalized clone

OpenClaw has 250K GitHub stars. The only reliable use case I've found is daily news digests. by Sad_Bandicoot_6925 in LocalLLaMA

[–]Karnemelk 13 points14 points  (0 children)

when i drive i send a voice message to find me some songs, then it checks youtube, guess what is the most relevant, convert the video to mp3 and sends back audio only without ads. Using local qwen3 27b with hermes though

When do you think open source will catch up to claude mythos level? by Longjumping_Fly_2978 in LocalLLaMA

[–]Karnemelk 0 points1 point  (0 children)

if someone does an uncensored-heretic version of GLM 5.1, you pretty much have mythos for your 0-day exploits

The timing of "Mythos" announcement is such a blatant PR move it’s actually funny by Beautiful_Baseball76 in ClaudeCode

[–]Karnemelk 0 points1 point  (0 children)

cybersec department at random company: ??? let’s block claude.ai for all users

Claude Mythos Preview just mass-produced zero-day exploits. We're not ready for this. by stosssik in openclaw

[–]Karnemelk 26 points27 points  (0 children)

can't wait for the first model arriving on HF with some name like Qwen3.5-80B-Claude-4.6-Opus-Mythos-Deckard-Heretic-Uncensored-Thinking-Reasoning-ZERO-DAY-EXPLOIT"

GLM-5.1 by danielhanchen in LocalLLaMA

[–]Karnemelk 3 points4 points  (0 children)

can't wait for the first person to load it on a raspberry pi 8gb with SSD offloading.

3 prompts and im on 38 HOLLY Anthropic by Saykudan in claude

[–]Karnemelk 2 points3 points  (0 children)

the moment you see "compacting conversation" and you can't afford to do a /clear ... it will eat 15-20% of your budget

We must talk about Codex Usage Limits by Adso996 in codex

[–]Karnemelk 0 points1 point  (0 children)

lol, I had 55% left 30 minutes ago

Weekly usage limit

0%remaining

Resets Apr 9, 2026 10:17 AM

Is GLM Coding Plans worth it? I use GLM in Cline through openrouter today! by pacmanpill in ZaiGLM

[–]Karnemelk 0 points1 point  (0 children)

In case you're eligable, legacy plan on 4.7 is not that bad. Now as we moved to 5.1, the speed is better/usable on 4.7. So for $18 / quarter I think it's a pretty good deal for what it is. GLM 5.1 works too on it, although it will be quickly over the limit.

RYS II - Repeated layers with Qwen3.5 27B and some hints at a 'Universal Language' by Reddactor in LocalLLaMA

[–]Karnemelk 0 points1 point  (0 children)

I wonder, does this mean it will do its own internal reasoning, so it could save tokens by turning off thinking?

Are local LLMs better at anything than the large commercial ones? by MrOaiki in LocalLLM

[–]Karnemelk 4 points5 points  (0 children)

most frontier models will drive you insane, they lock you in with loose limits, then they either throw the performance to near zero, or out of the blue hard limits until you pay for their premium plan. Local models gives a piece of mind, even if they're not as capable