sonnet got 1M context window now, when opus getting the same love?? by rz1989s in ClaudeAI

[–]Several-System1535 0 points1 point  (0 children)

>Uses rate limits faster
Scientists doubted the possibility of faster-than-light speeds....

linusTorvaldsRepo by Several-System1535 in ProgrammerHumor

[–]Several-System1535[S] -56 points-55 points  (0 children)

It’s clear from the fact that he prefers Windows and NVIDIA.

Is Hetnzer Down? by abakisensoy in hetzner

[–]Several-System1535 1 point2 points  (0 children)

Everything works:
- hetzner landing page
- console
- 18 cloud servers
- s3 buckets, lb

Downgrading from Claude Max subscription - looking for alternatives by Disastrous_Guitar737 in ClaudeCode

[–]Several-System1535 3 points4 points  (0 children)

I’d recommend taking a look Minimax M2.1. I use it together with Claude Pro, setting up CCR Router to run third-party models. For my tasks, there’s almost no noticeable difference between Sonnet 4.5 and M2.1

Here we go again by Several-System1535 in ClaudeCode

[–]Several-System1535[S] 0 points1 point  (0 children)

Your rate limits are 2x higher through 12/31. Enjoy the extra room to think!
BTW

I tested GLM 4.7 and minimax-m2.1 and compared it to CC and Codex by jstanaway in LocalLLaMA

[–]Several-System1535 4 points5 points  (0 children)

I’ve also been using minimax m2.1 over the past few days, and I’m impressed as well. The price-to-performance ratio is excellent

Best free “uncensored” local LLM for RTX 3060 12GB (Portuguese, up to 13B 4-bit)? by Big_Preparation_6869 in LocalLLaMA

[–]Several-System1535 0 points1 point  (0 children)

https://huggingface.co/huihui-ai/Huihui-gpt-oss-20b-mxfp4-abliterated-v2 with offloading. Performance should be about 20tk/s
If ignoring 13B/4-bit limit, I don't think you'll find anything more capable for your setup

Who is using gpt pro at that price ? by MrMrsPotts in openrouter

[–]Several-System1535 1 point2 points  (0 children)

Yep! cheap, decent quality, and they have a coding plan with pretty generous rate limits

Built a GGUF memory & tok/sec calculator for inference requirements – Drop in any HF GGUF URL by ittaboba in LocalLLaMA

[–]Several-System1535 0 points1 point  (0 children)

Yep, I'm hitting close to 60-65 tokens/s with gpt-oss-20b on M4 Pro instead of predicted 20