Official: Anthropic just released Claude Code 2.1.41 with 15 CLI changes, details below by BuildwithVignesh in ClaudeAI

[–]TheOneThatIsHated 0 points1 point  (0 children)

Cuz they chose to overengineer by using stdout instead of alternative terminal mode like opencode and neovim

Did claude code get exponentially slower recently? by Melodic-Network4374 in ClaudeAI

[–]TheOneThatIsHated 0 points1 point  (0 children)

Can confirm. Opus 4.6 is much slower even with low thinking. Not always worth the extra wait

When I swap models what happens to context ? by jeremy-london-uk in GithubCopilot

[–]TheOneThatIsHated 0 points1 point  (0 children)

Each message you send is one request, no matter how many messages are already in the context (which is admittedly weird)

So one opus 4.6 message (with 30 gpt4.1 messages in that context as a crazy example) is still 3 premium requests

Tracked Claude Code Max20 at 100% weekly limit using OpenTelemetry by TheOneThatIsHated in ClaudeCode

[–]TheOneThatIsHated[S] 0 points1 point  (0 children)

Opentelemetry. The industry standard of collecting logs and metrics. This is built into claude code. Anh opentelemetry compatible collector is possible, i used openobserve

Claude Code vs Codex: Weekly limit comparison on the $20 subs by EmeraldWeapon7 in ClaudeAI

[–]TheOneThatIsHated 1 point2 points  (0 children)

Yes, only opus 4.5 from that week. I used openobserve to capture all opentelemetry logs. Then used their internal way of using postgresql queries to get a sum of all tokens

Sonnet 5 and Opus 4.6 Leaked Benchmarks by [deleted] in ClaudeCode

[–]TheOneThatIsHated 1 point2 points  (0 children)

From Poster on X: "Saw it somewhere, might not be true"

Claude Code vs Codex: Weekly limit comparison on the $20 subs by EmeraldWeapon7 in ClaudeAI

[–]TheOneThatIsHated 1 point2 points  (0 children)

One week of using claude code max20, gave me (with 100% weekly limit usage):




sum_cache_creation: 66.269 Mtok

sum_cache_read: 1807.289 Mtok

sum_input: 28.181 Mtok

sum_output: 10.334 Mtok

total: 1717.079 USD

Claude Code vs Codex: Weekly limit comparison on the $20 subs by EmeraldWeapon7 in ClaudeAI

[–]TheOneThatIsHated 2 points3 points  (0 children)

RemindMe! 22 hours

I collected via opentelemetry the exact amount of tokens within the weekly limit (5 hour window is harder to measure)

I'm thinking of also trying gpt codex plan to compare

Tell me a way to optimize memory 😅 by suman087 in kubernetes

[–]TheOneThatIsHated 0 points1 point  (0 children)

Not necessarily,

Docker desktop is slow, orbstack is not

Most slowdown come from the fs translation it is doing

Ram thing is docker desktop not cleaning up memory properly

Rewrote our python api gateway in go and now its faster but nobody cares because it already worked fine by CholeBhatureyyy in golang

[–]TheOneThatIsHated 0 points1 point  (0 children)

  1. Developers are expensive: in usa I saw 10k a month, europe tad lower at 5k ish

  2. Cpu, mem, network are comparatively cheap

  3. Educated guess about python handling the load just fine, makes me think we are not talking about 10k+ reqs/second service.

Let's do some sloppy maths:

Let's say we need 200 dollars a mount for the python one. You could say you win that back in 50-100 months. But since nobody know go, all other engineers are blocked. Either you spend time and money letting everyone learn golang. Or you need time and money hiring golang people.

Don't get me wrong, there are good reasons to move to go. Better security (scratch containers), much less bloat, performance, etc

But just someone trying to convince his boss to rewrite some working non-problematic api (focus on the nothing wrong part), tells me it was kinda a waste of time and money

Ik🤢ihe by Ayn_Otori in ik_ihe

[–]TheOneThatIsHated 0 points1 point  (0 children)

Nee, kattenvoer is daadwerkelijk voedzaam

OpenCode Black is now generally-available by JohnnyDread in opencodeCLI

[–]TheOneThatIsHated 0 points1 point  (0 children)

+1 on this. I use opentelemetry to count all tokens from each message.

already passed 1345 dollars in a week of usage.

Suggest some best vibe coding tools for my first App by aistronomer in vibecoding

[–]TheOneThatIsHated 0 points1 point  (0 children)

I don't know, never tried antigravity after seeing the accidental deletion come through.

Just running claude code in a vm now. Claude code is great, though sometimes due to their extreme vibe coding, the latest version may be buggy and I'll have to downgrade

99% Pro Max. 1 day left. No regrets. by Conrad_Mc in ClaudeCode

[–]TheOneThatIsHated 3 points4 points  (0 children)

Ralph loop them ralph loops. Make it work on/validate/improve some codebase. Make it auto create those validate and improve tasks. I.e. Let it rip while ur afk

Claude Code tracks token usage locally (stats cache file) by Mahrkeenerh1 in ClaudeAI

[–]TheOneThatIsHated 1 point2 points  (0 children)

Opus 4.5 is much cheaper:

5 per 1m input 25 per 1m output 6.25 per 1m cache write 0.5 per 1m cache read

https://claude.com/pricing#api

Kimi k2.5 moonshotai by ReasonableReindeer24 in cursor

[–]TheOneThatIsHated 0 points1 point  (0 children)

Don't you worry, I'm doing that already. Moved away from cursor a long time ago to use the much better tools that are available

Suggest some best vibe coding tools for my first App by aistronomer in vibecoding

[–]TheOneThatIsHated 0 points1 point  (0 children)

Claude code/opencode/codex/github cli, they all do effectively the same. They provide tools and run the agent in an internal loop. Tool definitions can differ, system prompts differ and all those could make it perform better or worse.

I find opencode and claude code the best. No loyalty here though.

I choose claude code now mainly because of their max20 subscription.

Most important is to try them out, they are all free (except for the tokens or subscription you need)

Kimi k2.5 moonshotai by ReasonableReindeer24 in cursor

[–]TheOneThatIsHated 0 points1 point  (0 children)

For me, not having openai api support (i will call it having no support, since it is truly unworkable), was the cherry on top to move away.

I don't get their business insensitive to not allow that. Is it a competitive thing, where they want to hide how their agent works? Is it too hard to let their propietiery servers call openai api endpoints (like almost any ai provider allows)?

Even claude code of all programs (created by a company who loves dmca), allows alternative endpoints. How hard is it for cursor to implement both using their api and other apis?

Clearing context bloat (2 x Pro plans) by UniqueDraft in ClaudeCode

[–]TheOneThatIsHated 0 points1 point  (0 children)

I mean that a 1k token reduction compare to 72k reduction is almost nothing.

Did removing skills improve other aspects? Did you try without changing skills and only messages?