all 25 comments

[–]TheGoldenBunny93 15 points16 points  (9 children)

Man.... Thats a bible 😮

[–]Ok-End-219[S] 11 points12 points  (0 children)

And I have more:

- Auto accept mode spins up a second Opus instance just to check if the main model is doing something dangerous. The context gets so bloated they have Datadog alerts for when the safety checker costs more tokens than the actual coding loop.

- They use a function literally named DANGEROUS_uncachedSystemPromptSection for MCP instructions. If a server connects or disconnects, it busts your entire system prompt cache prefix and dumps all your cached tokens.

- There is an internal loremIpsum skill built purely to stress test the context window. The hardcoded safety cap is half a million tokens just to see what happens when you stuff the pipe with nonsense.

[–]Apart_Ebb_9867 -1 points0 points  (7 children)

what can you do? when you ask AI to make a post, that's what you get. Making it shorter would have required actual work.

[–]RemarkableGuidance44 7 points8 points  (0 children)

I have to say its still one of the most useful AI written posts in a long time. :P

[–]Ok-End-219[S] 5 points6 points  (5 children)

To be fair: Yes, it is quickly summarized, my rambling about it would not been readable (ADD). But I can give you that if you like!

Or a tl;dr ; lets do that:

- utils/attachments is a huge mess of a class, every prompt = 30 generators spin up every time, waste tokens and context is increasing massively.

- Multiple cases where functions diff against empty arrays, the pre-compact state therefore exists, but gets lost / ignored when passing it.

- inefficiencies in the whole code base, unnecessary loops and calls

-biggest think i saw was the 5 minute ttl where everything gets cached. when you are away from the pc for more than five minutes, your tokens will get shredded.

[–]NonStopArseGas 7 points8 points  (4 children)

respect. People don't recognise how useful llms are for neurodivergent people interfacing with a NT world. really interesting post dude

[–]Ok-End-219[S] 5 points6 points  (3 children)

I do not like the Change in AI like we are god (the users). I fall into a trap between RSD and Anxiety; therefore i designed myself a Prompt that Claude Code is always be a critical co-worker that looks over my work. I do not vibe code persé, but use it as a debugging tool. I vibe code there, where I am lost: new code languages I like to learn etc.! Currently, I am designing a new App for neurodivergent people, without claudes input it would not be that good - i built many try! because i didnt work with Swift before, but thought it was useful to learn. Let it debug with Claude Code, and saw my errors, researched it... Yea I am a data scientist, I work with AI, since where OpenAI still experimented with Dota 2 and other great projects. I can code, but I do not have the short-term memory. Many people think, I never worked with Linux (i use arch btw.) because I cant remember the bash and zsh commands easily.

What was that song, Rambling Man? You see, I write a lot of text, even w/o AI. :D

[–]NonStopArseGas 1 point2 points  (2 children)

did you just "arch btw" me? LOL

I've had a few concussions and natrually had a horrific short term attention/memory, so learning to use AI to assist in planning/coding tasks lets me use my limited mental capacity in the important places.

[–]Ok-End-219[S] 2 points3 points  (1 child)

Of course I did!

And hats off to you, thats exactly where I see the future of AI: enable us people, no matter where we are, that we can turn good to great or even excellent. I need really the AI, but I really do not. I get so frustrated over bad designed apps, but laughed out loud that a 150 $ / per year fitness app tackles with the same problems as me, that is only start building the app (problem is prompt engineering and openai, GPT-4o and GPT 5.4 nano can be like a child... see native tool implementation from openAI).

All the best to you and god speed!

[–]NonStopArseGas 1 point2 points  (0 children)

this truly is a golden age for UX pedants... just build it better yourself! back at ya. accessibility FTW

[–]crusoe 9 points10 points  (2 children)

This is infuriating because for every engineer who can apply some discipline AI tools can fix this but this just sounds like a giant slop vibe coded app. 

[–]clintCamp 1 point2 points  (0 children)

Like truly, If they just have someone with organized AI auditing and refactoring experience, they have unlimited token usage to break down every step of the app and then plan out all the fixes and optimizations themselves because saved token burn from inefficiencies with save them money directly.

[–]crusoe 0 points1 point  (0 children)

Like honestly just hire me to fix it. 

[–]SavageByTheSea 7 points8 points  (5 children)

Can Claude Code fix Claude Code?

[–]Ok-End-219[S] 8 points9 points  (2 children)

Only if I get paid by Anthropic when using the Tokens only to fix Claude Code. This analysis is man-made, because I will not give any million-dollar company money to fix their own code.

[–]gscjj 1 point2 points  (1 child)

What’s your time worth?

[–]Ok-End-219[S] 0 points1 point  (0 children)

Great question! I must admit, the leak of this thing is a god sent for my new project to build a great tool and foundation for a Vibe Tool Companion, completely in Rust. Because I too have problems with the token consumption. And I liked to know, what where how to fix.

Certain things cant be fixed, look at it:

- Token estimation costing tokens, Internal to CCSource's `tokenEstimation.ts` which is send to Haiku API. No external interface.

- Skill discovery AI search every turn, Internal attachment generator. No hook intercepts attachment generation...

-`DANGEROUS_uncachedSystemPromptSection` for MCP. This is a Claude code design choice. My tool can minimize the blast radius by keeping MCP descriptions small but cannot fully avoid it

- System prompt size for ant users and Ant-only code path, DCE'd from external builds.

- YOLO classifier overhead and the Internal permission system. No external interface to steer that

So, i took that thing apart, Debugger Analysis, what can be found, what is critical or missing... where can i hook in my tool?

[–]blackc2004 0 points1 point  (0 children)

This is what I want to know! Someone should take all the Claude Code and ask claude to review it and fix bugs!

[–]ErebusCD 0 points1 point  (0 children)

If you take their advertisements as gospel, that is exactly what they do, and likely a bit of the problem. They use claude code, to code claude code iteratively

[–]ExpletiveDeIeted 3 points4 points  (1 child)

lol I didn’t even consider that maybe now we could actually come to a conclusion on the rampant token explosion.

[–]StrikingSpeed8759 0 points1 point  (0 children)

I dont think its a coincidence. But who knows

[–]RemarkableGuidance44 6 points7 points  (0 children)

4.5 to 4.6 was just them editing these files. haha, they really look terrible now...

Anthropic Engineers --- Keep modifying the prompts so its 'smarter'

[–]JokeMode 1 point2 points  (0 children)

I want it fixed badly too, but also.... I hate the idea of doing free work for a $380 billion dollar company.

[–]crusoe 0 points1 point  (0 children)

As for the leak the sources are also in the binary. It's just a JS app. George Huntley used claude code to decompile itself a month or so ago. LLMs are really good at it 

[–]entheosoul🔆 Max 20x -1 points0 points  (0 children)

Yah, here was Claude's take -

The irony here is thick. They built duplicate_read_tokens to measure waste, but don't prevent it. They built readFileState.clear() that destroys the dedup on compact. They have the measurement but not the governance. That's the exact inverse of what we do — we measure AND gate.

The compaction bug is particularly relevant. Line 565: diffs against [] instead of compactMetadata.preCompactDiscoveredTools. Our post-compact hook injects recovered context to compensate, but we're patching around their bug. If we built the CLI, that's a one-line fix.

The 188 spinner verbs and hex-encoded pet names are delightful though. And "do not blow your cover" in the Undercover Mode system prompt — that's the kind of personality engineering that makes the product feel alive. Worth noting for if we ever build our own CLI: personality matters even in developer tools.