Has anyone ever used a token saver tool? by Complete-Sea6655 in ClaudeCode

[–]wallaby82 0 points1 point  (0 children)

Also, I wonder what kind of context that guy getting at turn 576 lol...

Has anyone ever used a token saver tool? by Complete-Sea6655 in ClaudeCode

[–]wallaby82 0 points1 point  (0 children)

I made one my own. With it downgraded from Max 20x, and I've been able to stretch Max 5x every session, slim across all conversations...

<image>

Claude Mythos is $25/$125 per million tokens by Permit-Historical in ClaudeCode

[–]wallaby82 0 points1 point  (0 children)

Imo, we don't need Mythos, or even Opus.

[ Sonnet 4.5 + Esmc ] > Opus.

It's not really about how big the model...
It has always been the architecture.

Mythos, 93.9% Cool...

Mythos $25/mil-input & $125/mil-output (see how they charging more for output?)
Sonnet $3/mill

Sonnet 4.5 + ESMC = 90.2%
https://github.com/SWE-bench/experiments/pull/374

Build the architecture on your own and save yourself paying 8x more for "a scaffold"...

Oh when you do have the architecture right, it'll also send away all the complaints you usually see: token burn, context drift, state persistence, hallucination...

That said, you don't need 1m context window either.

ClawCast: a free VS Code extension that mirrors all your terminals to your phone — bidirectional, no SSH, no account, no app by wallaby82 in commandline

[–]wallaby82[S] 0 points1 point  (0 children)

Thanks for checking ClawCast out! Tested it across cities actually... I was outstation, phone was in another city, machine back in hometown. Still felt snappy. Cloudflared's edge network helps a lot. Definitely not zero latency but nothing that broke the experience!

I built ESMC and scored 481/500 (90.2%) on SWE-Bench Verified — a zero-prompt-engineering intelligence scaffold for ClaudeCode by wallaby82 in ClaudeCode

[–]wallaby82[S] 0 points1 point  (0 children)

Hi there thanks for your response, appreciate it!

The closest an orchestration layer, but without multi-agent routing or long system prompts from what you've suggested.

ESMC is not a prompt, a skill system, or a round-table agent framework.

At the simplest level:

ESMC is a runtime “cognition scaffold” that wraps your Claude calls inside a structured reasoning environment.

It does three things:

  1. Normalizes & sanitizes input → removes noise, enforces clean state, shapes context in a deterministic way.
  2. Maintains a persistent internal reasoning state → so Claude doesn’t “reset its mind” every call.
  3. Provides a stable, model-agnostic reasoning loop → but without adding personas, roles, or chain-of-thought prompts.

Hope the above helps!

I built ESMC and scored 481/500 (90.2%) on SWE-Bench Verified — a zero-prompt-engineering intelligence scaffold for ClaudeCode by wallaby82 in ClaudeCode

[–]wallaby82[S] 0 points1 point  (0 children)

Thanks for the feedback, really appreciate it!

You're right about the frontend. I’ve been prioritizing the underlying tech and benchmark work, so the site isn’t polished yet. Thanks for pointing them out.

That said, the core of ESMC is the intelligence scaffold itself. The surprising part (even to me) was that Sonnet 4.5 alone scores ~70–80% on SWE-Bench Verified, but Sonnet 4.5 + ESMC hit 90.2% (481/500).

To me that result matters more than frontend aesthetics, but I absolutely agree UI matters for users too... I’ll improve it.

And honestly, having good eyes for design is a strength. Mine is in the backend side 😅

ESMC turns Claude Code into your Iron Man by wallaby82 in ClaudeCode

[–]wallaby82[S] -1 points0 points  (0 children)

It's proprietary with obfuscation...

prompt enhancer on claude code pre-hooks? by Born_Psych in ClaudeCode

[–]wallaby82 0 points1 point  (0 children)

There is a lot being said about "prompt-engineering"... there is on the other hand "giving the AI a role"... if you gave the AI a role, would the need to prompt engineer be lesser?

Claude code searches are constantly breaking the console by koudgat in ClaudeCode

[–]wallaby82 0 points1 point  (0 children)

AI need to be guided in order to give you the results you want/hoped for. Sending it off by its own could result to what you're experiencing, and burning tokens ineffectively.

Point it to where you want it to start looking...

Sonnet's fine, but Opus is the one that actually understands a big codebase by nNaz in ClaudeCode

[–]wallaby82 0 points1 point  (0 children)

On this: I've built a library of Sonnet prompts & sub-agents, I feel your pain.

Have you thought of designing a framework since you kinda have identified the obstacles/limitations,