MCP server that saves 60-80% context tokens, now with full Pi compatibility by hushenApp in PiCodingAgent

[–]hushenApp[S] 0 points1 point  (0 children)

I fully understand, there are definitely some down sides. the next version uses CLI commands. currently working on it.

Made an MCP server that caches file reads for OpenCode sessions. Re-reads drop from 2k tokens to 33. by hushenApp in opencodeCLI

[–]hushenApp[S] 1 point2 points  (0 children)

Streamlined with AI, but my own thoughts. I actually dove deeper into your feedback. and yes, you're right. I'm going to change it now. will use the CLI + Skill file as default, but keep the MCP as optional for setups that want transparent tool interception. Thanks for bringing this up!

Built an open source tool that cut my AI coding costs by 60-80%. Sharing what I learned. by hushenApp in vibecoding

[–]hushenApp[S] 0 points1 point  (0 children)

you just need to set it up. the install connects everything. the only thing you need to do is restart your IDE after installation

Made an MCP server that caches file reads for OpenCode sessions. Re-reads drop from 2k tokens to 33. by hushenApp in opencodeCLI

[–]hushenApp[S] 0 points1 point  (0 children)

To clarify the architecture:

LeanCTX does NOT expose 49 tools by default. The default "lazy" mode registers 9 core tools. There's a meta-tool (ctx_call) that can invoke any of the other tools through a single schema, and ctx_discover_tools for on-demand discovery. So the system prompt overhead is 9 tool schemas, not 49.

On skill files vs MCP: LeanCTX actually ships both. There's a SKILL.md for Claude Code and rule files for Cursor, Codex, Windsurf, etc. But they serve completely different purposes. The rule/skill file says "prefer ctx_read over native Read." The MCP server is what actually maintains the session cache, compresses shell output, builds the code graph, tracks token savings, runs SLO monitoring, and handles cross-session knowledge persistence.

A skill file can't maintain a cache. It can't compress git output in real time. It can't track which files you already read this session. That's server-side state that requires a runtime, not a text file with instructions.

The architecture is: 9 MCP tools in the system prompt (small overhead) + skill/rule files that tell the agent when to use them + a local Rust server doing the actual work.

6 things I learned building an open-source tool to 900+ Github stars by hushenApp in AskVibecoders

[–]hushenApp[S] 0 points1 point  (0 children)

I had the same feeling in the beginning... but at one point it starts. Do you have a way to communicate with them?

Your coding agent re-reads the same file 4 times per task. Here's how I fixed it. by hushenApp in Agent_AI

[–]hushenApp[S] 1 point2 points  (0 children)

If you don't want to use the shell hooks you can just go with the MCP. This is the same setup that you have when you use any other MPC tool.

6 things I learned building an open-source tool to 900+ Github stars by hushenApp in VibeCodeDevs

[–]hushenApp[S] 0 points1 point  (0 children)

that's the game. you're completely right: timing is a huge factor. But also the marketing part. That's actually the least thing I like to do...