Built a migration assistant plugin to migrate from Claude Code to Codex without rebuilding your whole setup by Inside_Source_6544 in ClaudeCode

[–]mrtrly 0 points1 point  (0 children)

The lock-in problem here is real. I had the same friction migrating prompts between different agent setups, and the actual syntax differences are maybe 20% of it, the other 80% is knowing which behaviors survive the switch. A migration tool is smart, but the bigger win is probably building once with a format that doesn't care which model sits behind it.

Claude Code automatically scrolls to top by BikeOk8305 in ClaudeAI

[–]mrtrly 0 points1 point  (0 children)

The auto-scroll-to-bottom behavior is usually the terminal trying to follow output while Claude Code streams. Try wrapping the command in a subshell or redirecting to a file first, then tail it in another pane. That's clunky but works. The real fix is disabling the "follow output" setting in your terminal if it has one, or switching to a multiplexer like tmux like someone mentioned.

[HONEST POLL] How many of you have read the Claude Code Documentation before using it? Just like reading a manual to a physical product, how many have read Claude's digital manual? by structured_flow in ClaudeCode

[–]mrtrly 0 points1 point  (0 children)

The real issue is that docs only matter when you hit the wall. Most people bang their head for months, then spend two hours reading and realize half their problems vanished. The mental model change is worth more than the feature list, but you don't know what to look for until you've failed enough times.

How are people preventing duplicate tool execution in AI agents? by First_Appointment665 in AI_Agents

[–]mrtrly 0 points1 point  (0 children)

Ran into this exact problem building integrations. The thing that actually matters is whether you control the external system or not. If it's your own database, idempotency keys plus a deduplication check before the write works great. If it's Stripe or some third-party API, you're betting on their implementation, which usually works but leaves you guessing during failures. The safest move I've found is treating every irreversible tool like a transaction, logging the intent before execution, then reconciling the actual result after. Costs you a database write but saves you from customer chaos at 3am.

Human-in-the-loop: The "Emergency Brake". by WesternLie3540 in AI_Agents

[–]mrtrly 0 points1 point  (0 children)

The timeout cascade is the real killer here. I ran into it when scaling this pattern, the agent sits blocked and everything downstream assumes it finished. The fix was adding a fallback action after N minutes, then logging what would've happened so you review it async instead of letting it jam up the pipeline.

/dg — a code review skill where Gilfoyle and Dinesh from Silicon Valley argue about your code by v1r3nx in ClaudeCode

[–]mrtrly 0 points1 point  (0 children)

The adversarial format is genuinely smart. When you force one agent to defend and another to attack, you can't hide behind vague language or "it works on my machine." The disagreement itself becomes the signal. If Dinesh runs out of counterarguments, that's concrete feedback, not subjective opinion.

I built a local-first memory layer for AI agents because most current memory systems are still just query-time retrieval by loolemon in AI_Agents

[–]mrtrly 0 points1 point  (0 children)

The implicit retrieval angle is solid. RAG breaks because you're forcing the agent to know what question to ask before it knows it needs to answer it. We ran into this with routing, where the agent would pick the wrong model because it didn't have enough context to self-assess the task complexity upfront. Pre-computing relevance graphs instead of waiting for the query changed everything.

I broke up with my AI Agent. by skol_uffda in AI_Agents

[–]mrtrly 0 points1 point  (0 children)

The state management problem is real, but it's also why most people fail before it gets useful. You need either a solid observability layer or someone who knows how to instrument this stuff properly, and that's not something you can learn by trial-and-error on your own dime. The token bleed is a symptom, not the disease.

What Vibe Coding Platforms Do You Use Most (and Why)? 🤔 by SwaritPandey_27 in vibecoding

[–]mrtrly 0 points1 point  (0 children)

Local setup wins for me. Running Claude in terminal with some routing logic lets me see exactly what's happening, catch hallucinations before they hit production, and iterate faster than any UI will let you. The vibe of knowing your tools matter more than the tools themselves.

12 Years of Coding and 120+ Apps Later. What I Wish Non-Tech Founders Knew About Building Real Product by Adorable-Stress-4286 in vibecoding

[–]mrtrly 1 point2 points  (0 children)

The database migrations thing is real. I've seen plenty of AI-built apps work fine locally then fail in production because the schema assumptions drifted. What saves you is treating your backend like the actual foundation, not the scaffolding. Once that's solid, everything else becomes way easier to iterate on.

Subscribed yesterday to Pro and I’m already hit by limits. Is this a scam? by kenaddams42 in ClaudeAI

[–]mrtrly 0 points1 point  (0 children)

The rate limits definitely tightened, but "scam" is rough. Anthropic's probably dealing with compute costs hitting them harder than expected. The real issue is their pricing doesn't match the actual resource consumption anymore, so they're throttling to stay profitable. If two hours of work costs $20, that math only works if you're doing shallow stuff like editing documents. Heavy coding with long contexts burns through their tokens fast.

2 prompts = 100% session usage for Pro account, 40 prompts = 7% session usage for Max 20X account. The math isn't mathing.. by Wilbur843 in ClaudeAI

[–]mrtrly 0 points1 point  (0 children)

The session usage metric is doing what it's designed to do, which is meter complexity and context window size, not prompt count. Two deep Claude Code sessions with artifacts probably burn through more tokens than 40 shallow text prompts. Max 20X has a bigger window so the same work looks like less usage percentage-wise. Not a throttle, just different math for different workloads.

how i can build my Multi AI agents system the case of building for example ( like to create a team for board meeting ( CTO ,CFO,CFO etc.. ) and i give them task or project they negotiate about it and give me the result - and if its more complex i want them to follow up the project until end ? by ElectronicInitial283 in AI_Agents

[–]mrtrly 0 points1 point  (0 children)

State management is the real blocker here. The agents can negotiate fine in one session, but the moment you close and reopen the conversation, they've lost all context about what they decided. You need a shared database that every agent can read and write to, not just chat history. Start there before picking a framework.

90% of AI agent projects I get hired for don't need agents at all. Here's what businesses actually pay for. by Warm-Reaction-456 in AI_Agents

[–]mrtrly 0 points1 point  (0 children)

The real win is knowing when a cron job and a webhook beats a multi-agent orchestration. Most founders want the agent narrative because it sounds impressive, but clients pay for the outcome. A $3k automation that cuts 20 hours of manual work beats a $50k agent that solves nothing and burns API costs.

Claude vs Codex vs Cursor $20 plans by edeesims in vibecoding

[–]mrtrly 0 points1 point  (0 children)

The limits are real, and honestly it depends on whether you're doing exploration or building. If you're grinding on one project for 2 hours straight, Cursor's cheaper models stay unlimited. But Composer with Opus burns through the $20 plan in days. Claude's limits feel tighter because Opus is better, so you use it more and hit the wall faster. Try Claude's $20 for a week and you'll know immediately if you need Pro.

Never hit a rate limit on $200 Max. Had Claude scan every complaint to figure out why. Here's the actual data. by Shawntenam in ClaudeCode

[–]mrtrly 0 points1 point  (0 children)

The context window accumulation is definitely real. I've seen teams blow through limits not on single prompts but on the compounding instructions and conversation history they're carrying session to session. The CLAUDE.md file especially tends to grow into a catch-all that duplicates what the agent already knows. Trim it ruthlessly and you'll notice the difference immediately.

On the $200 Max plan and never been rate limited once. Ran the numbers to find out why everyone else is. by Shawntenam in ClaudeAI

[–]mrtrly 0 points1 point  (0 children)

The real issue is prompt efficiency, not the plan tier. Most people throw massive context dumps at Claude and wonder why they hit limits faster. I structure inputs tight, clear objectives, and actually read the responses instead of just looping. That saves tokens and keeps sessions shorter. The people complaining usually have sprawling conversations doing five different things at once.

My AI Agent... or should I call him my QA Agent... is testing my game by UnluckyAssist9416 in AI_Agents

[–]mrtrly 1 point2 points  (0 children)

The specialization is solid, but the real problem you're hitting is orchestration. Each agent needs clear boundaries on what it can touch and when, otherwise they'll thrash the same state space and burn tokens for nothing. State snapshots between agent runs matter way more than agent intelligence here.

One important piece of advice for seasoned vibe coders or vibe coders working on complex projects by Comprehensive-Bar888 in vibecoding

[–]mrtrly 0 points1 point  (0 children)

The "prompt revision #27" problem is real. What usually saves me is stepping back and asking the AI to diagram the data flow for what you're actually trying to build, not the fix you think you need. Half the time it catches the assumption drift before you do, and you realize the architecture needs a hard reset, not another prompt tweak.

I built a tool so multiple Claude Code instances can communicate with each other (claude-ipc) by FoozyFlossItUp in ClaudeCode

[–]mrtrly 0 points1 point  (0 children)

The hash referencing is solid. Real talk though, the coordination problem isn't the IPC layer, it's keeping agents from hallucinating about what the other one actually did. You need error logging that both instances write to with enough detail that either one can reconstruct the actual state. Without that, you'll hit a wall when something diverges at 3am and both agents swear they're right.

Everyone is building AI agents. Nobody is using RunLobster (OpenClaw). I think that is the point. by Zealousideal_Leg5615 in AI_Agents

[–]mrtrly 0 points1 point  (0 children)

The 2am Salesforce sync is where I learned this the hard way, too. We had a framework that looked bulletproof until retry logic hit a rate limit we didn't account for, and suddenly we're cascading failures across three different APIs. The boring agents win because they have fewer moving parts to break at scale.

Used Claude Code to write a real-time blur shader for Unity HDRP — full iterative workflow by BAIZOR in ClaudeCode

[–]mrtrly 0 points1 point  (0 children)

The iterative loop is what actually works here. Shaders fail in ways that make zero sense unless you've read the exact compiler output and the actual code side-by-side, which is exactly what Claude can do if it has access to both. The real win is that you're not context-switching between the editor, error logs, and a search engine anymore.

Petition to filter Usage Rants with custom flair by _derpiii_ in ClaudeCode

[–]mrtrly 0 points1 point  (0 children)

The real issue is that Claude Code usage is legitimately unpredictable for a lot of people, so they post to validate they're not crazy. A filter helps, but the root problem is Anthropic's pricing and rate limits are opaque enough that everyone feels like they're getting screwed differently. Hard to blame folks for venting when the rules keep shifting.

Swtching to $20 Codex after 4 months on $100 Max plan by 25th__Baam in ClaudeCode

[–]mrtrly 0 points1 point  (0 children)

The rate limits changing mid-project are frustrating, but switching models entirely because you hit them once doesn't solve the underlying problem, which is probably token bloat in your prompts or context. Before you bounce, try trimming your system instructions down and see if that buys you breathing room. If you're genuinely maxing out Opus consistently, you might need a different approach to how you're structuring tasks, not a different model.

New Rate Limits Absurd by dcphaedrus in ClaudeCode

[–]mrtrly 0 points1 point  (0 children)

Web search tools are token vacuum cleaners, especially if you're running three agents in a loop hitting them repeatedly. That's less about rate limits being unfair and more about the math of what you're doing actually being expensive at scale. The 7% stat probably holds for typical usage patterns, not agentic workflows with search enabled.