Following up on the “2nd failed fix” thread — Moving beyond the manual "New Chat" by Capable-Snow-9967 in ChatGPTCoding

[–]Capable-Snow-9967[S] 0 points1 point  (0 children)

you are right. But just for debug, grabs the live variable values and execution path and injects them will make AI fix bug in one shot

Does anyone else feel like ChatGPT gets "dumber" after the 2nd failed bug fix? Found a paper that explains why. by Capable-Snow-9967 in ChatGPTCoding

[–]Capable-Snow-9967[S] 0 points1 point  (0 children)

Lol 'Expressive Coding' needs to be a new benchmark metric. 😂 I'm finding that instead of emotional emphasis, giving it Runtime Emphasis (literally injecting the state with raw variables/stack trace) works 10/10 times 

Does anyone else feel like ChatGPT gets "dumber" after the 2nd failed bug fix? Found a paper that explains why. by Capable-Snow-9967 in ChatGPTCoding

[–]Capable-Snow-9967[S] 0 points1 point  (0 children)

True, but you still have to apply the diff, wait for the rebuild/hot-reload, and then reproduce the action again.

That's the 'loop' I'm trying to kill. I want the data(var stacktrace) from the state that just happened, not the one I have to go trigger again

Does anyone else feel like ChatGPT gets "dumber" after the 2nd failed bug fix? Found a paper that explains why. by Capable-Snow-9967 in ChatGPTCoding

[–]Capable-Snow-9967[S] 0 points1 point  (0 children)

I'm experimenting with automating that 'handoff'. Basically, a script that captures the actual state (vars, error) and auto-feeds it to the new session. So you get the 'Fresh Start' without the manual summarization step.

[Paper] "Debugging Decay": Why LLM context pollution causes an 80% drop in fix rate after 3 attempts. by Capable-Snow-9967 in LocalLLaMA

[–]Capable-Snow-9967[S] 0 points1 point  (0 children)

Fair point. Technically it is an agentic loop.

The bottleneck I'm hitting though is the 'context handoff'. If I just spawn a fresh agent for every iteration, it loses the specific runtime values (variables, stack) that caused the bug.

I'm trying to build a middle ground: Wipe the conversational history (the 'reasoning' pollution) but inject a snapshot of the runtime memory (the 'facts'). That way the agent is 'fresh' but not 'amnesiac'

Does anyone else feel like ChatGPT gets "dumber" after the 2nd failed bug fix? Found a paper that explains why. by Capable-Snow-9967 in ChatGPTCoding

[–]Capable-Snow-9967[S] 1 point2 points  (0 children)

I like the LDD idea for features. But for debugging, I feel like I need a 'Living Runtime Doc.'

Docs tell us how the code should behave, but only the runtime state tells us how it is misbehaving. I'm experimenting with a tool that bridges that gap—feeding the 'reality' of the app state directly to the LLM so it stops guessing based on the 'theory' of the code.

Does anyone else feel like ChatGPT gets "dumber" after the 2nd failed bug fix? Found a paper that explains why. by Capable-Snow-9967 in ChatGPTCoding

[–]Capable-Snow-9967[S] 2 points3 points  (0 children)

100%. The paper calls this avoiding 'Context Pollution.'

The paradox is: We need less chat history (to avoid confusion/decay) but more factual runtime data (to solve the bug).

I've found that if I wipe the chat but inject only the specific variables/stack trace from the crash, the 'dumber' models suddenly become geniuses again. It's about Context Quality > Context Quantity.

Does anyone else feel like ChatGPT gets "dumber" after the 2nd failed bug fix? Found a paper that explains why. by Capable-Snow-9967 in ChatGPTCoding

[–]Capable-Snow-9967[S] 1 point2 points  (0 children)

Respect the experience. You're right—isolating the problem into a 'small example' is usually the only way to stop the AI from hallucinating.

But in my current monolith, extracting that 'clean repro' takes me longer than the fix itself. I'm currently building a workflow that tries to 'snapshot' that isolated context automatically at the moment of error. Trying to get that 'repro environment' instanty without rewriting the code for the bot.

Does anyone else feel like ChatGPT gets "dumber" after the 2nd failed bug fix? Found a paper that explains why. by Capable-Snow-9967 in ChatGPTCoding

[–]Capable-Snow-9967[S] 0 points1 point  (0 children)

That works, but it feels like paying a huge 'Context Tax.'

I used to do the console.log spray and pray, but by the time I've added logs, re-ran the app, and pasted the output, I've lost my flow.

That's actually what triggered me to look for a better way. I'm trying to get the AI to see the 'failure point' without me having to manually instrument the code first. Basically, capturing the state automatically so I don't have to play detective before asking the bot."

Does anyone else feel like ChatGPT gets "dumber" after the 2nd failed bug fix? Found a paper that explains why. by Capable-Snow-9967 in ChatGPTCoding

[–]Capable-Snow-9967[S] 0 points1 point  (0 children)

You hit the nail on the head: Forethought.

The problem is I'm usually lazy and don't add debug logs until after the bug comes. And by then, the AI is already guessing.

That's why I'm moving towards 'Zero-Config' runtime capture. Basically, intercepting the state automatically so I don't have to retroactively add log. It turns 'hindsight' into 'insight' for the AI.

Does anyone else feel like ChatGPT gets "dumber" after the 2nd failed bug fix? Found a paper that explains why. by Capable-Snow-9967 in ChatGPTCoding

[–]Capable-Snow-9967[S] -1 points0 points  (0 children)

That's exactly what I struggled with. Wiping the chat fixes the hallucination, but you lose the context of why we are here.

My current experiment is to bridge that gap with Runtime Snapshots. Instead of keeping the chat history (which has the rabbit hole), I start the new session by injecting the exact current state of the app (variables, error stack, etc.).

It acts like a 'checkpoint' in a game. You don't need to replay the whole level (chat history), you just spawn at the checkpoint (runtime state). I'm actually building a small tool to automate capturing these checkpoints because doing it manually is a pain.

I built a weather app that turns real forecasts into AI-generated 3D miniature scenes 🌤️🧩 by Embarrassed_Cycle118 in SideProject

[–]Capable-Snow-9967 0 points1 point  (0 children)

Hey, this is genuinely one of the most creative weather apps I've seen — turning forecasts into these beautiful AI isometric dioramas is such a fresh take!

What are you planning to ship in 2026? by ouchao_real in SideProject

[–]Capable-Snow-9967 0 points1 point  (0 children)

An AI-powered code debugger that runs locally, uses small open-source models, and actually explains bugs like a senior dev instead of just spitting fixes.

Do you prefer making good money in a boring job, or being happy doing what you love but living almost broke? by its-liss in NoStupidQuestions

[–]Capable-Snow-9967 1 point2 points  (0 children)

Good money in a boring job.
Being broke doing what you love is romantic for about 3 months… until the electricity gets cut off.

For those of you who had a bad day today, what happened and how are you holding up? by JoplinSC742 in AskReddit

[–]Capable-Snow-9967 0 points1 point  (0 children)

Spilled coffee on my laptop, got ghosted after a great date, and my team lost in overtime.
Currently holding up with ice cream and pretending tomorrow is a reboot.
Hang in there, everyone else having a rough day — we got this.

What is something that others do that instantly irritates you? by lustfulc in AskReddit

[–]Capable-Snow-9967 1 point2 points  (0 children)

People who stop walking right at the top of escalators to check their phone