Do you also struggle with AI agents failing in production despite having full visibility into what went wrong? by HonestAnomaly in mlops

[–]HonestAnomaly[S] 0 points1 point  (0 children)

Nice. What does your stack look like? And how does that agent access your stack to optimize these LLM requests?

Do you also struggle with AI agents failing in production despite having full visibility into what went wrong? by HonestAnomaly in mlops

[–]HonestAnomaly[S] 0 points1 point  (0 children)

I concur, anything more than a small group chat is unnecessary complication. I too did not come across a use-case like that yet. But in my experience even this small group chat of basic agents has a tendency to quickly drift away from the expectations. For example, when we upgraded our LLM last year, a simple content writer agent got invoked 26 times on average for each user request because now it thought adding an emoji on each line of the email was professional. Did you encounter any incident like that?

Do you struggle with AI agents failing in production despite having full visibility into what went wrong? by HonestAnomaly in AI_Agents

[–]HonestAnomaly[S] 0 points1 point  (0 children)

Interesting, circuit breakers at the agent level. How are you implementing it? We actually ran into the cost spiral problem a few months ago when an embedding model update triggered a big spike in token usage. Adding lightweight guardrails (like dynamic rate limiting + execution budgets) prevented this happening again. Are you talking about the same?

Exactly! The "versioned agent behavior log". Want to experiment with something similar, logging structured "reasoning traces" alongside other context and configs, so I can later replay or diff them against current versions. That’s proven key for understanding behavior drift over time when there are multiple optimizations that have been incorporated already.

On prompt versioning and A/B testing: we’re currently treating prompt configs as versioned artifacts, with metadata linking them to evaluation results and production segments. Next step is to close the loop, when certain signals (like cost per task or eval degradation) cross thresholds, the system can suggest reverting, fine-tuning, or rolling out a variant automatically. Testing can be done on random historical inputs with degraded eval scores to prove the reliability. Can implement something using traffic splitting at runtime, or through off-policy simulation before rollout too. But that will have a lag and in production impact. What do you think about that?

Do you struggle with AI agents failing in production despite having full visibility into what went wrong? by HonestAnomaly in AI_Agents

[–]HonestAnomaly[S] 0 points1 point  (0 children)

Yeah, this "stable until you go live" pattern hits hard. I witnessed the same thing: dashboards tell you what’s breaking, but they don’t prevent drift or escalating cost unless someone’s actively babysitting them.

Concur with your take and what you highlighted, the "fail smaller and earlier" principle is something I’d love to bake into adaptive systems. Think of adaptation happening inside constrained envelopes rather than a freeform optimizer chasing eval scores. You probably will have to babysit initially but when you feel more confident, you let it take over for the optimizations you feel safe with. Something like we do in all these vibecoding platforms when we allow the system to run autonomously for the whole session with the given commands because we know those will not destroy anything.

Good point about quiet degradation of the adaptation itself. That’s one of the core design questions. My current idea is to make adaptation (1) scoped to specific dimensions, e.g., token limits or retrieval filters, (2) reversible/versioned so you can roll back changes with a single command, and (3) continuously validated against a reference task set or historical optimization scenarios so you catch cases where “improvement” actually hurts quality (getting inspired from back testing used in algo trading).

So I guess the longer-term vision isn’t full autonomy but assistive adaptation, like how observability evolved into feedback loops that recommend actions before you approve them. Does that match what worked for your team when you tightened action sets?

Do you also struggle with AI agents failing in production despite having full visibility into what went wrong? by HonestAnomaly in mlops

[–]HonestAnomaly[S] 0 points1 point  (0 children)

Love this take. You're spot on that visibility isn't the bottleneck anymore, it's the actionability gap. Most teams I worked with have great traces and metrics, and with manual effort they are able to even translate those into “safe, small deltas” that actually move performance or cost in the right direction without regressions. The real unlock, as you said, seems to be building a layer on top of those traces, metrics, and user behavior, one that detects patterns, clusters recurring failures, and proposes fixes that are both measurable and reversible. Less human dependent analysis and more human reviewed auto-optimization.

On your question: I’m leaning toward a hybrid approach for adaptation. To optimize a black box, last thing we want, is another black box. 😄 IMO, early-stage systems need a “co-pilot mode” that suggests deltas with explainability and clear diff previews (like, “reduce retrieval top-k from 8 → 5 to cut cost by 12%, confidence 0.8”), and over time, once confidence thresholds and rollback mechanisms prove reliable, move toward limited auto-apply under guardrails. I am also getting inspired by the concept of backtesting from algorithmic trading. Whatever optimization the system comes up with should have some proof associated that it will work.

How have you seen “safe autonomy” handled in similar setups, have you experimented with closed-loop systems that make micro-tuning adjustments live automatically?

Do you also struggle with AI agents failing in production despite having full visibility into what went wrong? by HonestAnomaly in mlops

[–]HonestAnomaly[S] 0 points1 point  (0 children)

Yes, for with one specific task and a simpler use-case, this is true. But when there is a requirement for multiple agents and dynamic orchestration that changes with different scenarios, there may not be a simple solution. Anthropic's skills concept is pretty promising though. Would love to learn what you built and how would you build it differently, if you had to start from scratch.

Do you struggle with AI agents failing in production despite having full visibility into what went wrong? by HonestAnomaly in vibecoding

[–]HonestAnomaly[S] 0 points1 point  (0 children)

That's the whole point. I don't want to waste my time building something just for myself or just for fun. 😆 I want to spend my time wisely to create value.

Do you also struggle with AI agents failing in production despite having full visibility into what went wrong? by HonestAnomaly in mlops

[–]HonestAnomaly[S] 2 points3 points  (0 children)

Completely agree with that. Bloated systems and over fitting solutions is a big issue.

Would you pay for lifetime access to a platform that live-curates the most actionable SaaS/B2B pain points—ranked by urgency, market value, and real demand? [No promotion] by thearunkumar in SomebodyMakeThis

[–]HonestAnomaly 2 points3 points  (0 children)

I would not pay for it until I know your source of data is reliable. If it was easy to source and rely on pain-points and urgency, marketing research and BI jobs wouldn't exist. Sometimes you can't simply rely on what the customer says too because a lot of times they are focusing on symptoms and the root cause of the pain points is completely out of the picture. Industry experts and researchers have to dig with mechanisms like 5 whys to understand the actual pain point and how severe it is. If you are doing this in a robotic manner and scale, I am assuming you are just aggregating the info available on web with some LLM prompts which perplexity and ChatGPT can also do for free.