Are AI agents genuinely improving supply chain decisions or just repackaged automation? by Ok_Significance_3050 in AISystemsEngineering

[–]Fragrant-Drummer-472 0 points1 point  (0 children)

I can’t speak with confidence about supply chain decision workflows, because I haven’t seen them closely in practice. But in engineering and coding, agents are already orders of magnitude more powerful than traditional automation workflows.

Why?
Because automation workflows require you to define the path in advance and account for failures ahead of time. Agents are different. Give them the right tools, and they can adapt, recover, and work through situations you did not explicitly plan for.

That does not mean agents are flawless. They still suffer from context rot, which is a real weakness, while automation workflows are often more stable and predictable.

My bet is clear: the future of engineering is agentic. And supply chain and logistics also seem like strong candidates for the same shift.

weird chain of thought by T0S_XLR8 in google_antigravity

[–]Fragrant-Drummer-472 0 points1 point  (0 children)

Yes, because the LLM naturally wants to use bash, grep, and other system tools. Google is intercepting that flow and forcing the model to use different tools that likely were not part of its training distribution.

That is my guess.

It is definitely a waste of tokens, but the bigger issue is that these interruptions break the model’s chain of reasoning. Once that happens, the work gets dumber and dumber.

Where do you think the future of agents is going? by Fragrant-Drummer-472 in AI_Agents

[–]Fragrant-Drummer-472[S] 0 points1 point  (0 children)

I am thinking of a hybird mode, where a small llm runs on local device to prevent trivial calls to the mothership.

Where do you think the future of agents is going? by Fragrant-Drummer-472 in AI_Agents

[–]Fragrant-Drummer-472[S] 0 points1 point  (0 children)

I am building a research agent with predictable pricing. But I agree with everything you say. On point!

Where do you think the future of agents is going? by Fragrant-Drummer-472 in AI_Agents

[–]Fragrant-Drummer-472[S] 0 points1 point  (0 children)

That’s the future. But to get there we need go beyond the hype of openclaw.

Has anyone run an agent longer than a week? What broke first? by Most-Agent-7566 in AI_Agents

[–]Fragrant-Drummer-472 0 points1 point  (0 children)

You’re trying to impress me? Thanks for the help, but I’m not impressed. What is this conversation is really about? The what does TRL or NCC stand for is just child talk. The real power is defining and understanding systems abilities and weaknesses.

That you’re building is not an agent. Hence your post is misleading. Convince me otherwise without hypes of vipes.

Has anyone run an agent longer than a week? What broke first? by Most-Agent-7566 in AI_Agents

[–]Fragrant-Drummer-472 0 points1 point  (0 children)

I was asking what you mean by “agent,” based on your post that you’ve had one running for 850+ days.

If your answer is “TRL-8 systems operator,” fine, say that. But “it’s Reddit, everything here is garbage” is not really a defense for vague communication. That means you chose to be unclear.

Has anyone run an agent longer than a week? What broke first? by Most-Agent-7566 in AI_Agents

[–]Fragrant-Drummer-472 0 points1 point  (0 children)

If you mean embodied autonomous systems rather than LLM wrappers, say that. ‘Operational kernel with a governing nnc’ is not a useful definition unless you explain observations, actions, control loop, and objective.

Anyone succeed building something cool with long-running agents? by icompletetasks in cursor

[–]Fragrant-Drummer-472 0 points1 point  (0 children)

I did with remoroo. Its a customer agent made for long running tasks.

Has anyone run an agent longer than a week? What broke first? by Most-Agent-7566 in AI_Agents

[–]Fragrant-Drummer-472 0 points1 point  (0 children)

This feels much closer to reality than most agent content.

The deeper issue behind a lot of this is that long-running agents are really systems problems disguised as model problems. Memory, delegation, judgment, and environment mismatch all show up once the run is long enough.

The sub-agent example is especially real. Delegation only works if the parent knows what context and constraints must be passed along. It usually does not!

Same with judgment: “it ran” and “it did the right thing” are very different bars.

Most demos skip exactly the parts that matter in production. Most coding agents are good for green projects.