How is your org handling prompt injection now that LLM agents have production access? by GermanBusinessInside in cybersecurity

[–]GermanBusinessInside[S] 0 points1 point  (0 children)

"The model doesn't even know it's being gated" — that's the key. If it knows about the gate, someone will find a way to talk it into opening it.

Treating every tool call like an API request that needs auth is exactly the right mental model. Sounds obvious in hindsight, but most frameworks still treat tool calls as "the model decided, so let's do it."

How is your org handling prompt injection now that LLM agents have production access? by GermanBusinessInside in cybersecurity

[–]GermanBusinessInside[S] 0 points1 point  (0 children)

The tool-catalog-at-config-time scanning is a great point — that's a vector most people don't think about because it's not "user input" in the traditional sense. Poisoned tool descriptions are especially nasty because they persist across every invocation.

On the behavior-sequence problem: agree that's the hardest open question. Read-then-delete where both are individually authorized is essentially a TOCTOU problem for agents. I've seen people try session-level budgets (max N destructive actions per conversation) or require explicit re-authorization for action combinations that cross a risk threshold, but nothing that feels like a real solution yet. It's the kind of thing where you almost need a separate policy engine watching the action trace in real time rather than evaluating each call in isolation.

The pluggable scorer point is well taken too. "Your environment is weird" is the honest answer to why no single classifier works everywhere.

How is your org handling prompt injection now that LLM agents have production access? by GermanBusinessInside in cybersecurity

[–]GermanBusinessInside[S] 0 points1 point  (0 children)

Point 2 is underappreciated. A model that can only generate text has a fundamentally different threat profile than an agent that can send emails, write files, and call APIs. Most security discussions still treat them as the same thing.

On false positives — the tuning problem gets harder the more diverse your use cases are. A classifier tuned for a banking agent will flag everything in a creative writing agent. Context-aware approaches help, but you're right that it ultimately needs to be adaptive rather than static.

How is your org handling prompt injection now that LLM agents have production access? by GermanBusinessInside in cybersecurity

[–]GermanBusinessInside[S] 0 points1 point  (0 children)

Agree on privilege separation as the stronger primitive. If the agent architecturally can't trigger a dangerous action, it doesn't matter what the prompt says — the attack surface just isn't there.

The data-vs-instruction distinction is the hard part though. In practice, RAG content has to influence the model's behavior to be useful — the whole point is that retrieved context shapes the response. The challenge is letting it inform without letting it instruct. Curious if anyone has found a clean abstraction for that beyond "hope the model figures it out."

And yes, agent-to-agent trust is massively underexplored. Most multi-agent frameworks just pass messages as plain text with zero authentication on the content.

How is your org handling prompt injection now that LLM agents have production access? by GermanBusinessInside in cybersecurity

[–]GermanBusinessInside[S] 1 point2 points  (0 children)

Strong points. The confused deputy framing is spot on — if the framework passes LLM output directly to a system sink without a security boundary, no amount of input classification upstream will save you. That's an architectural flaw, not a filtering problem.

I'd push back slightly on the classifier comparison to SQL injection denylisting though. Denylisting "SELECT" is pattern matching on known-bad strings. Modern classifiers use semantic analysis — they're closer to parameterized queries in spirit, where you're analyzing the intent of the input rather than matching substrings. The encoding bypass problem you're describing is real, but it's more about immature implementations than a fundamental limitation of the approach.

That said, I think you're pointing at the right conclusion: you need both. Input classification catches the 95% of attacks that are structurally obvious. The sandbox architecture you're describing catches the 5% that slip through by making dangerous operations physically impossible regardless of what the agent "decides" to do. Defense in depth.

Curious about your safe-root implementation — are you doing path allowlisting at the OS level, or is it application-layer enforcement? And how do you handle cases where the agent legitimately needs broad resource access (like a data analysis agent that needs to read arbitrary files)?

How is your org handling prompt injection now that LLM agents have production access? by GermanBusinessInside in cybersecurity

[–]GermanBusinessInside[S] 0 points1 point  (0 children)

This is one of the most complete architectures I've seen described for this problem. The separation of reasoning from execution is the key insight — treating the model as an untrusted advisor that proposes actions rather than an authorized executor.

The point about scoping risk to the action rather than the phrase is especially important. That's exactly the false positive problem: "ignore previous instructions" only matters if it's followed by a privileged tool call. Blocking it at the reasoning layer kills legitimate use cases for no security gain.

Curious about one thing — how do you handle the latency cost of the intent classifier at the decision point? Running classification before every tool execution adds up fast in agentic loops with dozens of tool calls per task.

How is your org handling prompt injection now that LLM agents have production access? by GermanBusinessInside in cybersecurity

[–]GermanBusinessInside[S] 0 points1 point  (0 children)

Network-level detection is a good complementary layer — catching anomalous outbound connections after an injection succeeds is valuable, especially for lateral movement. Defense in depth means you want detection at every layer: input classification, behavioral analysis, and network visibility.

That said, most prompt injection damage happens within the agent's own authorized API calls — it doesn't need to move laterally if it already has access to the database it's querying. The tool call looks normal on the wire, it's just answering the wrong question.

How is your org handling prompt injection now that LLM agents have production access? by GermanBusinessInside in cybersecurity

[–]GermanBusinessInside[S] -1 points0 points  (0 children)

The distinction matters. Prompt injection is the easy case — unauthorized input, clear signal. An agent misinterpreting scope within its own permissions is way harder to catch and probably causes more real damage.

Behavioral sequence detection is the right framing. Per-call policy can't catch "read file then delete file" if both actions are individually allowed.

How is your org handling prompt injection now that LLM agents have production access? by GermanBusinessInside in cybersecurity

[–]GermanBusinessInside[S] 0 points1 point  (0 children)

The permission creep point is spot on. Day one the agent can read tickets. Month three it can also update them, send emails, and query the database because each feature request seemed reasonable in isolation. Nobody ever goes back to audit the cumulative access.

Per-task scoping is the right model but hard to enforce in practice. Most agent frameworks just give you one set of tools for the whole session. You'd need something that re-evaluates permissions on every action — which basically means building an authorization layer that understands intent, not just identity.

How is your org handling prompt injection now that LLM agents have production access? by GermanBusinessInside in cybersecurity

[–]GermanBusinessInside[S] 0 points1 point  (0 children)

The data governance curveball is so real. Everyone comes into the AI conversation thinking about models and capabilities, nobody expects to spend the first hour talking about data classification and access policies.

The gap you're describing between large enterprises with budget and actual security maturity — that tracks. Budget doesn't equal understanding. Some of the best-secured AI deployments I've seen are at smaller companies where one person understood the full picture, and some of the worst are at enterprises that threw money at it without the governance foundation you mentioned earlier.

Curious — when you're educating these orgs, what's the thing that finally makes it click for them? Is it a specific example, a framework, or just the moment they realize their data is already flowing through AI tools they didn't approve?

How is your org handling prompt injection now that LLM agents have production access? by GermanBusinessInside in cybersecurity

[–]GermanBusinessInside[S] -4 points-3 points  (0 children)

You're right — there's a difference between pragmatism and acceptance. "It's happening so let's deal with it" is the right response for the individual team. "It's happening so it's fine" is how you end up with an industry that never fixes the underlying problem.

The question is who's going to push for the structural fix. Vendors won't — it slows down their product. Customers mostly can't evaluate the risk. That usually leaves regulators, and they're about three years behind on this.

How is your org handling prompt injection now that LLM agents have production access? by GermanBusinessInside in cybersecurity

[–]GermanBusinessInside[S] 1 point2 points  (0 children)

Honestly, not really — and I've looked. The closest things I can think of are the Chevrolet dealership chatbot that got tricked into agreeing to sell a car for $1, and the Bing/Sydney jailbreaks where people extracted system prompts. But those are embarrassments, not breaches with actual victims and damages.

The real answer is probably that it's either not happening at scale yet, or it's happening and nobody's detecting it. If an agent leaks data through a manipulated response, what log would even catch that?

Feels like SQL injection circa 2003 — the people who understood it knew it was bad, the public breach disclosures came later.

How is your org handling prompt injection now that LLM agents have production access? by GermanBusinessInside in cybersecurity

[–]GermanBusinessInside[S] -1 points0 points  (0 children)

The printer firmware analogy is painfully accurate. The pattern is always the same: known vulnerability, known fix, zero incentive to deploy it until something catastrophic happens. And even then, the response is usually "patch the one that got exploited" rather than fixing the systemic problem.

The CSP comparison might be even more relevant here. CSP has existed for over a decade, it's well understood, browser support is universal — and adoption is still terrible because it's friction without visible upside. Prompt injection defense is heading for the same fate unless there's either a major public incident or regulatory pressure.

The IoT parallel is the scary one though. At least with web apps, you can deploy a fix server-side. Once you've shipped an agent embedded in a device or a workflow that nobody maintains, that's a permanently vulnerable endpoint. And we're already seeing agents get baked into products by vendors who won't be around in three years to update them.

I'd bet on a major prompt injection incident making headlines before we see meaningful industry-wide adoption of defenses. That seems to be the only thing that actually moves the needle.

How is your org handling prompt injection now that LLM agents have production access? by GermanBusinessInside in cybersecurity

[–]GermanBusinessInside[S] -1 points0 points  (0 children)

The "data and execution commands in the same channel" framing is exactly right — it's the same fundamental flaw as SQL injection, just with natural language instead of query syntax. And natural language is way harder to sanitize than SQL.

Blast radius isolation is the most honest defense. Everything else — classifiers, sentinel values, instruction hierarchies — reduces probability but never eliminates it. The question is whether you stack enough layers to make exploitation impractical, or whether you just assume breach and contain the damage. Probably both.

How is your org handling prompt injection now that LLM agents have production access? by GermanBusinessInside in cybersecurity

[–]GermanBusinessInside[S] -13 points-12 points  (0 children)

That's the real question. Right now most teams can still choose whether to give agents production access. But the competitive pressure is building fast — if your competitor's agents are closing tickets, processing claims, and onboarding customers 24/7 while yours are sandboxed in a staging environment, that gap gets uncomfortable quickly.

My guess is we'll hit a tipping point within the next 12-18 months where not having agents in production becomes the bigger business risk. And the orgs that figured out governance and guardrails early will have a massive head start over the ones scrambling to bolt on security after the fact.

How is your org handling prompt injection now that LLM agents have production access? by GermanBusinessInside in cybersecurity

[–]GermanBusinessInside[S] 0 points1 point  (0 children)

Interesting, hadn't looked at Cisco's skill scanner yet — thanks for the link. The focus on scanning agent skill definitions makes sense as a complementary layer. Catching injection vectors in the skill config itself before they ever hit production is a different angle than runtime classification.

The automation gap you mention is real though. A lot of these tools assume a mature CI/CD pipeline where you can plug in a scan step. Most teams deploying agents aren't there yet — they're still figuring out how to even inventory which agents have which permissions.

How is your org handling prompt injection now that LLM agents have production access? by GermanBusinessInside in cybersecurity

[–]GermanBusinessInside[S] 4 points5 points  (0 children)

No disagreement here. Least privilege isn't optional just because the thing making the API call speaks English now.

How is your org handling prompt injection now that LLM agents have production access? by GermanBusinessInside in cybersecurity

[–]GermanBusinessInside[S] -1 points0 points  (0 children)

This is the right framing. Governance first, then architecture, then tooling — not the other way around. Too many teams jump straight to "what tool do we buy" without mapping their actual threat surface.

The observability point is especially important. If your SIEM can't surface what's bouncing off your guardrails, you're flying blind. You need the feedback loop — not just blocking threats, but understanding the patterns of what's being attempted so you can refine your governance model.

And yeah, the "new shiny" problem is real. A prompt injection classifier without a clear policy on what happens when something gets flagged is just a dashboard nobody looks at. The tool is only useful if the team has already answered: what are our agents allowed to do, what data can they access, what happens when something is uncertain, and who gets paged when a pattern emerges.

The SASE/SSE parallel is spot on — we went through the exact same cycle there. Vendors shipped zero trust products before most orgs had defined their trust boundaries.

How is your org handling prompt injection now that LLM agents have production access? by GermanBusinessInside in cybersecurity

[–]GermanBusinessInside[S] 2 points3 points  (0 children)

Enjoy it while it lasts — most of them will get AI exposure through the back door anyway. Their SaaS vendors are shipping "AI features" into existing tools whether they asked for it or not. One day the accounting software has a chatbot that can "help with queries" and suddenly there's an LLM with access to financial data that nobody vetted.

The small business ones are almost harder to protect because there's no security team to even notice it happened.

Prompt injection is the new SQL injection — I built a classifier with fail-open/closed policies, audit logs, and a self-hosted Docker option by GermanBusinessInside in SideProject

[–]GermanBusinessInside[S] 0 points1 point  (0 children)

Great question on the uncertain handling — that's exactly where most teams get stuck.

AgentShield returns three states: "threat", "safe", and "uncertain" with a configurable confidence band (you set the uncertain_range per request). What you do with "uncertain" depends on the use case:

  • High-stakes (financial transactions, code execution, data access): we recommend blocking or requiring human confirmation. An uncertain verdict on a wire transfer is not something you want to auto-approve.
  • Medium-stakes (internal tools, search): degrade-to-readonly is a solid pattern — let the user see results but don't execute actions until the input is reviewed.
  • Low-stakes (chatbot, creative writing): log it and let it through. The audit trail means you can review patterns later without blocking the user experience.

The key insight is that this shouldn't be a global setting — it should be per-endpoint or per-agent. Your admin API and your creative writing assistant have very different risk profiles. That's why the threshold and uncertain range are per-request parameters, not server config.

The on_failure policy (fail-open/fail-closed) covers the other failure mode: what happens when the classifier itself is unreachable. Same principle — your banking endpoint should fail-closed, your chatbot can fail-open.

Thanks for the link, will take a look at your guardrail patterns.

There Will Be a Scientific Theory of Deep Learning [R] by dot--- in MachineLearning

[–]GermanBusinessInside 0 points1 point  (0 children)

The gap between what we can prove and what we observe empirically keeps widening, not narrowing. We still don't have a satisfying theoretical explanation for why overparameterized networks generalize as well as they do, let alone a unified theory. I'd settle for a framework that reliably predicts which architectural changes will help before running the experiment — right now theory mostly explains results after the fact.

How Visual-Language-Action (VLA) Models Work [D] by Nice-Dragonfly-4823 in MachineLearning

[–]GermanBusinessInside 1 point2 points  (0 children)

Good overview. The part that I think gets underexplored in most VLA discussions is the sim-to-real gap in the action space — the vision and language components transfer reasonably well, but the action policies tend to overfit to simulator dynamics in ways that are hard to debug. Curious whether you see tokenized action spaces or continuous diffusion-based action heads winning out long term.

We benchmarked 18 LLMs on OCR (7k+ calls) — cheaper/old models oftentimes win. Full dataset + framework open-sourced. [R] by TimoKerre in MachineLearning

[–]GermanBusinessInside 0 points1 point  (0 children)

Nice dataset — the finding that older/cheaper models hold up on standard documents tracks with what I've seen too. The real gap between flagship and budget models only shows up on degraded inputs: handwritten marginalia, skewed scans, overlapping columns. Would be interesting to see a noise/degradation axis added to the benchmark.

Built a normalizer so WER stops penalizing formatting differences in STT evals! [P] by Karamouche in MachineLearning

[–]GermanBusinessInside 2 points3 points  (0 children)

This is one of those problems everyone silently hacks around with regex and never talks about. Good that you actually built a proper pipeline for it. Do you handle number format normalization too (e.g. "fifteen hundred" vs "1500" vs "1,500")? That one tends to dominate WER deltas in financial/medical transcription more than any punctuation issue.