TL;DR
Agentic AI is outpacing both our vocabulary and our security models. The 80/20 approach to risk coverage works, until you consider that outliers are exactly where failures tend to originate. And the training data feeding these systems isn't representative of the real world yet, with a feedback loop that makes the gap harder to close over time. The good news is that none of these problems are unsolvable. But they do require intention.
The Gap Is Still Closable (For Now)
Not every AI governance conversation is worth your time. This one is. In Episode 3 of Curiouser & Curiouser, host Mo Sadek sits down with Laura Powell, Senior Director of Partnerships at LatticeFlow AI, to talk about something most teams avoid: the gap between knowing AI risk exists and actually doing something about it. Laura has spent years inside that gap, building privacy and responsible AI programs at large tech organizations, and her perspective is less about theory and more about what happens when nobody acts until something breaks.
The Incident That Hasn't Happened Yet
Here's an uncomfortable truth most people in governance already know: without a real incident to point to, the argument for controls doesn't get funded. People nod, ask what the bare minimum is, and move on. That's not cynicism. It's just how organizations work.
The problem is that by the time the incident happens, it's too late to retrofit the controls. Like when a car dealership chatbot got talked into selling a brand new SUV for one dollar, and the dealership had to honor it. Somewhere a developer looked at that scenario and decided it was an edge case not worth handling. That's the pattern. The systems being deployed today are orders of magnitude more complex, and the edge cases are not getting smaller.
Agentic AI: Complexity Before Clarity
That edge case problem gets significantly harder when you consider where AI is heading. The industry still doesn't have consistent language for what agentic systems actually are. A single LLM with tool access, a chain of models passing outputs to each other, a fully autonomous pipeline making decisions without human review. All of these get called "agentic AI." If you can't define what you're building, you can't govern it.
The security problem runs deeper than terminology though. In a multi-agent workflow, each agent starts its own chain of reasoning based on what it received from the previous one. By the time you're looking at a final output, you don't know which step introduced the variance or why. And the real issue is this: we don't have a handle on single LLM systems yet. Chaining them together and handing them autonomy before we do isn't a sequencing problem. It's a compounding one. Failure modes need to be engineered for before deployment, not discovered after something goes wrong.
The 80/20 Problem Has a Dangerous Edge
Even teams that take this seriously tend to default to an 80/20 approach: focus effort where it matters most and accept that you can't anticipate everything. That logic makes sense, until you factor in how these systems actually work. They're built on pattern recognition, and outliers get treated as noise. The edge cases that don't fit the pattern are exactly the ones the system is least equipped to handle, and often the ones most likely to trigger something catastrophic.
So the question isn't just how to cover 80% of risk efficiently. It's whether anyone has thought about what the system does when it encounters something entirely outside its training. Building in a rejection or escalation behavior for those moments isn't an optional refinement. It's the part most teams skip entirely, and it's where the compounding starts.
The Data Problem Is Already in Production
That compounding doesn't just happen in the architecture. It's baked into the data these systems are trained on. The systems being built today don't reflect the real world. They reflect the parts of the world that are already connected, already online, already well-represented. And because of how these systems learn, that skew doesn't stay flat over time.
Take food distribution. A grocery chain using ML to predict inventory needs will optimize for historical sales data and revenue. The result is that underserved areas with already limited access to fresh food keep getting the same products, not because anyone decided that was acceptable, but because the system was never asked to optimize for anything else. The system is working exactly as designed. The design is the problem.
Healthcare is where the stakes get highest. Models built on non-representative data make worse predictions for the patients who already have the fewest options. That's not a future scenario to prepare for. It's already happening, and the longer the feedback loop runs without correction, the harder it becomes to close the gap.
What to Take Away
None of these problems are waiting to arrive. They're already compounding, in the architecture of agentic systems, in the coverage gaps of risk frameworks, and in the training data feeding models that affect real people's lives. The analogues exist in privacy, in security, in every previous wave of technology adoption where governance came after the damage was done.
The window to act differently this time is still open. It won't stay that way indefinitely.
Stay curious.
What’s New from Alice
Building AI We Can Actually Trust
AI is moving from experimentation into production, and with that shift comes a harder question: how do we actually build systems people can trust? In this episode of Curiouser & Curiouser, Mo sits down with Laura Powell, Senior Director of Partnerships at LatticeFlow AI, to talk about what that actually requires. They cover why agentic AI is outpacing the frameworks meant to govern it, where the 80/20 approach to risk breaks down, and what biased training data is already doing in production today.
Distilling LLMs into Efficient Transformers for Real-World AI
This technical webinar explores how we distilled the world knowledge of a large language model into a compact, high-performing transformer—balancing safety, latency, and scale. Learn how we combine LLM-based annotations and weight distillation to power real-world AI safety.
Ungated Download Test
This resource gives you the clarity to ask the right questions, pressure-test your current approach, and take meaningful action before your customers or regulators do it for you. Download it now.

