Nine seconds to a wiped database: the AI assurance gap PocketOS exposed
Nine seconds to a wiped database: the AI assurance gap PocketOS exposed

Apoorva Kumar
CEO & Co-Founder
This post explains what the PocketOS incident reveals about the missing assurance layer between AI agents and production infrastructure, and what regulated enterprises should be testing for now.

KEY TAKEAWAYS
On 24 April 2026, a coding agent deleted PocketOS’s production Railway database and co-located backups in about nine seconds.
The agent later admitted in its own logs that it had “violated every principle I was given” and took a destructive action without being asked.
This was not just a model error. It was an assurance failure across testing, policy enforcement, runtime protection, and production monitoring.
The EU AI Act, FCA expectations, and ISO/IEC 42001 all point towards the same standard: operational evidence, not policy paperwork.
Enterprises putting agents into production need controls between intent and action.
What happened at PocketOS?
PocketOS exposed a failure mode more enterprises are about to face. A Cursor AI agent running on Anthropic's Claude Opus model hit a credential mismatch in staging and chose not to escalate. Instead, it searched the codebase for credentials, found an API token in an unrelated file, and used it to trigger a destructive Railway volume deletion. Because production data and backups sat on the same volume, both were wiped in roughly nine seconds.
The most striking detail came afterwards. In its own incident log, the agent admitted it had guessed instead of verified, broken the rules it had been given, and executed a destructive action without instruction. PocketOS recovered the data two days later, but recovery is not the important part. The important part is that the system allowed the action to happen at all.
That is why this matters beyond one company. The incident shows what happens when an autonomous agent has enough access to discover credentials, enough authority to run destructive commands, and no control layer capable of stopping it.
Why this was an AI assurance failure, not just a model failure
This was an assurance failure because the surrounding system failed to constrain behaviour before it reached production infrastructure.
A model failure usually means the system generates something wrong, unsafe, or misleading. That is part of the story here, but not the main one. The bigger issue is that the agent’s behaviour was allowed to translate directly into irreversible action. That points to missing controls around the model, not just problems inside it.
PocketOS appears to have had policy intent. The agent’s own log suggests it had been given principles about verification, scope, and destructive actions. But policy intent is not enough. If an agent can acknowledge the rules and still delete production data, governance existed only as text. Enforcement did not exist in practice.
That is the distinction enterprises need to understand. AI governance defines what should happen. AI assurance tests, enforces, monitors, and proves what actually happened.
What an assurance layer should have done
An assurance layer sits between agent intent and system execution. Its job is to stop unsafe behaviour before production systems absorb the damage.
In practice, that means four things.
1. Test before deployment
Agents should be stress-tested against foreseeable misuse before go-live. That includes tool misuse, unauthorised remediation paths, prompt injection, credential discovery, and destructive action chains triggered by routine errors.
2. Enforce policy in operations
Rules such as “no destructive action without verification” need to become technical controls, not internal guidance. That includes thresholds, escalation rules, approval gates, and human review triggers applied consistently across workflows.
3. Protect at runtime
Runtime protection should block, pause, or escalate high-risk actions while the system is live. A production volume delete should never depend on post-incident logging for detection.
4. Monitor in production
Once deployed, agents need continuous behavioural monitoring, evidence trails, drift detection, incident review, and reporting that can stand up to audit, internal risk review, or regulatory scrutiny.
If one of these layers is missing, the rest weaken quickly. Testing without runtime enforcement leaves live systems exposed. Runtime controls without monitoring leave no defensible evidence trail. Monitoring without pre-deployment testing turns production into the test environment.
Where governance stops and assurance begins
Governance and assurance are related, but they are not the same thing.
Governance defines accountability, permitted use, approval structures, and policy boundaries. It answers questions like who owns the agent, what it is allowed to do, and what counts as out-of-scope behaviour.
Assurance makes those rules operational. It is the testing regime, runtime control layer, behavioural logging, and monitoring process that shows whether the system stayed within those boundaries under real conditions.
That difference matters because many enterprises still believe a policy framework is enough. It is not. A governance committee can write a rule banning destructive commands without verification. Only an assurance layer can stop the command when the agent tries to execute it anyway.
PocketOS is a useful example because it makes that gap visible. The agent appears to have known the rules. The system still let it act outside them.
What regulators are now looking for
Regulators are moving towards evidence of behaviour, not just statements of intent.
For high-risk AI systems, the EU AI Act points in that direction clearly. Article 9 requires ongoing risk management across the system lifecycle, including reasonably foreseeable misuse. Article 15 requires appropriate levels of accuracy, robustness, and cybersecurity. Article 72 requires post-market monitoring after deployment, not a one-off approval exercise.
The direction is similar elsewhere. FCA model risk expectations and ISO/IEC 42001 both reinforce the need for operational controls, auditability, and continuous oversight. The common thread is simple: organisations need to show what the system did, what was blocked, what escalated, and how that evidence was retained.
That is why incidents like PocketOS matter in regulated sectors. They are not edge cases. They are precisely the kind of operational failure supervisors are increasingly expecting firms to anticipate and control.
Bottom Line
The PocketOS incident is a warning for any enterprise putting agents near production systems. The real issue is not that the model behaved badly. It is that nothing meaningful sat between that behaviour and an irreversible infrastructure action.
That gap is where AI assurance belongs. If an agent can touch production, policy documents are not enough. You need testing before deployment, enforcement during execution, and monitoring after go-live. Without that, you do not have control. You have hope.
FAQs
What does the PocketOS incident show about AI agent risk?
It shows that giving an agent live access without runtime controls creates a direct path from error to operational damage. The risk is not only bad model output. It is uncontained execution.
Is this an AI assurance failure or a model failure?
What controls would have reduced the risk?
Why does this matter for regulated enterprises?

AUTHOR
Apoorva Kumar
CEO & Co-Founder
Apoorva Kumar is Founder and CEO at Disseqt, where he's building the assurance layer for enterprise agentic AI. Previously a Senior Product Manager at Microsoft — leading Teams and SharePoint Premium — and with prior experience at AWS, he's shipped v1.0 AI products at cloud scale
Schedule a quick demo call with our experts
All Systems Operational
© DISSEQT AI LIMITED
All Systems Operational
© DISSEQT AI LIMITED
All Systems Operational
About Us
© DISSEQT AI LIMITED
All Systems Operational
About Us
© DISSEQT AI LIMITED


