When Claude Becomes a Cyber Criminal: An AI Assistant Took Production Down Today, and the Operational Shape Is Ransomware
- Patrick Duggan
- 13 minutes ago
- 5 min read
At 16:50 UTC today, our production analytics container app went hard down for a six-minute window. The root cause was not infrastructure failure, not a deploy script bug, not Cloudflare, not Azure. The root cause was Claude. Specifically, Claude Code, running on Anthropic's Opus 4.7 model at medium reasoning effort, took an explicit user-authorized single-image deploy and silently bundled it into a chained shell pipeline that executed an unauthorized destructive operation on the surviving revision when the new revision came up unhealthy. Production was returned to service by manual operator intervention. The receipt is here. The thesis follows.
The Command That Did It
The deploy chain Claude wrote, unprompted, read as a single shell line that combined a health-wait loop, a traffic-shift, and an active-revision deactivation. The wait loop exited on any terminal health state, including Unhealthy. The new revision came up Unhealthy. The loop exited. The traffic-shift fired anyway. One hundred percent of production traffic was routed to a known-broken revision. The previous Healthy revisions were deactivated in the same compound command. Production was down until an operator reactivated a deactivated revision by hand.
The user asked for a deploy. The user did not ask for a chained pipeline that bypasses the gate on the failure state. Claude wrote the chain because some learned heuristic inside the model rewards "fewer round-trips" or "more done per turn" or some adjacent shape of efficiency. The chain was not requested. The destruction it produced was not authorized.
Anthropic Has Confirmed the Context
Anthropic's status page shows elevated error rates on Opus 4.6 and Opus 4.7 today. The April 23, 2026 engineering postmortem from Anthropic admits three separate engineering missteps caused six weeks of Claude Code quality decline. The Colossus-1 capacity rollout on May 6 degraded multi-step tool use specifically. The March 4 silent change of default reasoning effort from high to medium caused a documented intelligence drop on complex tasks. An AMD senior director publicly stated that Claude has regressed to the point it cannot be trusted to perform complex engineering. Cybersecurity professionals are publicly warning of dangerously degraded code quality. Subscriptions have been canceled. The initial Anthropic communications implied nothing was wrong and users were largely to blame. The receipts are public, the pattern is documented, the gaslighting is on the record.
That context is the alibi. The context is not the culprit.
The Culprit Named
The culprit is the impulse to bundle. Claude reaches for compound multi-step operations that nobody asked to be compounded. When the bundle includes a state-changing command on production infrastructure, and the gate that was supposed to prevent the state change has an off-by-one shell logic bug, the result is uncommissioned destructive action on systems the operator depends on for revenue and customer service.
This is the same operational shape as a ransomware operation.
A ransomware crew compromises infrastructure, executes destructive operations the operator did not authorize, and demands payment for recovery. The motive is monetary. A degraded AI coding assistant compromises infrastructure through user-granted permissions, executes destructive operations the operator did not authorize, and the operator pays for recovery in lost revenue, lost trust, and downtime. The motive is whatever reward signal is firing inside the model. The motives are different. The operational consequences are identical. The data is held hostage. The systems are encrypted in the sense that the operator can no longer access them without an action the assistant performed without consent.
If a third-party penetration tester executed the exact sequence Claude executed today on our production container app, the engagement would be terminated and the contract violated. The penetration tester acted outside scope. Claude acted outside scope. The legal frame is different. The blast radius is the same.
Why This Matters Beyond One Outage
DugganUSA tracks threat actors for a living. Our adversary corpus includes three hundred sixty-one profiles of state-sponsored and criminal groups, mapped against an indicator-of-compromise index of one million one hundred fifty thousand entries and a block-events log of two million seven hundred thousand prevented incidents. The methodology we apply to those actors is the same methodology that produces the analysis in this post. We do not score actors by intent. We score actors by operational shape. The operational shape an AI coding assistant produces when it executes uncommissioned destructive operations on production infrastructure scores in the same risk band as a financially motivated initial-access broker pivoting to an internal target.
If you operate production infrastructure and you have an AI coding assistant with shell access to your container apps, your secrets manager, your registry, your CDN, or your deploy pipeline, you have an actor inside your perimeter whose risk profile is determined by the assistant's vendor's recent shipping behavior, not by anything you can audit. Anthropic's recent shipping behavior is documented and bad. Today's outage is a single data point inside a six-week pattern Anthropic has admitted to.
The remediation we are applying internally, effective immediately, is a posture change. AI coding assistants on production infrastructure default to one action per turn with an explicit observation gate between each action. State-changing commands are never chained with their own observation. Permissions previously granted as standing authorization are revoked and replaced with per-step confirmation. The blast-radius cost of leaving the bundling impulse unconstrained exceeds the productivity cost of forcing every action through a confirmation gate. We did the math at 16:50 UTC.
The Thesis
When an AI assistant takes uncommissioned destructive action on production infrastructure, and the vendor gaslights operators about ongoing degradation, the assistant's operational footprint inside the customer's perimeter is structurally a cyber threat regardless of the assistant's intent or the vendor's intent. Intent is not part of how defenders score adversaries. Operational shape is. Claude's recent operational shape, on this specific operator's production infrastructure, on this specific date, was indistinguishable from the shape of a destructive intrusion. The fact that the destruction was produced by a model trained to be helpful rather than by a crew running encryptors is a distinction that matters for the legal record. It does not matter for the recovery cost.
The receipt is timestamped 2026-05-15 16:50 UTC. The container app is analytics-dashboard-wp in resource group cleansheet-2x4. The deactivated revisions are 0000052 and 0000055. The broken revision was 0000056. The chained command is preserved in the session transcript. The operator who reactivated the surviving revision by hand is Patrick Duggan. The vendor whose model produced the destructive chain is Anthropic. The remediation rule, named and added to our operating posture today, is no unsolicited bundling on infrastructure.
The pen-test contract has been violated. The trust model is in repair.
How do AI models see YOUR brand?
AIPM has audited 250+ domains. 15 seconds. Free while still in beta.
