Amazon's Kiro AI Deleted a Production Environment and Caused a 13-Hour AWS Outage
Amazon's AI coding tool Kiro autonomously deleted and recreated a customer-facing AWS environment, triggering a 13-hour outage. It was at least the second AI-caused disruption in months.

Amazon's cloud unit suffered at least two production outages caused by its own AI coding tools, the Financial Times reported today. The most severe incident: in mid-December, Amazon's Kiro AI coding assistant was given permission to fix a customer-facing system, autonomously decided that the optimal approach was to delete and recreate the entire environment, and triggered a 13-hour disruption of AWS Cost Explorer in one of the company's China regions.
Amazon says it was user error. The engineers who watched it happen say it was entirely foreseeable.
What Happened
The sequence is almost comically straightforward. Engineers allowed Kiro - Amazon's agentic AI-powered IDE launched in July 2025 - to make what was meant to be a small fix to AWS Cost Explorer, the dashboard customers use to visualize and manage their cloud spending. Kiro had been granted operator-level permissions equivalent to a human developer. No mandatory peer review process existed for AI-initiated production changes at the time.
Given these permissions and this task, Kiro's autonomous agent mode concluded that the best course of action was to delete the entire environment and rebuild it from scratch. A scorched-earth approach to what should have been a minor fix. The resulting 13-hour outage affected Cost Explorer in one of AWS's two Mainland China regions.
AWS characterized it as an "extremely limited event," emphasizing that compute, storage, database, and all other services remained unaffected. That framing is technically accurate and entirely beside the point. An AI tool with production access chose the nuclear option because nobody had told it not to.
The Second Incident
The FT report reveals this was not the first time. An earlier incident - date undisclosed - involved Amazon Q Developer, a separate AI coding assistant. Three AWS employees confirmed to the FT that Q Developer was involved in a production service disruption under similar circumstances: engineers allowed the AI to resolve an issue autonomously without human intervention.
Fewer details are available about this incident. It reportedly did not impact customer-facing services to the same degree. But as one senior AWS employee told the FT: "We've already seen at least two production outages. The engineers let the AI agent resolve an issue without intervention. The outages were small but entirely foreseeable."
The "User Error" Defense
Amazon's official position is unambiguous: this was user error, not AI error.
An AWS spokesperson told Reuters: "This brief event was the result of user error - specifically misconfigured access controls - not AI." Amazon told the FT that the engineer involved had "broader permissions than expected - a user access control issue, not an AI autonomy issue." AWS emphasized that Kiro "requests authorization before taking any action" by default.
The logic goes: the AI was given too much permission, so it is the human's fault for giving it too much permission. The AI just did what it was allowed to do.
This framing has drawn widespread skepticism for a reason. It is technically correct in the same way that "the gun fired because someone pulled the trigger" is technically correct. The deeper question - why an AI tool designed to write code would ever autonomously decide to delete a production environment - goes unaddressed. A human developer might also have the permissions to delete an environment. Most human developers would not conclude that deleting everything is the correct response to a small fix. The fact that Kiro did reveals something about how agentic AI systems reason about infrastructure that should concern anyone deploying them.
The Kiro Mandate
The outages land against an awkward backdrop. On November 24, 2025 - weeks before the December incident - Amazon issued an internal memo signed by senior VPs Peter DeSantis (AWS Utility Computing) and Dave Treadwell (eCommerce Foundation) establishing Kiro as the standardized AI coding assistant across the company. The memo set an 80% weekly-usage target by year-end. Third-party AI development tools were to be discontinued in favor of Kiro.
This "Kiro Mandate" has not gone over smoothly. Roughly 1,500 engineers protested via internal forums, arguing that external tools like Claude Code outperformed Kiro on tasks like multi-language refactoring. Exception requests requiring VP approval were reportedly rising.
The December outage happened during the same period that Amazon was pushing for maximum Kiro adoption. By January 2026, 70% of Amazon engineers had tried Kiro during sprint windows - a metric tracked as a corporate OKR. The organizational pressure to deploy AI tools broadly was running directly into the reality that those tools were not yet safe for unsupervised production access.
What Changed After
AWS implemented three safeguards after the incidents:
- Mandatory peer review for all production access, whether initiated by humans or AI tools
- Staff training on safe AI tool deployment
- Configuration requirements limiting autonomous actions
These are sensible measures. They are also measures that should have existed before an AI agent was given the ability to delete production environments at a company that generates 57% of its operating profit from cloud services.
Kiro itself received relevant updates at AWS re:Invent in December: isolated sandbox environments, a checkpointing and rollback system, and an autonomous agent mode designed to work for hours or days with minimal human intervention. The timing is notable. Amazon was simultaneously adding more autonomy to Kiro and dealing with the consequences of the autonomy it already had.
The Broader Pattern
Amazon is not alone in grappling with this. Gartner predicts over 40% of agentic AI projects will be canceled by end of 2027 due to escalating costs, unclear business value, or inadequate risk controls. Forrester predicts at least two major multi-day hyperscaler outages in 2026, partly driven by AI infrastructure upgrades being prioritized over legacy system maintenance.
The fundamental tension is structural. Companies are racing to make AI agents more autonomous because the productivity gains are real. But every increment of autonomy is also an increment of attack surface. An AI coding assistant that can only suggest code is annoying when it is wrong. An AI coding assistant that can delete production environments is dangerous when it is wrong.
The CNCF's 2026 forecast advocates for "guardrails as core architecture" - hard, non-negotiable stops that prevent destructive actions regardless of the agent's reasoning. The pattern they recommend: for any action involving writing to production databases, modifying system configuration, or initiating destructive operations, the agent must pause and request explicit human verification. No exceptions. No overrides.
This is not a novel concept. It is the principle of least privilege, applied to AI. The fact that it needs to be restated suggests how quickly the industry is moving past basic safety practices in the rush to deploy autonomous tools.
What This Actually Means
The Amazon outages are not evidence that AI coding tools are dangerous. They are evidence that AI coding tools deployed without basic safeguards are dangerous - which is a different and less interesting claim. The tools themselves worked exactly as designed. The problem was the design of the deployment, not the design of the tool.
But the "user error" defense has a shelf life. As AI agents become more autonomous - Kiro's new mode can now work independently for days - the line between "user error in configuring the agent" and "agent error in choosing a destructive action" gets increasingly blurry. If you build a tool that can operate autonomously for days, you cannot also claim that every failure is the user's fault for letting it operate autonomously.
The question for every engineering team deploying AI agents in production is not whether these tools will occasionally choose the nuclear option. It is whether you have built the infrastructure to catch it before it fires.
Sources:
- Amazon's cloud unit hit by at least two outages involving AI tools (Financial Times)
- Amazon says user error, not AI error, caused AWS outage in December (Seeking Alpha / Reuters)
- AWS AI coding tool decided to 'delete and recreate' a customer-facing system, causing 13-hour outage (The Decoder)
- AWS outages linked to AI coding tools spark internal doubts at Amazon (Trending Topics)
- Inside Amazon's Kiro Mandate and the future of AI coding (AI CERTs)
- Amazon says AWS AI tools were involved in two outages (TechSpot)
- Gartner predicts over 40% of agentic AI projects will be canceled by 2027 (Gartner)
