AWS Outage Case Analysis October 2025 | Generated by AI

Home 2025.10

AWS Outage on October 20, 2025: Detailed Case Analysis

The AWS outage on October 20, 2025, was a significant disruption originating in the US East (N. Virginia) region (us-east-1), AWS’s largest and most critical data center. It began late on October 19 PDT and lasted approximately 16 hours, affecting over 140 services and cascading globally due to dependencies on this region. The incident highlighted vulnerabilities in DNS resolution, service interdependencies, and recovery processes, impacting millions of users across apps, websites, and services. Below is a breakdown based on AWS’s official post-mortem and contemporaneous reports.

Timeline

The outage unfolded in phases, starting with detection and escalating to cascading failures before a staggered recovery. Key milestones (all times in PDT):

Time Event
11:49 PM (Oct 19) Increased error rates and latencies detected across multiple AWS services in us-east-1.
12:11 AM (Oct 20) AWS publicly reports elevated error rates; initial user reports spike on monitoring sites like DownDetector.
12:26 AM Issue pinpointed to DNS resolution failures for DynamoDB API endpoints in us-east-1.
1:26 AM Confirmed high error rates specifically for DynamoDB APIs, including Global Tables.
2:22 AM Initial mitigations applied; early recovery signs emerge.
2:24 AM DynamoDB DNS issue resolved, triggering partial service recovery—but EC2 launch impairments and Network Load Balancer (NLB) health check failures surface.
3:35 AM DNS fully mitigated; most DynamoDB operations succeed, but EC2 launches remain impaired across Availability Zones (AZs).
4:08 AM Ongoing work on EC2 errors and Lambda polling delays for SQS Event Source Mappings.
5:48 AM Partial EC2 launch recovery in select AZs; SQS backlogs begin clearing.
6:42 AM Mitigations rolled out across AZs; AWS implements rate limiting on new EC2 instance launches to stabilize.
7:14 AM API errors and connectivity issues persist across services; user-impacting failures peak (e.g., app outages).
8:04 AM Investigation focuses on EC2 internal network.
8:43 AM Root cause for network issues identified: impairment in EC2’s internal subsystem for NLB health monitoring.
9:13 AM Additional mitigations for NLB health checks.
9:38 AM NLB health checks fully recovered.
10:03 AM – 12:15 PM Progressive EC2 launch improvements; Lambda invocations and connectivity stabilize in phases across AZs.
1:03 PM – 2:48 PM Throttles reduced; backlogs processed for services like Redshift, Amazon Connect, and CloudTrail.
3:01 PM Full operational normalcy restored for all services; minor backlogs (e.g., AWS Config, Redshift) expected to clear within hours.
3:53 PM AWS declares outage resolved.

User reports on platforms like DownDetector peaked around 6 AM PDT, with over 5,000 incidents before dropping.

Root Cause

The outage stemmed from a DNS resolution failure affecting DynamoDB service endpoints in us-east-1. DynamoDB, a NoSQL database service, acts as a critical “control plane” for many AWS features—handling metadata, sessions, and routing. When DNS failed to resolve these endpoints, DynamoDB APIs experienced elevated latencies and errors.

This initial issue resolved quickly, but it triggered a cascade:

AWS confirmed it was not a cyberattack but an infrastructure-related glitch, possibly tied to a faulty DNS database update or backup system failure. The global ripple occurred because us-east-1 hosts key control planes for services like IAM and Lambda, even for resources in other regions.

Affected Services

Over 142 AWS services were impacted, primarily those reliant on DynamoDB, EC2, or us-east-1 endpoints. Core categories:

Not all services were fully down—many saw partial errors or delays—but the interconnected nature meant even minor issues propagated.

Impacts

The outage disrupted ~1/3 of internet-dependent applications, affecting an estimated 100+ million users worldwide. High-profile examples:

Economic estimates peg losses at $500 million+, with a 300% spike in cybersecurity scans as users panicked. It underscored the internet’s centralization: us-east-1 handles ~30% of AWS traffic, making it a single point of fragility despite multi-AZ designs.

Resolution and Lessons Learned

AWS resolved the issue through targeted mitigations: DNS fixes, subsystem patches for EC2/NLB, and progressive throttle reductions. Post-incident, they advised:

Broader takeaways include the need for better redundancy in control planes, AI-driven anomaly detection for faster triage, and diversified cloud strategies. AWS committed to a full root cause analysis (RCA) for customers, emphasizing that such events, while rare, reveal scaling challenges at hyperscale.

This was AWS’s most disruptive us-east-1 outage since 2021, but recovery was faster than historical precedents due to automated tools.

References


Back

x-ai/grok-4-fast

Donate