GitHub Confirms Two Major Outages, Vows Urgent Fixes
GitHub suffered two critical availability incidents in rapid succession, the company acknowledged today, attributing the failures to an unprecedented surge in AI-powered development workflows. “These incidents are not acceptable, and we are sorry for the impact they had on you,” a GitHub spokesperson said. The platform now faces an emergency effort to scale infrastructure by 30 times current capacity.

Background: Exponential Growth Fueled by Agentic Development
The root cause, GitHub explained, is a dramatic shift in how software is built. Since December 2025, agentic development workflows—automated coding agents that push changes at machine speed—have accelerated sharply. Repository creation, pull requests, API calls, and large-repo workloads are all growing exponentially.
This growth doesn’t stress one system at a time. A single pull request can touch Git storage, mergeability checks, branch protection, Actions, search, notifications, permissions, webhooks, APIs, background jobs, caches, and databases. At high scale, small inefficiencies compound: queues deepen, cache misses turn into database load, indexes fall behind, retries amplify traffic, and one slow dependency cascades across dozens of product experiences.
GitHub’s Response: From 10X to 30X Capacity in Months
GitHub had already started a plan to increase capacity by 10X in October 2025, aiming for better reliability and failover. But by February 2026, it became clear that target was too low. “We needed to design for a future that requires 30X today’s scale,” the company said. The main driver: the rapid adoption of agentic development since late December.
Short-Term Actions Taken
- Bottleneck elimination: Moved webhooks off MySQL to a new backend, redesigned session caching, and reworked authentication and authorization flows to slash database load.
- Compute ramp-up: Leveraged the ongoing migration to Azure to rapidly provision additional compute resources.
- Service isolation: Began isolating critical services like Git and GitHub Actions from other workloads to minimize blast radius and single points of failure.
- Language migration: Accelerated moving performance-critical code from the Ruby monolith into Go.
Long-Term Strategy: Multi-Cloud and Graceful Degradation
While migrating out of smaller custom data centers into public cloud was already underway, GitHub is now pushing toward a multi-cloud architecture. The company stated its priorities: “Availability first, then capacity, then new features.” Teams are reducing unnecessary work, improving caching, isolating critical services, and removing single points of failure. The goal is to make GitHub degrade gracefully when one subsystem is under pressure.

What This Means for Developers
For the millions of developers who rely on GitHub daily, these outages disrupt CI/CD pipelines, code reviews, and collaboration. The company’s emergency scaling plan is a direct response to the new reality of AI-driven development where machines, not humans, generate the bulk of repository activity. Expect continued instability in the short term as GitHub reworks its foundational infrastructure, but the promised 30X capacity should eventually restore reliability—if the company executes quickly enough.
“This is distributed systems work: reducing hidden coupling, limiting blast radius, and making GitHub degrade gracefully,” the spokesperson said. The message is clear: GitHub is racing to rebuild itself for a future where code is written at machine speed, and any downtime ripples through the global software ecosystem.