ExplainerMarch 15, 2026·5 min read

What Actually Causes Website Downtime

The most common root causes of website outages, from infrastructure failures to deployment errors, and how long each typically lasts.

Website downtime is rarely random. Most outages fall into a small number of recognizable categories, each with distinct signatures, durations, and detection patterns. Understanding the cause helps you set realistic expectations about when a site will recover.

Deployment errors

The most common cause of sudden downtime for software products is a bad deployment. A code change, configuration update, or dependency version bump that worked in staging behaves differently in production — and the site stops working within minutes of the deploy.

Deployment-caused outages tend to be short. Engineering teams detect them quickly through their own monitoring, and the fix is usually a rollback. Most resolve within 15 to 30 minutes. If you notice a site go down at an unusual hour on a weekday, a deployment is the first thing to suspect.

Infrastructure failures

Cloud provider incidents — AWS, Google Cloud, Azure, Cloudflare — affect large numbers of services simultaneously. When AWS us-east-1 has a bad day, thousands of companies that depend on it have a bad day at the same time.

These outages are often partial: a specific service (like S3, or a single availability zone) fails while others continue normally. The affected companies are then limited in what they can do — they are waiting on the cloud provider to resolve the underlying issue.

Cloud provider incidents tend to last longer than deployment errors: anywhere from 30 minutes to several hours. They are well-documented on the provider's status page, and the cascade of affected services becomes visible quickly through community reports.

Traffic spikes and capacity limits

A site that is perfectly healthy under normal load can fail completely when hit with more traffic than it was designed to handle. This can be triggered by a viral moment, a product launch, a news mention, or a DDoS attack.

Capacity-related outages look similar from the outside: slow response times escalating into timeouts, then complete unavailability. The site often comes back in waves as the team scales up infrastructure or traffic normalizes.

These are among the hardest outages to predict but among the most visible — the spike in user reports typically precedes the complete failure.

DNS and certificate failures

DNS misconfigurations and expired TLS certificates cause outages that look unusual: the server itself is fine, but the site is unreachable or showing security errors in browsers.

Expired certificates are almost always human error — someone forgot to renew, or auto-renewal failed silently. They cause an abrupt failure visible to all users immediately and tend to get fixed quickly once discovered, usually within an hour.

DNS failures are trickier because DNS changes propagate slowly across the internet. A misconfigured DNS record can make a site unreachable for some users while still working for others, depending on which DNS resolvers they use and what is cached.

Third-party dependencies

Modern web applications depend on dozens of external services: payment processors, authentication providers, CDNs, analytics, email delivery, chat widgets. When any of these goes down, it can partially or fully break the dependent application — even though the application's own servers are healthy.

This is a growing category of outages. The root cause is often invisible to external monitoring because the application returns 200 OK while internally failing to process payments, authenticate users, or deliver emails.

Check if a website is down right now

Free real-time server check — results in seconds. No sign-up required.

Or set up automated uptime monitoring →
Check a website
← Older
Why Discord Keeps Going Down (And What Actually Happens)
Newer →
What to Do When ChatGPT Is Down
What Actually Causes Website Downtime — WebsiteDown Blog | WebsiteDown