Saturday, 7 Feb 2026
Subscribe
logo
  • Global
  • AI
  • Cloud Computing
  • Edge Computing
  • Security
  • Investment
  • Sustainability
  • More
    • Colocation
    • Quantum Computing
    • Regulation & Policy
    • Infrastructure
    • Power & Cooling
    • Design
    • Innovations
    • Blog
Font ResizerAa
Data Center NewsData Center News
Search
  • Global
  • AI
  • Cloud Computing
  • Edge Computing
  • Security
  • Investment
  • Sustainability
  • More
    • Colocation
    • Quantum Computing
    • Regulation & Policy
    • Infrastructure
    • Power & Cooling
    • Design
    • Innovations
    • Blog
Have an existing account? Sign In
Follow US
© 2022 Foxiz News Network. Ruby Design Company. All Rights Reserved.
Data Center News > Blog > Global Market > Top 11 network outages and application failures of 2025
Global Market

Top 11 network outages and application failures of 2025

Last updated: February 1, 2026 7:35 am
Published February 1, 2026
Share
Unhappy Programmer Caught In Maze Of Broken Software And Stress.
SHARE

Contents
Asana: February 5 & 6Slack: February 26X: March 10Zoom: April 16Google Cloud: June 12Classes realized in 2025

Asana: February 5 & 6

  • Period: Two consecutive outages, with the second lasting roughly 20 minutes
  • Signs: Service unavailability and degraded efficiency
  • Trigger: A configuration change overloaded server logs on February 5, inflicting servers to restart. A second outage with related traits occurred the next day.
  • Takeaways: “This pair of outages highlights the complexity of contemporary methods and the way it’s troublesome to check for each doable interplay state of affairs,” ThousandEyes reported. Following the incidents, Asana transitioned to staged configuration rollouts.

Slack: February 26

  • Period: 9 hours
  • Signs: Customers might log in and browse channels, however skilled points sending and receiving messages.
  • Trigger: Points with a upkeep motion of their database methods brought on an overload of heavy site visitors directed on the database.
  • Takeaways: “At first look, the whole lot appeared tremendous at Slack—community connectivity was good, there have been no latency points, and no packet loss,” in response to ThousandEyes. Solely by combining a number of diagnostic observations might investigators decide the true supply was the database system, later confirmed by Slack.

X: March 10

  • Period: A number of hours with numerous service downtimes
  • Signs: The platform appeared “down,” with customers experiencing connection failures just like a distributed denial-of-service (DDoS) assault.
  • Trigger: Community failures with vital packet loss and connection errors on the TCP signaling section occurred. “Connection errors usually point out a deeper drawback on the community layer,” in response to ThousandEyes.
  • Takeaways: ThousandEyes detected site visitors being dropped earlier than classes may very well be established. However there have been no seen BGP route modifications, which might usually happen throughout DDoS mitigation. “It was a network-level failure, however not what it might have first appeared,” ThousandEyes famous.
See also  2025 global network outage report and internet health check

Zoom: April 16

  • Period: Roughly two hours
  • Signs: All Zoom providers have been unavailable globally.
  • Trigger: Zoom’s title server (NS) data disappeared from the top-level area (TLD) nameservers, making the service unreachable regardless of wholesome infrastructure.
  • Takeaways: “Though the servers themselves have been wholesome all through and have been answering appropriately when queried straight, the DNS resolvers couldn’t discover them due to the lacking data,” ThousandEyes reported. The incident highlights how failures above a company’s Area Identify System (DNS) layer can utterly knock out providers.
  • Period: Greater than two hours
  • Signs: The appliance’s front-end loaded usually, however tracks and movies wouldn’t play correctly.
  • Trigger: Backend service points whereas community connectivity, DNS, and CDN “all appeared wholesome.”
  • Takeaways: “The very important indicators have been all good: connectivity, DNS, and CDN all appeared wholesome,” in response to ThousandEyes, which added that the incident illustrated how “server-side failures can quietly cripple core performance whereas giving the looks that the whole lot is working usually.”

Google Cloud: June 12

  • Period: Greater than two and a half hours
  • Signs: Customers couldn’t use Google to authenticate on third-party apps equivalent to Spotify and Fitbit; knock-on penalties impacted Cloudflare providers and downstream purposes.
  • Trigger: An invalid automated replace disrupted the corporate’s identification and entry administration (IAM) system.
  • Takeaways: “What you had was a three-tier cascade: Google’s failure led to Cloudflare issues, which affected downstream purposes counting on Cloudflare,” ThousandEyes defined, including that the incident is a “reminder to hint a fault all the way in which again to supply.”
  • Period: Multiple hour
  • Signs: Site visitors couldn’t attain quite a few web sites and apps that depend on Cloudflare’s 1.1.1.1 DNS resolver.
  • Trigger: A configuration error launched weeks earlier than was triggered by an unrelated change, prompting Cloudflare’s BGP route bulletins to fade from the worldwide web routing desk.
  • Takeaways: “With no legitimate routes, site visitors couldn’t attain Cloudflare’s 1.1.1.1 DNS resolver,” ThousandEyes reported, including that the incident highlights “how flaws in configuration updates don’t all the time set off a right away disaster, as a substitute storing up issues for later.”
  • Period: Greater than two hours
  • Signs: The corporate’s cell app, web site, and ATM machines all went down and failed concurrently.
  • Trigger: A shared backend dependency failed, affecting all buyer touchpoints, ThousandEyes estimated.
  • Takeaways: “The truth that three completely different channels with three completely different frontend applied sciences failed eliminates app or UI points,” ThousandEyes famous, explaining that this incident demonstrated “how a single failure can immediately disable each buyer touchpoint—and why it’s very important to examine all indicators earlier than reaching for treatments.”
  • Period: Each incidents lasted a number of hours
  • Signs: The primary outage affected EMEA area customers with slowdowns and failures; the second impacted customers worldwide with HTTP 503 errors and connection timeouts.
  • Trigger: The October 9 incident was brought on by software program defects that crashed edge websites within the EMEA area; the October 29 outage was triggered by a configuration change
  • Takeaways: “Collectively, these two outages illustrate an vital distinction: infrastructure failures are usually regional with solely sure clients affected, whereas configuration errors usually hit all areas concurrently,” in response to ThousandEyes.
  • Period: Greater than 15 hours for some clients
  • Signs: Lengthy, international service disruptions affected main clients, together with Slack, Atlassian, and Snapchat.
  • Trigger: Failure within the US-EAST-1 area, however international providers equivalent to IAM and DynamoDB World Tables trusted that regional endpoint, that means the outage propagated worldwide.
  • Takeaways: “The incident highlights how a failure in a single, centralized service can ripple outwards via dependency chains that aren’t all the time apparent from structure diagrams,” ThousandEyes famous.
  • Period: A number of hours of intermittent, international instability
  • Signs: Intermittent service disruptions slightly than an entire outage
  • Trigger: A nasty configuration file in Cloudflare’s Bot Administration system exceeded a hard-coded restrict, inflicting proxies to fail as they loaded the outsized file on staggered five-minute cycles.
  • Takeaways: “As a result of the proxies refreshed configurations on staggered five-minute cycles, we didn’t see a lights-on/lights-off outage, however intermittent, international instability,” ThousandEyes reported, noting that the incident revealed how distributed edge mixed with staggered updates can create intermittent points.
See also  How to get rid of unneeded files on Linux

Classes realized in 2025

ThousandEyes highlighted a number of takeaways for community operations groups trying to enhance their resilience in 2026:

Examine single signs as they are often deceptive. The true explanation for disruption can emerge from mixtures of indicators. “If the community appears wholesome however customers are experiencing points, the issue could be within the backend,” in response to ThousandEyes. “Simultaneous failures throughout channels level to shared dependencies, whereas intermittent failures might point out rollout or edge issues.”

Concentrate on speedy detection and response. The complexity of contemporary methods means it’s unrealistic to forestall each doable difficulty via testing alone. “As an alternative, deal with constructing speedy detection and response capabilities, utilizing strategies equivalent to staged rollouts and clear communication with stakeholders,” ThousandEyes acknowledged.

Source link

TAGGED: application, failures, Network, outages, Top
Share This Article
Twitter Email Copy Link Print
Previous Article How to avoid drowning in data at the expense of freshwater supplies How to avoid drowning in data at the expense of freshwater supplies
Next Article UK data centres are booming – but is the power running out? UK data centres are booming – but is the power running out?
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Your Trusted Source for Accurate and Timely Updates!

Our commitment to accuracy, impartiality, and delivering breaking news as it happens has earned us the trust of a vast audience. Stay ahead with real-time updates on the latest events, trends.
FacebookLike
TwitterFollow
InstagramFollow
YoutubeSubscribe
LinkedInFollow
MediumFollow
- Advertisement -
Ad image

Popular Posts

Delivery drones may soon take off in the US. Here’s why

Julep Toth, 9, appears to be like at a drone because it arrives at her…

August 14, 2025

Utah-based colocation data center prioritizes water conservation | DCN

Novva’s new Utah data center looks like it could double as the set of a…

February 6, 2024

As insurers look to be more agile, data mesh strategies take centerstage

In an period of enterprise the place each trade requires its stakeholders to be able…

March 4, 2024

Dexa Raises $6M in Seed Funding

Dexa, a NYC-based AI-powered search engine specifically built with multi-modal content in mind, raised $6M…

February 6, 2024

The day CrowdStrike broke the Internet, China was largely unaffected. Here’s why

On July 19, a routine software program replace from cybersecurity large CrowdStrike unleashed digital chaos…

July 22, 2024

You Might Also Like

A person watching a stream of videos on a tablet
Global Market

Ruckus makes some noise with preconfigured switches for AV-over-IP networks

By saad
SpaceX
Global Market

Musk’s million data centers in space won’t fly, say experts

By saad
Top 7 best AI penetration testing companies in 2026
AI

Top 7 best AI penetration testing companies in 2026

By saad
Is your Java estate audit-ready – or just hoping for the best?
Global Market

Is your Java estate audit-ready – or just hoping for the best?

By saad
Data Center News
Facebook Twitter Youtube Instagram Linkedin

About US

Data Center News: Stay informed on the pulse of data centers. Latest updates, tech trends, and industry insights—all in one place. Elevate your data infrastructure knowledge.

Top Categories
  • Global Market
  • Infrastructure
  • Innovations
  • Investments
Usefull Links
  • Home
  • Contact
  • Privacy Policy
  • Terms & Conditions

© 2024 – datacenternews.tech – All rights reserved

Welcome Back!

Sign in to your account

Lost your password?
We use cookies to ensure that we give you the best experience on our website. If you continue to use this site we will assume that you are happy with it.
You can revoke your consent any time using the Revoke consent button.