Skip to main content
Back to Insights
Case Study11 min readDecember 2024

The CrowdStrike Lesson: What Happens When System 1 Runs Wild

July 19, 2024. A single update. 8.5 million machines. $5.4 billion in damages. And a preview of what's coming.

On July 19, 2024, the world got a preview of what happens when speed wins over deliberation.

CrowdStrike, one of the world's largest cybersecurity companies, pushed a routine content update to its Falcon sensor software. Within hours, approximately 8.5 million Windows devices crashed. Blue screens of death cascaded across the globe.

The immediate impact:

  • • Airlines grounded flights
  • • Hospitals cancelled surgeries
  • • Banks went offline
  • • Emergency services lost dispatch systems
  • • Broadcasters went dark
  • • Retailers couldn't process payments

The cause? A faulty update that passed through automated validation but contained a logic error that human review would have caught. A single bad file, pushed at machine speed, scaled to machine catastrophe.

The SAFE-AI Manifesto, signed by 49 researchers, cites CrowdStrike as a warning:

"At the societal level, [aggressive release cycles] can lead to harm through failed systems."

They're being academic about it. Let me be direct: CrowdStrike is what happens when System 1 runs unsupervised. And in the age of AI, it's a preview – not an anomaly.


The Anatomy of the Failure

Let's be precise about what happened.

CrowdStrike's Falcon sensor runs at the kernel level – the deepest layer of the operating system. This gives it powerful security capabilities. It also means that when Falcon fails, the entire system fails. There's no graceful degradation. There's just a blue screen.

The Update

On July 19, CrowdStrike pushed a "Rapid Response Content" update. These updates are designed to be fast – responding to emerging threats in real-time. Speed is the feature. Speed is the selling point.

The Error

The update contained a logic error in a configuration file. The error caused Falcon to crash. Because Falcon runs at kernel level, crashing Falcon crashed Windows. Because the update pushed automatically to millions of endpoints, millions of machines crashed simultaneously.

The Response

The faulty update was live for approximately 78 minutes before CrowdStrike reverted it. But the damage was done. Affected machines couldn't boot. They couldn't receive the fix. Each one required manual intervention – a technician physically accessing the machine and deleting the bad file.

8.5M

machines requiring manual intervention

$5.4B

in estimated damages

And that's just the direct costs – not the flights missed, the surgeries delayed, the emergencies unresponded to, the trust eroded.


System 1 at Scale

The CrowdStrike failure is a textbook case of System 1 thinking at scale.

System 1, in Kahneman's framework, is fast, automatic, and pattern-based. It's how you catch a ball or recognise a face. It works brilliantly when situations match learned patterns. It fails catastrophically when they don't.

CrowdStrike's update pipeline was System 1:

  • • Automated testing
  • • Automated validation
  • • Automated deployment

Fast, efficient, scalable. It had processed thousands of updates successfully. The pattern suggested this one would work too.

But this update contained a novel error – one that didn't match the patterns the automated systems were checking for. The automated validation said "looks fine." The automated deployment said "push it everywhere."

No human looked at it. No one asked "what if this is wrong?" No one considered "what's the blast radius if this fails?"

System 1 ran at machine speed across 8.5 million endpoints. System 2 wasn't invited.


The Manifesto's Framework

The SAFE-AI Manifesto uses exactly this distinction to argue for human oversight in software development:

"Despite impressive progress, powerful AI algorithms, including Generative AI, often exhibit biases and can also provide unsafe recommendations... Furthermore, Generative AI can hallucinate, that is, produce nonsensical or factually incorrect responses."

CrowdStrike wasn't using generative AI for its update. But the principle is identical: automated systems operating at speed, without human deliberation, will eventually produce catastrophic failures.

The manifesto's first principle is "Strategic deliberation over speed and scale":

"Strategic deliberation involves practicing slower System 2 thinking before leveraging AI's speed and scale. It safeguards against costly, if not catastrophic failures due to the careless use of the awesome powers of AI."

CrowdStrike had the "awesome power" of instantaneous global deployment. They didn't have the safeguard of deliberation before using it.


The Speed Trap

Here's what makes this failure instructive: speed was the explicit design goal.

"Rapid Response Content" updates exist because cybersecurity threats move fast. When a new attack vector emerges, you want to push detection capabilities immediately. Waiting for extensive testing means leaving customers vulnerable.

This logic is seductive. It's the same logic driving vibe coding, aggressive release cycles, and AI-generated code pushed to production. Move fast. Ship constantly. Iterate in production.

The logic ignores a fundamental asymmetry:

The cost of delay is usually linear, but the cost of catastrophic failure is exponential.

If they had waited 2 hours:

Running additional validation, having a human review the changes – the "cost" would have been two hours of marginally increased threat exposure.

Instead, they pushed immediately:

And the cost was $5.4 billion, grounded airlines, cancelled surgeries, and a permanent case study in what not to do.

The manifesto quotes Sun Tzu:

"Strategy without tactics is the slowest route to victory. Tactics without strategy is the noise before defeat."

CrowdStrike had tactics – fast deployment, automated validation, global reach. They lacked strategy – deliberation about what could go wrong and how to prevent it. Noise before defeat.


The Blast Radius Problem

One of the most damning aspects of the CrowdStrike failure is the blast radius.

8.5 million machines failed simultaneously because the update pushed to everyone at once. There was no staged rollout. No canary deployment. No gradual ramp-up that would have caught the error before it spread.

This is System 1 thinking applied to deployment strategy: we've done this before, it worked before, it will work again. Push everywhere.

System 2 would ask: what's the worst case? How do we limit damage if something goes wrong? Can we structure deployment so that failure is contained rather than cascading?

These questions weren't asked – or if they were, the answers were overridden by the imperative to move fast.

"Accepting and proceeding with whatever code is suggested by AI amounts to letting System 1 take control. As a result, seemingly minor errors could lead to large financial losses and compliance violations. What is worse, if allowed to scale uncontrolled, AI-generated software has the capacity to cause great harm."

"Allowed to scale uncontrolled" is the key phrase. CrowdStrike's update scaled uncontrolled. The harm was great.


The Recovery Nightmare

Here's the detail that should terrify anyone building systems at scale: the recovery.

Because the faulty update caused machines to crash on boot, affected devices couldn't receive a fix over the network. They were stuck in a crash loop. The only solution was manual intervention – booting into safe mode, navigating to a specific directory, deleting the bad file.

8.5 million machines. Manual intervention.

  • • Some in data centres with physical access
  • • Many on employee desktops, scattered across the world
  • • Some on point-of-sale terminals, airport kiosks, hospital workstations

IT teams worked around the clock for days. Some organisations took weeks to fully recover. The remediation cost dwarfed the initial damage.

This is the hidden cost of speed-first deployment: when it fails, recovery scales linearly while failure scaled exponentially.

If they'd pushed to 1% first:

Waited an hour, discovered the error, remediation would have been 85,000 machines. Still painful. Not civilisation-disrupting.

Instead: 8.5 million

The difference between "incident" and "catastrophe." That difference is deliberation.


The AI Amplification

Now here's what keeps me up at night: CrowdStrike wasn't even using AI for code generation.

The failure was a human-written logic error that escaped automated testing. It was pushed by a conventional deployment pipeline optimised for speed. It cascaded through a conventional update mechanism designed for rapid response.

And it still caused $5.4 billion in damage.

What happens when we add AI to this equation?

AI will amplify every factor:

  • • AI can generate code faster than humans
  • • It can generate more code
  • • It can push updates more frequently
  • • It can operate with less human oversight

Every element that made CrowdStrike's failure possible – speed, scale, automation, global reach – gets amplified by AI.

"These risks are now dramatically amplified due to the speed and scale of AI."

CrowdStrike is the preview. AI at scale, without deliberation, is the feature film.


The Systemic Vulnerability

Step back from CrowdStrike specifically and look at the systemic picture.

We've built a world where single points of failure can cascade globally. A few companies provide infrastructure to millions of organisations. A few platforms underpin entire industries. A few updates can reach billions of devices simultaneously.

This is System 1 architecture – optimised for efficiency, speed, and scale. It works brilliantly when nothing goes wrong. It fails catastrophically when something does.

The same vulnerability exists everywhere:

  • • Cloud providers whose outages take down thousands of companies
  • • CDNs whose failures break major portions of the internet
  • • Payment processors whose glitches halt commerce
  • • AI systems whose bad outputs scale at machine speed

Every one of these is a CrowdStrike waiting to happen. Every one is System 1 at scale without System 2 safeguards.


What Deliberation Would Have Looked Like

Let's be concrete about what System 2 oversight would have meant for CrowdStrike:

  • Pre-deployment review.

    A human looks at every Rapid Response Content update before deployment. Not to catch every bug – that's what automated testing is for – but to ask: is this update appropriate? Is this the right time? Is there anything unusual about it? This adds minutes to the deployment process. It might have caught the logic error. Even if it didn't, it creates an accountability point.

  • Staged rollout.

    Push to 0.1% of endpoints. Wait 15 minutes. Check for anomalies. Push to 1%. Wait. Check. Push to 10%. Only after multiple successful stages, push globally. This adds hours to full deployment. It guarantees that catastrophic failures are caught before they become catastrophic.

  • Blast radius limits.

    Architecture that limits how much damage a single failure can cause. Kill switches that halt deployment if crash rates exceed thresholds. Rollback capabilities that don't require manual intervention. This requires upfront engineering investment. It pays off when things go wrong.

  • Pre-mortem thinking.

    Before any deployment, ask: "Imagine this fails catastrophically. What happened?" Work backward from imagined failure to identify risks and mitigations. This adds time to planning. It surfaces risks that pure speed optimization ignores.

None of these are revolutionary. They're basic System 2 practices. But they conflict with the imperative to move fast, so they get skipped.


The Cultural Problem

Here's the deeper issue: our industry culture celebrates speed and punishes deliberation.

Ship fast. Move fast. Break things. Iterate. Fail forward. These are the mantras of modern tech. They're not entirely wrong – speed matters, iteration works, shipping beats theorising.

But they've become religious rather than pragmatic. We've forgotten that speed is a means, not an end. We've forgotten that "break things" has costs. We've forgotten that some things shouldn't be broken.

CrowdStrike's culture, like most tech companies, almost certainly celebrated speed. Fast threat response. Rapid updates. Quick iteration. These were probably seen as competitive advantages.

And they were – until they weren't. Until speed without deliberation cost $5.4 billion and grounded air travel worldwide.

"Practitioners and industry leaders are urged to reestablish modeling as a vital component of fast-paced software development in the age of AI... Organizations should introduce conceptual model reviews alongside traditional code reviews to ensure deliberate reflection on system purpose, safety, and critical requirements."

Deliberate reflection. Not instead of speed. Alongside speed. The corrective, not the replacement.


The CrowdStrike Lesson

Here's what CrowdStrike teaches us:

Speed without deliberation is borrowed time

You can ship fast for a while. The odds catch up eventually. When they do, the bill comes due with interest.

Scale amplifies everything – including failures

The same infrastructure that lets you update 8.5 million machines in minutes lets you break 8.5 million machines in minutes. Power without wisdom is destruction.

Automated systems need human oversight for novel situations

Automation works for known patterns. It fails for unknown patterns. Humans are the circuit breaker that catches what automation can't.

Recovery cost often exceeds prevention cost by orders of magnitude

The hours "saved" by skipping staged rollouts are nothing compared to the weeks spent on manual remediation. Deliberation is cheap. Disaster recovery is expensive.

We've built fragile systems and called them efficient

Our global infrastructure is optimised for the happy path. We need to start optimising for the failure path too.


The Preview of AI

CrowdStrike happened without AI. It happened with conventional software, conventional testing, conventional deployment.

Now add AI to the picture.

AI-generated code that no one fully understands. Vibe-coded applications pushed to production. Automated decisions made at machine speed with machine confidence.

Everything that went wrong at CrowdStrike gets faster, more frequent, and less visible.

"With the twin pressures of agile and AI, the practice of conceptual modeling faces a perfect storm... Much danger exists in discarding modeling, as it has traditionally shielded software development from risks and vulnerabilities of developing a solution without careful and systematic deliberation."

Careful and systematic deliberation is what CrowdStrike lacked. It's what vibe coding lacks. It's what AI at scale, without oversight, will lack.

CrowdStrike was 8.5 million machines and $5.4 billion.

That was the preview.

Without deliberation, without System 2, without human oversight in the loop – the main feature is going to be worse.

CrowdStrikeSystem 1AI safetysoftware developmentdeliberationfailureinfrastructure
JL

Written by

Jason La Greca

Jason La Greca is the founder of Teachnology. He watched the CrowdStrike disaster unfold in real-time and saw exactly what happens when speed wins over thinking. Teachnology helps organisations build systems – and cultures – that balance speed with deliberation.

The SAFE-AI Manifesto referenced in this article was authored by Lukyanenko, Samuel, Tegarden, Larsen, and 45 additional researchers.

Connect on LinkedIn

Ready to build deliberation into your process?

Take the AI Readiness Assessment to understand your approach to speed vs deliberation.

Start Assessment

Need help building System 2 safeguards?

Learn how Teachnology Advisory helps organisations balance speed with oversight.

Explore Advisory