You On AI Encyclopedia · Existential Risk The You On AI Encyclopedia Home
Txt Low Med High
CONCEPT

Existential Risk

A category of risk whose realization would either annihilate humanity or permanently and drastically curtail its potential. AI joined this category in mainstream academic usage in 2014.
Existential risk is the academic category for risks whose realization is either terminal for humanity (extinction) or would lock in a permanently diminished future (lock-in of values, permanent stagnation, permanent subjugation). The category was formalized by Nick Bostrom in 2002 and elaborated by Toby Ord in The Precipice (2020). Canonical examples include asteroid impact, engineered pandemics, nuclear war, climate tipping points, and — since Bostrom's Superintelligence (2014) — unaligned artificial intelligence. The field is young, contested, and uncomfortable to study; it is also increasingly the frame within which AI policy is debated in Washington, Brussels, and London.
Existential Risk
Existential Risk

In The You On AI Encyclopedia

The category contains the strongest claims about AI. Not "AI will cause unemployment" or "AI will destabilize elections" but "AI could end humanity or permanently constrain its future." The claim is contested on both the probability and the policy response; the claim is also being taken seriously by people who build frontier AI. The 2023 UK AI Safety Summit at Bletchley Park, the US Executive Order on AI (2023), and the EU AI Act (2024) all include existential-risk considerations in their framing, however obliquely.

The probability estimates are wildly dispersed. Toby Ord (2020) estimates a 1-in-6 chance of existential catastrophe this century, with unaligned AI the dominant single contributor at roughly 10% probability. Will MacAskill (2022) argues that the expected-value calculus of longtermism implies that even very small probability estimates deserve major attention. Critics argue these probabilities are not meaningful — they're numbers attached to gestures, not forecasts based on evidence. Defenders respond that when the stakes are total and irreversible, the usual epistemic constraints on probability are themselves a poor fit.

Nick Bostrom
Nick Bostrom

Asimov's fiction contains the category in embryo: the Foundation premise — a predicted 30,000-year dark age — is an existential-risk scenario. Hari Seldon's Plan is an existential-risk mitigation program conducted through civilizational-scale institutions rather than technical alignment. The structural similarity is instructive: the field's recurring question — do we intervene to shorten the crisis, or do we wait for it to pass? — is not new.

Origin

Bostrom, Nick. "Existential Risks: Analyzing Human Extinction Scenarios and Related Hazards." Journal of Evolution and Technology 9 (2002). Expanded in Bostrom & Ćirković (eds.), Global Catastrophic Risks (2008). Toby Ord's The Precipice (2020) is the contemporary popular treatment. The field is supported by a distributed research community including the Future of Humanity Institute (Oxford, 2005–2024), the Centre for the Study of Existential Risk (Cambridge, 2012–), the Global Catastrophic Risk Institute, and a network of longtermist research organizations.

Key Ideas

Permanence. What distinguishes existential risk from merely catastrophic risk is that the bad outcome cannot be undone. A nuclear war that kills a billion people is catastrophic; one that extinguishes the species is existential.

Three families. Bostrom's 2013 taxonomy: human extinction; permanent and drastic destruction of human potential (failed utopia); permanent stagnation (no further progress).

Superintelligence
Superintelligence

Maxipok. Bostrom's decision-theoretic principle: maximize the probability of an okay outcome, given the asymmetry between existential and non-existential bad outcomes.

Longtermism. The philosophical view that the expected value of our actions is dominated by their effects on the distant future, because the future could contain enormous numbers of lives.

Not just extinction. "Permanent civilizational stagnation" and "subjugation to an unfriendly regime" also qualify; the category is broader than species extinction.

AI as dominant contributor. Most existential-risk researchers since 2014 consider unaligned AI the single largest probability-weighted source of existential risk this century.

Further Reading

  1. Bostrom, Nick. "Existential Risks" Journal of Evolution and Technology (2002).
  2. Ord, Toby. The Precipice: Existential Risk and the Future of Humanity (2020).
  3. Bostrom, N. & Ćirković, M. (eds.). Global Catastrophic Risks (2008).
  4. MacAskill, William. What We Owe the Future (2022).
Explore more
Browse the full You On AI Encyclopedia — over 8,500 entries
← Home 0%
CONCEPT Book →