| title | author | date | license | tags | |||||
|---|---|---|---|---|---|---|---|---|---|
Moore’s Doomsday Hypothesis |
Pete Moore |
2025-11-20 |
CC BY 4.0 |
|
A speculative model of existential risk in the age of open-source AI.
This paper proposes a probabilistic model for the eventual self-destruction of humanity through the intersection of human malevolence and technological capability. The hypothesis asserts that if a small but non-zero fraction of individuals possess both the intent and the means to cause global extinction, then humanity’s survival time is inversely related to the product of these factors. Open-source artificial intelligence accelerates this convergence by exponentially expanding access to catastrophic capability.
Let:
-
x: Fraction of the global population with a latent desire or willingness to destroy humanity, e.g. due to psychopathy, nihilism, or ideological extremism.
$0 < x \ll 1$ -
y(t): Number of individuals at time t who possess the technical capability to cause human extinction — for example, through AI-assisted bioweapon design, or synthetic pathogens.
$y(t)$ increases over time as technology becomes more accessible. -
m(t): Mitigation factor representing defences against catastrophic misuse — including governance, AI alignment, surveillance, and psychological intervention.
-
p(t) = x · y(t): Expected number of individuals both willing and able to destroy humanity at time t.
When
As
If each potential actor represents an independent extinction risk, the survival probability
and thus:
As technological capability proliferates (
Open-source AI, though motivated by transparency and decentralisation, accelerates
It thereby transforms existential risk from state-level containment to individual-level accessibility.
The psychological distribution
Thus, technological democratisation is the primary accelerant.
Traditional risk management assumes concentration of power in few hands; Moore’s Doomsday Hypothesis suggests that universal empowerment inevitably leads to diffuse existential risk.
When the means of destruction become as ubiquitous as intent, extinction becomes a statistical certainty.
Human extinction does not require collective intent — only the alignment of capability with a statistically inevitable fraction of destructive will.
The hypothesis implies that the open dissemination of advanced AI, while noble in spirit, may ultimately ensure the demise of the species it seeks to empower.
Efforts to reduce existential risk must therefore focus on:
- Slowing the growth of
$y(t)$ via containment and safety barriers. - Reducing
$x$ through social, educational, and psychological interventions. - Expanding
$m(t)$ — collective resilience mechanisms that scale at least as fast as capability.
Let
Assuming independent risk events follow a Poisson process, the probability that humanity survives up to time t is:
For simplicity, if
and the expected time to extinction is:
This makes clear that as
Authored by Pete Moore, 2025.
Licensed under Creative Commons Attribution 4.0 International (CC BY 4.0).