AI Safety Can Be a Pascal's Mugging Even if P(doom) Is High

AI Safety Can Be a Pascal's Mugging Even if P(doom) Is High

LessWrong
LessWrongApr 25, 2026

Key Takeaways

  • Pascal's mugging hinges on personal impact probability, not baseline risk
  • High p(doom) alone doesn't prevent AI safety from being a mugging
  • Your chance to influence AI outcomes is higher than a bajillion
  • Effective AI risk work requires estimating p(you avert doom)
  • Misreading probabilities can misguide funding and policy decisions

Pulse Analysis

Pascal’s mugging, a classic decision‑theory problem, asks whether a tiny probability of an enormous payoff justifies a disproportionate sacrifice. In AI safety circles, critics invoke the paradox to claim that investing heavily in alignment is irrational because an individual’s chance of preventing an existential catastrophe seems vanishingly small. The article clarifies that the mugging hinges on the probability *you* can affect the outcome, not on the overall likelihood of doom. This nuance separates the abstract risk of AI failure from the concrete influence any actor can wield.

Even when p(doom) is perceived as high, the personal impact probability remains the decisive factor. The author illustrates this with a thought experiment involving a stranger who promises to flip a divine coin, showing that a high baseline risk does not automatically raise the odds that your wallet‑exchange will change the world. However, the piece argues that most observers underestimate their network proximity to decision‑makers—presidents, lab CEOs, and AI leaders—meaning the chance of influencing policy or research is far greater than the "bajillion" myth suggests. This reframes the calculus: rather than dismissing AI safety as a futile gamble, we should assess how our actions ripple through the influence chain.

For investors, policymakers, and professionals, the takeaway is clear: risk allocation must be based on realistic estimates of marginal impact, not on sensationalized doom probabilities. By quantifying p(you avert doom), stakeholders can prioritize interventions—such as funding alignment research, supporting governance frameworks, or engaging in advocacy—that meaningfully shift the trajectory of AI development. This more granular approach promotes efficient resource deployment, bolsters the credibility of AI safety initiatives, and aligns with broader existential risk mitigation strategies.

AI safety can be a Pascal's mugging even if p(doom) is high

Comments

Want to join the conversation?