Pascal's mugging

From Wikipedia, the free encyclopedia
Jump to navigation Jump to search

In philosophy, Pascal's mugging is a thought-experiment demonstrating a problem in expected utility maximization. A rational agent should choose actions whose outcomes, when weighed by their probability, have higher utility. But some very unlikely outcomes may have very great utilities, and these utilities can grow faster than the probability diminishes. Hence the agent should focus more on vastly improbable cases with implausibly high rewards; this leads first to counter-intuitive choices, and then to incoherence as the utility of every choice becomes unbounded.

The name refers to Pascal's Wager, but unlike the wager, it does not require infinite rewards.[1] This sidesteps many objections to the Pascal's Wager dilemma that are based on the nature of infinity.[2]

Problem statement[edit]

In one description,[2] Blaise Pascal is accosted by a mugger who has forgotten their weapon. However, the mugger proposes a deal: the philosopher gives them his wallet, and in exchange the mugger will return twice the amount of money tomorrow. Pascal declines, pointing out that it is unlikely the deal will be honoured. The mugger then continues naming higher rewards, pointing out that even if it is just one chance in 1000 that they will be honourable, it would make sense for Pascal to make a deal for a 2000 times return. Pascal responds that the probability for that high return is even lower than one in 1000. The mugger argues back that for any low probability of being able to pay back a large amount of money (or pure utility) there exists a finite amount that makes it rational to take the bet—and given human fallibility and philosophical scepticism a rational person must admit there is at least some non-zero chance that such a deal would be possible. In one example, the mugger succeeds by promising Pascal 1,000 quadrillion happy days of life. Convinced by the argument, Pascal gives the mugger the wallet.

The term "Pascal's mugging" to refer to this problem was originally coined by Eliezer Yudkowsky in the Less Wrong forum.[3][2] In one of Yudkowsky's examples, the mugger succeeds by saying "give me five dollars, or I'll use my magic powers from outside the Matrix to run a Turing machine that simulates and kills people". Here, the number uses Knuth's up-arrow notation; writing the number out in base 10 would require enormously more writing material than there are atoms in the known universe.[3]

The supposed paradox results from two inconsistent views. On the one side, by multiplying an expected utility calculation, assuming loss of five dollars to be valued at , loss of a life to be valued at , and probability that the mugger is telling the truth at , the solution is to give the money if and only if . Assuming that is higher than , so long as is higher than , which is assumed to be true, it is considered rational to pay the mugger. On the other side of the argument, paying the mugger is intuitively irrational due to its exploitability. If the person being mugged agrees to this sequence of logic, then they can be exploited repeatedly for all of their money, resulting in a Dutch-book, which is typically considered irrational. Views on which of these arguments is logically correct differ.[3]

Moreover, in many reasonable-seeming decision systems, "Pascal's mugging" causes the expected utility of any action to fail to converge, as an unlimited chain of successively dire scenarios similar to Pascal's mugging would need to be factored in.[4][5]

Consequences and remedies[edit]

Philosopher Nick Bostrom argues that Pascal's mugging, like Pascal's wager, suggests that giving a superintelligent artificial intelligence a flawed decision theory could be disastrous.[6] Pascal's mugging may also be relevant when considering low-probability, high-stakes events such as existential risk or charitable interventions with a low probability of success but extremely high rewards. Common sense seems to suggest that spending effort on too unlikely scenarios is irrational.

One advocated remedy might be to only use bounded utility functions: rewards cannot be arbitrarily large.[4][7] Another approach is to use Bayesian reasoning to (qualitatively) judge the quality of evidence and probability estimates rather than naively calculate expectations.[8] Other approaches are to penalize the prior probability of hypotheses that argue that we are in a surprisingly unique position to affect large numbers of other people who cannot symmetrically affect us,[note 1] reject providing the probability of a payout first,[12] or abandon quantitative decision procedures in the presence of extremely large risks.[5]

See also[edit]


  1. ^ This 'leverage penalty' was first proposed by Robin Hanson in a comment on Yudkowsky's original statement of the problem;[9][10] Yudkowsky noted that this would imply refusing to believe in theories implying we can affect vastly many others, even in the face of what might otherwise look like overwhelming observational evidence for the theory.[11]


  1. ^ Olle Häggström (2016). Here Be Dragons. 3.10: Cryonics. ISBN 978-0198723547.{{cite book}}: CS1 maint: location (link)
  2. ^ a b c Bostrom, Nick (2009). "Pascal's mugging" (PDF). Analysis. 69 (3): 443–445. doi:10.1093/analys/anp062. JSTOR 40607655.
  3. ^ a b c Eliezer Yudkowsky, Pascal's Mugging: Tiny Probabilities of Vast Utilities. Less Wrong, 19 October 2007.
  4. ^ a b De Blanc, Peter. Convergence of Expected Utilities with Algorithmic Probability Distributions (2007), arXiv:0712.4318
  5. ^ a b Kieran Marray, Dealing With Uncertainty in Ethical Calculations of Existential Risk, Presented at The Economic and Social Research Council Climate Ethics and Climate Economics Workshop Series: Workshop Five - Risk and the Culture of Science, May 2016
  6. ^ Bostrom, Nick (2014). "Choosing the Criteria for Choosing". Superintelligence: Paths, Dangers, Strategies. Oxford: Oxford University Press. ISBN 978-0199678112. "Decision Theory" section.
  7. ^ Cowen, Tyler; High, Jack (1988). "Time, Bounded Utility, and the St. Petersburg Paradox". Theory and Decision. 25 (3): 219–223. doi:10.1007/BF00133163. S2CID 120584258.
  8. ^ Holden Karnofsky, Why We Can’t Take Expected Value Estimates Literally (Even When They’re Unbiased). GiveWell Blog August 18, 2011
  9. ^ Robin Hanson (21 October 2007), comment on Eliezer Yudkowsky's "Pascal's Mugging: Tiny Probabilities of Vast Utilities", LessWrong: "People have been talking about assuming that states with many people hurt have a low (prior) probability. It might be more promising to assume that states with many people hurt have a low correlation with what any random person claims to be able to effect."
  10. ^ Tomasik, Brian (June 2016). "How the Simulation Argument Dampens Future Fanaticism" (PDF). Center on Long-Term Risk. pp. 3–4. Archived (PDF) from the original on 2021-11-23.
  11. ^ Yudkowsky, Eliezer (2013-05-08). "Pascal's Muggle: Infinitesimal Priors and Strong Evidence". LessWrong. Archived from the original on 2016-06-27.
  12. ^ Baumann, Peter (2009). "Counting on numbers" (PDF). Analysis. 69 (3): 446–448. doi:10.1093/analys/anp061. JSTOR 40607656. Archived (PDF) from the original on 2019-11-21.