History of probability
|History of science|
Probability has a dual aspect: on the one hand the probability or likelihood of hypotheses given the evidence for them, and on the other hand the behavior of stochastic processes such as the throwing of dice or coins. The study of the former is historically older in, for example, the law of evidence, while the mathematical treatment of dice began with the work of Pascal and Fermat in the 1650s.
Probability is distinguished from statistics. (See history of statistics). While statistics deals with data and inferences from it, (stochastic) probability deals with the stochastic (random) processes which lie behind data or outcomes.
Probable and probability and their cognates in other modern languages derive from medieval learned Latin probabilis and, deriving from Cicero and generally applied to an opinion to mean plausible or generally approved. The mathematical sense of the term is from 1718. In the 18th century, the term chance was also used in the mathematical sense of "probability" (and probability theory was called Doctrine of Chances). This word is ultimately from Latin cadentia, i.e. "a fall, case". The English adjective likely is of Germanic origin, most likely from Old Norse likligr (Old English had geliclic with the same sense), originally meaning "having the appearance of being strong or able" "having the similar appearance or qualities, with a meaning of "probably" recorded from the late 14th century. Similarly, the derived noun likelihood had a meaning of "similarity, resemblance" but took on a meaning of "probability" from the mid 15th century.
Ancient and medieval law of evidence developed a grading of degrees of proof, probabilities, presumptions and half-proof to deal with the uncertainties of evidence in court. In Renaissance times, betting was discussed in terms of odds such as "ten to one" and maritime insurance premiums were estimated based on intuitive risks, but there was no theory on how to calculate such odds or premiums.
The mathematical methods of probability arose in the correspondence of Pierre de Fermat and Blaise Pascal (1654) on such questions as the fair division of the stake in an interrupted game of chance. Christiaan Huygens (1657) gave a comprehensive treatment of the subject.
- In ancient times there were games played using astragali, or Talus bone. The Pottery of ancient Greece was evidence to show :that there was a circle drawn on the floor and the astragali were tossed into this circle, much like playing marbles. In Egypt, :excavators of tombs found a game they called "Hounds and Jackals", which closely resembles the modern game "Snakes and Ladders". It seems that this is the early stages of the creation of dice.
- First dice game mentioned in literature of the Christian era was called Hazard. Played with 2 or 3 dice. Thought to have been brought :to Europe by the knights returning from the Crusades.
- Dante (1265-1321) mentions this game. A commentor of Dante puts further thought into this game: the thought was that with 3 dice, the lowest number you can get is 3, an ace for every die. Achieving a 4 can be done with 3 die by having a two on one die and aces on the other two dice.
- Cardano also thought about the throwing of three die. 3 dice are thrown: there are the same number of ways to throw a 9 as there are a 10. For a 9:(621) (531) (522) (441) (432) (333) and for 10: (631) (622) (541) (532) (442) (433). From this, Cardano found that the probability of throwing a 9 is less than that of throwing a 10. He also demonstrated the efficacy of defining odds as the ratio of favourable to unfavourable outcomes (which implies that the probability of an event is given by the ratio of favourable outcomes to the total number of possible outcomes ).
- In addition, the famous Galileo wrote about die-throwing sometime between 1613 and 1623. Essentially thought about Cardano's problem, about the probability of throwing a 9 is less than throwing a 10. Galileo had the following to say: Certain numbers have the ability to be thrown because there are more ways to create that number. Although 9 and 10 have the same number of ways to be created, 10 is considered by dice players to be more common than 9.
Jacob Bernoulli's Ars Conjectandi (posthumous, 1713) and Abraham de Moivre's The Doctrine of Chances (1718) put probability on a sound mathematical footing, showing how to calculate a wide range of complex probabilities. Bernoulli proved a version of the fundamental law of large numbers, which states that in a large number of trials, the average of the outcomes is likely to be very close to the expected value - for example, in 1000 throws of a fair coin, it is likely that there are close to 500 heads (and the larger the number of throws, the closer to half-and-half the proportion is likely to be).
The power of probabilistic methods in dealing with uncertainty was shown by Gauss's determination of the orbit of Ceres from a few observations. The theory of errors used the method of least squares to correct error-prone observations, especially in astronomy, based on the assumption of a normal distribution of errors to determine the most likely true value. In 1812, Laplace issued his Théorie analytique des probabilités in which he consolidated and laid down many fundamental results in probability and statistics such as the moment generating function, method of least squares, inductive probability, and hypothesis testing.
Towards the end of the nineteenth century, a major success of explanation in terms of probabilities was the Statistical mechanics of Ludwig Boltzmann and J. Willard Gibbs which explained properties of gases such as temperature in terms of the random motions of large numbers of particles.
The field of the history of probability itself was established by Isaac Todhunter's monumental History of the Mathematical Theory of Probability from the Time of Pascal to that of Lagrange (1865).
Probability and statistics became closely connected through the work on hypothesis testing of R. A. Fisher and Jerzy Neyman, which is now widely applied in biological and psychological experiments and in clinical trials of drugs, as well as in economics and elsewhere. A hypothesis, for example that a drug is usually effective, gives rise to a probability distribution that would be observed if the hypothesis is true. If observations approximately agree with the hypothesis, it is confirmed, if not, the hypothesis is rejected.
The theory of stochastic processes broadened into such areas as Markov processes and Brownian motion, the random movement of tiny particles suspended in a fluid. That provided a model for the study of random fluctuations in stock markets, leading to the use of sophisticated probability models in mathematical finance, including such successes as the widely used Black–Scholes formula for the valuation of options.
The twentieth century also saw long-running disputes on the interpretations of probability. In the mid-century frequentism was dominant, holding that probability means long-run relative frequency in a large number of trials. At the end of the century there was some revival of the Bayesian view, according to which the fundamental notion of probability is how well a proposition is supported by the evidence for it.
The mathematical treatment of probabilities, especially when there are infinitely many possible outcomes, was facilitated by Kolmogorov's axioms (1933).
- J. Franklin, The Science of Conjecture: Evidence and Probability Before Pascal, 113, 126.
- Franklin, The Science of Conjecture, ch. 2.
- Franklin, Science of Conjecture, ch. 11.
- Hacking, Emergence of Probability[page needed]
- Franklin, Science of Conjecture, ch. 12.
- Some laws and problems in classical probability and how Cardano anticipated them Gorrochum, P. Chance magazine 2012
- Salsburg, The Lady Tasting Tea.
- Bernstein, Against the Gods, ch. 18.
- Bernstein, Peter L. (1996). Against the Gods: The Remarkable Story of Risk. New York: Wiley. ISBN 0-471-12104-5.
- Daston, Lorraine (1988). Classical Probability in the Enlightenment. Princeton: Princeton University Press. ISBN 0-691-08497-1.
- Franklin, James (2001). The Science of Conjecture: Evidence and Probability Before Pascal. Baltimore, MD: Johns Hopkins University Press. ISBN 0-8018-6569-7.
- Hacking, Ian (2006). The Emergence of Probability (2nd ed). New York: Cambridge University Press. ISBN 978-0-521-86655-2.
- Hald, Anders (2003). A History of Probability and Statistics and Their Applications before 1750. Hoboken, NJ: Wiley. ISBN 0-471-47129-1.
- Hald, Anders (1998). A History of Mathematical Statistics from 1750 to 1930. New York: Wiley. ISBN 0-471-17912-4.
- Heyde, C. C.; Seneta, E. (eds) (2001). Statisticians of the Centuries. New York: Springer. ISBN 0-387-95329-9.
- McGrayne, Sharon Bertsch (2011). The Theory That Would Not Die: How Bayes' Rule Cracked the Enigma Code, Hunted Down Russian Submarines, and Emerged Triumphant from Two Centuries of Controversy. New Haven: Yale University Press. ISBN 9780300169690.
- von Plato, Jan (1994). Creating Modern Probability: Its Mathematics, Physics and Philosophy in Historical Perspective. New York: Cambridge University Press. ISBN 978-0-521-59735-7.
- Salsburg, David (2001). The Lady Tasting Tea: How Statistics Revolutionized Science in the Twentieth Century. ISBN 0-7167-4106-7
- Stigler, Stephen M. (1990). The History of Statistics: The Measurement of Uncertainty before 1900. Belknap Press/Harvard University Press. ISBN 0-674-40341-X.
- JEHPS: Recent publications in the history of probability and statistics
- Electronic Journ@l for History of Probability and Statistics/Journ@l Electronique d'Histoire des ProbabilitÃ©et de la Statistique
- Figures from the History of Probability and Statistics (Univ. of Southampton)
- Probability and Statistics on the Earliest Uses Pages (Univ. of Southampton)
- Earliest Uses of Symbols in Probability and Statistics on Earliest Uses of Various Mathematical Symbols