Inverse probability

In probability theory, inverse probability is an obsolete term for the probability distribution of an unobserved variable.

Today, the problem of determining an unobserved variable (by whatever method) is called inferential statistics, the method of inverse probability (assigning a probability distribution to an unobserved variable) is called Bayesian probability, the "distribution" of an unobserved variable given data is rather the likelihood function (which is not a probability distribution), and the distribution of an unobserved variable, given both data and a prior distribution, is the posterior distribution. The development of the field and terminology from "inverse probability" to "Bayesian probability" is described by Fienberg (2006).[1]

Ronald Fisher

The term "inverse probability" appears in an 1837 paper of De Morgan, in reference to Laplace's method of probability (developed in a 1774 paper, which independently discovered and popularized Bayesian methods, and 1812 book), though the term "inverse probability" does not occur in these.[1] Fisher uses the term in 1922, referring to "the fundamental paradox of inverse probability" as the source of the confusion between statistical terms that refer to the true value to be estimated, with the actual value arrived at by the estimation method, which is subject to error. ,[2] (See reprint in .[3]) Later Jeffreys uses the term in his defense of the methods of Bayes and Laplace, in 1939.[4] The term "Bayesian", which displaced "inverse probability", was introduced by Ronald Fisher around 1950.[5]

Inverse probability, variously interpreted, was the dominant approach to statistics until the development of frequentism in the early 20th century by Ronald Fisher, Jerzy Neyman and Egon Pearson.[1] Following the development of frequentism, the terms frequentist and Bayesian developed to contrast these approaches, and became common in the 1950s.

Details

In modern terms, given a probability distribution p(x|θ) for an observable quantity x conditional on an unobserved variable θ, the "inverse probability" is the posterior distribution p(θ|x), which depends both on the likelihood function (the inversion of the probability distribution) and a prior distribution. The distribution p(x|θ) itself is called the direct probability.

The inverse probability problem (in the 18th and 19th centuries) was the problem of estimating a parameter from experimental data in the experimental sciences, especially astronomy and biology. A simple example would be the problem of estimating the position of a star in the sky (at a certain time on a certain date) for purposes of navigation. Given the data, one must estimate the true position (probably by averaging). This problem would now be considered one of inferential statistics.

The terms "direct probability" and "inverse probability" were in use until the middle part of the 20th century, when the terms "likelihood function" and "posterior distribution" became prevalent.

See also

References

  1. 1 2 3 Fienberg, Stephen E. (2006). "When Did Bayesian Inference Become "Bayesian"?" (PDF). Bayesian Analysis 1 (1): 140. doi:10.1214/06-BA101.
  2. Fisher, R. A. (1922). "On the Mathematical Foundations of Theoretical Statistics". Philos. Trans. R. Soc. London, Ser. A 222A: 309–368.
  3. Kotz, S. (1992). Breakthroughs in Statistics Volume 1. Springer-Verlag.
  4. Jeffreys, Harold (1939). Theory of Probability (Third ed.). Oxford University Press.
  5. Stephen. E. Fienberg, (2006) When did Bayesian Inference become "Bayesian"? Bayesian Analysis, 1 (1), 140. See page 5.
This article is issued from Wikipedia - version of the Sunday, January 10, 2016. The text is available under the Creative Commons Attribution/Share Alike but additional terms may apply for the media files.