Bayesian Probability
Bayesian Probability
Contents
1Bayesian methodology
2Objective and subjective Bayesian probabilities
3History
4Justification of Bayesian probabilities
o 4.1Axiomatic approach
o 4.2Dutch book approach
o 4.3Decision theory approach
5Personal probabilities and objective methods for constructing priors
6See also
7References
8Bibliography
Bayesian methodology[edit]
Bayesian methods are characterized by concepts and procedures as follows:
The use of random variables, or more generally unknown quantities,[7] to model all
sources of uncertainty in statistical models including uncertainty resulting from lack of
information (see also aleatoric and epistemic uncertainty).
The need to determine the prior probability distribution taking into account the available
(prior) information.
The sequential use of Bayes' formula: when more data become available, calculate the
posterior distribution using Bayes' formula; subsequently, the posterior distribution becomes
the next prior.
While for the frequentist, a hypothesis is a proposition (which must be either true or false)
so that the frequentist probability of a hypothesis is either 0 or 1, in Bayesian statistics, the
probability that can be assigned to a hypothesis can also be in a range from 0 to 1 if the truth
value is uncertain.
History[edit]
Main article: History of statistics § Bayesian statistics
The term Bayesian derives from Thomas Bayes (1702–1761), who proved a special case of what
is now called Bayes' theorem in a paper titled "An Essay towards solving a Problem in the
Doctrine of Chances".[9] In that special case, the prior and posterior distributions were beta
distributions and the data came from Bernoulli trials. It was Pierre-Simon Laplace (1749–1827)
who introduced a general version of the theorem and used it to approach problems in celestial
mechanics, medical statistics, reliability, and jurisprudence.[10] Early Bayesian inference, which
used uniform priors following Laplace's principle of insufficient reason, was called "inverse
probability" (because it infers backwards from observations to parameters, or from effects to
causes).[11] After the 1920s, "inverse probability" was largely supplanted by a collection of
methods that came to be called frequentist statistics.[11]
In the 20th century, the ideas of Laplace developed in two directions, giving rise
to objective and subjective currents in Bayesian practice. Harold Jeffreys' Theory of
Probability (first published in 1939) played an important role in the revival of the Bayesian view of
probability, followed by works by Abraham Wald (1950) and Leonard J. Savage (1954). The
adjective Bayesian itself dates to the 1950s; the derived Bayesianism, neo-Bayesianism is of
1960s coinage.[12][13][14] In the objectivist stream, the statistical analysis depends on only the model
assumed and the data analysed.[15] No subjective decisions need to be involved. In contrast,
"subjectivist" statisticians deny the possibility of fully objective analysis for the general case.
In the 1980s, there was a dramatic growth in research and applications of Bayesian methods,
mostly attributed to the discovery of Markov chain Monte Carlo methods and the consequent
removal of many of the computational problems, and to an increasing interest in nonstandard,
complex applications.[16] While frequentist statistics remains strong (as seen by the fact that most
undergraduate teaching is still based on it [17][citation needed]), Bayesian methods are widely accepted and
used, e.g., in the field of machine learning.[18]
Axiomatic approach[edit]
Richard T. Cox showed that Bayesian updating follows from several axioms, including
two functional equations and a hypothesis of differentiability.[8] The assumption of differentiability
or even continuity is controversial; Halpern found a counterexample based on his observation
that the Boolean algebra of statements may be finite.[19] Other axiomatizations have been
suggested by various authors with the purpose of making the theory more rigorous.[7]