0% found this document useful (0 votes)
92 views

Bayesian Probability

Bayesian probability is an interpretation of probability that represents a state of knowledge or personal belief. It allows for reasoning with hypotheses of unknown truth by assigning probabilities. Bayesian probability uses prior probabilities that are updated based on new evidence into posterior probabilities using Bayes' theorem. The approach was pioneered by Thomas Bayes and Pierre-Simon Laplace.

Uploaded by

Veronica Gușan
Copyright
© © All Rights Reserved
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
92 views

Bayesian Probability

Bayesian probability is an interpretation of probability that represents a state of knowledge or personal belief. It allows for reasoning with hypotheses of unknown truth by assigning probabilities. Bayesian probability uses prior probabilities that are updated based on new evidence into posterior probabilities using Bayes' theorem. The approach was pioneered by Thomas Bayes and Pierre-Simon Laplace.

Uploaded by

Veronica Gușan
Copyright
© © All Rights Reserved
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 4

Bayesian probability 

is an interpretation of the concept of probability, in which, instead


of frequency or propensity of some phenomenon, probability is interpreted as reasonable
expectation[1] representing a state of knowledge[2] or as quantification of a personal belief.[3]
The Bayesian interpretation of probability can be seen as an extension of propositional logic that
enables reasoning with hypotheses;[4] that is, with propositions whose truth or falsity is unknown.
In the Bayesian view, a probability is assigned to a hypothesis, whereas under frequentist
inference, a hypothesis is typically tested without being assigned a probability.
Bayesian probability belongs to the category of evidential probabilities; to evaluate the probability
of a hypothesis, the Bayesian probabilist specifies a prior probability. This, in turn, is then updated
to a posterior probability in the light of new, relevant data (evidence).[5] The Bayesian
interpretation provides a standard set of procedures and formulae to perform this calculation.
The term Bayesian derives from the 18th-century mathematician and theologian Thomas Bayes,
who provided the first mathematical treatment of a non-trivial problem of statistical data
analysis using what is now known as Bayesian inference.[6]: 131  Mathematician Pierre-Simon
Laplace pioneered and popularized what is now called Bayesian probability.[6]: 97–98 

Contents

 1Bayesian methodology
 2Objective and subjective Bayesian probabilities
 3History
 4Justification of Bayesian probabilities
o 4.1Axiomatic approach
o 4.2Dutch book approach
o 4.3Decision theory approach
 5Personal probabilities and objective methods for constructing priors
 6See also
 7References
 8Bibliography

Bayesian methodology[edit]
Bayesian methods are characterized by concepts and procedures as follows:

 The use of random variables, or more generally unknown quantities,[7] to model all
sources of uncertainty in statistical models including uncertainty resulting from lack of
information (see also aleatoric and epistemic uncertainty).
 The need to determine the prior probability distribution taking into account the available
(prior) information.
 The sequential use of Bayes' formula: when more data become available, calculate the
posterior distribution using Bayes' formula; subsequently, the posterior distribution becomes
the next prior.
 While for the frequentist, a hypothesis is a proposition (which must be either true or false)
so that the frequentist probability of a hypothesis is either 0 or 1, in Bayesian statistics, the
probability that can be assigned to a hypothesis can also be in a range from 0 to 1 if the truth
value is uncertain.

Objective and subjective Bayesian probabilities [edit]


Broadly speaking, there are two interpretations of Bayesian probability. For objectivists, who
interpret probability as an extension of logic, probability quantifies the reasonable expectation that
everyone (even a "robot") who shares the same knowledge should share in accordance with the
rules of Bayesian statistics, which can be justified by Cox's theorem.[2][8] For
subjectivists, probability corresponds to a personal belief.[3] Rationality and coherence allow for
substantial variation within the constraints they pose; the constraints are justified by the Dutch
book argument or by decision theory and de Finetti's theorem.[3] The objective and subjective
variants of Bayesian probability differ mainly in their interpretation and construction of the prior
probability.

History[edit]
Main article: History of statistics §  Bayesian statistics
The term Bayesian derives from Thomas Bayes (1702–1761), who proved a special case of what
is now called Bayes' theorem in a paper titled "An Essay towards solving a Problem in the
Doctrine of Chances".[9] In that special case, the prior and posterior distributions were beta
distributions and the data came from Bernoulli trials. It was Pierre-Simon Laplace (1749–1827)
who introduced a general version of the theorem and used it to approach problems in celestial
mechanics, medical statistics, reliability, and jurisprudence.[10] Early Bayesian inference, which
used uniform priors following Laplace's principle of insufficient reason, was called "inverse
probability" (because it infers backwards from observations to parameters, or from effects to
causes).[11] After the 1920s, "inverse probability" was largely supplanted by a collection of
methods that came to be called frequentist statistics.[11]
In the 20th century, the ideas of Laplace developed in two directions, giving rise
to objective and subjective currents in Bayesian practice. Harold Jeffreys' Theory of
Probability (first published in 1939) played an important role in the revival of the Bayesian view of
probability, followed by works by Abraham Wald (1950) and Leonard J. Savage (1954). The
adjective Bayesian itself dates to the 1950s; the derived Bayesianism, neo-Bayesianism is of
1960s coinage.[12][13][14] In the objectivist stream, the statistical analysis depends on only the model
assumed and the data analysed.[15] No subjective decisions need to be involved. In contrast,
"subjectivist" statisticians deny the possibility of fully objective analysis for the general case.
In the 1980s, there was a dramatic growth in research and applications of Bayesian methods,
mostly attributed to the discovery of Markov chain Monte Carlo methods and the consequent
removal of many of the computational problems, and to an increasing interest in nonstandard,
complex applications.[16] While frequentist statistics remains strong (as seen by the fact that most
undergraduate teaching is still based on it [17][citation needed]), Bayesian methods are widely accepted and
used, e.g., in the field of machine learning.[18]

Justification of Bayesian probabilities[edit]


The use of Bayesian probabilities as the basis of Bayesian inference has been supported by
several arguments, such as Cox axioms, the Dutch book argument, arguments based on decision
theory and de Finetti's theorem.

Axiomatic approach[edit]
Richard T. Cox showed that Bayesian updating follows from several axioms, including
two functional equations and a hypothesis of differentiability.[8] The assumption of differentiability
or even continuity is controversial; Halpern found a counterexample based on his observation
that the Boolean algebra of statements may be finite.[19] Other axiomatizations have been
suggested by various authors with the purpose of making the theory more rigorous.[7]

Dutch book approach[edit]


The Dutch book argument was proposed by de Finetti; it is based on betting. A Dutch book is
made when a clever gambler places a set of bets that guarantee a profit, no matter what the
outcome of the bets. If a bookmaker follows the rules of the Bayesian calculus in the construction
of his odds, a Dutch book cannot be made.
However, Ian Hacking noted that traditional Dutch book arguments did not specify Bayesian
updating: they left open the possibility that non-Bayesian updating rules could avoid Dutch books.
For example, Hacking writes[20][21] "And neither the Dutch book argument, nor any other in the
personalist arsenal of proofs of the probability axioms, entails the dynamic assumption. Not one
entails Bayesianism. So the personalist requires the dynamic assumption to be Bayesian. It is
true that in consistency a personalist could abandon the Bayesian model of learning from
experience. Salt could lose its savour."
In fact, there are non-Bayesian updating rules that also avoid Dutch books (as discussed in the
literature on "probability kinematics"[22] following the publication of Richard C. Jeffreys' rule, which
is itself regarded as Bayesian[23]). The additional hypotheses sufficient to (uniquely) specify
Bayesian updating are substantial[24] and not universally seen as satisfactory.[25]

Decision theory approach[edit]


A decision-theoretic justification of the use of Bayesian inference (and hence of Bayesian
probabilities) was given by Abraham Wald, who proved that every admissible statistical procedure
is either a Bayesian procedure or a limit of Bayesian procedures.[26] Conversely, every Bayesian
procedure is admissible.[27]

Personal probabilities and objective methods for


constructing priors[edit]
Following the work on expected utility theory of Ramsey and von Neumann, decision-theorists
have accounted for rational behavior using a probability distribution for the agent. Johann
Pfanzagl completed the Theory of Games and Economic Behavior by providing an axiomatization
of subjective probability and utility, a task left uncompleted by von Neumann and Oskar
Morgenstern: their original theory supposed that all the agents had the same probability
distribution, as a convenience.[28] Pfanzagl's axiomatization was endorsed by Oskar Morgenstern:
"Von Neumann and I have anticipated ... [the question whether probabilities] might, perhaps more
typically, be subjective and have stated specifically that in the latter case axioms could be found
from which could derive the desired numerical utility together with a number for the probabilities
(cf. p. 19 of The Theory of Games and Economic Behavior). We did not carry this out; it was
demonstrated by Pfanzagl ... with all the necessary rigor".[29]
Ramsey and Savage noted that the individual agent's probability distribution could be objectively
studied in experiments. Procedures for testing hypotheses about probabilities (using finite
samples) are due to Ramsey (1931) and de Finetti (1931, 1937, 1964, 1970). Both Bruno de
Finetti[30][31] and Frank P. Ramsey[31][32] acknowledge their debts to pragmatic philosophy, particularly
(for Ramsey) to Charles S. Peirce.[31][32]
The "Ramsey test" for evaluating probability distributions is implementable in theory, and has kept
experimental psychologists occupied for a half century.[33] This work demonstrates that Bayesian-
probability propositions can be falsified, and so meet an empirical criterion of Charles S. Peirce,
whose work inspired Ramsey. (This falsifiability-criterion was popularized by Karl Popper.[34][35])
Modern work on the experimental evaluation of personal probabilities uses the
randomization, blinding, and Boolean-decision procedures of the Peirce-Jastrow experiment.
[36]
 Since individuals act according to different probability judgments, these agents' probabilities
are "personal" (but amenable to objective study).
Personal probabilities are problematic for science and for some applications where decision-
makers lack the knowledge or time to specify an informed probability-distribution (on which they
are prepared to act). To meet the needs of science and of human limitations, Bayesian
statisticians have developed "objective" methods for specifying prior probabilities.
Indeed, some Bayesians have argued the prior state of knowledge defines the (unique) prior
probability-distribution for "regular" statistical problems; cf. well-posed problems. Finding the right
method for constructing such "objective" priors (for appropriate classes of regular problems) has
been the quest of statistical theorists from Laplace to John Maynard Keynes, Harold Jeffreys,
and Edwin Thompson Jaynes. These theorists and their successors have suggested several
methods for constructing "objective" priors (Unfortunately, it is not clear how to assess the relative
"objectivity" of the priors proposed under these methods):
 Maximum entropy
 Transformation group analysis
 Reference analysis
Each of these methods contributes useful priors for "regular" one-parameter problems, and each
prior can handle some challenging statistical models (with "irregularity" or several parameters).
Each of these methods has been useful in Bayesian practice. Indeed, methods for constructing
"objective" (alternatively, "default" or "ignorance") priors have been developed by avowed
subjective (or "personal") Bayesians like James Berger (Duke University) and José-Miguel
Bernardo (Universitat de València), simply because such priors are needed for Bayesian practice,
particularly in science.[37] The quest for "the universal method for constructing priors" continues to
attract statistical theorists.[37]
Thus, the Bayesian statistician needs either to use informed priors (using relevant expertise or
previous data) or to choose among the competing methods for constructing "objective" priors.

You might also like