Bayes' Theorem is a theorem of probability theory originally stated by the Reverend Thomas Bayes. It can be seen as a way of understanding how the probability that a theory is true is affected by a new piece of evidence. It has been used in a wide variety of contexts, ranging from marine biology to the development of "Bayesian" spam blockers for email systems. In the philosophy of science, it has been used to try to clarify the relationship between theory and evidence. Many insights in the philosophy of science involving confirmation, falsification, the relation between science and pseudosience, and other topics can be made more precise, and sometimes extended or corrected, by using Bayes' Theorem. These pages will introduce the theorem and its use in the philosophy of science.

Begin by having a look at the theorem, displayed below. Then we'll look at the notation and terminology involved.

In this formula, T stands for a theory or hypothesis that we are interested
in testing, and E represents a new piece of evidence that seems to confirm or
disconfirm the theory. For any proposition S, we will use P(S) to stand for our
degree of belief, or "subjective probability," that S is true. In particular,
P(T) represents our best estimate of the probability of the theory we are
considering, prior to consideration of the new piece of evidence. It is known
as the *prior probability* of T.

What we want to discover is the probability that T is true supposing that our
new piece of evidence is true. This is a *conditional probability*, the
probability that one proposition is true provided that another proposition is
true. For instance, suppose you draw a card from a deck of 52, without showing
it to me. Assuming the deck has been well shuffled, I should believe that the
probability that the card is a jack, P(J), is 4/52, or 1/13, since there are
four jacks in the deck. But now suppose you tell me that the card is a face
card. The probability that the card is a jack, given that it is a face card, is
4/12, or 1/3, since there are 12 face cards in the deck. We represent this
conditional probability as P(J|F), meaning the probability that the card is a
jack *given that* it is a face card.

Using this idea of conditional probability to express what we want to use
Bayes' Theorem to discover, we say that P(T|E), the probability that T is true
given that E is true, is the *posterior probability* of T. The idea is
that P(T|E) represents the probability assigned to T *after* taking into
account the new piece of evidence, E. To calculate this we need, in addition to
the prior probability P(T), two further conditional probabilities indicating
how probable our piece of evidence is depending on whether our theory is or is
not true. We can represent these as P(E|T) and P(E|~T), where ~T is the
*negation* of T, i.e. the proposition that T is false.