Provides a summary of the rules of inductive reasoning, as advocated by E.T. Jaynes. Includes probability rules, and decision theory.
“Probability theory is nothing but common sense reduced to calculation” (Laplace).
Introduction
This module was adapted from E.T. Jaynes’ manuscript entitled: “Probability Theory with Applications to Science and Engineering – A Series of Informal Lectures”, 1974. The entire manuscript is available at
(External Link) .
A second and significantly expanded edition of this manuscript is available on Amazon. The first 3 chapters of the second edition are available here
(External Link) .
Deductive logic (boolean algebra)
Denote propositions by A, B, etc., their denials by
,
etc. Define the logical product and logical sum by
“Both A and B are true”
“At least one of the propositions, A, B are true”
Deductive reasoning then consists of applying relations such as
;
;
if
then
.
Inductive logic (probability theory)
Inductive logic is the extension of deductive logic, describing the reasoning of an idealized “robot”, who represents degrees of plausibility of a logical proposition by real numbers:
= probability of A, given B.
We use the original term “robot” advocated by Jaynes, it is intended to mean the use of inductive logic that follows a set of consistent rules that can be agreed upon. In this formulation of probability theory, conditional probabilities are fundamental. The elementary requirements of common sense and consistency determine these basic rules of reasoning (see Jaynes for the derivation).
In these rules, one can think of the proposition
being the prior information that is available to assign probabilities to logical propositions, but these rules are true without this interpretation.
Rule 1:
Rule 2:
Rule 3:
Rule 4: If
are mutually exclusive and exhaustive, and information
is indifferent to tem; i.e. if
gives no preference to one over any other then:
(principle of insufficient reason)
From rule 1 we obtain Bayes’ theorem:
From Rule 3, if
are mutually exclusive,
If in addition, the
are exhaustive, we obtain the chain rule:
Prior probabilities
The initial information available to the robot at the beginning of any problem is denoted by
.
is then the prior probability of
. Applying Bayes’ theorem to take account of new evidence
yields the posterior probability
. In a posterior probability we sometimes leave off the
for brevity:
Prior probabilities are determined by Rule 4 when applicable; or more generally by the principle of maximum entropy.
Decision theory
Enumerate the possible decisions
and introduce the loss function
representing the “loss” incurred by making decision
if
is the true state of nature. After accumulating new evidence E, make that decision
which minimizes the expected loss over the posterior distribution of
:
Choose the decision
which minimizes