Lewis's triviality result explained
In the mathematical theory of probability, David Lewis's triviality result is a theorem about the impossibility of systematically equating the conditional probability
with the probability of a so-called conditional event,
.
Conditional probability and conditional events
The statement "The probability that if
, then
, is 20%" means (put intuitively) that event
may be expected to occur in 20% of the outcomes where event
occurs. The standard formal expression of this is
, where the
conditional probability
equals, by definition,
.
Beginning in the 1960s, several philosophical logicians—most notably Ernest Adams and Robert Stalnaker—floated the idea that one might also write
, where
is the conditional event "If
, then
".
[1] That is, given events
and
, one might suppose there is an event,
, such that
could be counted on to equal
, so long as
.
Part of the appeal of this move would be the possibility of embedding conditional expressions within more complex constructions. One could write, say,
, to express someone's high subjective degree of confidence ("75% sure") that either
, or else if
, then
. Compound constructions containing conditional expressions might also be useful in the programming of automated decision-making systems.
[2] How might such a convention be combined with standard probability theory? The most direct extension of the standard theory would be to treat
as an event like any other, i.e., as a set of outcomes. Adding
to the familiar
Venn- or
Euler diagram of
and
would then result in something like Fig. 1, where
are probabilities allocated to the eight respective regions, such that
.
For
to equal
requires that
, i.e., that the probability inside the
region equal the
region's proportional share of the probability inside the
region. In general the equality will of course not be true, so that making it reliably true requires a new constraint on probability functions: in addition to satisfying Kolmogorov's
probability axioms, they must also satisfy a new constraint, namely that
for any events
and
such that
.
Lewis's result
pointed out a seemingly fatal problem with the above proposal: assuming a nontrivial set of events, the new, restricted class of
-functions will not be closed under conditioning, the operation that turns probability function
into new function
, predicated on event
's occurrence. That is, if
, it will not in general be true that
as long as
. This implies that if rationality requires having a well-behaved probability function, then a fully rational person (or computing system) would become irrational simply in virtue of learning that arbitrary event
had occurred.
Bas van Fraassen called this result "a veritable bombshell" (van Fraassen|1976}}|1976, p. 273).
Lewis's proof is as follows. Let a set of events be non-trivial if it contains two possible events,
and
, that are mutually exclusive but do not together exhaust all possibilities, so that
,
,
, and
. The existence of two such events implies the existence of the event
, as well, and, if conditional events are admitted, the event
. The proof derives a contradiction from the assumption that such a minimally non-trivial set of events exists.
- Consider the probability of
after conditioning, first on
and then instead on its complement
.
-
gives
PA((A\cupB) → A)=P(((A\cupB) → A)\capA)/P(A)
. But also, by the new constraint on
-functions,
PA((A\cupB) → A)=PA((A\cupB)\capA)/PA(A\cupB)=
P((A\cupB)\capA\midA)/P(A\cupB\midA)=1/1=1
. Therefore,
P(((A\cupB) → A)\capA)=P(A)
.
-
gives
PA'((A\cupB) → A)=P(((A\cupB) → A)\capA')/P(A')
. But also,
PA'((A\cupB) → A)=PA'((A\cupB)\capA)/PA'(A\cupB)=
P((A\cupB)\capA\midA')/P(A\cupB\midA')=0/P(A\cupB\midA')=0
. (The mutual exclusivity of
and
ensures that
.) Therefore,
P(((A\cupB) → A)\capA')=0
.
- Instantiate the identity
P(X\capY)+P(X\capY')=P(X)
as
P(((A\cupB) → A)\capA)+P(((A\cupB) → A)\capA')=
. By the results from Step 1, the left side reduces to
, while the right side, by the new constraint on
-functions, equals
P((A\cupB)\capA)/P(A\cupB)=P(A)/P(A\cupB)
. Therefore,
, which means that
, which contradicts the stipulation that
. This completes the proof.
Graphical version
A graphical version of the proof starts with Fig. 2, where the
and
from Fig. 1 are now disjoint and
has been replaced by
.
[3] By the assumption that
and
are possible,
and
. By the assumption that together
and
do not together exhaust all possibilities,
. And by the new constraint on probability functions,
P((A\cupB) → A)=P(A\midA\cupB)=
P(A\cap(A\cupB))/P(A\cupB)=P(A)/P(A\cupB)
, which means that
(1)
Conditioning on an event involves zeroing out the probabilities outside the event's region and increasing the probabilities inside the region by a common scale factor. Here, conditioning on
will zero out
and
and scale up
and
, to
and
, respectively, and so
(2)
which simplifies to
Conditioning instead on
will zero out
and
and scale up
and
, and so
(3)
which simplifies to
From (2), it follows that
, and since
is the scaled-up value of
, it must also be that
. Similarly, from (3),
. But then (1) reduces to
, which implies that
, which contradicts the stipulation that
.
Later developments
In a follow-up article, noted that the triviality proof can proceed by conditioning not on
and
but instead, by turns, on each of a finite set of mutually exclusive and jointly exhaustive events
He also gave a variant of the proof that involved not total conditioning, in which the probability of either
or
is set to 1, but partial conditioning (i.e., Jeffrey conditioning), by which probability is incrementally shifted from
to
.
Separately, pointed out that even without conditioning, if the number of outcomes is large but finite, then in general
, being a ratio of two outputs of the
-function, will take on more values than any single output of the function can. So, for instance, if in Fig. 1
are all multiples of 0.01 (as would be the case if there were exactly 100 equiprobable outcomes), then
must be a multiple of 0.01, as well, but
need not be. That being the case,
cannot reliably be made to equal
.
also argued that the condition
caused acceptable
-functions to be implausibly sparse and isolated from one another. One way to put the point: standardly, any weighted average of two probability function is itself a probability function, so that between any two
-functions there will be a continuum of weighted-average
-functions along which one of the original
-functions gradually transforms into the other. But these continua disappear if the added
condition is imposed. Now an average of two acceptable
-functions will in general not be an acceptable
-function.
Possible rejoinders
Assuming that
holds for a minimally nontrivial set of events and for any
-function leads to a contradiction. Thus
can hold for any
-function only for trivial sets of events—that is the triviality result. However, the proof relies on background assumptions that may be challenged. It may be proposed, for instance, that the referent event of an expression like “
” is not fixed for a given
and
, but instead changes as the probability function changes. Or it may be proposed that conditioning on
should follow a rule other than
.
But the most common response, among proponents of the
condition, has been to explore ways to model conditional events as something other than subsets of a universe set of outcomes. Even before Lewis published his result, had modeled conditional events as
ordered pairs of sets of outcomes. With that approach and others in the same spirit, conditional events and their associated combination and complementation operations do not constitute the usual
algebra of sets of standard probability theory, but rather a more exotic type of structure, known as a
conditional event algebra.
References
- Hájek . Alan . 1989 . Probabilities of conditionals – Revisited . Journal of Philosophical Logic . 18 . 4 . 423–428 . 10.1007/BF00262944 . 30226421. 31355969 .
- Book: Hájek, Alan . 1994 . Triviality on the cheap? . Eells . Ellery . Skyrms . Brian . Probability and Conditionals . Cambridge UP . 113–140 . 978-0521039338.
- Book: Hájek . Alan . Hall . Ned . 1994 . The hypothesis of the conditional construal of conditional probability . Eells . Ellery . Skyrms . Brian . Probability and Conditionals . Cambridge UP . 75–111 . 978-0521039338.
- Lewis . David . 1976 . Probabilities of conditionals and conditional probabilities . Philosophical Review . 85 . 3 . 297–315. 10.2307/2184045 . 2184045.
- Lewis . David . 1986 . Probabilities of conditionals and conditional probabilities II . Philosophical Review . 95 . 4 . 581–589 . 10.2307/2185051 . 2185051.
- Schay . Geza . 1968 . An algebra of conditional events . Journal of Mathematical Analysis and Applications . 24 . 2 . 334–344 . 10.1016/0022-247X(68)90035-8.
- Book: van Fraassen, Bas C. . 1976 . Probabilities of conditionals . Harper . W. . Hooker . C. . Foundations and Philosophy of Epistemic Applications of Probability Theory . Foundations of Probability Theory, Statistical Inference, and Statistical Theories of Science, Volume I . D. Reidel . 261–308 . 978-9027706171.
Notes and References
- Hájek and Hall (Hájek|Hall|1994}}|1994) give a historical summary. The debate was actually framed as being about the probabilities of conditional sentences, rather than conditional events. However, this is merely a difference of idiom, so long as sentences are taken to express propositions and propositions are thought of as sets of possible worlds.
- Reading "If
, then
" as "Not
, unless also
" makes compounding straightforward, since
becomes equivalent to the Boolean expression
. However, this has the unsatisfactory consequence that
; then "If
, then
" is assigned high probability whenever
is highly unlikely, even if
's occurrence would make
highly likely. This is a version of what in logic is called a paradox of material implication.
- A proof starting with overlapping
and
, as in Fig. 1, would use mutually exclusive events
and
in place of
and
.