In probability theory, Boole's inequality, also known as the union bound, says that for any finite or countable set of events, the probability that at least one of the events happens is no greater than the sum of the probabilities of the individual events. This inequality provides an upper bound on the probability of occurrence of at least one of a countable number of events in terms of the individual probabilities of the events. Boole's inequality is named for its discoverer, George Boole.[1]
Formally, for a countable set of events A1, A2, A3, ..., we have
infty | |
{P}\left(cup | |
i=1 |
Ai\right)\le
infty | |
\sum | |
i=1 |
{P}(Ai).
In measure-theoretic terms, Boole's inequality follows from the fact that a measure (and certainly any probability measure) is σ-sub-additive.
Boole's inequality may be proved for finite collections of
n
For the
n=1
P(A1)\leP(A1).
For the case
n
n | |
{P}\left(cup | |
i=1 |
Ai\right)\le
n | |
\sum | |
i=1 |
{P}(Ai).
Since
P(A\cupB)=P(A)+P(B)-P(A\capB),
n+1 | |
P\left(cup | |
i=1 |
Ai\right)=
n | |
P\left(cup | |
i=1 |
Ai\right)+P(An+1)
n | |
-P\left(cup | |
i=1 |
Ai\capAn+1\right).
Since
n | |
{P}\left(cup | |
i=1 |
Ai\capAn+1\right)\ge0,
by the first axiom of probability, we have
n+1 | |
P\left(cup | |
i=1 |
Ai\right)\leP
n | |
\left(cup | |
i=1 |
Ai\right)+P(An+1),
and therefore
n+1 | |
P\left(cup | |
i=1 |
Ai\right)\le
n | |
\sum | |
i=1 |
P(Ai)+P(An+1)=
n+1 | |
\sum | |
i=1 |
P(Ai).
For any events in
A1,A2,A3,...
P\left(cupiAi\right)\leq\sumiP(Ai).
One of the axioms of a probability space is that if
B1,B2,B3,...
P\left(cupiBi\right)=\sumiP(Bi);
this is called countable additivity.
If we modify the sets
Ai
Bi=Ai-
i-1 | |
cup | |
j=1 |
Aj
we can show that
infty | |
cup | |
i=1 |
Bi=
infty | |
cup | |
i=1 |
Ai.
by proving both directions of inclusion.
Suppose
x\in
infty | |
cup | |
i=1 |
Ai
x\inAk
k
i<k\impliesx\notinAi
x\inBk=Ak-
k-1 | |
cup | |
j=1 |
Aj
infty | |
cup | |
i=1 |
Ai\subset
infty | |
cup | |
i=1 |
Bi
Next suppose that
x\in
infty | |
cup | |
i=1 |
Bi
x\inBk
k
Bk=Ak-
k-1 | |
cup | |
j=1 |
Aj
x\inAk
infty | |
cup | |
i=1 |
Bi\subset
infty | |
cup | |
i=1 |
Ai
By construction of each
Bi
Bi\subsetAi
B\subsetA,
P(B)\leqP(A).
So, we can conclude that the desired inequality is true:
P\left(cupiAi\right)=P\left(cupiBi\right)=\sumiP(Bi)\leq\sumiP(Ai).
Boole's inequality may be generalized to find upper and lower bounds on the probability of finite unions of events.[2] These bounds are known as Bonferroni inequalities, after Carlo Emilio Bonferroni; see .
Let
S1:=
n | |
\sum | |
i=1 |
{P}(Ai), S2:=
\sum | |
1\lei1<i2\len |
{P}(A | |
i1 |
\cap
A | |
i2 |
), \ldots, Sk:=
\sum | |
1\lei1< … <ik\len |
{P}(A | |
i1 |
\cap … \cap
A | |
ik |
)
for all integers k in .
Then, when
K\leqn
K | |
\sum | |
j=1 |
(-1)j-1Sj\geq
n | |
P\left(cup | |
i=1 |
Ai\right)=
n | |
\sum | |
j=1 |
(-1)j-1Sj
holds, and when
K\leqn
K | |
\sum | |
j=1 |
(-1)j-1Sj\leq
n | |
P\left(cup | |
i=1 |
Ai\right)=
n | |
\sum | |
j=1 |
(-1)j-1Sj
holds.
The equalities follow from the inclusion–exclusion principle, and Boole's inequality is the special case of
K=1
Let
E=
n | |
cap | |
i=1 |
Bi
Bi\in\{Ai,
c\} | |
A | |
i |
i=1,...,n
E
E
i
E
Ai
If
E=
n | |
cap | |
i=1 |
c | |
A | |
i |
E
Otherwise, assume
E
L
Ai
E
P(E)
K | |
\sum | |
j=1 |
(-1)j-1{L\choosej}P(E)
to the left side of the inequality. However, by Pascal's rule, this is equal to
K | |
\sum | |
j=1 |
(-1)j-1\left({L-1\choosej-1}+{L-1\choosej}\right)P(E)
which telescopes to
\left(1+{L-1\chooseK}\right)P(E)\geqP(E)
Thus, the inequality holds for all events
E
E
K | |
\sum | |
j=1 |
(-1)j-1Sj\geq
n | |
P\left(cup | |
i=1 |
Ai\right)
The proof for even
K
Suppose that you are estimating 5 parameters based on a random sample, and you can control each parameter separately. If you want your estimations of all five parameters to be good with a chance 95%, what should you do to each parameter?
Tuning each parameter's chance to be good to within 95% is not enough because "all are good" is a subset of each event "Estimate i is good". We can use Boole's Inequality to solve this problem. By finding the complement of event "all five are good", we can change this question into another condition:
P(at least one estimation is bad) = 0.05 ≤ P(A1 is bad) + P(A2 is bad) + P(A3 is bad) + P(A4 is bad) + P(A5 is bad)
One way is to make each of them equal to 0.05/5 = 0.01, that is 1%. In another word, you have to guarantee each estimate good to 99%(for example, by constructing a 99% confidence interval) to make sure the total estimation to be good with a chance 95%. This is called the Bonferroni Method of simultaneous inference.