Conditional mutual information explained
In probability theory, particularly information theory, the conditional mutual information[1] [2] is, in its most basic form, the expected value of the mutual information of two random variables given the value of a third.
Definition
For random variables
,
, and
with
support sets
,
and
, we define the conditional mutual information as
This may be written in terms of the expectation operator:
I(X;Y|Z)=EZ[DKL(P(X,Y)|Z\|PX|Z ⊗ PY|Z)]
.
Thus
is the expected (with respect to
)
Kullback–Leibler divergence from the conditional joint distribution
to the product of the conditional marginals
and
. Compare with the definition of
mutual information.
In terms of PMFs for discrete distributions
For discrete random variables
,
, and
with
support sets
,
and
, the conditional mutual information
is as follows
} p_Z(z) \sum_ \sum_ p_(x,y|z) \log \fracwhere the marginal, joint, and/or conditional
probability mass functions are denoted by
with the appropriate subscript. This can be simplified as
In terms of PDFs for continuous distributions
For (absolutely) continuous random variables
,
, and
with
support sets
,
and
, the conditional mutual information
is as follows
} \bigg(\int_ \int_ \log \left(\frac\right) p_(x,y|z) dx dy \bigg) p_Z(z) dzwhere the marginal, joint, and/or conditional
probability density functions are denoted by
with the appropriate subscript. This can be simplified as
Some identities
Alternatively, we may write in terms of joint and conditional entropies as[3]
\begin{align}
I(X;Y|Z)&=H(X,Z)+H(Y,Z)-H(X,Y,Z)-H(Z)\\
&=H(X|Z)-H(X|Y,Z)\\
&=H(X|Z)+H(Y|Z)-H(X,Y|Z).
\end{align}
This can be rewritten to show its relationship to mutual information
usually rearranged as
the chain rule for mutual information
or
I(X;Y|Z)=I(X;Y)-(I(X;Z)-I(X;Z|Y)).
Another equivalent form of the above is
\begin{align}
I(X;Y|Z)&=H(Z|X)+H(X)+H(Z|Y)+H(Y)-H(Z|X,Y)-H(X,Y)-H(Z)\\
&=I(X;Y)+H(Z|X)+H(Z|Y)-H(Z|X,Y)-H(Z)
\end{align}.
Another equivalent form of the conditional mutual information is
\begin{align}
I(X;Y|Z)=I(X,Z;Y,Z)-H(Z)
\end{align}.
Like mutual information, conditional mutual information can be expressed as a Kullback–Leibler divergence:
I(X;Y|Z)=DKL[p(X,Y,Z)\|p(X|Z)p(Y|Z)p(Z)].
Or as an expected value of simpler Kullback–Leibler divergences:
} p(Z=z) D_[p(X,Y|z) \| p(X|z)p(Y|z) ],
} p(Y=y) D_[p(X,Z|y) \| p(X|Z)p(Z|y) ].
More general definition
A more general definition of conditional mutual information, applicable to random variables with continuous or other arbitrary distributions, will depend on the concept of regular conditional probability.[4]
Let
be a
probability space, and let the random variables
,
, and
each be defined as a Borel-measurable function from
to some state space endowed with a topological structure.
Consider the Borel measure (on the σ-algebra generated by the open sets) in the state space of each random variable defined by assigning each Borel set the
-measure of its preimage in
. This is called the
pushforward measure
The
support of a random variable is defined to be the
topological support of this measure, i.e.
Now we can formally define the conditional probability measure given the value of one (or, via the product topology, more) of the random variables. Let
be a measurable subset of
(i.e.
) and let
Then, using the
disintegration theorem:
akP(M|X=x)=\limU
{akP(\{X\inU\})}
rm{and} akP(M|X)=\intMdakP(\omega|X=X(\omega)),
where the limit is taken over the open neighborhoods
of
, as they are allowed to become arbitrarily smaller with respect to
set inclusion.
Finally we can define the conditional mutual information via Lebesgue integration:
I(X;Y|Z)=\int\Omegalog
l(
| dakP(\omega|X,Z)dakP(\omega|Y,Z) |
dakP(\omega|Z)dakP(\omega|X,Y,Z) |
r)
dakP(\omega),
where the integrand is the logarithm of a Radon–Nikodym derivative involving some of the conditional probability measures we have just defined.
Note on notation
In an expression such as
and
need not necessarily be restricted to representing individual random variables, but could also represent the joint distribution of any collection of random variables defined on the same
probability space. As is common in
probability theory, we may use the comma to denote such a joint distribution, e.g.
Hence the use of the semicolon (or occasionally a colon or even a wedge
) to separate the principal arguments of the mutual information symbol. (No such distinction is necessary in the symbol for
joint entropy, since the joint entropy of any number of random variables is the same as the entropy of their joint distribution.)
Properties
Nonnegativity
It is always true that
,for discrete, jointly distributed random variables
,
and
. This result has been used as a basic building block for proving other
inequalities in information theory, in particular, those known as Shannon-type inequalities. Conditional mutual information is also non-negative for continuous random variables under certain regularity conditions.
[5] Interaction information
Conditioning on a third random variable may either increase or decrease the mutual information: that is, the difference
, called the
interaction information, may be positive, negative, or zero. This is the case even when random variables are pairwise independent. Such is the case when: