In the mathematical theory of probability, the entropy rate or source information rate is a function assigning an entropy to a stochastic process.
For a strongly stationary process, the conditional entropy for latest random variable eventually tend towards this rate value.
A process
X
Hn(X1,X2,...Xn)
H(X):=\limn\tfrac{1}{n}Hn.
(an)n
a0=0
\Deltaak:=ak-ak-1
an={style\sum
n}\Delta | |
k=1 |
ak
n
n
While
X
H(X)
It can be thought of as a general property of stochastic sources - this is the subject of the asymptotic equipartition property.
A stochastic process also gives rise to a sequence of conditional entropies, comprising more and more random variables.For strongly stationary stochastic processes, the entropy rate equals the limit of that sequence
H(X)=\limnH(Xn|Xn-1,Xn-2,...X1)
H'(X)
Since a stochastic process defined by a Markov chain that is irreducible, aperiodicand positive recurrent has a stationary distribution, the entropy rate is independent of the initial distribution.
Pij
hi:=-\sumjPijlogPij
\displaystyleH(X)=\sumi\muihi,
\mui
In particular, it follows that the entropy rate of an i.i.d. stochastic process is the same as the entropy of any individual member in the process.
The entropy rate of hidden Markov models (HMM) has no known closed-form solution. However, it has known upper and lower bounds. Let the underlying Markov chain
X1:infty
Y1:infty
n\toinfty
The entropy rate may be used to estimate the complexity of stochastic processes. It is used in diverse applications ranging from characterizing the complexity of languages, blind source separation, through to optimizing quantizers and data compression algorithms. For example, a maximum entropy rate criterion may be used for feature selection in machine learning.[2]