In information theory, the error exponent of a channel code or source code over the block length of the code is the rate at which the error probability decays exponentially with the block length of the code. Formally, it is defined as the limiting ratio of the negative logarithm of the error probability to the block length of the code for large block lengths. For example, if the probability of error
Perror
e-n
n
\alpha
-lnPerror | |
n |
\alpha
n
The channel coding theorem states that for any ε > 0 and for any rate less than the channel capacity, there is an encoding and decoding scheme that can be used to ensure that the probability of block error is less than ε > 0 for sufficiently long message block X. Also, for any rate greater than the channel capacity, the probability of block error at the receiver goes to one as the block length goes to infinity.
Assuming a channel coding setup as follows: the channel can transmit any of
M=2nR
Let
n | |
X | |
i |
i
i
1
M
n | |
X | |
1 |
n | |
y | |
1 |
n | |
X | |
2 |
Perror 1\to=
\sum | |||||||
|
n\mid | |
Q(x | |
1 |
n) | |
x | |
2 |
>
n | |
p(y | |
1 |
\mid
n)). | |
x | |
1 |
The function
n\mid | |
1(p(y | |
1 |
n\mid | |
x | |
1 |
n)) | |
x | |
1 |
\left( |
| |||||||||||||||
|
\right)s
for
s>0
Perror 1\to\le
\sum | |||||||
|
n) | ||
Q(x | \left( | |
2 |
| ||||||||||||||||
|
\right)s.
Since there are a total of M messages, and the entries in the codebook are i.i.d., the probability that
n | |
X | |
1 |
M
n | |
X | |
1 |
Perror 1\to\leM\rho\left(
\sum | |||||||
|
n) | ||
Q(x | \left( | |
2 |
| ||||||||||||||||
|
\right)s\right)\rho
for any
0<\rho<1
n, | |
x | |
1 |
n | |
y | |
1 |
Perror 1\to\leM\rho
\sum | |||||||
|
\left(\sum | |||||||
|
n\mid | |
Q(x | |
1 |
n)] | |
x | |
1 |
1-s\rho\right)
\left(\sum | |||||||
|
n\mid | |
Q(x | |
1 |
n)] | |
x | |
2 |
s\right)\rho.
Choosing
s=1-s\rho
n | |
x | |
1 |
Perror 1\toany\leM\rho
\sum | |||||||
|
\left(\sum | |||||||
|
n\mid | |
Q(x | |
1 |
n)] | |
x | |
1 |
| ||||
\right)1+\rho.
Using the independence nature of the elements of the codeword, and the discrete memoryless nature of the channel:
Perror 1\toany\leM\rho
n | |
\prod | |
i=1 |
\sum | |
yi |
\left(\sum | |
xi |
Qi(xi)[pi(yi\mid
| ||||
x | ||||
i)] |
\right)1+\rho
Using the fact that each element of codeword is identically distributed and thus stationary:
Perror 1\toany\leM\rho\left(\sumy\left(\sumxQ(x)[p(y\mid
| ||||
x)] |
\right)1+\rho\right)n.
Replacing M by 2nR and defining
Eo(\rho,Q)=-ln\left(\sumy\left(\sumxQ(x)[p(y\midx)]1/(1+\rho)\right)1+\rho\right),
probability of error becomes
Perror\le\exp(-n(Eo(\rho,Q)-\rhoR)).
Q and
\rho
Er(R)=maxQmax\rhoEo(\rho,Q)-\rhoR.
The source coding theorem states that for any
\varepsilon>0
X
n
n
X1:n
n.(H(X)+\varepsilon)
X1:n
1-\varepsilon
Let
M=enR
M=m
n | |
X | |
1 |
n\mid | |
P(X | |
1 |
Am)
Am
m
n | |
X | |
1 |
m
n) | |
P(X | |
1 |
Thus, as an example of when an error occurs, supposed that the source sequence
n(1) | |
X | |
1 |
1
n(2) | |
X | |
1 |
n(1) | |
X | |
1 |
n(2)) | |
P(X | |
1 |
>
n(1)) | |
P(X | |
1 |
Let
Si
n(i) | |
X | |
1 |
P(Si)=
n(i)) | |
P(X | |
1 |
.
P(E)=\sumiP(E\midSi)P(Si).
P(E\midSi)
Let
Ai'
n(i') | |
X | |
1 |
n(i) | |
X | |
1 |
n(i')) | |
P(X | |
1 |
\geq
n(i)) | |
P(X | |
1 |
Xi,i'
i
i'
P(Ai')=P\left(Xi,i'cap
n(i')\right) | |
P(X | |
1 |
\geq
n(i))) | |
P(X | |
1 |
and using the fact that
P(Xi,i')=
1 | |
M |
P(Ai')=
1 | |
M |
n(i')) | |
P(P(X | |
1 |
\geq
n(i))) | |
P(X | |
1 |
.
A simple upper bound for the term on the left can be established as
\left[
n(i')) | |
P(P(X | |
1 |
\geq
n(i))) | |
P(X | |
1 |
\right]\leq\left(
| |||||||
|
\right)s
for some arbitrary real number
s>0.
n(i')) | |
P(P(X | |
1 |
>
n(i))) | |
P(X | |
1 |
1
0
n(i')) | |
P(X | |
1 |
\geq
n(i)) | |
P(X | |
1 |
,
| |||||||
|
\geq1
\left(
| |||||||
|
\right)s\geq0
for all possible source strings. Thus, combining everything and introducing some
\rho\in[0,1]
P(E\midSi)\leqP(cupiAi')\leq\left(\sumiP(Ai')\right)\rho\leq\left(
1 | |
M |
\sumi\left(
| |||||||
|
\right)s\right)\rho.
Where the inequalities follow from a variation on the Union Bound. Finally applying this upper bound to the summation for
P(E)
P(E)=\sumiP(E\midSi)P(Si)\leq\sumi
n(i)) | |
P(X | |
1 |
\left(
1 | |
M |
\sumi'\left(
| |||||||
|
\right)s\right)\rho.
Where the sum can now be taken over all
i'
P(E)\leq
1 | |
M\rho |
\sumi
n(i)) | |
P(X | |
1 |
1-s\left(\sumi'
n(i')) | |
P(X | |
1 |
s\right)\rho.
Now for simplicity let
1-s\rho=s
s=
1 | |
1+\rho |
.
s
i'
P(E)\leq
1 | |
M\rho |
\left(\sumi
n(i)) | |
P(X | |
1 |
| ||||
\right)1+.
M=enR
n(i) | |
X | |
1 |
P(E)\leq\exp\left(-n\left[\rhoR-ln\left(
\sum | |
xi |
| ||||
P(x | ||||
i) |
\right)(1+\rho)\right]\right).
The term in the exponent should be maximized over
\rho
Letting
E0(\rho)=ln\left(
\sum | |
xi |
| ||||
P(x | ||||
i) |
\right)(1+\rho),
Er(R)=max\rho\left[\rhoR-E0(\rho)\right].
R. Gallager, Information Theory and Reliable Communication, Wiley 1968