Additive white Gaussian noise (AWGN) is a basic noise model used in information theory to mimic the effect of many random processes that occur in nature. The modifiers denote specific characteristics:
Wideband noise comes from many natural noise sources, such as the thermal vibrations of atoms in conductors (referred to as thermal noise or Johnson–Nyquist noise), shot noise, black-body radiation from the earth and other warm objects, and from celestial sources such as the Sun. The central limit theorem of probability theory indicates that the summation of many random processes will tend to have distribution called Gaussian or Normal.
AWGN is often used as a channel model in which the only impairment to communication is a linear addition of wideband or white noise with a constant spectral density (expressed as watts per hertz of bandwidth) and a Gaussian distribution of amplitude. The model does not account for fading, frequency selectivity, interference, nonlinearity or dispersion. However, it produces simple and tractable mathematical models which are useful for gaining insight into the underlying behavior of a system before these other phenomena are considered.
The AWGN channel is a good model for many satellite and deep space communication links. It is not a good model for most terrestrial links because of multipath, terrain blocking, interference, etc. However, for terrestrial path modeling, AWGN is commonly used to simulate background noise of the channel under study, in addition to multipath, terrain blocking, interference, ground clutter and self interference that modern radio systems encounter in terrestrial operation.
The AWGN channel is represented by a series of outputs
Yi
i
Yi
Xi
Zi
Zi
N
Zi
Xi
Zi\siml{N}(0,N)
Yi=Xi+Zi.
The capacity of the channel is infinite unless the noise
N
Xi
(x1,x2,...,xk)
1 | |
k |
k | |
\sum | |
i=1 |
2 | |
x | |
i |
\leqP,
P
C=max\left\{I(X;Y):fs.t.E\left(X2\right)\leqP\right\}
where
f
X
I(X;Y)
\begin{align} &I(X;Y)=h(Y)-h(Y\midX)\\[5pt] ={}&h(Y)-h(X+Z\midX)\\[5pt] ={}&h(Y)-h(Z\midX) \end{align}
But
X
Z
I(X;Y)=h(Y)-h(Z)
Evaluating the differential entropy of a Gaussian gives:
h(Z)=
1 | |
2 |
log(2\pieN)
Because
X
Z
Y
E(Y2)=E((X+Z)2)=E(X2)+2E(X)E(Z)+E(Z2)\leqP+N
From this bound, we infer from a property of the differential entropy that
h(Y)\leq
1 | |
2 |
log(2\pie(P+N))
Therefore, the channel capacity is given by the highest achievable bound on the mutual information:
I(X;Y)\leq
1 | |
2 |
log(2\pie(P+N))-
1 | |
2 |
log(2\pieN)
Where
I(X;Y)
X\siml{N}(0,P)
Thus the channel capacity
C
C=
1 | log\left(1+ | |
2 |
P | |
N |
\right)
Suppose that we are sending messages through the channel with index ranging from
1
M
M
n
R
R=
logM | |
n |
A rate is said to be achievable if there is a sequence of codes so that the maximum probability of error tends to zero as
n
C
Consider a codeword of length
n
N
N
Each codeword vector has an associated sphere of received codeword vectors which are decoded to it and each such sphere must map uniquely onto a codeword. Because these spheres therefore must not intersect, we are faced with the problem of sphere packing. How many distinct codewords can we pack into our
n
n(P+N)
\sqrt{nN}
rn
(n(P+N))n/2 | |
(nN)n/2 |
=2(n/2)
By this argument, the rate R can be no more than
1 | |
2 |
log\left(1+
P | |
N |
\right)
In this section, we show achievability of the upper bound on the rate from the last section.
A codebook, known to both encoder and decoder, is generated by selecting codewords of length n, i.i.d. Gaussian with variance
P-\varepsilon
Received messages are decoded to a message in the codebook which is uniquely jointly typical. If there is no such message or if the power constraint is violated, a decoding error is declared.
Let
Xn(i)
i
Yn
U
P
V
Ej
(Xn(j),Yn)
(n) | |
A | |
\varepsilon |
i ≠ j
An error therefore occurs if
U
V
Ei
P(U)
P(V)
n
P(U)
P(V)
\varepsilon
Xn(i)
Xn(j)
i ≠ j
Xn(i)
Yn
P(Ej)=2-n(I(X;Y)-3\varepsilon)
(n) | |
P | |
e |
(n) | |
\begin{align} P | |
e |
&\leqP(U)+P(V)+\sumjP(Ej)\\ &\leq\varepsilon+\varepsilon+\sumj2-n(I(X;Y)-3\varepsilon)\\ &\leq2\varepsilon+(2nR-1)2-n(I(X;Y)-3\varepsilon)\\ &\leq2\varepsilon+(23n\varepsilon)2-n(I(X;Y)-R)\\ &\leq3\varepsilon \end{align}
Therefore, as n approaches infinity,
(n) | |
P | |
e |
R<I(X;Y)-3\varepsilon
Here we show that rates above the capacity
C=
1 | |
2 |
log\left(1+
P | |
N |
\right)
Suppose that the power constraint is satisfied for a codebook, and further suppose that the messages follow a uniform distribution. Let
W
\hat{W}
W\longrightarrowX(n)(W)\longrightarrowY(n)\longrightarrow\hat{W}
Making use of Fano's inequality gives:
H(W\mid\hat{W})\leq
(n) | |
1+nRP | |
e |
=n\varepsilonn
\varepsilonn → 0
(n) | |
P | |
e |
→ 0
Let
Xi
\begin{align} nR&=H(W)\\ &=I(W;\hat{W})+H(W\mid\hat{W})\\ &\leqI(W;\hat{W})+n\varepsilonn\\ &\leqI(X(n);Y(n))+n\varepsilonn\\ &=h(Y(n))-h(Y(n)\midX(n))+n\varepsilonn\\ &=h(Y(n))-h(Z(n))+n\varepsilonn\\ &\leq
n | |
\sum | |
i=1 |
h(Yi)-h(Z(n))+n\varepsilonn\\ &\leq
n | |
\sum | |
i=1 |
I(Xi;Yi)+n\varepsilonn \end{align}
Let
Pi
Pi=
1 | |
2nR |
\sumw
2 | |
x | |
i(w) |
where the sum is over all input messages
w
Xi
Zi
Yi
N
2) | |
E(Y | |
i |
=Pi+N
And, if
Yi
h(Yi)\leq
1 | |
2 |
log{2\pie}(Pi+N)
Therefore,
\begin{align} nR&\leq\sum(h(Yi)-h(Zi))+n\varepsilonn\\ &\leq\sum\left(
1 | |
2 |
log(2\pie(Pi+N))-
1 | |
2 |
log(2\pieN)\right)+n\varepsilonn\\ &=\sum
1 | |
2 |
log\left(1+
Pi | |
N |
\right)+n\varepsilonn \end{align}
We may apply Jensen's equality to
log(1+x)
1 | |
n |
n | |
\sum | |
i=1 |
1 | log\left(1+ | |
2 |
Pi | |
N |
\right)\leq
1 | log\left(1+ | |
2 |
1 | |
n |
n | |
\sum | |
i=1 |
Pi | |
N |
\right)
Because each codeword individually satisfies the power constraint, the average also satisfies the power constraint. Therefore,
1 | |
n |
n | |
\sum | |
i=1 |
Pi | |
N |
,
which we may apply to simplify the inequality above and get:
1 | log\left(1+ | |
2 |
1 | |
n |
n | |
\sum | |
i=1 |
Pi | |
N |
\right)\leq
1 | log\left(1+ | |
2 |
P | |
N |
\right).
Therefore, it must be that
R\leq
1 | |
2 |
log\left(1+
P | |
N |
\right)+\varepsilonn
\varepsilonn → 0
In serial data communications, the AWGN mathematical model is used to model the timing error caused by random jitter (RJ).
The graph to the right shows an example of timing errors associated with AWGN. The variable Δt represents the uncertainty in the zero crossing. As the amplitude of the AWGN is increased, the signal-to-noise ratio decreases. This results in increased uncertainty Δt.
When affected by AWGN, the average number of either positive-going or negative-going zero crossings per second at the output of a narrow bandpass filter when the input is a sine wave is
\begin{align} &
positivezerocrossings | |
second |
=
negativezerocrossings | |
second |
\\[8pt] ={}&f0\sqrt{
| ||||||||||||
SNR+1 |
where
ƒ0 = the center frequency of the filter,
B = the filter bandwidth,
SNR = the signal-to-noise power ratio in linear terms.
In modern communication systems, bandlimited AWGN cannot be ignored. When modeling bandlimited AWGN in the phasor domain, statistical analysis reveals that the amplitudes of the real and imaginary contributions are independent variables which follow the Gaussian distribution model. When combined, the resultant phasor's magnitude is a Rayleigh-distributed random variable, while the phase is uniformly distributed from 0 to 2.
The graph to the right shows an example of how bandlimited AWGN can affect a coherent carrier signal. The instantaneous response of the noise vector cannot be precisely predicted, however, its time-averaged response can be statistically predicted. As shown in the graph, we confidently predict that the noise phasor will reside about 38% of the time inside the 1σ circle, about 86% of the time inside the 2σ circle, and about 98% of the time inside the 3σ circle.