The Least mean squares filter solution converges to the Wiener filter solution, assuming that the unknown system is LTI and the noise is stationary. Both filters can be used to identify the impulse response of an unknown system, knowing only the original input signal and the output of the unknown system. By relaxing the error criterion to reduce current sample error instead of minimizing the total error over all of n, the LMS algorithm can be derived from the Wiener filter.
Given a known input signal
s[n]
x[n]
x[n]=
N-1 | |
\sum | |
k=0 |
hks[n-k]+w[n]
where
hk
w[n]
The model system
\hat{x}[n]
\hat{x}[n]=
N-1 | |
\sum | |
k=0 |
\hat{h}ks[n-k]
where
\hat{h}k
The error between the model and the unknown system can be expressed as:
e[n]=x[n]-\hat{x}[n]
The total squared error
E
E=
infty | |
\sum | |
n=-infty |
e[n]2
E=
infty | |
\sum | |
n=-infty |
(x[n]-\hat{x}[n])2
E=
infty | |
\sum | |
n=-infty |
(x[n]2-2x[n]\hat{x}[n]+\hat{x}[n]2)
Use the Minimum mean-square error criterion over all of
n
\nablaE=0
\partialE | |
\partial\hat{h |
i}=0
i=0,1,2,...,N-1
\partialE | |
\partial\hat{h |
i}=
\partial | |
\partial\hat{h |
i}
infty | |
\sum | |
n=-infty |
[x[n]2-2x[n]\hat{x}[n]+\hat{x}[n]2]
Substitute the definition of
\hat{x}[n]
\partialE | |
\partial\hat{h |
i}=
\partial | |
\partial\hat{h |
i}
infty | |
\sum | |
n=-infty |
[x[n]2-
N-1 | |
2x[n]\sum | |
k=0 |
\hat{h}ks[n-k]+
N-1 | |
(\sum | |
k=0 |
2 | |
\hat{h} | |
ks[n-k]) |
]
Distribute the partial derivative:
\partialE | |
\partial\hat{h |
i}=
infty | |
\sum | |
n=-infty |
[-2x[n]s[n-i]+
N-1 | |
2(\sum | |
k=0 |
\hat{h}ks[n-k])s[n-i]]
Using the definition of discrete cross-correlation:
Rxy(i)=
infty | |
\sum | |
n=-infty |
x[n]y[n-i]
\partialE | |
\partial\hat{h |
i}=-2Rxs[i]+
N-1 | |
2\sum | |
k=0 |
\hat{h}kRss[i-k]=0
Rearrange the terms:
Rxs[i]=
N-1 | |
\sum | |
k=0 |
\hat{h}kRss[i-k]
i=0,1,2,...,N-1
This system of N equations with N unknowns can be determined.
The resulting coefficients of the Wiener filter can be determined by:
W=
-1 | |
R | |
xx |
Pxs
Pxs
x
s
By relaxing the infinite sum of the Wiener filter to just the error at time
n
The squared error can be expressed as:
E=(d[n]-y[n])2
Using the Minimum mean-square error criterion, take the gradient:
\partialE | |
\partialw |
=
\partial | |
\partialw |
(d[n]-y[n])2
Apply chain rule and substitute definition of y[n]
\partialE | |
\partialw |
=2(d[n]-y[n])
\partial | |
\partialw |
(d[n]-
N-1 | |
\sum | |
k=0 |
\hat{w}kx[n-k])
\partialE | |
\partialwi |
=-2(e[n])(x[n-i])
Using gradient descent and a step size
\mu
w[n+1]=w[n]-\mu
\partialE | |
\partialw |
which becomes, for i = 0, 1, ..., N-1,
wi[n+1]=wi[n]+2\mu(e[n])(x[n-i])
This is the LMS update equation.