Convergence of the mean (first-order analysis) is insufficient
to guarantee desirable behavior of the LMS algorithm; the variancecould still be infinite. It is important to show that the variance
of the filter coefficients is finite, and to determine how closethe average squared error is to the minimum possible error using
an exact Wiener filter.
The minimum error is obtained using the Wiener filter
To analyze the average error in LMS, write
in terms of
, where
So we need to know
, which are the diagonal elements of the covariance
matrix of
, or
.
From the LMS update equation
we get
Note that
so
Note that the Patently False independence Assumption was invoked here.
To analyze
, we make yet another obviously false assumptioon that
and
are statistically independent. This is obviously false, since
. Otherwise, we get 4th-order terms in
in the product. These can be
dealt with, at the expense of a more complicated analysis, if aparticular type of distribution (such as Gaussian) is
assumed. See, for example
Gardner .
A questionable justification for this assumption is that as
,
becomes uncorrelated with
(if we invoke the original independence assumption),
which tends to randomize the error signal relative to
. With this assumption,
Now
so
Thus,
becomes
Now
if this system is stable and converges,
it converges to
So it is a diagonal matrix with all elements on the diagonal equal:
Then
Thus the error in the LMS adaptive filter after convergence is
is called the
misadjustment
factor . Oftern, one chooses
to select a desired
misadjustment factor, such as an error 10% higher than theWiener filter error.
2nd-order convergence (stability)
To determine the range for
for which
converges, we must determine
the
for which the matrix
difference equation converges.
The off-diagonal elements each evolve independently according to
These terms will decay to zero if
, or
The diagonal terms evolve according to
For the homoegeneous equation
for
positive,
will be strictly less than
for
or
or
This is a more rigorous bound than the first-order
bounds. Ofter engineers choose
a few times smaller than
this, since more rigorous analyses yield a slightly smallerbound.
is derived in some analyses assuming Gaussian
,
.