LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD...

29
LEAST MEAN-SQUARE (LMS) ADAPTİVE FİLTERİNG

Transcript of LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD...

Page 1: LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD is a deterministic algorithm, in the sense that p and.

LEAST MEAN-SQUARE (LMS)ADAPTİVE FİLTERİNG

Page 2: LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD is a deterministic algorithm, in the sense that p and.

Steepest Descent

The update rule for SD is

where

or

SD is a deterministic algorithm, in the sense that p and R are assumed to be exactly known.

In practice we can only estimate these functions.

Page 3: LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD is a deterministic algorithm, in the sense that p and.

Basic Idea

The simplest estimate of the expectations is To remove the expectation terms and replace them with the

instantaneous values, i.e.

Then, the gradient becomes

Eventually, the new update rule is

Noexpectations,Instantaneous

samples!

Page 4: LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD is a deterministic algorithm, in the sense that p and.

Basic Idea

However the term in the brackets is the error, i.e.

then

is the gradient of instead of as in SD.

Page 5: LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD is a deterministic algorithm, in the sense that p and.

Basic Idea

Filter weights are updated using instantaneous values

Page 6: LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD is a deterministic algorithm, in the sense that p and.

Update Equation forMethod of Steepest Descent

Update Equation forLeast Mean-Square

Page 7: LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD is a deterministic algorithm, in the sense that p and.

LMS Algorithm

Since the expectations are omitted, the estimates will have a high variance. Therefore, the recursive computation of each tap weight in the LMS

algorithm suffers from a gradient noise.

In contrast to SD which is a deterministic algorithm, LMS is a member of the family of stochastic gradient descent algorithms.

LMS has higher MSE (J(∞)) compared to SD (Jmin) (Wiener Soln.) as n→∞ i.e., J(n) →J(∞) as n→∞ Difference is called the excess mean-square error Jex(∞)

The ratio Jex(∞)/ Jmin is called the misadjustment. Hopefully, J(∞) is a finite value, then LMS is said to be stable in the

mean square sense. LMS will perform a random motion around the Wiener solution.

Page 8: LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD is a deterministic algorithm, in the sense that p and.

LMS Algorithm

Involves a feedback connection. Although LMS might seem very difficult to work due the

randomness, the feedback acts as a low-pass filter or performs averaging so that the randomness can be filtered-out.

The time-constant of averaging is inversely proportional to μ. Actually, if is chosen small enough, the adaptive process is made

to progress slowly and the effects of the gradient noise on the tap weights are largely filtered-out.

Computational complexity of LMS is very low → very attractive Only 2M+1 complex multiplications and 2M complex additions

per iteration.

Page 9: LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD is a deterministic algorithm, in the sense that p and.

LMS Algorithm

Page 10: LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD is a deterministic algorithm, in the sense that p and.

Canonical Model

LMS algorithm for complex signals/with complex coef.s can be represented in terms of four separate LMS algorithms for real signals with cross-coupling between them.

Write the input/desired signal/tap gains/output/error in the complex notation

Page 11: LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD is a deterministic algorithm, in the sense that p and.

Canonical Model Then the relations bw. these expressions are

Page 12: LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD is a deterministic algorithm, in the sense that p and.

Canonical Model

Page 13: LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD is a deterministic algorithm, in the sense that p and.

Canonical Model

Page 14: LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD is a deterministic algorithm, in the sense that p and.

Analysis of the LMS Algorithm

Although the filter is a linear combiner, the algorithm is highly non-linear and violates superposition and homogenity

Assume the initial condition , then

Analysis will continue using the weight-error vector inputoutput

Page 15: LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD is a deterministic algorithm, in the sense that p and.

Analysis of the LMS Algorithm We have

Let

Then the update eqn. can be written as

Analyse convergence in an average sense Algorithm run many times→study their ensemble average behavior

Page 16: LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD is a deterministic algorithm, in the sense that p and.

Analysis of the LMS Algorithm

Using

It can be shown that

Small step sizeassumption

Here we use expectation,however, actually it isthe ensemble average!.

Page 17: LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD is a deterministic algorithm, in the sense that p and.

Small Step Size Analysis

Assumption I: step size is small → LMS filter acts like a low-pass filter with very low cut-off frequency.

Assumption II: Desired response is described by a linear multiple regression model that is matched exactly by the optimum Wiener filter

where eo(n) is the irreducible estimation error and

Assumption III: The input and the desired response are jointly Gaussian.

Page 18: LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD is a deterministic algorithm, in the sense that p and.

Small Step Size Analysis

Applying the similarity transformation resulting from the eigendecom.

on

i.e.

Then, we have

where

We do not have this term in Wiener filtering!.

Components of v(n)are uncorrelated!

Page 19: LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD is a deterministic algorithm, in the sense that p and.

Small Step Size Analysis

Components of v(n) are uncorrelated:

first order difference equation

Solution: Iterating from n=0

natural componentof v(n)

forced componentof v(n)

stochastic force

Page 20: LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD is a deterministic algorithm, in the sense that p and.

Learning Curves

Two kinds of learning curves Mean-square error (MSE) learning curve

Mean-square deviation (MSD) learning curve

Ensemble averaging → results of many (→∞) realizations are averaged.

What is the relation bw. MSE and MSD?

for small

Page 21: LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD is a deterministic algorithm, in the sense that p and.

Learning Curves

under the assumptions of slide 17. Excess MSE

LMS performs worse than SD, there is always an excess MSE

for small

← use

Page 22: LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD is a deterministic algorithm, in the sense that p and.

Learning Curves

Mean-square deviation D is lower-upper bounded by the excess MSE.

They have similar response: decaying as n grows

or

Page 23: LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD is a deterministic algorithm, in the sense that p and.

Convergence For small

Hence, for convergence

The ensemble-average learning curve of an LMS filter does not exhibit oscillations, rather, it decays exponentially to the const. value

or

Jex(n)

Page 24: LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD is a deterministic algorithm, in the sense that p and.

Misadjustment

Misadjustment, define

For small , from prev. slide

or equivalently

but

then

Page 25: LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD is a deterministic algorithm, in the sense that p and.

Average Time Constant

From SD we know that

but

then

Page 26: LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD is a deterministic algorithm, in the sense that p and.

Observations

Misadjustment is directly proportional to the filter length M, for a fixed mse,av inversely proportional to the time constant mse,av

slower convergence results in lower misadjustment. Directly proportional to the step size

smaller step size results in lower misadjustment.

Time constant is inversely proportional to the step size

smaller step size results in slower convergence

Large requires the inclusion of k(n) (k≥1) into the analysis Difficult to analyse, small step analysis is no longer valid, learning curve becomes more noisy

Page 27: LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD is a deterministic algorithm, in the sense that p and.

LMS vs. SD Main goal is to minimise the Mean Square Error (MSE) Optimum solution found by Wiener-Hopf equations.

Requires auto/cross-correlations. Achieves the minimum value of MSE, Jmin.

LMS and SD are iterative algorithms designed to find wo. SD has direct access to auto/cross-correlations (exact measurements)

can approach the Wiener solution wo, can go down to Jmin.

LMS uses instantenous estimates instead (noisy measurements)

fluctuates around wo in a Brownian-motion manner, at most J(∞).

Page 28: LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD is a deterministic algorithm, in the sense that p and.

LMS vs. SD

Learning curves SD has a well-defined curve composed of decaying exponentials

For LMS, curve is composed of noisy- decaying exponentials

Page 29: LEAST MEAN-SQUARE (LMS) ADAPTIVE FILTERING. Steepest Descent The update rule for SD is where or SD is a deterministic algorithm, in the sense that p and.

Statistical Wave Theory

As filter length increases, M→∞ Propagation of electromagnetic disturbances along a

transmission line towards infinity is similar to signals on n infinitely long LMS filter.

Finite length LMS filter (transmission line) Corrections have to be made at the edges to tackle reflections, As length increases reflection region decreases compared to the

total filter. Imposes a limit on the step size to avoid instability as M→∞

If the upper bound is exceeded, instability is observed.

Smax: maximum componentof the PSD S(ω) of the tap inputs u(n).