0% found this document useful (0 votes)
108 views30 pages

Adaptive Filters

Adaptive Filters: Introduction: The conventional method of estimating a signal interrupted by additive interference is to pass it through a filter that tends to suppress the interference while leaving the signal relatively unchanged i.e. direct filtering
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
108 views30 pages

Adaptive Filters

Adaptive Filters: Introduction: The conventional method of estimating a signal interrupted by additive interference is to pass it through a filter that tends to suppress the interference while leaving the signal relatively unchanged i.e. direct filtering
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 30

Adaptive Filters: Introduction

The conventional method of estimating a signal interrupted by additive interference is to pass


it through a filter that tends to suppress the interference while leaving the signal relatively
unchanged i.e. direct filtering.
s + n Filter  sˆ

The design of such filters is the domain of optimal filtering, which originated with the
pioneering work of Wiener and was extended and enhanced by Kalman, Bucy and others.
Filters used for direct filtering can be either Fixed or Adaptive.
1. Fixed Filters - The design of fixed filters requires a priori knowledge of both the signal
and the interference, i.e. if we know the signal and interference beforehand; we can design a
filter that passes frequencies contained in the signal and rejects the frequency band occupied
by the interference.
2. Adaptive Filters - Adaptive filters, on the other hand, have the ability to adjust their
impulse response to filter out the correlated signal in the input. They require little or no a
priori knowledge of the signal and interference characteristics. (If the signal is narrowband
and interference broadband, which is usually the case, or vice versa, no a priori information is
needed; otherwise they require a signal (desired response) that is correlated in some sense to
the signal to be estimated.) Moreover adaptive filters have the capability of adaptively
tracking the signal under non-stationary conditions.
The concept of adaptive interference cancelling, an alternative method of estimating
signals corrupted by additive interference or interference. The method uses a "primary" input
containing the corrupted signal and a "reference" input containing interference correlated in
some unknown way with the primary interference. The reference input is adaptively filtered
and subtracted from the primary input to obtain the signal estimate. Adaptive filtering before
subtraction allows the treatment of inputs that are deterministic or stochastic, stationary or
time variable. Wiener solutions are developed to describe asymptotic adaptive performance
and output signal-to-interference ratio for stationary stochastic inputs, including single and
multiple reference inputs. These solutions show that when the reference input is free of signal
and certain other conditions are met interference in the primary input can be essentiany
eliminated without signal distortion. It is further shown that in treating periodic interference
the adaptive interference canceller acts as a notch filter with narrow bandwidth, infinite null,
and the capability of tracking the exact frequency of the interference; in this case the
canceller behaves as a linear, time-invariant system, with the adaptive filter converging on a
dynamic rather than a static solution. Experimental results are presented that illustrate the
usefulness of the adaptive interference cancelling technique in a variety of practical
applications. These applications include the cancelling of various forms of periodic
interference in electrocardiography.

~1~
The cancelling of periodic interference in speech signals, and the cancelling of broad-band
interference in the side-lobes of an antenna array. In further experiments it is shown that a
sine wave and Gaussian interference can be separated by using a reference input that is a
delayed version of the primary input. Suggested applications include the elimination of tape
hum or turntable rumble during the playback of recorded broad-band signals and the
automatic detection of very-low-level periodic signals masked by broad-band interference.
The purpose of adaptive interference cancellation is to improve the signal-to-interference
ratio (SIR) of a signal by removing interference from the signal that you receive. A typical
example of this application is the communication between a pilot, who is inside a jet aircraft,
and a ground control tower. A jet engine can produce interference of over 140 dB, but normal
human speech is below 50 dB. If you are in the ground control tower, you might have
difficulty hearing the pilot's speech clearly. In this situation, you can use adaptive filters to
remove the jet engine interference while retaining the pilot's speech.
The following figure shows a interference cancellation system.

Fig.1 Interference cancellation system


Subtracting interference from a received signal involves the risk of distorting the signal and if done
improperly, it may lead to an increase in the interference level. This requires that the interference
estimate ˆn should be an exact replica of n. If it were possible to know the relationship between n
and nˆ, or the characteristics of the channels transmitting interference from the interference source
to the primary and reference inputs are known, it would be possible to make nˆ a close estimate of n
by designing a fixed filter. However, since the characteristics of the transmission paths are not
known and are unpredictable, filtering and subtraction are controlled by an adaptive process. Hence
an adaptive filter is used that is capable of adjusting its impulse response to minimize an error signal,
which is dependent on the filter output. The adjustment of the filter weights, and hence the impulse
response, is governed by an adaptive algorithm. With adaptive control, interference reduction can
be accomplished with little risk of distorting the signal. Infect, Adaptive Interference Cancelling
makes possible attainment of interference rejection levels that are difficult or impossible to achieve
by direct filtering. The motivation for the topic “interference rejection using adaptive filter
techniques” we got by its inevitable applications. Its advantage lies in that, with no apriori
estimates of signal or interference, levels of interference rejection are attainable that would be
difficult or impossible to achieve by other signal processing methods of removing
interference. Its cost, inevitably, is that it needs two inputs - a primary input containing the
corrupted signal and a reference input containing interference correlated in some unknown
way with the primary interference. The reference input is adaptively filtered and subtracted

~2~
from the primary input to obtain the signal estimate. Adaptive filtering before subtraction
allows the treatment of inputs that are deterministic or stochastic, stationary or time-variable.
Adaptive Interference Cancellation algorithm has been applied to the removal of high-level
white interference from audio signals. Simulations and actual acoustically recorded signals
have been processed successfully, with excellent agreement between the results obtained
from simulations and the results obtained with acoustically produced data. A study of the
filter length required in order to achieve a desired interference reduction level in a hard-
walled room is presented. The performance of the algorithm in this application is described
and required modifications are suggested. A multichannel processing scheme is presented
which allows the adaptive filter to converge at independent rates in different frequency bands.
This is shown to be of particular use when the interfering interference is not white. Careful
implementation of the scheme allows the problem to be broken into several smaller ones
which can be handled by independent processors, thus allowing longer filter lengths to be
processed in real time.We have chosen this topic because in fixed coefficient transfer
function filters the coefficients was fixed it cannot be changed according to environmental
changes but adaptive filters can do so.

1.1 ADAPTIVE FILTER

An adaptive filter is a system with a linear filter that has a transfer function controlled by
variable parameters and a means to adjust those parameters according to an optimization
algorithm. Because of the complexity of the optimization algorithms, most adaptive filters
are digital filters. Adaptive filters are required for some applications because some
parameters of the desired processing operation (for instance, the locations of reflective
surfaces in a reverberant space) are not known in advance or are changing. The closed loop
adaptive filter uses feedback in the form of an error signal to refine its transfer function.
Generally speaking, the closed loop adaptive process involves the use of a cost function,
which is a criterion for optimum performance of the filter, to feed an algorithm, which
determines how to modify filter transfer function to minimize the cost on the next iteration.
The most common cost function is the mean square of the error signal.
As the power of digital signal processors has increased, adaptive filters have become much
more common and are now routinely used in devices such as mobile phones and other
communication devices, camcorders and digital cameras, and medical monitoring equipment.

An adaptive filter is defined by four aspects:

1. The signals being processed by the filter

2. The structure that defines how the output signal of the filter is computed from its input
signal

3. The parameters within this structure that can be iteratively changed to alter the filter’s
input-output relationship
~3~
4. The adaptive algorithm that describes how the parameters are adjusted from one time
instant to the next

Fig. 2 Block diagram of Adaptive filter

Figure shows a block diagram in which a sample from a digital input signal x(n) is fed into
advice, called an adaptive filter, that computes a corresponding output signal sample y(n)at
time n. For the moment, the structure of the adaptive filter is not important, except for the fact
that it contains adjustable parameters whose values affect how y (n) is computed. The output
signal is compared to a second signal d. (n), called the desired response signal, by subtracting
the two samples at time n. This difference signal, given by

e (n) = d(n)−y(n) 1

Is known as the error signal. The error signal is fed into a procedure which alters or adapts
the parameters of the filter from time n to time .n C 1/ in a well-defined manner. This process
of adaptation is represented by the oblique arrow that pierces the adaptive filter block in the
figure. As the time index is incremented, it is hoped that the output of the adaptive filter
becomes a better and better match to the desired response signal through this adaptation
process, such that the magnitude of e (n) decreases over time. In this context, what is meant
by “better” is specified by the form of the adaptive algorithm used to adjust the parameters of
the adaptive filter. In the adaptive filtering task, adaptation refers to the method by which the
parameters of the system are changed from time index n to time index n. The number and
types of parameters within this system depend on the computational structure chosen for the
system. We now discuss different filter structures that have been proven useful for adaptive
filtering tasks.

~4~
Adaptive Filter: Interference Mitigation techniques
Interference is a nuisance or disturbance during communication and it is unwanted. Noise
occurs because of many factors such as interference, delay, and overlapping. Noise problems
in the environment have gained attention due to the tremendous growth of technology that
has led to noisy engines, heavy machinery, high electromagnetic radiation devices and other
noise sources. For noise cancellation with the help of adaptive filter and employed for variety
of practical applications like the cancelling of various forms of periodic interference in
electrocardiography, the cancelling of periodic interference in speech signals, and the
cancelling of broad-band interference in the side-lobes of an antenna array. In sound signal or
speech signal, noise is very problematic because it will difficult to understanding of the
information. Speech is a very basic way for humans to convey information to one another
with a bandwidth of only 4 kHz; speech can convey information with the emotion of a human
voice. The speech signal has certain properties: It is a one-dimensional signal, with time as its
independent variable, it is random in nature, it is non-stationary, and i.e. the frequency
spectrum is not constant in time. Although human beings have an audible frequency range of
20Hz to 20 kHz, the human speech has significant frequency components only up to 4 kHz.
The most common problem in speech processing is the effect of interference noise in speech
signals. In the most of practical applications Adaptive filters are used and preferred over
fixed digital filters because adaptive filters have the property on the other hand, have the
ability to adjust their own parameters automatically and their design requires little or no a
priori knowledge of signal or noise characteristics. In this paper we have to use adaptive filter
for noise cancellation. The general configuration for an Adaptive filter system is shown in
Fig.1. It has two inputs: the primary input d(n), which represents the desired signal corrupted
with c undesired noise, and the reference signal x(n), which is the undesired noise to be
filtered out of the system. The goal of adaptive filtering systems is to reduce the noise
portion, and to obtain the uncorrupted desired signal. In order to achieve this, a reference of
the noise signal is needed and is called reference signal x(n). However, the reference signal is
typically not the same signal as the noise portion of the primary amplitude, phase or time.
Therefore the reference signal cannot be simply subtract from the primary signal to obtain the
desired portion at the output.

We have so many various techniques to study under adaptive filter techniques. Here we have
taken brief knowledge about

 Least mean square (LMS)


 Recursive least square (RLS),
 Minimum mean square error (MMSE).

The analysis of these three algorithms and result of these are shown below.

3.1) Least mean square error (LMS) Algorithm:-

~5~
Least mean square (LMS) algorithm is used to update the weight vector coefficient of the
adaptive filter or variable filter by using the mean square error.

x(i) wn y (i ) - + d(i)
Variable filter
+
e(n)

Updated Algorithm

fig. 5 Least mean square error(LMS )Algorithm

Where

x[n] = input signal

d[n] = desired signal

e[n] = error signal


^
y[n]  Output of the variable filter (estimated value)
Wn = transfer function of the variable filter

Main aim of the LMS algorithm is reduce the mean square error.

The mean square error is given by

n 2

 ( n )   e (i )
i 0 …………………………………………… (1)

It is the stochastic gradient descent method in the filter is only adapted based on the error
at the current time .

[Stochastic means non deterministic or randomly]

Gradient descent is the first order optimization algorithm. it is used to find the local minima
of a function.

The basic idea behind LMS filter is to approach the optimum filter weights (R-1P); by
updating the filter weight in a manner to converge the optimum filter weight.

Filter weight is calculated by the Wiener Hopf equation

Wopti = R-1P …………………………. (2)

Where
~6~
R = autocorrelation

P = cross correlation

For stationary processes the gradient descent adaptive filter converges to the solution to the
Wiener Hoff eqn. when

 2 
max

Where

µ = step size

 max = Greatest Eigen value of the autocorrelation matrix

Proof of Winner Hoff eqn.-

From eqn (1)

 (n)  E{ e(n) }
2
………………………. (3)
From fig.
^
e( n )  d ( n )  y ( n )
e(n)  d (n)  wn x( n)
T

………………… (4)

Put the value of e (n) in eqn. (3)

 (n)  E{( d (n)  wT x(n)) 2 }


 (n)  E{d (n) 2  2wT x(n)d (n)  ( wT x(n)) 2 }
 (n)  E{d (n) 2 }  2 EwT x(n)d (n)  E{( wT x(n)) 2 } ……………
(5)
For minimum value of the error. So necessary the different with respect to the weight vector
of eqn. (5)
 (n)
 0  2 x ( n) d ( n)  2 w x ( n) x ( n)  0
T *

w
 2 p  2 wR  0
2 wR  2 P
wR  P
woptim  R 1 P

~7~
The algorithm starts by assuming a small weight (zero is most cases) and each step, by
finding the gradient of the mean square error the weight are updated

Error

Weight vector

Graph no. 1 Gradient descent

Error functions are the quadratic function of the weight vector. If the mean square error
gradient is positive if implies, the error would keep increasing positively if the same weight is
used for further iteration which means we need to reduce the weights.

In the same way to if the gradient is negative we need to increase the weights so basic weight
updates eqn. is


wn 1  wn   ( n ) …………………………. (6)
2

Where
µ = step size
 (n) = mean square error
 = gradient operator
The (-ve) sign indicate that, we need to change the weights in opposite to that of the gradient
slope.

Derivation:-

d (n)

{w0,w1,w2,………….. wn}H _ +
^ +
(input) y[ n ] error e (n)
X (n)

Fig. 6 Adaptive Filter

The idea behind LMS filter is to use gradient descent to find filter weight w (n) which
minimize a cost function.
The cost function is given by

~8~
c(n)  E{ e(n) }
2
…………………….. (7a)
 (n)  E{ e(n) }
2
….………………… (7b)
Where e (n) is the error at the current sample ‘n’ and E {.} denotes the expected value.
Appling gradient descent means to take the partial derivatives with respect to the individual
entries of the filter coefficient vector.
w H  (n)  w H E{e(n)e * (n) }

H = denotes the hermitain transpose.

w H  (n)  2 E{w H e(n)e * (n)} ……………………


(8)
From fig. 4
e( n )  d ( n )  w x ( n )
H

Derivative with respect to weight vector.


So
w H e(n)  w H [d (n)  w x(n)]
H

w H e(n)  -x (n) ………………………… (9)


Put the value of eqn. (9) in the eqn. (8).
w H  (n)  2 E{e(n)e * (n)} ………………………
(10)
Substituted the value of eqn. (10) in the eqn. (6)

w( n 1)  wn  *
[-2E{x(n)x (n)}]
2
 wn  E{x(n)x * (n)}
w( n 1) …………………… (11)
This algorithm is not realizable until we know
E{x (n) x*(n)}
For most systems the expectation function E{x (n) x*(n)} must be approximation means the
expectation function E{x (n) x*(n)} is unknown.
Therefore it must be replaced with an estimation such as the sample means.
1 N 1
E{x(n)x * (n)}   x(n  i )e * (n  i ) ……………………
N i 0
(12)
Where N indicate the no. of sample we use for that estimate
Where N = I than we get
E{x (n) x*(n)} = x (n) e*(n)
For this case the update algorithm because
w( n1)  wn  x(n)e* (n) ………………….. (13)
th
This is known as LMS algorithm. But the k coefficient the algorithm because
w( n1) (k )  wn (k )  {x(n)x * (n - k)} …………………… (14)
th
The k coefficient requires only one multiplication and one addition. Therefore an LMS
adaptive filter having (p+1) coefficient requires (p+1) multiplication and (p+1) addition to
update the filter coefficient. In addition is necessary to compute the error
^
e( n )  d ( n )  y ( n )

~9~
And one multiplication is needed to from the product µe (n). Finally, (p+1) multiplication and
p addition are necessary to calculate the output y (n) of adaptive filter. Thus a total of (2p+3)
multiplication and (2p+2) addition per output point are required.
Table:-
The LMS algorithm for a pth order FIR adaptive filter:-
Parameter:
P = filter order
µ = step size
Initialization:-
For n=0, 1, 2, 3……..
y (n)  wnH x(n)
e( n )  d ( n )  y ( n )
wn 1  wn  x(n)e * (n)

Convergence of the LMS algorithm:-


Since wn is a vector of random variable, the convergence of the LMS algorithm must be
consider with in a statistical framework , therefore we will begin by assuming that x(n) and
d(n) are jointly wide sense stationary processes and will determine when the coefficients wn
converge in the mean to
w  R x1 rdx
lim E{wn }  w  R x1 rdx
n 

From eqn, (4) and (13) we get


w( n 1)  wn  [d(n) - w T x(n)]x * (n)
w( n 1)  wn  [ d * (n) - w T x * (n)]x(n)
w( n 1)  wn  d * (n)x(n) - w T x * (n)x(n)
Taking the expected value, we have
E[w( n 1) ]  E[wn ]  E[d * (n)x(n)] - E[w T x * (n)x(n)] ………………..
(14)
Eqn. (14) is not easy to evaluate it may be simplified considerably if we make the following
independence assumption.
Independence assumption:-
The data x (n) and the LMS weight vector wn are statistically independence eqn. (14)
E[ w( n 1) ]  E[ wn ]  rdx - E[ w T ]R x
or
E[ w( n 1) ]  ( I  rdx ) E[ wn ]  rdx

~ 10 ~
Property 3- the mean square error  (n) converges to a steady state value of

1
 ()   min   ex ()   min p
k
1 
k  0 2   k

And the LMS algorithm is said to converge in the mean square if and only if the step size 
satisfies the following two conditions

2
0 
max

p
k
 1
k  0 2   k

This is the condition that is required for the LMS algorithm to converge in the mean and last
equation guarantees that  () is positive. On further solving we will get

p
k
 2  
k 0
 ex ()   min k
p
k
1 
k  0 2   k

2  k  2 and equation may be simplified as


If   as is typically the case then
max

1 p
  k  1
2 k 0

2

tr ( R x )

2
If   it also follows that
max

~ 11 ~
1
 ()   min
1
1 tr ( R x )
2

And the excess mean square error given is approximately

1
tr ( R x )
2 1
 ex   min   min tr ( R x )
1 2
1  tr ( R x )
2

Thus for small  the excess mean square error is proportional to the step size  . Adaptive
filters may be described in terms of their misadjustment. Which is a normalized mean square
error that is defined as?

1
tr ( R x )
2 1
M   tr ( R x )
1 2
1  tr ( R x )
2

Definition- the misadjustment M is the ratio of the steady state excess mean square error to
the minimum mean square error,

 ex ()
M 
 min

Normalized LMS –

As the difficulty in LMS is the selection of the step size  . For stationary processes, the LMS
2
algorithm converges in the mean if 0<   , and converges in the mean square if 0<
 max
2
 , however since R x is generally unknown, then either max or R x must be
tr ( R x )
estimated in order to use these bounds. One way around this difficulty is to use the fact that
2
for stationary processes, tr(R x ) = (p+1) E{ x(n) }.

Therefore the condition for mean square convergence may be replaced with

2
0 
( p  1) E x(n)  2

Where power

  p 1 1  x(n  k )
p 2
^
2
E x ( n)
k 0

~ 12 ~
Which leads to this bound on step size for mean square convergence?

2
0  H
x ( n) x ( n)

A convenient way to incorporate this bound into the LMS adaptive filter is to use a step size
of the form

 
 ( n)  H
 2
x ( n) x ( n) x ( n)

Where  is a normalized step size with 0<   2 . Replacing  in the LMS weight vector
update equation with  (n) leads to the normalized LMS algorithm (NLMS) which is given
by

x ( n)
wn 1  wn   2
e( n )
x ( n)

2
Note that the effect of the normalized by x(n) is to alter the magnitude but not the
direction, of the estimated gradient vector. Therefore with the appropriate set of statistical
assumptions it may be shown that the normalized LMS converges in the mean square if 0< 
<2.

In the LMS algorithm, the correction that is applied to wn is proportional to the input vector
x (n).

Therefore when x (n) is large, the LMS algorithm experiences a problem with gradient
2
interference amplification .with the normalization of the LMS step size by x(n) in the
NLMS algorithm, however, this interference amplification problem is diminished. Although
the NLMS algorithm by passes the problem of interference amplification, we are now faced
with a similar problem that occurs when x(n) becomes too small. An alternative therefore,
is to use the following modification to the NLMS algorithm

x * ( n)
wn 1  wn   e( n )
  x ( n)
2

When  is some small positive number?


Compared with the LMS algorithm the NLMS algorithm requires additional computation to
2
evaluate the normalization term x(n) . However if this term evaluated recursively as
follows
x(n  1)
2
 x(n)  x(n  1)  x(n  p)
2 2 2

~ 13 ~
Then the extra computation involves only two operations one addition and one subtraction.

FLOW CHAT OF LMS ALGORITHM:-

~ 14 ~
3.2) Recursive Least Square Error (RLS) Algorithm:-
The Recursive Least Square Error (RLS) adaptive filter is algorithms which recursively find
the filter coefficient that minimize a weighted linear least square cost function relating to the
input signal.

Recursive least square (RLS) that aim to reduce the least square error.

n 2

 ( n )   e (i )
i 0

In a least square error required no statistical information about x (n) or d (n). So RLS input
signal are to be consider deterministic and it exhibits extremely fast convergence.

Minimizing a weighted least square error and derive an efficient algorithm for performing
this minimization known as Recursive Least Square.

Exponentially Weighted RLS:-

 wn (0) 
 w (1) 
 n 
Input x(i )  wn     y (i ) output
 
. 
 wn ( p )
 

Let us consider the design of an FIR adaptive wiener filter and find the filter coefficient
wn  wn (0), wn (1),........wn ( p)
T

That minimize at time n and the weighted least square error is

n 2

 (n)   n i e(i ) ………………… (1)


i 0

Where 0<   1 is an exponential weighting factor

i  0,1,2,3......n  n  i  n, n  1, n  2........0
i  0,1,2,3.........n  0  w n  n1  ........  0  1

~ 15 ~

x (i) Variable filter ( wn ) y (i ) - + d (i)
+
e (i)

wn

Updated Algorithm

Fig. 7 FIR Adaptive Winner filter

Output of the adaptive FIR filter


^
y(i)  wn n(i)
T

And error
^
e(i )  d (i )  y (i )
…………………………..
e(i )  d (i )  wn x (i )
T

(2)
^
Where d (i) is the desired signal and y (i ) filter output at time I using the latest set of filter
coefficient, wn (k )

*
Equation (2) is derivative with respect w (k ) n

e * (i ) 
 (d (i )  wn (k ) x(i  k )) *
T

 wn (k )  wn (k )
* *

e* (i )
  x* (i  k ) …………….. (3)
 wn (k )
*

*
Least square error is minimizing so eq. (1) is derivative with respect to w (k ) equal to zero
n
n
 (n) n
e (i )
*
 (n)   ni e(i )e * (i )   ni e(i ) 0
 wn (k )  wn (k )
* *
i 0 i 0

e* (i )
Put the of value from eq. (3)
 wn ( k )
*

n ………………………………. (4)
  n i e(i ) x * (i  k )  0
i 0

~ 16 ~
Where k= 0, 1, 2………p.

Note:-
T
Y (n) = wn x(n)

= wn (n) x(n), wn (n) x(n  1),........wk (n) x(n  k )

y (n)  wk (n) x(n  k )

n p


i 0
n i
{d (i )   wn (l ) x(n  l )}x * (i  k )  0
i 0
n n p


i 0
n i
d (i ) x (i  k )   
*

i 0
n i
 w (l ) x(n  l ) x
i 0
n
*
(i  k )  0

Interchanging the order of summation and rearranging terms we have


p
 n n i  n n i

i 0
wn (l ) 
 i 0
 x (i  l ) x *
(i  k )     d (i ) x (i  k )
 i 0
*
………………..

(5)

Eq. (5) express in matrix form

Rx (n) wn  rax (n) (This is the deterministic normal eqn.) ………………… (6)

Where Rx (n) is a (pX1) X (pX1) exponentially weighted deterministic autocorrelation


matrix for x (n)
n
R x (n)   ni x * (i ) x T (i )
i 0
n 1
  ni x * (i ) x T (i )  x * (i ) x T (i )
i 0
n 1
   ni 1 x * (i ) x T (i )  x * (i ) x T (i )
i 0

R x (n)  R x (n  1)  x * (i ) x T (i ) ……………………… (7)

Where the data vector x (i) is

x(i )  [ x(i), x(i  1),..............x(i  p)]T

And rdx (n) is the deterministic cross correlation between d (n) and x (n)

~ 17 ~
n
rdx (n)   n i x * (i )d (i )
i 0
n 1
  n i x * (i )d (i )  x * (i )d (i )
i 0
n 1
   n i 1 x * (i )d (i )  x * (i )d (i )
i 0

rdx (n)  rdx (n  1)  x* (i )d (i ) ………………… (8)

Derived the eqn. that define the optimum filter coefficient evaluation the minimum square
error from eqn. (1)

n 2

 ( n)    n i
e(i )
i 0
n
  n i e(i )e * (i )
i 0
n p
   e(i )[d (i )   wn (l ) x(i  l )]*
n i

i 0 i 0
n p n
   e(i )d (i )   w n (l ) n i e(i )x * (i  l )
n i * *

i 0 i 0 i 0

From eqn. (4) that the second term is zero and the minimum error is
n
{ (n)}min   n i e(i )d * (i )
i 0
n p
  n i {d (i )   wn (l ) x(i  l )}d * (i )
i 0 i 0
n p n
  n i d (i )   wn (l ) n i d * (i )x(i  l )
2

i 0 i 0 i 0

The minimum error in vector form


2
 (n) min  d (n)   r H
dx
( n) wn …………………(9)

2
d (n)  is the weighted norm of the vector?
Where

d (n) = [d(n),d(n-1)………d(0)] T

~ 18 ~
n
=  ni d (i )
2 2
d (n) 
i 0

And
n
r (n)   ni x(i  l )d * (i)
H
dx
i 0

Since R x (n) and rdx (n) both depends on n so scaling the deterministic normal eqn. directly
for each value of n. we will derive a recursive solution of the form

wn  wn1  wn1 ………………….


(10)

wn 1 is a correction that is applied to the solution at time n-1?


Where

From eqn. (6) we get

wn  R x1 (n)rdx ( x)

This recursion will be derived by first expressing rdx in terms of rdx (n  1) and then deriving
a recursion that allow us to evaluate R x1 ( n) in terms of R x1 ( n  1) and the new data vector.

First we will find the inverse autocorrelation R x1 ( n) is finding using Woodbury identity eqn.

A 1UV H A 1
H -1 -1
(A+UV ) =A -
1  V H A 1U

Where A= R x (n  1) and V=U=Y*(n)

Put the value of A, U and V

(Rx (n  1)) 1 x * (n)( x * (n) H (R x (n  1)) 1


(Rx (n  1)  x * (n)( x * (n) H ) 1  (R x (n  1)) 1 
1  x * (n)( x * (n) H (R x (n  1)) 1
2 R 1 x (n  1) x * (n)( x T (n) R 1 x (n  1)
(Rx (n  1)  x (n)( x (n))  ( R x (n  1)) 
* T 1 1 1

1  x * (n)( x T (n)(R x (n  1)) 1

We know that

[{x*(n)}H=xT(n)]

From eqn. (7)

R x (n)  R x (n  1)  x * (i ) x T (i )

~ 19 ~
So

2 R 1 x (n  1) x * (n)( x T (n) R 1 x (n  1)


[Rx(n)]-1= (1 R 1 x (n  1))  ………….(11)
1  x * (n)( x T (n)(R x (n  1)) 1

Rx-1(n) is denoted by p (n) for simplify notation so

P(n)=Rx-1(n)

And gain vector

1 p (n  1) x * (n)
g ( n)  ……………… (12)
1  1 p ( n  1) x * (n) x T ( n)

So eqn. (11) will be

p(n)   1 p(n  1)   1 p(n  1) g (n) x T (n)

p(n)  1 ( p(n  1)  p(n  1) g (n) x T (n)) ……………… (13)

From the eqn. (12)

g (n)[1  1 p(n  1) x * (n) x T (n)]  1 p(n  1) x * (n)

g (n)  1 g (n) p (n  1) x * (n) x T (n)  1 p (n  1) x * (n)

g (n)  1[ p(n  1)  g (n) p(n  1) x T (n)]x * (n) ……........... (14)

In eqn. (13) the multiple x*(n) on both side

x * (n) p(n)  1[ p(n  1)  g (n) p(n  1) x T (n)]x * (n) ………………(15)

From eqn. (15) and (14) we get

g ( n)  x * ( n) p ( n)

And thus the gain vector is the solution to the linear eqn.

Rx(n)g(n) =x*(n) …………… (16)

From eqn. (6) comparing the wn is replace by g(n) and rdx(n) is replace by x*(n).

To complete the recursive, we must derive the time update eqn. for the coefficient vector wn
with.

wn= rdx(n)p(n)

Put the value of rdx(n) from eqn. (8) then we get

~ 20 ~
w(n)  p(n)rdx (n  1)  d (n) p (n) x * (n)

But

g ( n)  x * ( n) p ( n)

w(n)  p(n)rdx (n  1)  d (n) g (n)

Put the value of p(n) from eqn. (13)

wn  [1 ( p (n  1)  p (n  1) g (n) x T (n))]rdx (n  1)  d (n) g (n)


wn  p (n  1)rdx (n  1)  p (n  1) g (n) x T (n))rdx (n  1)  d (n) g (n)
wn  wn 1  g (n)[d (n)  p (n  1) x T (n))rdx (n  1)]
wn  wn 1  g (n)[d (n)  wnT1 x(n)]

This may be written as

wn  wn 1  g (n) (n) …………………… (17)

Where

 (n)  d (n)  wn


T
1 x ( n)

This is a prior error and e (n) is the posterior error

e(n)  d (n)  wnT x(n)

Table – Recursive least square (for   1 )

Parameter:-

p=filter order

 = exponential weighted factor

 = value used to initialize p (0)

Initialization:-

W0=0

P (0) =  -1I

Computation:-

For n=1, 2, 3………..compute

Z (n) =p (n-1) x*(n)


~ 21 ~
z ( n)
g ( n) 
  x T ( n) z ( n)

 (n)  d (n)  wn


T
1 x ( n)

wn  wn 1  g (n) (n)

p(n)  1[( p(n  1)  g (n) z H (n))

Where  (n) is small , the current set of filter coefficient are close to their optimal value (in a
least square sense ) and only a small correction needs to be applied the coefficient .

Where  (n) is large the current set of filter coefficient are not performing must be applied to
the updated coefficient. The evolution of the gain vector g (n) and the inverse autocorrelation
matrix p (n), it is necessary to compute the product

Z (n) =p (n-1) x*(n)

This is known as filtered information vector and it is used to the calculate of both g (n) and p
(n).

First the initialization of the RLS algorithm since the RLS algorithm involves the recursive
updating of the vector wn and the inverse autocorrelation matrix p(n) initial condition for both
these terms are required.
1
 0 
p (0)     i x * (i ) x T (i )
i   p 

Evaluating the cross- correlation vector rdx(0) in the same manner.

0
rdx (0)  
i  p
i
d (i ) x * (i )

Than initialise

w0  p(0)rdx (0)

The advantage of this appropriate is that optimality is preserved at each step since the RLS
algorithm is initialized n=0 with the vector w0 that minimize the weighted least square error
 ( 0) .

Disadvantage:-

It is required the direct inversion of Rx(0), which required on the order of (p+1)3 operation.
And this approaches there will be a delay (p+1) sample before any updated are performed.

~ 22 ~
Another approach that many be used is to initialized the autocorrelation matrix

R x (0)  I

Where  is the small positive constant and I is the identity matrix.

p(0)   1 I

And to initialize the weight vector to zero w0=0

The disadvantage of this approach is that it introduces a bias in the least square solution.

The RLS algorithm require on the order of p2 operation, the evolution of z (n) requires (p+1)2
multiplication, computing the gain vector g (n) requires 2(p+1) multiplication, finding the a
prior error  (n) requires another p+1 multiplication and the update of the inverse
autocorrelation matrix p (n) require 2(p+1)2 multiplication for total 3(p+1)2+3(p+1).

Therefore, In RLS inverse in computational complexity over the LMS algorithm. But RLS
algorithm converges faster than the LMS algorithm and less sensitive to Eigen value.

~ 23 ~
FLOW CHART OF RLS ALGORITHM:-

START

Initialization of weight vector (W0)


and inverse auto correlation
matrix (P (0))

Input the value of  and x


(n)

Take the conjugate and


transpose of x (n)

Calculate the values of gain


vector Z (n)

Calculate the values of gain


vector (g (n))

Calculate the value of error e


(n)

Calculate the value of weight


vector Wn

If NO
n=N

YES
END

~ 24 ~
3.3) Minimum mean square error (MMSE) algorithm-:
A recurring theme in this text and in much of communication, control and signal
processing is that of making systematic estimates, predictions or decisions about some set of
quantities, based on information obtained from measurements of other quantities. This
process is commonly referred to as inference. Typically, inferring the desired information
from the measurements involves incorporating models that represent our prior knowledge or
beliefs about how the measurements relate to the quantities of interest. Inference about
continuous random variables and ultimately about random pro-cesses is the topic of this
chapter and several that follow. One key step is the introduction of an error criterion that
measures, in a probabilistic sense, the error between the desired quantity and our estimate of
it. Throughout our discussion in this and the related subsequent chapters, we focus primarily
on choosing our estimate to minimize the expected or mean value of the square of the error,
re-ferred to as a minimum mean-square-error (MMSE) criterion. We consider the MMSE
estimate without imposing any constraint on the form that the estimator takes. In Section 8.3
we restrict the estimate to be a linear combination of the measurements, a form of estimation
that we refer to as linear minimum mean-square-error (LMMSE) estimation. Later in the text
we turn from inference problems for continuous random variables to inference problems for
discrete random quantities, which may be numerically specified or may be non-numerical. In
the latter case especially, the various possible outcomes associated with the random quantity
are often termed hypotheses, and the inference task in this setting is then referred to as
hypothesis testing, i.e., the task of deciding which hypothesis applies, given measurements or
observations. The MMSE criterion may not be meaningful in such hypothesis testing
problems, but we can for instance aim to minimize the probability of an incorrect inference
regarding which hypothesis actually applies.

Estimation given a PDF:

Suppose x: Ω  R n is a random variable with pdf p x

One can predict or estimate the outcome as follows


• Given cost function c: R n × R n  R
^ ^
• pick estimate x to minimize Ec(x, x )

We will look at the cost function

^ ^ 2
C(x, x ) = x  x

Then the mean square error (MSE) is

^ 2 ^ 2
E( x  x ) =  x  x p x (x) dx

~ 25 ~
Minimizing the MSE

Let’s find the minimum mean-square error (MMSE) estimate of x; we need to solve

^ 2
min
^
E( x  x )
x

We have

^ 2 ^ ^
E( x  x )= E((x- x ) T ( x- x )

^T ^T ^
=E(x T x  2 x x  x x )

2 ^T ^T ^
=E x  2 x Ex  x x

^
Differentiating with respect to x gives the optimal estimate
^
x mmse = Ex

The MMSE estimate

The minimum mean-square error estimate of x is


^
^
x mmse= =ExEx
x mmse

Its mean square error is

^ 2

E( x  x mmse )=trace cov(x)

Since
2 2
^
E ( x  x mmse ) = E ( x  Ex )

The mean-variance decomposition

We can interpret this via the MVD. For any random variable z, we have
2 2 2
E z  E z  Ez  Ez

~ 26 ~
^
Applying this to z = x- x gives
2
2 2
^ ^
E ( x  x e ) = E ( x  Ex ) + E x Ex

• The first term is the variance of x; it is the error we cannot remove


^
• The second term is the bias of x ; the best we can do is make this zero.

~ 27 ~
4.3 Comparisons between LMS and RLS Algorithm:-

S.No. LMS Algorithm RLS Algorithm

Process Stochastic process Deterministic process

Error minimization Mean square error Least square error

Complexity Less More

Convergence speed Minimum Maximum

Error rejection Less More

Stability Less More

Signal to Interference Low High

Ratio (SIR)

Design Easier Difficult

Time taken Less More

Tracking Poor Good

~ 28 ~
References

1 B. Widrow, et al., “Adaptive Interference Cancelling: Principles and Applications”,


Proc. IEEE, vol. 63, pp.1692-1716, Dec. 1975.

2 Simon Haykin, Adaptive Filter Theory, Prentice Hall, II. Edition

3 John R. Glover, Jr., “Adaptive Interference Canceling Applied to Sinusoidal


Interferences”, IEEE Trans. ASSP, Vol. ASSP-25, No. 6, pp. 484-491, Dec.
1977.
4 Proakis, J.G. (1989). Digital communications, New York, McGraw
Hill. 5th Edition
5 Verdu, S. (1998). Multiuser detection. Cambridge University Press,
Cambridge, UK.
6 Wang, Xiaodong; and Poor, H.V. (2004). Wireless communication systems.
1st Ed, Pearson Education, Singapore.
7 Monsen P. (1984). MMSE equalization of interference on fading diversity
channels. IEEE Transaction on communications, 32(1), 5-12.

8 Hayes, Monson H. (1996). Statistical Digital Signal Processing and Modeling.


Wiley. ISBN 0-471-59431-8.

9 Haykin, Simon (2002). Adaptive Filter Theory. Prentice Hall. ISBN 0-13-048434-2.

10 Widrow, Bernard; Stearns, Samuel D. (1985). Adaptive Signal Processing. Englewood


Cliffs, NJ: Prentice Hall. ISBN 0-13-004029-0.

~ 29 ~
~ 30 ~

You might also like