spr052 ch13
spr052 ch13
Concern:
• To estimate the random process s(t) in terms of another related process x(ξ)
for a ≤ ξ ≤ b.
Theorem 13-1 The best linear estimator of s(t) in terms of {x(ξ) : a ≤ ξ ≤ b},
which is of the form b
ŝ(t) = h(α, t)x(α)dα
a
and which minimizes the MS error Pt = E[(s(t) − ŝ(t))2], satisfies
b
Rsx (t, s) = h(α, t)Rxx (α, s)dα for a ≤ s ≤ b.
a
Proof:
Pt = E[(s(t) − ŝ(t))2]
b b b
= E[s2(t)] + h(α, t)h(β, t)E[x(α)x(β)]dαdβ − 2 h(α, t)E[s(t)x(α)]dα
ba ba b a
= Rss(0) + h(α, t)h(β, t)Rxx(α, β)dαdβ − 2 h(α, t)Rsx(t, α)dα.
a a a
13-1 Introduction 13-2
Proof:
b
E[(s(t) − ŝ(t))x(ξ)] = E[s(t)x(ξ)] − h(α, t)E[x(α)x(ξ)]dα
b a
= Rsx(t, ξ) − h(α, t)Rxx(α, ξ)dα
a
= Rsx(t, ξ) − Rsx(t, ξ) = 0.
2
Orthogonality principle
• Linear estimator ŝ(t) that minimizes E[s(t) − ŝ(t), s(t) − ŝ(t)] = E[s(t) −
ŝ(t)2] should satisfy E[s(t) − ŝ(t), ŝ(t)] = 0.
• This may not be true for a non-linear estimator! (Note that the linear combi-
nation of {x(ξ), a ≤ ξ ≤ b} spans a hyperplane in an inner product space.)
Terminologies 13-4
Theorem 13-1 The best linear estimator of s(t) in terms of {s(ξ) : a ≤ ξ ≤ b},
which is of the form b
ŝ(t) = h(α, t)s(α)dα
a
and which minimizes the MS error P = E[(s(t) − ŝ(t))2], satisfies
b
Rss(t, s) = h(α, t)Rss(α, s)dα for a ≤ s ≤ b.
a
In addition,
b
Pt = Rss(0) − h(α, t)Rss(t, α)dα.
a
Theorem 13-1
E[(s(t) − ŝ(t))s(ξ)] = 0 for a ≤ ξ ≤ b.
Theorem 13-1 The best linear estimator of s(t) in terms of {xi (ξ) : a ≤ ξ ≤
b}ki=1, which is of the form
k b
ŝ(t) = hi(α, t)xi (α)dα
i=1 a
Proof: A different proof is used here. The optimal estimator should satisfy:
E[(s(t) − ŝ(t))xi (ξ)] = 0 for a ≤ ξ ≤ b and 1 ≤ i ≤ k.
Hence,
n
b
Rsxi (t, ξ) = h(α, t)Rx xi (α, ξ)dα.
=1 a
2
Theorem 13-1 Revisited 13-8
∂Rss(t1 − t2)
Rss (t1, t2) = = −Rss (t1 − t2)
∂t2
and
∂Rss (t1, t2)
Rss (t1, t2) = = −Rss (t1 − t2).
∂t1
Theorem 13-1 Revisited 13-9
Rss(λ)Rss (0) + Rss (λ)Rss (0) Rss(λ)
h
1 (a, a + λ) = h 1 (λ) = (0) + [R (0)]2
=
Rss(0)Rss ss Rss(0)
⇒
Rss (λ)Rss(0) − Rss (0)Rss(λ) Rss (λ)
2
h (a, a + λ) = h 2 (λ) = =
Rss(0)Rss (0) + [R (0)]2 (0)
Rss
ss
where it is reasonable to assume that Rss (0) = 0.
Rss(λ) Rss (λ)
⇒ ŝ(a + λ) = s(a) + s (a).
Rss(0) Rss(0)
Theorem 13-1 Revisited 13-10
Theorem 13-1 The best linear estimator of s(t) in terms of {x(ξ) : a ≤ ξ ≤ b},
which is of the form b
ŝ(t) = h(α, t)x(α)dα
a
and which minimizes the MS error Pt = E[(s(t) − ŝ(t))2], satisfies
b
Rsx (t, s) = h(α, t)Rxx (α, s)dα for a ≤ s ≤ b.
a
Concern
• To estimate, in the MS sense, s(t + λ) in terms of {s(t + kT )}N
k=−N with the
form
N
ŝ(t + λ) = ak s(t + kT ).
k=−N
we obtain
N
ak Rss(kT − nT ) = Rss(λ − nT ) for − N ≤ n ≤ N.
k=−N
In addition,
N
N
Pt = E s(t + λ) − ak s(t + kT ) s(t + λ) = Rss(0) − ak Rss(λ − kT ).
k=−N k=−N
Interpolation 13-13
where T = π/σ.
Example. Let Sss(ω) = 1 for |ω| < σ and zero, otherwise. Then,
∞ σ
1 1 sin(τ σ)
Rss(τ ) = Sss(ω)ejωτ dω = ejωτ dω = .
2π −∞ 2π −σ πτ
We thus derive for T σ = π that
N
sin((kT − nT )σ) sin((λ − nT )σ)
ak = for − N ≤ n ≤ N
π(kT − nT ) π(λ − nT )
k=−N
N
sin(π(k − n)) T sin(σ(λ − nT ))
⇔ ak = for − N ≤ n ≤ N
π(k − n) π(λ − nT )
k=−N
sin(σ(λ − nT ))
⇔ an = for − N ≤ n ≤ N
σ(λ − nT )
2
Example for Smoothing 13-14
Concern
• To estimate (real WSS) s(t) in terms of {x(ξ), −∞ < ξ < ∞} with WSS
x(t) = s(t) + v(t).
Using the orthogonality principle:
∞
E s(t) − h(α, t)x(t − α)dα x(t − ξ) = 0 for − ∞ < ξ < ∞,
−∞
or equivalently,
∞
Rsx(ξ) = h(α, t)Rxx (ξ − α)dα
−∞
∞
• ŝ[n] = k=1 h[k, n]s[n − k].
• Orthogonality principle: For all m ≥ 1,
∞
0 = E[(s[n] − ŝ[n])s[n − m]] = Rss[m] − h[k, n]Rss[m − k]
k=1
However,
∞
Rsx [m] = h[k]Rxx [m − k] only for m ≥ 1.
k=1
Solving Wiener-Höpf Equation Under Regularity 13-21
• By regularity,
∞
Rsi[m] = E[s[n]i[n − m]] = l[k]E{i[n − k]i[n − m]} = l[m].
k=0
• By noting that i[n] is the response of system 1/L[z] due to input s[n], and ŝ[n]
is the response of system L[z] − l[0] due to input i[n], we obtain:
1 l[0] limz↑∞ L[z]
H[z] = (L[z] − l[0]) = 1 − =1− .
L[z] L[z] L[z]
2
Solving Wiener-Höpf Equation Under Regularity 13-23
H[z]
1 i[n] -
s[n] - L[z] − L[0] - ŝ[n]
L[z]
5 − 4 cos(ω)
Example 13-3 (Slide 11-16) Sss[ω] =
10 − 6 cos(ω)
2z − 1
Then, L[z] = .
3z − 1
In this case,
limz↑∞ L[z] 2/3 2z − 2/3 −(1/6)z −1
H[z] = 1 − = 1− =1− = .
L[z] 2z − 1 2z − 1 1 − (1/2)z −1
3z − 1
Consequently,
1 1
ŝ[n] − ŝ[n − 1] = − s[n − 1]
2 6
or equivalently,
1 1
ŝ[n] = − s[n − 1] + ŝ[n − 1].
6 2
Kolmogorov-Szego MS Error Formula 13-25
Kolmogorov and Szego noted from the above result and Sss (ω) = |L[ejω ]|2 that
π π
1 1
P = l2[0] = exp log |L[ejω ]|2dω = exp log Sss[ω]dω .
2π −π 2π −π
This is named the Kolmogorov-Szego MS Error Formula.
Wide-Sense Markov of Order N 13-26
Concern
• To find the best linear estimator of s[n], in the MS sense, in terms of the
r-step-away entire past, i.e., {s[n − k]}k≥r .
∞
ŝ[n] = l[k]i[n − k]
k=r
is the best linear r-step predictor for a regular and stationary process
∞
r−1
s[n] = l[k]i[n − k] and P = l2[k].
k=0 k=0
Proof:
– A regular process can be represented as the response of a causal finite-energy
system due to a unit-power white-noise process i[n]. So,
∞
s[n] = l[k]i[n − k].
k=0
r-Step Predictor 13-28
∞
– Then, ŝ[n] = k=r h[k]s[n − k] can be written as
∞
ŝ[n] = g[k]i[n − k],
k=r
1
r−1
⇒ Hr [z] = 1 − l[k]z −k
L[z]
k=0
(1 − az −1 ) k −k
r−1
= 1− ba z
b
k=0
= 1 − (1 − az )(1 + az −1 + a2z −2 + · · · + ar−1z −(r−1) ) = ar z −r
−1
Analog Wiener-Höph Equation 13-30
Concern:
• To linearly estimate the random process s(t + λ) in terms of its entire past
{s(t − τ ), τ ≥ 0} in the MS sense.
• By regularity,
∞
Rsi (λ + τ ) = E[s(t)i(t − λ − τ )] = l(α)E{i(t − α)i(t − λ − τ )}dα = l(λ + τ ).
0
• By noting that i(t) is the response of system 1/L(s) due to input s(t), and
ŝ(t + λ) is the response of system l(τ + λ)1{τ ≥ 0} due to input i(t), we
obtain: ∞
1
H(ω) = l(τ + λ)e−jωτ dτ.
L(ω) 0
2
Solving Wiener-Höpf Equation Under Regularity 13-33
∞
−α|τ | −jωτ 4α2
⇒ Sss (ω) = 2αe e dτ = 2
−∞ α + ω2
4α2 4α2 2α 2α
⇒ Sss (s) = 2 = = = L(s)L(−s)
α + ω 2 ω=−js α2 − s2 (α + s) (α − s)
2α
⇒ L(s) =
α+s
2α
⇒ L(ω) =
α + jω
∞
1 2α −jωτ
⇒ l(τ ) = e dω = 2αe−ατ 1{τ ≥ 0}
2π −∞ α + jω
∞
1
⇒ H(ω) = l(τ + λ)e−jωτ dτ = e−αλ
L(ω) 0
⇒ h(τ ) = e−αλδ(τ )
∞
⇒ ŝ(t + λ) = h(τ )s(t − τ )dτ = e−αλ s(t).
0
Solving Wiener-Höpf Equation Under Regularity 13-34
and
∞ ∞
1 4α2
A(ω)H(ω)ejωτ dω = 2αe−α(|τ |+λ) = e −αλ jωτ
e dω
−∞ 2π −∞ (α2 + ω 2 )
= 2αe−α(|τ |+λ)
Concern
• To find the best linear estimator of s[n], in the MS sense, in terms of its N
most recent past, i.e., {s[n − k]}1≤k≤N .
• This is also named the forward predictor of order N .
Yule-Walker equations
• By orthogonality principle,
N
E s[n] − ak s[n − k] s[n − m] = 0 for 1 ≤ m ≤ N.
k=1
This yields
N
Rss[m] − ak Rss[m − k] = 0 for 1 ≤ m ≤ N
k=1
or equivalently,
Rss[1] Rss[0] Rss[−1] · · · Rss[1 − N ] a1
Rss[2] Rss[1] Rss[0] · · · Rss[2 − N ] a2
.. = .
.. ... ... ... ..
. .
Rss[N ] Rss[N − 1] Rss[N − 2] ··· Rss[0] aN
FIR Predictors 13-39
Hence,
% & % &
PN 0 · · · 0 Adj(DN+1) = 1 −a1 · · · −aN |DN+1|,
which implies PN |DN | = |DN+1|.
0, if for some k ≤ N, |Dk | = 0 and |Dk+1| = 0
• As a result, PN = |D |
N+1 , |DN | = 0.
|DN |
s[n] -
z −1 - -
z −1
? ?
−a1AA −aNAA
- ?-
⊕ - ?
⊕ - e[n]
are dependent on N ).
Implementation Structure of FIR Predictor 13-42
−k1
R B1
t - @
@
A −k1@@ @
R
@
-
z −1 r ⊕
- t -
C1
−k1
R B1 −k2
R
@ B2
t - @ @
@ @
A −k1@@ @ −k2@@ @
R
@ R
@
-
z −1 r ⊕
- t - -
z −1 r ⊕
- t -
C1 C2
– ěN [n] is the backward prediction error for predicting s[n − N ] in terms of
its most recent N futures. In other words,
N
(N)
ěN [n] = s[n − N ] − šN [n − N ] = s[n − N ] − ak s[n − N + k].
k=1
Implementation Structure of FIR Predictor 13-46
Derivation of kN
• From
(N−1) −1 (N−1)
ÊN−1 [z] = 1 − a1 z − · · · − aN−1 z −(N−1) ,
and
ĚN−1 [z] = z −N ÊN−1 [1/z],
we derive:
ÊN [z] = ÊN−1 [z] − kN z −1 ĚN−1 [z]
' (
(N−1) −1 (N−1) −(N−1)
= 1 − a1 z − · · · − aN−1 z
' (
−N (N−1) −(N−1) (N−1) −1
−kN z − a1 z − · · · − aN−1 z
' ( ' (
(N−1) (N−1) −1 (N−1) (N−1)
= 1 − a1 − kN aN−1 z − a2 − kN aN−2 z −2 − · · ·
' (
(N−1) (N−1)
− aN−1 − kN a1 z −(N−1) − kN z −N .
Comparing termwisely with
(N) (N)
ÊN [z] = 1 − a1 z −1 − · · · − aN z −N ,
we yield:
(N) (N−1) (N−1) (N)
ak = ak − kN aN−k for 1 ≤ k < N and aN = kN .
Implementation Structure of FIR Predictor 13-47
• It remains to solve kN :
Rss[0] Rss[1] Rss[2] ··· Rss[N ]
· · · Rss[N − 1]
% & ) * Rss[−1] Rss[0] Rss[1]
(N) (N)
PN 0 · · · 0 = 1 −a1 · · · −aN Rss[−2] Rss[−1] Rss[0] · · · Rss[N − 2]
... ... ... ...
Rss[−N ] Rss[1 − N ] Rss[2 − N ] ··· Rss[0]
implies
N−1
(N)
0 = Rss[N ] − ak Rss[N − k] − kN Rss[0]
k=1
'
N−1
(N−1) (N−1)
(
⇒ 0 = Rss[N ] − ak − kN aN−k Rss[N − k] − kN Rss[0]
k=1
N−1
(N−1)
Rss[N ] − k=1 ak Rss[N − k] 1
N−1
(N−1)
⇒ kN = N−1 (N−1) = Rss[N ] − ak Rss[N − k] ,
Rss[0] − k=1 aN−k Rss[N − k] PN−1
k=1
where the last step follows from the fact that Rss[N − k] = Rss[k − N ] (See Slide 13-39).
(N−1) N−1
• The above (blue-colored) formula gives kN from known PN−1 and {ak }k=1 .
Implementation Structure of FIR Predictor 13-48
Concern:
• A recursive algorithm to obtain kN and MS estimate error PN .
Levinson’s algorithm
(1)
• k1 = a1 = Rss[1]/Rss[0] and P1 = (1 − k12)Rss[0].
(N−1) N−1
• Assume that {ak }k=1 , kN−1 and PN−1 are known.
Then, it can be derived that
1
N−1
(N−1)
kN = Rss[N ] − ak Rss[N − k]
PN−1
k=1
2
PN = (1 − kN )PN−1
a(N−1) − kN a
(N−1)
N−k , 1 ≤ k ≤ N − 1
(N) k
ak =
k , k=N
N
Properties of FIR estimator 13-50
• P1 ≥ P2 ≥ · · · ≥ PN ≥ · · · ≥ 0.
• If PN > 0,
then |ki | < 1 for 1 ≤ i ≤ N ,
N (N)−k
and zi (the root of ÊN [z] = 1 − k=1 ak z ) satisfies |zi | < 1 for 1 ≤ i ≤ N .
k=0
In matrix form,
(0) (1) (n)
γ0 γ0 ··· γ0
% & % &
0
(1)
γ1 ···
(n)
γ1
i[0] i[1] · · · i[n] = s[0] s[1] · · · s[n] ... ...
. . . ...
(n)
0 0 · · · γn
% &
= s[0] s[1] · · · s[n] Γn+1
Remarks
% &
• This is similarly the Gram-Schmidt orthonormalization procedure for s[0] s[1] · · · s[n] .
• In terminologies, i[n] is called the Kalman innovations of s[n], and Γn+1 is
called the Kalman whitening filter of s[n].
• It can then be derived:
(0) (1) (n)
0 0 ··· 0
% & % & 0 1
(1)
···
(n)
1
s[0] s[1] · · · s[n] = i[0] i[1] · · · i[n] .. ... . . . ...
.
(n)
0 0 · · · n
% &
= i[0] i[1] · · · i[n] Ln+1
Kalman Innovations 13-54
Therefore,
ΓTn+1Rn+1Γn+1 = In+1,
where In+1 is the identity matrix.