H2-Optimal Control - Lec8
H2-Optimal Control - Lec8
H2-Optimal Control - Lec8
The H2 -Norm
Consider the LTI system with state-space description
x = Ax + Bw, z = Cx
and with transfer matrix
T (s) = C(sI A)1 B of dimension p q.
Here w is a disturbance input and z is an output of interest that is
desired to be small. A quantification of the effect of the input w onto
the output z is the so-called H2 -norm of the transfer matrix.
Definition 1 Let T have all its poles in the open left half-plane.
The H2 -norm of T is defined as
s Z
1
kT k2 :=
kT (i)k2F d
2
with k.kF denoting the Frobenius matrix norm.
2/82
Relation to Hardy-Spaces
The Hardy-space H2pq consists of all matrices S of dimension p q
whose elements are analytic functions in the open right half-plane s.th.
Z
1
2
kS(r + i)k2F d is finite.
kSk2 := sup
r>0 2
For all such functions one can show that the limit
T(i) := lim S(r + i)
r&0
F (i) =
eit F (t) dt.
0
Computation
It is a beautiful fact that the H2 -norm of a stable transfer matrix can
be computed algebraically on the basis of a state-space realization.
Theorem 2 Let A be Hurwitz and T (s) = C(sI A)1 B. Then:
1. kT k22 = tr CPc C T where APc + Pc AT + BB T = 0.
2. kT k22 = tr B T Po B where AT Po + Po A + C T C = 0.
R T
Proof of 2.: Recall that Po = 0 eA t C T CeAt dt. Using Parsevals theorem we infer
Z
1
2
tr
T (i) T (i) d =
kT k2 =
2
Z
= tr
[CeAt B]T [CeAt B] dt = tr B T Po B .
0
Inequality Characterization
The following characterizations of a bound on the H2 -norm will be useful
for controller synthesis proofs.
Lemma 3 A is Hurwitz and kT k22 < iff there exists X 0 with
AT X + XA + C T C 0 and tr(B T XB) .
Proof of if. Since AT X + XA 0 and X 0, A is Hurwitz. By
subtracting the Lyapunov equation for Po we infer AT (X Po ) +
(X Po )A 0 and thus Po X since A is Hurwitz. This implies
tr(B T Po B) tr(B T XB) < and thus kT k22 < .
Proof of only if. Consider the solution P of the perturbed Lyapunov
equation AT P + P A + C T C + I = 0 with > 0. Clearly P Po
for 0. Since tr(B T Po B) , we can hence fix some 0 with
tr(B T P0 B) . Then X = P0 does the job.
6/82
Deterministic Interpretation
Let ek be the standard unit vector of dimension q and let zk (.) denote
the response of
x = Ax, z = Cx, x(0) = Bek
Recall that this is just the response to an impulse in the k-th input.
Since zk (t) = CeAt Bek we infer
Z
Z
AT t T
At
T
T T
e C Ce dt Bek = eTk B T Po Bek .
zk (t) zk (t) dt = ek B
0
0
7/82
Random Vectors
Uncertain outcomes of experiments are modeled by random vectors
x = (x1 xn )T . Here x is a vector of n random variables x1 , . . . , xn
and is characterized by its distribution function Fx : Rn R which
admits the following interpretation: If (1 n )T Rn , the probability
for the event x1 1 , ..., xn n to happen is given by Fx (1 , ..., n ).
Fx (1 , ..., n ) has the density fx (1 , ..., n ) in case that
Z 1 Z n
Fx (1 , ..., n ) =
9/82
Wiener-Processes
In control the disturbance w on slide 2 is often considered as white
noise, which is associated with irregular signals having a flat spectrum.
Loosely speaking, this boils down to viewing w as the derivative of the
(normalized) Wiener-process or Brownian motion. It would take us too
far astray and it is not required for our purposes to develop the whole
theory of stochastic differential equations (based on Ito calculus).
Instead let us just collect some basic facts that are required in the sequel.
1) There exists a Wiener-process W (t) for t 0 with intensity 1:
Initialized at zero: W (0) = 0 with probability one.
Independent increments: For all 0 t1 t2 t3 t4 , the random
variables W (t2 ) W (t1 ) and W (t4 ) W (t3 ) are independent.
Gaussian increments: For all 0 t1 t2 , the increment W (t2 )
W (t1 ) is Gaussian with expectation 0 and variance t2 t1 = 1|t2 t1 |.
10/82
10
1 e 2t
2t
Wiener-Processes
2) For a square integrable real-valued function f on [a, b] (0 a b),
Z b
the random variable
f (t) dW (t)
a
k=1
Wiener-Processes
3) The integral is a Gaussian random variable. Just by direct calculation
for step-functions and taking limits, one proves for x, y L2 [a, b] that
Z b
E
x(t) dW (t) = 0,
a
Z
E
Z
x(t) dW (t)
Z b
y( ) dW ( ) =
x(t)y(t) dt.
E
x(t) dW (t)
y( ) dW ( ) = 0.
a
Wiener-Processes
The q-dimensional Wiener-process W = col(W1 , . . . , Wq ) is a vector
of q Wiener processes W1 , . . . , Wq (with intensity 1) that are pairwise
independent.
If X and Y are matrix-valued functions of dimension p q with square
integrable elements on [a, b] (0 a b), then the random vectors
Z b
Z b
x=
X(t) dW (t), y =
Y (t) dW (t)
a
of dimension p are defined elementwise. They both have mean zero and
and their correlation matrix is given by
Z b
T
E[xy ] =
X(t)Y (t)T dt.
a
W (t) =
W ( ) d =
dW ( ) for t 0.
0
15/82
At1
cov(, )e
AT t2
Z
+
t1
eA(t1 ) BB T eA
T (t )
2
for 0 t1 t2 .
The specialization to A = 0 and = 0 implies x(t) = BW (t) and thus
E[BW (t)] = 0 as well as E[BW (t)W (t)T B T ] = tBB T .
16/82
17/82
Example
300
200
0
100
2
4
0
10
0
5
Squared H2norm: 1
400
300
200
100
5
4
0
10
0
5
5
18/82
Colored Noise
Definition 7 We say that w is colored noise if there exists
B,
C)
with eig(A)
C such that w is the output of
(A,
x + B
W
, w = C x, x(0) = 0.
x = A
This is also expressed as w emerging through filtering white noise with
1 B.
=
+ W,
=
x
(0)
0
B
0 A
x
z = C DC
.
x
19/82
Remarks
R t A(t
dW ( ) for t 0, we can assume w.l.o.g.
)B
Since w(t)
= 0 Ce
B,
C)
is minimal, and that the coloring filter T (and not
that (A,
its realization) captures all properties of w.
This raises the question
about how to determine such coloring filters in practice.
The response of a linear system to colored noise is reduced to that
for white noise and for the series interconnection of the system and
the coloring filter.
In the theory of stochastic differential equations it is shown that our
definitions are consistent. More precisely, the response of
x = Ax + B w,
z = Cx + Dw,
x(0) =
can indeed be represented as it should, namely as
Z t
At
x(t) = Ce +
CeA(t ) B w(
) d + Dw(t)
for t 0,
0
20/82
Coloring Filters
Clearly E[w(t)]
C P eA C T
21/82
Proof
For t 0 and > 0 the filter state satisfies
Z t
T
B
T eAT (t) d =
E[
x(t + )
x(t) ] =
eA(t+ ) B
0
=e
T A
B e d eA P for t .
eA B
T
B
T eAT (t) d =
E[
x(t + )
x(t) ] =
eA(t+ ) B
0
Z
=
t+
T A
T
T
eA B
B e ( ) d P eA for t .
22/82
Coloring Filters
Theorem 10 The spectral density of w is given by
R(i)
= T(i)T(i) .
Hence R(i)
is Hermitian and positive semi-definite for all R.
Z
T
i
A
1
P C T =
e e
= C P
d + (iI A)
0
+ (iI A)
1 P C T .
= C P (iI A)
B
T = 0 also satisfies the equation
Now note that AP + P AT + B
P + P (iI A)
=B
B
T and thus
(iI A)
+ (iI A)
1 P = (iI A)
1 B
B
T (iI A)
.
P (iI A)
Plugging this into the above relation proves the formula.
23/82
Example
Consider the colored noise on slide 18. The filter has the realization
p 2p
T =
1
0
and hence P = 1. Therefore
p| |
R( ) = e
and R(i)
=
2p
2p
2p
.
= 2
p + i p i
+ p2
|T~(i)|
10
0.8
0.6
10
0.4
0.2
0
0.1
0.05
0.05
0.1
10
10
10
10
10
24/82
density R(i)
of the process under scrutiny. One then approximates the
experimentally determined spectral density by G(i) where G(s) is a
strictly proper real rational function without poles in C0 and with
G(i) = G(i) and G(i) < 0 for all R.
()
25/82
Example
Consider the transfer function G(s) =
1s2
.
s4 13s2 +36
We have
1 + 2
> 0 for R.
4 + 13 2 + 36
Hence, by Theorem 11, G has a spectral factorization. This can be seen
directly as follows. The numerator and denominator of G(s) can be
factorized as
G(i) =
z
y
27/82
z
P
y
u
K
A
BCK
Bw
A B
= BK C AK BK Dw .
C D
Cz Dz CK
0
Problem: Minimize the H2 -norm kC(sI A)1 Bk2 of the controlled
closed-loop system over all controllers which render A Hurwitz.
28/82
LQG-Control
One particular scenario is worth mentioning. Consider the system
1 + Bu
x = Ax + B1 W
1 . Suppose the measurewith control input u and process noise B1 W
2 where W1 and W2 are indements Cx are corrupted by white noise W
pendent Wiener processes. The measured output hence is
2.
y = Cx + D2 W
As in LQ-control, we are interested in keeping the linear combinations
C1 x and D1 u of the states and the controls small. Hence we choose
C1 x
z=
D1 u
as the performance output. The LQG-control goal is to find a stabilizing
controller which minimizes lim tr cov(z(t), z(t)), which equals
t
lim E[z(t)T z(t)] = lim E x(t)T C1T C1 x(t) + u(t)T D1T D1 u(t) .
t
29/82
LQG-Control
Let us define the following generalized plant
x = Ax + B1 0 w + Bu
C1
0
z =
x+
u
0
D1
y = Cx + 0 D2 w.
Finding a stabilizing output-feedback controller which minimizes the
asymptotic variance of z for a while noise w is the classical so-called
Linear-Quadratic-Gaussian (LQG) optimal control problem.
In view of our preparation this is merely a special case of H2 -control!
Note that the process and measurement noises are defined via the
Wiener-processes B1 W1 and D2 W2 with auto-covariances
E B1 W1 (t)W1 (t)T B1T = B1 B1T t, E D2 W2 (t)W2 (t)T D2T = D2 D2T t.
Hence B1 B1T and D2 D2T are the intensities of these processes.
30/82
LQG-Control
1 is not white but colored noise w1 , one absorbs the related coloring
If W
1 B
into the generalized plant and solves the
filter T (s) = C(sI
A)
H2 -problem for the weighted generalized plant
0 0
x
A B1 C
x
B
=
+
w+
u
x
x
0
B 0
0 A
C1 0
x
0
z =
+
u
0 0
x
D1
x
y = C 0
+ 0 D2 w.
x
The block-diagram of the weighted generalized plant is more instructive:
w1
z
P
y
w2
w1
u
31/82
32/82
min
F , ABF Hurwitz
Proof
Let > opt . Then there exists F for which A BF is Hurwitz and
such that k(Cz Dz F )(sI A + BF )1 Bw k22 < . Due to Lemma 3
we can choose X 0 with
(A BF )T X + X(A BF ) + (Cz Dz F )T (Cz Dz F ) 0
and tr(BwT XBw ) < . Exploiting DzT Cz = 0 and DzT Dz = I allows to
rearrange to
AT X + XA + CzT Cz XBB T X + (B T X F )T (B T X F ) 0.
Therefore AT X + XA + CzT Cz XBB T X 0 and thus P X by
the result on slide 68. This implies that
tr(BwT P Bw ) 4 tr(BwT XBw ) < .
Since > opt was arbitrary we conclude
tr(BwT P Bw ) opt .
34/82
Proof
Let us now choose F = B T P . Note that A BF = A BB T P is
Hurwitz, just because of the choice of P as the stabilizing solution of
the ARE. Moreover with the ARE we trivially have
AT P + P A + CzT Cz P BB T P + (B T P F )T (B T P F ) = 0.
As above this can be re-arranged to
(A BF )T P + P (A BF ) + (Cz Dz F )T (Cz Dz F ) = 0
Viewed as a Lyapunov equation, this shows that
k(Cz + Dz F )(sI A BF )1 Bw k22 = tr(BwT P Bw ).
This proves that tr(BwT P Bw ) is attained by the stabilizing state-feedback
gain F = B T P which indeed confirms
tr(BwT P Bw ) = opt
and optimality of F .
35/82
Example
For the following model of a DC-motor ([F] p.20)
0 1
0
x =
x+
u, x =
.
0 1
1
4
6
Reference (red) and output
10
10
2
0
2
0
36/82
Example
Assume that the control input is affected by colored noise with filter
134.2/(s + 10) and having H2 -norm 30 in order to clearly display its
effect. We get the following substantially deteriorated response:
Control action
200
0
200
0
4
6
Reference (red) and output
10
10
2
0
2
0
37/82
Example
With the coloring filter included in the system description, design an
optimal H2 -state-feedback gain for the cost function
2000||2 + 0.3|u|2 .
The feed-forward gain is adjusted appropriately. The cost has been tuned
so that the noise-free response resembles the one we obtained earlier:
Control action
200
0
200
0
4
6
Reference (red) and output
10
10
2
0
2
0
38/82
Example
The noisy closed-loop response shows that the effect of the noise onto
the to-be-tracked output is visibly reduced:
Control action
200
0
200
0
4
6
Reference (red) and output
10
10
2
0
2
0
39/82
Kalman Filtering
Consider again the full generalized plant
x = Ax + Bw w + Bu, z = Cz x + Dz u, y = Cx + Dw w.
If w = 0, an observer for this system is defined as
x = A
x + Bu + L(y y), z = Cz x + Dz u, y = C x
where L is taken with eig(A LC) C ; then the observer state
asymptotically reconstructs the systems state.
If w does not vanish and is a white-noise disturbance, the quantity
lim E (z(t) z(t))T (z(t) z(t))
(err)
t
can serve as a measure for how well z(t) approximates z(t) for t .
An observer for which the asymptotic variance (err) of z z is
minimized is called a Kalman Filter for the generalized plant.
40/82
min
L, ALC Hurwitz
Proof by Duality
Note that we have
kCz (sI A + LC)1 (Bw LDw )k2 =
= k(BwT DwT LT )(sI AT + C T LT )1 CzT k2 .
This observation reduces the problem to one of static state-feedback
control for the system
43/82
Remarks
Note that the optimal observer does not depend on Cz or Dz ! In
particular it is as well an optimal H2 -observer for the full state x
(with the choices Cz = I and Dz = 0) with optimal value tr(Q).
An optimal H2 -estimator is an unstructured LTI system with inputs
u and y, which generates an asymptotic state-estimate x such that
Cz x +Dz u is an optimal estimate of z in the H2 -sense. One can prove
that general estimators do not offer any benefit over observers!
In view of the stochastic interpretation of the H2 -norm, optimal H2 observers minimize the asymptotic variance of z z if w is white
noise. Then the optimal observer is the celebrated Kalman-Filter.
R.E. Kalman, A new approach to linear filtering and prediction problems, Journal of basic Engineering, 82 (1960) 35-45. (13560 citations
in Google Scholar as of February 7, 2013!)
44/82
Example
Consider again the two-compartment model ([AM] pp.85):
c =
y =
k0 k1 k1
k2
k2
1 0 c + 2 w 2
c+
0
1
w1 +
b0
0
u
Example
System and observer responses as well as errors for 1 = 2 = 0.01:
2
2
0
10
2
0
2
0
10
2
0
1
0
1
10
2
0
10
10
1
0
1
10
2
0
46/82
Example
System and observer responses as well as errors for 1 = 2 = 0.1:
4
2
0
2
0
10
4
0
10
10
10
1
0
1
2
1
2
0
3
0
10
4
0
2
5
10
3
0
47/82
Example
System and observer responses as well as errors for 1 = 2 = 1:
4
4
0
10
4
2
0
2
0
10
4
0
10
4
0
4
0
10
4
0
10
10
48/82
z
P
y
u
K
A
BCK
Bw
A B
= BK C AK BK Dw .
C D
Cz Dz CK
0
Problem: Minimize the H2 -norm kC(sI A)1 Bk2 of the controlled
closed-loop system over all controllers which render A Hurwitz.
49/82
Hypotheses
We derive a solution to this optimal synthesis problem in terms of AREs
under the following assumptions:
1. (A, B) is stabilizable and (A, C) is detectable.
Are required for the existence of a stabilizing controller.
Bw
0
T
T
Dw =
.
2. Dz Cz Dz = 0 I and
Dw
I
Its essential that Dz and Dw have full column and row rank. The
other properties are introduced to simplify the formulas.
3. (A, Cz ) has no unobservable and (A, Bw ) no uncontrollable modes
on the imaginary axis.
Are required for the existence of stabilizing solutions of AREs.
A fully general solution without hypotheses can be obtained with LMIs.
50/82
The controller can as well be seen as to provide an H2 -optimal estimate F xK of the unavailable but to-be-implemented signal F x.
The extra cost for this estimation is tr(F QF T ) if compared to the
cost for optimal state-feedback controller.
52/82
z
y
53/82
A iI B
Dz
Dz have full column rank R.
, Cz
Dze
Cze
Cze
If condition 3 fails choose Bwe , Dwe such that
A iI Bw Bwe
Dw Dwe ,
have full row column R.
C
Dw Dwe
These properties can be always achieved, as the following simple but
rough choices demonstrate:
Bwe
I 0
I 0
Cze Dze =
and
=
.
0 I
Dwe
0 I
54/82
=
=
=
=
Ax + Bw w + Bwe we + Bu
Cz x + Dzw w + Dz u
Cze x + Dze u
Cx + Dw w + Dwe we + Du
z
ze
y
P
w
we
u
The perturbed system satisfies the required hypothesis and one can find
an optimal H2 -controller K with optimal value . Now interconnect
K with the the original system P with closed-loop transfer matrix T .
One can show that K also stabilizes P and achieves an H2 -norm
level of at least :
kT k2 .
Moreover, converges monotonically to the optimal achievable H2 norm level for P (although an optimal controller might not exist).
55/82
Example
In the lectures we provide a demo on designing an output-feedback
tracking controller for the motor on slide 36. We obtain the following
responses for a pole-placement and LQG-synthesis with low noise-levels:
Control action
20
0
20
0
4
6
Reference (red) and output
10
10
10
10
2
0
2
0
Control action
20
0
20
0
4
6
Reference (red) and output
2
0
2
0
56/82
Example
For increased noise-levels we see the benefit of LQG-control:
Control action
20
0
20
0
4
6
Reference (red) and output
10
10
10
10
2
0
2
0
Control action
20
0
20
0
4
6
Reference (red) and output
2
0
2
0
57/82
Example
However, comparisons of this sort can be misleading:
Obviously the response of the pole-placement observer to non-zero initial conditions is faster than that of the LQG controller. This explains
its higher sensitivity to noise. Slowing down the observer poles does
not alter the tracking behavior (a lot), but it reduces the sensitivity
to noise.
For this simple example, one can obtain similar designs by poleplacement and LQG-synthesis after tuning. In practice, modern synthesis tools rather serve to reduce the time required for tuning a
controller, while the optimality properties are not that crucial.
LQG output-feedback controllers suffer from an essential deficiency:
There are no guarantees for robustness! This lead to the development
of dedicated tools for robust controller synthesis. You are now wellprepared to enter this exciting field of control.
58/82
Appendix
59/82
A1 0 B1
0 A2 B2 , eig(A1 ) C , eig(A2 ) C+ .
C1 C2 0
Therefore
G(s) = C1 (sI A1 )1 B1 + C2 (sI A2 )1 B2 .
Observe that
G(s)T = B2T (sI AT2 )1 C2T + B1T (sI AT1 )1 C1T .
Since G(i) = G(i)T , we have G(s) = G(s)T for s C0 and
hence for all s C (different from poles of G(s) and G(s)T ). Thus
the stable and anti-stable parts in the additive decomposition of G(s)
and G(s)T coincide.
59/82
A 0
B
0 AT C T with eig(A) C .
C BT
0
Observe that minimality of the realization is equivalent to (A, B) being
controllable and (A, C) being observable.
60/82
()
( )
T
A iI
0
CT
0
A iI B
( )
1
T
I
B
C
1s
Consider G(s) = s4 13s
2 +36 . Determine a minimal realization of G and
block-diagonalize the state-matrix. One then gets
0.06
2 1.73
.
0.13
0 3
J =
0.87 0.5
0
(ARE)
(ARI)
Proof: Step 1
The kernel N (P ) of P is A-invariant and contained in N (C).
Indeed, with x satisfying P x = 0, we infer from xT (ARE)x = 0 that
kCxk2 = 0 and thus Cx = 0. Then (ARE)x = 0 implies P Ax = 0.
This proves the claim.
Let the columns of T2 form a basis of N (P ) and expand to a non-singular
matrix T = (T1 T2 ). Then
P1 0
T
T PT =
and P1 0.
0 0
Due to AN (P ) N (P ) and N (P ) N (C) we also infer
A1 0
B1
1
1
T AT =
, T B=
, CT = C1 0 .
A21 A2
B2
W.l.o.g. (coordinate-change - check!) we can assume that the matrices
are already given in this form while X has no particular structure.
69/82
Proof: Step 2
By inspection, the ARE reads as AT1 P1 +P1 A1 P1 B1 B1T P1 +C1T C1 = 0.
This implies for Q1 = P11 (recall invertibility!) that
(A1 + Q1 C1T C1 )T = P1 (A1 B1 B1T P1 )P11 .
Therefore A + Q1 C1T C1 is anti-stable. (All eigenvalues are in the open
right half-plane.) We also have
A1 Q1 + Q1 AT1 B1 B1T + Q1 C1T C1 Q1 = 0.
Similarly, Y = X 1 satisfies AY + Y AT BB T + Y C T CY 0. Just
by considering the left-upper block, we infer for
Y1 Y12
Y =
Y21 Y2
that Y1 satisfies the strict algebraic Riccati inequality
A1 Y1 + Y1 AT1 B1 B1T + Y1 C1T C1 Y1 0.
70/82
With X2 0 we infer
P1 0
X1 X12 X21 X21 0
.
0 0
0
X2
I X12 X21
A congruence transformation with
finally leads to
0
I
P1 0
X1 X12
0 0
X21 X2
which finishes the proof.
71/82
Proof of Theorem 14
If opt denotes the optimal value of the H2 -problem, the first step is
2
to show tr(BwT P Bw + B T P QP B) opt
. Choose > opt . We can
then find AK , BK , CK which render A Hurwitz and such that kC(sI
A)1 Bk22 < 2 . By simply expanding the controller realization with
stable uncontrollable or unobservable modes, we can assume that AK
has at least the same dimension as A. Moreover, by slide 6 there exists
some X 0 with AT X + X A + C T C 0 and tr(B T X B) < 2 . With
Z := B T X B + I and suitably small > 0 we then infer
X XB
T
T
A X + X A + C C 0,
0, tr(Z) < 2 . (1)
BT X Z
A key step of the proof is to block-factorize X in the partition of A.
More precisely, let (X U ) denote the upper block row of X . We can
assume that U has full row rank and that the right-lower block of X is
non-singular (after perturbation if necessary). This assures:
72/82
Proof of Theorem 14
There exist X, U and Y , V such that
Y 0
I VT
X =
and S has full row rank.
X U
I 0
| {z }
| {z }
S
(2)
Let us first prepare some relations for the proof. In view of (1) we need
SX S T SX B
T
T
T
T
T
SA X S + SX AS + SC CS ,
BT X S T Z
which are equal to
T
SA R + RAS + SC CS ,
RS T RB
B T RT Z
.
(3)
RS =
Y 0
X U
I I
V 0
=
Y Y
Y X
Proof of Theorem 14
Moreover, all other blocks in (3) can be explicitly written as
Y 0
Bw
Y Bw
RB =
=
,
X U
BK Dw
XBw + U BK Dw
I I
T
CS = Cz Dz CK
= Cz + Dz CK V Cz
V 0
and
RAS =
=
Y 0
A BCK
I I
=
X U
V 0
BK C AK
Y 0
A + BCK V
A
=
=
X U
BK C + AK V BK C
Y (A + BCK V )
YA
X(A + BCK V ) + U BK C + U AK V XA + U BK C
.
74/82
Proof of Theorem 14
Therefore the matrices in (3) read as
Y
Y
Y Bw
T
R1 R21
X XBw + U BK Dw
, Y
R21 R2
T
Bw Y ()T
Z
(4)
with blocks
R1 = (A + BCK V )T Y +Y (A + BCK V )+(Cz + Dz CK V )T (Cz + Dz CK V ),
R2 = AT X + XA + U BK C + (U BK C)T + CzT Cz ,
R21 = AT Y + X(A + BCK V ) + U BK C + U AK V + CzT Cz .
With the equation for P and with = X P let us note that R2 and
R21 can be written as
R2 = AT + A + U BK C + (U BK C)T + P BB T P,
R21 = AT (Y P ) + A + (P + )BCK V + U BK C + U AK V + P BB T P.
75/82
Proof of Theorem 14
After these preparations we continue the proof. Since S has full row
rank, (1) clearly implies that R1 0, R2 0 and that the second
matrix in (4) is positive definite. By slide 34, R1 0 implies P Y .
On the other hand, by taking the Schur-complement we infer
X
XBw + U BK Dw
I
Y I Bw 0.
T
T
(XBw + U BK Dw )
Z
Bw
Combining the last two inequalities implies (recalling X P = ) that
Bw + U BK Dw
0.
(Bw + U BK Dw )T
Z BwT P Bw
With L = 1 U BK we have Bw + U BK Dw = (Bw LDw ) and
therefore Z BwT P Bw (Bw LDw )T (Bw LDw ) 0. This shows
tr(Bw LDw )T (Bw LDw ) < 2 tr(BwT P Bw )
()
()
76/82
Proof of Theorem 14
By slide 6 the inequalities (), () show that A LC is Hurwitz and
A LC Bw LDw
2
< 2 tr(BwT P Bw ).
BT P
0
2
This allows to apply our result on H2 -estimation on slide 41. Due the
choice of Q we can hence conclude
tr(B T P QP B) + tr(BwT P Bw ) < 2 .
Since > opt was arbitrary we finally get
2
.
tr(B T P QP B) + tr(BwT P Bw ) opt
(5)
(6)
satisfies
is easily seen to be the unique solution of (2). This renders all equations
on slides 73-75 satisfied.
79/82
References
[KK] H.W. Knobloch, H. Kwakernaak, Lineare Kontrolltheorie, Springer-Verlag
Berlin 1985
[AM] K.J. Astr
om, R.M. Murray, Feedback Systems: An Introduction for Scientists
and Engineers, Princeton University Press, Princeton and Oxford, 2009 (available
online for free with Wiki, just google.)
J.P. Hespanha, Linear Systems Theory, Princeton University Press, 2009
[S] E.D. Sontag, Mathematical Control Theory, Springer, New York 1998
[K] T. Kailath, Linear Systems, Prentice Hall, Englewood Cliffs, 1980
[F] B. Friedland, Control System Design: An Introduction to State-space Methods.
Dover Publications, 2005
W.J. Rugh, Linear System Theory, Prentice-Hall, 2 1998
R. Brockett, Finite dimensional linear systems, Wiley, 1970
W.M. Wonham, Linear multivariable control, a geometric approach, SpringerVerlag, 3 1985
81/82
References
J. Zabczyk, Mathematical Control Theory: An Introduction, Birkhauser, 2007
B.A. Francis, A course in H -control theory, Springer, 1987
H.K. Khalil, Nonlinear Systems, Prentice Hall, 3 2002
L. Arnold, Stochastic Differential Equations: Theory and Applications, Wiley, 1974
A. Packard, State-space Youla Notes, Course in Multivariable Control Systems,
UC Berkeley, 2008
J.M. Coron, Control and nonlinearity, Mathematical Surveys and Monographs,
2007
Y. Yuan, G.-B. Stan, S. Warnick, and J.M. Goncalves, Minimal realization of the
dynamical structure function and its application to network reconstruction, IEEE
Transactions on Automatic Control, http://www.bg.ic.ac.uk/research/g.stan/,
2012
J.S. Freudenberg with C.V. Hollot and D.P. Looze, A first graduate course in
feedback control, 2003
82/82