Introduction To Control Theory Including
Introduction To Control Theory Including
5
________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________
∂F d ⎛ ∂F ⎞
− ⎜⎜ ⎟=0 ⇒ p& 2 = − p1
11.1 Introduction
∂x 2 dt ⎝ ∂x& 2 ⎟
⎠
(11.10)
∂F d ⎛ ∂F ⎞
This chapter deals with the control with restrictions: is
− ⎜ ⎟=0 ⇒ p 2 = µ ( β − α − 2u )
bounded and might well be possible to have discontinuities.
∂u dt ⎝ ∂u& ⎠
(11.11)
∂F d ⎛ ∂F ⎞
To illustrate some of the basic concepts involved when
− ⎜ ⎟=0 2vµ = 0
∂v dt ⎝ ∂v& ⎠
controls are bounded and allowed to have discontinuities we (11.12)
start with a simple physical problem: Derive a controller such
that a car move a distance a with minimum time.
(11.12) ⇒ v = 0 or µ = 0 . We will consider these two
The motion equation of the car
cases.
d 2x
=u (i) µ =0
⎧p = 0
(11.1)
⇒⎨ 1 ⇒ be impossible.
dt 2
⎩ p2 = 0
(ii) v = 0
where
u = u (t ), − α ≤ u ≤ β (11.2) (11.5): v 2 = (u + α )( β − u ) ⇒ u = −α or u = β
Hence
represents the applied acceleration or deceleration (braking)
⎧β 0 ≤ t ≤τ
x& 2 = ⎨
and x the distance traveled. The problem can be stated as
⎩− α τ <t ≤T
minimize
T = 1 dt
∫ the switch taking place at time τ . Integrating using
T
(11.3)
o
boundary conditions on x 2
⎧β t 0 ≤ t ≤τ
subject to (10.11) and (10.12) and boundary conditions
⎧1 2
with inequality constraints of the form (11.2). We can
⎪⎪ β t 0 ≤ t ≤τ
change this constraint into an equality constraint by
x1 = ⎨ 2
introducing another control variable, v, where
⎪− 1 α (t − T ) 2 + a τ <t ≤T
(11.14)
v = (u + α )( β − u )
2
(11.5) ⎪⎩ 2
∫
Eliminating T gives the switching time as
T* = {1 + p1 ( x 2 − x&1 ) + p 2 (u − x& 2 ) + µ [v 2
T
2aα
−(u + α )( β − u )]}dt τ=
0
β (α + β )
(11.8) (11.15)
αβ
and v are (11.16)
∂F d ⎛ ∂F ⎞
− ⎜ ⎟=0 ⇒ p& 1 = 0
∂x1 dt ⎜⎝ ∂x&1 ⎟⎠
(11.9) The problem now is completely solved and the optimal
control is specified by
⎧β 0 ≤ t ≤τ ∂H
u=⎨ p& i = − (i = 1,2, L , n)
⎩− α τ <t ≤T ∂xi
(11.17) (11.23)
this is illustrated in Fig. 11.1 The Euler equations for the control variables, u i , do not
follow as in section 10.4 as it is possible that there are
β
control u discontinuities in u i , and so we cannot assume that the partial
derivaties ∂H / ∂u i exist. On the other hand, we can apply the
∂F
= 0 to obtain
∂x& i
free end point condition (9.23):
τ
p k (T ) = 0 k = q + 1, L , n
0
T time t (11.24)
−α
that is, the adjoint variable is zero at every end point where
the corresponding state variable is not specified. As before,
we refer to (11.24) as tranversality conditions.
Fig. 11.1 Optimal Control
∫ ∑ p x& }dt
Problem: We are seeking extremum values of the functional
∫
δJ * = δ {H −
n
J =
T
T
i i
i =1
f 0 (x, u, t )dt (11.18) o
0
subject to state equations The small change operator, δ , obeys the same sort of
x& i = f i(x, u, t ) (i = 1,2, L , n) (11.19) properties as the differential operator d / dx .
initial conditions x = x 0 and final conditions on x1 , x 2 , L , x q Assuming we can interchange the small change operator, δ ,
(q ≤ n) and subject to u ∈ U , the admissible control region.
and integral sign, we obtain
δJ * =
∫ [δ ( H − ∑ p x& )]dt
For example, in the previous problem, the admissible control T n
region is defined by i i
i =1
U = {u : −α ≤ u ≤ β }
=
T n
i i
i =1
As in section 10.4, we form the augmented functional
∑ ∫ ∑ ∑ p δ x& ]dt
0
∫
⎡ ⎤
J* = ⎢ f0 + pi ( f i − x& i )⎥ dt = [δH − x& i δ pi −
n T n n
T
⎢ ⎥
(11.20)
⎣ ⎦
i i
i =1 i =1 i =1
0 0
and define the Hamintonian Using chain rule for partial differentiation
H = f0 + ∑ δH = ∑
∂H
δu j + ∑
∂H
δxi + ∑ ∂p δp
∂H
n m n n
∂u j ∂xi
pi f i (11.21)
i =1
i
j =1 i =1 i =1 i
∫ ∑ ∂u
function of the state vector x, control vector u, and adjoint
⎡
⎢ ∂H
δJ * = δu j
T m
⎢
We can express J * as
⎣
∑ p x& ⎟⎟ dt
j =1
∫
0
T⎛ ⎞
∑ ⎜⎜⎝ ∂p δ p
j
J* = ⎜H − ⎛ ∂H ⎞⎤
n
∂H
since p& i = −∂H / ∂xi . Also, from (11.21) = f i = x& i
and evaluating the Euler equations for xi , we obtain as in
section 10.4 the adjoint equations ∂pi
using (11.19): x& i = f i(x, u, t ), (i = 1,2, L , n) . Thus We first illustrate its use by examining a simple problem. We
required to minimize
0 ⎢
T
∂u j
⎣ j =1 i =1 ⎦
0
∫ ∑ ∑
⎡ ⎤ subject to x&1 = x 2 , x& 2 = u where −α ≤ u ≤ β and x1 (T ) = a ,
= ⎢ ∂H
δu j − ( piδ xi )⎥⎥ dt
T m n
0 ⎢ x2 (T ) = 0 , x1 (0) = x 2 (0) = 0 .
d
∂u j
⎣ j =1 i =1 ⎦
dt
H = 1 + p1 x 2 + p 2 u
∑ ( p δ x ) + ∫ ⎜⎜ ∑
T ⎛⎜ m ∂H
⎞
⎟
T
δJ * = − δu j ⎟ dt
n
∂u j ⎟
(11.25)
⎝ ⎠
i i
i =1 j =1
u ∈U = [−α , β ] , the admissible control region.. Since H is
0 We must minimize H with respect to u, and where
0
∫ ∑
H
T ⎛⎜ ∂H
⎞
⎟
δJ * = ⎜ δu j ⎟ dt
m
0 ⎜ ∂u j ⎟
⎝ j =1 ⎠
∫∑
We thus obtain
(i) the control u = −α or u = β , that is, bang-bang control;
δJ * = [ H (x; u1 ,L, u j + δu j ,L, u m ; p) − H (x, u, p)] dt
T m
(ii) there is one and only one switch in the control.
j =1
Again, it is clear from the basic problem that initially u = β ,
0
H (x; u1 ,L, u j + δu j ,L, u m ; p) ≥ H (x, u, p) (11.27) In the last section we met the idea of a switch in a control.
The time (and position) of switching from one extremum
for all admissible δu j and for j = 1,L, m . So we have value of the control to another does of course depend on the
initial starting point in the phase plane. Byconsidering a
established that on the optimal control H is minimized with specific example we shall show how these switching
respect to the control variables, u1 , u 2 ,L, u m . This is known positions define a switching curve.
as Pontryagin’s minimum principle.
x 2 = − log | x1 − 1 | +C
x1 = 0, x 2 = 0 in minimum time; that is, we wish to minimize
we wish to find the optimal control which takes the system to
(11.32)
∫
where C is constant. The curves for different values of C are
J = 1dt
T
(11.30) illustrated in Fig.11.4
u =1
0
−1 ≤ u ≤ 1 (11.31)
⎧+1 if p1 + p 2 < 0
u=⎨
A
⎩− 1 if p1 + p 2 > 0
∂H x2 = log | x1 − 1 | +C
p& 1 = − = p1
(11.33)
∂x1 p1 = Ae
⇒
t
∂H p2 = B
, A and B are constant
p& 2 = − =0
and the curves are illustrated in Fig. 11.5.
∂x 2
sign change.
x1 = Ae −t + k
x2 = k t + B
, A and B are constants
The basic problem is to reach the origin from an arbitrary Following the usual procedure, we form the Hamiltonian
H = 1 + p1 x2 + p 2 u
initial point. All the possible trajectories are illustrated in
Figs. 11.4 and 11.5, and we can see that these trajectories are (11.39)
only two possible paths which reach the origin, namely AO
in Fig. 11.4 and BO in Fig. 11.5. We minimize H with respect to u, where −1 ≤ u ≤ 1 , which
gives
x2 B
u = −1 ⎧+1 if p2 < 0
u=⎨
⎩− 1 if p2 > 0
⎧& ∂H
⎪ p1 = − ∂x = 0 ⎧p = A
⎪
x1
−1 ⎨ ⇒ ⎨ 1
0
∂ ⎩ p 2 = − At + B
1
⎪ p& = −
1
= − p1
⎪⎩ 2
H
∂x2
⎧ log(1 + x1 ) x1 > 0
x2 = ⎨
for x1 x1
⎩− log(1 + x1 ) x1 < 0
(11.34) 0 0
for
{
x1 (0) = a, x 2 (0) = b to any point on the x2 axis, that is,
which takes the system from an arbitrary initial point
∫
T−{
J = 1.dt
T
(11.37) ( a, b ) A
0
subject to (11.35), (11.36), the above boundary conditions Fig. 11.8 Initial point below OA
−1 ≤ u ≤ 1 (11.38) points above the curve OA, there is only one trajectory,
x 2 (T ) − x 2 (0) = ±T
that is
T = | x 2 (T ) − x 2 (0) | (11.42)
x2
u = +1 u = −1
0 x1