0 Tlemcen Mio Contrib
0 Tlemcen Mio Contrib
net/publication/265935804
CITATIONS READS
48 4,110
2 authors:
All content following this page was uploaded by Benedetto Piccoli on 08 November 2017.
The aim of these notes is to give an introduction to the Theory of Optimal Control
for finite dimensional systems and in particular to the use of the Pontryagin Maximum
Principle towards the construction of an Optimal Synthesis. In Section 1, we introduce
the definition of Optimal Control problem and give a simple example. In Section 2 we
recall some basics of geometric control theory as vector fields, Lie bracket and con-
trollability. In Section 3, that is the core of these notes, we introduce Optimal Control
as a generalization of Calculus of Variations and we discuss why, if we try to write
the problem in Hamiltonian form, the dynamics makes the Legendre transformation
not well defined in general. Then we briefly introduce the problem of existence of
minimizers and state the Pontryagin Maximum Principle. As an application we con-
sider a classical problem of Calculus of Variations and show how to derive the Euler
Lagrange equations and the Weierstraß condition. Then we discuss the difficulties in
finding a complete solution to an Optimal Control Problem and how to attack it with
geometric methods. In Section 4 we give a brief introduction to the theory of Time
Optimal Synthesis on two dimensional manifolds developed in [14]. We end with a
bibliographical note and some exercises.
1 Introduction
Control Theory deals with systems that can be controlled, i.e. whose evolution can
be influenced by some external agent. Here we consider control systems that can be
defined as a system of differential equations depending on some parameters u ∈ U ⊆
Rm :
19
20 U. B OSCAIN AND B. P ICCOLI
The first problem one faces is the study of the set of points that can be reached,
from x0 , using open loop controls. This is also known as the controllability problem.
If controllability to a final point x1 is granted, one can try to reach x1 minimizing
some cost, thus defining an Optimal Control Problem:
Z T
min L(x(t), u(t)) dt, x(0) = x0 , x(T ) = x1 , (2)
0
of the Optimal Control Problem one should specify further: the time T fixed or free,
the set of admissible controls and admissible trajectories, etc. Moreover one can fix
an initial (and/or a final) set, instead than the point x0 (and x1 ).
Fixing the initial point x0 and letting the final condition x1 vary in some domain of
Rn , we get a family of Optimal Control Problems. Similarly we can fix x1 and let x0
vary. One main issue is to introduce a concept of solution for this family of problems
and we choose that of Optimal Synthesis. Roughly speaking, an Optimal Synthesis is
a collection of optimal trajectories starting from x0 , one for each final condition x1 .
As explained later, building an Optimal Synthesis is in general extremely difficult, but
geometric techniques provide a systematic method to attack the problem.
In Section 3.1 Optimal Control is presented as a generalization of Calculus of
Variations subjects to nonholonomic constraints.
Example Assume to have a point of unitary mass moving on a one dimensional line
and to control an external bounded force. We get the control system:
ẍ = u, x ∈ R, |u| ≤ C,
where x is the position of the point, u is the control and C is a given positive constant.
Setting x1 = x, x2 = ẋ and, for simplicity, C = 1, in the phase space the system is
written as:
ẋ1 = x2
ẋ2 = u.
One simple problem is to drive the point to the origin with zero velocity in minimum
time. From an initial position (x̄1 , x̄2 ) it is quite easy to see that the optimal strategy is
to accelerate towards the origin with maximum force on some interval [0, t] and then to
decelerate with maximum force to reach the origin at velocity zero. The set of optimal
trajectories is depicted in Figure 1.A: this is the simplest example of Optimal Synthesis
for two dimensional systems. Notice that this set of trajectories can be obtained using
the following feedback, see Figure 1.B. Define the curves ζ ± = {(x1 , x2 ) : ∓x2 >
0, x1 = ± x22 } and let ζ be defined as the union ζ ± ∪ {0}. We define A+ to be the
region below ζ and A− the one above. Then the feedback is given by:
+1 if (x1 , x2 ) ∈ A+ ∪ ζ +
u(x) = −1 if (x1 , x2 ) ∈ A− ∪ ζ −
0 if (x1 , x2 ) = (0, 0).
A N I NTRODUCTION TO O PTIMAL C ONTROL 21
x2
x2
−(
ζ u=−1
)
u(x)=−1
x1 x1
u(x)=+1
(u=+
1)
+
A B ζ
vector field can be seen as an operator from the set of smooth functions on M to R. If
x = (x1 , ..., xn ) is a local system of coordinates, we have:
n
X ∂
F (x) = Fi .
i=1
∂xi
The first definition we need is of the concept of control and of trajectory of a control
system.
Definition 1 A control is a bounded measurable function u(·) : [a, b] → U . A trajec-
tory of (1) corresponding to u(·) is a map γ(·) : [a, b] → M , Lipschitz continuous on
every chart, such that (1) is satisfied for almost every t ∈ [a, b]. We write Dom(γ),
Supp(γ) to indicate respectively the domain and the support of γ(·). The initial point
of γ is denoted by In(γ) = γ(a), while its terminal point T erm(γ) = γ(b)
Then we need the notion of reachable set from a point x0 ∈ M .
Definition 2 We call reachable set within time T > 0 the following set:
Rx0 (T ) := {x ∈ M : there exists t ∈ [0, T ] and a trajectory
γ : [0, t] → M of (1) such that γ(0) = x0 , γ(t) = x}. (3)
Computing the reachable set of a control system of the type (1) is one of the main
issues of control theory. In particular the problem of proving that R x0 (∞) coincide
with the whole space is the so called controllability problem. The corresponding local
property is formulated as:
Definition 3 (Local Controllability) A control system is said to be locally control-
lable at x0 if for every T > 0 the set Rx0 (T ) is a neighborhood of x0 .
Various results were proved about controllability and local controllability. We only
recall some definitions and theorems used in the sequel.
Most of the information about controllability is contained in the structure of the
Lie algebra generated by the family of vector fields. We start giving the definition of
Lie bracket of two vector fields.
Definition 4 (Lie Bracket) Given two smooth vector fields X, Y on a smooth mani-
fold M , the Lie bracket is the vector field given by:
[X, Y ](f ) := X(Y (f )) − Y (X(f )).
In local coordinates:
X ∂Yj ∂Xj
[X, Y ]j = Xi − Yi .
i
∂xi ∂xi
In matrix notation, defining ∇Y := ∂Yj /∂xi (j row, i column) and thinking to
(j,i)
a vector field as a column vector we have [X, Y ] = ∇Y · X − ∇X · Y .
A N I NTRODUCTION TO O PTIMAL C ONTROL 23
A very important class of families of vector fields are the so called Lie bracket
generating (or completely nonholonomic) systems for which:
LiexF = Tx M, ∀ x ∈ M. (5)
For instance analytic systems (i.e. with M and F analytic) are always Lie bracket
generating on a suitable immersed analytic submanifold of M (the so called orbit of
F). This is the well know Hermann-Nagano Theorem (see for instance [29], pp. 48).
If the system is symmetric, that is F = −F (i.e. f ∈ F ⇒ −f ∈ F), then the
controllability problem is more simple. For instance condition (5) with M connected
implies complete controllability i.e. for each x0 ∈ M , Rx0 (∞) = M (this is a
corollary of the well know Chow Theorem, see for instance [3]).
On the other side, if the system is not symmetric (as for the problem treated in
Section 4), the controllability problem is more complicated and controllability is not
guaranteed in general (by (5) or other simple conditions), neither locally. Anyway,
important properties of the reachable set for Lie bracket generating systems are given
by the following theorem (see [37] and [3]):
Theorem 1 (Krener) Let F be a family of smooth vector fields on a smooth man-
ifold M . If F is Lie bracket generating, then, for every T ∈]0, +∞], Rx0 (T ) ⊆
Clos(Int(Rx0 (T )). Here Clos(·) and Int(·) are taken with respect to the topology
of M .
Krener Theorem implies that the reachable set for Lie bracket generating systems has
the following properties:
• It has nonempty interior: Int(Rx0 (T )) 6= ∅, ∀ T ∈]0, +∞].
• Typically it is a manifold with or without boundary of full dimension. The
boundary may be not smooth, e.g. have corners or cuspidal points.
24 U. B OSCAIN AND B. P ICCOLI
f (x0 , u0 ) = 0,
rank[B, AB, A2 B, ..., An−1 B] = n, (6)
where A = ∂f /∂x)(x0 , u0 ) and B = ∂f /∂u)(x0 , u0 ),
Remark 3 Condition (6) is the well know Kalman condition that is a necessary and
sufficient condition for (global) controllability of linear systems:
In the local controllable case we get this further property of reachable sets:
Assume f of class C 1 with respect to x and continuous with respect to u. If the control
system is locally controllable at x0 then for every T , ε > 0 one has:
Proof. Consider x ∈ Rx0 (T ) and let ux : [0, T ] → U be such that the corresponding
trajectory starting from x0 reaches x at time T . Moreover, let Φt be the flux associated
to the time varying vector field f (·, u(t)) and notice that Φt is a diffeomorphism. By
local controllability at x0 , Rx0 (ε) is a neighborhood of x0 . Thus ΦT (Rx0 (ε)) is a
neighborhood of x and, using ΦT (Rx0 (ε)) ⊂ Rx0 (T + ε), we conclude.
3 Optimal Control
In this section we give an introduction to the theory of Optimal Control. Optimal Con-
trol can be seen as a generalization of the Classical Calculus of Variations
(min 0 L(x(t), ẋ(t))) to systems with nonholonomic constrains of the kind ẋ =
RT
f (x, u), u ∈ U .
The most important and powerful tool to look for an explicit solution to an Optimal
Control Problem is the well known Pontryagin Maximum Principle (in the following
PMP, see for instance [3, 29, 49]) that give a first order necessary condition for opti-
mality. PMP is very powerful for the following reasons:
• it generalizes the Euler Lagrange equations and the Weierstraß condition of Cal-
culus of Variations to variational problems with nonholonomic constrains;
• it provides a pseudo-Hamiltonian formulation of the variational problem in the
case in which the standard Legendre transformation is not well defined (as in
the case of Optimal Control, see below).
3.1 Introduction
In this Section we first recall the Euler Lagrange equations, how to transform them in
an Hamiltonian form and discuss in which case this transformation is applicable (the
Legendre transformation must be invertible).
A N I NTRODUCTION TO O PTIMAL C ONTROL 27
In Section 3.1.2 it is shown that for nonholonomic constraints (i.e. for an Opti-
mal Control problem with nontrivial dynamics) the Euler Lagrange equation cannot
be written in standard form and the Legendre transformation is never well defined.
Finally we explicate some connection with the Lagrangian formulation of mechanical
systems.
Here, for simplicity, the state space is assumed to be Rn and not an arbitrary
manifold.
∂L(x, u)
(x, u) → (x, p), where pi = Φi (x, u) := .
∂ui
This change of coordinates is well defined if it realizes a C 1 -diffeomorphism of
R2n into R2n , i.e. we must have:
0
2
11
∂ L(x, u)
det ∂Φ(x,u) ∂Φ(x,u) = det 6= 0, (13)
∂x ∂u . ∂ui ∂uj
for every (x, u) ∈ R2n . If the condition (13) is satisfied, the Legendre trans-
formation is said invertible. In this case the inverse transformation is u =
Φ−1 (x, p) and the Lagrangian is called regular.
In fact using carefully the chain rule we have for the first:
∂
pΦ−1 (x, p) − L(x, Φ−1 (x, p))
ẋ =
∂p
∂Φ−1 (x, p) ∂L(x, u) ∂Φ−1 (x, p)
= Φ−1 (x, p) + p − ,
∂p ∂u (x,Φ−1 (x,p)) ∂p
∂
pΦ−1 (x, p) − L(x, Φ−1 (x, p))
ṗ = −
∂x !
∂Φ−1 (x, p) ∂L ∂L(x, u) ∂Φ−1 (x, p)
= − p − − .
∂x ∂x ∂u (x,Φ−1 (x,p)) ∂x
Usually one considers also more general costs and targets (see below), but this is out
of the purpose of this introduction. The consequences of these generalizations are the
following:
• In the case of nonholonomic constraints the first order necessary condition can
not be written in the form (10). One can still write a Lagrange formulation of
the problem using Lagrange multipliers, but this procedure works only under
suitable regularity conditions (see the recent book [7] and references therein).
• One could try to write the problem in Hamiltonian form, but the dynamics ẋ =
f (x, u), u ∈ U (unless it is equivalent to the trivial one ẋ = u, u ∈ Rn ) renders
the Legendre transformation not well defined. This is simply consequence of the
fact that the ẋi ’s may be not independent (or equivalently f (x, U ) is a proper
subset of Rn ) and so operating with the Legendre transformation we may get
not enough pi ’s. In other words the Legendre transformation would be:
∂L(x, u)
(x, u) → (x, p), where pi = Φi (x, v) := , (16)
∂ u̇i
∂ 2 L(x,u)
but ∂ui ∂uj is not a n × n matrix (if U ⊂ Rm it is m × m) and so the Legendre
transformation can not be inverted.
There is a way of getting a well defined Legendre transformation also in the case
of Optimal Control, starting from the Lagrange equations written with Lagrange mul-
tipliers. But again this procedure works only under suitable regularity conditions, see
[7].
The Pontryagin Maximum Principle (that will be stated precisely in the next sec-
tion) permits to write the minimization problem directly in a pseudo-Hamiltonian form
(pseudo because the Hamiltonian still depends on the controls that must be determined
with a suitable maximization condition, see Definition 8, p. 33).
2
Calulus of Variation In standard L(x,v) in standard form, systems
if =0 but possible only if with
form u u holonomic
geneneralization
the Legendre transf
is invertible constraints
Figure 3:
in finite time T that can be fixed or free, depending on the problem. Notice that this
Optimal Control Problem is autonomous, hence we can always assume trajectories to
be defined on some interval of the form [0, T ].
Of great importance for applications are the so called control affine systems:
m
(18)
X
ẋ = F0 + ui F i , ui ∈ R,
i=1
or with cost equal to 1 (that is the minimum time problem) and bounded controls
|ui | ≤ 1. The term F0 in (18) is called drift term and, for the so called distributional
systems (or driftless), is equal to zero. Subriemannian problems are distributional
systems with quadratic cost. See also Exercise 3, p. 60. Single–input systems are
control affine systems with only one control, i.e. m = 1.
Definition 6 We call Px1 , x1 ∈ M , the Optimal Control Problem given by the dy-
namics (1) and the minimization (17) with TT = x1 .
Beside (H0), (see p. 21) we make the following basic assumptions on (1) and (17).
(H2) ψ is a C 1 function.
We are interested in solving the family of control problems {Px1 }x1 ∈M and for us a
solution is given by an Optimal Synthesis that is
Remark 6 One can invert the role of x0 and x1 letting x1 be fixed and x0 vary. More-
over, we can consider a generalization fixing an initial manifold SS, called source. In
this case the initial condition reads x0 ∈ SS.
3.2.1 Existence
In order to prove existence for Px1 , we first give the next:
Theorem 3 Assume f bounded, U compact and let the space of admissible controls
be the set of all measurable maps u(·) : [a, b] → U . If the set of velocities V (x) =
{f (x, u)}u∈U is convex, then the reachable set R(T ) is compact.
Theorem 4 Assume f bounded, U compact, and let the space of admissible controls
be the set of all measurable maps u(·) : [a, b] → U . Moreover assume that the set
{f (x, u), L(x, u)}u∈U is convex. If L ≥ C > 0 and x1 ∈ Rx0 (∞), then the problem
Px1 has a solution.
Sketch of the Proof. There exists T > 0 such that every trajectory ending at x 1 after
T is not optimal. Consider the augmented system obtained adding the extra variable
y such that:
ẏ = L(x, u).
Since f is bounded, all trajectories defined on [0, T ] have bounded costs. Thus apply-
ing the previous Theorem to the augmented system, we get that R (x0 ,0) (T ), the reach-
able set for the augmented system of (x, y), is compact. Therefore there exists a min-
imum for the function (x, y) → y + ψ(x) on the set R(x0 ,0) (T ) ∩ {(x, y) : x = x1 }.
The trajectory, reaching such a minimum, is optimal.
i) the final time is fixed or free (for fixed final time see Exercise 4, p. 61),
iii) the cost contains only the running part, only the final part, or both,
Here we state a version in which: i) the final time is free, ii) the source is zero dimen-
sional and the target TT is a smooth submanifold of M of any dimension, iii) there are
both running and final cost, iv) the source and the target do not depend on time. Let
us introduce some notation.
For every (x, p, λ0 , u) ∈ T ∗ M × R × U we define:
and
H(x, p, λ0 ) = max{H(x, p, λ0 , u) : u ∈ U }.
(PMP2) for a.e. t ∈ [0, T ], we have H(γ(t), λ(t), λ0 ) = H(γ(t), λ(t), λ0 , u(t)) = 0;
(PMP3) for every v ∈ Tγ(T ) TT , we have hλ(T ), vi = λ0 h∇ψ(γ(T )), vi (transversal-
ity condition).
In this case we say that (γ, λ) is an extremal pair.
Remark 9 Notice that the dynamic ẋ = f (x, u) and equation (PMP1) can be written
in the pseudo-Hamiltonian form:
∂H
ẋ(t) = (x(t), p(t), λ0 , u(t)),
∂p
∂H
ṗ(t) = − (x(t), p(t), λ0 , u(t)).
∂x
Remark 10 Notice that the couple (λ, λ0 ) is defined up to a positive multiplicative fac-
tor, in the sense that if the triple (γ, λ, λ0 ) represents an extremal, than the same hap-
pens for the triple (γ, αλ, αλ0 ), α > 0. If λ0 6= 0, usually one normalizes (λ(·), λ0 )
by λ0 = −1/2 or λ0 = −1.
34 U. B OSCAIN AND B. P ICCOLI
Remark 12 In the case in which we have also a source: γ(0) ∈ SS, we have also a
transversality condition at the source:
where u(x, p) is one of the values of the control realizing the maximum. Now if the
set of controls U is an open subset of Rm and H is smooth with respect to u then
u(x, p) is one of the solutions of:
∂H(x, p, u)
= 0. (21)
∂u
A useful fact is the following. If moreover (21) permits to find in a unique way u
as a smooth function of (x, p), then we can write the Hamiltonian equations with H
instead of H getting a true Hamiltonian system with no dependence on the control. In
fact in this case:
∂H(x, p) ∂H(x, p, u(x, p)) ∂H(x, p, u)
= =
∂x ∂x ∂x (x,p,u(x,p))
∂H(x, p, u) ∂u(x, p) ∂H(x, p, u)
+ =
∂u (x,p,u(x,p)) ∂x ∂x (x,p,u(x,p))
∂H(x, p) ∂H(x, p, u(x, p)) ∂H(x, p, u)
= =
∂p ∂p ∂p (x,p,u(x,p))
∂H(x, p, u) ∂u(x, p) ∂H(x, p, u)
+ =
∂u (x,p,u(x,p)) ∂p ∂p (x,p,u(x,p))
where the last equality follows from (21). Then one gets the controls plugging the
solution of this Hamiltonian system in the expression u(x, p). This nice P situation
happens always for normal extremals for distributional systems (ẋ = i ui Fi (x))
with quadratic cost (min 0 i ui dt). See Section 3.4, p. 38 for an example.
RT P 2
∂ H̄
ẋ(t) = (x(t), p(t), u(t)),
∂p
∂ H̄
ṗ(t) = − (x(t), p(t), u(t)),
∂x
∂ H̄
(x(t), p(t), u(t)) = 0,
∂u
where H̄(x, p, u) := hp, f (x, u)i, p(t) 6= 0 and the constraint u ∈ U ∈ Rm is
changed in u ∈ Int(U ) 6= ∅. Endpoint singular trajectories that are also extremals
are called endpoint singular extremals.
Remark 14 Notice that, although endpoint singular trajectories do not depend on the
cost and on the constraints on the control set, endpoint singular extremals (that for our
minimization problem are the interesting ones) do depend.
The name endpoint singular trajectories comes from the fact that they are singu-
larities of the endpoint mapping that, fixed an initial point and a time T , associates to
a control function u(·), defined on [0, T ], the end point of the corresponding trajectory
γ:
E x0 ,T : u(·) 7→ γ(T ).
By singularity of the end point mapping, we mean a control at which the Fréchet
derivative of E x0 ,T is not surjective. For more details see [11]. Roughly speaking,
this means that the reachable set, locally, around the trajectory, does not contain a
neighborhood of the end point γ(T ).
In the case of a minimum time problem for a control affine system (18) with
|ui | ≤ 1, endpoint singular trajectories satisfy < p(t), Fi (γ(t)) >= 0. Endpoint sin-
gular extremals, are just endpoint singular trajectories for which there exists λ 0 ≤ 0
satisfying < p(t), F0 > +λ0 = 0 and corresponding to admissible controls.
In Section 4, we consider control systems of the form ẋ = F (x) + uG(x) where
|u| ≤ 1. In this case, under generic conditions, endpoint singular extremals are arcs
of extremal trajectories corresponding to controls not constantly equal to +1 or −1.
Remark 15 Abnormal extremals do not depend on the cost, but depend on the con-
straint u ∈ U .
In other problems (e.g. minimum time for control affine systems) the two defi-
nitions are different, but coincide for some very special class of trajectories. These
trajectories are usually called singular exceptional (see [13]).
Remark 17 (High Order Conditions) PMP is used in Synthesis Theory to attempt a
finite dimensional reduction of the minimization problem, as explained in Step 2 of
Section 3.5, p. 41. Of course high order conditions can be very useful for further
restriction of candidate optimal trajectories.
There are several possible higher order variations. For instance the high order
principle of Krener [38] and the envelope theory developed by Sussmann [52, 53].
Other high order conditions are based on different techniques: symplectic geometric
methods, conjugate points, Generalized Index Theory. For a list of references see [14].
High order conditions are out of the purpose of this notes. See for instance [14]
and references therein.
Here, together with (H0) and (H1), we assume also L(x, u) ∈ C 1 and we are looking
for Lipschitz minimizers. For this problem there are not abnormal extremals and end-
point singular trajectories. Moreover one can deduce from PMP the Euler Lagrange
equations and the Weierstraß condition.
Indeed from (22) the Hamiltonian reads:
H(x, p, λ0 , u) = hp, ui + λ0 L(x, u). (23)
In this case H̄(x, p, u) = hp, ui thus ∂ H̄/∂u implies p ≡ 0. Hence there are not
endpoint singular trajectories (cfr. Definition 9, p. 35). Moreover (PMP2) with λ 0 = 0
implies (cfr. Remark 10) p = 0 contradicting the fact that (p, λ0 ) 6= 0. Therefore
there are not abnormal extremals and it is convenient to normalize λ 0 = −1 (in the
following we drop the dependence from λ0 ):
H(x, p, u) = hp, ui − L(x, u). (24)
Since there are is no constraint on the control set (u ∈ Rn ) and L is differentiable with
respect to u, the maximization condition (PMP2) implies a.e.:
∂H ∂L
(x(t), p(t), u(t)) = 0 ⇒ p(t) = (x(t), u(t)), (25)
∂u ∂u
A N I NTRODUCTION TO O PTIMAL C ONTROL 37
ẋ(t) = ∂H
∂p (x(t), p(t), u(t)) = u(t), a.e., (26)
ṗ(t) = − ∂H ∂L
∂x (x(t), p(t), u(t)) = + ∂x (x(t), u(t)), a.e.
The first equation is the dynamics, while the second together with (25) give the Euler
Lagrange equations in the case of a Lipschitz minimizer. (We recall that in general
(∂L/∂u)(x(t), u(t)) is not absolutely continuous, but it coincides a.e. with an abso-
lutely continuous function, whose derivative is a.e. equal to (∂L/∂x)(x(t), u(t))). If
x(.) ∈ C 2 (that implies u(.) ∈ C 1 ) and ∂L/∂u is C 1 , then p(t) is C 1 and one gets the
Euler Lagrange equations in standard form:
d ∂L ∂L
(x(t), u(t)) = (x(t), u(t)),
dt ∂u ∂x
but it is well known that they hold everywhere also for x(.) ∈ C 1 and for a less regular
L (cfr. for instance [22, 26]).
Remark 18 Notice that, in the case of Calculus of Variations, the condition to get a
true Hamiltonian system in the sense of Remark 13 (i.e. ∂H(x, p, u)/∂u solvable with
respect to u) is the same condition of invertibility of the Legendre transformation:
2
∂ L(x, u)
det 6= 0.
∂ui ∂uj
From PMP one gets also the stronger condition known as Weierstrass condition. From
the maximization condition:
we have 0 L(x(t), ẋ(t)) dt ≤ 0 L(x̄(t), x̄˙ (t)) dt. A classical example of a weak
RT RT
local minimizer that is not a strong local minimizer is x(.) ≡ 0 for the minimization
problem: R1
minx(.)∈C 1 [0,1] 0 (ẋ2 − ẋ4 ) dt
x(0) = 0, x(1) = 0.
For more details on the Weierstraß condition see any book of Calculus of Variations,
for instance [22, 26].
a.e. in [a, b], for some c ∈ R. The cost (30) is “nicer” than the cost (29) (is smooth and
quadratic) and usually, in Riemannian geometry and in control theory, one deals with
(30). The only thing to keep in mind is that, in (30), the final time must be fixed other-
wise there is no minimum (cfr. Exercise 3). Thus if we want to find minimizers via the
PMP (in this case one can use also Euler-Lagrange equations), then the Hamiltonian
must be equal to a (strictly) positive constant (cfr. Exercises 4).
If the vector fields {Fj (x)} are not always linearly independent, then the corre-
sponding metric gx presents singularities at points x where span{Fj (x)} is a proper
subset of Tx M . In this case the problem can not be treated as a problem of calculus
of variations, since there are nonholonomic constraints at the singular points.
An example is given by Grusin’s metric on the plane (x1 , x2 ), that is the metric
corresponding to the frame F1 = (1, 0)T and F2 = (0, x1 )T . In the language of
control theory, this translates into the following problem:
ẋ = u1 F1 (x) + u2 F2 (x),
x1 1 0
x= ∈ R2 , F1 (x) = , F2 (x) = .
x2 0 x1
Z T
u1 (t)2 + u2 (t)2 dt.
min
u(.) 0
The vector field F2 is zero on the x2 axis. In other words, on the x2 axis, every
trajectory of the system has velocity parallel to the x1 axis. Therefore the Grusin’s
metric explodes when one is approaching the x2 axis (see Figure 4).
Let us compute the geodesics starting from the origin using PMP. In this case the
Hamiltonian is (cfr. equation 20, p. 33):
There are not non trivial abnormal extremals. In fact let x(.) be an extremal trajectory
and p(.) the corresponding covector. If λ0 = 0, then it is possible to maximize (32)
with respect to the control only if
(otherwise the supremum is infinity). Being λ0 = 0, p(t) can not vanish. It follows
that (33) can hold only in the region where the two vector fields are linearly dependent
(i.e. on the axis x1 = 0). But the only admissible trajectories, whose support belongs
to this set, are trivial (i.e. γ(t) = (0, const)). We then normalize λ0 = −1/2.
In this case, since the controls belong to R, the maximization condition:
possible velocities
x2 F1
F1
F2 F2
x1
F1 F1
possible velocities
u1 = p1 ,
u2 = p 2 x1 ,
∂H
ṗ1 = − = −p22 x1 , (37)
∂x1
∂H
ṗ2 = − = 0. (38)
∂x2
From (38), we have p2 = a ∈ R. Equations (35),(37) are the equation for an harmonic
oscillator whose solution is:
x1 = A sin(at + φ0 ),
p1 = aA cos(at + φ0 ).
Since we are looking for geodesics starting from the origin we have φ 0 = 0 and
integrating (36), with x2 (0) = 0, we have:
aA2 A2
x2 (t) = t− sin(2at). (39)
2 4
In the expressions of x1 (t), x2 (t), p2 (t) there are still the two real constants A and a.
Requiring that the geodesics are parameterized by arclength (u21 (t) + u22 (t) = 1), that
corresponds to fix the level 1/2 of the Hamiltonian, we have:
1 1 1
= H = (p21 (t) + p22 (t)x21 (t)) = a2 A2 .
2 2 2
This means that A = ±1/a. Finally the formulas for extremal trajectories and controls
are given by the following two families (parametrized by a ∈ R):
x1 (t) = ± a1 sin(at)
1
t − 4a12 sin(2at)
x2 (t) = 2a
(40)
u1 (t) = p1 (t) = ± cos(at)
u2 (t) = ax1 (t) = ± sin(at)
-0.3
a=−1.3 a=−1.3
A B
Figure 5: Geodesics and minimum time front (for t = 1) for the Grusin’s metric
Step 1. Use PMP and high order conditions to study the properties of optimal trajec-
tories.
Step 2. Use Step 1. to obtain a finite dimensional family of candidate optimal trajec-
tories.
Step 1.
We stated the PMP that in some cases gives many information on optimal trajec-
tories. Beside high order maximum principle and envelopes, there are various higher
order conditions that can be used to discard some extremal trajectories that are not lo-
cally optimal. These conditions come from: symplectic geometric methods, conjugate
points, degree theory, etc.
Step 3.
Once a finite dimensional reduction is obtained, one may construct a synthesis in
the following way. Assume that on the compact set R(τ ), there is a bound on the
number of arcs (corresponding to controls ui ) that may form, by concatenation, an
optimal trajectory. Then one can construct, by induction on n, trajectories that are
concatenations of n arcs and cut the not optimal ones. The latter operation produces
some special sets, usually called cut loci or overlaps, reached optimally by more than
one trajectory.
The above procedure is done on the base space M , however extremals admit lifts
to T ∗ M . Thus another possibility is to construct the set of extremals in the cotan-
gent bundle and project it on the base space. In this case, projection singularities are
responsible for singularities in the synthesis.
Step 4.
Even if a finite dimensional reduction is not obtained, one can still fix a finite
dimensional family of extremal trajectories and construct a synthesis on some part of
the state space. If the synthesis is regular enough then there are sufficiency theorems
ensuring the optimality of the synthesis.
These sufficiency theorems fit well also inside the framework of viscosity solution
to the corresponding Hamilton–Jacobi–Bellman equation.
The above approach is quite powerful, but in many cases it is not known how to reach
the end, namely to produce an Optimal Synthesis. For the problem treated in Section
4, we are able not only to construct an Optimal Synthesis under generic assumptions,
but also to give a topological classification of singularities of the syntheses and of the
syntheses themselves, see [14].
Remark 19 For general problems (not necessarily with bound on the control, e.g. with
quadratic cost), Steps 1 and 4 are still used, while Step 3 is not based on a finite
dimensional reduction.
exits. The most famous example is that given by Fuller in [25]. The author considered
the infinite horizon Optimal Control problem:
The function p4 (.) is the so called switching function (cfr. Section 4.1, p. 46). In fact
from (PMP2) it follows that:
• if p4 (t) > 0 (resp < 0) for every t ∈ [a, b], then u ≡ 1 (resp. u ≡ −1) on
[a, b]. In this case the corresponding trajectory x(.)|[a,b] is an arc of clothoid in
the (x, y) space.
• if p4 (t) ≡ 0 for every t ∈ [a, b], then u ≡ 0 in [a, b]. In this case the trajectory
x(.)|[a,b] is a straight line in the (x, y) space.
The main feature of this highly nongeneric problem is that an optimal trajectory can-
not contain points where the control jumps from ±1 to 0 or from 0 to ±1. In other
words a singular arc cannot be preceded or followed by a bang arc.
In [58] it is proved the presence of Fuller phenomenon. More precisely, there exist
extremal trajectories x(.) defined on some interval [a, b], which are singular on some
interval [c, b], where a < c < b, and such that p4 (.) does not vanish identically on any
subinterval of [a, c]. Moreover the set of zeros of p4 (.) consists of c together with an
increasing sequence {tj }∞ j=1 of points converging to c. At each tj , x(.) is switches the
control from +1 to −1 or viceversa.
An optimal path can thus have at most a finite number of switchings only if it is
a finite concatenation of clothoids, with no singular arc. Existence of optimal trajec-
tories (not only extremal) presenting chattering was proved by Kostov and Kostova in
[36]. More precisely if the distance between the initial and final point is big enough,
then the shortest path can not be a finite concatenation of clothoids.
and the problem of reaching every point of the plane in minimum time starting from
the origin.
To guarantee completeness of the vector fields and to be able to compute deriva-
tives of arbitrary order, we assume the following:
(H) F, G are C ∞ vector fields, uniformly bounded with all their partial derivatives on
R2 .
46 U. B OSCAIN AND B. P ICCOLI
The PMP takes a particularly simple form for (42) and one can easily see that (un-
der generic conditions) controls are always either bang-bang, that is corresponding to
constant control u = +1 or u = −1, or singular (i.e. corresponding to trajectories
whose switching functions vanishes, see Definition 12 below and Remark 22). The
latter can happen only on the set where the vector field G and the Lie bracket [F, G]
are parallel (this generically is a one dimensional submanifold of the plane).
Moreover one can also predict (see Section 4.2) which kind of “switchings” (See
Definition 12) can happen on each region of the plane. We first need to state the PMP
for the system (42) and introduce some notations.
Theorem 6 (PMP for the minimum time problem for planar control affine sys-
tems)
Define for every (x, p, u) ∈ R2 × (R2 )∗ × [−1, 1]:
and:
Remark 21 Notice that, since the Lagrangian cost is constantly equal to 1, the condi-
tion (λ(t), λ0 ) 6= (0, 0), for all t ∈ [0, T ], given in Definition 8, p. 33, now becomes
λ(t) 6= 0 for all t. In fact λ(t) ≡ 0 with the condition (PMP2) of Definition 8 implies
λ0 = 0. To see what happens for vanishing Lagrangians cfr. Exercise 8, p. 61).
From the PMP one immediately get that the switching function describes when the
control switches from +1 to -1 and viceversa. In fact we have:
Lemma 2 Let (γ, λ) : [0, τ ] → R2 × (R2 )∗ be an extremal pair and φ(·) the corre-
sponding switching function. If φ(t) 6= 0 for some t ∈]0, τ [, then there exists ε > 0
such that γ corresponds to a constant control u = sgn(φ) on ]t − ε, t + ε[.
Proof. There exists ε > 0 such that φ does not vanish on ]t − ε, t + ε[. Then from
condition iii) of PMP we get u = sgn(φ).
Since φ(·) is absolutely continuous and λ(t) · [F, G](γ(t)) is continuous, we deduce
that φ is C 1 .
Notice that if φ(·) has no zeros then u is a.e. constantly equal to +1 or −1. Next
we are interested in determining when the control may change sign or may assume
values in ] − 1, +1[. For this purpose we give first the following definition:
48 U. B OSCAIN AND B. P ICCOLI
Definition 12 Let u(.) : [a, b] → [−1, 1] be a control for the control system (42).
• u(.) is said to be a bang control if for almost every t ∈ [a, b], u(t) is constant
and belongs to a point of the set {+1, −1}.
• A switching time of u(.) is a time t ∈ [a, b] such that for every ε > 0, u(.) is not
bang on (t − ε, t + ε) ∩ [a, b].
Remark 22 A singular extremal in the sense above is also an endpoint singular ex-
tremal in the sense of Section 3.2.3, p. 34. In fact for these trajectories the Hamil-
tonian is independent from the control. In the following we use the term singular
trajectories with the same meaning of endpoint singular extremal.
Remark 23 On any interval where φ has no zeroes (respectively finitely many zeroes)
the corresponding control is bang (respectively bang-bang).
Singular trajectories are studied in details below. In Figure 6 we give an example
to illustrate the relationship between an extremal trajectory and the corresponding
switching function. The control ϕ, corresponding to the singular arc, is computed
below, see Lemma 6, p. 51.
φ( t)
of points where F and G are parallel and the set of points where G is parallel to
[F, G]. These loci are fundamental in the construction of the optimal synthesis. In
fact, assuming that they are smooth one dimensional submanifolds of R 2 , we have the
following:
• on each connected component of M \ (∆−1 A (0) ∪ ∆B (0)), every extremal
−1
∆B (x)
f (x) = − . (46)
∆A (x)
Proof. We have:
∆B (x) = Det G(x), [F, G](x) = Det(G(x), f (x)F (x) + g(x)G(x))
= f (x)Det G(x), F (x) = −f (x)∆A (x).
Now from PMP, we have H(γ(t̄), λ(t̄)) = λ(t̄) · F (γ(t̄)) ≥ 0. Hence φ̇ > 0, since
F (γ(t̄)), and G(γ(t̄)) are independent. This proves that φ has at most one zero with
positive derivative at the switching time and gives the desired conclusion.
We are now interested in understanding what happens at points that are not ordinary.
Definition 14 A point x ∈ R2 is called an non ordinary point if x ∈ ∆−1
A (0) ∪
∆−1
B (0).
In the following we study some properties of singular trajectories in relation to non
ordinary points on which ∆B = 0.
Definition 15 • An non ordinary arc is a C 2 one-dimensional connected embedded
submanifold S of R2 with the property that every x ∈ S is a non ordinary point.
• A non ordinary arc is said isolated if there exists a set Ω satisfying the following
conditions:
(C1) Ω is an open connected subset of R2 .
(C2) S is a relatively closed subset of Ω.
A N I NTRODUCTION TO O PTIMAL C ONTROL 51
The following Lemmas describes the relation between turnpikes, anti-turnpikes and
singular trajectories. In Lemma 6 we compute the control corresponding to a trajec-
tory whose support is a turnpike or an anti-turnpike. In Lemma 7 we prove that if this
control is admissible (that is the turnpike or the anti-turnpike is regular, see Defini-
tion 16) then the corresponding trajectory is extremal and singular. In Lemma 8 we
show that anti-turnpike are locally not optimal.
Lemma 6 Let S be a turnpike or an anti-turnpike and γ : [c, d] → R2 a trajectory of
(42) such that γ(c) = x0 ∈ S. Then γ(t) ∈ S for every t ∈ [c, d] iff γ corresponds to
the feedback control (called singular):
∇∆B (x) · F (x)
ϕ(x) = − , (47)
∇∆B (x) · G(x)
Proof. Assume that γ([c, d])) ⊂ S and let u be the corresponding control, that is
γ̇(t) = F (γ(t)) + u(t)G(γ(t)), for almost every t. From ∆B (γ(t)) = 0, for a.e. t we
have:
d
0 = ∆B (γ(t)) = ∇∆B · (F (γ(t)) + u(t)G(γ(t))).
dt
This means that at the point x = γ(t) we have to use control ϕ(x) given by (47).
Definition 16 (regular turnpike or anti-turnpike) We say that a turnpike or anti-
turnpike S is regular if |ϕ(x)| < 1 for every x ∈ S.
Lemma 7 Let (γ, λ) : [0, t̄] → R2 be an extremal pair that verifies γ(t̄) = x, x ∈ S
where S is a turnpike or an anti-turnpike, and λ(t̄) · G(γ(t̄)) = 0. Moreover let
γ 0 : [0, t0 ] → R2 (t0 > t̄) be a trajectory such that:
52 U. B OSCAIN AND B. P ICCOLI
• γ 0 |[0,t̄] = γ,
• γ 0 ([t̄, t0 ]) ⊂ S.
Proof. Choose an open set Ω containing γ([c, d]) such that ∆A 6= 0 on Ω and define
the differential form ω on Ω by ω(F ) = 1, ω(G) = 0. Let γ1 : [c, d1 ] → Ω be any
trajectory such that γR1 (c) = γ(c), γ1R(d1 ) = γ(d), γ1 (t) ∈
/ SR for every t ∈]c, d1 [.
Notice that d − c = γ ω, d1 − c = γ1 ω. Hence d − d1 = γ −1 ∗γ ω where γ1−1
1
is γ1 run backward. Assume that γ1−1 ∗ γ is oriented counterclockwise, being similar
the opposite case. Then by Stokes’ Theorem, d − d1 = A dω where A is the region
R
−1
∆ Α(0)
+1
sing.
−1
−1 ∆Β(0)
relation with the sign of f , and an example of extremal trajectory containing a singular
arc.
Remark 24 The hypothesis F (x0 ) = 0, i.e. x0 is a stable point for F , is very natural.
In fact, under generic assumptions, it guarantees local controllability. Moreover if we
reverse the time, we obtain the problem of stabilizing in minimum time all the points
of M to x0 . For this time-reversed problem, the stability of x0 guarantees that once
reached the origin, it is possible to stay.
In [46, 47] (see also [14]) it was proved that this problem with F and G generic,
admits an optimal synthesis in finite time T . More precisely, if we define the reachable
set in time T :
an optimal synthesis in time T , for the control system (42) under the additional hy-
pothesis F (x0 ) = 0, is a family of time optimal trajectories Γ = {γx : [0, bx ] 7→ R2 ,
54 U. B OSCAIN AND B. P ICCOLI
Figure 8:
x ∈ R(T ) : γ(0) = x0 , γ(bx ) = x}. Moreover it was also proved that there exists a
stratification of R(T ) (roughly speaking a partition of R(T ) in manifolds of different
dimensions, see [14] for more details) such that the optimal synthesis can be obtained
from a feedback u(x) satisfying:
• on strata of dimension 2, u(x) = ±1,
• on strata of dimension 1, called frame curves (FC for short), u(x) = ±1 or
u(x) = ϕ(x), where ϕ(x) is defined by (47).
The strata of dimension 0 are called frame points (FP). Every FP is an intersection of
two FCs. A FP x, which is the intersection of two frame curves F1 and F2 is called an
(F1 , F2 ) Frame Point. In [47] (see also [14]), it is provided a complete classification
of all types of FPs and FCs, under generic conditions. All the possible FCs are:
• FCs of kind Y (resp. X), corresponding to subsets of the trajectories γ + (resp.
γ − ) defined as the trajectory exiting x0 with constant control +1 (resp. con-
stant control −1);
• FCs of kind C, called switching curves, i.e. curves made of switching points;
• FCs of kind S, i.e. singular trajectories;
• FCs of kind K, called overlaps and reached optimally by two trajectories com-
ing from different directions;
• FCs which are arcs of optimal trajectories starting at FPs. These trajectories
“transport” special information.
The FCs of kind Y, C, S, K are depicted in Fig. 8.
There are eighteen topological equivalence classes of FPs. They are showed in Figure
9. A detailed description can be found in [14, 47].
A N I NTRODUCTION TO O PTIMAL C ONTROL 55
Y Y
X
Y
Y Y
γ γ
0 or A
K K
(C,C) (C,S) 1 (C,S) 2 γ
2 0
γ S
C γ γ 0
0 or A
γ
C 0
C S C
(C,K) 1 (C,K) 2 (S,K) 1
K
K K S
γ
0
C γ C
k
(S,K) (S,K) 3 K
(K,K)
γ 2
K
k
γ
k
S K S K
Hence:
Since ∆−1B (0) = ∅ there are not singular trajectories (cfr. Theorem 8, p. 52). More-
over, in the region x2 > 0 (resp. x2 < 0), the only switchings admitted are from
control +1 to control −1 (resp. from control −1 to control +1, cfr. Theorem 7).
The trajectories corresponding to control +1 (resp. −1) and starting from the point
(x10 , x20 ) are the two parabolas:
x1 = x10 + x20 t ± 12 t2
(50)
x2 = x20 ± t.
To compute the switching times, we have to solve the equation for the covector ṗ =
−p(∇F + u∇G):
p1 = p10
(51)
p2 = p20 − p10 t
The switching function is then φ(t) = p(t)·G(x(t)) = p20 −p10 t. It is clear that every
time optimal trajectory has at most one switching because, for every initial conditions
p10 and p20 , the function φ has at most one zero. After normalizing p210 + p220 = 1, the
switchings are parametrized by an angle belonging to the interval [0, 2π]. The time
optimal synthesis is showed in Figure 10 A. In Figure 10 B it is presented the time
optimal synthesis for the same control system, but starting from a ball of radius 1/2.
Computing the explicit expression of the switching curves is an exercise for the reader
(see Exercise 9, p. 61).
A N I NTRODUCTION TO O PTIMAL C ONTROL 57
x2 x2
+1 +1
x1 x1
−1
−1
A B
Figure 10:
A more rich example presenting the FPs of kind (X, Y ), (Y, K)1 , (Y, S) is the
following:
ẋ1 = u
.
ẋ2 = x1 + 12 x21
Let us build the time optimal synthesis for a fixed time T > 2. The X– and Y –
trajectories can be described giving x2 as a function of x1 and are, respectively, cubic
polynomials of the following type:
x31 x2
x2 = − − 1 +α α∈R
6 2
x31 x2
x2 = + 1 +α α ∈ R. (52)
6 2
With a straightforward computation we obtain:
0
[F, G] =
−1 − x1
then the system is locally controllable and:
1 0
∆B (x) = det = −1 − x1 . (53)
0 −1 − x1
From equation (53) it follows that every turnpike is subset of {(x1 , x2 ) ∈ R2 : x1 =
−1}. Indeed, the synthesis contains the turnpike:
1
S = (x1 , x2 ) : x1 = −1, x2 ≤ − .
3
58 U. B OSCAIN AND B. P ICCOLI
Given b, consider the trajectories γ1 : [0, b] 7→ R2 for which there exists t0 ∈ [0, b]
such that γ1 |[0, t0 ] is a Y –trajectory and γ1 |[t0 , b] is an X–trajectory, and the trajecto-
ries γ2 : [0, b] 7→ R2 , b > 2, for which there exists t1 ∈ [2, b] such that γ2 |[0, t1 ] is an
X–trajectory and γ2 |[t1 , b] is a Y –trajectory. For every b > 2, these trajectories cross
each other in the region of the plane above the cubic (52) with α = 0 and determine
an overlap curve K that originates from the point (−2, − 32 ). We use the symbols
x+− (b, t0 ) and x−+ (b, t1 ) to indicate, respectively, the terminal points of γ1 and γ2
above. Explicitly we have:
(2t0 − b)3 (2t0 − b)2 t3
x+−
1 = 2t0 − b x+−
2 =− − + t20 + 0 (54)
6 2 3
(b − 2t1 )3 (b − 2t1 )2 t3
x−+
1 = b − 2t1 x−+
2 = + − t21 + 1 . (55)
6 2 3
Now the equation:
as b varies in [2, +∞[, describes the set K. From (54), (55) and (56) it follows:
t0 = b − t 1 t1 − 2t21 + (2 + 3b)t1 + (−b2 − 2b) = 0.
Bibliographical Note
For the problem of controllability, classical results are found in the papers by Krener
[37], and Lobry [41]. For the problem of controllability on Lie groups see the papers
by Jurdjevic-Kupka [30, 31], Jurdjevic-Sussmann [32], Gauthier-Bornard [27], and
Sachkov [60].
The issue of existence for Optimal Control as well as for Calculus of Variations is
a long standing research field. For a review of classical available results we refer to
[22, 23, 26].
A N I NTRODUCTION TO O PTIMAL C ONTROL 59
Y
K
X
A
B
S
A=(−1,−1/3)
B=(−2,−2/3)
The terms Abnormal Extremals and Singular Extremals (in our case called End-
point Singular Extremals) are used with several different meanings. For some results
see [1, 2, 11, 12, 13, 44, 45].
The first sufficiency theorem, for extremal synthesis, was given by Boltyanskii in
[10]. Then various generalization appeared, [20, 21, 48]. In particular the result of
[48], differently from the previous ones, can be applied to systems presenting Fuller
phenomena.
In Step 3 of Section 3.5, an alternative method for the construction of the Optimal
Synthesis is indicated. Namely, to construct all extremals in the cotangent bundle.
This second method is more involved, but induces a more clear understanding of the
fact that projection singularities are responsible for singularities in the Optimal Syn-
thesis. This approach was used in [15, 16, 33, 34].
For sufficiency theorems, in the framework of viscosity solutions to Hamilton–
Jacobi–Bellman equation (mentioned in Step 4 of Section 3.5), see [6].
Dubins’ car problem was originally introduced by Markov in [43] and studied by
Dubins in [24]. In particular Dubins proved that every minimum time trajectory is con-
catenation of at most three arcs, each of which is either an arc of circle or a straight
line. If we consider the possibility of non constant speed and admit also backward
motion, then we obtain the model proposed by Reed and Shepp [50]. A family of
time optimal trajectories, that are sufficiently reach to join optimally any two points,
was given in [59]. Now the situation is more complicated since there are 46 possible
60 U. B OSCAIN AND B. P ICCOLI
combination of straight lines and arcs of circles. Then a time optimal synthesis was
built by Soueres and Laumond in [51].
Time optimal trajectories for the system (41) were studied mainly by Boissonnat,
Cerezo, Kostov, Kostova, Leblond and Sussmann, see [8, 9, 35, 36, 58].
The concept of synthesis was introduced in the pioneering work of Boltianski [10].
Properties of extremal trajectories for single input affine control systems on the plane
were first studied by Baitman in [4, 5] and later by Sussmann in [55, 56].
The existence of an optimal synthesis for analytic systems and under generic con-
ditions for C ∞ systems was proved by Sussmann in [57] and by Piccoli in [46] respec-
tively.
Finally, a complete classification of generic singularities and generic optimal syn-
thesis is given by Bressan and Piccoli in [19, 47].
Exercises
Exercise 1 Consider the control system on the plane:
ẋ sin(x) G1 (x, y)
= +u , u ∈ R, (58)
ẏ 0 G2 (x, y)
where G1 , G2 ∈ C 1 (R2 , R) and G1 (x, y), G2 (x, y) > 0 for every (x, y) ∈ R2 . Find
the set of points where Theorem 2, p. 24 permits to conclude that there is local con-
trollability. Is it true that in all other points the system is not locally controllable?
Find where the system is locally controllable and prove that it is (globally) control-
lable.
[Hint: to prove global controllability, find a trajectory connecting each couple of
points.]
Exercise 3 Prove that, if the controls take values in the whole Rm , then there exists
no minimum for the minimum time problem of system (18). Moreover, show that, for
a distributional problem with quadratic cost (the so called subriemannian problem), if
we do not fix the final time, then there exists no minimum.
A N I NTRODUCTION TO O PTIMAL C ONTROL 61
Exercise 4 Prove that if we fix the final time in the PMP, then condition (PMP2)
becomes
(PMP2bis) For a.e. t ∈ [0, T ],
Exercise 9 Compute the time optimal synthesis for the control system on the plane
ẋ1 = x2 , ẋ2 = u, |u| ≤ 1 with source the unit closed ball.
References
[1] A.A. Agrachev, A. Sarychev, “Abnormal Sub-Riemannian Geodesics: Morse
Index and Rigidity,” Ann. Inst. H. Poincaré Anal. Non Linéaire Vol. 13, No. 6
(1996), pp. 635-690.
[2] A.A. Agrachev, A. Sarychev, “ On Abnormal Extremals for Lagrange Variational
Problems”, J. Math. Systems Estim. Control, Vol. 8, No. 1 (1998), pp. 87-118.
62 U. B OSCAIN AND B. P ICCOLI
[15] U. Boscain and B. Piccoli, “Extremal syntheses for generic planar systems”,
Journal of Dynamical and Control Systems Vol.7 No.2 (2001), pp. 209-258.
[16] U. Boscain and B. Piccoli, “On automaton recognizability of abnormal ex-
tremals”, SIAM J. Control and Opt., Vol. 40 No.5 pp. 1333-1357. (2002).
A N I NTRODUCTION TO O PTIMAL C ONTROL 63
[17] U. Boscain and B. Piccoli, “Morse Properties for the Minimum Time Function
on 2-D Manifolds”, Journal of Dynamical and Control Systems Vol. 7, No. 3,
pp. 385-423, (2001).
[18] A. Bressan and B. Piccoli, “Structural Stability for Time–Optimal Planar Syn-
theses”, Dynamics of Continuous, Discrete and Impulsive Systems, No.3 (1997),
pp. 335-371.
[19] A. Bressan and B. Piccoli, “A Generic Classification of Time Optimal Planar
Stabilizing Feedbacks”, SIAM J. Control and Optimization, Vol.36 No.1 (1998),
pp. 12-32.
[20] P. Brunovsky, “Existence of Regular Syntheses for General Problems,” J. Diff.
Equations 38 (1980), pp. 317-343.
[21] P. Brunovsky, “Every Normal Linear System Has a Regular Time–Optimal Syn-
thesis”, Math. Slovaca 28, 1978, pp. 81-100.
[22] G. Buttazzo, M. Giaquinta, S. Hildebrandt, “One-dimensional variational prob-
lems : an introduction” Oxford University Press, 1998.
[23] L. Cesari, “Optimization-theory and applications: problems with ordinary dif-
ferential equations”. New York : Springer-Verlag, c1983.
[24] Dubins L.E. “On curves of minimal length with a constraint on average curvature
and with prescribed initial and terminal position and tangents”, Amer. J. Math,
V.79 (1957), pp.497-516.
[25] A.T. Fuller, “Relay Control Systems Optimized for Various Performance Crite-
ria”, Automatic and remote control, Proc. first world congress, IFAC Moscow,
vol. 1, Butterworths, 1961, pp. 510-519.
[26] M. Giaquinta, S. Hildebrandt, “Calculus of Variations” Berlin, Springer-Verlag,
(1996).
[27] J.P. Gauthier and G. Bornard. “Controlabilite des sytemes bilineaires”, SIAM J.
Control and Optimization, 20:377-384, 1982.
[28] A. Isidori. “Nonlinear control systems: an introduction”, Springer-Verlag,
(1985).
[29] V. Jurdjevic, “Geometric Control Theory”, Cambridge University Press, (1997).
[30] V. Jurdjevic and I.K. Kupka, “Control Systems subordinated to a group action:
accessibility”, J. Diff. Eq. Vol.39, pp. 186–211.
[31] V. Jurdjevic and I.K. Kupka, “Control Systems on Semisimple Lie Groups and
Their Homogeneous Spaces”, Ann. Inst. Fourier, Vol.31, pp. 151–179.
64 U. B OSCAIN AND B. P ICCOLI
[37] A.J. Krener, “A generalization of Chow’s and the bang bang theorem to non-
linear control problems”. SIAM J. Control Optimization 12 (1974), pp. 43-51.
[38] A.J. Krener,“The high order maximal principle and its application to singular
extremals”, SIAM J. Control Optimization 15 (1977), no. 2, pp. 256–293.
[40] M.M. Lee, L. Markus, “Foundations of Optimal Control Theory”, Whiley, New
York, 1967.
[43] A.A. Markov, “Some examples of the solution of a special kind of problem in
greatest and least quantities,” (in Russian) Soobshch. Karkovsk. Mat. Obshch. 1,
1887, pp. 250–276.
[48] B. Piccoli and H.J. Sussmann, “Regular Synthesis and Sufficiency Conditions
for Optimality”, SIAM J. Control and Optimization, Vol. 39 No. 2 pp. 359-410,
(2000).
[50] J.A. Reeds and L.A. Shepp, “Optimal Path for a car that goes both forwards and
backwards”, Pacific J. Math, Vol. 145, pp. 367-393, (1990).
[51] P. Soueres and J.P. Laumond, Jean-Paul, “Shortest paths synthesis for a car-like
robot”, IEEE Trans. Automat. Control 41 (1996), no. 5, pp. 672–688.
[52] H.J. Sussmann, “Envelopes, conjugate points, and optimal bang-bang ex-
tremals”, in Algebraic and geometric methods in nonlinear control theory, 325–
346, Math. Appl., 29, Reidel, Dordrecht, (1986).
[53] H.J. Sussmann, “Envelopes, higher–order optimality conditions and Lie Brack-
ets”, in Proc. 1989 I.E.E.E, Conf. Decision and Control.
[54] S. Lojasiewicz, H.J. Sussmann, “Some examples of reachable sets and optimal
cost functions that fail to be subanalytic”. SIAM J. Control Optim. Vol. 23, no.
4, pp. 584–598 (1985).
[55] H.J. Sussmann, “The structure of time-optimal trajectories for single-input sys-
tems in the plane: the general real analytic case”, SIAM J. Control Optim. 25
No. 4 (1987), pp. 868–904
[57] H.J. Sussmann, “Regular synthesis for time optimal control of single–input real–
analytic systems in the plane,” SIAM J. Control and Opt., 25 (1987), pp.1145-
1162.
[58] H.J. Sussmann, “The Markov-Dubins Problem with Angular Acceleration Con-
trol”, Proceedings of the 36th IEEE, Conference on Decision and Control, San
Diego, CA, Dec. 1997. IEEE Publications, New York, 1997, pp. 2639-2643.
66 U. B OSCAIN AND B. P ICCOLI
[59] H.J. Sussmann, G. Q. Tang, “Shortest Paths for the Reeds-Shepp Car: A Worked
Out Example of the Use of Geometric Techniques in Nonlinear Optimal Con-
trol” Rutgers Center for Systems and Control Technical Report 91-10, Septem-
ber 1991, to appear in SIAM J. Control.
[60] Y. Sachkov, “Controllability of Invariant Systems on Lie Groups and Homoge-
neous Spaces,” J. Math. Sci., Vol.100, n.4, pp.2355-2427, (2000).
[61] M.I. Zelikin and V.F. Borisov, “Theory of chattering control. With applications to
astronautics, robotics, economics, and engineering”, Systems & Control: Foun-
dations & Applications, Birkhäuser, Boston, 1994.
U. B. B. P.
SISSA, IAC-CNR,
via Beirut 2-4 viale Del Policlinico 137,
34014 Trieste, Italy 00161 Roma, Italy
[email protected] [email protected]