0% found this document useful (0 votes)
20 views28 pages

Direct Methods SLP Report

This document summarizes a student project report on direct methods for solving optimal control problems numerically. It discusses direct shooting, multiple shooting, and direct collocation methods. Examples of applying these methods to problems like the Bryson-Denham problem and a double integrator bang-bang problem are also presented.

Uploaded by

Carolina Ribeiro
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
20 views28 pages

Direct Methods SLP Report

This document summarizes a student project report on direct methods for solving optimal control problems numerically. It discusses direct shooting, multiple shooting, and direct collocation methods. Examples of applying these methods to problems like the Bryson-Denham problem and a double integrator bang-bang problem are also presented.

Uploaded by

Carolina Ribeiro
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 28

Department of Aerospace Engineering

Indian Institute of Technology, Bombay

Direct Methods for Optimal Control


AE 419: Supervised Learning Project
by

Nakul Randad (180010047)

under the supervision of

Prof. Debasish Chatterjee

May 8, 2021
Certificate

I certify that this Supervised Learning Project Report titled “Direct Methods for

Optimal Control” by “Nakul Randad” is approved by me for submission. To the

best of my knowledge, the report represents the research and work carried out by

the candidate. I confirm that the investigations were conducted in accord with the

ethics policies and integrity standards of Indian Institute of Technology, Bombay

and that the research data are presented honestly and without prejudice

Date: May 8, 2021 Prof. Debasish Chatterjee


Table of Contents

Table of Contents

Abstract iii

Acknowledgement iv

Lists of Figures v

1 Introduction 1
1.1 Problem Formulation . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.2 Indirect Methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.3 Direct Methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3

2 Direct Shooting Methods 4


2.1 Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
2.2 Single Shooting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
2.3 Multiple Shooting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
2.4 Discussions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7

3 Direct Collocation Methods 8


3.1 Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
3.1.1 h-methods and p-methods . . . . . . . . . . . . . . . . . . . . . 8
3.2 Forms of Collocation Methods . . . . . . . . . . . . . . . . . . . . . . 9
3.2.1 Local Collocation . . . . . . . . . . . . . . . . . . . . . . . . . 9
3.2.2 Global Collocation . . . . . . . . . . . . . . . . . . . . . . . . . 10
3.3 Pseudospectral Method . . . . . . . . . . . . . . . . . . . . . . . . . . . 11
3.4 Problem Formulation . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12

4 Optimal Control Problems 15


4.1 Bryson-Denham Problem . . . . . . . . . . . . . . . . . . . . . . . . . 15
4.1.1 Description . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
4.1.2 Single Shooting . . . . . . . . . . . . . . . . . . . . . . . . . . 16
4.1.3 Multiple Shooting . . . . . . . . . . . . . . . . . . . . . . . . . 16
4.1.4 Direct Collocation . . . . . . . . . . . . . . . . . . . . . . . . . 17
4.2 Double Integrator Bang-Bang Problem . . . . . . . . . . . . . . . . . . 17
4.2.1 Description . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
4.2.2 Single Shooting . . . . . . . . . . . . . . . . . . . . . . . . . . 18
4.2.3 Multiple Shooting . . . . . . . . . . . . . . . . . . . . . . . . . 19
4.2.4 Direct Collocation . . . . . . . . . . . . . . . . . . . . . . . . . 19

i
Table of Contents

4.3 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20

References 21

ii
Abstract

Abstract

In this report, we study the numerical methods for optimal control. In particular,
the direct methods are discussed with an objective to showcase an approach for
general optimal control problems. These methods, which include single shooting,
multiple shooting and collocation methods, are relatively simple to follow and ef-
fectively solve a wide variety of optimization problems. We illustrate each of the
methods by working through example problems.

Keywords: Optimal Control, Trajectory Optimization, Direct Methods, Single Shoot-


ing, Multiple Shooting, Direct Collocation, Pseudospectral Method.

iii
Acknowledgements

Acknowledgements

I would like to express my deep gratitude to Prof Debasish Chatterjee, my re-


search supervisor, for his patient guidance, encouragement and useful critiques
throughout the duration of this project. I would also like to thank Siddhartha
Ganguly, Ph.D. candidate at IITB, for the assistance in gathering resources, for-
matting the report and the valuable discussions pertaining to the topic.

iv
List of Figures

List of Figures

1.1 Different types of Direct Methods . . . . . . . . . . . . . . . . . . . . 3

2.1 Basic algorithm for single shooting: ref. [1] . . . . . . . . . . . . . . 4


2.2 Schematic of direct multiple shooting: ref. [1] . . . . . . . . . . . . . 6
2.3 Single shooting vs multiple shooting: ref. [2] . . . . . . . . . . . . . . 7

3.1 Comparison between h-method and p-method: ref. [2] . . . . . . . . . 9


3.2 Differences between LGL, LGR, and LG collocation points: ref. [3] 11
3.3 Distribution of nodes and collocation points for both the global and
local approaches (N = 20): ref. [4] . . . . . . . . . . . . . . . . . . . . 12

4.1 Bryson-Denham problem by single shooting . . . . . . . . . . . . . . 16


4.2 Bryson-Denham problem by multiple shooting . . . . . . . . . . . . . 16
4.3 Bryson-Denham problem by direct collocation . . . . . . . . . . . . . 17
4.4 Double integrator bang-bang problem by single shooting . . . . . . . 18
4.5 Double integrator bang-bang problem by multiple shooting . . . . . . 19
4.6 Double integrator bang-bang problem by direct collocation . . . . . . 19

v
Introduction

Chapter 1

Introduction

Optimal control theory is a branch of mathematical optimization that deals with


finding a control for a dynamical system over a period of time such that an ob-
jective function is optimized. Because of the complexity of most applications, op-
timal control problems are most often solved numerically.

1.1 Problem Formulation

In general, an objective function can include two terms: a boundary objective Φ(·)
and a path integral along the entire trajectory, with the integrand L (·). A problem
with both terms is said to be in Bolza form. A problem with only the integral
term is said to be in Lagrange form, and a problem with only a boundary term is
said to be in Mayer form [5].

Z tf
L x(t), u(t),t
 
J := Φ x(t0 ),t0 , x(t f ),t f + dt (1.1)
t0

Here we determine the state (equivalently, the trajectory), x(t) ∈ Rn , the control
u(t) ∈ Rm , the initial time, t0 ∈ R, and the terminal time, t f ∈ R (where t ∈ [t0 ,t f ]
is the independent variable) that optimizes (generally minimizes) the objective func-
tion J in (1.1). The optimization is subject to a variety of limits and constraints,
detailed in (1.2) – (1.4).

1
Introduction

The first, and perhaps most important, of these constraints is the system dynamics,
which are typically nonlinear and describe how the system changes in time:

ẋ(t) = f (t, x(t), u(t)) (1.2)

Where f : R × Rn × Rm −→ Rn is a continuously differentiable map in x, u and t


which constitutes the dynamics corresponding to (1.2). Next is the path constraint,
which enforces restrictions along the trajectory.

h(t, x(t), u(t)) ≤ 0 (1.3)

Another important type of constraint is a nonlinear boundary constraint, which


puts restrictions on the initial and final states of the system.

g(t0 ,t f , x(t0 ), x(t f )) ≤ 0 (1.4)

Numerical methods for solving optimal control problems are divided into two ma-
jor classes: indirect methods (optimize, then discretize) and direct methods (dis-
cretize, then optimize) [1].

1.2 Indirect Methods

Indirect Methods use the necessary conditions of optimality of the infinite problem
to derive a boundary value problem (BVP) in ordinary differential equations (ODE).
The BVP is numerically solved to determine candidate optimal trajectories called
extremals. Each of the computed extremals is then examined to see if it is a local
minimum, maximum, or a saddle point. Of the locally optimizing solutions, the
particular extremal with the lowest cost is chosen.

Here we seek a solution to the (closed system of) conditions of optimality which
can be derived using the well-known calculus of variations and the Euler-Lagrange

2
Introduction

differential equations, and Pontryagin’s maximum principle (a necessary condition),


or by solving the Hamilton–Jacobi–Bellman equation (a sufficient condition) [6].

1.3 Direct Methods

Direct methods to continuous optimal control finitely parameterize the infinite di-
mensional decision variables, notably the controls u(t) and sometimes both state
x(t) and controls u(t), such that the original problem is approximated by a finite
dimensional nonlinear program (NLP). This NLP can then be addressed by struc-
turally exploiting numerical NLP solution methods. For this reason, the approach
is often characterized as “First discretize, then optimize.” We will discuss direct
single shooting, direct multiple shooting, and direct collocation in this report.

Figure 1.1: Different types of Direct Methods

One of the most important advantages of direct compared to indirect methods


is that they can easily treat inequality constraints, like the inequality path con-
straints in the formulation above. This is because structural changes in the active
constraints during the optimization procedure are treated by well developed NLP
methods that can deal with inequality constraints and active set changes. All direct
methods are based on a finite dimensional parameterization of the control trajec-
tory, but differ in the way the state trajectory is handled.

3
Direct Shooting Methods

Chapter 2

Direct Shooting Methods

2.1 Overview

All shooting methods use an embedded ordinary differential equation (ODE) or


differential-algebraic equation (DAE) solver in order to eliminate the continuous
time dynamic system. They do so by first parameterizing the control function
u(t), e.g. by polynomials, by piecewise constant functions, or, more generally, by
piecewise polynomials. [1]

2.2 Single Shooting

Figure 2.1: Basic algorithm for single shooting: ref. [1]

4
Direct Shooting Methods

We denote the finite control parameters by the vector q [7], and the resulting
control function by u(t; q). The most widespread parameterization are piecewise
constant controls, for which we choose a fixed grid 0 = t0 < t1 < . . . < tN = t f , and
N parameters qi ∈ Rnu , i = 0, . . . , N − 1, and then we set
u(t; q) = qi if t ∈ [ti ,ti + 1].
Thus, the dimension of the vector q = (q0 , . . . , qN−1 ) is Nnu . In single shooting,
which is a sequential approach, we then regard the states x(t) on [t0 ,t f ], keeping
t0 fixed at zero, as dependent variables that are obtained by a forward integration
of the dynamic system, starting at x0 and using the controls u(t; q). This can be
performed using methods like Euler forward, Euler backward, Crank-Nicolson, Runge-
Kutta, Hermite-Simpson and many more depending upon the required accuracy. We
denote the resulting trajectory as x(t; q). In order to discretize inequality path con-
straints, we choose a grid, typically the same as for the control discretization, at
which we check the inequalities. Thus, in single shooting, we transcribe the OCP
(1.1) into the following NLP, that is visualized as
Z tf
L x(t; q), u(t; q),t dt
 
min Φ x(t0 ; q),t0 , x(t f ; q),t f +
q∈RNnu t0

s.t. h(ti , x(ti ; q), u(ti ; q)) ≤ 0 i = 0, . . . , N − 1 (2.1)

g(t0 ,tN , x(t0 ; q), x(tN ; q)) ≤ 0

2.3 Multiple Shooting

Multiple shooting works by breaking up a trajectory into some number of seg-


ments, and using single shooting to solve for each segment. As the segments get
shorter, the relationship between the decision variables and the objective function
and constraints becomes more linear. In multiple shooting, the end of one segment
will not necessarily match up with the start of the next. This difference is known
as a defect, and it is added to the constraint vector. Adding all of the segments
will increase the number of decision variables (the start of each segment) and the
number of constraints (defects).

5
Direct Shooting Methods

Figure 2.2: Schematic of direct multiple shooting: ref. [1]

The direct multiple shooting method performs first a piecewise control discretiza-
tion on a grid, exactly as we did in single shooting, i.e. we set
u(t) = qi for t ∈ [ti ,ti+1 ].
But then, it solves the ODE separately on each interval [ti ,ti+1 ], starting with arti-
ficial initial values si :
ẋi (t; si , qi ) = f (xi (t; si , qi ), qi ) for t ∈ [ti ,ti+1 ]

xi (ti ; si , qi ) = si .
Thus, we obtain trajectory pieces xi (t; si , qi ). Likewise, we numerically compute
the integrals
Z ti+1
L xi (t; si , qi ), u(t; qi ),t

li (si , qi ) := dt
ti

Finally, we choose a grid at which we check the inquality path constraints; here
we choose the same as for the controls and states, but note that a much finer
sampling would be possible as well, which, however, requires continuous output
from the integrator. Thus, the NLP that is solved in multiple shooting is given

6
Direct Shooting Methods

by:
N−1

min Φ s0 ,t0 , sN ,t f + Σi=0 li (si , qi )
s,q

s.t. h(ti , si , qi ) ≤ 0 i = 0, . . . , N
(2.2)
g(t0 ,tN , s0 , sN ) ≤ 0

xi (ti+1 ; si , qi ) − si+1 = 0 i = 0, . . . , N − 1

2.4 Discussions

Figure 2.3: Single shooting vs multiple shooting: ref. [2]

In both methods the state trajectory is stored as the result of a simulation. Notice
that multiple shooting is just like a series of single shooting methods, with a de-
fect constraint added to make the trajectory continuous. Multiple shooting results
in a higher dimensional non-linear program, but it is sparse and more linear than
the program produced by single shooting. Therefore it is important, we can and
should employ a sparsity exploiting NLP solver.
Single shooting works well enough for simple problems, but it will almost cer-
tainly fail on problems that are more complicated. This is because the relationship
between the decision variables and the objective and constraint functions is not
well approximated by the linear (or quadratic) model that the NLP solver uses [2].
This is not a huge problem for multiple shooting as on generating shorter seg-
ments, the relationship becomes more linear.

7
Direct Collocation Methods

Chapter 3

Direct Collocation Methods

3.1 Overview

Arguably the most powerful methods for solving general optimal control problems
are direct collocation methods. A direct collocation method is a state and con-
trol parameterization method where the state and control are approximated using a
specified functional form. The time derivative of the state is approximated by dif-
ferentiating the interpolating polynomial and constraining the derivative to be equal
to the vector field at a finite set of collocation points.

The two most common forms of collocation are local collocation and global collo-
cation [1]. The methodology classification for direct collocation is h-methods and
p-methods.

3.1.1 h-methods and p-methods

Originally, direct collocation methods were developed as h-methods (for example,


Euler or Runge-Kutta methods) where the time interval is divided into a mesh and
the state is approximated using the same fixed-degree polynomial in each mesh
interval. Convergence in an h-method is then achieved by increasing the number
and placement of the mesh points. [8]

8
Direct Collocation Methods

For p-methods, we divide the time horizon [t0 ,t f ] into (fixed) mesh intervals. The
state is approximated in each interval by an N th order polynomial. Convergence
of the p method was then achieved by increasing the degree of the polynomial
approximation. For problems whose solutions are smooth and well-behaved, a p
Gaussian quadrature collocation method has a simple structure and converges at
an exponential rate. The most well developed p Gaussian quadrature methods
are those that employ either Legendre-Gauss (LG) points, Legendre-Gauss-Radau
(LGR) points, or Legendre-Gauss-Lobatto (LGL) points. [8]

Figure 3.1: Comparison between h-method and p-method: ref. [2]

3.2 Forms of Collocation Methods

3.2.1 Local Collocation

In a local approach, the time interval is divided into a large number of subinter-
vals called segments or finite elements and a small number of collocation points
are used within each segment. The segments are then linked via continuity condi-
tions on the state, the independent variable, and possibly the control. The rationale
for using local collocation is that a local method provides so-called local support
(i.e., the discretization points are located so that they support the local behavior

9
Direct Collocation Methods

of the dynamics) and is both computationally simple and efficient. [4]

A local collocation method follows a procedure in which the time interval [t0 ,t f ]
is divided into S subintervals [ts−1 ,ts ], (s = 1, . . . , S) where tS = t f . In order
to ensure continuity in the state across subintervals, the following compatibility
constraint is enforced at the interface of each subinterval:
x(t −i ) = x(t +i ), (s = 2, . . . , S − 1)
In the context of optimal control, local collocation has been employed using one
of two categories of discretization: Runge-Kutta methods and orthogonal colloca-
tion methods. [1]

3.2.2 Global Collocation

As a global collocation method, the number of segments is set to unity (i.e.,


S = 1) and the accuracy is assessed as a function of the number of collocation
points N. Thus global collocation methods need to apply p-methods using a single
interval. Global polynomial methods use a high-degree global polynomial (Cheby-
shev or Lagrange polynomials) for parameterization, and a set of orthogonal nodes
associated with the family of the polynomial for discretization.

The three most commonly used set of collocation points are Legendre-Gauss (LG),
Legendre-Gauss-Radau (LGR), and Legendre-Gauss-Lobatto (LGL) points. These
three sets of points are obtained from the roots of a Legendre polynomial and/or
linear combinations of a Legendre polynomial and its derivatives. All three sets
of points are defined on the domain [−1, 1], but differ significantly in that the
LG points include neither of the endpoints, the LGR points include one of the
endpoints, and the LGL points include both of the endpoints. [3]

10
Direct Collocation Methods

Figure 3.2: Differences between LGL, LGR, and LG collocation points: ref. [3]

3.3 Pseudospectral Method

A pseudospectral method is a form of orthogonal collocation, i.e., in a pseudospec-


tral method the state and control are approximated using orthogonal polynomial
and collocation is performed at chosen points, i.e., the collocation points are the
roots of an orthogonal polynomial (or linear combinations of such polynomials and
their derivatives).

Because pseudospectral methods are commonly implemented via orthogonal col-


location, the terms pseudospectral and orthogonal collocation are interchangeable.
One advantage to pseudospectral methods is that for smooth problems, pseudospec-
tral methods typically have faster convergence rates than other methods, exhibiting
“spectral accuracy”. [4]

11
Direct Collocation Methods

Figure 3.3: Distribution of nodes and collocation points for both the global and local
approaches (N = 20): ref. [4]

The results obtained in [4] suggest that, except in special circumstances, global or-
thogonal collocation is preferable to local orthogonal collocation. [4] also suggests
that, the global Gauss pseudospectral method (GPM) is much more accurate than
the local GPM for a given number of total collocation points. Furthermore, for
a desired accuracy, the global approach is computationally more efficient than the
local approach in smooth problems. For non-smooth problems the local and global
approach are quite similar in terms of accuracy.

3.4 Problem Formulation

For this illustration we employ local collocation with Legendre-Gauss (LG) collo-
cation points. The state at a set of collocation points (in addition to the beginning
of the interval) enters in the NLP as variables. An example of a choice of time

12
Direct Collocation Methods

points are the Legendre points of order d = 3 :

τ = [0, 0.112702, 0.500000, 0.887298]

Keeping the same control discretization as in single and multiple shooting:

u(t) = uk , for t ∈ [tk ,tk+1 ], k = 0, . . . , N − 1

the complete list of time points, with hk := tk+1 − tk , is:

tk, j := tk + hk τ j , for k = 0, . . . , N − 1 and j = 0, . . . , d

as well as the final time tN,0 = t f . Also let xk, j denote the states at these time
points. On each control interval, we shall define a Lagrangian polynomial basis:
d
τ − τr
L j (τ) = ∏
r=0,r6= j τ j − τr

Since the Lagrangian basis satisfies:



 1, if j = r ;
L j (τr ) =
 0, otherwise.
we can approximate the state trajectory approximation as a linear combination of
these basis functions:
d  
t − tk
x̃k (t) = ∑ Lr xk,r
r=0 hk

In particular we get approximations of the state time derivative at each collocation


point (not including τ0 ):
d d
1 1
ẋ˜k (tk, j ) = ∑ L̇r (τ j ) xk,r = ∑ Cr, j xk,r , (3.1)
hk r=0 hk r=0

as well as an approximation of the state at the end of the control interval:


d d
x̃k+1,0 = ∑ Lr (1) xk,r = ∑ Dr xk,r , (3.2)
r=0 r=0

where Cr, j := L̇r (τ j ) and Dr := Lr (1). Plugging in the approximation of the state
derivative (3.1) into the ODE gives us a set of collocation equations that needs to

13
Direct Collocation Methods

be satisfied for every state at every collocation point:

d
hk f (tk, j , xk, j , uk ) − ∑ Cr, j xk,r = 0, k = 0, . . . , N − 1, j = 1, . . . , d (3.3)
r=0

And the approximation of the end state (3.2) gives us a set of continuity equations
that must be satisfied for every control interval:
d
xk+1,0 − ∑ Dr xk,r = 0, k = 0, . . . , N − 1, (3.4)
r=0

These two sets of equations [(3.3) and (3.4)] take the place of the continuity equa-
tion (represented by the integrator call) in direct multiple shooting.

14
Optimal Control Problems

Chapter 4

Optimal Control Problems

4.1 Bryson-Denham Problem

4.1.1 Description

The Bryson-Denham optimal control problem is a benchmark test problem for op-
timal control algorithms. Consider the system

ẋ(t) = v(t)
(4.1)
v̇(t) = u(t)

The parameter u(t) ∈ R (acceleration) is adjusted over the time horizon from a
starting time of zero to a final time of one. The variable x(t) ∈ R is the position
and v(t) ∈ R is the velocity. Consider the optimal control problem

1 1
Z
minimize u(t)2 dt
2 0




 underlying system dynamics (4.1),



x(0) = x(1) = 0,
 (4.2)
subject to
v(0) = −v(1) = 1,







x(t) ≤ l, where l = 1 .

9

15
Optimal Control Problems

4.1.2 Single Shooting

Figure 4.1: Bryson-Denham problem by single shooting

4.1.3 Multiple Shooting

Figure 4.2: Bryson-Denham problem by multiple shooting

16
Optimal Control Problems

4.1.4 Direct Collocation

Figure 4.3: Bryson-Denham problem by direct collocation

4.2 Double Integrator Bang-Bang Problem

4.2.1 Description

Consider the system ẍ = u, u ∈ [−1, 1] which can represent a car with posi-
tion x ∈ R and with bounded acceleration u acting as the control (negative accel-
eration corresponds to braking). Let us study the problem of parking the car at
the origin, i.e., bringing it to rest at x = 0, in minimal time. It is clear that the
system can indeed be brought to rest at the origin from every initial condition.
However, since the control is bounded, we cannot do this arbitrarily fast (we are
ignoring the trivial case when the system is initialized at the origin). Thus we
expect that there exists an optimal control u∗ which achieves the transfer in the
smallest amount of time. [6]

17
Optimal Control Problems

We know that the dynamics of the double integrator are equivalently described by
the state-space equations
ẋ1 = x2
(4.3)
ẋ2 = u

where x1 ∈ R is the displacement of car and x2 ∈ R is the velocity. We assume


that the initial values and final values x1 (t0 ), x2 (t0 ), x1 (t f ) and x2 (t f ) are given
(which are (10, 0, 0, 0) respectively).

4.2.2 Single Shooting

Figure 4.4: Double integrator bang-bang problem by single shooting

18
Optimal Control Problems

4.2.3 Multiple Shooting

Figure 4.5: Double integrator bang-bang problem by multiple shooting

4.2.4 Direct Collocation

Figure 4.6: Double integrator bang-bang problem by direct collocation

19
Optimal Control Problems

4.3 Conclusions

Except for the single-shooting method used to solve bang-bang double integrator
problem, all the methods gave decently correct results for each of the problems.
The non-smoothness of the optimal control function in bang-bang problems might
have caused single-shooting method to fail.

Looking at these direct methods for optimal control, we can summarise them in
the following way: When we went from direct single shooting to direct multiple
shooting we essentially traded non-linearity for problem size. The NLP in sin-
gle shooting is small, but often highly nonlinear, whereas the NLP for multiple-
shooting is larger, but less nonlinear and with a sparsity structure that can be
exploited efficiently. Direct collocation is to take one more step in the same di-
rection, adding even more degrees of freedom. The resulting NLP is even larger,
but has even more structure that can be exploited.

Direct methods are easier to implement with effective incorporation of state and
control constraints. But, this comes at the cost of accuracy and that is why it is
important to choose an appropriate method based on the problem and requirements.

20
References

References

[1] Anil V. Rao. A survey of numerical methods for optimal control. Advances in
the Astronautical Sciences, 135, 2010.

[2] Matthew Kelly. Transcription methods for trajectory optimization: a beginners


tutorial. arXiv: Optimization and Control, 2017.

[3] Divya Garg et al. Direct trajectory optimization and costate estimation of gen-
eral optimal control problems using a radau pseudospectral method. AIAA Guid-
ance, Navigation, and Control Conference, 2009.

[4] Anil V. Rao Geoffrey T. Huntington. Comparison of global and local colloca-
tion methods for optimal control. Journal of Guidance, Control, and Dynamics,
31, 2008.

[5] Matthew Kelly. An introduction to trajectory optimization: How to do your


own direct collocation. Society for Industrial and Applied Mathematics, 59, 2017.

[6] Daniel Liberzon. Calculus of Variations and Optimal Control Theory: A Concise
Introduction. Princeton University Press, 2012.

[7] Moritz Diehl. Lecture Notes on Optimal Control and Estimation,


2014. URL: https://syscop.de/files/2015ss/events/tempo/
optimalcontrolDiehl.pdf. Accessed: 25/04/2021.

[8] Michael. A. Patterson Anil V. Rao and William W. Hager. A ph collocation


scheme for optimal control. Optimal Control Applications and Methods, 36(4),
2014.

21

You might also like