0% found this document useful (0 votes)
36 views

Distributed Cont

This paper proposes a distributed receding horizon control approach for stabilizing a group of dynamically coupled nonlinear systems. Each subsystem solves its own optimal control problem to determine the input, while exchanging information with neighboring subsystems. Conditions for feasibility and stability of the distributed approach are provided, requiring sufficient control updates, limited dynamic coupling between subsystems, and consistency between the trajectories each subsystem computes and those assumed by its neighbors. The approach is demonstrated for a set of coupled Van der Pol oscillators modeling a walking robot experiment. Numerical results show the distributed approach performs well and has lower computational cost than a centralized approach.

Uploaded by

Baba Jee
Copyright
© Attribution Non-Commercial (BY-NC)
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
36 views

Distributed Cont

This paper proposes a distributed receding horizon control approach for stabilizing a group of dynamically coupled nonlinear systems. Each subsystem solves its own optimal control problem to determine the input, while exchanging information with neighboring subsystems. Conditions for feasibility and stability of the distributed approach are provided, requiring sufficient control updates, limited dynamic coupling between subsystems, and consistency between the trajectories each subsystem computes and those assumed by its neighbors. The approach is demonstrated for a set of coupled Van der Pol oscillators modeling a walking robot experiment. Numerical results show the distributed approach performs well and has lower computational cost than a centralized approach.

Uploaded by

Baba Jee
Copyright
© Attribution Non-Commercial (BY-NC)
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 7

Distributed Receding Horizon Control of Coupled

Nonlinear Oscillators: Theory and Application


William B. Dunbar
AbstractThis paper extends recent results on distributed
receding horizon control (RHC) of dynamically coupled nonlin-
ear systems subject to decoupled input constraints. Motivating
examples of such systems include chains of coupled oscilla-
tors and supply chain management systems. Conditions for
feasibility and stability of the distributed RHC algorithm are
stated, with substantially less conservative requirements then
previously derived. The conditions are shown to be satised
for a set of coupled Van der Pol oscillators that model a
walking robot experiment. Numerical experiments show good
performance and demonstrate the computational savings over
centralized RHC.
Keywords: receding horizon control, model predictive
control, distributed control, decentralized control, large scale
systems.
I. INTRODUCTION
The problem of interest is to design a distributed controller
for a set of dynamically coupled nonlinear subsystems that
are required to perform stabilization in a cooperative way.
Examples of such situations where distributed control is
desirable include certain large scale process control systems
[18] and supply chain management systems [2], [7]. The
control approach advocated here is receding horizon control
(RHC). In RHC, the current control action is determined
by solving a nite horizon optimal control problem online
at every update. In continuous time formulations, each opti-
mization yields an open-loop control trajectory and the initial
portion of the trajectory is applied to the system until the next
update. A survey of RHC, also known as model predictive
control, is given by Mayne et al. [12]. Advantages of RHC
are that a large class of performance objectives, dynamic
models and constraints can in principle be accommodated.
In this paper, subsystems that are dynamically coupled are
referred to as neighbors. The work presented is an extension
of a recent work [5]. As in [5], each subsystem is assigned
its own optimal control problem, optimizes only for its
own control at each update, and exchanges information
with neighboring subsystems. The primary motivations for
pursuing such a distributed implementation are to enable
the autonomy of the individual subsystems and reduce
the computational burden of centralized implementations.
The requirement of distributed control in the presence of
constraints is particularly true in the case of supply chain
problems [4], since stages within a chain employ decentral-
ized decision making. In comparison to [5], the theoretical
conditions which guarantee feasibility and stability are much
W. B. Dunbar is an Assistant Professor of Computer Engineering, Uni-
versity of California, Santa Cruz, 95064, USA. [email protected].
less conservative, and shown to be satised in the case of a
set of coupled nonlinear oscillators.
Previous work on distributed RHC of dynamically coupled
systems include Jia and Krogh [10] and Acar [1]. These
papers address coupled liner time-invariant subsystem dy-
namics with quadratic separable cost functions. State and
input constraints are not included, aside from a stability
constraint in [10] that permits state information exchanged
between the subsystems to be delayed by one update period.
In another work, Jia and Krogh [11] solve a min-max prob-
lem for each subsystem, where again coupling comes in the
dynamics and the neighboring subsystem states are treated
as bounded disturbances. Stability is obtained by contracting
each subsystems state at every sample period, until the
objective set is reached. As such, stability does not depend
on information updates between neighbors, although such
updates may improve performance. More recently, Venkat et
al. [16], [17] have designed a distributed model predictive
control (MPC) algorithm for coupled LTI subsystems and
compared it to centralized and decentralized alternatives. In
their formulation, subsystems are coupled solely through the
control inputs. Consequently, feasibility and stability analysis
is leveraged by the diagonally decoupled and linear form of
the state dynamics, for which the state solution can be carried
out analytically given the set of all control trajectories.
Section II begins by dening the nonlinear coupled
subsystem dynamics and control objective. In Section III,
distributed optimal control problems are dened for each
subsystem, and the distributed RHC algorithm is stated.
Feasibility and stability results are then given in Section
IV. Key requirements are that the receding horizon updates
happen at a sufcient rate, the amount of dynamic coupling
remain below a quantitative threshold, and each distributed
optimal state trajectory satisfy a consistency constraint. The
consistency constraint ensures that the computed state trajec-
tory of each subsystem is not too far from the trajectory that
each neighbor assumes for that subsystem, at each receding
horizon update. In Section V, the theory is applied to the
problem of regulating a set of coupled Van der Pol oscillators
that capture the thigh and knee dynamics of a walking robot
experiment [8]. Fnally, Section VI provides conclusions. To
keep the paper conference length, all proofs are omitted and
can be found in [6].
II. SYSTEM DESCRIPTION AND OBJECTIVE
In this section, the system dynamics and control objective
are dened. For any vector x R
n
, x
P
denotes the P-
weighted 2-norm, dened by x
2
P
= x
T
Px, and P is any
positive-denite real symmetric matrix. Also,
max
(P) and

min
(P) denote the largest and smallest eigenvalues of P,
respectively. Often, the notation x is understood to mean
x(t) at some instant of time t R. THe objective is to
stabilize a group of N
a
2 dynamically coupled agents
toward the origin in a cooperative and distributed way using
RHC. For each agent i {1, ..., N
a
}, the state and control
vectors are denoted z
i
(t) R
n
and u
i
(t) R
m
, respectively,
at any time t t
0
R. The dimension of every agents
state (control) are assumed to be the same, for notational
simplicity and without loss of generality. The concatenated
vectors are denoted z = (z
1
, ..., z
Na
) and u = (u
1
, ..., u
Na
).
The dynamic coupling between the agents is topologically
identied by a directed graph G = (V, E), where V =
{1, ..., N
a
} is the set of nodes (agents) and E V V
is the set of all directed edges between nodes in the graph.
The set E is dened in the following way. If any components
of z
j
appear in the dynamic equation for agent i, for some
j V, it is said that j is an upstream neighbor of agent
i, and N
u
i
V denotes the set of upstream neighbors of
any agent i V. The set of all directed edges is dened as
E = {(i, j) V V | j N
u
i
, i V}. For every i V,
it is assumed that z
i
appears in the dynamic equation for i,
and so i N
u
i
for every i V. In the language of graph
theory, then, every node has a self-loop edge in E. Note that
j N
u
i
does not necessarily imply i N
u
j
.
It will also be useful to reference the set of agents for
which any of the components of z
i
arises in their dynamical
equation. This set is referred to as the downstream neighbors
of agent i, and is denoted N
d
i
. The set of all directed edges
can be equivalently dened as E = {(i, j) V V | i
N
d
j
, i V}. Note that j N
u
i
if and only if i N
d
j
,
for any i, j V. It is assumed in this paper that the
graph G is connected. Consequently, for every i V, the
set
_
N
d
i
N
u
i
_
\ {i} = , and every agent is dynamically
coupled to at least one other agent. It is also assumed that
agents can receive information directly from each and every
upstream neighbor, and agents can transmit information
directly to each and every downstream neighbor. The coupled
time-invariant nonlinear system dynamics for each agent
i V are given by
z
i
(t) = f
i
(z
i
(t), z
i
(t), u
i
(t)), t t
0
, (1)
where z
i
= (z
j1
, ..., z
j
l
), l = |N
u
i
|, denotes the concate-
nated vector of the states of the upstream neighbors of i.
Each agent i is also subject to the decoupled input constraints
u
i
(t) U, t t
0
. The set U
N
is the N-times Cartesian
product U U. In concatenated vector form, the system
dynamics are
z(t) = f(z(t), u(t)), t t
0
, (2)
given z(t
0
), and f = (f
1
, ..., f
Na
).
Assumption 1: The following holds: (a) f is C
2
and 0 =
f(0, 0); (b) system (2) has a unique solution for any z(t
0
)
and any piecewise right-continuous control u : [t
0
, )
U
Na
; (c) U R
m
is compact, containing the origin in its
interior.
Consider now the linearization of (1) around the origin,
denoting A
il
= f
i
/z
l
(0, 0) and B
i
= f
i
/u
i
(0, 0). As
in many RHC formulations, a feedback controller for which
the closed-loop system is asymptotically stabilized inside a
neighborhood of the origin will be utilized.
Assumption 2: For every agent i V, there exists a
decoupled static feedback u
i
= K
i
z
i
such that A
di
A
ii
+
B
i
K
i
is Hurwitz, and the closed-loop linear system z = A
c
z
is asymptotically stable, where A
c
[f
z
(0, 0) + f
u
(0, 0)K]
and K = diag(K
1
, ..., K
Na
).
The decoupled linear feedbacks above are referred to as
terminal controllers. Associated with the closed-loop lin-
earization, denote the block-diagonal Hurwitz matrix A
d
=
diag(A
d1
, ..., A
dNa
) and the off-diagonal matrix A
o
= A
c

A
d
. Assumption 2 inherently presumes decoupled stabiliz-
ability and that the coupling between subsystems in the
linearization is sufciently weak, as discussed and quantied
in the survey paper [15]. The terminal controllers will be
employed in a prescribed neighborhood of the origin.
III. DISTRIBUTED RECEDING HORIZON CONTROL
In this section, N
a
separate optimal control problems
are dened and the distributed RHC algorithm. In every
distributed optimal control problem, the same constant pre-
diction horizon T (0, ) and constant update period
(0, T] are used. In practice, the update period
(0, T] is typically the sample interval. By the distributed
implementation presented here, additional conditions on
are required, as quantied in the next section. Denote the
update time t
k
= t
0
+ k, where k N = {0, 1, 2, ...}. In
the following implementation, every distributed RHC law is
updated globally synchronously, i.e., at the same instant of
time t
k
for the k
th
-update.
At each update, every agent optimizes only for its own
predicted open-loop control, given its current state. Since the
dynamics of each agent i depend on states z
i
, that agent
will presume some trajectories for z
i
over each prediction
horizon. To that end, prior to each update, each agent i
receives an assumed state trajectory z
j
from each upstream
neighbor j N
u
i
. Likewise, agent i transmits an assumed
state trajectory z
i
to every downstream neighbor j N
d
i
,
prior to each update. By design, then, the assumed state
trajectory for any agent is the same in the distributed optimal
control problem of every downstream neighbor. Since the
models are used with assumed trajectories for upstream
neighbors, there will be a discrepancy, over each optimization
time window, between the predicted open-loop trajectory and
the actual trajectory that results from every agent applying
the predicted control. This discrepancy is identied by using
the following notation. Recall that z
i
(t) and u
i
(t) are the
actual state and control, respectively, for each agent i V
at any time t t
0
. Associated with update time t
k
, the
trajectories for each agent i V are denoted
z
p
i
(t; t
k
) the predicted state trajectory,
z
i
(t; t
k
) the assumed state trajectory,
u
p
i
(t; t
k
) the predicted control trajectory,
where t [t
k
, t
k
+ T]. Consistent with the ordering of z
i
,
let z
i
(; t
k
) be the assumed open-loop state trajectories of
the upstream neighbors of i. For any agent i V, then, the
predicted state trajectory satises
z
p
i
(t; t
k
) = f
i
(z
p
i
(t; t
k
), z
i
(t; t
k
), u
p
i
(t; t
k
)), (3)
for all t [t
k
, t
k
+ T], given z
p
i
(t
k
; t
k
) = z
i
(t
k
). The
assumed state trajectory for each agent i V is given by
z
i
(t; t
k
) =
_
z
p
i
(t; t
k1
), t [t
k
, t
k1
+ T)
z
K
i
(t), t [t
k1
+ T, t
k
+ T]
(4)
where z
K
i
is the solution to z
K
i
(t) = A
di
z
K
i
(t) with initial
condition z
K
i
(t
k1
+ T) = z
p
i
(t
k1
+ T; t
k1
). By con-
struction, each assumed state trajectory z
i
is the remainder
of the previously predicted trajectory, concatenated with the
closed-loop linearization response that ignores coupling. The
collective actual state trajectories for the agents over any
update window [t
k
, t
k+1
) is given by
z(t) = f(z(t), u
p
(t; t
k
)), t [t
k
, t
k+1
), (5)
given z(t
k
). While the actual and predicted state trajectories
do have the same initial condition z
i
(t
k
) for each i V,
they typically diverge over each update window [t
k
, t
k+1
],
and z
p
(t
k+1
; t
k
) = z(t
k+1
) in general. The reason is that,
while the predicted state trajectories in (3) are based on
the assumption that neighbors continue along their previous
trajectory, neighbors in fact compute and employ their own
updated predicted control trajectory. Therefore, the actual
state evolves according to (5). The challenge then is to
generate a distributed RHC algorithm that has feasibility
and stability properties in the presence of the discrepancy
between predicted and actual state trajectories. A desirable
property of any RHC algorithm is to have feasible state and
control trajectories at any update, as the trajectories can be
used to preempt the optimization algorithm used to solve the
optimal control problem. In many formulations, the feasible
state trajectory is the remainder of the previous trajectory
concatenated with the response under a terminal controller
[3], [12], [13]. While z
i
(; t
k
) is such a trajectory, it cannot
be used since z
i
(t
k
; t
k
) = z
i
(t
k
). Still, a feasible control
trajectory exists. Indeed, a contribution of this paper is to
show that a feasible control is the remainder of the previous
control trajectory concatenated with the terminal controller,
with the corresponding feasible state trajectory starting from
the true state at each update time. The feasible state and
control trajectories at any update t
k
are denoted z
i
(; t
k
) and
u
i
(; t
k
), respectively. The feasible state trajectory satises

z
i
(t; t
k
) = f
i
( z
i
(t; t
k
), z
i
(t; t
k
), u
i
(t; t
k
)) , (6)
given initial condition z
i
(t
k
; t
k
) = z
i
(t
k
), and the feasible
control is given by
u
i
(t; t
k
) =
_
u
p
i
(t; t
k1
), t [t
k
, t
k1
+ T)
K
i
z
i
(t; t
k
), t [t
k1
+ T, t
k
+ T]
. (7)
The feasible control trajectory u
i
is the remainder of the
previously predicted control trajectory, concatenated with the
linear control applied to the nonlinear model and based on
the decoupled linear responses for each upstream neighbor.
In the next section, feasibility and stability will be proven.
In each local optimal control problem, for any agent i V
at update time t
k
, the cost function J
i
(z
i
(t
k
), u
p
i
(; t
k
)) is
given by
_
t
k
+T
t
k
z
p
i
(s; t
k
)
2
Qi
+u
p
i
(s; t
k
)
2
Ri
ds +z
p
i
(t
k
+ T; t
k
)
2
Pi
,
where Q
i
= Q
T
i
> 0, R
i
= R
T
i
> 0 and P
i
= P
T
i
>
0. The matrix P
i
= P
T
i
> 0 is chosen to satisfy the
Lyapunov equation P
i
A
di
+ A
T
di
P
i
=

Q
i
, where

Q
i
=
Q
i
+ K
T
i
R
i
K
i
. Denoting P = diag(P
1
, ..., P
Na
) and

Q =
diag(

Q
1
, ...,

Q
Na
), it follows that PA
d
+ A
T
d
P =

Q and

Q > 0. Decoupled terminal state constraints will be included


in each local optimal control problem. A lemma used to
dene the terminal state constraint sets and to guarantee that
the terminal controllers are stabilizing inside the sets is now
presented. The proof of the lemma utilizes an assumption
that limits the amount of coupling between neighboring
subsystems in the linearization.
Assumption 3: PA
o
+ A
T
o
P

Q/2.
Lemma 1: Suppose that Assumptions 13 hold. There
exists a positive constant (0, ) such that the set


_
z R
nNa
| z
P

_
, is a positively invariant
region of attraction for both the closed-loop linearization
z(t) = A
c
z(t) and the closed-loop nonlinear system z(t) =
f(z(t), Kz(t)). Additionally, Kz U
Na
for all z

.
The parameter (0, ) that satised the conditions of
the lemma can be found numerically, as described in [6],
[13]. In each local optimal control problem, the terminal state
constraint set for each i V is

i
()
_
z
i
R
n
| z
i

Pi
/
_
N
a
_
. (8)
By construction, if z
1
()
Na
(), then the
decoupled controllers can stabilize the system to the origin,
since z
i

2
Pi


2
Na
, i V implies

iV
z
i

2
Pi

2
.
Suppose then that at some time t

t
0
, z
i
(t

)
i
() for
every i V. Then, from Lemma 1, stabilization is achieved
if every agent employs their decoupled static feedback con-
troller K
i
z
i
(t) for all time t t

. Thus, the objective of the


RHC law is to drive each agent i to the set
i
(). Once all
agents have reached these sets, they switch to their decoupled
controllers for stabilization. The collection of local optimal
control problems is now dened.
Problem 1: Let (0, ) satisfy the conditions in
Lemma 1, and let q {1, 2, 3, ...} be any positive integer.
For each agent i V and at any update time t
k
, k 1:
Given: z
i
(t
k
), z
i
(t; t
k
) and z
i
(t; t
k
), t [t
k
, t
k
+ T];
Find: the control trajectory u
p
i
(; t
k
) : [t
k
, t
k
+T] U that
minimizes J
i
(z
i
(t
k
), u
p
i
(; t
k
)), subject to the constraints
z
p
i
(t; t
k
)
2
Pi
z
i
(t; t
k
)
2
Pi
+ a, (9)
z
p
i
(t; t
k
) z
i
(t; t
k
)
2
Pi
z
i
(t; t
k
) z
i
(t; t
k
)
2
Pi
+ b, (10)
for all t [t
k
, t
k
+ T], where a = [/(8T

N
a
)]
2
, b =
[/(2(q + 1)N
a
)]
2
, z
p
i
(; t
k
) satises the dynamic equation
(3) and the terminal constraint z
p
i
(t
k
+ T; t
k
)
i
(/2),
with
i
dened in (8).
Equation (9) is utilized to prove that the distributed RHC
algorithm is stabilizing. Equation (10) is referred to as the
consistency constraint, which requires that each predicted
trajectory remain close to the assumed trajectory (that neigh-
bors assume for that agent). In particular, the predicted
trajectory z
p
i
must remain nearly as close to the assumed
trajectory z
i
as the feasible trajectory z
i
, with an added
margin of freedom parameterized by (/2(q + 1)N
a
)
2
. In
the analysis that follows, the consistency constraint (10)
is a key equation in proving that z
i
is a feasible state
trajectory at each update. The constant q {1, 2, 3, ...} is
a design parameter, and the choice for q will be motivated
in Section IV. Before stating the distributed RHC algorithm,
an assumption (standard in centralized implementations) is
made to facilitate the initialization phase.
Assumption 4: Given z(t
0
) at initial time t
0
, there exists
a feasible control u
p
i
(; t
0
) U, [t
0
, t
0
+ T], for each
agent i V, such that the solution to the full system z() =
f(z(), u
p
(; t
0
)), denoted z
p
(; t
0
), satises z
p
i
(t
0
+T; t
0
)

i
(/2) and results in a bounded cost J
i
(z
i
(t
0
), u
p
i
(; t
0
))
for every i V. Moreover, each agent i V has access to
u
p
i
(; t
0
).
Let Z R
nNa
denote the set of initial states for which there
exists a control satisfying the conditions in Assumption 4.
The control algorithm is now stated.
Algorithm 1: The distributed receding horizon control law
for any agent i V is as follows:
Data: z(t
0
), u
p
i
(; t
0
) satisfying Assumption 4, T
(0, ), (0, T], q {1, 2, 3, ...}.
Initialization: At time t
0
, if z(t
0
)

, then apply the


terminal controller u
i
(t) = K
i
z
i
(t), for all t t
0
. Else:
Controller:
1) Over any interval [t
k
, t
k+1
), k N:
a) At any time t [t
k
, t
k+1
), if z(t)

, then
apply the terminal controller u
i
(t

) = K
i
z
i
(t

),
for all t

t. Else:
b) Compute z
i
(; t
k+1
) according to (4) and trans-
mit it to every downstream neighbor l N
d
i
.
c) Receive z
j
(; t
k+1
) from every upstream neighbor
j N
u
i
and assemble z
i
(; t
k+1
).
d) Apply u
p
i
(; t
k
), [t
k
, t
k+1
).
2) At any time t
k+1
, k N:
a) Measure z
i
(t
k+1
).
b) Compute z
i
(; t
k+1
) according to (6).
c) Solve Problem 1, yielding u
p
i
(; t
k+1
).
Part 1(a) of Algorithm 1 presumes that the every agent can
obtain the full state z(t). This requirement is a theoretical
artifact needed when employing dual-mode control, so that
switching occurs only when the conditions of Lemma 1 are
satised. In the next section, it is shown that the distributed
RHC policy drives the state z(t
l
) to

after a nite number


of updates l, and the state remains in

for all future


time. If

is sufciently small for stability purposes, then


agents do not need access to the full state at any update,
since RHC can be employed for all time. The next section
states the theoretical results showing that the distributed RHC
algorithm is feasible at every update and stabilizing.
IV. THEORETICAL RESULTS
In this section, the feasibility and stability results are
states, with all proofs provided in [6]. A desirable property
of the implementation is that the existence of a feasible
solution to Problem 1 at update k = 0 implies the existence
of a feasible solution for any subsequent update k 1. If
Assumption 4 holds true, the rst result of this section is
that a feasible control solution to Problem 1 for any i V
and at any time t
k
, k 1, is u
p
i
(; t
k
) = u
i
(; t
k
), with
u
i
dened by (7). The corresponding feasible state trajec-
tory dened by (6) is z
p
i
(; t
k
) = z
i
(; t
k
). The feasibility
result requires a local Lipschitz property on the collective
dynamics. In vector form, the collective set of differential
equations for the predicted trajectories (using (3) for each
i V) is denoted z
p
(t; t
k
) = F(z
p
(t; t
k
), z(t; t
k
), u
p
(t; t
k
)),
t [t
k
, t
k
+T], where F : R
nNa
R
nNa
R
mNa
R
nNa
,
z
p
= (z
p
1
, ..., z
p
Na
), z = ( z
1
, ..., z
Na
) and u
p
= (u
p
1
, ..., u
p
Na
).
By denition, the function F satises F(z, z

, u) = f(z, u)
whenever z = z

.
Assumption 5: Given P and R, there exist positive con-
stants and such that the Lipschitz bound
F(z, z

, u)
P
z
P
+ z

P
+ u,
holds for all z, y Z, and u U
Na
.
More generally, the Lipschitz bound would take the form

F(z, z

, u)
P
z
P
+

z

P
+ u for some positive
constants ( ,

, ). Thus, Assumption 5 presumes that one
can identify the Lipschitz constants ( ,

, ), and that the
differential equation f (or F) is already normalized by , so
that =

/ and = / . The local Lipschitz constant
represents a normalized measure of the amount of coupling
in the collective dynamic model. The feasibility result is now
stated.
Theorem 1: Suppose that Assumptions 15 hold, z(t
0
)
Z and the following parametric conditions hold:
e
(1+
K
)
ln
_
(q + 1)
2
/q
2

min
iV
{
min
(P
1/2
i

Q
i
P
1/2
i
)}
e
(1+
K
)

r/(r + 1)
c(q + 1)

N
a
, (11)
2T(r + 1) exp [T + (1 + +
K
)] 1, (12)
where constants c and
K
are denes as c =
1/(10
max
(

Q
1/2
P

Q
1/2
))+
1/2
max
(P
1/2
A
T
o
PA
o
P
1/2
),

K
=
1/2
max
(P
1/2
K
T
KP
1/2
), and constants q, r
{1, 2, 3, ...} are chosen design parameters. Then, for every
agent i V, the control and state pair ( u
i
(; t
k
), z
i
(; t
k
)),
dened by equations (6) and (7), is a feasible solution to
Problem 1 at every update k 1.
The purpose of the design parameters q, r {1, 2, 3, ...}
is to shift the bounds on and , as necessary to nd
feasible values for a specic problem. The larger the chosen
value of q, the smaller the lower and upper bounds on . for
example. The ability to shift the feasible range for is useful
for design purposes, as will be demonstrated in Section V.
Also, larger values of q reduce the margin in the consistency
constraint (10) that bounds how much the predicted state
can deviate from the assumed state. Equation (12) places
an upper bound on the Lipschitz coupling constant . By
increasing the design parameter r, one can increase the upper
bound on at the price of requiring a tighter bound on .
The utility of being able to choose r is also demonstrated in
Section V. Parametric conditions which ensure the stability
of the closed-loop system (5) are now stated.
Theorem 2: Suppose that Assumptions 15 hold, z(t
0
)
Z, conditions (11)(12) are satised, and the following
parametric conditions hold
T 8, (q + 1) 2
T

. (13)
Then, by application of Algorithm 1, the closed-loop system
(5) is asymptotically stabilized to the origin.
The feasibility and stability results in this paper are related
to those of Michalska and Mayne [13], who demonstrated
robust feasibility and stability in the presence of model
error by placing parametric bounds on (combinations of)
the update period and a Lipschitz constant. While there is
no model error here, bounds are likewise derived to ensure
robustness to the bounded discrepancy between what agents
do, and what their neighbors believe they will do.
V. COUPLED OSCILLATORS
In this section, the example of three coupled Van der Pol
oscillators is considered for application of the distributed
RHC algorithm. The three oscillators modeled here are
physically meaningful in that they capture the thigh and knee
dynamics of a walking robot experiment [8].In the following,

1
[/2, /2] is the relative angle between the two
thighs,
2
[/2, /2] is the right knee angle (relative to
the right thigh), and
3
[/2, /2] is the left knee angle
(relative to left thigh). The controlled equations of motion in
units of (rad/sec)
2
are

1
(t) = 0.1
_
1 5.25
2
1
(t)

1
(t)
1
(t) + u
1
(t)

2
(t) = 0.01
_
1 p
2
(
2
(t)
2e
)
2
_

2
(t) 4(
2
(t)
2e
)
+ 0.057
1
(t)

1
(t) + 0.1(

2
(t)

3
(t)) + u
2
(t)

3
(t) = 0.01
_
1 p
3
(
3
(t)
3e
)
2
_

3
(t) 4(
3
(t)
3e
)
+ 0.057
1
(t)

1
(t) + 0.1(

3
(t)

2
(t)) + u
3
(t),
subject to |u
i
(t)| 1, t 0, i = 1, 2, 3.
Two-phase biped locomotion is generated by these
equations with zero control (open-loop) and time-
varying parameter values, given by (
2e
,
3e
, p
2
, p
3
)(t) =
(0.227, 0.559, 6070, 192) for t [0, ), and equal to
(0.559, 0.226, 226, 5240) for t [, 2). Figure 1
shows the resulting open-loop stable limit cycle response,
starting from the initial position (40, 3, 3) degrees, with

i
(0) = 0 for i = 1, 2, 3. Through perturbation analysis
and the method of harmonic balance, the limit cycle
is closely approximated by
lc
1
(t) = (50/180) cos(t),

lc
2
(t) =
2e
+ (3/180
2e
) cos(2t), and

lc
3
(t) =
3e
+ (3/180
3e
) cos(2t). The chosen
initial condition demonstrates the attractivity of the stable
0 5 10 15
1
0.5
0
0.5
1
time (sec)
(
r
a
d
)

3
Fig. 1. Open-loop stable limit cycle, showing the angular positions starting
from (40, 3, 3) degrees with zero initial angular velocity.
limit cycle. For example, note that the amplitude of
1
(t)
starts at 0.70 radians and approaches 0.87 radians, the
amplitude of
lc
1
(t). While the robot has 6 total degrees of
freedom when walking in accordance with the limit cycle
response above, the remaining degrees of freedom (including
two ankles and one free foot) can be derived from the three
primary degrees of freedom,
1
,
2
and
3
[8]. With zero
control, there are two equilibrium conditions. One is the
limit cycle dened above, and the other is the unstable xed
point (
1
,
2
,
3
) = (
1e
,
2e
,
3e
) with
1e
=

i
= 0 for
i = 1, 2, 3. A reasonable control objective is to command
torque motors (controls u
i
) to drive the three angles from
the stable limit cycle response to the xed point; that is,
to stably bring the robot to a stop. To do so within one
half-period of the limit cycle response means that one set of
parameter values (
2e
,
3e
, p
2
, p
3
) can be considered in the
model. As such, for control purposes, these parameters are
assumed to take on the values (0.227, 0.559, 6070, 192).
In this way, discontinuous dynamic equations are also
avoided. Now, through a change of variables, the dynamics
and input constraints satisfy the conditions of Assump. 1.
Denoting z
i
= (
i

ie
,

i
), the dynamics are linearized
around (z
i
, u
i
) = (0, 0). The matrix A
11
has unstable
eigenvalues 0.05 j, and the matrices A
22
and A
33
are
unstable with eigenvalues 0.0552j. For all three oscillators,
the dynamics are linearly controllable around the origin. In
accordance with Assumption 2, the following gain matrices
are used to stabilize the linearized dynamics: K
1
= [3.6 5.3],
K
2
= K
3
= [2.0 5.0]. The resulting closed-loop matrix A
c
has eigenvalues (1.1, 4.1, 3, 2.4 0.5j, 2). For the
cost function J
i
, the chosen weights are Q
i
= diag(30, 30)
and R
i
= 0.1, i = 1, 2, 3. Then, each P
i
is calculated accord-
ing to the Lyapunov equation on page 3. Since the maximum
eigenvalue of PA
o
+ A
T
o
P

Q/2 is 11, Assumption 3
is satised. The constraint parameter = 0.2 satises the
conditions of Lemma 1, with the calculated for described in
[6]. In accordance with Assumption 4, a centralized optimal
control problem is solved at initial time t
0
= 0. In this
problem, and in the centralized RHC implementation, the
sum of the three cost functions J
1
+ J
2
+ J
3
is minimized,
enforcing terminal state and input constraints with a horizon
time of T = 6 seconds. The initial condition is kept the same
as that shown in Figure 1.
To solve the centralized optimal control problem, and
each of the distributed optimal control problems, the same
approach is used. The computer with MATLAB 7.0 software
has a 2.4 GHz Intel Pentium(R) 4 CPU, with 512 MB of
RAM. In the spirit of the Nonlinear Trajectory Generation
package developed by Milam et al. [14], a collocation
technique is employed within MATLAB. First, each angular
position trajectory
i
(t) is parameterized as a C
2
[t
k
, t
k
+T]
6-th order B-spline polynomial. The constraints and cost
functions are evaluated at 121 breakpoints over each 6 second
time window. The resulting nonlinear programming problem
is solved using the fmincon function, generating the 27 B-
spline coefcients for each position
i
(t). Using the concept
of differential atness, the control inputs u
i
are not param-
eterized as polynomials for which the coefcients must also
be calculated. Instead, each control input is dened in terms
of the parameterized positions
i
(t) and their derivatives
through the dynamics. With an update period of = 0.15
seconds, the centralized RHC state and control response is
shown in Figure 2. The position and control trajectories are
0 1 2 3 4 5 6
0.3
0.2
0.1
0
0.1
0.2
0.3
0.4
0.5
0.6
time (sec)
(
r
a
d
)
CRHC

1,C

2,C

3,C
0 1 2 3 4 5 6
1
0.5
0
0.5
1
time (sec)
(
r
a
d
/
s
e
c
2
)
CRHC
u
1,C
u
2,C
u
3,C
Fig. 2. The centralized RHC response, showing the angular position
trajectories
i,C
(left plot) and the control trajectories u
i,C
(right plot),
for each i = 1, 2, 3.
denoted
i,C
and u
i,C
, respectively. Note that the positions are
brought suitably close to their xed point values (shown by
dashed lines) within a half-period of seconds, validating
the assumption that the model parameters (
2e
,
3e
, p
2
, p
3
)
are constant over the time horizon of 6 seconds.
With an initially feasible solution available, the distributed
RHC algorithm can be employed. Before presenting the
results, the theoretical conditions are evaluated. In total, the
parametric equations that must be satised are (11)(13). In
accordance with Assumption 5, the Lipschitz parameters are
identied as

F(z, z

, u)
P
4z
P
+0.1z

P
+1u. To
facilitate calculation of an update period that satises the
parametric conditions, time scaling is introduced to normal-
ize the horizon time from T = 6 seconds to 1 second. For
the dynamics

F, let (t) = t/T [0, 1] such that
d
d
z() =
T

F(z(), z

(), u()) for all [0, 1]. Now, the scaled


dynamics satisfy T

F(z, z

, u)
P
4Tz
P
+0.1Tz

P
+
Tu. To get into the normalized form, the dynamics are
scaled as F =

F/(4T). Then, the normalized Lipschitz
bounds become F(z, z

, u)
P
z
P
+ z

P
+ u,
where = 0.1/4 = 0.025 and = 1/4 = 0.25. Choosing
the design parameters q = 90 and r = 6, the lower bound in
(11) is 0.025 and the upper bound is 0.028. So, the update
period (for the time-scaled dynamics) is chosen to be =
0.025 seconds. Also, the left hand side of (12) is 0.997, so
the inequality is satised. Lastly, equation (13) is a sufcient
condition for stability, and it is satised for the values T = 1,
= 0.025 and q = 90. Therefore, the parametric conditions
of the theory guaranteeing feasibility and stability of the
distributed RHC algorithm are satised. Scaling time back
to a planning horizon of 6 seconds corresponds to an update
period of = 0.15 seconds, and this is the update period used
in the centralized and distributed RHC implementations.
Distributed RHC is implemented precisely according to
Algorithm 1, with one modication to Problem 1. In the
optimization code, the constants on the right-hand side of
constraints (9) and (10) are set to a = b = 0.1. The actual
constants in (9) and (10) are small enough ( 10
7
) to
cause feasibility problems in each distributed optimization
code. The value of 0.1, on the other hand, worked quite
well. Of course, the constants dened in constraints (9)
and (10) are derived based on the sufcient conditions of
the theory, and are likely to be conservative. The closed-
loop position and control trajectories generated by applying
the distributed RHC algorithm are shown in Figure 3. The
0 1 2 3 4 5 6
0.3
0.2
0.1
0
0.1
0.2
0.3
0.4
0.5
0.6
time (sec)
(
r
a
d
)
DRHC

1,D

2,D

3,D
0 1 2 3 4 5 6
1
0.5
0
0.5
1
time (sec)
(
r
a
d
/
s
e
c
2
)
DRHC
u
1,D
u
2,D
u
3,D
Fig. 3. The distributed RHC response, showing the angular position
trajectories
i,D
(left plot) and the control trajectories u
i,D
(right plot),
for each i = 1, 2, 3. The response is quite close to the centralized RHC
response shown in Figure 2.
position and control trajectories for this closed-loop solution
are denoted
i,D
and u
i,D
, respectively. While the algorithm
and theory suggest switching to the terminal controllers
once z(t)

, the distributed receding horizon controllers


are employed for all time in these results. To compute the
actual closed-loop response between RHC updates requires
numerical integration of the dynamic equations (see (5)).
Also, to calculate each z
i
, as required in part 2(b) of
Algorithm 1, requires numerical integration of equation (6).
In all cases, numerical integration was performed using the
ode23 function in MATLAB.
The centralized and distributed RHC responses are quite
close, with the distributed position responses showing
slightly more overshoot than the centralized counterparts,
particularly for angles
2
and
3
. The closeness in the two
responses can be attributed in part to the weak coupling in
the dynamics as quantied by the coefcient = 0.025. For
weakly coupled dynamics, the error introduced by relying
on z
j
for neighbors has less of an impact on the closed-loop
response, than for systems with dynamics that are strongly
inuence by neighboring responses. Application of the the-
ory to systems with stronger dynamic coupling would be
useful in identifying difference between centralized RHC and
the distributed RHC algorithm presented here. A hypothesis
worth testing is that, even in the stronger coupling case,
if the update period is sufciently small, the distributed
RHC response is likely to be close to the centralized RHC
response. The intuition behind this hypothesis is that the error
introduced by relying on z
j
for neighbors is likely smaller
for smaller update periods.
To compare the computational burden of the centralized
problem and the distributed problems, the cputime func-
tion is used in MATLAB. The centralized optimal control
problem has 81 variables to solve for at each RHC update.
The computational time for each RHC update, corresponding
to the response shown in Figure 2, is shown in the top
plot in Figure 4. Each distributed optimal control prob-
lem has 27 variables to solve for, where each problem is
solved in parallel. The computational time for each RHC
update per agent, corresponding to the responses shown in
Figure 3, is shown in the bottom plot in Figure 4 From
0 10 20 30 40
300
400
500
Computation times
C
R
H
C

(
s
e
c
)
0 10 20 30 40
6
8
10
12
RHC update number
D
R
H
C

(
s
e
c
)
1
2
3
Fig. 4. Comparison of computation times, at each receding horizon
update, to solve the centralized optimal control problem (top plot), and
the distributed optimal control problems in parallel (bottom plot). The
computation times correspond to the responses shown in Figures 2, 3.
the gure, the distributed optimal control problems were
solved between 43 and 58 times faster than the centralized
optimal control problem, over all updates. On average, each
distributed problem was solved 50 times faster, than the
single centralized problem. Clearly, for this example, there
is substantial savings in being able to solve the distributed
problems in parallel. The savings are also consistent with the
computational complexity comparison reported in [6]
VI. CONCLUSIONS
In this paper, a distributed implementation of RHC is
developed for the case of dynamically coupled nonlinear
systems subject to decoupled input constraints. A central ele-
ment to the feasibility and stability analysis is that the actual
and assumed responses of each agent are not too far from one
another, as quantied by a consistency constraint. Parametric
bounds on the receding horizon update period are identied.
Also, conditions that bound the amount of dynamic coupling,
parameterized by a Lipschitz constant, are also identied.
While the theoretical results are sufcient, the proposed
algorithm with minor relaxations is shown to be applicable
to the problem of distributed control of coupled nonlinear
oscillators. In the numerical results, the time it takes to solve
the distributed optimal control problems in parallel is two
orders of magnitude less than the time it takes to solve a
corresponding centralized optimal control problem, under-
lining the computational savings incurred by employing the
distributed algorithm. The closed-loop response generated by
the distributed algorithm is also quite close to a centralized
RHC implementation. While it makes sense to compare
centralized RHC with the distributed implementation for
the academic coupled oscillator example considered here,
centralized RHC is often not a viable option in applications
[9] (for example, in supply chain management) where the
distributed RHC algorithm may prove relevant.
REFERENCES
[1] L. Acar. Boundaries of the receding horizon control for interconnected
systems. Journal of Optimization Theory and Applications, 84(2),
1995.
[2] M. W. Braun, D. E. Rivera, W. M. Carlyle, and K. G. Kempf.
Application of model predictive control to robust management of
multiechelon demand networks in semiconductor manufacturing. Sim-
ulation, 79(3):139156, 2003.
[3] H. Chen and F. Allg ower. A quasi-innite horizon nonlinear model
predictive scheme with guaranteed stability. Automatica, 14(10):1205
1217, 1998.
[4] S. Chopra and P. Meindl. Supply Chain Management, Strategy,
Planning, and Operations - Second Edition. Prentice-Hall, 2004.
[5] W. B. Dunbar. A distributed receding horizon control algorithm for
dynamically coupled nonlinear systems. In Proc. of the IEEE Con-
ference on Decision and Control / IEE European Control Conference,
Seville, Spain, 2005.
[6] W. B. Dunbar. Distributed receding horizon control of dynami-
cally coupled nonlinear systems. Technical Report ucsc-crl-06-03,
Computer Engineering, University of California, Santa Cruz, 2006.
Accepted to IEEE Trans. Aut. Cntr., July, 2006.
[7] W. B. Dunbar and S. Desa. Distributed nonlinear model predictive
control for dynamic supply chain management. In Proceedings of
the International Workshop on Assessment and Future Directions of
NMPC, Freudenstadt-Lauterbad, Germany, August, 2005.
[8] M. S. Dutra, A. C. de Pina Filho, and V. F. Romano. Modeling of a
bipedal locomotor using coupled nonlinear oscillators of Van der Pol.
Biol. Cybern., 88:286292, 2003.
[9] Yu-Chi Ho. On centralized optimal control. IEEE Trans. Auto. Contr.,
50(4):537538, 2005.
[10] D. Jia and B. H. Krogh. Distributed model predictive control. In
Proceedings of the IEEE American Control Conference, 2001.
[11] D. Jia and B. H. Krogh. Min-max feedback model predictive control
for distributed control with communication. In Proceedings of the
IEEE American Control Conference, 2002.
[12] D. Q. Mayne, J. B. Rawlings, C. V. Rao, and P. O. M. Scokaert. Con-
trained model predictive control: Stability and optimality. Automatica,
36:789814, 2000.
[13] H. Michalska and D. Q. Mayne. Robust receeding horizon control
of contrained nonlinear systems. IEEE Trans. Auto. Contr., 38:1623
1632, 1993.
[14] M. B. Milam, K. Mushambi, and R. M. Murray. A new computational
approach to real-time trajectory generation for constrained mechanical
systems. In Proceedings of the Conference on Decision and Control,
2000.
[15] N. R. Sandell, P. Varaiya, M. Athans, and M. G. Safanov. Survey
of decentralized control methods of large scale systems. IEEE Trans.
Auto. Contr., 23(2):108128, 1978.
[16] A. N. Venkat, J. B. Rawlings, and S. J. Wright. Stability and
optimality of distributed model predictive control. In Proc. of the
IEEE Conference on Decision and Control / IEE European Control
Conference, Seville, Spain, 2005.
[17] A. N. Venkat, J. B. Rawlings, and S. J. Wright. Plant-wide optimal
control with decentralized MPC. In Proceedings of the IFAC Dynamics
and Control of Process Systems Conference, Boston, MA, July, 2004.
[18] G-Y. Zhu and M. A. Henson. Model predicitive control of intercon-
nected linear anad nonlinear processes. Ind. Eng. Chem. Res., 41:801
816, 2002.

You might also like