(2001) Optimal Control by Least Squares Support Vector Machines
(2001) Optimal Control by Least Squares Support Vector Machines
Networks
Abstract
Support vector machines have been very successful in pattern recognition and function estimation problems. In this paper we introduce the
use of least squares support vector machines (LS-SVM's) for the optimal control of nonlinear systems. Linear and neural full static state
feedback controllers are considered. The problem is formulated in such a way that it incorporates the N-stage optimal control problem as well
as a least squares support vector machine approach for mapping the state space into the action space. The solution is characterized by a set of
nonlinear equations. An alternative formulation as a constrained nonlinear optimization problem in less unknowns is given, together with a
method for imposing local stability in the LS-SVM control scheme. The results are discussed for support vector machines with radial basis
function kernel. Advantages of LS-SVM control are that no number of hidden units has to be determined for the controller and that no centers
have to be speci®ed for the Gaussian kernels when applying Mercer's condition. The curse of dimensionality is avoided in comparison with
de®ning a regular grid for the centers in classical radial basis function networks. This is at the expense of taking the trajectory of state
variables as additional unknowns in the optimization problem, while classical neural network approaches typically lead to parametric
optimization problems. In the SVM methodology the number of unknowns equals the number of training data, while in the primal space
the number of unknowns can be in®nite dimensional. The method is illustrated both on stabilization and tracking problems including
examples on swinging up an inverted pendulum with local stabilization at the endpoint and a tracking problem for a ball and beam system.
q 2001 Elsevier Science Ltd. All rights reserved.
Keywords: Neural optimal control; Support vector machines; Radial basis functions
parameters can be chosen such that a bound on the general- optimal trajectory that one seeks. Furthermore, standard
ization error is minimized, expressed in terms of the VC neural network approaches always work in a primal weight
dimension. One has the possibility to use polynomials, space, while in SVM methodologies the computations are
splines, radial basis function (RBF) networks or multilayer done in a dual space such that the number of unknowns
perceptrons as kernels. Although one preferably applies equals the number of training data points (and not the
Mercer's condition, this is not absolutely necessary as is number of weights in the primal space, which can be in®nite
shown e.g. in Suykens and Vandewalle (1999a), where dimensional).
standard multilayer perceptron classi®ers with a ®xed We illustrate the LS-SVM control method on a number
number of hidden units are trained using a SVM method- of simulation examples including swinging up an inverted
ology. Being based on the structural risk minimization prin- pendulum with local stabilization at the endpoint and a
ciple and capacity concept with pure combinatorial tracking problem for a ball and beam system. In Suykens
de®nitions, the quality and complexity of the SVM solution et al. (1994), methods for realizing a transition between
does not depend directly on the dimensionality of the input two states with local stabilization at the target point was
space. discussed. This was illustrated on swinging up an inverted
In the SVM control method we make use of a least pendulum with local stabilization at the endpoint. For the
squares version of support vector machines. Originally, full static state feedback case a Linear Quadratic Regula-
Vapnik's epsilon insensitive loss function has been tor (LQR) (Franklin, Powell, & Workman, 1990) was
employed for the function estimation problem. A least designed for balancing the pole in its upright position.
squares interpretation on the other hand has been given by This result was used in order to impose a set of
Saunders, Gammerman, and Vovk (1998) for function esti- constraints on the interconnection weights of a multilayer
mation problems, Suykens and Vandewalle (1999b) for perceptron controller. A drawback of the approach was
classi®cation, and MuÈller et al. (1997) for time±series that the number of hidden units for the neural controller
prediction. In this case, the problem formulation involves had to be chosen ad hoc. In this paper we present a solu-
equality instead of inequality constraints and the support tion by SVM control which takes into account an LQR
values are proportional to the errors at the data points, design in the LS-SVM control with RBF kernel. The
which simpli®es the problem. Sparseness gets lost when number of hidden units is determined here by the number
the function estimation corresponds to this form of ridge N in the N-stage optimal control problem formulation.
regression, but this can be imposed afterwards by doing Instead of local linear LQR design one has also the possi-
pruning based upon the support value spectrum (Suykens, bility to apply robust linear control methods (Boyd &
Lukas, & Vandewalle, 2000a,b). Barratt, 1991) such as H1 control and m theory instead,
In the optimal control method by LS-SVM's, the N-stage in order to take into account uncertainties (parametric
optimal control problem and the optimization problem uncertainties, unmodelled dynamics) and noise.
related to the LS-SVM controller are incorporated within For the example of a ball and beam system (Hauser,
one problem formulation. A model is assumed to be avail- Sastry, & Kokotovic, 1992) a tracking problem with SVM
able for the plant, the controller is designed based upon this control is discussed. Again a LQR design is incorporated in
model and afterwards applied to the plant, assuming that the order to impose local stability of the origin for the autono-
certainty equivalence principle holds (similar to Nguyen mous closed-loop system. Imposing robust local stability for
Widrow's emulator approach). The solution is characterized the closed-loop system has also been successful in the appli-
by a set of nonlinear equations. However, the set of cation of NLq neural control theory, introduced in Suykens,
nonlinear equations will typically contain a large number De Moor, and Vandewalle (1997) and Suykens et al. (1996),
of unknowns. Therefore an alternative formulation for LS- to the control of a real-life ball and beam system (Verrelst
SVM control is given in less unknowns. This formulation et al., 1998). Within NLq theory suf®cient conditions for
also enables to incorporate local stability constraints. A main global asymptotic stability and input/output stability with
difference between standard neural network approaches ®nite L2-gain are available, which can also be employed in
(Bishop (1995) MLP, RBF) and LS-SVM control is that in order to impose robust local stability of the origin for the
the former one solves a parametric optimization problem in closed-loop system. A similar approach is followed in the
the unknown interconnection weights while in the latter also LS-SVM control for the ball and beam system, where robust
the state vector sequence is part of the unknown parameter local stability can be realized based upon the work of Sezer
vector in the optimization problem. However, standard and Siljak (1998).
methodologies suffer from problems like the choice of the This paper is organized as follows. In Section 2 we
number of hidden units needed in order to accomplish a formulate the N-stage optimal control problem. In
given control task. More speci®cally, in the case of RBF Section 3 we review work on support vector machines. In
networks one has a curse of dimensionality when one de®nes Section 4 we discuss optimal control by least squares
a regular grid for the centers (hidden units) in state space, as support vector machines. In Section 5 we discuss an alter-
explained e.g. in the method of Sanner and Slotine (1992). In native formulation together with stability constraints. In
the LS-SVM control case, the centers will follow from the Section 6 we present examples.
J.A.K. Suykens et al. / Neural Networks 14 (2001) 23±35 25
2. The N-stage optimal control problem 3. The support vector method of function estimation
In the N-stage optimal control problem one aims at In this section we review basic ideas of the support vector
solving the following problem (Bryson & Ho, 1969): method of function estimation, that are essential for its intro-
duction within the neural control problem as will be consid-
X
N
ered in the sequel. For further details on SVM's we refer to
minJN xk ; uk r xN11 1 h x k ; uk 1
k1
Cherkassky and Mulier (1998), Haykin (1994), SchoÈlkopf et
al. (1999), Smola (1999), Smola, SchoÈlkopf, and Muller
subject to the system dynamics (1998), Vapnik (1995, 1998a,b), Vapnik, Golowich, and
Smola (1997), SchoÈlkopf et al. (1997) and Smola and
xk11 f xk ; uk ; k 1; ¼; N x1 given; SchoÈlkopf (1998).
Consider regression in the following set of functions
where r ´ and h ´; ´ are positive de®nite functions. A typi-
cal choice is the quadratic cost h xk ; uk xTk Qxk 1 F X WT w X 1 B 5
uTk Ruk ; r xN11 xTN11 QxN11 withQ QT . 0; R RT .
0: The functions r ´; h ´; ´; f ´; ´ are assumed to be twice with given training data {Xi ; Yi }M
i1 where M denotes the
continuously differentiable. xk [ Rn denotes the state number of training data, Xi [ Rm are the input data and
vector, uk [ R is the input of the system. The methods Yi [ R are the output data. The nonlinear mapping w :
discussed in this paper are not restricted to single input Rm ! Rnh maps the input data into a so-called high dimen-
systems. sional feature space (which can be in®nite dimensional) and
In order to ®nd the optimal control law, one constructs the W [ Rnh ; B [ R: In the support vector method one aims at
Lagrangian minimizing the empirical risk
1 XM
X
N
Remp W; B uY 2 WT w Xi 2 Bu1 6
LN xk ; uk ; lk JN xk ; uk 1 lTk xk11 2 f xk ; uk 2 M i1 i
k1
subject to elements of the structure Sn, de®ned by the
with Lagrange multipliers lk [ Rn : The conditions for inequality WT W # cn : The loss function employs
optimality are given by (Fletcher, 1987; Gill, Murray, & Vapnik's 1 -insensitive model:
Wright, 1981) (
0; if uY 2 F Xu # 1
8
> 2LN 2h
2f T
uY 2 F Xu1
>
>
>
> 2xk
2xk
1 lk21 2
2xk
lk 0; k 2; ¼; N adjoint equation uY 2 F Xu 2 1; otherwise:
>
>
>
>
>
>
>
2LN
2r
1 lN 0 adjoint final condition
7
< 2xN11 2xN11
>
> 2LN 2h 2f The estimation problem is formulated then as the optimiza-
>
> 2 lTk 0; k 1; ¼; N variational condition
>
>
>
>
2uk 2uk 2uk tion problem:
>
>
>
> 2LN (M )
:
2 lk
xk11 2 f xk ; uk 0; k 1; ¼; N system dynamics:
p 1 T
X p X M
min J 1 W; j ; j W W 1 g j i 1 j i 8
3 W;B;j p ;j 2 i1 i1
For the case of a quadratic cost function subject to linear subject to the constraints
system dynamics with in®nite time horizon, the optimal 8
>
> Yi 2 WT w Xi 2 B # 1 1 j ip ; i 1; ¼; M
control law can be represented by full static state feedback >
>
>
< 2Yi 1 WT w Xi 1 B # 1 1 j i ;
control. However, in general, the optimal control law cannot i 1; ¼; M
be represented by state feedback as the optimal control law > j p $ 0;
>
> i i 1; ¼; M
may also depend on the co-state. Nevertheless, one is often >
>
:
interested in ®nding a suboptimal control strategy of this j i $ 0; i 1; ¼; M
form. In the context of neural control, Saerens et al.
where j i ; j ip are slack variables
P and pg is a positive real
(1993) considered, in addition to Eq. (1),
constant. One obtains W M i1 ai 2 ai w Xi where
uk g x k 4 api ; ai are obtained by solving a quadratic program and are
the Lagrange multipliers related to the ®rst and second set of
where g ´ is parametrized by a neural network architecture, constraints. The data points corresponding to non-zero
and discussed the link with the backpropagation algorithm. values for api 2 ai are called support vectors. Typically,
In this paper, we will also consider a control law given in many of these values are equal to zero. Finally, one obtains
Eq. (4), by relating Eq. (4) to support vector machines. the following model in the dual space
Because SVM methodology is not a parametric modelling
X
M
approach, this is less straightforward in comparison with F X api 2 ai K Xi ; X 9
standard neural networks such as MLP's and RBF's. i1
26 J.A.K. Suykens et al. / Neural Networks 14 (2001) 23±35
applied to the plant is assumed to be wT w xk : In the The set of nonlinear Eq. (21) is of the form
linear control case one has w xk xk with nh n: In
F1 xk ; xN11 ; uk ; w; ek ; lk ; ak 0 23
the sequel we will employ RBF kernels. The support
vector method, as it has been originally introduced, for k 1; ¼; N with x1 given and can be numerically solved
exploits Mercer's condition such that one does not in the unknown variables. In Appendix A a formulation
have to construct w ´: Although the use of this kernel without ek variables is given. In comparison with function
function is preferred, one could also evaluate w ´ expli- estimation or classi®cation problems, one looses the advan-
citly, as it has been demonstrated in Suykens and tage of solving a quadratic program or a linear least squares
Vandewalle (1999a) for multilayer perceptron classi®ers. problem. Nevertheless, one is still able to exploit some of
For the RBF case, Eq. (18) becomes then the interesting SVM features.
As in SVM theory, Mercer's
Xnh
1 Pcondition can be applied with-
uk wi exp 2 2 xk 2 ci 22 1 ek 19 in Eq. (21), by replacing w k ak w xk in the equations. For
s
i1 kernels satisfying Mercer's condition one can impose
where ci [ Rn are chosen centers and s is a chosen
K xk ; xl w xk T w xl : 24
constant. One can take for example the set {c i }ni1 h
N
equal to {xk }k1 in order to avoid additional unknown For RBF kernels one takes (Vapnik, 1995)
parameters for the centers which is also motivated by
the work on regularization networks by Poggio and K xk ; xl exp 2hixk 2 xl i22 25
Girosi (1990). In standard SVM theory for static func- with h a positive real constant. One sees that Eq. (25) does not
tion estimation problems s can be selected so as to contain the centers ci as in Eq. (19). After this elimination of w,
minimize an upper bound on the generalization error. a set of nonlinear equations of the form
These bounds are not applicable in the context of
SVM control due to the fact that the input patterns to F2 xk ; xN11 ; uk ; lk ; ak 0 26
the activation function are not independent from each
for k 1; ¼; N with x1 given is obtained. More speci®cally,
other. Hence s should be chosen as hoc (typically avoid
after exploiting Mercer's condition, one has
values which are too small) or could be taken as an
8
additional unknown within the cost function J. >
> 2h 2f T XN
2K xk ; xl
>
> 1 lk21 2 lk 2 ak al k 2; ¼; N
In order to ®nd the optimal control law we construct the >
>
> 2x k 2x k l1
2xk
>
>
Lagrangian >
> 2 r
>
> 1 lN 0
>
> 2x
1 T 1 XN >
< N11
L x k ; uk ; w; ek ; lk ; ak JN xk ; uk 1 w w1g e2 2h 2f : 27
2 2 k1 k >
>
> 2uk
2 lTk
2uk
1 ak 0; k 1; ¼; N
>
>
>
>
>
> xk11 2 f xk ; uk 0; k 1; ¼; N
X
N X
N >
>
>
1 lTk xk11 2 f xk ; uk 1 T
ak uk 2 w w xk 2 ek : >
>
>
XN
>
: uk 2 al K xl ; xk 2 ak =g 0; k 1; ¼; N
k1 k1
l1
20
For RBF kernels one has
The conditions for optimality are given by
8 2K xk ; xl
22h xk 2 xl exp 2hixk 2 xl i22 :
>
>
>
2L
2h
1 lk21 2
2f T
lk 2 ak
2
wT w xk 0; k 2; ¼; N adjoint equation 28
>
>
>
>
>
2xk 2xk 2xk 2xk 2xk
>
> 2L 2r
>
> 1 lN 0; adjoint final condition
>
>
>
>
>
>
2xN11 2xN11 The actual control signal applied to the plant becomes
>
> 2L 2h 2f
>
>
> 2 lTk 1 ak 0; k 1; ¼; N variational condition
>
>
>
>
<
2uk 2uk 2uk
X
N
2L XN
a k w xk 0 uk al K x l ; x k 29
> 2w w 2
>
support vectors
>
> k1 l1
>
> 2L
>
>
>
> gek 2 ak 0 k 1; ¼; N support values
>
>
>
>
>
2ek
where {xl }Nl1 ; {al }Nl1 are obtained from solving the set of
>
> 2L
>
>
>
>
> 2lk
xk11 2 f xk ; uk 0; k 1; ¼; N system dynamics nonlinear Eq. (27) and xk is the actual state vector at time k. The
>
>
>
>
>
:
2L
uk 2 wT w xk 2 ek 0; k 1; ¼; N SVM control:
data {xl }Nl1 are used as support vector data for the control
2ak
signal. Furthermore, note that ek has been taken equal to zero
21 in Eq. (29).
For RBF kernels one has
X
2 T 1 2 5. Alternative formulation and local stability
w w xk 22 wi exp 2 2 ixk 2 ci i2
2xk s
i 22
In this section we give an alternative formulation to the
2
£ xk 2 ci =s : LS-SVM control approach in less unknowns, together with a
28 J.A.K. Suykens et al. / Neural Networks 14 (2001) 23±35
method for imposing local stability at the origin for the be formulated as:
autonomous closed-loop system.
X
N
Based upon Eqs. (21) and (27) we consider the optimiza- min JN xk ; xeq ; uk 1 l a2k 32
tion problem xk ;ak ;Q
k1
X
N subject to
min JN xk ; xrk ; uk 1l a2k 30 8 !
xk ; a k >
> XN
eq eq
k1 < xk11 f xk ; al K xl ; x ; xk ; x ; x1 given
l1
>
>
subject to : T
A PA 2 P , 0
8
>
> x f x k ; uk x1 given; where P QT Q; A 2f =2x^k ux^k xeq and Z , 0 denotes a
< k11
X
N negative de®nite symmetric matrix.
> al K xl ; xrl ; xk ; xrk :
: uk
> Imposing robust local stability for the closed-loop system
l1 has been successfully applied within the context of NLq neural
control theory (Suykens et al., 1997, 1996). Within NLq theory
In this formulation a reference state vector xrk is considered suf®cient conditions for global asymptotic stability and input/
and g ! 1 is taken P with respect to Eq. (27). A regulariza- output stability with ®nite L2-gain are available, which can also
tion by the term k a2k is included (see also Smola (1999)) be employed in order to impose robust local stability of the
where l is a positive real constant. Instead of solving a set origin for the closed-loop system. Based upon NLq stability
of nonlinear equations one solves then a constrained criteria, dynamic backpropagation can be modi®ed by impos-
nonlinear optimization problem with less unknowns than ing matrix inequality constraints as in Eq. (32). For Eq. (32)
in the approaches of the previous section. From the formu- also robust local stability can be imposed, e.g. based upon
lation (30) the difference between standard neural network (Sezer & Siljak, 1998).
controllers and LS-SVM control is clear. In the former one
solves a parametric optimization problem in the unknown
interconnection weights while in the latter also the state 6. Simulation examples
vector sequence is part of the unknown parameter vector
in the optimization problem. Furthermore, standard neural 6.1. Example 1
network approaches always work in the primal weight
space, while in SVM methodology the computations are Here we illustrate the LS-SVM optimal control method of
done in the dual space such that the number of unknowns Section 4 on an example reported in Narendra and Mukho-
equals the number of training data points (and not the padhyay (1997). Given the nonlinear system
8
number of weights in the primal space, which can be in®nite > uk 1 x2;k
>
> x1;k11 0:1x1;k 1 2
dimensional). >
< 1 uk 1 x2;k 2
For RBF kernels the parameter h in Eq. (25) can be taken ! 33
as an additional unknown to the cost function (30). When >
> u2k
>
> x2;k11 0:1x2;k 1 uk 2 1
applying an optimization method the constraints in Eq. (30) : 1 1 x21;k 1 x22;k
will only hold with a certain tolerance. These small numer-
ical errors may lead to differences between the state vectors we consider a state vector tracking problem with
as solution to Eq. (30) and the simulation of the closed-loop h xk ; uk xk 2 xrk T Q xk 2 xrk 1 uTk Ruk ; r xN11
system with LS-SVM control, especially in the control of xN11 2 xrN11 T Q xN11 2 xrN11 where xrk is the reference
unstable systems. For control of stable systems this problem trajectory to be tracked. We aim at tracking the ®rst
will be less critical. A simulation of the closed-loop system state variable and choose Q diag{1; 0:001}; R 1
with LS-SVM control for xrk sin 2pk=20; cos 2pk=20 with k 1; N and
N 20: The given initial state is x1 0; 0: As control
! law is taken wT w xk ; xrk ; which results only into a
X
N
x^k11 f x^k ; al K xl ; xrl ; x^k ; xrk x^1 x1 given slight modi®cation of Eqs. (21) and (27). In Fig. 1,
l1 simulation results are shown for the method (27) that
31 makes use of Mercer's condition. The set of nonlinear
equations was solved using Matlab's optimization tool-
is always needed in order to validate the results, where box (function leastsq) with unknowns xk, uk, l k, a k, s
{xl }Nl1 and {al }Nl1 are obtained as a solution to Eq. (30). (variables ek eliminated) for g 100: The unknowns
In order to impose local stability at a ®xed equilibrium were randomly initialized with zero mean and standard
point x eq, one locally linearizes the autonomous closed-loop deviation 0.3. The plots show the simulation results for
P
simulation model (31) by evaluating 2f =2x^k at x eq. A LS- the closed-loop system x^ k11 f x^k ; Nl1 al K xl ; x^k with
SVM control design with local stability constraint can then RBF kernel. The controller is generalizing well with
J.A.K. Suykens et al. / Neural Networks 14 (2001) 23±35 29
Fig. 1. (Top) Optimal control by a least squares support vector machine with RBF kernel and application of Mercer's condition. (Full line) ®rst state variable to
be tracked; (Dashed line) actual state variable by closed-loop simulation of the system. The data k 1; ¼; 20 were used for training of the controller with the
origin as initial state. (Bottom) simulation result for another randomly chosen initial state x1. The LS-SVM controller shows a good generalization performance
with respect to other initial states and beyond the time horizon.
The state variables x1, x2, x3, x4 are respectively position and with
velocity of the cart, angle of the pole with the vertical and
rate of change of the angle. The input signal u is the force X
N
where uk is assumed to be constant in the time intervals [kh, Here we discuss the LS-SVM control method of Section 5
(k 1 1)h] (zero order hold). The control law is taken as on a ball and beam system, which has been described in
Hauser et al. (1992).
uk Llqr 2 LD xk 1 usvm
k 38 The continuous time system description of the ball and
J.A.K. Suykens et al. / Neural Networks 14 (2001) 23±35 31
Fig. 2. Swinging up an inverted pendulum by a LS-SVM controller with local stabilization in its upright position. Around the target point the controller is
behaving as a LQR controller. (Top) inverted pendulum system; (Bottom) simulation result which visualizes the several pole positions in time.
beam system (Fig. 5) is given by Eq. (34) with objective we consider tracking of the ball position for a refer-
2 3 2 3 ence input d t 2 cos pt=5:
x2 0 As for the inverted pendulum example, a LQR design is
6 7 6 7
6 B x x2 2 G sin x 7 607 incorporated in the LS-SVM control law. Here we impose
6 1 4 3 7 6 7
F x 6
6
7;
7 G x 66 7
7 43 local stability for the autonomous closed-loop system,
6 x 7 607 which is motivated by the application of NLq neural control
4 4 5 4 5
0 1 theory (Suykens et al., 1997, 1996), to the control of a real-
life ball and beam system (Verrelst et al., 1998). We proceed
where x x1 ; x2 ; x3 ; x4 r; r_; u; u_ with r the ball position in a similar way as for the inverted pendulum system. The
and u the beam angle and B M= Jb =R2b 1 M where M, Jb, continuous time state space description (43) is linearized
Rb are the mass, moment of inertia and radius of the ball, around the origin, for which a LQR controller is designed
respectively. For the control input one has t 2Mrr_u_ 1 with Q I; R 1 in the LQR cost function. Then the
MGr cos u 1 Mr2 1 J 1 Jb u where t; G; J denote the continuous time model is discretized by using a fourth
torque applied to the beam, the acceleration of gravity and order Runge±Kutta integration rule with constant step h
the moment of inertia of the beam, respectively. As control 0:3 into a discrete time model similar to Eq. (37). The ®rst
32 J.A.K. Suykens et al. / Neural Networks 14 (2001) 23±35
Fig. 3. (Continued) (Top) state variables with respect to time of the closed-loop simulation model with LS-SVM controller: x1(t) (±); x2(t) (± ±); x3 (±.); x4(t)
(:). The state vector data before the vertical line were used in the training process as support vector data with an RBF kernel; (Bottom) control signal uk.
component of the reference state vector xr1;k is derived from the autonomous closed-loop system) and
d(t) according to this step size. The control law is taken as
" # 2usvm
k
xk LD : 46
uk Llqr 2 LD r 1 usvm
k 44 2xk xk 0
x1;k
Note that Eq. (46) also depends on the reference xr1;k : A zero
with
" # " #! initial state has been taken. SQP was applied for constrained
X
N xl xk nonlinear optimization of Eq. (30) with a similarly chosen
usvm
k al K ; 45 initial unknown parameter vector as for the inverted pendu-
l1 xr1;l xr1;k
lum example. Simulation results of the closed-loop simula-
where Llqr is the resulting feedback matrix from LQR (for tion model are shown in Fig. 5.
J.A.K. Suykens et al. / Neural Networks 14 (2001) 23±35 33
Acknowledgements
Fig. 5. Tracking control of a ball and beam system by a LS-SVM controller: (Top) ball and beam system; (Bottom) reference input (±) and position of the ball
(± ±). The state vector data before the vertical line were used in the training process as support vector data with an RBF kernel. Shown is the closed-loop
simulation result.
with the National Fund for Scienti®c Research FWO Ð The conditions for optimality are given by
Flanders. 8
>
> 2L0 2h 2f T 2f 2uk T
>
> 1 lk21 2 lk 2 lk 0; k 2; ¼; N
>
> 2xk 2xk 2xk 2uk 2xk
>
>
>
> 2L0 2r
Appendix A. Formulation without error variables >
> 1 lN 0
>
< 2xN11 2xN11
>
> 2L0 X
N
2f
A variation on the problem Eqs. (17) and (18) without the >
>
> w2 lTk w xk 0
>
> 2w 2uk
use of the variables ek is >
>
>
k1
>
> 2L0
>
: xk11 2 f xk ; wT w xk 0; k 1; ¼; N:
minJ0 xk ; w JN xk ; uk w; xk 1 1
2 wT w A1 2l k
A3
subject to
( Note that the support values are directly related here to the
xk11 f xk ; uk w; xk ; k 1; ¼; N x1 given Lagrange multipliers lk : The set of nonlinear Eq. (A3) is of
uk w T w xk k 1; ¼; N: the form
X
N Mercer condition can be applied as to Eqs. (A3) and (A4),
L0 xk ; w; lk JN xk ; uk w; xk 1 1
2 wT w 1 lTk xk11 yielding a set of equations of the form
k1
F4 xk ; xN11 ; lk 0 A5
2 f xk ; wT w xk :
A2 by elimination of w.
J.A.K. Suykens et al. / Neural Networks 14 (2001) 23±35 35
References SchoÈlkopf, B., Sung, K. -K., Burges, C., Girosi, F., Niyogi, P., Poggio, T., &
Vapnik, V. (1997). Comparing support vector machines with Gaussian
Bishop, C. M. (1995). Neural networks for pattern recognition, Oxford: kernels to radial basis function classi®ers. IEEE Transactions on Signal
Oxford University Press. Processing, 45 (11), 2758±2765.
Boyd, S., & Barratt, C. (1991). Linear controller design, limits of perfor- SchoÈlkopf, B., Burges, C. J. C., & Smola, A. J. (1999). Advances in kernel
mance, Englewood Cliffs, NJ: Prentice-Hall. methods Ð Support vector learning, Cambridge, MA: MIT Press.
Bryson, A. E., & Ho, Y. C. (1969). Applied optimal control, Waltham, MA: Sezer, M. E., & Siljak, D. D. (1988). Robust stability of discrete systems.
Blaisdel. International Journal of Control, 48 (5), 2055±2063.
Cherkassky, V., & Mulier, F. (1998). Learning from data: concepts, theory Smola, A. (1999). Learning with Kernels. PhD thesis, published by Birlin-
and methods, New York: Wiley. ghoven: GMD.
Fletcher, R. (1987). Practical methods of optimization, New York: Wiley. Smola, A., & SchoÈlkopf, B. (1998). On a kernel-based method for pattern
Franklin, G. F., Powell, J. D., & Workman, M. L. (1990). Digital control of recognition, regression, approximation and operator inversion. Algor-
dynamic systems, Reading, MA: Addison-Wesley. ithmica, 22, 211±231.
Gill, P. E., Murray, W., & Wright, M. H. (1981). Practical optimization, Smola, A., SchoÈlkopf, B., & MuÈller, K. -R. (1998). The connection between
London: Academic Press. regularization operators and support vector kernels. Neural Networks,
Golub, G. H., & Van Loan, C. F. (1989). Matrix computations, Baltimore, 11 (4), 637±649.
MD: Johns Hopkins University Press. Suykens, J. A. K., & Vandewalle, J. (1999a). Training multilayer percep-
Hauser, J., Sastry, S., & Kokotovic, P. (1992). Nonlinear control via tron classi®ers based on a modi®ed support vector method. IEEE Trans-
approximate input±output linearization: the ball and beam example. actions on Neural Networks, 10 (4), 907±912.
IEEE Transactions on Automatic Control, 37 (3), 392±398. Suykens, J. A. K., & Vandewalle, J. (1999b). Least squares support vector
Haykin, S. (1994). Neural Networks: a Comprehensive Foundation, (2nd machine classi®ers. Neural Processing Letters, 9 (3), 293±300.
ed.). Englewood Cliffs: Macmillan. Suykens, J. A. K., De Moor, B., & Vandewalle, J. (1994). Static and
Ljung, L. (1987). System Identi®cation: Theory for the User, New York: dynamic stabilizing neural controllers, applicable to transition between
Prentice-Hall. equilibrium points. Neural Networks, 7 (5), 819±831.
MuÈller, K. R., Smola, A. J., RaÈtsch, G., SchoÈlkopf, B., Kohlmorgen, J., & Suykens, J. A. K., De Moor, B., & Vandewalle, J. (1997). NLq theory: a
Vapnik, V. (1997). Predicting time series with support vector machines. neural control framework with global asymptotic stability criteria.
In W. Gerstner, A. Germond, M. Hasler & J. -D. Nicoud, Proceedings Neural Networks, 10 (4), 615±637.
of the International Conference on Arti®cial Neural Networks Suykens, J. A. K., Lukas, L., & Vandewalle, J. (2000). Sparse approxima-
ICANN'97LNCS 1327 (pp. 999±1004). Berlin: Spinger. tion using least squares support vector machines. IEEE International
Narendra, K. S., & Mukhopadhyay, S. (1997). Adaptive control using Symposium on Circuits and Systems ISCAS 2000, Geneva, Switzerland,
neural networks and approximate models. IEEE Transactions on Neural May 28±31 (pp. II-757±760).
Networks, 8 (3), 475±485. Suykens, J. A. K., Lukas, L., & Vandewalle, J. (2000). Sparse Least
Narendra, K. S., & Parthasarathy, K. (1991). Gradient methods for the Squares Support Vector Machine Classi®ers. 8th European Sympo-
optimization of dynamical systems containing neural networks. IEEE sium on Arti®cial Neural Networks ESANN 2000, Bruges Belgium,
Transactions on Neural Networks, 2 (2), 252±262. (pp. 37±42).
Nguyen, D., & Widrow, B. (1990). Neural networks for self-learning Suykens, J. A. K., Vandewalle, J., & De Moor, B. (1996). Arti®cial neural
control systems. IEEE Control Systems Magazine, 10 (3), 18±23. networks for modelling and control of non-linear systems, Boston:
Parisini, T., & Zoppoli, R. (1994). Neural networks for feedback feedfor- Kluwer Academic.
ward nonlinear control systems. IEEE Transactions on Neural Vapnik, V. (1995). The nature of statistical learning theory, New York:
Networks, 5 (3), 436±449. Springer.
Plumer, E. S. (1996). Optimal control of terminal processes using neural Vapnik, K. (1998a). Statistical learning theory, New York: Wiley.
networks. IEEE Transactions on Neural Networks, 7 (2), 408±418. Vapnik, V. (1998b). The support vector method of function estimation. In J.
Poggio, T., & Girosi, F. (1990). Networks for approximation and learning. A. K. Suykens & J. Vandewalle, Nonlinear modeling: advanced black-
Proceedings of the IEEE, 78 (9), 1481±1497. box techniques (pp. 55±85). Boston: Kluwer Academic.
Saerens, M., Renders, J. -M., & Bersini, H. (1993). Neural controllers based on Vapnik, V., Golowich, S., & Smola, A. (1997). Support vector method for
backpropagation algorithm. In M. M. Gupta & N. K. Sinha, IEEE Press function approximation, regression estimation and signal processing.
book on intelligent control: theory and practic New York: IEEE Press. Advances in Neural Information Processing Systems, Vol. 9.
Sanner, R. M., & Slotine, J. -J. E. (1992). Gaussian networks for direct Cambridge, MA: MIT Press.
adaptive control. IEEE Transactions on Neural Networks, 3 (6), 837± Verrelst, H., Van Acker, K., Suykens, J., Motmans, B., De Moor, B., &
863. Vandewalle, J. (1998). Application of NLq neural control theory to a
Saunders, C., Gammerman, A., & Vovk, V. (1998). Ridge regression learning ball and beam system. European Journal of Control, 4, 148±157.
algorithm in dual variables. Proceedings of the 15th International Confer- Werbos, P. (1990). Backpropagation through time: what it does and how to
ence on Machine Learning ICML-98, Madison, WI (pp. 515±521). do it. Proceedings of the IEEE, 78 (10), 1150±1560.