Iterative Learning Control
Iterative Learning Control
Chapter 9
Iterative Learning Control
Kevin L. Moore
9.1 Introduction
Problems in control system design may be broadly classied into two categories: stabilization and performance. In the latter category a typical problem is to force the output response of a (dynamical) system to follow a
desired trajectory as close as possible, where \close" is typically dened
relative to a specic norm or some other measure of optimality. Although
control theory provides numerous tools for attacking such problems, it is
not always possible to achieve a desired set of performance design requirements. This may be due to the presence of unmodelled dynamics or parametric uncertainties that are exhibited during actual system operation or
to the lack of suitable design techniques for a particular class of systems
(e.g., there is not a comprehensive theory of linear quadratic optimal design
technique for nonlinear systems).
Iterative learning control is a relatively new addition to the control engineer's toolkit that, for a particular class of problems, can be used to
overcome some of the traditional diculties associated with performance
design of control systems. Specically, iterative learning control, or ILC, is
a technique for improving the transient response and tracking performance
of processes, machines, equipment, or systems that execute the same tra-
426
K. L. Moore
jectory, motion, or operation over and over. The classic example of such a
process is a robotic manipulator performing spot welding in a manufacturing assembly line. For instance, such a manipulator might be programmed
to wait in its home position until a door panel is moved into place. It then
carries out a series of welds at pre-dened locations, after which it returns
to its home position until the door panel is removed. The entire process is
then repeated. Although robotic operations and manufacturing present obvious examples of situations in which a machine or process must execute a
given trajectory over and over, there are numerous other problems that can
be viewed from the framework of repetitive operations. In these situations,
iterative learning control can be used to improve the system response. The
approach is motivated by the observation that if the system controller is
xed and if the system's operating conditions are the same each time it
executes, then any errors in the output response will be repeated during
each operation. These errors can be recorded during system operation and
can then be used to compute modications to the input signal that will be
applied to the system during the next operation, or trial, of the system.
In iterative learning control renements are made to the input signal after
each trial until the desired performance level is reached. Research in the
eld of iterative learning control focuses on the algorithms that are used to
update the input signal. Note that in describing the technique of iterative
learning control we use the word iterative because of the recursive nature
of the system operation and we use the word learning because of the renement of the input signal based on past performance in executing a task
or trajectory.
In this paper we give an overview of the eld of iterative learning control.
We begin with a detailed description of the ILC technique, followed by two
illustrative examples that give a avor of the nature of ILC algorithms and
their performance. This is followed by a topical classication of some of
the literature of ILC and a discussion of the connection between ILC and
other common control paradigms, including conventional feedback control,
optimal control, adaptive control, and intelligent control. Next, we give a
summary of the major algorithms, results, and applications of ILC given in
the literature. This discussion also considers some emerging research topics
in ILC. As an example of some of the new directions in ILC theory, we
present some of our recent results that show how ILC can be used to forced
a desired periodic motion in an initially non-repetitive process: a gas-metal
arc welding system. The paper concludes with summary comments on the
past, present, and future of ILC.
427
uk
System
uk+1
Learning
Controller
- yk
Memory Memory
Memory
yd
428
K. L. Moore
In this context, ILC is an iterative technique for nding u (t) for the case
in which all the signals are assumed to be dened on the nite interval
0 tf ]. The ILC approach is to generate a sequence of inputs, uk (t) in such
a way that the sequence converges to u . That is, we seek a sequence
uk+1 (t) = fL(uk (t0 ) yk (t0 ) yd(t0 ) t)
= fL(uk (t0 ) fS (uk (t0 )) yd (t0 ) t) t0 2 0 tf ]
such that
lim u (t) = u (t) for all t 2 0 tf ]:
k!1 k
Some remarks about this problem include:
1. In a successful ILC algorithm the next input will be computed so
that the performance error will be reduced on the next trial. The
issue is usually quantied by saying that the error should converge,
with convergence measured in the sense of some norm.
2. It is worth repeating that we have dened our signals with two variables, k and t. The trial is indexed with the integer k, while time is
described by the variable t, which may be continuous or discrete.
3. The general algorithm shown introduces a new variable t0 . This reects the fact that after the trial is complete there is eectively no
causality restriction in the ILC operator fL. Thus one may use information about what happened after the input uk (t0 ) was applied when
constructing the input uk+1 (t0 ). The only place this is not possible
is at t = tf . Although we can assume t0 2 0 tf ], realistically we only
need t0 2 t tf ] when computing uk+1 (t).
4. The previous remark is emphasized in Figure 9.2, which illustrates
the distinction between conventional feedback and iterative learning
control. In Figure 9.2(a) it is shown how the ILC approach preserves
information about the eect of the input at each instant during the
iteration and uses that information to compute corrections to the
control signal during the next trial. Figure 9.2(b) shows that in a
conventional feedback strategy the error from the current time step
is used by the controller to compute the input for the next time step.
However, the eect of this decision is not preserved from one trial to
the next.
5. It is usually assumed implicitly that the initial conditions of the system are reset at the beginning of each trial (iteration) to the same
value. This has always been a key assumption in the formulation of
the ILC problem.
6. It is also usually assumed by denition that the trial length tf is xed.
Note, however, that it is possible to allow tf ! 1. This is often done
for analysis purposes.
Trial (k-1)
Error
...
Trial k
...
t-1
t+1
t-1
t+1
...
Trial (k+1)
...
t-1
t+1
t-1
t+1
429
...
...
t-1
t+1
t-1
t+1
t-1
t+1
t-1
t+1
Input
...
...
...
...
...
...
(a)
Error
...
...
t-1
t+1
t-1
t+1
...
...
t-1
t+1
t-1
t+1
...
...
Input
...
...
...
...
...
...
(b)
430
K. L. Moore
431
4.5
3.5
2.5
1.5
0
10
15
20
25
30
35
432
K. L. Moore
4
2
0
0
10
15
20
25
30
35
20
25
30
35
20
25
30
35
20
25
30
35
(a)
4
2
0
0
10
15
(b)
4
2
0
0
10
15
(c)
10
5
0
0
10
15
(d)
FIGURE 9.4. ILC algorithm behavior: (a) desired output and initial
output (b) desired output and output on the 5th trial (c) desired
output and output on 10th trial (d) input signal on 10th trial.
C (x) =
433
v
s v
g
? m1 H I2
; HHH
;
l
; 1 l2 HH m2
;I
l1= l = 0:3m
1
m1 = 3:0kg
m2 = 1:5kg
FIGURE 9.5. Two-joint manipulator.
To describe the learning controller for this system, rst dene the vectors
yk = (xTk x_ Tk xTk )T
yd = (xTd x_ Td xTd )T
to represent the complete system output and desired trajectory, respectively at the k-th trial. The learning controller is then dened by
434
K. L. Moore
The eld of iterative learning control has a relatively small, but steadily
growing literature. Table 9.1 gives a topical classication of general results in the eld and Table 9.2 lists references that are specic to robotics
and other applications. The ordering of the references in these two tables is
roughly chronological and references may be listed more than once, depending on their coverage of topics. Also, the classication of a given paper into
a particular category(ies) reects this author's impression of the paper and
is not necessarily the only possibility. Many of the references listed were
obtained from the Engineering Index and the INSPEC electronic databases
using a search strategy dened by: \control" AND \learning" AND \iterative." This is, of course, a very restrictive search strategy and it is quite
likely that we have missed some papers. Unfortunately, the large number of
conferences and journals available today make it impossible to be aware of
every contribution on the subject of iterative learning control, which is discussed in a large number of elds, from robotics, to articial intelligence, to
classical control, to neural networks. Further, the phrase \iterative learning
435
General/tutorial
Linear systems
Adaptive/identication
Discrete-time
Direct learning
Frequency-domain
General
Multivariable
Non-minimum phase
Norm-optimal
Periodic paramters
Time-delay systems
Time-invariant
Time-varying
Two-dimensional systems
Repetitive control
General
Linear
Discrete
Nonlinear/stochastic
Convergence
Linear
High-gain/nonlinear
Robustness
Linear
Nonlinear
Initial conditions
No-reset
Inversion
Papers
160, 169, 140, 162]
96, 177, 178, 23, 145]
183, 185, 160, 191, 143]
162, 134, 72, 70]
222, 241, 202, 5, 10]
23, 190]
112, 113, 239]
244, 115, 243, 246]
87, 146, 108]
98, 153, 66, 1, 136]
103, 144, 68, 237]
229, 222, 157, 158]
83, 26, 142]
73, 184]
4, 196]
8, 7, 10]
174, 101]
217, 90, 241, 91, 181]
15, 14, 12, 17, 21]
167, 166, 163, 152, 25]
150, 179, 182]
15, 11, 89, 91, 121]
75, 126, 139, 6, 178]
9, 3, 195]
140, 145]
156, 80, 81, 223, 211, 228]
128, 98, 97, 209, 3, 203]
223, 43, 27]
151, 211, 122]
21, 88, 99]
91, 210, 130]
111, 21, 24, 168]
176, 141, 149, 148]
202, 210, 58, 61]
85, 86, 41, 247]
33, 218, 32, 31]
240, 46, 247]
200, 219]
129, 40, 39, 34]
208, 170]
242, 245, 177, 135, 41]
39, 58, 36, 170, 42]
21, 217, 64, 100, 36]
34, 100, 42]
116, 165, 234, 230]
30, 224, 51]
101, 199, 141, 44, 45]
84, 215, 154, 161, 141]
201, 104, 135, 48, 106, 238, 28]
76, 180, 49, 123, 52]
235, 221, 236, 187]
436
K. L. Moore
5
0
-5
0
5
(a)
10
5
(b)
10
5
(c)
10
5
(d)
10
5
0
-5
0
5
0
-5
0
5
0
0
FIGURE 9.6. System response for joint angle 1 : (a) desired output and
initial output (b) desired output and output after 2nd trial (c) desired
output and output after 4th trial (d) desired output and output after
8th trial.
control" has only recently become the standard phrase used to describe the
ILC approach and many of the early papers dealing with ILC do not have
all three of these terms in their title. It is inevitable that some work will
be inadvertently omitted. Thus, the citation list has been limited to those
works with which this author is familiar. Note also that Tables 9.1 and 9.2
list only those papers dealing with the ILC approach as we have dened it.
As we discuss below, learning methods in control cover a broader spectrum
than ILC. References dealing with other learning methods and concepts
are not included.
At the risk of oending those who are left out and at the risk of appearing
to lose impartiality, it is also possible to discuss the literature by author.
Such a classication is useful, however, as it gives an idea of the level of
interest in the eld. The concept of iterative learning control in the sense
of Figure 9.1 was apparently rst introduced by Uchiyama in 1978 229].
Because this was a Japanese language publication it was not widely known
437
Elastic joints
Flexible links
Cartesian coordinates
Neural networks
Cooperating manipulators
Hybrid force/position control
Nonholonomic
Applications
Vehicles
Chemical processing
Machining/manufacturing
Mechanical systems
Miscellaneous
Nuclear reactor
Robotics demonstrations
Papers
15, 13, 16, 55, 235, 11]
109, 222, 157, 158, 19, 54]
111, 112, 77, 82, 113, 164]
56, 24, 174, 155, 168, 205]
22, 95, 248, 249, 204, 146]
96, 67, 102, 242, 207]
125, 117, 194, 124, 118, 18]
114, 59]
212, 213, 148, 57, 147]
225, 78]
230, 30, 224]
254, 216]
110, 105, 149]
175]
232, 231, 138]
251, 29, 50]
132, 37, 133]
137, 119, 47, 250, 79, 120]
228, 131, 187, 188]
198, 66, 65, 68]
179, 253, 159, 189]
127, 186, 93, 74]
92, 35, 62, 38]
192, 63, 197]
107]
110, 109, 19, 205, 248, 204]
146, 105, 96, 207, 59, 175]
78, 194, 30]
in the West until the idea was developed by the Japanese research group
of Arimoto, Kawamura, and Miyazaki, particularly through the middle to
late 1980's 15, 13, 16, 14, 110, 11, 12, 17, 109, 111, 114, 112, 113, 85],
86, 153, 18, 115]. This author was involved in several new ILC results in the
late 1980's to early 1990's 164, 165, 160, 167, 166, 168] 163, 161, 234, 169,
170], including the book Iterative Learning Control for Deterministic Systems 162]. Also during this period, a research group at the Dipartimento di
Infomatica, Sistemistica in Genoa, Italy, including Bondi, Lucibello, Ulivi,
Oriolo, Panzieeri and others, was quite active 24, 146, 59, 175, 147, 149,
150, 148]. Other early researchers included: Mita, et al. 156, 157, 158],
Craig, et al. 55, 54, 56], and Hideg, et al. 87, 108, 88, 89, 90, 91]. Also
of note is the work of Tomizuka, et al. in the area of repetitive control
223, 43, 228, 105, 98, 47, 97, 209]. Other active researchers in ILC include
Horowitz, et al. 205, 154, 95, 96], Sadegh, et al. 205, 204, 203, 78], Saab
201, 199, 202, 200], and C.H. Choi, et al. 250, 104, 48, 49]. A group at
KAIST, centered around Professor Bien, one of the pioneers in ILC research, has made numerous contributions 174, 21, 22, 101, 129, 130, 181].
Another leading research group is based around Professor Longman at
438
K. L. Moore
Columbia University 155, 184, 183, 185, 26, 141, 143, 134, 67, 102, 198],
68, 142, 99, 136, 237, 103, 144, 66, 65, 140, 145] and Professor Phan of
Princeton 192, 190, 189, 187, 188, 186, 191, 70]. Recently the work of
Amann, Owens, and Rogers has produced a number of new ideas 8, 9, 7,
177, 178, 5, 10, 4, 176]. Finally, we mention a very active research group in
Singapore centered around Jian-Xin Xu and Yangquan Chen 242, 245, 241,
41, 40, 239, 39, 240], 83, 36, 37, 38, 35, 62, 63, 42, 138, 219, 244, 243, 246],
33, 34, 218, 32, 31]. It is interesting to note that at the 2nd Asian Control
Conference, held in Seoul in July 1997, over thirty papers were devoted
to iterative learning control. This was nearly ve percent of all the papers
presented. Thus we see that ILC has grown from a single idea to a very
active area of research.
R-+ E- C
-
U- P
Y-
439
Given P and R, nd C so that the closed-loop has a prescribed set of poles, frequency characteristics, or steadystate error properties.
This problem is obviously dierent from the ILC approach, which, as
we have noted, is not a feedback methodology. The iterative learning
controller cannot aect the system poles.
2. Optimal control: Most optimal control problems can be generally described as:
min
kE k
C
subject to: P and R given and to constraints on U .
In optimal control we conduct an a priori design, based on a model
of the system. If the plant changes relative to the model then the
optimal controller will cease to be optimal. Further, the optimal controller operates in a feedback loop. Note, however, that in the case
of a stable plant it may be possible to design an ILC system that
produces the same output as an optimal controller, because both
methods are concerned with the minimization of a measure of the
error. The dierence is that the ILC algorithm achieves this by injecting the optimal input U into the system, as opposed to forming
it by processing the error in real-time. This can be illuminated by
referring to Figure 9.8, which emphasizes in a dierent way how the
ILC scheme can be viewed as an open-loop control strategy. The key
point is that ILC is a way to derive the signal U , using information
about past behavior of the system. Figure 9.8 points out the obvious
absence of an explicit controller in the ILC approach.
R-+ E-
U- P
Y-
440
K. L. Moore
441
and optimal control 193] have all used the term learning control to describe
their work. Most of these references, however, refer to learning in the sense
of
adapting or changing controller parameters on-line, as opposed to the oline learning of ILC. Other authors have considered the general problems of
learning from a broader view than either ILC or adaptation of parameters.
71, 227] are early works on the subject. 233] gives a recent mathematical perspective. Several researchers have also considered learning control
as a special case of learning in general, in the context of intelligent systems 94, 214, 53, 252]. A particularly visionary work in this regard is 2].
We mention one other area in which the phrase \learning control" arises:
reinforcement learning control. Reinforcement learning controllers are sophisticated stochastic search engines and are very much an ILC technique
in the sense that we have dened ILC. They work by evaluating the outcome of an action after the action and its eect are over. The next action
is then chosen based on the outcome of the previous action. Because of the
stochastic nature of this approach we will not discuss it further, but refer
the reader to 20, 69, 220, 161]. Likewise we will not consider any explicitly
stochastic learning controllers 206, 60].
Finally, after saying what ILC is not, we should say what ILC is. Terms
used to describe the process of Figure 9.1 include \betterment process"
(Arimoto's original description), \iterative control," \repetitive control,"
\training," and \iterative learning control." As in the discussion about
the term \learning," one must be careful to dene what is meant in any
particular usage of a word or phrase. For instance, the term \repetitive
control" is used to mean ILC but is also used to describe the control of
periodic systems (see the discussion in the next section). Also, in 248]
the term \virtual reference" is introduced to describe the optimal input
signal derived by the learning controller. This emphasizes the fact that
ILC algorithms produce the output of an optimal prelter and what we are
doing in essence is to compute a \virtual" or \not-real" input to try to fool
the system into going to where we want it to go.
442
K. L. Moore
the reader to 162], 96], 177], 23], and 145], each of which contains a
signicant amount of tutorial information.
The rst learning control scheme proposed by Arimoto, et al. involved the
derivative of the error ek (t) = yd (t) ; yk (t) 15]. Specically, the algorithm
had the form
uk+1 = uk + ;e_ k :
This is the continuous-time version of the algorithm we gave in the linear
example above:
uk+1 (t) = uk (t) + ;ek (t + 1):
For the case a linear time-invariant (LTI) system, with signals dened over
the interval 0 tf ], and with a state-space description (A,B,C), Arimoto, et
al. showed that if CB > 0 and if the induced operator norm kI ; CB ;ki
satises
kI ; CB ;ki < 1
and some initial condition requirements are met, then
lim y (t) ! yd (t)
k!1 k
in the sense of the -norm, dened as
kx(t)k = sup fe;t 1max
jxi jg
ir
0ttf
Z
uk+1 = uk + "ek + ;e_ k + # ek dt:
This algorithm essentially forms a PID-like system for processing the error
from the previous cycle, while maintaining a linear eect on the past input
signal.
Much of the early work in ILC focused on linear operators, with learning
laws of the form:
uk+1 = Tu uk + Te (yd ; yk )
443
where Tu and Te are both linear operators. This is really the most general
linear algorithm we might consider because it allows separate weighting
of both the current input and the current error. In 162] we proved the
following sucient condition:
Theorem For the LTI plant yk = Ts uk , the proposed LTI learning control algorithm
uk+1 = Tu uk + Te (yd ; yk )
converges to a xed point u (t) if
kTu ; Te Ts ki < 1:
The xed point u is given by
u (t) = (I ; Tu + Te Ts );1 Te yd(t)
and the resulting xed point of the error is given by
e (t) = klim
(yk ; yd) = (I ; Ts (I ; Tu + Te Ts );1 Te )yd(t)
!1
for t 2 t0 tf ].
The gist of almost all of the ILC results is related to proper selection of
the learning operators Tu and Te for specic classes of systems. In the
remainder of this section we will discuss some of these results.
Optimization-Based Approaches
One approach taken by some authors is to pick Te so as to force the convergence to follow some gradient of the error. For example, in 222] the
discrete-time learning control algorithm
uk+1 (t) = uk (t) + Gek (t + 1)
is used, with the gain G optimized using gradient methods to minimize the
quadratic cost of the error
J = 12 eTk (i + 1)Qek (i + 1):
between successive trials. The authors consider several techniques for choosing G, specically using the steepest-descent, Gauss-Newton, and NewtonRaphson methods. The rst two result in a constant gain G, giving a learning controller with exactly the same form as Arimoto, et al. For the NewtonRaphson method the result is a time-varying gain Gk which is dierent for
each trial. In 73] the update algorithm has the form
uk+1 = uk +
k Tp ek
where Tp is the adjoint operator of the system and
k is a time-varying
gain computed from the error and the adjoint to provide a steepest-descent
minimization of the error at each step of the iteration. This work also
explicitly considered multivariable systems.
444
K. L. Moore
Norm-Optimal ILC
A more recent approach to ILC algorithm design has been developed by
Amann, et al. 8, 7, 5, 10]. They compute the change in the input so as to
minimize the cost
445
kH ki = kH k1 = max
(
i
j =1
jhij j):
hN hN ;1 hN ;2 : : : h1
446
K. L. Moore
447
448
K. L. Moore
for systems with time delay properties. In a result that parallels a number
of results about the performance of ILC, 210] proves that the there exist
no bounded operator Te such that the update law uk+1 = uk + Tuek results
in exponential convergence for linear systems. However, the paper shows
that by introducing a digital controller it is possible to achieve exponential
convergence in exchange for certain non-zero residuals. Much of the most
well-developed work regarding convergence is by Amann, et al. Be careful
to note, however, that the ILC update law used in much of this work is
not the same as the one we used in most of this paper. Rather, Amann's
work uses current cycle feedback along with previous cycle feedback. One
of the main approaches to studying convergence in ILC is to invoke highgain results. In the robotics example given above, the convergence proof
is based on the fact that when the adaptive gain k gets big enough then
the ILC algorithm will converge. Amann explores this carefully in several
papers (see 8], for example). Other references in which a high-gain condition is invoked are listed in Table 9.1. One of the new directions for the
study of convergence is to consider modication to the ILC update law.
For instance, in 21] the authors use errors from more than one previous
cycle to update the input. By doing this they show that convergence can
be improved. This is called higher-order ILC.
Associated with the question of convergence is that of robustness. Generally, the results that are available consider a convergent ILC algorithm
and study how robust its convergence properties are to various types of disturbances or uncertainties. Much recent work has focused on mis-match in
initial condition from trial to trial. 86] computed error bounds due to the
eects of state and output noise as well as initial condition mis-match. In
129] a very good analysis of this eect is given and it is shown how undesirable eects due to mismatch can be overcome by utilizing a previous cycle
pure error term in the learning controller. 202] considers robustness and
convergence for a \D-type" ILC algorithm. A condition is given for a class
of linear systems to ensure global robustness to state disturbances, measurement noise, and initialization errors. Another approach to robustness
has been to combine current cycle feedback with previous cycle feedback
41, 40]. In particular, in 40] a typical \D-type" ILC algorithm is combined with a scheme for learning the initial condition. This information is
then used to oset the eect of the initial condition mis-match. For linear
systems, 139] shows how to use H1 theory in a two-dimensional systems
theory framework to address modelling uncertainty.
Most of the basic results described above were derived primarily for linear
systems. However, researchers have also considered the learning control
problem for dierent classes of nonlinear systems. Of particular interest
to many researchers are the classes of nonlinear systems representative of
449
robotic manipulator models. However, ILC for nonlinear systems has also
been considered independent of robotics.
Suppose now that our system is dened by
uk+1 = uk + (yd ; yk ):
This is a simple linear learning control algorithm, with unity weighting on
both the current input and the current error. If yk (t) = fP (uk ), where fP ()
is continuous on the interval of interest, then convergence is guaranteed if
kI ; fP0 (u)k < 1
450
K. L. Moore
u. This result gives a class of systems for which the ILC algorithm will
converge.
However, the previous result is very general and is not always easily
checked. Less general results can be obtained by restricting the plant and
the ILC algorithms to have a more specic structure. Let the plant be
described by
x_ k = a(xk t) + bp (t)uk
yk = c(xk t) + dp (t)uk
with a(x t) and c(x t) Lipschitz in their arguments. Let the ILC update
law be
v_ k = Ac (t)vk + Bc (t)ek
uk+1 = Cc (t)xk + Dc (t)ek + uk :
For this setup, convergence can be shown if 215]
kI ; dp (t)Dc (t)k < 1:
Notice that this result depends only on the direct transmission terms from
the plant and the learning controller.
A similar results was given in 84]. For the system
451
above give an idea of the types of analysis and results available for nonlinear ILC. Interested readers can consult the references listed in Table 9.1 for
additional information. We have previously noted in 162] that there is not
a unifying theory of iterative learning control for nonlinear systems. This
does not seem to have changed as of this writing. Results from contraction
mapping or Lyapunov techniques usually provide sucient, but general
conditions for convergence, which must be applied on a case-by-case basis.
To obtain more useful results it is often necessary to restrict our attention
to more specic systems, as in the examples above. One example of this is
the problem of learning control for robotic manipulators. By assuming that
the nonlinear system has the standard functional form typical of a manipulator, researchers have been able to establish specic learning controllers
that will converge. We mention as a nal comment that many researchers
have begun to use neural networks as one tool for nonlinear ILC. In most
papers, for instance 234], by this author, the results show an approach
that works, but there is little analysis to say why it works. Consequently,
it is our view that the more general problem of ILC for nonlinear systems
is still an open area of research.
We have noted several times that robotics is the natural application area
for ILC. Arimoto, et al.'s original work included a discussion of learning
control for robotics and others have independently proposed similar learning and adaptive, motivated by robotic control problems 55, 56, 54]. It is
beyond the scope of this single paper to discuss ILC for robotics in any
suitable detail. The interested reader can refer to the references in Table
9.2 to identify papers dealing with dierent aspects of ILC and robotics.
In particular, we recommend 96], which contains a good summary of ILC
for robotics. Also, to get a avor for the types of algorithms that are used,
the reader may refer back to the representative example of an ILC algorithm applied to a simulated two-joint manipulator that was given earlier
in the paper. See also 24], which rst gave the high-gain feedback, modelreference approach to learning control that motivated the adaptive result
presented in the example. It should be clear from Table 9.2 that researchers
have considered a wide-variety of problems in robotics. It is also interesting to point out the large number of actual demonstrations of ILC using
robots. Indeed, in 248] an incredible demonstration of an ILC algorithms
is reported in which a robot arm iteratively learns to catch a ball in a cup
(the Japanese Kendama game).
In addition to robotics, iterative learning control has been applied to an
increasing number of applications, as shown in Table 9.2. Most of these
have been to problems that involve repetitive or iterative operations. For
instance, in chemical processing, ILC has been applied to batch reactor
control problems, which is inherently an iterative activity 251, 50, 132].
452
K. L. Moore
Other applications emphasize learning on-line based on rejection of periodic or repeatable errors that occur during normal step changes or other
operational procedures 232, 231, 127, 107]. Many of these consider periodic
systems and apply ILC to solve disturbance rejection or tracking problems,
including applications for controlling peristaltic pump used in dialysis 93],
coil-to-coil control in a rolling mill 74], vibration suppression 92], and a
nonlinear chemical process 29]. Motion control of non-robotic systems has
also attracted a signicant amount of research eort for ILC applications,
including a scanner driver for a plain paper copier 253], a servo system
for a VCR 131], a hydraulic servo for a machining system 47], an inverted pendulum 179], a cylindrical cutting tool 79], CNC machine tools
228, 119, 250, 120], and an optical disk drive 159]. 137] also describes an
application to self-tuning a piezo-actuator. Following the next section we
apply some new approaches to learning control to a gas-metal arc welding
problem 170].
In this section we will discuss three specic emerging areas in ILC research.
The rst is the intersection between repetitive control and ILC, which has
led to the idea of \no-reset" or \continuous" ILC. The second area is the
development of ILC algorithms that do not t the standard form of uk+1 =
Tu uk + Te ek : We end with what has been called \direct learning control."
Repetitive Control and No-Reset/Continuous ILC
Strictly speaking, repetitive control is concerned with cancelling an unknown periodic disturbance or tracking an unknown periodic reference signal 96]. The solutions that have been developed in the literature tend to
focus on the internal model principle to produce a periodic controller 80].
As such, the repetitive controller is a feedback controller as opposed to
the ILC scheme, which ultimately acts as a feedforward controller. Other
dierences include the fact that the ILC algorithms act on a nite horizon whereas the repetitive controllers are continuous and the fact that the
typical assumption of ILC is that each trial starts over at the same initial
condition whereas the repetitive control system does not start with such
an assumption. Despite these dierences however, it is instructive to consider the repetitive control strategy (a good summary of repetitive control
design is found in 128]), because the intersection between the strict denitions of repetitive control and ILC is a fertile ground for research. It is
also interesting to note that the topical breakdown between the two elds
is very similar, with progress reported in the eld of repetitive control for
linear systems 156, 80, 81], multivariable systems 203], based on modelmatching and 2-D systems theory 3], for discrete-time systems explicitly
97, 223, 43, 27], using frequency domain results 98, 211], nonlinear sys-
453
tems 151], and stochastic systems 43, 122]. Also note that the connections
between ILC and repetitive control are not new. In particular, see many
of the works by Longman, et al. (145], for example), which make it clear
that ILC and repetitive control are really the same beast.
To illustrate how ILC and repetitive control are similar, we consider a
simulated vibration suppression problem. Consider a simple system dened
by
y(t) = Kp (u(t) + Tdd(t))
where we dene
;Td s
Td = Gd (s) = K dse + 1 :
d
Thus the system consists of a constant gain plant with an additive disturbance that is delayed and attenuated before it is applied to the plant. We
suppose d(t) is periodic (sinusoidal) with a known frequency. Suppose we
wish to drive the output of the system to zero. It is clear that after all
transients have died away the output will be periodic with the same period
as the disturbance. This suggests a way to apply an ILC-like algorithm to
the problem. We simply view each cycle of the output as a trial and apply
the standard ILC algorithm. Because the output is periodic we can expect
to have no problem with initial conditions at the beginning of each \trial."
The typical ILC update would be:
But, this is simply a delay factor, as depicted in Figure 9.9, and this last
equation is what has been called \... the familiar repetitive control law..." in
203]. Thus our ILC algorithm reduces in this case to a repetitive controller.
Regardless of the implementation of the algorithm, the resulting controller
easily suppresses the disturbance from the output, as shown in Figure 9.10.
From this example one can see that ILC applied to a periodic, continuoustime situation is equivalent to repetitive control. This has been addressed
recently by several researchers. The idea is called \no-reset" ILC in 208]
because the system never actually starts, stops, resets, and then repeats.
Rather, the operation is continuous. This suggests calling an ILC approach
to such a control system to be a \continuous" iterative learning controller.
Below we give an example of an extended version of the idea for a system
454
K. L. Moore
- + E - Tez;N
oS+
{ 6S
S
S
qj
D
- +? -
-Y
455
1
0
-1
0
10
15
20
25
(a)
30
35
40
45
50
10
15
20
25
(b)
30
35
40
45
50
10
15
20
25
(c)
30
35
40
45
50
0.5
0
-0.5
0
0.5
0
-0.5
0
FIGURE 9.10. System response for continuous ILC example: (a) disturbance input (b) output signal (c) ILC produced input signal.
Thus we have simply added the error from the current cycle to the
ILC algorithm. In such a scheme it is easy to show for a plant Tp and
with Tu = I the sucient condition for convergence becomes:
k(I + TpTefb );1 (I ; TpTeff )k < 1:
Thus we see a combination of the normal ILC convergence condition
with a more common feedback control type expression. Amann, et
al. have given a slightly dierent form of the current error feedback
algorithm:
uk+1 (t) = Tu uk (t0 ) + Te ek+1 (t0 ):
This expression does not explicitly use past cycle error information,
but does keep track of the previous inputs. Most of the results on
norm-optimal ILC by Amann, et al. use this form of an update algorithm. An obvious advantage of either of these current error feedback
algorithms is that it is possible to ensure stability in both trial convergence and with respect to time relative to the closed-loop system.
Others to consider current cycle feedback include 36, 170, 42].
456
K. L. Moore
Trial (k-1)
Error
...
...
t-1
Trial k
...
t t+1
Trial (k+1)
...
t-1
...
t t+1
...
t-1
t+1
Input
...
...
t-1 t
...
...
t-1 t
t+1
...
...
t-1 t t+1
t+1
(a)
Error
...
...
t-1 t t+1
...
...
...
t-1 t t+1
...
t-1
t+1
Input
...
...
t-1 t
t+1
...
...
t-1 t
t+1
...
...
t-1 t t+1
(b)
FIGURE 9.11. Two new ILC strategies: (a) higher-order learning control (b) current cycle feedback.
can be combined to give what may be the most general form of an ILC
algorithm:
uk+1 (t) = Tfb ek+1 (t) + Tu uk (t0 ) + Te1 ek (t0 ) + Te2 ek;1 (t0 ) +
This algorithm was considered by Xu in 36, 42]. With such an algorithm
it may be possible to take exploit the advantages of both the current error
feedback and the higher-order elements. This will certainly be an important
area for additional research.
Direct Learning
One nal category that we would like to address is what has recently been
called \direct learning" 239]. The concern in this area of study is that
457
once an ILC algorithm has converged and is then presented with a new
trajectory to follow it will lose any memory of the previous trajectory, so
that if it is presented again it will have to re-learn. Of course, in practice we
may keep track of the optimal input U corresponding to dierent desired
outputs, but at a more fundamental level, this represents a shortcoming of
the ILC approach. Arimoto, et al. considered this in some earlier papers
112, 113]. Also, in 162] an approach was given for \learning with memory."
Recently, however, the problem was posed in two interesting ways:
1. Can we generate the optimal inputs for signals that have the same
shape and magnitude, but dierent time scales by learning only one
signal 239, 243]?
2. Can we generate the optimal inputs for signals that have the same
shape and time scales, but dierent magnitudes 246]?
At this time there are no denitive conclusions that can be drawn, but this
will be an important research area in the future (see also 244, 115]).
458
K. L. Moore
In 171] the following simplied model of the gas-metal arc welding process
(GMAW) is given:
x_ 1 = x2
x_ 2 = m1(t) (;2:5x1 ; 10;5x2 + F + I 2 )
m_ = k2 I + k5 I 2 Ls
Reset condition:
m(t) =
m(t)
if m 25
m(1 ; 0:8( 1+e;110x2 ) + 0:125) otherwise
459
30
mass
20
10
0
0
10
20
30
40
50
60
70
80
90
100
10
20
30
40
50
time
60
70
80
90
100
0.6
0.4
vel
0.2
0
-0.2
-0.4
25
20
15
10
5
0
10
15
20
25
30
35
40
460
K. L. Moore
461
(a) If the actual mass resets before the desired mass, then reset the
desired mass and continue, using a typical ILC update algorithm.
(b) If the desired mass resets before the actual mass, then (1) set
the inputs and the desired mass to some xed waveform (e.g., a
nominal constant)& (2) continue until the actual mass resets& (3)
reset the desired mass when the actual mass resets and continue
using a typical ILC update algorithm.
(c) If the actual mass on previous trials has always reset at a time
less that the desired reset time, then the rst time the system
goes past the longest time the system had ever previously run
the ILC algorithm will not have an error signal to use. To handle
this all errors should be initialized at zero and updated only as
data becomes available.
What is happening in item (b) is that if the actual system has not
reset by the time we want it to, we simply suspend the ILC algorithm.
That is, it should not be doing anything at time greater than the
period of the desired output waveform. Thus, we just provide the
system a nominal input and wait until the actual mass resets. We
then continue, computing the errors for our ILC algorithm from the
beginning of the past trial. Likewise, in item (c) we wish to ensure
that, if the duration of the previous trial was shorter than desired,
but the current trial's length is longer than the previous trial's length,
then we do not compute changes to the input for times greater than
the previous trial (because then you would actually be in the current
trial).
4. ILC algorithm to adjust the slope: The nal piece of our controller is
the use of a standard ILC algorithm to update the stickout based on
errors from the previous trial. We do not adjust the current, which
is dedicated to controlling the velocity using current (present) error
feedback.
Algorithmically, the ILC procedure can be represented in terms of a new
quantity, called the \duration." The duration, denoted as td k , is the length
of trial k. Using this notation, and dening the desired mass to be md (t),
the desired trial length to be tdd , and the starting time of each trial to be
ts k , the ILC algorithm can be written as:
462
K. L. Moore
'(t)(m(t))1=2
I (t) = I8
Ls (t ; tdk;1 )
>
>
>
< +k4 (m_ d (t ; td k;1 + 1)
Ls (t) = > ;m_ (t ; tdk;1 + 1)) if (t ; tsk )
tdk;1
>
>
: Ls (t ; 1)
otherwise
It should be emphasized again that the desired mass waveform is dened
according to when the actual trial ends relative to the desired trial length.
If the actual trial ends then the desired mass is reset to its initial value as
the next trial begins. If the trial lasts longer than the desired length then
the desired mass is set as follows (and, the ILC algorithm for stickout is
discontinued until the next trial begins):
md (t) =
md(t ; ts k ) if (t ; ts k ) tdd
mdmax
otherwise
463
40
20
0
1
50
100
150
200
250
300
50
100
150
200
250
300
50
100
150
200
250
300
50
100
150
200
250
300
50
100
150
200
250
300
50
100
150
200
250
300
0.5
0
20
0
-20
20
edot
0
-20
4
2
0
2
Ls
0
-2
time
464
K. L. Moore
pret ILC in terms of other control paradigms and in the larger context of
learning in general. Looking to the future, it seems clear to this author that
there a number of areas of research in ILC that promise to be important.
These include:
1. Integrated Higher-Order ILC/Current-Cycle Feedback: We have noted
that by including current-cycle feedback together with higher-order
past-cycle feedback it is possible to simultaneously stabilize and achieve the desired performance, including possible improvements in convergence rates from the system. However, more work is needed to
understand this approach.
2. Continuous ILC/Repetitive Control: This is one of the most important areas for future research. The last example presented above
shows the value of such an approach. What is needed now is to understand how the technique can be made more general. Also it must also
be reconciled with the more strict denitions of repetitive and periodic control. However, a particular vision of this author is that ILC
can be used in a continuous situation when the goal is to produce
a periodic response so as to produce what can be called a nonlinear, phase-locked, frequency-locked loop. This can lead to signicant
applications in motor control (pulse-width modulated systems) and
communications.
3. Robustness and Convergence Analysis: What is still needed is more
rigorous and more conclusive results on when algorithms will converge
and how robust this convergence will be. Such information can be
used to develop comprehensive theories of ILC algorithm design.
4. System-Theoretic Analysis: In the same vein as robustness and convergence, more work is needed to characterize the capabilities of ILC
algorithms. One such analysis is extend the 2-D analysis that some
authors have applied to linear systems to the nonlinear case in order
to provide combined global convergence and stability results. Another
is to explore the connections with other optimal control methodologies. As an example, if one poses the problem of L2 minimization of
error on a xed interval 0 tf ] and solves it using ILC, one would
expect the resulting u (t) to be the same as the u (t) that results
from solving a standard linear quadratic regulator problem on the
same interval. It would also be interesting to consider the same type
of comparison for mixed-sensitivity problems that do not have analytical solutions to see if it is possible in such cases to derive the
optimal input using ILC. These same comments also apply to ILC
for nonlinear systems.
5. Connections to More General Learning Paradigms: One of the important areas of research for ILC in the future will be developing
465
1] Hyun-Sik Ahn, Soo-Hyung Lee, and Do-Hyun Kim. Frequencydomain design of iterative learning controllers for feedback systems.
In IEEE International Symposium on Industrial Electronics, volume 1, pages 352{357, Athens, Greece, July 1995.
2] J. Albus. Outline for a theory of intelligence. IEEE Transactions on
Systems, Man, and Cybernetics, 21(3):473{509, May/June 1991.
3] David M. Alter and Tsu-Chin Tsao. Two-dimensional exact model
matching with application to repetitive control. Journal of Dynamic
Systems, Measurement, and Control, 116:2{9, March 1994.
4] N. Amann and D. H. Owens. Non-minimum phase plants in iterative
learning control. In Second International Conference on Intelligent
Systems Engineering, pages 107{112, Hamburg-Harburg, Germany,
September 1994.
5] N. Amann and D. H. Owens. Iterative learning control for discrete
time systems using optimal feedback and feedforward actions. In
466
K. L. Moore
467
468
K. L. Moore
29] Chien Chong Chen, Chyi Hwang, and Ray Y. K. Yang. Optimal periodic forcing of nonlinear chemical processes for performance improvements. The Canadian Journal of Chemical Engineering, 72:672{682,
August 1994.
30] Peter C. Y Chen, James K. Mills, and Kenneth C. Smith. Performance improvement of robot continuous-path operation through iterative learning using neural networks. Machine Learning, 23:191{220,
May-June 1996.
31] Y. Chen and H. Dou. Robust curve identication by iterative learning. In Proceedings of the 1st Chinese World Congress on Intelligent
Control and Intelligent Automation, Beijing, China, 1993.
32] Y. Chen, M. Sun, and H. Dou. Dual-staged P-type iterative learning
control schemes. In Proceedings of the 1st Asian Control Conference,
Tokyo, Japan, 1994.
33] Y. Chen, C. Wen, and M. Sun. Discrete-time iterative learning control of uncertain nonlinear feedback systems. In Proceedings of the
2nd Chinese World Congress on Intelligent Control and Intelligent
Automation, Xi'an, China, 1997.
34] Y. Chen, C. Wen, and M. Sun. A high-order iterative learning controller with initial state learning. In Proceedings of the 2nd Chinese
World Congress on Intelligent Control and Intelligent Automation,
Xi'an, China, 1997.
35] Y. Chen, C. Wen, J.-X. Xu, and M. Sun. extracting projectile's
aerodynamic drag coecient curve via high-order iterative learning
identication. In Proceedings of 35th IEEE Conference on Decision
and Control, Kobe, Japan, December 1996.
36] Y. Chen, J.-X. Xu, and T.H. Lee. Feedback-assisted high-order iterative learning control of uncertain nonlinear discrete-time systems. In
Proceedings of the International Conference on Control, Automation,
Robotics, and Vision, Singapore, December 1996.
37] Y. Chen, J.-X. Xu, T.H. Lee, and S. Yamamoto. Comparative studies
of iterative learning control schemes for a batch chemical process.
In Proceedings of the IEEE Singapore Int. Symposium on Control
Theory and Applications, Singapore, July 1997.
38] Y. Chen, J.-X. Xu, T.H. Lee, and S. Yamamoto. An iterative learning
control in rapid thermal processing. In Proceedings of IASTED Int.
Conference on Modeling, Simulation, and Optimization, Singapore,
August 1997.
469
39] Y. Chen, J.-X. Xu, and T.-H. Tong. An iterative learning controller
using current iteration tracking error information and initial state
learning. In Proceedings of the 35th IEEE Conference on Decision
and Control, Kobe, Japan, December 1996.
40] Yangquan Chen, Changyun Wen, Jian-Xin Xu, and Mingxuan Sun.
Initial state learning method for iterative learning control of uncertain
time-varying systems. In Proceedings of the 35th IEEE Conference
on Decision and Control, volume 4, pages 3996{4001, Kobe, Japan,
December 1996.
41] Yangquan Chen, Jian-Xin Xu, and Tong Heng Lee. Current iteration
tracking error assisted iterative learning control of uncertain nonlinear discrete-time systems. In Proceedings of the 35th IEEE Conference on Decision and Control, volume 3, pages 3038{3043, Kobe,
Japan, December 1996.
42] Yangquan Chen, Jian-Xin Xu, and tong Heng Lee. Current iteration tracking error assisted high-order iterative learning control of
discrete-time uncertain nonlinear systems. In Proceedings of the 2nd
Asian Control Conference, Seoul, Korea, July 1997.
43] Kok Kia Chew and Masayoshi Tomizuka. Steady state and stochastic performance of a modied discrete-time prototype repetitive controller. Journal of Dynamic Systems, Measurement, and Control,
112:35{41, March 1990.
44] Chiang-Ju Chien. A discrete iterative learning control of nonlinear
time- varying systems. In Proceedings of the 35th IEEE Conference
on Decision and Control, volume 3, pages 3056{3061, Kobe, Japan,
December 1996.
45] Chiang-Ju Chien. On the iterative learning control of sampled-data
systems. In Proceedings of the 2nd Asian Control Conference, Seoul,
Korea, July 1997.
46] Chiang-Ju Chien and Jing-Sin Liu. P-type iterative learning controller for robust output tracking of nonlinear time-varying systems.
International Journal of Control, 64(2):319{334, May 1996.
47] Tsu chin Tsao and Masayoshi Tomizuka. Robust adaptive and repetitive digital tracking control and application to a hydraulic servo for
noncircular machining. Journal of Dynamic Systems, Measurement,
and Control, 116:24{32, March 1994.
48] Chong-Ho Choi and Tae-Jeong Jang. Iterative learning control for
a general class of nonlinear feedback systems. In Proceedings of the
1995 American Control Conference, pages 2444{2448, Seattle, WA,
June 1995.
470
K. L. Moore
61]
62]
63]
64]
65]
66]
67]
68]
69]
471
on Systems, Man, and Cybernetics, volume 4, pages 3790{3794, Vancouver, BC, Canada, October 1995.
Tae-Yong Doh, Jung-Hoo Moon, Kyung Bog Jin, and Myung Jin
Chung. An iterative learning control for uncertain systems using
structured singular value. In Proceedings of the 2nd Asian Control
Conference, Seoul, Korea, July 1997.
H. Dou, Z. Zhou, Y. Chen, J.-X. Xu, and J. Abbas. Robust motion control of electrically stimulated human limb via discrete-time
high-order iterative learning scheme. In Proceedings of the International Conference on Control, Automation, Robotics, and Vision,
Singapore, December 1996.
Huifang Dou, Zhaoying Zhou, Yangquan Chen, Jian-Xin Xu, and
James J. Abbas. Robust control of functional neurouscular stimulation system by discrete-time iterative learning scheme. In Proceedings
of the 2nd Asian Control Conference, Seoul, Korea, July 1997.
Huifang Dou, Zhaoying Zhou, Mingxuan Sun, and Yangquan Chen.
Robust high-order P-type iterative learning control for a class of uncertain nonlinear systems. In Proceedings of the IEEE International
Conference on Systems, Man, and Cybernetics, volume 2, pages 923{
928, Beijing, China, October 1996.
H. Elci, R.W. Longman, M.Q. Phan, J. Juang, and R. Ugoletti. Automated learning control through model updating for precision motion
control. In ASME Winter Meeting, Adaptive Structures and Materials Systems Symposium, Chicago, Illinois, November 1994.
H. Elci, R.W. Longman, M.Q. Phan, J. Juang, and R. Ugoletti. Discrete frequency-based learning control for precision motion control.
In Proceedings of the 1994 IEEE Conference on Systems, Man, and
Cybernetics, San Antonio, Texas, October 1994.
H. Elci, M. Phan, and R. W. Longman. Experiments in the use of
learning control for maximum precision robot trajectory tracking. In
Proceedings of the 1994 Conference on Information Science and Systems, pages 951{958, Princeton University, Department of Electrical
Engineering, Princeton, NJ, 1994.
H. Elci, H. N. Juang R. W. Longman, M. Phan, and R. Ugoletti.
Automated learning control through model updating for precision
motion control. Adaptive Structures and Composite Materials: Analysis and Applications, ASME, AD-45/MD-54:299{314, 1994.
J. A. Franklin. Renement of robot motor skills through reinforcement learning. In Proceedings of 27th Conference on Decision and
Control, pages 1096{1101, Austin, Texas, December 1988.
472
K. L. Moore
70] James A. Frueh and Minh Q. Phan. System identication and inverse
control using input-output data from repetitive trials. In Proceedings
of the 2nd Asian Control Conference, Seoul, Korea, July 1997.
71] K.S. Fu. Learning control systems | review and outlook. IEEE
Transactions on Automatic Control, 15:210{215, April 1970.
72] Mitsuo Fukuda and Seiichi Shin. Model reference learning control
with a wavelet network. In Proceedings of the 2nd Asian Control
Conference, Seoul, Korea, July 1997.
73] K. Furuta and M. Yamakita. The design of a learning control system
for multivariable systems. In Proceedings of IEEE International Symposium on Intelligent Control, pages 371{376, Philadelphia, Pennsylvania, January 1987.
74] S. S. Garimella and K. Srinivasan. Application of iterative learning
control to coil-to-coil control in rolling. In Proceedings of the 1995
American Control Conference, Seattle, WA, June 1995.
75] Zheng Geng, Robert Carroll, and Jahong Xie. Two-dimensional
model and algorithm analysis for a class of iterative learning control
systems. International Journal of Control, 52(4):833{862, December
1989.
76] D. Gorinevsky. An approach to parametric nonlinear least square
optimization and application to task-level learning control. IEEE
Transactions on Automatic Control, 42(7):912{927, 1997.
77] Y.L. Gu and N.K. Loh. Learning control in robotic systems. In
Proceedings of IEEE International Symposium on Intelligent Control,
pages 360{364, Philadelphia, Pennsylvania, January 1987.
78] Kennon Guglielmo and Nader Sadegh. Theory and implementation
of a repetitive robot controller with Cartesian trajectory description.
Journal of Dynamic Systems, Measurement, and Control, 118:15{21,
March 1996.
79] Reed D. Hanson and Tsu-Chin Tsao. Compensation for cutting
force induced bore cylindricity dynamic errors - a hybrid repetitive
servo/iterative learning process feedback control approach. In Proceedings of the Japan/USA Symposium on Flexible Automation, volume 2, pages 1019{1026, Boston, MA, July 1996.
80] S. Hara, T. Omata, and M. Nakano. Synthesis of repetitive control systems and its application. In Proceedings of 24th Conference
on Decision and Control, pages 1387{1392, Ft. Lauderdale, Florida,
December 1985.
473
81] S. Hara, Y. Yamamoto, T. Omata, and M. Nakano. Repetitive control system: A new type servo system for periodic exogenous signals.
IEEE Transactions on Automatic Control, 33(7):659{668, July 1988.
82] E. Harokopos. Optimal learning control of mechanical manipulators
in repetitive motions. In Proceedings of IEEE International Symposium on Intelligent Control, pages 396{401, Philadelphia, Pennsylvania, January 1987.
83] H. Hashimoto and J.-X. Xu. Learning control systems with feedback. In Proceedings of the IEEE Asian Electronics Conference, Hong
Kong, September 1987.
84] J.E. Hauser. Learning control for a class of nonlinear systems. In
Proceedings of 26th Conference on Decision and Control, pages 859{
860, Los Angeles, California, December 1987.
85] G. Heinzinger, D. Fenwick, B. Paden, and F. Miyazaki. Robust learning control. In Proceedings of the 28th Conference on Decision and
Control, pages 436{440, Tampa, Florida, December 1989.
86] Greg Heinzinger, Dan Fenwick, Brad Paden, and Fumio Miyazaki.
Stability of learning control with disturbances and uncertain initial
conditions. IEEE Transactions on Automatic Control, 37(1), January
1992.
87] L. Hideg and R. Judd. Frequency domain analysis of learning systems. In Proceedings of the 27th Conference on Decision and Control,
pages 586{591, Austin, Texas, December 1988.
88] L. M. Hideg. Stability and convergence issues in iterative learning
control. In Proceedings Intelligent Engineering Systems Through Articial Neural Networks in Engineering Conference, volume 4, pages
211{216, St. Louis, MO, November 1994.
89] L. M. Hideg. Stability of LTV MIMO continuous-time learning control schemes: a sucient condition. In Proceedings of the 1994 IEEE
International Symposium on Intelligent Control, pages 285{290, 1995.
90] L. M. Hideg. Time delays in iterative learning control schemes. In
Proceedings of the 1995 IEEE International Symposium on Intelligent
Control, pages 5{20, Monterey, CA, August 1995.
91] L. M. Hideg. Stability and convergence issues in iterative learning
control - II. In Proceedings of the 1996 IEEE International Symposium on Intelligent Control, pages 480{485, Dearborn, MI, September
1996.
474
K. L. Moore
475
476
K. L. Moore
477
124] Tae-Yong Kuc and Chae-Wook Chung. An optimal learning control of robot manipulators. In Proceedings of the 2nd Asian Control
Conference, Seoul, Korea, July 1997.
125] Tae-Yong Kuc, Jin S. Lee, and Byung-Hyun Park. Tuning convergence rate of a robust learning controller for robot manipulators. In
Proceedings of the 1995 IEEE Conference on Decision and Control,
volume 2, pages 1714{1719, New Orleans, LA, 1995.
126] Jerzy E. Kurek and Marek B. Zaremba. Iterative learning control
sysnthesis based on 2-D system theory. IEEE Transactions on Automatic Control, 38(1), January 1993.
127] R. J. LaCarna and J. R. Johnson. A learning controller for the
megawatt load-frequency control problem. IEEE Transactions on
Systems, Man, and Cybernetics, 10(1):43{49, January 1980.
128] G. F. Ledwich and A. Bolton. Repetitive and periodic controller
design. In IEE Proceedings Part D, Control Theory and Applications,
volume 140, pages 19{24, January 1993.
129] Hak-Sung Lee and Zeungnam Bien. Study on robustness of iterative
learning control with non-zero initial error. International Journal of
Control, 64:345{359, June 1996.
130] Hak-Sung Lee and Zeungnam Bien. Robustness and convergence of a
pd-type iterative learning controller. In Proceedings of the 2nd Asian
Control Conference, Seoul, Korea, July 1997.
131] Ju Jang Lee and Jong Woon Lee. Design of iterative learning controller with VCR servo system. IEEE Transactions on Consumer
Electronics, 39:13{24, February 1993.
132] K. S. Lee, S. H. Bang, S. Yi, J. S. Son, and S. C. Yoon. Iterative
learning control of heat up phase for a batch polymerization reactor.
Journal of Process Control, 6(4):255{262, August 1996.
133] Kwang Soon Lee and Jay H. Lee. Constrained model-based predictive control combined with iterative learning for batch or repetitive
processes. In Proceedings of the 2nd Asian Control Conference, Seoul,
Korea, July 1997.
134] S. C. Lee, R. W. Longman, and M. Phan. Direct model reference
learning and repetitive control. Advances in the Astronautical Sciences, 85, March 1994.
135] T. H. Lee, J. H. Nie, and W. K. Tan. Developments in learning
control enhancements for nonlinear servomechanisms. Mechatronics,
5(8):919{936, December 1995.
478
K. L. Moore
479
480
K. L. Moore
159] Jung-Ho Moon, Moon-Noh Lee, Myung Jin Chung, Soo Yul Jung,
and Dong Ho Shin. Track-following control for optical disk drives
using an iterative learning scheme. IEEE Transactions on Consumer
Electronics, 42(2):192{198, May 1996.
160] K.L. Moore. Design Techniques for Transient Response Control.
PhD thesis, Texas A&M University, College Station, Texas, 1989.
161] K.L. Moore. A reinforcement-learning neural network for the control of nonlinear systems. In Proceedings of 1991 American Control
Conference, Boston, Massachusetts, June 1991.
162] K.L. Moore. Iterative Learning Control for Deterministic Systems.
Springer-Verlag, Advances in Industial Control, London, U.K., 1993.
163] K.L. Moore, S.P. Bhattacharyya, and M. Dahleh. Some results on
iterative learning. In Proceedings of 1990 IFAC World Congress,
Tallin, Estonia, USSR, August 1990.
164] K.L. Moore, M. Dahleh, and S.P. Bhattacharyya. Learning control for robotics. In Proceedings of 1988 International Conference on Communications and Control, pages 976{987, Baton Rouge,
Louisiana, October 1988.
165] K.L. Moore, M. Dahleh, and S.P. Bhattacharyya. Articial neural
networks for learning control. TCSP Research Report 89{011, Department of Electrical Engineering, Texas A&M University, College
Station, Texas, July 1989.
166] K.L. Moore, M. Dahleh, and S.P. Bhattacharyya. Iterative learning
for trajectory control. In Proceedings of 28th IEEE Conference on
Decision and Control, Tampa, Florida, December 1989.
167] K.L. Moore, M. Dahleh, and S.P. Bhattacharyya. Some results on iterative learning control. TCSP Research Report 89{004, Department
of Electrical Engineering, Texas A&M University, College Station,
Texas, January 1989.
168] K.L. Moore, M. Dahleh, and S.P. Bhattacharyya. Adaptive gain
adjustment for a learning control method for robotics. In Proceedings
of 1990 IEEE International Conference on Robotics and Automation,
Cincinnati, Ohio, May 1990.
169] K.L. Moore, M. Dahleh, and S.P. Bhattacharyya. Iterative learning
control: A survey and new results. Journal of Robotic Systems, 9(5),
July 1992.
481
170] K.L. Moore and Anna Mathews. Iterative learning control for systems with non-uniform trial length with applications to gas-metal arc
welding. In Proceedings of the 2nd Asian Control Conference, Seoul,
Korea, July 1997.
171] D.S. Naidu, K.L. Moore, R. Yender, and J. Tyler. Gas metal arc
welding control: Part I { modeling and analysis. In Proceedings of
the 2nd World Congress of Nonlinear Analysts, Athens, Greece, July
1996.
172] K. Narendra and M. Thathachar. Learning automata | a survey.
IEEE Transactions on Systems, Man, and Cybernetics, 4(4):323{334,
July 1974.
173] K. Narendra and M Thathacher. Learning Automata. Prentice-Hall,
Englewood Clis, New Jersey, 1989.
174] S.-R. Oh, Z. Bien, and I. Suh. An iterative learning control method
with application for the robot manipulator. IEEE Journal of Robotics
and Automation, 4(5):508{514, October 1988.
175] Giuseppe Oriolo, Stefano Panzieri, and Giovanni Ulivi. Iterative
learning controller for nonholonomic robots. In IEEE International
Conference on Robotics and Automation, volume 3, pages 2676{2681,
Minneapolis, MN, 1996.
176] D. H. Owens. Iterative learning control-convergence using high gain
feedback. In Proceedings of the 1992 Conference on Decision and
Control, volume 4, page 3822, Tucson, AZ, December 1992.
177] D. H. Owens, N. Amann, and E. Rogers. Iterative learning control-an
overview of recent algorithms. Applied Mathematics and Computer
Science, 5(3):425{438, 1995.
178] D. H. Owens, N. Amann, and E. Rogers. Systems structure in iterative learning control. In Proceedings of International Conference on
System Structure and Control, pages 500{505, Nantes, France, July
1995.
179] H. Ozaki, C. J. Lin, T. Shimogawa, and H. Qiu. Control of mechatronic systems by learning actuator reference trajectories described
by B-spline curves. In Proceedings of the IEEE International Conference on Systems, Man, and Cybernetics, volume 5, pages 4679{4684,
Vancouver, BC, Canada, October 1995.
180] B. H. Park and Jin S. Lee. Continuous type fuzzy learning control
for a class of nonlinear dynamics system. In Proceedings of the 2nd
Asian Control Conference, Seoul, Korea, July 1997.
482
K. L. Moore
483
484
K. L. Moore
485
214] Y. Suganuma and M. Ito. Learning control and knowledge representation. In Proceedings of IEEE International Symposium on Intelligent Control, pages 354{359, Philadelphia, Pennsylvania, January
1987.
215] T. Sugie and T. Ono. An iterative learning control law for dynamical
systems. Automatica, 27:729{732, 1991.
216] Dong Sun and Xiaolun Shi Yunhui Liu. Adaptive learning control
for cooperation of two robots manipulating a rigid object with model
uncertainties. Robotica, 14(4):365{373, July-August 1996.
217] M Sun, Y. Chen, and H. Dou. Robust higher order repetitive learning
control algorithm for tracking control of delayed systems. In Proceedings of the 1992 Conference on Decision and Control, Tucson, AZ,
December 1992.
218] M. Sun, B. Huang, X. Zhang, and Y. Chen. Robust convergence of
D-type learning controller. In Proceedings of the 2nd Chinese World
Congress on Intelligent Control and Intelligent Automation, Xi'an,
China, 1997.
219] Mingxuan Sun, Baojian Huang, Xuezhi Zhang, Yangquan Chen, and
Jian-Xin Xu. Selective learning with a forgetting factor for trajectory
tracking of uncertain nonlinear systems. In Proceedings of the 2nd
Asian Control Conference, Seoul, Korea, July 1997.
220] R. Sutton. Learning to predict by the methods of temporal dierences. Machine Learning, 3:9{44, 1988.
221] S.Wang. Inversion of nonlinear dynamical systems. Technical Report, Dept. of Electrical and Computer Engineering, University of
California, Davis, California, 1988.
222] M. Togai and O. Yamano. Analysis and design of an optimal learning
control scheme for industrial robots: A discrete system approach. In
Proceedings of 24th Conference on Decision and Control, pages 1399{
1404, Ft. Lauderdale, Florida, December 1985.
223] Masayoshi Tomizuka, Tsu Chin Tsao, and Kok Kia Chew. Analysis
and synthesis of discrete-time repetitive controllers. Journal of Dynamic Systems, Measurement, and Control, 111:353{358, September
1989.
224] S. K. Tso, Y. H. Fung, and N. L. Lin. Tracking improvement for robot
manipulator control using neural-network learning. In Proceedings of
the 2nd Asian Control Conference, Seoul, Korea, July 1997.
486
K. L. Moore
487
237] Y. Wang and R. W. Longman. Use of non-causal digital signal processing in learning and repetitive control. Advances in the Astronautical Sciences, 90:649{668, 1996.
238] Chen Wen and Liu Deman. Learning control for a class of nonlinear systems with exogenous disturbances. International Journal of
Systems Science, 27(12):1453{1459, December 1996.
239] J.-X. Xu. Direct learning of control input proles with dierent time
scales. In Proceedings of the 35th IEEE Conference on Decision and
Control, Kobe, Japan, December 1996.
240] J.-X. Xu and Z. Qu. Robust learning control for a class of non-linear
systems. In Proceedings of the 35th IEEE Conference on Decision
and Control, Kobe, Japan, December 1996.
241] Jian-Xin Xu, Y. Dote, Xiaowei Wang, and Coming Shun. On the
instability of iterative learning control due to sampling delay. In
Proceedings of the 1995 IEEE IECON 31st International Conference
on Industrial Electronics, Control, and Instrumentation, volume 1,
pages 150{155, Orlando, FL, November 1995.
242] Jian-Xin Xu, Lee Tong Heng, and H. Nair. A revised iterative learning control strategy for robotic manipulators. In Proceedings 1993
Asia-Pacic Workshop on Advances in Motion Control, pages 88{93,
Singapore, July 1993.
243] Jian-Xin Xu and Yanbin Song. Direct learning control of high-order
systems for trajectories with dierent time scales. In Proceedings of
the 2nd Asian Control Conference, Seoul, Korea, July 1997.
244] Jian-Xin Xu and Yanbin Song. Direct learning control scheme with
an application to a robotic manipulator. In Proceedings of the 2nd
Asian Control Conference, Seoul, Korea, July 1997.
245] Jian-Xin Xu, Xiao-Wei Wang, and Tong Heng Lee. Analysis of
continuous iterative learning control systems using current cycle
feedback. In Proceedings of 1995 American Control Conference ACC'95, volume 6, pages 4221{4225, Seattle, WA, June 1995.
246] Jian-Xin Xu and Tao Zhu. Direct learning control of trajectory tracking with dierent time and magnitude scales for a class of nonlinear
uncertain systems. In Proceedings of the 2nd Asian Control Conference, Seoul, Korea, July 1997.
247] W. Xu, C. Shao, and T Chai. Learning control for a class of constrained mechanical systems with uncertain disturbances. In Proceedings of the 1997 American Control Conference, Albuquerque, NM,
June 1997.
488
K. L. Moore