Model-Reference Adaptive Control: Nhan T. Nguyen
Model-Reference Adaptive Control: Nhan T. Nguyen
Nhan T. Nguyen
Model-
Reference
Adaptive
Control
A Primer
Advanced Textbooks in Control and Signal
Processing
Series editors
Michael J. Grimble, Glasgow, UK
Michael A. Johnson, Oxford, UK
Linda Bushnell, Seattle, WA, USA
More information about this series at http://www.springer.com/series/4045
Nhan T. Nguyen
Model-Reference Adaptive
Control
A Primer
123
Nhan T. Nguyen
Intelligent Systems
NASA Ames Research Center
Moffett Field, CA
USA
MATLAB® and Simulink® are registered trademarks of The MathWorks, Inc., 1 Apple Hill Drive,
Natick, MA 01760-2098, USA, http://www.mathworks.com.
vii
viii Series Editors’ Foreword
This textbook is developed from a series of lecture notes for graduate courses that the
author has taught. The textbook includes new research materials that the author
developed during his research at NASA Ames Research Center. The intended readers
of this book are masters-level graduate students and beginning doctoral students. The
author notes that there exist many excellent advanced adaptive control textbooks that
provide in-depth rigorous mathematical control theory aimed at doctoral and post-
doctoral researchers. This textbook is aimed at applied control theory and applica-
tions intended to provide the readers a sufficient working understanding of adaptive
control theory through applications. Examples and problem sets are used to reinforce
the materials and aid the understanding through applications of adaptive control
techniques. A solutions manual for many of the problems, available free of charge to
instructors who adopt this textbook for use in teaching their courses, can be
downloaded from http://www.springer.com/book/9783319563923. During the
course of teaching, the author feels that such a textbook could help beginning
graduate students to better understand and appreciate the richness of adaptive control
without feeling being overwhelmed by the mathematics of real analysis required in
adaptive control theory. It is with this objective in mind that the author decided to
compose this manuscript.
Adaptive control is a well-researched subject, and as such, the subject is enriched
by a voluminous body of research literature. Many new advancements in adaptive
control theory are still being developed in the present time. The textbook does not
attempt to incorporate all of the latest developments in adaptive control theory, for
this would require an extensive endeavor which would be exceedingly beyond the
scope of this textbook. Rather, the book tries to provide a foundation in the
model-reference adaptive control with basic well-accepted adaptive control tech-
niques. Parameter estimation by least-squares techniques and uncertainty approxi-
mation by neural networks are covered. The second half of the book is devoted to
the subject of robust adaptive control. Robustness issues with model-reference
adaptive control are discussed. Standard methods to improve robustness are covered
in the remainder of the second half of the textbook. These include well-established
methods, such as the dead zone, projection, r modification, and e modification.
ix
x Preface
More recent robust adaptive control methods are also covered. These include the
author’s work in optimal control modification and bi-objective optimal control
modification, as well as two recent well-known adaptive control methods: the
adaptive loop recovery and L1 adaptive control. The author well recognizes that this
is by no means a complete exposition of all the latest advancements in robust
adaptive control theory. The readers will be referred to a list of references provided
in the bibliography for other adaptive control methods. Finally, this book includes a
chapter on applications. In particular, the applications on adaptive flight control are
exhibited from the author’s body of research in this field.
Acknowledgements
1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ... 1
1.1 Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ... 2
1.2 Verification and Validation Challenges for Adaptive Flight
Control Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ... 6
1.2.1 Verification by Simulation of Adaptive Flight
Control Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
1.2.2 Adaptive Control Metrics . . . . . . . . . . . . . . . . . . . . . 8
1.3 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
2 Nonlinear Systems . . . . . . . . . . . . . . . ......... . . . . . . . . . . . . . . 17
2.1 Equilibrium and Linearization . ......... . . . . . . . . . . . . . . 21
2.2 Local Stability and Phase Plane Analysis . . . . . . . . . . . . . . . . 24
2.3 Other Nonlinear Behaviors . . . ......... . . . . . . . . . . . . . . 27
2.4 Summary . . . . . . . . . . . . . . . . ......... . . . . . . . . . . . . . . 28
2.5 Exercises . . . . . . . . . . . . . . . . ......... . . . . . . . . . . . . . . 29
Reference . . . . . . . . . . . . . . . . . . . . . . ......... . . . . . . . . . . . . . . 30
3 Mathematical Preliminaries . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
3.1 Vector and Matrix Norms . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
3.1.1 Vector Norms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
3.1.2 Matrix Norms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35
3.2 Compact Set . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37
3.3 Existence and Uniqueness . . . . . . . . . . . . . . . . . . . . . . . . . . . 38
3.3.1 Cauchy Theorem . . . . . . . . . . . . . . . . . . . . . . . . . . . 38
3.3.2 Global Lipschitz Condition . . . . . . . . . . . . . . . . . . . . 39
3.3.3 Local Lipschitz Condition . . . . . . . . . . . . . . . . . . . . . 40
xi
xii Contents
Dr. Nhan Nguyen is a senior research scientist at NASA Ames Research Center
with over 31 years of experience. He is the technical group lead of the Advanced
Control and Evolvable Systems research group in the Intelligent Systems Division
with 22 researchers conducting aerospace flight control research. He is currently
leading a project team to conduct research and development of advanced aeroser-
voelastic control technologies for flexible aircraft. He served as the Project Scientist
in the NASA Integrated Resilient Aircraft Control (IRAC) Project from 2007 to
2010 with an annual budget of $21 million that funded advanced research in
adaptive control conducted by NASA, industry, and academia. His prior position
was the Deputy Principal Investigator of the IRAC project from 2005 to 2006. His
work on optimal control modification adaptive control culminated in a successful
flight validation experiment on NASA F-18 research aircraft in 2010. He has
received several prestigious awards including the NASA Exceptional Scientific
Achievement Medal, NASA Exceptional Achievement Medal, two-time NASA
Ames Honors for Excellence as Engineer, and more than 50 other NASA career
awards. He has four US patents and two other patents in pending. He has over 15
NASA invention disclosures. He has published over 200 peer-reviewed technical
publications. He is the chair of the American Institute of Aeronautics and
Astronautics (AIAA) Intelligent Systems Technical Committee with 50 profes-
sional members. He has received two AIAA Distinguished Services Awards.
xvii
Chapter 1
Introduction
Abstract This chapter provides a brief summary of the recent advancements made in
model-reference adaptive control theory. Adaptive control is a promising technology
that can improve the performance of control systems in the presence of uncertainty
due to a variety of factors such as degradation and modeling uncertainty. During the
past decade, the adaptive control research community has produced several advance-
ments in adaptive control theory along with many novel adaptive control methods
as a result of the increased government research funding. Many of these new adap-
tive control methods have added new capabilities in terms of improved performance
and robustness that further increase the viability of model-reference adaptive control
as a future technology. Flight test validation of adaptive control on full-scale air-
craft and unmanned arial vehicles has increased the confidence in model-reference
adaptive control as a possible new flight control technology for aerospace vehicles
in the near future. In spite of the five decades of research in adaptive control, the
fact still remains that currently no adaptive control system has ever been deployed
on any safety-critical or human-rated production systems. Many technical problems
remain unresolved. As a nonlinear control method, the lack of well-accepted met-
rics for adaptive control system design presents a major hurdle for certification. The
development of certifiable adaptive control systems is viewed as a major techni-
cal challenge for the adaptive control research community to address in the current
research.
• To recognize the fact that, in spite of the many advancements in the field, adoption
of adaptive control technologies is not widespread and its applications are limited
in scope; and
• To recognize that verification, validation, and certification are viewed as significant
technical barriers that need to be further researched in order to increase the con-
fidence in and the trustworthiness of adaptive control technologies, which would
enable future adoption.
1.1 Overview
Adaptive control is a study of a class of nonlinear control methods that deal with
systems with uncertainty. The uncertainty can be due to unforeseen internal changes
in system dynamics or external disturbances. An adaptive control system can be
broadly described as any control system that has the ability to adjust control design
parameters such as control gains online based on inputs received by the plant in
order to accommodate system uncertainty as illustrated in Fig. 1.1. The adjustable
parameters are called adaptive parameters, and the adjusting mechanism, which is
described by a set of mathematical equations, is called an adaptive law. In most
cases, a typical adaptive law is nonlinear. This nonlinearity makes adaptive control
inherently difficult to design and analyze since many traditional design and analysis
methods for linear time-invariant (LTI) control systems such as bode plot, phase and
gain margins, eigenvalue analysis are not applicable.
Adaptive control has a long history dated back to the early 1950s when there was
an extensive research interest in designing advanced autopilots for high-performance
aircraft that operated over a wide range of flight conditions [1]. After a significant
research and development effort, gain-scheduling control was gaining acceptance
over adaptive control due to the inherent difficulty associated with nonlinearity of
adaptive control. On the other hand, gain-scheduling control can take advantage of
classical control techniques in the design selection of suitable control gains that are
scheduled as a function of aircraft operating conditions.
The 1960s saw the emergence of the modern control theory and Lyapunov stabil-
ity theory that also contributed to adaptive control theory. Whitaker et al. developed
model-reference adaptive control with the sensitivity method and MIT rule. During
this development phase, there was a lack of stability proofs and good understand-
ing of the properties of adaptive control. In 1967, an adaptive flight controller was
flight-tested by NASA on one of the three X-15 experimental hypersonic aircrafts
[2, 3]. Several successful test flights were made before an ill-fated test flight occurred
that resulted in a fatal crash of one of these vehicles. This event and the technical
difficulties led to a diminished interest in adaptive control.
1.1 Overview 3
In the 1970s, the Lyapunov stability theory was established as a foundation for
model-reference adaptive control. This represented a breakthrough in adaptive con-
trol as it led to the development of improved model-reference adaptive control
schemes with the accompanied Lyapunov stability theory. The excitements soon
gave way to a realization in the 1980s that, in spite of the Lyapunov stability theory,
adaptive control can exhibit unstable behaviors in the presence of small disturbances
or unmodeled dynamics [4]. This led to the recognition that model-reference adap-
tive control was sensitive to system modeling accuracy and the resultant mismatch
between a real system and a plant model of the system. This lack of robustness
led to the development of the σ modification [5] and e modification to improve the
stability properties of adaptive control [6]. These robust modification schemes rep-
resent a new class of “robust adaptive control.”
In the 1990s and up to the present time, research in adaptive control remains
highly active. Neural networks were introduced as a mechanism for adaptation
[7–12]. This led to the development of a class of adaptive control called “intelli-
gent control” or “neural network adaptive control” which utilizes neural networks to
approximate model uncertainties, but the basic framework remains to be the same as
model-reference adaptive control [13–21].
The following decade saw a resurgence of interest in adaptive control research.
This also coincided with a period of increased research funding made possible by
NASA [22] and other US government agencies. In particular, NASA has been an
active key player in adaptive control technology development as illustrated on the
time line in Fig. 1.2. During this period, the adaptive control research community
produced several advancements in adaptive control theory along with many novel
adaptive control methods as a result of the increased research funding. The list of
these new advancements is long, and it would not do justice to include all of these
new methods in this limited space. The reader is referred to a subset of this list from
which those novel adaptive control methods that were funded by NASA during this
decade can be found. These methods include: adaptive control based on retrospective
cost optimization by Santillo and Bernstein [23, 24], adaptive control with reference
model modification by Stepanyan and Krishnakumar [25, 26], adaptive loop recov-
4 1 Introduction
ery by Calise and Yucelen [27, 28], bounded linear stability analysis metric-driven
adaptive control by Nguyen [29, 30], combined/composite model-reference adaptive
control by Lavretsky [31, 32], derivative-free model-reference adaptive control by
Yucelen and Calise [33, 34], hybrid adaptive control by Nguyen [35–37], K modifica-
tion by Kim et al. [38], Kalman filter modification by Yucelen and Calise [39, 40], L1
adaptive control by Hovakimyan and Cao [41–43], least-squares adaptive control by
Nguyen [44, 45], concurrent learning least-squares adaptive control by Chowdhary
and Johnson [46, 47], modified state observer adaptive control by Balaskrishnan [48],
multivariable model-reference adaptive control by Guo and Tao [49], optimal control
modification [50, 51] and multiobjective optimal control modification by Nguyen
[52, 53], Q modification by Volyanskyy et al. [54, 55], and parameter-dependent
Riccati equation adaptive control by Kim et al. [56]. Many of these new adaptive
control methods have added new capabilities in terms of improved performance and
robustness that further increase the viability of model-reference adaptive control as
a future technology.
control was also evaluated for fight testing on a Beechcraft Bonanza fly-by-wire
testbed in 2010 [67]. These flight experiments and many others that followed have
increased the confidence in model-reference adaptive control as a possible new flight
control technology for aerospace vehicles in the near future. They also revealed that
further flight validation for adaptive control is still needed to mature this technology.
Research in adaptive control continues at the present time. It would be well beyond
the scope of this book to include all current research activities. The interested reader is
referred to the references for more information on numerous applications for aircraft
[68–79], spacecraft [21, 80–83], UAVs [15, 47, 65, 66, 84], aerospace structures
[85, 86], robotic systems [18, 87], munition systems [88], fluid systems [89], and
many others.
The current research in adaptive control generally lacks the ability to deal with
integrated effects that exist in many different system designs and operations. These
include but are not limited to: complex uncertainties due to unmodeled dynamics
[4, 90], gross changes in system dynamics resulting from unintended operations and
structural damage [36, 76, 91], unknown component failures and anomalies [80,
92–94], high degree of design complexity [59], novel actuators and sensors [79, 95],
multiphysics interactions [85, 86, 96–98], and many other effects.
Adaptive control plays a major role in the aerospace domain. In aerospace vehicles
with damaged structures or failed flight control surfaces or operating in off-nominal
flight events, the vehicles can exhibit numerous coupled effects such as aerodynam-
ics, vehicle dynamics, structural dynamics, and propulsion. These coupled effects
can impose a wide range of uncertainties on the performance of a flight control sys-
tem. Thus, even though an adaptive control system may be shown to be stable in a
nominal flight condition, it may fail to provide enough stability in the presence of
these uncertainties [99, 100]. For example, conventional aircraft flight control sys-
tems incorporate aeroservoelastic (ASE) notch filters to prevent control signals from
exciting aeroelastic wing modes. However, if changes in aircraft dynamics are signif-
icant, the frequencies of the aeroelastic modes may change enough to render the ASE
notch filters ineffective. This could cause the control signals to excite wing modes
which could lead to problems for a pilot to control an aircraft. Another problem
faced by adaptive control is the ability of adaptive control systems to accommodate
slow or degraded flight control actuators such an impaired flight control surface or
engine as a flight control effector [79, 101]. The dissimilar actuator rates due to slow
actuator dynamics can cause problems with adaptive control and can potentially lead
to pilot-induced oscillations (PIO) [102].
To adequately deal with these coupled effects, integrated approaches in adaptive
control research need to be developed. These integrated approaches will be required to
develop new fundamental multidisciplinary methods in adaptive control and system
modeling. Unmodeled dynamics are a significant source of uncertainty and a causal
factor of instability of adaptive control systems in high-gain adaptation situations.
Future research in adaptive control should leverage fundamental understanding of the
structures of these secondary dynamics by incorporating multidisciplinary methods
into the design of an adaptive control system. With a better understanding of the
6 1 Introduction
In spite of the five decades of research in adaptive control, the fact still remains that
currently no adaptive control system has ever been deployed on any safety-critical
or human-rated production systems such as passenger transport aircraft [99, 103–
106]. It should be noted that adaptive control has been successfully developed and
deployed in munition production systems [88]. The problem lies in the difficulty
with the certification of adaptive control systems since existing certification methods
for linear time-invariant (LTI) control systems cannot readily be used for nonlinear
adaptive control systems. Research to address the notion of metrics for adaptive
control began to appear during the first decade of the twenty-first century [14, 107–
115]. The goal of this research is to develop performance and stability metrics for
adaptive control similar to overshoot, settling time, phase and gain margins for LTI
control systems. These metrics, if accepted, could pave a path toward certification
that would potentially lead to the adoption of adaptive control as a future control
technology for safety-critical and human-rated production systems.
Development of certifiable adaptive control systems represents a major challenge
to overcome. Adaptive control systems with learning algorithms will never become
part of the future unless it can be proven that they are highly safe and reliable.
Rigorous methods for adaptive control software verification and validation must
therefore be developed to ensure that adaptive control system software failures will
not occur to verify that the adaptive control system functions as required, to eliminate
unintended functionality, and to demonstrate that certification requirements imposed
by regulatory bodies such as the Federal Aviation Administration (FAA) can be
satisfied [104, 105].
The ability of an adaptive control system to modify a pre-designed flight control
system is a strength and a weakness at the same time. On the one hand, the abil-
ity of adaptive control systems to accommodate system degradation is considered
as a major advantage since traditional gain-scheduling control methods tend to be
less capable of handling off-nominal flight conditions outside their design operating
envelopes. On the other hand, serious problems with adaptive control can still ex-
ist with regard to unmodeled dynamics and high-gain adaptation. Adaptive control
systems are sensitive to these potential problems as well as many others including
actuator dynamics and exogenous disturbances. To be certifiable, an adaptive flight
control system must be able to prove that these effects as well as other factors such
as time delay, system constraints, and measurement noise are handled in a globally
satisfactory manner.
1.2 Verification and Validation Challenges for Adaptive Flight Control Systems 7
Simulation is an integral part of the verification of adaptive control systems [104, 105,
116, 117]. Many aspects of adaptive control systems, in particular convergence and
stability, can only be analyzed with simulations that provide enough detail and fidelity
to model significant nonlinear dynamics. For example, stall upset of an aircraft cannot
be expressed as a linear model since this effect is highly nonlinear and unsteady.
Simulation provides a fairly rapid way to accomplish the following tasks:
• Evaluation and comparison of different adaptive control algorithms;
• Tuning control gains and weight update laws;
• Determination of progress of adaptation at each step;
• Evaluation of the effects of process and measurement noise on convergence of
adaptive parameters;
• Determination of stability boundaries;
• Validation with actual flight computer hardware;
• Conducting piloted evaluation of adaptive control in a flight simulator;
• Simulating ad hoc techniques of improving the adaptation process, such as adding
persistent excitation to improve identification and convergence, or stopping the
adaptation process after the tracking error converges to within a specified tolerance
or after a specified number of iterations.
Simulations differ primarily in the fidelity with which the plant is modeled. Higher
fidelity simulations require more complicated mathematical models of the adaptive
control system and also a greater use of actual (and expensive) controller hardware.
The behavior of simple linear models is compared to that of higher fidelity nonlinear
models when they are available to ensure that analyses performed using the linear
models still apply. In order to be cost effective, the lowest fidelity testbed is usually
used as much as possible.
The lowest fidelity simulations are usually run on a desktop computer. This simu-
lation typically includes the control laws and a linear or nonlinear plant of the system
dynamics. A linear model is most often used in early control law design and analysis
or to calculate linear gain and phase margins. Changes to the plant model can be
simulated by changing the system transfer function from one matrix to another with
varying frequency. By varying the amount of change, the stability boundaries of the
system can be determined. Concomitant with this process is an evaluation of the sys-
tem tuning parameters that are used in the adaptive control algorithms. The desktop
simulation environment provides a quick way to compare different adaptive control
algorithms and controller architectures. Only the most promising designs need to be
simulated using higher fidelity simulations.
Higher fidelity simulation testbeds use actual flight hardware (or even aircraft) in
the simulation of the control loop and are often run in dedicated computing environ-
ments with a cockpit and out-the-window graphics [117, 118]. These simulations
may include a cockpit to interface with the pilot and can either be fixed-base or
8 1 Introduction
In spite of the many advancements made in adaptive control research and the poten-
tial benefits of adaptive control systems, the absence of verification and validation
methods for adaptive control systems remains a major hurdle to the implementation
of adaptive control in safety-critical or human-rated production systems. This hurdle
can be traced to the lack of performance and stability metrics for adaptive control.
The development of verifiable metrics for adaptive control is an important aspect
of adaptive control research in order to mature adaptive control technology for use
in future safety-critical and human-rated production systems. Stability metrics of
adaptive control are an important consideration for assessing system robustness to
unmodeled dynamics, time delay, high-gain learning, and exogenous disturbances.
Therefore, it is imperative to define a proper set of stability and performance metrics
for adaptive systems as a first step in the development of reliable verification and
validation methods that could lead to the certification of adaptive control software.
Another potential benefit of metrics for adaptive control is the consideration for
metrics-driven adaptive control. Metrics-driven adaptive control is the notion that
adaptation in certain cases should be driven by the need to trade off between stability
and performance to maintain operational safety of a control system [30]. Research in
this area has resulted in some initial analysis methods for computing stability metrics
online that could be used to adjust the adaptive parameters of an adaptive control
system to improve stability margins of the closed-loop system [29].
Generally, the classical gain and phase margins for LTI systems are not applicable
to adaptive control systems which are nonlinear. Some results began to appear in the
literature on the definition and evaluation of candidate metrics for adaptive control
systems [14, 109, 110, 112, 115]. The use of a parameter sensitivity as a metric for
the output of neural networks has been investigated [103]. Stability metrics based
on the Lyapunov analysis and passivity theory have been studied [108]. Metrics for
evaluation of stability and robustness of adaptive control systems can also be obtained
from optimization frameworks [107, 119]. The notion of time-delay margin as a
stability metric has been studied by many authors [120, 121].
1.2 Verification and Validation Challenges for Adaptive Flight Control Systems 9
1.3 Summary
Adaptive control is a well-researched subject over the past several decades. Adap-
tive control is a promising technology that could provide improved performance of
control systems in the presence of uncertainty due to a variety of factors such as degra-
dation and modeling uncertainty. During the past decade, advancements in adaptive
control theory along with many novel adaptive control methods have been made by
researchers in the field. Many of these new adaptive control methods have added new
capabilities in terms of improved performance and robustness that further increase
the viability of model-reference adaptive control as a future technology. Flight test
validation of adaptive control on full-scale aircraft and unmanned arial vehicles has
increased the confidence in model-reference adaptive control as a possible new flight
control technology for aerospace vehicles in the near future.
In spite of this, adaptive control is not well-accepted for applications in safety-
critical or human-rated production systems. Many technical problems remain unre-
solved. As a nonlinear control method, the lack of well-accepted metrics for adaptive
control system design as opposed to linear control systems presents a major hurdle
for certification.
References
1. Åström, K. J., & Wittenmark, B. (2008). Adaptive control: Dover Publications Inc.
2. Jenkins, D. R. (2000). Hypersonics Before the Shuttle: A Concise History of the X-15 Research
Airplane, NASA SP-2000-4518.
3. Staff of the Flight Research Center. (1971). Experience with the X-15 Adaptive Flight Control
System, NASA TN D-6208.
4. Rohrs, C. E., Valavani, L., Athans, M., & Stein, G. (1985). Robustness of continuous-time
adaptive control algorithms in the presence of unmodeled dynamics. IEEE Transactions on
Automatic Control, AC-30(9), 881–889.
5. Ioannou, P., & Kokotovic, P. (1984). Instability analysis and improvement of robustness of
adaptive control. Automatica, 20(5), 583–594.
6. Narendra, K. S. & Annaswamy, A. M. (1987). A new adaptive law for robust adaptation
without persistent excitation. IEEE Transactions on Automatic Control, AC-32(2), 134–145.
10 1 Introduction
29. Nguyen, N., Bakhtiari-Nejad, M., & Huang, Y. (2007). Hybrid adaptive flight control with
bounded linear stability analysis. In AIAA Guidance, Navigation, and Control Conference,
AIAA-2007-6422, August 2007.
30. Bakhtiari-Nejad, M., Nguyen, N., & Krishnakumar, K. (2009). Adjustment of adaptive gain
with bounded linear stability analysis to improve time-delay margin for metrics-driven adap-
tive control. In AIAA Infotech@Aerospace Conference, AIAA-2009-1801, April 2009.
31. Lavretsky, E. (2009). Combined/composite model reference adaptive control. In AIAA Guid-
ance, Navigation, and Control Conference, AIAA-2009-6065, August 2009.
32. Lavretsky, E. (2009). Combined/composite model reference adaptive control. IEEE Transac-
tions on Automatic Control, 54(11), 2692–2697.
33. Yucelen, T., & Calise, A. J. (2010). Derivative-free model reference adaptive control. In AIAA
Guidance, Navigation, and [11, 5] Control Conference, AIAA-2009-5858, August 2010.
34. Yucelen, T., & Calise, A. J. (2011). Derivative-free model reference adaptive control. AIAA
Journal of Guidance, Control, and Dynamics, 34(4), 933–950.
35. Nguyen, N., Krishnakumar, K., Kaneshige, J., & Nespeca, P. (2006). Dynamics and adaptive
control for stability recovery of damaged asymmetric aircraft. In AIAA Guidance, Navigation,
and Control Conference, AIAA-2006-6049, August 2006.
36. Nguyen, N., Krishnakumar, K., Kaneshige, J., & Nespeca, P. (2008). Flight dynamics model-
ing and hybrid adaptive control of damaged asymmetric aircraft. AIAA Journal of Guidance,
Control, and Dynamics, 31(3), 751–764.
37. Nguyen, N. (2011). Hybrid adaptive flight control with model inversion adaptation. In Ad-
vances in Flight Control Systems. Croatia: Intech Publishing. ISBN 978-953-307-218-0.
38. Kim, K., Yucelen, T., & Calise, A. J. (2010). K-modification in adaptive control. In AIAA
Infotech@Aerospace Conference, AIAA-2010-3321, April 2010.
39. Yucelen, T., & Calise, A. J. (2009). A Kalman filter optimization approach to direct adaptive
control. In AIAA Guidance, Navigation, and Control Conference, AIAA-2010-7769, August
2009.
40. Yucelen, T., & Calise, A. J. (2010). A Kalman filter modification in adaptive control. AIAA
Journal of Guidance, Control, and Dynamics, 33(2), 426–439.
41. Cao, C., & Hovakimyan, N. (2007). Guaranteed transient performance with L1 adaptive
controller for systems with unknown time-varying parameters and bounded disturbances:
Part I. In American Control Conference, July 2007.
42. Cao, C., & Hovakimyan, N. (2008). Design and analysis of a novel L1 adaptive control
architecture with guaranteed transient performance. IEEE Transactions on Automatic Control,
53(2), 586–591.
43. Hovakimyan, N., & Cao, C. (2010). L1 Adaptive control theory: Guaranteed robustness with
fast adaptation. Society for Industrial and Applied Mathematics.
44. Nguyen, N., Burken, J., & Ishihara, A. (2011). Least-squares adaptive control using Chebyshev
orthogonal polynomials. In AIAA Infotech@Aerospace Conference, AIAA-2011-1402, March
2011.
45. Nguyen, N. (2013). Least-squares model reference adaptive control with Chebyshev orthog-
onal polynomial approximation. AIAA Journal of Aerospace Information Systems, 10(6),
268–286.
46. Chowdhary, G., & Johnson, E. (2010). Least squares based modification for adaptive control.
In IEEE Conference on Decision and Control, December 2010.
47. Chowdhary, G., & Johnson, E. (2011). Theory and flight-test validation of a concurrent-
learning adaptive controller. AIAA Journal of Guidance, Control, and Dynamics, 34(2), 592–
607.
48. Balakrishnan, S. N., Unnikrishnan, N., Nguyen, N., & Krishnakumar, K. (2009). Neuroad-
aptive model following controller design for non-affine and non- square aircraft systems. In
AIAA Guidance, Navigation, and Control Conference, AIAA-2009-5737, August 2009.
49. Guo, J., & Tao, G. (20011). A Multivariable MRAC design for aircraft systems under failure
and damage conditions. In American Control Conference, June 2011.
12 1 Introduction
50. Nguyen, N., Krishnakumar, K., & Boskovic, J. (2008). An optimal control modification to
model-reference adaptive control for fast adaptation. In AIAA Guidance, Navigation, and
Control Conference, AIAA 2008-7283, August 2008.
51. Nguyen, N. (2012). Optimal control modification for robust adaptive control with large adap-
tive gain. Systems and Control Letters, 61(2012), 485–494.
52. Nguyen, N. (2014). Multi-objective optimal control modification adaptive control method for
systems with input and unmatched uncertainties. In AIAA Guidance, Navigation, and Control
Conference, AIAA-2014-0454, January 2014.
53. Nguyen, N., & Balakrishnan, S. N. (2014). Bi-objective optimal control modification adaptive
control for systems with input uncertainty. IEEE/CAA Journal of Automatica Sinica, 1(4),
423–434.
54. Volyanskyy, K. Y., & Calise, A. J. (2006). A novel q-modification term for adaptive control.
In American Control Conference, June 2006.
55. Volyanskyy, K. Y., Haddad, W. M., & Calise, A. J. (2009). A new neuroadaptive control
architecture for nonlinear uncertain dynamical systems: Beyond σ - and e-modifications. IEEE
Transactions on Neural Networks, 20, 1707–1723.
56. Kim, K., Yucelen, T., Calise, A.J., & Nguyen, N. (2011). Adaptive output feedback control
for an aeroelastic generic transport model: A parameter dependent Riccati equation approach.
In AIAA Guidance, Navigation, and Control Conference, AIAA-2011-6456, August 2011.
57. Bosworth, J., & Williams-Hayes, P. S. (2007). Flight test results from the NF-15B IFCS project
with adaptation to a simulated stabilator failure. In AIAA Infotech@Aerospace Conference,
AIAA-2007-2818, May 2007.
58. Williams-Hayes, P. S. Flight test implementation of a second generation intelligent flight
control system. In NASA TM-2005-213669.
59. Hanson, C., Johnson, M., Schaefer, J., Nguyen, N., & Burken, J. (2011). Handling qualities
evaluations of low complexity model reference adaptive controllers for reduced pitch and roll
damping scenarios. In AIAA Guidance, Navigation, and Control Conference, AIAA-2011-
6607, August 2011.
60. Hanson, C., Schaefer, J., Johnson, M., & Nguyen, N. Design of low complexity model refer-
ence adaptive controllers. In NASA-TM-215972.
61. Nguyen, N., Hanson, C., Burken, J., & Schaefer, J. (2016). Normalized optimal control modi-
fication and flight experiments on NASA F/A-18 aircraft. AIAA Journal of Guidance, Control,
and Dynamics.
62. Schaefer, J., Hanson , C., Johnson, M., & Nguyen, N. (2011). Handling qualities of model
reference adaptive controllers with varying complexity for pitch-roll coupled failures. In IAA
Guidance, Navigation, and Control Conference, AIAA-2011-6453, August 2011.
63. Campbell, S., Kaneshige, J., Nguyen, N., & Krishnakumar, K. (2010). An adaptive control
simulation study using pilot handling qualities evaluations. In AIAA Guidance, Navigation,
and Control Conference, AIAA-2010-8013, August 2010.
64. Campbell, S., Kaneshige, J., Nguyen, N., & Krishnakumar, K. (2010). Implementation and
evaluation of multiple adaptive control technologies for a generic transport aircraft simulation.
In AIAA Infotech@Aerospace Conference, AIAA-2010-3322, April 2010.
65. Gregory, I. M, Cao, C., Xargay, E., Hovakimyan, N., & Zou, X. (2009). L1 adaptive control
design for NASA AirSTAR flight test vehicle. In AIAA Guidance, Navigation, and Control
Conference, AIAA-2009-5738, August 2009.
66. Kitsios, I., Dobrokhodov, V., Kaminer, I., Jones, K., Xargay, E., Hovakimyan, N., Cao, C.,
Lizarraga, M., Gregory, I., Nguyen, N., & Krishnakumar, K. (2009). Experimental validation
of a metrics driven L1 adaptive control in the presence of generalized unmodeled dynamics.
In AIAA Guidance, Navigation, and Control Conference, AIAA-2009-6188, August 2009.
67. Steck, J., Lemon, K., Hinson, B., Kimball, D., & Nguyen, N. (2010). Model reference adap-
tive fight control adapted for general aviation: Controller gain simulation and preliminary
flight testing on a bonanza fly-by-wire Testbed. In AIAA Guidance, Navigation, and Control
Conference, AIAA-2010-8278, August 2010.
References 13
68. Bhattacharyya, S., Krishnakumar, K., & Nguyen, N. (2012). Adaptive autopilot designs for
improved tracking and stability. In AIAA Infotech@Aerospace Conference, AIAA-2012-2494,
June 2012.
69. Burken, J., Nguyen, N., & Griffin, B. (2010). Adaptive flight control design with optimal
control modification for F- 18 aircraft model. In AIAA Infotech@Aerospace Conference,
AIAA-2010-3364, April 2010.
70. Campbell, S., Nguyen, N., Kaneshige, J., & Krishnakumar, K. (2009). Parameter estimation
for a hybrid adaptive flight controller. In AIAA Infotech@Aerospace Conference, AIAA-2009-
1803, April 2009.
71. Chen, S., Yang, Y., Balakrishnan, S. N., Nguyen, N., & Krishnakumar, K. (2009). SNAC
convergence and use in adaptive autopilot design. In International Joint Conference on Neural
Networks, June 2009.
72. Eberhart, R. L., & Ward, D. G. (1999). Indirect adaptive flight control system interactions.
International Journal of Robust and Nonlinear Control, 9, 1013–1031.
73. Hinson, B., Steck, J., Rokhsaz, K., & Nguyen, N. (2011). Adaptive control of an elastic general
aviation aircraft. In AIAA Guidance, Navigation, and Control Conference, AIAA-2011-6560,
August 2011.
74. Lemon, K., Steck, J., Hinson, B., Rokhsaz, K., & Nguyen, N. (2011). Application of a six
degree of freedom adaptive controller to a general aviation aircraft. In AIAA Guidance, Nav-
igation, and Control Conference, AIAA-2011-6562, August 2011.
75. Nguyen, N., Tuzcu, I., Yucelen, T., & Calise, A. (2011). Longitudinal dynamics and adaptive
control application for an aeroelastic generic transport model. In AIAA Atmospheric Flight
Mechanics Conference, AIAA-2011-6319, August 2011.
76. Rajagopal, K., Balakrishnan, S.N., Nguyen, N., & Krishnakumar, K. (2010). Robust adap-
tive control of a structurally damaged aircraft. In AIAA Guidance, Navigation, and Control
Conference, AIAA-2010-8012, August 2010.
77. Reed, S., Steck, J., & Nguyen, N. (2011). Demonstration of the optimal control modification
for general aviation: Design and simulation. In AIAA Guidance, Navigation, and Control
Conference, AIAA-2011-6254, August 2011.
78. Reed, S., Steck, J., & Nguyen, N. (2014). Demonstration of the optimal control modification
for 6-DOF control of a general aviation aircraft. In AIAA Guidance, Navigation, and Control
Conference, AIAA-2014-1292, January 2014.
79. Stepanyan, V., , Nguyen, N., & Krishnakumar, K. (2009). Adaptive control of a transport
aircraft using differential thrust. In AIAA Guidance, Navigation, and Control Conference,
AIAA-2009-5741, August 2009.
80. Boskovic, J., Jackson, J., Mehra, R., & Nguyen, N. (2009). Multiple-model adaptive fault-
tolerant control of a planetary lander. AIAA Journal of Guidance, Control, and Dynamics,
32(6), 1812–1826.
81. Lam, Q., & Nguyen, N. (2013). Pointing control accuracy and robustness enhancement of
an optical payload system using a direct adaptive control combined with an optimal control
modification. In AIAA Infotech@Aerospace Conference, AIAA-2013-5041, August 2013.
82. Lefevre B., & Jha, R. (2009). Hybrid adaptive launch vehicle ascent flight control. In AIAA
Guidance, Navigation, and Control Conference, AIAA-2009-5958, August 2009.
83. Swei, S., & Nguyen, N. (2016). Adaptive estimation of disturbance torque for orbiting space-
craft using recursive least squares method. In AIAA Infotech@Aerospace Conference, AIAA-
2016-0399, January 2016.
84. Hovakimyan, N., Kim, N., Calise, A. J., Prasad, J. V. R., & Corban, E. J. (2001). Adaptive
output feedback for high-bandwidth control of an unmanned helicopter. In AIAA Guidance,
Navigation and Control Conference, AIAA-2001-4181, August 2001.
85. Nguyen, N., Swei, S., & Ting, E. (2015). Adaptive linear quadratic gaussian optimal control
modification for flutter suppression of adaptive wing. In AIAA Infotech@Aerospace Confer-
ence, AIAA 2015-0118, January 2015.
86. Kumar, M., Rajagopal, K., Balakrishnan, S. N., & Nguyen, N. (2014). Reinforcement learning
based controller synthesis for flexible aircraft wings. IEEE/CAA Journal of Automatica Sinica,
1(4), 435–448.
14 1 Introduction
87. Ishihara, A., Al-Ali, K., Kulkarni, N., & Nguyen, N. (2009). Modeling Error Driven Robot
Control. In AIAA Infotech@Aerospace Conference, AIAA-2009-1974, April 2009.
88. Sharma, M., Calise, A., & Corban, J. E. (2000). Application of an adaptive autopilot design
to a family of guided munitions. In AIAA Guidance, Navigation, and Control Conference,
AIAA-2000-3969, August 2000.
89. Nguyen, N., Bright, M., & Culley, D. (2007). Adaptive feedback optimal control of flow
separation on stators by air injection. AIAA Journal, 45(6),
90. Nguyen, N. (2013). Adaptive control for linear uncertain systems with unmodeled dynam-
ics revisited via optimal control modification. In AIAA Guidance, Navigation, and Control
Conference, AIAA-2013-4988, August 2013.
91. Lemaignan, B. (2005). Flying with no flight controls: handling qualities analyses of the
baghdad event. In AIAA Atmospheric Flight Mechanics Conference, AIAA-2005-5907, August
2005.
92. Nguyen, N. (2012). Bi-objective optimal control modification adaptive control for systems
with input uncertainty. In AIAA Guidance, Navigation, and Control Conference, AIAA-2012-
4615, August 2012.
93. Nguyen, N. (2010). Optimal control modification adaptive law for time-scale separated sys-
tems. In American Control Conference, June 2010.
94. Stepanyan, V., & Nguyen, N. (2009). Control of systems with slow actuators using time scale
separation. In AIAA Guidance, Navigation, and Control Conference, AIAA-2009-6272, August
2009.
95. Nguyen, N., & Stepanyan, V. (2010). Flight-propulsion response requirements for directional
stability and control. In AIAA Infotech@Aerospace Conference, AIAA-2010-3471, April 2010.
96. Tuzcu, I., & Nguyen, N. (2009). Aeroelastic modeling and adaptive control of generic transport
model. In AIAA Atmospheric Flight Mechanics, AIAA-2010-7503, August 2010.
97. Tuzcu, I., & Nguyen, N. (2010). Modeling and control of generic transport model. In 51st
AIAA/ASME/ASCE/AHS/ASC Structures, Structural Dynamics, and Materials Conference,
AIAA-2010-2622, April 2010.
98. Yucelen, T., Kim, K., Calise, A., & Nguyen, N. (2011). Derivative-free output feedback
adaptive control of an aeroelastic generic transport model. In AIAA Guidance, Navigation,
and Control Conference, AIAA-2011-6454, August 2011.
99. Nguyen, N., & Jacklin, S. (2010). Stability, convergence, and verification and validation
challenges of neural net adaptive flight control. In Applications of Neural Networks in High
Assurance Systems Studies in Computational Intelligence (Vol. 268, pp. 77–110). Berlin:
Springer.
100. Nguyen, N., & Jacklin, S. (2007). Neural net adaptive flight control stability, verification
and validation challenges, and future research. In International Joint Conference on Neural
Networks, August 2007.
101. Nguyen, N., Ishihara, A., Stepanyan, V., & Boskovic, J. (2009). Optimal control modification
for robust adaptation of singularly perturbed systems with slow actuators. In AIAA Guidance,
Navigation, and Control Conference, AIAA-2009-5615, August 2009.
102. Gilbreath, G. P. (2001). Prediction of Pilot-Induced Oscillations (PIO) due to Actuator Rate
Limiting Using the Open-Loop Onset Point (OLOP) Criterion. In M.S. Thesis, Air Force
Institute of Technology, Wright-Patterson Air Force Base, Ohio, 2001.
103. Schumann, J., & Liu, Y. (2007). Tools and methods for the verification and validation of
adaptive aircraft control systems. In IEEE Aerospace Conference, March 2007.
104. Jacklin, S. (2009). Closing the certification gaps in adaptive flight control software. In AIAA
Guidance, Navigation, and Control Conference, AIAA-2008-6988, August 2008.
105. Jacklin, S. A., Schumann, J. M., Gupta, P. P., Richard, R., Guenther, K., & Soares, F. (2005).
Development of advanced verification and validation procedures and tools for the certification
of learning systems in aerospace applications. In AIAA Infotech@aerospace Conference,
September 2005.
106. Jacklin, S. A. (2015). Survey of verification and validation techniques for small satellite
software development. In Space Tech Expo Conference, May 2015.
References 15
107. Crespo, L. G., Kenny, S. P., & Giesy, D. P. (2008). Figures of merit for control verification.
In AIAA Guidance, Navigation, and Control Conference, AIAA 2008-6339, August 2008.
108. Hodel, A. S., Whorton, M., & Zhu, J. J. (2008). Stability metrics for simulation and flight-
software assessment and monitoring of adaptive control assist compensators. In AIAA Guid-
ance, Navigation, and Control Conference, AIAA 2008-7005, August 2008.
109. Ishihara, A., Ben-Menahem, S., & Nguyen, N. (2009). Time delay margin computation via
the Razumikhin method for an adaptive control system. In AIAA Guidance, Navigation, and
Control Conference, AIAA-2009-5969, August 2009.
110. Ishihara, A., Ben-Menahem, S., & Nguyen, N. (2009). Time delay margin estimation for direct
adaptive control of a linear system. In IASTED International Conference on Identification,
Control and Applications, August 2009.
111. Ishihara, A., Nguyen, N., & Stepanyan, V. (2010). Time delay margin estimation for adaptive
outer- loop longitudinal aircraft control. In AIAA Infotech@Aerospace Conference, AIAA-
2010-3455, April 2010.
112. Nguyen, N., & Summers, E. (2011). On time delay margin estimation for adaptive control and
robust modification adaptive laws. In AIAA Guidance, Navigation, and Control Conference,
AIAA-2011-6438, August 2011.
113. Nguyen, N., & Boskovic, J. (2008). Bounded linear stability margin analysis of nonlinear
hybrid adaptive control. In American Control Conference, June 2008.
114. Rajagopal, K., Balakrishnan, S. N., Nguyen, N., & Krishnakumar, K. (2013). Time delay mar-
gin analysis of modified state observer based adaptive control. In AIAA Guidance, Navigation,
and Control Conference, AIAA-2013-4755, August 2013.
115. Stepanyan, V., Krishnakumar, K., Nguyen, N., & Van Eykeren, L. (2009). Stability and per-
formance metrics for adaptive flight control. In AIAA Guidance, Navigation, and Control
Conference, AIAA-2009-5965, August 2009.
116. Belcastro, C., & Belcastro, C. (2003). On the validation of safety critical aircraft systems, part I:
Analytical and simulation methods. In AIAA Guidance, Navigation, and Control Conference,
AIAA-2003-5559, August 2003.
117. Belcastro, C., & Belcastro, C. (2003). On the validation of safety critical aircraft systems,
Part II: Analytical and simulation methods. In AIAA Guidance, Navigation, and Control
Conference, AIAA-2003-5560, August 2003.
118. Duke, E. L., Brumbaugh, R. W., & Disbrow, D. (1989). A rapid prototyping facility for flight
research in advanced systems concepts. In IEEE Computer, May 1989.
119. Matsutani, M., Jang, J., Annaswamy, A., Crespo, L. G., Kenny, S. P. (2008). An adaptive
control technology for safety of a GTM-like aircraft. In NASA CR-2008-1, December 2008.
120. Annaswamy, A., Jang, J., & Lavretsky, E. (2008). Stability margins for adaptive controllers
in the presence of time-delay. In AIAA Guidance, Navigation, and Control Conference, AIAA
2008-6659, August 2008.
121. Li, D., Patel, V. V., Cao, C., Hovakimyan, N., & Wise, K. (2007). Optimization of the time-
delay margin of L1 adaptive controller via the design of the underlying filter. In AIAA Guid-
ance, Navigation, and Control Conference, AIAA 2007-6646, August 2007.
Chapter 2
Nonlinear Systems
Nonlinear systems are abundant in nature. In fact, most real-world systems are inher-
ently nonlinear. Linear systems in some regards are viewed as idealization of non-
linear systems in subspaces of nonlinear solutions. Nonlinear systems possess many
complex behaviors that are not observed in linear systems. Multiple equilibrium
points, limit cycle, finite escape time are illustrative of some of the complex behav-
iors of nonlinear systems. The learning objectives of this chapter are the following:
• To develop an understanding of equilibrium concepts and linearization; and
• To be able to assess local stability of equilibrium points.
Consider a typical control block diagram as shown in Fig. 2.1.
y = h (x, u, t) (2.2)
where y (t) = y1 (t) y2 (t) . . . yl (t) ∈ Rl is the system output and h () =
h 1 () h 2 () . . . h l () ∈ Rl is a nonlinear output function.
Equations (2.1) and (2.2) together form the system state-space representation.
When m = l = 1, the system is said to be a single-input, single-output (SISO) sys-
tem. When m = 1 and l = 1, the system is said to be a multiple-input, multiple-output
(MIMO) system. The system can also be single-input, multiple-output (SIMO); or
multiple-input, single-output (MISO).
When the system is explicitly dependent on time t as in Eqs. (2.1) and (2.2), the
system is said to be non-autonomous or time-varying. When the explicit dependency
on time is not present, then the system is said to be autonomous or time-invariant.
A special case of nonlinear systems is a linear affine-in-control form which is
given by
ẋ = f (x, t) + g (x, t) u (2.3)
y = h (x, t) (2.4)
where the matrices A (t), B (t), C (t), and D (t) are functions of t.
When these matrices are constant matrices, then the following system is linear
time-invariant (LTI):
ẋ = Ax + Bu (2.7)
y = C x + Du (2.8)
2 Nonlinear Systems 19
LTI systems are much easier to analyze since there are a wealth of design and
analysis tools for such systems. As a result, many control system designs are still
based on a LTI representation.
The general solution for a LTI state-space system is given by
t
A(t−t0 )
x =e x0 + e A(t−τ ) Bu (τ ) dτ (2.9)
t0
t
y = Ce A(t−t0 ) x0 + C e A(t−τ ) Bu (τ ) dτ + Du (t) (2.10)
t0
Example 2.1:
• The following second-order spring-mass-damper system
m ẍ + c ẋ + k (x) x = bu
where m, c, and b are constant, k (x) is a nonlinear spring function, and u (x, ẋ) =
f (x, ẋ) is a nonlinear state feedback control is classified as an autonomous and
nonlinear time-invariant system.
• The following spring-mass-damper system
m ẍ + c ẋ + k (x) x = bu
m ẍ + c ẋ + k (t) x = bu
amplitude saturation and rate limiting when the actuator command exceeds certain
limits. These nonlinear sources can make even a LTI control system to become
nonlinear. Generally, actuator saturation should be avoided in order for a linear
system to retain its linearity. Thus, actuators are sized appropriately to provide a
bandwidth much larger than the control bandwidth of the system and the control
gains are designed so as not to result in an actuator amplitude saturation or rate
limiting. When properly designed, the actuator command will follow closely the
actuator output. In an ideal case, the effect of the actuator system may be neglected
during an initial design process.
Example 2.2: Consider a second-order plant
X (s) b
G (s) = = 2
U (s) s + 2ζ ωn s + ωn2
where sat () is called a saturation function, rate () is called a rate limiter function,
and u̇ c (t) is the actuator slew rate which is computed as
u c (t) − u c (t − Δt)
u̇ c (t) =
Δt
The saturation and rate limiter functions are nonlinear discontinuous functions
since the input does not map linearly with the output when the amplitude or rate
limit is exceeded.
The actuator command u c (t) is the output of the feedback controller. In this
example, the goal is to enable x (t) to track a sinusoidal input
r (t) = a sin ωt
ki
Uc (s) = k p + + kd s E (s)
s
Note that the system closed-loop transfer function for an ideal case with no actu-
ator amplitude saturation is
Even though the controller is linear, the response with the actuator saturation
becomes quite nonlinear.
2.2 shows the closed-loop response for G (s) =
Figure
5
s 2 +5s+6
, U (s) = 15 + 8
s
+ 5s E (s) with the amplitude limit at ±1 for r (t) = sin t.
0.8
0.6
0.4
0.2
r, x
−0.2
−0.4
−0.6
−0.8
r x
−1
0 2 4 6 8 10
t
All nonlinear autonomous systems possess a set of equilibrium points about which
the systems exhibit certain stability properties. Denoting x ∗ as an equilibrium point,
then for an autonomous system, the equilibrium points are the real roots of
f x∗ = 0 (2.12)
A system that is initially at an equilibrium point will remain there at all times
since an equilibrium is a constant solution of an autonomous nonlinear system.
Unlike linear systems which can have only one equilibrium point, nonlinear sys-
tems can have multiple isolated equilibrium points. Under certain conditions, the
behavior of a nonlinear system can be studied by linearization which can reveal
some important local stability properties of the system at these equilibrium points.
Linearization generally cannot predict the behavior of a nonlinear system far from
an equilibrium point about which the linearization is performed.
22 2 Nonlinear Systems
ẋ = f (x) (2.13)
x = x ∗ + x̃ (2.14)
where x̃ (t) is a small perturbation in the solution of x (t) about the equilibrium x ∗ .
Then, Eq. (2.13) can be written as
ẋ ∗ + x̃˙ = f x ∗ + x̃ (2.15)
Example 2.3 Consider a rotating pendulum without friction in Fig. 2.3. The equa-
tion of motion is given by
g
θ̈ + sin θ − ω2 sin θ cos θ = 0
l
This can be cast in a state-space form by setting x1 (t) = θ (t) and x2 (t) = θ̇ (t) as
ẋ1 x2
=
ẋ2 − gl sin x1 + ω2 sin x1 cos x1
The equilibrium points can be found by setting ẋ1 (t) = 0 and ẋ2 (t) = 0. This gives
the following solution:
2.1 Equilibrium and Linearization 23
⎧
⎨cos−1 lωg 2 , ω ≥ gl
x1∗ =
⎩0, π , ω < gl
x2∗ = 0
Physically, this means that one of the equilibrium points occurs when the angular
speed ω exceeds a certain value and the pendulum will be suspended by an angle
as the centrifugal force exerted on the pendulum is in balance with its weight. The
other two equilibrium points occur when the angular speed ω is low enough and the
pendulum is at either the bottom or the top in the vertical plane. Thus, this nonlinear
system has three equilibrium points.
The equation of motion can be linearized by computing its Jacobian matrix as
J= 0 1
− gl cos x1 + ω2 cos2 x1 − sin2 x1 0
40
0.5 Nonlinear EOM
ω = 0.8(g/l)
20 Linear EOM
θ, deg
−20
−40
0 5 10 15 20
t, sec
80
Nonlinear EOM
ω = 1.2(g/l)0.5
Linear EOM
60
θ, deg
40
20
0 5 10 15 20
t, sec
g
Fig. 2.4 Comparison between nonlinear and linear solutions, l =1
Global stability of a nonlinear system over its entire solution domain is difficult to
analyze. Linearization can provide information on the local stability in a region about
an equilibrium point. For a LTI system, there is only one equilibrium point which
is the origin. For a linearized system, the origin corresponds to an equilibrium point
of the nonlinear system about which the linearization is performed. Recall that a
LTI system is absolutely stable if the eigenvalues of the transition matrix A all have
negative real part. That is,
x1 x1
2. Saddle point occurs when both λ1 and λ2 are real and have opposite signs. Because
one of the eigenvalue corresponds to an unstable pole, half of the trajectories
diverge from the center of a saddle point as shown in Fig. 2.6. The system is
always on the verge of instability.
Saddle Point
x2
x
1
3. Stable or unstable focus occurs when both λ1 and λ2 are a complex conjugate
pair. When Re (λ1 ) < 0 and Re (λ2 ) < 0, the focus is a stable focus and the
trajectories spiral toward the focus as illustrated in Fig. 2.7. When Re (λ1 ) > 0
and Re (λ2 ) > 0, the focus is an unstable focus and the trajectories spiral outward
from the focus as shown in Fig. 2.7.
26 2 Nonlinear Systems
x x
1 1
4. Center occurs when both λ1 and λ2 are purely imaginary. All trajectories encircle
the center point at the origin with concentric-level curves as illustrated in Fig. 2.8.
Center
x2
x
1
The phase plane analysis of a nonlinear system is related to that of its linearized
system because the local behaviors of the nonlinear system can be approximated
by the behaviors of its linearized systems in the vicinity of the equilibrium points.
Because a nonlinear system can have multiple equilibrium points, one important fact
to note is that the trajectories of the nonlinear solution can exhibit unpredictable
behaviors. For example, if the solution starts with an initial condition near a stable
equilibrium point, it does not necessarily guarantee that the ensuing trajectories will
converge to a stable node or focus if there exists an unstable equilibrium point in the
solution set.
Example 2.4: From Example 2.3, the eigenvalues of the linearized systems are
g g2
∗ −1
λ1,2 J x1 = cos = ±i ω 2−
lω2 l 2 ω2
∗ g
λ1,2 J x1 = 0 = ±i − ω2
l
2.2 Local Stability and Phase Plane Analysis 27
∗ g
λ1,2 J x1 = π = ± + ω2
l
The first two equilibrium points are centers, and the last equilibrium point is a
saddle point. The phase portrait is shown in Fig. 2.9.
1
2
0
x
−1
−2
−3
−4
−4 −3 −2 −1 0 1 2 3 4
x
1
g g
Fig. 2.9 Phase portrait of θ̈ + l sin θ − ω2 sin θ cos θ = 0, l = 1, ω = 0.8 gl
Nonlinear systems differ from linear systems in that they can exhibit many complex
behaviors [1]. One feature which has been discussed is the existence of multiple
isolated equilibrium points. There are many other complex behaviors that are unique
to nonlinear systems, such as finite escape time, limit cycle, chaos.
1. Finite escape time: An unstable linear system becomes unbounded as time
approaches to infinity. For nonlinear systems, an interesting phenomenon exists.
It is possible for a nonlinear system to become unbounded in a finite interval of
time. The solution then exhibits a finite escape time.
ẋ = x 2
1
x (t) = −
t −1
which is defined only for t ∈ [0, 1). The solution has a finite escape time at t = 1.
28 2 Nonlinear Systems
2. Limit cycle: For some nonlinear systems, a limit cycle is a periodic nonlinear
solution represented by a closed trajectory in the phase plane such that all tra-
jectories in its vicinity either converge to it or diverge from it. A limit cycle can
be classified as stable, unstable, or neutrally stable depending on the behaviors
of the trajectories of the solution in its vicinity. A stable limit cycle and a stable
equilibrium point are the only two types of “regular” attractors.
Example 2.6: The Van der Pol oscillator described by
ẍ − μ 1 − x 2 ẋ + x = 0
1
2
0
x
−1
−2
−3
−4
−5
−5 0 5
x1
Fig. 2.10 Stable limit cycle of Van der Pol oscillator, μ = 0.5
2.4 Summary
Nonlinear systems can have multiple isolated equilibrium points. Under certain con-
ditions, the behavior of a nonlinear system can be studied by linearization which
can reveal some important local stability properties of the nonlinear system at these
equilibrium points. Linearization generally cannot predict the behavior of a nonlin-
ear system far from an equilibrium point about which the linearization is performed.
Nonlinear systems differ from linear systems in that they can exhibit many complex
behaviors, such as finite escape time, limit cycle, chaos. Global stability of a nonlin-
ear system over its entire solution domain is difficult to analyze. Linearization can
provide information on the local stability in a region about an equilibrium point. The
phase plane analysis of a nonlinear system is related to that of its linearized system
because the local behaviors of the nonlinear system can be approximated by the
behaviors of its linearized systems in the vicinity of the equilibrium points. Because
a nonlinear system can have multiple equilibrium points, one important fact to note
is that the trajectories of the nonlinear system can exhibit unpredictable behaviors.
2.5 Exercises 29
2.5 Exercises
1. Consider the following PID control system with the actuator amplitude saturation
and rate limiting:
5
G (s) =
s 2 + 5s + 6
with k p = 15, ki = 8, and kd = 5 as the control gains. The actuator has both
amplitude and rate limits between −1 and 1.
a. Compute the characteristic roots of the ideal closed-loop system without con-
sideration for actuator amplitude saturation and rate limiting.
b. Construct a Simulink model for a sinusoidal input r (t) = sin t. Plot the input,
the ideal output without the actuator amplitude saturation and rate limiting,
and the actual output for a simulation time t = 10 sec. Also plot the actuator
command signal u c (t) and the control input signal to the plant u (t).
c. Comment on the effect of rate limiting.
2. Given
θ̈ + cθ̇ + 2 sin θ − 1 = 0
a. Let x0 = 1. Does the solution have a finite escape time? If so, determine it.
b. Repeat part (a) with x0 = −1.
c. Comment on the effect of initial condition on the stability of the system.
30 2 Nonlinear Systems
Reference
Abstract This chapter presents some basic mathematical fundamentals for adaptive
control theory. Vector and matrix norms are defined. The existence and uniqueness of
a solution of a nonlinear differential equation are stated by the Cauchy theorem and
the Lipschitz condition. Positive-valued functions are an important class of functions
in adaptive control theory. Positive definiteness of a real-valued function is defined.
The properties of a positive-definite matrix are given.
1. Positivity
x ≥ 0 ∀x ∈ Rn (3.1)
2. Positive definiteness
x = 0 (3.2)
if and only if x = 0.
3. Homogeneity
αx = |α| x (3.3)
x is said to be a semi-norm if it satisfies all the above conditions except the positive
definiteness condition.
A p-norm of a vector x ∈ Cn , where Cn is a complex number space that general-
izes Rn , is defined as n 1/ p
x p = |xi | p
(3.5)
i=1
for p = 1, 2, . . . , ∞.
In special cases when p = 1, 2, ∞, the 1-, 2-, and infinity norms of x are defined
as
n
x1 = |xi | (3.6)
i=1
n
x2 = |xi |2 (3.7)
i=1
The 2-norm is also called the Euclidean norm which is a “distance” measure of
the vector x from the origin.
The inner product in the Euclidean space Rn is defined as
n
x, y
= xi yi (3.9)
i=1
In R2 and R3 spaces, the inner product is also known as the dot product. The inner
product is also a measure in the Euclidean space since
x, x
= x22 (3.10)
3.1 Vector and Matrix Norms 33
Example 3.1 Verify that x2 , where x ∈ Rn , satisfies the norm conditions.
Firstly, x2 can be expressed as
αx2 = (αx1 )2 + (αx2 )2 + · · · + (αxn )2 = |α| x12 + x22 + · · · + xn2 = |α| x2
Taking the square root of both sides yields the triangle inequality
provided the integral exists, in which case x ∈ L p , the space with the L p norm.
34 3 Mathematical Preliminaries
∞
n
x2 = |xi (t)|2 dt (3.14)
t0 i=1
where the notation sup denotes the supremum over all t that yields the largest value
of the argument.
for all t ≥ 0.
5
|x |+|x |
1 2
2 2
4.5 x1+x2
4 max(x1,x2)
∞
Integrand of ||x||1, ||x| , ||x||
3.5
2
2.5
1.5
0.5
0
0 1 2 3 4 5
t
The integrands of the L1 , L2 , and L∞ norms of x (t) are shown in Fig. 3.1.
The L1 , L2 , and L∞ norms of x (t) can be evaluated analytically as
∞ −t 5
x1 = 2e + e−2t dt =
0 2
∞ 9 3
x2 = 4e−2t + e−4t dt = =
0 4 2
3.1 Vector and Matrix Norms 35
x∞ = 2
Example 3.3 The function x (t) = sin t does not have a L1 or L2 norm because the
integral of sin t for t ∈ [0, ∞) does not exist. However, the L∞ norm of x (t) exists
and is equal to 1. Thus, x (t) ∈ L∞ .
In special cases when p = 1, 2, ∞, then the 1-, 2-, and infinity norms of A are
defined as
m
A1 = max ai j (3.17)
1≤ j≤n
i=1
A2 = λmax (A∗ A) (3.18)
n
A∞ = max ai j (3.19)
1≤i≤m
j=1
where the trace operator is the sum of the diagonal elements of a square matrix, that
is,
n
trace (A) = aii (3.21)
i=1
where A and B are any matrices of appropriate dimensions and ρ is the spectral
radius operator of a square matrix which is defined as
2
√
A1 = max ai j = max (|1 + i| + 2, |−1|) = max 2 + 2, 1 = 3.4142
1≤ j≤2
i=1
2
√
A∞ = max ai j = max (|1 + i| , 2 + |−1|) = max 2, 3 = 3
1≤i≤2
j=1
Thus,
ρ (A) ≤ A
is a compact set that represents a closed rectangular region in R2 as shown in Fig. 3.2.
On the other hand, the complementary set S c such that S ∪ S c = R2 (the union
of S and S c is R2 )
S c = x ∈ R2 : |x1 | > a, |x2 | > b
A nonlinear differential equation may or may not have a unique solution that is
dependent continuously on an initial condition. Certain requirements are imposed
on a nonlinear differential equation to ensure the existence of a unique solution. The
existence and uniqueness are fundamental requirements for any nonlinear differential
equations.
|t − t0 | ≤ T, x − x0 ≤ R (3.27)
where T > 0 and R > 0 are some positive constants, then there exists t1 > t0 such
that at least one solution of Eq. (3.26) exists and is continuous over the time interval
[t0 , t1 ].
Geometrically, this simply means that continuity of f (x, t) in a closed region
that contains the initial condition ensures that there exists at least one continuous
solution that lies
therein.
This is illustrated in Fig. 3.3. The gradient of a continuous
solution of x t; t0, x0 starting from an initial condition x (t0 ) = x0 is the value of
the function f (x, t) at any point on the trajectory of x (t). For the gradient to exist,
f (x, t) must be a continuous function.
1
ẋ =
x −1
for all x1 , x2 ∈ Rn and t ∈ [t0 , t1 ], then Eq. (3.26) has a unique solution for all
t ∈ [t0 , t1 ].
Inequality (3.28) is called the Lipschitz condition and the constant L is called a
Lipschitz constant. When x1 and x2 are any two points in the Euclidean space Rn ,
then f (x, t) is said to be globally Lipschitz. Geometrically, the Lipschitz condition
essentially imposes the continuity and boundedness of the partial derivative or the
Jacobian of f (x, t) with respect to x (t) as illustrated in Fig. 3.4.
subject to x (0) = 0, not only has a solution x (t) = t 3 but also has a zero solution
x (t) = 0. Thus, this equation does not have a unique solution.
2
The function f (x) = 3x 3 does not satisfy the global Lipschitz condition since
the derivative of f (x)
f (x) = 2x − 3
1
is not bounded at x = 0.
Example 3.8
• The function f (x) = x is globally Lipschitz since it has continuous and bounded
derivative for all x (t) ∈ R. The Lipschitz constant L is equal to 1.
• The function f (x) = sin x is globally Lipschitz since its derivative is continuous
and bounded with a Lipschitz constant L = 1 for all x (t) ∈ R.
• The function f (x) = x 2 is not globally Lipschitz since its derivative, f (x) = 2x,
is unbounded for x (t) ∈ R as x (t) → ±∞.
Note that the global Lipschitz condition is very restrictive since a function with this
condition must have a bounded derivative for all values of x (t) ∈ Rn as x (t) → ±∞.
Many nonlinear physical models fail to satisfy this condition since physically there
are usually upper and lower limits placed on variables of most if not all physical
models. These limits thus does not require x (t) to be in a domain that occupies
the entire Euclidean space Rn , but rather in a subset of it. This leads to a notion of
local Lipschitz condition, which is less restrictive and can be easily satisfied by any
nonlinear differential equations.
for some x1 and x2 in a finite neighborhood of x0 and t ∈ [t0 , t1 ], then Eq. (3.26) has
a unique solution in a finite time interval t ∈ [t0 , t0 + δ], where t0 + δ < t1 .
The local Lipschitz condition now only imposes the continuity and boundedness
of the partial derivative of f (x, t) with respect to x (t) over a finite neighborhood of
x0 as a subset of Rn instead of over all x (t) ∈ Rn . Therefore, it is a less restrictive
condition that can easily be met. In most cases, the local Lipschitz condition is a
reasonable assumption for many continuous nonlinear functions.
Example 3.9 The differential equation
2
ẋ = 3x 3
subject to x (0) = 1 has a unique solution x (t) = (t + 1)3 for a finite neighborhood
of x0 , for example, x (t) ∈ [1, 2]. f (x) satisfies the local Lipschitz condition since
the derivative of f (x) is now continuous and bounded for all x (t) ∈ [1, 2].
Example 3.10 The function f (x) = x 2 is locally Lipschitz for any finite neighbor-
hood of x (t) ∈ [x1 , x2 ], where x1 and x2 are finite-valued.
The Lyapunov stability theory for nonlinear systems often invokes a class of functions
called positive-definite functions. One such function is a quadratic function
V (x) = x P x (3.30)
A positive (negative) (semi)-definite matrix is defined the same way as a positive
(negative) (semi-)definite function.
A positive-definite matrix P has the following properties:
• P is a symmetric matrix, that is,
P = P (3.32)
• All eigenvalues of P are real and positive, that is, λ (P) > 0.
• A scalar multiplication with a constant causes the product to have the same sign
definiteness as the constant, that is, α P > 0 if α > 0 and α P < 0 if α < 0.
• Let A be any arbitrary matrix of appropriate dimension, then
A P A > 0 (3.33)
P+Q>0 (3.34)
which can be illustrated geometrically in Fig. 3.5, where λmin (.) and λmax (.)
denote the minimum and maximum eigenvalues of a matrix, respectively.
The above properties apply for a negative definite matrix by reversing the inequality
sign.
Example 3.11
• The identity matrix I is a positive-definite matrix.
• The matrix
53
P=
32
which, at a first glance, does not seem to be positive definite. By completing the
square, it can be shown that V (x) > 0 as follows:
2 2
3 3 3 √
V (x) = − √ x1 + √ x1 + 2 √ x1
5x12 2x2 + 2x22
2 2 2
2
1 3 √
= x12 + √ x1 + 2x2 > 0
2 2
qii = 0 (3.38)
44 3 Mathematical Preliminaries
V (x) = x Qx = 0 (3.39)
is identically zero.
A=M+N (3.40)
where M is called the symmetric part of A and N is called the anti-symmetric part
of A such that
1
M= A + A (3.41)
2
1
N= A − A (3.42)
2
If V (x) is a quadratic function formed with A, then
1
V (x) = x Ax = x A + A x (3.43)
2
3.5 Summary 45
3.5 Summary
3.6 Exercises
n
x1 = |xi |
i=1
and verify the answers with MATLAB using the function “norm.”
Is C a compact set? Write the set notation for the complementary set C c . Plot
and illustrate the region in R2 that represents C .
5. For each of the following equations, determine if f (x) is locally Lipschitz at
x = x0 or globally Lipschitz:
46 3 Mathematical Preliminaries
√
a. ẋ = x 2 + 1, x0 = 0.
−x 3 , x0 = 1.
b. ẋ = √
c. ẋ = x 3 + 1, x0 = 0.
Chapter 4
Lyapunov Stability Theory
with an initial condition x (t0 ) = x0, where f (x) is locally Lipschitz in some
subset D of Rn and the solution x t; t0, x0 exists and is unique in a region
B R = {x (t) ∈ Rn : x < R} ⊂ D of an equilibrium point x ∗ . The region B R
can be thought of as a hypersphere in Rn with the origin at x = x ∗ . Colloquially, it
is often referred to in the literature as a ball B R . Since x ∗ is a constant vector, for
convenience, the autonomous system can be transformed by shifting the equilibrium
point to the origin at x = 0. Let y (t) = x (t) − x ∗ , then
ẏ = f y + x ∗ g (y) (4.2)
Stability concept essentially implies that, given a system with an initial condition
close to the origin, the trajectory of the system can be kept arbitrarily close to it.
Figure 4.1 illustrates the stability concept.
Note that instability for linear systems means that the solution grows exponentially
as t → ∞ due to unstable poles in the right half plane, resulting in unbounded
signals. For nonlinear systems, instability of an equilibrium does not always lead to
unbounded signals. For example, the Van der Pol oscillator in Example 2.6 has a
stable limit cycle that encloses an unstable equilibrium point at the origin. So, the
equilibrium point in theory is unstable and the system cannot stay arbitrarily close to
it. If we choose any arbitrary circle B R to be completely inside the limit cycle, then
no matter how close the initial condition is to the origin, the trajectory of the system
will eventually escape the circle B R as illustrated in Fig. 4.2. However, the trajectory
tends to the limit cycle and remains there as t → ∞.
50 4 Lyapunov Stability Theory
1.5
0.5
Br
2
0
x
−0.5
−1
B
−1.5 R
−2
−2 −1 0 1 2
x
1
The stability concept in the Lyapunov sense does not explicitly imply that the tra-
jectory of a nonlinear system will eventually converge to the origin. For example, an
ideal spring-mass system without friction will display a sinusoidal motion forever if
it is subject to a disturbance. So, the system is stable in the Lyapunov sense but does
not converge to the origin. Asymptotic stability is a stronger stability concept than
the Lyapunov stability concept and is defined as follows:
All trajectories starting within B R will eventually converge to the origin. The
origin is then said to be attractive. For a second-order system, both stable focus and
stable node are attractive. The largest such region is called a region of attraction,
defined as
R A = x (t) ∈ D : lim x (t) = 0 (4.5)
t→∞
It is noted that the asymptotic stability concept in the definition above is a local
concept for any initial condition that lies within the ball B R . If an equilibrium point
of a system is asymptotically stable for all initial conditions x0 ∈ Rn , then the
4.1 Stability Concepts 51
ẋ = −x 2
Note that the equilibrium point is unstable if x0 < 0 and has a finite escape time
at t = −1/x0 . So, the equilibrium is asymptotically stable for all x (t) ∈ R+ but not
asymptotically stable in the large.
This definition gives a local version of the exponential stability concept for some
initial condition x0 close to the origin. If the origin is exponentially stable for all
initial conditions x0 ∈ Rn , then the equilibrium point is said to be exponentially
stable in the large. The constant β is called the rate of convergence.
It is noted that exponential stability implies asymptotic stability, but the converse
is not true.
Example 4.3 The differential equation
ẋ = −x 1 + sin2 x
if x (t) > 0.
The solution is bounded from below and above as shown in Fig. 4.3 by
1
. 2
x=−x(1+sin x)
0.9
.
x=−x
.
x=−2x
0.8
0.7
0.6
0.5
x
0.4
0.3
0.2
0.1
0
0 0.5 1 1.5 2 2.5 3 3.5 4
t
4.2.1 Motivation
m ẍ + c ẋ + kx = 0 (4.7)
where m is the mass, c > 0 is the viscous friction coefficient, and k is the spring
constant.
4.2 Lyapunov’s Direct Method 53
where x1 (t) = x (t) and x2 (t) = ẋ (t). The system has an equilibrium point at
(0, 0), that is, it is at rest with zero displacement and velocity
The spring-mass-damper system possesses two types of energy: (1) kinetic energy
and (2) potential energy. The kinetic energy of any moving point mass is given by
1 2 1 1
T = mv = m ẋ 2 = mx22 (4.9)
2 2 2
The potential energy for a spring is given by
1 2 1
U= kx = kx12 (4.10)
2 2
The energy of the spring-mass-damper system is the sum of the kinetic energy
and potential energy. Thus, the energy function of the system is defined as
1 2 1 2
E = T +U = mx + kx (4.11)
2 2 2 1
Note that the energy function is a quadratic positive-definite function.
The friction force also does work on the mass. This type of work is called a
non-conservative work which is usually due to a dissipative force, as opposed to
a conservative work of which potential energy is one form. The work function is
defined in general as
W = F.d x (4.12)
The total energy of the system is the sum of the energy and the work done. Thus,
1 2 1 2
E+W = mx + kx + cx22 dt (4.14)
2 2 2 1
According to the first law of thermodynamics, the total energy of a closed system
is neither created or destroyed. In other words, the total energy is conserved and is
equal to a constant. Thus,
E + W = const (4.15)
or equivalently
Ė + Ẇ = 0 (4.16)
This energy conservation law can be easily verified for the spring-mass-damper
system as
d c k
Ė + Ẇ = mx2 ẋ2 +kx1 ẋ1 + = mx2 − x2 − x1 +kx1 x2 +cx22 = 0
cx22 dt
dt m m
(4.17)
The time derivative of the energy function is evaluated as
c k
Ė = mx2 ẋ2 + kx1 ẋ1 = mx2 − x2 − x1 + kx1 x2 = −cx22 ≤ 0 (4.18)
m m
for c > 0.
The reason Ė is only negative semi-definite is because Ė can be zero for any
x1 = 0.
Thus, for a dissipative system, the time derivative of the positive definite energy
function is a negative semi-definite function. That is,
Ė ≤ 0 (4.19)
The equilibrium point is then stable. Thus, stability of a dynamical system can
be studied by examining the time derivative of the energy function. The Lyapunov
stability theory is motivated by the concept of energy. In fact, the energy function
is a Lyapunov function. Whereas the energy function is unique for a given physical
system, a Lyapunov function can be any positive-definite function that satisfies the
negative (semi-)definiteness of its time derivative.
Lyapunov’s direct method is a powerful tool for assessing stability of an equilib-
rium of a nonlinear system directly without solving the system’s dynamical equation.
4.2 Lyapunov’s Direct Method 55
The motivation of the method is based on the energy concept of a mechanical system.
From the spring-mass-damper example, the following observations are made:
• The energy function is positive definite.
• The time rate of the energy function is negative semi-definite in which case the
equilibrium is stable.
Aleksandr Mikhailovich Lyapunov (1857–1918) recognized that stability of a system
can be proven without developing a true knowledge of the system energy using a
class of positive-definite functions, known as Lyapunov functions, provided they can
be found.
Definition 4.4 A function V (x) is said to be a Lyapunov function if it satisfies the
following conditions:
• V (x) is positive definite; i.e.,
V (x) > 0 (4.20)
∂V ∂V
V̇ (x) = ẋ = f (x) ≤ 0 (4.21)
∂x ∂x
or
V̇ (x) < 0 (4.22)
Geometrically, a Lyapunov function may be illustrated by a bowl-shaped surface
as shown in Fig. 4.5. The Lyapunov function is a level curve on the bowl starting
at the top and progressing downward toward the bottom of the bowl. The value of
the Lyapunov function thus decreases toward zero at the bottom of the bowl which
represents a stable equilibrium.
Example 4.4 For the spring-mass-damper system, the energy function is clearly a
Lyapunov function. Suppose one chooses another Lyapunov candidate function
Then,
c k k c
V̇ (x) = 2x1 ẋ1 +2x2 ẋ2 = 2x1 x2 +2x2 − x2 − x1 = 2x1 x2 1 − −2 x22
m m m m
Note that V̇ (x) is not negative (semi-)definite because of the term with the product
x1 x2 . So, this candidate function is not a Lyapunov function.
In many systems, finding a Lyapunov function is not trivial. Unfortunately, there
is no straightforward way to obtain a Lyapunov function. Perhaps, the most obvious
Lyapunov function for any system is the energy function, but it is not always easy to
identify such a function for a nonlinear system unless one knows the physics of the
system.
Example 4.5 Consider a pendulum with viscous friction whose motion is described
by
ml 2 θ̈ + cθ̇ + mgl sin θ = 0
It is not obvious what a Lyapunov candidate function would look like for this
system. One can try
V (x) = x12 + x22
1 2 2
T = ml x2
2
4.2 Lyapunov’s Direct Method 57
U = mgl (1 − cos x1 )
1 2 2
E = T +U = ml x2 + mgl (1 − cos x1 ) > 0
2
Thus, it is not surprising that the energy function can always be used for a
Lyapunov function, provided such a function can be found.
In summary, it can be seen that the Lyapunov’s direct method is a powerful tech-
nique for studying stability of an equilibrium point of a nonlinear system.
It is important to note that the Lyapunov’s direct method only gives a sufficient
condition for stability. Failure of a Lyapunov candidate function to satisfy the stability
condition does not imply that the equilibrium is unstable. It simply means that a good
Lyapunov candidate function may not have been identified. An exception to this rule
is the energy function which provides both the necessary and sufficient conditions
for stability.
Example 4.6 For the pendulum in Example 4.5, V̇ (x) is negative semi-definite, so
the equilibrium is locally stable.
Example 4.7 Consider the system
ẋ1 x13 + x1 x22 − x1
=
ẋ2 x23 + x12 x2 − x2
Observing V̇ (x), one can conclude that V̇ (x) < 0 for all x ∈ B R where
B R = x (t) ∈ D ⊂ R2 : x12 + x22 < 1
k m
p12 = 1 ⇒ p12 =
m k
c m m m
p12 − p22 = −1 ⇒ p22 = ( p12 + 1) = +1
m c c k
c k c k c m k
p11 − p12 − p22 = 0 ⇒ p11 = p12 + p22 = + +
m m m m k c c
4.2 Lyapunov’s Direct Method 59
The matrix c
+ m
+ k m
mk
P= k c c
m
k
m
c k
+1
ẋ = Ax
where
0 1
A=
− mk − mc
V̇ (x) = ẋ P x + x P ẋ
A P + P A < 0
which is called a linear matrix inequality (LMI) that can be solved for P.
Alternatively, one can write the LMI as a linear matrix equation
A P + P A = −Q
A
P + P A = −2I
60 4 Lyapunov Stability Theory
There are numerical methods that can be used to solve the Lyapunov equation.
For a low matrix dimension less than 4, the equation can be solved analytically. For
this example, one can set up
0 − mk p11 p12 p11 p12 0 1 −2 0
+ =
1 − mc p12 p22 p12 p22 − mk − mc 0 −2
then expand
−2 p12 mk p11 −
p12 m − pc 22
c
m
k
−2 0
=
p11 − p12 m − p22 m 2 p12 − p22 m
c k
0 −2
and
V̇ (x) ≤ −βV (x) (4.24)
subject to x (0) = 1.
Choose a Lyapunov candidate function
V (x) = x 2 = x2
So,
V (0) = x02 = 1
V̇ (x) is computed as
V̇ (x) = 2x ẋ = −2x 2 1 + sin2 x ≤ −2 x2 = −2V (x) < 0
4.2 Lyapunov’s Direct Method 61
or
x 2 ≤ e−2t
which is equivalent to
|x| ≤ e−t
unboundedness of V (x) ensures that all trajectories in the large when x → ∞
will be attracted to the origin. The global stability Lyapunov condition can be stated
by the following Barbashin–Krasovskii theorem [2, 4]:
ẋ = −ax
where a > 0, whose equilibrium point at the origin is asymptotically stable in the
large. Suppose a Lyapunov candidate function is chosen as
x2
V1 (x) = >0
1 + x2
2x ẋ 2x 3 ẋ −2ax 2
V̇1 (x) = − = 2 → 0
1 + x2 1 + x2
2
1 + x2
as x → ∞, which implies that the origin is not asymptotically stable as x → ∞.
Therefore, the origin is stable but not asymptotically stable in the large, which is a
contradiction.
Now, suppose another Lyapunov candidate function is chosen as
V2 (x) = x 2 > 0
which implies that the origin is indeed asymptotically stable in the large.
4.2 Lyapunov’s Direct Method 63
Asymptotic stability requires that V̇ (x) < 0. Yet, for the spring-mass-damper sys-
tem, if the energy function is chosen as a Lyapunov function, then V̇ (x) ≤ 0, even
though the solution of the system is clearly asymptotic in the presence of a dissipa-
tive force due to the viscous friction. The LaSalle’s invariant set theorem can resolve
this apparent contradiction when an asymptotically stable equilibrium point of an
autonomous system only satisfies the Lyapunov condition V̇ (x) ≤ 0.
Example 4.11
• An equilibrium point is an invariant set because by definition x (t) = x ∗ is a
constant solution of an autonomous system so that
x (t) = x (t0 ) = x ∗ ∈ M , ∀t ≥ t0
• The limit cycle of the Van der Pol oscillator is an invariant set since any point on
the limit cycle will remain on it for all future time.
• The entire Euclidean space Rn is a trivial invariant set since all trajectories must
be in some subspace that belongs in Rn .
The LaSalle’s invariant set theorem is stated as follows:
which represents a circular region that includes the origin. Therefore, the origin is
unstable.
Let R be a set of all points where V̇ (x) = 0. Then,
R = x (t) ∈ B R : g (x) = x12 + x22 − 1 = V (x) − 1 = 0
in fact represents a bounded solution x (t) since all trajectories either inside or outside
of S c will move toward R and remain in R as illustrated in Fig. 4.6. Thus, R is an
invariant set. One can also verify this by taking the time derivative of the function
g (x) that represents the trajectories in the set R
d
ġ (x) = (V − 1) = V̇ (x) = 0, ∀x (t) ∈ R
dt
The Lyapunov function can be solved analytically by noting that
V̇ = −2V (V − 1)
which leads to
dV
= −2dt
V (V − 1)
V0
V (t) =
V0 − (V0 − 1) e−2t
lim V (t) = 1
t→∞
1.8
1.6
1.4
0.8
0.6
0.4
0.2
0
0 0.5 1 1.5 2 2.5 3 3.5 4
t
which is in fact the set R. Therefore, the set R is a bounding set of all possible
solutions x (t). Thus, x (t) ∈ L∞ ; i.e., x (t) is bounded.
Example 4.13 Consider the spring-mass-damper system. Choose the Lyapunov func-
tion to be the energy function
1 2 1 2
V (x) = mx + kx
2 2 2 1
Then,
V̇ (x) = mx2 ẋ2 + kx1 ẋ1 = −cx22 ≤ 0
Since V̇ (x) is negative semi-definite, then the origin is only stable in the Lyapunov
sense, but not asymptotically stable as one would expect for a spring-mass-damper
system with friction. Let R be a set of all points where V̇ (x) = 0. Then,
R = x (t) ∈ R2 : V̇ (x) = 0 ⇒ x2 = 0
is a collection of all points that lie on the x1 -axis. It follows that any point on this
axis must satisfy
m ẋ2 + kx1 = 0
or
k
ẋ2 = ẍ1 = − x1
m
If x1 (t) = 0, then ẍ1 (t) = 0 with sgn (ẍ1 ) = −sgn (x1 ) in R, where sgn () is the
sign function which returns 1 if the argument is positive, 0 if the argument is zero,
or -1 if the argument is negative. This means that a point on this axis cannot remain
in R because the acceleration ẍ1 causes the point to move toward the origin, unless
it is already at the origin.
Another way to find an invariant set is to evaluate the derivative of the function
that describes R and set it equal to zero. Hence,
ẋ2 = 0
solution x = 0. Then, the origin is asymptotically stable. Moreover, if V (x) > 0 is
a positive definite radially unbounded function and R = x (t) ∈ Rn : V̇ (x) = 0 ,
then the origin is asymptotically stable in the large.
The Lyapunov equation has a connection to the optimal control theory. In particular,
the Lyapunov equation can be viewed as a special case of the Riccati equation for a
Linear Quadratic Regulator (LQR) optimal control problem. Consider the following
LTI system:
ẋ = Ax (4.27)
∂ H
λ̇ = − = −Qx − A
λ (4.30)
∂x
subject to the transversality (terminal time) condition
λ tf = 0 (4.31)
Ṗ x + P Ax = −Qx − A
P x (4.33)
68 4 Lyapunov Stability Theory
Ṗ + P A + A
P + Q = 0 (4.34)
subject to P t f = 0.
Contrasting this with the differential Riccati equation
Ṗ + P A + A P − P B R −1 B P + Q = 0 (4.35)
the differential Lyapunov equation is a special case of the differential Riccati equation
for R → ∞.
Note that the differential Lyapunov equation is defined backward in time with the
transversality condition given at the final time. By transforming into a time-to-go
variable, τ = t f − t, then
dP
= P A + A
P + Q (4.36)
dτ
subject to P (0) = 0 in the time-to-go coordinate.
If A is Hurwitz and let t f → ∞ which corresponds to an infinite time horizon
solution, then the time-varying solution of the differential Lyapunov equation tends
to a constant solution of the algebraic Lyapunov equation
P A + A P + Q = 0 (4.37)
which is positive definite for Q > 0 and requires that A be Hurwitz since the solution
must be a stable solution such that
lim e Aτ = 0 (4.39)
τ →∞
The differential Lyapunov equation in time-to-go can be solved using any numer-
ical technique for solving differential equations such as the Euler or Runge–Kutta
method. For example, the Euler method for solving the Lyapunov equation is as
follows:
Pi+1 = Pi +
τ Pi A + A
Pi + Q
4.2 Lyapunov’s Direct Method 69
1.2
0.8
0.6
P
0.4
0.2
0
p11 p12=p21 p22
−0.2
0 2 4 6 8 10
τ=tf−t
Most of the concepts for Lyapunov stability for autonomous systems can be applied
to non-autonomous systems with some additional considerations [2, 4].
Consider a non-autonomous system
ẋ = f (x, t) (4.40)
Otherwise, the system does not have a “true” equilibrium point in the Lyapunov
sense.
Example 4.15
• The system
ẋ = g (t) h (x)
• The system
ẋ = g (t) h (x) + d (t)
does not have a true equilibrium point since h (x) would have been a function of
t to satisfy ẋ (t) = 0, which is a contradiction unless g (t) = αd (t) for some
constant α.
Definition 4.9 The equilibrium point x ∗ = 0 is said to be stable (in the sense of
Lyapunov) if, for any R > 0, there exists some r (R, t0 ) > 0 such that
Note that the difference in this definition as compared to that for an autonomous
system is the ball of radius r that encloses x0 now may depend on the initial time t0 .
Thus, the stability of the origin may also be dependent on the initial time.
The concept of uniform stability is an additional consideration for non-autonomous
systems. Uniform stability implies that the radius r = r (R) is not dependent on
the initial time, and so are the stability properties of the equilibrium point. For
autonomous systems, stability is independent of the initial time. This property is
highly desirable since it eliminates the need for examining the effect of the initial
time on the stability of a non-autonomous system.
V (0, t) = 0 (4.43)
4.3 Stability of Non-Autonomous Systems 71
and
0 < V (x, t) ≤ W (x) (4.44)
∂V ∂V
V̇ (x, t) = f (x, t) + ≤0 (4.45)
∂x ∂t
then the origin is said to be uniformly stable in the Lyapunov sense. Moreover,
if V̇ (x, t) < 0, then the origin is said to be uniformly asymptotically stable, and
additionally, if the region B R is extended to the entire Euclidean space Rn , then the
origin is said to be uniformly asymptotically stable in the large.
When a non-autonomous system does not have an equilibrium point, stability of such
a system is defined by the notion of boundedness [2, 4].
Moreover, the solution is said to be uniformly ultimately bounded if, for any
R > 0, there exists some r > 0 independent of R and the initial time t0 such that
The uniform ultimate boundedness concept simply means that the solution may
not be uniformly bounded initially according to Definition 4.11 but eventually
becomes uniformly ultimately bounded after some time has passed. The constant
R is called a bound if the solution is uniformly bounded or an ultimate bound if the
solution is uniformly ultimately bounded.
The Lyapunov’s direct method can be applied to a non-autonomous system
according to the following theorem:
Theorem 4.6 Given a Lyapunov function V (x, t) for all x ≥ R and t ∈ [0, ∞),
then the solution of the non-autonomous system (4.40) is said to be uniformly
bounded if there exist functions ϕ1 (x) ∈ K R and ϕ2 (x) ∈ K R such that [2]
• ϕ1 (x) ≤ V (x, t) ≤ ϕ2 (x)
• V̇ (x, t) ≤ 0
for all x ≥ R and t ∈ [0, ∞). In addition, if there exists a function ϕ3 (x) ∈ K R
such that
• V̇ (x, t) ≤ −ϕ3 (x)
for all x ≥ R and t ∈ [0, ∞), then the solution is said to be uniformly ultimately
bounded.
Example 4.17 Consider
ẋ = −x + 2 sin t
subject to x (0) = x0 .
The system does not have an equilibrium. The solution is
x → sin t − cos t
so that √
x ≤ 2=R
independent of r . √
The solution then is also uniformly ultimately bounded with an
ultimate bound of 2 as shown in Fig. 4.9.
4.3 Stability of Non-Autonomous Systems 73
3
R (Bound for t ≥0)
Lyapunov Ultimate Bound
2
1
r R (Ultimate Bound for t ≥T)
0
x
−1
−2
−3
0 5 10 15 20
t
The Lyapunov’s direct method is now applied to determine the uniform bound-
edness of a solution. Consider a Lyapunov candidate function for this system
V (x) = x 2 > 0
Note that one can always find ϕ1 (x) ∈ K R and ϕ2 (x) ∈ K R for this
Lyapunov candidate function such that ϕ1 (x) ≤ V (x) ≤ ϕ2 (x), for example,
ϕ1 (x) = ax 2 with a < 1 and ϕ2 (x) = bx 2 with b > 1.
Then,
V̇ (x) = 2x ẋ = 2x (−x + 2 sin t) ≤ −2x 2 + 4 x
We see that
V̇ (x) ≤ −2V (x) + 4 V (x)
√
Let W (t) = V (t) = x. Then,
V̇ √
Ẇ = √ = − V + 2 ≤ −W + 2
2 V
W ≤ (x0 − 2) e−t + 2
Thus,
lim x = lim W ≤ 2 = R
t→∞ t→∞
74 4 Lyapunov Stability Theory
Choose
ϕ3 (x) = 2 x2 − 4 x
The LaSalle’s invariant set theorem can be used to show asymptotic stability of
an equilibrium point for an autonomous system when V̇ (x) is only negative semi-
definite. This theorem cannot be used for non-autonomous systems. Therefore, it can
be much more difficult to show asymptotic stability for a non-autonomous system
than for an autonomous system. Barbalat’s lemma is a mathematical tool that can be
used to address this situation to some extent [4].
Firstly, the concept of uniform continuity needs to be introduced. A mathematical
formal definition of uniform continuity of a function is given as follows:
Definition 4.12 The function f (t) ∈ R is uniformly continuous on a set D if, for
any > 0, there exists some δ () > 0 such that
The following statements are equivalent to the definition of uniform continuity:
4.3 Stability of Non-Autonomous Systems 75
with x(t0 ) = x0 and a continuous bounded input u (t). The system is exponentially
stable with the solution
t
x = e−A(t−t0 ) x0 + e A(t−τ ) Bu (τ ) dτ
t0
Thus, x (t) is a continuous bounded signal with a bounded derivative ẋ (t) for all
t ∈ [0, ∞). Therefore, x (t) is uniformly continuous. Any output signal
y = C x + Bu
is also uniformly continuous if u (t) has a bounded derivative. The system is then
said to be bounded-input-bounded-output (BIBO) stable.
The Barbalat’s lemma is now stated as follows:
Example 4.19
• The function f (t) = e−t has a finite limit as t → ∞. To determine the uniform
2
not the second derivative f¨ (t) is bounded for all t ∈ [0, ∞). The second derivative
f¨ (t) = −2e−t + 4t 2 e−t , in fact, is bounded because the exponential term e−t
2 2 2
decreases at a much faster rate than the power term t 2 . Therefore, limt→∞ f˙ (t) =
−2te−t = 0 using the L’Hospital rule.
2
0. In fact, one can verify lim
t→∞
• The function f (t) = 1t sin t 2 which tends to zero as t → ∞ but whose derivative
f˙ (t) = − t12 sin t 2 + 2 cos t 2 does not have a limit as t → ∞. Thus, it can be
seen that even if the limit of a differentiable function f (t) exists and is finite as
t → ∞, it does not necessarily imply that limt→∞ f˙ (t) = 0 since f˙ (t) may not
be uniformly continuous, that is, f˙ (t) has a bounded derivative
or equivalently
f¨ (t) is bounded. Therefore, the function f (t) = 1t sin t 2 does not satisfy the
Barbalat’s lemma.
• The function f (t) = sin (ln t) whose derivative f˙ (t) = 1t cos (ln t) tends to zero
but f (t) does not have a finite limit as t → ∞. Thus, limt→∞ f˙ (t) = 0 does
not necessarily imply that the limit of a differentiable function f (t) exists and is
finite. Therefore, the converse of the Barbalat’s lemma is not true.
The Barbalat’s lemma is now extended to the Lyapunov’s direct method to examine
the asymptotic stability of a non-autonomous system by the following Lyapunov-like
lemma [4]:
ẋm = −axm
θ̇ = −bew (t)
where e (t) = xm (t) − x (t) is called a tracking error, described by the tracking error
equation
ė = ẋm − ẋ = −ae + bθ̃ w (t)
V (e, θ ) = e2 + θ̃ 2
Then,
V̇ e, θ̃ = 2e −ae + bθ̃ w (t) + 2θ̃ [−bew (t)] = −2ae2 ≤ 0
∞
V e (t → ∞) , θ̃ (t → ∞) − V e (t0 ) , θ̃ (t0 ) = V̇ e, θ̃ dt
t0
∞
= −2a e2 (t) dt = −2a e22
t0
V e (t → ∞) , θ̃ (t → ∞) = V e (t0 ) , θ̃ (t0 ) − 2a e22
= e2 (t0 ) + θ̃ 2 (t0 ) − 2 e22 < ∞
So, V e, θ̃ has a finite limit as t → ∞. Since e2 exists, therefore e (t) ∈
L2 ∩ L∞ .
Next, V̇ e, θ̃ must be shown to be uniformly continuous. This can be done by
examining the derivative of V̇ e, θ̃ to see if it is bounded. V̈ e, θ̃ is computed as
78 4 Lyapunov Stability Theory
V̈ e, θ̃ = −4ae −ae + bθ̃ w (t)
Since e (t) ∈ L2 ∩ L∞ and θ̃ (t) ∈ L∞ by the virtue of V̇ e, θ̃ ≤ 0, and
w (t) ∈ L∞ by assumption, then V̈ e, θ̃ ∈ L∞ . Therefore, V̇ e, θ̃ is uniformly
continuous. It follows from the Barbalat’s lemma that V̇ e, θ̃ → 0 and hence
e (t) → 0 as t → ∞. Note that one cannot conclude that the system is asymptotically
stable since only e (t) → 0 as t → ∞, but θ̃ (t) is only bounded.
4.4 Summary
The Lyapunov stability theory is the foundation of nonlinear systems and adap-
tive control theory. Various stability concepts for autonomous and non-autonomous
systems are introduced. The Lyapunov’s direct method is an indispensable tool
for analyzing stability of nonlinear systems. Barbashin–Krasovskii theorem pro-
vides a method for global stability analysis. LaSalle’s invariant set theorem provides
another complementary tool for analyzing systems with invariant sets. Stability of
non-autonomous systems involves the concepts of uniform stability, uniform bound-
edness, and uniform ultimate boundedness. Barbalat’s lemma is an important math-
ematical tool for analyzing stability of adaptive control systems in connection with
the concept of uniform continuity of a real-valued function.
4.5 Exercises
1. Given
ẋ1 x1 x12 + x22 − 1 − x2
=
ẋ2 x1 + x2 x12 + x22 − 1
a. Determine all the equilibrium points of the system and linearize the system
about the equilibrium points to classify the types of equilibrium points.
b. Use the Lyapunov candidate function
subject to x (0) = 1.
a. Determine the upper and lower bound solutions.
b. Use the Lyapunov candidate function
V (x) = x 2
to determine the type of Lyapunov stability and the upper bound of V (x)
as an explicit function of time.
3. Use the Lyapunov candidate function
4. Given
ẋ = Ax
A P + P A = −2I
where
0 1
A=
−4 4
6. Given ⎡ ⎤
0 1 0
A = ⎣ −1 −1 −2 ⎦
1 0 −1
Determine whether or not A is Hurwitz. If so, compute P using the Euler method
to integrate the differential Lyapunov equation
dP
= P A + A
P + I
dτ
subject to P (0) = 0, where τ is time-to-go. Plot all six elements of P on the
same plot and verify the result at the final time-to-go with the MATLAB function
“lyap.”
7. Use the Lyapunov’s direct method to determine an ultimate bound of the solution
x (t) for the following equation:
ẋ = −x + cos t sin t
V (x) = x 2
Plot f (t) for t ∈ [0, 5]. Determine whether or not the limit of f (t) exists as
t → ∞ and f˙ (t) is uniformly continuous. If so, use the Barbalat’s lemma to
show that f˙ (t) → 0 as t → ∞ and verify by taking the limit of f˙ (t) as t → ∞.
11. Consider the following adaptive control system:
ė = −e + θ x
θ̇ = −xe
where e (t) = xm (t) − x (t) is defined as the tracking error between a given
explicit reference time signal xm (t) which is assumed to be bounded, i.e.,
xm (t) ∈ L∞ , and the state variable x (t). Show that the adaptive system is
stable and that e (t) → 0 as t → ∞.
References
When designing a controller for a system, a control designer typically would like
to know how the system behaves physically. This knowledge is usually captured
in the form of a mathematical model. For many real-world applications, modeling
of physical systems can never be perfect as systems may have parameter variations
due to nonlinearity, parameter uncertainty due to modeling inaccuracy or imprecise
measurements, uncertainty in exogenous disturbances coming from the operating
environment, or other sources of uncertainty. The role of a modeling specialist is
to reduce the system uncertainty as much as practicable. The control designer then
uses a mathematical model of the system to design a controller which may incorpo-
rate performance measures and stability margins to account for any residual system
uncertainty that cannot be completely accounted for.
In situations when the system uncertainty may become significant beyond a level
of desired tolerance that can adversely affect the performance of a controller, adaptive
control can play an important role in reducing the effects of the system uncertainty on
the controller performance. Situations that may warrant the use of adaptive control
could include unintended consequences of off-nominal modes of operation such
as system failures or highly uncertain operating conditions, and complex system
behaviors that can result in an increase in the complexity and hence cost of the
modeling efforts. In this chapter, the learning objectives are as follows:
where k x (t) and kr (t) are adjustable control gains. The mechanism to adjust these
control gains is via an adaptive law. Thus, a direct adaptive control in effect adjusts a
feedback control mechanism of a control system directly to cancel out any unwanted
system uncertainty so that the performance of the control system can be regained in
the presence of significant system uncertainty.
In contrast, an indirect adaptive controller achieves the same objective by adjusting
the control gains in an indirect way, which may be expressed as
u = k x ( p (t)) x + kr ( p (t)) r (5.2)
where p (t) are system parameters that are estimated online to update the control
gains.
ẍ + 2ζ ωn ẋ + ωn2 x = u
Choose
ωn2 + k p = ωm2 ⇒ k p = ωm2 − ωn2
2ζ ωn + kd = 2ζm ωm ⇒ kd = 2 (ζm ωm − ζ ωn )
kr = ωn2
Adaptive control can deal with either linear or nonlinear plants with various types of
uncertainty which can be structured uncertainty, unstructured uncertainty, or unmod-
eled dynamics.
1. Structured uncertainty is a source of uncertainty with uncertain parameters but
known functional characteristics. It is also often referred to as parametric uncer-
tainty.
Example 5.2 A linear spring-mass-damper system with an uncertain spring constant
m ẍ + c ẋ + k ∗ x = u
ẏ = c2 x + c3 y
where y (t) is an internal system state that is not modeled, and ci , i = 1, 2, 3 are
parameters.
4. Matched uncertainty is a type of structured uncertainty that can be matched by
the control input for a class of MIMO linear affine-in-control systems of the form
ẋ = f (x) + B u + Θ ∗ Φ (x) (5.3)
5.1 Composition of a Model-Reference Adaptive Control System 87
−1
where B B B is the right pseudo-inverse of a full-rank non-square “wide”
−1
matrix B ∈ R × R with n < m and rank (B) = n, or where B B B
n m
=
B −1 for a full-rank square matrix B.
6. Control input uncertainty is a type of uncertainty that exists in the control input
matrix for a class of MIMO linear affine-in-control systems of the form
should capture all important performance specifications such as rise time and settling
time, as well as robustness specifications such as phase and gain stability margins.
Example 5.8 For Example 5.1, the reference model for an adaptive control system
could be selected to be a second-order system as
where xm (t) is a model-reference signal that only depends on the reference command
input r (t).
The tracking error is defined as
e = xm − x (5.8)
5.1.3 Controller
ẋ = ax + b [u + f (x)] (5.9)
p
f (x) = θi∗ φi (x) = Θ ∗ Φ (x) (5.10)
i=1
where Θ ∗ = θ1 θ2 . . . θ p ∈ R p is an unknown constant vector, and Φ (x) =
φ1 (x) φ2 (x) . . . φ p (x) ∈ R p is a vector of known bounded basis functions.
ẋm = am xm + bm r (5.11)
where the superscript ∗ denotes ideal constant values which are unknown.
Upon substituting into the plant model, we get the ideal closed-loop plant
ẋ = a + bk ∗ x + bkr∗r (5.13)
Comparing the ideal closed-loop plant to the reference model, the ideal gains k x∗
and kr∗ can be determined by the following model matching conditions:
a + bk x∗ = am (5.14)
bkr∗ = bm (5.15)
5.2 Direct MRAC for First-Order SISO Systems 91
It turns out that the solutions for k x∗ and kr∗ always exist since there are two
independent equations with two unknowns.
The actual adaptive controller is an estimate of the ideal controller with a goal
that in the limit the adaptive controller approaches the ideal controller. Let
be the adaptive controller, where k x (t), kr (t), and Θ (t) are the estimates of k x∗ , kr∗ ,
and Θ ∗ , respectively.
The adaptive controller is a direct adaptive controller since k x (t), kr (t), and Θ (t)
are estimated directly without the knowledge of the unknown system parameters a,
b, and Θ ∗ .
Now, define the estimation errors as
Let e (t) = xm (t) − x (t) be the tracking error. Then, the closed-loop tracking
error equation is established as
where γx > 0 and γr > 0 are called the adaptation (or learning) rates for k x (t) and
kr (t), and Γ = Γ > 0 ∈ R p × R p is a positive-definite adaptation rate matrix for
Θ (t).
92 5 Model-Reference Adaptive Control
V̇ e, k̃ x , k̃r , Θ̃ is evaluated as
2k̃ x k̃˙ x 2k̃r ˜k˙r −1 ˙
V̇ e, k̃ x , k̃r , Θ̃ = 2eė + |b| + + 2Θ̃ Γ Θ̃
γx γr
˙
k̃ ˙
k̃
x r
= 2am e2 + 2k̃ x −ebx + |b| + 2k̃r −ebr + |b|
γx γr
+ 2Θ̃ ebΦ (x) + |b| Γ −1 Θ̃˙ (5.23)
Since b = |b| sgnb, then V̇ e, k̃ x , k̃r , Θ̃ ≤ 0 if
k̃˙ x
− exsgnb + =0 (5.24)
γx
k̃˙r
− er sgnb + =0 (5.25)
γr
Because k x∗ , kr∗ , and Θ ∗ are constant, therefore from Eqs. (5.17)–(5.19) k̃˙ x = k̇ x ,
k̃˙r = k̇r , and Θ̃˙ = Θ̇. Thus, the following adaptive laws are obtained:
k̇ x = γx xesgnb (5.27)
Then,
V̇ e, k̃ x , k̃r , Θ̃ = 2am e2 ≤ 0 (5.30)
Since V̇ e, k̃ x , k̃r , Θ̃ ≤ 0, then e (t), k x (t), kr (t), and Θ (t) are bounded. Then,
lim V e, k̃ x , k̃r , Θ̃ = V e0 , k̃ x0 , k̃r0 , Θ̃0 + 2am e 22 (5.31)
t→∞
= e0 , k̃ x (0)
where e (0) = k̃ x0 , k̃r (0) = k̃r0 , and Θ̃ (0) = Θ̃0 .
So, V e, k̃ x , k̃r , Θ̃ has a finite limit as t → ∞. Since e 2 exists, therefore
e (t) ∈ L2 ∩ L∞ , but ė ∈ L∞ .
5.2 Direct MRAC for First-Order SISO Systems 93
V̇ e, k̃ x , k̃r , Θ̃ can be shown to be uniformly continuous by examining its deriv-
ative to see whether it is bounded. V̈ e, k̃ x , k̃r , Θ̃ is computed as
V̈ e, k̃ x , k̃r , Θ̃ = 4am eė = 4am e am e − bk̃ x x − bk̃r r + bΘ̃ Φ (x) (5.32)
Since e (t), k x (t), kr (t), and Θ (t) are bounded by the virtue that V̇ e, k̃ x , k̃r , Θ̃
≤ 0, x (t) is bounded because e (t) and xm (t) are bounded, r (t) is a bounded ref-
erence
command signal, and Φ (x) is bounded because x (t) is bounded; therefore,
V̈ e, k̃ x , k̃r , Θ̃ is bounded. Thus, V̇ e, k̃ x , k̃r , Θ̃ is uniformly continuous. It fol-
lows from the Barbalat’s lemma that V̇ e, k̃ x , k̃r , Θ̃ → 0, hence e (t) → 0 as
t → ∞. The tracking error is asymptotically stable, but the whole adaptive control
system is not asymptotically stable since k x (t), kr (t) and Θ (t) can only be shown
to be bounded.
If a and b are known, then the gains k x and kr need not be estimated since they are
known and can be found from the model matching conditions
am − a
kx = (5.33)
b
bm
kr = (5.34)
b
Then, the adaptive controller is given by
Θ̇ = −Γ Φ (x) eb (5.37)
Proof To show that the adaptive law is stable, choose a Lyapunov candidate function
94 5 Model-Reference Adaptive Control
V e, Θ̃ = e2 + Θ̃ Γ −1 Θ̃ > 0 (5.38)
Then,
V̇ e, Θ̃ = 2eė + 2Θ̃ Γ −1 Θ̃˙ = 2am e2 + 2ebΘ̃ Φ (x) − 2Θ̃ ebΦ (x) = 2am e2 ≤ 0
(5.39)
The Barbalat’s lemma can be used to show that the tracking error is asymptotically
stable, i.e., e (t) → 0 as t → ∞.
am − a
kx = = −2
b
bm
kr = =1
b
The adaptive controller is given by
u = −2x + r − θ x
θ̇ = −γ xeb
0.02 1
0 0.5
x, xm
e −0.02 0
−0.04 −0.5
x xm
−0.06 −1
0 10 20 30 40 50 0 10 20 30 40 50
t t
2 0.1
0 0.05
u
θ
−1 θ
*
θ
−2 0
0 10 20 30 40 50 0 10 20 30 40 50
t t
The results are shown in Fig. 5.3. It can be seen that e (t) → 0 and x (t) → xm (t)
as t → ∞. The estimate θ (t) also converges to the correct value of the uncertain
parameter θ ∗ , although the convergence is quite gradual. The convergence rate can
be increased by increasing the adaptation rate γ , but a large value of γ can lead to
an increase in the sensitivity of the control system to noise and unmodeled dynamics
that can lead to instability. In other words, a larger value of γ results in a better
tracking performance but at the same time degrades the system robustness. In a
practical design, the adaptation rate must be chosen carefully to maintain a sufficient
robustness while achieving a desired level of performance.
Consider the system in Sect. 5.2.1 with a and b unknown, but sign of b is known.
From the model matching conditions, if a and b can be estimated, then the gain k x
and kr can be obtained. Therefore, the objective of indirect adaptive control is to
estimate system parameters which are then used to update the control gains. Hence,
indirect adaptive control is essentially a parameter identification method.
Let
am − â (t)
k x (t) = (5.40)
b̂ (t)
bm
kr (t) = (5.41)
b̂ (t)
96 5 Model-Reference Adaptive Control
Let ã (t) = â (t) − a and b̃ (t) = b̂ (t) − b be the estimation errors. Now, the plant
model is expressed as
ẋ = ax + b̂ − b̃ u + Θ ∗ Φ (x) (5.42)
Then, substituting Eqs. (5.16), (5.40), and (5.41) into Eq. (5.42) yields
am − â bm
∗
ẋ = ax + b̂ x+ r − Θ Φ (x) + Θ Φ (x)
b̂ b̂
am − â bm
∗
− b̃ x+ r − Θ Φ (x) + Θ Φ (x)
b̂ b̂
am − â bm
= (am − ã) x + bm r − bΘ̃ Φ (x) − b̃ x+ r (5.43)
b̂ b̂
Let
ū = k x (t) x + kr (t) r (5.44)
The Lyapunov’s direct method is now used to find the adaptive laws as follows:
Proof Choose a Lyapunov candidate function
ã 2 b̃2
V e, ã, b̃, Θ̃ = e2 + + + |b| Θ̃ Γ −1 Θ̃ > 0 (5.46)
γa γb
where γa >0 and γb >0 are the adaptation rates for â (t) and b̂ (t), respectively.
Then, V̇ e, ã, b̃, Θ̃ is evaluated as
Then,
V̇ e, ã, b̃, Θ̃ = 2am e2 ≤ 0 (5.51)
Since V̇ e, ã, b̃, Θ̃ ≤ 0, then e (t), â (t), b̂ (t), and Θ (t) are bounded. Then,
lim V e, k̃ x , k̃r , Θ̃ = V e0 , ã0 , b̃0 , Θ̃0 + 2am e 22 (5.52)
t→∞
It should be noted that the possibility of b̂ (t) = 0 does exist, and in such a case,
the control gains will “blow up.” Thus, indirect MRAC may not be as robust as direct
MRAC. To prevent this from occurring, the adaptive law for b̂ (t) must be modified
so that the adaptation can take place in a closed subset of R that does not include
b̂ (t) = 0. One technique for modification is the projection method which assumes a
priori knowledge of b [2].
Suppose a lower bound on b is known, i.e., 0 < b0 ≤ |b|, then the adaptive law
can be modified by the projection method as
⎧ " " " " " "
" "
⎨ " " " " d "b̂"
−γb ūe if "b̂" > b0 or if "b̂" = b0 and ≥0
b̂˙ = dt (5.54)
⎩
0 otherwise
98 5 Model-Reference Adaptive Control
b̃˙
V̇ e, ã, b̃, Θ̃ = 2am e + 2b̃ ūe +
2
γb
⎧ " " " " " "
" "
⎪
⎨2am e2 ≤ 0 " " " " d "b̂"
if "b̂" ≥ b0 or if "b̂" = b0 and dt
≥0
= " " " "
" " (5.55)
⎪
⎩a e2 + 2b̃ūe if ""b̂"" = b and d "b̂" < 0
m 0 dt
" "
" " " "
" " d "b̂"
Consider the second case when "b̂" = b0 and dt < 0 for which the sign definite-
" "
" "
d "b̂"
ness of V̇ e, ã, b̃, Θ̃ is still undefined. The condition dt < 0 gives
" "
" "
d "b̂"
˙
= b̂sgnb = −γb ūesgnb < 0 ⇒ ūesgnb > 0 (5.56)
dt
" "
" "
Since "b̂" = b0 , then
" "
" "
2b̃ūe = 2 b̂ − b ūe = 2 "b̂" sgnb − |b| sgnb ūe = 2 (b0 − |b|) ūesgnb (5.57)
5.3 Indirect MRAC for First-Order SISO Systems 99
Since |b| ≥ b0 , which implies |b| = b0 + δ > 0 where δ ≥ 0, then ūesgnb > 0
implies
2b̃ūe = 2 (b0 − b0 − δ) ūesgnb = −2δ ūesgnb ≤ 0 (5.58)
Therefore,
V̇ e, ã, b̃, Θ̃ = 2am e2 + 2b̃ūe = 2am e2 − 2δ ūesgnb ≤ 2am e2 ≤ 0 (5.59)
Using the usual Barbalat’s lemma, one can conclude that e (t) → 0 as t → ∞.
where ζ and ωn are unknown and f (y, ẏ) = Θ ∗ Φ (y, ẏ) is defined in a manner
similar to Eq. (5.10).
Let x1 (t) = y (t), x2 = ẏ (t), and x (t) = x1 (t) x2 (t) ∈ R2 . The state-space
form of the system is
ẋ = Ax + B u + Θ ∗ Φ (x) (5.61)
where
0 1 0
A= , B = (5.62)
−ωn2 −2ζ ωn b
Let xm 1 (t) = ym (t), xm 2 = ẏm (t), and xm (t) = xm 1 (t) xm 2 (t) ∈ R2 . A
reference model is given by
ẋm = Am xm + Bm r (5.63)
Comparing the ideal closed-loop plant to the reference model, the model matching
conditions are
A + B K x∗ = Am (5.65)
Bkr∗ = Bm (5.66)
Note that in general, one cannot always assume that K x∗ and kr∗ exist because A,
Am , B, and Bm may have different structures that do not allow the solutions for K x∗
and kr∗ to be determined. In most cases, if A and B are known, then K x∗ and kr∗ can
be designed by any standard non-adaptive control techniques to stabilize the closed-
loop system and enable it to follow a command. Then, Am and Bm can be computed
from A, B, K x∗ , and kr∗ .
Example 5.10 A second-order SISO system and a reference model are specified as
0 1 0 0 1 0
A= , B= , Am = , Bm =
−1 −1 1 −16 −2 2
Now, suppose
1 1
Am =
−16 −2
Then, the solution of K x∗ is the same (verify!), but the model matching condition
is not satisfied since
0 1
A + BK = = Am
−16 −2
Thus, it is important to state an explicit assumption that there exist constant but
unknown K x∗ and kr∗ such that the model matching conditions are satisfied. For a
second-order SISO system, the model matching conditions are satisfied if Am and
Bm have the same structures as those of A and B, respectively.
A full-state feedback adaptive controller is designed as
Let K̃ x (t) = K x (t) − K x∗ , k̃r (t) = kr (t) − kr∗ , and Θ̃ (t) = Θ (t) − Θ ∗ be the
estimation errors, then the closed-loop plant becomes
⎛ ⎞ ⎛ ⎞
⎜ ⎟ ⎜ ⎟
ẋ = ⎝ A + B K x∗ +B K̃ x ⎠ x + ⎝ Bkr∗ +B k̃r ⎠ r − B Θ̃ Φ (x) (5.68)
Am Bm
P Am + A
m P = −Q (5.71)
2k̃r k̃˙r
V̇ e, K̃ x , k̃r , Θ̃ = ė Pe + e P ė + |b| 2 K̃ x Γx−1 K̃˙ x + −1 ˙
+ 2Θ̃ ΓΘ Θ̃
γr
(5.72)
Substituting the tracking error equation yields
V̇ e, K̃ x , k̃r , Θ̃ = e P Am + A
m P e + 2e P B − K̃ x x − k̃r r + Θ̃ Φ (x)
−1 ˙ 2k̃r k̃˙r −1 ˙
+ |b| 2 K̃ x Γx K̃ x + + 2Θ̃ ΓΘ Θ̃ (5.73)
γr
or
V̇ e, K̃ x , k̃r , Θ̃ = −e Qe + 2 |b| K̃ x −xe P̄sgnb + Γx−1 K̃˙ x
k̃˙r
+ 2 |b| k̃r −r e P̄sgnb +
γr
+ 2 |b| Θ̃ Φ (x) e P̄sgnb + ΓΘ−1 Θ̃˙ (5.76)
It follows that
V̇ e, K̃ x , k̃r , Θ̃ = −e Qe ≤ −λmin (Q) e 22 ≤ 0 (5.80)
Since V̇ e, K̃ x , k̃r , Θ̃ ≤ 0, therefore e (t), K x (t), kr (t), and Θ (t) are bounded.
Then,
lim V e, K̃ x , k̃r , Θ̃ = V e0 , K̃ x0 , k̃r0 , Θ̃0 − λmin (Q) e 22 (5.81)
t→∞
So, V e, K̃ x , k̃r , Θ̃ has a finite limit as t → ∞. Since e 2 exists, therefore
e (t) ∈ L2 ∩ L∞ , but ė ∈ L∞ .
V̇ e, K̃ x , k̃r , Θ̃ can be shown to be uniformly continuous by examining its
derivative to see whether it is bounded, where
V̈ e, K̃ x , k̃r , Θ̃ = −ė Qe − e Q ė = −e Q A + A Q e
− 2e Q Am e − B K̃ x x − B k̃r r + B Θ̃ Φ (x) (5.82)
5.4 Direct MRAC for Second-Order SISO Systems 103
Since e (t), K x (t), kr (t), and Θ (t) are bounded by the virtue of V̇ e, K̃ x , k̃r , Θ̃
≤ 0, x (t) is bounded because e (t) and xm (t) are bounded, r (t) is a bounded
reference
command signal, and Φ (x) is bounded
x (t) is bounded, therefore
because
V̈ e, K̃ x , k̃r , Θ̃ is bounded. Thus, V̇ e, K̃ x , k̃r , Θ̃ is uniformly continuous. It
follows from the Barbalat’s lemma that V̇ e, K̃ x , k̃r , Θ̃ → 0, hence e (t) → 0 as
t → ∞. Therefore, the tracking error is asymptotically stable.
bm
kr∗ = =4
b
Let Q = I , then the solution of the Lyapunov equation (5.71) is
$ % $ %
3 1 1
P= 2 8
1 5
⇒ P̄ = 8
5
8 16 16
Let Γx = diag γx1 , γx2 and ΓΘ = diag γθ1 , γθ2 . Then, the adaptive laws are
$ % $ %
x1 1
k̇ x1 γ 0
K̇ x =
= Γx xe P̄ sgn (b) = x1 e1 e2 8
k̇ x2 0 γx2 x2 5
16
1
1 5 γx1 x 1
= e1 + e2
8 16 γx2 x 2
104 5 Model-Reference Adaptive Control
1 5
k̇r = γr r e P̄sgnb = e1 + e2 γr r
8 16
$ %
θ̇1 γ θ1 0 1 18
Θ̇ = = −ΓΘ Φ (x) e P̄sgnb = − e1 e2
θ̇2 0 γ θ2 x12 5
16
1 5 γ θ1
=− e1 + e2
8 16 γ 2
θ2 x 1
0.1 0.5
0.05
1
0 0
e
−0.05
−0.1 −0.5
0 20 40 60 0 20 40 60
t t
2 x 4 x
x m x m
1 1 2 2
1 2
1
2
x , xm
2 m
x ,x
0 0
1
−1 −2
−2 −4
0 20 40 60 0 20 40 60
t t
For simulations, all adaptation rates are chosen to be 100 and all initial conditions
are set to zero. The results are shown in Figs. 5.4 and 5.5.
It is noted that the plant follows the reference model very well only after a short
time, but the adaptive gains K x (t) and kr (t) and the adaptive parameter Θ (t) are
converging much more slowly. When they converge, some of them do not converge
to their true values, such as k x2 (t) and kr (t). This is one of the properties of MRAC
whereby there is no assurance on the convergence of adaptive parameters to their
true values.
5.4 Direct MRAC for Second-Order SISO Systems 105
2
K ,k,Θ
r
0
x
−2
−4
kx kx kr θ1 θ2
1 2
−6
0 50 100 150 200
t
If A and B are known, then it is assumed that there exist K x and kr that satisfy the
model matching conditions
A + B K x = Am (5.83)
Bkr = Bm (5.84)
u = K x x + kr r − Θ Φ (x) (5.85)
ė = Am e + B Θ̃ Φ (x) (5.87)
106 5 Model-Reference Adaptive Control
Θ̇ = −Γ Φ (x) e P B (5.91)
Then,
V̇ e, Θ̃ = −e Qe ≤ −λmin (Q) e 2 (5.92)
Therefore, e (t) and Θ (t) are bounded. Using the same argumentwith the
Barbalat’s lemma as in the previous sections, one can conclude that V̇ e, Θ̃ is
uniformly continuous so V̇ e, Θ̃ → 0 as t → ∞. Therefore, the tracking error is
asymptotically stable with e (t) → 0 as t → ∞.
Indirect MRAC for second-order systems is similar to that for first-order systems.
Consider the second-order system in Sect. 5.4.1 with A and B unknown, but sign
of b is known. Assuming that there exist K x and kr that satisfy the model matching
conditions, and furthermore that Am and Bm have the same structures as those of A
and B, respectively, then A and B can be estimated. Let
0 1 0 0 1 0
A= , B= , Am = , B=
−ωn2 −2ζ ωn b −ωm2 −2ζm ωm bm
(5.93)
5.5 Indirect MRAC for Second-Order SISO Systems 107
where  (t), B̂ (t), ω̂n (t), and ζ̂ (t) are estimates of A, B, ωn , and ζ , respectively.
Let à (t) =  (t) − A and B̃ (t) = B̂ (t) − B be the estimation errors. Now, the
plant model is expressed as
ẋ = Â − Ã x + B̂ − B̃ u + Θ ∗ Φ (x) (5.98)
Then, substituting Eqs. (5.67), (5.96), and (5.97) into Eq. (5.98) yields
ẋ = Â − Ã x + B̂ K x x + kr r − Θ Φ (x) + Θ ∗ Φ (x)
− B̃ K x x + kr r − Θ Φ (x) + Θ ∗ Φ (x)
⎛ ⎞
Let
ū = K x (t) x + kr (t) r (5.100)
B̃ B̃
V e, Ã, B̃, Θ̃ = e Pe + trace ÃΓ A−1 Ã + + |b| Θ̃ ΓΘ−1 Θ̃ (5.102)
γb
108 5 Model-Reference Adaptive Control
trace C D = C D = D C (5.104)
n
C D = DC = ci di (5.105)
i=1
& '
C D = ci d j , i, j = 1, 2, . . . , n (5.106)
The trace operator is the sum of all the diagonal elements. Therefore,
j
trace C D = ci di = C D = D C (5.107)
i=1
and
2 e P B Θ̃ Φ (x) = 2Θ̃ Φ (x) e P B = 2Θ̃ Φ (x) e P̄ |b| sgn (b) (5.110)
since the terms e P B and e P B̃ are scalar quantities (verify!), where P̄ is defined
previously. Therefore,
5.5 Indirect MRAC for Second-Order SISO Systems 109
B̃˙
V̇ e, Ã, B̃, Θ̃ = −e Qe + trace 2 Ã xe P + Γ A−1 Ã˙ +2 B̃ Peū +
γb
+ 2 |b| Θ̃ Φ (x) e P̄sgnb + ΓΘ−1 Θ̃˙ (5.111)
It follows that e (t), Ã (t), B̃ (t), and Θ̃ (t) are bounded since
V̇ e, Ã, B̃, Θ̃ = −e Qe ≤ −λmin (Q) e 2 (5.115)
V e, Ã, B̃, Θ̃ has a finite limit as t → ∞ since
( ∞
V (t → ∞) = V (t0 ) − λmin (Q) e 2 dt < ∞ (5.116)
t0
It can be shown that V̇ e, Ã, B̃, Θ̃ is uniformly continuous because
V̈ e, Ã, B̃, Θ̃ is bounded. Then, applying the Barbalat’s lemma, one can conclude
that the tracking error is asymptotically stable with e (t) → 0 as t → ∞.
Let
Āˆ = 0 1  = 0 1
0 1
= −ω̂n2 −2ζ̂ ω̂n (5.117)
−ω̂n2 −2ζ̂ ω̂n
and since
0
b̂ = 0 1 B̂ = 0 1 (5.118)
b̂
then the adaptive laws can be expressed in terms of the estimates of the unknown
quantities ωn and ζ as
˙
Āˆ = −Γ A xe P
0
= −Γ A xe P̄ (5.119)
1
110 5 Model-Reference Adaptive Control
b̂˙ = −γb 0 1 Peū = −γb P̄ eū = −γb ūe P̄ (5.120)
Let
γω 0
ΓA = >0 (5.121)
0 γζ
Then,
d 2
−ω̂n = −γω x1 e P̄ (5.122)
dt
or
γω x1 e P̄
ω̂˙ n = (5.123)
2ω̂n
and
d
−2ζ̂ ω̂n = −2ω̂n ζ̂˙ − 2ζ̂ ω̂˙ n = −γζ x2 e P̄ (5.124)
dt
or
γζ x2 ω̂n − γω x1 ζ̂ e P̄
ζ̂˙ = (5.125)
2ω̂n2
To prevent the possibility of ω̂n (t) = 0 or b̂ (t) = 0 that will cause the adaptive
laws to blow up, both the adaptive laws for estimating ω̂n (t) and b̂ (t) need to be
modified by the projection method according to
)
γω x1 e P̄
if ω̂n > ω0 > 0 or if ω̂n = ω0 and ω̂˙ n ≥ 0
ω̂˙ n = 2ω̂n (5.126)
0 otherwise
⎧ " " " " " "
" "
⎨ " " " " d "b̂"
−γb ūe P̄ if "b̂" > b0 or if "b̂" = b0 and ≥0
b̂˙ = dt (5.127)
⎩
0 otherwise
In the modified adaptive law for ω̂n (t), it is assumed that ω̂n (t) is always a positive
quantity for a physically realizable system.
ẋ = Ax + BΛ [u + f (x)] (5.128)
5.6 Direct MRAC for MIMO Systems 111
ẋm = Am xm + Bm r (5.131)
Firstly, it must be assumed that there exist ideal control gains K x∗ and K r∗ such that
the following model matching conditions are satisfied:
A + BΛK x∗ = Am (5.132)
BΛK r∗ = Bm (5.133)
Then,
14 1
K x∗ = (BΛ)−1 (Am − A) =
−15 −1
2 −1
K r∗ = (BΛ) −1
Bm =
0 1
Define an adaptive controller as
Proof To derive the adaptive laws, choose the following Lyapunov candidate func-
tion:
V e, K̃ x , K̃ r , Θ̃ = e Pe + trace |Λ| K̃ x Γx−1 K̃ x + trace |Λ| K̃ r Γr−1 K̃ r
+ trace |Λ| Θ̃ ΓΘ−1 Θ̃ (5.137)
V̇ e, K̃ x , K̃ r , Θ̃ is evaluated as
V̇ e, K̃ x , K̃ r , Θ̃ = −e Qe + 2e P −BΛ K̃ x x − BΛ K̃ r r + BΛΘ̃ Φ (x)
+ 2trace |Λ| K̃ x Γx−1 K̃˙ x
+ 2trace |Λ| K̃ r Γr−1 K̃˙ r + 2trace |Λ| Θ̃ ΓΘ−1 Θ̃˙
(5.138)
Utilizing the trace property trace C D = D C and Λ = sgnΛ |Λ| where
sgnΛ = diag(sgnλ1 , sgnλ2 , . . . , sgnλm ), then notice that
5.6 Direct MRAC for MIMO Systems 113
e P BΛ K̃ x x = e P BsgnΛ |Λ| K̃ x x = trace |Λ| K̃ x xe P BsgnΛ (5.139)
e P BΛ K̃ r r = trace |Λ| K̃ r r e P BsgnΛ (5.140)
e P BΛΘ̃ Φ (x) = trace |Λ| Θ̃ Φ (x) e P BsgnΛ (5.141)
Then,
V̇ e, K̃ x , K̃ r , Θ̃ = −e Qe + 2trace |Λ| K̃ x −xe P BsgnΛ + Γx−1 K̃˙ x
+ 2trace |Λ| K̃ r −r e P BsgnΛ + Γr−1 K̃˙ r
+ 2trace |Λ| Θ̃ Φ (x) e P BsgnΛ + ΓΘ−1 Θ̃˙ (5.142)
K̇ r = Γr r e P BsgnΛ (5.144)
It follows that e (t), K̃ x (t), K̃ r (t), and Θ̃ (t) are bounded since
V̇ e, K̃ x , K̃ r , Θ̃ = −e Qe ≤ −λmin (Q) e 2 ≤ 0 (5.146)
Using the usual argument with the Barbalat’s lemma, the tracking error is asymp-
totically stable with e (t) → 0 as t → ∞.
Example 5.13 Let x (t) = x1 (t) x2 (t) , u (t) = u 1 (t) u 2 (t) , Φ (x) =
2 2
x1 x2 , A is unknown, Λ is unknown but Λ > 0 so sgnΛ = I , and B is known
and given by
11
B=
01
where ζm = 0.5, ωm = 2, bm = 4, and r (t) = sin 2t. For simulation purposes, the
true A, Λ, and Θ ∗ matrices are
114 5 Model-Reference Adaptive Control
4
1 1 0 0.2 0
A= , Λ = 5 4 Θ∗ =
−1 −1 0 5 0 −0.1
$ %−1
4 4
0 −5
K r∗ = (BΛ)−1 Bm = 5 5
=
0 45 4 5
Let Γx = diag γx1 , γx2 , Γr = γr , and ΓΘ = diag γθ1 , γθ2 . Then, the adaptive
laws are
$ %
γx1 0 x1 23 13
K̇ x = Γx xe P B sgnΛ = e1 e2 8
0 γx2 x2 1 7
8 16
I
$ 3 13 %
γx1 x 1 e1 + 18 e2 γx1 x1 8 e1 + 7
e
16 2
= 23
γx2 x 2 e + 18 e2 γx2 x2 13
2 1
e +
8 1
7
e
16 2
$ %
3 13
K̇ r
= Γr r e P BsgnΛ = γr r e1 e2 2 8
1 7
= γr r 23 e1 + 18 e2 γr r 13
8 e1 +
7
16 e2
8 16
2 $ %
x1 23 13
γ θ1 0
Θ̇ = −ΓΘ Φ (x) e P BsgnΛ = − e1 e2 8
0 γ θ2 x22 1 7
8 16
$ %
γθ1 x12 23 e1 + 18 e2 γθ1 x12 13 e + 16
8 1
7
e2
=−
γθ2 x2 2 e1 + 8 e2 γθ2 x2 8 e1 + 16 e2
2 3 1 2 13 7
For simulation, all adaptation rates are chosen to be 10 and all initial conditions
are set to zero. The results are shown in Figs. 5.6, 5.7, 5.8, and 5.9.
It can be seen that the tracking error tends to zero so that x (t) follows xm (t). Also
note that K r (t) and some elements of K x (t) do not converge to their ideal values.
5.6 Direct MRAC for MIMO Systems 115
0.5 2
2
e 0 0
e
−1
−0.5 −2
0 10 20 30 40 0 10 20 30 40
t t
2 x 4 x
x m x m
1 1 2 2
1 2
1
2
m
x , xm
x ,x
0 0
1
2
−1 −2
−2 −4
0 10 20 30 40 0 10 20 30 40
t t
0
x
K
−1
−2
−3
kx kx kx kx
11 12 21 22
−4
0 20 40 60 80 100
t
2
Kr
−2
−4
kr k
r
1 2
−6
0 20 40 60 80 100
t
1.5
0.5
Θ
−0.5
−1
−1.5
θ θ θ θ
11 12 21 22
−2
0 20 40 60 80 100
t
A + B K x = Am (5.148)
B K r = Bm (5.149)
Let Θ̃ (t) = Θ (t) − Θ ∗ be the estimation error, then the closed-loop plant model
is ⎛ ⎞
ẋ = ⎝ A + B K x ⎠ x + B K r r − B Θ̃ Φ (x) (5.151)
Am Bm
Then,
V̇ e, Θ̃ = −e Qe + 2e P B Θ̃ Φ (x) + 2trace Θ̃ Γ Θ̃˙
= −e Qe + 2trace Θ̃ Φ (x) e P B + Γ Θ̃˙ (5.154)
Using the Barbalat’s lemma, the tracking error can be shown to be asymptotically
stable with e (t) → 0 as t → ∞.
118 5 Model-Reference Adaptive Control
For the system in Sect. 5.6.1 with A and Λ unknown but B and sign of Λ known,
assuming that BΛ ∈ Rn × Rm is invertible and n ≤ m, then there exist K x∗ and K r∗
that satisfy the model matching conditions such that
K r∗ = (BΛ)−1 Bm (5.157)
If  (t) and Λ̂ (t) are estimates of A and Λ, then the estimates of K x∗ and K r∗ are
given by
−1
K x (t) = B Λ̂ (t) Am − Â (t) (5.158)
−1
K r (t) = B Λ̂ (t) Bm (5.159)
−1
Note that if n < m, then B Λ̂ (t) is defined by the right pseudo-inverse
−1
Λ̂ (t) B B Λ̂ (t) Λ̂ (t) B .
Let the adaptive controller be
Let à (t) =  (t) − A and Λ̃ (t) = Λ̂ (t) − Λ be the estimation errors. Then, the
closed-loop plant model is expressed as
ẋ = Ax + B Λ̂ − Λ̃ K x x + K r r − Θ̃ Φ (x)
= A + Am − Â x + Bm r − B Λ̃ (K x x + K r r ) − BΛΘ̃ Φ (x) (5.161)
Let
ū = K x x + K r r (5.162)
Then,
V̇ e, Ã, B̃, Θ̃ = −e Qe + 2e P Ãx + B Λ̃ū + BΛΘ̃ Φ (x)
+ 2trace ÃΓ A−1 Ã˙ + 2trace Λ̃ΓΛ−1 Λ̃˙ + 2trace |Λ| Θ̃ ΓΘ−1 Θ̃˙ (5.165)
we obtain
V̇ e, Ã, B̃, Θ̃ = −e Qe + 2trace à xe P + Γ A−1 Ã˙
+ 2trace Λ̃ ūe P B + ΓΛ−1 Λ̃˙
+ 2trace |Λ| Θ̃ Φ (x) e P Bsgn (Λ) + ΓΘ−1 Θ̃˙ (5.169)
5.8 Summary
In situations when the system uncertainty may become significant beyond a level of
desired tolerance that can adversely affect the performance of a controller, adaptive
control can play an important role in reducing the effects of the system uncertainty on
the controller performance. Situations that may warrant the use of adaptive control
could include unintended consequences of off-nominal modes of operation such
as system failures or highly uncertain operating conditions, and complex system
behaviors that can result in an increase in the complexity and hence cost of the
modeling efforts.
There are generally two classes of adaptive control: (1) direct adaptive control
and (2) indirect adaptive control. Adaptive control architectures that combine both
types of adaptive control are also frequently used and are referred to as compos-
ite, combined, or hybrid direct–indirect adaptive control. Adaptive control can deal
with either linear or nonlinear plants with various types of uncertainty which can be
structured uncertainty, unstructured uncertainty, or unmodeled dynamics. Matched
uncertainty is a type of structured uncertainty that can be matched by the control
input for a class of MIMO linear affine-in-control systems. Adaptive control sys-
tems can be designed to provide cancellation of a matched uncertainty. When an
uncertainty cannot be matched, it is called an unmatched uncertainty. Adaptive con-
trol systems can be designed to accommodate unmatched uncertainty, but cannot
cancel the unmatched uncertainty in general. Control input uncertainty is a type of
uncertainty that exists in the control input matrix for a class of MIMO linear affine-
in-control systems. Control input uncertainty can be in the amplitude or in the sign or
both. When the control input uncertainty is in the amplitude, a control saturation can
occur and may worsen the performance of a controller. When the input uncertainty is
in the sign, a control reversal can occur and potentially can cause instability. Control
input uncertainty in general presents more challenges to adaptive control designers.
A reference model is used to specify a desired response of an adaptive control
system to a command input. It is essentially a command shaping filter to achieve
a desired command following. Since adaptive control is formulated as a command
following or tracking control, the adaptation is operated on the tracking error between
the reference model and the system output. A reference model must be designed
properly for an adaptive control system to be able to follow.
5.8 Summary 121
Various direct and indirect model-reference adaptive control techniques for first-
order and second-order SISO systems and MIMO systems are presented. MRAC
can be shown to achieve asymptotic tracking, but it does not guarantee that adaptive
parameters converge to their true values. The Lyapunov stability theory shows that
adaptive parameter estimation errors are only bounded but not asymptotic.
5.9 Exercises
ẋm = am xm + bm r
where am < 0 and bm are known, and r (t) is a bounded command signal.
a. Design and implement in Simulink a direct adaptive controller that enables
the plant output x (t) to track the reference model signal xm (t), given b = 2,
am = −1, bm = 1, and r (t) = sin t. For adaptation rates, use γx = 1 and
γ = 1. For simulation purposes, assume a = 1 and θ ∗ = 0.2 for the
unknown parameters. Plot e (t), x (t), xm (t), u (t), and θ (t) for t ∈ [0, 50].
b. Show by the Lyapunov stability analysis that the tracking error is asymptot-
ically stable, i.e., e (t) → 0 as t → ∞.
c. Repeat part (a) for r (t) = 1 (t) where 1 (t) is the unit step function. Plot the
same sets of data as in part (a). Comment on the convergence of k x (t) and
θ (t) to the ideal values k x∗ and θ ∗ .
2. Consider the following first-order plant
ẋ = ax + b u + θ ∗ φ (x)
where ζ and ωn are unknown, but b is known. Show by applying the Barbalat’s
lemma that the tracking error is asymptotically stable.
Design a direct adaptive controller for a second-order system using the following
information: b = 1, ζm = 0.5, ωm = 2, bm = 4, r (t) = sin 2t, and
1
Φ (y) =
y2
where a , b, and θ ∗ are unknown, but b > 0 is known, and φ (t) = sin 2t − cos 4t.
Design an indirect adaptive controller in Simulink by estimating a, b, and θ ∗ so
that the plant follows a reference model
ẋm = am xm + bm r
design a direct adaptive controller in Simulink for the MIMO system to follow a
second-order SISO system specified by
ẋm = Am x + Bm r
and all initial conditions are assumed to be zero. Use Γx = ΓΘ = 10I . Plot the
time histories of e (t), x (t) vs. xm (t), K x (t), and Θ (t) for t ∈ [0, 100].
References
1. Åström, K. J., & Wittenmark, B. (2008). Adaptive control. New York: Dover Publications Inc.
2. Ioannu, P. A., & Sun, J. (1996). Robust adaptive control. Upper Saddle River: Prentice-Hall Inc.
3. Lavretsky, E. (2009). Combined/composite model reference adaptive control. IEEE Transactions
on Automatic Control, 54(11), 2692–2697.
4. Ishihara, A., Al-Ali, K., Kulkarni, N., & Nguyen, N. (2009). Modeling Error Driven Robot
Control, AIAA Infotech@Aerospace Conference, AIAA-2009-1974.
Chapter 6
Least-Squares Parameter Identification
y = Θ ∗ Φ (x) (6.1)
ŷ = Θ Φ (x) (6.2)
= ŷ − y = Θ Φ (x) − y (6.3)
6.1 Least-Squares Regression 127
1
N
J (Θ) = i (6.4)
2 i=1 i
∂ J ∂i N N
= ∇ JΘ (Θ) =
i = Φ (xi ) Φ (xi ) Θ − yi = 0 (6.5)
∂Θ i=1
∂Θ i=1
Θ = A−1 B (6.6)
where
N
A= Φ (xi ) Φ (xi ) (6.7)
i=1
N
B= Φ (xi ) yi (6.8)
i=1
p
y = θ0 + θ1 x + · · · + θ p x p = θ j x j = Θ Φ (x)
j=0
where Θ = θ0 θ1 . . . θ p and Φ (x) = 1 x . . . x p .
AΘ = B
128 6 Least-Squares Parameter Identification
where
⎡ ⎤
1
N ⎢
⎥
p
N xi
⎢ ⎥
A= Φ (xi ) Φ (xi ) = ⎢ .. ⎥ 1 xi · · · xi
⎣ . ⎦
i=1 i=1
p
xi
⎡ N N N p ⎤
i=1 1 i=1 x i ··· xi
⎢ N xi Ni=1 p+1 ⎥ N
N 2
···
⎢ i=1 i=1 x i i=1 x i ⎥ j+k
=⎢ .. .. .. .. ⎥= xi
⎣ . . . . ⎦ i=1 jk
N p N p+1 N 2 p
i=1 x i i=1 x i ··· i=1 x i
⎡ ⎡ N⎤ ⎤
1 yi N
⎢ N xi yi ⎥
i=1
N N ⎢ ⎥xi j
⎢ ⎥ ⎢ i=1 ⎥
B= Φ (xi ) yi = ⎢ ⎥ yi = ⎢
.. .. ⎥= xi yi
⎣ ⎦ ⎣
. . ⎦
i=1 i=1
p N p
i=1 j
xi i=1 x i yi
where ⎡ N N 2 ⎤ ⎡ N ⎤
N x xi i=1 yi
N Ni=1 2i i=1
A = ⎣ i=1 xi xi N
xi3 ⎦ , B = ⎣ i=1
N
x y ⎦
N 2 i=1 N 3
i=1
N 4
N 2i i
i=1 x i i=1 x i i=1 x i i=1 x i yi
When all available data are used in the least-squares regression method, it is some-
times called a batch least-squares method. This is usually when there are sufficient
data over a given time interval and the estimates of the unknown coefficients are not
needed immediately at each time step.
When the estimates of the unknown coefficients are needed at each time step, Θ can
be estimated recursively using each pair of data (xi , yi ) at each time step.
Consider the following cost function:
1
J (Θ) = (6.9)
2
6.2 Convex Optimization and Least-Squares Gradient Method 129
Note that J (Θ) is convex since
1
[α + (1 − α) 1 ] [α + (1 − α) 1 ]
2
1 1
= α 2 + α (1 − α) 1 + (1 − α)2 1 1
2 2
1 2 1
= α − 2 1 + α 1 + (1 − α)2 1 1
(6.12)
2 2
1 1 1 1
α − 2 1 + α 1 + (1 − α) 1 1 ≤ α + (1 − α) 1 1 (6.13)
2 2 2 2
Θ̇ = −Γ ∇ JΘ (Θ) (6.20)
Example 6.2 For Example 6.1, the least-squares gradient method is expressed as
⎡ ⎤ ⎛⎡ ⎤⎡ ⎤ ⎡ ⎤⎞
θ̇0 1 x ··· xp θ0 y
⎢ θ̇1 ⎥ ⎜⎢ x x 2 · · · x p+1 ⎥ ⎢ θ1 ⎥ ⎢ x y ⎥⎟
⎢ ⎥ ⎜⎢ ⎥⎢ ⎥ ⎢ ⎥⎟
Θ̇ = ⎢ . ⎥ = −Γ ⎜⎢ . .. . . .. ⎥ ⎢ .. ⎥ − ⎢ .. ⎥⎟
⎣ .. ⎦ ⎝⎣ .. . . . ⎦ ⎣ . ⎦ ⎣ . ⎦⎠
θ̇ p p p+1
x x ··· x 2p
θp p
x y
Then,
V̇ Θ̃ = 2trace Θ̃ Γ −1 Θ̃˙ = −2trace Θ̃ Φ (x) Φ (x) Θ̃
= −2Φ (x) Θ̃ Θ̃ Φ (x) = −2 = −2
2 ≤ 0 (6.25)
Note that V̇ Θ̃ can only be negative semi-definite because V̇ Θ̃ can be zero
when Φ (x) = 0 independent of Θ̃ (t). One can establish that V Θ̃ has a finite
limit as t → ∞ since
∞
V (t → ∞) = V (t0 ) − 2
2 dt < ∞ (6.26)
t0
which implies
∞
2
2 dt = V (t0 ) − V (t → ∞) < ∞ (6.27)
t0
t+T t+T
α0 I ≤ Φ(x)Φ (x) dτ ≤ Φ (x) Φ(x)dτ I = β0 I (6.32)
t t
Then, if the persistent excitation condition is satisfied, V̇ Θ̃ is expressed as
2α0 V Θ̃
V̇ Θ̃ ≤ − (6.33)
λmax Γ −1
Thus, the least-squares gradient method is exponentially stable with the rate of
convergence of λ α0Γ −1 .
max ( )
Example 6.3 Consider a scalar estimation error equation
θ̃˙ = −γ φ 2 (t) θ̃
As t → ∞, θ̃˙ (t) → 0 but θ̃ (t) → e−γ = 0. Thus, φ (t) is not PE and does not
guarantee a parameter convergence. The only way to ensure that θ̃ (t) → 0 is for
γ → ∞.
Consider another signal with zero value in some time interval such as
1 0≤t ≤1
φ (t) =
0 t >1
Then,
e−γ t 0 ≤ t ≤ 1
θ̃ (t) =
e−γ t > 1
Notice that Φ (t) Φ (t) is singular at any instant of time t. The PE condition is
evaluated as
Now, T must be chosen such that the PE condition is satisfied. Let T = 2π which
is the period of the signal. Then, Φ (t) is PE since
# $
t+T 2π cos (t) − cos (t + 2π )
Φ (τ ) Φ (τ ) dτ =
t cos (t) − cos (t + 2π ) 1 − 14 sin 2 (t + 2π ) + 41 sin 2t
# $ # $ # $
2π 0 2 0 1 0
= =π ≥π
0 π 0 1 0 1
∂ J t
∇ JΘ (Θ) = = Φ (x) Φ (x) Θ − y dτ = 0 (6.35)
∂Θ t0
# t $−1 t
Θ= Φ (x) Φ (x) dτ Φ (x) y dτ (6.36)
t0 t0
't
assuming the inverse of t0 Φ (x) Φ (x) dτ exists. Note that the matrix Φ (x) Φ (x)
is
' talways singular and is not invertible. However, if the PE condition is satisfied, then
t0 Φ (x) Φ (x) dτ is invertible.
Introducing a matrix R (t) = R (t) > 0 ∈ Rm × Rm where
# t $−1
R= Φ (x) Φ (x) dτ (6.37)
t0
then t
R −1 Θ = Φ (x) y dτ (6.38)
t0
d R −1
R −1 Θ̇ + Θ = Φ (x) y (6.39)
dt
From Eq. (6.37), we obtain
d R −1
= Φ (x) Φ (x) (6.40)
dt
Therefore,
Θ̇ = −RΦ (x) Φ (x) Θ − y = −RΦ (x) (6.41)
d R −1
Ṙ R −1 + R =0 (6.42)
dt
Thus,
Ṙ = −RΦ (x) Φ (x) R (6.43)
Both Eqs. (6.41) and (6.43) constitute the well-known recursive least-squares
(RLS) parameter identification method. The time-varying matrix R (t) is called
the covariance matrix, and the RLS formula is similar to the Kalman filter where
Eq. (6.43) is a differential Riccati equation for a zero-order plant model. Compar-
ing Eqs. (6.21)–(6.41), R (t) plays the role of Γ as a time-varying adaptation rate
matrix and Eq. (6.43) is effectively an adaptive law for the time-varying adapta-
tion rate matrix.
136 6 Least-Squares Parameter Identification
Let Θ̃ (t) = Θ (t) − Θ ∗ be the estimation error. Since (t) = Θ̃ (t) Φ (x), then
Then,
d R −1
V̇ Θ̃ = trace 2Θ̃ R −1 Θ̃˙ + Θ̃ Θ̃
dt
= trace −2Θ̃ Φ (x) Φ (x) Θ̃ + Θ̃ Φ (x) Φ (x) Θ̃
= −trace Θ̃ Φ (x) Φ (x) Θ̃ = − = −
2 ≤ 0 (6.46)
One can establish that V Θ̃ has a finite limit as t → ∞ since
∞
V (t → ∞) = V (t0 ) −
2 dt < ∞ (6.47)
t0
• Φ̇ (x) ∈ L∞ .
!' "−1
• t0 Φ (x) Φ (x) dτ
t
is invertible which implies Φ (x) is PE.
6.4 Recursive Least-Squares 137
If these conditions are satisfied, then using the Barbalat’s lemma, it can be shown
that (t) → 0 as t → ∞. In addition, Θ̃ (t) → 0 as t → ∞ and the parameter
convergence is achieved.
Note that there are various versions of the RLS method. One popular version is
the RLS method with normalization where the adaptive law for R (t) is modified as
follows:
RΦ (x) Φ (x) R
Ṙ = − (6.49)
1 + n2
RΦ (x) Φ (x) R
Ṙ = β R − (6.51)
1 + n2
Consider the following MIMO system with a matched uncertainty in Sect. 5.7 with
A ∈ Rn × Rn unknown, but B ∈ Rn × Rm known with n ≤ m [4]:
ẋ = Ax + B u + Θ ∗ Φ (x) (6.52)
ẋm = Am xm + Bm r (6.53)
where K r = B −1 Bm is known.
Then, the closed-loop plant model is given by
ẋ = A + Am − Â x + Bm r + B Θ ∗ − Θ Φ (x) (6.56)
where
ū = K x x + K r r (6.57)
Thus, we see that the plant modeling error (t) between the true plant and the
actual plant actually is what really affects the tracking error e (t). So, if we can
minimize the plant modeling error, then the tracking error will also be minimized.
Let Ω̃ (t) = Ã (t) B Θ̃ (t) ∈ Rn × Rn+l and Ψ (x) = x Φ (x) ∈
Rn+1 where à (t) =  (t) − A and Θ̃ (t) = Θ (t) − Θ ∗ . Then, the plant modeling
error is expressed as
= Ω̃ Ψ (x) (6.61)
ė = Am e + Ω̃ Ψ (x) (6.62)
∂ J
Ω̇ = −Γ = −Γ Ψ (x) (6.64)
∂ Ω̃
For the standard RLS adaptive law, the constant adaptation rate matrix Γ is
replaced by the covariance matrix R (t) as follows:
We now combine the least-squares gradient adaptive law with the MRAC adap-
tive law to obtain the following adaptive law which we call combined least-squares
gradient MRAC adaptive law:
Ω̇ = −Γ Ψ (x) + e P (6.67)
Proof To prove stability of the combined RLS MRAC adaptive laws, we choose a
Lyapunov candidate function
140 6 Least-Squares Parameter Identification
V e, Ω̃ = e Pe + trace Ω̃ R −1 Ω̃ (6.69)
Then, V̇ e, Ω̃ is evaluated as
* +
dR −1
V̇ e, Ω̃ = −e Qe + 2e P + trace −2Ω̃ Ψ (x) + e P + Ω̃ Ω̃
dt
= −e Qe + 2e P − 2 + e P + Ψ (x) Ω̃ Ω̃ Ψ (x)
Note that V̇ e, Ω̃ ≤ 0 as opposed to V̇ e, Ω̃ < 0 because the condition of
Ψ (x) on which (t) depends on has not been imposed. Therefore, e (t) ∈ L2 ∩ L∞ ,
(t) ∈ L2 ∩ L∞, ė (t) ∈ L∞ , and Ω̃ ∈ L∞ . Applying the Barbalat’s lemma, we
evaluate V̈ e, Ω̃ as
! "
V̈ e, Ω̃ = −2e Q (Am e + ) − Ω̃˙ Ψ (x) + Ω̃ Ψ̇ (x)
! "
= −2e Q (Am e + ) − − (Pe + ) Ψ (x) RΨ (x) + Ω̃ Ψ̇ (x)
(6.71)
But, by definition
# t $−1
R= Ψ (x) Ψ (x) dτ (6.72)
t0
If Ψ̇ (x) ∈ L∞ and if Ψ (x) satisfies the PE condition, then V̈ e, Ω̃ ∈ L∞ .
Therefore, V̇ e, Ω̃ is uniformly continuous. It follows that e (t) → 0, (t) → 0,
and Ω̃ (t) → 0 as t → ∞. Since Ω̃ (t) tends to zero exponentially because of
the PE condition, e (t) also tends to zero exponentially. Thus, the system achieves
exponential stability.
α0 λmax (P)
Choose Q and β0 from Eq. (6.32) such that λmin (Q) = β0
. Then,
6.5 Indirect Adaptive Control with Least-Squares Parameter Identification 141
α0 λmax (P) & &2 α0 V e, Ω̃
& &
V̇ e, Ω̃ ≤ −
e
2 − α0 &Ω̃ & ≤ − (6.74)
β0 β0
α0
Thus, the system is exponentially stable with the convergence rate of 2β0
.
Note that for the combined least-squares gradient MRAC adaptive laws, it can be
shown that
V̇ e, Ω̃ = −e Qe − 2 ≤ −λmin (Q)
e
2 − 2
2 (6.75)
V̈ e, Ω̃ is evaluated as
! "
V̈ e, Ω̃ = −2e Q (Am e + ) − 2 − (Pe + ) Ψ (x) Γ Ψ (x) + Ω̃ Ψ̇ (x)
(6.76)
If Ψ̇ (x) ∈ L∞ , then V̈ e, Ω̃ is bounded. Therefore, V̇ e, Ω̃ is uniformly
continuous. It follows that e (t) → 0 and (t) → 0 as t → ∞. The rate of con-
vergence of (t) is twice as large as that of the combined RLS MRAC adaptive
laws. In addition, if Ψ (x) satisfies the PE condition, then the exponential parameter
convergence is achieved.
Example 6.5 Consider a first-order system with a matched uncertainty
ẋ = ax + b u + θ ∗ x 2
ẋm = am xm + bm r
= âx + bū − ẋ
am − â bm
ū = k x (t) x + kr r = x+ r
b b
Simulation results are shown in Figs. 6.1 and 6.2.
It can be seen that â (t) and θ (t) tends to their true values, and the tracking error
e (t) also tends to zero. The covariance matrix R (t) which acts as time-varying
adaptation rate matrix Γ converges to zero as t → ∞.
For comparison, the combined least-squares gradient MRAC method is used
# $ # $# $
â˙ γ11 0 x ( + e)
Ω̇ = =−
bθ̇ 0 γ22 x 2 ( + e)
where γa = γθ = 10.
The simulation results are shown in Fig. 6.3.
The tracking error tends to zero asymptotically much faster than the combined
RLS MRAC method. The combined least-squares gradient MRAC method gener-
ally provides a much better parameter convergence than the combined RLS MRAC
method.
0.2 1
x xm
0.1 0.5
x, xm
0 0
e
−0.1 −0.5
−0.2 −1
0 10 20 30 40 0 10 20 30 40
t t
1.5 0.3
1 0.2
0 0
−0.5 −0.1
0 10 20 30 40 0 10 20 30 40
t t
10
r
11
r
12
r
21
r22
R 5
−5
0 5 10 15 20 25 30 35 40
t
0.05 1
x xm
0 0.5
x, xm
−0.05 0
e
−0.1 −0.5
−0.15 −1
0 10 20 30 40 0 10 20 30 40
t t
1.5 0.3
1 0.2
0 0
−0.5 −0.1
0 10 20 30 40 0 10 20 30 40
t t
μ > 0.
with
V̇ e, Ω̃ is evaluated as
V̇ e, Ω̃ = −e Qe + 2e P + μtrace −2Ω̃ Ψ (x) + Ω̃ Ψ (x) Ψ (x) Ω̃
= −e Qe + 2e P − μ ≤ −λmin (Q)
e
2
+ 2λmax (P)
e
− μ
2 (6.78)
V̇ e, Ω̃ is now not negative semi-definite unconditionally. We note that
# $2
λ2max (P) λmax (P)
−
e
+2λmax (P)
e
−δ
= −
2 2 2
e
− δ
≤ 0
δ2 δ
(6.79)
Therefore, V̇ e, Ω̃ can be expressed as
# $
λ2 (P)
V̇ e, Ω̃ ≤ − λmin (Q) − max 2
e
2 − μ − δ 2
2 (6.80)
δ
√
Then, V̇ e, Ω̃ ≤ 0 if √λλmax (P) < δ < μ. Furthermore, if Ψ (x) satisfies the
min (Q)
PE condition, then from Eq. (6.32) we have
# $
λ2max (P) & &2
2 & &
V̇ e, Ω̃ ≤ − λmin (Q) −
e
2
− α0 μ − δ &Ω̃ & (6.81)
δ2
From Sect. 6.3, for the least-squares gradient method, if Ψ̇ (x) ∈ L∞ , then the
plant modeling error is asymptotically stable with (t) → 0 as t → ∞. This implies
that the tracking error is also asymptotically stable with e (t) → 0 as t → ∞.
6.5 Indirect Adaptive Control with Least-Squares Parameter Identification 145
However, for the RLS method, asymptotic tracking additionally requires the PE
condition, but it is important to recognize that the PE condition cannot be verified in
a closed-loop system. So, if the PE condition is not satisfied, then the plant modeling
error is only bounded for the RLS method. This implies that the tracking error is
only bounded but is not asymptotically stable.
Example 6.6 Consider Example 6.5
ẋ = ax + b u + θ ∗ x 2
with R (0) = I .
The simulation results are shown in Fig. 6.4.
0 1.5
−0.2
1
m
x, x
−0.4
e
0.5
−0.6
x xm
−0.8 0
0 10 20 30 40 0 10 20 30 40
t t
1 0.4
0.5 0.2
<
a
0 0
−0.5 −0.2
0 10 20 30 40 0 10 20 30 40
t t
For comparison, the least-squares gradient adaptive control method is also imple-
mented as # $ # $# $
â˙ γ 0 x
Ω̇ = = − 11
bθ̇ 0 γ22 x 2
with Γ = I .
146 6 Least-Squares Parameter Identification
The simulation results are shown in Fig. 6.5. It can be seen that the least-squares
gradient adaptive control achieves asymptotic tracking with e (t) → 0 as t → ∞.
Both the RLS and least-squares gradient adaptive control do not achieve a parameter
convergence with this reference command signal.
0 1.5
−0.2
1
x, xm
−0.4
e
0.5
−0.6
x xm
−0.8 0
0 20 40 0 20 40
t t
1 0.4
0.5 0.2
<
a
0 0
−0.5 −0.2
0 20 40 0 20 40
t t
x (t) − x (t − Δt)
x̂˙ (t) = (6.83)
Δt
to estimate ẋ (t) at a current time step using the current and past state information.
The disadvantage of numerical differentiation is the introduction of noise into the
signal.
Another approach is to sample data sufficiently and then use them to construct an
at least C 1 smooth time function using a cubic or B-spline method over a finite time
interval. This curve is then differentiated at the spline knots to find the estimated time
derivative values. This derivative estimate would tend to result in lower noise than the
backward finite-difference method [11]. In either case, the derivative computation
will introduce a noise source. Least-squares methods are generally robust to noise,
6.6 Estimation of Time Derivative Signals 147
If  (t) → A and Θ (t) → Θ ∗ , then x̂˙ (t) → ẋ (t). Thus, x̂˙ (t) is in fact an
estimate of ẋ (t). The estimate x̂˙ (t) only depends on the information of the current
state, the control, and the adaptive parameters. The predictor model thus can be used
to provide the estimate of the signal ẋ (t) without differentiation.
As the adaptation proceeds, the predictor model should converge to the plant
model with x̂ (t) → x (t) + e p (t) as t → ∞, where e p (t) = x̂ (t) − x (t) is the
predictor error whose dynamics are described by
If the predictor error converges to zero, i.e., e p (t) → 0, then p (t) → (t).
Therefore, the signal p (t) could be used in least-squares adaptive control in lieu of
the signal (t).
6.7 Summary
6.8 Exercises
3. Determine if the following functions are persistently exciting (PE), and if so,
determine T and α.
6.8 Exercises 149
where a and θ ∗ are unknown, but b = 2, and φ (t) = sin t. For simulation
purposes, use a = 1 and θ ∗ = 0.2. The reference model is given by
ẋm = am xm + bm r
References
1. Ioannu, P. A., & Sun, J. (1996). Robust adaptive control. Upper Saddle River: Prentice-Hall
Inc.
2. Slotine, J.-J., & Li, W. (1991). Applied nonlinear control. Upper Saddle River: Prentice-Hall
Inc.
3. Bobal, V. (2005). Digital self-tuning controllers: algorithms, implementation, and applications.
London: Springer.
4. Nguyen, N. (2013). Least-squares model reference adaptive control with chebyshev orthogonal
polynomial approximation. AIAA Journal of Aerospace Information Systems, 10(6), 268–286.
5. Chowdhary, G., & Johnson, E. (2011). Theory and flight-test validation of a concurrent-learning
adaptive controller. AIAA Journal of Guidance, Control, and Dynamics, 34(2), 592–607.
6. Chowdhary, G., Yucelen, T., & Nguyen, N. (2013). Concurrent learning adaptive control of the
aeroelastic generic transport model. In AIAA Infotech@Aerospace Conference, AIAA-2013-
5135.
7. Guo, L. (1996). Self-convergence of weighted least-squares with applications to stochastic
adaptive control. IEEE Transactions on Automatic Control, 41(1), 79–89.
8. Lai, T., & Wei, C.-Z. (1986). Extended least squares and their applications to adaptive control
and prediction in linear systems. IEEE Transactions on Automatic Control, 31(10), 898–906.
9. Suykens, J., Vandewalle, J., & deMoor, B. (1996). Artificial neural networks for modeling and
control of non-linear systems. Dordrecht: Kluwer Academic Publisher.
10. Zou, A., Kumar, K., & Hou, Z. (2010). Attitude control of spacecraft using chebyshev neural
networks. IEEE Transactions on Neural Networks, 21(9), 1457–1471.
11. Nguyen, N., Krishnakumar, K., Kaneshige, J., & Nespeca, P. (2008). Flight dynamics model-
ing and hybrid adaptive control of damaged asymmetric aircraft. AIAA Journal of Guidance,
Control, and Dynamics, 31(3), 751–764.
Chapter 7
Function Approximation and Adaptive
Control with Unstructured Uncertainty
y = f (x) (7.1)
where x (t) ∈ R p is the input and f (x) ∈ Rn is an unknown function but assumed
to be bounded function.
Any sufficiently smooth function f (x) ∈ C q can be expanded as a Taylor series
about some x = x̄
p
∂ f (x̄) 1 ∂ f (x̄)
p p
f (x) = f (x̄)+ (xi − x̄i )+ (xi − x̄i )(x j − x̄ j )+· · · (7.2)
i=1
∂ xi 2 i=1 j=1 ∂ xi x j
Then, ŷ (t) → f (x) as q → ∞. This means that any sufficiently smooth function
can be approximated by a polynomial of q-th degree. Then, the function approxima-
tion error could be made sufficiently small on a compact domain of x (t) such that
supx∈D ε∗ (x) ≤ ε0∗ for all x (t) ∈ D ⊂ R p .
Another type of polynomials that can be used for function approximation is a
class of orthogonal polynomials which form a true basis that spans a Hilbert inner
product space with the following inner product definition:
b
w (x) pi (x) p j (x) d x = δi j (7.6)
a
where δi j = 1 if i = j and δi j = 0 if i
= j.
There are many orthogonal polynomial classes. The Chebyshev polynomials are
one such polynomial class that is frequently used in function approximation. In
particular, the first few Chebyshev polynomials of the first kind are shown as follows:
T0 (x) = 1
T1 (x) = x
T2 (x) = 2x 2 − 1
T3 (x) = 4x 3 − 3x (7.7)
T4 (x) = 8x 4 − 8x 2 + 1
..
.
Tn+1 (x) = 2x Tn (x) − Tn−1 (x)
over a compact domain x (t) ∈ [−1, 1], where x (t) = sin 10t, with a 4th -degree
Chebyshev polynomial using the least-squares gradient method with ΓΘ = 2I and
Θ (0) = 0.
The approximation results are shown in Fig. 7.1 for 1000 s which shows an excel-
lent parameter convergence after only less than 100 s. The root mean square error is
3.8 × 10−4 .
154 7 Function Approximation and Adaptive Control with Unstructured Uncertainty
Θ
0
−1
θ1 θ2 θ3 θ4 θ5
−2
0 200 400 600 800 1000
t
2
y, ΘTΦ(x)
0 y
ΘTΦ(x)
−1
−1 −0.5 0 0.5 1
x
Neural network is a concept pioneered in the 1960s in the artificial intelligence fields
as an attempt to represent a simple model of how a human brain neuron connectivity
works. Neural networks have been used in many applications such as classifiers,
pattern recognition, function approximation, and adaptive control [1, 3–12]. The
terminology “intelligent control” has been used loosely to describe adaptive control
with neural networks. A neural network is a network of connectivity of neurons
that form various layers of neurons to describe a complex input–output mapping.
Figure 7.2 is a two-layer neural network that performs input–output mapping. A
neural network with connectivity progressing from inputs to an output is called
a feedforward neural network. A neural network with a closed-loop connectivity
between some elements of the neural network is called a recurrent neural network.
where W is called a weight matrix and b is a constant vector called a bias vector.
y = W Φ (x) = w2 x 2 + w1 x + w0
The most common activation functions are sigmoidal and radial basis functions
(RBF) since these functions possess a universal approximation property, which
156 7 Function Approximation and Adaptive Control with Unstructured Uncertainty
σ(x)
0.5
0
0
x
implies that any nonlinear function can be approximated by these functions. Note
that a polynomial in some sense can be viewed as a universal approximator since
any sufficiently smooth function can be represented by a Taylor series.
A sigmoidal function or also called a logistic function as shown in Fig. 7.5 is
described by
1
σ (x) = (7.10)
1 + e−ax
ψ (x) = e−ax
2
(7.12)
ψ(x)
0.5
0
0
x
or “off-line” neural network is one whose coefficients are estimated a priori from data
or model information. An “on-line learning” neural network is one whose coefficients
are recursively estimated usually from a gradient method such as “back propagation”
in real time.
Consider a sigmoidal neural network approximation. The gradient method can be
applied to derive least-squares learning laws for a neural network to minimize the
following cost function:
1 1
J= ε ε= Θ Φ W x̄ − y Θ Φ W x̄ − y (7.14)
2 2
Taking the partial derivatives of J with respect to the weight matrices and bias
vectors yields
∂ J ∂ε
∇ JΘ =
=
ε = Φ W x̄ ε (7.15)
∂Θ ∂Θ
∂J ∂ W x̄ ∂ε
∇ JW = =
ε = x̄ε V σ W x̄ (7.16)
∂W ∂W ∂ W x̄
where σ W x̄ = diag σ W1 x̄ , σ W2 x̄ , . . . , σ Wm x̄ is a diagonal
matrix with
ae−ax
σ (x) = (7.17)
(1 + e−ax )2
Then, the least-squares gradient learning laws for the neural network are given by
Θ̇ = −ΓΘ Φ W x̄ ε (7.18)
Ẇ = −ΓW x̄ε V σ W x̄ (7.19)
158 7 Function Approximation and Adaptive Control with Unstructured Uncertainty
over a compact domain x (t) ∈ [−1, 1], where x (t) = sin 10t.
y (t) is approximated by the following sigmoidal neural network:
ŷ = v0 + v1 σ w01 + w1 x + v2 σ w02 + w2 x + v3 σ w03 + w3 x
⎤ ⎡
v0 ⎡ ⎤
v1
⎢ v1 ⎥
Θ=⎢ ⎥ , V = ⎣ v2 ⎦ , W = w01 w02 w03
⎣ v2 ⎦ w1 w2 w3
v3
v3
⎡ ⎤
1
1 ⎢ σ w01 + w1 x ⎥
x̄ = , Φ W x = ⎢ ⎥
⎣ σ w02 + w2 x ⎦ ,
x
σ w03 + w3 x
⎡ ⎤
σ w01 + w1 x 0 0
σ W x̄ = ⎣ ⎦
0 σ w02 + w2 x 0
0 0 σ w03 + w3 x
Note that the initial conditions of Θ (t) and W (t) are usually set by a random
number generator. Otherwise, if they are zero, then V (t) and W (t) will have identical
elements, that is, v1 = v2 = v3 , w01 = w02 = w03 , and w1 = w2 = w3 . For this
example, let ΓΘ = ΓW = 10I . The neural network approximation is shown in
Fig. 7.7.
5
Θ, W
−5
0 200 400 600 800 1000
t
2
y, Θ Φ(W x)
|
T
1
T
0 y
Θ Φ(W −
T T
x)
−1
−1 −0.5 0 0.5 1
t
Note that the neural network approximation converges much more slowly than the
Chebyshev polynomial approximation in Example 7.1. Some of the neural network
weights are still converging even after 1000 s. The root mean square error is 6.9×10−4
which is about twice that for the Chebyshev polynomial approximation. The neural
network has ten weighting coefficients while the Chebyshev polynomial has only
five coefficients.
ẋ = Ax + B [u + f (x)] (7.20)
ẋm = Am xm + Bm r (7.23)
We assume that there exist ideal control gains K x∗ and K r∗ that satisfy the model
matching conditions
A + B K x∗ = Am (7.24)
B K r∗ = Bm (7.25)
where K r = K r∗ is known.
Let  (t) be an estimate of A, then
Let K̃ x (t) = K x (t) − K x∗ and Θ̃ (t) = Θ (t) − Θ ∗ be the estimation errors, then
ẋ = Am + B K̃ x x + Bm r − B Θ̃ Φ (x) − Bε∗ (7.29)
ẋd = Am x + Bm r (7.30)
The tracking error equation can be expressed in terms of the plant modeling error
as
ė = ẋm − ẋ = ẋm − ẋd + ẋd − ẋ = Am e + ε (7.33)
7.3 Adaptive Control with Unstructured Uncertainty 161
Note that K x (t) and Θ (t) require the matrix inverse B −1 . This sometimes can
cause some issues which will be discussed later.
The estimation error equation for the RLS adaptive law is obtained as
Ω̃˙ = −RΨ (x) Ψ (x) Ω̃ + ε∗ B (7.37)
μ > 0.
with
V̇ e, Ω̃ is evaluated as
V̇ e, Ω̃ = −e Qe + 2e Pε − με ε + με∗ B Bε∗
≤ −λmin (Q) e2 + 2λmax (P) e ε − μ [[ε]] 2 + μ B2 ε0∗2
(7.42)
λ2max (P)
2λmax (P) e ε ≤ e2 + δ 2 ε2 (7.43)
δ2
then V̇ e, Ω̃ can be expressed as
λ2max (P)
V̇ e, Ω̃ ≤ − λmin (Q) − e2 −(μ−δ 2 )Ψ (x)2 Ω̃2 +μ B2 ε0∗2
δ2
(7.44)
λmax (P) √
Thus, V̇ e, Ω̃ ≤ 0 if λ (Q) < δ < μ and is further bounded by
√
min
λ2max (P)
V̇ (e, Ω̃) ≤ − λmin (Q) − e2 + μB2 ε0∗2 (7.45)
δ2
7.3 Adaptive Control with Unstructured Uncertainty 163
Then, the largest lower bound of e is obtained by letting V̇ (e, Ω̃) ≤ 0 as
√
μ B ε0∗
e ≥ =p (7.46)
λ2max (P)
λmin (Q) − δ2
Similarly, V̇ e, Ω̃ is also bounded by
V̇ e, Ω̃ ≤ −(μ − δ 2 )Ψ (x)2 Ω̃2 + μB2 ε0∗2 (7.47)
Then, the largest lower bound of Ω̃ is obtained as
√
μBε0∗
Ω̃ ≥ =α (7.48)
Ψ0 μ − δ 2
Let D be a subset that contains all trajectories of e (t) and Ω̃ (t). Let Rα ⊂ D
and Rβ ⊂ D be two subsets in D defined by the lower bounds and upper bounds of
e and Ω̃ , respectively, where
! "
Rα = e (t) ∈ Rn , Ω̃ (t) ∈ Rn+l × Rn : e ≤ p, Ω̃ ≤ α (7.49)
! "
Rβ = e (t) ∈ Rn , Ω̃ (t) ∈ Rn+l × Rn : e ≤ ρ, Ω̃ ≤ β (7.50)
where ρ and β are the upper bounds of e and Ω̃ , respectively.
Then, we see that V̇ e, Ω̃ ≤ 0 for all e (t) ∈ D − Rα , that is, a region outside
Rα . Let Sα be the smallest subset that encloses Rα such that Rα ⊆ Sα . That is, Sα
is a hypersphere that circumscribes Rα which is given by
! "
Sα = e (t) ∈ Rn , Ω̃ (t) ∈ Rn+l × Rn : V e, Ω̃ ≤ λmax (P) p 2 + μλmax R −1 α 2
(7.51)
where λmax (R −1 ) is the largest eigenvalue of R −1 (t) for all t.
Let Sβ be the largest subset inside Rβ that encloses Sα such that Sβ ⊆ Rβ .
That is, Sβ is an inscribed hypersphere in Rβ which is given by
! "
Sβ = e (t) ∈ Rn , Ω̃ (t) ∈ Rn+l × Rn : V e, Ω̃ ≤ λmin (P) ρ 2 + μλmin R −1 β 2
(7.52)
where λmin (R −1 ) is the smallest eigenvalue of R −1 (t) for all t.
We see that Sα ⊆ Sβ . Then, the ultimate bound of e can be found by setting
Sα = Sβ [13]. Thus,
164 7 Function Approximation and Adaptive Control with Unstructured Uncertainty
λmin (P) ρ 2 + μλmin R −1 β 2 = λmax (P) p 2 + μλmax R −1 α 2 (7.53)
Since e ≤ ρ and Ω̃ ≤ β in Sβ , it follows that [13]
λmin (P) e2 ≤ λmax (P) p 2 + μλmax R −1 α 2 (7.54)
2
μλmin R −1 Ω̃ ≤ λmax (P) p 2 + μλmax R −1 α 2 (7.55)
Then, the ultimate bounds of e and Ω̃ are obtained as
#
λmax (P) p 2 + μλmax R −1 α 2
e ≤ =ρ (7.56)
λmin (P)
#
λmax (P) p 2 + μλmax R −1 α 2
Ω̃ ≤ =β (7.57)
μλmin R −1
The bounding sets for uniform ultimate boundedness are illustrated in Fig. 7.8.
In the RLS adaptive laws, the matrix inverse B −1 is used to compute the adaptive
parameters K (t) and Θ (t). This sometimes can cause issues when B −1 is small
7.3 Adaptive Control with Unstructured Uncertainty 165
Note that K x (t) and Θ (t) no longer require the matrix inverse B −1 .
The estimation error equation for the modified RLS adaptive law is obtained as
Ω̃˙ = −RΨ (x) Ψ (x) Ω̃ + ε∗ B B (7.62)
The tracking error equation can be expressed in terms of the plant modeling
error as
ė = ẋm − ẋ = ẋm − ẋd + ẋd − ẋ = Am e + ε = Am e + B Ω̃ Ψ (x) + Bε∗ (7.63)
with μ
> 0.
V̇ e, Ω̃ is evaluated as
V̇ e, Ω̃ = −e Qe + 2e P B Ω̃ Ψ (x) + 2e P Bε∗
− 2με ε + 2με Bε∗ + μηΨ (x) Ω̃ Ω̃ Ψ (x)
≤ −λmin (Q) e2 + 2 P B e Ψ (x) Ω̃ + 2 P B e ε0∗
2
− μ 2λmin B B − η Ψ (x)2 Ω̃
+ 2μλmax B B Ψ (x) Ω̃ ε0∗ (7.70)
' (
P B2
V̇ e, Ω̃ ≤ − λmin (Q) − e2 + 2P Beε0 ∗
δ2
) * 2
− μ 2λmin B B − η − δ 2 Ψ (x)2 Ω̃
+ 2μλmax B B Ψ (x) Ω̃ ε0∗ (7.72)
Thus, V̇ e, Ω̃ ≤ 0 if √P B < δ < μ[2λmin (B B) − η]. Upon completing
λmin (Q)
the square, we obtain
2
V̇ e, Ω̃ ≤ −c1 (e − c2 )2 + c1 c22 − c3 Ψ (x)2 Ω̃ − c4 + c3 c42 (7.73)
P B2 P Bε0∗
where c1 = λmin (Q) − δ2
, c2 = c1
, c3 = μ 2λmin (B B) − η − δ 2 , and
μλ (B B)ε∗
c4 = max c3 0
.
Then, the largest lower bound of e is obtained from
which yeilds #
c3 c42
e ≥ c2 + c22 + =p (7.75)
c1
168 7 Function Approximation and Adaptive Control with Unstructured Uncertainty
Similarly, the largest lower bound of Ω̃ is obtained from
2
V̇ e, Ω̃ ≤ −c3 Ψ (x)2 Ω̃ − c4 + c1 c22 + c3 c42 ≤ 0 (7.76)
which yields 2
c1 c2
c4 + + c42
Ω̃ ≥ c3
=α (7.77)
Ψ0
!
Thus, V̇ e, Ω̃ ≤ 0 outside a compact set defined by Rα = e (t) ∈ Rn , Ω̃ (t) ∈
"
Rn+l × Rn : e ≤ p, Ω̃ ≤ α . Therefore, the tracking error is uniformly
ultimately bounded with the ultimate bound
#
λmax (P) p 2 + μλmax (R −1 )α 2
e ≤ =ρ (7.78)
λmin (P)
Note that since n > m, then η can be made independent of the B matrix by scaling
the modified RLS adaptive laws by the left pseudo-inverse as
−1
Ω̇ = −RΨ (x) ε B B B (7.79)
Then, the estimation error equation for the scaled RLS adaptive law becomes
Ω̃˙ = −RΨ (x) Ψ (x) Ω̃ + ε (7.80)
Least-squares gradient adaptive laws can be designed to not require the use of the
matrix inverse B −1 . The modified least-squares gradient adaptive laws are given by
K̇ x = Γx xε B (7.81)
with μ >0.
Then, V̇ e, K̃ x , Θ̃ is evaluated as
V̇ e, K̃ x , Θ̃ = −e Qe + 2e Pε + μtrace 2 K̃ x xε B + μtrace −2Θ̃ Φ (x) ε B
λ2max (P)
2λmax (P)eε ≤ e2 + δ 2 ε2 (7.88)
δ2
and completing the square, we obtain
170 7 Function Approximation and Adaptive Control with Unstructured Uncertainty
λ2max (P)
V̇ e, K̃ x , Θ̃ ≤ − λmin (Q) − e2
δ2
μBε0∗ 2 μ2 B2 ε0∗2
− (2μ − δ ) ε −
2
+ (7.89)
2μ − δ 2 (2μ − δ 2 )
√
Then, V̇ e, K̃ x , Θ̃ ≤ 0 if √λλmax (P) < δ < 2μ. The largest lower bounds of
min (Q)
e and ε and then obtained as
μ B ε0∗
e ≥ & =p (7.90)
λ2max (P)
(2μ − δ 2 ) λmin (Q) − δ2
2μBε0∗
ε ≥ =α (7.91)
2μ − δ 2
Since V̇ e, K̃ x , Θ̃ ≤ 0 outside a compact set, the closed-loop adaptive system
is uniformly ultimately bounded.
ẋ = ax + b [u + f (x)]
where a and f (x) are unknown, but b = 1. For simulation purposes, a = 1 and
f (x) = 0.2 sin 2x + cos 4x + e−x .
2
ẋm = am xm + bm r
where k x (t) and Θ (t) are computed by the least-squares gradient adaptive law as
k̇ −xεb
Ω̇ = x = −Γ
Θ̇ Φ (x) εb
7.3 Adaptive Control with Unstructured Uncertainty 171
ε = âx + bū − ẋ
â = am − bk x
ū = k x x + kr r
The tracking error for q = 1, 2, 3, 4 is shown in Figs. 7.9 and 7.10. Note that the
tracking error improves for q ≥ 2. Even the tracking error improves, the function
f (x) does not seem to be well approximated as shown in Fig. 7.11. This is also due
to k x (t) and Θ (t) not quite converging.
Now, suppose the Chebyshev orthogonal polynomials are used instead. Then,
The simulation results are as shown in Fig. 7.12. For q = 1, the result is the
same as that for the regular polynomial. However, it can be seen that the tracking
error significantly reduces for q = 2 with the Chebyshev polynomial and is even
smaller than that for q = 4 with the regular polynomial. For q = 4, the Chebyshev
polynomial approximation results in a very small tracking error. The parameter con-
vergence of the Chebyshev polynomial coefficients is excellent for q = 4 as shown
in Fig. 7.13. The unknown function is very well approximated by a fourth-degree
Chebyshev polynomial as shown in Fig. 7.14.
1 1
0.5 0.5
0 0
e
q=1 q=2
−0.5 −0.5
−1 −1
0 50 100 0 50 100
t t
1 1
0.5 0.5
0 0
e
q=3 q=4
−0.5 −0.5
−1 −1
0 50 100 0 50 100
t t
Fig. 7.9 Tracking error with regular q-th degree polynomial approximation
172 7 Function Approximation and Adaptive Control with Unstructured Uncertainty
2
1
1
kx, Θ
k ,Θ
0
0 q=2
x
q=1 −1
−1
−2 −2
0 50 100 0 50 100
t t
2 2
1 1
k ,Θ
k ,Θ
0 0 q=4
q=3
x
x
−1 −1
−2 −2
0 50 100 0 50 100
t t
1.5
1
ax+bf(x), ax+bΘTΦ(x)
0.5
0
<
−0.5
−1
ax+bf(x)
^
ax+bΘ
T
Φ(x) q=1
−1.5
^ T
ax+bΘ Φ(x) q=2
^ T
−2 ax+bΘ Φ(x) q=3
T
^
ax+bΘ Φ(x) q=4
−2.5
−1 −0.5 0 0.5 1
x
Fig. 7.11 Function approximation at t = 100 with regular q-th degree polynomial
1 1
0.5 0.5
0 0
e
q=1 q=2
−0.5 −0.5
−1 −1
0 50 100 0 50 100
t t
1 1
0.5 0.5
0 0
e
q=3 q=4
−0.5 −0.5
−1 −1
0 50 100 0 50 100
t t
Fig. 7.12 Tracking error with Chebyshev q-th degree polynomial approximation
7.3 Adaptive Control with Unstructured Uncertainty 173
2
1
1 q=2
k ,Θ
k ,Θ
0
0
x
q=1 −1
−1
−2 −2
0 50 100 0 50 100
t t
2 2
1 1
q=3 q=4
kx, Θ
k ,Θ
0 0
x
−1 −1
−2 −2
0 50 100 0 50 100
t t
1.5
1
ax+bf(x), ax+bΘTΦ(x)
0.5
< 0
−0.5
ax+bf(x)
−1 ^ax+bΘTΦ(x) q=1
^ T
ax+bΘ Φ(x) q=2
−1.5 ^ T
ax+bΘ Φ(x) q=3
T
^
ax+bΘ Φ(x) q=4
−2
−1 −0.5 0 0.5 1
x
Fig. 7.14 Function approximation at t = 100 with Chebyshev q-th degree polynomial
ε = −B K̃ x x + BΘ Φ W x̄ − B Θ − Θ̃
× Φ W x̄ + Φ W x̄ W ∗ x̄ − W x̄ + · · · + Bε∗ (7.94)
Combining the higher-order terms with the neural network approximation error
ε∗ (x), then
ε = −B K̃ x x + B Θ̃ Φ W x̄ + BΘ Φ W x̄ W̃ x̄ + Bδ (7.95)
where δ (x) is the combined approximation error due to the Taylor series truncation
and the neural network approximation.
This becomes
ε = −B K̃ x x + B Θ̃ Φ W x̄ + BV σ W x̄ W̃ x̄ + Bδ (7.96)
The least-squares gradient adaptive laws for K x (t), Θ (t), and W (t) are then
given by
K̇ x = Γx xε B (7.97)
Θ̇ = −ΓΘ Φ W x̄ ε B (7.98)
Ẇ = −ΓW x̄ε BV σ W x̄ (7.99)
Then,
V̇ e, K̃ x , Θ̃, W̃ = −e Qe + 2e Pε − 2με ε + 2με Bδ
≤ −λmin (Q) e2 + 2λmax (P) e ε − 2μ ε2
+ 2μ B ε δ0 (7.103)
Using the results in Sect. 7.3.3, V̇ e, K̃ x , Θ̃, W̃ ≤ 0 outside a compact set
defined by the following lower bounds of e and ε:
μ B δ0
e ≥ & =p (7.104)
λ2max (P)
(2μ − δ 2 ) λmin (Q) − δ2
2μBδ0
ε ≥ =α (7.105)
2μ − δ 2
By now, it can be seen that there are at least two approaches to adaptive control. One
approach is based on the Lyapunov method which is the basis for MRAC and another
approach is based on the least-squares adaptive control. The underlying principle of
both the MRAC and the least-squares adaptive control is to reduce system uncertainty
176 7 Function Approximation and Adaptive Control with Unstructured Uncertainty
ẋ = Ax + B [u + f (x)] (7.106)
Then, it can be shown that all the adaptive laws developed in Sects. 5.6 and 5.7
could be used to estimate Θ ∗ with some caveats as explained later. For example, if
A and B are known, then Eq. (5.155) from Sect. 5.6.2 can be used to update Θ (t) as
Θ̇ = −Γ Φ (x) e P B (7.108)
Since A and B are known, assuming there exist constant control gain matrices K x
and K r that satisfy the model matching conditions
A + B K x = Am (7.109)
B K r = Bm (7.110)
Then,
7.3 Adaptive Control with Unstructured Uncertainty 177
V̇ e, Θ̃ = −e Qe + 2e P B Θ̃ Φ (x) + Bε∗ − 2trace Θ̃ Φ (x) e P B
= −e Qe + 2e P B Θ̃ Φ (x) + Bε∗ − 2e P B Θ̃ Φ (x)
= −e Qe + 2e P Bε∗ ≤ −λmin (Q) e2 + 2 P B e ε0∗ (7.115)
Thus, V̇ e, Θ̃ ≤ 0 if
2 P B ε0∗
− λmin (Q) e2 + 2 P B e ε0∗ ≤ 0 ⇒ e ≥ =p (7.116)
λmin (Q)
At a first glance, it may be tempted to conclude that the tracking error is uniformly
ultimately bounded, but this conclusion would be incorrect since the tracking error is
only bounded. The reason for the least-squares adaptive laws to result in a uniformly
ultimately bounded tracking error is that any least-squares method guarantees that the
approximation error is at least bounded. This implies that the parameter estimation
error Θ̃ (t) is also bounded. In addition, if the uncertainty is structured and the basis
function is PE, then the parameter estimation error tends to zero, which implies a
parameter convergence. If both the tracking error and parameter estimation error can
be shown to be bounded, then the system
is uniformly ultimately bounded.
For MRAC, the condition for V̇ e, Θ̃ ≤ 0 in Eq. (7.116) does not say anything
about the parameter estimation error Θ̃ (t). Let
) *
Br = e (t) ∈ Rn : e ≤ p (7.117)
toward Br . However, if Θ̃ (t) grows too large inside Br , the small approximation
error assumption may be violated. Consequently, the boundary of Br may also grow
and therefore e (t) can also grow. The possibility of a parameter drift that can lead to
a point where the system becomes unstable may exist when an adaptive law does not
provide a mathematically guaranteed bound on adaptive parameters. As a result, the
standard MRAC as developed in Chap. 5 for systems with unstructured uncertainty
is non-robust. Instability due to a parameter drift when adaptive parameters are not
bounded is a well-known issue with adaptive control. In this regard, least-squares
adaptive control offers an advantage over the standard MRAC since the adaptive
parameters are bounded. Robust modification of MRAC is another way to address
this issue. In summary, it is generally not a good practice to use the standard MRAC
for systems with unstructured uncertainty.
Example 7.5 Consider Example 7.4. The MRAC laws are given by
k̇ x = γx xeb
Θ̇ = −ΓΘ Φ (x) eb
1 1
0.5 0.5
0 0
e
q=1 q=2
−0.5 −0.5
−1 −1
0 50 100 0 50 100
t t
1 1
0.5 0.5
0 0
e
q=3 q=4
−0.5 −0.5
−1 −1
0 50 100 0 50 100
t t
2
1
1
kx, Θ
k ,Θ
0
0
x
q=1
−1 −1
q=2
−2 −2
0 50 100 0 50 100
t t
2 2
1 1
k ,Θ
k ,Θ
0 0
x
x
−1 −1
q=3 q=4
−2 −2
0 50 100 0 50 100
t t
2
ax+bf(x), ax+bΘ Φ(x)
1
T
< 0
−1
ax+bf(x)
^ax+bΘTΦ(x) q=1
^ax+bΘTΦ(x) q=2
−2
^ T
ax+bΘ Φ(x) q=3
T
^
ax+bΘ Φ(x) q=4
−3
−1 −0.5 0 0.5 1
x
Fig. 7.18 Function approximation at t = 100 with Chebyshev q-th degree polynomial
Note that
e P B Θ̃ Φ W x̄ = trace Θ̃ Φ W x̄ e P B (7.122)
e P BV σ W x̄ W̃ x̄ = trace W̃ x̄e P BV σ W x̄ (7.123)
Hence,
V̇ e, Θ̃, W̃ = −e Qe + 2e P Bδ + 2trace Θ̃ Φ W x̄ e P B + ΓΘ−1 Θ̃˙
+ 2trace W̃ x̄e P BV σ W x̄ + ΓW−1 W̃˙ (7.124)
Setting the trace terms to zero yields the neural network adaptive laws
Θ̇ = −ΓΘ Φ W x̄ e P B (7.125)
Ẇ = −ΓW x̄e P BV σ W x̄ (7.126)
Then,
V̇ e, Θ̃, W̃ = −e Qe + 2e P B ε̄ ≤ −λmin (Q) e + 2 P B e δ0 (7.127)
Therefore, V̇ e, Θ̃, W̃ ≤ 0 if
2 P B δ0
e ≥ =p (7.128)
λmin (Q)
k̇ x = γx xeb
7.3 Adaptive Control with Unstructured Uncertainty 181
Θ̇ = −ΓΘ Φ W x̄ eb
Ẇ = −ΓW x̄ebV σ W x̄
The neural network topology is shown in Fig. 7.19. The simulation results are
shown in Fig. 7.20.
2 1
1 0
x, xm
e
0 −1
x xm
−1 −2
0 50 100 0 50 100
t t
1 2
0
0
kx, Θ
W
−1
−2
−2
−3 −4
0 50 100 0 50 100
t t
The neural network adaptive control appears to reduce the tracking error a bit
better than the direct MRAC with Chebyshev polynomial approximation. However,
there is a large transient at the beginning. This is probably due to the initial conditions
of Θ (t) and W (t) being chosen arbitrarily, and the poor initial approximation that
causes the tracking error to grow before the learning becomes effective in reducing
the tracking error. The neural network adaptive control does not seem to perform as
well as the least-squares gradient method with the Chebyshev polynomials.
In the context of least-squares adaptive control, the plant modeling error ε (t)
represents a modeling error between the true plant and the desired plant. Modeling
errors in dynamical systems usually arise from the presence of system uncertainty.
In model-reference adaptive control, the tracking error e (t) is exclusively used as a
control signal to drive the adaptation so as to cancel out undesired effects of system
uncertainty on closed-loop systems. Philosophically, the tracking error is a mani-
festation of the plant modeling error which in turn is a direct consequence of the
system uncertainty. Thus, the tracking error is a result of the plant modeling error,
but not vice versa. One can separate control and system identification as two related
dynamical actions. A control action is generally designed to reduce the tracking
error to ensure that a closed-loop system follows a reference model as closely as
possible. The objective of adaptive control is not so much to estimate the uncertainty
itself but rather to achieve a desired reference model, regardless whether or not a
parameter convergence is achieved. Example 7.5 clearly demonstrates that model-
reference adaptive control does a rather poor job on parameter estimation even with
the Chebyshev polynomials.
On the other hand, a system identification action uses the plant modeling error
to estimate the uncertainty directly by least-squares estimation methods. The goal
of a system identification is first and foremost to achieve a parameter convergence.
The system identification action can achieve a better parameter convergence than
MRAC. System identification is usually performed in an open-loop process without
directly influencing a control action. Both the system identification and control can
also be combined together to provide a very effective adaptation strategy. Using
parameter estimates, least-squares adaptive control can be formulated to cancel out
effects of system uncertainty by feeding back the parameter estimates into an adaptive
control law. In so doing, least-squares adaptive control achieves bounded tracking and
bounded adaptive parameters, whereas MRAC can only achieve bounded tracking.
7.4 Summary
In many physical applications, there is no clear certainty about the structure between
the input and output of a process. In systems with unstructured uncertainty, the map-
ping between the input and output is usually not known. Polynomial approximation
by least-squares methods is a well-known regression method. Regular polynomials
are frequently used in least-squares data regression. Other types of polynomials can
7.4 Summary 183
also be used. One class of polynomials that can offer better function approximation
over the regular polynomials is orthogonal polynomials. Orthogonal polynomials
form true basis functions that span the Euclidean space. In particular, the Chebyshev
orthogonal polynomials are generally considered to be optimal that can accurately
approximate real-valued functions with the smallest degree for the same accuracy as
a regular polynomial approximation.
Neural networks have been used in many applications such as classifiers, pat-
tern recognition, function approximation, and adaptive control. A neural network is
a nonlinear function approximation that models a complex input–output nonlinear
mapping. Two basic types of neural networks are discussed: sigmoidal or logistic
functions and radial basis functions. Various adaptive control methods for unstruc-
tured uncertainty are presented. Both least-squares and model-reference adaptive
control with polynomial and neural network function approximation are shown to
be able to handle adaptive control systems with unstructured uncertainty. Least-
squares adaptive control, in general, exhibits better performance and robustness than
MRAC. Robustness of least-squares adaptive control is achieved through the nat-
ural parameter identification process which ensures parameter boundedness, whereas
MRAC cannot guarantee boundedness of adaptive parameters. This leads to robust-
ness issues, such as a parameter drift.
7.5 Exercises
1. Approximate
y = 0.1 sin 0.4x + cos2 2x
where
1
σ (x) =
1 + e−x
ẋ = ax + b [u + f (x)]
where a and f (x) are unknown, but b = 2. For simulation purposes, a = 1 and
f (x) = 0.1 sin 0.4x + cos2 2x.
The reference model is given by
ẋm = am xm + bm r
References
1. Lee, T., & Jeng, J. (1998). The chebyshev-polynomials-based unified model neural networks
for function approximation. IEEE Transactions on Systems, Man, and Cybernetics, Part B:
Cybernetics, 28(6), 925–935.
2. Mason, J., & Handscomb, D. (2002). Chebyshev polynomials. Boca Raton: Chapman and
Hall/CRC.
3. Cybenko, G. (1989). Approximation by superpositions of a sigmoidal function. Mathematics
of Control Signals Systems, 2, 303–314.
4. Suykens, J., Vandewalle, J., & deMoor, B. (1996). Artificial neural networks for modeling and
control of non-linear systems. Dordrecht: Kluwer Academic Publisher.
5. Wang, X., Huang, Y., & Nguyen, N. (2010). Robustness quantification of recurrent neural
network using unscented transform. Elsevier Journal of Neural Computing, 74(1–3).
6. Calise, A. J., & Rysdyk, R. T. (1998). Nonlinear adaptive flight control using neural networks.
IEEE Control System Magazine, 18(6), 1425.
7. Ishihara, A., Ben-Menahem, S., & Nguyen, N. (2009). Protection ellipsoids for stability analysis
of feedforward neural-net controllers. In International Joint Conference on Neural Networks.
8. Johnson, E. N., Calise, A. J., El-Shirbiny, H. A., & Rysdyk, R. T. (2000). Feedback linearization
with neural network augmentation applied to X-33 attitude control. AIAA Guidance, Naviga-
tion, and Control Conference, AIAA-2000-4157.
9. Kim, B. S., & Calise, A. J. (1997). Nonlinear flight control using neural networks. Journal of
Guidance, Control, and Dynamics, 20(1), 26–33.
10. Lam, Q., Nguyen, N., & Oppenheimer, M. (2012). Intelligent adaptive flight control using
optimal control modification and neural network as control augmentation layer and robustness
enhancer. In AIAA Infotech@Aerospace Conference, AIAA-2012-2519.
11. Lewis, F. W., Jagannathan, S., & Yesildirak, A. (1998). Neural network control of robot manip-
ulators and non-linear systems. Boca Raton: CRC.
12. Zou, A., Kumar, K., & Hou, Z. (2010). Attitude control of spacecraft using chebyshev neural
networks. IEEE Transactions on Neural Networks, 21(9), 1457–1471.
13. Khalil, H. K. (2001). Nonlinear systems. Upper Saddle River: Prentice-Hall.
Chapter 8
Robustness Issues with Adaptive Control
ẋ = Ax + Bu + w (8.1)
ẋm = Am + Bm r (8.2)
A + B K x∗ = Am (8.3)
B K r = Bm (8.4)
Suppose the disturbance is not accounted for, then one can design an adaptive
controller as
u = K x (t) x + K r r (8.5)
K̇ x = Γx xe P B (8.6)
P Am + A
m P = −Q (8.7)
ẋ = (A + B K x ) x + B K r r + w (8.8)
ė = Am e − B K̃ x x − w (8.9)
Then,
V̇ e, K̃ x = −e Qe − 2e Pw ≤ −λmin (Q) e2 + 2λmax (P) e w0 (8.11)
where w0 = maxw∞.
Therefore, V̇ e, K̃ x ≤ 0 if
2λmax (P) w0
e ≥ =p (8.12)
λmin (Q)
ẋ = ax + bu + w (8.13)
u = k x (t) x (8.14)
k̇ x = −γx x 2 b (8.15)
It can be shown that it is possible for a bounded response of x (t) to exist that
would result in an unbounded signal k x (t).
x = (1 + t)n
The disturbance w (t) that results in this particular solution of x (t) can be found by
(1 + t)2n+1 − 1
w = ẋ − ax − bu = n (1 + t)n−1 − a (1 + t)n + b γx b − k x (0) (1 + t)n
2n + 1
As w (t) is sufficiently large, k x (t) tends to a large negative value in order to keep
x (t) close to zero. This high-gain control can cause k x (t) to become unbounded. In
the limit as x (t) → 0 to achieve the ideal asymptotic tracking property of MRAC,
k x (t) → ∞. While Example 8.1 shows that x (t) can remain bounded even though
k x (t) is unbounded, in practice, a high-gain control with a large value of k x (t) can be
1 0
−5
0.5
u
x
−10
0 −15
0 2500 5000 0 2500 5000
t t
15 0
10 −50
kx
w
5 −100
0 −150
−5 −200
0 2500 5000 0 2500 5000
t t
problematic since real plants can include other closed-loop behaviors that can lead to
instability when k x (t) becomes large. Thus, a parameter drift can lead to instability
in practical applications.
A transfer function of a SISO system is called minimum phase if all of its poles and
zeros are in the left half plane. If any of the zeros lie in the right half plane, then
the system is called non-minimum phase. A non-minimum phase system is more
difficult to control since the existence of the unstable zeros can limit the values of
stable feedback gains. Increasing feedback gains beyond certain limits will cause the
closed-loop poles to become unstable.
y (s) s+1
=
u (s) (s + 2) (s + 3)
y=x
is minimum phase. For a feedback with u (s) = k x x (s) + kr r (s), the closed-loop
transfer function
y (s) kr
G (s) = 2
r (s) s + (5 − k x ) s + 6 − k x
x (s) s−1
=
u (s) (s + 2) (s + 3)
kr
G (s) =
s 2 + (5 − k x ) s + 6 + k x
If the system exhibits a non-minimum phase behavior, then the adaptation can
result in instability due to MRAC attempting to perform an unstable pole-zero can-
cellation in order to achieve asymptotic tracking.
ẋ = ax + bu − 2z + w
ż = −z + u
y=x
x (s) s−1
=
u (s) (s − a) (s + 1)
If the same adaptive controller is used for the same disturbance in Example 8.1
with n = −1, then the disturbance w (t) is bounded and the response of the closed-
loop system in Example 8.1 is completely bounded. However, this is no longer the
case for the non-minimum phase system in this example. Let a = −1, b = 1, γx = 1,
x (0) = 1, and k x (0). The closed-loop system is unstable as shown in Fig. 8.2.
In fact, the closed-loop system is unstable even when the disturbance is zero and
γx = 10 as shown in Fig. 8.3.
Now, suppose the plan is minimum phase with
ẋ = ax + u + 2z + w
ż = −z + u
y=x
192 8 Robustness Issues with Adaptive Control
4 0
3
−20
2
u
x
−40
1
0 −60
0 5 10 0 5 10
t t
0.8 0
0.6 −5
kx
w
0.4 −10
0.2 −15
0 −20
0 5 10 0 5 10
t t
6 0
4 −500
u
x
2 −1000
0 −1500
0 5 10 0 5 10
t t
1 0
0.5 −100
x
w
0 −200
k
−0.5 −300
−1 −400
0 5 10 0 5 10
t t
The closed-loop system now is stable with the same adaptive controller as shown
in Fig. 8.4.
1 0.5
0
0.5
−0.5
u
x
0
−1
−0.5 −1.5
0 5 10 0 5 10
t t
1 0
0.5
−1
kx
w
0
−2
−0.5
−1 −3
0 5 10 0 5 10
t t
Many real systems have latency which results in a time delay at the control input.
Time delay is caused by a variety of sources such as communication bus latency,
computational latency, transport delay, etc. Time-delay systems are a special class
of non-minimum phase systems which exhibit some similar behaviors with respect
to adaptive control. Model-reference adaptive control of time-delay systems is chal-
lenging as no well-established theory is available. Instability can result when an
adaptive parameter exceeds a certain limit.
Consider a time-delay MIMO system with a matched uncertainty
ẋ = Ax + B u (t − td ) + Θ ∗ Φ (x) (8.17)
Θ̇ = −Γ Φ (x) e P B (8.19)
When a time delay is present, there exists an upper limit on the value of the adap-
tation rate matrix Γ that ensures stable adaptation. Thus, the adaptation rate can no
longer be chosen arbitrary but rather by a careful consideration of the effect of the
time delay present in the system. Unfortunately, no well-established theory currently
exists that would enable a systematic design of adaptive control for time-delay sys-
tems. As a result, the design is relied upon extensive Monte-Carlo simulations to
ensure that a proper adaptation rate is selected.
Consider a time-delay SISO system
ẋ = ax + bu (t − td ) (8.20)
x (s) be−td s
= (8.21)
u (s) s−a
is non-minimum phase since there exists a multiplicity of zeros in the right half plane
due to the term e−td s .
The system has a feedback control
u = kx x (8.22)
The feedback gain k x is designed to be robust in the presence of the time delay to
ensure stability of the closed-loop system.
The closed-loop transfer function
kr
G (s) = (8.23)
s − a − bk x e−td s
is not absolutely stable but conditionally stable for k xmin < k x < − ab .
The system pole can be computed by letting s = σ + jω. Then,
the real part and imaginary parts are equated to zero so that the following equations
are obtained:
− a − bk x cos ωtd = 0 (8.26)
The solutions of these equations yield the jω-axis cross-over frequency from
which k xmin can be determined as
ω= b2 k x2min − a 2 (8.28)
a + bk xmin cos b2 k x2min − a 2 td = 0 (8.29)
Thus, the closed-loop system is stable for all values of k xmin < k x < − ab .
Suppose the constant a is unknown and the adaptive controller in Eqs. (8.14) and
(8.15) is used to regulate x (t) to zero. Without the time delay, the adaptation rate can
be arbitrarily chosen. However, in the presence of the time delay, there exists an upper
limit on γx that ensures stability of the closed-loop system. Thus, the closed-loop
adaptive system is stable if γx ≤ γxmax such that
t
k x = k x (0) − γxmax x 2 bdτ > k xmin (8.30)
0
1 2
0
0.5
γ =7 γ =7.6
x −2 x
x
0
−4
−0.5 −6
0 5 10 0 5 10
t t
0 0
−1 −20
γ =7 γx=7.6
kx
kx
−2 x −40
−3 −60
−4 −80
0 5 10 0 5 10
t t
Non-minimum phase systems are a special class of systems with unmodeled dynam-
ics. Unmodeled dynamics are dynamics that actually exist in a plant but are not
accounted for in a control design. This is a very common situation since most con-
trol designs are based on mathematical models of some real plants that most likely
do not fully capture all the complex physics. In some cases, the neglected dynamics
do not present problems in a design, but in other cases, controllers that are designed
based on slightly inaccurate mathematical models can lead to catastrophic conse-
quences. Therefore, it is a good practice to include stability margins in any control
design to account for effects of unmodeled dynamics. Traditional gain and phase
margins are well-established design criteria for linear control design for SISO sys-
tems. For adaptive control, there is no such equivalent stability margin nor is there
any well-established theory for analysis of stability margins of adaptive systems.
Consider the system
ẋ = Ax + Bu + Δ (x, z, u) (8.31)
ż = f (x, z, u) (8.32)
y=x (8.33)
where a < 0, b > 0, ζ > 0, ω > 0, and u c (t) is the actuator command.
Suppose we want to design an adaptive controller to track a first-order reference
model
ẋm = am xm + bm r (8.36)
where am < 0.
Without knowing or ignoring the actuator dynamics, the adaptive controller is
designed to be of the form
where k x (t) and kr (t) are adjusted by the following MRAC adaptive laws:
k̇ y = γx ye (8.38)
k̇r = γr r e (8.39)
y (s) bωn2
=
(8.40)
u c (s) (s − a) s 2 + 2ζ ωn s + ωn2
We introduce a time delay at the input u c (t). Then, the closed-loop transfer
function is obtained as
bωn2 kr e−ωtd
G (s) =
(8.41)
(s − a) s 2 + 2ζ ωn s + ωn2 − bωn2 k y e−ωtd
8.4 Unmodeled Dynamics 197
Let s = jω. Then, the characteristic equation of the closed-loop transfer function
is computed as
− jω3 −(2ζ ωn − a) ω2 + ωn2 − 2aζ ωn jω−aωn2 −bωn2 k y (cos ωtd − j sin ωtd ) = 0
(8.42)
Separating the real part and imaginary part leads to the following equations:
or alternatively as
−1 ω3 − ωn2 − 2aζ ωn ω
φ = ωtd = tan (8.46)
− (2ζ ωn − a) ω2 − aωn2
The feedback gain k y affects the cross-over frequency of the closed-loop system
according to the following equation:
ω6 + a 2 + 4ζ 2 ωn2 − 2ωn2 ω4 + ωn2 + 4a 2 ζ 2 − 2a 2 ωn2 ω2 + a 2 − b2 k 2y ωn4 = 0
(8.48)
We now return to the adaptive control design. Suppose the reference command
signal is specified as
r = r0 sin ω0 t (8.49)
Since the reference command signal provides frequency input that matches the
cross-over frequency at zero phase margin, the closed-loop adaptive system will be
unstable.
1 5
0.5
0
0
u
x
−5
−0.5
−1 −10
0 5 10 0 5 10
t t
−0.5 4
−1
3
ky
r
−1.5
k
2
−2
−2.5 1
0 5 10 0 5 10
t t
1 4
0.5 2
0 0
u
x
−0.5 −2
−1 −4
0 20 40 0 20 40
t t
0 3
−1 2.5
ky
−2 2
k
−3 1.5
−4 1
0 20 40 0 20 40
t
Fig. 8.7 Stable response with reference command r = sin 3t, k y (0) = −1
8.4 Unmodeled Dynamics 199
0.5 2
0 0
u
x
−1
−0.5 −2
0 20 40 0 20 40
t t
−5 2
−5.5
ky
r
1.5
k
−6
−6.5 1
0 20 40 0 20 40
t t
Fig. 8.8 Incipient instability of MRAC due to unmodeled dynamics with reference command
r = sin 3t, k y (0) = −6.2 at zero phase margin
5 100
50
0 0
u
x
−50
−5 −100
0 10 20 0 10 20
t t
0 20
15
−5
y
kr
10
k
−10
5
−15 0
0 10 20 0 10 20
t t
The adaptive laws are initialized with k y (0) = −0.65 and kr (0) = 1.14. The
adaptation rates are chosen to be γ√ x = 1 and γr = 1. The zero phase-margin
frequency is computed to be ω0 = 259 = 16.1 rad/s. The reference command
is given by
r = 0.3 + 1.85 sin 16.1t
bandwidth ω, that is, ωn ω. Then, the closed-loop transfer function in Eq. (8.41)
may be approximated by
bkr e−ωtd
G (s) = (8.50)
s − a − bk y e−ωtd
ẋ = ax + bu (8.53)
Differentiating yields
ẍ − a + bk p ẋ − bki x = −bki xm (8.56)
Equating the real and imaginary parts to zero results in the following equations:
Multiplying Eq. (8.59) by bki and Eq. (8.60) by bk p ω and then adding the resulting
equations together yield
202 8 Robustness Issues with Adaptive Control
abk p + bki ω2 + b2 k 2p ω2 + b2 ki2 cos ωtd = 0 (8.65)
The closed-loop plant is then stable for all t ≤ td where td is called a time-delay
margin which is computed as
1 −1 abk p + bki ω2
td = cos − 2 2 2 (8.66)
ω b k p ω + b2 ki2
ω → ωn → ∞ (8.67)
1 ω4
td → cos−1 n4 → 0 (8.68)
ω ωn
φ = ωtd → 0 (8.69)
Thus, as the integral gain increases, the closed-loop response becomes more oscil-
latory and the time-delay margin and phase margin of the closed-loop plant decrease.
In the limit, these margins tend to zero as ki → ∞. On the other hand, the tracking
performance is improved with increasing the integral gain. Clearly, the two require-
ments cannot be simultaneously met. Thus, a control design typically strives for a
balance between these two competing requirements by a judicious choice of ki .
Suppose the plant becomes uncertain with a being an unknown parameter. A
reference model is specified by
ẋm = am xm + bm r (8.70)
u = k x (t) x + kr r (8.71)
k̇ x = γx x (xm − x) b (8.72)
where kr = bbm .
Differentiating the closed-loop plant yields
ẍ − [a + bk x (t)] ẋ + b2 γx x 3 = b2 γx x 2 xm (8.73)
Contrasting Eq. (8.73) with Eq. (8.56), one can conclude that the effect of adaptive
control is very similar to that of the linear integral control. In fact, let
8.5 Fast Adaptation 203
30
20
ω
10
0 0 1 2 3
10 10 10 10
−ki
0.8
0.6
td
0.4
0.2
0 0 1 2 3
10 10 10 10
−k
i
Fig. 8.10 Cross-over frequency and time-delay margin as function of integral gain
204 8 Robustness Issues with Adaptive Control
1.5
1
x
0
0 2 4 6 8 10
t
1.5
1
x
0
0 2 4 6 8 10
t
Fig. 8.11 Responses of adaptive systems and linear systems with fast adaptation
1.5
1
x
0.5
Adaptive Controller
0
0 2 4 6 8 10
t
1.5
1
x
0.5
Linear PI Controller
0
0 2 4 6 8 10
t
Fig. 8.12 Responses of adaptive systems and linear systems with fast adaptation and time delay
8.6 Summary 205
8.6 Summary
dynamics, instability can occur. Initialization of adaptive parameters can also affect
the closed-loop stability if the initial condition coincides with a near zero phase
margin.
Fast adaptation refers to the notion that a large adaptation rate can be used to
achieve fast reference model following. In the limit, as the adaptation rate increases
to a theoretically infinite value, robustness of MRAC as measured by the time-delay
margin tends to zero. As with the case with non-minimum phase, time delay, and
unmodeled dynamics, MRAC can cause instability when a large adaptation rate is
used for fast adaptation.
8.7 Exercises
ẋ = ax + bu + w
k̇ x = −γx x 2 b
x = t (1 + t) p
ÿ − ẏ + y = u (t − td )
u = kd∗ ẏ
8.7 Exercises 207
u = Kx x
K̇ x = −Γx x x P B
to give a 60◦ phase margin. Also compute the ideal feedback gain k ∗y corresponding
to this phase margin. Implement in Simulink the Rohrs’ counterexample using
the same initial conditions k y (0) and kr (0) with γ y = γr = 1 and Δt = 0.001 s.
Plot the time histories of y (t), u (t), k y (t), and kr (t) for t ∈ [0, 60] s.
References
1. Ioannu, P. A., & Sun, J. (1996). Robust adaptive control. Upper Saddle River: Prentice-Hall,
Inc.
2. Rohrs, C. E., Valavani, L., Athans, M., & Stein, G. (1985). Robustness of continuous-time
adaptive control algorithms in the presence of unmodeled dynamics. IEEE Transactions on
Automatic Control, AC-30(9) 881–889.
Chapter 9
Robust Adaptive Control
mal control modification can produce good tracking performance for both minimum
phase and non-minimum phase reference models.
Parameter drift can occur when noise is present in a signal. If the signal-to-noise
ratio is sufficiently small, adaptive control would attempt to reduce noise instead of
the tracking error. As a result, adaptive parameters would continue to be integrated,
thereby leading to a parameter drift. Thus, the dead-zone method can be used to stop
the adaptation process when the tracking error is reduced below a certain level [5].
The dead-zone method for the adaptive law in Eq. (8.19) is stated as
−Γ Φ (x) e P B e > e0
Θ̇ = (9.1)
0 e ≤ e0
Proof When e > e0 , the adaptation process is active. The usual result of the
Lyapunov stability analysis was given in Sect. 8.1 with a bounded tracking error.
When e ≤ e0 , the adaptation is shut off to prevent the potential parameter drift as
e (t) → 0. Then, the following Lyapunov candidate function is used:
V (e) = e Pe (9.2)
ė = Am e + B Θ̃ Φ (x) − w (9.3)
V̇ (e) ≤ 0 requires
or
λ
max (Q) e + 2λmax (P) w0 λmax (Q) e0 + 2λmax (P) w0
Θ̃ ≤ ≤ (9.6)
2 P B Φ (x) 2 P B Φ0
9.1 Dead-Zone Method 213
Since e ≤ e0 , it implies that Φ (x) ≤ Φ0 is bounded. Thus, Θ̃ is also
bounded from above. The adaptive law with the dead-zone method therefore is robust
and can prevent a parameter drift.
The difficulty with the dead-zone method usually lies with the selection of e0 which
is generally by trial-and-error. If the threshold is set too large, then the adaptation
process may be prematurely terminated by the dead-zone method. On the other hand,
if it is set too low, then the integral action of the adaptive law can lead to a parameter
drift.
Example 9.1 For Example 8.1 with k x (t) unbounded for n = − 12 5
, using the dead-
zone method with |x| > 0.2, the value of k x (t) is limited to −54.2195. Since
w (t) → 0 as t → ∞, x (t) will also tend to zero with the dead-zone method.
However, it tends to zero at a slower rate than that with MRAC. The response of
the closed-loop system with the dead-zone method is compared to that with MRAC
as shown in Fig. 9.1. A time delay is injected at the input. The closed-loop system
can tolerate up to td = 0.0235 s with MRAC and td = 0.0285 s with the dead-zone
method. This indicates that the dead-zone method is more robust than MRAC.
1
Deadband w/ |e|>0.2
MRAC
0.5
x
0
0 20 40 60 80 100
t
0
Deadband w/ |e|>0.2
−20 MRAC
kx
−40
−60
−80
0 20 40 60 80 100
t
The projection method is used to limit adaptive parameters based on a priori knowl-
edge of the uncertainty. This may be a contradiction since uncertainty is usually
not known a priori. However, an adaptive controller can be designed under a cer-
tain assumption of the uncertainty. As long as the uncertainty does not violate the
assumption, then the projection method is guaranteed to provide robustness.
214 9 Robust Adaptive Control
g (Θ) = 0 (9.9)
The second term is the augmented term which does not change the cost function
since g (Θ) = 0.
The necessary condition of optimality is obtained as
∂J
∇ JΘ (Θ) = =0 (9.11)
∂Θ
This yields
∇ f Θ (Θ) + λ∇gΘ (Θ) = 0 (9.12)
∇ gΘ (Θ) ∇ f Θ (Θ)
λ=− (9.14)
∇ gΘ (Θ) ∇gΘ (Θ)
Thus, the adjoint variable λ (t) is negative of the projection of the vector ∇ JΘ (Θ)
onto the vector ∇gΘ (Θ).
Now, the gradient update law for Θ (t) is given by
The projection thus restricts of the optimal solution to only permissible values of
Θ (t) that lie on the constraint set g (Θ) = 0.
For example, consider the following constraint:
g (Θ) = Θ − Θ ∗ Θ − Θ ∗ − R2 = 0 (9.18)
From Fig. 9.2, when Θ (t) moves away from S , it follows that
Θ − Θ ∗ ∇gΘ (Θ) > 0 (9.22)
Using the projection method for MRAC, the adaptive law is expressed as
Θ̇ = Pro Θ, −Γ Φ (x) e P B
⎧
⎪
⎪ −Γ Φ (x) e PB if g (Θ) < 0 or if g (Θ) = 0
⎪
⎪
⎪
⎨ and − Φ (x) e P B ∇gΘ (Θ) ≤ 0
= ∇gΘ (Θ) ∇ gΘ (Θ)
⎪
⎪ −Γ I − Φ (x) e PB if g (Θ) ≥ 0
⎪
⎪ ∇ gΘ (Θ) ∇gΘ (Θ)
⎪
⎩
and − Φ (x) e P B ∇gΘ (Θ) > 0
(9.24)
Then,
∇gΘ (Θ) ∇ gΘ (Θ)
V̇ e, Θ̃ = −e Qe − 2e Pw + 2Θ̃ Φ (x) e P B (9.26)
∇ gΘ (Θ) ∇gΘ (Θ)
The direction of Θ (t) away from S is defined by − Φ (x) e P B ∇gΘ (Θ) >
0. Then, let ∇ gΘ (Θ) Φ (x) e P B = −c0 < 0 where c0 > 0. It follows from
Eq. (9.22) that
So,
V̇ e, Θ̃ ≤ −λmin (Q) e2 + 2λmax (P) e w0 − 2c0
≤ −λmin (Q) e2 + 2λmax (P) e w0 (9.29)
Thus, V̇ e, Θ̃ ≤ 0 if
2λmax (P) w0
e ≥ =p (9.30)
λmin (Q)
Since g (Θ) ≥ 0, then Θ̃ is bounded from below by Θ̃ ≥ R. We want the
trajectory of Θ̃ (t) to return to the set S . Thus, V̇ e, Θ̃ ≥ 0 inside a compact set
S = e , Θ̃ : e ≤ p and Θ̃ ≤ R (9.31)
but V̇ e, Θ̃ ≤ 0 outside S to ensure that both e (t) and Θ̃ (t) are uniformly
ultimately bounded.
The ultimate bounds can be found from the following inequalities:
2
λmin (P) e2 ≤ λmin (P) e2 + λmin Γ −1 Θ̃
≤ V e, Θ̃ ≤ λmax (P) p 2 + λmax Γ −1 R 2 (9.32)
218 9 Robust Adaptive Control
2
2
λmin Γ −1 Θ̃ ≤ λmin (P) e2 + λmin Γ −1 Θ̃ ≤ V e, Θ̃
≤ λmax (P) p 2 + λmax Γ −1 R 2 (9.33)
Thus,
λmax (P) p 2 + λmax Γ −1 R 2
p ≤ e ≤ =ρ (9.34)
λmin (P)
λmax (P) p 2 + λmax Γ −1 R 2
R ≤ Θ̃ ≤ =β (9.35)
λmin Γ −1
n
2
g (Θ) = θi − θi∗ − R 2 ≤ 0 (9.36)
i=1
Then,
θimin = θi∗ − R ≤ θi ≤ θi∗ + R = θimax (9.37)
The problem is that the range of θi which is equal to θimax − θimin is determined
by a single parameter R when in practice each θi could have a different range. For
example, if a priori bounds on the adaptive parameters can be determined such that
Then, R is determined as
n
R = min (ri ) (9.39)
i=1
This would then result in an over-constrained convex set wherein one or more
adaptive parameters will be over-constrained. Thus, a convex scalar function may
not be sufficiently flexible for handling more than one adaptive parameter. In such
situations, a convex vector function could be used.
9.2 Projection Method 219
where m ≤ n.
The augmented cost function is expressed as
m
J (Θ) = f (Θ) + g (Θ) λ = f (Θ) + λi gi (θi ) (9.41)
i=1
where λ (t) ∈ Rm .
Then,
∇ f Θ (Θ) + ∇ g (Θ) λ = 0 (9.42)
The same projection method can be used when Θ (t) is a matrix for each element
θi j (t).
The challenge with the projection method is to be able to determine θi∗ and ri∗ a
priori. Since technically θi∗ is an unknown ideal parameter, the projection method
has the same shortcoming as the dead-zone method in that a priori knowledge of the
uncertainty must be established. This illustrates an important point that robustness
can never be attained for any adaptive control design if the uncertainty is completely
unknown. Thus, with some knowledge on the bounds of the adaptive parameters, the
projection method can be used as an effective robust modification to MRAC.
Example 9.2 For Example 8.4, the largest negative feedback gain that the system
can tolerate is k xmin = −3.8069. If a priori knowledge of this limiting feedback gain
is known, then the projection method can be used to enforce the constraint on k x (t)
during the adaptation process. Suppose this constraint is specified as
k x − k ∗ ≤ r
x
where k x∗ = 0 and r = 3.
Using an adaptation rate γx = 7.6, which, in Example 8.4, would result in insta-
bility, the closed-loop system is now stable with k x (t) → −3 when the adaptation
is shut off as shown in Fig. 9.3.
0.5
x
−0.5
0 2 4 6 8 10
t
−1
x
−2
k
−3
−4
0 2 4 6 8 10
t
9.3 σ Modification
control [1]. Since then, researchers began to examine ways to modify the standard
MRAC to improve robustness. A number of modifications were developed in the
1980s, among which the σ modification is perhaps the simplest modification method
that demonstrates the potential for improved robustness. The σ modification method
was developed by Ioannu and Kokotovic [2]. It is widely used in adaptive control
practice due to its simplicity.
Consider a MIMO system with a parametric matched uncertainty and an unknown
bounded disturbance
ẋ = Ax + B u + Θ ∗ Φ (x) + w (9.48)
ẋm = Am xm + Bm r (9.49)
A + B K x = Am (9.50)
B K r = Bm (9.51)
u = K x x + K r r − Θ Φ (x) (9.52)
ė = Am e + B Θ̃ Φ (x) − w (9.55)
Evaluating V̇ e, Θ̃ yields
V̇ e, Θ̃ = −e Qe + 2e P B Θ̃ Φ (x) − 2e Pw + trace −2Θ̃ Φ (x) e P B + σ Θ
(9.56)
2
V̇ e, Θ̃ ≤ −λmin (Q) e2 +2 e λmax (P) w0 −2σ Θ̃ +2σ Θ̃ Θ0 (9.58)
λmax (P) w0 2
V̇ e, Θ̃ ≤ −λmin (Q) e −
λmin (Q)
! "2
λ 2
(P) w02
Θ0 σ Θ02
+ max − 2σ Θ̃ − + (9.59)
λmin (Q) 2 2
9.3 σ Modification 223
Thus, V̇ e, Θ̃ ≤ 0 if
λmax (P) w0 λ2max (P) w02 σ Θ02
e ≥ + + =p (9.60)
λmin (Q) λ2min (Q) 2λmin (Q)
or
Θ λ2max (P) w02 Θ2
0
Θ̃ ≥ + + 0 =α (9.61)
2 2σ λmin (Q) 4
V̇ e, Θ̃ ≤ 0 outside a compact set S defined by
! "2
λmax (P) w0 2
2 Θ0
S = e , Θ̃ : λmin (Q) e − + 2σ Θ̃ −
λmin (Q) 2
#
λ2 (P) w02 σ Θ02
≤ max + (9.62)
λmin (Q) 2
Therefore, the solution is uniformly ultimately bounded with the ultimate bounds
determined by
λmax (P) p 2 + λmax Γ −1 α 2
p ≤ e ≤ =ρ (9.63)
λmin (P)
λmax (P) p 2 + λmax Γ −1 α 2
α ≤ Θ̃ ≤ =β (9.64)
λmin Γ −1
We see that the ultimate bound for the tracking error increases due to the σ modifi-
cation term. This is a consequence of the trade-off between the tracking performance
and robustness. Another interesting
observation is that, if σ = 0, then the standard
MRAC is recovered and Θ̃ is unbounded if w0 is not zero. This is consistent with
the parameter drift behavior in the standard MRAC.
If the adaptive controller is a regulator type and the disturbance is zero, then the
closed-loop system is autonomous and is expressed as
ẋ = Am x − B Θ − Θ ∗ Φ (x) (9.65)
Θ̇ = −Γ −Φ (x) x P B + σ Θ (9.66)
1
Θ̄ = Φ (x̄) x̄ P B (9.68)
σ
Example 9.3 Consider the parameter drift example in Example 8.1. Let
x = (1 + t)n
d γx σ t
e k x = −eγx σ t γx x 2 b
dt
The solution of this equation is
eγx σ t (1 + t)2n − 1 2neγx σ t (1 + t)2n−1 − 2n
eγx σ t k x − k x (0) = −γx b −
γx σ γx2 σ 2
2n (2n − 1) eγx σ t (1 + t)2n−2 − 2n (2n − 1)
+ − ···
γx3 σ 3
which yields
! "
−γx σ t 1 2n 2n (2n − 1)
kx = e k x (0) − γx b − + 2 2− + ···
γx σ γx σ γx3 σ 3
(1 + t)2n 2n (1 + t)2n−1 2n (2n − 1) (1 + t)2n−2
− γx b − + − ···
γx σ γx2 σ 2 γx3 σ 3
9.3 σ Modification 225
1.2 0
1.1 −5
u
x
1 −10
0.9 −15
0 10 20 0 10 20
t t
15 0
10
−5
x
w
k
−10
0
−5 −15
0 10 20 0 10 20
t t
Note that not only x (t) is bounded for all n < 0, but k x (t) is also bounded. Thus,
the σ modification has eliminated the parameter drift issue.
The response of the closed-loop plant for the same disturbance in this example
with n = − 125
, x (0) = 1, k x (0) = 0, γx = 10, and σ = 0.1 is shown in Fig. 9.4.
Note that with MRAC, k x (t) is unbounded but now becomes bounded with the σ
modification.
where A is unknown.
The plant is designed with the following adaptive controller:
u = K x (t) x + K r r (9.70)
ẍ = A ẋ + B K̇ x x + B K x ẋ + B K r ṙ (9.72)
Then,
ẍ − (A + B K x ) ẋ − B B Pex + σ K x Γx x = B K r ṙ (9.73)
The second term on the left-hand side acts as a damping term due to the propor-
tional control action. The third term is effectively a nonlinear integral control. Without
226 9 Robust Adaptive Control
k̇ x = γx (xeb − σ k x ) (9.74)
The third term in the left-hand side is the nonlinear integral control
Thus, the σ modification changes the ideal integral control action of the adaptive
controller.
Suppose γx → ∞ and r (t) is a constant reference command signal, then k x (t)
tends to an equilibrium
x̄ ēb
k̄ x → (9.77)
σ
This effectively reduces the integral gain ki to zero. Then, the closed-loop system
tends to ! "
x̄ ēb2
ẋ = a + x + bkr r (9.78)
σ
which yields ⎛ ⎞
x̄m ⎝ 4σ k̄ x ⎠
x̄ = 1+ 1− (9.82)
2 b x̄m2
The σ modification therefore causes x (t) to not follow xm (t). So, the ideal prop-
erty of asymptotic tracking of MRAC is not preserved. This is a trade-off with
improved robustness.
The equilibrium value of k x (t) can be determined from the state equation
ẋ = (a + bk x ) x + bkr r (9.83)
ẋm = am xm + bm r (9.85)
am x̄m a
k̄ x = − (9.87)
b x̄ b
Because the tracking is not asymptotic and the adaptive parameter does not tend
to its ideal value, a switching modification to the σ modification has been proposed
by Ioannu [3]. The switching σ modification turns on the σ modification when the
adaptive parameter exceeds a certain threshold and switches off when the adaptive
parameter falls below the threshold. When the σ modification is switched off, the
adaptive law will attempt to restore its ideal property of asymptotic tracking. In
practice, the switching can cause transients and high-frequency oscillations that could
present a problem in itself.
228 9 Robust Adaptive Control
As with the dead-zone method, the selection of the adaptive parameter threshold
is usually not guided by systematic design approaches, but rather by trial-and-error.
Example 9.4 Consider Example 8.7. Keeping everything the same but using a value
of σ = 0.1, the responses of the closed-loop system with and without a time delay of
0.0020 s are shown in Fig. 9.5. Note that the two responses are very similar, whereas
the response with the standard MRAC with a time delay of 0.0020 s shows that the
closed-loop system is unstable. All high-frequency oscillations are no longer present.
1.5 1.5
x xm x xm
1 1
x
0.5 0.5
td = 0 t =0.002 sec
d
0 0
0 5 10 0 5 10
t t
0 1
−0.5
0
x
−1
k
−1
−1.5 td = 0 t =0.002 sec
d
−2 −2
0 5 10 0 5 10
t t
Fig. 9.5 Responses due to σ modification with and without time delay
The ideal feedback gain is computed to be k x∗ = −2. The equilibrium state is then
computed from
x̄ 3 − x̄ 2 − 0.1x̄ − 0.1 = 0
9.4 e Modification
2
V̇ e, Θ̃ ≤ −λmin (Q) e2 + 2 e λmax (P) w0 − 2μ e P B Θ̃
+ 2μ e P B Θ̃ Θ0 (9.94)
Θ
0
Θ̃ = (9.96)
2
Then,
μ e P B Θ02
V̇ e, Θ̃ ≤ −λmin (Q) e2 + 2 e λmax (P) w0 + (9.97)
2
V̇ e, Θ̃ ≤ 0 if
2
V̇ e, Θ̃ ≤ 2 e λmax (P) w0 − 2μ e P B Θ̃ + 2μ e P B Θ̃ Θ0
(9.99)
Then, V̇ e, Θ̃ ≤ 0 if
Θ Θ02 λmax (P) w0
0
Θ̃ ≥ + + =α (9.100)
2 4 μ P B
Therefore, V̇ e, Θ̃ ≤ 0 outside a compact set S defined by
S = e , Θ̃ : e ≤ p and Θ̃ ≤ α (9.101)
Therefore, both e and Θ̃ are uniformly ultimate bounded with the ultimate
bounds determined from Eqs. (9.63) and (9.64). Thus, the e modification results in
bounded tracking. As can be seen, the duality of tracking performance and robustness
cannot usually be simultaneously met.
Example 9.5 Consider the parameter drift example in Example 8.1. The response
of the closed-loop plant for the same disturbance in this example with p = − 12 5
,
x (0) = 1, k x (0) = 0, γx = 10, and μ = 0.1 is shown in Fig. 9.6. The e modification
produces all bounded signals of the closed-loop system.
9.4 e Modification 231
1.2 0
1.1 −5
u
x
1 −10
0.9 −15
0 10 20 0 10 20
t t
15 0
10
−5
x
w
k
−10
0
−5 −15
0 10 20 0 10 20
t t
Consider the SISO plant in Sect. 8.5, the e modification adaptive law for k x (t) is
where x̄m is the equilibrium reference state for a constant reference command signal
r (t).
Then,
x̄ 2 − x̄m ± μk̄ x x̄ ± μk̄ x x̄m = 0 (9.106)
which yields ⎡ ) ⎤
*
x̄m ± μk̄ x ⎣ * 4μ k̄ x̄
1 + +1 ∓ 2 ⎦
x m
x̄ = (9.107)
2 x̄m ± μk̄ x
where the upper sign is for eb > 0 and the lower sign is for eb < 0.
Using the equilibrium value of k x (t), the equilibrium state can be determined
explicitly as a function of the equilibrium reference state x̄m from the following
232 9 Robust Adaptive Control
equation:
This equation leads to an interesting observation. Since there are multiple roots,
only one of these roots is a valid solution. One possible solution is x̄ = x̄m , which
corresponds to asymptotic tracking. The other solution could be one of the other roots,
depending on the nature of the equilibrium. The closed-loop system is linearized as
x̃˙ = a + bk̄ x x̃ + b x̄ k̃ (9.109)
k̃˙ x = γx b x̄m − 2 x̄ ± μk̄ x x̃ ∓ γx bμ (x̄m − x̄) k̃ x (9.110)
Example 9.6 Consider Example 8.7 with a = 1, b = 1, and am = −1, the responses
of the closed-loop system with a constant reference command signal r (t) = 1 for
μ = 0.2 and μ = 0.8 are shown in Fig. 9.7.
For μ = 0.2, the solution of x (t) tends to the equilibrium reference state x̄m = 1,
and k x (t) tends to its the equilibrium value at k x∗ = −2. The roots of the polynomial
of x̄ for μ = 0.2 assuming eb < 0
x̄ 3 − 1.2 x̄ 2 + 0.2 = 0
are 1, 0.5583, and −0.3583. The eigenvalues of the Jacobian are (−0.5000 ±
17.3133i), (−4.3332, 46.7162), and (4.6460, 133.9710) for the respective roots.
Therefore, the equilibrium state is x̄ = 1 since it is the only stable equilibrium. The
tracking is therefore asymptotic as shown in Fig. 9.7.
For μ = 0.8, x (t) → x̄ = 1.3798, and k x (t) → k¯x = −1.7247. The roots of the
polynomial of x̄ for μ = 0.8 assuming eb < 0
x̄ 3 − 1.8x̄ 2 + 0.8 = 0
are 1.3798, 1, and −0.5798. The eigenvalues of the Jacobian are (−2.4781,
−150.1650), (−0.5000 ± 17.3133i), and (2.4523, 631.1908) for the respective
roots. The first equilibrium is a stable node and the second equilibrium is a stable
focus. The stable node is more attractive because all trajectories converge exponen-
tially faster to the first equilibrium than to the second equilibrium. Therefore, the
equilibrium state is x̄ = 1.3798. The tracking is no longer asymptotic as shown in
9.4 e Modification 233
Fig. 9.7. The equilibrium value of k x (t) is computed to be –1.7247 which agrees
with the simulation result.
Now, consider a time-varying reference command signal r (t) = sin 2t − cos 4t.
The responses of the closed-loop system μ = 0.2 and μ = 0.8 are shown in Fig. 9.8.
The tracking error is no longer asymptotic as can be seen. The closed-loop system is
non-autonomous, so it does not have a true equilibrium. As μ increases, the tracking
performance is degraded, while robustness increases. For small values of μ, it appears
that asymptotic tracking is almost achieved.
1.5 1.5
1 1
x
x
0.5 μ=0.2 0.5 μ=0.8
x x x x
m m
0 0
0 5 10 0 5 10
t t
0 0
−0.5
−1 μ=0.2
μ=0.8
kx
−1
k
−2
−1.5
−3 −2
0 5 10 0 5 10
t t
2 2
x x x x
m m
1 1
x, xm
x, xm
0 0
μ=0.2 μ=0.8
−1 −1
0 5 10 0 5 10
t t
2 2
0 0
kx
x
k
−2 −2
μ=0.2 μ=0.8
−4 −4
0 5 10 0 5 10
t t
The solution of the optimization must satisfy the state equation (9.114). There-
fore, the state equation imposes a dynamical constraint on the optimization. Thus,
the optimization is referred to as a constrained optimization. A standard problem
definition in optimal control can usually be stated as follows:
9.5 Optimal Control Modification 235
where λ (t) ∈ Rn is the Lagrange multiplier and is often called an adjoint or co-state
vector.
Note that the augmented term is zero, so in effect the cost function does not
change. We define a Hamiltonian function as
The standard tool in optimal control is calculus of variations. The cost function is
perturbed by a small variation δx in the state vector and variation δu in the control
vector. These variations result in the cost function being perturbed by a small variation
such that
, tf
J + δJ = H (x + δx, u + δu) − λ (ẋ + δ ẋ) dt (9.118)
t0
Then,
, tf
J + δJ = H (x, u) − λ ẋ dt
t
-0 ./ 0
J
236 9 Robust Adaptive Control
,
tf
∂H ∂H
+ δx + δu + λ̇ δx dt − λ t f δx t f + λ (t0 ) δx (t0 )
t0 ∂x ∂u
(9.121)
∂H
= ∇ Hu = 0 (9.124)
∂u
Since the initial condition x (t0 ) = x0 is specified for the initial value problem,
there is no variation in x (t) at t = t0 , i.e., δx (t0 ) = 0. At the final time t = t f ,
then
x t f is generally unknown, and so δx t f = 0. Therefore, the optimality enforces
a transversality condition
λ tf = 0 (9.125)
with the optimal control u ∗ (t) obtained from the optimality condition (9.124) when
there is no constraint on u (t), or more generally from the celebrated Pontryagin’s
minimum principle
u ∗ (t) = arg min H (x, u) (9.127)
u∈U
which states that the optimal control u ∗ (t) which belongs to an admissible set of
control U is one that minimizes H (x, u).
The Pontryagin’s maximum principle can deal with constraints on the control
such as u min ≤ u (t) ≤ u max . So, in determining an optimal control, we must also
evaluate the Hamiltonian with the bounding values of u (t) if u (t) is constrained.
1
∇ Hu = −2u + 1 = 0 ⇒ u =
2
The two-point boundary value problem is so-called because half of the boundary
conditions are at the initial time t = t0 and the other half are at the final time t = t f .
The solution method is generally quite complicated and usually must be solved
by numerical methods such as the shooting method. For a general optimization
problem, the gradient method has a certain appeal in that the system equations are
solved exactly at each iteration, with the control variable u ∗ (t) being perturbed
from one iteration to the next to “home in” on the optimal solution. The solution
method usually starts with an initial guess of the control. This then allows the state
vector to be computed by integrating the state equation forward in time using the
specified initial conditions. Once the state vector is computed, the adjoint equation is
integrated backward in time using the final time transversality conditions. The control
is then updated for the next iteration. The whole solution process is iterated until a
convergence on the control is obtained. A schematic of the gradient optimization is
shown in Fig. 9.9. The gradient method is given by
This update is also called the steepest descent method. Visualizing a bowl-shaped
Hamiltonian function H , the gradient ∇ Hu defines the magnitude and direction of
the local slope of the Hamiltonian function as illustrated in Fig. 9.10. Perturbing the
control by some function of ∇ Hu moves the control toward the bottom of the bowl. A
proper selection of the step size is critical for a rapid convergence to the minimizing
control. If ε is too small, the convergence may require a large number of iterations.
On the other hand, if ε is too large, the convergence may not occur at all. Therefore,
the effectiveness of a gradient method is predicated upon a judicious choice of the
weighting matrix ε. If the weighting matrix ε is the inverse of the Hessian matrix
∇ 2 Hu of the Hamiltonian function, then the gradient method is known as the second-
order gradient or Newton–Raphson method.
−1
u i+1 = u i − ∇ 2 Hu ∇ Hu (9.129)
The continuous time version of the steepest descend method is the gradient search
method
u̇ = −Γ ∇ Hu (9.130)
Now, consider a Linear Quadratic Regulator (LOR) optimal control design for a
MIMO system
ẋ = A p x + B p u (9.131)
y = Cx (9.132)
Then,
ė = y − r = C x − r (9.134)
Let z (t) = e (t) x (t) , then the augmented plant is expressed as
ż = Az + Bu − Dr (9.136)
We design an optimal control that minimizes the following linear quadratic cost
function: ,
1 tf
J = lim z Qz + u Ru dt (9.137)
t→∞ 2 0
1 1
H (z, u) = z Qz + u Ru + λ (Az + Bu − Dr ) (9.138)
2 2
The adjoint equation and the necessary condition of optimality are obtained as
∇ Hu = Ru + B λ = 0 ⇒ u ∗ = −R −1 B λ (9.140)
The optimal control problem can be solved by the “sweep” method [18] which
assumes an adjoint solution of λ (t) of the form
λ = Wz + Vr (9.141)
Then, substituting the assumed solution of λ (t) into the adjoint equation with the
optimal control u ∗ (t), we get
240 9 Robust Adaptive Control
⎡ ⎤
Ẇ z + W ⎣ Az − B R −1 B (W z + V r ) − Dr ⎦ + V̇ r = −Qz − A (W z + V r )
- ./ 0
ż
(9.142)
Separating terms by z (t) and r (t) yields the following equations:
Ẇ + W A + A W − WBR−1 B W + Q = 0 (9.143)
V̇ + A V − W B R −1 B V − W D = 0 (9.144)
subject to the transversality conditions W t f = V t f = 0.
Equation (9.143) is the celebrated differential Riccati equation in optimal control.
Since t f → ∞, this is called an infinite-time horizon control problem. Equations
(9.143) and (9.144) have the following algebraic solutions:
W A + A W − W B R −1 B W + Q = 0 (9.145)
−1
V = A − W B R −1 B WD (9.146)
u = K z z + Kr r (9.147)
K z = −R −1 B W (9.148)
−1
K r = −R −1 B A − W B R −1 B WD (9.149)
ż = Am z + Bm r (9.150)
where
Am = A − B R −1 B W (9.151)
−1
Bm = −B R −1 B A − W B R −1 B WD − D (9.152)
Given the MIMO system in Sect. 9.3, the optimal control modification seeks to min-
imize the L2 norm of the tracking error bounded away from the origin with a cost
9.5 Optimal Control Modification 241
function [9] , tf
1
J = lim (e − Δ) Q (e − Δ) dt (9.153)
t f →∞ 2 0
where Δ (t) represents an unknown lower bound of the tracking error, subject to
tracking error dynamics described by
ė = Am e + B Θ̃ Φ (x) − w (9.154)
The cost function J is convex and represents the weighted norm square measured
from a point on the trajectory of e (t) to the normal surface of a hypersphere BΔ =
{e (t) ∈ Rn : e ≤ Δ} ⊂ D ⊂ Rn as illustrated in Fig. 9.11. The cost function
is designed to provide robustness by not seeking the ideal property of asymptotic
tracking of MRAC whereby e (t) → 0 as t → ∞, but rather a bounded tracking
whereby the tracking error tends to some lower bound Δ (t) away from the origin.
By not requiring asymptotic tracking, the adaptation therefore can be made more
robust. Thus, the tracking performance can be traded with robustness by a suitable
design of the optimal control modification adaptive law.
λ̇ = −∇ He = −Q (e − Δ) − A
mλ (9.156)
with the transversality condition λ t f → ∞ = 0 since e (0) is known.
Treating Θ̃ (t) as a control variable, the necessary condition is obtained as
The optimal control modification adaptive law is then expressed by the gradient
update law as
Θ̃˙ = −Γ ∇ HΘ̃ = −Γ Φ (x) λ B (9.158)
Note that the adaptive law is dependent upon the adjoint vector which is governed
by the adjoint equation. To eliminate λ (t), we use the “sweep” method by assuming
the following adjoint solution:
λ = Pe + SΘ Φ (x) (9.159)
d Θ Φ (x)
Ṗe + P Am e + B Θ − Θ ∗ Φ (x) − w + ṠΘ Φ (x) + S
dt
= −Q (e − Δ) − Am Pe + SΘ
Φ (x) (9.160)
Separating the equation that contains e (t), Θ (t) Φ (x), and the remaining terms
result in
Ṗ + P Am + A mP + Q = 0 (9.161)
Ṡ + A
mS + PB = 0 (9.162)
∗ d Θ Φ (x)
QΔ + P BΘ Φ (x) + Pw − S =0 (9.163)
dt
subject to the transversality conditions P t f → ∞ = 0 and S t f → ∞ = 0.
Equation (9.161) is the differential Lyapunov equation. Equation (9.163) shows
that the unknown lower bound Δ (t) of the tracking error is a function of the para-
metric uncertainty Θ ∗ and the unknown disturbance w (t). Thus, as long as the
uncertainty and disturbance exist, the lower bound will be finite, thereby causing the
tracking error to be bounded away from the origin as opposed to tending to the origin
asymptotically as required by MRAC. Robustness thereby is achieved.
For the infinite-time horizon optimal control problem as t → ∞, the solutions
to both Eqs. (9.161) and (9.162) tend to their constant solutions at t = 0 which are
given by
P Am + A mP + Q = 0 (9.164)
A
mS + PB = 0 (9.165)
λ = Pe − A−
m P BΘ Φ (x) (9.167)
Substituting the adjoint solution into the gradient update law in Eq. (9.158) yields
the “ideal” optimal control modification adaptive law
Θ̇ = −Γ Φ (x) e P − Φ (x) Θ B P A−1
m B (9.168)
The first term in the adaptive law is the familiar MRAC. The second term is the
damping term of the optimal control modification. So, this analysis is able to show
the connection between adaptive control and optimal control.
In any design, the ability to adjust a controller to obtain a desired tracking per-
formance and robustness is important. Therefore, suboptimal solutions may provide
a more flexible, yet practical approach to a control design by allowing a trade-off
between the optimality and other design considerations. Thus, to enable the optimal
control modification adaptive law to be sufficiently flexible for a control design, a
modification parameter ν > 0 is introduced as a gain to allow for the adjustment
of the optimal control modification term. The role of the modification parameter ν
is important. If the tracking performance is more desired in a control design than
robustness, then ν could be selected to be a small value. In the limit when ν = 0,
the standard MRAC is recovered and asymptotic tracking performance is achieved
but at the expense of robustness. On the other hand, if robustness is a priority in a
design, then a larger value of ν should be chosen.
Thus, the solution of S is modified as
S = −ν A−
m PB (9.169)
C=M+N (9.172)
where
1
M = M = C + C (9.173)
2
1
N = −N = C − C (9.174)
2
x C x = x M x + x N x (9.175)
x C x = x M x (9.176)
1 1 −
M= P A−1 − −1
m + Am P = − Am Q Am < 0 (9.177)
2 2
Therefore, x B P A−1 1 − −1
m Bx = − 2 x B Am Q Am Bx < 0. Thus, the optimal
control modification term is negative definite with respect to Θ (t).
Consider a SISO system in Sect. 8.1. The optimal control modification adaptive
law is given by
k̇ x = −γx x 2 b − νx 2 am−1 b2 k x (9.178)
Since am < 0, the modification term is negative with respect to k x (t). The feed-
back gain k x (t) can be computed as
k̇ x
= −γ x 2 b (9.179)
1 − νam−1 bk x
1 1 − νam−1 bk x
− ln = k x∗ − k x∗ (0) (9.182)
νam−1 b 1 − νam−1 bk x (0)
This yields
1 1 − νam−1 bk x (0)
kx = − exp −νam−1 b k x∗ − k x (0) (9.183)
νam−1 b νam−1 b
In the case of the parameter drift, k x∗ (t) → −∞ for b > 0 with the standard
MRAC, therefore −νam−1 bk x∗ (t) → −∞. This implies that k x (t) is bounded with
the optimal control modification.
The optimal control modification exhibits some nice properties that make it easy
to analyze. Since k x (t) is bounded, the equilibrium value of k x (t) can be analytically
obtained by letting γx → ∞ which yields
1
k̄ x = (9.184)
νam−1 b
The equilibrium value of k x (t) can also be obtained from the solution of k x (t) as
k x∗ (t) → −∞ which yields the same result.
The equilibrium value of k x (t) thus is inversely proportional to the optimal control
modification parameter ν.
k x (t) can then be expressed as
∗
k x (0) k x − k x (0)
k x = k̄ x − k̄ x 1 − exp − (9.185)
k̄ x k̄ x
The first case is when the open-loop plant is stable, i.e., a < 0. Then, the optimal
control modification parameter ν can take on any positive value. Thus, if the plant
246 9 Robust Adaptive Control
Example 9.8 Consider the parameter drift example in Example 8.1, k x∗ (t) is given
by
(1 + t)2n+1 − 1
k x∗ − k x∗ (0) = −γx b
2n + 1
k x∗ (t) → −∞ if n > − 21 .
With the optimal control modification, k x (t) is obtained as
1 1 − νam−1 bk x (0) (1 + t)2n+1 − 1
kx = − exp γx νam−1 b2
νam−1 b νam−1 b 2n + 1
1.2
−5
1.1
u
x
−10
1
0.9 −15
0 10 20 0 10 20
t t
15 0
10
x
w
5 −5
k
−5 −10
0 10 20 0 10 20
t t
The optimal control modification can be shown to achieve stable and bounded track-
ing by the Lyapunov’s direct method as follows:
Then,
V̇ e, Θ̃ = −e Qe − 2e Pw + 2νΦ (x) Θ̃ B P A−1
m B Θ̃ Φ (x)
But, B P A−1
m B < 0. So,
Let c1 = λmin (Q), c2 = λmaxc(P)w 1
0
, c3 = λmin B A− −1
m Q Am B , and c4 =
m B Θ0
B P A−1
c3
. Then, upon completing the square, we get
2
V̇ e, Θ̃ ≤ −c1 (e − c2 )2 +c1 c22 −νc3 Φ (x)2 Θ̃ − c4 +νc3 c42 Φ (x)2
(9.194)
V̇ e, Θ̃ ≤ 0 implies
248 9 Robust Adaptive Control
νc3 c42 Φ (x)2
e ≥ c2 + c22 + =p (9.195)
c1
c1 c22
Θ̃ ≥ c4 + c42 + =α (9.196)
νc3 Φ (x)2
Note that the lower bounds p and α are dependent on Φ (x). Therefore, to prove
boundedness, we also need to show that Φ (x) is bounded. If Φ (x) is a bounded
function such as sin x, sigmoidal or radial basis functions, then Φ (x) ≤ Φ0 and
the solution is uniformly ultimately bounded. The ultimate bounds can be established
as
λmax (P) p 2 + λmax Γ −1 α 2
e ≤ ρ = (9.197)
λmin (P)
λmax (P) p 2 + λmax Γ −1 α 2
Θ̃ ≤ β = (9.198)
λmin Γ −1
ė = Am e − BΘ ∗ Φ (x) (9.199)
V (e) = e Pe (9.200)
Then,
λ2 (Q) e2 2
V̇ e, Θ̃ ≤ −c1 (e − c2 )2 + c1 c22 − νc3 min Θ̃ − c4
4 P B2 Θ02
λ2min (Q) e2
+ νc3 c42 (9.203)
4 P B2 Θ02
! "
ν λ2 (Q) e2
2
V̇ e, Θ̃ ≤ −c1 1 − e2 + 2c1 c2 e − νc3 min Θ̃ − c4
νmax 4 P B2 Θ02
(9.205)
! "5 62
ν c2
V̇ e, Θ̃ ≤ −c1 1 − e − ν
νmax 1 − νmax
c1 c22 νc3 λ2min (Q) e2 2
+ ν − Θ̃ − c4 (9.206)
1 − νmax 4 P B2 Θ02
The lower bound of e for which V̇ e, Θ̃ ≤ 0 is determined from
! "
ν
− c1 1 − e2 + 2c1 c2 e ≤ 0 (9.207)
νmax
which yields
2c2
e ≥ ν = p (9.208)
1 − νmax
The lower bound of Θ̃ is determined by setting e = c2
ν that renders
1−
νmax
c1 c22 νc3 λ2min (Q) c22 2
V̇ Θ̃ ≤ ν − 2 Θ̃ − c4 (9.209)
1− ν
4 1 − νmax P B2 Θ02
νmax
250 9 Robust Adaptive Control
Therefore, V̇ Θ̃ ≤ 0 implies
)
*
* 4c1 1 − ν P B2 Θ 2
+ νmax 0
Θ̃ ≥ c4 + =α (9.210)
νc3 λmin (Q)
2
Then, the closed-loop system is uniformly ultimately bounded with the ultimate
bounds given by Eqs. (9.197) and (9.198).
2. The closed-loop plant with the nominal controller has no stability guarantee. The
If ν = 0, then from Eqs. (9.195) and (9.196),
uncertainty can be destabilizing.
we see that e is bounded but Θ̃ is unbounded. This illustrates the parameter
drift behavior of MRAC. On the other hand, if ν → ∞, then Θ̃ is bounded,
but e can be unbounded. Thus, there exists a maximum value νmax such that
ν < νmax which corresponds to the largest value of Φ (x) ≤ Φ0 that renders
the largest ultimate bound of e. From Eq. (9.194), the largest ultimate bound
of e is determined from
V̇ e, Θ̃ ≤ −c1 (e − c2 )2 + c1 c22 + νc3 c42 Φ (x)2 (9.211)
But,
Therefore,
ϕ (x , xm , Q, ν, w0 , Θ0 ) = − c1 − νc3 c42 x2 + 2 (c1 c2 + c5 xm ) x
+ 2c1 c2 xm − c1 xm 2 (9.219)
x = ϕ −1 (xm , Q, ν, w0 , Θ0 )
7
c2 + c5 xm + (c1 c2 + c5 xm )2 + 4 c1 + νc3 c42 2c1 c2 xm − c1 xm 2
=
c1 − νc3 c42
(9.220)
for any 0 < ν < νmax such that c1 − νc3 c42 > 0 which yields
c1 λmin (Q) λmin B A− −1
m Q Am B
νmax = = 2 2 (9.221)
c3 c42 B P A−1
m B
Θ
0
Note that νmax is now dependent on the upper bound of the parametric uncertainty
Θ0 and is a conservative estimate since the negative definite term with ||Θ̃||2 in
Eq. (9.194) is neglected. As the bound of the uncertainty increases, νmax must
be reduced to ensure the stability of the closed-loop system. Thus, the stability
of the optimal control modification depends on the bound of the uncertainty. The
adaptive law is guaranteed to be stable if a priori knowledge of the bound of the
uncertainty exists.
control modification adaptive law is more restrictive. However, the lack of a well-
defined limit on the modification parameter can sometimes be a source of stability
issues. Numerical simulations using the Rohrs counterexample have shown that both
the σ modification and e modification as well as the optimal control modification all
exhibit limiting values on the modification parameters σ , μ, and ν, respectively.
Is should be noted that the limiting value νmax derived from the Lyapunov analysis,
in general, is a conservative estimate of the true limiting value νmax . Consider the
following first-order SISO system with an adaptive regulator controller using the
optimal control modification:
ẋ = ax + b u + θ ∗ x (9.222)
u = k x x − θ (t) x (9.223)
θ̇ = −γ −x 2 b − νx 2 am−1 b2 θ (9.224)
Note that the adaptive law implicitly uses p = 1 where p is the solution of the
scalar Lyapunov equation
2 pam = −q (9.226)
1
θ̄ = − (9.227)
νam−1 b
4b2 am−2 b2
νmax = =4 (9.229)
b4 am−2
If the uncertainty is destabilizing but the closed-loop plant with the nominal
(non-adaptive) controller u (t) = k x x (t) is stable, then 0 < bθ ∗ < −am . Then, the
closed-loop system is still stable for any ν > 0. Yet, using Eq. (9.221), ν should be
limited by
9.5 Optimal Control Modification 253
! "2
4am2 b2 am−2 2am
νmax = = (9.230)
b4 am−2 θ ∗2 bθ ∗
Finally, if the uncertainty is destabilizing and the closed-loop plant with the nom-
inal controller is unstable, then bθ ∗ > −am . The closed-loop plant is stable if
am
am + + bθ ∗ < 0 (9.231)
ν
Then, ν is limited by
am
νmax = − (9.232)
am + bθ ∗
m 2 2
Let bθ ∗ = −αam where α > 1, then 2a bθ ∗
→ α2 and − am +bθ am
∗ → α−1 .
1
2 2
Since α < α−1 1
for all α > 1, therefore the most conservative value of νmax is
established by the Lyapunov analysis which is equal to
4
νmax = (9.234)
α2
On the other hand, the least conservative value of νmax is the true value of νmax
which is equal to
1
νmax = (9.235)
α−1
While the optimal control modification is defined according to Eq. (9.170), there
are other variances of the optimal control modification. These variances are:
Θ̇ = −Γ Φ (x) e P + νΦ (x) Θ B A− −1
m Q Am B (9.236)
Θ̇ = −Γ Φ (x) e P + νΦ (x) Θ B R B (9.237)
Θ̇ = −Γ Φ (x) e P B + νΦ (x) Θ R (9.238)
Θ̇ = −Γ Φ (x) e P B + νΦ (x) Θ (9.239)
u = K x x + K r r − Θ (t) x (9.241)
ẋm = Am xm + Bm r (9.242)
where Am = A + B K x and Bm = B K r .
The optimal control modification adaptive law for Θ (t) is given by
Θ̇ = −Γ xe P B − νx x Θ B P A−1
m B (9.243)
It has been shown that MRAC is non-robust with fast adaptation. The time-delay
margin of MRAC tends to zero as the adaptation rate tends to infinity. The optimal
control modification exhibits a linear asymptotic property as Γ → ∞ since the
equilibrium value Θ (t) is independent of x (t) if xm (t) = 0, as has been shown for
simple SISO systems [20, 21]. The equilibrium value of Θ (t) x (t) is given by
1 − −1
Θ̄ x = B Am P B B Pe (9.244)
ν
Then, the closed-loop system tends to the following asymptotic linear system as
Γ → ∞ in the limit:
1 −1 ∗
ẋ = Am + B B A− P B B P + BΘ x
ν m
1 −1
− B B A− m PB B P xm + Bm r (9.245)
ν
This system is stable for all ν > 0 if Am + BΘ ∗ is Hurwitz and stable for
ν < νmax if Am + BΘ ∗ is an unstable matrix.
9.5 Optimal Control Modification 255
Since P > 0 and Q > 0, −P −1 Q < 0. The closed-loop poles of the ideal system
with Θ ∗ = 0 are all negative real. Therefore, the closed-loop system has the best
stability margin. The ideal system is exponentially stable with no high-frequency
oscillations. By choosing Q appropriately, the closed-loop system is guaranteed to
be stable if −P −1 Q + BΘ ∗ is Hurwitz.
The linear asymptotic property of the optimal control modification is quite useful
since it can be used for stability analysis using many available linear analysis tools.
Another advantage that comes with the linear asymptotic property is the scaled input–
output behavior of the closed-loop system. That is, if r (t) is scaled by a multiplier
c, then x (t) is scaled by the same amount. To see this, we express the asymptotic
closed-loop system as a transfer function
1 − −1 ∗
sx = Am + B B Am P B B P + BΘ x
ν
1 −1
− B B A−
m PB B P xm + Bm r (9.248)
ν
From the reference model, we get
Therefore,
−1
1 − −1 ∗
x = s I − Am − B B Am P B B P − BΘ
ν
1 − −1 −1
× − B B Am P B B P (s I − Am ) + I Bm r (9.250)
ν
If x (t) = x0 (t) is the response due to r (t) = r0 (t) and if r (t) is scaled by a
constant multiplier c, i.e., r (t) = cr0 (t), then it follows that x (t) is also scaled by
the same multiplier c, i.e., x (t) = cx0 (t). The scale input–output behavior makes
the optimal control modification more predictable than the standard MRAC which
does not possess this linear asymptotic property.
256 9 Robust Adaptive Control
If r (t) is a constant signal, then the equilibrium value of x (t) can be found by
setting s = 0 as t → ∞. Then,
−1
1 −1 ∗
x̄ = − Am + B B A− P B B P + BΘ
ν m
1 − −1
× B B Am P B B P A−1 + I Bm r (9.251)
ν m
The linear asymptotic property also affords another advantage in that the stability
margins of the system in the limit can be computed.
Consider a first-order time-delay SISO system with an optimal control modifica-
tion adaptive controller
ẋ = ax + b u (t − td ) + θ ∗ x (9.255)
9.5 Optimal Control Modification 257
u = k x x + kr r − θ (t) x (9.256)
θ̇ = −γ xeb − νx 2 am−1 b2 θ (9.257)
This is the expected result for MRAC. For any 0 < ν < νmax , the optimal
control modification yields a non-zero time-delay margin with fast adaptation. This
is a robustness property of any robust adaptive control with the ability to provide
sufficient stability margins under fast adaptation. For a given time delay td and a priori
knowledge of θ ∗ , the modification parameter ν thus can be computed to guarantee
stability of the closed-loop system.
Choose ν = 0.1 < 1. The time-delay margin for the closed-loop system with the
optimal control modification is calculated as
$
1
ω= − 1 = 9.9499 rad/s
ν2
1
td = 7 cos−1 ν = 0.1478 sec
1
ν2
−1
Since the time-delay margin td decreases with increasing the adaptation rate γ ,
therefore, for finite γ , the time-delay margin estimated with γ → ∞ is the lowest
estimate of the time-delay margin for any finite γ < ∞. In other words, the estimated
time-delay margin td = 0.1478 s using the linear asymptotic property provides a
lower bound on the time-delay margin. The actual time-delay margin for any arbitrary
adaptation rate γ should be greater than this value in theory. We see that, for even
a small value of ν, a significant increase in the time-delay margin can be achieved
as compared to the time-delay margin of 0.0020 s for the standard MRAC with
γx = 500. In the limit as ν → 0, td → 0 corresponding to the standard MRAC as
γx → ∞.
The steady-state error is estimated to be
−1 ! 1 "
−1 am ∗
ē = −am + am + + bθ + 1 bm r = −0.0185
ν ν
ē ē
θ̄ = = = 0.1818
νam−1 b x̄ νam−1 b (x̄m − ē)
9.5 Optimal Control Modification 259
We choose γ = 100 and θ (0) = 0. The results of the simulation with the optimal
control modification are shown in Fig. 9.13. The closed-loop system is completely
stable. The tracking error e (t)and the adaptive parameter θ (t) converge to –0.0185
and 0.1818, respectively, at 10 s. Thus, the simulation results agree exactly with the
equilibrium values of ē and θ̄ .
1.5 0
−0.005
1
x, xm
−0.01
e
0.5
x −0.015
xm
0 −0.02
0 5 10 0 5 10
t t
1 0.4
0.3
0
0.2
u
−1
0.1
−2 0
0 5 10 0 5 10
t t
When a time delay equal to 0.1477 s is injected at the input, the closed-loop system
is on the verge of instability as shown in Fig. 9.14. In fact, for γ = 1000 using a
time step Δt = 0.0001, the closed-loop system becomes unstable at 0.1478 s as
predicted. Thus, the numerical evidence of the time-delay margin is in agreement
with the analytical prediction.
To illustrate the linear asymptotic property, we consider the asymptotic linear
system in the limit as follows:
xm − x
ẋ = ax + bθ ∗ x − b
νam−1 b
Figure 9.15 shows the closed-loop response with the optimal control modification
for r (t) = 1 + sin 2t + cos 4t. The closed-loop response follows the asymptotic
linear system almost perfectly as both responses are on top of each other.
To illustrate the scaled input–output linear behavior of the optimal control modi-
fication, the reference command signal r (t) = 1 + sin 2t + cos 4t is doubled so that
the new command is r1 (t) = 2 + 2 sin 2t + 2 cos 4t. Let x1 (t) be the closed-loop
response to r1 (t). Then, the response of x1 (t) is twice that of x (t) as shown in
Fig. 9.16.
260 9 Robust Adaptive Control
3 2
x x
m
2 1
x, xm
1 0
e
0 −1
−1 −2
0 5 10 0 5 10
t t
20 40
10 20
0 0
u
θ
−10 −20
−20 −40
0 5 10 0 5 10
t t
1.6
1.4
1.2
|
x, x
0.8
0.6
0.4
0.2
x
−
x
0
0 2 4 6 8 10
t
4
x x
1
3
1
x, x
0
0 2 4 6 8 10
t
2.005
2
x1/x
1.995
1.99
0 2 4 6 8 10
t
Δė = Am Δe (9.270)
Then, the closed-loop plant follows the reference model exactly and the stability
of the reference model is preserved.
This leads to a hypothesis that if the quantity Θ (t) Φx (x) is minimized, as-
ymptotic tracking is achieved, and the stability margin of the reference model is
preserved. This can be established by minimizing the following cost function:
1
J= Φ (x) ΘΘ Φx (x) (9.271)
2 x
as an unconstrained optimization.
The gradient of the cost function is computed as
Adding the standard MRAC to the update law results in the adaptive loop recovery
modification.
The original stability proof is based on a singular perturbation approach [6, 22].
The reader is referred to the original work [6, 22] for the formal proof. It can be shown
that the stability proof of the optimal control modification can also be applied to the
adaptive loop recovery modification since both adaptive laws have their damping
terms proportional to some positive-definite function of x (t). The alternate proof is
as follows:
2
V̇ e, Θ̃ ≤ −λmin (Q) e2 + 2λmax (P) e w0 − 2η Φx (x)2 Θ̃
+ 2η Φx (x)2 Θ̃ Θ0 (9.276)
λmax (P) w0
Let c1 = λmin (Q) and c2 = . Upon completing the square, we get
c1
! "2
Θ0 η
V̇ e, Θ̃ ≤ −c1 (e − c2 )2 + c1 c22 − 2η Φx (x)2 Θ̃ − + Φx (x)2 Θ02
2 2
(9.277)
Then, V̇ e, Θ̃ ≤ 0 implies
η Φx (x)2 Θ02
e ≥ c2 + c22 + =p (9.278)
2c1
Θ Θ02 c1 c22
0
Θ̃ ≥ + + =α (9.279)
2 4 2η Φx (x)2
9.6 Adaptive Loop Recovery Modification 263
Note that the lower bounds p and α are dependent on Φx (x). Therefore, to
prove boundedness, we need to show that Φx (x) is bounded.
There exists a maximum value ηmax such that η < ηmax for which V̇ e, Θ̃ ≤ 0.
Let
It follows that the closed-loop system is also uniformly ultimately bounded for
any 0 < η < ηmax .
Consider a special case when Φ (x) belongs to a class of regressor functions such
that Φ (x) ≤ x2 or Φx (x) ≤ 2 x. Then, it can be shown that the adaptive
loop recovery modification is stable and bounded for any 0 ≤ η < ηmax with a
conservative estimate of ηmax given by
λmin (Q)
ηmax = (9.282)
2Θ02
Consider another special case when Φ (x) belong to a class of regressor func-
tions such that Φ (x) ≤ x or Φx (x) ≤ 1. Then, the adaptive loop recovery
modification is unconditionally stable.
Since the possibility of Φx (x) being unbounded exists, the adaptive loop recovery
modification only works for any Φ (x) such that Φx (x) is bounded. For example, if
Φ (x) = x p , then the adaptive law is valid only for p ≥ 1. If p = 1, then Φx (x) = 1,
and the adaptive loop recovery modification becomes the σ modification.
Example 9.10 Consider a first-order SISO plant
ẋ = ax + b u + θ ∗ φ (x)
u = k x x + kr r − θ (t) φ (x)
Consider the fast adaptation condition when γ → ∞, then the asymptotic behav-
ior of the adaptive loop recovery modification is given by
(xm − x) bx 2
θ̄ x 2 = −
4η
(xm − x) bx 2
ū = k x x + kr r +
4η
Let r (t) be a constant command signal. Then, the equilibrium value of x (t) can
be determined from
−am b2 x̄ 3 + 4ηbam θ ∗ − b2 bm r x̄ 2 + 4ηam2 x̄ + 4ηam bm r = 0
1.5 0
−0.1
1
x, xm
−0.2
e
0.5
−0.3
x x
m
0 −0.4
0 5 10 0 5 10
t t
2 0.4
1 0.3
0 0.2
u
−1 0.1
−2 0
0 5 10 0 5 10
t t
9.6 Adaptive Loop Recovery Modification 265
Consider a special case when the asymptotic behavior of the adaptive loop re-
covery modification results in a linear controller for a certain function φ (x). The
adaptive law is expressed in general as
θ̇ = −γ φ (x) eb + ηφx2 (x) θ (9.283)
where c > 0, as γ → ∞.
This results in the differential equation
dφ
= cd x (9.285)
φ
Then, the adaptive loop recovery modification and the optimal control modifica-
tion have the same linear asymptotic property if ηc2 = −νb2 am−1 . Note that even
though the adaptive controller tends to a linear controller asymptotically, the closed-
loop plant is still nonlinear since φ (x) is nonlinear.
Example 9.11 Consider a time-delay first-order SISO system
ẋ = ax + b u (t − td ) + θ ∗ e−x
The adaptive controller with the adaptive loop recovery modification is given by
u = k x x + kr r − θ (t) e−x
θ̇ = −γ e−x eb + ηe−2x θ
eb
u → k x x + kr r +
η
Even though the adaptive controller is linear in the limit, the closed-loop plant is
nonlinear due to the nonlinear uncertainty since
266 9 Robust Adaptive Control
! "
b b
ẋ = ax + b kx − x (t − td ) + kr r (t − td ) + xm (t − td ) + θ ∗ e−x
η η
1.5 0.2
0
1
m
x, x
−0.2
e
0.5
−0.4
x xm
0 −0.6
0 5 10 0 5 10
t t
1 3
0 2
u
−1 1
−2 0
0 5 10 0 5 10
t t
Fig. 9.18 Closed-loop response with adaptive loop recovery modification with td = 0.1 s
The L1 adaptive control has gained a considerable attention in recent years due to
its ability to achieve robustness with fast adaptation for a given a priori bound on
the uncertainty. The L1 adaptive control was developed by Hovakimyan and Cao in
2006 [8, 23, 24]. Since then, it has been widely used in adaptive control practice.
The underlying principle of the L1 adaptive control is the use of fast adaptation for
improved transient or tracking performance coupled with a low-pass filter to suppress
high-frequency responses for improved robustness. As a result, the L1 adaptive
control can be designed to achieve stability margins under fast adaptation for a given
a priori bound on the uncertainty. Theoretical proofs of guaranteed robustness bounds,
stability margins, and tracking performance in transient and steady-state responses
for the L1 adaptive control are well-established [8].
Consider the following linear SISO uncertain system with a matched uncertainty:
ẋ = Am x + B u + Θ ∗ (t) (t) x + d (t) (9.287)
y = Hx (9.288)
(9.289)
! " ! " 4
dmax + dmin dmax − dmin 2
Sd = d (t) ∈ R : gd (d) = d − − ≤0
2 2
(9.290)
A state-predictor model is defined as
x̂˙ = Am x̂ + B u + Θ (t) x + d̂ (t) (9.291)
268 9 Robust Adaptive Control
with the initial condition x̂ (0) = x0 , where x̂ (t), Θ (t) and d̂ (t) are estimates of
x (t), Θ ∗ (t), and d (t), respectively.
We define the state-predictor error as e p (t) = x̂ (t) − x (t). Then, the estimates of
Θ ∗ (t), and d (t) are computed from the following adaptive laws with the projection
method:
⎧
⎪
⎨−Γ xe p P B if g (Θ) < 0 or if
g (Θ) = 0
Θ̇ = Pro Θ, −Γ xe p P B = and − xe p P B ∇gΘ (Θ) ≤ 0
⎪
⎩
0 otherwise
⎧ (9.292)
⎪
⎪ ⎨−γd e p P B if gd d̂ < 0 or ifgd d̂ = 0
˙
d̂ = Pro d̂, −γd e p P B =
⎪ and − ep P B∇gd d̂ ≤ 0
⎪
⎩
0 otherwise
(9.293)
with the initial conditions Θ (0) = Θ0 and d̂ (0) = d̂0 .
The adaptive laws can be shown to be stable as follows:
d̃ 2
V e p , Θ̃, d̃ = ep Pe p + Θ̃ Γ −1 Θ̃ + (9.295)
γd
The projection
method guarantees that Θ (t) and d̂ (t) stay inside their constraint
sets. Then, V̇ e p , Θ̃, d̃ is evaluated as
2d̃ ḋ
V̇ e p , Θ̃, d̃ = −ep Qe p − 2Θ̃ Γ −1 Θ̇ ∗ −
γd
2 −1
2 d̃ δ
≤ −λmin (Q) e p + 2λmax Γ Θ̃ Δ + (9.296)
γd
This implies
θ̃i ≤ max θimin − θi∗ , θimax − θi∗ ≤ max θimin , θimax + max θi∗ (9.299)
d̃ ≤ max (|dmin − d| , |dmax − d|) ≤ max (|dmin | , |dmax |) + max |d| = d̃0
(9.300)
Let Θ̃0 = max max θimin , θimax + max θi∗ . Then,
2
V̇ e p , Θ̃, d̃ ≤ −λmin (Q) e p + 2λmax Γ −1 Θ̃0 Δ + 2γd−1 d̃0 δ (9.301)
V̇ e p , Θ̃, d̃ ≤ 0 implies
2λmax Γ −1 Θ̃0 Δ + 2γd−1 d̃0 δ
e p ≥ =p (9.302)
λmin (Q)
There exists an upper bound of e p along with the upper bounds of Θ̃ and
d̃ that establish a compact set outside which V̇ e p , Θ̃, d̃ ≤ 0. Therefore, all
signals are bounded.
where k > 0 is a feedback gain, D (s) is a proper transfer function and w (s) is the
Laplace transform of
1
kr = − (9.305)
H A−1
m B
ẋ = Am x + Bkr r (9.307)
270 9 Robust Adaptive Control
The L1 reference controller is then given by the filtered version of the ideal
controller
u m (s) = C (s) u ∗ (s) (9.308)
k D (s)
C (s) = (9.309)
1 + k D (s)
We require C (0) = 1. Thus, D (s) can be any suitable transfer function that
1
satisfies lims→0 D(s) = 0.
The design of the L1 adaptive control then essentially comes down to the design
of the feedback gain k and the transfer function D (s) which defines the low-pass
filter C (s).
The closed-loop reference system formed by the L1 reference controller is then
obtained as
sx (s) = Am x (s) + B C (s) u ∗ (s) + kr r (s) − u ∗ (s) (9.310)
n
∗
L = Θ ∗ 1 = max θ
i (9.311)
i=1
Then, for the closed-loop reference system to be stable, the L1 norm of the system
transfer function G (s) must satisfy the following condition:
G (s) = (s I − Am )−1 B [1 − C (s)]1 L < 1 (9.312)
This condition is the stability condition on the unit circle for discrete-time which
is equivalent to the Hurwitz stability condition of the matrix Am + B [1 − C (s)] Θ ∗
for continuous time.
The choices of k and D (s) can greatly influence the performance and stability of
the L1 adaptive control. For example, consider the following transfer function:
1
D (s) = (9.313)
s
which satisfies the ideal closed-loop transfer function
k
C (0) = =1 (9.314)
s + k s=0
k
u = − w (s) (9.315)
s
or, in the time domain,
u̇ = −ku − k Θ (t) x + d̂ (t) − kr r (9.316)
Consider the case when Θ ∗ and d are constant. Then, the closed-loop reference
system becomes
ẋ Am + BΘ ∗ B x Bd
= + (9.317)
u̇ −kΘ ∗ −k u −k (d − kr r )
1
Choose D (s) = . Then, the L1 adaptive controller is defined as
s
k
u = − w (s) (9.321)
s
θ̇ = Pro θ, −γ xe p b (9.322)
d̂˙ = Pro d̂, −γd e p b (9.323)
s − am − bθ ∗ −b
det (s I − Ac ) = = s 2 + k − am − bθ ∗ s − kam = 0
kθ ∗ s +k
(9.325)
Comparing the characteristic equation to that of a second-order system
s 2 + 2ζ ωn s + ωn2 = 0 (9.326)
then
ωn2 = −kam (9.327)
2ζ ωn = k − am − bθ ∗ (9.328)
which yields
7
k = am + bθ ∗ − 2ζ 2 am + 2ζ am − (am + bθ ∗ ) + ζ 2 am (9.330)
2. The open-loop plant with uncertainty is unstable with am + bθ ∗ > 0. Then, the
closed-loop poles are computed as
k − am − bθ ∗ 4kam
s=− 1± 1+ (9.332)
2 (k − am − bθ ∗ )2
Example 9.12 Consider the parameter drift example in Example 8.1. Let am = −1,
w (t) 5
b = 1, θ ∗ = 2, and d (t) = with p = − .
b 12
1
Choose D (s) = . Since am + bθ ∗ = 1 > 0, we choose k = 2 > 1. Then, the
s
L1 adaptive regulator controller is specified by
9.7 L1 Adaptive Control 273
u̇ = −2u − 2 θ x + d̂
⎧
⎪−γ xe p b if θmin < θ < θmax or if θ = θmin
⎨
θ̇ = Pro θ, −γ xe p b = and θ̇ ≥ 0 or if θ = θmax and θ̇ ≤ 0
⎪
⎩
0 otherwise
⎧
⎨−γd e p b if dmin ≤ d̂ ≤ dmax or if d̂ = dmin
⎪
˙
d̂ = Pro d̂, −γd e p b = and d̂˙ ≥ 0 or if d̂ = dmax and d̂˙ ≤ 0
⎪
⎩
0 otherwise
3 5
x
2 x^ 0
<
x, x
1 −5
u
0 −10
−1 −15
0 50 100 0 50 100
t t
15 4
10 3
<
d, d
5 2
θ
0 d 1
d^
−5 0
0 50 100 0 50 100
t t
3 5
x
2 x^ 0
<
x, x
1 −5
u
0 −10
−1 −15
0 50 100 0 50 100
t t
15 6
10
4
<
d, d
5
θ
2
0 d
d^
−5 0
0 50 100 0 50 100
t t
We choose θmin = 0, θmax = 3, dmin = −2, and dmax = 11. To ensure fast
adaptation, γ and γd must be chosen to be large values. The time step must be chosen
sufficiently small if γ and γd are large. Choose Δt = 0.001 and γ = γd = 100. The
initial conditions are x (0) = 0, x̂ (0) = 0, θ (0) = 0, and d̂ (0) = 0. The response of
the closed-loop system with the L1 adaptive controller is shown in Fig. 9.19. Notice
that θ (t) → θmax because the upper bound of θ (t) is set too small. Nonetheless,
x (t) → ε as t → ∞ and d̂ (t) → d (t) + δ.
Suppose the upper bound is increased to θmax = 10. Then, the response with the
new upper bound of θ (t) is shown in Fig. 9.20. The response shows that θ (t) →
θ̄ = 5.1011 < θmax . Note that θ (t) does not tend to θ ∗ .
9.8 Normalization
Fast adaptation can lead to loss of robustness of MRAC. Fast adaptation is usually
associated with the use of a large adaptation rate. However, this is not the only effect
of fast adaptation. When the amplitude of the regressor function Φ (x) is large, the
effect is equivalent. Normalization is a technique that can be used to achieve more
robust adaptation for large amplitude inputs [5]. The objective of the normalization
is to reduce the effect of fast adaptation based on the amplitude of the regressor
function. Normalized adaptation can achieve a significant increase in the time-delay
margin of a closed-loop adaptive system. The normalized MRAC is given by
Γ Φ (x) e P B
Θ̇ = − (9.333)
1 + Φ (x) RΦ (x)
where c0 > 0.
Therefore,
c0 λmax (R) Φ (x)2
V̇ e, Θ̃ ≤ −λmin (Q) e2 + 2 e λmax (P) w0 +
m 2 (x)
(9.338)
Note that Φ(x)
2
m 2 (x)
≤ 1
λmin (R)
. Then, it follows that V̇ e, Θ̃ ≤ 0 if
c0 λmax (R)
− c1 e2 + 2c1 c2 e + ≤0 (9.339)
λmin (R)
or
c0 λmax (R)
e ≥ c2 + c2 + =p (9.340)
c1 λmin (R)
For the same g (Θ), 2e P B Θ̃ Φ (x) = −c0 < 0 where c0 > 0. Then,
∇gΘ (Θ) ∇ gΘ Θ̃ Θ̃
= (9.344)
∇ gΘ (Θ) ∇gΘ (Θ) Θ̃ Θ̃
Therefore,
c0 λmin (R) Φ (x)2
V̇ e, Θ̃ ≤ −λmin (Q) e2 + 2 e λmax (P) w0 −
n 2 (x)
≤ −λmin (Q) e2 + 2 e λmax (P) w0 (9.345)
Note that the normalization technique will result in poorer tracking as λmax (R)
increases, but at the same time improve robustness in the presence of large amplitude
inputs or fast adaptation.
For the normalized MRAC without the projection method, if a disturbance does
not exist, the tracking error is asymptotic, but in the presence of a disturbance, the
Lyapunov stability proof does not guarantee boundedness of the adaptive parameters.
Therefore, the normalized MRAC without the projection method is not robust to the
parameter drift.
In lieu of the projection method, the normalization technique can be used in
conjunction with any one of the robust modification schemes such as the σ modifica-
tion, e modification, and optimal control modification. For example, the normalized
σ modification is given by
Γ Φ (x) e P B + σ Θ
Θ̇ = − (9.347)
1 + Φ (x) RΦ (x)
9.8 Normalization 277
Since there is no projection, R and σ have to be chosen such that the closed-loop
system is stable. This can be determined from the Lyapunov stability analysis as
follows:
Φ (x)2
Then, evaluating V̇ e, Θ̃ and recognizing that 2 ≤ 1
λmin (R)
yield
m (x)
V̇ e, Θ̃ = −e Qe + 2e P B Θ̃ Φ (x) − 2e Pw
5 6
2Θ̃ Φ (x) e P B + σ Θ
− trace
1 + Φ (x) RΦ (x)
≤ −λmin (Q) e2 + 2 e λmax (P) w0
2
2λmax (R) e P B Θ̃ Φ (x) 2σ Θ̃ 2σ Θ̃ Θ0
+ − 2 +
λmin (R) n (x) m 2 (x)
(9.349)
Utilizing the inequality 2 a b ≤ a2 + b2 , V̇ e, Θ̃ is bounded by
2
λmax (R) P B Θ̃ Φ (x)2
V̇ e, Θ̃ ≤ −c1 (e − c2 )2 + c1 c22 +
λmin (R)
2
− c3 Θ̃ − c4 + c3 c42 (9.350)
λmin (R)
and m 4 (x)
≤ 1, then
ẋ = ax + bu (9.353)
u = k x (t) x + kr r (9.354)
γx x (xm − x) b
k̇ x = (9.355)
1 + Rx 2
γx x 2 b γx x 2 x m b
ẍ − [a + bk x (t)] ẋ + b x =b (9.356)
1 + Rx 2 1 + Rx 2
The effect of the normalized MRAC is to reduce the nonlinear integral control
action of MRAC. The effective nonlinear integral gain of the normalized MRAC is
given by
γx x 2 b
ki (x) = (9.357)
1 + Rx 2
γx b
0 < ki (x) ≤ (9.359)
R
Thus, the maximum closed-loop frequency is independent of x (t). So, any large
amplitude input will not affect the nonlinear integral control action. Nonetheless,
ki (x) can still tend to a large value with γx .
Example 9.13 Consider the following first-order SISO system:
ẋ = ax + b u + θ ∗ x 2
u = k x x + kr r − θ (t) x 2
γ x 2 eb
θ̇ = −
1 + Rx 4
ẋm = am xm + bm r
1.5 1.5
x x x x
m m
1 1
x, xm
x, xm
0.5 0.5
γ=1000 R=0 γ=1000 R=100
0 0
0 5 10 15 20 0 5 10 15 20
t t
40 6
20 4
θ
0 2
γ=1000 R=0 γ=1000 R=100
−20 0
0 5 10 15 20 0 5 10 15 20
t t
Fig. 9.21 Closed-loop responses with standard MRAC and normalized MRAC
ẋ = ax + bu + w
where
(1 + t)2 p+1 − 1
w = p (1 + t) p−1
− a (1 + t) + b γx bp
− k x (0) (1 + t) p
2p + 1
280 9 Robust Adaptive Control
u = k x (t) x
where k x (t) is computed by the normalized MRAC without the projection method
γx x 2 b
k̇ x = −
1 + Rx 2
Let a = 1, b = 1, n = − 12 5
, γx = 10, and R = 1. The closed-loop response
with the normalized MRAC is shown in Fig. 9.22. Notice that the response exhibits a
parameter drift as k x (t) → −∞. Thus, the normalized MRAC without the projection
method is not robust to the parameter drift.
1.5
1
x
0.5
0
0 10 20 30 40 50
t
−20
x
k
−40
−60
0 10 20 30 40 50
t
Fast adaptation is needed to reduce the tracking error rapidly when a large uncertainty
is present in a system. However, in most situations, when the adaptation process
has achieved sufficiently a desired level of tracking performance, fast adaptation is
usually no longer needed. Maintaining fast adaptation even when after the adaptation
has achieved its objective can result in persistent learning. At best, persistent learning
would do nothing to further improve the tracking performance once the adaptation has
achieved its objective. At worst, persistent learning reduces robustness of an adaptive
controller which is not highly desired. Therefore, the adaptation rate can be reduced
by a covariance adjustment method. The adjustment allows for an arbitrarily large
initial adaptation rate to be used to enable fast adaptation. The covariance adjustment
then adjusts the adaptation rate toward a lower value to achieve improved robustness
9.9 Covariance Adjustment of Adaptation Rate 281
as the adaptation proceeds. By reducing the adaptation rate, improved robustness can
be achieved with the covariance adjustment, while the tracking performance during
the initial adaptation is retained [5, 16, 25].
The standard MRAC with the covariance adjustment of the adaptation rate is
described by
Θ̇ = −Γ (t) Φ (x) e P B (9.360)
βΓ − ηΓ Φ (x) Φ (x) Γ
Γ˙ = (9.363)
1 + Φ (x) RΦ (x)
The projection method can be used to bound the adaptive parameter. Thus, the
MRAC with the projection method and covariance adjustment of the adaptation rate
is described by
Θ̇ = Pro Θ, −Γ (t) Φ (x) e P B (9.364)
The stability proof of the covariance adjustment of the adaptation rate is as follows:
when g (Θ) < 0 or g (Θ) = 0 and − Φ (x) e P B ∇gΘ
Consider the case
(Θ) ≤ 0. V̇ e, Θ̃ is evaluated as
! −1 "
dΓ
V̇ e, Θ̃ = −e Qe − 2e Pw + trace Θ̃ Θ̃ (9.366)
dt
But,
Γ Γ −1 = I (9.367)
282 9 Robust Adaptive Control
Then,
dΓ −1
Γ˙ Γ −1 + Γ =0 (9.368)
dt
So,
dΓ −1
= −Γ −1 Γ˙ Γ −1 = ηΦ (x) Φ (x) (9.369)
dt
Therefore,
V̇ e, Θ̃ = −e Qe − 2e Pw + ηΦ (x) Θ̃ Θ̃ Φ (x) ≤ −λmin (Q) e2
2
+ 2 e λmax (P) w0 + η Θ̃ Φ (x)2 (9.370)
Since Θ̃ ≤ β where β is the a priori bound from the projection method, then
the parameter η is chosen to ensure that V̇ e, Θ̃ ≤ 0 or equivalently the following
inequality is satisfied:
ϕ x , xm , Q, η, w0 , Θ̃ = −c1 x2 + 2 (c1 c2 + c5 xm ) x
+ 2c1 c2 xm − c1 xm 2 + ηβ 2 Φ (x)2 ≤ 0
(9.371)
Note that if a disturbance does not exist, then the tracking error converges as-
ymptotically to zero, but the adaptive parameter is bounded. In the presence of a
disturbance, a parameter drift can still exist even though the adaptation rate is con-
tinuously adjusted toward zero. In theory, the adaptive parameter would stop drift-
ing when Γ (t) → 0 as t → ∞. In practice, the adaptive parameter can reach a
sufficiently large value that can cause robustness issues. Therefore, the covariance
adjustment method needs to be used in conjunction with the projection method or
any robust modification schemes to prevent a parameter drift.
When used in conjunction with a robust modification scheme, the covariance
adjustment method can further improve robustness. For example, the optimal control
modification with the covariance adjustment is described by [25]
Θ̇ = −Γ (t) Φ (x) e P − νΦ (x) Θ B P A−1
m B (9.372)
2
V̇ e, Θ̃ ≤ −c1 (e − c2 )2 +c1 c22 −νc3 Φ (x)2 Θ̃ − c4 +νc3 c42 Φ (x)2
(9.374)
9.9 Covariance Adjustment of Adaptation Rate 283
η
where c1 , c2 , c4 are defined in Sect. 9.5, c3 = λmin B A− −1
m Q Am B 1 − ηmax ,
− −1
and 0 ≤ η < ηmaxwith ηmax = νλmin B Am Q Am B .
Thus, V̇ e, Θ̃ ≤ 0 outside a compact set. Therefore, the closed-loop adaptive
system is completely bounded.
Consider a first-order SISO system with an adaptive controller with the covariance
adjustment of the adaptation rate
ẋ = ax + bu (9.375)
u = k x (t) x + kr r (9.376)
k̇ x = γx xeb (9.377)
This yields
γx (0)
γx = :t (9.380)
1 + γx (0) η 0 x 2 (τ ) dτ
γx (0) xeb
k̇ x = :t (9.381)
1 + γx (0) η 0 x 2 (τ ) dτ
γx (0) x 2 b
ki (x) = :t (9.383)
1 + γx (0) η 0 x 2 (τ ) dτ
:t
As t → ∞, ki (x) → 0 because 0 x 2 (τ ) dτ → ∞ since r (t) is a constant
command signal for which x (t) ∈
/ L2 . Therefore, the effect of fast adaptation is
entirely eliminated.
284 9 Robust Adaptive Control
The effect of the covariance adjustment of the adaptation rate is to gradually shut
down the adaptation process. A problem associated with the covariance adjustment is
that, once the adaptation rate is reduced to a value near zero, the adaptation process has
to be restarted manually if a new uncertainty becomes present, whereas the adaptation
rate with any other adaptive control methods will be finite and the adaptation process
can be made to always remain active. It is possible to design a resetting adaptation
process using the covariance adjustment. The adaptation process can be reset with
a new initial value of the adaptation rate whenever a threshold criterion is satisfied.
For example, a covariance adjustment resetting algorithm could be expressed as
Γ˙ with Γ (te ) = Γe t ≥ te when e (t) > e0 at t = te
Γ˙ = (9.384)
Γ˙ with Γ (0) = Γ0 otherwise
The threshold should be chosen judiciously so that the trigger would occur appro-
priately to prevent false triggering. Also, when a switching action occurs with a large
change in the value of the initial condition of Γ (t), to prevent transient behaviors,
a filter can be used to phase in the new initial condition of Γ (t). For example, if a
first-order filter is used, then
Γ˙ = −λ (Γ − Γe ) (9.385)
for t ∈ [te , te + Δt], where λ > 0, Γ (te ) = Γ (t) at t = te is computed from the
previous covariance adjustment, and Γe is the new reset initial condition of Γ (t).
Example 9.15 Consider the same first-order SISO system in Example 9.13
ẋ = ax + b u + θ ∗ x 2
The MRAC adaptive law with the covariance adjustment without the projection
method is given by
θ̇ = −γ x 2 eb
γ̇ = −ηγ 2 x 4
Let γ (0) = 1000 and η = 0.1. The response of the closed-loop system is shown
in Fig. 9.23. Comparing to Fig. 9.21, the response with the covariance adjustment is
free of high-frequency oscillations. Both the tracking error and adaptive parameter
estimation error tend to zero asymptotically. The value of the adaptation rate at
t = 20 s is γ (t) = 0.6146 which is substantially less than the initial value of
γ (0) = 1000.
9.9 Covariance Adjustment of Adaptation Rate 285
2 1
1.5
0
1
e
x
−1
0.5
0 −2
0 10 20 0 10 20
t t
4
20 10
2
10 10
θ
γ
0
0 10
−2
−10 10
0 10 20 0 10 20
t t
Example 9.16 Consider the same parameter drift example in Example 9.14 with
the covariance adjustment without the projection method
k̇ x = −γx x 2 b
γ̇x = −ηγx2 x 2
2.5
1.5
x
0.5
0 100 200 300 400 500
t
0
x
−5
k
−10
0 100 200 300 400 500
t
Let γx (0) = 10 and η = 1. The response is shown in Fig. 9.24. Note that even
though the value of k x (t) appears to reach a steady-state value, in fact, k x (t) is just
drifting very slowly because the adaptation rate is very small and never reaches zero
in a finite time.
286 9 Robust Adaptive Control
A nominal fixed gain controller is designed to stabilize the plant with no uncer-
tainty and enable it to track a reference command r (t)
ū = K x x + K r r (9.388)
ẋ = Am x + Bm r (9.389)
ẋm = Am xm + Bm r (9.390)
u = ū + ΔK x x + ΔK r r − Ω x (9.391)
9.10 Optimal Control Modification for Systems with Control Input Uncertainty 287
u = ū − Θ Φ (x, r ) (9.392)
where Θ (t) = −ΔK x (t) + Ω (t) −ΔK r (t) ∈ Rm × Rn+q and Φ (x, r ) =
x r ∈ Rn+q .
It is assumed that there exist constant and unknown matrices ΔK x∗ and ΔK r∗ such
that the following model matching conditions are satisfied:
A + BΛ K x + ΔK x∗ = Am (9.393)
BΛ K r + ΔK r∗ = Bm (9.394)
Defining the tracking error as e (t) = xm (t) − x (t), then the tracking error
equation becomes
ė = Am e + B Λ̂Θ̃ Φ (x, r ) + Bε (9.396)
where
ε = −Λ̃Θ̃ Φ (x, r ) (9.397)
If the sign of Λ is known, then the standard MRAC law for adjusting Θ (t) is
given by
Θ̇ = −ΓΘ Φ (x, r ) e P BsgnΛ (9.398)
This MRAC law, while providing asymptotic tracking, does not provide robustness
to potential unmodeled dynamics. The optimal control modification adaptive law can
be used to provide robust adaptation as [21]
Θ̇ = −ΓΘ Φ (x, r ) e P − νΦ (x, r ) Θ Λ̂ B P A−1
m B Λ̂ (9.399)
It is noted that the adaptive law depends on the estimate of Λ which needs to be
computed. Toward this end, a predictor model of the plant is defined as
x̂˙ = Am x̂ + (A − Am ) x + B Λ̂ u + Ω x (9.400)
where
ε p = −Λ̃Ω̃ x (9.402)
Λ̂˙ = −ΓΛ u + Ω x ep P − ν u + x Ω Λ̂ B P A−1
m B (9.404)
2
V̇ e, e p Π̃ ≤ −c1 (e − c2 )2 + c1 c22 − c1 e p − c3
2
+ c1 c32 − νc4 Π̃ − c5 + νc4 c52 (9.410)
Thus, V̇ e, e p , Π̃ ≤ 0 outside a compact set S defined by
2
S = e (t) , e p (t) , Π̃ (t) : c1 (e − c2 )2 + c1 e p − c3
2 #
+νc4 Π̃ − c5 ≤ c1 c22 + c1 c32 + νc4 c52 (9.411)
This implies
νc4 c52
e ≥ c2 + c22 + c32 + =p (9.412)
c1
νc4 c52
e p ≥ c3 + c22 + c32 + =q (9.413)
c1
290 9 Robust Adaptive Control
c1 c22 + c1 c32
Π̃ ≥ c5 + c52 + =α (9.414)
νc4
There exists Ψ0 such that Ψ (x, r ) ≤ Ψ0 for any 0 < ν < νmax that satisfies the
following inequalities:
2
φ x p , xm , Q, ν, ε p0 , Π0 = −c1 x p + 2 (c1 c3 + λmax (Q) x) x p
+ 2c1 c3 x − c1 x2 + c1 c22
+ νc4 (Ψ (x, r )) c52 ≤ 0 (9.416)
Then, the lower bounds which are dependent on Ψ (x, r ) also exist. Since
V̇ e, e p , Π̃ ≤ 0 outside the compact set S , V e, e p , Π̃ ≤ V0 , where V0 is the
smallest upper bound of V e, e p , Π̃ which is given by
V0 = λmax (P) p 2 + q 2 + λmax ΓΘ−1 + λmax ΓΩ−1 + λmax ΓΛ−1 α 2 (9.417)
Then,
λmin (P) e2 ≤ V e, e p , Π̃ ≤ V0 (9.418)
2
λmin (P) e p ≤ V e, e p , Π̃ ≤ V0 (9.419)
Consider the following MIMO system with a control input uncertainty and a matched
uncertainty:
ẋ = Ax + BΛ u + Θ ∗ Φ (x) + w (9.422)
ẋm = Am xm + Bm r (9.423)
A + BΛK x∗ = Am (9.425)
BΛK r∗ = Bm (9.426)
If Λ is unknown but sign of Λ is known, then the standard MRAC adaptive laws
are given by
K̇ r = Γr r e P BsgnΛ (9.428)
where w̃ (t) = ŵ (t) − w (t) is the disturbance estimation error, and ε p (t) ∈ Rm is
the residual estimation error of the predictor model
9.11 Bi-Objective Optimal Control Modification for Systems … 293
Θ̇ = −ΓΘ Φ (x) e P + νu Λ̂ B P A−1
m + ep W
− η u + 2Θ Φ (x) Λ̂ B + ŵ W A−1
m B Λ̂ (9.441)
! ; # "
Λ̂˙ = −ΓΛ u + Θ Φ (x) e W − η u + 2Θ Φ (x) Λ̂ B + ŵ W A−1 B
p m
(9.442)
ŵ˙ = −γw ep W − η u + 2Θ Φ (x) Λ̂ B + ŵ W A−1
m (9.443)
P Am + A
m P = −Q (9.444)
W Am + A
m W = −R (9.445)
subject to Eqs. (9.434) and (9.437), where Δ1 (t) and Δ2 (t) represent the unknown
lower bounds of the tracking error and the predictor error, respectively.
The cost functions J1 and J2 are combined into the following bi-objective cost
function:
J = J1 + J2 (9.448)
Proof Using the optimal control framework, the Hamiltonian of the cost function is
defined as
1 1
H= (e − Δ1 ) Q (e − Δ1 ) + e p − Δ2 R e p − Δ2
2 2
+ λ Am e + B Λ̂ − K̃ x x − K̃ r r + Θ̃ Φ (x) − w + Bε
+ μ Am e p + B Λ̃ u + Θ Φ (x) + B Λ̂Θ̃ Φ (x) + ŵ − w + Bε p
(9.449)
λ̇ = −∇ He = −Q (e − Δ1 ) − A
mλ (9.450)
μ̇ = −∇ Hep = −R e p − Δ2 − A
mμ (9.451)
9.11 Bi-Objective Optimal Control Modification for Systems … 295
subject to the transversality conditions λ t f → ∞ = 0 and μ t f → ∞ = 0 since
both e (0) and e p (0) are given.
Treating K̃ x (t), K̃ r (t), Θ̃ (t), Λ̃ (t), and ŵ (t) as control variables, then the opti-
mal control solutions are obtained by the following gradient-based adaptive laws:
Θ̇ = Θ̃˙ = −ΓΘ ∇ HΘ̃ = −ΓΘ Φ (x) λ + μ B Λ̂ (9.454)
Λ̂˙ = Λ̃˙ = −ΓΛ ∇ HΛ̃ = −ΓΛ u + Θ Φ (x) μ B (9.455)
Substituting the adjoint solutions back into the adjoint equations yields
Ṗe + P Am e + P B Λ̂ −K x x − K r r + Θ Φ (x)
− P B Λ̂ −K x∗ x − K r∗r + Θ ∗ Φ (x) − Pw + P Bε
+ Ṡ Λ̂ −K x x − K r r + Θ Φ (x)
d Λ̂ −K x x − K r r + Θ Φ (x)
+S
dt
= −Q (e − Δ1 ) − A
m Pe − Am S Λ̂ −K x x − K r r + Θ Φ (x)
(9.459)
Ẇ e p + W Am e p + W B Λ̂ u + Θ Φ (x) − W BΛ u + Θ Φ (x) + W B Λ̂Θ Φ (x)
− W B Λ̂Θ ∗ Φ (x) + W ŵ − W w + W Bε p
d Λ̂ u + 2Θ Φ (x)
+ Ṫ Λ̂ u + 2Θ Φ (x) + T + V̇ ŵ
dt
+ V ŵ˙ = −R e p − Δ2 − A
m W e p − Am T Λ̂ u + 2Θ Φ (x) − Am V ŵ
(9.460)
296 9 Robust Adaptive Control
Ṗ + P Am + A
mP + Q = 0 (9.461)
Ṡ + A
mS + PB = 0 (9.462)
QΔ1 + P B Λ̂ −K x∗ x − K r∗r + Θ ∗ Φ (x) + Pw
d Λ̂ −K x x − K r r + Θ Φ (x)
− P Bε − S =0 (9.463)
dt
Ẇ + W Am + A
mW + R = 0 (9.464)
Ṫ + A
mT + W B = 0 (9.465)
V̇ + A
mV + W = 0 (9.466)
RΔ2 + W BΛ u + Θ Φ (x) + W B Λ̂Θ ∗ Φ (x) + W w − W Bε p
d Λ̂ u + 2Θ Φ (x)
−T − V ŵ˙ = 0 (9.467)
dt
subject
to the
transversality
conditions
P t f → ∞ = 0, S t f → ∞ = 0,
W t f → ∞ = 0, T t f → ∞ = 0, and V t f → ∞ = 0.
The infinite-time horizon solutions of P (t) and W (t) as t f → ∞ tend to their
equilibrium solutions at t = 0 and are given by the algebraic Lyapunov equations
P Am + A
mP + Q = 0 (9.468)
W Am + A
mW + R = 0 (9.469)
The solutions of S (t), T (t), and V (t) also tend to their equilibrium solutions
A
mS + PB = 0 (9.470)
A
mT + W B = 0 (9.471)
A
mV + W = 0 (9.472)
As with any control design, performance and robustness are often considered as
two competing design requirements. Increasing robustness tends to require a com-
promise in performance and vice versa. Thus, to enable the bi-objective optimal
control modification adaptive laws to be sufficiently flexible for a control design, the
9.11 Bi-Objective Optimal Control Modification for Systems … 297
modification parameters ν > 0 and η > 0 are introduced as design free parameters
to allow for the adjustment of the bi-objective optimal control modification terms in
the adaptive laws.
Thus, the solutions of S (t), T (t), and V (t) are given by
S = −ν A−
m PB (9.473)
T = −η A−
m WB (9.474)
V = −η A−
m W (9.475)
Using the expression of u (t), the adjoint solutions are then obtained as
λ = Pe + ν A−
m P B Λ̂u (9.476)
μ = W e p − η A− −
m W B Λ̂ u + 2Θ Φ (x) − η Am W ŵ (9.477)
Substituting the adjoint solutions into the gradient-based adaptive laws yields the
bi-objective optimal control modification adaptive laws in Eqs. (9.452)–(9.456).
We note that K x (t) and K r (t) are adapted based on the tracking error, Λ̂ (t) and
ŵ (t) are adapted based on the predictor error, and Θ (t) is adapted based on both
the tracking error and the predictor error.
The bounds on Δ1 (t) and Δ2 (t) as t f → ∞ can be estimated by
1
Δ1 ≤ P B Λ̂ −K x∗ x − K r∗ r + Θ ∗ Φ (x) + λmax (P) w0 + P B ε0
λmin (Q)
⎤
d Λ̂ −K x x − K r r + Θ Φ (x)
+ ν A−
m P B
⎦
(9.478)
dt
1
Δ2 ≤ W BΛ u + Θ Φ (x) + W B Λ̂ Θ ∗ Φ (x) + λmax (W ) w0
λmin (R)
d Λ̂ u + 2Θ Φ (x)
+η − W ŵ
˙
+ W B ε p0 + η A−
m W B A m
dt
(9.479)
which are dependent upon the modification parameters, control effectiveness uncer-
tainty, matched uncertainty, and residual tracking error and predictor error.
298 9 Robust Adaptive Control
− 2e Pw + 2e P Bε
− e
p Re p + 2e p W B Λ̃ u + Θ Φ (x) + Λ̂Θ̃ Φ (x)
+ 2e
p W w̃ + 2e p W Bε p
+ 2trace K̃ x x e P + νu Λ̂ B P A−1
m B Λ̂
+ 2trace K̃ r r e P + νu Λ̂ B P A−1
m B Λ̂
− 2trace Θ̃ Φ (x) e P + νu Λ̂ B P A−1m + ep W
; "
− η u + 2Θ Φ (x) Λ̂ B + ŵ W A−1 m B Λ̂
− 2trace Λ̃ u + Θ Φ (x)
! ; # " "
e
pW −η u + 2Θ Φ (x) Λ̂ B + ŵ W A−1 m B
! ; # "
− 2ep W w̃ + 2 η u + 2Θ Φ (x) Λ̂ B + ŵ W A−1 m w̃
− ν ũ Λ̂ B A− −1 ∗ −1
m Q Am B Λ̂ũ + 2νu Λ̂ B P Am B Λ̂ũ
− ηΨ (x, r ) Ω̃ B̄ A− −1
m R Am B̄ Ω̃ Ψ (x, r )
+ 2ηΨ (x, r ) Ω ∗ B̄ W A−1
m B̄ Ω̃ Ψ (x, r ) (9.488)
⎤ ⎡
x
Let K (t) = K x (t) K r (t) −Θ (t) ∈ Rm × Rn+q+ p and z (x, r ) = ⎣ r ⎦
Φ (x)
∈Rn+q+ p
. Then, V̇ e, e p , Ω̃, ũ is bounded by
V̇ e, e p , K̃ , Ω̃ ≤ −λmin (Q) e2 + 2 e λmax (P) w0 + 2 e P B ε0
2
− λmin (R) e p + 2 e p W B ε p0
2 2 2
+ 2γw−1 Ω̃ δ0 − νλmin B A− −1
m Q Am B z (x, r ) Λ̂ K̃
2
+ 2ν z (x, r )2 B P A−1
m B Λ̂ K̃ K 0
− ηλmin B̄ A− −1 2 2
m R Am B̄ Ψ (x, r ) Ω̃
2
+ 2η B̄ W A−1
m B̄ Ψ (x, r ) Ω̃ Ω0 (9.489)
− m B̄ Ω0
B̄ W A−1 γ −1 δ
B̄ Am R A−1m B̄ Ψ (x, r ) , and c8 = λmin ( B̄ A−
2
+ w 0 2 . Then,
m R Am B̄ )
−1
ηc7 B̄
2
V̇ e, e p , K̃ , Ω̃ ≤ −c1 (e − c2 )2 + c1 c22 − c3 e p − c4 + c3 c42
2 2 2
− νc5 Λ̂ K̃ − c6 + νc5 c62 Λ̂
2
− ηc7 Ω̃ − c8 + ηc7 c82 (9.490)
Note that Λ̃ ≤ Ω̃ . Then,
2 2 2
Λ̂ = Λ + Λ̃ ≤ Λ + Ω̃ (9.491)
Let Ω̃ = c8 . Then, V̇ e, e p , K̃ , Ω̃ is bounded by
9.11 Bi-Objective Optimal Control Modification for Systems … 301
2
V̇ e, e p , K̃ , Ω̃ ≤ −c1 (e − c2 )2 + c1 c22 − c3 e p − c4
2
+ c3 c42 − νc5 Λ20 K̃ − c6 + νc5 c62 Λ20 + ηc7 c82 (9.492)
;
S = e (t) , e p (t) , K̃ (t) , Ω̃ (t) : c1 (e − c2 )2
2 2
+ c3 e p − c4 + νc5 Λ20 K̃ − c6
2 #
+ ηc7 Ω̃ − c8 ≤ c1 c22 + c3 c42 + νc5 c62 Λ20 + ηc7 c82 (9.493)
This implies
c3 c42 + νc5 c62 Λ20 + ηc7 c82
e ≥ c2 + c22 + =p (9.494)
c1
c1 c22 + νc5 c62 Λ20 + ηc7 c82
e p ≥ c4 + c42 + =q (9.495)
c3
c1 c22 + c3 c42 + ηc7 c82
K̃ ≥ c6 + c62 + =α (9.496)
νc5 Λ20
c1 c22 + c3 c42 + νc5 c62 Λ20
Ω̃ ≥ c8 + c82 + =β (9.497)
ηc7
There exist z 0 and Ψ0 such that z (x, r ) ≤ z 0 and Ψ (x, r ) ≤ Ψ0 for any
0 < ν < νmax and 0 < η < ηmax that satisfy the following inequalities:
2
φ x p , xm , R, η, ẇ0 , ε p0 , B0 , Ω0 = −c3 x p + 2 (c3 c4 + λmax (R) x) x p
+ 2c3 c4 x − c3 x2
+ c1 c22 + νc5 (z (x, r )) c62 Λ20
+ ηc7 (Ψ (x, r )) c82 ≤ 0 (9.499)
Then, the lower bounds which are dependent on z (x, r ) and Ψ (x, r )
also exist. Since V̇ e, e p , K̃ , Ω̃ ≤ 0 outside the compact set S , limt→∞ V
e, e p , K̃ , Ω̃ ≤ V0 , where V0 is the smallest upper bound of V e, e p , K̃ , Ω̃
which is given by
V0 = λmax (P) p 2 + λmax (W ) q 2 + λmax Γx−1 α 2
+ λmax Γr−1 α 2 + λmax ΓΘ−1 α 2 + β 2
+ λmax ΓΛ−1 β 2 + γw−1 β 2 (9.500)
Then,
λmin (P) e2 ≤ V e, e p , K̃ , Ω̃ ≤ V0 (9.501)
2
λmin (W ) e p ≤ V e, e p , K̃ , Ω̃ ≤ V0 (9.502)
ẋm = am xm + bm r
using the bi-objective optimal control modification method with the following pre-
dictor model:
x̂˙ = am x̂ + (a − am ) x + bλ̂ u (t − td ) + θ x 2 + ŵ
λ̂˙ = −γλ u + θ x 2 e p − η u + 2θ x 2 λ̂b + ŵ am−1 b
ŵ˙ = −γw e p − η u + 2θ x 2 λ̂b + ŵ am−1
x, xm
0
−1
x x
m
−2
0 10 20 30 40 50 60
t
0
<
x, x
−2
x x^
−4
0 10 20 30 40 50 60
t
10
kx
−10
0 10 20 30 40 50 60
t
5
kr
−5
0 10 20 30 40 50 60
t
0
θ
−5
0 10 20 30 40 50 60
t
The adaptive parameters k x (t), kr (t), and θ (t) are shown in Fig. 9.26. These
adaptive parameters show a parameter convergence to their corresponding ideal val-
ues. The parameter convergence is also facilitated by having a persistently exciting
reference command signal r (t) = sin t.
The estimates of the control input effectiveness λ̂ (t) and disturbance ŵ (t) are
shown in Fig. 9.27. It can be seen that λ̂ (t) converges nicely to the ideal value of
−1. Because of the parameter convergence of λ̂ (t), the bi-objective optimal control
modification adaptive laws remain stable in light of the sign reversal of the control
input effectiveness. Without accurate estimation of λ̂ (t), it will be a challenge for
any adaptive control method to be able to maintain stability. It is also noted that this
parameter convergence is achieved even with a time delay of 0.1 s. The bi-objective
optimal control modification remains stable up to a time delay of 0.108 s.
For comparison, we redesign the adaptive controller with the optimal control
modification with the predictor model according to Sect. 9.10. The optimal control
modification adaptive laws are
9.11 Bi-Objective Optimal Control Modification for Systems … 305
0
<
λ
−5
−10
0 10 20 30 40 50 60
t
0
<
w, w
−2
w ^
w
−4
0 10 20 30 40 50 60
t
Fig. 9.27 Estimation of control input effectiveness and disturbance with bi-objective optimal con-
trol modification
k̇ x = γx x e + νxk x λ̂bam−1 bλ̂
k̇r = γr r e + νr kr λ̂bam−1 bλ̂
θ̇ = −γθ x 2 e p − ηx 2 θ λ̂bam−1 bλ̂
λ̂˙ = −γλ u + θ x 2 e p − η u + θ x 2 λ̂bam−1 b
ŵ˙ = −γw e p − ηŵam−1
Keeping the same initial conditions, it can be seen that the optimal control modi-
fication performs worse than the bi-objective optimal control modification as shown
in Figs. 9.28, 9.29 and 9.30. Even though the predictor model does an excellent job
of estimating x (t), the tracking performance is much worse than that with the bi-
objective optimal control modification. The adaptive parameters also suffers from
a poor parameter convergence with the optimal control modification as shown in
Figs. 9.29 and 9.30.
By setting ν = 0 and η = 0, one can recover the standard MRAC adaptive
laws with the predictor model. Comparing robustness of the three controllers, the
optimal control modification has the largest time-delay margin of 0.116 s, whereas
the standard MRAC has the smallest time-delay margin of 0.082 s.
306 9 Robust Adaptive Control
x, xm
0
−1
−2
0 10 20 30 40 50 60
t
1
<
x, x
−1
x x^
−2
0 10 20 30 40 50 60
t
10
x
0
k
−10
0 10 20 30 40 50 60
t
10
r
0
k
−10
0 10 20 30 40 50 60
t
0
θ
−1
0 10 20 30 40 50 60
t
< −1
λ
−2
−3
0 10 20 30 40 50 60
t
0.5
0
<
w, w
−0.5
−1
w ^
w
−1.5
0 10 20 30 40 50 60
t
Fig. 9.30 Estimation of control input effectiveness and disturbance with optimal control modifi-
cation
9.12 Adaptive Control for Singularly Perturbed Systems … 307
When there exists a timescale separation between a fast plant and a slow actuator
which prevents the plant from following a reference model, a singular perturbation
approach can be used to separate the time scale between the plant and actuator
and then modify the adaptive law to account for the slow actuator in the singularly
perturbed system [32]. The singular perturbation approach transforms the original
system into a reduced-order system in slow time. A model matching condition is
applied to the reduced-order system and the reference model in the transformed slow
time coordinate that results in changes in the actuator command to accommodate
the slow actuator dynamics. The resulting control signal can then track the reference
model better than if the actuator command is not modified.
Consider a linear plant as
ẋ = Ax + B u + Θ ∗ x (9.505)
u̇ = εG (u − u c ) (9.506)
u c = K x∗ x + K r∗r − u ad (9.508)
The control gain matrices K x∗ and K r∗ can be chosen to satisfy the model matching
conditions A + B K x∗ = Am and B K r∗ = Bm , and the adaptive signal u ad (t) is chosen
to be
u ad = Θ Φ (x) (9.509)
τ = εt (9.511)
u̇ i = εG (u i − u c ) = 0 (9.517)
Differentiating Eq. (9.519) with respect to the slow time variable and then substi-
tuting the actuator model into the result yield
d x0 −1
= − A + BΘ ∗ BG (u 0 − u c ) (9.520)
dτ
From Eq. (9.514), we have
u 0 = − B̄ −1 Ax0 − Θ ∗ x0 (9.521)
−1
where B̄ −1 = B B B is the right pseudo-inverse of B.
Hence, we obtain the following reduced-order plant model constrained by the
slow actuator dynamics:
d x0 −1
= A + BΘ ∗ BG B̄ −1 Ax0 + Θ ∗ x0 + u c (9.522)
dτ
Using the matrix inversion lemma, we obtain
−1 −1 ∗ −1
A + BΘ ∗ = A−1 − A−1 B I + Θ ∗ A−1 B Θ A (9.523)
−1 ∗ −1
Let Ψ ∗ = A−1 B I + Θ ∗ A−1 B Θ A . Then,
d x0
= A−1 BG B̄ −1 Ax0 + −Ψ ∗ BG B̄ −1 A + A−1 − Ψ ∗ BGΘ ∗ x0
dτ
+ A−1 − Ψ ∗ BGu c (9.524)
dx
= As x + Bs Λ u c + Θs∗ x (9.525)
dτ
where As = A−1 BG B̄ −1 A, Bs = A−1 BG, Bs Λ = A−1 − Ψ ∗ BG, and
Bs ΛΘs∗ = −Ψ ∗ BG B̄ −1 A+ A−1 − Ψ ∗ BGΘ ∗ .
If As is Hurwitz and Θs∗ = 0, then the Tikhonov’s theorem guarantees that the
reduced solution with ε > 0 converges to the solution of the original system as
ε → 0 [33].
Because of the slow actuator dynamics, the time scale of the response of the
plant cannot exceed that of the actuator. Thus, if the reference model is faster than
the actuator model, the tracking error cannot be guaranteed to be small even with
adaptive control due to the model mismatch. A possible solution is to revise the
reference model to match the actuator-constrained plant model, or alternatively to
redesign the actuator command to reduce the tracking error.
In slow time, the reference model is expressed as
d xm 1
= (Am xm + Bm r ) (9.526)
dτ ε
We make the following choice for the actuator command signal:
u c = K x x + K r r − u ad (9.527)
where K x and K r are nominal control gains for the normal plant without the slow
actuator dynamics and u ad (t) is of the form
1
Am = As + Bs Λ K x + ΔK x∗ (9.529)
ε
1
Bm = Bs Λ K r + ΔK r∗ (9.530)
ε
Then, the closed-loop system becomes
dx 1
= (Am xm + Bm r ) − Bs ΛΩ̃ Φ (x, r ) (9.531)
dτ ε
de d xm dx 1
= − = Am e + ε Bs ΛΩ̃ Φ (x, r ) (9.532)
dτ dτ dτ ε
Since Am is Hurwitz and if Ω̃ (t) is bounded, then the Tikhonov’s theorem guar-
antees that the reduced solution with ε > 0 converges to the solution of the original
system as ε → 0.
Transforming into the real time, the tracking error equation becomes
However, this assumption is not always guaranteed since Λ can be of mixed sign.
The bi-objective optimal control modification adaptive law can be used to estimate
Λ using the following predictor model:
x̂˙ = Am x̂ + (ε As − Am ) x + ε Bs Λ̂ u c + Θs x (9.535)
Then, the bi-objective optimal control modification adaptive laws are expressed
as
Ω̇ = −εΓΩ Φ (x, r ) e P + ενu Λ̂ B P A−1
m + ep W
−εη u + 2Θs Φ (x) Λ̂ Bs W A−1 m Bs Λ̂ (9.536)
Θ̇s = −εΓΘs x ep W − εη u + 2Θs Φ (x) Λ̂ Bs W A−1
m Bs Λ̂ (9.537)
Λ̂˙ = −εΓΛ u + Θs Φ (x) ep W − εη u + 2Θs Φ (x) Λ̂ Bs W A−1
m Bs
(9.538)
where e p (t) = x̂ (t) − x (t).
ẋ = ax + bu + θ ∗ x + w (t)
u̇ = εg (u − u c )
where a < 0, g < 0, ε > 0, |εg| < |a|, and w (t) is a disturbance signal.
The reference model is
312 9 Robust Adaptive Control
ẋm = am xm + bm r
where am < 0.
The actuator command for the slow actuator dynamics is designed as
u c = k x x + kr r − Ω Φ (x, r )
where k x = amb−a , kr = bbm , and Φ (x, r ) = x r .
Note that, if the actuator dynamics are fast, then the actuator command is
u c = k x x + kr r − θ x
The optimal control modification adaptive law for the slow actuator dynamics is
Ω̇ = −εΓ Φ (x, r ) esgnλ − ενΦ (x, r ) Ωbs am−1 bs
where bs = bg
a
and λ = a
a+bθ ∗
, and for the fast actuator dynamics is
θ̇ = −Γ x e − νxθ bam−1 b
If a and g are nominally of the same order of magnitude, then we note that, for
the slow actuator dynamics, the effective adaptive gain is reduced by ε for a similar
performance as that for the fast actuator dynamics.
For the numerical example, we let a = −1, b = 1, θ ∗ = 0.1, g = −1, ε = 0.1,
am = −5, bm = 1, r (t) = sin t, and w (t) = 0.05 sin 10t. The responses due to the
standard MRAC adaptive law and optimal control modification adaptive law with the
singular perturbation approach are shown in Fig. 9.31. The response for the standard
MRAC exhibits more initial transients than that for the optimal control modification
using the same adaptive gain.
Figure 9.32 is the plot of the control input and actuator command with the singular
perturbation approach. As can be seen, the actuator command signal is quite large
relative to the control input. This is due to the fact that the actuator dynamics are
slow so a large actuator command does not translate into the same control input in
a finite time. The effectiveness of the optimal control modification is demonstrated
by the reduced amplitude of oscillations in the control input significantly over that
due to the standard MRAC.
Figure 9.33 shows the responses due to the unmodified actuator command for the
fast actuator dynamics. As can be seen, the control input is insufficient to allow the
plant to follow the reference model even with adaptive control.
9.12 Adaptive Control for Singularly Perturbed Systems … 313
1
MRAC Γ=1000
0.5
x, xm
0
−0.5
x xm
−1
0 10 20 30 40 50 60
t
1
Optimal Control Modification Γ=1000 ν=0.5
0.5
m
x, x
−0.5
x xm
−1
0 10 20 30 40 50 60
t
Fig. 9.31 Responses for slow actuator dynamics with MRAC and optimal control modification
40
MRAC Γ=1000
20
u, uc
−20
u u
m
−40
0 10 20 30 40 50 60
40
Optimal Control Modification Γ=1000 ν=0.5
20
u, uc
−20
u u
m
−40
0 10 20 30 40 50 60
Fig. 9.32 Control and actuator command for slow actuator dynamics with MRAC and optimal
control modification
0.4
MRAC Γ=1000
0.2
x, xm
−0.2
x x
m
−0.4
0 10 20 30 40 50 60
t
0.4
Optimal Control Modification Γ=1000 ν=0.5
0.2
x, xm
−0.2
x x
m
−0.4
0 10 20 30 40 50 60
t
Fig. 9.33 Responses for slow actuator dynamics due to unmodified actuator command
314 9 Robust Adaptive Control
ż = f (x, z, u) (9.540)
y = Cx (9.541)
where z (t) is an internal state vector, Δ (x, z, u) is the plant model error that is
unknown and not accounted for, ż (t) is the unmodeled dynamics, and y (t) is the
plant output vector.
When the tracking error based on only the output signal y (t) is used for MRAC,
such a class of adaptive control is called output feedback adaptive control. If a
stable first-order reference model ym (t) is specified for the output y (t) to track, then
model-reference adaptive control is feasible if the plant transfer function satisfies a
certain property called strictly positive real (SPR) condition. Therefore, an essential
requirement for output feedback adaptive control is the SPR condition that the transfer
function must satisfy. In Chap. 8, we see that the standard MRAC is unstable for non-
minimum phase plants (see Example 8.3). Non-minimum phase systems do not have
SPR transfer functions. Therefore, output feedback adaptive control of systems with
non-SPR transfer functions can be challenging.
The strictly positive real condition is defined as follows:
Definition 9.1 Consider a proper SISO transfer function
Z (s)
G (s) = (9.542)
R (s)
where Z (s) and R (s) are polynomials of degrees m and n, respectively, with n > m.
The relative degree is defined as n − m. If the relative degree of G (s) is not greater
than 1 and the real part of G (s) is at least positive for all σ ≥ 0 where s = σ + jω,
that is, (G (s)) ≥ 0 for σ ≥ 0, then the transfer function is said to be positive real
(PR). Furthermore, if G (s − ε) is positive real for some ε > 0, then G (s) is strictly
positive real (SPR) [30].
Example 9.19 • Consider the transfer function in Example 8.2
s+1
G (s) =
s2 + 5s + 6
We see that ((G (s))) > 0 for all σ ≥ 0 and ω ≥ 0. If we let σ = −ε and
ω = 0, then (G (s)) ≥ 0 for 0 < ε ≤ 1. So, G (s) is SPR and is a minimum
phase, stable transfer function with relative degree 1.
• Consider the transfer function
3
s σ + 5σ 2 + σ ω2 + 6σ + 5ω2 + jω −σ 2 − ω2 + 6
G (s) = 2 = 2
s + 5s + 6 σ 2 + 5σ − ω2 + 6 + ω2 (2σ + 5)2
We see that (G (s)) ≥ 0 for all σ ≥ 0 and ω ≥ 0. So, G (s) is only PR but not
SPR. Note that G (s) is a stable transfer function with a zero on the jω axis.
• The transfer function in Example 8.2
s−1
G (s) =
+ 5s + 6
s2
3
σ + 4σ 2 + σ ω2 + σ + 6ω2 − 6 + jω −σ 2 + 2σ − ω2 + 11
= 2
σ 2 + 5σ − ω2 + 6 + ω2 (2σ + 5)2
is not SPR since (G (s)) can be negative for some values of σ ≥ 0 and ω ≥ 0.
Note that G (s) is non-minimum phase.
• The transfer function
s+1
G (s) =
s 2 − 5s + 6
3
σ − 4σ 2 + σ ω2 + σ − 6ω2 + 6 + jω −σ 2 − 2σ − ω2 + 11
= 2
σ 2 − 5σ − ω2 + 6 + ω2 (2σ − 5)2
is not SPR since (G (s)) can be negative for some values of σ ≥ 0 and ω ≥ 0.
Note that G (s) is unstable.
• The transfer function
1
G (s) =
s 2 + 5s + 6
2
σ + 5σ − ω2 + 6 − jω (2σ + 5)
= 2
σ 2 + 5σ − ω2 + 6 + ω2 (2σ + 5)2
is not SPR since (G (s)) can be negative for some values of σ ≥ 0 and ω ≥ 0.
Note that G (s) has relative degree 2.
y (s) bωn2
= (9.543)
u (s) (s − a) s 2 + 2ζ ωn s + ωn2
where a < 0 is unknown, b is unknown, but b > 0 is known, and ζ and ωn are
known.
The system has relative degree 3 and therefore is not SPR.
The transfer function of the reference model is specified as
ym (s) bm
= (9.544)
r (s) s − am
The reference model is SPR with relative degree 1. Since the relative degree of
the reference model is less than the relative degree of the plant, perfect tracking is not
possible. Adaptive control of systems with relative degrees greater than 1 is generally
much more difficult to handle since the model reference cannot be chosen to be SPR
[30].
The controller is given by
k̇ y = γx ye (9.546)
k̇r = γr r e (9.547)
r = r0 sin ω0 t (9.548)
√
where r0 = 1 and ω0 = 30 rad/s is the frequency at which the closed-loop transfer
function has zero phase margin.
The closed-loop system is unstable as shown in Fig. 9.34.
The underlying cause of instability is the lack of robustness of the closed-loop
system. Changing either the initial condition of k y (t) and or the frequency in the
reference command signal can result in stabilization of the closed-loop system if it
has a sufficient phase margin.
9.13 Optimal Control Modification for Linear Uncertain Systems … 317
The optimal control modification adaptive law can be designed to handle linear
systems with unmodeled dynamics by utilizing the linear asymptotic property.
Suppose the open-loop plant is expressed in general by the following transfer
function:
y (s) Z p (s)
W p (s) = k p (9.549)
u (s) R p (s)
where k p is a high-frequency gain, and Z p (s) and R p (s) are monic Hurwitz poly-
nomials of degrees m p and n p , respectively, and n p − m p > 0 is the relative degree
of the plant.
The reference model is given by a transfer function
ym (s) Z m (s)
Wm (s) = km (9.550)
r (s) Rm (s)
where km is a high-frequency gain, and Z m (s) and Rm (s) are monic Hurwitz poly-
nomials of degrees m m and n m , respectively, and n m − m m > 0 is the relative degree
of the reference model.
Let n p − m p > n m − m m . So the SPR condition is no longer possible to ensure
tracking of the reference model. Stability of the closed-loop system cannot also be
guaranteed with the standard MRAC.
1 5
0.5
0
0
u
x
−5
−0.5
−1 −10
0 5 10 0 5 10
t t
−0.5 4
−1
3
y
−1.5
k
k
2
−2
−2.5 1
0 5 10 0 5 10
t t
Suppose the adaptive controller is redesigned with the optimal control modifica-
tion as
u = k y y + kr r (9.551)
where
k̇ y = γ y ye − νy 2 k y (9.552)
k̇r = γr r e − νr 2 kr (9.553)
318 9 Robust Adaptive Control
Using the linear asymptotic property of the optimal control modification, the
asymptotic value of the adaptive controller u can be computed as γ y → ∞ and
γr → ∞ in the limit. Then,
2ym − 2y
ū = (9.554)
ν
The asymptotic closed-loop transfer function can now be computed as
2ym − 2y 2bm r 2y
ū = = − (9.556)
ν ν (s − am ) ν
as γ y → ∞ and γr → ∞.
Then, the asymptotic closed-loop transfer function is obtained as
ȳ 2bωn2 bm
G (s) = (9.557)
r 2bωn2
ν (s − am ) (s − a) s 2 + 2ζ ωn s + ωn2 + ν
Note that the closed-loop transfer function has a relative degree 4 while the transfer
function of the reference model has a relative degree 1. This prevents the output y (t)
from tracking ym (t).
The characteristic equation of the transfer function G (s) with an input time delay
is computed as
2bωn2 −td s
s 3 + (2ζ ωn − a) s 2 + ωn2 − 2aζ ωn s − aωn2 + e =0 (9.558)
ν
Substituting s = jω yields
2bωn2
− jω3 −(2ζ ωn − a) ω2 + ωn2 − 2aζ ωn jω−aωn2 + (cos ωtd − j sin ωtd ) = 0
ν
(9.559)
This results in two frequency equations
9.13 Optimal Control Modification for Linear Uncertain Systems … 319
2bωn2
− (2ζ ωn − a) ω2 − aωn2 + cos ωtd = 0 (9.560)
ν
2bωn2
− ω3 + ωn2 − 2aζ ωn ω − sin ωtd = 0 (9.561)
ν
We then obtain the cross-over frequency and phase margin as functions of the
modification parameter ν in the following expressions:
! "
2 4 2 2 2 4b2
ω + a + 4ζ ωn − 2ωn ω + ωn + 4a ζ − 2a ωn ω + a − 2 ωn4 = 0
6 2 2 2 2 2 2 2
ν
2 (9.562)
−ω 3
+ ω − 2aζ ωn ω
φ = ωtd = tan−1 n
(9.563)
(2ζ ωn − a) ω2 + aωn2
<
Note that the cross-over frequency at zero phase margin ω0 = ωn2 − 2aζ ωn is
still the same as that without modification (see Sect. 8.4). However, this cross-over
frequency corresponds to a modification parameter νmin . By choosing a modification
ν > νmin that meets a certain phase margin requirement, the closed-loop adaptive
system can be made robustly stable.
Example 9.20 Consider Example 8.5 with the following parameters: a = −1, b =
1, am = −2, bm = 2, ωn = 5 rad/s,√ and ζ = 0.5. The cross-over frequency at zero
phase margin is computed to ω0 = 30 rad/s. The cross-over frequency and phase
margin as a function of the modification parameter ν are shown in Fig. 9.35.
The modification parameter ν corresponding to zero phase margin and 90o phase
margin are νmin = ν0 = 0.3226 and νmax = ν90o = 0.9482, respectively. So, the
adaptive systems with the unmodeled dynamics can be stabilized with the selection of
the modification parameter ν between these two values. The< cross-over frequencies
√
margin and 90o phase margin are ω0 = ωn2 − 2aζ ωn = 30
at the zero phase 7
7
aω2
rad/s and ω90o = − 2ζ ωn n−a = 25 6
rad/s, respectively. So, robustness is achieved
with the optimal control modification by reducing the cross-over frequency of the
closed-loop system.
Suppose a 45o phase margin is selected. Then, the cross-over frequency can be
computed from Eq. (9.563) as the root of the following equation:
ω3 + (2ζ ωn − a) ω2 + 2aζ ωn − ωn2 ω + aωn2 = 0
which yields ω45o = 3.7568 rad/s. Then, the modification parameter ν is computed
from Eq. (9.562) to be ν45o = 0.5924.
The asymptotic closed-loop transfer function with ν45o = 0.5924 is computed to
be
100
ν45o
G (s) =
s 4 + 8s 3 + 42s 2 + 85 + 50
ν45o
s + 50 + 100
ν
320 9 Robust Adaptive Control
ω, rad/sec 15
10
0
0 0.2 0.4 0.6 0.8 1
ν
100
50
φ, deg
−50
−100
0 0.2 0.4 0.6 0.8 1
ν
For the simulations, the time step is chosen to be 1 × 10−6 sec to permit the
adaptation rates γx and γr to be set at 1√× 106 to simulate fast adaptation. The
reference command signal is r (t) = sin 30t. The asymptotic response ȳ (t) and
control signal ū (t) evaluated analytically agree very well with the simulation results
of y (t) and u (t), respectively, as shown in Fig. 9.36. The control signal u (t) exhibits
sharp spikes due to the adaptive gain kr (t) but otherwise tracks the asymptotic control
signal ū (t) very well. Thus, the linear asymptotic property of the optimal control
modification is demonstrated to be able to facilitate the stability analysis of linear
uncertain systems with unmodeled dynamics, time delay, or non-minimum phase
behaviors. In this study, the optimal control modification can handle systems with
relative degrees greater than 1 with relative ease.
1 _ 4 _
y y u u
0.5 2
|
u, u
y, y
|
0 0
−0.5 −2
−1 −4
0 5 10 0 5 10
t t
100 100
50 50
y
kr
k
0 0
−50 −50
0 5 10 0 5 10
t t
Fig. 9.36 Closed-loop response with optimal control modification (ν = 0.5924, γx = γr = 106 )
9.13 Optimal Control Modification for Linear Uncertain Systems … 321
Example 9.21 Consider the Rohrs’ counterexample problem in Example 8.6 [1].
The cross-over frequency and phase margin as a function of the modification para-
meter ν are shown in Fig. 9.37. The cross-over frequency 7 at the zero phase margin
√
and 90 phase margin are ω0 = 259 rad/s and ω90o = 229
o
31
rad/s, respectively.
The corresponding modification parameters are ν0 = 0.1174 and ν90o = 1.3394.
We select ν45o = 0.4256 corresponding to a 45o phase margin and a cross-over
frequency ω45o = 7.5156 rad/s for the optimal control modification adaptive laws.
The closed-loop response is shown in Fig. 9.38. The closed-loop adaptive system is
robustly stable.
All the three robust modification schemes, namely σ modification, e modification,
and optimal control modification exhibit minimum values of the modification para-
meters at which the plant in the Rohrs counterexample begins to stabilize. The σ and
e modification parameters can be found by trial-and-error. In contrast, the modifica-
tion parameter ν is found analytically by taking advantage of the linear asymptotic
property of the optimal control modification method.
30
ω, rad/sec
20
10
0
0 0.2 0.4 0.6 0.8 1
ν
100
50
φ, deg
−50
0 0.2 0.4 0.6 0.8 1
ν
Fig. 9.37 Asymptotic phase margin and cross-over frequency of Rohrs counterexample with
optimal control modification
0.4 4
0.3
2
0.2
u
y
0
0.1
0 −2
0 50 100 0 50 100
t t
0.5 1.5
0 1
y
r
k
k
−0.5 0.5
−1 0
0 50 100 0 50 100
t t
Fig. 9.38 Closed-loop response of Rohrs counterexample with optimal control modification
(ν = 0.4256, γx = γr = 1)
322 9 Robust Adaptive Control
ẋ = ax + bu + gz (9.564)
ż = hz + lx + mu (9.565)
y=x (9.566)
where z (t) is the unmeasurable state with internal dynamics, a unknown but all the
other parameters are known, and h < 0.
The objective is to design an output feedback adaptive controller to track the
following reference model:
Z m (s) bm r
ym = Wm (s) r = km r= (9.567)
Rm (s) s − am
y Z p (s) b (s − h) + gm
= W p (s) = k p = (9.568)
u R p (s) (s − a) (s − h) − gl
where k p = b.
Note that Wm (s) is SPR with a relative degree 1. The plant also has a relative
degree 1 and is assumed to be stable. So, R p (s) = (s − a) (s − h) − gl is Hurwitz.
9.14 Adaptive Control of Non-Minimum Phase Plants … 323
where λ < 0 is a chosen parameter, and θ1∗ , θ2∗ , and θ3∗ are unknown constants.
u ∗ (s) is obtained explicitly as
b−1 −θ1∗ (s − λ) − θ2∗ y ∗ + b−1 bm (s − λ) r
u∗ = (9.570)
s − λ + b−1 θ3∗
This results in
θ3∗ = b (λ − h) + gm (9.573)
y∗ bm (s − λ)
= 2 (9.576)
r s − a + h − θ1∗ s + ah − gl − λθ1∗ + θ2∗
324 9 Robust Adaptive Control
We want the ideal closed-loop plant to track the reference model. So, the ideal
closed-loop transfer function must be equal to the reference model transfer function
Wm (s). Thus,
bm (s − λ) bm
= (9.577)
s 2 − a + h − θ1∗ s + ah − gl − λθ1∗ + θ2∗ s − am
a + h − θ1∗ = λ + am (9.578)
θ1∗ = a − am + h − λ (9.580)
θ2∗ = gl − ah + λ (a + h − λ) (9.581)
b−1 θ2 y b−1 θ3 u
u = −b−1 θ1 y − − + b−1 bm r (9.582)
s−λ s−λ
where θ1 (t), θ2 (t), and θ3 (t) are the estimates of θ1∗ , θ2∗ , and θ3∗ , respectively.
Let θ̃1 (t) = θ1 (t) − θ1∗ , θ̃2 (t) = θ2 (t) − θ2∗ , and θ̃3 (t) = θ3 (t) − θ3∗ . Then, the
output y (s) is expressed as
5 6
Wm (s) θ̃2 y θ̃3 u
y = Wm (s) r − θ̃1 y + + (9.583)
bm s−λ s−λ
Define the tracking error as e (t) = ym (t)−y (t). Then, the tracking error equation
is obtained as
ė = am e + Θ̃ Φ (t) (9.584)
where Θ̃ (t) = θ̃1 (t) θ̃2 (t) θ̃3 (t) and Φ (t) = φ1 (t) φ2 (t) φ3 (t) .
The adaptive law is given by
Θ̇ = −Γ Φ (t) e (9.585)
where
φ1 = y (9.586)
Therefore, all signals are bounded, and the tracking error is asymptotically stable.
For non-minimum phase plants, because the standard MRAC attempts to seek as-
ymptotic tracking by performing a pole-zero cancellation in the right half plane, the
resulting adaptive controller will become unbounded. Thus, if the adaptive law can
be modified to seek only bounded tracking instead of asymptotic tracking, then this
would prevent a pole-zero cancellation in the right half plane. The plant then can be
stabilized.
For non-minimum phase plants, we consider two possible adaptive controllers.
1. We use the same adaptive controller in Eq. (9.582), but with the optimal control
modification adaptive law
Θ̇ = −Γ Φ (t) e − νΦ (t) Θam−1 (9.589)
am y [νbm − am Wm (s)] r
u= + (9.591)
νb νb
Comparing the ideal controller with the asymptotic linear controller, we see that
the adaptive controller does not attempt to cancel the unstable zero of W p (s)
since it has a stable pole at s = am due to the term Wm (s). Otherwise, the pole-
zero cancellation would take place in the right half plane since Z p (s) is unstable.
Therefore, the stability of the adaptive controller is no longer affected by the
non-minimum phase behavior of W p (s). The stability of the closed-loop plant is
then determined by a proper selection of the modification parameter ν such that
ν > 0 and the closed-loop transfer function is stable.
The asymptotic closed-loop transfer function is expressed as
am y [νbm − am Wm (s)] r
u= + (9.597)
νb νb
Note that this asymptotic linear controller is the same as that with the first adaptive
controller. Thus, even though both adaptive controllers are different, the closed-
loop plants for both controllers behave the same in the limit.
We now will formalize the proof of the optimal control modification for the non-
minimum phase plant with relative degree 1 described by Eqs. (9.564) and (9.565)
with a unknown but all the other parameters are known and h < 0. The adaptive
controller is given by Eqs. (9.594) and (9.595).
The closed-loop plant becomes
ẏ = am y + bm r + bk̃ y y + gz (9.598)
ż = hz + l + mk ∗y y + m k̃ y y + mkr r (9.599)
ė = am e − bk̃ y y − gz (9.600)
Let ε (t) = z m (t) − z (t) be the internal state tracking error. Then, the internal
state tracking error equation is given by
ε̇ = hε + l + mk ∗y e − m k̃ y y (9.602)
b2
Using the inequality 2 a b ≤ δ 2 a2 + δ2
, we get
V̇ e, ε, k̃ y ≤ −2α |am | e2 + 2α |g| z m e − 2β |h| ε2
! "
2 ε2
+ αg + β l + mk y ∗
δ1 e + 2
2
δ1
! 2 ε2 "
+ β |m| δ22 y2 k̃ y + 2
δ2
2
− 2ανb2 am−1 y2 k̃ y
+ 2ανb2 am−1 k ∗y y2 k̃ y (9.605)
2
Note that the negative definite term −2ανb2 am−1 y2 k̃ y of the optimal
control modification can be made to dominate the positive definite term
2
β |m| δ22 y2 k̃ y to enable V̇ e, ε, k̃ y to be negative definite.
|αg+β (l+mk ∗y )|
Let c1 = 2α |am | − αg + β l + mk ∗y δ12 , c2 = 2β |h| − δ12
− β|m|
δ22
,
2 −1 ανb2 |am−1 ||k ∗y |
c3 = 2ανb am − β |m| δ2 , and c4 =
2
c3
. Then,
328 9 Robust Adaptive Control
2
V̇ e, ε, k̃ y ≤ −c1 e2 + 2α |g| z m e − c2 ε2 − c3 y2 k̃ y − c4 + c3 c42 y2
(9.606)
We note that y ≤ e + 2 e ym + ym . The ultimate boundsof ym
2 2 2
2
V̇ e, ε, k̃ y ≤ −c5 (e − c6 )2 − c2 ε2 − c3 y2 k̃ y − c4 + c7 (9.607)
ν, α, β, δ1 , and δ2 are
chosen such that c2 > 0, c3 > 0, and c5 > 0. Then, it
follows that V̇ e, ε, k̃ y ≤ 0 if
$
c7
e ≥ c6 + =p (9.608)
c5
$
c7
ε ≥ =q (9.609)
c2
c7
k̃ y ≥ c4 + 2 = κ (9.610)
c3 p + c y r0
Thus, the closed-loop non-minimum phase system is stable with output feedback
adaptive control with the optimal control modification. The ultimate bound of e is
then obtained as
$
β
e ≤ p 2 + q 2 + γ y−1 κ 2 (9.611)
α
It should be noted that the aforementioned output feedback adaptive control de-
sign is feasible only if the knowledge of the plant transfer function is mostly known
with the exception of the parameter a. This requirement imposes a severe restriction
on the design of output feedback adaptive control for non-minimum phase plants
since in many cases such knowledge may not exist if the plants are uncertain. An-
other drawback with this approach is that the desired tracking performance is not
guaranteed. In practical situations, this poor tracking performance can be a major
issue.
ẋ = ax + bu + gz
9.14 Adaptive Control of Non-Minimum Phase Plants … 329
ż = −z + u
y=x
s+1+g
W p (s) =
(s − a) (s + 1)
The system is minimum phase if g > −1. Consider a minimum phase plant with
g = 2. Then, we design the output feedback adaptive controller according to Eq.
(9.582). Let λ = −1, Γ = I , and r (t) = 1. Figure 9.39 shows the response of the
closed-loop plant which tracks the reference model very well in the limit as t → ∞.
Consider a non-minimum phase plant with g = −2. We use both adaptive con-
trollers given by Eqs. (9.582) and (9.594).
The asymptotic closed-loop transfer function is given by
y (s − 1) [νbm (s − am ) − am bm ]
=
r (s − am ) [ν (s − a) (s + 1) − am (s − 1)]
ȳ bm (ν + 1)
=
r νa − am
For a = −2 and am = −1, the transfer function is stable for ν > 0.5. Choose
ν = 2 and γ y = 1. The steady-state closed-loop output is equal to ȳ (t) = −1.
Figure 9.40 shows the stable closed-loop responses of the non-minimum phase plant
with both the adaptive controllers. The adaptive controller 1 with the adaptive pa-
rameter Θ (t) has a faster response than the adaptive controller 2 with the adaptive
parameter k y (t). Both adaptive controllers tend to the same steady-state closed-loop
response.
The closed-loop plant tracks the reference model very poorly as expected, but
the response is stable and tends to an equilibrium value ȳ = −1 which agrees with
the analytical result from the steady-state closed-loop transfer function. The poor
tracking performance is typical of a non-minimum phase system with the output
generally exhibiting an opposite response to the input.
As can be seen from Example 9.22, while the optimal control modification can
stabilize a non-minimum phase plant, the tracking performance is quite unaccept-
able. The issue at hand is the requirement for a non-minimum phase plant to track
330 9 Robust Adaptive Control
1.5
m
y, y
0.5
y
ym
0
0 5 10 15 20
t
0.5
u
0
0 5 10 15 20
t
0.5
m
y, y
0
y − Controller 1
−0.5 y − Controller 2
y
m
−1
0 5 10 15 20
t
1.5
1
u
0.5 u − Controller 1
u − Controller 2
0
0 5 10 15 20
t
Fig. 9.40 Closed-loop response of non-minimum phase plant with optimal control modification
a minimum phase reference model with the same relative degree. This is a very de-
manding and perhaps unrealistic requirement. As a consequence, MRAC attempts to
seek asymptotic tracking which results in an unstable pole-zero cancellation. If the
reference model could be redesigned so that the unstable pole-zero cancellation can-
not occur while the tracking performance can still be met, then the output feedback
adaptive control design would be more acceptable. One such approach is to design
an observer state feedback adaptive control using the Luenberger observer design.
Consider a MIMO system as
ẋ1 A11 A12 x1 B1
ẋ = Ax + Bu ⇔ = + u (9.612)
ẋ2 A21 A22 x2 B2
x1
y = C x = C1 0 = C1 x1 (9.613)
x2
9.14 Adaptive Control of Non-Minimum Phase Plants … 331
P A + A P = −Q (9.614)
B P = C (9.615)
where x̂ (t) is the observer state which estimates the plant state x (t), Â (t) is the
estimate of A, ŷ (t) = C x̂ (t) is the observer output which estimates the plant output
y (t), and L is the Kalman filter gain matrix computed using A0 .
A full-state feedback controller can be designed to enable the output y (t) to track a
reference command signal r (t). For example, we can use the LQR method to design
the full-state feedback controller using the following cost function for tracking a
constant reference command signal r (t):
,
1 tf
J = lim (C x − r ) Q (C x − r ) + u Ru dt (9.617)
t f →∞ 2 0
K x∗ = −R −1 B W (9.618)
−1
K r = −R −1 B A − W B R −1 B C Q (9.619)
W A + A W − W B R −1 B W + C QC = 0 (9.620)
u = K x (t) x̂ + K r r (9.621)
ė = Am e − LCe p − Ã x̂ − B K̃ x x̂ (9.622)
ė p = A p + ΔA e p − Ã x̂ (9.623)
P Am + A
m P = −Q (9.627)
W A p + Ap W = −R (9.628)
The stability of the observer output feedback adaptive control using the optimal
control modification is provided as follows:
Proof Choose a Lyapunov candidate function
V e, e p , K̃ x , Ã = e Pe + ep W e p + trace K̃ x Γx−1 K̃ x + trace ÃΓ A−1 Ã
(9.629)
Then, V̇ e, e p , K̃ x , Ã is evaluated as
V̇ e, e p , K̃ x , Ã = − e Qe − ep R̄e p − 2e P LCe p + 2ν x̂ K x B P A−1
m B K̃ x x̂
+ ην x̂ Â P A−1 + 2c3 e e p
2
m à x̂ ≤ −c1 e − c2 e p
2
2 2 2
− νc4 x̂ K̃ x − c5 + νc4 c52 x̂
2 2 2
− ηc6 x̂ Ã − c7 + ηc6 c72 x̂ (9.630)
where R̄ = R − W ΔA − ΔA W , c1 = λmin (Q), c2 = λmin R̄ , c3 = P LC,
m B
K x∗ B P A−1
c4 = λmin B A− −1
m Q Am B , c5 = c4
, c6 = λmin A− −1
m Q Am , c7 =
m
ΔA P A−1
.
c6
2
We utilize the inequality 2 a b ≤ a2 + b 2
and also note that x̂ ≤
e + 2 e xm + xm . Then, V̇ e, e p , K̃ x , Ã is bounded by
2 2
V̇ e, e p , K̃ x , Ã ≤ − c1 − c3 − νc4 c52 − ηc6 c72 e2 + 2 νc4 c52
2 2 2 2
+ηc6 c72 e xm − (c2 − c3 ) e p − νc4 x̂ K̃ x − c5 − ηc6 x̂
2
à − c7 + νc4 c52 + ηc6 c72 xm 2 (9.631)
Note that the ultimate bound of xm can be expressed as xm ≤ cx r0 . Let
c8 = c1 − c3 − νc4 c52 − ηc6 c72 , c9 = ( 4 5 c8 6 7 ) x 0 , c10 = c2 − c3 , and c11 =
νc c2 +ηc c2 c r
c8 c92 + νc4 c52 + ηc6 c72 cx2 r02 . Then, V̇ e, e p , K̃ x , Ã ≤ 0 outside the compact set
;
S = e (t) ∈ Rn , e p (t) ∈ Rn , K̃ x (t) ∈ Rn×m , Ã (t) ∈ Rn×n : c8 (e − c9 )2
#
2 2
2 2
2
+ c10 e p + νc4 x̂ K̃ x − c5 + ηc6 x̂ Ã − c7 ≤ c11 (9.632)
where A∗m is not established from the non-SPR plant. Then, the model matching
condition cannot be satisfied since there exists no solution of K x∗ . To show this, we
see that suppose K x∗ exists and can be solved using the pseudo-inverse of B with
m < p < n as
−1 ∗
K x∗ = B B B Am − A (9.634)
But −1 ∗
A + B K x∗ = A + B B B B Am − A = A∗m (9.635)
For the standard MRAC, the tracking error equation in the presence of the
mismatch between the reference model and the non-SPR plant is established as
ė = A∗m e + A∗m − Am x̂ − LCe p − Ã x̂ − B K̃ x x̂ (9.636)
Because the optimal control modification only seeks bounded tracking, so the
model matching condition is not satisfied. Using the linear asymptotic property of
the optimal control modification, the asymptotic value of K x (t) and  (t) can be
computed from Eqs. (9.625) and (9.626) by letting Γ → ∞. Then, K x (t) → K̄ x
and  (t) → Ā for a constant reference command signal r (t). From the linear
asymptotic property, we get
1 ∗− −1
K̄ x x̂ = − B Am P B B Pe (9.637)
ν
1
Ā x̂ = − P −1 A∗ Pe + W e p (9.638)
η m
9.14 Adaptive Control of Non-Minimum Phase Plants … 335
P A∗m + A∗
m P = −Q (9.639)
We redefine the estimation errors as K̃ x (t) = K x (t) − K̄ x and à (t) =  (t) − Ā.
Then, the error equations are established as
1 ∗− −1 1 −1 ∗
ė = A∗m + B B Am P B B P + P Am P e
ν η
! "
1 −1 ∗
− LC − P Am W e p + A∗m x̂ − Ã x̂ − B K̃ x x̂ (9.640)
η
! "
1 −1 ∗ 1 −1 ∗
ė p = P Am Pe + A p + P Am W e p + A x̂ − Δ Ã x̂ (9.641)
η η
The standard MRAC results in instability due to the mismatch between the refer-
ence model and the non-SPR plant if A + B K x∗ = Am = A∗m and P Am + A m P ≮ 0,
whereas a stable adaptation of the plant can be achieved with the optimal control
modification.
Proof Choose thesame Lyapunov candidate function in Eq. (9.627). Then, for the
standard MRAC, V̇ e, e p , K̃ x , Ã is evaluated as
∗
V̇ e, e p , K̃ x , Ã = e P A∗m + A∗
m P e + e P Am − Am x̂
+x̂ A∗
m − Am Pe
−2e P LCe p − ep R̄e p (9.642)
Note that P Am + A
m P is not necessarily negative definite. Therefore,
∗ 2
V̇ e, e p , K̃ x , Ã ≤ P Am + A
m P e + P Am − Am
2 e + xm 2
2 2
+ c3 e2 + e p − c2 e p (9.644)
Thus, V̇ e, e p , K̃ x , Ã 0. Therefore, the tracking error is unbounded and the
closed-loop system is unstable.
336 9 Robust Adaptive Control
On the other hand, for the optimal control modification, V̇ e, e p , K̃ x , Ã is
evaluated as
; −1 − #
1 1
V̇ e, e p , K̃ x , Ã = e − Q + P B B A∗− m P B + B ∗−
A m P B B
P e
η ν
1 ∗
+ 2e P A∗m xm − 2e P LC − Am + P A∗m P −1 W + A W e p
η
1
+ e
p − R̄ + W P
−1 ∗
Am + A∗m P −1 W e p + 2e p W Ax m
η
+ 2ν x̂ K x B P A−1 −1
m B K̃ x x̂ + ην x̂ Δ Â P Am Δ Ã x̂ (9.645)
where c17 = c12 − c14 − νc4 c52 − ηc6 c72 , c18 = 15 ( 4 5c17 6 7 ) x 0 , c19 = c13 − c14 ,
c + νc c2 +ηc c2 c r
c20 = cc16 , and c21 = c17 c18
2
+ c19 c20
2
+ νc4 c52 + ηc6 c72 cx2 r02 .
19
Choose L, Q, R, ν, and η such that c17 > 0 and c19 > 0. Then, V̇ e, e p , K̃ x , Ã ≤ 0
outside a compact set. Therefore, the closed-loop system is uniformly ultimately
bounded with the optimal control modification.
Using the linear asymptotic property of the optimal control modification,
L, Q,ν
and η can also be chosen such that the closed-loop matrix formed by ẋ (t) , x̂˙ (t)
9.14 Adaptive Control of Non-Minimum Phase Plants … 337
⎡ −1 ⎤
A 1
BB A∗− PB B P
⎢ 1 −1 ∗
ν m −1 ⎥
Ac = ⎣ LC − η P Am W −LC + ν1 B B A∗−
m PB B P ⎦ (9.648)
+ η1 P −1 A∗
m (P + W )
is Hurwitz.
To demonstrate the observer state feedback adaptive control, we return to Example
9.22.
Example 9.23 The non-minimum phase plant in Example 9.22 is defined by the
following matrices:
a g b
A= , B= ,C= 10
0 −1 1
Figure 9.42 shows the unstable closed-loop response with the standard MRAC
which no longer tracks the reference model. The adaptive parameters are unbounded
and drifting as t → ∞.
The optimal control modification is then used with ν = 0.13 selected. Figure 9.43
shows the closed-loop response with the optimal control modification which is able
to track the ideal reference model very well. All the control and adaptive parameter
signals are bounded.
338 9 Robust Adaptive Control
Fig. 9.41 Closed-loop output response to LQR non-minimum phase reference model with observer
state feedback MRAC
Fig. 9.42 Closed-loop output response to ideal minimum phase reference model with observer
state feedback MRAC
Example 9.23 illustrates the significance of the reference model in adaptive con-
trol of non-minimum phase plants. If the reference model can be redesigned to be
non-minimum phase, then the standard MRAC can achieve asymptotic tracking, al-
beit with a non-minimum phase behavior. Otherwise, even with the observer state
feedback adaptive control design, if the reference model is minimum phase, instabil-
ity will still result with the standard MRAC. On the other hand, the optimal control
modification can produce bounded tracking for both types of reference models.
We have simplified the analysis of the Luenberger observer state feedback adaptive
control. There are some excellent textbooks that cover in-depth the subject of output
feedback adaptive control. The reader is referred to the textbooks by Ioannu [5] and
by Narendra and Annaswamy [35]. The textbook by Lavretsky and Wise presents a
9.14 Adaptive Control of Non-Minimum Phase Plants … 339
Fig. 9.43 Closed-loop output response to ideal minimum phase reference model with optimal
control modification
9.15 Summary
results in asymptotic linear plants in the limit which can be exploited in adaptive
control system design and analysis using many existing linear control techniques.
For systems with input uncertainty, the bi-objective optimal control modification
provides an adaptation mechanism based on both the tracking error and the predictor
error using a predictor model that approximates the plant dynamics. For systems with
slow first-order actuator dynamics, a singular perturbation method is developed by
scaling the adaptive law to allow the plant to track a reference model. An output feed-
back adaptive control approach based on the optimal control modification method
is shown to be able to stabilize first-order plants with unmodeled dynamics having
relative degree greater than 1 by taking advantage of the linear asymptotic prop-
erty. This approach is extended to non-minimum phase plants with relative degree 1.
For these non-minimum phase plants, the optimal control modification adaptive law
can be designed by a suitable selection of the modification parameter to ensure the
closed-loop stability by preventing an unstable pole-zero cancellation typically oc-
curred with the standard MRAC. The potential degraded tracking performance of the
output feedback adaptive control for non-minimum phase systems can be addressed
by observer state feedback adaptive control using the Luenberger observer design.
9.16 Exercises
ÿ + 2ζ ωn ẏ + ωn2 y = bu (t − td )
where b = 1, td = 13 s, and ζ and ωn are unknown but their actual values are
−0.5 and 1 rad/s, respectively.
The system is designed to track a second-order reference model
u = K x (t) x + kr r
where x (t) = y (t) ẏ (t) and K x (t) = k p (t) kd (t) .
a. Calculate the fixed-gain values of k pmin and kdmin to achieve a phase margin
of 60o and a time-delay margin of 1/3 s.
b. Define a convex set described by an ellipse that contains k p (t) and kd (t)
! "2 ! "2
kp kd
g k p , kd = + −1≤0
a b
9.16 Exercises 341
where a and b are to be determined from k pmin and kdmin . Design a projection
method for the adaptive controller to ensure robustness in the presence of
time delay. Write down the adaptive law. Implement the adaptive controller
in Simulink using the following information: y (0) = 0, ẏ (0) = 0, K x (0) =
0, and Γx = 0.2I with a time step Δt = 0.001 s. Plot the time histories of
y (t), u (t), k p (t), and kd (t) for t ∈ [0, 600] s. What happens when the
projection method is removed from the adaptive law?
2. Implement in Simulink the σ modification and e modification for the Rohrs’
counterexample with the reference command
using the same initial conditions k y (0) and kr (0) with γx = γr = 1, σ = 0.2,
and μ = 0.2 and Δt = 0.001 s. Plot the time histories of y (t), u (t), k y (t),
and kr (t) for t ∈ [0, 100] s. Experiment with different values of σ and μ and
determine by trial-and-error the values of σ and μ at which the system begins
to stabilize.
3. Consider a first-order SISO system
ẋ = ax + bu + w
y = Cx
Design a reference model for tracking the output y (t) with a constant refer-
ence command r (t) using the optimal control approach and the following cost
342 9 Robust Adaptive Control
function:
,
1 tf
J = lim (C x − r ) Q (C x − r ) + u Ru dt
t f →∞ 2 0
Derive the expressions for the optimal control gain matrices K x and K r for the
closed-loop system
ẋ = (A + B K x ) x + B K r r
Given
1 2 2
ẋ = + u
1 −1 1
y= 10
ÿ − ẏ + y = u (t − td )
u = Kx x
where x (t) = y (t) ẏ (t) ∈ R2 and K x (t) = k p (t) kd (t) , to achieve an
ideal reference model
ÿm + 6 ẏm + ym = 0
a. Express the optimal control modification adaptive law for K x (t). Let Γ →
∞ and Q = I , calculate the equilibrium values of K x (t) as a function of
the modification parameter ν.
b. Determine numerically the value of the modification parameter ν to achieve
the maximum time-delay margin to within 0.001. Compute the equilibrium
values of K x (t) corresponding to this modification parameter ν. Implement
the adaptive controller in Simulink with this modification parameter using
the following information: Γ = 10I , y (0) = 1, ẏ (0) = 0, and K x (0) = 0
with a time step Δt = 0.001 s. Plot the time histories of x (t), u (t), and
K x (t) for t ∈ [0, 10] s.
9.16 Exercises 343
This disturbance will cause a parameter drift when the standard MRAC is used
in a regulator design.
An adaptive controller is designed as
u = kr r − θ (t) x − ŵ (t)
ẋm = am xm + bm r
ÿ + 2ζ ωn ẏ + ωn2 y = bu (t − td )
ẋm = am xm + bm r
References
1. Rohrs, C. E., Valavani, L., Athans, M., & Stein, G. (1985). Robustness of continuous-time
adaptive control algorithms in the presence of unmodeled dynamics. IEEE Transactions on
Automatic Control, AC-30(9), 881–889.
2. Ioannou, P., & Kokotovic, P. (1984). Instability analysis and improvement of robustness of
adaptive control. Automatica, 20(5), 583–594.
3. Ioannu, P. A. and Sun, J., (1996). Robust Adaptive Control, Prentice-Hall, Inc., 1996
4. Narendra, K. S., & Annaswamy, A. M. (1987). A new adaptive law for robust adaptation
without persistent excitation. IEEE Transactions on Automatic Control, AC-32(2), 134–145.
5. Ioannu, P. A., & Sun, J. (1996). Robust adaptive control. Upper Saddle River: Prentice-Hall,
Inc.
6. Calise, A. J., & Yucelen, T. (2012). Adaptive loop transfer recovery. AIAA Journal of Guidance,
Control, and Dynamics, 35(3), 807–815.
7. Lavretsky, E. (2009). Combined/composite model reference adaptive control. IEEE Transac-
tions on Automatic Control, 54(11), 2692–2697.
8. Hovakimyan, N., & Cao, C. (2010). L1 Adaptive control theory: Guaranteed robustness with
fast adaptation. Philadelphia: Society for Industrial and Applied Mathematics.
9. Nguyen, N. (2012). Optimal control modification for robust adaptive control with large adaptive
gain. Systems and Control Letters, 61(2012), 485–494.
10. Yucelen, T., & Calise, A. J. (2011). Derivative-free model reference adaptive control. AIAA
Journal of Guidance, Control, and Dynamics, 34(4), 933–950.
11. Nguyen, N., Krishnakumar, K., & Boskovic, J. (2008). An optimal control modification to
model-reference adaptive control for fast adaptation. In AIAA Guidance, Navigation, and Con-
trol Conference, AIAA 2008-7283, August 2008.
12. Campbell, S., Kaneshige, J., Nguyen, N., & Krishnakumar, K. (2010). An adaptive control
simulation study using pilot handling qualities evaluations. In AIAA Guidance, Navigation,
and Control Conference, AIAA-2010-8013, August 2010.
13. Campbell, S., Kaneshige, J., Nguyen, N., & Krishnakumar, K. (2010). Implementation and
evaluation of multiple adaptive control technologies for a generic transport aircraft simulation.
In AIAA Infotech@Aerospace Conference, AIAA-2010-3322, April 2010.
346 9 Robust Adaptive Control
14. Hanson, C., Johnson, M., Schaefer, J., Nguyen, N., & Burken, J. (2011). Handling qualities
evaluations of low complexity model reference adaptive controllers for reduced pitch and roll
damping scenarios. In AIAA Guidance, Navigation, and Control Conference, AIAA-2011-6607,
August 2011.
15. Hanson, C., Schaefer, J., Johnson, M., & Nguyen, N. Design of Low Complexity Model Ref-
erence Adaptive Controllers, NASA-TM-215972.
16. Nguyen, N., Hanson, C., Burken, J., & Schaefer, J. (2016). Normalized optimal control modi-
fication and flight experiments on NASA F/A-18 Aircraft. AIAA Journal of Guidance, Control,
and Dynamics.
17. Schaefer, J., Hanson, C., Johnson, M., & Nguyen, N. (2011). Handling qualities of model
reference adaptive controllers with varying complexity for pitch-roll coupled failures. AIAA
Guidance, Navigation, and Control Conference, AIAA-2011-6453, August 2011.
18. Bryson, A. E., & Ho, Y. C. (1979). Applied optimal control: Optimization, estimation, and
control. John Wiley & Sons, Inc.
19. Bryson, A. E., & Ho, Y. C. (1979). Applied optimal control: Optimization, estimation, and
control. New Jersey: Wiley Inc.
20. Nguyen, N. (2013). Adaptive control for linear uncertain systems with unmodeled dynam-
ics revisited via optimal control modification. In AIAA Guidance, Navigation, and Control
Conference, AIAA-2013-4988, August 2013.
21. Nguyen, N. (2010). Asymptotic linearity of optimal control modification adaptive law with
analytical stability margins. In AIAA Infotech@Aerospace Conference, AIAA-2010-3301, April
2010.
22. Calise, A. J., Yucelen, T., Muse, J., & Yang, B. (2009). A loop recovery method for adaptive
control. In AIAA Guidance, Navigation, and Control Conference, AIAA-2009-5967, August
2009.
23. Cao, C., & Hovakimyan, N. (2007). Guaranteed transient performance with L1 adaptive con-
troller for systems with unknown time-varying parameters and bounded disturbances: Part I.
In American Control Conference, July 2007.
24. Cao, C., & Hovakimyan, N. (2008). Design and analysis of a novel L1 adaptive control ar-
chitecture with guaranteed transient performance. IEEE Transactions on Automatic Control,
53(2), 586–591.
25. Nguyen, N., Burken, J., & Hanson, C. (2011). Optimal control modification adaptive law with
covariance adaptive gain adjustment and normalization. In AIAA Guidance, Navigation, and
Control Conference, AIAA-2011-6606, August 2011.
26. Nguyen, N. (2014). Multi-objective optimal control modification adaptive control method for
systems with input and unmatched uncertainties. In AIAA Guidance, Navigation, and Control
Conference, AIAA-2014-0454, January 2014.
27. Nguyen, N. and Balakrishnan, S. N. (2014). Bi-objective optimal control modification adaptive
control for systems with input uncertainty. IEEE/CAA Journal of Automatica Sinica, 1(4), 423–
434.
28. Nguyen, N. (2012). Bi-objective optimal control modification adaptive control for systems with
input uncertainty. In AIAA Guidance, Navigation, and Control Conference, AIAA-2012-4615,
August 2012.
29. Nguyen, N. (2013). Bi-objective optimal control modification adaptive law for unmatched
uncertain systems. In AIAA Guidance, Navigation, and Control Conference, AIAA-2013-4613,
August 2013.
30. Slotine, J.-J., & Li, W. (1991). Applied nonlinear control. Upper Saddle River: Prentice-Hall,
Inc.
31. Nguyen, N., Krishnakumar, K., Kaneshige, J., & Nespeca, P. (2008). Flight dynamics model-
ing and hybrid adaptive control of damaged asymmetric aircraft. AIAA Journal of Guidance,
Control, and Dynamics, 31(3), 751–764.
32. Nguyen, N., Ishihara, A., Stepanyan, V., & Boskovic, J. (2009). Optimal control modification
for robust adaptation of singularly perturbed systems with slow actuators. In AIAA Guidance,
Navigation, and Control Conference, AIAA-2009-5615, August 2009.
References 347
33. Kokotovic, P., Khalil, H., & O’Reilly, J. (1987). Singular perturbation methods in control:
analysis and design. Philadelphia: Society for Industrial and Applied Mathematics.
34. Ardema, M. (1981). Computational singular perturbation method for dynamical systems. AIAA
Journal of Guidance, Control, and Dynamics, 14, 661–663.
35. Narendra, K. S., & Annaswamy, (2005). Stable adaptive systems. New York: Dover Publica-
tions.
36. Lavretsky, E., & Wise, K. (2012). Robust and adaptive control. Berlin: Springer.
Chapter 10
Aerospace Applications
Abstract This chapter presents several adaptive control applications with a partic-
ular focus on aerospace flight control applications. Two relatively simple pendulum
applications are used to illustrate a nonlinear dynamic inversion adaptive control
design method for tracking a linear reference model. The rest of the chapter presents
several adaptive flight control applications for rigid aircraft and flexible aircraft.
The chapter concludes with an application of the optimal control modification to a
F-18 aircraft model. The flight control applications for rigid aircraft include the σ
modification, e modification, bi-objective optimal control modification, least-squares
adaptive control, neural network adaptive control, and hybrid adaptive control which
combines model-reference adaptive control with a recursive least-squares parameter
identification method. An adaptive control design for a flexible aircraft is presented
using a combined adaptive law with the optimal control modification and adaptive
loop recovery modification to suppress the dynamics of the aircraft flexible modes.
An adaptive linear quadratic gaussian (LQG) design based on the optimal control
modification for flutter suppression is presented. By taking advantage of the linear
asymptotic property, the adaptive flutter suppression can be designed to achieve a
closed-loop stability with output measurements.
In some instances, the intelligent flight control system generated large commands
that caused load limit excursions. A subsequent modification to simplify the neu-
ral network demonstrated an improvement in the intelligent flight control system
in simulations. However, during the final phase of the flight test experiment, mixed
results were obtained. Lateral pilot-induced oscillations were experienced during an
in-flight simulated stabilator failure. Consequently, a further study was conducted at
the conclusion of the IFCS program in 2006. It was noted during this study that the
normalization method performed well in simulations.
During 2010 to 2011, NASA conducted a follow-on flight test program of adaptive
control onboard a F/A-18A aircraft (tail number 853) shown in Fig. 10.2. The adaptive
controllers were a simplified MRAC with Φ (x) = x and the optimal control modifi-
cation with and without normalization [5–8]. Prior to the flight test program, several
adaptive controllers were evaluated by eight NASA test pilots in the Advanced Flight
Concept Simulator at NASA Ames Research Center in 2009 [9, 10]. The evaluation
study showed that the optimal control modification performed well across different
failure scenarios [7, 9]. NASA Armstrong Flight Research Center conducted a fur-
ther investigation of the optimal control modification and simulation experiments in
a F-18 flight simulator [11]. The simulation results supported the improved perfor-
mance with the optimal control modification in a simplified adaptive control design.
In addition, the normalization method was also implemented in conjunction with the
optimal control modification based on the previous study during the IFCS program
[12]. The flight test program showed improvements in the adaptation process under
most failure scenarios [5–8].
The use of adaptive control potentially could help reduce the development cost
of production flight control systems for aerospace vehicles. Typically, in aerospace
production systems, accurate plant models must be developed to ensure high con-
fidence in a flight control design. The model development of a typical aerospace
system in general can be quite expensive since extensive experimental validation in
wind tunnel facilities and in flight is often required. Adaptive control may permit
the use of less accurate plant models since the adaptation process can accommodate
10 Aerospace Applications 351
the uncertainty in the plants. For this to happen, certification barriers for adaptive
control must be overcome.
In aerospace applications, the ability of an adaptive controller to modify a pre-
designed control system is viewed as a strength and a weakness at the same time.
On the one hand, the premise of being able to accommodate system failures is a
major advantage of adaptive control. On the other hand, potential problems with
adaptive control exist with regard to robustness. Therefore, in spite of the many
recent advances in adaptive control, currently there are no adaptive controllers that
have yet been certified for mission-critical or safety-critical systems. This is not to
say that adaptive control is generally not considered as a viable option. Quite the
contrary, in a number of major aerospace systems, adaptive control has been viewed
as an option in the recent years.
The development of certifiable adaptive flight control systems represents a major
challenge to overcome. Adaptive control will never become part of the future unless it
can be proven that it is is highly safe and reliable. Technical challenges are continued
to be addressed by the research community.
To demonstrate the potential benefits of adaptive control, this chapter presents
some example adaptive control applications with a particular focus on aerospace
flight control. The learning objectives of this chapter are:
• To be able to design adaptive controllers for a variety of applications;
• To be able to design nonlinear dynamic inversion adaptive controllers for nonlinear
systems;
• To develop a familiarity with various aircraft applications and a basic understand-
ing of aircraft flight control;
• To learn how to design adaptive controllers for systems with partial-state informa-
tion using the Kalman filter; and
• To be able to combine both direct MRAC and indirect least-squares parameter
estimation to improve effectiveness of adaptive controllers.
a slight perturbation will cause the pendulum to swing down to the stable equilibrium
in the vertically downward position.
1 1
m L 2 θ̈ − mgL sin θ + cθ̇ = u (t − td ) (10.1)
3 2
where m is the mass of the pendulum, L is the length, g is the gravity constant, c is the
damping coefficient which is assumed to be unknown, θ (t) is the angular position
of the pendulum, u (t) is the control input which represents the motor torque, and td
is a time delay which represents the motor actuator dynamics.
Let x1 (t) = θ (t), x2 (t) = θ̇ (t), and x (t) = x1 (t) x2 (t) . Then,
where
x2
f (x) = 3g (10.3)
2L
sin x1
0
B= 3 (10.4)
m L2
ẋ = Am x + Bm r + B c̃x2 (10.7)
10.1 Inverted Pendulum 353
ė = Am e − B c̃x2 (10.8)
ĉ˙ = γ x2 e P B (10.9)
However, in the presence of time delay, the adaptive law may be modified to
include the optimal control modification as
ĉ˙ = γ x2 e P B + νx22 ĉB P A−1
m B (10.10)
ν can be chosen to guarantee stability in the presence of the time delay by using
the linear asymptotic property of the optimal control modification which yields
1 − −1
B ĉx2 = − B Am P B Pe (10.11)
ν
as γ → ∞.
The closed-loop plant with the time delay is expressed as
ẋ = f (x) − f (x (t − td )) + Am x (t − td ) + Bm r (t − td ) + B ĉ (t − td ) x2 (t − td ) − Bcx2
(10.12)
The nonlinear term f (x) does create a problem since linear stability analysis
cannot be used. However, consider the case when td is small, then
x2 − x2 (t − td )
f (x) − f (x (t − td )) = 3g (10.13)
2L
3g
sin x1 − 2L sin x1 (t − td )
Therefore,
x2 − x2 (t − td )
f (x) − f (x (t − td )) ≈ 3g (10.15)
t x cos x1
2L d 2
where
0 1
A (t) = 3g (10.17)
0 2L t
Example 10.1 Let m = 0.1775 slug, L = 2 ft, c = 0.2 slug-ft2 /s, td = 0.05 s, and
θ (0) = θ̇ (0) = 0. The reference model is specified by
20 20
.
.
, m , deg/sec
15
, m , deg
10 m
10
. 0
5 .
0 -10
0 5 10 0 5 10
t, sec t, sec
1 0.1
0
c, lb-ft-sec
0
u, lb-ft
-0.1
-1 <
-0.2
-2 -0.3
0 5 10 0 5 10
t, sec t, sec
1 1 1
(m 1 + 3m 2 ) L 21 θ̈1 + m 2 L 1 L 2 θ̈2 cos (θ2 − θ1 ) − m 2 L 1 L 2 θ22 sin (θ2 − θ1 )
3 2 2
1
+ (m 1 + 2m 2 ) gL 1 sin θ1 + (c1 + c2 ) θ̇1 − c2 θ̇2 = u 1 (t − td ) + u 2 (t − td )
2
(10.18)
1 1 1
m 2 L 22 θ̈2 + m 2 L 1 L 2 θ̈1 cos (θ2 − θ1 ) + m 2 L 1 L 2 θ12 sin (θ2 − θ1 )
3 2 2
1
+ m 2 gL 2 sin θ2 + c2 θ̇2 = u 2 (t − td ) (10.19)
2
where m 1 and m 2 are the masses of the linkages, L 1 and L 2 are the lengths, c1 and c2
are the friction coefficients at the joints which are assumed to be unknown, g is the
gravity constant, θ1 (t) and θ2 (t) are the angular positions of the double pendulum,
u 1 (t) and u 2 (t) are the control variables which represent the motor torques at the
joints, and td is the time delay due to the motor actuator dynamics.
These equations can be recast as
(m 1 +3m 2 )L 21 m 2 L 1 L 2 cos(θ2 −θ1 )
3 2
θ̈1
m 2 L 1 L 2 cos(θ2 −θ1 ) m 2 L 22 θ̈2
2
3
p(x1 )
m 2 L 1 L 2 θ22 sin(θ2 −θ1 )
− (m 1 +2m 22)gL 1 sin θ1
= 2 2
m L L θ sin(θ −θ )
− 2 1 2 12 2 1 − m 2 gL22 sin θ2
f (x1 )
11 u 1 (t − td ) −c1 − c2 c2 θ̇1
+ + (10.20)
01 u 2 (t − td ) 0 −c2 θ̇2
C D
356 10 Aerospace Applications
Let x1 (t) = θ1 (t) θ2 (t) , x2 (t) = θ̇1 (t) θ̇2 (t) , and u (t) =
u 1 (t) u 2 (t) . Then, the equations of motion become
where K p and K d are the proportional and derivative control gain matrices such that
−ωm2 1 0
Kp = = diag −ωm2 1 , −ωm2 2 (10.23)
0 −ωm2 2
−2ζm 1 ωm 1 0
Kd = = diag −2ζm 1 ωm 1 , −2ζm 2 ωm 2 (10.24)
0 −2ζm 2 ωm 2
Let e (t) = xm1 (t) − x1 (t) xm2 (t) − x2 (t) , then the tracking error equation
is obtained as
ė = Am e + B Θ̃ Φ (x1 , x2 ) (10.32)
where ⎡ ⎤
0 0 1 0
⎢ 0 0 0 1 ⎥
Am = ⎢
⎣ −ωm2
⎥
⎦ (10.33)
1
0 −2ζm 1 ωm 1 0
0 −ωm 2 2
0 −2ζm 2 ωm 2
⎡ ⎤
0 0
⎢0 0⎥
B=⎢
⎣1
⎥ (10.34)
0⎦
0 1
Due to the presence of the time delay which represents the motor actuator dynamics,
the adaptive law for Θ (t) should include a robust modification scheme or the pro-
jection method. For example, the adaptive law for Θ (t) with the σ modification is
given by
Θ̇ = −Γ Φ (x1 , x2 ) e P B + σ Θ (10.35)
358 10 Aerospace Applications
150 300
θ , θm , deg
θ , θ , deg
100 200
2
m
50 100
2
θ1 θm θ2 θm
1 2
0 0
0 5 10 0 5 10
t, sec t, sec
200 15
10
0
u, lb−ft
Θ
5
−200
0
u1 u2
−400 −5
0 5 10 0 5 10
t, sec t, sec
The responses of the closed-loop system with the standard MRAC and σ modifi-
cation are shown in Figs. 10.6 and 10.7. The closed-loop response with the standard
MRAC is very poor due to the time delay. The adaptive control signals are very large,
and the adaptive parameters appear to be drifting. In other words, the closed-loop
system is on the verge of instability. On the other hand, the σ modification produces
a very good closed-loop response. The adaptive control signals exhibit some ini-
tial high-frequency oscillations but then settle to steady-state values. The adaptive
parameters all converge to zero.
150 300
θ , θm , deg
θ2, θm , deg
100 200
1
50 100
1
θ1 θm θ2 θm
1 2
0 0
0 5 10 0 5 10
t, sec t, sec
40 2
1
20
u, lb−ft
0
0
−1
u1 u2
−20 −2
0 5 10 0 5 10
t, sec t, sec
The longitudinal dynamics of an aircraft are described by the airspeed V , the angle
of attack α, and the pitch rate q as shown in Fig. 10.8. There are two longitudinal
modes: the phugoid mode and the short-period mode. The short-period mode involves
the angle of attack and pitch rate that affect the aircraft longitudinal stability.
The equations of motion for a symmetric flight are given by [14]
I yy q̇ = Cm q̄ S c̄ + T z e (10.37)
where m (t) is the aircraft mass, q̄ (t) is the dynamic pressure, S is the wing reference
area, c̄ is the mean aerodynamic chord, T (t) is the engine thrust, V (t) is the airspeed,
α (t) is the angle of attack, θ (t) is the pitch attitude, q (t) = θ̇ (t) is the pitch rate,
I yy (t) is the mass moment of inertia about the aircraft pitch axis, C L (t) is the lift
coefficient, and Cm (t) is the pitching moment coefficient about the aircraft center
of gravity. Note that the aircraft mass and inertia properties as well as aerodynamic
coefficients are time-varying due to the fuel burn.
The motion can be decomposed into a trim motion and a perturbed motion. The
trim motion is a motion when the aircraft is in an equilibrium for a steady-level flight
for which V (t) = V̄ , α (t) = ᾱ, and q (t) = q̄ = 0. The perturbed motion is a small
amplitude motion about the trim condition. Because the amplitudes are assumed to
be small, linearization of the equations of motion can be performed to derive the
linear equations of motion for the perturbed motion. The equations of motion are
linearized as
360 10 Aerospace Applications
α̇ c̄ q c̄
m V̄ α̇ = − C L α α + C L α̇ + CLq + C L δe δe q̄ S + mg sin γ̄ (α − θ ) + m V̄ q
2 V̄ 2 V̄
(10.38)
α̇ c̄ q c̄
I yy q̇ = Cm α α + Cm α̇ + Cm q + Cm δe δe q̄ S c̄ (10.39)
2 V̄ 2 V̄
where γ̄ = θ̄ − ᾱ is the trim flight path angle; δe (t) is the elevator control surface
deflection; C L α , C L α̇ , C L q , and C L δe are stability and control derivatives of C L ; and
Cm α , Cm α̇ , Cm q , and Cm δe are stability and control derivatives of Cm .
Assuming γ̄ = 0 and neglecting C L α̇ and C L q which are generally small, then the
equations of motion can be written as
⎡ ⎤ ⎡ Zα ⎤⎡ ⎤
α̇ V̄
0 1 α
⎣ θ̇ ⎦ = ⎣ 0 0 1 ⎦⎣θ ⎦
q̇ Mα + Mα̇V̄Z α 0 Mq + Mα̇ q
ẋ A x
⎛ ⎞
⎡ ⎤ ⎡ ⎤ ⎡ ⎤
Z δe ⎜ α ⎟ wα
⎢ V̄ ⎥⎜⎜
∗ ⎣ ⎦⎟
⎟
+ λ⎣ 0 δ (t −
⎦ ⎜e
d t ) + θ ∗
0 θ θ + ⎣ wθ ⎦
q ⎟
α q
M Z
Mδe + α̇V̄ δe ⎝ ⎠ wq
u(t−td )
Θ ∗
B x
(10.40)
C L α q̄ S C L q̄ S
where Z α = − m
and Z δe = δme are the normal force derivatives (positive
C q̄ S c̄ Cm q̄ S c̄ Cm α̇ q̄ S c̄2 Cm q q̄ S c̄2
down); Mα = mIαyy , Mδe = δIeyy , Mα̇ = 2I , and Mq = 2I are the
yy V̄ yy V̄
∗ ∗
pitching moment derivatives (positive nose up); θα and θq are the uncertainty in
α (t) and q (t) due to failures; λ is the uncertainty in the control effectiveness of the
elevator control surface; and td is a time delay which represents the elevator control
surface actuator dynamics.
Suppose an adaptive pitch attitude controller is to be designed to enable the pitch
attitude to follow a reference model specified as
For perfect tracking, the pitch attitude should have the same dynamics. Therefore,
where Θ ∗ = θα∗ 0 θq∗ and x (t) = α (t) θ (t) q (t) .
Therefore, the elevator control surface deflection δe (t) can be designed by invert-
ing the pitch rate equation. This yields
δe = kα α + kθ (θ − r ) + kq q − Θ (t) x (10.44)
where
Mα + Mα̇ Z α
kα = − V̄
Mα̇ Z δe
(10.45)
Mδe + V̄
ωm2
kθ = − Mα̇ Z δe
(10.46)
Mδe + V̄
2ζm ωm + Mq + Mα̇
kq = − Mα̇ Z δe
(10.47)
Mδe + V̄
ė = Am e + B Θ̃ x (10.49)
The optimal control modification can be used for the adaptive law to take advan-
tage of the linear asymptotic property that allows the modification parameter ν to
be computed to guarantee stability of the closed-loop system in the presence of time
delay.
The adaptive controller is designed using the bi-objective optimal control modi-
fication adaptive laws as follows:
u = K x x + kr r + u ad (10.50)
where
u ad = ΔK x (t) x + Δkr (t) r − Θ (t) x (10.51)
362 10 Aerospace Applications
Then, ΔK x (t), Δkr (t), and Θ (t) are computed from the following bi-objective
optimal control modification adaptive laws:
Δ K̇ x = Γ K x x e P + νu ad
λ̂B P A−1
m B λ̂ (10.52)
Δk̇r = γkr r e P + νu ad
λ̂B P A−1
m B λ̂ (10.53)
!
Θ̇ = −ΓΘ x e P + νu ad
λ̂B P A−1
m + e
p P − η u + 2Θ Φ (x) λ̂B + ŵ P A−1
m B λ̂
(10.54)
" #
λ̂˙ = −γλ u + Θ x ep P − η u + 2Θ Φ (x) λ̂B + ŵ P A−1
m B
" # (10.55)
ŵ˙ = −γw ep P − η u + 2Θ Φ (x) λ̂B + ŵ P A−1
m (10.56)
Example 10.3 The short-period mode dynamics of a transport aircraft at Mach 0.8
and 30,000 ft are described by
ẋ = Ax + Bλ u (t − td ) + Θ ∗ x
The adaptive flight control architecture for the bi-objective optimal control mod-
ification is illustrated in Fig. 10.9 [15].
Figure 10.10 shows the aircraft response due to the baseline controller. With no
adaptation, the closed-loop plant does not track the reference model well.
Figure 10.11 is the plot of the aircraft response with the standard MRAC for the
adaptation rates Γx = ΓΘ = 50I and γr = 50. The attitude angle command tracking
has improved considerably. However, there are large initial transients in the pitch
rate response as well as high-frequency oscillations.
Figure 10.12 shows the aircraft response with the bi-objective MRAC for the
adaptation rates Γx = ΓΘ = 50I and γr = γλ = γw = 50 by setting ν = η = 0 in the
bi-objective optimal control modification adaptive laws. The closed-loop becomes
unstable after 14 s. The instability of the adaptive laws is consistent with the theory
which shows that η cannot be zero when an external disturbance w (t) exists due to
the term c8 in the stability theorem. Moreover, it is also consistent with the MRAC
theory which establishes that the standard MRAC generally exhibits a parameter drift
in the presence of a disturbance. To prevent a parameter drift, the disturbance estimate
ŵ (t) must be bounded by setting η > 0. If the disturbance w (t) is not estimated by
setting γw = 0, then the bi-objective MRAC is stable as shown in Fig. 10.13 since
the term c8 becomes bounded when the adaptive law for ŵ (t) is not present. The
bi-objective MRAC with γw = 0 has a very similar tracking performance to that with
the standard MRAC.
Figure 10.14 shows the aircraft response with the bi-objective optimal control
modification for the same adaptation rates with ν = η = 0.4. The closed-loop
response with the bi-objective optimal control modification is significantly improved
with a very good tracking performance and no high-frequency oscillations.
Fig. 10.9 Adaptive flight control architecture for bi-objective optimal control modification
364 10 Aerospace Applications
3
α α
m
α, deg
0
−3
0 5 10 15 20 25 30 35 40
t, sec
5
θ θm
θ, deg
0
−5
0 5 10 15 20 25 30 35 40
t, sec
5
qm
q, deg/sec
q
0
−5
0 5 10 15 20 25 30 35 40
t, sec
3
α αm
α, deg
−3
0 5 10 15 20 25 30 35 40
t, sec
5
θ θ
m
θ, deg
−5
0 5 10 15 20 25 30 35 40
t, sec
5
qm
q, deg/sec
q
0
−5
0 5 10 15 20 25 30 35 40
t, sec
3
α αm
α, deg
−3
0 5 10 15 20 25 30 35 40
t, sec
5
θ θ
m
θ, deg
−5
0 5 10 15 20 25 30 35 40
t, sec
5
qm
q, deg/sec
q
0
−5
0 5 10 15 20 25 30 35 40
t, sec
3
α α
m
α, deg
0
−3
0 5 10 15 20 25 30 35 40
t, sec
5
θ θm
θ, deg
0
−5
0 5 10 15 20 25 30 35 40
t, sec
5
qm
q, deg/sec
q
0
−5
0 5 10 15 20 25 30 35 40
t, sec
3
α α
m
α, deg
−3
0 5 10 15 20 25 30 35 40
t, sec
5
θ θm
θ, deg
−5
0 5 10 15 20 25 30 35 40
t, sec
5
qm
q, deg/sec
q
0
−5
0 5 10 15 20 25 30 35 40
t, sec
Figure 10.15 compares the elevator deflections produced by all the various con-
trollers. The elevator deflections with the standard MRAC and bi-objective MRAC
with γw = 0 exhibit significant high-frequency oscillations and high amplitudes.
The elevator deflection with the bi-objective MRAC with γw > 0 is in full satura-
tion before the controller goes unstable at 14 s. In contrast, the bi-objective optimal
control modification produces a well-behaved control signal for the elevator deflec-
tion with no discernible saturation or high-frequency oscillations. The amplitude of
the elevator deflection produced by the bi-objective optimal control modification is
roughly double that for the baseline controller to account for the 50% reduction in
the control effectiveness of the elevator.
366 10 Aerospace Applications
δ , deg
0
e
Baseline
−5
0 5 10 15 20 25 30 35 40
t, sec
δ , deg
0
e
MRAC
−5
0 5 10 15 20 25 30 35 40
t, sec
5
δ , deg
0 Bi−Objective MRAC
e
−5
0 5 10 15 20 25 30 35 40
t, sec
5
δ , deg
0
e
Bi−Objective MRAC, γ =0
−5 w
0 5 10 15 20 25 30 35 40
t, sec
5
δ , deg
0
e
The elevator input is designed with the following proportional-integral (PI) control
law:
δe = kα α + kθ (θ − r ) + kq q − fˆ (α) (10.59)
10.4 Recursive Least-Squares and Neural Network Pitch … 367
where x (t) = α (t) θ (t) q (t) , K x = kα kθ kq , and fˆ (α) is the function
approximation of the unstructured uncertainty f (α).
The plant modeling error is computed as ε (t) = ẋd (t) − ẋ (t) = Am x (t) +
Bm r (t) − ẋ (t), where ẋ (t) is estimated by a backward finite-difference method.
The uncertainty is modeled with the first four terms of the Chebyshev orthogonal
polynomials
fˆ (α) = Θ Φ (α) = θ0 + θ1 α + θ2 2α 2 − 1 + θ3 4α 3 − 3α (10.60)
Note that the adaptive laws are implemented with the scaling of the B matrix by
−1
B B to keep the RLS parameter η < 2 (see Sect. 7.3.2).
As discussed in Sect. 6.6, the signal ẋ (t) may not be necessarily available and
therefore needs to be estimated. The predictor model of the plant can be implemented
to estimate the signal ẋ (t) as follows:
x̂˙ = Am x̂ + (A − Am ) x + B u + Θ Φ (α) (10.64)
Then, the predictor-based least-squares gradient and RLS adaptive laws are given
by
−1
Θ̇ = −Γ Φ (α) εp B B B (10.65)
−1
Θ̇ = −RΦ (α) εp B B B (10.66)
Θ̇ = −Γ Φ (α) e P B (10.67)
Example 10.4 The numerical model of the short-period dynamics for a transport
aircraft at Mach 0.8 and 30,000 ft is given by
⎡ ⎤ ⎡ ⎤⎡ ⎤ ⎡ ⎤
α̇ −0.7018 0 0.9761 α −0.0573
⎣ θ̇ ⎦ = ⎣ 0 0 1 ⎦⎣θ ⎦ + ⎣ 0 ⎦ [δe + f (α)]
q̇ −2.6923 0 −0.7322 q −3.5352
The least-squares gradient and RLS adaptive controllers are implemented with
η = 0.2 and Γ = R (0) = I . The aircraft longitudinal responses with the nominal
controller and with the least-squares gradient and RLS adaptive controllers are shown
in Figs. 10.16, 10.17 and 10.18.
20
α, deg
0
α αm
−20
0 10 20 30 40 50 60
t, sec
50
θ θm
θ, deg
−50
0 10 20 30 40 50 60
t, sec
50
q, deg/sec
q qm
0
−50
0 10 20 30 40 50 60
t, sec
The aircraft response with the nominal controller is seen to be extremely poor with
the maximum angle of attack of almost 14o which is well into stall. The benefit of
the least-squares gradient adaptive control is clearly demonstrated by the simulation
results which show a very good tracking performance. However, the RLS adaptive
control does not perform well. This could be due to the slow parameter convergence of
the RLS adaptive control which can improve stability robustness of adaptive control
in the presence of time delay or unmodeled dynamics at the expense of the tracking
performance.
5
α, deg
0
α α
m
−5
0 10 20 30 40 50 60
t, sec
5
θ θ
m
θ, deg
−5
0 10 20 30 40 50 60
t, sec
10
q, deg/sec
q qm
0
−10
0 10 20 30 40 50 60
t, sec
5
α, deg
0
α αm
−5
0 10 20 30 40 50 60
t, sec
10
θ θ
m
θ, deg
−10
0 10 20 30 40 50 60
t, sec
10
qm
q, deg/sec
q
0
−10
0 10 20 30 40 50 60
t, sec
α, deg
0
α αm
−5
0 10 20 30 40 50 60
t, sec
10
θ θm
θ, deg
0
−10
0 10 20 30 40 50 60
t, sec
10
qm
q, deg/sec
q
0
−10
0 10 20 30 40 50 60
t, sec
Fig. 10.19 Aircraft response with predictor-based least-squares gradient adaptive control
5
α, deg
0
α αm
−5
0 10 20 30 40 50 60
t, sec
10
θ θ
m
θ, deg
−10
0 10 20 30 40 50 60
t, sec
10
q
q, deg/sec
q m
−10
0 10 20 30 40 50 60
t, sec
Fig. 10.20 Aircraft response with predictor model-based RLS adaptive control (η = 0.2)
Figures 10.19 and 10.20 illustrate the use of the predictor-based plant modeling
error ε p (t) in the least-squares gradient and RLS adaptive control instead of the
true plant modeling error ε (t). The predictor-based least-squares gradient adaptive
controllers provide a very good tracking performance as compared to its counterpart
using the true plant modeling error. The pitch rate response exhibits some small
initial high-frequency oscillations. The predictor-based RLS adaptive controller is
observed to provide much better tracking after 20 s than the original RLS adaptive
controller, but overall both the RLS adaptive controllers perform worse than the least-
squares gradient adaptive controllers. The results indicate that the predictor model
can provide a good estimation of the signal ẋ (t) without differentiation which can
introduce noise in practical applications.
10.4 Recursive Least-Squares and Neural Network Pitch … 371
α, deg
0
α αm
−5
0 10 20 30 40 50 60
t, sec
10
θ θ
m
θ, deg
0
−10
0 10 20 30 40 50 60
t, sec
10
q
q, deg/sec
q m
−10
0 10 20 30 40 50 60
t, sec
5
α, deg
0
α αm
−5
0 10 20 30 40 50 60
t, sec
10
θ θm
θ, deg
−10
0 10 20 30 40 50 60
t, sec
10
q
q, deg/sec
q m
−10
0 10 20 30 40 50 60
t, sec
0 0
LS Gradient RLS
−5 −5
δ , deg
δ deg
e
e
−10 −10
−15 −15
0 20 40 60 0 20 40 60
t, sec t, sec
0 20
MRAC Neural Network MRAC
−5 10
δ , deg
δ , deg
−10 0
e
e
−15 −10
−20 −20
0 20 40 60 0 20 40 60
t, sec t, sec
To illustrate the issue of robustness and show that the RLS adaptive controller
is actually better able to handle time delay or unmodeled dynamics than the least-
squares gradient adaptive controller or MRAC, a numerical evidence of the time-
delay margin is computed for each of the four adaptive controllers using a time step
of 0.001 s. The results are shown in Table 10.1.
The RLS adaptive controller has the best time-delay margin than the other three
adaptive controllers. The standard MRAC has very poor robustness which is a well-
known fact [17]. Generally, the standard MRAC has to be modified to improve
robustness using the projection method or various modification techniques, such
as the σ modification [18], e modification [3], optimal control modification [19],
and adaptive loop recovery modification [20]. The time-delay margin for the neural
network MRAC varies with the initialization of the neural network weights. It should
be noted that both the predictor-based least-squares adaptive controllers suffer a
significant reduction in the time-delay margin by a factor of more than 2.
10.4 Recursive Least-Squares and Neural Network Pitch … 373
The aircraft responses due to a 0.020 s time delay for the four adaptive controllers
are shown in Figs. 10.24, 10.25, 10.26, and 10.27. As shown, the least-squares gra-
dient adaptive controller maintains a very good tracking performance with a 0.020 s
time delay. The MRAC and neural network MRAC exhibit high-frequency oscilla-
tions. The RLS adaptive controller exhibits low-frequency transients as it is much
more robust than the other three adaptive controllers. Thus, overall, the least-squares
gradient adaptive controller seems to perform the best among all of the adaptive
controllers.
5
α, deg
0
α α
m
−5
0 10 20 30 40 50 60
t, sec
5
θ θ
m
θ, deg
−5
0 10 20 30 40 50 60
t, sec
10
qm
q, deg/sec
q
0
−10
0 10 20 30 40 50 60
t, sec
Fig. 10.24 Aircraft response with least-squares gradient with 0.020 s time delay
10
α, deg
0
α αm
−10
0 10 20 30 40 50 60
t, sec
10
θ θm
θ, deg
−10
0 10 20 30 40 50 60
t, sec
10
q
q, deg/sec
q m
−10
0 10 20 30 40 50 60
t, sec
Fig. 10.25 Aircraft response with RLS with 0.020 s time delay (η = 0.2)
374 10 Aerospace Applications
10
α, deg
0
α αm
−10
0 10 20 30 40 50 60
t, sec
10
θ θ
m
θ, deg
0
−10
0 10 20 30 40 50 60
t, sec
100
q
q, deg/sec
q m
−100
0 10 20 30 40 50 60
t, sec
Fig. 10.26 Aircraft response with MRAC with 0.020 s time delay
10
α, deg
0
α α
m
−10
0 10 20 30 40 50 60
t, sec
10
θ θ
m
θ, deg
−10
0 10 20 30 40 50 60
t, sec
50
qm
q, deg/sec
q
0
−50
0 10 20 30 40 50 60
t, sec
Fig. 10.27 Aircraft response with neural network MRAC with 0.020 s time delay
reduce the fuel consumption. Structural flexibility of airframes can also cause signif-
icant aeroelastic interactions that can degrade vehicle stability margins, potentially
leading to degraded flying qualities. There exists a trade-off between the desire of
having lightweight, flexible structures for weight savings and the need for maintain-
ing sufficient robust stability margins in consideration of flight dynamics, stability,
and control.
It is of interest to examine the effect of aeroelasticity on the short-period mode
of an aircraft. The aeroelastic effect on an aircraft is assumed to be contributed only
by the flexible wing structures. Aeroelastic contributions by the fuselage and tail
empennage are assumed to be negligible. For simplicity, only the first symmetric
bending mode (1B) and first symmetric torsion mode (1T) are considered. Quasi-
steady aerodynamics assumption is invoked to simplify the model. The coupled
aeroelastic flight dynamic model of an aircraft that accounts for interactions between
wing bending and torsion on aircraft performance and stability in the pitch axis can
be expressed in the following state-space form [21]:
⎡ C L q c̄
CLw
⎡ ⎤ − 1 CL 1 mV∞
q∞ S − 2V∞ − 1 1
α̇ ⎢ m αα α C m αα C m αα c̄
⎢ q̇ ⎥ ⎢ 1 m qα
Cm α + m αα 1 c̄
mq m qα C L q c̄
+ m αα 2V∞ − mV ∞ 1 m w1
+
m qα C L w1
⎢ ⎥ ⎢ CLα
⎢ ẇ1 ⎥ ⎢
m qq m qq 2V∞ q∞ S m qq c̄ m αα c̄
⎢ ⎥ ⎢ 0 0 0
⎢ θ̇1 ⎥ = ⎢
⎢ ⎥ ⎢ 0 0 0
⎣ ẅ1 ⎦ ⎢ ⎢
⎣ − m w1 w h w1 α − m w1 w h w1 q − m w1 w kw1 w1
θ̈1 1 1 1 1 1 1
− mθ 1
h θ1 α − mθ 1
h θ1 q − m θ1 θ kθ1 w1
1 θ1 1 θ1 1 1
C L ẇ C L c̄ ⎤
θ̇1
− m1ααC L θ1 − 1 1
− 1 ⎡ ⎤
m αα V̄∞ C mc̄αα 2V̄∞ C c̄ ⎥ α
Cm ẇ m qα C L ẇ1 m θ̇ m qα L θ̇1 ⎥
1 m qα
Cm θ1 + m αα C L θ1 1 1
+ m αα 1 1
+ m αα 2V̄ ⎥ ⎢ q⎥
m qq m qq V̄∞ V̄∞ m qq 2 V̄∞ ⎥⎢ ⎥
∞
⎥⎢w ⎥
0 1 0 ⎥⎢ ⎥
⎥⎢ θ ⎥
0 0 1 ⎥⎢ ⎥
⎥ ⎣ ẇ ⎦
− m w1 w kw1 θ1 − m w1 w cw1 w1 − m w1 w cw1 θ1 ⎥
1 1 1 1 1 1
⎦ θ̇
− mθ 1
k θ1 θ1 − m θ θ c θ 1 w1 1
− m θ1 θ cθ1 θ1
1 θ1 1 1 1 1
⎡ ⎤
− 1 C − 1 C
m αα L δe m αα L δ
⎢ 1 m qα m qα ⎥
⎢ m qq Cm δe + m αα C L δe 1
Cm δ + m αα CLδ ⎥
⎢ m qq ⎥
⎢ 0 0 ⎥ δe
+⎢
⎢
⎥
⎥ δf
⎢ 0 0 ⎥
⎢ 1 ⎥
⎣ 0 g
m w1 w1 w1 δ1 ⎦
0 − m w1 w gθ1 δ1
1 1
(10.71)
where δ f (t) is a symmetric flap control surface deflection on the wing and m αα , m qα ,
and m qq are defined as
m V̄∞ C L α̇ c̄
m αα = + (10.72)
q∞ S 2 V̄∞
Cm α̇ c̄
m qα = − (10.73)
2 V̄∞
376 10 Aerospace Applications
IY Y
m qq = (10.74)
q∞ S c̄
The subscripts w1 and θ1 denote the first bending and first torsion quantities,
respectively. The symbols m, c, and k denote the generalized mass, generalized
damping which includes both structural and aerodynamic damping, and generalized
stiffness, respectively. The subscripts α and q denote the angle of attack and pitch
rate, respectively. The symbols h and g denote the generalized forces acting on the
wing structure due to the rigid aircraft states α (t) and q (t) and due to the symmetric
flap control surface deflection δ f (t), respectively.
The eigenvalues of the short-period mode of the rigid aircraft can be computed
from the 2 × 2 upper left matrix partition of the A matrix. These eigenvalues are
stable as shown
λ S P = −0.8580 ± 1.5380i
The eigenvalues of the flexible aircraft are also stable, but with a reduced damping
in the 1 T mode, as shown
λ S P = −0.5077 ± 0.5229i
The computed frequencies and damping ratios of the short-period mode, and the
1B and 1 T modes with 50% fuel remaining are shown in Table 10.2. The short-period
mode can be seen to be significantly affected by the aeroelastic coupling of the rigid
10.5 Adaptive Control of Flexible Aircraft 377
aircraft flight dynamics with the 1B and 1 T modes. The frequency of the coupled
short-period mode is significant reduced, but the damping increases.
Table 10.2 Frequencies and damping ratios at Mach 0.8 and 30,000 ft for flexible aircraft
Mode Short period 1B 1T
Uncoupled frequency, 1.761 8.4641 15.3337
rad/s
Coupled frequency, 0.7288 8.9648 15.2424
rad/s
Uncoupled damping 0.4872 0.2476 0.1433
ratio
Coupled damping ratio 0.6966 0.3556 0.0954
The frequencies and damping ratios as a function of the airspeed at the same
altitude of 30,000 ft are shown in Figs. 10.28 and 10.29. Generally, the frequencies
of the short-period mode and 1B mode increase with increasing the airspeed, while
the frequency of the 1 T mode decreases precipitously with increasing the airspeed.
The divergence speed is the airspeed at which the torsion modal frequency becomes
zero.
The damping ratios for both the short-period mode and 1B mode generally increase
with increasing airspeed. The damping ratio for the 1 T mode increases with increas-
ing the airspeed up to Mach 0.7 and thereafter decreases rapidly. The flutter speed is
the airspeed at which the damping ratio of any aeroelastic modes becomes zero. It is
apparent that the 1 T mode would exhibit a zero damping at a flutter speed of about
Mach 0.85. The low damping ratio of the 1 T mode can be a problem for aircraft sta-
bility. Uncertainty in aeroelastic properties can adversely affect the performance and
stability of a flexible aircraft. Active feedback control can potentially help improve
the stability margins of the aeroelastic modes.
25
Short−Period Mode
1st Bending Mode
1st Torsion Mode
20
Frequency, rad/sec
15
10
0
0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9
Mach Number
0.8
Short−Period Mode
0.7 1st Bending Mode
1st Torsion Mode
0.6
0.5
Damping Ratio
0.4
0.3
0.2
0.1
−0.1
−0.2
0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9
Mach Number
xr = C x (10.76)
where x (t) ∈ Rn is a state vector that is composed of a rigid aircraft state vector
xr (t) ∈ Rnr and an elastic wing state vector xe (t) ∈ Rn e =n−nr , u (t) ∈ Rm is a
control vector, A ∈ Rn × Rn and B ∈ Rn × Rm are constant and known matrices,
Θ ∗ ∈ R p×m is a constant and unknown matrix that represents a matched parametric
uncertainty in the rigid aircraft state, and Φ (xr ) ∈ R p is a vector of known regressor
functions of the rigid aircraft states.
Assuming that there is a sufficient frequency separation between the “slow” rigid
aircraft dynamics and “fast” elastic wing dynamics, then the fast and slow dynamics
can be decoupled using the standard singular perturbation method. The fast dynamics
of the elastic wing modes are assumed to approach the equilibrium solution infinitely
fast. Therefore, the rigid aircraft dynamics with approximately zero-order elastic
wing dynamics can be obtained by setting ẋe (t) = ε (x), where ε (x) is a small
parameter [22]. Thus,
ẋr Arr Ar e xr Br
= + u + Θ ∗ Φ (xr ) (10.77)
ε Aer Aee xe Be
10.5 Adaptive Control of Flexible Aircraft 379
where
A p = Arr − Ar e A−1
ee Aer (10.80)
B p = Br − Ar e A−1
ee Be (10.81)
Δ (x) = Ar e A−1
ee ε (x) (10.82)
The term Δ (x) represents the effect of unmodeled dynamics of the elastic wing
modes. The reduced-order plant matrix A p is assumed to be Hurwitz. Otherwise, an
output feedback adaptive control design would be necessary.
The objective is to design an adaptive control that enables the rigid aircraft state
vector xr (t) to tracks a reference model
ẋm = Am xm + Bm r (10.83)
where Am ∈ Rnr ×Rnr is a known Hurwitz matrix, Bm ∈ Rnr ×Rr is a known matrix,
and r (t) ∈ Rr is a piecewise continuous bounded reference command vector.
The adaptive controller is designed with
u = K x xr + K r r − Θ Φ (xr ) (10.84)
where Θ (t) is an estimate of Θ ∗ and it is assumed that K x and K r can be found such
that the following model matching conditions are satisfied:
A p + B p K x = Am (10.85)
B p K r = Bm (10.86)
Defining the tracking error as e (t) = xm (t) − xr (t), then the tracking error
equation becomes
ė = Am e + B Θ̃ Φ (xr ) − Δ (x) (10.87)
Θ̇ = −Γ Φ (xr ) e P B (10.88)
is not robust. As the adaptive gain Γ increases, the adaptive law becomes increasingly
sensitive to the unmodeled dynamics Δ (x) that can lead to instability [17].
To improve robustness to unmodeled dynamics, we use the optimal control modifi-
cation adaptive law to estimate the unknown parameter Θ ∗ [19]. The optimal control
modification adaptive law is given by
Θ̇ = −Γ Φ (xr ) e P B − νΦ (xr ) Φ (xr ) Θ B P A−1
m B (10.89)
P Am + A
m P = −Q (10.90)
Alternatively, the adaptive loop recovery modification adaptive law [20] can be
used to adjust Θ (t) as
dΦ (xr ) dΦ (xr )
Θ̇ = −Γ Φ (xr ) e P B + η Θ (10.91)
d xr d xr
Example 10.6 The reduced-order model of the flexible aircraft in Example 10.5
that retains only the rigid aircraft states is given by
α̇ −0.2187 0.9720 α −0.0651 α
= + δe + θα∗ θq∗
q̇ −0.4053 −0.8913 q −3.5277 q
Δα (x) f (t)
+ + α
Δq (x) f q (t)
The time-varying functions f α (t) and f q (t) are disturbances due to a moderate
vertical wind gust modeled by the Dryden turbulence model [14] with a vertical
velocity amplitude of about 10 ft/s and a pitch rate amplitude of 1.5 deg/s as shown
in Fig. 10.30.
A desired reference model of the pitch attitude is given by
where ζm = 0.85 and ωm = 1.5 rad/s are chosen to give a desired handling charac-
teristic.
Let xr (t) = α (t) θ (t) q (t) , u (t) = δe (t), and Θ ∗ = θα∗ 0 θq∗ .
A nominal
as ū (t) = K x x (t) + kr r (t) where K x =
controller is designed
− b13 a31 ωn2 2ζ ωn + a33 = − 0.1149 0.6378 0.4702 and kr = b13 ωn2 =
−0.6378. The closed-loop eigenvalues are −0.2112 and −1.2750 ± 0.7902i. The
nominal closed-loop plant is then chosen to be the reference model as
⎡ ⎤ ⎡ ⎤⎡ ⎤ ⎡ ⎤
α̇m −0.2112 −0.0415 0.9414 αm 0.0415
⎣ θ̇m ⎦ = ⎣ 0 0 1 ⎦ ⎣ θm ⎦ + ⎣ 0 ⎦ r
q̇m 0 −2.2500 −2.5500 qm 2.2500
ẋm Am xm Bm
0.5
w /V , deg
0
∞ a
−0.5
−1
0 20 40 60 80 100
t, sec
1
qa, deg/sec
−1
−2
0 20 40 60 80 100
t, sec
An adaptation rate Γ = 100I is used for the adaptive controller with the input
function chosen as Φ (xr ) = 1 α θ q whereby the bias input is needed to handle
the time-varying wind gust disturbances.
For the optimal control modification, the modification parameter is set to ν =
0.2. For the adaptive loop recovery modification, the modification parameter is set
to η = 0.2. Also, the Jacobian of the input function dΦ (xr ) /d xr is simply an
382 10 Aerospace Applications
identity matrix, thereby making the adaptive loop recovery modification effectively
a σ modification adaptive law [18].
A pitch attitude doublet is commanded. The response of the flexible aircraft with-
out adaptive control is shown in Fig. 10.31. It is clear that the aircraft response does
not track well with the reference model. Moreover, the high-frequency oscillations
in the pitch rate response due to the aeroelastic interactions with the bending and
torsion modes are highly objectionable from the pilot handing quality perspective.
5
α α
m
α, deg
−5
0 20 40 60 80 100
t, sec
5
θ θm
θ, deg
−5
0 20 40 60 80 100
t, sec
5
qm
q, deg/sec
q
0
−5
0 20 40 60 80 100
t, sec
5
α α
m
α, deg
−5
0 20 40 60 80 100
t, sec
5
θ θ
m
θ, deg
−5
0 20 40 60 80 100
t, sec
5
qm
q, deg/sec
q
0
−5
0 20 40 60 80 100
t, sec
Fig. 10.32 Longitudinal response of flexible aircraft with standard MRAC (Γ = 100I )
5
α α
m
α, deg
0
−5
0 20 40 60 80 100
t, sec
5
θ θ
m
θ, deg
0
−5
0 20 40 60 80 100
t, sec
5
q
q, deg/sec
q m
0
−5
0 20 40 60 80 100
t, sec
Fig. 10.33 Longitudinal response of flexible aircraft with optimal control modification (OCM)
(Γ = 100I , ν = 0.2)
5
α α
m
α, deg
−5
0 20 40 60 80 100
t, sec
5
θ θm
θ, deg
−5
0 20 40 60 80 100
t, sec
5
q
q, deg/sec
q m
0
−5
0 20 40 60 80 100
t, sec
Fig. 10.34 Longitudinal response of flexible aircraft with adaptive loop recovery modification
(ALR) (Γ = 100I , η = 0.2)
In contrast, Fig. 10.33 shows that the optimal control modification adaptive law
is able to suppress the large initial transient of the pitch rate and the amplitude of
the high-frequency oscillations. The response of the aircraft due to the adaptive loop
recovery modification adaptive law as shown in Fig. 10.34 is very similar to that with
the optimal control modification adaptive law.
The aeroelastic wing tip bending deflection and torsional twist are shown in
Figs. 10.35 and 10.36 for the four different controllers: baseline nominal controller,
standard MRAC, optimal control modification, and adaptive loop recovery modi-
fication. The aeroelastic wing is modeled to be rather flexible to demonstrate the
aeroelastic effects on adaptive control. The rigid aircraft pitch attitude command
and wind gust result in a bending deflection amplitude of 5 ft and a torsional twist
amplitude of about 3 deg at the wing tip. The aeroelastic deflections are quite signif-
icant since the flight condition at Mach 0.8 is approaching the flutter speed at Mach
384 10 Aerospace Applications
0.85. It is noted that the standard MRAC results in a very large initial transient of
the torsional twist. This large torsional twist is clearly not realistic and in practice
would result in excessive wing loading and wing stall. These effects are not taken
into account in the simulations. Nonetheless, this illustrates the undesirable behavior
of the standard MRAC in the flight control implementation for flexible aircraft.
10 10
Wing Tip Deflection, ft
0 0
−5 −5
Baseline MRAC Γ=100
−10 −10
0 50 100 0 50 100
t, sec t, sec
10 10
Wing Tip Deflection, ft
0 0
−5 −5
OCM Γ=100 ν=0.2 ALR, Γ=100 η=0.2
−10 −10
0 50 100 0 50 100
t, sec t, sec
Figure 10.37 is the plot of the elevator deflections for the four controllers. The
standard MRAC produces a significant control saturation during the initial transient.
This saturation leads to an undesirable rigid aircraft response and aeroelastic deflec-
tions. Both the optimal control modification and adaptive loop recovery modification
adaptive laws produce quite similar elevator deflections. Also, it is observed that the
elevator deflection for the baseline controller has a lower-frequency content than
those corresponding to the optimal control modification and adaptive loop recovery
modification adaptive laws. This indicates that the adaptive controllers effectively
account for the aeroelastic wing dynamics in the elevator deflection commands to
suppress the responses of the aeroelastic wing modes, whereas the baseline controller
has no knowledge of the effect of the aeroelastic wing dynamics.
This study shows that adaptive control can be used to accommodate uncertainty for
flexible aircraft. The effect of aeroelasticity is captured in the reduced-order model
as unmodeled dynamics. The results show that the standard MRAC is neither suffi-
ciently robust nor able to produce well-behaved control signals. Excessive torsional
twist and control saturation due to the standard MRAC are noted. Both the optimal
control modification and adaptive loop recovery modification adaptive laws are seen
to be more effective in reducing the tracking error while maintaining the aeroelastic
deflections to within reasonable levels.
10.6 Adaptive Linear Quadratic Gaussian Flutter Suppression Control 385
4 10
4 4
Wing Tip Twist, deg
0 0
−2 −2
OCM Γ=100 ν=0.2 ALR, Γ=100 η=0.2
−4 −4
0 50 100 0 50 100
t, sec t, sec
4 20
2 10
δ , deg
δe, deg
0 0
e
−2 −10
Baseline MRAC Γ=100
−4 −20
0 50 100 0 50 100
t, sec t, sec
10 10
5 5
δe, deg
δe, deg
0 0
−5 −5
OCM Γ=100 ν=0.2 ALR, Γ=100 η=0.2
−10 −10
0 50 100 0 50 100
t, sec t, sec
In the recent years, NASA has developed a new type of aircraft flight control sur-
faces called the Variable Camber Continuous Trailing Edge Flap (VCCTEF) [23–25].
The VCCTEF, illustrated in Fig. 10.38, is a wing shaping control device designed to
reshape a flexible aircraft wing in-flight to improve the aerodynamic performance
while suppressing any adverse aeroelastic interactions. It employs three chordwise
flap segments to provide a variable camber to change the wing shape for increas-
ing the aerodynamic performance. The flap is also made up of individual spanwise
sections which enable different flap settings at each flap spanwise position. This
results in the ability to control the wing shape as a function of the wing span, thereby
resulting in a change to the wing twist to establish the best lift-to-drag ratio at any
aircraft gross weight or mission segment. The individual spanwise flap sections are
connected with a flexible elastomer material to form a continuous trailing edge with
no flap gaps in the wing planform for drag and noise reduction purposes.
Consider a notional aircraft equipped with highly flexible wings. The flexibility
of modern transport wings can cause a reduction in the flutter margin which can
compromise the stability of an aircraft. A flutter analysis is conducted to examine
the effect of the increased flexibility of a conventional aircraft wing based on a NASA
Generic Transport Model (GTM) [26]. The baseline stiffness of the GTM wing is
reduced by 50%. This configuration is referred to as the flexible wing GTM. The
flutter analysis computes the aeroelastic frequencies and damping ratios over a range
of the airspeed as shown in Figs. 10.39 and 10.40 [27]. When the damping ratio of
an aeroelastic mode encounters a zero crossing, a flutter speed is noted. The lowest
flutter speed establishes a flutter boundary.
10.6 Adaptive Linear Quadratic Gaussian Flutter Suppression Control 387
12
10
Frequency, Hz
8
0
0 0.5 1 1.5
Mach @ 35K ft
0.8
0.6
Aerodynamic Damping
0.4
0.2
−0.2
−0.4
0 0.5 1 1.5
Mach @ 35K ft
Table 10.3 shows the flutter speed prediction at 35,000 ft for the stiff wing GTM
and the flexible wing GTM. The critical flutter mode is an antisymmetric mode. The
flutter boundary is significantly reduced by 31% with the flexible wing GTM.
The FAA (Federal Aviation Administration) requires the aircraft certification to
demonstrate a flutter margin of at least 15% above the dive speed which is normally
determined from flight testing. For a maximum operating Mach 0.8, the dive speed
may be estimated to be about 15% over the maximum operating Mach, or Mach 0.92.
Thus, the flutter clearance for this notional aircraft would require a minimum flutter
speed of Mach 1.06 at 35,000 ft. The stiff wing GTM meets this flutter clearance,
but the flexible wing GTM does not. To increase the flutter speed, active flutter sup-
pression is an option. Currently, active flutter suppression has not been certified for
transport aircraft, but this situation may change as the FAA has begun to investigate
388 10 Aerospace Applications
certification requirements for active flutter suppression control for commercial trans-
port aircraft.
y = C1 η + C2 μ + D1 δ + D2 δ̇ + D3 δ̈ (10.94)
where μ (t) is a fast state vector that contains the generalized aeroelastic states
for high-frequency aeroelastic modes, η (t) is a slow state vector that contains the
generalized aeroelastic states for lower-frequency aeroelastic modes as well as the
rigid aircraft states, δ (t) is the control surface deflection vector, and y (t) is the output
vector.
Note that the aeroelastic states are dependent on the velocity and acceleration of
the control surface deflection vector δ (t). This dependency is usually neglected in
rigid aircraft flight dynamics, but can be significant in aeroservoelasticity especially
for flutter suppression.
The quality of fast and slow states can be examined by the eigenvalues of the
partitioned matrices A11 and A22 . Since μ (t) is a fast state vector, then we can write
ε
A11
<
A22
, where ε is a small quantity.
In general, an ASE state-space model contains both the rigid aircraft modes, which
usually have low frequencies, and the aeroelastic modes which are at much higher
frequencies than the rigid aircraft modes. Flutter modes are usually associated with
those aeroelastic modes in the low-frequency range. As a result, high-frequency
aeroelastic modes normally do not participate in a flutter response. In a control
design, it is usually expedient to remove high-frequency modes in order to simplify a
controller design. By removing the high-frequency aeroelastic modes, the ASE state-
space model is reduced in the model order. We will employ the singular perturbation
method to reduce the order of an ASE state-space model [28]. The advantage of the
singular perturbation method is that the physical modes are properly retained for a
flutter suppression control design.
Using the singular perturbation approach, the fast and slow dynamics can be
decoupled. To decouple the fast and slow states, we perform a time scale separation
10.6 Adaptive Linear Quadratic Gaussian Flutter Suppression Control 389
by applying the singular perturbation method. Toward that end, we consider the usual
slow or stretched time transformation
τ = εt (10.95)
dμ
ε = A21 η + A22 μ + B21 δ + B22 δ̇ + B23 δ̈ (10.97)
dτ
The Tikhonov’s theorem can be used to approximate the solution of the singularly
perturbed system with the solution of a “reduced-order” system by setting ε = 0 [22].
Thus, the reduced-order system is given by
where η0 (τ ) and μ0 (τ ) are the “outer” solution of the singularly perturbed system.
The term “outer” is in connection with the concept of “inner” or “boundary layer”
and “outer” solutions which have the origin in the boundary layer theory due to
Prandtl. The “inner” or “boundary layer” solution for this system is obtained from
η = η0 + ηi − η M AE (10.102)
μ = μ0 + μi − μ M AE (10.103)
where η M AE (t) and μ M AE (t) are correction terms by a matched asymptotic expan-
sion method applied to both the inner and outer solutions [29]. The outer solution is
in fact the asymptotic solution of the original system as t → ∞.
Since the asymptotic behavior of a closed-loop system is an important consider-
ation for stability implication, the outer solution of the singularly perturbed system
is of significant importance. Thus, we obtain the outer solution as the reduced-order
model using only the outer solution of the slow state vector η0 (t) as
390 10 Aerospace Applications
⎛ ⎞
⎜ ⎟
η̇0 = ⎝ A11 − A12 A−1 A21 ⎠ η0
22
Ā11
⎡δ⎤
B11 − A12 A−1 −1 −1
22 B21 B12 − A12 A22 B22 B13 − A21 A22 B23 ⎣ ⎦
+
δ̇ + Δη̇0
B̄11 B̄12 B̄13 δ̈
(10.104)
⎛ ⎞ ⎛ ⎞ ⎛ ⎞
⎜ ⎟ ⎜ ⎟ ⎜ ⎟
y = ⎝C1 − C2 A−1 A21 ⎠ η0 + ⎝ D1 − C2 A−1 B21 ⎠ δ + ⎝ D2 − C2 A−1 B22 ⎠ δ̇
22
22
22
C̄1 D̄1 D̄2
⎛ ⎞
⎜ ⎟
+ ⎝ D3 − C2 A−1 B23 ⎠ δ̈ + Δy (10.105)
22
D̄3
We will use the slow state vector η0 (t) as the approximation of the actual slow
state vector η (t).
Now, consider a simplified second-order actuator model
ẋ = Ax + Bu + Δ (10.109)
y = C x + Du (10.110)
Example 10.7 To illustrate the model reduction, the ASE state-space models with
22 modes at various flight conditions are developed for the flexible wing GTM roll
dynamics coupled to the wing antisymmetric modes [30]. The open-loop ASE model
has two unstable modes at Mach 0.86 and an altitude of 10,000 ft. Using the model
reduction method above, it can be shown in Table 10.4 that a reduced-order model
using only the first 8 modes can capture all the unstable modes and approximate the
first 6 modes of the higher-order ASE model quite well [28].
Table 10.4 Antisymmetric modes of ESAC wing @ Mach 0.86 and Altitude 10,000 ft
Mode n = 6 n=7 n=8 n = 22 (Full)
Rigid −2.7392 −2.7395 −2.7385 −2.7385
1 2.7294 ± 19.8683i 2.7512 ± 19.8529i 2.7804 ± 19.8561i 2.7842 ± 19.8513i
2 −0.1553 ± 24.3565i −0.1557 ± 24.3562i −0.1547 ± 24.3553i −0.1549 ± 24.3552i
3 −6.3434 ± 24.0892i −6.3272 ± 24.0739i −6.4220 ± 23.9949i −6.4174 ± 23.9920i
4 −0.3902 ± 37.1580i −0.3782 ± 37.1461i 0.0571 ± 37.4423i 0.0584 ± 37.4846i
5 −20.0160 ± −20.2813 ± −20.4217 ± −20.4833 ±
32.3722i 32.3013i 32.4999i 32.5445i
Suppose we introduce an uncertainty into the nominal ASE state-space model as
y = C x + Du (10.112)
where A ∈ Rn × Rn , B ∈ Rn × Rm , C ∈ R p × Rn , D ∈ R p × Rm with p ≥ m,
ΔA = δ A A, and ΔB = δ B B are known perturbations of the A and B matrices, and
δ A and δ B are assumed to be small parameters that represent multiplicative model
variations.
Observer state feedback control based on output measurements is generally used
for flutter suppression control since the full-state information is not measurable. The
idea is to design an adaptive augmentation controller that is robust to the perturbation
in the nominal plant model due to the model variation. Toward that end, we assume
the pair (A, B) is controllable, and the pair (A, C) is observable.
In general, output feedback adaptive control of MIMO systems can be quite dif-
ficult. The SPR condition for a system transfer function matrix is required as is the
case with a SISO system for stable adaptive control. In general, the system transfer
function matrix from u (t) to y (t) can be non-square, hence non-SPR, if the number
of inputs and the number of outputs are different. The Kalman–Yakubovich lemma
[31] can be used to determine whether the transfer function G (s) = C (s I − A)−1 B
is SPR by the following conditions:
392 10 Aerospace Applications
P A + A P = −Q (10.113)
P B = C (10.114)
B P B = B C = C B > 0 (10.115)
where x̂ (t) is the observer state vector, L is the Kalman filter gain, and ŷ (t) is the
observer output given by
ŷ = C x̂ + Du (10.117)
where
A + B K x∗ = Am (10.121)
B K r∗ = Bm (10.122)
B K y∗ = −L (10.123)
Let e p (t) = x (t) − x̂ (t) be the state estimation error. Then, the state estimation
error equation is computed as
ė p = A p e p + ΔA e p + x̂ + ΔB K x x̂ + K y Ce p + K r r (10.126)
where A p = A − LC is Hurwitz.
394 10 Aerospace Applications
Let e (t) = xm (t) − x̂ (t) be the tracking error, then the tracking equation is
expressed as
ė = Am e − B K̃ x x̂ − B K̃ y Ce p − B K̃ r r (10.127)
The adaptive laws for K x (t), K y (t), and K r (t) using the optimal control modi-
fication [28] are given by
K̇ x = Γx x̂ e P + νx x̂ K x B P A−1
m B (10.128)
K̇ y = Γ y y − ŷ e P + ν y y − ŷ K y B P A−1
m B (10.129)
K̇ r = Γr r e P + νr r K r B P A−1
m B (10.130)
P Am + A
mP + Q = 0 (10.131)
where Q = Q > 0.
As discussed in Sect. 9.14, if the reference model is SPR and the plant is non-
SPR, then an adaptive control design with the standard MRAC is infeasible unless
the reference model is modified to be non-SPR from the ideal control design for
the non-SPR plant. The optimal control modification has been shown to be able
to handle non-SPR plants with either SPR or non-SPR reference models. For an
adaptive augmentation design, the robust baseline controller can also serve to reduce
the sensitivity of the closed-loop plant to the non-minimum phase behavior if a robust
modification or the projection method is used.
Consider an adaptive augmentation regulator design with the following adaptive
controller:
u = K̄ x x̂ + ΔK x x̂ + K y y − ŷ (10.132)
where K̄ x is obtained from the Linear Quadratic Regulator (LQR) design of the
full-state equation, and the adaptive laws for ΔK x (t) and K y (t) are given by
Δ K̇ x = −Γx x̂ x̂ P − νx ΔK x B P A−1
m B (10.133)
K̇ y = −Γ y y − ŷ x̂ P − ν y y − ŷ K y B P A−1
m B (10.134)
$ $2 $ $2 $ $
$2
$
V̇ x̂, e p , Δ K̃ x , K̃ y ≤ − c1 $x̂ $ − νx c2 $x̂ $ $Δ K̃ x $
$ $2 $$ $ $ $2
$
+ 2νx c2 c3 $x̂ $ $Δ K̃ x $ − (c4 − 2c7 ) $e p $
$ $2 $ $ $2
$ $ $2 $ $ $
$ $ $$ $
− ν y c5 $e p $ $ K̃ y $ + 2ν y c5 c6 $e p $ $ K̃ y $ + 2c8 $x̂ $ $e p $
$ $$ $$ $
$
$ $ $2 $ $ $
$
+ 2c9 $x̂ $ $e p $ $Δ K̃ x $ + 2c10 $e p $ $ K̃ y $ (10.138)
B P A−1
m B
ΔK x
∗
where c1 = λmin (Q), c2 = λmin B A− −1
m Q Am B , c3 = c2
, c4 =
m B
K y
B P A−1 ∗ $ $
λmin (R), c5 = c2
C
, c6 =
2
c5
, c7 = $W ΔA p $, c8 =
W ΔAm
,
c9 =
W ΔB
, and c10 = c9
C
.
Using the inequality 2
a
b
≤
a
2 +
b
2 , we obtain
396 10 Aerospace Applications
$ $2 $ $2 $$
$2
$
V̇ x̂, e p , Δ K̃ x , K̃ y ≤ − (c1 − c8 ) $x̂ $ − (νx c2 − c9 ) $x̂ $ $Δ K̃ x $
$ $2 $$ $ $ $2
$
+ 2νx c2 c3 $x̂ $ $Δ K̃ x $ − (c4 − 2c7 − c8 − c9 ) $e p $
$ $2 $ $ $2
$ $ $2 $$ $
$
− ν y c5 $e p $ $ K̃ y $ + 2 ν y c5 c6 + c10 $e p $ $ K̃ y $
(10.139)
νx c2 − c9 > 0 (10.142)
2
ν y c5 c6 + c10
c4 − 2c7 − c8 − c9 − >0 (10.143)
ν y c5
Then, V̇ x̂, e p , Δ K̃ x , K̃ y ≤ 0. Therefore, the adaptive regulator design is stable.
Note that this adaptive augmentation controller is designed to tolerate the uncer-
tainty due to the plant model variation, but does not attempt to accommodate the
uncertainty by canceling the uncertainty. The goal is to enable the adaptive control
design to have a sufficient robustness due to the plant model variation. In the event
the plant model variation is zero, then c7 = 0, c8 = 0, c9 = 0, and c10 = 0. This
implies νx ≤ ccc1 2 and ν y ≤ ccc4 2 . On the other hand, if the plant model variation is
2 3 5 6
large, it is possible that there are no suitable values of νx and ν y that can stabilize the
plant.
The Lyapunov stability analysis shows how the modification parameters νx and
ν y could be selected to ensure a stable adaptation. As discussed in Sect. 9.5.3, the
Lyapunov stability analysis often results in a high degree of conservatism that could
render the selection of the modification parameters infeasible. Therefore, in practice,
10.6 Adaptive Linear Quadratic Gaussian Flutter Suppression Control 397
the linear asymptotic property of the optimal control modification can be invoked to
find suitable values of νx and ν y to ensure the closed-loop stability. In the limit as
Γ → ∞, the following adaptive parameters tend to:
1 − −1
BΔK x → B B Am P B B P (10.144)
νx
1 − −1
BKy y → B B Am P B B P x̂ (10.145)
νy
Example 10.8 The ASE state-space model in Example 10.7 contains 45 states,
64 outputs, and 4 inputs. The states include the aircraft roll rate p (t) and two
generalized states for each of the 22 aeroelastic modes. The outputs include the
aircraft airspeed V (t), angle of attack α (t), sideslip angle β (t), aircraft angu-
lar rates p (t) q (t) r (t) , aircraft position x (t) y (t) h (t) , aircraft attitude
φ (t) θ (t) ψ (t) , accelerations in the three axes N x , N y , Nz at the forward and
aft locations of the wing tip, and the four hinge moments of the control surfaces.
The control surfaces are the four outboard third camber segments of the VCCTEF.
All these surfaces are blended together to provide a single input to address the
relative constraints imposed by the elastomer material. For the flutter suppression
398 10 Aerospace Applications
control design, only the two Nz acceleration measurements are used as the outputs.
Figure 10.41 illustrates the input and output locations [28].
All the control surfaces are not entirely independent in their motions due to the
physical constraints imposed by the elastomer material. This material has certain
displacement and rate limits. Thus, any adjacent control surfaces will also have
relative displacement and rate limits. These limits are in addition to the normal
position and rate limits that each of the control surfaces is subjected to. Thus, these
relative constraints can cause challenges in a control design of this system.
Consider the following relative displacement and rate constraints:
|δi+1 − δi | ≤ Δδ (10.149)
' '
'δ̇i+1 − δ̇i ' ≤ Δδ̇ (10.150)
where i = 1, 2, 3.
For the VCCTEF design, the relative motion between any adjacent flap sections
is allowed to be within two degrees. The rate constraint imposed by the elastomer
material is not yet defined and thus is assumed to be large. The actuator dynamics
are modeled as a second-order system. This actuator model is highly simplified since
it does not take into account the hinge moment which is a function of the states and
the dynamics of the elastomer material in the overall actuator model.
To address the relative displacement limit, a concept of virtual control is introduced
[36]. The control surface deflections are described by a shape function. This shape
function can be any reasonable shape function with a smooth and gradual slope.
One simple function is a linear function. The control surface deflections are then
parametrized as a linear function
iδv
δi = (10.151)
4
where i = 1, 2, 3, 4 such that δ1 (t) is the inboard flap and δ4 (t) is the outboard flap,
and δv (t) is the virtual control surface deflection.
Since the inboard flap section δ1 (t) cannot deflect more than two degrees relative
to the stationary flap adjacent to it, then δv (t) ≤ 8 deg. Also, the outboard flap
deflection δ4 (t) is the same as the virtual control surface deflection. Thus, one can
think that the outboard flap δ4 (t) is a master control input, while the other three
control surfaces are slave control inputs since their motions are dependent on the
master control input.
Thus, the virtual control derivatives are computed as
(
4
i Bi jk
B jk = (10.152)
i=1
4
where B jki is the control derivative of mode jth with respect to the displacement
(k = 1), velocity (k = 2), and acceleration (k = 3) of flap section ith.
10.6 Adaptive Linear Quadratic Gaussian Flutter Suppression Control 399
100
50
jω (seconds−1)
0
−50
−100
−150
−50 −40 −30 −20 −10 0 10 20 30 40 50
−1
σ (seconds )
400
300
200
jω (seconds−1)
100
−100
−200
−300
−400
−500
−100 −80 −60 −40 −20 0 20 40 60 80 100
−1
σ (seconds )
sensor noises for the baseline full-state feedback LQR controller and the output
feedback LQG controller with and without the adaptive augmentation controller. All
the controllers are able to maintain good performance in the presence of the process
and sensor noises.
Figures 10.50 and 10.51 show the root locus plots of the closed-loop transfer
functions of the accelerometers with the adaptive augmentation controller using the
final gain matrix. As can be seen, the closed-loop transfer functions are completely
stable. Figures 10.52 and 10.53 are the frequency response plots for the open-loop and
closed-loop transfer functions of the accelerometers. The largest frequency response
is due to mode 4. The closed-loop frequency response is significantly less than the
open-loop frequency response, indicating the effectiveness of the aeroelastic mode
suppressioncontroller.
Next, the plant model variation is introduced into the aeroservoelastic state-space
model by specifying ΔA = 0.05A and ΔB = −0.1B. The process and sensor
10.6 Adaptive Linear Quadratic Gaussian Flutter Suppression Control 401
0.12
LQR
0.1 LQG
LQG + Adaptive
0.08
0.06
0.02
−0.02
−0.04
−0.06
−0.08
0 2 4 6 8 10
t, sec
Fig. 10.44 Roll rate response for LQR and LQG controllers without process and sensor noises
7
LQR
6 LQG
LQG + Adaptive
5
Mode 1 Generalized Displacement
−1
−2
−3
0 2 4 6 8 10
t, sec
Fig. 10.45 Mode 1 generalized displacement response for LQR and LQG controllers without
process and sensor noises
1.5
LQR
LQG
LQG + Adaptive
1
Mode 4 Generalized Displacement
0.5
−0.5
−1
0 2 4 6 8 10
t, sec
Fig. 10.46 Mode 4 generalized displacement response for LQR and LQG controllers without
process and sensor noises
402 10 Aerospace Applications
0.12
LQR
0.1 LQG
LQG + Adaptive
0.08
0.06
0.02
−0.02
−0.04
−0.06
−0.08
0 2 4 6 8 10
t, sec
Fig. 10.47 Roll rate response for LQR and LQG controllers with process and sensor noises
7
LQR
6 LQG
LQG + Adaptive
5
Mode 1 Generalized Displacement
−1
−2
−3
0 2 4 6 8 10
t, sec
Fig. 10.48 Mode 1 generalized displacement response for LQR and LQG controllers with process
and sensor noises
1.5
LQR
LQG
LQG + Adaptive
1
Mode 4 Generalized Displacement
0.5
−0.5
−1
−1.5
0 2 4 6 8 10
t, sec
Fig. 10.49 Mode 4 generalized displacement response for LQR and LQG controllers with process
and sensor noises
10.6 Adaptive Linear Quadratic Gaussian Flutter Suppression Control 403
200
150
100
jω (seconds−1)
50
−50
−100
−150
−200
−250
−50 −40 −30 −20 −10 0 10 20 30 40 50
−1
σ (seconds )
400
300
200
jω (seconds−1)
100
−100
−200
−300
−400
−500
−100 −80 −60 −40 −20 0 20 40 60 80 100
σ (seconds−1)
Open−Loop
50 Closed−Loop
Magnitude (dB)
−50
1080
900 Open−Loop
Closed−Loop
Phase (deg)
720
540
360
180
0
−180
−2 0 2 4
10 10 10 10
ω (rad/s)
Open−Loop
50 Closed−Loop
Magnitude (dB)
0
−50
1080
Open−Loop
900 Closed−Loop
Phase (deg)
720
540
360
180
0
−2 0 2 4
10 10 10 10
ω (rad/s)
noises are also included. The output feedback LQG controller without the adaptive
augmentation controller is unstable as shown in Figs. 10.54, 10.55, and 10.56. On the
other hand, the adaptive augmentation controller is able to stabilize the aeroelastic
modes in the presence of the plant model variation. The closed-loop plant matrix
without the adaptive augmentation controller is in fact unstable. Stability of the
closed-loop plant is restored in the presence of the adaptive augmentation controller.
0.15
LQG
LQG + Adaptive
0.1
Roll Rate p, rad/sec
0.05
−0.05
−0.1
0 2 4 6 8 10
t, sec
Fig. 10.54 Roll rate response with and without adaptive augmentation controller
10.6 Adaptive Linear Quadratic Gaussian Flutter Suppression Control 405
20
LQG
LQG + Adaptive
15
−5
−10
−15
−20
0 2 4 6 8 10
t, sec
Fig. 10.55 Mode 1 generalized displacement response with and without adaptive augmentation
controller
2.5
LQG
2 LQG + Adaptive
1.5
Mode 4 Generalized Displacement
0.5
−0.5
−1
−1.5
−2
−2.5
0 2 4 6 8 10
t, sec
Fig. 10.56 Mode 4 generalized displacement response with and without adaptive augmentation
controller
Figure 10.57 is the plot of the time history of the virtual control command for
the output feedback LQG controller with and without the adaptive augmentation
controller. The largest amplitude of the stabilizing virtual control command for the
adaptive augmentation controller is 6.22o . The linear mapping between the virtual
control command and the physical control commands results in 1.56o which meets
the physical constraint of 2o on the relative deflection of the VCCTEF.
406 10 Aerospace Applications
6
LQG
LQG + Adaptive
4
−2
−4
−6
−8
0 2 4 6 8 10
t, sec
Fig. 10.57 Virtual control command with and without adaptive augmentation controller
Flight control is an important aircraft system that provides many critical functions to
enable safe operations of an aircraft. A pilot controls an aircraft by moving a control
input device such as a control yoke, rudder pedal, or side stick. A pilot command
signal is sent to a flight control system which processes the signal further and converts
it into an actuator command signal. This signal is then sent to an actuator system that
drives the position of a flight control surface such as the aileron, elevator, or rudder.
This flight control feature is called a pilot command tracking task. Another important
flight control feature is to enhance aircraft stability via a stability augmentation
system (SAS). The SAS provides the additional damping to the rigid aircraft modes
by feedback control. A typical SAS on an aircraft is a yaw damper.
Connections between control input devices to flight control surfaces can be made
via mechanical linkages which exist in older aircraft. Hydraulic flight control systems
use hydraulic fluid circuits in conjunction with partial mechanical flight control sys-
tems to drive flight control surfaces. Hydraulic flight control systems are still found
in many aircraft in operation nowadays. Modern aircraft usually employ fly-by-wire
(FBW) flight control systems which replace mechanical systems with electronics. A
pilot command is converted into an electrical signal transmitted to a flight control
computer for processing. The flight control computer computes actuator command
signals and transmits these signals to hydraulic actuators to drive flight control sur-
faces.
A typical flight control system consists of an inner-loop stability augmentation
system and an outer-loop autopilot system as illustrated in the Fig. 10.58. The com-
mand input into a flight control system can be accomplished via a pilot’s direct
mechanical input or an autopilot. When engaged, the autopilot generates an attitude
command of the attitude angles or an altitude command for a selected autopilot task
such as cruise or landing. This command is processed as a feedforward signal and
then is summed with the SAS to form a control signal. The SAS uses feedback of
10.7 Adaptive Flight Control 407
aircraft’s sensed angular rate variables, such as the roll rate, pitch rate, and yaw rate
to provide the additional damping to the aircraft dynamics. The control signal is
transmitted to flight control actuators that drive flight control surfaces to change the
aircraft dynamics.
For rate command attitude hold (ACAH) flight control, a first-order reference
model is used to filter the rate command r (t) into a reference acceleration command
ẋm (t) as
ẋm = −Ω (xm − r ) (10.155)
where Ω = diag ω p , ωq , ωr is a diagonal matrix of the reference model frequencies.
The nominal controller is then specified as
) t
ū = K p (xm − x) + K i (xm − x) dτ (10.156)
0
K2 p =2 diag
where 2ζ p ω p , 2ζq ωq , 2ζr ωr is the proportional gain matrix and K i =
diag ω p , ωq , ωr2 is the integral gain matrix [37].
For transport aircraft, typical values of the reference model frequencies ω p , ωq ,
and ωr are 3.5, 2.5, and 2.6, respectively; and typical values of the reference model
damping ratios ζ p , ζq , and ζr are √12 .
10.7 Adaptive Flight Control 409
The adaptive signal u ad (t) is designed to estimate the uncertainty in the aircraft
plant dynamics.
The desired acceleration is the ideal response of the nominal aircraft plant dynam-
ics. Therefore, the ideal aircraft dynamics can be expressed as
ẋd = C x + Du + E z (10.158)
The dynamic inversion controller then computes the actuator command from the
closed-loop plant model of the ideal aircraft dynamics [37]
u c = D −1 (ẋd − C x − E z) (10.159)
Substituting u c (t) for u (t) into the aircraft model, then the closed-loop plant
model of the aircraft is obtained as
Then,
) t
ẋ = ẋm + K p (xm − x) + K i (xm − x) dτ − Θ̃ Φ (x, u c , z) (10.162)
0
*t
Let e (t) = 0 xm (t) − x (t) dτ xm (t) − x (t) . Then,
ė = Am e + B Θ̃ Φ (x, u c , z) (10.164)
where
410 10 Aerospace Applications
0 I
Am = (10.165)
−K i −K p
0
B= (10.166)
I
Then, the adaptive law can be designed with any robust adaptive control scheme
such as the e modification
$ $
Θ̇ = −Γ Φ (x, u c , z) e P B + μ $e P B $ Θ (10.167)
Note that actuator dynamics can affect the closed-loop stability of an aircraft.
Therefore, a suitable selection of the modification parameter should be made to ensure
sufficient robustness. An actuator system must have a higher frequency bandwidth
than the frequency of a given mode in consideration. Roughly speaking, the frequency
bandwidth of an actuator should be at least several times greater than the highest
frequency of the aircraft modes. For example, the short-period mode is about 2.5 rad/s
and then the frequency of the actuator for the elevator could be about 10 times larger
or 25 rad/s. The damping ratio should be sufficiently large to provide a well-damped
control surface deflection.
For example, a second-order actuator system is modeled as
Consequently, the pilot would tend to apply a greater command to try to overcome
the lag in the actuator dynamics. This actuator lag and the pilot positive feedback can
result in an adverse consequence known as pilot-induced oscillations (PIO), which
could lead to a catastrophic loss of control.
Example 10.9 Adaptive flight control is an effective method for controlling a dam-
aged aircraft. A transport aircraft with wing damage is shown in Fig. 10.60. The
wing damage is modeled as a 28% loss of the left wing [37]. Since the damage is
asymmetric, the motion of the damaged aircraft is fully coupled in the roll, pitch,
and yaw axes.
A level flight condition of Mach 0.6 at 15,000 ft is selected. The remaining right
aileron control surface is the only roll control effector available. The reference model
is specified
√ by ω p = 2.0 rad/s, ωq = 1.5 rad/s, ωr = 1.0 rad/s, and ζ p = ζq = ζr =
1/ 2.
The pilot pitch rate command is simulated with a series of ramp input longitu-
dinal stick command doublets, corresponding to the reference pitch angle ±3.81o
from trim. The tracking performance of the nominal controller without adaptation
is compared against the standard MRAC, the e modification with μ = 0.1, and the
optimal control modification with ν = 0.1. The adaptive gains are selected to be
as large as possible within the numerical stability limit of the adaptive laws. This
results in Γ = 60 for the standard MRAC, Γ = 800 for the e modification, and
Γ = 2580 for the optimal control modification. Thus, it can be seen that the opti-
mal control modification can tolerate a much larger adaption rate than the standard
MRAC. This large adaptation rate allows the optimal control modification to achieve
fast adaptation to better adapt to uncertainty than the standard MRAC.
The aircraft angular rate responses are shown in Fig. 10.61. The nominal controller
without adaptation produces a rather poor tracking of the reference pitch rate. Both the
standard MRAC and e modification improve the tracking performance significantly.
412 10 Aerospace Applications
The optimal control modification (OCM) is able to provide a better tracking than
both the standard MRAC and e modification. Due to the asymmetric wing damage,
the roll axis is most affected. With the nominal controller, there is a significant roll
rate of as high as 20o /s. Both the standard MRAC and e modification reduce the
maximum amplitude of the roll rate to about 10o /s. The optimal control modification
further reduces the roll rate to a maximum value of about 4o /s. All the three adaptive
controllers significantly reduce the yaw rate to a very low level. The e modification
is observed to perform slightly better than the standard MRAC and optimal control
modification in the yaw axis.
20 20 20 20
p, deg/sec
p, deg/sec
p, deg/sec
p, deg/sec
0 0 0 0
q, deg/sec
q, deg/sec
q, deg/sec
0 0 0 0
r, deg/sec
r, deg/sec
r, deg/sec
0.2 0.2 0.2 0.2
0 0 0 0
−0.2 Nom. Ref. −0.2 MRAC Ref. −0.2 e Mod. Ref. −0.2 OCM Ref.
−0.4 −0.4 −0.4 −0.4
0 10 20 30 40 0 10 20 30 40 0 10 20 30 40 0 10 20 30 40
t, sec t, sec t, sec t, sec
The aircraft attitude responses are shown in Fig. 10.62. When there is no adapta-
tion, the reference pitch attitude could not be followed accurately. With adaptive con-
trol, the tracking is much improved and the optimal control modification produces
the pitch angle tracking better than both the standard MRAC and e modification.
Without adaptation, as expected the damaged aircraft exhibits a rather severe roll
behavior with a large bank angle between −30o and 20o . Both the standard MRAC
and e modification reduce the bank angle significantly. However, the optimal control
modification shows a drastic improvement in the arrest of the roll motion with the
bank angle maintained very close to the trim value. All the three adaptive controllers
produce similar angle-of-attack responses. With the optimal control modification,
the sideslip angle is reduced to near zero, while both the standard MRAC and e
modification still produce considerable sideslip angle responses.
10.7 Adaptive Flight Control 413
15 15 15 15
Nom. Ref. MRAC Ref. e Mod. Ref. OCM Ref.
10 10 10 10
θ, deg
θ. deg
θ, deg
θ, deg
5 5 5 5
0 0 0 0
0 10 20 30 40 0 10 20 30 40 0 10 20 30 40 0 10 20 30 40
t, sec t, sec t, sec t, sec
20 20 20 20
0 0 0 0
φ, deg
φ, deg
φ, deg
φ, deg
−20 −20 −20 −20
Nom. MRAC e Mod. OCM
−40 −40 −40 −40
0 10 20 30 40 0 10 20 30 40 0 10 20 30 40 0 10 20 30 40
t, sec t, sec t, sec t, sec
10 10 10 10
α, deg
α, deg
α, deg
α, deg
5 5 5 5
β, deg
β, deg
β, deg
0 0 0 0
Fig. 10.62 Closed-loop responses of pitch angle, bank angle, angle of attack, and sideslip angle
The control surface deflections are shown in Fig. 10.63. Because of the wing
damage, the damaged aircraft has to be trimmed with a rather large aileron deflection.
This causes the roll control authority to be severely limited. Therefore, a roll control
saturation is present in all cases. The elevator deflection is nominally similar for all
the four controllers and is well within its control authority. The rudder deflection
produced by the baseline controller is quite significant. Generally, it is desired to
keep the rudder deflection as small as possible in normal operation. Typically, the
rudder deflection limit is reduced as the airspeed and altitude increase. Both the
standard MRAC and e modification reduce the baseline rudder deflection to some
extent, but the optimal control modification is observed to produce the smallest rudder
deflection.
40 40 40 40
30 30 30 30
δa, deg
δa, deg
δ , deg
δ , deg
a
20 20 20 20
Nom. MRAC e Mod. OCM
10 10 10 10
0 20 40 0 20 40 0 20 40 0 20 40
t, sec t, sec t, sec t, sec
2 2 2 2
0 0 0 0
δe, deg
δe, deg
δe, deg
δ , deg
e
−2 −2 −2 −2
Nom. MRAC e Mod. OCM
−4 −4 −4 −4
0 20 40 0 20 40 0 20 40 0 20 40
t, sec t, sec t, sec t, sec
5 5 5 5
0 0 0 0
δr, deg
δ , deg
δ , deg
δ , deg
r
−5 −5 −5 −5
Nom. MRAC e Mod. OCM
−10 −10 −10 −10
0 20 40 0 20 40 0 20 40 0 20 40
t, sec t, sec t, sec t, sec
In the adaptive flight control design in Sect. 10.7, MRAC is used to estimate the
plant uncertainty due to ΔC, ΔD, and ΔE. However, it is well known that MRAC
is designed to drive the tracking error to zero, but is not designed with the primary
objective of parameter estimation. Least-squares methods on the hand are well-
known techniques for parameter estimation. It is possible to combine both the least-
squares parameter estimation and MRAC into a control architecture. This approach
for flight control is called hybrid adaptive flight control [37].
The hybrid adaptive flight control architecture is shown in Fig. 10.64. The archi-
tecture includes an indirect adaptive control that uses a least-squares parameter esti-
mation technique to estimate the true plant model of the aircraft in conjunction with
a direct MRAC designed to reduce the tracking error.
In the hybrid adaptive flight control approach, the dynamic inversion controller
computes the actuator command from a predictor or estimated plant model of the
aircraft dynamics rather than the ideal plant model. The estimated plant model is
expressed as
x̂˙ = C + ΔĈ x + D + Δ D̂ u + E + Δ Ê z (10.170)
Setting the desired acceleration ẋd (t) equal to x̂˙ (t), the actuator command is
computed from the estimated plant model by the dynamic inversion controller as
10.8 Hybrid Adaptive Flight Control 415
−1
u c = D + Δ D̂ ẋd − C + ΔĈ x − E + Δ Ê z (10.171)
The actuator command now depends on the estimates of the plant matrices ΔC,
ΔD, and ΔD which are computed from the indirect adaptive control using a least-
squares parameter estimation method. The plant modeling error ε (t) is computed by
subtracting the plant model from the estimated plant model as
ε = ẋd −ẋ = ΔĈ − ΔC x+ Δ D̂ − ΔD u c + Δ Ê − ΔE z = Θ̃ Φ (x, u c , z)
(10.172)
The recursive least-squares (RLS) indirect adaptive laws for computing Θ (t) are
given by
Θ̇ = −RΦ (x, u c , z) ε (10.173)
where R (t) = R (t) > 0 is the covariance matrix that acts as a time-varying
adaptation rate matrix.
Alternatively, the least-squares gradient indirect adaptive law can be used by
setting R (t) to be a constant.
The matrices ΔĈ (t), Δ D̂ (t), and Δ Ê (t) are computed from Θ (t) at each time
step and then are used in the dynamic inversion controller to compute the actuator
command.
Using Eq. (10.157), the closed-loop plant is expressed as
) t
ẋ = ẋm + K p (xm − x) + K i (xm − x) dτ − u ad − ε (10.175)
0
u ad = ΔΘ Φ (x, u c , z) (10.176)
ė = Am e + B (u ad + ε) (10.177)
The direct MRAC law for computing ΔΘ (t) can be selected to be any robust
modification scheme such as the e modification given by
$ '
ΔΘ̇ = −Γ Φ (x, u c , z) e P B + μ $e P B ' ΔΘ (10.178)
The hybrid adaptive flight control thus uses the RLS indirect adaptive control to
estimate the plant uncertainty. The goal of the RLS indirect adaptive control is to
416 10 Aerospace Applications
drive the plant modeling error to zero. This information is then used in the dynamic
inversion controller to compute the actuator command. The control signal from the
actuator command produces a closed-loop aircraft response which is then compared
with the reference model to form the tracking error. However, because the plant mod-
eling error is reduced by the RLS indirect adaptive control, the resulting tracking error
will be smaller than if the RLS indirect adaptive control is not present. The MRAC
law further drives the residual tracking error to zero. Because the residual tracking
error is small, the adaptation rate for the direct MRAC can be set to a small value to
improve robustness. The RLS parameter estimation method is inherently robust, so,
together with the direct MRAC, the hybrid adaptive flight control architecture can
provide a very effective and robust adaptation mechanism.
Example 10.10 For the same wing-damage aircraft simulation in Example 10.9,
the hybrid adaptive flight control is implemented using the RLS indirect adaptive
control in conjunction with the direct adaptive control using the e modification. The
covariance matrix is initialized with R (0) = 100I . The closed-loop response of the
damaged aircraft due to the hybrid adaptive flight control is shown in Fig. 10.65.
The closed-loop response is exceedingly well-behaved. Both the roll and yaw rates
are much smaller than those with the e modification alone. The pitch rate and pitch
attitude both track the reference model extremely well. The bank angle and sideslip
angle are reduced close to zero as shown in Fig. 10.66. The saturation in the aileron is
still present, but the rudder deflection is much smaller than that with the e modification
alone as shown in Fig. 10.67. Thus, the performance of the hybrid adaptive flight
control is much better than the e modification alone.
10
p, deg/sec
0
p p
m
−10
0 5 10 15 20 25 30 35 40
t, sec
5
q, deg/sec
0
q qm
−5
0 5 10 15 20 25 30 35 40
t, sec
0.2
r, deg/sec
0
r r
m
−0.2
0 5 10 15 20 25 30 35 40
t, sec
10 2
8 0
θ, deg
φ, deg
6 −2
4 −4
θ θ
m
2 −6
0 20 40 0 20 40
t, sec t, sec
10 0.5
8
0
α, deg
β, deg
6
−0.5
4
2 −1
0 20 40 0 20 40
t, sec t, sec
Fig. 10.66 Closed-loop responses of pitch angle, bank angle, angle of attack, and sideslip angle
40
δ , deg
20
a
0
0 5 10 15 20 25 30 35 40
t, sec
5
δe, deg
−5
0 5 10 15 20 25 30 35 40
t, sec
−1
δ , deg
−1.5
r
−2
0 5 10 15 20 25 30 35 40
t, sec
Adaptive flight control can be used to provide consistent handling qualities and
restore stability of aircraft under off-nominal flight conditions such as those due to
failures or damage. Suppose an aircraft plant is described by
where Θi∗ is an unknown, constant ideal weight matrix, and Φ (x, z) is the input
regressor function vector chosen to be
Φ (x, z) = x px q x r x z u (x, z) (10.182)
The inner-loop rate feedback control is designed to improve aircraft rate response
characteristics such as the short-period mode and Dutch roll mode. The reference
model is a second-order model that specifies desired handling qualities with good
damping and natural frequency characteristics as
2
s + 2ζ p ω p s + ω2p φm = g p δlat (10.183)
2
s + 2ζq ωq s + ωq2 θm = gq δlon (10.184)
2
s + 2ζr ωr s + ωr2 βm = gr δr ud (10.185)
where φm (t), θm (t), and βm (t) are the reference bank, pitch, and sideslip angles;
ω p , ωq , and ωr are the natural frequencies for desired handling qualities in the roll,
pitch, and yaw axes; ζ p , ζq , and ζr are the desired damping ratios; δlat (t), δlon (t), and
δr ud (t) are the lateral stick input, longitudinal stick input, and rudder pedal input;
and g p , gq , and gr are the input gains.
Let pm (t) = φ̇m (t), qm (t) = θ̇m (t), and rm (t) = −β̇m (t), then the reference
model can be represented as
) t
ẋm = −K p xm − K i xm (τ ) dτ + Gr (10.186)
0
where xm (t) = pm (t) qm (t) rm (t) , K p = diag 2ζ p ω p , 2ζq ωq , 2ζr ωr , K i = diag
2 2 2
ω p , ωq , ωr = Ω 2 , G = diag g p , gq , gr , and r (t) = δlat (t) δlon (t) δr ud (t) .
For the roll axis, the reference model could also be a first-order model
s + ω p pm = g p δlat (10.187)
10.9 Adaptive Flight Control for F-18 Aircraft with Optimal Control Modification 419
Assuming the pair (A11 , B1 ) is controllable and the outer-loop state vector z (t)
is stabilizable, then the nominal PI feedback controller, defined by ū (t), is given by
) t
ū = K p (xm − x) + K i [xm (τ ) − x (τ )] dτ (10.188)
0
In a more general case when the control vector has more inputs than the number
of states to be controlled, then an optimal control allocation strategy using a pseudo-
inverse method is used to compute the dynamic inversion controller as
−1
u = B1 B1 B1 (ẋm − A11 x − A12 z + ū − u ad ) (10.191)
*t
Let e (t) = 0 xm (t) − x (t) dτ xm (t) − x (t) be the tracking error, then
the tracking error equation is given by
ė = Am e + B Θ1 Φ (x, z) − f 1 (x, z) (10.192)
where
0 I
Am = (10.193)
−K i −K p
0
B= (10.194)
I
−2
B P A−1
m B = −cK i <0 (10.197)
Then, the optimal control modification adaptive law for a nominal PI feedback
controller is specified by
420 10 Aerospace Applications
Θ̇1 = −Γ Φ (x, z) e P B + cνΦ (x, z) Φ (x, z) Θ1 K i−2 (10.198)
where
K −1
PB = c p
(10.202)
K d−1 I + K −1
p
then the optimal control modification adaptive law for a nominal proportional con-
troller is specified by
Θ̇3 = −cΓ Φ (x, z) e K −1 −2
p + νΦ (x, z) Φ (x, z) Θ3 K p (10.204)
Figure 10.69 shows the tracking errors in the roll, pitch, and yaw axes along with
the adaptation weights. With the adaptive controller, the roll rate tracking error is
essentially reduced to zero. Both the pitch rate tracking error and yaw rate tracking
error are reduced to no more than about 2◦ /s with the adaptive controller. The weights
exhibit some large initial transients but then converge to their steady-state values
quickly after 28 s.
Figure 10.70 shows the control surface deflections with the nominal controller and
the adaptive controller. The actuator models are high-fidelity fourth-order models
with time delays. The control surface deflections with the adaptive controller are
well-behaved and smaller in amplitudes than the control surface deflections with the
nominal controller.
Fig. 10.68 Longitudinal states due to an A matrix failure (Cm α shift at 13 s) with and without
optimal control modification
Fig. 10.69 Tracking errors due to an A matrix failure (Cm α shift at 13 s) with and without optimal
control modification
422 10 Aerospace Applications
Fig. 10.70 Control surfaces due to an A matrix failure (Cm α shift at 13 s) with and without optimal
control modification
Fig. 10.71 Longitudinal states due to a B matrix failure (Stabilator Jammed at 2.5o at 13 s) with
and without optimal control modification
10.9 Adaptive Flight Control for F-18 Aircraft with Optimal Control Modification 423
Fig. 10.72 Lateral-directional states due to a B matrix failure (Stabilator Jammed at 2.5o at 13 s
with and without optimal control modification
Figure 10.73 shows the tracking errors in the roll, pitch, and yaw axes along with
the adaptation weights. The roll rate tracking error and pitch rate tracking error are
generally smaller with the adaptive controller than with the nominal controller. The
yaw rate tracking error is about the same for both the nominal controller and the
adaptive controller. The weights converge after about 20 s.
Fig. 10.73 Tracking errors due to a B matrix failure (Stabilator Jammed at 2.5o at 13 s) with and
without optimal control modification
is inserted at 2 s. Figure 10.74 shows the pitch rate tracking error for Γ = 0.5 and
Γ = 50. Increasing the adaptation rate Γ from 0.5 to 50 results in a significantly
smaller tracking error. This is as expected. The response is nicely damped with
Γ = 50 without any high-frequency oscillations. The weights converge faster with
increasing the adaptation rate Γ . Thus, the simulation shows that a larger adaptation
rate can produce better tracking with the optimal control modification without causing
any high-frequency oscillations.
Fig. 10.74 Pitch rate error and weights due to an A-matrix failure (Cm α Shift at 2 s) with optimal
control modification (Γ = 0.5 and Γ = 50 with fixed ν = 1)
The modification parameter ν provides the damping to the adaptive controller. The
same simulation scenario with the A matrix failure is conducted with the modification
parameter changed from ν = 0.25 to ν = 1 while keeping the adaptation rate Γ = 5
constant. Figure 10.75 shows the pitch rate tracking errors for ν = 0.25 and ν = 1.
Increasing the modification parameter ν from 0.25 to 1 adds more damping to the
adaptive controller. As a result, the pitch rate tracking error is well-damped with
ν = 1 without any oscillations. All the weights converge faster with increasing the
modification parameter ν. Thus, this simulation shows that a larger modification
parameter can help improved the performance of the adaptive controller.
This F-18 aircraft simulation study validates the effectiveness of adaptive control
in general and the optimal control modification in particular in improving the flight
control performance of a degraded aircraft. The study coupled with a piloted high-
fidelity F-18 flight simulator experiment established the basis for a flight test program
on NASA F/A-18 research aircraft (tail number 853) in 2010 and 2011.
10.10 Summary 425
Fig. 10.75 Pitch rate error and weights due to an A-matrix failure (Cm α shift at 2 s) with optimal
control modification (fixed Γ = 5 with ν = 0.25 and ν = 1)
10.10 Summary
Adaptive control enjoys many applications in aerospace. It has a long history dated
back to the 1950s with the first adaptive control deployed on NASA X-15 hyper-
sonic vehicles. Applications of adaptive control in other domains are numerous.
Adaptive flight control applications for aircraft longitudinal dynamics using the
bi-objective optimal control modification and least-squares adaptive control with
Chebyshev polynomial approximation and neural networks are illustrated. Adap-
tive flight control for flexible aircraft with aeroelastic interactions using the optimal
control modification and adaptive loop recovery modification is presented. An adap-
tive LQG control method is developed for a flutter suppression of a flexible wing.
A general adaptive flight control architecture based on a dynamic inversion with
adaptive augmentation control is outlined. Robust modification schemes based on
σ modification, e modification, and optimal control modification are used in the
adaptive augmentation control to improve robustness. Hybrid adaptive flight control
is a robust adaptive flight control architecture that blends both direct MRAC and
indirect adaptive control using least-squares parameter identification together. This
technique can significantly improve the performance of a flight control system under
off-nominal operation of an impaired or damaged aircraft. Finally, an adaptive flight
control design for a F/A-18A NASA aircraft based on the optimal control modifica-
tion is implemented in simulations. This implementation was later used in the actual
flight testing in 2010 and 2011.
426 10 Aerospace Applications
10.11 Exercises
1 1
m L 2 θ̈ − mgL sin θ + cθ̇ = u (t − td )
3 2
a. Expand sin θ using the Taylor series expansion about θ (t) = 0 for the first
two terms. Then, express the equation of motion in the form of
ẋ = Ax + B u (t − td ) + Θ ∗ Φ (x)
where x1 (t) = θ (t), x2 (t) = θ̇ (t), x (t) = x1 (t) x2 (t) , Φ (x) is com-
prised of the function in the nonlinear term of the Taylor series expansion of
sin θ and the function in the damping term, and Θ ∗ is a vector of parameters
associated with Φ (x) which are assumed to be unknown.
b. Let m = 0.1775 slug, L = 2 ft, c = 0.2 slug-ft2 /s, td = 0.05 sec, and θ (0) =
θ̇ (0) = 0. Using the equation of motion in part (a), design an adaptive
controller using the optimal control modification to enable the closed-loop
plant to track a reference model specified by
the adaptive controller with the feedback gain values and the reference
model.
b. Implement the adaptive controller in Simulink using the following infor-
mation: x (0) = 0 and Θ (0) = 0 with a time step Δt = 0.01 s for: 1) the
nominal controller, 2) the standard MRAC with Γ = 500, and 3) the optimal
control modification with Γ = 500 and ν = 0.5. The reference command
signal r (t) is a pitch attitude doublet specified in the following plot.
6
2
r, deg
−2
−4
−6
0 5 10 15 20 25 30
t, sec
For each controller, plot the time histories of each of the elements of x (t)
and xm (t) on the same plot, and u (t) for t ∈ [0, 30] s. Plot in units of deg
for α (t), θ (t), and δe (t), and deg/s for q (t). Comment on the simulation
results.
References
1. Bosworth, J., & Williams-Hayes, P. S. (2007). Flight test results from the NF-15B IFCS project
with adaptation to a simulated stabilator failure. In AIAA Infotech@Aerospace Conference,
AIAA-2007-2818, May 2007.
2. Williams-Hayes, P. S. Flight test implementation of a second generation intelligent flight control
system. In: NASA TM-2005-213669.
3. Narendra, K. S., & Annaswamy, A. M. (1987). A new adaptive law for robust adaptation
without persistent excitation. IEEE Transactions on Automatic Control, AC-32(2), 134–145.
4. Calise, A. J., & Rysdyk, R. T. (1998). Nonlinear adaptive flight control using neural networks.
IEEE Control System Magazine, 18(6):14–25.
5. Hanson, C., Johnson, M., Schaefer, J., Nguyen, N., & Burken, J. (2011). Handling qualities
evaluations of low complexity model reference adaptive controllers for reduced pitch and roll
damping scenarios. In AIAA Guidance, Navigation, and Control Conference, AIAA-2011-6607,
August 2011.
6. Hanson, C., Schaefer, J., Johnson, M., & Nguyen, N. Design of low complexity model reference
adaptive controllers. In: NASA-TM-215972.
7. Nguyen, N., Hanson, C., Burken, J., & Schaefer, J. (2016). Normalized optimal control modifi-
cation and flight experiments on NASA F/A-18 aircraft. In: AIAA Journal of Guidance, Control,
and Dynamics.
8. Schaefer, J., Hanson, C., Johnson, M., & Nguyen, N. (2011). Handling qualities of model
reference adaptive controllers with varying complexity for pitch-roll coupled failures. In AIAA
Guidance, Navigation, and Control Conference AIAA-2011-6453, August 2011.
428 10 Aerospace Applications
9. Campbell, S., Kaneshige, J., Nguyen, N., & Krishnakumar, K. (2010) An adaptive control
simulation study using pilot handling qualities evaluations. In AIAA Guidance, Navigation,
and Control Conference, AIAA-2010-8013, August 2010.
10. Campbell, S., Kaneshige, J., Nguyen, N., & Krishnakumar, K. (2010). Implementation and
evaluation of multiple adaptive control technologies for a generic transport aircraft simulation.
In AIAA Infotech@Aerospace Conference, AIAA-2010-3322, April 2010.
11. Burken, J., Nguyen, N., & Griffin, B. (2010). Adaptive flight control design with optimal control
modification for f- 18 aircraft model. In AIAA Infotech@Aerospace Conference, AIAA-2010-
3364, April 2010.
12. Nguyen, N., Burken, J., & Hanson, C. (2011). Optimal control modification adaptive law with
covariance adaptive gain adjustment and normalization. In AIAA Guidance, Navigation, and
Control Conference, AIAA-2011-6606, August 2011.
13. Nguyen, N., & Summers, E. (2011). On time delay margin estimation for adaptive control and
robust modification adaptive laws. In AIAA Guidance, Navigation, and Control Conference,
AIAA-2011-6438, August 2011.
14. Nelson, R. C. (1989). Flight stability and automatic control. New York: McGraw-Hill
15. Nguyen, N., & Balakrishnan, S. N. (2014). Bi-objective optimal control modification adaptive
control for systems with input uncertainty. IEEE/CAA Journal of Automatica Sinica, 1(4),
423–434.
16. Nguyen, N. (2013). Least-squares model reference adaptive control with chebyshev orthogonal
polynomial approximation. AIAA Journal of Aerospace Information Systems, 10(6), 268–286.
17. Rohrs, C. E., Valavani, L., Athans, M., & Stein, G. (1985). Robustness of continuous-time
adaptive control algorithms in the presence of unmodeled dynamics. IEEE Transactions on
Automatic Control, AC-30(9), 881–889.
18. Ioannou, P., & Kokotovic, P. (1984). Instability analysis and improvement of robustness of
adaptive control. Automatica, 20(5), 583–594.
19. Nguyen, N. (2012). Optimal control modification for robust adaptive control with large adaptive
gain. Systems & Control Letters, 61(2012), 485–494.
20. Calise, A. J., & Yucelen, T. (2012). Adaptive loop transfer recovery. AIAA Journal of Guidance,
Control, and Dynamics, 35(3), 807–815.
21. Nguyen, N., Tuzcu, I., Yucelen, T., & Calise, A. (2011). Longitudinal dynamics and adaptive
control application for an aeroelastic generic transport model. In AIAA Atmospheric Flight
Mechanics Conference, AIAA-2011-6319, August 2011.
22. Kokotovic, P., Khalil, H., & O’Reilly, J. (1987). Singular perturbation methods in control:
analysis and design. Philadelphia: Society for Industrial and Applied Mathematics.
23. Nguyen, N. (2010). Elastically Shaped Future Air Vehicle Concept,” NASA Innovation Fund
Award 2010 Report, October 2010, Submitted to NASA Innovative Partnerships Program.
http://ntrs.nasa.gov/archive/nasa/casi.ntrs.nasa.gov/20110023698.pdf.
24. Nguyen, N., Trinh, K., Reynolds, K., Kless, J., Aftosmis, M., Urnes, J., & Ippolito, C., Elas-
tically shaped wing optimization and aircraft concept for improved cruise efficiency. In AIAA
Aerospace Sciences Meeting, AIAA-2013-0141, January 2013.
25. Urnes, J., Nguyen, N., Ippolito, C., Totah, J., Trinh, K., & Ting, E., A mission adaptive variable
camber flap control system to optimize high lift and cruise lift to drag ratios of future N+3
transport aircraft. In AIAA Aerospace Sciences Meeting, AIAA-2013-0214, January 2013.
26. Jordan, T., Langford, W., Belcastro, C., Foster, J., Shah, G., Howland, G., et al. (2004). Devel-
opment of a dynamically scaled generic transport model testbed for flight research experiments.
In AUVSI Unmanned Unlimited, Arlington, VA.
27. Nguyen, N., Ting, E., Nguyen, D., & Trinh, K., Flutter analysis of mission-adaptive wing with
variable camber continuous trailing edge flap. In 55th AIAA/ASME/ASCE/AHS/ASC Structures,
Structural Dynamics, and Materials Conference, AIAA-2014-0839, January 2014.
28. Nguyen, N., Swei, S., & Ting, E. (2015). Adaptive linear quadratic gaussian optimal control
modification for flutter suppression of adaptive wing. In AIAA Infotech@Aerospace Confer-
ence, AIAA 2015-0118, January 2015.
References 429
29. Ardema, M. (1981). Computational singular perturbation method for dynamical systems. AIAA
Journal of Guidance, Control, and Dynamics, 14, 661–663.
30. Dykman, J., Truong, H., & Urnes, J. (2015). Active control for elastic wing structure dynamic
modes. In 56th AIAA/ASCE/AHS/ASC Structures, Structural Dynamics, and Materials Con-
ference, AIAA-2015-1842, January 2015.
31. Slotine, J.-J., & Li, W. (1991) Applied Nonlinear Control. Englewood Cliffs: Prentice-Hall,
Inc.
32. Hsu, L., Teixeira, M. C. M., Costa, R. R., & Assunc̃ao, E. (2011). Necessary and sufficient con-
dition for generalized passivity, passification and application to multivariable adaptive systems.
In 18th International Federation of Automatic Control World Congress, August 2011.
33. Misra, P. (1993). Numerical algorithms for squaring-up non- square systems part II: general
case. In American Control Conference, June 1993.
34. Qu, Z., Wiese, D., Annaswamy, A. M., & Lavretsky, E. (2014). Squaring-up method in the
presence of transmission zeros. In 19th International Federation of Automatic Control World
Congress, August 2014.
35. Lavretsky, E., & Wise, K. (2012). Robust and adaptive control. Berlin: Springer.
36. Nguyen, N., & Urnes, J. (2012). Aeroelastic modeling of elastically shaped aircraft concept via
wing shaping control for drag reduction. In: AIAA Atmospheric Flight Mechanics Conference,
AIAA-2012-4642, August 2012.
37. Nguyen, N., Krishnakumar, K., Kaneshige, J., & Nespeca, P. (2008). Flight dynamics model-
ing and hybrid adaptive control of damaged asymmetric aircraft. AIAA Journal of Guidance,
Control, and Dynamics, 31(3), 751–764.
Suggested Exam Questions
1. For the following systems, determine the equilibrium points. Use the Lyapunov’s
direct method to determine the type of Lyapunov stability for each of the equi-
librium points. Determine all the invariant sets and the values of the Lyapunov
function on the sets. If an equilibrium point is stable, conclude if it is asymptot-
ically stable, and if so, show whether or not it is also exponentially stable.
a.
ẋ1 (x2 − x1 ) x12 + x22 − 1
=
ẋ2 − (x1 + x2 ) x12 + x22 − 1
b.
ẋ1 x22 − x2
=
ẋ2 −x1 − x2 + 1
c.
ẋ1 x2
=
ẋ2 −x1 − (1 + sin x1 ) x2
2. Linearize the systems in problem 2 and determine the types of equilibrium points.
Plot phase portraits.
3. Given the following system
ẋ1 −2 + sin2 x1 1 − sin x1 cos x2 x1
=
ẋ2 −1 + sin x1 cos x2 −2 − cos2 x2 x2
Determine the stability of this system about the origin using the Lyapunov can-
didate function
1
V (x) = x x
2
If asymptotically stable, determine if the origin is exponentially stable and find
the rate of convergence of x, where x (t) = x1 (t) x2 (t) .
P A + A P = −I
V (x) = x P x
C D ≤ C D
ẋm = am xm + bm r
Suggested Exam Questions 433
where am < 0 and bm are known, and r (t) is a bounded command signal.
a. Design a direct adaptive controller that enables the plant output x (t) to track
the reference model signal xm (t). Show by Lyapunov stability analysis that
the tracking error is asymptotically stable; i.e., e (t) → 0 as t → ∞.
b. Implement the adaptive controller in Simulink, given b = 2, am = −1,
bm = 1, and r (t) = sin t. For adaptation rates, use γx = 1 and γ = 1.
For simulation purposes, assume a = 1 and θ ∗ = 0.2 for the unknown
parameters. Plot e (t), x (t), xm (t), u (t), and θ (t) for t ∈ [0, 50].
c. Repeat part (b) for γx = 10 and γ = 10. Plot the same sets of data as in
part (b). Comment on the simulation results for parts (b) and (c) regarding
the tracking of the reference model, the quality of the signal in terms of the
relative frequency content, and the convergence of k x (t) and θ (t) as the
adaptation rates increase.
d. Repeat part (b) for r (t) = 1 (t) where 1 (t) is the unit-step function. Plot
the same sets of data as in part (b). Comment on the convergence of k x (t)
and θ (t) to the ideal values k x∗ and θ ∗ .
6. Given a first-order nonlinear system
ẋ = ax + Bu + cx 2
where x (t) ∈ R, u (t) ∈ R2 , a is an unknown constant, B = 1 2 is known,
and c is an unknown constant.
The reference model is specified as
ẋm = am xm + bm r
Determine K x∗ , K r∗ , and Θ ∗ . Write down the adaptive laws for K x (t) and Θ (t).
Implement the controller in Simulink. Use γx = γΘ = 1. Assume all initial
conditions to be zero and a = 1, c = 0.2 for simulation purpose. Plot e (t), x (t)
versus xm (t), K x (t), and Θ (t) for t ∈ [0, 40].
7. The symmetric sigmoidal function
1 − e−x
σ (x) =
1 + e−x
0.8
0.6
0.4
0.2
B(u)u
−0.2
−0.4
−0.6
−0.8
−1
−10 −5 0 5 10
u
Saturation occurs when a control actuator ceases to be effective. When there are
more inputs than commands, a control allocation strategy should be developed
to allocate redundant control effectors in an optimal manner so as to produce an
output that tracks a command. Define y (u) as the output of a control allocator
as
y = V σ W u
1
J (u) =
2
where = y − r and r ∈ Rn are a command vector for which an optimal control
vector u is to be found to minimize the cost function.
Given r = 1 and
0.75 1.2 0.8
V = , W =
0.5 0.5 1.5
Write a MATLAB code to compute u using the steepest descent method with
an adaptation rate ε = 0.1 and a number of iteration of n = 1000. Indicate the
final value of u and plot u.
8. Adaptive control can be used for disturbance rejection. Disturbances are usually
time signals that may have multiple frequency contents. Unlike unstructured
uncertainty in the form of an unknown function f (x), an unknown function of
Suggested Exam Questions 435
ẋ = ax + b [u + f (t)]
where a and f (t) are unknown, but b = 2. For simulation purpose, a = 1 and
f (t) = 0.1 sin 2.4t − 0.3 cos 5.1t + 0.2 sin 0.7t.
The reference model is given by
ẋm = am xm + bm r
ẋ = −2x − z + u + w
ż = −3z + 4u
y=x
where x (t) is the plant output, z (t) is an internal state, and w (t) = 1 is a
constant disturbance.
u = k x (t) x
k̇ x = −γ x 2 + σ k x
Find the minimum value of the modification parameter σmin to within 0.01
by finding the roots of a polynomial in terms of k̄ x for which one or more
roots satisfy the values of k x in part (a). Calculate k̄ x and x̄.
436 Suggested Exam Questions
ẋm = am xm + bm r
u = k x x + kr r − θ (t) x 2
1
1 1 1
m L 2 4 − 3 cos2 θ θ̈ − mgL sin θ + m L 2 θ̇ 2 sin 2θ + cθ̇ = L cos θu (t − td )
12 2 8 2
where m is the mass of the pendulum, L is the length, g is the gravity constant, c
is the damping coefficient which is assumed to be unknown, θ (t) is the angular
position, u (t) is the control input which represents the horizontal force at point
O, and td is a time delay which represents the motor actuator dynamics.
a. Let x1 (t) = θ (t), x2 (t) = θ̇ (t), and x (t) = x1 (t) x2 (t) . Derive the
expressions for the nonlinear dynamic inversion adaptive controller and the
σ modification adaptive law to estimate the unknown coefficient c in order
to enable the closed-loop plant to track a reference model specified by
ẋm = Am xm + Bm r
by finding the general expressions and the numerical values for k p , ki , and
kq to enable the aircraft to track a reference model of the angle of attack
where ζm = 0.75
t and ωm = 1.5 rad/s.
b. Let z (t) = 0 (α (t) − r (t)) dτ , provide the general expression and the
numerical value for the reference model of the aircraft as
ẋm = Am xm + Bm r
where x (t) = z (t) α (t) q (t) .
c. Let Θ ∗ = 0 θα∗ θq∗ . Design an adaptive angle-of-attack controller using
the optimal control modification to enable the closed-loop plant to track
the reference model. Express the adaptive controller and the adaptive law.
Given Q = 100I , select the modification parameter to guarantee stability of
the closed-loop plant by using the linear asymptotic property of the optimal
control modification and the following formulas to compute the crossover
frequency and time delay margin for MIMO systems. Plot ν versus td for
ν ∈ [0, 5] and determine ν to within 0.01 for td = 0.01 s
For a general time delay system
ẋ = Ax + Bu (t − td )
ω = μ (− j A) + B K x
1 μ (A)
td = cos−1
ω μ (−B K x )
2
r, deg
−2
−4
−6
0 5 10 15 20 25 30
t, sec
Plot the time histories of each of the elements of x (t) and xm (t) on the same
plot, and u (t) for t ∈ [0, 30] s. Plot in the units of deg–sec for z (t), deg for
α (t) and δe (t), and deg/s for q (t).
a. Design an ACAH hybrid adaptive flight controller for the pitch axis to enable
the aircraft to follow a reference model
by providing the expressions for the hybrid adaptive controller, the least-
squares gradient parameter estimation of ΔAqα , ΔAqq , ΔBq , and the optimal
control modification adaptive law to handle the residual tracking error.
b. Implement the hybrid adaptive flight controller in Simulink using the same
aircraft parameters from Exam Problem 12 and the following additional
information: td = 0.02 s, ζq = 0.75, ωq = 2.5 rad/s, ΔAαα = 0.1616/s,
ΔAqα = 2.1286 /s2 , ΔAqq = 0.5240 /s, ΔBα = −0.0557 /s, ΔBq =
−2.5103 /s2 , α (0) = 0, q (0) = 0, Δ Âqα (0) = 0, Δ Âαα (0) = 0,
Δ B̂q (0) = 0, R = 1000I , Γ = 1000I , and ν = 0.1 with a time step
Δt = 0.001 s. The reference command signal r (t) is a pitch rate doublet
specified in the following plot.
3
1
r, deg/sec
−1
−2
−3
0 5 10 15 20 25 30
t, sec
c. Simulate three cases: (1) nominal controller, (2) only direct MRAC, and (3)
hybrid adaptive control with both direct MRAC and indirect least-squares
gradient adaptive control. For each case, plot the time histories of each of
the elements of α (t), θ (t) , q (t) and qm (t) on the same plot, and u (t) for
t ∈ [0, 30] s. In addition, plot Θ (t) for case 2; and Δ Âqα (t), Δ Âqq (t),
Δ B̂q (t) for case 3. Plot in the units of deg for α (t), θ (t), and δe (t), and
deg/s for q (t).
Index
G
Generic Transport Model M
Damaged Aircraft, 412–419 Model Matching Condition, 90, 100, 111
Flexible Aircraft, 377–380, 382 Model-Reference Adaptive Control
Longitudinal Dynamics, 364, 370 Composition, 85–90
Direct, 85, 89–96, 98–107, 109–118
Indirect, 95–99, 105–110, 117–120
H Neural Network, 179–183, 369–370
Hamiltonian Function, 235, 239, 241, 242, Nonlinear Integral Control, 197–204
294 Overview, 83–85
Hybrid Adaptive Flight Control, 414, 416 Robustness Issues, 186–204
Time-Delay Margin, 257
K
Kalman-Yakubovich Lemma, 331, 394 N
Negative Definiteness, 24–27
Neural Network
L Aircraft Adaptive Control, 370
L1 Adaptive Control, 267–274 Learning Law, 156–158
Lagrange Multiplier, 235 Radial Basis Function, 155–157
LaSalle’s Invariant Set Theorem, 63–66 Sigmoidal Function, 155–158, 369–370
Least-Squares Non-Autonomous Systems, 18–20, 69–79
Index 443
T Unmatched, 87
Time-Delay Margin, 202, 203, 257, 258 Unstructured, 86, 176
Time-Delay Systems, 193–195 Uniform Boundedness, 71
Trace Operator, 35, 108 Uniform Continuity, 74–76
Tranversality Condition, 236, 239–241 Uniform Ultimate Boundedness, 72–75
Two-Point Boundary Value Problem, 236, Unmodeled Dynamics
237 Definition, 86, 95
Robustness Issues, 195–200
U
Uncertainty
Control Input, 88, 286, 291 V
Matched, 86, 87, 90 Van der Pol Oscillator, 28
Parametric, 85–87 Verification and Validation, 6–9
Structured, 86, 90 Virtual Control, 399–400