A Lagrangian Network For Kinematic Control of Redundant Robot Manipulators
A Lagrangian Network For Kinematic Control of Redundant Robot Manipulators
A Lagrangian Network For Kinematic Control of Redundant Robot Manipulators
Abstract—A recurrent neural network, called the Lagrangian problem for serial-link manipulators. The difficulties are com-
network, is presented for the kinematic control of redundant pounded by the requirement of real-time solution in sensor-
robot manipulators. The optimal redundancy resolution is de- based robotic operations. The most direct way to solve (2)
termined by the Lagrangian network through real-time solution
to the inverse kinematics problem formulated as a quadratic is to derive a closed-form solution from (1). Unfortunately,
optimization problem. While the signal for a desired velocity obtaining a closed-form solution is difficult for most manipula-
of the end-effector is fed into the inputs of the Lagrangian tors due to their nonlinearity of . Moreover, the solution is
network, it generates the joint velocity vector of the manipulator often not unique for kinematically redundant manipulators due
in its outputs along with the associated Lagrange multipliers. The to their redundancy. Making use of the relation between joint
proposed Lagrangian network is shown to be capable of asymp-
totic tracking for the motion control of kinematically redundant velocity and Cartesian velocity is a common indirect
manipulators. approach to the inverse kinematics problem. The velocity
vectors and have the following linear relation:
Index Terms— Asymptotic stability, kinematic control, kine-
matically redundant manipulators, optimization method, recur- (3)
rent neural networks.
where is the Jacobian matrix and can be rank-
I. INTRODUCTION deficient sometimes. In a kinematically redundant manipulator,
. This indirect approach begins with the desired
control [24], [31], [38], [41]–[43], [45], [49], [50]–[54]. Unlike By setting the partial derivatives of to zero,
feedforward neural networks, most recurrent neural networks the Lagrange necessary condition gives rise to the following
do not need off-line supervised learning and thus are more time-varying algebraic equations:
suitable for real-time robot control in uncertain environments.
In this paper, a recurrent neural network, called the La- (8)
grangian network, is presented for the kinematic control of (9)
redundant manipulators. The proposed Lagrangian network
explicitly minimizes a weighted norm of the joint velocity Multiplying both sides of (8) by 1, then rewriting (8) and
vector while keeping the linear relation between and (9) in a combined matrix form, we have
satisfied. It is shown to be asymptotically stable and capable
of solving the inverse kinematics problem in real time. The (10)
operating characteristics and performance of the Lagrangian
network are demonstrated by use of simulation results. III. NETWORK DESCRIPTION
The rest of this paper is organized in five sections. In
Section II, the inverse kinematics problem is formulated as A recurrent neural network is presented in [28] for solving
a time-varying quadratic programming problem with equal- the quadratic programming problem with equality constraints
ity constraints. In Section III, the energy function and dy- like (5) and (6). While the original neural network was used for
namic equation of the Lagrangian network are described. In solving static quadratic programming with equality constraints,
Section IV, the Lagrangian network is proven to have the it can be easily extended to time-varying problems. Let the
capability of asymptotic tracking. In Section V, the simulation state vectors of output neurons and hidden neurons be denoted
results of the Lagrangian network to control a seven-degree-of- by and , an -vector representing estimated
freedom (DOF) manipulator are discussed. Finally, Section VI and an -vector representing estimated , respectively.
concludes the paper with final remarks. The dynamic equations of the Lagrangian network can be
expressed by the following time-varying linear differential
equations:
II. PROBLEM FORMULATION
In order to determine for given through (3), the (11)
joint velocity vector needs to be computed. One way to (12)
determine is as follows [9]:
where and are positive diagonal
(4) capacitive matrices. The positive diagonal capacitive matrices
and are used to precondition the system matrix and
where is the pseudoinverse of , is an
scale the convergence rate of the Lagrangian network.
-vector of desired velocity, is an identity matrix, and
Equation (11) shows that the symmetric connection weight
is an -vector of arbitrary time-varying variables. This
matrix among the neurons represented by is . Equa-
method entails the computation of time-varying pseudoinverse
tion (11) also shows that the time-varying connection weight
. Since is underdetermined in a kinematically
matrix from the neurons represented by to the neurons
redundant manipulator, another way to determine without
represented by is . Equation (12) shows that
the need for computing the pseudoinverse is to solve the
the time-varying connection weight matrix from the neurons
following time-varying quadratic programming problem with
represented by to the neurons represented by is
equality constraints:
. Equation (12) also shows that the external input
minimize (5) vector to the hidden layer is . Fig. 1 illustrates the
kinematic control process based on the Lagrangian network.
subject to (6)
In this context, the desired velocity vector is input into
where the superscript denotes the transpose operator and the Lagrangian network, and at the same time the Lagrangian
is an symmetric positive-definite weighting matrix. network outputs the computed joint velocity vector .
In particular, if is an identity matrix, then the objective
function to be minimized is equivalent to the 2-norm of joint IV. STABILITY ANALYSIS
velocity . If is the inertia matrix, then the objective Written in a combined format, the Lagrangian network can
function to be minimized is the kinetic energy. be described as the following time-varying linear dynamic
The Lagrangian of the time-varying quadratic programming system:
problem subject to equality constraints described in (5) and
(6) is defined as follows: (13)
where diag ,
(7) , and
(15)
where . Then is a real matrix. Furthermore
Since is a constant positive definite symmetric matrix and
is a diagonal matrix with positive diagonal entries, (15)
implies that
(16)
Choose a vector as
as (17)
(18)
(19)
(20) Since
(a) (b)
(c) (d)
(e) (f)
(g) (h)
(i) (j)
(k) (l)
Fig. 5. Transient behaviors of the Lagrangian network and the PA-10 manipulator.
WANG et al.: LAGRANGIAN NETWORK FOR KINEMATIC CONTROL 1129
VI. CONCLUSIONS
The proposed Lagrangian network provides a new parallel
distributed model for real-time kinematic control of redundant
robot manipulators. Compared with the supervised learning
neural networks for robot kinematic control, the present ap-
proach eliminates the need for training and produces much
smaller position error and velocity error. Compared with other
recurrent neural networks, the present approach explicitly
minimizes the norm of joint velocity and performs better in
terms of joint velocity norm as well as position and velocity
errors. The simulation results have shown the capability and
characteristics of the proposed Lagrangian network for the
kinematic control of redundant manipulators. Implemented in
a dedicated hardware such as ASIC, the Lagrangian network
is suitable for the real-time kinematic control of redundant
robots, especially mobile ones.
APPENDIX I
The homogeneous transformation matrix , which spec-
ifies the position and orientation of the end-effector with
respect to the base coordinate system, is the chain product of
Fig. 6. Trajectories of the poles of the Lagrangian network on the complex
plane. successive coordinate transformation matrices of , for
1130 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 10, NO. 5, SEPTEMBER 1999
0s1 q114 0 c1 q234 c1 p23 s1 s2 p33 0 c2 p32 q423 p43 0 q433 p42 0q422 p53 + q432 p52 q522 p63 0 q532 p62 0
c1 q114 0 s1 q234 s1 p23 c2 p31 0 c1 s2 p33 q433 p41 0 q413 p43 0q432 p51 + q412 p53 q532 p61 0 q512 p63 0
J= 0 0s1 p22 0 c1 p21 c1 s2 p32 0 s1 s2 p31 q413 p42 0 q423 p41 0q412 p52 + q422 p51 q512 p62 0 q522 p61 0
0 0 s1 c1 s2 q413 0q412 q512 0q612 (21)
0 c1 s1 s2 q423 0q422 q522 0q622
1 0 c2 q433 0q432 q532 0q632
, expressed as
and
Man, Cyben., vol. 16, pp. 93–101, 1986. [35] J. Hakala, G. Fahner, and R. Eckmiller, “Rapid learning of inverse robot
[6] R. V. Mayorga, A. K. C. Wong, and N. Milano, “A fast procedure for kinematics based on connection assignment and topographical encoding
manipulator inverse kinematics evaluation and pseudoinverse robust- (CATE),” in Proc. IJCNN, Seattle, WA, vol. II, 1991, pp. 1536–1541.
ness,” IEEE Trans. Syst., Man, Cybern., vol. 22, pp. 790–798, 1992. [36] C. Kazakiewicz, T. Ogiso, and N. Myyake, “Partitioned neural-network
[7] T. Isabe, K. Nagasaka, and S. Yamamoto, “A new approach to kinematic architecture for inverse kinematic calculation of a 6 DOF robot manip-
control of simple manipulators,” IEEE Trans. Syst., Man, Cybern., vol. ulators,” in Proc. IJCNN, Seattle, WA, vol. III, 1991, pp. 2001–2006.
22, pp. 1116–1124, 1992. [37] C. Cox, R. Saeks, M. Lothers, R. Pap, and C. Thomas, “A neurocon-
[8] C. A. Klein and C.-H. Huang, “Review of pseudoinverse control for use troller for robotic applications,” in Proc. IEEE Int. Conf. Syst., Man,
with kinematically redundant manipulators,” IEEE Trans. Syst., Man, Cybern., 1992, pp. 712–716.
Cybern., vol. 13, pp. 245–250, 1983. [38] Y. Li and N. Zeng, “A neural-network-based inverse kinematics solution
[9] D. R. Baker and C. W. Wampler II, “On the inverse kinematics of in robotics,” in Neural Networks in Robotics, G. A. Bekey and K. Y.
redundant manipulators,” Int. J. Robot. Res., vol. 7, pp. 3–21, 1988. Goldberg, Eds. Boston, MA: Kluwer, 1993, pp. 97–111.
[10] J. J. Hopfield and D. W. Tank, “Neural computation of decisions in [39] T. C. Hsia and Z. Mao, “Obstacle avoidance inverse kinematics solution
optimization problems,” Biol. Cybern., vol. 52, pp. 141–152, 1985. of redundant manipulators by neural networks,” in Proc. IEEE Int. Conf.
[11] D. W. Tank and J. J. Hopfield, “Simple neural optimization networks: Robot. Automat., 1993, p. 1014.
An A/D converter, signal decision circuit, and a linear programming [40] J. M. I. Zannatha, D. F. Bassi, and R. A. Garcia, “Position and dif-
circuit,” IEEE Trans. Circuits Syst., vol. 33, pp. 533–541, 1986. ferential kinematic neural control of robot manipulators: A comparison
[12] J. Wang, “Analysis and design of a recurrent neural network for linear between two schemes,” in Proc. IEEE Int. Conf. Syst., Man, Cybern.,
programming,” IEEE Trans. Circuits Syst. I, vol. 40, pp. 613–618, 1993. 1993, pp. 479–484.
[13] J. Wang, “A deterministic annealing neural network for convex pro- [41] D. H. Rao, M. M. Gupta, and P. N. Nikiforuk, “On-line learning of robot
gramming,” Neural Networks, vol. 7, pp. 629–641, 1994. inverse kinematic transformations,” in Proc. IJCNN, Nagoya, Japan,
[14] A. Cichocki and R. Unbehauen, “Neural networks for solving systems 1993, pp. 2827–2830.
of linear equations and related problems,” IEEE Trans. Circuits Syst., [42] S. Lee and R. M. Kil, “Redundant arm kinematic control with recurrent
vol. 39, pp. 124–138, 1992. loop,” Neural Networks, vol. 7, pp. 643–655, 1994.
[15] A. Cichocki and R. Unbehauen, Neural Networks for Optimization and [43] G. Wu and J. Wang, “A recurrent neural network for manipulator inverse
Signal Processing. New York: Wiley, 1993. kinematics computation,” in Proc. IEEE ICNN, Orlando, FL, 1994, vol.
[16] Y. Xia and J. Wang, “A general methodology for designing globally con- V, pp. 2715–2720.
vergent optimization neural networks,” IEEE Trans. Neural Networks, [44] Y. Kuroe, Y. Nakai, and T. Mori, “A new neural-network learning of
vol. 9, pp. 1331–1343, 1998. inverse kinematics of robot manipulator,” in Proc. IEEE ICNN, Orlando,
[17] J. Wang, “A recurrent neural network for real-time matrix inversion,” FL, 1994, vol. V, pp. 2819–2824.
Appl. Math. Comput., vol. 55, pp. 89–100, 1993. [45] K. Mathia, R. Saeks, and G. G. Lendaris, “Linear Hopfield networks,
[18] , “Recurrent neural networks for solving linear matrix equations,” inverse kinematics and constrained optimization,” in Proc. IEEE Int.
Comput. Math. Applicat., vol. 26, no. 9, pp. 23–34, 1993. Conf. Syst., Man, Cybern., 1994, pp. 1269–1273.
[19] J. Wang and H. Li, “Solving simultaneous linear equations using [46] R. Gazit and B. Widrow, “Back propagation through links: A new
recurrent neural networks,” Inform. Sci., vol. 76, pp. 255–277, 1994. approach to kinematic control of serial manipulators,” in Proc. IEEE
[20] J. Wang, “Recurrent neural networks for computing pseudoinverses of Int. Symp. Intell. Contr., 1995, pp. 99–104.
rank deficient matrices,” SIAM J. Sci. Comput., vol. 18, pp. 1497–1493, [47] B.-L. Lu and K. Ito, “Regulation of inverse kinematics for redundant
1997. mainulators using neural network inversions,” in Proc. IEEE ICNN,
[21] G. A. Bekey, “Robotics and neural networks,” in Neural Networks for Perth, Australia, 1995, pp. 2726–2731.
Signal Processing, B. Kosko, Ed. Englewood Cliffs, NJ: Prentice-Hall, [48] A. Ramdane-Cherif, V. Perdereau, and M. Drouin, “Optimization
1992, pp. 161–185. schemes for learning the forward and inverse kinematic equations
[22] A. Guez and Z. Ahmad, “Solution to the inverse problem in robotics by with neural network,” in Proc. IEEE ICNN, Perth, Australia, 1995, pp.
neural networks,” in Proc. IEEE ICNN, San Diego, CA, vol. II, 1988, 2732–2737.
[49] H. Ding and S. P. Chan, “A real-time planning algorithm for obstacle
pp. 617–624.
[23] G. Josin, D. Charney, and D. White, “Robot control using neural avoidance of redundant robots,” J. Intell. Robot. Syst., vol. 12, pp. 1–15,
networks,” in Proc. IEEE ICNN, San Diego, CA, 1988, pp. 625–631. 1996.
[24] J. Guo and V. Cherkassy, “A solution to the inverse kinematics [50] D. Wang and A. Zilouchian, “Solutions of kinematics of robot manip-
problem in robotics using neural network processing,” in Proc. IJCNN, ulators using Kohonen self-organizing neural network,” in Proc. IEEE
Int. Symp. Intell. Contr., Istanbul, Turkey, 1997, pp. 251–255.
Washington, D.C., 1989, vol. II, pp. 299–304.
[51] J. Wang, Q. Hu, and D.-C. Jiang, “A two-layer recurrent neural network
[25] R. Eckmiller, J. Beckmann, and M. Lades, “Neural kinematics net for
for pseudoinverse control of kinematically redundant manipulators,” in
a redundant robot arm,” in Proc. IJCNN, Washington, D.C., 1989, vol.
Proc. IEEE Conf. Decision Contr., San Diego, CA, 1997, pp. 2507–2512.
II, pp. 333–339. [52] A. F. R. Araujo and M. Vieira, “Associative memory used for trajec-
[26] A. Guez and Z. Ahmad, “Accelerated convergence in the inverse
tory generation and inverse kinematics problem,” in Proc. IEEE/INNS
kinematics via multilayer feedforward networks,” in Proc. IJCNN,
IJCNN, Anchorage, AK, 1998, pp. 2057–2062.
Washington, D.C., 1989, vol. II, pp. 341–344. [53] H. Ding and S. K. Tso, “Minimum infinity-norm kinematic solution for
[27] Z. Ahmad and A. Guez, “On the solution to the inverse kinematic prob- redundant robots using neural networks,” in Proc. IEEE Int. Conf. Robot.
lem,” in Proc. IEEE Int. Conf. Robot. Automat., 1990, pp. 1692–1697. Automat., Leuven, Belgium, 1998, pp. 1719–1724.
[28] F. J. Arteage-Bravo, “Multilayer backpropagation network for learn- [54] A. F. R. Araujo and H. D’Arbo, “A partially recurrent neural network to
ing the forward and inverse kinematics equations,” in Proc. IJCNN, perform trajectory planning, inverse kinematics, and inverse dynamics,”
Washington, D.C., 1990, vol. II, pp. 319–322. in Proc. IEEE Int. Conf. Syst., Man, Cybern., San Diego, CA, 1998, pp.
[29] H. W. Werntges, “Delta rule-based neural network for inverse kinemat- 1784–1789.
ics: Error gradient reconstruction replaces the teacher,” in Proc. IJCNN,
Washington, D.C., 1990, vol. III, pp. 415–420.
[30] J. Hakala, R. Steiner, and R. Eckmiller, “Quasilocal solution for inverse
kinematics of a redundant robot arm,” in Proc. IJCNN, San Diego, CA,
1990, vol. III, pp. 321–326.
[31] S. Lee and R. M. Kil, “Robot kinematic control based on bidirectional Jun Wang (S’89–M’90–SM’93) received the B.S.
mapping neural network,” in Proc. IJCNN, San Diego, CA, vol. III, degree in electrical engineering and the M.S. degree
1990, pp. 327–335. in systems engineering from Dalian University of
[32] S. Kieffer, V. Morellas, and M. Donath, “Neural network learning of Technology, China. He received the Ph.D. degree
the inverse kinematic relationships for a robot arm,” in Proc. IEEE Int. in systems engineering from Case Western Reserve
Conf. Robot. Automat., 1991, pp. 2418–2425. University, Cleveland, OH.
[33] C. L. P. Chen and A. D. McAulay, “Robot kinematics learning com- Presently, he is an Associate Professor at the Chi-
putations using polynomial neural networks,” in Proc. IEEE Int. Conf. nese University of Hong Kong. Prior to this position,
Robot. Automat., Sacramento, CA, 1991, pp. 2638–2643. he was an Associate Professor at the University
[34] F. Pourboghrat, “Neural networks for learning inverse kinematics of of North Dakota, Grand Forks, ND. His current
redundant manipulators,” in Proc. IJCNN, Seattle, WA, vol. II, 1991, research interests include neural networks and their
pp. 1004–1006. applications in control and manufacturing systems.
1132 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 10, NO. 5, SEPTEMBER 1999
Qingni Hu received the B.S. degree in mechanical Danchi Jiang received the B.S. degree in mathemat-
engineering in 1982 and the M.S. degree in com- ics from Wuhan University, Wuhan, China, the M.S.
puter science in 1988, both from Dalian University degree in control systems and applications from
of Technology, China. East China Normal University, Shanghai, China,
He is currently an Associate Professor at Dalian and the Ph.D. degree in systems engineering from
University of Technology. His research interests Australian National University, Canberra, Australia.
include image processing, computer graphics, and This work was done while he was a Postdoctoral
neural networks. Research Fellow at the Chinese University of Hong
Kong. His present research interests include engi-
neering software development, neural computing,
system optimization, and control system design.