Nonlinear Algebra Matrix
Nonlinear Algebra Matrix
ABSTRACT
A matrix v e c t o r formalism is d e v e l o p e d for systematizing the m a n i p u l a t i o n o f sets o f non-
linear algebraic e q u a t i o n s . In this formalism all m a n i p u l a t i o n s are p e r f o r m e d b y multiplication
with specially c o n s t r u c t e d t r a n s f o r m a t i o n matrices. For m a n y i m p o r t a n t classes o f nonlinear-
ities, algorithms based on this formalism are p r e s e n t e d for rearranging a set o f e q u a t i o n s so
that their solution m a y be o b t a i n e d b y numerically searching along a single variable. T h e o r y
developed proves that all solutions are obtained.
THEOREM 3-1
fn=0 I f (1) P is a transformation matrix
(2) F is an equation vector
Definition then
A set o f equations F 2 -- 0 is said to be equivalent F=0 C PF~_0
(N) to a set o f equations F 1 = 0 if and only if the [IPI=0]
solution sets o f F 2 = 0 are identical with the solu-
tion sets o f F 1 = 0. In symbolic notation
Proof
The p r o o f consists o f two parts. First it is proved
F 1 = 0 ~ F 2 ~ 0. that PF = 0 is d o m i n a n t t o F = 0. Second it is prov-
ed that any additional solution sets, if t h e y exist,
Definition must satisfy I PI = 0.
A set o f equations F 2 = 0 is said to be subordinate Clearly every solution set o f F _- 0 is also a solution
(D) to a set o f equations F 1 = 0, if and only if the set o f PF = 0; thus, PF = 0 is d o m i n a n t to F = 0.
solution sets o f F 2 = 0 are a subset o f the solution If there exists any additional solution set, i. e., a
solution set which satisfies PF = 0 but which does
sets of F 1 = 0 . In symbolic notation not satisfy F = 0, let it be d e n o t e d b y 7- Substitut-
F 1 = 0 D F 2 -__ 0. ing 7 into PF = 0 and formally using Cramer's rule
yields
Definition IP ( ' / ) I F j ( 7 ) = 0 j = 1, 2 , . . . , n
A set o f equations F 2 = 0 is said~to be dominant
(C) to a set o f equations F 1 = 0, if and only if the As IP('/)[ =/=0, one has a system o f homogeneous
equations with a solution different from null vector
solution sets o f F 1 = 0 are a subset o f the solution
0. Hence IP(~/)] - 0.
sets of F 2 = 0. In symbolic notation
FI=0CF2=0. Corollary
If (1) P is a transformation matrix
Definition (2) F is an equation vector
If a set o f equations F 2 = 0 is d o m i n a n t to a set o f (3) IPI ~ 0 for any set o f values o f the variables
equations F 1 = 0, any solution sets o f F 2 = 0 which contained in P,
are not solution sets o f F 1 = 0 are called additional then
solution sets (with respect to F 1 = 0). The notation F=0~PF=0
A n y transformation matrix P has a formal inverse
F1 = 0 C F2 = 0
p - 1 with the property
If=01
is defined to mean that F 2 = 0 is d o m i n a n t to pp-1 = p-1p = I
F 1 = 0 and if any additional solution sets exist, where I is the unit matrix. The construction o f p - 1
they must satisfy f = O, where f is a continuous 'is the same whether P contains elements which are
function. constants or functions. The inverse o f any transform-
ation matrix P can be always constructed w i t h o u t
Definition
assigning a set o f values to the variables contained
If a set o f equations F 2 = 0 is subordinate to a set
inP.
o f equations F 1 = 0, any solution sets o f F 1 = 0 There m a y exist sets o f values for the variables con-
which are n o t solution sets o f F 2 = 0 are called tained in P such that I P l - - 0 . Thus p - 1 usually con-
missing solution sets (with respect to F 1 = 0). The tains elements which are discontinuous functions.
Journal o f Computational and Applied Mathematics, volume I, no 1, 1975. 27
The following Theorem deals with the special matrix Since ID I =# 0 identically, D-1 exists. Premultiplying
p-1 whose elements may not be continuous func- on the left by D -1 yields
tions, but whose inverse P has elements which are Q = D-1p
continuous functions.
In order to illustrate the application of Theorem
THEOREM 3-2 3-3, consider the following case when the Q matrix
If (1) P is a transformation matrix is used for the premultiplication of F = 0. Then
(2) F is an equation vector QF=0~D-1pF=0 C PF=0 ~ F=0
then [IDI=O] [IN=O]
F=O D p-1F= 0
according to Theorems 3-3, 3-2 and 3-1.
[IPI]=0
Q = [qiJ] =
[,,]
Any matrix Q can be represented by linear form matrices, an important difference between
sets of linear equations and sets of nonlinear equa-
tions should be noted. The linear form coefficient
and augmented matrices B and A are unique for a
set of fLxed linear equations, but are not unique for
where qij may not be continuous, but where qij and a set of fLxed nonlinear equations. The non-unique-
q'.. are constants or continuous functions. The q'.. are ness of B, for example, is illustrated with the follow-
B . . . B ing set of nonlinear equations.
restricted to be not Identically zero.
Let the elements of the diagonal matrix D=[diirij ] 2
be given by XlX 2 + x 2 + x 2 x 3 - 12 = 0 ] -
n
XlX 2 + x 2 x 3 - x22x 3 + 4 ~t (4.1-1)
d i i = k =II 1 qik
' = qil
' qi2 . . . . q i n
xI + x22 x3 2
The elements dii are continuous since the q~k are
continuous. The linear form coefficient matrix B may take a
The ij th element of DQ, (DQ)ij, is number of forms, three of which are
n
I! 2
(DQ)ij = r =I
Z dii ~ir qrj = dii qij 1 x2 t/]Ix22 1 x2211x22 1 x2-I
2
1
0
\
C 0
Xs
0
S rOW
(4.1.2-2)
4.2. Rank, linear dependence, and nonlinear depend-
ence
A few useful concepts are singled out for special
emphasis here. The concepts of rank, linear depend-
ence, and nonlinear dependence help to clarify the
algorithms of solution described in the next section.
1 0 4.2.1. RANK
C) ",iJ
0 Xt t row
In linear algebra the concept of rank permits elegant
OI statement of the conditions under which a set of
linear equations has a solution. This concept of rank
can readily be extended to sets of nonlinear equa-
tions represented in linear form. Since linear form
matrices may contain one or more unknowns, it is
Thus the nonlinear equations have been reduced con- (5) Perform a column operation on A' for each vari-
ditiona~y to a set o f n linear equations with n-I un- able contained in A', and delete the column o f
knowns. The Theorem now follows directly from zeros created. If x 2 appears in A' operate on and
classical linear algebra. then delete the second column, etc. If the result-
ing matrix contains more than R + I columns,
4.2. 3. NONLINEAR DEPENDENCE continue column operation and subsequent dele-
The concept o f linear dependence presented in Sec- tion on arbitrarily selected columns until exactly
tion 4.2. 2 can be extended to nonlinear dependence. R + I columns remain. Through step (5) strict
equivalence has been maintained.
Definition (6) Using the generalized Cramer's rule (see Theorem
A set o f m row vectors X 1, X2, . . . , X m is said to 4-3, below) solve for the R variables which cor-
be nonlinearly dependent if there exists a set of con- respond to the first R columns in terms o f the
Y y2 1 (4.3-4)
Pnl Pn2
F2
r
=0~
i -Y
[(2-y2)
Y I
=1
_y3
+ y
y2 y y2 + y
1 -1 (4.3-5)
LPIn .P2n Pnn Pnl Pn2 "'" Pnn Fn
y2 -y
1ili02
PI z = 1 (2-y2) = y4_2y2_y
\o IPI
=0
y2
1 -1
y y2 + y
~
f
y=0
y = -0.618
y=-I y=+l y = 1.618
fkl
f12 " " " f l m - 1
[i1 f
f
lm
[xo
ist
f, f, f, f ~
nl n2" " " nm-1 nm 1
Thas the roots y = 0 and y = -1 may correspond (4.4-1)
to additional solution sets. These two values are now
where Xi (i = 1, 2 . . . . . n) are the n row vectors fil'
checked by means of Theorem 4- !. Substituting
f'i2 . . . .. f'im-l" If there exists a set of functions or
y = 0 into (4.3-4) yields
constants @i (i = I, 2 . . . . . m) as defined in Section
I: 0 01 0 1 (4.3-9)
4.2.3, i. e.,
~1 X1 + @2X2 + • • + @mXm = 0,
L1 -1 -2J then
Since the conditional rank of the coefficient matrix
of (4.3-9) is one, and the conditional rank of the X1 f'lm' -X1 f'lm
augmented matrix is two, y = 0 is not a solution set
of the original problem (4.3-2). Substituting y = -1 " C
into (4.3-4) yields Xk [@k'=ol Xk-1 fk-lm
0 @1f,lm+@2f2m •
, +" "+@mf,nm
t
-hi - 11
-1
- 11
-1
Since the conditional rank of the coefficient matrix
of (4.3-10) equals the conditional rank of the aug-
mented matrix of (4.3-10) which equals one, y = - 1
is part of a one-fold infinity of solutions of the
(4.3-10)
iX!o Xk+l
Xn
f'nm
(4.4-2)
I
4.4. Singular elimination
Singular elimination is a flexible elimination tech- (4.4-4)
nique which in favorable cases is more powerful than whose determinant is @k"
systematic elimination. Occasionally singular elimina- Clearly is @k is a non-zero constant the matrix on
X1
X2
fi2 . , •
fi 2 "'"
f,
m2 k
=0
[¢1 ¢ 2 " " ¢ r - 1 ] . = 0, i. e., ~bX = 0.
flm-1 . . . fi m-1
" " "
b [
ram-1 i
Cm
Xr
L Xr+lJ (4.4-9)
(4.4-5) Taking the transpose of (4.4-9) yields
(4.4-10)
= Ok+l ..... Om = 0 and Ok = any non-zero
function or constant, is not useful for further reduc- 0 0
tion by singular elimination. Premultiplying (4.3-4) by (4.4-10) yields the
equivalent matrix
[: y:l[
Example
The same problem solved by systematic elimination
in Section 4.3 is now treated by singular elimination. y2 y2 ~" 0
~ (y+l)(y-Yl)(y2-y-11
xy 2 + y +yz =0"1
-I y2_~ i -1 y2-2 J
xy + y 2 z + I ~f (4.3-2)
x +y2-z-2 = so that all of the solution sets of the original prob-
lem must satisfy
The linear form augmented matrix corresponding to
(y+l) (y-l) (y2-y-1) = 0.
these equations is
This result is identical with the results obtained when
y2 1 y 0 using systematic elimination, as shown in Section 4.3.
y 0 y2 (4.3-3)
4. 5. Triang,11ar elimination
i y -i
Triangular elimination is an alternate reduction tech-
Operating on the second column yields nique for treating linear form matrices. As with
y2 y y ] singular elimination this technique can also reduce n
equations in n variables to I equations in I variables,
Y y2 1 J (4.3-4) where I is the number of variables contained in the
1 -1 y2_2 linear form matrix. Triangular elimination is a repet-
itive procedure mechanically similar to the Gauss
A set of 0's for carrying out the singular elimination
reduction in linear algebra and may therefore be the
is defined by best of the linear form reduction techniques for
computer implementation.
y2 Y _t Algorithm
[¢i 02 ¢31 Y y2 = 0
1 -1 The algorithm for performing triangular elimination
is as follows :
Before solving for the O's, it is convenient to re- (1) The augmented linear form matrix which re-
arrange the above matrix equation to presents the set of n equations in n unknowns
is written down.
(2) The linear form matrix rearrangement rule (Sec-
1 -1 1 tion 4.1.1) is used to minimize the number of
[03 02 01 ] y y2 = 0,
variables which appear in the matrix.
y2 y (3) Perform any number of column operations (Sec-
tion 4.1.2). It is usually advantageous to perform
Taking the transpose and assuming 02 = - 1 yields
I column operations on the columns correspond-
the following matrix associated with the unknowns ing to the I variabhs which appear in the matrix.
03 and OI (4) Rearrange the rows such that the 1, 1 element of
the matrix becomes the simplest possible non-
_ y y2 zero function, preferably a non-zero constant.
Denote this column operated augmented matrix
Reducing this matrix by a process similar to Gauss- by A~1] .
Jordan reduction yields
1 0 ....
I
so that
0 1
411J 1 41J ....
01 = 1 411 ,i 1 ....
02 = -1
03 = y_y2
nl .... tim nxm
_L(~?
)roduces the result
d nxm
yields
1 fl] ] 0 0 o 4f f[21
23
4lj ,~] 0 f[1]
21 f[l] fill
22 23 ,2t~ -~3f4f Iti '3(f ,~2j f[21
3m
_f[1131O~ fill fll]32 fi~ ]
31
Cf f[21
_-'fnl
[ 1] fill L[l] ~[1]
nm
nlTl
12 13 "'"
23 "'"
o f;2 f;.3 f,
2m i o4]] ,[~1
5'3
4~
f" "" 3m
2 3 f, 4 J" (4.5-6)
3m
0 0 f"n3 "'" f"nm~
0 f' f' f~ where (4.5-4) C (4.5-6).
- n2 n3 nl-n (4.5-3)
where .[21 o]
t22 =
(4.5-1) ,,,C (4.5-3)
lql]=0, (8) Continuing in this manner it is possible to trans-
form the original matrix (4.5-1) into a dominant
matrix which has only zeros in its first m-1 col-
(6) Rearrange the 2nd through nth rows of the umns below the principle diagonal. Thus
matrix on the fight hand side of (4.5-3) such W"
that the 2, 2 element becomes the simplest poss- fill f[1] f[ll f[l]
ible non-zero function, preferably a non-zero 12 13 "'" lm
constant. Denote this matrix by A[2]. C
C
'"]~.~ 4]? ,~]l... 4~ ~,iI]f[2]
22 "'"
• .. ,i~]] 31 32 "'" f[m-ll
m-lm-1
- Ol
-
0 f[21224 ~ ] •.. 4f" fill fill ~ ] fill
nl n2 "'" nm
f[2] f[2] f[2] (4.5- 4)
32 33 "'" 3n
o ff ~[3]... f[2]
nlTl
C "'4.
rim:l] .
m-xm-i
f[m-1 ]
m-lm
fire1
mm
I y2+y
y2+y
l_(y2_2)y
y_y2 (y2_2)
matrix
L f[m]
(4.5-7)
mm --I 0 0 1 -1 y2_1
If in step (3) I column operations are performed,
then the solution sets of the matrix on the right l~ 1 0 y2+y l_(y2_2)y
hand side of (4.5-7) are found by solving the -1 1_ y2+y y_y2(y2_2 )
n - m + t = I equations containing t variables
£[m~ = 0
.rD_rD_ y2-i
f[m] = 0 (4.5-8) "" 0 y2+y i-(y2-2)y (4.5-11)
m+lm
t0 0 -(y+ l)(y-l)(y2-y-1)
fim] = 0 Since it has been possible to maintain equivalence
throughout these transformations, the solution sets
and substituting back into the matrix on the right of the original problem must satisfy
hand side of (4.5-7), providing none of the diagonal
dements Ln the first m-1 columns of the matrix on (y + 1) (y- 1)(y2-y- 1) = 0 (4.5-12)
the right hand side of (4.5-7) are zero. i. e.,
(9) Any solution set of (4.5-8) which makes one or
more of the diagonal elements of the matrix on y = 1 y = -1 y = 1.618 y = -0.618
~che right hand side of (4.5-7) zero is tested by Substituting into the matrix on the right hand side
substituting it back into the original column of (4.5-11) yields the four solution sets expressed by
operated augmented matrix A[cl]. Such solution (4.3-11) and (4.3-12).
sets of (4.5-8) are accepted or rejected on the
basis of rank according to Theorem 4-1. 9. BIBLIOGRAPHY
I. Uspensky, J. V., "Theory of equations", First Edition,
Example McGraw-Hill Book Company, Inc., N e w York (1948).
The same problem solved by systematic elimination 2, Burnside, W. S., Panton, A. W., "Theory of equations",
in Section 4.3 and singular elimination in Section Third Edition, Hodges, Figgs, and Co., Dublin (1892).
4.4 y2 y 3. Aitken, A. C., "Determinants and matrices", Ninth Edi-
tion, Oliver and Boyd, Section 41, pp. 96-9? (1958).
4. Griffiths,L. W., "Introduction to the theory of equa-
tions", Second Edition, John Wiley and Sons, N e w York,
t_ 1 -1 y2_ 2 (4.3-4) pp. 218-220 (1947).
5. Chrystal, G., "Algebra", Third Edition, A d a m and Charles
is now treated by triangular elimination. Black, London (1892).
First, the order of the rows is changed to place the 6. Wu, C. L., "Nonlinear matrix algebra and engineering
third row first applications", Ph.D. Thesis, Case Institute of Technology
(1964).
1 -1 y2_ 2-
y y2 (4.5-9)
y2 Y Y
A
Second, premultiplying by a transformation matrix