Algebraic and Matrix Form
Algebraic and Matrix Form
1 INTRODUCTION
2 PREMLIMINARIES
3 DIRECTED GRAPHS
4 RELATIONS
5 GROUPS
6 TREES
7 CONCLUSION
8 REFERENCES
1
Algebraic and Matrix form
Introdution:
Linear algebra is a very useful subject and its basic concepts arose and were
used in different areas of mathematics and its applications. It is therefore not
simplifying that the subject had its roots in such diverse field as number theory,
geometry, abstract algebra, analysis and physics. Among the elementary concepts of
linear algebra are linear equations, matrices determinants, linear transformations,
linear independence, dimensions, bilinear forms, quadratic forms and vector space.
Since these concepts are closely interconnected, several usually appear in a given
context and it’s often impossible to disengage them.
Linear algebra is the branch of the mathematics concerning vector spaces and
linear mappings between such spaces. It includes the study of lines, planes and
subspaces but is also concerned with properties of points with coordinates that
satisfy a linear equation form a hyper plane in an n- dimensional space. The
condition under which a set of n hyper planes intersect in a single point is an
important focus of study in linear algebra.
In chapter 2, linear and quadratic forms are satisfied, we examine “forms” that
is functions from a vector space V to its field, which are either linear or quadratic.
The linear forms comprise the dual space of V.
2
In chapter 4, introduce and discuss the problem of finding canonical form s
for matrices with respect to similarity via representation of linear transformation.
3
CHAPTER - I
BASIC DEFINITION
Definition
That is,
( , )= ( , )+ ( , )
( , , ,
(c , , , ∀ and c F
Definition
(x) = XT AX = ∑ , ij xi xj
Definition
Example
1 2
( x1, x2 ) = (x1)2 + 6x1x2 + 2(x2)2
4 2
4
Definition
Definition
⋮ ⋮ …… ⋮
k = 1, 2,…,n
Definition
G1 Associative law :
∀ , , in G
5
G2 Unit element:
G contains at least one element 1, called a right unit element with the property
1 = ∀ in G.
G3 Inverse element:
For each unit element 1 and every element in G the equation = 1 has a
solution in G.
Definition
Example
6
The multiplication tables for the set follows,
f1 f2 f3 f4 f5 f6
f1 f1 f2 f3 f4 f5 f6
f 2 f1 f2 f3 f4 f5 f6
f 3 f1 f2 f3 f4 f5 f6
f 4 f1 f2 f3 f4 f5 f6
f 5 f1 f2 f3 f4 f5 f6
f 6 f1 f2 f3 f4 f5 f6
The cosets of N in G are N and f4N, which form a cyclic group of two
elements.
Definition
The set of all vectors in V such that A = 0 is called the null space of A and
denoted by N(A).
7
Definition
Definition
The selection of a basis { 1, 2, ..., n } for a vector space V over F and the
designation of vectors by there – coordinates, the linear transformation →A
on V assumes the concrete form X → AX, where A = ( A ; )
V X Vn(F)
A AX
8
Definition
Definition
Definition
Definition
Definition
m-1
The companion matrix of a monic polynomial f( ) = a0 + a1 +...+ am-1 +
m
is the matrix
9
0 0 0 …0
1 0 0 …0
0 1 0 …0
C(f) = ……………………
……………………
……………………
0 0 0 …1
Which has the negatives of the coefficient of f in the last column, i’s in the
diagonal just below the principal diagonal and 0’s elsewhere.
10
CHAPTER – II
Definition:
f = ( a1,a2,. . ., an )
V = ( x1,x2, . . . , xn)T
Definition:
Linear forms can be added and multiplied by scalars in the obvious way
(c f) V = c f(V)
So they form a vector space, which is called the dual space of V and is denoted
by V*.
11
Proposition
Proof:
If (V1,. . . ,Vn) is a basis for V and a1, . . . an are any scalars, then there is a
unique linear map f with the property that f(Vi) = ai for i=1,2, . . . n.
1
fi(Vj) =
0
Then ( f1,. . . . fn) form a basis for V* , the linear form f defined in the preceding
is a1f1 + . . . +anfn.
Then basis is called the dual basis of V* corresponding to the given basis for
V.
∴ It has n elements.
Dim(V*) = n = dim(V)
Definition:
The Kronecker delta ij for i,j ∈ { 1,2, . . . n } is defined by the rule that
1
ij =
0
12
2.1.1 Adjoint
Definition:
Proposition:
Let : V → W be a linear map. Choose bases B for V and C for W and let A
be the matrix representing relative to the bases. Let B* and C* denote the dual
bases of V* and W* corresponding to B and C.
*
Then the matrix representing relative to the basis C* and B* is the transpose
of A, that is AT.
Proposition:
Let B and B' be bases for V and B* and (B)* the dual bases of the dual space.
Then
*, (B')* = ( , B')-1
Proof:
If P = PB,B' has (i,j) entry Pi.j and Q = *, (B')* has (i,j) entry Qi.j
We’ve =∑
=∑
And so = ( )
13
= ∑ ∑
=∑ ∑
=∑
We’ve I = QTp
QT = p-1
Q = (p -1)T
Q = (pT)-1 as required.
Definition:
∑ ∑
Proposition:
Definition:
Two symmetric matrices A . A' over a field k are congruent if A' = PTAP for
some invertible matrix P.
14
Proposition:
Two symmetric matrices are congruent iff they represent the same quadratic
form with respect to different bases.
Theorem:
Proof:
∑ ∑
Now, assume that the theorem is true for forms in n-1 variables,
Case (i):
15
y1 = x1 + ∑ ⁄
Then we’ve
By induction,
Case (ii)
= )
So that, = ( = (
We obtain a new form for q which does contain a non-zero diagonal term.
Case (iii)
All aij are zero. Now q is the zero form and there is nothing to prove
Example
16
(x+y+2z)2 = x2 + y2 + 4z2 + 2xy + 4yz + 4xz.
= (x+y+2z)2- 4yz
= u2 + v2 – w2
u = x +y + 2z
v =y–z
w =y+z
a11 = coefficient of x2 =1
a22 = coefficient of y2 =1
a33 = coefficient of z2 =4
a12 = ( coefficient of xy ) = = 1
17
a13 = ( coefficient of xz ) = = 2
a21 = ( coefficient of yx ) = = 1
a23 = ( coefficient of yz ) = 0
a31 = ( coefficient of zx ) = = 2
a32 = ( coefficient of zy ) = 0
1 1 2
A= 1 1 0
2 0 4
1 1 2
~ 0 0 2 R2 → R2 – R1 , R3 → R3 – 2R1
0 2 0
1 2 1
~ 0 2 0 C3 ~ C2
0 0 2
1 2 1
~ 0 1 0 2 , R3 → 2
0 0 1
1 0 0
~ 0 1 0 C3 → C2 – 2C1 , C3 → C3 – C1
0 0 1
1 0 0
A' ~ 0 1 0
0 0 1
By a linear substitution. But this is still not a “canonical form for congruence”.
Definition:
b(V,cw) = cb(V,w)
b is symmetric
b(v,w) = b(w,v) ∀ v,w ∈ V
b) Let q : V → k be a function, q is a quadratic form,
-q(cv) = c2q(v) ∀ c ∈ k, v ∈ V
Is a bilinear form on V.
Remarks:
The bilinear form in the second part is symmetric, and the division by 2 in the
definition is permissible because of our assumption that the characteristics of k is
not 2. The 1st equation
19
(cx)2 = c2x2
((x+y)2 – x2-y2 ) = xy
Note:
Theorem:
0
0 0
Proof:
r = rank (PTAP)
= rank (A).
20
CHAPTER - III
Definition;
That is,
( , ) = ( , )+ ( , )
( , , ,
(c , , , ∀ and c F
Theorem:
, ( X,Y)
= ∑, ij xi yj
= xT Ay
21
Proof:
Define as follows
X= Y=
Then X, Y ) = ,
Thus, , X, Y)
=f(∑ i
i, ∑ j
j)
=∑, i
+( i, j) yj
=∑ i
i j
Definition:
Two bilinear forms are equivalent iff they are representation of one and the
same bilinear function.
Theorem:
Two bilinear forms are equivalent iff their respective matrices A and B are
equivalent ( A ~ B) or alternatively iff the forms have the same rank.
Proof:
= PT = Q
X = P TU Y = QV
( , ) = , (X, Y)
= XT AY
= (PT U)T A (Q V)
= P UT AQV
= UT (PAQ) V
= (U, V)
Conversely, Suppose that f and g are the two bilinear forms with
f (x,y) = XTAY
g (x,y) = XTBY
B = PAQ
23
where P and Q are non-singular.
Thus f and g are equivalent since both are representation of the bilinear
function f.
Theorem:
Proof:
X = BU ∴ B is non singular.
24
Theorem:
Example
Reduce the form f = 2x'y' – 3x'y2 + x'y3 – x2y' + 5x2y3 – 6x3y' + 3x3y2 +
19x3y3 to the form uivi by a substitution having rational coefficient.
The matrix of f is
2 3 1
A= 1 0 5
6 3 19
PAQ = diag(Ir,0)
1 0 5
≃ 2 3 1 R1 ⟷ R2
6 3 19
1 0 5
≃ 0 3 11 R2 → R2 – 2R1, R3 → R3 – 6R1
0 3 11
25
1 0 5
≃ 0 3 11 R1 → R1 (-1), R3 → R3 + R2
0 0 0
1 0 5
≃ 0 11
1 3 R2 ⟶ 3
0 0 0
1 0 0
≃ 0 1 11
3 C3 → C3 + 5C1
0 0 0
1 0 0
≃ 0 1 0 C3 → C3 + 11 3 C2
0 0 0
0 1 0
P⟶ 1 0 0
0 0 1
0 1 0
≃ 1 0 0 R2 → R2 – 2R1
0 0 1
0 1 0
≃ 1 2 0 R3 → R3 – 6R1
0 6 1
0 1 0
≃ 1 2 0 R1 → R1 (-1)
0 6 1
0 1 0
≃ 1 2 0 R3 → R3 + R2
1 4 1
0 1 0
≃ 1 2
3 3 0 R2 ⟶ 3
1 4 1
26
1 0 5
Q:I⟶ 0 1 11
3
0 0 1
Theorem:
Proof:
A = ½ ( A + AT ) + ½ ( A - AT )
= B +C
Where B = ½ ( A + AT )
C = ½ ( A - AT )
∴ (A AT ) T = A AT
=B1 - C1
= 2B1
27
And A – AT = B1 + C1 - B1 + C1
= 2C1
Lemma:
Dk ( B )= Dk ( A ) ∀k
Proof :
Consequently,
It sis easily seen that they are adequate to reduce a symmetric matric of
positive rank to an essential diagonal from which, although not unique, will play
the role of the preferred form in our exposition.
28
Theorem:
Proof:
0
(a) Or A(a) = ,a≠0
0
It is only necessary to show that the matrix A (a) has his property.
The row transformation R12 (1) and its column analogue, followed by R21
( 1 2) and its column analogue.
Theorem:
A symmetric matric A of rank r >0 has at least one non-zero principal minor
of order r.
29
Proof:
0
A(a) = a≠0
0
Since Dr ( A* ) 0.
Dk ( B )= Dk ( A ) for all K.
30
Theorem:
Proof :
Where each A,
0
A(a) = 0
0
Dr (B) 0.
Dr (A*) 0 and if
r (A*) > r
31
QUADRATIC FUNCTIONS AND FORMS OVER THE REAL FIELD
Theorem:
Proof:
With a change in of the basis vector we may assume d1,d2, . . . dP > 0 and
dP+1 . . . dr < 0
P = PT
To prove,
(w1)2 + (w2)2 + . . . + ( )2 - ( +1 2
) - . . . – (w1)2
32
Hence that (1) and (2) are equivalent forms.
Now,
w1 = . . . = = =...= =0
33
CHAPTER - IV
Theorem:
The one – one corresponding between the set of all linear function on V to
W and the set of all m x n matrix over F. [ where m = d (w), n = d (V) ] that is
defined when basis are chosen for v and w is an isomorphism between the vector
space of linear function and that of matrices.
Proof:
(BA) = B (A )
When basis are chosen for the spaces involved, then BA corresponds to BA
relative to the same bases.
A applied to X products Y
34
The word “image” stems from the interpretation of a linear transformation of
the vector into the vector
Theorem:
Proof:
Then A = iff
= X, = Y
To show that,
(ii) 1 ≠ 2 implies A 1 ≠A 2
Define = A-1Y
35
Then A =
(ii) A 1 =A 2
Where i = Xi , i = 1,2, . . .
Conversely,
Theorem:
Proof:
A (B ) = (AB)
=I =n
B (A = (BA)
= (I ) =
The 1st result asserts that any vector is the image A of some vector = B
36
The 2nd result asserts that if A = then B = , which ⇒ that is the
image of only one vector.
∴ A is one – one.
Y=AX
We call this the alias interpretation of AX=Y, the vector with co-ordinates X
is re-labelled with co-ordinates Y.
We call this the alibi interpretation of AX=Y, the vector with co-ordinates X
is carried into the vector with co-ordinates Y=AX.
Basis and basis are given a linear function Aon V is defined by the
equation A =
A(
37
Groups of transformation in general.
Theorem: (Cayley)
Proof:
The set S of objects upon which the correspondence shall be defined the set
G of group elements {a, b,. . . }.
ax = ax all x in G
( a b)x = a( b x)
= a (bx)
= a (bx)
= (ab) x
= ab x
38
a x = ex = x ∀ x in G.
The presence in G of an inverse a-1 for each small a in G ensures the presence in of
an inverse for a.
( a a
-1
) x = (a a -1) x = x = e x ∀ x in G.
a→ a
a= b
⇒a e =b e (or)
⇒a=b
a b= ab
Hence proved
Theorem:
39
Proof:
Then s +t =d(V)
∵ Ci(A i)=0
A Ci i=0 (or)
Ci I is in N(A)
Hence proved
Theorem:
i) A is non-singular.
ii) R(A) =V.
iii) N(A) = 0
40
Proof:
A 1 =A 2
A( 1 - 2)=0
Consequently
(ii) ⇒(iii)
1 - 2=0
⇒ 1 = 2
V= S1⊕S2
of two subspaces. So that every vector has a unique representation in the form
= 1 + 2 i in Si
41
It is accidental that
V = R(E) ⨁ N(E)
In general it need not even be true that these sub spaces are disjoint
Theorem:
Proof:
A2 = AA
= A (A )
= A ( 1)
= 1
=A
And A is idempotent.
42
V= S1⊕S2
= A + (I-A)
= 1+ 2
A 1 = A2 = A = 1
A 2 = A (I-A)
= (A-A2)
=0
Thus is in S.
V= S1+S2
Hence V= S1⊕S2
Theorem:
Proof:
43
Extend this to a basis { 1 2,… n } for E.
= i i=1,2,…,r
It is linearly independent
If ci i = 0,
Then A ci i = ci i = 0.
So that, c1=c2=….=cr=0
A1 i = i i=1,2,3…n.
0 1
0
44
Thus A1 is an interchange of the two basis vectors, and A1D is the projection
followed by A1-1=A1
45
CHAPTER - V
Lemma
If f and g are any two members of F [ ] such that g/f, then N (g) ⊆ N (f)
Proof:
= h(A) (g(A))
=0
So that is in N (f)
Lemma
Proof:
Then N (d) ⊆ D
If f and g are any two members of F [ ] such that g/f, then N (g) ⊆ N(f)
46
If is any vector in D.
So that is in N (d).
Lemma:
Proof:
Let S denote the right – hand member, which is the subspace of V consisting
of all sums.
=∑ i i= N (fi)
Then N (h) ⊇ S.
Since
h = hifi i = 1, 2,…r.
∑ i (A) hi (A) = I
I =
47
=∑ i (A) hi (A))
= Si (A) (h (A))
=0
Thus N (h) ⊆ S.
Theorem
Proof:
So that N (f) is the sum of the various null spaces N (fi). Let
{ f1 f2 … fr } denote a finite set of polynomial in F[ ] and h a l.e.m then
0 = N(fk) ∩ N( f i) or
48
0 = N(fk) ∩ N(fi)
∑ iA
i
=0
f(A) ai I =0
That is f (A) = 0
Theorem
49
Proof:
The remaining requirements for an ideal are quickly verified so that every
ideal E in F ( ) is a principal ideal. If E ≠ (0), the generator of E may be
characterized within associates as a polynomial of least degree in E.
Theorem:
The set of characteristics value of A coincides with the set of distinct roots
in F of the minimum function, m ( ) of A.
Proof:
Then A or
(A-c I) =0
⇒ 0 = m (A)
= m (C)
50
Conversely,
If C is a root in F of m ( ),
m( ) = q( ) ( -c)
= q(A) 0
The equation,
= (A-cI)
Lemma:
Proof:
Theorem:
51
which divides its successor, hi is precisely the ith invariant factor of M. Two
polynomial matrices are equivalent iff they have the same invariant factors.
Proof:
Conversely,
If the invariant factors of M and N coincide, they are equivalent to one and
the same diagonal matrix which displays three factors, and hence equivalent to
each other.
Theorem:
Proof:
Let P denote any irreducible factor of any one of the fi’s are arrange the fi’s
according to ascending power of P.
g1 ... gr
Then the highest power of P which divides di (M) = di (D) has exponent
k1 + k2 + … + ki
52
Hence the highest power of P which divides hi = di / di-1 is pki so that for each
I such that ki > 0. Pki is an elementary divisor of D and M.
A repetition of this argument for each prime which divides some fi shows
that every prime power factor of every fi is an elementary divisor of M.
Theorem:
a matrix A with non-trivial similarity factors hk, hk+1, … hn is similar to the matrix
(R1)
Proof:
A (Am-1 ) = Am
m-1
=(- 0 I– 1A -…- m-1A )
m-1
= (- 0 1A -…- m-1A
Since h(A) = 0 + 1A + … + Am =0
53
Thus in space S=( 1A 1 … Am-1 )
A is represented by c( h).
Theorem
Proof:
Hence in particular,
hi(A) i = hi(Ai) I =0
It follows that hi (Ai) = 0 on Si. Since hi (Ai) is seen to carry each basis
vector Aij I for Si into zero.
∴ mi = hi
det ( I – c (hi)).
h( =∑
⋯⋯⋯
⋯⋯⋯
⋯⋯⋯
det ( I – c(h)) = det ⋯⋯⋯⋯⋯⋯⋯⋯⋯⋯⋯⋯⋯
⋯⋯⋯⋯
⋯⋯
times each row is added to the preceding one, beginning with the last row
the desired conclusion is immediate.
Theorem
Proof:
Conversely,
Hence m = hn.
hkhk+1…hn = f.
55
CHAPTER – VI
Definition
= { w ∈ V : w.u = 0 ∀ u ∈ U }
Proposition
Proof:
Now choose a basis for U and extend it to a basis for V. Then apply the Gram-
Schmidt process to this basis to obtain an orthogonal basis (v1… vn). Since the
process only modifies vectors by adding multiples of earlier vectors, the first r
vectors in the resulting basis will form an orthogonal basis for U. The last n-r vectors
will be orthogonal to U, and so lie in ; and they are clearly linearly independent.
Now suppose that w ∈ and w = civi where (v1…vn) is the orthogonal basis we
constructed. Then ci = w.vi = 0 for i = 1,…,r; so w is a linear combination of the last
n-r basis vectors, which thus form a basis of . Hence dim( ) = n - r as required.
Since we have a basis for V which is a disjoint union bases for U and
56
Definition:
Proposition:
Proof:
We know that V = Ker( ) ⊕ Im( we only have to show that these two
subspaces are orthogonal. So take v ∈ Ker( , so that (v) = 0 and w ∈ Im( , so
that w = (u) for some u ∈ V. Then
Theorem
Theorem:
57
Proof:
= (A ̅ )T v
= ̅ Av
= ̅ T ( v)
Now since has a real eigenvalues, we can choose a real eigenvectors v, and
we can assume that |v| = 1.
58
Let U be the subspace = { u ∈ V : v.u = 0 }. This is a subspace of V of
dimension n-1. We claim that : U → U. for take u ∈ U. then
Remark:
The theorem is almost a canonical form for real symmetric relations under the
relation of orthogonal congruence. If we require that the eigenvalues occur in
decreasing order down the diagonal, then the result is a true canonical form each
matrix is orthogonally similar to a unique diagonal matrix with this property.
Corollary:
Proof:
This follows from the theorem, but is easily proved directly. If (v) = v and
(w) = w, then
59
Example
10 2 2
Let A = 2 13 4
2 4 13
10 2 2
2 13 4 = (x-9)2 (x-18)
2 4 13
10 2 2 18
2 13 4 = 18
2 4 13 18
10 2 2 9
2 13 4 = 9
2 4 13 9
60
1⁄3 0 4⁄3√2
P = 2⁄3 1⁄√2 1⁄3√2
2⁄3 1⁄√2 1⁄3√2
18 0 0
T
P AP = 0 9 0
0 0 9
You might like to check that the orthogonal matrix in example in the last
chapter of the notes also diagonalises A.
Proposition:
Proof:
Then
SIMULTANEOUS DIAGONALISATION
61
Theorem:
Proof:
And
As required.
The diagonal entries of D are the eigenvalues of C, that is, the roots of the
equation det (x I - C) = 0. Now we have
And det (P1T) = det (P1) is non-zero; so the polynomials det (x A - B) and
det (x I - C) are non-zero multiples of each other and so have the same roots.
62
Theorem:
Proof:
Statement (b) is just a translation of (a) into matrix terms; so we prove (a).
V = U1⊕ … ⊕ Ur,
We can claim that maps Ui to Ui. for take (u) = iu. Then
iu )= I ,
Now is a self- adjoint linear map on the inner product space Ui, and so by
the spectral theorem again, Ui has an orthonormal bases consisting of eigenvectors
of . But these vectors are also eigenvectors of , since they belong to Ui.
63
Finally, since we have an orthogonal decomposition, putting together all these
bases gives us an orthonormal bases of V consisting of simultaneous eigenvectors
of and .
Remark:
This theorem easily extends to an arbitrary set of real symmetric matrices such
that any two commute. For a finite set, the proof is by induction on the number of
matrices in the set, based on the proof just given. For an infinite set, we use the fact
that they span a finite-dimensional subspace of the space of all real symmetric
matrices; to diagonalise all the matrices in our set, it suffices to diagonalise the
matrices in a bases.
64
Conclusion
Finally, it may review a theory of methods and techniques most suitable for
the study, simply by looking at the topic and by evaluating their suitability and
effectiveness.
65
Bibliography
66