MTH 2202 Linear Algebra II Lecture Notes 2020-2021 (ICIE)
MTH 2202 Linear Algebra II Lecture Notes 2020-2021 (ICIE)
LINEAR ALGEBRA II
(MTH 2202)
LECTURE NOTES
2020/2021 SESSION
(Course Lecturer)
0
Chapter 1
Introduction
1.1 Preamble
Linear Algebra is one of the most important mathematical tools used in the world today. It is
an excellent example of mathematical abstraction, the process which starts with the
observation that apparently different mathematical objects share common properties. The
common rules which these objects obey are then identified, and formalized as axioms. Then
an abstract object which satisfies these axioms is studied, with no further assumption being
made. Of course any deduction which we can draw from these axioms then holds in any
mathematical object which satisfies them, and so we have proved many apparently different
results all at the same time. Although linear algebra is properly thought of as a branch of pure
Mathematics, it is immensely widely used in all branches of science and engineering, and in
the social sciences. Indeed, solving systems of linear equations with very large numbers of
unknowns is so crucially important in some industries that mathematicians are employed to
find faster and faster (computer-based) methods of solution.
1
1.3 Recommended Texts
S.I. Grossman, Elementary Linear Algebra, (Second Edition), Wadsworth California,
1984
R. Larson, Elementary Linear Algebra, (7th Edition), Brooks/Cole, USA, 2013
D.Poole, Linear Algebra: A modern Introduction, (4th Edition), Cengage Learning,
USA, 2015.
K. Singh, Linear Algebra Step by Step, Oxford University Press, United Kingdom,
2014.
R.Kaye, R. Wilson, Linear Algebra, Oxford University Press, 2003.
S. Lipschutz, Schaum’s Outline Series Theory and Problems of Linear Algebra,
(Fifth Edition), McGraw-Hill, USA, 2013.
2
Chapter 2
2.1 Introduction
In this chapter we will study system of linear equations and describe methods for finding all
solutions (if any) to a system of m equations in n unknowns (variables).
2.2 Definitions
A linear equation is an equation of the forma1 x1 a2 x2 ... am xn b
where a1 , a2 ,..., am , b and x1 , x2 ,..., xn are variables. The scalars a j are coefficients and the
scalar b is the constant term. The number a1 is the leading coefficient, and x1 is the leading
variable.
A system of linear equations, or linear system, is a set of one or more linear equations in
the same variables, such as
a11 x1 a12 x2 ... a1n xn b1
a21 x1 a22 x2 ... a2 n xn b2
...
am1 x1 am 2 x2 ... amn xn bm .
3
The matrix
a11 . . . a1n
. .
A . .
. . . . .
am1 . . . amn
b1
.
b .
.
bm
system.
A solution of the system is an n-tuple (c1 , c2 ,..., cn ) such that letting x j c j for each
j satisfies every equation. The solution set of the system is the set of all solutions. It is
possible for a system of linear equations to have exactly one solution, infinitely many
solutions, or no solution. A system of linear equations is consistent if there exists at least one
solution; otherwise it is inconsistent.
Two systems of linear equations are equivalent if they have the same solution set.
4
discuss a determinant method for solving system of linear equations when the number of
unknowns is equal to the number of the equations generally known as Cramer’s methods.
Definition 2.1 Let A be a matrix with m rows. When a row of A is not zero, its first nonzero
entry is the leading entry of the row. The matrix A is in row echelon form (REF) when the
following two conditions are met:
1. Any zero rows are below all nonzero rows.
2. For each nonzero row i, i ≤ m − 1, either row i + 1 is zero or the leading entry of row
i + 1 is in a column to the right of the column of the leading entry in row i.
The matrix A is in reduced row echelon form (RREF) if it is in row echelon form and the
following third condition is also met:
3. If is the leading entry in row i, then , and every entry of column k other
than is zero
Example 2.1 The following matrices are in row echelon form (REF):
1 2 3 1 1 6 4
i.
0 1 5 ii.
0 1 2 8
0 0 1 0 0 0 1
1 1 6 4
0 1 2 8 1 2
ii. iv. 0 1
0 0 0 1
5
Example 2.2 The following matrices are in reduced row echelon form (RREF):
1 0 0 1 0 0 0
i.
0 1 0 ii.
0 1 0 0
0 0 1 0 0 0 1
1 0 1 0 0 5
iii. 0 1 iv. 0 0 1 2
Question: How can we convert a given matrix to row echelon form (REF) and reduced row
echelon form (RREF)?
Answer: We apply the elementary row operation.
6
To solve the system of linear equations
a11 x1 a12 x2 ... a1n xn b1
a21 x1 a22 x2 ... a2n xn b2
...
am1 x1 am 2 x2 ... amn xn bm .
we write the system as an augmented matrix and row-reduce the matrix to its echelon form.
After reducing the augmented form, we observe one of the following
Example 2.3 Solve the following systems using the Gaussian elimination method.
2 x1 4 x2 6 x3 18 2 x1 4 x2 6 x3 18
1. 4 x1 5 x2 6 x3 24 2. 4 x1 5 x2 6 x3 24
3 x1 x2 2 x3 4 2 x1 7 x2 12 x3 30
2 x1 4 x2 6 x3 18
3. 4 x1 5 x2 6 x3 24
2 x1 7 x2 12 x3 40
Solution
7
2 4 6 18 1 2 3 9 R2' R2 4 R1 1 2 3 9
R1 12 R3' R3 3 R1
4 5 6 24 4 5 6 24 0 3 6 12
3 1 2 4 3 1 2 4 0 5 11 23
1 2 3 9 1 2 3 9
R3' R3 5 R2
R2' R2 ( 13 )
0 1 2 4 0 1 2 4
0 5 11 23 0 0 1 3
x2 2 x3 4
x2 2(3) 4
x2 2
x1 2 x2 3x3 9
x1 2( 2) 3(3) 9
x1 4.
2 4 6 18 1 2 3 9 R2' R2 4 R1 1 2 3 9
R1 12 R3' R3 2 R1
4 5 6 24 4 5 6 24 0 3 6 12
2 7 12 30 2 7 12 4 0 3 6 12
1 2 3 9 1 2 3 9
R3' R3 3 R2
R2' R2 ( 13 )
0 1 2 4 0 1 2 4
0 3 6 12 0 0 0 0
x1 2 x2 3x3 9
x2 2 x3 4
This system has infinitely many solutions, because it has two equations and three
unknowns. That is the number of unknowns are more than that of the equation.
8
x1 x3 1 (Equation1 - 2Equation2)
x1 1 x3
From Equation1, we have
x2 4 2 x3
The solution set is (1 x3 , 4 2 x3 , x3 ).
For instance, if x3 0, we have (1,4,0), if x3 10, we have (11,-16,10) e.t.c.
2 4 6 18 1 2 3 9 R2' R2 4 R1 1 2 3 9
R1 12 R3' R3 2 R1
4 5 6 24 4 5 6 24 0 3 6 12
2 7 12 40 2 7 12 40 0 3 6 22
1 2 3 9 1 2 3 9
R3' R3 3 R2
R2' R2 ( 13 )
0 1 2 4 0 1 2 4
0 3 6 22 0 0 0 10
0 x1 0 x2 0 x3 10
4. The last equation now reads which is not possible. Thus the
system has no solution.
2 x1 3 x2 x3 a
x1 x2 3 x3 b
3 x1 7 x2 5 x3 c
Use Gaussian elimination method to find the condition on a,b,c such that the system is
inconsistent.
Solution
2 3 1 a 1 1 3 b R2' R2 2 R1 1 1 3 b
R1 R2 R3' R3 3 R1
1 1 3 b 2 3 1 a 0 5 7 a 2b
3 7 5 c 3 7 5 c 0 10 14 c 3b
9
1 1 3 b 1 1 3 b
a 2b
R2' R2 ( 15 ) R3' R3 10 R2
0 1 7
5 5 0 1 7
5
a 2b
5
0 10 14 c 3b 0 0 0 c b 2a
It follows from the last row that the system will be consistent if
c b 2a 0
cb
a .
2
x1 2 x2 3 x3 9
x1 3 x2 4
2 x1 5 x2 5 x3 17
Solution:
1 2 3 9 R2' R2 R1
1 2 3 9 1 2 3 9
R3' R3 2 R1 R3' R3 R2
1 3 0 4 0 1 3 5 0 1 3 5
2 5 5 17 0 1 1 1 0 0 2 4
1 2 3 9 1 0 9 19 R1' R1 9 R3 1 0 0 1
R1' R1 2 R2
R2' R2 ( 12 ) R2' R2 3 R3
0 1 3 5 0 1 3 5 0 1 0 1
0 0 1 2 0 0 1 2 0 0 1 2
The matrix is now in reduced row-echelon form (RREF). Converting back to a system
x1 1, x2 1 and x3 2
of linear equations, we have .
2 x1 x2 3x3 a
3 x1 x2 5 x3 b
5 x1 5 x2 21x3 c
10
2.4 Homogeneous system of equations
The general m n system of linear equation is called homogeneous if all the constants
b1 , b2 ,..., bm are zero. That is, the general homogeneous system is given by
For the general homogeneous system, there are two possibilities for its solution since
x1 x2 ... xn 0 is always a solution (called the trivial solution or zero solution). These
possibilities are either the zero solution is the only solution or there are an infinite number of
solutions in addition to the zero solution. Solutions other than the zero solution are called
nontrivial solutions.
Gaussian and Gauss-Jordan elimination methods described earlier are used to solve the
homogeneous system of equations.
2 x1 4 x2 6 x3 0
4 x1 5 x2 6 x3 0
3x1 x2 2 x3 0
Solution: We use Gauss-Jordan’ method as follows:
2 4 6 0 1 2 3 0 R2' R2 4 R1 1 2 3 0
R1' R1 ( 15 ) R3' R3 3 R1
4 5 6 0 4 5 6 0 0 3 6 0
3 1 2 0 3 1 2 0 0 5 11 0
1 2 3 0 R1' R1 2 R2 1 0 1 0 1 0 1 0
R3' R3 5 R2
R2' R2 ( 13 ) R3' R3 ( 1)
0 1 2 0 0 1 2 0 0 1 2 0
0 5 11 0 0 0 1 0 0 0 1 0
1 0 0 0
R1' R1 R3
R2' R2 2 R3
0 1 0 0
0 0 1 0
Thus the system has the trivial solution (0,0,0) as the only solution.
11
Example 2.7 Solve the following homogeneous systems
x1 2 x2 x3 0
3x1 3x2 2 x3 0
x1 11x2 6 x3 0
Solution: We use Gauss-Jordan’ method as follows:
1 2 1 0 R2' R2 3 R1
1 2 1 0 1 2 3 0
R3' R3 R1 R2' R2 ( 19 )
3 3 2 0 0 9 5 0 0 1 9 0
5
1 11 6 0 0 9 5 0 0 9 5 0
1 2 3 0
R3' R3 9 R3
0 1 59 0
0 0 0 0
The augmented matrix is now in reduced echelon form and, we can clearly see that, there are
( 19 x3 , 95 x3 , x3 ) x3 0
infinitely number of solutions given by . For instance, if we obtain the
trivial solution. If
x3 1 we obtain the solution ( 19 x3 , 95 x3 ,1)
x1 x2 x3 0
4 x1 2 x2 7 x3 0
Solution: We use Gauss-Jordan’ method as follows:
1 1 1 0 R2' R2 4 R1 1 1 1 0
4 2 7 0 0 6 11 0
R2' R2 ( 16 ) 1 1 1 0 R1' R1 R2 1 0 56 0
0 1 6 0 0 1 6 0
11 11
Thus, there are infinitely many solutions given by ( 6 x3 , 6 x3 , x3 ) . This follows from the fact
5 11
that the linear system contains two equations and three unknowns.
12
2.5. Determinant Method (Cramer’s rule)
In this section we will discuss the method of solving systems of linear equations with the
same number of unknowns as equation.
Ax b (2.5.2)
1
We suppose that det A 0 . The system (2.5.2) has a unique solution given by x A b. We
can develop a method for finding that solution without row reduction and without computing
A1.
That is Ai . , is the matrix obtained by replacing the ith column of A with b. Finally, let
D1 det A1 , D2 det A2 ,..., Dn det An .
Theorem 2.5.1: Let A be an n n matrix and suppose that det A 0 . Then the unique
solution to the system Ax b is given by
D1 D D D
x1 , x2 2 ,..., xi i ,..., xn n
D D D D
13
Example 2.9 Solve the following system using Cramer’s rule
x1 x2 x3 5
x1 2 x2 3 x3 1
2 x1 x2 x3 3
Solution
1 1 1
D 1 2 3 5 5 5 5
2 1 1
Because D 0 then the system has a unique solution. We now compute D1 , D2 and D3 .
5 1 1 1 5 1 1 1 5
D1 1 2 3 20, D2 1 1 3 10, D1 1 2 1 15
2 1 1 1 3 1 2 1 3
D3
x1 D1 4, x2 D2 2, x3 3
D D D
Exercise 2.2
1 0 0 0
0 1 0 0
(i) 0 1 1 2 (ii)
0 0 0 0 1 0 2 1
2 0 1 3 1 0 2 1
(ii) 0 1 1 4 (iv) 0 1 3 4
0 0 0 1 0 0 1 0
14
0 0 1 0 0 1 0 0 0
(v) 0 0 0 1 0 (iv) 0 0 0 1
0 0 0 2 0 0 0 0 0
2. Use Gaussian and Gauss-Jordan elimination methods to find all solutions, if any, to
the given systems
2 x1 x2 3x3 5 x1 2 x2 3x3 11
(i) 3 x1 2 x2 2 x3 5 (ii) 4 x1 x2 x3 4
5 x1 3x2 x3 16 2 x1 x2 3x3 10
x1 x2 x3 7 x1 x2 5 x3 3
(iii) 4 x1 x2 5 x3 4 (iv) x1 2 x3 1
2 x1 2 x2 3 x3 0 2 x1 x2 x3 0
2 x y z 2w 6
3x 4 y w 1
(v) 4 x 12 y 7 z 20w 22 (vi) x 5 y 2 z 6w 3
3 x 9 y 5 z 28w 30 5x 2 y z w 3
3, Determine the value of k, a, b and c (if possible) such that the system in the unknown
(a) a unique solution (b) no solution (c) more than one solution.
kx y z 1 x 2 y kz 1
(i) x ky z 1 (ii) 2 x ky 8 z 3
x y kz 1
x y 2 x y 0
y z 2 y z 0
(iii) x z2 (iii) x z0
ax by cz 0 ax by cz 0
15
4. Determine whether each of the following system has a nontrivial solution
x 3y 2z 0 x 3y 2z 0
(i) x 8 y 8z 0 (ii) 2x 3 y z 0
3x 2 y 4 z 0 3x 2 y 2 z 0
5. Find the value of for which the homogeneous linear system has nontrivial solution
2 x1 x2 x3 5 x1 x2 x3 7
(i) x1 2 x2 3 x3 0 (ii) 2 x1 5 x3 4
4 x1 x2 x3 1 3 x2 x3 2
16
Chapter 3
Change of basis
Learning Outcomes:
By the end of this chapter you should be able to
Write vectors in terms of the standard basis.
Write vectors in terms of other bases.
Compute the transition matrix
2 1 0 n
Recall that, in R we wrote vectors in terms of the standard basis , . In R we define
0 1
the standard basis e1 , e2 ,..., en where
e1 (1, 0,0,...),
e2 (0,1, 0,...),
...
en (0,0,0,...,1).
These bases are most commonly used because it is relatively easy to work with them. But it
sometimes happens that some other bases are more convenient. There are obviously many
bases to choose from since in an n-dimensional vector space any n linearly independent
vectors form a basis. In this chapter we shall see how to change from one basis to another by
computing a certain matrix.
1 0
Example 3.1. Let u1 and u2 . Then B1 u1 , u2 is the standard basis in R .
2
0 1
1 1
Let v1 and v2 . The set B2 v1, v2 is also a basis in R .
2
3 2
x 2
Let x 1 be a vector in R . This notation means that
x2
17
x 1 0
x 1 x1 x2 x1u1 x2u2 .
x2 0 1
That is, x can be written in terms of the vectors in the basis B1 . This is denoted as
x
( x) B1 1 .
x2
2
Since B2 is another basis in R , there are scalars c1 and c2 such that
c
( x) B2 1 .
c2
How do we find the scalars c1 and c2 ? To find the scalars c1 and c2 , we write the old basis
1 1 1
u1 25 35 52 v1 35 v2 (3.2)
0 3 2
and
0 1 1
u2 15 15 15 v1 51 v2 . (3.3)
1 3 2
That is,
2 1
(u1 ) B2 53 and (u2 ) B2 15 .
5 5
x x1u1 x2u2 x1 52 v1 35 v2 x2 15 v1 15 v2
52 x1 15 x2 v1 53 x1 15 x2 v2 .
18
c1 2
5 x1 15 x2
c2 53 x1 15 x2
or
c 2 x1 15 x2 25 1
x1
( x) B2 1 = 5 3 3
5
.
c2 5 x1 5 x2 5 x2
1 1
5
( x) B2 A( x) B1 (3.4)
2 1
where A 53 5
is called the transition matrix from B1 to B2 .
5
1
5
This example can easily be generalized. Let B1 = u1 , u 2 ,..., u n and B2 = v1 , v2 ,..., vn be two
base for an n-dimensional real vector space V. Let x V , then x can be written in terms of
the two bases:
and
x c1v1 c2 v2 ... cn vn
(3.6)
b1
b2
( x) B1
:
bn
To denote the representation of x in terms of the basis B1 . This is unambiguous because the
19
c1
c2
( x) B2
:
cn
( w1 w2 ) B1 ( w1 ) B1 ( w2 ) B1
n
That is, in the new notation we can add vectors just as we add vectors in in R . Moreover, it
is easy to show that
( w) B ( w) B
1 1
B2 uj B1 vi ' s
Now, since is a basis, each in can be written as a linear combination of the .
a1 j , a2 j ,..., anj
Thus, there exists a unique set of scalars such that for j 1, 2,..., n
u j a1 j v1 a2 j v2 ... anj vn
(3.7)
or
a1 j
a2 j
(u j ) B2
:
anj (3.8)
Definition 3.1: The n n matrix A whose columns are given by (3.8) is called the transition
Theorem 3.1 Let B1 and B2 be bases for a vector space V. Let A be the transition matrix
( x) B2 A( x) B1 . (3.9)
Proof
20
We shall use the representation of x given in (3.5) and (3.6)
Thus,
Theorem 3.2 If A is the transition matrix from B1 to B2 , then A1 is the transition matrix from
B2 to B1 .
Proof Let C be the transition matrix from B1 to B2 Then from (9), we have
( x) B1 C ( x) B2 (3.10)
( x) B1 CA( x) B1 (3.11)
But (3.11) hold only for every x in V only if CA=I. Thus, C A1 , and the result follows.
21
Now we give a simple procedure for determining the transition matrix from standard basis to
basis B2
1 3 0
Example 3.2 In R let B1 be the standard basis and let B2 0 , 1 , 1 . If
3
2 0 2
x
x y R 3 , write x in terms of the vectors in B2 .
z
1 3 0
0 1 1 80
2 0 2
1 3 0
C 0 1 1
2 0 2
22
2 6 3
1
A C 2 2 1
1
8
2 6 1
Therefore, if
x 2 6 3 x
1
( x) B1 y then ( x) B2 8 2 2 1 y .
z 2 6 1 z
For example, if
1 2 6 3 1 2 14
( x) B1 2 then ( x) B2 18 2 2 1 2 18 2 14 .
4 2 6 1 4 14 7
4
Example 3.3 In P2 (the set of polynomials of degree 2) the standard basis is B1 1, x, x 2 .
Another basis is B2 4 x 1, 2 x 2 x,3 x 2 3 . If p a0 a1 x a2 x 2 , write p in terms of the
polynomial in B2 .
c1 3c3 0
4c1 c2 0
2c2 3c3 0
23
1 0 3
4 1 0 27 0
0 2 3
1 0 3 3 6 3
C 4 1 0 , so that A C 1 1
27 12 3 12
0 2 3 8 2 1
a0
Is the transition matrix from B1 to B2 . Since (a0 a1 x a2 x ) B1 a1 ,we have
2
a
2
3 6 3 4 15 27
21
(5 x 3 x 4) B1 271 12 3 12 3 27
2
8 2 1 31
5 27
or 5 x 2 3 x 4 15
27
(4 x 1) 27
21
(2 x 2 x) 27
31
(3 x 2 3).
3 2 2 5 b
Example 3.4 Let B1 , and B2 , be two bases in 2 . If ( x) B1 1 ,
1 1 4 3 b2
write x in terms of the vectors in B2 .
24
Solution: This problem is more difficult compared to the other two examples, because
neither basis is the standard basis. We must write the vectors in B1 as linear combinations of
the vectors in B2 . That is, we must find constants a11 , a21 , a12 , a22 such that
3 2 5 2 2 5
a11 a21 and a12 a22 .
1 4 3 1 4 3
14 1
A 1
26
10 10
14 1 b1 261 (14b1 b2 )
and ( x) B1 1
10 .
10 10 b2 26 (b1 b2 )
26
7
For example let x (in standard basis). Then
4
7 3 2 3 2
b1 b2 3
4 B1 1 1 1 1
So that
7 3
4 B1 1
and
7 14 1 3 2641
1
26 20
4 B 2 10 10 1 26
That is
7 2 20 5
26
41
26
4 4 3
25
Exercise 3.1
x
1. In the following problems write 2 in terms of the given basis
y
1 1 5 3 a b
(a) , (b) , (c) , where ad dc 0
1 1 7 4 c d
x
2. Write y 3 in terms of the given basis
z
1 1 1 2 1 3 a b c
(a) 0 , 1 , 1 (b) 1 , 4 , 2 (c) 0 , d , e where adf 0
0 0 1 3 5 4 0 0 f
2 1 2
6. In 2 suppose that x , where B1 , . Write x in terms of
1 B1 1 3
0 5
the basis B2 , .
3 1
2 1 0 1
7. In , x 1 where B1 1 , 1 , 0 . Write x in terms of
3
4 0 1 1
B1
3 1 0
B2 0 , 2 , 1 .
0 1 5
26
Chapter 4
4.1 Eigenspace
Definition 4.1. Let V be a finite-dimensional vector space over , and let T : V V be a
linear transformation. A non-zero vector v V such that T (v) v for some is said to
be an eigenvector of T with eigenvalue . When is an eigenvalue of T, the set of all
eigenvectors of T belonging to is called the -eigenspace of T denoted by
V v V : T (v) v .
Definition 4.2. Let A be an n n matrix with real entries. The number is called eigenvalue
if there is a non-zero vector v such that
Av v
The non-zero vector v is called an eigenvector of A corresponding to the eigenvalue .
Proof Suppose that v, w V ; that is T (v) v and T ( w) w . Then for any scalars a, b ,
of V .
27
Example 4.1
1. Let V 2 . Let T : V V be the linear transformation defined by
M ( ) det( M I )
where I is the n n identity matrix. When V is a finite-dimensional vector space over and
T : V V is a linear transformation, A is the matrix of T with respect to some basis of V, we
define the characteristic polynomial of T denoted by T ( )
T ( ) det( A I ).
28
COUNTING MULTIPLICITIES: every n n matrix has n eigenvalues.
We now give the following three-step procedure for calculating eigenvalue and
corresponding eigenvectors.
i. Find T ( ) det( A I ) .
Example 4.2
1. Let V 2 , T : V V be the linear transformation defined by
T (a, b)t (3a b, a 3b) for all ( a, b) V . Then it is easy to check that T has
3 1
matrix A with respect to the basis (1,0) , (0,1) . Hence the characteristics
t t
1 3
polynomial of T is
3 1
T ( ) 2 6 8.
1 3
1 2, 2 4.
( A 1 I )v 0
3 1 1 0 x x
2 0; u .
1 3 0 1 y y
1 1 x
0
1 1 y
x y 0
x y 0
x 1, y 1 is the eigenvector corresponding to the eigenvalue 1 2.
29
Similarly, to find the eigenvector for the eigenvalue 2 4 , we solve
( A 2 I )v 0
3 1 1 0 x x
4 0; u .
1 3 0 1 y y
1 1 x
0
1 1 y
x y 0
x y 0
x 1, y 1 is the eigenvector corresponding to the eigenvalue 2 4.
2. Let S : V V be the linear transformation with S (a, b)t (b, a)t for all
0 1
(a, b)t V . With respect to the basis above, S has matrix A Hence S has
1 0
characteristic polynomial
1
S ( ) 2 1.
1
Since this polynomial has no root in , there is no eigenvector of S in V associated
to any eigenvalue. It should be clear from this example that the presence (or absence)
of eigenvalues depends very much on the field. If we had been working with the
complex field , then the polynomial 2 1 would have had the roots i and i.
Definition 4.4 Two n n matrices A and B are said to similar if there exists an invertible
n n matrix C such that
B C 1 AC.
Theorem 4.3 If A and B are similar n n matrices, then A and B have the same
characteristic equation and, therefore, have the same eigenvalues.
Proof Since A and B are similar, B C 1 AC and
30
det( B I ) det(C 1 AC I ) det C 1 AC C 1 ( I )C
det C 1 ( A I )C det(C 1 ) det( A I ) det(C )
det(C 1 ) det(C ) det( A I ) det(C 1C ) det( A I )
det I det( A I ) det( A I ).
These mean that A and B have the same characteristic equation and, since eigenvalues are
roots of the equation, they have the same eigenvalues.
This is very simple kind of matrix, and it would make our lives easy if we could choose a
basis with respect to which the linear transformation in question is diagonal.
n
Proof Let A aij be the matrix of T with respect to S. Then A is diagonal if and only if
i , j 1
T (v j ) aij v j a jj v j , j 1, 2,..., n.
i
Our next result gives a necessary and sufficient condition for the existence of a basis of
eigenvectors of a linear transformation.
dim V n.
i 1
i
31
Corollary 4.1 Let T : V V be a linear transformation, where V is a finite-dimensional
vector space of dimension n. Suppose that T has n distinct eigenvalues. Then there is basis of
V which consists of eigenvectors of T
T ( a, b, c)t 5a b c, a 5b c, a b,5c
t
for all (a, b, c )t V . With respect to the
5 1 1
1 5 1 ( A, say)
1 1 5
The characteristic polynomial of T is
T ( ) (4 )2 (7 ),
so that the eigenvalues of T are 4 and 7.
We proceed to find the eigenspace of T. Now T (( a, b, c )t ) (4a, 4b, 4c )t if and only if
5a b c, a 5b c, a b 5c 4a, 4b, 4c
t t
which happens if and only if a b c 0 . It
follows that 1, 1, 0 , 0,1, 1 is a basis for the 4-eigenspace of T.
t t
This happens only when a b c , so the 7-eigenspace of T has a basis 1,1,1 , and is one
t
diagonalizable.
Now with respect to this basis, the matrix of T is
32
4 0 0
B 0 4 0
0 0 7
1 0 1
P 1 1 1 .
0 1 1
Check that P 1 AP B.
1 4
Example 4.4 Let A .
2 3
Solution
1 4
A ( ) 0
2 3
2 4 5 0
1 1, 2 5.
( A 1 I )v 0
1 4 1 0 x x
(1) 0; u .
2 3 0 1 y y
33
2 4 x
0
2 4 y
2x 4 y 0
2x 4 y 0
( A 1 I )v 0
1 4 1 0 x x
5 0; u .
2 3 0 1 y y
4 4 x
0
3 3 y
4 x 4 y 0
3x 3 y 0
x 1, y 1 is the eigenvector corresponding to the eigenvalue 1 5.
2 1
P
1 1
The diagonal matrix D PAP 1 is a matrix whose entries in the main diagonal are the
eigenvalues of A. That is
1 0
D
0 5
1
2 1 1 4 2 1
1 1 0
D PAP
1 1 2 3 1 1 0 5
34
4.3 Powers of matrices
Theorem 4.6 If an n n matrix A is diagonalizable with P 1 AP D where D is a diagonal
matrix then
Am PD m P 1 .
PDP 1 P P 1 AP P 1
PP 1 A PP 1 IAI A.
Thus we have our result for m = 1 which is A PDP 1 . This means that we can factorize
matrix A into three matrices P, D and P 1 .
Ak PD k P 1
Step 3: We need to prove the result for m = k + 1, that is we need to prove Ak 1 PD k 1P 1 .
Ak 1 Ak A
PD k P 1P DP 1
PD k I DP 1
PD k DP 1
PD k 1 P 1.
35
1 2 3
Example 4.5 Find A , where A 0 2 5 .
5
0 0 3
1 2 7 2 4 13
1
Solution: Using the methods described above P 0 1 10 , P 0 2 10 . and
1
0 0 2
2 0 0 1
1 0 0
1
D P AP 0 2 0 . Using the above result, we have
0 0 3
A5 PD 5 P 1
5
1 2 7 1 0 0 2 4 13
1
0 1 10 0 2 0 0 2 10
0 0 2 2
0 0 3 0 0 1
1 2 7 1 0 0 2 4 13
5
1
0 1 10 0 25 0 0 2 10
2
0 0 2 0 0 35 0 0 1
1 2 7 1 0 0 2 4 13
1
0 1 10 0 32 0 0 2 10
2
0 0 2
0 0 243 0 0 1
1 62 537
Thus, A 0 32 1055 .
5
0 0 243
36
Exercise 4.1
2. For each matrix, find all eigenvalues and linearly independent eigenvectors:
2 2 4 2 5 1
i. A ii. B iii. C
1 3 3 3 1 3
Find invertible matrices P1 , P2 and P3 such that P11 AP1 , P21BP2 and P31CP3 are
diagonal.
4. For each matrix, find all eigenvalues and a basis for each eigenspace:
3 1 1 1 2 2 1 1 0
i. A 2 4 2 ii. B 1 2 1 iii. C 0 1 0
1 1 3 1 1 4 0 0 1
7. For each of the following matrix, find all eigenvalues and a basis of each
eigenspace. Hence determine which matrix can be diagonalized.
1 3 3 3 1 1
(i) A 3 5 3 (ii) A 7 5 1
6 6 4 6 6 2
37
Chapter 5
1 1
A
0 1
Let p ( ) 2 2 1 : Then one can check by straightforward computation that p ( A) 0 .
Thus p (T ) 0 .
38
Exercise 5.1 Prove that if the linear transformation T above satisfies the polynomial p ( x ) ,
then p ( ) 0 for each eigenvalue of T. (Hint: Consider the effect of p (T ) on an
eigenvector with eigenvalue ).
Theorem 5.1 (THE CAYLEY-HAMILTON). Let V be a finite-dimensional vector space
over a field F, and let T : V V be a linear transformation with characteristic polynomial
p ( ) . Then p (T ) 0 .
In other words, If the Cayley-Hamilton states that every matrix is a zero of its characteristic
polynomial.
Proof
Let A be an arbitrary n-square matrix and let A ( ) be the characteristic polynomial; say
A ( ) A I n an 1 n 1 ... a1 a0 .
Now let B ( x) denote the classical adjoint of the matrix A I .The elements of B ( ) are cofactors
of the matrix A I and hence are polynomials in x of degree not exceeding n 1 . Thus
B( ) Bn 1 n 1 ... B1 B0
Where the Bi ’s are n-square matrices which are independent of x. It follows from the
following property “For any square matrix A, A (adj A) (adjA) A A I ”
that
( A I ) B( ) A I I Or
( A I )( Bn 1 n 1 ... B1 B0 ) ( n an 1 n 1 ... a1 a0 ) I
Removing parenthesis and equating the coefficients of corresponding powers of ,
Bn 1 I
Bn 2 ABn 1 an 1 I
Bn 3 ABn 2 an 2 I
.............................
B0 AB1 a1 I
AB0 a0 I
Multiplying the above matrix equations by An , An 1 ,..., A, I respectively,
An Bn 1 An
An 1 Bn 2 An Bn 1 an1 An 1
An 2 Bn 3 An 1Bn 2 an 2 An 2
.............................
AB0 A2 B1 a1 A
AB0 a0 I
Adding the above matrix equations,
0 An an 1 An 1 ... a1 A a0 I
In other words, ( A) 0 .
39
Example 5.1 Let V 2 . Let T :V V be the linear transformation with
0 1
T (( a, b)t ) (b, a )t for all ( a, b) V . Then the matrix A corresponding to T is A
1 0
1
with A I . Thus, T has characteristic polynomial 1 ; and it follows that
2
1
2
0 1 1 0 1 0 1 0 0 0
T I A I
2 2
.
1 0 0 1 0 1 0 1 0 0
1 1 4
Example 5.2 Let A 3 2 1 then the characteristic polynomial of A is
2 1 1
3 2 2 5 6 (verify). Now you can also verify that
6 1 1 11 3 22
3
A 7 0 11 , A 29 4 17 and
2
3 1 8 16 3 5
11 3 22 6 1 1 1 1 4 1 0 0 0 0 0
A 2 A 5 A 6 I 29 4 17 2 7 0 11 5 3 2 1 6 0 1 0 0 0 0
3 2
16 3 5 3 1 8 2 1 1 0 0 1 0 0 0
The Cayley-Hamilton theorem is useful in calculating the inverse of a matrix. If A1 exists
and ( A) 0 then A1 ( A) 0 . To illustrate, if
( A) An an 1 An 1 ... a1 A a0 I 0
and
A1 ( A) An 1 an 1 An 2 ... a2 A a1 I a0 A1 0
Thus
1
A1
a0
An 1 an 1 An 2 ... a2 A a1 I
40
1 1 4
Example 5.2 Let A 3 2 1 . Then from Example 5.1, the characteristic polynomial of
2 1 1
A is A ( ) 3 2 2 5 6 . Therefore, n 3, a2 2, a1 5, a0 6 and
1
A1
6
A2 2 A 5 I
6 1 1 2 2 8 5 0 0
1
7 0 11 6 4 2 0 5 0
6
3 1 8 4 2 2 0 0 5
1 3 7
1
1 9 13
6
1 3 5
for all (a, b, c )t V . With respect to the standard basis (1,0,0)t ,(0,1,0)t ,(0,0,1)t , T has
matrix
5 1 1
1 5 1 ( A, say)
1 1 5
The characteristic polynomial of T is
T ( ) ( 4)2 ( 7),
so that the eigenvalues of T are 4 and 7. So we have T (T ) 0 . But we see that
(T 4 I )(T 7 I ) 0 already, as
1 1 1 2 1 1 0 0 0
1 1 1 1 2 1 0 0 0 .
1 1 1 1 1 2 0 0 0
41
Definition 5.1 Let T : V V be a linear transformation, where V is a finite-dimensional real
vector space. The minimal polynomial of T is the monic polynomial mT ( x) of least degree
such that mT (T ) 0 .
Remark: You may have noticed that the above definition implicitly assumes that the
minimum polynomial is unique. This assumption may be justified by:
Lemma 5.1 Let V,T, mT ( x) be as above. Let q(x) be any polynomial such that q(T) = 0:
Remark: Lemma 5.1 makes it fairly routine to determine the minimum polynomial of T :
42
Lemma 5.1 may be used to reduce the work involved somewhat. For if _ 1 ,...., n are the
distinct eigenvalues of T, then the minimal polynomial of T is divisible by
n
i 1
( x i ) t ( x ) , so we need only check those factors q(x) of p(x) which are themselves
divisible by t(x).
2 2 5
Example 5.4. Find the minimal polynomial of the following matrix A 3 7 15
1 2 4
Solution: The characteristics polynomial of A is ( x) ( x 1) 2 ( x 3) The minimal
1 2 5 1 2 5 0 0 0
f ( A) ( A I )( A 3I ) 3 6 15 3 4 15 0 0 0 .
1 2 5 1 2 7 0 0 0
Thus f ( x) mT ( x) ( x 1)( x 3) x 2 4 x 3 is the minimal polynomial of A.
Exercise 5.2
1. Find the minimum polynomial of
2 1 0
0
0 2 0 0
.
0 0 1 1
0 0 2 4
a 0
a
(i) A (ii) B 0 a
0 0 0
a 0 0
0 a 0
(iii) C
0 0 a
0 0 0
43