Diagonal Ization
Diagonal Ization
1.1. Square matrices. We are mainly interested in square matrices. A matrix is square
if
Pnn = m. The trace of a square matrix A is the sum of its diagonal elements, trace (A) =
i=1 aii .
1
2
For general n the method is the same that for matrices of order 3, expanding the determi-
nant by a row or a column and reducing in this way the order of the determinants that must
be computed. For a determinant of order 4 one has to compute 4 determinants of order 3.
Definition 1.7. Given a matrix A of order n, the complementary minor of element aij is
the determinant of order n − 1 which results from the deletion of the row i and the column
j containing that element. The adjoint Aij of the element aij is the minor multiplied by
(−1)i+j .
According to this definition, the determinant of matrix A can be defined as
|A| = ai1 Ai1 + ai2 Ai2 + · · · + ain Ain (by row i)
or, equivalently
|A| = a1j A1j + a2j A2j + · · · + anj Anj (by column j).
Example 1.8. Find the value of the determinant
1 2 0 3
4 7 2 1
.
1 3 3 1
0 2 0 7
Answer: Expanding the determinant by the third column, one gets
1 2 0 3
1 2 3 1 2 3
4 7 2 1
= (−1)3+2 2 1 3 1 + (−1)3+3 3 4 7 1 .
1 3 3 1
0 2 7 0 2 7
0 2 0 7
The main properties of the determinants are the following. It is assumed that the
matrices A and B in each of the following laws are square of order n and λ ∈ R.
(1) |A| = |AT |.
(2) |λA| = λn |A|.
(3) |AB| = |A||B|.
(4) A matrix A is regular if and only if |A| =6 0; in this case |A−1 | = |A|
1
.
(5) If in a determinant two rows (or columns) are interchanged, the value of the deter-
minant is changed in sign.
(6) If two rows (columns) in a determinant are identical, the value of the determinant is
zero.
(7) If all the entries in a row (column) of a determinant are multiplied by a constant λ,
then the value of the determinant is also multiplied by this constant.
(8) In a given determinant, a constant multiple of the elements in one row (column) may
be added to the elements of another row (column) without changing the value of the
determinant.
The next result is very useful to check if a given matrix is regular or not.
Theorem 1.9. A square matrix A has an inverse if and only |A| =
6 0.
4
2. Diagonalization of matrices
Definition 2.1. Two matrices A and B of order n are similar if there exists a matrix P
such that
B = P −1 AP.
Definition 2.2. A matrix A is diagonalizable if it is similar to a diagonal matrix D, that
is, there exists D diagonal and P invertible such that D = P −1 AP .
Of course, D diagonal means that every element out of the diagonal is null
λ1 0 . . . 0
0 λ2 . . . 0
D= ... .. . . . , λ1 , . . . , λn ∈ R.
. . ..
0 0 . . . λn
Proposition 2.3. If A is diagonalizable, then for all m ≥ 1
(2.1) Am = P Dm P −1 ,
where
λm 0 ... 0
1
m
m
0 λ 2 ... 0
D = ..
.. .. .. .
. . . .
0 0 . . . λm
n
In matrix form !
1 1
gt+1 2 9 gt
= .
bt+1 1 8 bt
2 9
Obviously
!4
1 1
g5 2 9 g1
= .
b5 1 8 b1
2 9
If the matrix were diagonalizable and we could find matrices P and D, then the computation
of the powers of the matrix would be easy using Proposition 2.3. We will come back to this
example afterwards.
Definition 2.5. Let A be a matrix of order n. We say that λ ∈ R is an eigenvalue of A and
that u ∈ Rn , u 6= 0, is an eigenvector of A associated to λ if
Au = λu.
The set of eigenvalues of A, σ(A) = {λ1 , . . . , λk }, is called the spectrum of A. The set of all
eigenvectors of A associated to the same eigenvalue λ, including the null vector, is denoted
S(λ), and is called the eigenspace or proper subspace associated to λ.
The following result shows that an eigenvector can only be associated to a unique eigen-
value.
Proposition 2.6. Let 0 6= u ∈ S(λ) ∩ S(µ). Then λ = µ.
Proof. Suppose 0 6= u ∈ S(λ) ∩ S(µ). Then
Au = λu
Au = µu.
Subtracting both equations we obtain 0 = (λ−µ)u and, since 0 6= u, we must have λ = µ.
Recall that for an arbitrary matrix A, the rank of the matrix is the number of linearly
independent columns or rows (both numbers necessarily coincide). It is also given by the
order of the largest non null minor of A.
Theorem 2.7. The real number λ is an eigenvalue of A if and only if
|A − λIn | = 0.
Moreover, S(λ) is the set of solutions (including the null vector) of the linear homogeneous
system
(A − λIn )u = 0,
and hence it is a vector subspace, which dimension is
dim S(λ) = n − rank(A − λIn ).
6
The solutions are λ = 3 (simple root) and λ = 2 (double root). To find S(3) = {u ∈ R3 :
(B − 3I3 )u = 0} we compute the solutions to
−1 1 0 x 0
(B − 3I3 )u = 0 −2 −1 y = 0 ,
0 2 1 z 0
which are x = y and z = −2y, and hence S(3) =< (1, 1, −2) >. To find S(2) we solve the
system
0 1 0 x 0
(B − 2I3 )u = 0 −1 −1 y = 0 ,
0 2 2 z 0
from which x = y = 0 and hence S(2) =< (1, 0, 0) >.
So, the eigenvalues are λ1 = 1, λ2 = 2 and λ3 = 3. We now compute the eigenvectors. The
eigenspace S(1) is the solution of the homogeneous linear system whose associated matrix is
C − λI3 with λ = 1. That is, S(1) is the solution of the following homogeneous linear system
0 2 0 x 0
0 2 0 y = 0
1 1 2 z 0
On the other hand, S(2) is the set of solutions of the homogeneous linear system whose
associated matrix is C − λI3 with λ = 2. That is, S(2) is the solution of the following
−1 2 0 x 0
0 0 0 y = 0
1 1 1 z 0
So,
S(2) = {(2y, y, −3y) : y ∈ R} =< (2, 1, −3) >
Finally, S(3) is the set of solutions of the homogeneous linear system whose associated matrix
is A − λI3 with λ = 3. That is, S(3) is the solution of the following
−2 2 0 x 0
0 −1 0 y = 0
1 1 0 z 0
and we obtain
S(3) = {(0, 0, z) : z ∈ R} =< (0, 0, 1) >
We now start describing the method to diagonalize a matrix. Fix a square matrix A. Let
λ1 , λ2 , . . . , λk
be distinct real roots of the characteristic polynomial pA (λ) an let mk be the multiplicity
of each λk (Hence mk = 1 if λk is a simple root, mk = 2 if it is double, etc.). Note that
m1 + m2 + · · · + mk = n.
The following result estates that the number of independent vectors in the subspace S(λ)
can never be bigger than the multiplicity of λ.
Proposition 2.12. For each j = 1, . . . , k
1 ≤ dim S(λj ) ≤ mj .
The following theorem gives necessary and sufficient conditions for a matrix A to be
diagonalizable.
Theorem 2.13. A matrix A is diagonalizable if and only if the two following conditions
hold.
(1) Every root, λ1 , λ2 , . . . , λk of the characteristic polynomial pA (λ) is real.
(2) For each j = 1, . . . , k
dim S(λj ) = mj .
Corollary 2.14. If the matrix A has n distinct real eigenvalues, then it is diagonalizable.
Theorem 2.15. If A is diagonalizable, then the diagonal matrix D is formed by the eigen-
values of A in its main diagonal, with each λj repeated nj times. Moreover, a matrix P
such that D = P −1 AP has as columns independent eigenvectors selected from each proper
subspace S(λj ), j = 1, . . . , k.
9
We can wonder what happens in the long run, that is, supposing that the course lasts
forever (oh no!). In this case
! !
2 2
1 0
lim An = P ( lim Dn )P −1 = P P −1 = 11
9
11
9
,
n→∞ n→∞ 0 0 11 11
to find that the stationary distribution of probabilities is
g∞ 0.1818
= .
b∞ 0.8182