0% found this document useful (0 votes)
101 views26 pages

Chapter 2. Jordan Forms: Lecture Notes For MA1212

The document discusses Jordan forms and related concepts: 1) It defines generalised eigenvectors, Jordan chains, and invariant subspaces. Generalised eigenvectors are vectors in the null space of (A - λI)j for some j. Jordan chains are vectors related to an eigenvalue where applying A - λI repeatedly yields another vector or zero. 2) A matrix is in Jordan form if it is block diagonal with Jordan blocks, which are matrices with the eigenvalue on the diagonal and ones above it. 3) The column spaces and null spaces of (A - λI)j give invariant subspaces, and Jordan chains span invariant subspaces. Finding Jordan forms involves finding eigenvalues and relating vectors via
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
101 views26 pages

Chapter 2. Jordan Forms: Lecture Notes For MA1212

The document discusses Jordan forms and related concepts: 1) It defines generalised eigenvectors, Jordan chains, and invariant subspaces. Generalised eigenvectors are vectors in the null space of (A - λI)j for some j. Jordan chains are vectors related to an eigenvalue where applying A - λI repeatedly yields another vector or zero. 2) A matrix is in Jordan form if it is block diagonal with Jordan blocks, which are matrices with the eigenvalue on the diagonal and ones above it. 3) The column spaces and null spaces of (A - λI)j give invariant subspaces, and Jordan chains span invariant subspaces. Finding Jordan forms involves finding eigenvalues and relating vectors via
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 26

Chapter 2.

Jordan forms
Lecture notes for MA1212

P. Karageorgis

[email protected]

1 / 26
Generalised eigenvectors

Definition 2.1 – Generalised eigenvector


Suppose λ is an eigenvalue of the square matrix A. We say that v
is a generalised eigenvector of A with eigenvalue λ, if v is a nonzero
element of the null space of (A − λI)j for some positive integer j.

An eigenvector of A with eigenvalue λ is a nonzero element of the


null space of A − λI, so it is also a generalised eigenvector of A.
We shall denote by N (A − λI)j the null space of (A − λI)j .

Theorem 2.2 – Null spaces eventually stabilise


Let A be a square matrix and let λ be an eigenvalue of A. Then the
null spaces N (A − λI)j are increasing with j and there is a unique
positive integer k such that N (A − λI)j = N (A − λI)k for all j ≥ k.

2 / 26
Column space

Definition 2.3 – Column space


The column space of a matrix A is the span of the columns of A. It
consists of all vectors y that have the form y = Ax for some vector x.

The column space of A is usually denoted by C(A). The dimension of


the column space of A is also known as the rank of A.
To find a basis for the column space of a matrix A, we first compute
its reduced row echelon form R. Then the columns of R that contain
pivots form a basis for the column space of R and the corresponding
columns of A form a basis for the column space of A.
The dimension of the column space is the number of pivots and the
dimension of the null space is the number of free variables, so

rank + nullity = number of columns.

3 / 26
Column space: Example
We find a basis for the column space of the matrix
 
1 2 4 5 4
A = 3 1 7 2 3 .
2 1 5 1 5
The reduced row echelon form of this matrix is given by
 
1 0 2 0 0
R = 0 1 1 0 7 .
0 0 0 1 −2
Since the pivots of R appear in the 1st, 2nd and 4th columns, a basis
for the column space of A is formed by the corresponding columns
     
1 2 5
v1 = 3 , v2 = 1 , v4 = 2 .
2 1 1
4 / 26
Invariant subspaces

Definition 2.4 – Invariant subspace


Suppose A is an n × n complex matrix and U is a subspace of Cn . We
say that U is A-invariant, if Au ∈ U for each vector u ∈ U .

The one-dimensional invariant subspaces correspond to eigenvectors.


The matrix A might represent a reflection of the xy-plane along a line
through the origin. In that case, A has two one-dimensional invariant
subspaces. One is spanned by a vector which is parallel to the line of
reflection and one is spanned by a vector which is perpendicular to it.
Similarly, A might represent a rotation in R3 around some axis. Such
a matrix has an one-dimensional invariant subspace along the axis of
rotation and a two-dimensional invariant subspace perpendicular to it.
Both the null spaces N (A − λI)j and the column spaces C(A − λI)j
give rise to A-invariant subspaces. We shall mostly study the former.
5 / 26
Coordinate vectors

Theorem 2.5 – Coordinate vectors


Suppose B is invertible with columns v1 , v2 , . . . , vn and A is n × n.
Then the kth column of B −1 AB lists the coefficients that one needs
in order to express Avk as a linear combination of v1 , v2 , . . . , vn .

That is, the kth column of B −1 AB is the coordinate vector of Avk


with respect to the basis v1 , v2 , . . . , vn . This is true because
n
X n
X
Avk = c i vi ⇐⇒ ABek = ci Bei
i=1 i=1
n
X
⇐⇒ B −1 ABek = c i ei .
i=1

Suppose, for instance, that Avk is a linear combination of v1 and v2 .


Then the kth column of B −1 AB is a linear combination of e1 and e2 ,
so its first two entries are nonzero and its other entries are zero.
6 / 26
Invariant subspaces and blocks

Consider a 3 × 3 matrix A which has two invariant subspaces

U = Span{v1 , v2 }, V = Span{v3 }

such that v1 , v2 , v3 form a basis of R3 . We can then write


 
 Av1 = a1 v1 + a2 v2 
Av2 = b1 v1 + b2 v2
Av3 = cv3
 

for some scalars a1 , a2 , b1 , b2 , c and this implies the identity


 
  a 1 b1
B = v1 v2 | v3 =⇒ B −1 AB =  a2 b2 .
c

Thus, we get a 2 × 2 block for the 2-dimensional invariant subspace


and an 1 × 1 block for the 1-dimensional invariant subspace.
7 / 26
Jordan chains

Definition 2.6 – Jordan chain


Suppose λ is an eigenvalue of the square matrix A. We say that the
vectors v1 , v2 , . . . , vk form a Jordan chain, if they are nonzero with
 
vi+1 when i < k
(A − λI)vi = .
0 when i = k

The last vector in a Jordan chain is simply an eigenvector of A.


The first vector in a Jordan chain of length k is a vector that lies in
the null space N (A − λI)k but not in the null space N (A − λI)k−1 .
The span of a Jordan chain is an A-invariant subspace because

Avi = (λI + A − λI)vi = λvi + vi+1

when i < k and since Avk = λvk . In particular, each Avi except for
the last one is a linear combination of precisely two vectors.
8 / 26
Jordan chains: Example 1
N1 r r r dim N1 = 3

N2 r r r dim N2 − dim N1 = 3

N3 r dim N3 − dim N2 = 1

Let λ be an eigenvalue of A and let Nj = N (A − λI)j for each j. As


we already know, these null spaces are increasing with j. Assume, for
instance, that dim N1 = 3, dim N2 = 6 and dim N3 = 7.
We draw a diagram by placing 3 dots in the first row, 6 − 3 = 3 dots
in the second row and 7 − 6 = 1 dot in the third row.
The dots in this diagram represent linearly independent vectors and if
a dot represents v, then the dot right above it represents (A − λI)v.
Reading the diagram vertically, we conclude that there is one Jordan
chain of length 3 as well as two Jordan chains of length 2.
9 / 26
Jordan chains: Example 2
N1 r r r r dim N1 = 4

N2 r r dim N2 − dim N1 = 2

N3 r dim N3 − dim N2 = 1

Let λ be an eigenvalue of A and let Nj = N (A − λI)j for each j. In


this example, we assume dim N1 = 4, dim N2 = 6 and dim N3 = 7.
The corresponding diagram includes 4 dots in the first row, 6 − 4 = 2
dots in the second row and 7 − 6 = 1 dot in the third row.
Reading the diagram vertically, we get one Jordan chain of length 3,
one Jordan chain of length 2 and two Jordan chains of length 1.
Jordan chains of length 1 are just eigenvectors of A. To find a Jordan
chain of length 3, one needs to find a vector v that lies in N3 but not
in N2 . Such a vector generates the chain v, (A − λI)v, (A − λI)2 v.
10 / 26
Jordan blocks and Jordan form
Definition 2.7 – Jordan blocks and Jordan form
A Jordan block with eigenvalue λ is a square matrix whose entries are
equal to λ on the diagonal, equal to 1 right below the diagonal and
equal to 0 elsewhere. A Jordan form is a block diagonal matrix that
consists entirely of Jordan blocks.

Some typical examples of Jordan blocks are


 
  λ
  λ
J1 = λ , J2 = , J3 =  1 λ  .
1 λ
1 λ
Two typical examples of Jordan forms are
   
2 1
 1 2   1 1 
J = , J′ =  .
 1 2   1 
2 2
11 / 26
Jordan chains ←→ Jordan blocks

Theorem 2.8 – Jordan chains and Jordan blocks


Suppose A is an n × n complex matrix and let B be a matrix whose
columns form a basis of Cn consisting entirely of Jordan chains of A.
Then J = B −1 AB is a matrix in Jordan form whose kth Jordan block
has the same size and the same eigenvalue as the kth Jordan chain.

For instance, suppose A is 4 × 4 with eigenvalues λ = 0, 0, 3, 3. If A


has a Jordan chain of length 2 with λ = 0 and two Jordan chains of
length 1 with λ = 3, then the Jordan form of A is
 
0
 1 0 
J = .
 3 
3
The Jordan form of a square matrix is unique (up to a permutation of
its blocks). There might be several blocks with the same eigenvalue.
12 / 26
Jordan form: Example 1

We compute the Jordan form of the matrix


 
3 −1
A= .
1 1

The characteristic polynomial is f (λ) = λ2 − 4λ + 4 = (λ − 2)2 , so


the only eigenvalue is λ = 2. Moreover, it is easy to check that

dim N (A − 2I) = 1, dim N (A − 2I)2 = 2.

The corresponding diagram for the Jordan chains is •• and we need to


find a Jordan chain of length 2. Pick a vector v1 that lies in the null
space of (A − 2I)2 but not in the null space of A − 2I.
Letting v2 = (A − 2I)v1 now gives a Jordan chain v1 , v2 and so
 
  2
B = v1 v2 =⇒ J = B −1 AB = .
1 2

13 / 26
Jordan form: Example 2, page 1

We compute the Jordan form of the matrix


 
4 −5 2
A = 1 −2 2 .
2 −6 5
In this case, the characteristic polynomial is given by

f (λ) = −λ3 + 7λ2 − 15λ + 9 = −(λ − 1)(λ − 3)2 ,

so there are two eigenvalues that need to be treated separately.


When it comes to the eigenvalue λ = 1, one finds that

dim N (A − I) = 1, dim N (A − I)2 = 1.

The corresponding diagram for the Jordan chains is • and it only


contains one Jordan chain of length 1. In fact, a Jordan chain of
length 1 is merely an eigenvector v1 with eigenvalue λ = 1.
14 / 26
Jordan form: Example 2, page 2

When it comes to the eigenvalue λ = 3, one finds that

dim N (A − 3I) = 1, dim N (A − 3I)2 = 2.

The corresponding diagram for the Jordan chains is •• and we need to


find a Jordan chain of length 2. Pick a vector v2 that lies in the null
space of (A − 3I)2 but not in the null space of A − 3I. Such a vector
gives rise to a Jordan chain v2 , v3 with v3 = (A − 3I)v2 .
In particular, A has a Jordan chain v1 with eigenvalue λ = 1 and also
a Jordan chain v2 , v3 with eigenvalue λ = 3, so its Jordan form is
 
  1
B = v1 | v2 v3 =⇒ J = B −1 AB =  3 .
1 3

The chosen vectors v1 , v2 are by no means unique. In fact, there are


infinitely many matrices B such that B −1 AB is in Jordan form.
15 / 26
Jordan basis and Jordan form
Definition 2.9 – Jordan basis
A Jordan basis for an n × n complex matrix A is a basis of Cn that
consists entirely of Jordan chains of A.

1 Finding the Jordan form. Determine the various eigenvalues λ and


apply the following steps for each λ. Compute the numbers
dj = dim N (A − λI)j
until they stabilise and draw a diagram for the Jordan chains. The
lengths of these chains represent the sizes of the Jordan blocks.
2 Finding a Jordan basis. Consult the diagram of Jordan chains for
each eigenvalue λ and worry about the longest chains first. To find a
chain of length k > 1, pick a vector that lies in the kth null space but
not in the previous one and repeatedly multiply it by A − λI. Once
you have a Jordan chain, you may proceed similarly to find the next
longest chain. Chains of length 1 are merely eigenvectors of A.
16 / 26
Number of Jordan chains

Theorem 2.10 – Number of Jordan chains


Consider the diagram of Jordan chains for an eigenvalue λ which has
multiplicity m as a root of the characteristic polynomial of A.
1 The total number of dots in the diagram is equal to m.
2 The total number of Jordan chains is equal to dim N (A − λI).

The first number is also known as the algebraic multiplicity of λ. The


second number is also known as the geometric multiplicity of λ.
When m ≤ 3, one only needs to know these two numbers to find the
diagram of Jordan chains (and thus the sizes of the Jordan blocks).
For instance, the diagram for a simple eigenvalue λ contains only one
dot, so each simple eigenvalue λ contributes a single 1 × 1 block.
Similarly, a triple eigenvalue λ such that dim N (A − λI) = 2 has a
diagram containing 3 dots but only two chains. Such an eigenvalue
will necessarily contribute one 2 × 2 block and one 1 × 1 block.
17 / 26
Direct sums

Theorem 2.11 – Linear independence of Jordan chains


Suppose γ1 , γ2 , . . . , γm are Jordan chains of a square matrix A. If the
last vectors of the Jordan chains are linearly independent, then all the
vectors that belong to the Jordan chains are linearly independent.

Definition 2.12 – Direct sum


Let U, V be subspaces of a vector space W . Their sum U + V is the
set of all vectors w which have the form w = u + v for some u ∈ U
and some v ∈ V . If it happens that U ∩ V = {0}, then we say that
the sum is direct and we denote it by U ⊕ V .

Theorem 2.13 – Basis of a direct sum


One may obtain a basis for the direct sum U ⊕ V by appending a basis
of V to a basis of U . In particular, dim(U ⊕ V ) = dim U + dim V .
18 / 26
Primary decomposition theorem

Theorem 2.14 – Primary decomposition theorem


Given an n × n complex matrix A, one can write

Cn = N (A − λ1 I)k1 ⊕ · · · ⊕ N (A − λp I)kp ,

where λ1 , λ2 , . . . , λp are the distinct eigenvalues of A and each ki is


the exponent at which the null spaces N (A − λi I)j stabilise.

This theorem is mostly of theoretical value. It ensures that one can


always find a basis of Cn by looking at the various null spaces.
Each of the null spaces has a basis which consists entirely of Jordan
chains. Since the above sum is direct, we may then merge all these
Jordan chains to obtain a Jordan basis for the given matrix A.
Letting B denote the matrix whose columns form a Jordan basis, we
conclude that B is invertible, while J = B −1 AB is in Jordan form.
19 / 26
Similar matrices

Definition 2.15 – Similar matrices


A square matrix A is said to be similar to a square matrix C, if there
exists an invertible matrix B such that C = B −1 AB.

Theorem 2.16 – Similarities of similar matrices


If two square matrices A and C are similar, then
1 They have the same characteristic polynomial and eigenvalues.
2 They have the same rank, nullity, trace and determinant.
3 The matrices An and C n are similar for any positive integer n.

Theorem 2.17 – Similarity test


Two square matrices are similar if and only if their Jordan forms are
the same (up to a permutation of their Jordan blocks).
20 / 26
Properties of Jordan blocks

Theorem 2.18 – Properties of Jordan blocks


Suppose that J is a k × k Jordan block with eigenvalue λ.
1 The entries of (J − λI)j are equal to 1, if they lie j steps below
the diagonal, and they are equal to 0, otherwise.
2 One has (J − λI)j = 0 if and only if j ≥ k.

Loosely speaking, the powers of J − λI are obtained by shifting its


entries downwards one step at a time. For instance, one has
   
0 0
1 0  2
0 0 
J − λI =  1 0  =⇒ (J − λI) = 1 0 0  .
  

1 0 1 0 0
When it comes to a Jordan block J, it is easy to see that the null
spaces N (J − λI)j are increasing until they eventually stabilise.
21 / 26
Powers of Jordan blocks

Theorem 2.19 – Powers of Jordan blocks


Let J be a Jordan block with eigenvalue λ 6= 0. Then
 the entries of
n n n n−1
its nth power J  are equal to λ on the diagonal, 1 λ right below
the diagonal, n2 λn−2 two steps below the diagonal, and so on.

For instance, if J is a 2 × 2 Jordan block with eigenvalue λ 6= 0, then


   n 
λ n λ
J= =⇒ J = .
1 λ nλn−1 λn
Jordan blocks with eigenvalue λ = 0 are somewhat different, as
   
0 0
J = 1 0  =⇒ J 2 = 0 0  =⇒ J 3 = 0.
1 0 1 0 0
In particular, powers of such matrices may be computed as before by
shifting their entries downwards one step at a time.
22 / 26
Powers of a square matrix
1 Given a complex square matrix A, we first find its Jordan form J as
well as a matrix B such that J = B −1 AB is in Jordan form.
2 Letting J1 , J2 , . . . , Jk denote the Jordan blocks of J, one has
   n 
J1 J1
 J2 
n
 J2n 
J = =⇒ J =  .
   
. .  . .
 .   . 
Jk n
Jk
The powers Jin are easy to compute, as those are powers of Jordan
blocks and we have explicit formulas for computing them.
3 The equation above determines the powers of the Jordan form J. To
find the powers of the original matrix A, we note that

J = B −1 AB =⇒ J n = B −1 An B
=⇒ An = BJ n B −1 .
23 / 26
Matrices and polynomials

Theorem 2.20 – Matrices and polynomials


Let A be a square matrix and let g be a polynomial.
1 The matrix A is a root of g if and only if its Jordan form J is a
root of g. In other words, one has g(A) = 0 ⇐⇒ g(J) = 0.
2 If the matrix A is a root of g, then every eigenvalue λ of A is a
root of g. In other words, one has g(A) = 0 =⇒ g(λ) = 0.

Theorem 2.21 – Cayley-Hamilton theorem


Every square matrix is a root of its characteristic polynomial.

Definition 2.22 – Minimal polynomial


The minimal polynomial m(λ) of a square matrix A is defined as the
monic polynomial of smallest degree that has A as a root.
24 / 26
Minimal polynomial

Theorem 2.23 – Properties of the minimal polynomial


Let A be a square matrix and let m(λ) be its minimal polynomial.
1 The minimal polynomial divides every polynomial that has A as a
root. In particular, it divides the characteristic polynomial f (λ).
2 If λ1 , λ2 , . . . , λp are the distinct eigenvalues of A, then

m(λ) = (λ − λ1 )k1 (λ − λ2 )k2 · · · (λ − λp )kp ,

where each exponent ki is given by the size of the largest Jordan


block that corresponds to the eigenvalue λi .
3 The matrix A is diagonalisable if and only if the factors of m(λ)
are all linear, namely if and only if ki = 1 for each i.

For instance, a square matrix A such that A3 = A is diagonalisable


because its minimal polynomial divides λ3 − λ = λ(λ − 1)(λ + 1).
25 / 26
Minimal polynomial: Example

Suppose A is a matrix with characteristic and minimal polynomials

f (λ) = −(λ − 2)2 (λ − 4)3 , m(λ) = (λ − 2)(λ − 4)2 .

Then A is a 5 × 5 matrix that has two distinct eigenvalues. Let us


find its Jordan form by looking at each eigenvalue separately.
When it comes to the eigenvalue λ1 = 2, we have k1 = 1 and so the
largest Jordan block is 1 × 1. Since λ1 = 2 is a double eigenvalue, we
must thus have two 1 × 1 Jordan blocks with this eigenvalue.
When it comes to the eigenvalue λ2 = 4, we have k2 = 2 and so the
largest Jordan block is 2 × 2. Since λ2 = 4 is a triple eigenvalue, it
contributes one 2 × 2 Jordan block and one 1 × 1 Jordan block.
In other words, the Jordan form of A consists of the blocks
 
    4  
2 , 2 , , 4 .
1 4

26 / 26

You might also like