Primary Decomposition Theorem
Primary Decomposition Theorem
net/publication/311589069
CITATIONS READS
0 19,587
2 authors:
All content following this page was uploaded by Rajesh Singh on 13 December 2016.
Table of Contents
1. Learning Outcomes
2. Prerequisites
3. Preliminaries
3.1. Diagonalizable Operator
3.2. Minimal Polynomial
3.3. Invariant Subspaces
3.4. Direct Sums
4. Decomposition of an Operator
4.1. Primary Decomposition Theorem
4.2. Decomposition into Diagonalizable and Nilpotent
operator
5. Solved Problems
6. Applications
7. Summary
8. Exercises
9. References
" 𝑴𝒂𝒕𝒉𝒆𝒎𝒂𝒕𝒊𝒄𝒔 𝒊𝒔 𝒏𝒐𝒕 𝒋𝒖𝒔𝒕 𝒂 𝒓𝒆𝒂𝒅𝒊𝒏𝒈 𝒔𝒖𝒃𝒋𝒆𝒄𝒕. 𝒀𝒐𝒖 𝒎𝒖𝒔𝒕 𝒖𝒔𝒆 𝒚𝒐𝒖𝒓 𝒃𝒓𝒂𝒊𝒏𝒔
1.LEARNING OUTCOMES
We believe that at the end of this chapter the reader will become well versed with the
concept of decompositions, the motivation behind studying various types of
decompositions and will gain a good knowledgein primary decompositions, diagonalizable
operators and nilpotent operators. We have also given an application of this concept to
differential equations to make the reader realize that by studying such abstract concepts
we develop an effective unifying approach to tackle problems from different fields.
2. PREREQUISITES
We expect that the reader knows the definition of all the terms listed below and is
familiar with all basic concepts associated with these terms.
3. PRELIMINARIES
In this section we touch upon some concepts which we feel the reader is already familiar
with and hence we will only be stating definitions and results without going into the
details associated with them. The idea is just to make the text self-contained. Taking the
liberty that the reader knows above mentioned concepts we start with diagonalizable
operators.
Note that if 𝔙′ is any other basis of 𝑉, then there exists an invertible matrix 𝑃 such that
𝑇 𝔅′ = 𝑃−1 𝑇 𝛽 𝑃.
Note:
If𝑇 is a diagonalizable operator and 𝛽 = 𝑣1 , 𝑣2 , … , 𝑣𝑛 is an ordered basis of𝑉such that
the matrix of 𝑇 w.r.t basis 𝛽 is a diagonal matrix
𝑐1 0 ⋯ 0
𝑐2
𝑖. 𝑒., 𝑇 𝛽 = 0 ⋯ 0 .
⋮ ⋮ ⋱ ⋮
0 0 ⋯ 𝑐𝑛
Then 𝑐1 , 𝑐2 , … , 𝑐𝑛 are characteristic values of 𝑇 and each 𝑣𝑖 1 ≤ 𝑖 ≤ 𝑛 is an eigen vector
of 𝑇 corresponding to eigen value 𝑐𝑖 i.e.,
𝑇𝑣𝑖 = 𝑐𝑖 𝑣𝑖 ∀ 𝑖 = 1,2, … , 𝑛.
Theorem 3.1.4 Let𝑃 and 𝑄 be two diagonalizable operators vector space𝑉. Then𝑃 and
𝑄commute if and only if they are simultaneously diagonalizable.
Let Tbe a linear operator on a finite-dimensional vectorspace 𝑉over the field 𝐹. The
minimal polynomialfor𝑇is the (unique)monic generator of the ideal of polynomials over
𝐹which annihilate 𝑇.
The minimal polynomial 𝑝for the linear operator 𝑇is uniquely determined by these three
properties :
1. 𝑝is a monic polynomial over the scalar field F.
2. 𝑝 𝑇 = 0.
3. No polynomial over 𝐹 which annihilates 𝑇has smaller degree than𝑝.
If any (and hence all) of the conditions of the last lemma hold, then we say that the sum
𝑊 = 𝑊1 + . . . + 𝑊𝑘 is director that 𝑊is the direct sumof 𝑊1 , . . . , 𝑊𝑘 and we write
𝑊 = 𝑊1 ⊕. . .⊕ 𝑊𝑘 .
diagonalizable and if 𝑐1 , … . , 𝑐𝑘 are the distinct characteristicvalues of T, then there exist linear
(i). 𝑇 = 𝑐1 𝐸1 + ⋯ + 𝑐𝑘 𝐸𝑘 ;
(ii). 𝐼 = 𝐸1 + ⋯ + 𝐸𝑘 ;
(iii). 𝐸𝑖 𝐸𝑗 = 0, whenever 𝑖 ≠ 𝑗;
𝑟1 𝑟2 𝑟𝑘
𝑓 𝑥 = 𝑥 − 𝑐1 𝑥 − 𝑐2 … 𝑥 − 𝑐𝑘
4. Decomposition of an Operator
In this chapter we try to decompose a linear operator 𝑇 defined on a finite-dimensional
vector space 𝑉 into a direct sum of operators which are elementary in some sense.
We already know from the invariant direct sums theory that if 𝑇 is a diagonalizable linear
operator on a finite dimensional space 𝑉 and 𝑐1 , … . , 𝑐𝑘 are distinct characteristic values of
𝑇, then there exist linear operators 𝐸1 , … . , 𝐸𝑘 on 𝑉 such that 𝑇 = 𝑐1 𝐸1 + ⋯ + 𝑐𝑘 𝐸𝑘 where 𝐸𝑖 is
the projection of vector space 𝑉 onto the space of characteristic vectors associated with
the characteristic value 𝑐𝑖 . Thus if the minimal polynomial 𝑝of 𝑇 can be factorized over
the scalar field 𝐹 into a product of distinct monic polynomials of degree 1i.e.,
𝑝 = 𝑥 − 𝑐1 𝑥 − 𝑐2 … 𝑥 − 𝑐𝑘
then
𝑉 = 𝐾𝑒𝑟 𝑇 − 𝑐1 𝐼 ⊕ 𝐾𝑒𝑟 𝑇 − 𝑐2 𝐼 ⊕ … ⊕ 𝐾𝑒𝑟 𝑇 − 𝑐𝑘 𝐼 𝑎𝑛𝑑
𝑇 = 𝑐1 𝐸1 + ⋯ + 𝑐𝑘 𝐸𝑘
where𝐸𝑖 is the projection of vector space 𝑉 onto the null space 𝐾𝑒𝑟 𝑇 − 𝑐𝑖 𝐼 of 𝑇 − 𝑐𝑖 𝐼 .
But what if 𝑇 is not diagonalizable? In other words, how to decompose a linear operator
𝑇 defined on a finite-dimensional vector space 𝑉 whose minimal polynomial cannot be
factorized over the field 𝐹 into a product of distinct monic polynomials of degree 1 ? If 𝑇
is not diagonalizable, then for the minimal polynomial 𝑝 of 𝑇 we have two possibilities:
1 𝑝 = 𝑝1 𝑟1 . . . 𝑝𝑘 𝑟 𝑘 , where 𝑝𝑖 's are distinct irreducible monic polynomials over 𝐹 and𝑟𝑖 's
are positive integers.
2 𝑝is an irreducible monic polynomial over 𝐹.
In this chapter we will discuss the first case where the minimal polynomial can be
decomposed further over the scalar field 𝐹. In fact we will show that if the minimal
polynomial 𝑝 of a linear operator is of the form
𝑝 = 𝑝1 𝑟1 … 𝑝𝑘 𝑟 𝑘
where 𝑝𝑖 's are distinct irreducible monic polynomials over 𝐹 and𝑟𝑖 's are positive integers.
Then the vector space 𝑉 is expressible as
𝑟1 𝑟2 𝑟𝑘
𝑉 = 𝐾𝑒𝑟 𝑝1 𝑇 ⊕ 𝐾𝑒𝑟 𝑝2 𝑇 ⊕ … ⊕ 𝐾𝑒𝑟 𝑝𝑘 𝑇
3 1 −1
𝐴= 2 2 −1
2 2 0
3−𝑥 1 −1
𝑑𝑒𝑡 𝐴 − 𝑥𝐼 = 2 2−𝑥 −1 = − 𝑥 3 − 5𝑥 2 + 8𝑥 − 4 = −(𝑥 − 1)(𝑥 − 2)2 .
2 2 −𝑥
Thus 1 and 2 are the two characteristic values of the operator 𝑇. Now let us try to find
the characteristic vectors of 𝑇 associated with characteristic values 1and 2. Consider
2 1 −1
𝐴−𝐼 = 2 1 −1
2 2 −1
It is easy to observe that rank of 𝐴 − 𝐼 is 2 and hence dimension of the null space of
𝑇 − 𝐼 is 1. Thus the space of characteristic vectors associated with characteristic value 1
has dimension 1. Now
2 1 −1 𝑥1 2𝑥1 + 𝑥2 − 𝑥3
(𝐴 − 𝐼)𝑥 = 2 1 𝑥
−1 2 = 2𝑥1 + 𝑥2 − 𝑥3 .
2 2 𝑥
−1 3 2𝑥1 + 2𝑥2 − 𝑥3
Clearly, the vector 𝛼 = (1,0,2) satisfies the equation (𝐴 − 𝐼)𝑥 = 0 and hence is a
characteristic vector of 𝑇 associated with characteristic value 1. Also, the vector 𝛼 =
(1,0,2) spans the null space of 𝑇 − 𝐼. Now consider
1 1 −1
𝐴 − 2𝐼 = 2 0 −1
2 2 −2
1 1 −1 1 1 −1
2
𝐴 − 2𝐼 = 2 0 −1 2 0 −1
2 2 −2 2 2 −2
1+2−2 1+0−2 −1 − 1 + 2
= 2+0−2 2+0−2 −2 + 0 + 2
2+4−4 2+0−4 −2 − 2 + 4
1 −1 0
= 0 0 0
2 −2 0
2 2
Clearly, rank of 𝐴 − 2𝐼 is 1 and hence nullity of 𝐴 − 2𝐼 is 2. Now
1 −1 0 𝑥1 𝑥1 −𝑥2 0
2
𝐴 − 2𝐼 𝑥 = 0 0 0 𝑥2 = 0 0 0
2 −2 0 𝑥3 2𝑥1 −2𝑥2 0
The vectors (0,0,1) and (1,1,0) are linearly independent and satisfy the equation 𝐴−
2𝐼 2 𝑥 = 0. Hence the null space of (𝑇 − 2𝐼)2 is spanned by the vectors (0,0,1)and (1,1,0). The
vectors = (1,0,2), 𝛽 = (0,0,1) and 𝛾 = (1,1,0) are linearly independent and hence form the
basis of vector space ℝ3 . Thus we observe that the null space of (𝑇 − 𝐼) and the null
space of (𝑇 − 2𝐼)2 together span the vector space 𝑉 i.e.,
𝑝 = 𝑝1 𝑟1 . . . 𝑝𝑘 𝑟 𝑘
where the 𝑝𝑖 are distinct irreducible monic polynomials over 𝐹 and𝑟𝑖 are positive integers.
Let 𝑊𝑖 = 𝐾𝑒𝑟 𝑝𝑖 (𝑇)𝑟 𝑖 be null space of 𝑝𝑖 (𝑇)𝑟 𝑖 , 𝑖 = 1, . . . , 𝑘. Then
(i). 𝑉 = 𝑊1 . . . 𝑊𝑘 ;
(ii). each 𝑊𝑖 is invariant under 𝑇;
(iii). if𝑇𝑖 is the operator induced on 𝑊𝑖 by 𝑇, then the minimal polynomial for 𝑇𝑖 is 𝑝𝑖 𝑟 𝑖 .
Then clearly, the polynomials 𝑓1 , . . . , 𝑓𝑘 are relatively prime. Therefore there exist
polynomials 𝑔1 , . . . , 𝑔𝑘 such that
𝑓1 𝑔1 +. . . +𝑓𝑘 𝑔𝑘 = 1
𝑞= 𝑝𝑚 𝑟𝑚 .
𝑚 ≠𝑖,𝑗
𝐸𝑖 2 = 𝐸𝑖 𝑓𝑜𝑟𝑎𝑙𝑙𝑖 (1 ≤ 𝑖 ≤ 𝑘).
Let 𝑊𝑖 be the null space of 𝑝𝑖 (𝑇)𝑟 𝑖 , 𝑖 = 1, . . . , 𝑘. We claim that 𝑅𝑎𝑛𝑔𝑒𝐸𝑖 = 𝑊𝑖 . Let
𝑣𝜖𝑅𝑎𝑛𝑔𝑒𝐸𝑖 . Then 𝑣 = 𝐸𝑖 (𝑢) for some 𝑢𝜖𝑉 and therefore
𝐸𝑖 (𝑣) = 𝐸𝑖 2 (𝑢) = 𝐸𝑖 (𝑢) = 𝑣.
Consider
𝑟𝑖 𝑟𝑖
𝑝𝑖 𝑇 𝑣 = 𝑝𝑖 𝑇 𝐸𝑖 𝑣
𝑟𝑖
= 𝑝𝑖 𝑇 𝐸𝑖 𝑣
𝑟𝑖
= 𝑝𝑖 𝑇 𝑓𝑖 𝑇 𝑔𝑖 𝑇 𝑣
= 𝑝 𝑇 𝑔𝑖 𝑇 𝑣
=0 𝑓𝑜𝑟𝑎𝑙𝑙𝑖 (1 ≤ 𝑖 ≤ 𝑘).
𝑟𝑖
Thus 𝑣𝜖𝑊𝑖 and hence 𝑅𝑎𝑛𝑔𝑒𝐸𝑖 ⊆ 𝑊𝑖 . Conversely, let 𝑣𝜖𝑊𝑖 .Then 𝑝𝑖 𝑇 𝑣 = 0. Since 𝑝𝑖 𝑟 𝑖
divides 𝑓𝑗 𝑔𝑗 for 𝑗 ≠ 𝑖, we have 𝑓𝑗 𝑔𝑗 = 𝑠𝑝𝑖 𝑟 𝑖 for some polynomial 𝑠. Thus
𝐸𝑗 𝑣 = 𝑓𝑗 𝑇 𝑔𝑗 𝑇 𝑣
𝑟𝑖 𝑟
= 𝑠 𝑇 𝑝𝑖 𝑇 𝑣 ∵ 𝑓𝑗 𝑔𝑗 = 𝑠𝑝𝑖 𝑖
=0 𝑓𝑜𝑟𝑎𝑙𝑙𝑗 ≠ 𝑖.
Also, since 𝐸1 +. . . +𝐸𝑘 = 𝐼 and 𝐸𝑗 (𝑣) = 0 𝑓𝑜𝑟𝑎𝑙𝑙𝑗 ≠ 𝑖,wehave 𝐸𝑖 (𝑣) = 𝑣. Hence 𝑣𝜖𝑅𝑎𝑛𝑔𝑒𝐸𝑖 and
consequently,𝑊𝑖 ⊆ 𝑅𝑎𝑛𝑔𝑒𝐸𝑖 . Thus 𝑅𝑎𝑛𝑔𝑒𝐸𝑖 = 𝑊𝑖 .
Since for each 𝑖 (1 ≤ 𝑖 ≤ 𝑘), 𝑅𝑎𝑛𝑔𝑒𝐸𝑖 = 𝑊𝑖 and 𝐸1 +. . . +𝐸𝑘 = 𝐼therefore we must have
𝑉 = 𝑊1 +. . . +𝑊𝑘 .
For any 𝑣𝜖𝑉, let
𝑣 = 𝐸1 𝑢1 + ⋯ + 𝐸𝑘 𝑢𝑘
= 𝐸1 (𝑤1 )+. . . +𝐸𝑘 (𝑤𝑘 ).
Then for any 𝑖 (1 ≤ 𝑖 ≤ 𝑘),
𝐸𝑖 𝑣 = 𝐸𝑖 2 𝑢𝑖 = 𝐸𝑖 𝑢𝑖
𝑎𝑛𝑑𝐸𝑖 (𝑣) = 𝐸𝑖 2 (𝑤𝑖 ) = 𝐸𝑖 (𝑤𝑖 )
Thus 𝐸𝑖 (𝑢𝑖 ) = 𝐸𝑖 (𝑤𝑖 ) 𝑓𝑜𝑟𝑎𝑙𝑙𝑖 (1 ≤ 𝑖 ≤ 𝑘). Hence each element of 𝑉 is uniquely expressible as
sum of elements of 𝑊𝑖 ′𝑠. Therefore
𝑉 = 𝑊1 . . . 𝑊𝑘 .
Now since we have
𝑇𝐸𝑖 = 𝑇𝑓𝑖 𝑇 𝑔𝑖 𝑇
= 𝑓𝑖 𝑇 𝑔𝑖 𝑇 𝑇
= 𝐸𝑖 𝑇.
Therefore
𝑇 𝑊𝑖 = 𝑇 𝑅𝑎𝑛𝑔𝑒 𝐸𝑖 ⊆ 𝑅𝑎𝑛𝑔𝑒 𝐸𝑖 = 𝑊𝑖 ∵ 𝑇𝐸𝑖 = 𝐸𝑖 𝑇
It follows that for each 𝑖, 𝑊𝑖 is invariant under 𝑇.
Finally consider the operator 𝑇𝑖 induced on 𝑊𝑖 by 𝑇. Since
𝑟𝑖 𝑟𝑖
𝑝𝑖 𝑇 𝑣 = 0 𝑓𝑜𝑟𝑎𝑙𝑙𝑣𝜖𝑊𝑖 = 𝐾𝑒𝑟 𝑝𝑖 𝑇
𝑟𝑖
⇒ 𝑝𝑖 𝑇𝑖 𝑊𝑖 =0 .
Thus the minimal polynomial𝑔 of 𝑇𝑖 divides the polynomial 𝑝𝑖 𝑟 𝑖 .
Conversely, as 𝑔 is a minimal polynomial of 𝑇𝑖 , so𝑔(𝑇𝑖 ) = 0. Since 𝑔(𝑇) vanishes on 𝑊𝑖 and
𝑓𝑖 (𝑇) vanishes outside 𝑊𝑖 , therefore 𝑔(𝑇)𝑓𝑖 (𝑇) = 0. Thus polynomial 𝑔𝑓𝑖 is divisible by the
minimal polynomial 𝑝 of 𝑇. Hence in particular, 𝑝𝑖 𝑟 𝑖 divides 𝑔𝑓𝑖 . But then it implies that 𝑝𝑖 𝑟 𝑖
divides 𝑔 as 𝑝𝑖 𝑟 𝑖 is relatively prime with 𝑓𝑖 . Therefore the minimal polynomial 𝑔 for 𝑇𝑖 is
𝑝𝑖 𝑟 𝑖 . ∎
Do You Know?
We have not used the fact that 𝑉is finite-dimensional while constructing the proof.
Thus it is not necessary that vector space 𝑉 should be finite-dimensional for the
primary decomposition theorem to hold.
Also, we can drop the condition that the polynomial 𝑝 needs to be minimal and can
still obtain parts (i) and (ii).
Remark:
1. If 𝑇 is a linear operator as in Theorem 4.1.1 then 𝑇 can be decomposed as follows
𝑇 = 𝑇1 + ⋯ + 𝑇𝑘
where𝑇𝑖 = 𝑇𝐸𝑖 , 𝑖 = 1, … , 𝑘.
2. Since 𝐸𝑖 is the projection of 𝑉 onto 𝑊𝑖 and 𝑊𝑖 is the null space of 𝑝𝑖 (𝑇)𝑟 𝑖 , therefore
for each 𝑖 1 ≤ 𝑖 ≤ 𝑘 ,
𝑝𝑖 (𝑇)𝑟 𝑖 𝐸𝑖 = 0.
Thus, in particular, if 𝑝𝑖 = 𝑥 − 𝑐𝑖 , 𝑖 = 1, … , 𝑘 then
𝑟𝑖
𝑇 − 𝑐𝑖 𝐼 𝐸𝑖 = 0.
Also, since 𝑝 is a minimal polynomial of 𝑇, therefore for each 𝑖 1 ≤ 𝑖 ≤ 𝑘 , 𝑟𝑖 is the
least such integer.
6−𝑥 −3 −2
𝑑𝑒𝑡 𝐴 − 𝑥𝐼 = 4 −1 − 𝑥 −2 = − 𝑥 3 − 2𝑥 2 + 𝑥 − 2 = − 𝑥 2 + 1 𝑥 − 2 .
10 −5 −3 − 𝑥
Clearly, 𝑝 𝑥 = 𝑥 2 + 1 𝑥 − 2 is the minimal polynomial for 𝑇. Here 𝑝1 𝑥 = 𝑥 2 + 1and
𝑝2 𝑥 = 𝑥 − 2. Let 𝑊𝑖 be the null space of 𝑝𝑖 𝑇 , 𝑖 = 1,2 i.e.,
𝑊1 = 𝐾𝑒𝑟 𝑇 2 + 𝐼 𝑎𝑛𝑑 𝑊2 = 𝐾𝑒𝑟 𝑇 − 2𝐼 .
The vectors 𝛼 = (0,0,1) and 𝛽 = (1,1,0) are linearly independent and satisfy the equation
𝑇2 + 𝐼 𝑥 = 0 .
Hence the null space𝑊1 of 𝑇 2 + 𝐼 is spanned by the vectors 0,0,1 and 1,1,0 .Now let us
calculate the null space 𝑊2 of 𝑇 − 2𝐼. Consider
𝑇 − 2𝐼 𝑥 = 0
6 −3 −2 1 0 0 𝑥1
⇒ 4 −1 −2 − 2 0 1 0 𝑥2 = 0
10 −5 −3 0 0 1 𝑥3
4 −3 −2 𝑥1
⇒ 4 −3 −2 𝑥2 = 0
10 −5 −5 𝑥3
⇒ 2𝑥1 = 𝑥3 𝑎𝑛𝑑 𝑥2 = 0 .
Clearly, the vector 𝛾 = 1, 0, 2 satisfy the equation 𝑇 − 2𝐼 𝑥 = 0 and hence spans the null
space 𝑊2 of 𝑇 − 2𝐼. ThusΒ1 = 0,0,1 , 1,1,0 is the basis for 𝑊1 and Β2 = 1, 0, 2 is the
basis for 𝑊2 .Now,
6 −3 −2 0 −2 0 1 1
𝐴𝛼 = 4 −1 −2 0 = −2 = −3 0 − 2 1 + 0 0
10 −5 −3 1 −3 1 0 2
6 −3 −2 1 3 0 1 1
𝐴𝛽 = 4 −1 −2 1 = 3 = 5 0 + 3 1 + 0 0
10 −5 −3 0 5 1 0 2
6 −3 −2 1 2 0 1 1
𝐴𝛾 = 4 −1 −2 0 = 0 = 0 0 + 0 1 + 2 0 .
10 −5 −3 2 4 1 0 2
Let 𝑇𝑖 is the operator induced on 𝑊𝑖 by 𝑇 and 𝐴𝑖 be the matrix of 𝑇𝑖 w.r.t basis Β𝑖 . Then
−3 5 0 0 0 0
𝐴1 = −2 3 0 𝑎𝑛𝑑 𝐴2 = 0 0 0 . █
0 0 0 0 0 2
𝑑𝑛 𝑝 𝑥
𝐷𝑛 +1 𝑝 𝑥 = =0 ∀ 𝑝 𝑥 ∈𝑉 ℝ
𝑑𝑥 𝑛+1
𝑖. 𝑒., 𝐷𝑛 +1 = 0.
FAQ
Question :Let 𝑉 be a finite dimensional vector space over a field 𝐹. Does there
exist a natural number 𝑚 such that 𝑁 𝑚 = 0 for all nilpotent operators𝑁on 𝑉?
Answer :If 𝑉 is an 𝑛-dimensional vector space and 𝑁 is a nilpotent operator on 𝑉,
then there exists a positive integer 𝑘 ≤ 𝑛 such that 𝑁 𝑘 = 0. In particular, we
always have 𝑁 𝑛 = 0.
𝑝 = (𝑥 − 𝑐1 )𝑟1 … (𝑥 − 𝑐𝑘 )𝑟 𝑘
𝐷 = 𝑐1 𝐸1 + ⋯ + 𝑐𝑘 𝐸𝑘 .
Then obviously, 𝐷 is a diagonalizable operator (cf. Theorem 3.3.3). We will call the
operator 𝐷 as the diagonalizable part of 𝑻.
Proof: Let 𝑊𝑖 be the null space of (𝑇 − 𝑐𝑖 𝐼)𝑟 𝑖 and 𝐸𝑖 be the projection of 𝑉 onto the space
𝑊𝑖 . Consider the operator 𝐷 on vector space given by
𝐷 = 𝑐1 𝐸1 + ⋯ + 𝑐𝑘 𝐸𝑘 .
Then obviously, 𝐷 is a diagonalizable operator (cf. invariant direct sum Theorem 3.3.3).
Now consider another operator 𝑁 = 𝑇 − 𝐷.Since
𝐷 = 𝑐1 𝐸1 + ⋯ + 𝑐𝑘 𝐸𝑘 ,
therefore,
𝑁 = 𝑇 − 𝑐1 𝐼 𝐸1 + ⋯ + 𝑇 − 𝑐𝑘 𝐼 𝐸𝑘 .
𝑁 2 = 𝑇 − 𝑐1 𝐼 2 𝐸1 + ⋯ + 𝑇 − 𝑐𝑘 𝐼 2 𝐸𝑘
and in general we have
𝑁 𝑟 = 𝑇 − 𝑐1 𝐼 𝑟 𝐸1 + ⋯ + 𝑇 − 𝑐𝑘 𝐼 𝑟 𝐸𝑘 . (𝑟 ≥ 1)
𝑟𝑖
Also, since 𝑇 − 𝑐𝑖 𝐼 𝐸𝑖 = 0, 𝑖 = 1, … , 𝑘 , therefore we shall have
𝑁𝑟 = 0 𝑤𝑒𝑛𝑒𝑣𝑒𝑟𝑟 ≥ max 𝑟𝑖 .
1≤𝑖≤𝑘
(i). 𝑇 =𝐷+𝑁
(ii). 𝐷𝑁 = 𝑁𝐷
The diagonalizable operator 𝐷 and the nilpotent operator are uniquely determined by (i)
and (ii) and each of them is a polynomial in 𝑇.
Proof:As already observed in Theorem 4.2.4 and the remark following it that we can
write 𝑇 = 𝐷 + 𝑁 where 𝐷 is a diagonalisable operator and 𝑁 is nilpotent operator, and
further,𝐷 and 𝑁 are polynomials in 𝑇 which commute with each other.
Now suppose there exists diagonalisable operator 𝐷′ and nilpotent operator 𝑁′ such that
𝑇 = 𝐷′ + 𝑁′ and 𝐷 ′ 𝑁 ′ = 𝑁 ′ 𝐷′ .We claim that 𝐷 = 𝐷′ and 𝑁 = 𝑁 ′ .
Consider
𝐷′ 𝑇 = 𝐷′ 𝐷′ + 𝑁 ′ = 𝐷′ 𝐷′ + 𝐷′ 𝑁 ′ = 𝐷′ 𝐷′ + 𝑁 ′ 𝐷′ = 𝐷′ + 𝑁 ′ 𝐷′ = 𝑇𝐷′ .
Since 𝐷 and 𝐷 ′ are both diagonalizable and they commute, therefore by Theorem 3.1.4𝐷
and 𝐷′ are simultaneously diagonalizable. Hence 𝐷 − 𝐷′ is diagonalizable operator. Since
𝑇 = 𝐷 + 𝑁 = 𝐷′ + 𝑁′, therefore we have 𝐷 − 𝐷′ = 𝑁′ − 𝑁. Thus (𝑁 ′ − 𝑁) is a diagonalizable
operator.
Since 𝑁 and 𝑁′ are nilpotent operators, therefore we can choose smallest positive integer
𝑚 such that 𝑁 𝑗 = 0 and 𝑁′𝑗 = 0 for all 𝑗 ≥ 𝑚. Again as 𝑁 and 𝑁 ′ commute, we can expand
(𝑁 ′ − 𝑁)2𝑚 binomially as
2𝑚
′ 2𝑚 2𝑚 2𝑚 −𝑘
𝑁 −𝑁 = 𝑁′ −𝑁 𝑘
𝑘
𝑘=0
Proof: For any linear operator 𝑇, the minimal polynomial for 𝑇 decomposes over an
algebraically closed field 𝐹 into a product of linear polynomials. Therefore by Theorem
4.2.5 it follows. ∎
nilpotent operator 𝑁 which commute. These operators 𝐷 and 𝑁 are unique and each is a
polynomial in 𝑇.
Example 4.2.8 Again consider the linear operator 𝑇 of Example 4.1given by the matrix
3 1 −1
𝐴= 2 2 −1 .
2 2 0
2
𝑊2 = 𝐾𝑒𝑟 𝑇 − 2𝐼 = 𝑠𝑝𝑎𝑛 0, 0, 1 , 1, 1, 0 .
Since
1,0,0 = 1 1, 0, 2 − 2 0, 0, 1 + 0 1, 1, 0
0,1,0 = −1 1, 0, 2 + 2 0, 0, 1 + 1 1, 1, 0
0, 0, 1 = 0 1, 0, 2 + 1 0, 0, 1 + 0 1, 1, 0
The projection 𝐸1 of 𝑉 onto the space 𝑊1 acts on the standard ordered basis as follows:
𝐸1 1, 0, 0 = 1 1, 0, 2 = 1 1, 0, 0 + 0 0, 1, 0 + 2 0, 0, 1
𝐸1 0, 1, 0 = − 1, 0, 2 = − 1, 0, 0 + 0 0, 0, 1 − 2 0, 0, 1
𝐸1 1, 0, 0 = 0 1, 0, 2 = 0 1, 0, 0 + 0 0, 0, 1 + 0 0, 0, 1 .
1 −1 0 0 1 0
Therefore 𝐸1 = 0 0 0 . Similarly, it can be easily seen that 𝐸2 = 0 1 0 . Thus the
2 −2 0 −2 2 1
diagonalizable operator 𝐷 of 𝑇 is given by
1 −1 0 0 2 0 1 1 0
𝐷 = 1𝐸1 + 2𝐸2 = 0 0 0 + 0 2 0 = 0 2 0 .
2 −2 0 −4 4 2 −2 2 2
3 1 −1 1 1 0 2 0 −1
𝑁 =𝐴−𝐷 = 2 2 −1 − 0 2 0 = 2 0 −1
2 2 0 −2 2 2 4 0 −2
0 0 0
Observe that 𝑁 2 = 0 0 0 . █
0 0 0
Thus, from all the above results we see that to study linear operators having minimal
polynomial of the first type, it is enough to focus on nilpotent operators. But for linear
operators with minimal polynomial of second type, we need to find out some way other
than characteristic values and vectors. For example consider the linear operator𝑇 on 𝑅2
given by the matrix
0 −1
𝐴=
1 0
−𝑥 −1
𝑑𝑒𝑡(𝐴 − 𝑥𝐼) = = 𝑥2 + 1
1 −𝑥
Clearly, this polynomial has no real roots and hence the operator 𝑇 has no characteristic
value. How to deal with such operators ?In the next chapters we will show that these
two problems can be simultaneously handled.
Now let us look at some of the problems based on the discussions above. These
problems will take us deeper into this concept of decomposition.
5. SOLVED PROBLEMS
𝑇 =𝐷+𝑁
where𝐷 is the diagonalizable part of 𝑇 and 𝐷 and 𝑁 both commute. Now since both 𝐷 and
𝑁 commute, therefore by using simple binomial expansions we can write
𝑔 𝑇 = 𝑔 𝐷 + 𝑁 𝐷, 𝑁
𝑘
𝑁 𝐷, 𝑁 = 𝑁 𝑘 𝐷, 𝑁 𝑘
=0 𝑠𝑖𝑛𝑐𝑒 𝐷 𝑎𝑛𝑑 𝑁 𝑐𝑜𝑚𝑚𝑢𝑡𝑒
𝑔 𝑇 = 𝑔 𝐷 + 𝑁 𝐷, 𝑁
Problem 5.2. Let 𝑉 be a finite-dimensional vector space over the field 𝐹, and let 𝑇 be a
linear operator on 𝑉 such that 𝑟𝑎𝑛𝑘 𝑇 = 1. Prove that either 𝑇 is diagonalizable or 𝑇 is
nilpotent, not both.
Now if dim 𝑉 = 1, then obviously 𝑇 𝑛 = 𝑐 𝑛 𝐼 for some scalar 𝑐(≠ 0) 𝜖𝐹 and for all 𝑛 ∈ ℕ. Hence
𝑇 is diagonalizable and 𝑇 is not nilpotent. Thereforelet dim 𝑉 > 1. Then dim 𝐾𝑒𝑟 𝑇 ≥ 1 and
hence 0 is a characteristic value of 𝑇.
Now from Theorem 4.2.2, it is clear that 𝑇 cannot be both diagonalizable and nilpotent.
If 𝑇 is nilpotent, then we have nothing to prove. Thus let 𝑇be not nilpotent.
Claim:𝑇(𝑥) ≠ 0.
Suppose on the contrary, 𝑇(𝑥) = 0 and let 𝑦𝜖𝑉 be arbitrary. Now since 𝑇(𝑦) 𝜖𝑅𝑎𝑛𝑔𝑒 𝑇 ,
therefore 𝑇(𝑦) = 𝛼𝑥 for some 𝛼𝜖𝐹. But then we have
Since 𝑦𝜖𝑉 is arbitrary, we have 𝑇 2 = 0 which contradicts the fact that 𝑇 is not nilpotent.
Hence our assumption is wrong and 𝑇(𝑥) ≠ 0.
Now as 𝑇(𝑥) 𝜖𝑅𝑎𝑛𝑔𝑒 𝑇 , there exists 𝛽 ≠ 0 𝜖𝐹 such that 𝑇(𝑥) = 𝛽𝑥.Then 𝛽is a characteristic
value of 𝑇. We will show that 𝛽 is the only non-zero characteristic value of 𝑇. On the
contrary. suppose𝛾 ≠ 𝛽 is a non-zero characteristic value of 𝑇. There exists 𝑧 ≠ 0 ∈ 𝑉
such that 𝑇 𝑧 = 𝛾𝑧. Since 𝑇 𝑧 ∈ 𝑅𝑎𝑛𝑔𝑒 𝑇 = 𝑠𝑝𝑎𝑛 𝑥 , 𝑧, 𝑥 is a linearly dependent set, a
contradicton to the fact that 𝑧 and 𝑥 are characteristic vectors of distinct non-zero
characteristic values. Thus 𝛽 is the only non-zero characteristic value of 𝑇.Let𝑊1 be the
null space of 𝑇 − 𝛽𝐼 and 𝑊2 be the null space of 𝑇 − 0𝐼 = 𝑇.Since
𝑉 = 𝑅𝑎𝑛𝑔𝑒 𝑇 ⊕ 𝐾𝑒𝑟 𝑇 = 𝑊1 ⊕ 𝑊2
it follows that 𝑇 = 𝛽𝐸1 + 0𝐸2 , where 𝐸𝑖 be the projection of 𝑉onto𝑊𝑖 , 𝑖 = 1,2.Hence 𝑇is
diagonalizable operator. ∎
Problem 5.3. Let𝑉be a finite-dimensional vector space over the field 𝐹, and let𝑇be a
linear operator on𝑉. Suppose that𝑇commutes with every diagonalizable linear operator
on𝑉. Prove that 𝑇is a scalar multiple of the identity operator.
Proof: Suppose𝑇is not a scalar multiple of identity operator, then there exist linearly
independent vectors𝑥and𝑦in𝑉such that𝑇 𝑥 = 𝑦.First we will construct a linear operator 𝑆
on 𝑉which is diagonalizable and hence commutes with 𝑇. Let 𝑊1 = {𝛼𝑥 ∶ 𝛼 𝜖 𝐹}and
𝑊2 = {𝛽𝑦 ∶ 𝛽 𝜖 𝐹}. Then obviously, 𝑊1 ∩ 𝑊2 = 0 as 𝑥and𝑦 are linearly independent vectors.
Now let 𝑊0 be the subspace of𝑉 such that 𝑊0 is complementaryto 𝑊1 + 𝑊2 = 𝑊1 𝑊2 .
Then we have vector space𝑉as
𝑉 = 𝑊0 𝑊1 𝑊2 .
𝑆 𝑣 = 𝑆 𝑣0 + 𝑣1 + 𝑣2 = 𝑣1 + 2𝑣2
which implies that 𝑦 = 0, a contradiction to the fact that 𝑦 ≠ 0. Hence our assumption is
wrong. Thus 𝑇 is a scalar multiple of an identity operator. ∎
Now since 𝑁 is a nilpotent operator, there exists a positive integer 𝑘 such that 𝑁 𝑘 = 0.
Thus 𝑁 satisfies the polynomial 𝑥 = 𝑥 𝑘 . Hence 𝑔 must divide . Therefore 0 is the only
root of 𝑔. Consequently, 0 is the only root of 𝑓 repeated 𝑛-times. Thus characteristic
polynomial for 𝑁 is 𝑥 𝑛 . ∎
In the next section we discuss one application of the primary decomposition theorem.
We will see how we use the primary decomposition theorem to decompose a complicated
differential equation into much simpler differential equations.
6. APPLICATION
Let 𝑉 denote the space of all 𝑛 times continuously differentiable functions 𝑓 on ℂ (set of
complex numbers) which satisfy the differential equation
𝐷𝑛 𝑓 + 𝑎𝑛−1 𝐷𝑛 −1 𝑓 + ⋯ + 𝑎1 𝐷𝑓 + 𝑎0 𝑓 = 0 (𝐴)
𝑝 𝑥 = 𝑥 𝑛 + 𝑎𝑛−1 𝑥 𝑛−1 + ⋯ + 𝑎1 𝑥 + 𝑎0
then clearly, we have
𝑉 = 𝑓𝜖𝐶𝑛 ∶ 𝑝 𝐷 𝑓 = 0
= 𝑛𝑢𝑙𝑙𝑠𝑝𝑎𝑐𝑒𝑜𝑓𝑝 𝐷 .
𝑝 = (𝑥 − 𝑐1 )𝑟1 … (𝑥 − 𝑐𝑘 )𝑟 𝑘
where𝑐1 , … , 𝑐𝑘 are distinct complex numbers. If 𝑊𝑖 is the null space of (𝐷 − 𝑐𝑖 𝐼)𝑟 𝑖 , then
since the primary decomposition theorem holds for an arbitrary vector space and for any
monic polynomial annihilating the given linear operator, therefore we have
𝑉 = 𝑊1 ⊕ … ⊕ 𝑊𝑘 .
Thus if 𝑓 satisfies the differential equation (A), then we can express 𝑓 uniquely as
𝑓 = 𝑓1 + ⋯ + 𝑓𝑘
where𝑓𝑖 satisfies the differential equation (𝐷 − 𝑐𝑖 𝐼)𝑟 𝑖 𝑓𝑖 = 0. Thus, the problem of finding
solutions for the differential equation of the form aboveis reduced to the problem of
finding solutions for much simpler differential equations of the form
𝑟𝑖
𝐷 − 𝑐𝑖 𝐼 𝑓 = 0.
Hence by the primary decomposition theorem, we have reduced our original differential
equation into simpler and easily solvable differential equations.
𝐷 − 𝑐𝐼 𝑘 𝑓 = 0 (𝐵)
we need to know something about differential equations ,i.e., we should know a little
more than the fact that 𝐷 is a linear operator. It is very easy to show by induction on 𝑘
that if 𝑓 𝜖 𝐶𝑘 , then
𝐷 − 𝑐𝐼 𝑘 𝑓 = 𝑒 𝑐𝑡 𝐷𝑘 𝑒 −𝑐𝑡 𝑓 .
𝑔 𝑡 = 𝑏0 + 𝑏1 𝑡 + ⋯ + 𝑏𝑘−1 𝑡 𝑘−1 .
𝑓 = 𝑒 𝑐𝑡 𝑏0 + 𝑏1 𝑡 + ⋯ + 𝑏𝑘−1 𝑡 𝑘−1 .
Hence we see that the functions 𝑒 𝑐𝑡 , 𝑡𝑒 𝑐𝑡 , … , 𝑡 𝑘−1 𝑒 𝑐𝑡 span the space of solutions of (B). Now
since 1, 𝑡, … , 𝑡 𝑘−1 are linearly independent and the exponential functions has no zeroes,
therefore the functions 𝑒 𝑐𝑡 , 𝑡𝑒 𝑐𝑡 , … , 𝑡 𝑘−1 𝑒 𝑐𝑡 are linearly independent and hence form a basis
for the space of solutions of (B).
Returning to the differential equation (A), from the above discussions we see that the set
𝔅= { 𝑒 𝑐𝑖 𝑡 , 𝑡𝑒 𝑐𝑖 𝑡 , … , 𝑡 𝑟 𝑖 −1 𝑒 𝑐𝑖 𝑡 }
𝑖=1
of𝑛 functions forms a basis for the space 𝑉 of solutions of (A). Thus in particular, we
have that 𝑉 is finite-dimensional and hence has dimension equal to the degree of the
polynomial 𝑝.
7. SUMMARY
6. EXERCISES
1. Let 𝑇be a linear operator on 𝑅3 which is represented in the standard ordered basis by
the matrix
6 −3 −2
4 −1 −2
10 −5 −3
Express the minimal polynomial 𝑝for 𝑇 in the form 𝑝 = 𝑝1 𝑝2 , where 𝑝1 and 𝑝2 are monic
and irreducible over the field of real numbers. Let 𝑊𝑖 be the null space of 𝑝𝑖 (𝑇). Find bases
𝔅𝑖 for the spaces 𝑊1 and 𝑊2 .If 𝑇𝑖 is the operator induced on𝑊𝑖 by 𝑇, find the matrix of 𝑇𝑖 in
the basis 𝔅𝑖 (above).
3. If 𝑉is the space of all polynomials of degree less than or equal to 𝑛over a field 𝐹,
prove that the differentiation operator on 𝑉is nilpotent.
4.Let 𝑉 be the space of 𝑛 × 𝑛matrices over a field𝐹, and let 𝐴 be a fixed 𝑛 × 𝑛matrix
over 𝐹. Define a linear operator 𝑇 on 𝑉 by 𝑇 𝐵 = 𝐴𝐵 − 𝐵𝐴. Prove that if 𝐴 is a nilpotent
matrix, then 𝑇 is a nilpotent operator.
9. REFERENCES
[1] Kenneth Hoffman and Ray Kunze, Linear Algebra (Second Edition), Pearson
Education Inc., India 2005.
[2] Roger A. Horn and Charles R. Johnson, Matrix Analysis (Second Edition),