La 2
La 2
Duc Vu
Spring 2021
This is math 115B – Linear Algebra which is the second course of the undergrad
linear algebra at UCLA – continuation of 115A(H). Similar to 115AH, this class is
instructed by Professor Elman, and we meet weekly on MWF from 2:00 pm to 2:50
pm. There is no official textbook used for the class. You can find the previous linear
algebra notes (115AH) with other course notes through my github. Any error in this
note is my responsibility and please email me if you happen to notice it.
Contents
2
Duc Vu (Spring 2021) List of Theorems
List of Theorems
10.12Fundamental Theorem of Arithmetic (Polynomial Case) . . . . . . . . . . . 44
11.6 Cayley-Hamilton . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 47
12.10Fundamental Theorem of Algebra . . . . . . . . . . . . . . . . . . . . . . . . 53
14.2 Primary Decomposition . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 58
22.5 Orthogonal Decomposition . . . . . . . . . . . . . . . . . . . . . . . . . . . . 93
22.6 Best Approximation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 93
24.8 Schur . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 99
25.5 Spectral Theorem for Normal Operators . . . . . . . . . . . . . . . . . . . . 103
26.4 Spectral Theorem for Hermitian Operators . . . . . . . . . . . . . . . . . . 104
27.9 Singular Value . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 110
28.5 Singluar Value - Linear Transformation Form . . . . . . . . . . . . . . . . . 114
28.7 Polar Decomposition . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 114
30.4 Minimax Principle . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 123
3
Duc Vu (Spring 2021) List of Definitions
List of Definitions
1.1 Field . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
1.3 Ring . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
1.6 Vector Space . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
2.3 Subspace . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
2.8 Span . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
2.9 Direct Sum . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
3.1 Independent Subspace . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
3.6 Complementary Subspace . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
6.5 T-invariant . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24
6.9 Projection . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25
7.9 Dual Space . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
8.8 Annihilator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36
9.7 Transpose . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39
9.11 Row/Column Rank . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
9.13 Polynomial Division . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
9.16 Polynomial Degree and Leading Coefficient . . . . . . . . . . . . . . . . . . 41
10.1 Greatest Common Divisor . . . . . . . . . . . . . . . . . . . . . . . . . . . . 42
10.6 Irreducible Polynomial . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
12.2 Triangularizability . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 50
12.4 Splits . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 51
12.8 Algebraically Closed . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 53
15.2 Jordan Block Matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 63
15.3 Nilpotent . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 63
16.1 Sequence of Generalized Eigenvectors . . . . . . . . . . . . . . . . . . . . . . 66
16.3 Jordan Canonical Form . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67
16.4 Jordan Basis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67
18.5 Companion Matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 76
19.2 T-Cyclic . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 78
19.7 Equivalent Matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 81
22.1 Inner Product Space . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 91
22.2 Sesquilinear Map . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 91
22.8 Adjoint . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 94
23.2 Isometry . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 95
24.1 Unitary Operator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 98
24.4 Unitary Matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 98
25.1 Hermitian(Self-Adjoint) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 101
27.1 Positive/Negative (Semi-) Definite . . . . . . . . . . . . . . . . . . . . . . . 106
27.8 Pseudodiagonal . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 109
28.1 Singular Value Decomposition . . . . . . . . . . . . . . . . . . . . . . . . . . 113
4
Duc Vu (Spring 2021) 1 Lec 1: Mar 29, 2021
§1.1 Ve c t o r S p a c e s
Notation: if ? : A × B → B is a map (= function) write a ? b for ?(a, b), e.g., + : Z × Z → Z
where Z = the integer.
• Addition: + : F × F → F
• Multiplication: · : F × F → F
if ∀a, b, c ∈ F , we have
A1) (a + b) + c = a + (b + c)
A2) ∃ 0 ∈ F 3 a + 0 = a = 0 + a
A4) a + b = b + a
M1) (a · b) · c = a · (b · c)
M4) x · y = y · x
D1) a · (b + c) = a · b + a · c
D2) (a + b) · c = a · c + b · c
Example 1.2
Q, R, C are fields as is
+ 0 1 0 1
0 0 1 0 0 0
1 1 0 1 0 1
5
Duc Vu (Spring 2021) 1 Lec 1: Mar 29, 2021
Fact 1.1. Let p > 0 be a prime number in Z. Then ∃ a field Fpn having pn elements write
|Fpn | = pn ∀n ∈ Z+ .
• +:R×R→R
• ·:R×R→R
satisfying A1) – A4), M1), M2), D1), D2), then R is called a RING.
A ring is called
M 3’) a · b = 0 =⇒ a = 0 or b = 0
3. Let F be a field
F [t] := {polys coeffs in F }
with usual +, · of polys, is a domain but not a field. So if f ∈ F [t]
f = a0 + a1 t + . . . + an tn
where a0 , . . . , an ∈ F .
(when does ab = dc ?)
5. If F is a field
f
F (t) := |f, g ∈ F [t], g 6= 0 (rational function)
g
6
Duc Vu (Spring 2021) 1 Lec 1: Mar 29, 2021
√ √
Example 1.5√(Cont’d from above) 6. Q[ −1] := α + β −1 ∈ C|α, β ∈ Q < C.
Then Q[ −1] is a field and
√ na √ o
Q( −1) := |a, b ∈ Q[ −1], b 6= 0
b√
= Q[ −1]
na √ o
= |a, b ∈ Z[ −1], b 6= 0
b
√ √
where Z[ √−1] := α + β −1 ∈ C, α, β ∈ Z < C. How to show this? – ratio-
nalize (Z[ −1] is a domain not a field, F [t] < F (t) if F is a field so we have to
be careful).
7. F a field
Mn F := {n × n matrices entries in F }
is a ring under +, · of matrices.
1 0
1 Mn F = In = n × n identity matrix . . .
0 1
0 ... 0
0 Mn F = 0 = 0n = n × n zero matrix ... ..
.
0 ... 0
Mn R = {n × n matrices entries in R}
8. Let ∅ =
6 I ⊂ R be a subset, e.g., [α, β] , α < β ∈ R. Then
for all x ∈ I.
7
Duc Vu (Spring 2021) 1 Lec 1: Mar 29, 2021
• + : V × V → V – Addition
• · : F × V → V – Scalar multiplication
if ∀x, y, z ∈ V ∀α, β ∈ F .
1. (x + y) + z = x + (y + z)
4. x + y = y + x
5. 1F · x = x.
6. (α · β) · x = α · (β · x)
7. (α + β) · x = α · x + β · x
8. α · (x + y) = α · x + α · y
Remark 1.7. The usual properties we learned in 115A hold for V a vector space over F , e.g.,
0F V = 0V , general association law,. . .
8
Duc Vu (Spring 2021) 2 Lec 2: Mar 31, 2021
§2.1 Ve c t o r S p a c e s ( C o nt ’ d )
Example 2.1
The following are vector space over F
Fcn(S, V ) := {f : S → V | f a fcn}
f + g : x 7→ f (x) + g(x)
αf : x 7→ αf (x)
i.e.
with 0 by 0(x) = 0V ∀x ∈ S.
+
|{z} F × F : F × F → F and |{z}
· F ×F :F ×F →F
on R
on R
| {z } | {z }
restrict dom restrict dom
· F ×R
:F ×R→R
9
Duc Vu (Spring 2021) 2 Lec 2: Mar 31, 2021
§2.2 Subspaces
αw1 , w1 + w2 ∈ W
with 0W = 0V is a vector space over F under +|W ×W and ·|F ×W i.e., closed under the
operation on V .
Theorem 2.4
Let V be a vector space over F, ∅ =
6 W ⊆ V a subset. Then W is a subspace of V iff
∀α ∈ F , ∀w1 , w2 ∈ W , αw1 + w2 ∈ W .
Example 2.6
Let n > 1 in Z. Then n is a product of primes unique up to order.
10
Duc Vu (Spring 2021) 2 Lec 2: Mar 31, 2021
Example 2.7
Let V be a finite dimensional inner product space over R (or C) and T : V → V a
hermitian (=self adjoint) operator. Then ∃ an ON basis for V consisting of eigenvectors
for T . In particular, T is diagonalizable. This means
ET (λi ) := {v ∈ V | T v = λi v} =
6 0 eigenspace of λi ; λ1 , . . . , λr the distinct eigenvalues
of T . So
T E (λ ) : ET (λi ) → ET (λi )
T i
T ET (λi )
= λi 1ET (λi )
Goal: Generalize this to V any finite dimensional vector space over F , any F , and T : V → V
linear. We have many problems to overcome in order to get a meaningful result, e.g.,
2. F 6= R or C is possible.
11
Duc Vu (Spring 2021) 2 Lec 2: Mar 31, 2021
when almost all zero means only finitely many wi 6= 0. Warning: In a vector space/F
we can only take finite linear combination of vectors. So
! ( )
X [ [
Wi = Span Wi = finite linear combos of vectors in Wi
i∈I i∈I i∈I
w ∈ W1 ⊕ . . . ⊕ Wn means ∃! wi ∈ Wi 3 w = w1 + . . . + wn
12
Duc Vu (Spring 2021) 3 Lec 3: Apr 2, 2021
§3.1 D i r e c t S u m s ( C o nt ’ d )
Theorem 3.2
Let V be a vector space over F, Wi ⊆ V, i ∈ I subspaces, W ⊆ V a subspace. Then
the following are equivalent:
L
1. W = i∈I Wi
P
2. W = i∈I Wi and ∀i
X
Wi ∩ Wj = 0 := {0}
j∈I\{i}
P
3. W = i∈I Wi and the Wi , i ∈ I, are independent.
L P
Proof. 1) =⇒ 2) Suppose W = i∈I Wi . Certainly, W = i∈I Wi . Fix i and suppose
that X
∃x ∈ Wi ∩ Wj
j∈I\{i}
So X
0V = 0W = wi − wj
j6=i
But X
0W = 0Wk 0Wk = 0V ∀k ∈ I
I
By uniqueness of 1), wi = 0 so x = 0.
2) =⇒ 3) Let wi ∈ Wi , i ∈ I, almost all zero satisfy
X
wi = 0
i∈I
13
Duc Vu (Spring 2021) 3 Lec 3: Apr 2, 2021
a contradiction. So wi =P0 ∀i
3) =⇒ 1) Suppose v ∈ i∈I Wi and ∃wi , wi0 ∈ Wi , i ∈ I, almost all 0 3
X X
wi = v = wi0
i∈I i∈I
Corollary 3.3
Then
V = W I1 ⊕ W I2
Example 3.4
Let V be a vector space over F .
1. If ∅ =
6 S ⊆ V is a subset, then
X
F v = Span(S)
v∈S
the span of S. So
2. If ∅ =
6 S is linearly indep. (i.e. meaning every finite nonempty subset of S is
linearly indep.), then M
Span(S) = Fs
s∈S
14
Duc Vu (Spring 2021) 3 Lec 3: Apr 2, 2021
L
Example 3.5 (Cont’d from above) 3. If S is a basis for V , then V = s∈S F s.
P
4. If ∃ a finite set S ⊆ V 3 V = Span(S), then V = s∈S F s and ∃ a subset
B ⊆ S that is a basis for V , i.e., V is a finite dimensional vector space over F
and dim V = dimF V = |B| is indep. of basis B for V .
So
W1 + W2 = W1 ⊕ W2 ⇐⇒ W1 ∩ W2 = ∅
Warning: be very careful if you wish to generalize this.
V = W ⊕ W 0, W 0 ⊆ V a subspace
Example 3.7
Let B0 be a basis of W . Extend B0 to a basis B for V (even works if V is not finite
dimensional). Then
M
W0 = F v is a complement of W in V
B\B0
15
Duc Vu (Spring 2021) 3 Lec 3: Apr 2, 2021
§3.2 Q u o t i e nt S p a c e s
Idea: Given a surjective map f : X → Y and “nice”, can we use properties of Y to obtain
properties of X?
Example 3.8
Let V = R3 , W = X − Y plane. Let X = plane parallel to W intersecting the z-axis
at γ.
z
V = R3
γ
v
X
W
x
So
X = {(α, β, γ)|α, β ∈ R}
= {(α, β, 0) + (0, 0, γ)|α, β ∈ R}
= W + γ e3
|{z}
(0,0,1)
= {(α + x, β + y, γ) | α, β ∈ R}
= W + γe3
It follows if v, v 0 ∈ V , then
W + v = W + v 0 =⇒ v − v 0 ∈ W
v 0 ∈ X =⇒ v 0 = w + v some w ∈ W
hence
v0 − v ∈ W
So for arbitrary v, v 0 ∈ V , we have the conclusion W + v = W + v 0 ⇐⇒ v − v 0 ∈ W .
We can also write W + v as v + W .
16
Duc Vu (Spring 2021) 4 Lec 4: Apr 5, 2021
§4.1 Q u o t i e nt S p a c e s ( C o nt ’ d )
Recall from the last example of the last lecture, we have
[
V = W +v
v∈V
If v, v 0 ∈ V , then
0 6= v 00 ∈ (W + v) ∩ (W + v 0 )
means
W + v − W + v 00 = W + v 0
This means either W + v = W + v 0 or W + v ∩ W + v 0 = ∅, i.e., planes parallel to the
xy-plane partition V into a disjoint unions of planes.
Let
S := {W + v| v ∈ V }
the set of these planes. We make S into a vector space over R as follows: ∀v, v 0 ∈ V, ∀α ∈ R
define
(W + v) + (W + v 0 ) := W + (v + v 0 )
α · (W + v) := W + αv
0S := W
S → {(0, 0, γ)| γ ∈ R} → R
by
W + v 7→ (0, 0, γv ) 7→ γ
both maps are bijection and, in fact, linear isomorphism. So
S∼
= {(0, 0, γ)| γ ∈ R} ∼
=R
V → S by (α, β, γ) 7→ W + γe3
a surjection.
We can now generalize this.
Construction: Let V be a vector space over F, W ⊆ V a subspace. Define ≡ mod W
called congruent mod W on V as follows: if x, y ∈ V, then
x≡y mod W ⇐⇒ x − y ∈ W ⇐⇒ ∃w ∈ W 3 x = w + y
17
Duc Vu (Spring 2021) 4 Lec 4: Apr 5, 2021
1. x ≡ x mod W
2. x ≡ y mod W =⇒ y ≡ x mod W
x := {y ∈ V | y ≡ x mod W }
We can also write x as [x]W if W is not understood. Also, x ⊆ V is a subset and not an
element of V called a coset of V by W . We have
x = {y ∈ V | y ≡ x mod W }
= {y ∈ V | y = w + x for some w ∈ W }
= {w + x| w ∈ W } = W + x = x + W
Example 4.1
0V = W + 0V = W .
y ∈ x = W + x ⇐⇒ x ∈ y = W + y
and
x≡y mod W ⇐⇒ x = y ⇐⇒ W + x = W + y
and check
x ∩ y = ∅ ⇐⇒ (W + x) ∩ (W + y) = ∅ ⇐⇒ x 6≡ y mod W
This means the W + x partition V , i.e.,
[
V = (W + x) with (W + x) ∩ (W + y) = ∅ if x = (W + x) 6= (W + y) = y
V
Let
V := V /W := {x| x ∈ V } = {W + x| x ∈ V }
a collection of subsets of V .
18
Duc Vu (Spring 2021) 5 Lec 5: Apr 7, 2021
§5.1 Q u o t i e nt S p a c e s ( C o nt ’ d )
Suppose we have W ⊆ V a subspace. For x, y, z, v ∈ V
Then
(x + z) − (y + v) = (x − y) + (z − v) ∈ W
| {z } | {z }
∈W ∈W
So
x+z mod y + v mod W
and if α ∈ F
αx − αy = α(x − y) ∈ W ∀x, y ∈ V
So
αx ≡ αy mod W
Therefore, V = V /W . If (+) holds, then for all x, y, z, v ∈ V and α ∈ F , we have
x+z =y+v ∈V
αx = αy ∈ V
19
Duc Vu (Spring 2021) 5 Lec 5: Apr 7, 2021
Proposition 5.1
Let V be a vector space over F , W ⊆ V a subspace, V = V /W . Let B0 be a basis for
W and
B1 = {vi | i ∈ I, vi − vj ∈
/ W if i 6= j}
where vi 6= vj if i 6= j or w + vi 6= w + vj if i 6= j.
Let
C = {vi = W + vi | i ∈ I, vi ∈ B1 }
If C is a basis for V = V /W , then B0 ∪ B1 is a basis for V (compare with the proof
of the Dimension Theorem).
Proof. Hw 2 # 3.
§5.2 L i n e a r Tr a n s f o r m a t i o n
A review of linear of linear transformation can be found here.
Now, we consider
GLn F := {A ∈ Mn F | det A =
6 0}
The elements in GLn F in the ring Mn F are those having a multiplicative inverse. If R is a
commutative ring, determinants are still as before but
Example 5.2
Let V be a vector space over F, W ⊆ V a subspace. Recall
V = V /W = {v = W + v| v ∈ V }
0V = 0V = W
v1 + v2 = v1 + v2
αv1 = αv1
Then
– : V → V /W = V by v 7→ v = W + v
is an epimorphism with ker– = W .
Recall from 115A(H) that the most important theorem about linear transformation is
Universal Property of Vector Spaces. As a result, we can deduce the following corollary
20
Duc Vu (Spring 2021) 5 Lec 5: Apr 7, 2021
Corollary 5.3
Let V, W be vector space over F with bases B, C respectively. Suppose there exists a
bijection f : B → C , i.e., |B| = |C |. Then V ∼
= W.
21
Duc Vu (Spring 2021) 6 Lec 6: Apr 9, 2021
§6.1 L i n e a r Tr a n s f o r m a t i o n ( C o nt ’ d )
Theorem 6.1
Let T : V → W be linear. Then ∃X ⊆ V a subspace s.t.
V = ker T ⊕ X with X ∼
= im T
Proof. Let B0 be a basis for ker T . Extend B0 to a basis B for V by the Extension
Theorem. Let B1 = B \ B0 , so B = B0 ∨ B1 (B = B0 ∪ B1 and B0 ∩ B1 = ∅) and let
M
X= Fv
B1
L
As ker T = B0 F v, we have
V = ker T ⊕ X
and we have to show
X∼
= im T
Then
X X
0W = T αv v , i.e. αv v ∈ ker T
v∈B1 B1
Hence X X
αv v = βv v ∈ ker T almost all βv ∈ F = 0
B1 B0
As B1 αv v − B0 βv v = 0 and B = B0 ∪ B1 is linearly indep., αv = 0 ∀v. This proves
P P
the above claim.
Let C = {T v| v ∈ B1 }. By the claim
B1 → C by v 7→ T v is 1 − 1
and onto as C is linearly indep. Lastly, we must show C spans im T . Let w ∈ im T . Then
∃x ∈ V 3 T x = w. Then
X X
w = Tx = T αv v + T αv v
B0 B1
X X X
= αv T v + αv T v = αv T v
B0 B1 B1
22
Duc Vu (Spring 2021) 6 Lec 6: Apr 9, 2021
Remark 6.2. Note that the proof is essentially the same as the proof of the Dimension
Theorem.
Corollary 6.4
If V is a finite dimensional vector space over F , W ⊆ V a subspace, then
Proof. − : V → V /W by v 7→ v = W + v is an epi.
T : V → Z be linear
W = ker T
V = V /W
− : V → V /W by v 7→ v = W + v linear
∀x, y ∈ V we have
x = y ∈ V ⇐⇒ x ≡ y mod W ⇐⇒ x − y ∈ W ⇐⇒ T (x − y) = 0Z
so
T : V → Z is also injective
As T is linear, let α ∈ F, x, y ∈ V , then
as needed. Therefore,
T : V → Z by x 7→ T (x)
23
Duc Vu (Spring 2021) 6 Lec 6: Apr 9, 2021
T
V Z
−
T
V / ker T = V
V T Z
inclusion map
−
T
V / ker T = V im T
24
Duc Vu (Spring 2021) 6 Lec 6: Apr 9, 2021
Example 6.6
Let T : V → V be linear.
Remark 6.7. Let V be a finite dimensional vector space over F , T : V → V linear. Suppose
that
V = W1 ⊕ . . . ⊕ Wn
with each Wi T -invariant, i = 1, . . . , n and Bi an ordered basis for Wi , i = 1, . . . , n. Let
B = B1 ∪ . . . ∪ Bn be a basis of V ordered in the obvious way.
Then the matrix representation of T in the B basis is
h i
T W1 0
B1
[T ]B =
..
. h
i
0 T Wn
Bn
Example 6.8
Suppose that T : V → V is diagonalizable, i.e., there exists a basis B of eigenvectors
of T for V . Then, T : V → V , M
V = ET (λi )
each ET (λi ) is T -invariant.
T ET (λi )
= λi 1ET (λi )
25
Duc Vu (Spring 2021) 6 Lec 6: Apr 9, 2021
3. If P is a projection, so is 1V − P .
If T : V → V is linear, then
V = ker T ⊕ X with X ∼
= im T
Lemma 6.11
Let P : V → V be a projection. Then
V = ker P ⊕ im P
Moreover, if v ∈ im P , then
Pv = v
i.e.
P im P
: im P → im P is 1im P
In particular, if V is a finite dimensional vector space over F, B1 an ordered basis for
ker P , B2 an ordered basis for im P , then B = B1 ∪ B2 is an ordered basis for V and
0
..
.
[0]B1 0 0
[P ]B = =
0 [1im P ]B2 1
. .
.
1
P (v − P v) = P v − P 2 v = P v − P v = 0
Hence
v = (v − P v) + P v ∈ ker P + im P
ker P ∩ im P = 0 and P im P
= 1im P . Let v ∈ im P. By definition, P w = v for some
w ∈ V . Therefore,
Pv = PPw = Pw = v
Hence
P im P
= 1im P
If v ∈ ker P ∩ im P , then
v = Pv = 0
26
Duc Vu (Spring 2021) 7 Lec 7: Apr 12, 2021
§7.1 P r o j e c t i o n ( C o nt ’ d )
Lemma 7.1
Let V be a vector space over F , W, X ⊆ V subspaces. Suppose
V =W ⊕X
W = ker P (*)
X = im P
∃!w ∈ W, x ∈ X 3 v = w + x
Define
P : V → V by v 7→ x
To show P 2 = P , we suppose v ∈ V satisfies v = w + x, for unique w ∈ W , x ∈ X. Then check P
is linear
P v = P w + P x = P x = 1X x = x and well
defined
so
P 2v = P x = x = P v ∀v ∈ V
hence P 2 = P .
Uniqueness: Any P satisfying (*) takes a basis for W to 0 and fix a basis of X. Therefore,
P is unique by the UPVS.
Remark 7.2. Compare the above to the case that V is an inner product space over F , W ⊆ V
is a finite dimensional subspace and P : V → V by v →
7 vW , the orthogonal projection of P
onto W .
Proposition 7.3
Let V be a vector space over F , W, X ⊆ V subspaces s.t. V = W ⊕ X, P : V → V the
projection along W onto X, and T : V → V linear. Then the following are equivalent:
2. P T = T P .
27
Duc Vu (Spring 2021) 7 Lec 7: Apr 12, 2021
PTw = TPw = T0 = 0
T w ∈ ker P = W so W is T -invariant.
X is T -invariant, X = im P , P |X = 1X . So if x ∈ X
T x = T P x = P T x ∈ im P = X
So X is T -invariant.
1) =⇒ 2) Let v ∈ V . Then ∃! w ∈ W , x ∈ X s.t.
v =w+x
P T v = P T (w + x) = P T w + P T x
= 0 + Tx = TPx = TPw + TPx = TPv
V = W1 ⊕ W2
V = Wi ⊕ W1 ⊕ . . . ⊕ Ŵi ⊕ . . . ⊕ Wn
|{z}
omit
Construction: Let
V = W1 ⊕ . . . ⊕ Wn
as above. Define
PWi : V → V
to be the projection along W1 ⊕ . . . ⊕ Ŵi ⊕ . . . ⊕ Wn , i.e.
28
Duc Vu (Spring 2021) 7 Lec 7: Apr 12, 2021
e) 1V = PW1 + . . . + PWn .
T PWi = PWi T, i = 1, . . . , n
Hence
Remark 7.5. Compare this to the case that T is diagonalizable and the Wi are the eigenspaces.
Question 7.1. Let V be a real or complex finite dimensional inner product space, T :
V → V hermitian. What can you replace ⊕ by? What if V is a complex finite dimensional
inner product space and T : V → V is normal.
i) Pi − Pj = δij Pi , i = 1, . . . , n
ii) 1V = P1 + . . . + Pn
iii) Wi = im Pi , i = 1, . . . , n
Then
V = W1 ⊕ . . . ⊕ Wn
Pi = PWi i = 1, . . . , n
29
Duc Vu (Spring 2021) 7 Lec 7: Apr 12, 2021
Example 7.6
Let En = {e1 , . . . , en } be the standard basis for Rn and in the above e1 = v0 . . . Then
fe1 : Rn → R satisfies
If v = (α1 , . . . , αn ) in Rn
n
X
v= αi ei
i=1
so
n
!
X
fe1 (v) = fe1 αi ei
i=1
n
X n
X
= αi fe1 (ei ) = αi δii = α1
i=1 i=1
So fv0 : V → F is a surjective linear transformation. Since this is true for all v0 ∈ B, for
each v ∈ B, ∃! fv : V → F s.t.
(
1 if v = v 0
fv (v 0 ) = δv,v0 = ∀v 0 ∈ B
0 if v 6= v 0
Hence
!
X X
fv0 (x) = fv0 αv v = αv fv0 (v)
v∈B B
X
= αv δv,v0 = αv0
B
30
Duc Vu (Spring 2021) 7 Lec 7: Apr 12, 2021
Example 7.7
B = En standard basis for Rn
(
1 if ei = ej
fei (ej ) = δei ,ej = δi,j =
0 6 ej
if ei =
Example 7.8
inner product space over R, B = {v1 , . . . , vn } an
Let V be a finite dimensional P
orthonormal basis. Then if x = B αi vi , then
αi = hx, vi i
Take
X X
hx, vi i = h αj vj , vi i = αj hvj , vi i
X X
= αj δij kvi k2 = αj δij = αi
Definition 7.9 (Dual Space) — Let V be a vector space over F . A linear transformation
f : V → F is called a linear functional. Set
V ∗ := L(V, F ) := {f : V → F |f is linear}
31
Duc Vu (Spring 2021) 7 Lec 7: Apr 12, 2021
Proposition 7.10
Let V, W be a vector space over F . Then
L(V, W ) := {T : V → W |T linear}
is a vector space over F . Moreover, if V, W are finite dimensional vector spaces over F
dim V = dim V ∗
so
V ∼
=V∗
Proof. 115A.
Example 7.11
Let V be a vector space over F . Then the following are linear functionals
1. 0 : V → F
32
Duc Vu (Spring 2021) 8 Lec 8: Apr 14, 2021
§8.1 D u a l S p a c e s ( C o nt ’ d )
4. α < β ∈ R, then
Z β
I : C [α, β] → R by f 7→ f
α
eγ : C [α, β] → R by f 7→ f (γ)
Lemma 8.2
Let V be a vector space over F , B a basis for V ,
so
f v0 (v) = δv0 ,v ∀v ∈ B
the set of coordinate functions relative to B. Then B ∗ ⊆ V ∗ is linearly indep.
Proof. Suppose
X
0 = 0V ∗ = βvf v, βv ∈ F almost all 0
v∈B
33
Duc Vu (Spring 2021) 8 Lec 8: Apr 14, 2021
Corollary 8.3
Let V be a vector space over F with basis B. Then the linear transformation
DB : V → V ∗ induced by B → B ∗ by v 7→ f v
is a monomorphism.
In particular, if V is a finite dimensional vector space over F , then B ∗ is a basis for
V ∗ and
DB : V → V ∗ is an isomorphism
Remark 8.4. 1. If V = R∞
f := {(α1 , α2 , . . .) |αi ∈ R almost all 0}, then by HW1 # 4,
L : V → (V ∗ )∗ =: V ∗∗
Lemma 8.5
Let V be a vector space over F , v ∈ V . Then
Lv : V ∗ → F by f 7→ Lv (f ) := f (v)
Lv ∈ V ∗∗
Theorem 8.6
The “natural” map
L : V → V ∗∗ by v 7→ L(v) := Lv
is a monomorphism.
34
Duc Vu (Spring 2021) 8 Lec 8: Apr 14, 2021
So
L(αv + w) = αL(v) + L(w)
L is monic. Suppose v 6= 0. To show Lv = L(v) 6= 0. By example 2,
∃0 6= f ∈ V ∗ 3 f (v) 6= 0
So
Lv f = f (v) 6= 0
so Lv = L(v) 6= 0 and L is monic.
Corollary 8.7
If V is a finite dimensional vector space over F , then L : V → V ∗∗ is a natural
isomorphism.
1. v = w ⇐⇒ Lv = Lw
2. ∀f ∈ V ∗ f (v) = f (w) ⇐⇒ Lv f = Lw f
Lvi ∈ V ∗∗
B ∗∗ := {Lv1 , . . . , Lvn }
35
Duc Vu (Spring 2021) 8 Lec 8: Apr 14, 2021
Pn Pn
is the dual basis of B ∗ for V ∗∗ . So we have if x = i=1 αi vi ∈V, g = i=1 βi fi ∈ V ∗.
n
X n
X
x= αi vi = fi (x)vi
i=1 i=1
Xn Xn n
X
g= βi fi = Lvi (g)fi = g(vi )fi
i=1 i=1 i=1
i.e.
n
X
x= fi (x)vi ∀x ∈ V
i=1
Xn
g= g(vi )fi ∀g ∈ V ∗
i=1
S ◦ := {f ∈ V ∗ | f (x) = 0 ∀x ∈ S}
= {f ∈ V ∗ | f |S = 0} ⊆ V ∗
Remark 8.9. Many people write hv, f i for f (v) in the above even though f ∈
/ v.
36
Duc Vu (Spring 2021) 9 Lec 9: Apr 16, 2021
§9.1 D u a l S p a c e s ( C o nt ’ d )
Lemma 9.1
Let V be a vector space over F , ∅ =
6 S ⊆ V a subset. Then
1. S ◦ ⊆ V ∗ is a subspace.
0 = f (x) = Lx f, so Lx ∈ (S ◦ )◦ = S ◦◦
Theorem 9.2
Let V be a finite dimensional vector space over F , S ⊆ V a subspace. Then
37
Duc Vu (Spring 2021) 9 Lec 9: Apr 16, 2021
Corollary 9.3
Let V be a finite dimensional vector space over F , S ⊆ V a subspace. Then S = S ◦◦ .
§9.2 T h e Tr a n s p o s e
Construction: Fix T : V → W linear. For every S : W → X, we have a composition
S ◦ T : V → X is linear
So T :→ W linear induces a map
T ? : L(W, X) → L(V, X)
by
S 7→ S ◦ T
Proposition 9.5
Let V, W, X be vector spaces over F , T : V → W linear. Then
T ? : L(W, X) → L(V, X)
is linear.
Corollary 9.6
Let T : V → W be linear. Then
T ∗ : W ∗ → V ∗ by f 7→ f ◦ T
is linear.
38
Duc Vu (Spring 2021) 9 Lec 9: Apr 16, 2021
Lemma 9.8
Let T : V → W be linear. Then
Theorem 9.9
Let V, W be finite dimensional vector space over F , T : V → W linear. Then
Proof. Consider:
Suppose
B = {v1 , . . . , vn } , B ∗ = {f1 , . . . , fn }
fi (vj ) = δij ∀i, j
So
C = {w1 , . . . , wn } , C ∗ = {g1 , . . . , gn }
gi (wj ) = δij ∀i, j
39
Duc Vu (Spring 2021) 9 Lec 9: Apr 16, 2021
Let h i
A = [T ]B,C , B = T>
C ∗ ,B ∗
So
Bkj = Ajk ∀j, k
So we just proved. . .
Theorem 9.10
Let V, W be finite dimensional vector space over F , T : V → W linear, B, B ∗ ordered
dual bases for V, V ∗ and C , C ∗ ordered dual bases for W, W ∗ . Then
h i >
T> = [T ]B,C
C ∗ ,B ∗
Definition 9.11 (Row/Column Rank) — Let A ∈ F m×n . The row (column) rank of A
is the dimension of the span of the rows (columns) of A.
A : F n×1 → F m×1 by v 7→ A · v
A = [A]En,1 ,Em,1
where En,1 , Em,1 are the standard bases for F n×1 and F m×1 respectively.
Corollary 9.12
Let A ∈ F m×n . Then
row rank A = column rank A
and we call this common number the rank of A.
§9.3 Po l y n o m i a l s
40
Duc Vu (Spring 2021) 9 Lec 9: Apr 16, 2021
Lemma 9.14
If f |g and f |h in F [t], then f |gk + hl in F [t] for all k, l ∈ F [t].
Proof. By definition,
g = f g1 , h = f h1 , g1 , h1 ∈ F [t]
So
gk + hl = f g1 k + f h1 l = f (g1 k + h1 l)
in F [t].
We can define the degree of 0 ∈ F [t] to be the symbol −∞ or just do not define it at
all.
Remark 9.17. Let f, g ∈ F [t] \ {0}. Then
So
deg(f g) = deg f + deg g
41
Duc Vu (Spring 2021) 10 Lec 10: Apr 19, 2021
§10.1 Po l y n o m i a l s ( C o nt ’ d )
Division Algorithm: Let 0 6= f ∈ F [t], g ∈ F [t]. Then
∃! q, r ∈ F [t]
satisfying
g = fq + r with r = 0 or deg r < deg f
Definition 10.1 (Greatest Common Divisor) — Let f, g ∈ F [t] \ {0}. We say d in F [t]
is a gcd (greatest common divisor) of f, g if
i) d is monic.
Theorem 10.5
Let f, g ∈ F [t] \ {0}. Then a gcd of f, g exists and is unique write gcd(f,g) for the gcd
of f, g. Moreover, we have an equation
Proof. The existence and (?) follow from the Euclidean Algorithm. Let f, g ∈ F [t] \ {0}.
Then iteration of the Division Algorithm produces equations in F [t], if f + g ∈ F [t],
42
Duc Vu (Spring 2021) 10 Lec 10: Apr 19, 2021
rn = f k + gl ∈ F [t]
and if e|r1 → e|r2 → . . . → e|rn then (lead rn )−1 rn is the gcd of f and g in F [t] if a = lead f
f = gh ∈ F [t], then 0 6= g ∈ F or 0 6= h ∈ F
Example 10.7
If f ∈ F [t], deg f = 1, then f is irreducible.
Remark 10.8. If f, g ∈ F [t] \ F with f irreducible, then either f and g are relatively prime
or f |g since only a, af, 0 6= a ∈ F can divide f .
Proof. Suppose f × g where × means does not divide. Then f and g are relatively prime.
By the Euclidean Algorithm, there exists an equation
1 = f k + gl ∈ F [t]
Hence
h = f hk + ghl ∈ F [t]
As f |f hk and f |ghl in F [t], f |h by the lemma.
Remark 10.11. Euclid’s lemma is the key idea. The “correct” generalization of “prime” is the
conclusion of Euclid’s lemma. This generalization is profound as, in general, there is difference
between the two conditions “irreducible” and “prime”, although not for Z or F [t].
We know that any positive integer is a product of positive primes unique up to order n. If
we allow n < 0 such is unique up to ±1.
43
Duc Vu (Spring 2021) 10 Lec 10: Apr 19, 2021
g = ape11 . . . perr
unique up to order.
Remark 10.13. Applying the Euclidean Algorithm is relatively fast to compute, (for f |g
takes ≤ deg f steps to get a gcd). Factoring into the irreducible is not.
44
Duc Vu (Spring 2021) 11 Lec 11: Apr 21, 2021
§11.1 M i n i m a l Po l y n o m i a l s
We use the following theorem from 115A, Matrix Theory Theorem.
T n = T ◦| .{z
. . ◦} T, n≥0
n
Lemma 11.2
Let V be a finite dimensional vector space over F , f, g, h ∈ F [t], B an ordered basis
for V , T : V → V linear. Then
2. If f = gh ∈ F [t], then
f (T ) = g(T )h(T )
Pn
Proof. • By MTT, if g = i=0 ai t
i
∈ F [t], then
" n # n
X X
i
ai T i B
[g(T )]B = ai T =
i=0 B i=0
X
= ai [T ]iB = g ([T ]B )
• Left as exercise.
Lemma 11.3
Let V be a finite dimensional vector space over F , T : V → V linear. Then ∃q ∈
F [t]\{0} 3 q(T ) = 0 and if a = lead q, then q0 := a−1 q is moinc and satisfies q0 (T ) = 0
q ∈ ker eT := {f ∈ F [t]| f (T ) = 0}
45
Duc Vu (Spring 2021) 11 Lec 11: Apr 21, 2021
So
2
1V , T, T 2 , . . . , T n ∈ L(V, V )
are linearly dependent. So ∃a0 , . . . , an2 ∈ F not all 0 s.t.
n 2
X
ai T i = 0
i=0
Pn2 i
Then q = i=0 ai t works.
Theorem 11.4
Let V be a finite dimensional vector space over F , T : V → V linear. Then ∃!0 6=
qT ∈ F [t] monic called the minimal polynomial of T having the following properties:
1. qT (T ) = 0
Proof. By the lemma, ∃0 6= q ∈ F [t] monic s.t. q(T ) = 0. Among all such q, choose one
with deg q minimal.
Claim 11.1. q works.
Let g 6= 0 in F [t] satisfy g(T ) = 0. To show q|g ∈ F [t]. Write g = qh + r in F [t] with
r = 0 or deg r < deg q. Then
Corollary 11.5
Let V be a finite dimensional vector space over F , B an ordered basis for V1 and
T : V → V linear. Then
qT = q[T ]B
In particular, if A, B ∈ Mn F are similar write A ∼ B. Then
qA = qB
46
Duc Vu (Spring 2021) 11 Lec 11: Apr 21, 2021
Note:By the theorem, if V is a finite dimensional vector space over F g ∈ F [t] g = 6 0, and
deg g < deg qT , then q(T ) 6= 0.
Goal: Let V be a finite dimensional vector space over F , B an ordered basis of V , T : V → V
linear. Call
tI − [T ]B the characteristics matrix of T relative to B
Recall the characteristics polynomial fT of T is defined to be
qT |fT , hence fT (T ) = 0
Remark 11.7. 1. There exists a determinant proof of this – essentially Cramer’s rule.
2. A priori we only know deg qT ≤ n2 , where n = dim V .
3. fT is independent of B depends on properties of det : Mn F [t] → F [t]
= det tI − P AP −1
Proposition 11.8
Let V be a finite dimensional vector space over F , T : V → V linear. Then qT and fT
have the same roots in F , the eigenvalues of T .
qT = (t − λ)h ∈ F [t]
So
0 = qT (T ) = (T − λ1V )h(T )
As
0 ≤ deg h < deg qT , we have h(T ) 6= 0
Since h(T ) 6= 0∃0 6= v ∈ V s.t.
w = h(T )v 6= 0
Then
0 = qT (T )v = (T − λ1V ) h(T )v = (T − λ1V ) w
47
Duc Vu (Spring 2021) 11 Lec 11: Apr 21, 2021
eT : F [t] → L(V, V )
given by X X
ai ti 7→ ai T i
so eT is not only a linear transformation but a ring homomorphism, i.e., it also follows that
We know that
dimF F [t] = ∞
which has {1, t, . . . , tn , . . .} is a basis for F [t] and
by MTT. So
0 < ker eT := {f ∈ F [t]|eT f = f (T ) = 0}
is a vector space over F and a subspace of F [t]. This induces a linear transformation
eT : V / ker eT → im eT = F [T ]
48
Duc Vu (Spring 2021) 11 Lec 11: Apr 21, 2021
by
f · v := f (T )v
qT (T ) = 0
This turns V into what is called an F [t]-module, i.e., V via (*) satisfies the axioms of a
vector space over F but the scalars F [t] are now a ring rather than only a field.
49
Duc Vu (Spring 2021) 12 Lec 12: Apr 23, 2021
§12.1 Tr i a n g u l a r i z a b i l i ty
Proposition 12.1
Let V be a finite dimensional vector space over F , T : V → V linear, W ⊆ V a
T -invariant subspace. Then T induces a linear transformation
T : V /W → V /W by T (v) := T (v)
where v = W + v, V = V /W and
qT |qT ∈ F [t]
qT |qT ∈ F [t]
∗ ∗
[T ]B ∼ [T ]C
50
Duc Vu (Spring 2021) 12 Lec 12: Apr 23, 2021
and A11 , . . . , Ann are all the eigenvalues of T (not necessarily distinct) and hence roots of qT .
Example 12.5
If V is a finite dimensional vector space over F , T : V → V linear and T is triangular-
izable, then fTsplits in F [t].
0 1
Note: ∈ M2 R is not triangularizable as it has no eigenvalues.
−1 0
Theorem 12.6
Let V be a finite dimensional vector space over F , T : V → V linear. Then T is
triangularizable if and only if qT splits in F [t].
51
Duc Vu (Spring 2021) 12 Lec 12: Apr 23, 2021
hence
n−1
X n−1
X
0 = T vj − Aij v i = T vj − Aij vi
i=j i=j
1 ≤ j ≤ n − 1 in V = V /W . Therefore,
n−1
X
T vj − Aij vi ∈ ker− = W = F vn
i=j
by definition as W = ker− : V → V /W .
In particular, ∃Anj ∈ F , 1 ≤ j ≤ n − 1 satisfying
n−1
X
T vj − Aij vi = Anj vn
i=j
So
n
X
T vj = Aij vn 1≤j ≤n−1
i=j
T vn = λvn
A11 , . . . , Ann are the (not necessarily distinct) eigenvalues of T and hence roots of qT .
Let λi = Aii , i = 1, . . . , n. We have
n
X n
X
T vj = Aij vi = λj vj + Aij vi , 1≤j ≤n−1
i=1 i=j+1
T vn = λn vn
So
n
X
(T − λj 1V )vj = Aij vi ∈ Span (vj+1 , . . . , vn ) ∀1 ≤ j ≤ n − 1 (*)
i=j+1
Now
(T − λn 1V )vn = 0
52
Duc Vu (Spring 2021) 12 Lec 12: Apr 23, 2021
So
(T − λn 1V )vn−1 ∈ Span(vn ) by (∗)
This implies
(T − λn 1V )(T − λn−1 1V )vn−1 = 0
By induction, we may assume that
(T − λn 1V ) . . . (T − λj 1V )vj = 0
So by (*),
(T − λn 1V ) . . . (T − λj 1V )(T − λj−1 1V )vj−1 = 0
Therefore,
fT (T )vi = (T − λn 1V ) . . . (T − λi 1V )vi = 0
for i = 1, . . . , n. As B is a basis for V , fT (T ) = 0. Thus qT |fT ∈ F [t]. In particular, qT
splits in F [t].
Corollary 12.7
Let V be a finite dimensional vector space over F , T : V → V a triangularizable linear
operator. Then
qT |fT ∈ F [t]
In particular,
fT (T ) = 0
1. T is triangularizable.
2. qT |fT
3. fT (T ) = 0
53
Duc Vu (Spring 2021) 13 Lec 13: Apr 26, 2021
§13.1 Tr i a n g u l a r i z a b i l i ty ( C o nt ’ d )
Remark 13.1. Let V be a finite dimensional vector space over F , T : V → V linear, B an
ordered basis for V , A = [T ]B . So qA = qT and fA = fT .
viewing
A : F n×1 → F n×1 linear
Thus,
fA (A) = 0
Hence fT (T ) = 0 and qT = qA |fA = fT . So qT |fT in F [t]. Thus, if we knew such an F̃
exists in general, we would have proven the Cayley-Hamilton Theorem in general, i.e., if V
is a finite dimensional vector space over F and T : V → V linear, then
qT |fT ∈ F [t]
fT (T ) = 0
This is, in fact, true (and proven in Math 110C). Of course, assuming FTA, this proves
Cayley-Hamilton for all fields F ⊆ C.
f
q= = pe11 . . . pei i . . . perr
pei i
54
Duc Vu (Spring 2021) 13 Lec 13: Apr 26, 2021
1V = PW1 + PW2
PWi PWj = δij PWj
PWi T = T PWi , i = 1, 2
T = T PW1 + T PW2 = T |W1 + T |W2
By hw 4 # 6
qT = lcm (qT |W1 , qT |W2 )
This easily extends to more blocks.
Lemma 13.3
Let f ∈ F [t], T : V → V linear. Then ker f (T ) is T -invariant.
f (T ) T v = T f (T )v = 0
so this is immediate.
Lemma 13.4
Let g, h ∈ F [t] \ F be relatively prime. Set f = gh ∈ F [t]. Suppose T : V → V is
linear and f (T ) = 0. Then
subspaces of V and
V = ker g(T ) ⊕ ker h(T ) (+)
55
Duc Vu (Spring 2021) 13 Lec 13: Apr 26, 2021
Proof. By the lemma we just proved, we need only show (+). Since g, h are relatively
prime, there exists equation
1 = gk + hl ∈ F [t]
Hence
1V = g(T )k(T ) = h(T )l(T )
as linear operators on V i.e. ∀v ∈ V
Since f (T ) = 0 we have
Therefore,
g(T )k(T )v ∈ ker h(T )
and
0 = f (T )l(T )v = g(T )h(T )l(T )v
so
h(T )l(T )v ∈ ker g(T )
It follows by (*), ∀v ∈ V
where
V = ker g(T ) + ker h(T )
By (*), if v ∈ ker g(T ) ∩ ker h(T ), then
Hence
V = ker g(T ) ⊕ ker h(T )
as needed.
56
Duc Vu (Spring 2021) 14 Lec 14: Apr 28, 2021
§14.1 P r i m a r y D e c o m p o s i t i o n ( C o nt ’ d )
Proposition 14.1
Let V be a finite dimensional vector space over F , T : V → V linear, g, h ∈ F [t] \ F
monic and relatively prime. Suppose that
qT = gh ∈ F [t]
and
g = qT ker g(T )
and h = qT ker h(T )
Proof. By the last lemma in last lecture, we need only prove the last statement. By
definition, we have
g(T ) ker g(T ) = 0 and h(T ) ker h(T ) = 0
So by definition,
qT ker q(T )
|g and qT ker h(T )
|h ∈ F [t]
As g and h are relatively prime, by the FTA, so are
qT ker g(T )
and qT ker h(T )
Therefore, we have
f := lcm qT ,q
ker g(T ) T ker h(T )
= qT
ker q(T )qT
ker h(T )
Since
V = ker g(T ) ⊕ ker h(T )
f (T )v = 0 ∀v ∈ V
Hence
qT |f ∈ F [t]
By (+) and FTA
f |gh = qT
As both f and qT are monic,
f = qT
Applying FTA again, we conclude that
g = qT ker g(T )
and h = qT ker h(T )
We now generalize the proposition to an important result that decomposes a finite dimen-
sional vector space over F relative to a linear operator T : V → V .
57
Duc Vu (Spring 2021) 14 Lec 14: Apr 28, 2021
V = W1 ⊕ . . . ⊕ Wr (*)
i) Each Wi is T -invariant, i = 1, . . . , r
Moreover, any direct sum decomposition (*) of V satisfying i), ii), iii) is uniquely
determined by T and the p1 , . . . , pr up to order. If in addition, this is the case, then
Wi = ker pei i (T ) i = 1, . . . , r
Proof. We induct on r.
• r = 1 is immediate
• r > 1 By TFA, pe11 and g = pe22 . . . perr are relatively prime, so by the Proposition
V = W1 ⊕ V 1
where
Let
T1 = T |V1 : V1 → V1
By induction on r, we may assume all of the following:
V1 = W2 ⊕ . . . ⊕ Wr
Wi = ker pei i (T1 ) and is T1 -invariant
qT1 |W = pei i for i = 2, . . . , r
i
58
Duc Vu (Spring 2021) 14 Lec 14: Apr 28, 2021
Note:
r
X
ker pei i (T1 ) ∩ ker pj (T1 ) = 0 ∀i > 0
j=2
j6=i
i.e.,
v ∈ ker g(T ) = V1
So
T v = T |V1 v = T1 v
and
0 = pei i (T )v = pei i (T1 )v
as needed.
Let v ∈ ker pei i (T1 ), i > 1. By definition, v ∈ V1 , so
Wi = ker pei i (T ), i = 1, . . . , r
V1 = W2 ⊕ . . . ⊕ Wr
V = W1 ⊕ V 1
59
Duc Vu (Spring 2021) 14 Lec 14: Apr 28, 2021
It follows that
qT |V1 = pe22 . . . perr = g
Moreover, we have an equation
1 = pe11 k + gl ∈ F [t]
So
1V = pe11 (T )k(T ) + g(T )l(T ) (+)
pei i = qT |W , i = 2, . . . , r by (ii)
i
We have
pe22 (T ) . . . perr (T )v1 = 0
Hence by (+)
Therefore,
v = w1 + v 1 = w1 ∈ W 1
and it follows that ker pe11 (T ) ⊆ W1 as needed.
60
Duc Vu (Spring 2021) 14 Lec 14: Apr 28, 2021
61
Duc Vu (Spring 2021) 15 Lec 15: Apr 30, 2021
§15.1 P r i m a r y D e c o m p o s i t i o n ( C o nt ’ d )
Recall: Let V be a finite dimensional vector space over F , T : V → V linear is called
diagonalizable if there exists an ordered basis B for V consisting of eigenvectors of T . By
hw 2 # 2, this is equivalent to M
V = ET (λ)
λ
Theorem 15.1
Let V be a finite dimensional vector space over F , T : V → V linear. Then T is
diagonalizable iff qT splits in F [t] and has no repeated roots in F . If this is the case,
then
Yr
qT = (t − λi ), λ1 , . . . , λr the distinct roots of qT
i=1
V = V1 ⊕ . . . ⊕ Vr
SO T is diagonalizable.
“ =⇒ ” Let B = {v1 , . . . , vn } be an ordered basis for V consisting of eigenvectors of T
with λi the eigenvalue of vi and ordered s.t.
(T − λi 1V ) vj = T vj − λi vj = (λj − λi ) vj , j = 1, . . . , n
So
r
Y
(T − λi 1V ) vj = 0 for j = 1, . . . , n
i=1
i.e.,
r
Y
(T − λi 1V ) vanishes on a basis for V
i=1
hence vanishes on all of V . It follows that
r
Y
qT | (t − λi ) ∈ F [t]
i=1
In particular, qT splits in F [t] and has no multiple roots in F by FTA. As every eigenvalue
of T is a root of fT , we have
t − λi |qT , i = 1, . . . , r
62
Duc Vu (Spring 2021) 15 Lec 15: Apr 30, 2021
§15.2 J o r d a n B l o ck s
Note: fJn (λ) = det (tI − Jn (λ)) = (t − λ)n ∈ F [t], so splits with just one root of multiplic-
ity.
Example 15.4
J = Jn (0) is nilpotent and has qJ = tm for some m. In fact, qJ = tn – why?
In fact, let A ∈ Mn F , A : F n×1 → F n×1 linear with A ∼ N with
A = P N P −1 , some P ∈ GLn F,
we have
n
An = P N P −1 = P N P −1 P N P −1 . . . P N P −1 = P N n P −1 = 0
N ei = ei+1 , i≤n−1
N en = 0
N 2 ei = N − N ei = ei+2 , i≤n−2
..
.
63
Duc Vu (Spring 2021) 15 Lec 15: Apr 30, 2021
Lemma 15.6
Let J = Jn (λ) ∈ Mn F . Then
2. dim EJ (λ) = 1
3. tJ = qJ = (t − λ)n
4. fJ (J) = 0
Proof. Let
N = J − λI ∈ Mn F
the characteristics matrix of J
0 ... 0
.. ..
N n−1 = . . ∈ Mn F
0 0
1 0 ... 0
dim ker N = 1
and
ker N = ET (λ)
This gives 2) as fT = (t − λ)n , 1) is clear.
Remark 15.7. Jn (λ) has only a line as an eigenspace, so among triangulariazable operator
away from being diagonalizable when n ≥ 1.
64
Duc Vu (Spring 2021) 15 Lec 15: Apr 30, 2021
Proposition 15.8
Let A ∈ Mn F be triangularizable. Suppose fA = (t − λ)n for some λ ∈ F . Then A is
diagonalizable iff qA = (t − λ) iff A = λI.
Proof. If qA = t − λ, then A = λI as
[T ]B = Jn (λ)
Then by definition
So
Eλ (λ) = F vn
v1 , . . . , vn−1 are not eigenvectors, but do satisfy
65
Duc Vu (Spring 2021) 16 Lec 16: May 3, 2021
§16 L e c 1 6 : M ay 3 , 2 0 2 1
§16.1 J o r d a n B l o ck s ( C o nt ’ d )
We let
Lemma 16.2
Let V be a vector space over F , T : V → V linear, 0 6= vn ∈ ET (λ), v1 , . . . , vn an
ordered sequence of generalized eigenvectors of T of length n, gn (λ) = {v1 , . . . , vn }.
Then
So
(T − λ1V )k vi = 0 ∀k > n − i
Suppose
α1 v1 + . . . + αn vn = 0, αi ∈ F not all 0
Choose the least k s.t. αk 6= 0. Then
66
Duc Vu (Spring 2021) 16 Lec 16: May 3, 2021
As vn 6= 0, αk = 0, a contradiction.
So 1) follows and 1) → 2).
Definition 16.4 (Jordan Basis) — Let V be a finite dimensional vector space over F ,
T : V → V linear. An ordered basis B for V is called a Jordan basis (if it exists) for
V relative to T if B is the union
Proposition 16.5
Let V be a finite dimensional vector space over F , T : V → V linear. Then V has a
Jordan basis relative to T ⇐⇒ T has a matrix representation in Jordan canonical
form (JCF).
Proof. Let wi = gri (vi,ri , λi ) in (?). The only thing to show is: Wi is T -invariant, but this
follows from our computation.
with
ri1 ≤ ri2 ≤ . . . ≤ ri ni , 1≤i≤m
67
Duc Vu (Spring 2021) 16 Lec 16: May 3, 2021
e.g.
1 0
0 1
J1 (1)
1 0
J1 (1)
[T ]B = 1 1 =
J2 (1)
0 2 0 0
J3 (2)
1 2 0
0 1 2
Let
Wij = Span gri ,j (vij , λi ) ∀i, j
These are all T -invariant. We have
Y
fT = (t − λi )rij
i,j
and
Y
qT = lcm ((t − λi )rij |j = 1, . . . , ni )
i
Y
= (t − λi )rini
i
So
qT |fT and fT (T ) = 0
Also
qT |Wij = fT |Wij = (t − λi )rij
for all 1 ≤ j ≤ nj , 1 ≤ i ≤ m. There are called the elementary divisors of T
W11 ⊕ . . . ⊕ W
d ij ⊕ . . . ⊕ Wm,nm
|{z}
omit
Then
(
Pjl if i = k and j = l
Pij Pkl = δik δjl Pjl =
0 otherwise
1V = P11 + . . . + Pmnm
T Pij = Pij T
T = T P11 + . . . + T Pmnm = T W11
+ ... + T Wmnm
Abusing notation
λ1 , . . . , λm are the distinct eigenvalues of T
Let
Wi = Wi1 ⊕ . . . ⊕ Wini i = 1, . . . , m
68
Duc Vu (Spring 2021) 16 Lec 16: May 3, 2021
As ri1 ≤ . . . ≤ rini ,
(T − λi 1V )rini Wij
= 0, 1 ≤ j ≤ ni
(T − λi 1V )ri ni −1 Wij
6= 0
showing
qT |Wi = (t − λi )rini
So
V = W1 ⊕ . . . ⊕ Wm
is the unique primary decomposition of V relative to T .
Note: The Jordan canonical form of T above is completely determined by the elementary
divisors of T .
§16.2 J o r d a n C a n o n i c a l Fo r m
Theorem 16.6
Let V be a finite dimensional vector space over F , T : V → V linear. Suppose that
qT splits in F [t]. Then there exists a Jordan basis B for V relative to T . Moreover,
[T ]B is unique up to the order of the Jordan blocks. In addition, all such matrix
representations are similar.
qT = (t − λ)r
Suppose that
qT = (t − λ1 )r1 . . . (t − λm )rm ∈ F [t]
λ1 , . . . , λm distinct. Set
Wi = ker (T − λi 1V )ri , i = 1, . . . , m
V = W1 ⊕ . . . ⊕ Wm
Wi is T -invariant, i = 1, . . . , n
qT |W = (t − λi )ri , i = 1, . . . , m
i
69
Duc Vu (Spring 2021) 17 Lec 17: May 5, 2021
§17 L e c 1 7 : M ay 5 , 2 0 2 1
§17.1 J o r d a n C a n o n i c a l Fo r m ( C o nt ’ d )
Proof. (Cont’d from Lec 16) Reduction 2: We may assume that qT = tr , i.e., λ = 0.
Suppose that we have proven the case for λ = 0. Let S = T − λ1V , T as in Reduction 1.
Then
S r = (T − λ1V )r = 0 and S r−1 = (T − λ1V )r−1 6= 0
Therefore,
qS = t r
if B is a Jordan basis for V relative to S, then
[S]B = [T ]B − λI
[T ]B = [S]B + λI
is a JCF with diagonal entries λ and B is also a Jordan basis for V relative to T . Reduction 2
now follows easily. We turn to
Existence: We have reduced to the case
qT = t r , i.e., T r = 0, T r−1 6= 0
• dim V = 1 is immediate.
im T < V
T im T
: im T → im T linear
As T r = 0, certainly (T |im T )r = 0, so
T im T
is also nilpotent
and
qT im T
qT ∈ F [t]
since
qT T im T
= 0 = qT (T )
So qT |im T splits in F [t] and
q = ts , for some s ≤ r
T
im T
70
Duc Vu (Spring 2021) 17 Lec 17: May 5, 2021
Thus we have
T ri wi = 0, i = 1, . . . , m
ri
q =t , i = 1, . . . , m
T
Wi
Since wi ∈ Wi ⊆ im T ,
∃vi ∈ V 3 T vi = wi , i = 1, . . . , m
So we also have
T ri +1 vi = T ri T vi = T ri wi = 0
and
T ri vi = T ri −1 T vi = T ri −1 wi 6= 0
Therefore, vi , T vi , . . . , T ri vi is an ordered sequence of generalized eigenvalues for T
in V , and, in particular, linearly independent. For each i = 1, . . . , m, let
Vi = Span {vi , T vi , . . . , T ri vi }
So
Xri
Vi = αj T j vi |αj ∈ F
j=0
In particular,
B0 = {v1 , T v1 , . . . , T ri v1 , . . . , vm , T vm , . . . , T rm vm }
is a basis for V 0 .
71
Duc Vu (Spring 2021) 18 Lec 18: May 7, 2021
§18 L e c 1 8 : M ay 7 , 2 0 2 1
§18.1 J o r d a n C a n o n i c a l Fo r m ( C o nt ’ d )
Proof. (Cont’d) Suppose ui ∈ Vi , i = 1, . . . , m satisfies
u1 + . . . + um = 0 (1)
ui = fi (T )vi
wi = T vi i = 1, . . . , m
W1 + . . . + Wm = W1 ⊕ . . . ⊕ Wm (*)
We have
fi (T )wi = 0, i = 1, . . . , m
Hence
fi (T ) = 0 on Wi , i = 1, . . . , m
Thus
tri = q fi ∈ F [t], i = 1, . . . , m
T
Wi
fi = tgi ∈ F [t], i = 1, . . . , m
deg gi < deg fi , i = 1, . . . , m if fi 6= 0
Since
fi (T ) = T gi (T ) = gi (T )T
and
wi = T vi , i = 1, . . . , m
(2) now becomes
g1 (T )w1 + . . . + gm (T )wm = 0 (3)
Since each Wi is T -invariant, by (*)
gi (T )wi = 0, hence gi (T ) = 0 on Wi
72
Duc Vu (Spring 2021) 18 Lec 18: May 7, 2021
tri = q gi ∈ F [t]
T
Wi
gi = tri hi ∈ F [t], i = 1, . . . , m
So
fi = tri +1 hi ∈ F [t], i = 1, . . . , m
Thus we have
ui = fi (T )vi = hi (T )T ri +1 vi = 0, i = 1, . . . , m
This establishes claim 1. As
wi = T vi ∈ Wi , i = 1, . . . , m
We have
T V 0 = T V1 ⊕ . . . ⊕ T Vm
= W1 ⊕ . . . ⊕ Wm = T V (?)
T Vi = Wi , i = 1, . . . , m
Therefore,
T V0
=T V1
+ ... + T Vm
73
Duc Vu (Spring 2021) 18 Lec 18: May 7, 2021
let B2 = B1 . In either case, B2 is a linearly independent set. Continuing in this way, since
B0 ∪ C spans V , we get a spanning set of V
B = B0 ∪ {uj1 , . . . , ujr } ⊆ V
with
Tuji = 0
for some uji constructed above, 1 ≤ i ≤ s.
Using claim 1, we have
qT = tr , T r = 0, T r−1 6= 0
the corresponding Jordan canonical form. Prior computation showed for each i, 1 ≤ i ≤ m,
(
rank Jrji (0) = ri − j
if j < ri
dim ker Jrji (0) = j
and (
rank Jrji (0) = 0
if j ≥ ri
dim ker Jrji (0) = ri
Clearly, for each i, j
Jr1 (0)
[T ]jB =
..
.
Jrjm (0)
as [T ]B is in block form. So by (*),
m
mj = rank [T ]jB =
X
rank Jrji (0)
i=1
74
Duc Vu (Spring 2021) 18 Lec 18: May 7, 2021
Corollary 18.1
Let A ∈ Mn F . If qA ∈ F [t] splits in F [t], then A is similar to a matrix in JCF unique
up to the order of the Jordan blocks.
Corollary 18.2
Let F be an algebraically closed field, e.g., F = C. Then every A ∈ Mn F is similar
to a matrix in JCF unique up to the order of the Jordan blocks and for every V , a
finite dimensional vector space over F , and T : V → V linear, V has a Jordan basis
relative to T . Moreover, the Jordan blocks of [T ]B are completely determined by the
elementary divisors (minimal polys) that correspond to the Jordan blocks.
Theorem 18.3
Let F be an algebraically closed field, e.g., F = C, A, B ∈ Mn F . Then, the following
are equivalent
1. A ∼ B
75
Duc Vu (Spring 2021) 18 Lec 18: May 7, 2021
Corollary 18.4
Let F be an algebraically closed field. Then A ∼ A> .
Example 18.6
C (tn ) = Jn (0).
f = g ⇐⇒ C(f ) = C(g)
Lemma 18.7
Let g ∈ F [t] \ F be moinc. Then
fC(g) = g
• n = 1 is immediate
76
Duc Vu (Spring 2021) 18 Lec 18: May 7, 2021
Lemma 18.8
Let g ∈ F [t] \ F be monic. Then
qC(g) = fC(g) = g
In particular,
fC(g) (C(g)) = 0
77
Duc Vu (Spring 2021) 19 Lec 19: May 10, 2021
§19 L e c 1 9 : M ay 1 0 , 2 0 2 1
§19.1 C o m p a n i o n M a t r i x ( C o nt ’ d )
Remark 19.1. If C is a companion matrix in Mn F , viewing
then
B = e1 , Ce1 , . . . , C n−1 e1
V = Span {v, T v, . . . , T n v, . . .} = F [T ]v
Proposition 19.3
Let V be a finite dimensional vector space over F , n = dim V , T : V → V linear.
Suppose there exists a T -cyclic vector v for V , i.e., V = F [T ]v. Then all of the
following are true
ii) [T ]B = C(fT )
iii) fT = qT
T j v ∈ Span v, T v, . . . , T j−1 v
say
T j v = αj−1 T j−1 v + αj−2 T j−2 v + . . . + α1 T v + α0 v (*)
for α0 , . . . , αj−1 ∈ F . Take T of (*), to get
78
Duc Vu (Spring 2021) 19 Lec 19: May 10, 2021
T N v ∈ Span v, T v, . . . , T j−1 v
∀N ≥ j
It follows that
v = F [T ]v = Span v, T v, . . . , T j−1 v
So
n = dim V ≤ j, hence n = j
This proves i).
ii) The computation proving i) shows
B = v, T v, . . . , T n−1 v
Example 19.4
Let V be a finite dimensional vector space over F , dim V = n, T : V → V linear s.t.
there exists an ordered basis B with
[T ]B = Jn (λ)
B = v, Sv, . . . , S n−1 v
where X
det D = sgn σD1σ(1) . . . Dnσ(n)
79
Duc Vu (Spring 2021) 19 Lec 19: May 10, 2021
§19.2 S m i t h N o r m a l Fo r m
We say that A ∈ F [t]m×n is in Smith Normal Form (SNF) if A is the zero matrix or if A is
the matrix of the form
q1 0 . . .
0 q2
.. . .
.
.
q r
0
..
.
0
with q1 |q2 |q3 | . . . |qr in F [t] and all monic, i.e., there exists a positive integer r satisfying
r ≤ min(m, n) and q1 |q2 |q3 | . . . |qr monic in F [t] s.t. Aii = qi for 1 ≤ i ≤ r and Aij = 0
otherwise.
We generalize Gaussian elimination, i.e., row (and column) reduction for matrices with
entries in F to matrices with entries in F [t]. The only difference arises because most
elements of F [t] do not have multiplicative inverses.
Let A ∈ Mn (F [t]). We say that A is an elementary matrix of
i) Type I: if there exists λ ∈ F [t] and l 6= k s.t.
1 if i = j
Aij = λ if (i, j) = (k, l)
0 otherwise
Remark 19.5. Let A ∈ F [t]m×n . Multiplying A on the left (respectively right) by a suitable
size elementary matrix of
a) Type I is equivalent to adding a multiple of a row (respectively column) of A to another
row (respectively column) of A.
b) Type II is equivalent to interchanging two rows (respectively columns) of A.
c) Type III is equivalent to multiplying a row (respectively column) of A by an element in
F [t] having a multiplicative inverse.
80
Duc Vu (Spring 2021) 19 Lec 19: May 10, 2021
Let
GLn (F [t]) := {A|A is invertible}
Warning: A matrix in Mn (F [t]) having det(A) 6= 0 may no longer be invertible, i.e.,
have an inverse. What is true is that GLn (F [t]) = {A|0 6= det(A) ∈ F }, equivalently
GLn (F [t]) consists of those matrices whose determinant have a multiplicative inverse in
F [t].
Theorem 19.8
Let A ∈ F [t]m×n . Then A is equivalent to a matrix in Smith Normal Form. Moreover,
there exist matrices P ∈ GLm (F [t]) and Q ∈ GLn (F [t]), each a product of matrices
of Type I, Type II, Type III, s.t. P AQ is in SNF.
Remark 19.9. The SNF derived by this algorithm is, in fact, unique. In particular, the monic
polynomials q1 |q2 |q3 | . . . |qr arising in the SNF of a matrix A are unique and are called the
invariant factor of A. This is proven using results about determinant.
81
Duc Vu (Spring 2021) 20 Lec 20: May 12, 2021
§20 L e c 2 0 : M ay 1 2 , 2 0 2 1
§20.1 R a t i o n a l C a n o n i c a l Fo r m
If A, B ∈ F [t]m×n then A ≈ B if and only if they have the same SNF if and only if they
have the same invariant factors. So what good is the NSF relative to linear operators on
finite dimensional vector spaces?
Let A, B ∈ Mn (F ). Then A ∼ B if and only if tI − A ≈ tI − B in Mn (F [t]) and this is
completely determined by the SNF hence the invariant factors of tI − A and tI − B. Now
the SNF of tI − A may have some of its invariant factors 1, and we shall drop these.
Let V be a finite dimensional vector space over F with B an ordered basis. Let T : V → V
be a linear operator. If one computes the SNF of tI − [T ]B , it will have the form
1 0 ... ... 0
0 1 0
.. . . ..
.
. .
q 1
q2
.. .. .
. . ..
0 ... ... qr
with q1 |q1 | . . . |qr are all the monic polynomials in F [t] \ F . These are called the invariant
factors of T . They are uniquely determined by T . The main theorem is that there exists
an ordered basis B for V s.t.
C(q1 ) 0 ... 0
0 C(q2 ) . . . 0
[T ]B = .
. . . ..
. . .
0 . . . C(qr )
and this matrix representation is unique. This is called the rational canonical form or RCF
of T . Moreover, the minimal polynomial qt of T is qr . The algorithm computes this as
well as all invariant factors of T . The characteristic polynomial fT of T is the product
of q1 . . . qr . This works over any field F , even if qT does not split. The basis B gives a
subspaces V = W1 ⊕ . . . ⊕ Wr where fT |Wi = qT |Wi = qi
decomposition of V into T -invariant
and if dim(Wi ) = ni then Bi = vi , T vi , . . . , T ni −1 vi is a basis for Wi .
Let V be a finite dimensional vector space over F with B an ordered basis. Let T : V → V
be a linear operator. Suppose that qT splits over F . Then we know that there exists a
Jordan canonical form of T .
Question 20.1. How do we compute it?
We use the Smith Normal Form of tI −[T ]B to compute the invariant factors q1 |q1 | . . . |qr
of T just as one does to compute the RCF of T . We then factor each qi . Suppose this
factorization is
qi = (t − λ1 )r1 . . . (t − λm )rm
in F [t] with λ1 , . . . , λm distinct. Note that qi+1 has this as a factor so it has the form
82
Duc Vu (Spring 2021) 20 Lec 20: May 12, 2021
nv := dim F [T ]v ≤ dim V
As F [T ]v is T -invariant,
T |F [T ]v Bv = C fT |F [T ]v
and
qT |F [T ]v = fT |F [T ]v
We want to show that V can be decomposed as a direct sum of T -cyclic subspaces of V .
The SNF of the characteristic matrix
tI − [T ]C
i) V = W1 ⊕ . . . ⊕ Wr , ni = dim Wi , i = 1, . . . , r
ii) Wi is T -invariant, i = 1, . . . , r
v) qT = qr
83
Duc Vu (Spring 2021) 20 Lec 20: May 12, 2021
A ∈ GLn F [t] ⇐⇒ A = I
since
q1 0
..
.
qr
0
..
0 .
means 0 . . . 0 · q1 . . . qr ∈ F \ {0} if there are any 0’s on the diagonal, which is inseparable.
Lemma 20.2
Let g ∈ F [t] \ F be monic of degree n. Then
1 0
..
It − C(q) ≈
.
1
0 q
Corollary 20.3
Let V be a finite dimensional vector space over F , T : V → V linear q1 | . . . |qr the
invariants of T in F [t]. Then
C(q1 ) 0
tI −
..
.
0 C(qr )
Certainly, if there exists an ordered basis B for V a finite dimensional vector space over F ,
T : V → V linear s.t. [T ]B is in RCF, then everything in goal falls out. So by the above,
the goal will follow if we prove the following
84
Duc Vu (Spring 2021) 20 Lec 20: May 12, 2021
Theorem 20.4
Let A0 , B0 ∈ Mn F , A = tI − A0 and B = tI − B0 in Mn F [t], the corresponding
characteristic matrices. Then the following are equivalent
Lemma 20.5
Let A ≈ B in Mn F [t]. Then ∃P, Q ∈ GLm F [t] each products of elementary matrices
s.t. A = P BQ.
Proof. P ∈ GLn F [t] iff its SNF = I which we get using elementary matrices.
For the second lemma, we need the “division algorithm” by “linear polys” in Mn F [t]. If we
were in F [t], we know if f, g ∈ F [t], f 6= 0,
AQ + R 6= QA + R
but the same argument to get (*) for polys, will give a right and left remainder.
Notation: Let Ai ∈ Mr F , i = 0, . . . , n and let
An tn + An−1 tn−1 + . . . + A0
denote
An (tn I) + . . . + A0 I ∈ Mn F [t]
So if
A = (αij )
then
Atn = (αij tn )
i.e., two matrix polynomials are the same iff all their corresponding entries are equal, i.e.,
85
Duc Vu (Spring 2021) 20 Lec 20: May 12, 2021
Lemma 20.6
Let A0 ∈ Mn F , A = tI − A0 ∈ Mn F [t] and
0 6= P = P (t) ∈ Mn F [t]
i) P = AM + R
ii) P = N A + S
86
Duc Vu (Spring 2021) 21 Lec 21: May 14, 2021
§21 L e c 2 1 : M ay 1 4 , 2 0 2 1
§21.1 R a t i o n a l C a n o n i c a l Fo r m ( C o nt ’ d )
Recall from last lecture,
Lemma 21.1
Let A0 ∈ Mn F , A = tI − A0 ∈ Mn F [t] and
0 6= P = P (t) ∈ Mn F [t]
i) P = AM + R
ii) P = N A + S
Proof. i) Let
m = max deg Plk , Plk 6= 0
l,k
Every entry in
has deg = m or is zero and the tm -coefficient of (APm−1 )ij is αij . Thus, P − APm−1
has polynomial entries of degree at most m − 1 (or = 0). Apply the same argument
to P − APm−1 (replacing m by m − 1 in (*)) to produce a matrix Pm−2 in Mn F [t]
s.t. all the polynomial entries in (P − APm−1 ) − APm−2 have degree at most m − 2
(or = 0). Continuing this way, we construct matrices Pm−3 , . . . , P0 satisfying if
M := Pm−1 + Pm−2 + . . . + P0
then
R := P − AM
has only constant entries, i.e., R ∈ Mn F . So
P = AM + R
as needed.
87
Duc Vu (Spring 2021) 21 Lec 21: May 14, 2021
Theorem 21.2
Let A0 , B0 ∈ Mn F , A = tI − A0 , B = tI − B0 in Mn F [t]. Then
A ≈ B ∈ Mn F [t] ⇐⇒ A0 ∼ B0 ∈ Mn F
Proof. “ ⇐= ” If
B0 = P A0 P −1 , P ∈ GLn F,
then
P (tI − A0 ) P −1 = P tP −1 − P A0 P −1 = tI − B0 = B
So B = P AP −1 and B ≈ A.
“ =⇒ ” Suppose there exist P1 , Q1 ∈ GLn F [t], hence each a product of elementary matrices
by Lemma 20.5, satisfying
B = tB − B0 = P1 AQ1 = P1 (tI − A0 ) Q1
i) P1 = BP2 + R, P2 ∈ Mn F [t], R ∈ Mn F
ii) Q1 = Q2 B + S, Q2 ∈ Mn F [t], S ∈ Mn F
iii) P1 A = BQ−1
Thus, we have
i)
B = P1 AQ1 = (BP2 + R)AQ1 = BP2 AQ1 + RAQ1
iv) ii)
= BP2 P1−1 B + RAQ1 = BP2 P1−1 B + RA(Q2 B + S)
= BP2 P1−1 B + RAQ2 B + RAS
i.e., we have
By i)
R = P1 − BP2
Plugging this into RAQ2 B, yields
i)
RAQ2 B = (P1 − BP2 )AQ2 B = P1 AQ2 B − BP2 AQ2 B
iii)
= BQ−1
−1
1 Q2 B − BP2 AQ2 B = B Q1 Q2 − P2 AQ2 B
i.e.
88
Duc Vu (Spring 2021) 21 Lec 21: May 14, 2021
Let
T = P2 P1−1 + Q−1
1 Q2 − P2 AQ2
Then
viii) Every entry of B = tI − B0 is zero or has deg ≤ 1 and every entry of RAS =
R(tI − A0 )S has is zero or has deg ≤ 1.
Comparing coefficients of the matrix of polys BT B = B − RAS using vii), viii) shows
Tm = 0
Hence
T =0
So vii) becomes
I = RS
B0 = RA0 S
89
Duc Vu (Spring 2021) 21 Lec 21: May 14, 2021
Theorem 21.3
Let A0 , B0 ∈ Mn F , A = tI − A0 , B = tI − B0 in Mn F [t]. Then the following are
equivalent
i) A0 ∼ B0
ii) A ≈ B
V = ker q1 (T ) ⊕ . . . ⊕ ker qn (T )
qr = qT
fT = q1 . . . qr
Qr
Note: If qi = j=1 (t − λi )ej is an invariant factor, then
Je1 (λ1 ) 0
C(qi ) ∼
..
.
0 Jer (λr )
Corollary 21.4
Let A, B ∈ Mn F , F ⊆ K a subfield. Then A ∼ B in Mn F iff A ∼ B in Mn K.
90
Duc Vu (Spring 2021) 22 Lec 22: May 17, 2021
§22 L e c 2 2 : M ay 1 7 , 2 0 2 1
h, i = h, iV : V × V → F
satisfies
2. Let v1 , v2 ∈ V \ {0},
hv2 , v1 i
w= v1
kv1 k2
is called the orthogonal projection of v2 on v1 and v = v2 − w is orthogonal to w, i.e.
hv, wi = 0, write v ⊥ w.
Example 22.3
If F ⊆ R, then any sesquilinear map is linear and V † = V ∗ .
91
Duc Vu (Spring 2021) 22 Lec 22: May 17, 2021
hv,wi i
Then, the wi -coordinate of vw is kwi k2
∈ F . Hence
fi = p wi :V →F
kwi k2
92
Duc Vu (Spring 2021) 22 Lec 22: May 17, 2021
S ⊥ := {x ∈ V | x ⊥ s ∀s ∈ S} ⊆ V
a subspace.
Note: The sesquilinear map
p : V → V ∗ by v 7→ pv
induces an injective sesquilinear map
p S⊥
: S⊥ → S◦
and we have
S ⊆ S ⊥⊥ := (S ⊥ )⊥
If S is a subspace, S ∩ S ⊥ = 0 so
S + S⊥ = S ⊕ S⊥
write
S + S⊥ = S ⊥ S⊥
called an orthogonal direct sum and if V is finite dimensional then
S = S ⊥⊥
e.g., if v ∈ V , then
ker pv = (F v)⊥
so
V = F v ⊥ (F v)⊥
More generally, we have the following crucial result.
V = S ⊥ S⊥
i.e., if v ∈ V
∃!s ∈ S, s⊥ ∈ S ⊥ 3 v = s + s⊥
In particular, s = vS . If V is finite dimensional, then
93
Duc Vu (Spring 2021) 22 Lec 22: May 17, 2021
V = W1 ⊕ . . . ⊕ Wn
with
wi ⊥ wj ∀wi ∈ Wi , wj ∈ Wj , i 6= j
We call V an orthogonal direct sum or orthogonal decomposition of V .
V = Wi ⊥ Wi⊥
and
Wi⊥ = W1 ⊥ . . . Ŵi ⊥ . . . ⊥ Wn
|{z}
omit
Wi⊥ = W1 ⊥ . . . ⊥ Ŵi ⊥ . . . ⊥ Wn
ker Pi = Wi⊥
im Pi = Wi
Pi Pj = δij Pj ∀i, j
1V = P1 + . . . + Pn
Pi (v) = vWi
So
v = vW1 + . . . + vWn
is a unique decomposition of v relative to (*).
Theorem 22.9
Let V, W be finite dimensional inner product space over F , T : V → W linear. Then
the adjoint T ∗ : W → V exists.
94
Duc Vu (Spring 2021) 23 Lec 23: May 19, 2021
§23 L e c 2 3 : M ay 1 9 , 2 0 2 1
§23.1 I n n e r P r o d u c t S p a c e s ( C o nt ’ d )
Corollary 23.1
Let V, W be finite dimensional inner product space over F , T : V → W linear. Then
T = T ∗∗ := (T ∗ )∗
and
hT ∗ w, viV = hw, T viW ∀w ∈ W ∀v ∈ V
Proof. We have
hT v, T v 0 iW = hv, v 0 iV ∀v, v 0 ∈ V
Remark 23.3. Let T : V → W linear of inner product space over F . If T preserves inner
products, then T is monic.
T v = 0 ⇐⇒ kT vk = 0 ⇐⇒ hT v, T vi = 0 ⇐⇒ hv, vi = 0
95
Duc Vu (Spring 2021) 23 Lec 23: May 19, 2021
Theorem 23.4
Let V, W be finite dimensional inner product space over F with dim V = dim W and
T : V → W linear. Then the following are equivalent
2. T is an isometry.
kT vi k = kvi k i = 1, . . . , n
= hT x, T yi
Corollary 23.5
Let V, W be finite dimensional inner product space over F both having orthonormal
basis. Then V is isometric to W if and only if dim V = dim W .
Theorem 23.6
Let V, W be inner product space over F , T : V → W linear. Then T preserves inner
products iff T preserves lengths, i.e., kT vkW = kvkV for all v ∈ V .
96
Duc Vu (Spring 2021) 23 Lec 23: May 19, 2021
“ ⇐= ” Let x, y ∈ V and
√
hx, yiV = α + β −1
√
hT x, T yiW = γ + δ −1
2α = 2γ =⇒ α = γ
√
So we are done if F ⊆ R. Suppose F * R, then there exists 0 6= µ ∈ R s.t. µ −1 ∈ F .
Then
√ √ √
hx, −1µyiV = − −1µhx, yiV = −µ −1α + βµ
√ √ √
hT x, −1µT yiW = − −1µhT x, T yiW = −µ −1γ + δµ
Analogous to (*),
βµ = δµ, so β = δ
Hence hx, yiV = hT x, T yiW .
97
Duc Vu (Spring 2021) 24 Lec 24: May 21, 2021
§24 L e c 2 4 : M ay 2 1 , 2 0 2 1
§24.1 I n n e r P r o d u c t S p a c e s ( C o nt ’ d )
Proposition 24.2
Let V be an inner product space over F , T : V → V linear. Suppose that T ∗ exists.
Then, T is an isometry if and only if T ∗ = T −1 , i.e., T T ∗ = 1V = T ∗ T .
hT x, yi = hT x, 1V yi = hT x, T T −1 yi = hx, T −1 yi
Remark 24.3. Let V be a finite dimensional inner product space over F , T : V → V linear.
1. T is monic iff T is epic iff T is an iso of vector space over F .
2. T is unitary ⇐⇒ T ∗ T = 1V ⇐⇒ T T ∗ = 1V
3. T is unitary ⇐⇒ T ∗ is unitary as T ∗∗ = T
Un F := {A ∈ GLn F | AA∗ = I}
98
Duc Vu (Spring 2021) 24 Lec 24: May 21, 2021
Remark 24.5. 1. Let F ⊆ C, F = F , F n×1 , F 1×n inner product space over F via the
dot product. If A ∈ Mn F , then
linear and sn,1 the ordered standard basis. Then A is unitary iff
i) The columns of A form an ordered orthonormal basis for F n×1
ii) The rows of A form an ordered orthonormal basis for F 1×n
2. If T : V → V is linear, V an inner product space over F with dim V = n, B, C ordered
orthonormal bases for V , then T is unitary iff [T ]B,C is unitary.
Lemma 24.6
Let V be an inner product space over F , T : V → V linear, W ⊆ V a subspace.
Suppose that T ∗ exists. Then the following is true: If W is T -invariant, then W ⊥ is
T ∗ -invariant.
hw, T ∗ vi = hT w, vi = 0
Lemma 24.7
Let V be a finite dimensional inner product space over F , T : V → V linear. Then
the following is true: If λ is an eigenvalue of T , then λ is an eigenvalue of T ∗ .
S ∗ = T ∗ − λ1V : V → V linear
Then ∀w ∈ V ,
0 = h0, wi = hSv, wi = hv, S ∗ wi
Hence v ⊥ im S ∗ and v ∈
/ im S ∗ as v 6= 0. By the Dimension Theorem,
99
Duc Vu (Spring 2021) 24 Lec 24: May 21, 2021
By expansion,
f fT ∈ F [t]
T
(F v)⊥
100
Duc Vu (Spring 2021) 25 Lec 25: May 24, 2021
§25 L e c 2 5 : M ay 2 4 , 2 0 2 1
§25.1 S p e c t r a l T h e o r y ( C o nt ’ d )
T T ∗ = T ∗T
hT v, wi = hv, T wi ∀v, w ∈ V
Lemma 25.2
Let V be an inner product space over F , λ ∈ F , 0 6= v ∈ V , T : V → V a normal
operator. Then
v ∈ ET (λ) ⇐⇒ v ∈ ET ∗ (λ)
Corollary 25.3
Let V be an inner product space over F , T : V → V normal, λ 6= µ eigenvalue of T .
Then, ET (λ) and ET (µ) are orthogonal. In particular,
X 1
ET (λ) = ET (λ)
λ
λ
101
Duc Vu (Spring 2021) 25 Lec 25: May 24, 2021
Proposition 25.4
Let V be a finite dimensional inner product space over F , F = R or C, T : V → V
linear, B an ordered orthonormal basis for V s.t. [T ]B is upper triangular. Then, T
is normal if and only if [T ]B is diagonal.
Proof. “ ⇐= ” If
λ1 0
[T ]B =
..
.
0 λn
then
λ1 0
[T ∗ ]B = [T ]∗B =
..
.
0 λn
So
|λ1 |2
0
[T T ∗ ]B = [T ]B [T ∗ ]B =
..
.
0 |λn |2
= [T ∗ ]B [T ]B
= [T ∗ T ]B
By definition,
n
X n
X
T ∗ v1 = (A∗ )i1 vi = A1i vi
i=1 i=1
So
A1i = 0 ∀i > 1
Hence,
A1i = 0 ∀i > 1
In particular,
A12 = 0
By the lemma,
T v2 = A22 v2 , hence T ∗ v2 = A22 v2
The same argument shows A2i = 0, i 6= 2, i.e.,
A2i = 0, i 6= 2
102
Duc Vu (Spring 2021) 25 Lec 25: May 24, 2021
Proof. This follows immediately by Schur’s theorem, FTA, and the above proposition.
Remark 25.6. Let V be a finite dimensional inner product space over R, T : V → V linear.
Suppose that fT ∈ R[t] splits. Then T is normal iff ∃ an orthonormal basis B for V consisting
of eigenvectors for T .
By Schur’s theorem, T is triangularizable via an orthonormal basis for V . The same result
follows by the proposition in the case F = R.
Theorem 25.7
Let V be a finite dimensional inner product space over F , F = R or C, T : V → V
linear with fT ∈ F [t] splits. Then, T is normal if and only if ∃g ∈ F [t] s.t. T ∗ = g(T ).
103
D u c Vu ( S p r i n g 2 0 2 1 ) 26 L e c 2 6 : M ay 2 6 , 2 0 2 1
§26 L e c 2 6 : M ay 2 6 , 2 0 2 1
§26.1 S p e c t r a l T h e o r y ( C o nt ’ d )
Remark 26.1. A rotation Tθ : R2 → R2 by ∠θ, 0 < θ < 2π, θ 6= π has no eigenvalues, but is
normal (with R2 an inner product space over R via the dot product) as it is unitary.
Lemma 26.2
Let V be an inner product space over F , T : V → V hermitian. If λ is an eigenvalue
of T , then λ ∈ F ∩ R.
Lemma 26.3
Let V be a finite dimensional inner product space over F with F = R or C, T : V → V
hermitian. Then fT ∈ F [t] splits in F [t].
Proof. By previous result, we can assume that F = R. Let B be an orthonormal basis for
V . Then
A := [T ]B = [T ∗ ]B = [T ]∗B = A∗
in Mn R ⊆ Mn C, n = dim V . As
A : Cn×1 → Cn×1 is Hermitian
fA splits with real roots by Lemma 26.2. (and FTA), i.e.,
Y
fA = (t − λi ) ∈ C[t], λi ∈ R ∀i
Q
So fT = fA = (t − λi ) ∈ R[t] splits.
104
Duc Vu (Spring 2021) 26 Lec 26: May 26, 2021
Theorem 26.5
If 0 6= V is a finite dimensional inner product space over R, T : V → V hermitian,
then T has an eigenvalue.
The proof in Axler’s book is very nice, and he does not use determinant theory. He uses
the following arguments
Lemma 26.6
Let q = t2 + bt + c in R[t], b2 < 4c, i.e., q is an irreducible monic quadratic polynomial
in R[t]. If V is a finite dimensional inner product space over R and T : V → V is
Hermitian, then q(T ) is an isomorphism.
So q(T )v 6= 0.
As all the qi (T ) are isomorphism, at least one of the (T − λi 1V ) is not injective, i.e., λi is
an eigenvalue.
105
Duc Vu (Spring 2021) 27 Lec 27: May 28, 2021
§27 L e c 2 7 : M ay 2 8 , 2 0 2 1
§27.1 Po s i t i ve ( S e m i - ) D e f i n i t e O p e r a t o r s
Let V be a finite dimensional inner product space over F , where F = R or C, T : V → V
hermitian, B = {v1 , . . . , vn } an orthonormal basis of eigenvectors of T , i.e.,
T vi = λi vi , i = 1, . . . , n
So λi ∈ R, i = 1, . . . , n. Suppose v ∈ V . Then
Xn
v= αi vi , αi ∈ F ∀i
i=1
and
Xn n
X
hT v, vi = h T (αi vi ), αj vj i
i=1 j=1
Xn n
X
=h λi αi vi , αj vj i
i=1 j=1
Xn
= λi αi αj hvi , vj i (*)
i,j=1
X n
= λi αi αj δij
i,j=1
X n
= λi |αi |2
i=1
hT v, vi > 0 ∀0 6= v ∈ V
hT v, vi ≥ 0 ∀0 6= v ∈ V
Proposition 27.2
Let V be a finite dimensional inner product space over F , F = R or C, T : V → V
hermitian. Then T is positive semi-definite (respectively positive) if and only if all
eigenvalues of T are non-negative (respectively positive).
106
Duc Vu (Spring 2021) 27 Lec 27: May 28, 2021
Question 27.1. What does this say about the 2nd derivative test for C 2 function, f : S → R
at a critical point in the interior of S?
Theorem 27.3
Let V be a finite dimensional inner product space over F , F = R or C, T : V → V
hermitian. Then T is non-negative (respectively positive) iff ∃S : V → V non-negative
s.t.
T = S2
i.e., T has a square root (respectively, and S is invertible).
So
B 2 = [T ]B
By MTT, ∃S : V → V linear s.t. [S]B = B. So
[T ]B = B 2 = [S]2B = S 2 B
[S ∗ ]B = [S]∗B = B ∗ = B = [S]B
107
Duc Vu (Spring 2021) 27 Lec 27: May 28, 2021
Theorem 27.4
Let V be a finite dimensional inner product space over F , F = R or C and T : V → V
hermitian. Suppose that T is non-negative. Then T has a unique square root S, i.e.,
S : V → V non-negative s.t. S 2 = T .
V = ET (λ1 ) ⊥ . . . ⊥ ET (λr )
T = λ1 Pλ1 + . . . + λr Pλr
Pλi Pλj = δij Pλi Pλj , ∀i, j
1 V = P λ1 + . . . + P λr
V = ES (µ1 ) ⊥ . . . ⊥ ES (µs ), µi ≥ 0, i = 1, . . . , s
S = µ1 Pµ1 + . . . + µs Pµs
Pµi Pµj = δij Pµi , ∀i, j
1V = Pµ1 + . . . + Pµs
In particular,
As T = S 2 ,
µ21 Pµ1 + . . . + µ2s Pµs = λ1 Pλ1 + . . . + λr µr
So by uniqueness, we must have s = r and changing the order if necessary
Lemma 27.5
Let V, W be finite dimensional inner product space over F , F = R or C, T : V → W
linear. Then T ∗ T : V → V is hermitian and non-negative.
Remark 27.6. If in the definition of positive operator, etc, we omit V being finite dimensional
but assume T ∗ exists, then we would still have T ∗ T hermitian.
(T ∗ T )∗ = (T ∗ T ∗∗ )∗ = T ∗ T
108
Duc Vu (Spring 2021) 27 Lec 27: May 28, 2021
So
kT vk2W
λ= ≥0
kvk2V
as kvk2V 6= 0.
Corollary 27.7
Let V be a finite dimensional inner product space over F , F = R or C, T : V → V linear.
Then T is non-negative (respectively positive) iff ∃S : V → V linear (respectively an
isomorphism) s.t. T = S ∗ S.
Notation:
• F = R or C, A ∈ F m×n
• A = A(1) . . . A(m)
• UN (F ) = U ∈ GLN F | U ∗ = U −1
109
Duc Vu (Spring 2021) 27 Lec 27: May 28, 2021
µ1 ≥ . . . ≥ µ r > 0
and
r = rank(A)
λ1 ≥ . . . ≥ λr > 0
C ∗ C = (AU )∗ (AU ) ∈ Mn F
Write
λi = µ2i , µi > 0, 1≤i≤r
So
µ1 ≥ . . . ≥ µ r > 0
Set
µ1 0
..
.
µr
∈ Mn F
B=
0
..
.
0 0
110
Duc Vu (Spring 2021) 27 Lec 27: May 28, 2021
Hence
C = C (1) . . . C (r) 0 . . . 0
111
Duc Vu (Spring 2021) 28 Lec 28: Jun 2, 2021
§28.1 Po s i t i ve ( S e m i - ) D e f i n i t e O p e r a t o r s ( C o nt ’ d )
Proof. (Cont’d) Recall, we have proven so far
C = C (1) . . . C (r) 0 . . . 0
Since both Sm,1 and B are orthonormal bases, X ∈ Um (F ). Let D be the pseudo-diagonal
matrix
µ1 0
..
.
∈ F m×n
D := µr
0
..
0 .
as in the statement of the theorem. Then
µ1
..
.
(1) (m)
XD = X ... X
µr
0
..
.
= µ1 X (1) . . . µr X (r) 0 . . . 0
= C = AU
Hence
X ∗ AU = D
as needed.
112
Duc Vu (Spring 2021) 28 Lec 28: Jun 2, 2021
A = XDU ∗ , U ∈ Un F, X ∈ Um F
µ1 0
..
.
∈ F m×n
D= µr (*)
0
..
0 .
µ1 ≥ . . . ≥ µ r > 0 ∈ R
Corollary 28.2
The singular values of A ∈ F m×n , F = R or C are unique (including multiplicity) up
to order.
Corollary 28.4
The singular values of A ∈ F m×n , F = R or C are the same as the singular values of
A∗ ∈ F n×m .
113
Duc Vu (Spring 2021) 28 Lec 28: Jun 2, 2021
B = {v1 , . . . , vn } for V
C = {w1 , . . . , wm } for W
µ1 ≥ . . . ≥ µ r > 0 ∈ R
satisfying (
µi wi , i = 1, . . . , r
T vi =
0, i>r
Conversely, suppose the above conditions are all satisfied. Then vi is an eigenvector for
T ∗ T with eigenvalue µ2i for i = 1, . . . , r and eigenvalue 0 for i = r + 1, . . . , n. In particular,
µ1 , . . . , µr are uniquely determined.
A = Ũ N
√
here N ↔ r, Ũ ↔ e −1θ for n = 1.
Since
(XU ∗ )∗ (XU ∗ ) = U X ∗ XU ∗ = U U ∗ = I
XU ∗ ∈ Un F also. Let Ũ = XU ∗ ∈ Un F , N = U DU ∗ which completes the proof.
114
Duc Vu (Spring 2021) 28 Lec 28: Jun 2, 2021
T x = vW = PW (v)
Solution:
hx, T ∗ yiX = hT x, yiV ,
we have
W ⊥ = (im T )⊥ = ker T ∗
Since
v − vW ∈ W ⊥ = (im T )⊥ (by the OR Decomposition Theorem)
and
T ∗ v = T ∗ vW
So if x is a best approximation of v via T , then
T ∗T x = T ∗v (*)
T x − v ∈ ker T ∗ = (im T )⊥ = W ⊥
In particular,
vW = PW v = PW (T x − (T x − v))
= PW (T x) − PW (T x − v)
= Tx + 0 = Tx
T ∗ T : X → X is an isomorphism
and
PW = T (T ∗ T )−1 T ∗ : V → V (+)
115
Duc Vu (Spring 2021) 28 Lec 28: Jun 2, 2021
T ∗T x = T ∗v (??)
and
T (T ∗ T )−1 T ∗ v = T (T ∗ T )−1 T ∗ T x
= T x = vW = PW (v)
showing (+). This proves the claim and also shows that the x in (??) is a best approximation
to v via T .
116
Duc Vu (Spring 2021) 29 Lec 29: Jun 4, 2021
§29.1 L e a s t S q u a r e s ( C o nt ’ d )
Claim 29.1. Let v ∈ V . Then ∃!x ∈ X an optimal approximation to v via T . Moreover,
this x is characterized by
T ∗ T x = T ∗ v = T ∗ T x0
Therefore,
x − x0 ∈ ker T ∗ T =: Y
It follows if x is a best approximation to v via T , then any other is of the form x + y, y ∈ Y .
We also have for such x + y
So
kx00 + yk2 = kx00 k2 + kyk2 ≥ kx00 k2 ∀y ∈ Y
by the Pythagorean Theorem. Hence, x00
= PY ⊥ (x) is the unique optimal approximation.
This proves the claim above.
Let A = T : F n×1 → F m×1 , A ∈ F m×n , v ∈ F m×1 with F = R or C. Let
µ1
..
.
∗ ∈ F m×n
A = XDU , D = µr
0
..
.
and
µ1 ≥ . . . ≥ µ r > 0 ∈ R
be an SVD. Let’s define
µ−1
1
..
.
† ∈ F n×m
µ−1
D :=
r
0
..
.
Then
A† := U D† X ∗ ∈ F n×m
is called the Moore-Penrose generalized pseudoinverse of A. Then the following are true
117
Duc Vu (Spring 2021) 29 Lec 29: Jun 4, 2021
i) rank(A) = rank(A† )
and we have
D† v ∈ Span {ej ∈ Sn,1 |Djj 6= 0} = Y ⊥ ,
and PY D† v = 0.
iii) This follows as in (ii) for if rank(A) = n, then (A∗ A)−1 A∗ v is the unique optimal
best approximation to Az = v.
AX = B for X ∈ F n×1
118
Duc Vu (Spring 2021) 29 Lec 29: Jun 4, 2021
Example 29.1
Let F = R or C. Given data (x1 , y1 ), . . . , (xn , yn ) in F 2 , find the best line relative to
this data, i.e., find
y = λx + b, λ = slope
Let
x1 1 y1
.. .. , λ
A= . X= , Y = ...
. b
xn 1 yn
To solve AX = Y , we want the optimal solution
x1 1 y1
.. .. λ ..
. . = .
b
xn 1 yn
X = (A∗ A)−1 A∗ Y
§29.2 R ay l e i g h Q u o t i e nt
Let F = R or C, A ∈ Mn F . The euclidean norm of A is defined by
kAvk
kAk := max
06=v∈F n×1 kvk
is defined by
hAv, vi
R(v) :=
kvk2
Rayleigh quotients are used to approximate eigenvalues of hermitian A ∈ Mn F .
Theorem 29.2
Let F = R or C, A ∈ Mn F hermitian. Then,
λ1 ≥ . . . ≥ λn ∈ R
119
Duc Vu (Spring 2021) 29 Lec 29: Jun 4, 2021
Pn
i) Let v ∈ F n×1 and v = i=1 αi vi , αi ∈ F , i = 1, . . . , n. Then
n n
hAv, vi X X
R(v) = 2
=h αi λ i vi , αj vj i/kvk2
kvk
i=1 j=1
Pn Pn
i,j=1 i i j ij i j i
hv
λ α α δ , v 2
i=1 λi − |αi |
= =
kvk2 kvk2
By the Pythagorean Theorem
n
X
|αi |2 = kvk2
i=1
So Pn 2
i=1 λ1 |αi | λ1 kvk2
R(v) ≤ = = λ1
kvk2 kvk2
ii) Prove similarly.
Corollary 29.3
Let F = R or C, A ∈ Mn F . Then kAk < ∞. Moreover, if µ is the largest singular
value of A, then
kAk = µ
Proof. Consider:
kAvk2 hAv, Avi hA∗ Av, vi
0≤ = =
kvk2 kvk2 kvk2
for all v 6= 0. Since A∗ A is non-negative, the result follows.
We know that the singular value of A ∈ F m×n are the same as for A∗ ∈ F n×m if F = R or
C. Therefore,
Corollary 29.4
Let A ∈ GLn F, F = R or C, µ the smallest singular value of A. Then
1
kA−1 k = √
µ
120
Duc Vu (Spring 2021) 30 Additional Materials: Jun 04, 2021
§30.1 C o n d i t i o n a l N u mb e r
Let F = R or C, A ∈ GLn F , b 6= 0 in F n×1 . Suppose Ax = b.
Problem 30.1. What happens if we modify x a bit, i.e., by δx ∈ F n×1 . Then we get a
new equation
A(x + δx) = b + δb, δb ∈ F n×1
and we would like to understand the variance in b.
Since A is linear,
A(x + δx) = b + A(δx)
i.e.
A(δx) = δb or δx = A−1 (δb)
and we know, therefore, that
Therefore,
1 kAk
≤ as kxk =
6 0 (b 6= 0)
kxk kbk
kδxk kA−1 kkδbk kAk kδbk
=⇒ ≤ · = kAkkA−1 k
kxk 1 kbk kbk
Similarly,
1 kδbk kδxk
−1
≤
kAkkA k kbk kxk
We call the number kAkkA−1 k the Conditional Number of A and denote it cond(A).
Theorem 30.1
Let F = R or C, A ∈ GLn F , b 6= 0 in F n×1 . Then
kδbk kδxk
1. 1
cond(A) kbk ≤ kxk ≤ cond(A) kδbk
kbk
121
Duc Vu (Spring 2021) 30 Additional Materials: Jun 04, 2021
Remark 30.3. If there is an error SA of A, things would get more complicated. For example,
A + δA may no longer be invertible.
There exist conditions that can control this. For example, if A + SA ∈ GLn F , F = R or C,
it is true that
kδxk kδAk
≤ cond(A)
kx + δxk kAk
One almost never computes cond(A), as error arises trying to compute it as we need to
compute the singular values. However, in some cases, remarkable estimates can be found.
§30.2 Mini-Max
Let F = R or C, A ∈ Mn F . We want a method to compute its eigenvalues if A is hermitian.
Since A is hermitian, by the Spectral Theorem,
λ1 0
U ∗ AU =
.. , U ∈ Un F
.
0 λn
where A = [A]Sn,1 .
B = {v1 , . . . , vn } is an ordered orthonormal basis of eigenvectors for V = F n×1 satisfying
Avi = λi vi
So
vi = the ith column of U ∗
We let the order be s.t.
λ1 ≥ . . . ≥ λn
As (F v1 )⊥ is A-invariant, A (F v1 )⊥
has maximum eigenvalue λ2 obtained from v2 , i.e.,
y
max RU (y) ≥ λn−1
hy,wi=0
122
Duc Vu (Spring 2021) 30 Additional Materials: Jun 04, 2021
So
min max
y
RU (y) ≥ λn−1
w6=0
hy,wi=0
RU (ỹ) = λ2
Let w = e1 . Then
max RU (y) = λ2
y
hy,e1 i
So
min max
y
RU (y) = λ2
w6=0
hy,wi=0
and
min max
y
RU (y) = λ3
w1 ,w2 6=0
hy,w1 i=0
hy,w2 i=0
Proceed inductively.
λ1 ≥ . . . ≥ λn
Then
min max RA (x) = λk
z1 ,...,zk 6=0 hx,z1 i=0
..
.
hx,zk i=0
§30.3 U n i q u e n e s s o f S m i t h N o r m a l Fo r m
Consult Professor Elman’s notes.
123