Math 5610 Fall 2018 Notes of 9/24/18 Review: The Significance of Orthogonal Matrices
Math 5610 Fall 2018 Notes of 9/24/18 Review: The Significance of Orthogonal Matrices
Notes of 9/24/18
A = U ΣV T (5)
where
• U is m × m orthogonal, i.e., U −1 = U T ,
• V is n × n orthogonal, i.e., V −1 = V T , and
• Σ is m × n diagonal. Specifically,
σ1 0 . . . 0
0 σ2 . . . 0
. .. . . ..
.
. . . .
Σ= 0 0 . . . σn (6)
0 0 ... 0
. .. ..
.. . .
0 0 ... 0
where
σ1 ≥ σ2 ≥ . . . ≥ σn ≥ 0. (7)
Note that Σ is a matrix! The capital Greek letter
Σ in this context has nothing to do with the
summation symbol−2− . The σi are the singular
−2−
I first learned about the singular value decom-
position in an excellent talk by Cleve Moler that
I understood only in retrospect. At the time of
that first exposure the talk was utterly wasted
on me because the whole time I kept thinking
What is he summing there? However, the nota-
tion A = U ΣV T is well established.
AT Avj = V ΣT U T T
| {zU} Σ |V {zv}j
=I =ej
= V ΣT Σej
(10)
= V Sej
= σj2 V ej
= σj2 vj
2
(14)
we have, after dividing by σ 2 + wT w , that
kA1 k22 ≥ (σ 2 + wT w). (15)
But
σ 2 = kAk22 = kA1 k22 ≥ σ 2 + wT w, (16)
and so we must have w = 0. An obvious induc-
tion argument completes the proof of the theo-
rem.
Σ = U T AV (17)
References.
The SVD is discussed in many textbooks on nu-
merical analysis, or numerical linear algebra. The
most comprehensive discussion is in the author-
itative monograph
• Gene H. Golub and Charles F. van Loan, Ma-
trix Computations, 4th ed., The Johns Hopkins
University Press, 2013, ISBN 10: 1-4214-0794-9.
Some Applications.
Note that the following list in no way is meant
to be complete.
Rank Determination.
−3−
Software to compute the SVD is available, for
example, at
http://www.netlib.org/
−4−
The round off unit ǫ, also called the machine
epsilon, is the smallest number that can be rep-
resented on a computer such that the system rec-
ognizes 1 + ǫ as being larger than 1. On many
systems, including our Unix systems, ǫ equals
approximately 2 × 10−16 .
σ1
kAk2 kA−1 k2 = . (23)
σn
kAk2 = σ1 (24)
1
kA−1 k2 = (25)
σn
Ax = b (26)
Ax = U ΣV T x = b. (27)
Σz = c (28)
where
z = V T x and c = U T b. (29)
This is a diagonal linear system that can be
analyzed easily.
c
σ1 1
.. ...
. 0 z1
c
σm .. r
. c
0 r+1
.. zr ..
0 .
= .
zr+1
0 . cn
..
0 0 0 ... 0 0 cn+1
. .. .. .. .. zn .
.. . . . . .
.
0 0 0 ... 0 0 cm
(30)
Recalling (19) we distinguish three cases:
1. r = n and
cn+1 = . . . = cm = 0. (31)
x = V z. (34)
Data Compression.
It’s an easy exercise to see that
n
X
T
A = U ΣV = σi ui viT (38)
i=1
AT Ax = AT b. (39)
y T Ax
σ1 = maxm . (40)
y ∈ IR kyk2 kxk2
x ∈ IRn