stml094 Endmatter
stml094 Endmatter
Volume 94
Analysis and
Linear Algebra:
The Singular Value
Decomposition
and Applications
James Bisgard
10.1090/stml/094
Analysis and
Linear Algebra:
The Singular Value
Decomposition
and Applications
ST UD ENT MATH EM AT IC A L LI BR ARY
Volume 94
Analysis and
Linear Algebra:
The Singular Value
Decomposition
and Applications
James Bisgard
EDITORIAL COMMITTEE
John McCleary Kavita Ramanan
Rosa C. Orellana John Stillwell (Chair)
Preface xi
Pre-Requisites xv
Notation xvi
Acknowledgements xvii
Chapter 1. Introduction 1
§1.1. Why Does Everybody Say Linear Algebra is “Useful”? 1
§1.2. Graphs and Matrices 4
§1.3. Images 7
§1.4. Data 9
§1.5. Four “Useful” Applications 9
vii
viii Contents
Bibliography 209
Index 215
Preface
xi
xii Preface
uniqueness theorem for initial value problems and the inverse function
theorem). But many of those results require yet more machinery, and so
students have to wait to see why the Banach Fixed Point Theorem is use-
ful until we have that machinery. On the other hand, after having been
told that math is useful for several years, students can be understandably
dubious when being told that what they’re learning is useful.
For the student: What should you get out of this book? First, a
better appreciation of the “applicability” of the analytic tools you have,
as well as a sense of how many of the basic ideas you know can be gen-
eralized. On a more itemized level, you will see how linear algebra and
analysis can be used in several “data science” type problems: determin-
ing how close a given set of data is to a given subspace (the “best” sub-
space problem), how to solve least squares problems (the Moore-Penrose
pseudo-inverse), how to best approximate a high rank object with a lower
rank one (low rank approximation and the Eckart-Young-Mirsky Theo-
rem), and how to find the best transformation that preserves angles and
distances to compare a given data set to a reference one (the orthogonal
Procrustes problem). As you read the text, you will find exercises — you
should do them as you come to them, since they are intended to help
strengthen and reinforce your understanding, and many of them will be
helpful later on!
For the student and instructor: What is the topic here? The ex-
traordinary utility of linear algebra and analysis. And there are many,
many examples of that usefulness. One of the most “obvious” exam-
ples of the utility of linear algebra comes from Google’s PageRank Algo-
rithm, which has been covered extremely well by Langville and Meyer,
in [22] (see also [3]). Our main topic is the Singular Value Decomposi-
tion (SVD). To quote from Golub and Van Loan [13], Section 2.4, “[t]he
practical and theoretical importance of the SVD is hard to overestimate.”
There is a colossal number of examples of SVD’s usefulness. (See for
example the Netflix Challenge, which offered a million dollar prize for
improving Netflix’s recommendations by 10% and was won by a team
which used the SVD.) What then justifies (at least in this author’s mind)
another book? Most of the application oriented books do not provide
proofs (see my interest in “why is that true?”) of the foundational parts,
commonly saying “. . . as is well known . . . ” Books that go deeply into the
proofs tend more to the numerical linear algebra side of things, which
are usually oriented to the (incredibly important) questions of how
Preface xiii
Pre-Requisites
It is assumed that readers have had a standard course in linear alge-
bra and are familiar with the ideas of vector spaces (over ℝ), subspaces,
bases, dimension, linear independence, matrices as linear transforma-
tions, rank of a linear transformation, and nullity of a linear transforma-
tion. We also assume that students are familiar with determinants, as
well as eigenvalues and how to calculate them. Some familiarity with
linear algebra software is useful, but not essential.
In addition, it is assumed that readers have had a course in basic
analysis. (There is some debate as to what such a course should be
called, with two common titles being “advanced calculus” or “real anal-
ysis.”) To be more specific, students should know the definition of infi-
mum and supremum for a non-empty set of real numbers, the basic facts
about convergence of sequences, the Bolzano-Weierstrass Theorem (in
the form that a bounded sequence of real numbers has a convergent sub-
sequence), and the basic facts about continuous functions. (For a much
more specific background, the first three chapters of [31] are sufficient.)
Any reader familiar with metric spaces at the level of Rudin [32] is defi-
nitely prepared, although exposure to metric space topology is not neces-
sary. We will work with a very particular type of metric space: normed
vector spaces, and Chapter 2 provides a background for students who
xvi Preface
may not have seen it. (Even students familiar with metric spaces may
benefit by reading the sections in Chapter 2 about convexity and coer-
civity.)
Notation
If 𝐴 is an 𝑚 × 𝑛 real matrix, a common way to write the Singular Value
Decomposition is 𝐴 = 𝑈Σ𝑉 𝑇 , where 𝑈 and 𝑉 are orthogonal (so their
columns form orthonormal bases), and the only non-zero entries in Σ
are on the main diagonal. (And 𝑉 𝑇 is the transpose of 𝑉.) With this
notation, if 𝑢𝑖 are the columns of 𝑈, 𝑣𝑗 are the columns of 𝑉, and the
diagonal entries of Σ are 𝜎 𝑘 , we will have 𝐴𝑣 𝑖 = 𝜎 𝑖 𝑢𝑖 and 𝐴𝑇 𝑢𝑖 = 𝜎 𝑖 𝑣 𝑖 .
Thus, 𝐴 maps the 𝑣 to the 𝑢, which means that 𝐴 is mapping vector space
𝒱 into a vector space 𝒰. However, I prefer to preserve alphabetical order
when writing domain and co-domain, which means 𝐴 ∶ 𝒱 → 𝒰 feels
awkward to me. One solution would be to simply reverse the role of 𝑢
and 𝑣 and write the Singular Value Decomposition as 𝐴 = 𝑉Σ𝑈 𝑇 , which
would be at odds with just about every single reference and software out
there and make it extraordinarily difficult to compare to other sources
(or software). On the other hand, it is very common to think of 𝑥 as the
inputs and 𝑦 as outputs for a function (and indeed it is common to write
𝑓(𝑥) = 𝑦 or 𝐴𝑥 = 𝑦 in linear algebra), and so I have chosen to write the
Singular Value Decomposition as 𝐴 = 𝑌 Σ𝑋 𝑇 . From this point of view,
the columns of 𝑋 will form an orthonormal basis for the domain of 𝐴,
which makes 𝐴𝑥𝑖 fairly natural. Similarly, the columns of 𝑌 will form
an orthonormal basis for the codomain of 𝐴, which hopefully makes
𝐴𝑥𝑖 = 𝜎 𝑖 𝑦 𝑖 feel natural.
𝑇
Elements of ℝ𝑛 will be written as [𝑥1 𝑥2 . . . 𝑥𝑛 ] , where the
superscript 𝑇 indicates the transpose. Recall that if 𝐴 is an 𝑚 × 𝑛 matrix
with 𝑖𝑗th entry given by 𝑎𝑖𝑗 , then 𝐴𝑇 is the transpose of 𝐴, which means
𝐴𝑇 is 𝑛 × 𝑚 and the 𝑖𝑗th entry of 𝐴𝑇 is 𝑎𝑗𝑖 . In particular, this means that
elements of ℝ𝑛 should be thought of as column vectors. This means that
𝑥 = [𝑥1 𝑥2 . . . 𝑥𝑛 ]𝑇 is equivalent to
𝑥
⎡ 1⎤
⎢ 𝑥2 ⎥
𝑥 = ⎢ ⎥.
⎢⋮⎥
⎣𝑥𝑛 ⎦
Acknowledgements xvii
Acknowledgements
I would like to first thank CWU’s Department of Mathematics and all
my colleagues there for making it possible for me to go on sabbatical,
during which I wrote most of this book. Without that opportunity, this
book would certainly not exist. Next, my thanks to all of the students
who made it through my year long sequence in “applicable analysis”:
Amanda Boseck, John Cox, Raven Dean, John-Paul Mann, Nate Minor,
Kerry Olivier, Christopher Pearce, Ben Squire, and Derek Wheel. I’m
sorry that it has taken me so long to finally get around to writing the
book. I’m even more sorry that I didn’t know any of this material when I
taught that course. There is also a large group of people who read various
parts of the early drafts: Dan Curtis, Ben Freeman, James Harper, Ralf
Hoffmann, Adrian Jenkins, Mary Kastning, Dominic Klyve, Brianne and
George Kreppein, Mike Lundin, Aaron Montgomery, James Morrow,
Ben Squire, Mike Smith, Jernej Tonejc, and Derek Wheel. Their feed-
back was very useful, and they caught many typos and mistakes. In
particular, Brianne Kreppein, Aaron Montgomery, and Jernej Tonejc
deserve special thanks for reading the entire draft. Any mistakes and
typos that remain are solely my fault! Next, Ina Mette at AMS was fan-
tastic about guiding me through the process. On my sabbatical, there
were many places I worked. I remember working out several important
details in the Suzallo Reading Room at the University of Washington,
and a few more were worked out at Uli’s Bierstube in Pike Place Mar-
ket. My family was also a source of inspiration. My parents Carl and
xviii Preface
James Bisgard
Bibliography
209
210 Bibliography
[19] Wolfgang Kabsch, A discussion of the solution for the best rotation
to relate two sets of vectors, Acta Crystallographica A34 (1978), 827–
828.
[20] Dan Kalman, A singularly valuable decomposition: The SVD of a
matrix, College Math. J. 27 (1996), no. 1, 2–23.
[21] Tosio Kato, A short introduction to perturbation theory for linear op-
erators, Springer-Verlag, New York-Berlin, 1982. MR678094
[22] Amy N. Langville and Carl D. Meyer, Google’s PageRank and be-
yond: the science of search engine rankings, Princeton University
Press, Princeton, NJ, 2006. MR2262054
[23] J. Lawrence, J. Bernal, and C. Witzgall, A purely algebraic justifica-
tion of the Kabsch-Umeyama algorithm, Journal of Research of the
National Institute of Standards and Technology 124 (2019), 1–6.
[24] Peter D. Lax, Linear algebra and its applications, 2nd ed., Pure and
Applied Mathematics (Hoboken), Wiley-Interscience [John Wiley
& Sons], Hoboken, NJ, 2007. MR2356919
[25] R. Lee and L. Carter, Modeling and forecasting US mortality, J.
American Statistical Assoc. 87 (1992), 659–671.
[26] Carla D. Martin and Mason A. Porter, The extraordinary
SVD, Amer. Math. Monthly 119 (2012), no. 10, 838–851, DOI
10.4169/amer.math.monthly.119.10.838. MR2999587
[27] Jiří Matoušek, Thirty-three miniatures, Student Mathematical Li-
brary, vol. 53, American Mathematical Society, Providence, RI,
2010. Mathematical and algorithmic applications of linear algebra.
MR2656313
[28] Elizabeth S. Meckes and Mark W. Meckes, Linear algebra, Cam-
bridge University Press, Cambridge, 2018.
[29] L. Mirsky, Symmetric gauge functions and unitarily invariant
norms, Quart. J. Math. Oxford Ser. (2) 11 (1960), 50–59, DOI
10.1093/qmath/11.1.50. MR114821
[30] Cleve Moler and Donald Morrison, Singular value analysis of
cryptograms, Amer. Math. Monthly 90 (1983), no. 2, 78–87, DOI
10.2307/2975804. MR691178
212 Bibliography
213
Index
215
216 Index
topology
normed vector space, 32
trace of a matrix, 19
transpose, xvi
zero mapping, 29
Selected Published Titles in This Series
94 James Bisgard, Analysis and Linear Algebra: The Singular Value
Decomposition and Applications, 2021
93 Iva Stavrov, Curvature of Space and Time, with an Introduction to
Geometric Analysis, 2020
92 Roger Plymen, The Great Prime Number Race, 2020
91 Eric S. Egge, An Introduction to Symmetric Functions and Their
Combinatorics, 2019
90 Nicholas A. Scoville, Discrete Morse Theory, 2019
89 Martin Hils and François Loeser, A First Journey through Logic, 2019
88 M. Ram Murty and Brandon Fodden, Hilbert’s Tenth Problem, 2019
87 Matthew Katz and Jan Reimann, An Introduction to Ramsey Theory,
2018
86 Peter Frankl and Norihide Tokushige, Extremal Problems for Finite
Sets, 2018
85 Joel H. Shapiro, Volterra Adventures, 2018
84 Paul Pollack, A Conversational Introduction to Algebraic Number
Theory, 2017
83 Thomas R. Shemanske, Modern Cryptography and Elliptic Curves, 2017
82 A. R. Wadsworth, Problems in Abstract Algebra, 2017
81 Vaughn Climenhaga and Anatole Katok, From Groups to Geometry
and Back, 2017
80 Matt DeVos and Deborah A. Kent, Game Theory, 2016
79 Kristopher Tapp, Matrix Groups for Undergraduates, Second Edition,
2016
78 Gail S. Nelson, A User-Friendly Introduction to Lebesgue Measure and
Integration, 2015
77 Wolfgang Kühnel, Differential Geometry: Curves — Surfaces —
Manifolds, Third Edition, 2015
76 John Roe, Winding Around, 2015
75 Ida Kantor, Jiřı́ Matoušek, and Robert Šámal, Mathematics++,
2015
74 Mohamed Elhamdadi and Sam Nelson, Quandles, 2015
73 Bruce M. Landman and Aaron Robertson, Ramsey Theory on the
Integers, Second Edition, 2014
72 Mark Kot, A First Course in the Calculus of Variations, 2014
71 Joel Spencer, Asymptopia, 2014
STML/94