Tensor
Tensor
Differential Forms
Contents
Preface v
Introduction v
Organization vi
Notational Conventions x
Acknowledgments xi
iv Contents
Preface
Introduction
For most math undergraduates one’s first encounter with differential forms is the change
of variables formula in multivariable calculus, i.e. the formula
𝜕𝑓𝑖
𝐽𝑓 (𝑥) ≔ [ (𝑥)] ,
𝜕𝑥𝑗
As for the “𝑑𝑥” and “𝑑𝑦”, their presence in (1) can be accounted for by the fact that in single-
variable calculus, with 𝑈 = (𝑎, 𝑏), 𝑉 = (𝑐, 𝑑), 𝑓 ∶ (𝑎, 𝑏) → (𝑐, 𝑑), and 𝑦 = 𝑓(𝑥) a 𝐶1 function
with positive first derivative, the tautological equation 𝑑𝑦
𝑑𝑥
= 𝑑𝑓
𝑑𝑥
can be rewritten in the form
𝑑(𝑓⋆ 𝑦) = 𝑓⋆ 𝑑𝑦
and (1) can be written more suggestively as
(2) ∫ 𝑓⋆ (𝜙 𝑑𝑦) = ∫ 𝜙 𝑑𝑦
𝑈 𝑉
One of the goals of this text on differential forms is to legitimize this interpretation of equa-
tion (1) in 𝑛 dimensions and in fact, more generally, show that an analogue of this formula
is true when 𝑈 and 𝑉 are 𝑛-dimensional manifolds.
Another related goal is to prove an important topological generalization of the change
of variables formula (1). This formula asserts that if we drop the assumption that 𝑓 be a
bijection and just require 𝑓 to be proper (i.e., that pre-images of compact subsets of 𝑉 to be
compact subsets of 𝑈) then the formula (1) can be replaced by
where deg(𝑓) is a topological invariant of 𝑓 that roughly speaking counts, with plus and
minus signs, the number of pre-image points of a generically chosen point of 𝑉.1
This degree formula is just one of a host of results which connect the theory of differ-
ential forms with topology, and one of the main goals of this book will explore some of the
other examples. For instance, for 𝑈 an open subset of 𝐑2 , we define 𝛺0 (𝑈) to be the vector
1It is our feeling that this formula should, like formula (1), be part of the standard calculus curriculum,
particularly in view of the fact that there now exists a beautiful elementary proof of it by Peter Lax (see [6, 8, 9]).
v
Draft: March 28, 2018
vi Preface
space of 𝐶∞ functions on 𝑈. We define the vector space 𝛺1 (𝑈) to be the space of formal
sums
(4) 𝑓1 𝑑𝑥1 + 𝑓2 𝑑𝑥2 ,
where 𝑓1 , 𝑓2 ∈ 𝐶 (𝑈). We define the vector space 𝛺2 (𝑈) to be the space of expressions of
∞
the form
(5) 𝑓 𝑑𝑥1 ∧ 𝑑𝑥2 ,
where 𝑓 ∈ 𝐶 (𝑈), and for 𝑘 > 2 define 𝛺𝑘 (𝑈) to the zero vector space.
∞
Organization
Chapter 1: Multilinear algebra
As we mentioned above one of our objectives is to legitimatize the presence of the 𝑑𝑥
and 𝑑𝑦 in formula (1), and translate this formula into a theorem about differential forms.
However a rigorous exposition of the theory of differential forms requires a lot of algebraic
preliminaries, and these will be the focus of Chapter 1. We’ll begin, in Sections 1.1 and 1.2, by
reviewing material that we hope most of our readers are already familiar with: the definition
of vector space, the notions of basis, of dimension, of linear mapping, of bilinear form, and
Draft: March 28, 2018
Organization vii
of dual space and quotient space. Then in Section 1.3 we will turn to the main topics of
this chapter, the concept of 𝑘-tensor and (the future key ingredient in our exposition of
the theory of differential forms in Chapter 2) the concept of alternating 𝑘-tensor. Those 𝑘
tensors come up in fact in two contexts: as alternating 𝑘-tensors, and as exterior forms, i.e.,
in the first context as a subspace of the space of 𝑘-tensors and in the second as a quotient
space of the space of 𝑘-tensors. Both descriptions of 𝑘-tensors will be needed in our later
applications. For this reason the second half of Chapter 1 is mostly concerned with exploring
the relationships between these two descriptions and making use of these relationships to
define a number of basic operations on exterior forms such as the wedge product operation
(see §1.6), the interior product operation (see §1.7) and the pullback operation (see §1.8).
We will also make use of these results in Section 1.9 to define the notion of an orientation for
an 𝑛-dimensional vector space, a notion that will, among other things, enable us to simplify
the change of variables formula (1) by getting rid of the absolute value sign in the term
| det 𝐽𝑓 |.
viii Preface
will have some important de Rham theoretic applications when we get to Chapter 5. To de-
scribe this result let 𝑈 be a connected open set in 𝐑𝑛 and 𝜔 = 𝑓𝑑𝑥1 ∧ ⋯ ∧ 𝑑𝑥𝑛 a compactly
supported 𝑛-form on 𝑈. We will prove:
Theorem 11. The integral ∫𝑈 𝜔 = ∫𝑈 𝑓𝑑𝑥1 ⋯ 𝑑𝑥𝑛 of 𝜔 over 𝑈 is zero if and only if 𝜔 = 𝑑𝜈
where 𝜈 is a compactly supported (𝑛 − 1)-form on 𝑈.
An easy corollary of this result is the following weak version of the degree theorem:
Corollary 12. Let 𝑈 and 𝑉 be connected open subsets of 𝐑𝑛 and 𝑓 ∶ 𝑈 → 𝑉 a proper mapping.
Then there exists a constant 𝛿(𝑓) with the property that for every compactly supported 𝑛-form
𝜔 on 𝑉
∫ 𝑓⋆ 𝜔 = 𝛿(𝑓) ∫ 𝜔 .
𝑈 𝑉
Thus to prove the degree theorem it suffices to show that this 𝛿(𝑓) is the deg(𝑓) in formula
(3) and this turns out not to be terribly hard.
The degree formula has a lot of interesting applications and at the end of Chapter 3 we
will describe two of them. The first, the Brouwer fixed point theorem, asserts that if 𝐵𝑛 is
the closed unit ball in 𝐑𝑛 and 𝑓 ∶ 𝐵𝑛 → 𝐵𝑛 is a continuous map, then 𝑓 has a fixed point,
i.e. there is a point 𝑥0 ∈ 𝐵𝑛 that gets mapped onto itself by 𝑓. (We will show that if this were
not true the degree theorem would lead to an egregious contradiction.)
The second is the fundamental theorem of algebra. If
𝑝(𝑧) = 𝑎0 + 𝑎1 𝑧 + ⋯ + 𝑎𝑛−1 𝑧𝑛−1 + 𝑧𝑛
is a polynomial function of the complex variable 𝑧, then it has to have a complex root 𝑧0
satisfying 𝑝(𝑧0 ) = 0. (We’ll show that both these results are more-or-less one line corolaries
of the degree theorem.)
Chapter 4: Forms on Manifolds
In the previous two chapters the differential forms that we have been considering have
been forms defined on open subsets of 𝐑𝑛 . In this chapter we’ll make the transition to forms
defined on manifolds; and to prepare our readers for this transition, include at the beginning
of this chapter a brief introduction to the theory of manifolds. (It is a tradition in under-
graduate courses to define 𝑛-dimensional manifolds as 𝑛-dimensional submanifolds of 𝐑𝑁 ,
i.e., as 𝑛-dimensional generalizations of curves and surfaces in 𝐑3 , whereas the tradition in
graduate level courses is to define them as abstract entities. Since this is intended to be a text
book for undergraduates, we’ll adopt the first of these approaches, our reluctance to doing
so being somewhat tempered by the fact that, thanks to the Whitney embedding theorem,
these two approaches are the same.) In section Section 4.1 we will review a few general facts
about manifolds, in section Section 4.2, define the crucial notion of the tangent space 𝑇𝑝 𝑋
to a manifold 𝑋 at a point 𝑝 and in sections Sections 4.3 and 4.4 show how to define dif-
ferential forms on 𝑋 by defining a 𝑘-form to be, as in § 2.4, a function 𝜔 which assigns to
each 𝑝 ∈ 𝑋 an element 𝜔𝑝 of 𝛬𝑘 (𝑇𝑝⋆ 𝑋). Then in Section 4.4 we will define what it means
for a manifold to be oriented and in Section 4.5 show that if 𝑋 is an oriented 𝑛-dimensional
manifold and 𝜔 a compactly supported 𝑛-form, the integral of 𝜔 is well-defined. Moreover,
we’ll prove in this section a manifold version of the change of variables formula and in Sec-
tion 4.6 prove manifold versions of two standard results in integral calculus: the divergence
theorem and Stokes theorem, setting the stage for the main results of Chapter 4: the mani-
fold version of the degree theorem (see §4.7) and a number of applications of this theorem
(among them the Jordan–Brouwer separation theorem, the Gauss–Bonnet theorem and the
Index theorem for vector fields (see §§4.8 and 4.9).
Draft: March 28, 2018
Organization ix
(𝜈, 𝜔) ↦ ∫ 𝜈 ∧ 𝜔
𝑋
gives rise to a pairing
𝐻𝑘 (𝑋) × 𝐻𝑐𝑛−𝑘 (𝑋) → 𝐑 ,
and that if 𝑋 is connected this pairing is non-degenerate, i.e., defines a bijective linear trans-
formation
𝐻𝑐𝑛−𝑘 (𝑋) ⥲ 𝐻𝑘 (𝑋)⋆ .
This results in the Poincaré duality theorem and it has some interesting implications which
we will explore in Sections 5.5 to 5.7. For instance in Section 5.5 we will show that if 𝑌 and
𝑍 are closed oriented submanifolds of 𝑋 and 𝑍 is compact and of codimension equal to
Draft: March 28, 2018
x Preface
Appendices
In Appendix a we review some techniques in multi-variable calculus that enable one to
reduce global problems to local problems and illustrate these techniques by describing some
typical applications. In particular we show how these techniques can be used to make sense
of “improper integrals” and to extend differentiable maps 𝑓 ∶ 𝑋 → 𝐑 on arbitrary subsets
𝑋 ⊂ 𝐑𝑚 to open neighborhoods of these sets.
In Appendix b we discuss another calculus issue: how to solve systems of equations
{ 𝑓 (𝑥) = 𝑦1
{ 1
{ ⋮
{
𝑓
{ 𝑁 (𝑥) = 𝑦𝑁
Notational Conventions
Below we provide a list of a few of our common notational conventions.
≔ used to define a term; the term being defined appears to the left of the colon
⊂ subset
→ denotes an map (of sets, vector spaces, etc.)
↪ inclusion map
↠ surjective map
⥲ a map that is an isomorphism
∖ difference of sets: if 𝐴 ⊂ 𝑋, then 𝑋 ∖ 𝐴 is the complement of 𝐴 in 𝑋
Draft: March 28, 2018
Acknowledgments xi
Acknowledgments
To conclude this introduction we would like to thank Cole Graham and Farzan Vafa
for helping us correct earlier versions of this text, a task which involved compiling long list
of typos and errata. (In addition a lot of pertinent comments of theirs helped us improve
the readability of the text itself.)
Draft: March 28, 2018
Draft: March 28, 2018
CHAPTER 1
Multilinear Algebra
1.1. Background
We will list below some definitions and theorems that are part of the curriculum of a
standard theory-based course in linear algebra.1
Definition 1.1.1. A (real) vector space is a set 𝑉 the elements of which we refer to as vectors.
The set 𝑉 is equipped with two vector space operations:
(1) Vector space addition: Given vectors 𝑣1 , 𝑣2 ∈ 𝑉 one can add them to get a third vector,
𝑣1 + 𝑣2 .
(2) Scalar multiplication: Given a vector 𝑣 ∈ 𝑉 and a real number 𝜆, one can multiply 𝑣 by
𝜆 to get a vector 𝜆𝑣.
These operations satisfy a number of standard rules: associativity, commutativity, dis-
tributive laws, etc. which we assume you’re familiar with. (See Exercise 1.1.i.) In addition
we assume that the reader is familiar with the following definitions and theorems.
(1) The zero vector in 𝑉 is the unique vector 0 ∈ 𝑉 with the property that for every vector
𝑣 ∈ 𝑉, we have 𝑣 + 0 = 0 + 𝑣 = 𝑣 and 𝜆𝑣 = 0 if 𝜆 ∈ 𝐑 is nonzero.
(2) Linear independence: A (finite) set of vectors, 𝑣1 , …, 𝑣𝑘 ∈ 𝑉 is linearly independent if
the map
(1.1.2) 𝐑𝑘 → 𝑉 , (𝑐1 , …, 𝑐𝑘 ) ↦ 𝑐1 𝑣1 + ⋯ + 𝑐𝑘 𝑣𝑘
is injective.
(3) A (finite) set of vectors 𝑣1 , …, 𝑣𝑘 ∈ 𝑉 spans 𝑉 if the map (1.1.2) is surjective.
(4) Vectors 𝑣1 , …, 𝑣𝑘 ∈ 𝑉 are a basis of 𝑉 if they span 𝑉 and are linearly independent; in
other words, if the map (1.1.2) is bijective. This means that every vector 𝑣 can be written
uniquely as a sum
𝑛
(1.1.3) 𝑣 = ∑ 𝑐𝑖 𝑣𝑖 .
𝑖=1
(5) If 𝑉 is a vector space with a basis 𝑣1 , …, 𝑣𝑘 , then 𝑉 is said to be finite dimensional, and 𝑘
is the dimension of 𝑉. (It is a theorem that this definition is legitimate: every basis has to
have the same number of vectors.) In this chapter all the vector spaces we’ll encounter
will be finite dimensional. We write dim(𝑉) for the dimension of 𝑉.
(6) A subset 𝑈 ⊂ 𝑉 is a subspace if it is vector space in its own right, i.e., for all 𝑣, 𝑣1 , 𝑣2 ∈ 𝑈
and 𝜆 ∈ 𝐑, both 𝜆𝑣 and 𝑣1 + 𝑣2 are in 𝑈 (where the addition and scalar multiplication
is given as vectors of 𝑉).
(7) Let 𝑉 and 𝑊 be vector spaces. A map 𝐴 ∶ 𝑉 → 𝑊 is linear if for 𝑣, 𝑣1 , 𝑣2 ∈ 𝑈 and 𝜆 ∈ 𝐑
we have
(1.1.4) 𝐴(𝜆𝑣) = 𝜆𝐴𝑣
1Such a course is a prerequisite for reading this text.
1
Draft: March 28, 2018
and
𝐴(𝑣1 + 𝑣2 ) = 𝐴𝑣1 + 𝐴𝑣2 .
(8) Let 𝐴 ∶ 𝑉 → 𝑊 be a linear map of vector spaces. The kernel of 𝐴 is the subset of 𝑉
defined by
ker(𝐴) ≔ { 𝑣 ∈ 𝑉 | 𝐴(𝑣) = 0 } ,
i.e, is the set of vectors in 𝑉 which 𝐴 sends to the zero vector in 𝑊. By equation (1.1.4)
and item 7, the subset ker(𝐴) is a subspace of 𝑉.
(9) By (1.1.4) and (7) the set-theoretic image im(𝐴)of 𝐴 is a subspace of 𝑊. We call im(𝐴)
the image 𝐴 of 𝐴. The following is an important rule for keeping track of the dimensions
of ker 𝐴 and im 𝐴:
(1.1.5) dim(𝑉) = dim(ker(𝐴)) + dim(im(𝐴)) .
(10) Linear mappings and matrices: Let 𝑣1 , …, 𝑣𝑛 be a basis of 𝑉 and 𝑤1 , …, 𝑤𝑚 a basis of 𝑊.
Then by (1.1.3) 𝐴𝑣𝑗 can be written uniquely as a sum,
𝑚
(1.1.6) 𝐴𝑣𝑗 = ∑ 𝑐𝑖,𝑗 𝑤𝑖 , 𝑐𝑖,𝑗 ∈ 𝐑 .
𝑖=1
The 𝑚 × 𝑛 matrix of real numbers, [𝑐𝑖,𝑗 ], is the matrix associated with 𝐴. Conversely,
given such an 𝑚 × 𝑛 matrix, there is a unique linear map 𝐴 with the property (1.1.6).
(11) An inner product on a vector space is a map
𝐵∶ 𝑉 × 𝑉 → 𝐑
with the following three properties:
(a) Bilinearity: For vectors 𝑣, 𝑣1 , 𝑣2 , 𝑤 ∈ 𝑉 and 𝜆 ∈ 𝐑 we have
𝐵(𝑣1 + 𝑣2 , 𝑤) = 𝐵(𝑣1 , 𝑤) + 𝐵(𝑣2 , 𝑤)
and
𝐵(𝜆𝑣, 𝑤) = 𝜆𝐵(𝑣, 𝑤) .
(b) Symmetry: For vectors 𝑣, 𝑤 ∈ 𝑉 we have 𝐵(𝑣, 𝑤) = 𝐵(𝑤, 𝑣).
(c) Positivity: For every vector 𝑣 ∈ 𝑉 we have 𝐵(𝑣, 𝑣) ≥ 0. Moreover, if 𝑣 ≠ 0 then
𝐵(𝑣, 𝑣) > 0.
Remark 1.1.7. Notice that by property (11b), property (11a) is equivalent to
𝐵(𝑤, 𝜆𝑣) = 𝜆𝐵(𝑤, 𝑣)
and
𝐵(𝑤, 𝑣1 + 𝑣2 ) = 𝐵(𝑤, 𝑣1 ) + 𝐵(𝑤, 𝑣2 ) .
Example 1.1.8. The map (1.1.2) is a linear map. The vectors 𝑣1 , …, 𝑣𝑘 span 𝑉 if its image of
this map is 𝑉, and the 𝑣1 , …, 𝑣𝑘 are linearly independent if the kernel of this map is the zero
vector in 𝐑𝑘 .
The items on the list above are just a few of the topics in linear algebra that we’re assum-
ing our readers are familiar with. We have highlighted them because they’re easy to state.
However, understanding them requires a heavy dollop of that indefinable quality “mathe-
matical sophistication”, a quality which will be in heavy demand in the next few sections of
this chapter. We will also assume that our readers are familiar with a number of more low-
brow linear algebra notions: matrix multiplication, row and column operations on matrices,
transposes of matrices, determinants of 𝑛 × 𝑛 matrices, inverses of matrices, Cramer’s rule,
recipes for solving systems of linear equations, etc. (See [10, §§1.1 & 1.2] for a quick review
of this material.)
Draft: March 28, 2018
It is easy to see that the operations on 𝑉/𝑊 from Definition 1.2.3 are well-defined. For
instance, suppose 𝑣1 + 𝑊 = 𝑣′1 + 𝑊 and 𝑣2 + 𝑊 = 𝑣′2 + 𝑊. Then 𝑣1 − 𝑣′1 and 𝑣2 − 𝑣′2 are in
𝑊, so
(𝑣1 + 𝑣2 ) − (𝑣′1 + 𝑣′2 ) ∈ 𝑊 ,
and hence (𝑣1 + 𝑣2 ) + 𝑊 = (𝑣′1 + 𝑣′2 ) + 𝑊.
Definition 1.2.6. Let 𝑉 be a vector space and 𝑊 ⊂ 𝑉 a subspace. Define a map quotient
map
(1.2.7) 𝜋 ∶ 𝑉 → 𝑉/𝑊
by setting 𝜋(𝑣) ≔ 𝑣 + 𝑊. It is clear from equations (1.2.4) and (1.2.5) that 𝜋 is linear, and
that it maps 𝑉 surjectively onto 𝑉/𝑊.
Observation 1.2.8. Note that the zero vector in the vector space 𝑉/𝑊 is the zero coset
0 + 𝑊 = 𝑊. Hence 𝑣 ∈ ker(𝜋) if and only if 𝑣 + 𝑊 = 𝑊, i.e., 𝑣 ∈ 𝑊. In other words,
ker(𝜋) = 𝑊.
In the Definitions 1.2.3 and 1.2.6, 𝑉 and 𝑊 do not have to be finite dimensional, but if
they are then by equation (1.1.5) we have
dim(𝑉/𝑊) = dim(𝑉) − dim(𝑊) .
We leave the following easy proposition as an exercise.
Proposition 1.2.9. Let 𝐴 ∶ 𝑉 → 𝑈 a linear map of vector spaces. If 𝑊 ⊂ ker(𝐴) there exists
a unique linear map 𝐴♯ ∶ 𝑉/𝑊 → 𝑈 with the property that 𝐴 = 𝐴♯ ∘ 𝜋, where 𝜋 ∶ 𝑉 → 𝑉/𝑊
is the quotient map.
The dual space of a vector space
Definition 1.2.10. Let 𝑉 be a vector space. Write 𝑉⋆ the set of all linear functions ℓ ∶ 𝑉 → 𝐑.
Define a vector space structure on 𝑉⋆ as follows: if ℓ1 , ℓ2 ∈ 𝑉⋆ , then define the sum ℓ1 + ℓ2
to be the map 𝑉 → 𝐑 given by
(ℓ1 + ℓ2 )(𝑣) ≔ ℓ1 (𝑣) + ℓ2 (𝑣) ,
which is clearly is linear. If ℓ ∈ 𝑉⋆ is a linear function and 𝜆 ∈ 𝐑, define 𝜆ℓ by
(𝜆ℓ)(𝑣) ≔ 𝜆 ⋅ ℓ(𝑣) ;
then 𝜆ℓ is clearly linear.
The vector space 𝑉⋆ is called the dual space of 𝑉.
Suppose 𝑉 is 𝑛-dimensional, and let 𝑒1 , …, 𝑒𝑛 be a basis of 𝑉. Then every vector 𝑣 ∈ 𝑉
can be written uniquely as a sum
𝑣 = 𝑐1 𝑒1 + ⋯ + 𝑐𝑛 𝑒𝑛 , 𝑐𝑖 ∈ 𝐑 .
Let
(1.2.11) 𝑒𝑖⋆ (𝑣) = 𝑐𝑖 .
If 𝑣 = 𝑐1 𝑒1 + ⋯ + 𝑐𝑛 𝑒𝑛 and 𝑣′ = 𝑐1′ 𝑒1 + ⋯ + 𝑐𝑛′ 𝑒𝑛 then 𝑣 + 𝑣′ = (𝑐1 + 𝑐1′ )𝑒1 + ⋯ + (𝑐𝑛 + 𝑐𝑛′ )𝑒𝑛 , so
𝑒𝑖⋆ (𝑣 + 𝑣′ ) = 𝑐𝑖 + 𝑐𝑖′ = 𝑒𝑖⋆ (𝑣) + 𝑒𝑖⋆ (𝑣′ ) .
This shows that 𝑒𝑖⋆ (𝑣) is a linear function of 𝑣 and hence 𝑒𝑖⋆ ∈ 𝑉⋆ .
Claim 1.2.12. If 𝑉 is an 𝑛-dimensional vector space with basis 𝑒1 , …, 𝑒𝑛 , then 𝑒1⋆ , …, 𝑒𝑛⋆ is a
basis of 𝑉⋆ .
Draft: March 28, 2018
Claim 1.2.15. The linear map 𝐴 is defined, in terms of 𝑓1⋆ , …, 𝑓𝑚⋆ and 𝑒1⋆ , …, 𝑒𝑛⋆ by the trans-
⋆
Exercise 1.2.ix. Let 𝑊 be a subspace of a finite dimensional vector space, 𝑉. From the
inclusion map, 𝜄 ∶ 𝑊⟂ → 𝑉⋆ , one gets a transpose map,
𝜄⋆ ∶ (𝑉⋆ )⋆ → (𝑊⟂ )⋆
and, by composing this with (1.2.16), a map
𝜄⋆ ∘ ev ∶ 𝑉 → (𝑊⟂ )⋆ .
Show that this map is onto and that its kernel is 𝑊. Conclude from Exercise 1.2.viii that
there is a natural bijective linear map
𝜈 ∶ 𝑉/𝑊 → (𝑊⟂ )⋆
with the property 𝜈 ∘ 𝜋 = 𝜄⋆ ∘ ev. In other words, 𝑉/𝑊 and (𝑊⟂ )⋆ are two descriptions of the
same object. (This shows that the “quotient space” operation and the “dual space” operation
are closely related.)
Exercise 1.2.x. Let 𝑉1 and 𝑉2 be vector spaces and 𝐴 ∶ 𝑉1 → 𝑉2 a linear map. Verify that for
the transpose map 𝐴⋆ ∶ 𝑉2⋆ → 𝑉1⋆ we have:
ker(𝐴⋆ ) = im(𝐴)⟂
and
im(𝐴⋆ ) = ker(𝐴)⟂ .
Exercise 1.2.xi. Let 𝑉 be a vector space.
(1) Let 𝐵 ∶ 𝑉 × 𝑉 → 𝐑 be an inner product on 𝑉. For 𝑣 ∈ 𝑉 let
ℓ𝑣 ∶ 𝑉 → 𝐑
be the function: ℓ𝑣 (𝑤) = 𝐵(𝑣, 𝑤). Show that ℓ𝑣 is linear and show that the map
(1.2.17) 𝐿 ∶ 𝑉 → 𝑉⋆ , 𝑣 ↦ ℓ𝑣
is a linear mapping.
(2) If 𝑉 is finite dimensional, prove that 𝐿 bijective. Conclude that if 𝑉 has an inner product
one gets from it a natural identification of 𝑉 with 𝑉⋆ .
Hint: Since dim 𝑉 = dim 𝑉⋆ it suffices by equation (1.1.5) to show that ker(𝐿) = 0.
Now note that if 𝑣 ≠ 0 ℓ𝑣 (𝑣) = 𝐵(𝑣, 𝑣) is a positive number.
Exercise 1.2.xii. Let 𝑉 be an 𝑛-dimensional vector space and 𝐵 ∶ 𝑉 × 𝑉 → 𝐑 an inner
product on 𝑉. A basis, 𝑒1 , …, 𝑒𝑛 of 𝑉 is orthonormal if
1, 𝑖=𝑗
(1.2.18) 𝐵(𝑒𝑖 , 𝑒𝑗 ) = {
0, 𝑖≠𝑗.
(1) Show that an orthonormal basis exists.
Hint: By induction let 𝑒1 , …, 𝑒𝑘 be vectors with the property (1.2.18) and let 𝑣 be a
vector which is not a linear combination of these vectors. Show that the vector
𝑘
𝑤 = 𝑣 − ∑ 𝐵(𝑒𝑖 , 𝑣)𝑒𝑖
𝑖=1
1
is nonzero and is orthogonal to the 𝑒𝑖 ’s. Now let 𝑒𝑘+1 = 𝜆𝑤, where 𝜆 = 𝐵(𝑤, 𝑤)− 2 .
Draft: March 28, 2018
(2) Let 𝑒1 , …𝑒𝑛 and 𝑒1′ , …𝑒𝑛′ be two orthonormal bases of 𝑉 and let
𝑛
𝑒𝑗′ = ∑ 𝑎𝑖,𝑗 𝑒𝑖 .
𝑖=1
Show that
𝑛
1, 𝑗 = 𝑘
(1.2.19) ∑ 𝑎𝑖,𝑗 𝑎𝑖,𝑘 = {
𝑖=1 0, 𝑗 ≠ 𝑘 .
(3) Let 𝐴 be the matrix [𝑎𝑖,𝑗 ]. Show that equation (1.2.19) can be written more compactly
as the matrix identity
(1.2.20) 𝐴𝐴⊺ = id𝑛
where id𝑛 is the 𝑛 × 𝑛 identity matrix.
(4) Let 𝑒1 , …, 𝑒𝑛 be an orthonormal basis of 𝑉 and 𝑒1⋆ , …, 𝑒𝑛⋆ the dual basis of 𝑉⋆ . Show that
the mapping (1.2.17) is the mapping, 𝐿𝑒𝑖 = 𝑒𝑖⋆ , 𝑖 = 1, …𝑛.
1.3. Tensors
Definition 1.3.1. Let 𝑉 be an 𝑛-dimensional vector space and let 𝑉𝑘 be the set of all 𝑘-tuples
(𝑣1 , …, 𝑣𝑘 ), where 𝑣1 , …, 𝑣𝑘 ∈ 𝑉, that is, the 𝑘-fold direct sum of 𝑉 with itself. A function
𝑇 ∶ 𝑉𝑘 → 𝐑
is said to be linear in its 𝑖th variable if, when we fix vectors, 𝑣1 , …, 𝑣𝑖−1 , 𝑣𝑖+1 , …, 𝑣𝑘 , the map
𝑉 → 𝐑 defined by
(1.3.2) 𝑣 ↦ 𝑇(𝑣1 , …, 𝑣𝑖−1 , 𝑣, 𝑣𝑖+1 , …, 𝑣𝑘 )
is linear.
If 𝑇 is linear in its 𝑖th variable for 𝑖 = 1, …, 𝑘 it is said to be 𝑘-linear, or alternatively
is said to be a 𝑘-tensor. We write L𝑘 (𝑉) for the set of all 𝑘-tensors in 𝑉. We will agree that
0-tensors are just the real numbers, that is L0 (𝑉) ≔ 𝐑.
Let 𝑇1 , 𝑇2 ∶ 𝑉𝑘 → 𝐑 be linear functions. It is clear from (1.3.2) that if 𝑇1 and 𝑇2 are
𝑘-linear, so is 𝑇1 + 𝑇2 . Similarly if 𝑇 is 𝑘-linear and 𝜆 is a real number, 𝜆𝑇 is 𝑘-linear. Hence
L𝑘 (𝑉) is a vector space. Note that for 𝑘 = 1, “𝑘-linear” just means ‘linear”, so L1 (𝑉) = 𝑉⋆ .
Definition 1.3.3. Let 𝑛 and 𝑘 be positive integers. A multi-index of 𝑛 of length 𝑘 is a 𝑘-tuple
𝐼 = (𝑖1 , …, 𝑖𝑘 ) of integers with 1 ≤ 𝑖𝑟 ≤ 𝑛 for 𝑟 = 1, …, 𝑘.
Example 1.3.4. Let 𝑛 be a positive integer. The multi-indices of 𝑛 of length 2 are in bijection
the square of pairs of integers integers
(𝑖, 𝑗) , 1 ≤ 𝑖, 𝑗 ≤ 𝑛 ,
2
and there are exactly 𝑛 of them.
We leave the following generalization as an easy exercise.
Lemma 1.3.5. Let 𝑛 and 𝑘 be positive integers. There are exactly 𝑛𝑘 multi-indices of 𝑛 of length
𝑘.
Now fix a basis 𝑒1 , …, 𝑒𝑛 of 𝑉. For 𝑇 ∈ L𝑘 (𝑉) write
(1.3.6) 𝑇𝐼 ≔ 𝑇(𝑒𝑖1 , …, 𝑒𝑖𝑘 )
for every multi-index 𝐼 of length 𝑘.
Draft: March 28, 2018
§1.3 Tensors 9
Proposition 1.3.7. The real numbers 𝑇𝐼 determine 𝑇, i.e., if 𝑇 and 𝑇′ are 𝑘-tensors and 𝑇𝐼 =
𝑇𝐼′ for all 𝐼, then 𝑇 = 𝑇′ .
Proof. By induction on 𝑛. For 𝑛 = 1 we proved this result in § 1.1. Let’s prove that if this
assertion is true for 𝑛 − 1, it is true for 𝑛. For each 𝑒𝑖 let 𝑇𝑖 be the (𝑘 − 1)-tensor
(𝑣1 , …, 𝑣𝑛−1 ) ↦ 𝑇(𝑣1 , …, 𝑣𝑛−1 , 𝑒𝑖 ) .
Then for 𝑣 = 𝑐1 𝑒1 + ⋯ 𝑐𝑛 𝑒𝑛
𝑛
𝑇(𝑣1 , …, 𝑣𝑛−1 , 𝑣) = ∑ 𝑐𝑖 𝑇𝑖 (𝑣1 , …, 𝑣𝑛−1 ) ,
𝑖=1
A tensor of the form (1.3.11) is called a decomposable 𝑘-tensor. These tensors, as we will
see, play an important role in what follows. In particular, let 𝑒1 , …, 𝑒𝑛 be a basis of 𝑉 and
𝑒1⋆ , …, 𝑒𝑛⋆ the dual basis of 𝑉⋆ . For every multi-index 𝐼 of length 𝑘 let
𝑒𝐼⋆ = 𝑒𝑖⋆1 ⊗ ⋯ ⊗ 𝑒𝑖⋆𝑘 .
Then if 𝐽 is another multi-index of length 𝑘,
1, 𝐼 = 𝐽
(1.3.12) 𝑒𝐼⋆ (𝑒𝑗1 , …, 𝑒𝑗𝑘 ) = {
0, 𝐼 ≠ 𝐽
by (1.2.13), (1.3.10) and (1.3.11). From (1.3.12) it is easy to conclude:
Theorem 1.3.13. Let 𝑉 be a vector space with basis 𝑒1 , …, 𝑒𝑛 and 0 ≤ 𝑘 ≤ 𝑛 be an integer.
The 𝑘-tensors 𝑒𝐼⋆ of (1.3.12) are a basis of L𝑘 (𝑉).
for constants, 𝐶𝐼 ∈ 𝐑. Then by (1.3.14) with 𝑇′ = 0, 𝐶𝐽 = 0, so the 𝑒𝐼⋆ ’s are linearly indepen-
dent. □
As we noted in Lemma 1.3.5, there are exactly 𝑛𝑘 multi-indices of length 𝑘 and hence
𝑛 basis vectors in the set {𝑒𝐼⋆ }𝐼 , so we have proved
𝑘
Permutations
Definition 1.4.1. Let 𝛴𝑘 be the 𝑘-element set 𝛴𝑘 ≔ {1, 2, …, 𝑘}. A permutation of order
𝑘 is a bijecton 𝜎 ∶ 𝛴𝑘 ⥲ 𝛴𝑘 . Given two permutations 𝜎1 and 𝜎2 , their product 𝜎1 𝜎2 is the
composition of 𝜎1 ∘ 𝜎2 , i.e., the map,
𝑖 ↦ 𝜎1 (𝜎2 (𝑖)) .
For every permutation 𝜎, one denotes by 𝜎−1 the inverse permutation given by the inverse
bijection of 𝜎, i.e., defined by
𝜎(𝑖) ≔ 𝑗 ⟺ 𝜎−1 (𝑗) = 𝑖 .
Let 𝑆𝑘 be the set of all permutations of order 𝑘. One calls 𝑆𝑘 the permutation group of 𝛴𝑘 or,
alternatively, the symmetric group on 𝑘 letters.
It is easy to check the following.
Lemma 1.4.2. The group 𝑆𝑘 has 𝑘! elements.
Definition 1.4.3. Let 𝑘 be a positive integer. For every 1 ≤ 𝑖 < 𝑗 ≤ 𝑘, let 𝜏𝑖,𝑗 be the permu-
tation
{𝑗, ℓ = 𝑖
{
𝜏𝑖,𝑗 (ℓ) ≔ {𝑖, ℓ = 𝑗
{
{ℓ, ℓ ≠ 𝑖, 𝑗 .
The permuation 𝜏𝑖,𝑗 is called a transposition, and if 𝑗 = 𝑖 + 1, then 𝜏𝑖,𝑗 is called an elementary
transposition.
Theorem 1.4.4. Every permutation in 𝑆𝑘 can be written as a product of (a finite number of)
transpositions.
Proof. We prove this by induction on 𝑘. The base case when 𝑘 = 2 is obvious.
For the induction step, suppose that we know the claim for 𝑆𝑘−1 . Given 𝜎 ∈ 𝑆𝑘 , we
have 𝜎(𝑘) = 𝑖 if and only if 𝜏𝑖,𝑘 𝜎(𝑘) = 𝑘. Thus 𝜏𝑖,𝑘 𝜎 is, in effect, a permutation of 𝛴𝑘−1 . By
induction, 𝜏𝑖,𝑘 𝜎 can be written as a product of transpositions, so
𝜎 = 𝜏𝑖,𝑘 (𝜏𝑖,𝑘 𝜎)
can be written as a product of transpositions. □
Theorem 1.4.5. Every transposition can be written as a product of elementary transpositions.
Proof. Let 𝜏 = 𝜏𝑖𝑗 , 𝑖 < 𝑗. With 𝑖 fixed, argue by induction on 𝑗. Note that for 𝑗 > 𝑖 + 1
𝜏𝑖𝑗 = 𝜏𝑗−1,𝑗 𝜏𝑖,𝑗−1 𝜏𝑗−1,𝑗 .
Now apply the inductive hypothesis to 𝜏𝑖,𝑗−1 . □
Corollary 1.4.6. Every permutation can be written as a product of elementary transpositions.
The sign of a permutation
Definition 1.4.7. Let 𝑥1 , …, 𝑥𝑘 be the coordinate functions on 𝐑𝑘 . For 𝜎 ∈ 𝑆𝑘 we define
𝑥𝜎(𝑖) − 𝑥𝜎(𝑗)
(1.4.8) (−1)𝜎 ≔ ∏ .
𝑖<𝑗 𝑥𝑖 − 𝑥𝑗
Notice that the numerator and denominator in (1.4.8) are identical up to sign. Indeed,
if 𝑝 = 𝜎(𝑖) < 𝜎(𝑗) = 𝑞, the term, 𝑥𝑝 − 𝑥𝑞 occurs once and just once in the numerator and
once and just once in the denominator; and if 𝑞 = 𝜎(𝑖) > 𝜎(𝑗) = 𝑝, the term, 𝑥𝑝 − 𝑥𝑞 , occurs
Draft: March 28, 2018
once and just once in the numerator and its negative, 𝑥𝑞 − 𝑥𝑝 , once and just once in the
numerator. Thus
(−1)𝜎 = ±1 .
In light of this, we call (−1)𝜎 the sign of 𝜎.
Claim 1.4.9. For 𝜎, 𝜏 ∈ 𝑆𝑘 we have
(−1)𝜎𝜏 = (−1)𝜎 (−1)𝜏 .
That is, the sign defines a group homomorphism 𝑆𝑘 → {±1}.
Proof. By definition,
𝑥𝜎𝜏(𝑖) − 𝑥𝜎𝜏(𝑗)
(−1)𝜎𝜏 = ∏ .
𝑖<𝑗 𝑥𝑖 − 𝑥𝑗
We write the right hand side as a product of
𝑥𝜏(𝑖) − 𝑥𝜏(𝑗)
∏ = (−1)𝜏
𝑖<𝑗 𝑥𝑖 − 𝑥𝑗
and
𝑥𝜎𝜏(𝑖) − 𝑥𝜎𝜏(𝑗)
(1.4.10) ∏
𝑖<𝑗 𝑥𝜏(𝑖) − 𝑥𝜏(𝑗)
For 𝑖 < 𝑗, let 𝑝 = 𝜏(𝑖) and 𝑞 = 𝜏(𝑗) when 𝜏(𝑖) < 𝜏(𝑗) and let 𝑝 = 𝜏(𝑗) and 𝑞 = 𝜏(𝑖) when
𝜏(𝑗) < 𝜏(𝑖). Then
𝑥𝜎𝜏(𝑖) − 𝑥𝜎𝜏(𝑗) 𝑥𝜎(𝑝) − 𝑥𝜎(𝑞)
=
𝑥𝜏(𝑖) − 𝑥𝜏(𝑗) 𝑥𝑝 − 𝑥𝑞
(i.e., if 𝜏(𝑖) < 𝜏(𝑗), the numerator and denominator on the right equal the numerator and
denominator on the left and, if 𝜏(𝑗) < 𝜏(𝑖) are negatives of the numerator and denominator
on the left). Thus equation (1.4.10) becomes
𝑥𝜎(𝑝) − 𝑥𝜎(𝑞)
∏ = (−1)𝜎 . □
𝑝<𝑞 𝑥 𝑝 − 𝑥𝑞
We’ll leave for you to check that if 𝜏 is a transposition, (−1)𝜏 = −1 and to conclude from
this:
Proposition 1.4.11. If 𝜎 is the product of an odd number of transpositions, (−1)𝜎 = −1 and
if 𝜎 is the product of an even number of transpositions (−1)𝜎 = +1.
Alternation
Definition 1.4.12. Let 𝑉 be an 𝑛-dimensional vector space and 𝑇 ∈ L𝑘 (𝑣) a 𝑘-tensor. For
𝜎 ∈ 𝑆𝑘 , define 𝑇𝜎 ∈ L𝑘 (𝑉) to be the 𝑘-tensor
(1.4.13) 𝑇𝜎 (𝑣1 , …, 𝑣𝑘 ) ≔ 𝑇(𝑣𝜎−1 (1) , …, 𝑣𝜎−1 (𝑘) ) .
Proposition 1.4.14.
(1) If 𝑇 = ℓ1 ⊗ ⋯ ⊗ ℓ𝑘 , ℓ𝑖 ∈ 𝑉⋆ , then 𝑇𝜎 = ℓ𝜎(1) ⊗ ⋯ ⊗ ℓ𝜎(𝑘) .
(2) The assignment 𝑇 ↦ 𝑇𝜎 is a linear map L𝑘 (𝑉) → L𝑘 (𝑉).
(3) If 𝜎, 𝜏 ∈ 𝑆𝑘 , we have 𝑇𝜎𝜏 = (𝑇𝜎 )𝜏 .
Draft: March 28, 2018
But as 𝜏 runs over 𝑆𝑘 , 𝜏𝜎 runs over 𝑆𝑘 , and hence the right hand side is (−1)𝜎 Alt(𝑇).
To prove (2), note that if 𝑇 ∈ A 𝑘 (𝑉)
Alt 𝑇 = ∑ (−1)𝜏 𝑇𝜏 = ∑ (−1)𝜏 (−1)𝜏 𝑇 = 𝑘!𝑇 .
𝜏∈𝑆𝑘 𝜏∈𝑆𝑘
Draft: March 28, 2018
We will use this alternation operation to construct a basis for A 𝑘 (𝑉). First, however, we
require some notation.
Definition 1.4.18. Let 𝐼 = (𝑖1 , …, 𝑖𝑘 ) be a multi-index of length 𝑘.
(1) 𝐼 is repeating if 𝑖𝑟 = 𝑖𝑠 for some 𝑟 ≠ 𝑠.
(2) 𝐼 is strictly increasing if 𝑖1 < 𝑖2 < ⋯ < 𝑖𝑟 .
(3) For 𝜎 ∈ 𝑆𝑘 , write 𝐼𝜎 ≔ (𝑖𝜎(1) , …, 𝑖𝜎(𝑘) ).
Remark 1.4.19. If 𝐼 is non-repeating there is a unique 𝜎 ∈ 𝑆𝑘 so that 𝐼𝜎 is strictly increasing.
Let 𝑒1 , …, 𝑒𝑛 be a basis of 𝑉 and let
𝑒𝐼⋆ = 𝑒𝑖⋆1 ⊗ ⋯ ⊗ 𝑒𝑖⋆𝑘
and
𝜓𝐼 = Alt(𝑒𝐼⋆ ) .
Proposition 1.4.20.
(1) 𝜓𝐼𝜎 = (−1)𝜎 𝜓𝐼 .
(2) If 𝐼 is repeating, 𝜓𝐼 = 0.
(3) If 𝐼 and 𝐽 are strictly increasing,
1, 𝐼 = 𝐽
𝜓𝐼 (𝑒𝑗1 , …, 𝑒𝑗𝑘 ) = {
0, 𝐼 ≠ 𝐽 .
Proof. To prove (1) we note that (𝑒𝐼⋆ )𝜎 = 𝑒𝐼⋆𝜎 ; so
Alt(𝑒𝐼⋆𝜎 ) = Alt(𝑒𝐼⋆ )𝜎 = (−1)𝜎 Alt(𝑒𝐼⋆ ) .
To prove (2), suppose 𝐼 = (𝑖1 , …, 𝑖𝑘 ) with 𝑖𝑟 = 𝑖𝑠 for 𝑟 ≠ 𝑠. Then if 𝜏 = 𝜏𝑖𝑟 ,𝑖𝑠 , 𝑒𝐼⋆ = 𝑒𝐼⋆𝜏 so
𝜓𝐼 = 𝜓𝐼𝜏 = (−1)𝜏 𝜓𝐼 = −𝜓𝐼 .
To prove (3), note that by definition
𝜓𝐼 (𝑒𝑗1 , …, 𝑒𝑗𝑘 ) = ∑(−1)𝜏 𝑒𝐼⋆𝜏 (𝑒𝑗1 , …, 𝑒𝑗𝑘 ) .
𝜏
But by (1.3.12)
1, 𝐼𝜏 = 𝐽
(1.4.21) 𝑒𝐼⋆𝜏 (𝑒𝑗1 , …, 𝑒𝑗𝑘 ) = {
0, 𝐼𝜏 ≠ 𝐽 .
Thus if 𝐼 and 𝐽 are strictly increasing, 𝐼𝜏 is strictly increasing if and only if 𝐼𝜏 = 𝐼, and
(1.4.21) is nonzero if and only if 𝐼 = 𝐽. □
By (1.4.23) the 𝜓𝐼 ’s, 𝐼 strictly increasing, are a spanning set of vectors for A 𝑘 (𝑉), and by
(1.4.25) they are linearly independent, so we have proved:
Proposition 1.4.26. The alternating tensors, 𝜓𝐼 , 𝐼 strictly increasing, are a basis for A 𝑘 (𝑉).
Thus dim A 𝑘 (𝑉) is equal to the number of strictly increasing multi-indices 𝐼 of length 𝑘.
We leave for you as an exercise to show that this number is equal to the binomal coefficient
𝑛 𝑛!
(1.4.27) ( )≔
𝑘 (𝑛 − 𝑘)!𝑘!
if 1 ≤ 𝑘 ≤ 𝑛. □
Hint: Show that every strictly increasing multi-index of length 𝑘 determines a 𝑘 element
subset of {1, …, 𝑛} and vice-versa.
Note also that if 𝑘 > 𝑛 every multi-index
𝐼 = (𝑖1 , …, 𝑖𝑘 )
of length 𝑘 has to be repeating: 𝑖𝑟 = 𝑖𝑠 for some 𝑟 ≠ 𝑠 since the 𝑖𝑝 ’s lie on the interval 1 ≤ 𝑖 ≤ 𝑛.
Thus by Proposition 1.4.17
𝜓𝐼 = 0
for all multi-indices of length 𝑘 > 0 and
A 𝑘 (𝑉) = 0 .
Exercises for §1.4
Exercise 1.4.i. Show that there are exactly 𝑘! permutations of order 𝑘.
Hint: Induction on 𝑘: Let 𝜎 ∈ 𝑆𝑘 , and let 𝜎(𝑘) = 𝑖, 1 ≤ 𝑖 ≤ 𝑘. Show that 𝜏𝑖,𝑘 𝜎 leaves 𝑘
fixed and hence is, in effect, a permutation of 𝛴𝑘−1 .
Exercise 1.4.ii. Prove that if 𝜏 ∈ 𝑆𝑘 is a transposition, (−1)𝜏 = −1 and deduce from this
Proposition 1.4.11.
Exercise 1.4.iii. Prove assertion 2 in Proposition 1.4.14.
Exercise 1.4.iv. Prove that dim A 𝑘 (𝑉) is given by (1.4.27).
Exercise 1.4.v. Verify that for 𝑖 < 𝑗 − 1
𝜏𝑖,𝑗 = 𝜏𝑗−1,𝑗 𝜏𝑖,𝑗−1 , 𝜏𝑗−1,𝑗 .
Draft: March 28, 2018
Exercise 1.4.vi. For 𝑘 = 3 show that every one of the six elements of 𝑆3 is either a transpo-
sition or can be written as a product of two transpositions.
Exercise 1.4.vii. Let 𝜎 ∈ 𝑆𝑘 be the “cyclic” permutation
𝜎(𝑖) ≔ 𝑖 + 1 , 𝑖 = 1, …, 𝑘 − 1
and 𝜎(𝑘) ≔ 1. Show explicitly how to write 𝜎 as a product of transpositions and compute
(−1)𝜎 .
Hint: Same hint as in Exercise 1.4.i.
Exercise 1.4.viii. In Exercise 1.3.vii show that if 𝑇 is in A 𝑘 (𝑉), 𝑇𝑣 is in A 𝑘−1 (𝑉). Show in
addition that for 𝑣, 𝑤 ∈ 𝑉 and 𝑇 ∈ A 𝑘 (𝑉) we have (𝑇𝑣 )𝑤 = −(𝑇𝑤 )𝑣 .
Exercise 1.4.ix. Let 𝐴 ∶ 𝑉 → 𝑊 be a linear mapping. Show that if 𝑇 is in A 𝑘 (𝑊), 𝐴⋆ 𝑇 is in
A 𝑘 (𝑉).
Exercise 1.4.x. In Exercise 1.4.ix show that if 𝑇 is in L𝑘 (𝑊) then Alt(𝐴⋆ 𝑇) = 𝐴⋆ (Alt(𝑇)),
i.e., show that the Alt operation commutes with the pullback operation.
and the terms on the right are redundant, and hence in I2 (𝑉). Next let 𝑘 be arbitrary and
𝜎 = 𝜏𝑖,𝑖+1 . If 𝑇1 = ℓ1 ⊗ ⋯ ⊗ ℓ𝑖−2 and 𝑇2 = ℓ𝑖+2 ⊗ ⋯ ⊗ ℓ𝑘 . Then
𝑇 − (−1)𝜎 𝑇 = 𝑇1 ⊗ (ℓ𝑖 ⊗ ℓ𝑖+1 + ℓ𝑖+1 ⊗ ℓ𝑖 ) ⊗ 𝑇2
is in I𝑘 (𝑉) by Proposition 1.5.3 and the computation above.
The general case: By Theorem 1.4.5, 𝜎 can be written as a product of 𝑚 elementary transpo-
sitions, and we’ll prove (1.5.5) by induction on 𝑚.
We have just dealt with the case 𝑚 = 1.
The induction step: the“𝑚 − 1” case implies the “𝑚” case. Let 𝜎 = 𝜏𝛽 where 𝛽 is a product of
𝑚 − 1 elementary transpositions and 𝜏 is an elementary transposition. Then
𝑇𝜎 = (𝑇 𝛽 )𝜏 = (−1)𝜏 𝑇 𝛽 + ⋯
= (−1)𝜏 (−1)𝛽 𝑇 + ⋯
= (−1)𝜎 𝑇 + ⋯
where the “dots” are elements of I𝑘 (𝑉), and the induction hypothesis was used in the second
line. □
Corollary 1.5.6. If 𝑇 ∈ L𝑘 (𝑉), then
(1.5.7) Alt(𝑇) = 𝑘!𝑇 + 𝑊 ,
𝑘
where 𝑊 is in I (𝑉).
Proof. By definition Alt(𝑇) = ∑𝜎∈𝑆𝑘 (−1)𝜎 𝑇𝜎 , and by Proposition 1.5.4, 𝑇𝜎 = (−1)𝜎 𝑇 + 𝑊𝜎 ,
with 𝑊𝜎 ∈ I𝑘 (𝑉). Thus
Alt(𝑇) = ∑ (−1)𝜎 (−1)𝜎 𝑇 + ∑ (−1)𝜎 𝑊𝜎 = 𝑘!𝑇 + 𝑊
𝜎∈𝑆𝑘 𝜎∈𝑆𝑘
𝜎
where 𝑊 = ∑𝜎∈𝑆𝑘 (−1) 𝑊𝜎 . □
Exercise 1.5.iii. Show that if 𝑇 is a symmetric 𝑘-tensor, then for 𝑘 ≥ 2, then 𝑇 is in I𝑘 (𝑉).
Hint: Let 𝜎 be a transposition and deduce from the identity, 𝑇𝜎 = 𝑇, that 𝑇 has to be in
the kernel of Alt.
Exercise 1.5.iv (a warning). In general 𝒮 𝑘 (𝑉) ≠ I𝑘 (𝑉). Show, however, that if 𝑘 = 2 these
two spaces are equal.
Exercise 1.5.v. Show that if ℓ ∈ 𝑉⋆ and 𝑇 ∈ L𝑘−2 (𝑉), then ℓ ⊗ 𝑇 ⊗ ℓ is in I𝑘 (𝑉).
Exercise 1.5.vi. Show that if ℓ1 and ℓ2 are in 𝑉⋆ and 𝑇 ∈ L𝑘−2 (𝑉), then ℓ1 ⊗𝑇⊗ℓ2 +ℓ2 ⊗𝑇⊗ℓ1
is in I𝑘 (𝑉).
Exercise 1.5.vii. Given a permutation 𝜎 ∈ 𝑆𝑘 and 𝑇 ∈ I𝑘 (𝑉), show that 𝑇𝜎 ∈ I𝑘 (𝑉).
Exercise 1.5.viii. Let W(𝑉) be a subspace of L𝑘 (𝑉) having the following two properties.
(1) For 𝑆 ∈ 𝒮 2 (𝑉) and 𝑇 ∈ L𝑘−2 (𝑉), 𝑆 ⊗ 𝑇 is in W(𝑉).
Draft: March 28, 2018
Prove that this operation has properties (2)–(4) of Proposition 1.4.17 and, as a substitute
for (1), has the property: Sym(𝑇)𝜎 = Sym(𝑇).
(2) Let 𝜙𝐼 = Sym(𝑒𝐼⋆ ), 𝑒𝐼⋆ = 𝑒𝑖⋆1 ⊗ ⋯ ⊗ 𝑒𝑖⋆𝑛 . Prove that { 𝜙𝐼 | 𝐼 is non-decreasing } form a basis
of 𝑆𝑘 (𝑉).
(3) Conclude that dim(𝒮 𝑘 (𝑉)) is equal to the number of non-decreasing multi-indices of
length 𝑘: 1 ≤ 𝑖1 ≤ 𝑖2 ≤ ⋯ ≤ ℓ𝑘 ≤ 𝑛.
(4) Compute this number by noticing that the assignment
(𝑖1 , …, 𝑖𝑘 ) ↦ (𝑖1 + 0, 𝑖2 + 1, …, 𝑖𝑘 + 𝑘 − 1)
is a bijection between the set of these non-decreasing multi-indicesand the set of in-
creasing multi-indices 1 ≤ 𝑗1 < ⋯ < 𝑗𝑘 ≤ 𝑛 + 𝑘 − 1.
Proof. For every 𝑇 ∈ L𝑘 (𝑉), 𝑇 = (−1)𝜎 𝑇 + 𝑊 for some 𝑊 ∈ I𝑘 (𝑉) by Proposition 1.5.4.
Therefore since 𝜋(𝑊) = 0
𝜋(𝑇𝜎 ) = (−1)𝜎 𝜋(𝑇) .
𝜎
In particular, if 𝑇 = ℓ1 ⊗ ⋯ ⊗ ℓ𝑘 , 𝑇 = ℓ𝜎(1) ⊗ ⋯ ⊗ ℓ𝜎(𝑘) , so
𝜋(𝑇𝜎 ) = ℓ𝜎(1) ∧ ⋯ ∧ ℓ𝜎(𝑘) = (−1)𝜎 𝜋(𝑇)
= (−1)𝜎 ℓ1 ∧ ⋯ ∧ ℓ𝑘 . □
In particular, for ℓ1 and ℓ2 ∈ 𝑉⋆
(1.6.9) ℓ1 ∧ ℓ2 = −ℓ2 ∧ ℓ1
and for ℓ1 , ℓ2 and ℓ3 ∈ 𝑉⋆
ℓ1 ∧ ℓ2 ∧ ℓ3 = −ℓ2 ∧ ℓ1 ∧ ℓ3 = ℓ2 ∧ ℓ3 ∧ ℓ1 .
More generally, it is easy to deduce from equation (1.6.8) the following result (which we’ll
leave as an exercise).
Draft: March 28, 2018
(1.6.11) 𝜔1 ∧ 𝜔2 = (−1)𝑟𝑠 𝜔2 ∧ 𝜔1 .
Hint: It suffices to prove this for decomposable elements i.e., for 𝜔1 = ℓ1 ∧ ⋯ ∧ ℓ𝑟 and
𝜔2 = ℓ1′ ∧ ⋯ ∧ ℓ𝑠′ . Now make 𝑟𝑠 applications of (1.6.9).
Let 𝑒1 , …, 𝑒𝑛 be a basis of 𝑉 and let 𝑒1⋆ , …, 𝑒𝑛⋆ be the dual basis of 𝑉⋆ . For every multi-
index 𝐼 of length 𝑘,
Theorem 1.6.13. The elements (1.6.12), with 𝐼 strictly increasing, are basis vectors of 𝛬𝑘 (𝑉⋆ ).
Proof. The elements 𝜓𝐼 = Alt(𝑒𝐼⋆ ), for 𝐼 strictly increasing, are basis vectors of A 𝑘 (𝑉) by
Proposition 1.4.26; so their images, 𝜋(𝜓𝐼 ), are a basis of 𝛬𝑘 (𝑉⋆ ). But
= ∑ (−1) (−1) 𝜎 𝜎
𝜋(𝑒𝐼⋆ ) = 𝑘!𝜋(𝑒𝐼⋆ ) . □
𝜎∈𝑆𝑘
Exercise 1.6.ii. Verify the multiplication law in equation (1.6.11) for the wedge product.
Exercise 1.6.iii. Given 𝜔 ∈ 𝛬𝑟 (𝑉⋆ ) let 𝜔𝑘 be the 𝑘-fold wedge product of 𝜔 with itself, i.e.,
let 𝜔2 = 𝜔 ∧ 𝜔, 𝜔3 = 𝜔 ∧ 𝜔 ∧ 𝜔, etc.
(1) Show that if 𝑟 is odd then for 𝑘 > 1, 𝜔𝑘 = 0.
(2) Show that if 𝜔 is decomposable, then for 𝑘 > 1, 𝜔𝑘 = 0.
Hint: As in freshman calculus, prove this binomial theorem by induction using the iden-
tity: ( 𝑘ℓ ) = ( 𝑘−1
ℓ−1
) + ( 𝑘−1
ℓ
).
on the 𝑘 − 1-tuple of vectors, 𝑣1 , …, 𝑣𝑘−1 , i.e., in the 𝑟th summand on the right, 𝑣 gets in-
serted between 𝑣𝑟−1 and 𝑣𝑟 . (In particular the first summand is 𝑇(𝑣, 𝑣1 , …, 𝑣𝑘−1 ) and the last
summand is (−1)𝑘−1 𝑇(𝑣1 , …, 𝑣𝑘−1 , 𝑣).)
It is clear from the definition that if 𝑣 = 𝑣1 + 𝑣2
(1.7.2) 𝜄𝑣 𝑇 = 𝜄 𝑣 1 𝑇 + 𝜄 𝑣 2 𝑇 ,
and if 𝑇 = 𝑇1 + 𝑇2
(1.7.3) 𝜄𝑣 𝑇 = 𝜄𝑣 𝑇1 + 𝜄𝑣 𝑇2 ,
and we will leave for you to verify by inspection the following two lemmas.
Lemma 1.7.4. If 𝑇 is the decomposable 𝑘-tensor ℓ1 ⊗ ⋯ ⊗ ℓ𝑘 then
𝑘
(1.7.5) 𝜄𝑣 𝑇 = ∑ (−1)𝑟−1 ℓ𝑟 (𝑣)ℓ1 ⊗ ⋯ ⊗ ℓ𝑟̂ ⊗ ⋯ ⊗ ℓ𝑘 ,
𝑟=1
where the “hat” over ℓ𝑟 means that ℓ𝑟 is deleted from the tensor product.
Lemma 1.7.6. If 𝑇1 ∈ L𝑝 (𝑉) and 𝑇2 ∈ L𝑞 (𝑉)
(1.7.7) 𝜄𝑣 (𝑇1 ⊗ 𝑇2 ) = 𝜄𝑣 𝑇1 ⊗ 𝑇2 + (−1)𝑝 𝑇1 ⊗ 𝜄𝑣 𝑇2 .
We will next prove the important identity.
Lemma 1.7.8. Let 𝑉 be a vector space and 𝑇 ∈ L𝑘 (𝑉). Then for all 𝑣 ∈ 𝑉 we have
(1.7.9) 𝜄𝑣 (𝜄𝑣 𝑇) = 0 .
Proof. It suffices by linearity to prove this for decomposable tensors and since (1.7.9) is
trivially true for 𝑇 ∈ L1 (𝑉), we can by induction assume (1.7.9) is true for decomposable
tensors of degree 𝑘 − 1. Let ℓ1 ⊗ ⋯ ⊗ ℓ𝑘 be a decomposable tensor of degree 𝑘. Setting
𝑇 ≔ ℓ1 ⊗ ⋯ ⊗ ℓ𝑘−1 and ℓ = ℓ𝑘 we have
𝜄𝑣 (ℓ1 ⊗ ⋯ ⊗ ℓ𝑘 ) = 𝜄𝑣 (𝑇 ⊗ ℓ) = 𝜄𝑣 𝑇 ⊗ ℓ + (−1)𝑘−1 ℓ(𝑣)𝑇
by (1.7.7). Hence
𝜄𝑣 (𝜄𝑣 (𝑇 ⊗ ℓ)) = 𝜄𝑣 (𝜄𝑣 𝑇) ⊗ ℓ + (−1)𝑘−2 ℓ(𝑣)𝜄𝑣 𝑇 + (−1)𝑘−1 ℓ(𝑣)𝜄𝑣 𝑇 .
But by induction the first summand on the right is zero and the two remaining summands
cancel each other out. □
We’ll now show how to define the operation 𝜄𝑣 on 𝛬𝑘 (𝑉⋆ ). We’ll first prove:
Lemma 1.7.11. If 𝑇 ∈ L𝑘 (𝑉) is redundant then so is 𝜄𝑣 𝑇.
Proof. Let 𝑇 = 𝑇1 ⊗ ℓ ⊗ ℓ ⊗ 𝑇2 where ℓ is in 𝑉⋆ , 𝑇1 is in L𝑝 (𝑉) and 𝑇2 is in L𝑞 (𝑉). Then by
equation (1.7.7)
𝜄𝑣 𝑇 = 𝜄𝑣 𝑇1 ⊗ ℓ ⊗ ℓ ⊗ 𝑇2 + (−1)𝑝 𝑇1 ⊗ 𝜄𝑣 (ℓ ⊗ ℓ) ⊗ 𝑇2 + (−1)𝑝+2 𝑇1 ⊗ ℓ ⊗ ℓ ⊗ 𝜄𝑣 𝑇2 .
However, the first and the third terms on the right are redundant and
𝜄𝑣 (ℓ ⊗ ℓ) = ℓ(𝑣)ℓ − ℓ(𝑣)ℓ
by equation (1.7.5). □
Now let 𝜋 be the projection (1.5.12) of L𝑘 (𝑉) onto 𝛬𝑘 (𝑉⋆ ) and for 𝜔 = 𝜋(𝑇) ∈ 𝛬𝑘 (𝑉⋆ )
define
(1.7.12) 𝜄𝑣 𝜔 = 𝜋(𝜄𝑣 𝑇) .
To show that this definition is legitimate we note that if 𝜔 = 𝜋(𝑇1 ) = 𝜋(𝑇2 ), then 𝑇1 − 𝑇2 ∈
I𝑘 (𝑉), so by Lemma 1.7.11 𝜄𝑣 𝑇1 − 𝜄𝑣 𝑇2 ∈ I𝑘−1 and hence
𝜋(𝜄𝑣 𝑇1 ) = 𝜋(𝜄𝑣 𝑇2 ) .
Therefore, (1.7.12) does not depend on the choice of 𝑇.
By definition, 𝜄𝑣 is a linear map 𝛬𝑘 (𝑉⋆ ) → 𝛬𝑘−1 (𝑉⋆ ). We call this the interior product
operation. From the identities (1.7.2)–(1.7.12) one gets, for 𝑣, 𝑣1 , 𝑣2 ∈ 𝑉, 𝜔 ∈ 𝛬𝑘 (𝑉⋆ ), 𝜔1 ∈
𝛬𝑝 (𝑉⋆ ), and 𝜔2 ∈ 𝛬2 (𝑉⋆ )
(1.7.13) 𝜄(𝑣1 +𝑣2 ) 𝜔 = 𝜄𝑣1 𝜔 + 𝜄𝑣2 𝜔
(1.7.14) 𝜄𝑣 (𝜔1 ∧ 𝜔2 ) = 𝜄𝑣 𝜔1 ∧ 𝜔2 + (−1)𝑝 𝜔1 ∧ 𝜄𝑣 𝜔2
(1.7.15) 𝜄𝑣 (𝜄𝑣 𝜔) = 0
and
𝜄𝑣1 𝜄𝑣2 𝜔 = −𝜄𝑣2 𝜄𝑣1 𝜔 .
Moreover if 𝜔 = ℓ1 ∧ ⋯ ∧ ℓ𝑘 is a decomposable element of 𝛬𝑘 (𝑉⋆ ) one gets from (1.7.5)
𝑘
𝜄𝑣 𝜔 = ∑ (−1)𝑟−1 ℓ𝑟 (𝑣)ℓ1 ∧ ⋯ ∧ ℓ𝑟̂ ∧ ⋯ ∧ ℓ𝑘 .
𝑟=1
In particular if 𝑒1 , …, 𝑒𝑛 is a basis of 𝑉, 𝑒1⋆ , …, 𝑒𝑛⋆ the dual basis of 𝑉⋆ and 𝜔𝐼 = 𝑒𝑖⋆1 ∧ ⋯ ∧ 𝑒𝑖⋆𝑘 ,
1 ≤ 𝑖1 < ⋯ < 𝑖𝑘 ≤ 𝑛, then 𝜄𝑒𝑗 𝜔𝐼 = 0 if 𝑗 ∉ 𝐼 and if 𝑗 = 𝑖𝑟
Now let 𝜔 be an element of 𝛬𝑘 (𝑊⋆ ) and let 𝜔 = 𝜋(𝑇) where 𝑇 is in L𝑘 (𝑊). We define
(1.8.2) 𝐴⋆ 𝜔 = 𝜋(𝐴⋆ 𝑇) .
Claim 1.8.3. The left hand side of equation (1.8.2) is well-defined.
Proposition 1.8.8. Write id𝑉 ∶ 𝑉 → 𝑉 for the identity map. Then det(id𝑉 ) = 1.
We’ll leave the proof as an exercise. As a hint, note that id⋆𝑉 is the identity map on
𝛬 (𝑉⋆ ).
𝑛
Proof. Let 𝑊 be the image of 𝐴. Then if 𝐴 is not onto, the dimension of 𝑊 is less than 𝑛,
so 𝛬𝑛 (𝑊⋆ ) = 0. Now let 𝐴 = 𝑖𝑊 𝐵 where 𝑖𝑊 is the inclusion map of 𝑊 into 𝑉 and 𝐵 is the
mapping, 𝐴, regarded as a mapping from 𝑉 to 𝑊. Thus if 𝜔 is in 𝛬𝑛 (𝑉⋆ ), then by (2)
𝐴⋆ 𝜔 = 𝐵⋆ 𝑖⋆𝑊 𝜔
and since 𝑖⋆𝑊 𝜔 is in 𝛬𝑛 (𝑊⋆ ) it is zero. □
We will derive by wedge product arguments the familiar “matrix formula” for the deter-
minant. Let 𝑉 and 𝑊 be 𝑛-dimensional vector spaces and let 𝑒1 , …, 𝑒𝑛 be a basis for 𝑉 and
𝑓1 , …, 𝑓𝑛 a basis for 𝑊. From these bases we get dual bases, 𝑒1⋆ , …, 𝑒𝑛⋆ and 𝑓1⋆ , …, 𝑓𝑛⋆ , for 𝑉⋆
and 𝑊⋆ . Moreover, if 𝐴 is a linear map of 𝑉 into 𝑊 and [𝑎𝑖,𝑗 ] the 𝑛 × 𝑛 matrix describing
𝐴 in terms of these bases, then the transpose map, 𝐴⋆ ∶ 𝑊⋆ → 𝑉⋆ , is described in terms of
these dual bases by the 𝑛 × 𝑛 transpose matrix, i.e., if
𝑛
𝐴𝑒𝑗 = ∑ 𝑎𝑖,𝑗 𝑓𝑖 ,
𝑖=1
then
𝑛
𝐴⋆ 𝑓𝑗⋆ = ∑ 𝑎𝑗,𝑖 𝑒𝑖⋆ .
𝑖=1
(See §1.2.) Consider now 𝐴⋆ (𝑓1⋆ ∧ ⋯ ∧ 𝑓𝑛⋆ ). By (1),
𝐴⋆ (𝑓1⋆ ∧ ⋯ ∧ 𝑓𝑛⋆ ) = 𝐴⋆ 𝑓1⋆ ∧ ⋯ ∧ 𝐴⋆ 𝑓𝑛⋆
= ∑ (𝑎1,𝑘1 𝑒𝑘⋆1 ) ∧ ⋯ ∧ (𝑎𝑛,𝑘𝑛 𝑒𝑘⋆𝑛 ) .
1≤𝑘1 ,…,𝑘𝑛 ≤𝑛
Thus,
𝐴⋆ (𝑓1⋆ ∧ ⋯ ∧ 𝑓𝑛⋆ ) = ∑ 𝑎1,𝑘1 …𝑎𝑛,𝑘𝑛 𝑒𝑘⋆1 ∧ ⋯ ∧ 𝑒𝑘⋆𝑛 .
If the multi-index, 𝑘1 , …, 𝑘𝑛 , is repeating, then 𝑒𝑘⋆1 ∧ ⋯ ∧ 𝑒𝑘⋆𝑛 is zero, and if it is not repeating
then we can write
𝑘𝑖 = 𝜎(𝑖) 𝑖 = 1, …, 𝑛
for some permutation, 𝜎, and hence we can rewrite 𝐴⋆ (𝑓1⋆ ∧ ⋯ ∧ 𝑓𝑛⋆ ) as
𝐴⋆ (𝑓1⋆ ∧ ⋯ ∧ 𝑓𝑛⋆ ) = ∑ 𝑎1,𝜎(1) ⋯ 𝑎𝑛,𝜎(𝑛) (𝑒1⋆ ∧ ⋯ ∧ 𝑒𝑛⋆ )𝜎 .
𝜎∈𝑆𝑛
But
(𝑒1⋆ ∧ ⋯ ∧ 𝑒𝑛⋆ )𝜎 = (−1)𝜎 𝑒1⋆ ∧ ⋯ ∧ 𝑒𝑛⋆
so we get finally the formula
(1.8.10) 𝐴⋆ (𝑓1⋆ ∧ ⋯ ∧ 𝑓𝑛⋆ ) = det([𝑎𝑖,𝑗 ])𝑒1⋆ ∧ ⋯ ∧ 𝑒𝑛⋆
where
(1.8.11) det([𝑎𝑖,𝑗 ]) = ∑ (−1)𝜎 𝑎1,𝜎(1) ⋯ 𝑎𝑛,𝜎(𝑛)
𝜎∈𝑆𝑛
summed over 𝜎 ∈ 𝑆𝑛 . The sum on the right is (as most of you know) the determinant of the
matric [𝑎𝑖,𝑗 ].
Notice that if 𝑉 = 𝑊 and 𝑒𝑖 = 𝑓𝑖 , 𝑖 = 1, …, 𝑛, then 𝜔 = 𝑒1⋆ ∧ ⋯ ∧ 𝑒𝑛⋆ = 𝑓1⋆ ∧ ⋯ ∧ 𝑓𝑛⋆ ,
hence by (1.8.6) and (1.8.10),
det(𝐴) = det[𝑎𝑖,𝑗 ] .
Draft: March 28, 2018
and
∑ (−1)𝜎 𝑎𝜎(1),1 …𝑎𝜎(𝑛),𝑛
𝜎∈𝑆𝑛
are the same. Show that the second sum is identical with
−1
∑ (−1)𝜎 𝑎𝜎−1 (1),1 …𝑎𝜎−1 (𝑛),𝑛 .
𝜎∈𝑆𝑛
§1.9 Orientations 29
1.9. Orientations
Definition 1.9.1. We recall from freshman calculus that if ℓ ⊂ 𝐑2 is a line through the
origin, then ℓ ∖ {0} has two connected components and an orientation of ℓ is a choice of
one of these components (as in the Figure 1.9.1 below).
More generally, if 𝐿 is a one-dimensional vector space then 𝐿 ∖ {0} consists of two com-
ponents: namely if 𝑣 is an element of 𝐿 ∖ {0}, then these two components are
𝐿1 = { 𝜆𝑣 | 𝜆 > 0 }
and
𝐿2 = { 𝜆𝑣 | 𝜆 < 0 } .
Definition 1.9.2. Let 𝐿 be a one-dimensional vector space. An orientation of 𝐿 is a choice of
one of a connected component of 𝐿 ∖ {0}. Usually the component chosen is denoted 𝐿+ , and
called the positive component of 𝐿∖{0} and the other component 𝐿− the negative component
of 𝐿 ∖ {0}.
Definition 1.9.3. Let (𝐿, 𝐿+ ) be an oriented one-dimensional vector space. A vector 𝑣 ∈ 𝐿
is positively oriented if 𝑣 ∈ 𝐿+ .
Definition 1.9.4. Let 𝑉 be an 𝑛-dimensional vector space. An orientation of 𝑉 is an orien-
tation of the one-dimensional vector space 𝛬𝑛 (𝑉⋆ ).
One important way of assigning an orientation to 𝑉 is to choose a basis, 𝑒1 , …, 𝑒𝑛 of 𝑉.
Then, if 𝑒1⋆ , …, 𝑒𝑛⋆ is the dual basis, we can orient 𝛬𝑛 (𝑉⋆ ) by requiring that 𝑒1⋆ ∧ ⋯ ∧ 𝑒𝑛⋆ be
in the positive component of 𝛬𝑛 (𝑉⋆ ).
Definition 1.9.5. Let 𝑉 be an oriented 𝑛-dimensional vector space. We say that an ordered
basis (𝑒1 , …, 𝑒𝑛 ) of 𝑉 is positively oriented if 𝑒1⋆ ∧ ⋯ ∧ 𝑒𝑛⋆ is in the positive component of
𝛬𝑛 (𝑉⋆ ).
Suppose that 𝑒1 , …, 𝑒𝑛 and 𝑓1 , …, 𝑓𝑛 are bases of 𝑉 and that
𝑛
(1.9.6) 𝑒𝑗 = ∑ 𝑎𝑖,𝑗, 𝑓𝑖 .
𝑖=1
Then by (1.7.10)
𝑓1⋆ ∧ ⋯ ∧ 𝑓𝑛⋆ = det[𝑎𝑖,𝑗 ]𝑒1⋆ ∧ ⋯ ∧ 𝑒𝑛⋆
so we conclude:
Draft: March 28, 2018
As a special case of this theorem suppose dim 𝑊 = 𝑛 − 1. Then the choice of a vector
𝑣 ∈ 𝑉 ∖ 𝑊 gives one a basis vector 𝜋(𝑣) for the one-dimensional space 𝑉/𝑊 and hence if 𝑉
is oriented, the choice of 𝑣 gives one a natural orientation on 𝑊.
Definition 1.9.12. Let 𝐴 ∶ 𝑉1 → 𝑉2 a bijective linear map of oriented 𝑛-dimensional vector
spaces. We say that 𝐴 is orientation-preserving if, for 𝜔 ∈ 𝛬𝑛 (𝑉2⋆ )+ , we have that 𝐴⋆ 𝜔 is in
𝛬𝑛 (𝑉1⋆ )+ .
Example 1.9.13. If 𝑉1 = 𝑉2 then 𝐴⋆ 𝜔 = det(𝐴)𝜔 so 𝐴 is orientation preserving if and only
if det(𝐴) > 0.
The following proposition we’ll leave as an exercise.
Proposition 1.9.14. Let 𝑉1 , 𝑉2 , and 𝑉3 be oriented 𝑛-dimensional vector spaces and 𝐴 𝑖 ∶ 𝑉𝑖 ⥲
𝑉𝑖+1 for 𝑖 = 1, 2 be bijective linear maps. Then if 𝐴 1 and 𝐴 2 are orientation preserving, so is
𝐴2 ∘ 𝐴1 .
Draft: March 28, 2018
§1.9 Orientations 31
CHAPTER 2
Differential Forms
The goal of this chapter is to generalize to 𝑛 dimensions the basic operations of three
dimensional vector calculus: divergence, curl, and gradient. The divergence and gradient
operations have fairly straightforward generalizations, but the curl operation is more subtle.
For vector fields it does not have any obvious generalization, however, if one replaces vector
fields by a closely related class of objects, differential forms, then not only does it have a
natural generalization but it turns out that divergence, curl, and gradient are all special cases
of a general operation on differential forms called exterior differentiation.
We leave the following generalization of the product rule for differentiation as an exer-
cise.
Lemma 2.1.11. Let 𝑈 be an open subset of 𝐑𝑛 , 𝒗 a vector field on 𝑈, and 𝑓1 , 𝑓2 ∈ 𝐶1 (𝑈). Then
𝐿𝒗 (𝑓1 ⋅ 𝑓2 ) = 𝐿𝒗 (𝑓1 ) ⋅ 𝑓2 + 𝑓1 ⋅ 𝐿𝒗 (𝑓2 ) .
We now discuss a class of objects which are in some sense “dual objects” to vector fields.
Definition 2.1.12. Let 𝑝 ∈ 𝐑𝑛 . The cotangent space to 𝐑𝑛 at 𝑝 is the dual vector space
𝑇𝑝⋆ 𝐑𝑛 ≔ (𝑇𝑝 𝐑)⋆ .
An element of 𝑇𝑝⋆ 𝐑𝑛 is called a cotangent vector to 𝐑𝑛 at 𝑝.
Definition 2.1.13. Let 𝑈 be an open subset of 𝐑𝑛 . A differential one-form, or simply one-
form on 𝑈 is a function 𝜔 which assigns to each point 𝑝 ∈ 𝑈 a vector 𝜔𝑝 in 𝑇𝑝⋆ 𝐑𝑛 .
Examples 2.1.14.
(1) Let 𝑓 ∶ 𝑈 → 𝐑 be a 𝐶1 function. Then for 𝑝 ∈ 𝑈 and 𝑞 = 𝑓(𝑝) one has a linear map
(2.1.15) 𝑑𝑓𝑝 ∶ 𝑇𝑝 𝐑𝑛 → 𝑇𝑞 𝐑
and by making the identification 𝑇𝑞 𝐑 ⥲ 𝐑 by sending (𝑞, 𝑣) ↦ 𝑣, the differential 𝑑𝑓𝑝
can be regarded as a linear map from 𝑇𝑝 𝐑𝑛 to 𝐑, i.e., as an element of 𝑇𝑝⋆ 𝐑𝑛 . Hence the
assignment
𝑝 ↦ 𝑑𝑓𝑝
defines a one-form on 𝑈 which we denote by 𝑑𝑓.
(2) Given a one-form 𝜔 and a function 𝜙 ∶ 𝑈 → 𝐑 the pointwise product of 𝜙 with 𝜔 defines
one-form 𝜙𝜔 on 𝑈 by (𝜙𝜔)𝑝 ≔ 𝜙(𝑝)𝜔𝑝 .
(3) Given two one-forms 𝜔1 and 𝜔2 their pointwise sum defines a one-form 𝜔1 + 𝜔2 by
(𝜔1 + 𝜔2 )𝑝 ≔ (𝜔1 )𝑝 + (𝜔2 )𝑝 .
(4) The one-forms 𝑑𝑥1 , …, 𝑑𝑥𝑛 play a particularly important role. By equation (2.1.15)
𝜕
(2.1.16) (𝑑𝑥𝑖 ) ( ) = 𝛿𝑖,𝑗
𝜕𝑥𝑗 𝑝
i.e., is equal to 1 if 𝑖 = 𝑗 and zero if 𝑖 ≠ 𝑗. Thus (𝑑𝑥1 )𝑝 , …, (𝑑𝑥𝑛 )𝑝 are the basis of
𝑇𝑝⋆ 𝐑𝑛 dual to the basis (𝜕/𝜕𝑥𝑖 )𝑝 . Therefore, if 𝜔 is any one-form on 𝑈, 𝜔𝑝 can be written
uniquely as a sum
𝑛
𝜔𝑝 = ∑ 𝑓𝑖 (𝑝)(𝑑𝑥𝑖 )𝑝 , 𝑓𝑖 (𝑝) ∈ 𝐑 ,
𝑖=1
and 𝜔 can be written uniquely as a sum
𝑛
(2.1.17) 𝜔 = ∑ 𝑓𝑖 𝑑𝑥𝑖
𝑖=1
where 𝑓𝑖 ∶ 𝑈 → 𝐑 is the function 𝑝 ↦ 𝑓𝑖 (𝑝). We say that 𝜔 is a 𝐶∞ one-form or smooth
one-form if the functions 𝑓1 , …, 𝑓𝑛 are 𝐶∞ .
We leave the following as an exercise.
Lemma 2.1.18. Let 𝑈 be an open subset of 𝐑𝑛 . If 𝑓 ∶ 𝑈 → 𝐑 is a 𝐶∞ function, then
𝑛
𝜕𝑓
𝑑𝑓 = ∑ 𝑑𝑥𝑖 .
𝑖=1 𝜕𝑥𝑖
Draft: March 28, 2018
Suppose that 𝒗 is a vector field and 𝜔 a one-form on 𝑈 ⊂ 𝐑𝑛 . Then for every 𝑝 ∈ 𝑈 the
vectors 𝒗(𝑝) ∈ 𝑇𝑝 𝐑𝑛 and 𝜔𝑝 ∈ (𝑇𝑝 𝐑𝑛 )⋆ can be paired to give a number 𝜄𝒗(𝑝) 𝜔𝑝 ∈ 𝐑, and
hence, as 𝑝 varies, an 𝐑-valued function 𝜄𝒗 𝜔 which we will call the interior product of 𝒗 with
𝜔.
Example 2.1.19. For instance if 𝒗 is the vector field (2.1.6) and 𝜔 the one-form (2.1.17),
then
𝑛
𝜄𝒗 𝜔 = ∑ 𝑓𝑖 𝑔𝑖 .
𝑖=1
Thus if 𝒗 and 𝜔 are 𝐶∞ , so is the function 𝜄𝒗 𝜔. Also notice that if 𝜙 ∈ 𝐶∞ (𝑈), then as we
observed above
𝑛
𝜕𝜙 𝜕
𝑑𝜙 = ∑
𝑖=1 𝜕𝑥𝑖 𝜕𝑥𝑖
so if 𝒗 is the vector field (2.1.6) then we have
𝑛
𝜕𝜙
𝜄𝒗 𝑑𝜙 = ∑ 𝑔𝑖 = 𝐿𝒗 𝜙 .
𝑖=1 𝜕𝑥𝑖
Exercises for §2.1
Exercise 2.1.i. Prove Lemma 2.1.18
Exercise 2.1.ii. Prove Lemma 2.1.11
Exercise 2.1.iii. Let 𝑈 be an open subset of 𝐑𝑛 and 𝒗1 and 𝒗2 vector fields on 𝑈. Show that
there is a unique vector field 𝒘, on 𝑈 with the property
𝐿𝒘 𝜙 = 𝐿𝒗1 (𝐿𝒗2 𝜙) − 𝐿𝒗2 (𝐿𝒗1 𝜙)
∞
for all 𝜙 ∈ 𝐶 (𝑈).
Exercise 2.1.iv. The vector field 𝒘 in Exercise 2.1.iii is called the Lie bracket of the vector
fields 𝒗1 and 𝒗2 and is denoted by [𝒗1 , 𝒗2 ]. Verify that the Lie bracket is skew-symmetric, i.e.,
[𝒗1 , 𝒗2 ] = −[𝒗2 , 𝒗1 ] ,
and satisfies the Jacobi identity
[𝒗1 , [𝒗2 , 𝒗3 ]] + [𝒗2 , [𝒗3 , 𝒗1 ]] + [𝒗3 , [𝒗1 , 𝒗2 ]] = 0 .
(And thus defines the structure of a Lie algebra.)
Hint: Prove analogous identities for 𝐿𝒗1 , 𝐿𝒗2 and 𝐿𝒗3 .
𝑛
Exercise 2.1.v. Let 𝒗1 = 𝜕/𝜕𝑥𝑖 and 𝒗2 = ∑𝑗=1 𝑔𝑗 𝜕/𝜕𝑥𝑗 . Show that
𝑛
𝜕𝑔𝑗 𝜕
[𝒗1 , 𝒗2 ] = ∑ .
𝑗=1 𝜕𝑥𝑖 𝜕𝑥𝑗
Exercise 2.1.vi. Let 𝒗1 and 𝒗2 be vector fields and 𝑓 a 𝐶∞ function. Show that
[𝒗1 , 𝑓𝒗2 ] = 𝐿𝒗1 𝑓𝒗2 + 𝑓[𝒗1 , 𝒗2 ] .
Exercise 2.1.vii. Let 𝑈 be an open subset of 𝐑𝑛 and let 𝛾 ∶ [𝑎, 𝑏] → 𝑈, 𝑡 ↦ (𝛾1 (𝑡), …, 𝛾𝑛 (𝑡))
𝑛
be a 𝐶1 curve. Given a 𝐶∞ one-form 𝜔 = ∑𝑖=1 𝑓𝑖 𝑑𝑥𝑖 on 𝑈, define the line integral of 𝜔 over
𝛾 to be the integral
𝑛 𝑏
𝑑𝛾
∫ 𝜔 ≔ ∑ ∫ 𝑓𝑖 (𝛾(𝑡)) 𝑖 𝑑𝑡 .
𝛾 𝑖=1 𝑎 𝑑𝑡
Draft: March 28, 2018
∫ 𝜔 = 𝑓(𝛾(𝑏)) − 𝑓(𝛾(𝑎)) .
𝛾
In particular conclude that if 𝛾 is a closed curve, i.e., 𝛾(𝑎) = 𝛾(𝑏), this integral is zero.
Exercise 2.1.viii. Let 𝜔 be the 𝐶∞ one-form on 𝐑2 ∖ {0} defined by
𝑥 𝑑𝑥 − 𝑥2 𝑑𝑥1
𝜔 = 1 22 ,
𝑥1 + 𝑥22
and let 𝛾 ∶ [0, 2𝜋] → 𝐑2 ∖{0} be the closed curve 𝑡 ↦ (cos 𝑡, sin 𝑡). Compute the line integral
∫𝛾 𝜔, and note that ∫𝛾 𝜔 ≠ 0. Conclude that 𝜔 is not of the form 𝑑𝑓 for 𝑓 ∈ 𝐶∞ (𝐑2 ∖ {0}).
Proof. The identity 𝐿𝒗 𝜙 = 0 implies that 𝜙 is constant on 𝛾(𝑡), but if 𝜙(𝑝) = 𝑐 this implies
that the curve 𝛾(𝑡) lies on the compact subset 𝜙−1 (𝑐) ⊂ 𝑈, hence it cannot “run off to infinity”
as in scenario (2) or “run off the boundary” as in scenario (3). □
supp(𝒗) = { 𝑞 ∈ 𝑈 | 𝒗(𝑞) ≠ 0 } .
We say that 𝒗 is compactly supported if supp(𝒗) is compact.
We now show that compactly supported vector fields are complete.
Theorem 2.2.15. If a vector field 𝒗 is compactly supported, then 𝒗 is complete.
Proof. Notice first that if 𝒗(𝑝) = 0, the constant curve, 𝛾0 (𝑡) = 𝑝, −∞ < 𝑡 < ∞, satisfies the
equation
𝑑
𝛾 (𝑡) = 0 = 𝒗(𝑝) ,
𝑑𝑡 0
so it is an integral curve of 𝒗. Hence if 𝛾(𝑡), −𝑎 < 𝑡 < 𝑏, is any integral curve of 𝒗 with the
property 𝛾(𝑡0 ) = 𝑝 for some 𝑡0 , it has to coincide with 𝛾0 on the interval (−𝑎, 𝑎), and hence
has to be the constant curve 𝛾(𝑡) = 𝑝 on (−𝑎, 𝑎).
Now suppose the support supp(𝒗) of 𝒗 is compact. Then either 𝛾(𝑡) is in supp(𝒗) for all
𝑡 or is in 𝑈 ∖ supp(𝒗) for some 𝑡0 . But if this happens, and 𝑝 = 𝛾(𝑡0 ) then 𝒗(𝑝) = 0, so 𝛾(𝑡)
has to coincide with the constant curve 𝛾0 (𝑡) = 𝑝, for all 𝑡. In neither case can it go off to ∞
or off to the boundary of 𝑈 as 𝑡 → 𝑏. □
One useful application of this result is the following. Suppose 𝒗 is a vector field on
𝑈 ⊂ 𝐑𝑛 , and one wants to see what its integral curves look like on some compact set 𝐴 ⊂ 𝑈.
Let 𝜌 ∈ 𝐶0∞ (𝑈) be a bump function which is equal to one on a neighborhood of 𝐴. Then
the vector field 𝒘 = 𝜌𝒗 is compactly supported and hence complete, but it is identical with
𝒗 on 𝐴, so its integral curves on 𝐴 coincide with the integral curves of 𝒗.
If 𝒗 is complete then for every 𝑝, one has an integral curve 𝛾𝑝 ∶ 𝐑 → 𝑈 with 𝛾𝑝 (0) = 𝑝,
so one can define a map
𝑓𝑡 ∶ 𝑈 → 𝑈
by setting 𝑓𝑡 (𝑝) ≔ 𝛾𝑝 (𝑡). If 𝒗 is 𝐶 , this mapping is 𝐶∞ by the smooth dependence on initial
∞
data theorem, and, by definition, 𝑓0 = id𝑈 . We claim that the 𝑓𝑡 ’s also have the property
(2.2.16) 𝑓𝑡 ∘ 𝑓𝑎 = 𝑓𝑡+𝑎 .
Draft: March 28, 2018
Indeed if 𝑓𝑎 (𝑝) = 𝑞, then by the reparametrization theorem, 𝛾𝑞 (𝑡) and 𝛾𝑝 (𝑡 + 𝑎) are both
integral curves of 𝒗, and since 𝑞 = 𝛾𝑞 (0) = 𝛾𝑝 (𝑎) = 𝑓𝑎 (𝑝), they have the same initial point,
so
𝛾𝑞 (𝑡) = 𝑓𝑡 (𝑞) = (𝑓𝑡 ∘ 𝑓𝑎 )(𝑝)
= 𝛾𝑝 (𝑡 + 𝑎) = 𝑓𝑡+𝑎 (𝑝)
for all 𝑡. Since 𝑓0 is the identity it follows from (2.2.16) that 𝑓𝑡 ∘ 𝑓−𝑡 is the identity, i.e.,
𝑓−𝑡 = 𝑓𝑡−1 ,
so 𝑓𝑡 is a 𝐶∞ diffeomorphism. Hence if 𝒗 is complete it generates a one-parameter group 𝑓𝑡 ,
−∞ < 𝑡 < ∞, of 𝐶∞ -diffeomorphisms of 𝑈.
If 𝒗 is not complete there is an analogous result, but it is trickier to formulate precisely.
Roughly speaking 𝒗 generates a one-parameter group of diffeomorphisms 𝑓𝑡 but these dif-
feomorphisms are not defined on all of 𝑈 nor for all values of 𝑡. Moreover, the identity
(2.2.16) only holds on the open subset of 𝑈 where both sides are well-defined.
We devote the remainder of this section to discussing some “functorial” properties of
vector fields and one-forms.
Definition 2.2.17. Let 𝑈 ⊂ 𝐑𝑛 and 𝑊 ⊂ 𝐑𝑚 be open, and let 𝑓 ∶ 𝑈 → 𝑊 be a 𝐶∞ map. If 𝒗
is a 𝐶∞ -vector field on 𝑈 and 𝒘 a 𝐶∞ -vector field on 𝑊 we say that 𝒗 and 𝒘 are 𝑓-related
if, for all 𝑝 ∈ 𝑈 we have
𝑑𝑓𝑝 (𝒗(𝑝)) = 𝒘(𝑓(𝑝)) .
Writing
𝑛
𝜕
𝒗 = ∑ 𝑣𝑖 , 𝑣𝑖 ∈ 𝐶𝑘 (𝑈)
𝑖=1 𝜕𝑥𝑖
and
𝑚
𝜕
𝒘 = ∑ 𝑤𝑗 , 𝑤𝑗 ∈ 𝐶𝑘 (𝑉)
𝑗=1 𝜕𝑦𝑗
this equation reduces, in coordinates, to the equation
𝑛
𝜕𝑓𝑖
𝑤𝑖 (𝑞) = ∑ (𝑝)𝑣𝑗 (𝑝) .
𝑗=1 𝜕𝑥𝑗
We leave the following assertions as easy exercises (but provide some hints).
Theorem 2.2.20. For 𝑖 = 1, 2, let 𝑈𝑖 be an open subset of 𝐑𝑛𝑖 , 𝒗𝑖 a vector field on 𝑈𝑖 and
𝑓 ∶ 𝑈1 → 𝑈2 a 𝐶∞ -map. If 𝒗1 and 𝒗2 are 𝑓-related, every integral curve
𝛾 ∶ 𝐼 → 𝑈1
of 𝒗1 gets mapped by 𝑓 onto an integral curve 𝑓 ∘ 𝛾 ∶ 𝐼 → 𝑈2 of 𝒗2 .
Proof sketch. Theorem 2.2.20 follows from the chain rule: if 𝑝 = 𝛾(𝑡) and 𝑞 = 𝑓(𝑝)
𝑑 𝑑
𝑑𝑓𝑝 ( 𝛾(𝑡)) = 𝑓(𝛾(𝑡)) . □
𝑑𝑡 𝑑𝑡
Corollary 2.2.21. In the setting of Theorem 2.2.20, suppose 𝒗1 and 𝒗2 are complete. Let
(𝑓𝑖,𝑡 )𝑡∈𝐑 ∶ 𝑈𝑖 ⥲ 𝑈𝑖 be the one-parameter group of diffeomorphisms generated by 𝒗𝑖 . Then
𝑓 ∘ 𝑓1,𝑡 = 𝑓2,𝑡 ∘ 𝑓.
Proof sketch. To deduce Corollary 2.2.21 from Theorem 2.2.20 note that for 𝑝 ∈ 𝑈, 𝑓1,𝑡 (𝑝)
is just the integral curve 𝛾𝑝 (𝑡) of 𝒗1 with initial point 𝛾𝑝 (0) = 𝑝.
The notion of 𝑓-relatedness can be very succinctly expressed in terms of the Lie differ-
entiation operation. For 𝜙 ∈ 𝐶∞ (𝑈2 ) let 𝑓⋆ 𝜙 be the composition, 𝜙 ∘ 𝑓, viewed as a 𝐶∞
function on 𝑈1 , i.e., for 𝑝 ∈ 𝑈1 let 𝑓⋆ 𝜙(𝑝) = 𝜙(𝑓(𝑝)). Then
𝑓⋆ 𝐿𝒗2 𝜙 = 𝐿𝒗1 𝑓⋆ 𝜙 .
To see this, note that if 𝑓(𝑝) = 𝑞 then at the point 𝑝 the right hand side is
(𝑑𝜙)𝑞 ∘ 𝑑𝑓𝑝 (𝒗1 (𝑝))
by the chain rule and by definition the left hand side is
𝑑𝜙𝑞 (𝒗2 (𝑞)) .
Moreover, by definition
𝒗2 (𝑞) = 𝑑𝑓𝑝 (𝒗1 (𝑝))
so the two sides are the same. □
Another easy consequence of the chain rule is:
Theorem 2.2.22. For 𝑖 = 1, 2, 3, let 𝑈𝑖 be an open subset of 𝐑𝑛𝑖 , 𝒗𝑖 a vector field on 𝑈𝑖 , and
for 𝑖 = 1, 2 let 𝑓𝑖 ∶ 𝑈𝑖 → 𝑈𝑖+1 be a 𝐶∞ -map. Suppose that 𝒗1 and 𝒗2 are 𝑓1 -related and that 𝒗2
and 𝒗3 are 𝑓2 -related. Then 𝒗1 and 𝒗3 are (𝑓2 ∘ 𝑓1 )-related.
In particular, if 𝑓1 and 𝑓2 are diffeomorphisms writing 𝒗 ≔ 𝒗1 we have
(𝑓2 )⋆ (𝑓1 )⋆ 𝒗 = (𝑓2 ∘ 𝑓1 )⋆ 𝒗 .
The results we described above have “dual” analogues for one-forms.
Construction 2.2.23. Let 𝑈 ⊂ 𝐑𝑛 and 𝑉 ⊂ 𝐑𝑚 be open, and let 𝑓 ∶ 𝑈 → 𝑉 be a 𝐶∞ -map.
Given a one-form 𝜇 on 𝑉 one can define a pullback one-form 𝑓⋆ 𝜇 on 𝑈 by the following
method. For 𝑝 ∈ 𝑈, by definition 𝜇𝑓(𝑝) is a linear map
𝜇𝑓(𝑝) ∶ 𝑇𝑓(𝑝) 𝐑𝑚 → 𝐑
and by composing this map with the linear map
𝑑𝑓𝑝 ∶ 𝑇𝑝 𝐑𝑛 → 𝑇𝑓(𝑝) 𝐑𝑛
we get a linear map
𝜇𝑓(𝑝) ∘ 𝑑𝑓𝑝 ∶ 𝑇𝑝 𝐑𝑛 → 𝐑 ,
Draft: March 28, 2018
Exercise 2.2.iii. Let 𝒗 be a complete vector field on 𝑈 and 𝑓𝑡 ∶ 𝑈 → 𝑈, the one parameter
group of diffeomorphisms generated by 𝒗. Show that if 𝜙 ∈ 𝐶1 (𝑈)
𝑑 ⋆
𝐿𝒗 𝜙 = 𝑓 𝜙| .
𝑑𝑡 𝑡 𝑡=0
Exercise 2.2.iv.
(1) Let 𝑈 = 𝐑2 and let 𝒗 be the vector field, 𝑥1 𝜕/𝜕𝑥2 − 𝑥2 𝜕/𝜕𝑥1 . Show that the curve
𝑡 ↦ (𝑟 cos(𝑡 + 𝜃), 𝑟 sin(𝑡 + 𝜃)) ,
for 𝑡 ∈ 𝐑, is the unique integral curve of 𝒗 passing through the point, (𝑟 cos 𝜃, 𝑟 sin 𝜃),
at 𝑡 = 0.
𝑛
(2) Let 𝑈 = 𝐑𝑛 and let 𝒗 be the constant vector field: ∑𝑖=1 𝑐𝑖 𝜕/𝜕𝑥𝑖 . Show that the curve
𝑡 ↦ 𝑎 + 𝑡(𝑐1 , …, 𝑐𝑛 ) ,
for 𝑡 ∈ 𝐑, is the unique integral curve of 𝒗 passing through 𝑎 ∈ 𝐑𝑛 at 𝑡 = 0.
Draft: March 28, 2018
𝑛
(3) Let 𝑈 = 𝐑𝑛 and let 𝒗 be the vector field, ∑𝑖=1 𝑥𝑖 𝜕/𝜕𝑥𝑖 . Show that the curve
𝑡 ↦ 𝑒𝑡 (𝑎1 , …, 𝑎𝑛 ) ,
for 𝑡 ∈ 𝐑, is the unique integral curve of 𝒗 passing through 𝑎 at 𝑡 = 0.
Exercise 2.2.v. Show that the following are one-parameter groups of diffeomorphisms:
(1) 𝑓𝑡 ∶ 𝐑 → 𝐑 , 𝑓𝑡 (𝑥) = 𝑥 + 𝑡
(2) 𝑓𝑡 ∶ 𝐑 → 𝐑 , 𝑓𝑡 (𝑥) = 𝑒𝑡 𝑥
(3) 𝑓𝑡 ∶ 𝐑2 → 𝐑2 , 𝑓𝑡 (𝑥, 𝑦) = (𝑥 cos(𝑡) − 𝑦 sin(𝑡), 𝑥 sin(𝑡) + 𝑦 cos(𝑡)).
Exercise 2.2.vi. Let 𝐴 ∶ 𝐑𝑛 → 𝐑𝑛 be a linear mapping. Show that the series
∞
(𝑡𝐴)𝑛 𝑡2 𝑡3
exp(𝑡𝐴) ≔ ∑ = id𝑛 +𝑡𝐴 + 𝐴2 + 𝐴3 + ⋯
𝑛=0 𝑛! 2! 3!
converges and defines a one-parameter group of diffeomorphisms of 𝐑𝑛 .
Exercise 2.2.vii.
(1) What are the infinitesimal generators of the one-parameter groups in Exercise 2.2.v?
(2) Show that the infinitesimal generator of the one-parameter group in Exercise 2.2.vi is
the vector field
𝜕
∑ 𝑎𝑖,𝑗 𝑥𝑗
1≤𝑖,𝑗≤𝑛 𝜕𝑥 𝑖
Exercise 2.3.iv. Let 𝑈 be an open subset of 𝐑𝑛 . Show that every (𝑛 − 1)-form 𝜔 ∈ 𝛺𝑛−1 (𝑈)
can be written uniquely as a sum
𝑛
̂ 𝑖 ∧ ⋯ ∧ 𝑑𝑥𝑛 ,
∑ 𝑓𝑖 𝑑𝑥1 ∧ ⋯ ∧ 𝑑𝑥
𝑖=1
Proof. We prove this by induction on 𝑘. For the base case note that by (3)
(2.4.6) 𝑑(𝑑𝑓1 ) = 0
∞
for every function 𝑓1 ∈ 𝐶 (𝑈).
For the induction step, suppose that we know the result for 𝑘 − 1 functions and that
we are given functions 𝑓1 , …, 𝑓𝑘 ∈ 𝐶∞ (𝑈). Let 𝜇 = 𝑑𝑓2 ∧ ⋯ ∧ 𝑑𝑓𝑘 . Then by the induction
hypothesis 𝑑𝜇 = 0 Combining (2.4.3) with (2.4.6) we see that
𝑑(𝑑𝑓1 ∧ 𝑑𝑓2 ∧ ⋯ ∧ 𝑑𝑓𝑘 ) = 𝑑(𝑑𝑓1 ∧ 𝜇)
= 𝑑(𝑑𝑓1 ) ∧ 𝜇 + (−1) 𝑑𝑓1 ∧ 𝑑𝜇
=0. □
Example 2.4.7. As a special case of Lemma 2.4.5, given a multi-index 𝐼 = (𝑖1 , …, 𝑖𝑘 ) with
1 ≤ 𝑖𝑟 ≤ 𝑛 we have
(2.4.8) 𝑑(𝑑𝑥𝐼 ) = 𝑑(𝑑𝑥𝑖1 ∧ ⋯ ∧ 𝑑𝑥𝑖𝑘 ) = 0 .
Recall that every 𝑘-form 𝜔 ∈ 𝛺𝑘 (𝑈) can be written uniquely as a sum
𝜔 = ∑ 𝑓𝐼 𝑑𝑥𝐼 , 𝑓𝐼 ∈ 𝐶∞ (𝑈)
𝐼
where the multi-indices 𝐼 are strictly increasing. Thus by (2.4.3) and (2.4.8)
(2.4.9) 𝑑𝜔 = ∑ 𝑑𝑓𝐼 ∧ 𝑑𝑥𝐼 .
𝐼
This shows that if there exists an operator 𝑑 ∶ 𝛺∗ (𝑈) → 𝛺∗+1 (𝑈) with Properties 2.4.2, then
𝑑 is necessarily given by the formula (2.4.9). Hence all we have to show is that the operator
defined by this equation (2.4.9) has these properties.
Proposition 2.4.10. Let 𝑈 be an open subset of 𝐑𝑛 . There is a unique operator 𝑑 ∶ 𝛺∗ (𝑈) →
𝛺∗+1 (𝑈) satisfying Properties 2.4.2.
Proof. The property (1) is obvious.
To verify (2) we first note that for 𝐼 strictly increasing equation (2.4.8) is a special case
of equation (2.4.9) (take 𝑓𝐼 = 1 and 𝑓𝐽 = 0 for 𝐽 ≠ 𝐼). Moreover, if 𝐼 is not strictly increasing
it is either repeating, in which case 𝑑𝑥𝐼 = 0, or non-repeating in which case there exists a
permuation 𝜎 ∈ 𝑆𝑘 such that 𝐼𝜎 is strictly increasing. Moreover,
(2.4.11) 𝑑𝑥𝐼 = (−1)𝜎 𝑑𝑥𝐼𝜎 .
Hence (2.4.9) implies (2.4.8) for all multi-indices 𝐼. The same argument shows that for any
sum ∑𝐼 𝑓𝐼 𝑑𝑥𝐼 over multi-indices 𝐼 of length 𝑘, we have
(2.4.12) 𝑑(∑𝐼 𝑓𝐼 𝑑𝑥𝐼 ) = ∑ 𝑑𝑓𝐼 ∧ 𝑑𝑥𝐼 .
𝐼
(As above we can ignore the repeating multi-indices 𝑑𝑥𝐼 = 0 if 𝐼 is repeating, and by (2.4.11)
we can replace the non-repeating multi-indices by strictly increasing multi-indices.)
Suppose now that 𝜔1 ∈ 𝛺𝑘 (𝑈) and 𝜔2 ∈ 𝛺ℓ (𝑈). Write 𝜔1 ≔ ∑𝐼 𝑓𝐼 𝑑𝑥𝐼 and 𝜔2 ≔
∑𝐽 𝑔𝐽 𝑑𝑥𝐽 , where 𝑓𝐼 , 𝑔𝐽 ∈ 𝐶∞ (𝑈). We see that the wedge product 𝜔1 ∧ 𝜔2 is given by
(2.4.13) 𝜔1 ∧ 𝜔2 = ∑ 𝑓𝐼 𝑔𝐽 𝑑𝑥𝐼 ∧ 𝑑𝑥𝐽 ,
𝐼,𝐽
(Notice that if 𝐼 = (𝑖1 , …, 𝑖𝑘 ) and 𝐽 = (𝑗𝑖 , …, 𝑖ℓ ), we have 𝑑𝑥𝐼 ∧ 𝑑𝑥𝐽 = 𝑑𝑥𝐾 , where 𝐾 ≔
(𝑖1 , …, 𝑖𝑘 , 𝑗1 , …, 𝑗ℓ ). Even if 𝐼 and 𝐽 are strictly increasing, 𝐾 is not necessarily strictly in-
creasing. However, in deducing (2.4.14) from (2.4.13) we have observed that this does not
matter.)
Now note that by equation (2.2.8)
𝑑(𝑓𝐼 𝑔𝐽 ) = 𝑔𝐽 𝑑𝑓𝐼 + 𝑓𝐼 𝑑𝑔𝐽 ,
and by the wedge product identities of §1.6,
or
or, finally,
𝑑𝜔1 ∧ 𝜔2 + (−1)𝑘 𝜔1 ∧ 𝑑𝜔2 .
Thus the the operator 𝑑 defined by equation (2.4.9) has (2).
Let is now check that 𝑑 satisfies (3). If 𝜔 = ∑𝐼 𝑓𝐼 𝑑𝑥𝐼 , where 𝑓𝐼 ∈ 𝐶∞ (𝑈), then by
definition, 𝑑𝜔 = ∑𝐼 𝑑𝑓𝐼 ∧ 𝑑𝑥𝐼 and by (2.4.8) and (2.4.3)
so it suffices to check that 𝑑(𝑑𝑓𝐼 ) = 0, i.e., it suffices to check (2.4.6) for zero forms 𝑓 ∈
𝐶∞ (𝑈). However, by (2.2.3) we have
𝑛
𝜕𝑓
𝑑𝑓 = ∑ 𝑑𝑥
𝑗=1 𝜕𝑥𝑗 𝑗
so by equation (2.4.9)
𝑛 𝑛 𝑛
𝜕𝑓 𝜕2 𝑓
𝑑(𝑑𝑓) = ∑ 𝑑 ( ) 𝑑𝑥𝑗 = ∑ (∑ 𝑑𝑥𝑖 ) ∧ 𝑑𝑥𝑗
𝑗=1 𝜕𝑥𝑗 𝑗=1 𝑖=1 𝜕𝑥𝑖 𝜕𝑥𝑗
𝜕2 𝑓
= ∑ 𝑑𝑥𝑖 ∧ 𝑑𝑥𝑗 .
1≤𝑖,𝑗≤𝑛 𝜕𝑥𝑖 𝜕𝑥𝑗
Notice, however, that in this sum, 𝑑𝑥𝑖 ∧ 𝑑𝑥𝑗 = −𝑑𝑥𝑗 ∧ 𝑑𝑥𝑖 and
𝜕2 𝑓 𝜕2 𝑓
=
𝜕𝑥𝑖 𝜕𝑥𝑗 𝜕𝑥𝑗 𝜕𝑥𝑖
so the (𝑖, 𝑗) term cancels the (𝑗, 𝑖) term, and the total sum is zero. □
By (3) every exact form is closed, but the converse is not true even for 1-forms (see
Exercise 2.1.iii). In fact it is a very interesting (and hard) question to determine if an open
set 𝑈 has the following property: For 𝑘 > 0 every closed 𝑘-form is exact.2
Some examples of spaces with this property are described in the exercises at the end of
§2.6. We also sketch below a proof of the following result (and ask you to fill in the details).
Lemma 2.4.16 (Poincaré lemma). If 𝜔 is a closed form on 𝑈 of degree 𝑘 > 0, then for every
point 𝑝 ∈ 𝑈, there exists a neighborhood of 𝑝 on which 𝜔 is exact.
Proof. See Exercises 2.4.v and 2.4.vi. □
Exercises for §2.4
Exercise 2.4.i. Compute the exterior derivatives of the following differential forms.
(1) 𝑥1 𝑑𝑥2 ∧ 𝑑𝑥3
(2) 𝑥1 𝑑𝑥2 − 𝑥2 𝑑𝑥1
𝑛
(3) 𝑒−𝑓 𝑑𝑓 where 𝑓 = ∑𝑖=1 𝑥𝑖2
𝑛
(4) ∑𝑖=1 𝑥𝑖 𝑑𝑥𝑖
𝑛 ̂𝑖 ∧ ⋯ ∧ 𝑑𝑥𝑛
(5) ∑𝑖=1 (−1)𝑖 𝑥𝑖 𝑑𝑥1 ∧ ⋯ ∧ 𝑑𝑥
Exercise 2.4.ii. Solve the equation 𝑑𝜇 = 𝜔 for 𝜇 ∈ 𝛺1 (𝐑3 ), where 𝜔 is the 2-form:
(1) 𝑑𝑥2 ∧ 𝑑𝑥3
(2) 𝑥2 𝑑𝑥2 ∧ 𝑑𝑥3
(3) (𝑥12 + 𝑥22 )𝑑𝑥1 ∧ 𝑑𝑥2
(4) cos(𝑥1 )𝑑𝑥1 ∧ 𝑑𝑥3
Exercise 2.4.iii. Let 𝑈 be an open subset of 𝐑𝑛 .
(1) Show that if 𝜇 ∈ 𝛺𝑘 (𝑈) is exact and 𝜔 ∈ 𝛺ℓ (𝑈) is closed then 𝜇 ∧ 𝜔 is exact.
Hint: Equation (2.4.3).
(2) In particular, 𝑑𝑥1 is exact, so if 𝜔 ∈ 𝛺ℓ (𝑈) is closed 𝑑𝑥1 ∧ 𝜔 = 𝑑𝜇. What is 𝜇?
Exercise 2.4.iv. Let 𝑄 be the rectangle (𝑎1 , 𝑏1 ) × ⋯ × (𝑎𝑛 , 𝑏𝑛 ). Show that if 𝜔 is in 𝛺𝑛 (𝑄),
then 𝜔 is exact.
Hint: Let 𝜔 = 𝑓𝑑𝑥1 ∧ ⋯ ∧ 𝑑𝑥𝑛 with 𝑓 ∈ 𝐶∞ (𝑄) and let 𝑔 be the function
𝑥1
𝑔(𝑥1 , …, 𝑥𝑛 ) = ∫ 𝑓(𝑡, 𝑥2 , …, 𝑥𝑛 )𝑑𝑡 .
𝑎1
Show that 𝜔 = 𝑑(𝑔𝑑𝑥2 ∧ ⋯ ∧ 𝑑𝑥𝑛 ).
Exercise 2.4.v. Let 𝑈 be an open subset of 𝐑𝑛−1 , 𝐴 ⊂ 𝐑 an open interval and (𝑥, 𝑡) product
coordinates on 𝑈 × 𝐴. We say that a form 𝜇 ∈ 𝛺ℓ (𝑈 × 𝐴) is reduced if 𝜇 can be written as a
sum
(2.4.17) 𝜇 = ∑ 𝑓𝐼 (𝑥, 𝑡)𝑑𝑥𝐼 ,
𝐼
(i.e., with no terms involving 𝑑𝑡).
(1) Show that every form, 𝜔 ∈ 𝛺𝑘 (𝑈 × 𝐴) can be written uniquely as a sum:
(2.4.18) 𝜔 = 𝑑𝑡 ∧ 𝛼 + 𝛽
where 𝛼 and 𝛽 are reduced.
2For 𝑘 = 0, 𝑑𝑓 = 0 does not imply that 𝑓 is exact. In fact “exactness” does not make much sense for zero
forms since there are not any “(−1)-forms”. However, if 𝑓 ∈ 𝐶∞ (𝑈) and 𝑑𝑓 = 0 then 𝑓 is constant on connected
components of 𝑈 (see Exercise 2.2.iii).
Draft: March 28, 2018
is effectively a closed 𝑘-form on 𝑈. Prove that if every closed 𝑘-form on 𝑈 is exact, then
every closed 𝑘-form on 𝑈 × 𝐴 is exact.
Hint: Let 𝜔 be a closed 𝑘-form on 𝑈 × 𝐴 and let 𝛾 be the form (2.4.22).
Exercise 2.4.vi. Let 𝑄 ⊂ 𝐑𝑛 be an open rectangle. Show that every closed form on 𝑄 of
degree 𝑘 > 0 is exact.
Hint: Let 𝑄 = (𝑎1 , 𝑏1 ) × ⋯ × (𝑎𝑛 , 𝑏𝑛 ). Prove this assertion by induction, at the 𝑛th stage
of the induction letting 𝑈 = (𝑎1 , 𝑏1 ) × ⋯ × (𝑎𝑛−1 , 𝑏𝑛−1 ) and 𝐴 = (𝑎𝑛 , 𝑏𝑛 ).
Draft: March 28, 2018
We also leave for you to prove the following two assertions, both of which are special
cases of (6).
Example 2.5.4. If 𝒗 = 𝜕/𝜕𝑥𝑟 and 𝜔 = 𝑑𝑥𝐼 = 𝑑𝑥𝑖1 ∧ ⋯ ∧ 𝑑𝑥𝑖𝑘 then
𝑘
(2.5.5) 𝜄𝒗 𝜔 = ∑(−1)𝑖−1 𝛿𝑖,𝑖𝑟 𝑑𝑥𝐼𝑟
𝑖=1
where
1, 𝑖 = 𝑖𝑟
𝛿𝑖,𝑖𝑟 ≔ {
0, 𝑖 ≠ 𝑖𝑟 .
and 𝐼𝑟 = (𝑖1 , …, 𝚤𝑟̂ , …, 𝑖𝑘 ).
Draft: March 28, 2018
𝑛
Example 2.5.6. If 𝒗 = ∑𝑖=1 𝑓𝑖 𝜕/𝜕𝑥𝑖 and 𝜔 = 𝑑𝑥1 ∧ ⋯ ∧ 𝑑𝑥𝑛 then
𝑛
(2.5.7) ̂ 𝑟 ⋯ ∧ 𝑑𝑥𝑛 .
𝜄𝒗 𝜔 = ∑ (−1)𝑟−1 𝑓𝑟 𝑑𝑥1 ∧ ⋯ 𝑑𝑥
𝑟=1
By combining exterior differentiation with the interior product operation one gets an-
other basic operation of vector fields on forms: the Lie differentiation operation. For zero-
forms, i.e., for 𝐶∞ functions, we defined this operation by the formula (2.1.16). For 𝑘-forms
we define it by a slightly more complicated formula.
Definition 2.5.8. Let 𝑈 be an open subset of 𝐑𝑛 , 𝒗 a vector field on 𝑈, and 𝜔 ∈ 𝛺𝑘 (𝑈). The
Lie derivative of 𝜔 with respect to 𝒗 is the 𝑘-form
(2.5.9) 𝐿𝒗 𝜔 ≔ 𝜄𝒗 (𝑑𝜔) + 𝑑(𝜄𝒗 𝜔) .
Notice that for zero-forms the second summand is zero, so (2.5.9) and (2.1.16) agree.
Properties 2.5.10. Let 𝑈 be an open subset of 𝐑𝑛 , 𝒗 a vector field on 𝑈, 𝜔 ∈ 𝛺𝑘 (𝑈), and
𝜇 ∈ 𝛺ℓ (𝑈). The Lie derivative enjoys the following properties:
(1) Commutativity with exterior differentiation: we have
(2.5.11) 𝑑(𝐿𝒗 𝜔) = 𝐿𝒗 (𝑑𝜔) .
(2) Interaction with wedge products: we have
(2.5.12) 𝐿𝒗 (𝜔 ∧ 𝜇) = 𝐿𝒗 𝜔 ∧ 𝜇 + 𝜔 ∧ 𝐿𝒗 𝜇
From Properties 2.5.10 it is fairly easy to get an explicit formula for 𝐿𝒗 𝜔. Namely let 𝜔
be the 𝑘-form
𝜔 = ∑ 𝑓𝐼 𝑑𝑥𝐼 , 𝑓𝐼 ∈ 𝐶∞ (𝑈)
𝐼
and 𝒗 the vector field
𝑛
𝜕
𝒗 = ∑ 𝑔𝑖 , 𝑔𝑖 ∈ 𝐶∞ (𝑈) .
𝑖=1 𝜕𝑥𝑖
By equation (2.5.12)
𝐿𝒗 (𝑓𝐼 𝑑𝑥𝐼 ) = (𝐿𝒗 𝑓𝐼 )𝑑𝑥𝐼 + 𝑓𝐼 (𝐿𝒗 𝑑𝑥𝐼 )
and
𝑘
𝐿𝒗 𝑑𝑥𝐼 = ∑ 𝑑𝑥𝑖1 ∧ ⋯ ∧ 𝐿𝒗 𝑑𝑥𝑖𝑟 ∧ ⋯ ∧ 𝑑𝑥𝑖𝑘 ,
𝑟=1
and by equation (2.5.11)
𝐿𝒗 𝑑𝑥𝑖𝑟 = 𝑑𝐿𝒗 𝑥𝑖𝑟
so to compute 𝐿𝒗 𝜔 one is reduced to computing 𝐿𝒗 𝑥𝑖𝑟 and 𝐿𝒗 𝑓𝐼 .
However by equation (2.5.12)
𝐿𝒗 𝑥𝑖𝑟 = 𝑔𝑖𝑟
and
𝑛
𝜕𝑓
𝐿𝒗 𝑓𝐼 = ∑ 𝑔𝑖 𝐼 .
𝑖=1 𝜕𝑥𝑖
We leave the verification of Properties 2.5.10 as exercises, and also ask you to prove (by the
method of computation that we have just sketched) the following divergence formula.
𝑛
Lemma 2.5.13. Let 𝑈 ⊂ 𝐑𝑛 be open, let 𝑔1 , …, 𝑔𝑛 ∈ 𝐶∞ (𝑈), and set 𝒗 ≔ ∑𝑖=1 𝑔𝑖 𝜕/𝜕𝑥𝑖 . Then
𝑛
𝜕𝑔𝑖
(2.5.14) 𝐿𝒗 (𝑑𝑥1 ∧ ⋯ ∧ 𝑑𝑥𝑛 ) = ∑ ( ) 𝑑𝑥1 ∧ ⋯ ∧ 𝑑𝑥𝑛 .
𝑖=1 𝜕𝑥𝑖
Draft: March 28, 2018
• Conclude that
(2.5.15) 𝜔 − 𝑑𝜄𝒗 𝜈 = 𝜄𝒗 𝑑𝜈 .
• Suppose 𝑑𝜔 = 0. Conclude from equation (2.5.15) and (5) of Properties 2.5.3 that
the form 𝛽 = 𝜄𝒗 𝑑𝜈 satisfies 𝑑𝛽 = 𝜄(𝒗)𝛽 = 0.
• By part (3), 𝛽 is effectively a closed form on 𝐑𝑛−1 , and by induction, 𝛽 = 𝑑𝛼. Thus
by equation (2.5.15)
𝜔 = 𝑑𝜄𝒗 𝜈 + 𝑑𝛼 .
(4) We observed in §1.7 that the operation (2.6.1) commutes with wedge-product, hence if
𝜔1 ∈ 𝛺𝑘 (𝑉) and 𝜔2 ∈ 𝛺ℓ (𝑉)
𝑑𝑓𝑝⋆ (𝜔1 )𝑞 ∧ (𝜔2 )𝑞 = 𝑑𝑓𝑝⋆ (𝜔1 )𝑞 ∧ 𝑑𝑓𝑝⋆ (𝜔2 )𝑞 .
In other words
(2.6.8) 𝑓⋆ 𝜔1 ∧ 𝜔2 = 𝑓⋆ 𝜔1 ∧ 𝑓⋆ 𝜔2 .
(5) Let 𝑊 be an open subset of 𝐑𝑘 and 𝑔 ∶ 𝑉 → 𝑊 a 𝐶∞ map. Given a point 𝑝 ∈ 𝑈, let
𝑞 = 𝑓(𝑝) and 𝑤 = 𝑔(𝑞). Then the composition of the map
(𝑑𝑓𝑝 )⋆ ∶ 𝛬𝑘 (𝑇𝑞⋆ ) → 𝛬𝑘 (𝑇𝑝⋆ )
and the map
(𝑑𝑔𝑞 )⋆ ∶ 𝛬𝑘 (𝑇𝑤⋆ ) → 𝛬𝑘 (𝑇𝑞⋆ )
is the map
(𝑑𝑔𝑞 ∘ 𝑑𝑓𝑝 )⋆ ∶ 𝛬𝑘 (𝑇𝑤⋆ ) → 𝛬𝑘 (𝑇𝑝⋆ )
by formula (1.7.5). However, by the chain rule
(𝑑𝑔𝑞 ) ∘ (𝑑𝑓)𝑝 = 𝑑(𝑔 ∘ 𝑓)𝑝
so this composition is the map
𝑑(𝑔 ∘ 𝑓)𝑝⋆ ∶ 𝛬𝑘 (𝑇𝑤⋆ ) → 𝛬𝑘 (𝑇𝑝⋆ ) .
Thus if 𝜔 ∈ 𝛺𝑘 (𝑊)
𝑓⋆ (𝑔⋆ 𝜔) = (𝑔 ∘ 𝑓)⋆ 𝜔 .
Let us see what the pullback operation looks like in coordinates. Using multi-index
notation we can express every 𝑘-form 𝜔 ∈ 𝛺𝑘 (𝑉) as a sum over multi-indices of length 𝑘
(2.6.9) 𝜔 = ∑ 𝜙𝐼 𝑑𝑥𝐼 ,
𝐼
the coefficient 𝜙𝐼 of 𝑑𝑥𝐼 being in 𝐶∞ (𝑉). Hence by (2.6.5)
𝑓⋆ 𝜔 = ∑ 𝑓⋆ 𝜙𝐼 𝑓⋆ (𝑑𝑥𝐼 )
where 𝑓⋆ 𝜙𝐼 is the function of 𝜙 ∘ 𝑓.
What about 𝑓⋆ 𝑑𝑥𝐼 ? If 𝐼 is the multi-index, (𝑖1 , …, 𝑖𝑘 ), then by definition
𝑑𝑥𝐼 = 𝑑𝑥𝑖1 ∧ ⋯ ∧ 𝑑𝑥𝑖𝑘
so
𝑓⋆ (𝑑𝑥𝐼 ) = 𝑓⋆ (𝑑𝑥𝑖1 ) ∧ ⋯ ∧ 𝑓⋆ (𝑑𝑥𝑖𝑘 )
by (2.6.8), and by (2.6.7), we have
𝑓⋆ 𝑑𝑥𝑖 = 𝑑𝑓⋆ 𝑥𝑖 = 𝑑𝑓𝑖
where 𝑓𝑖 is the 𝑖th coordinate function of the map 𝑓. Thus, setting
𝑑𝑓𝐼 ≔ 𝑑𝑓𝑖1 ∧ ⋯ ∧ 𝑑𝑓𝑖𝑘 ,
we see that for each multi-index 𝐼 we have
(2.6.10) 𝑓⋆ (𝑑𝑥𝐼 ) = 𝑑𝑓𝐼 ,
and for the pullback of the form (2.6.9)
(2.6.11) 𝑓⋆ 𝜔 = ∑ 𝑓⋆ 𝜙𝐼 𝑑𝑓𝐼 .
𝐼
Draft: March 28, 2018
We will use this formula to prove that pullback commutes with exterior differentiation:
(2.6.12) 𝑑(𝑓⋆ 𝜔) = 𝑓⋆ (𝑑𝜔) .
To prove this we recall that by (2.3.6) we have 𝑑(𝑑𝑓𝐼 ) = 0, hence by (2.3.1) and (2.6.10) we
have
𝑑(𝑓⋆ 𝜔) = ∑ 𝑑(𝑓⋆ 𝜙𝐼 ) ∧ 𝑑𝑓𝐼 = ∑ 𝑓⋆ (𝑑𝜙𝐼 ) ∧ 𝑓⋆ (𝑑𝑥𝐼 )
𝐼 𝐼
⋆ ⋆
= 𝑓 ∑ 𝑑𝜙𝐼 ∧ 𝑑𝑥𝐼 = 𝑓 (𝑑𝜔) .
𝐼
A special case of formula (2.6.10) will be needed in Chapter 4: Let 𝑈 and 𝑉 be open
subsets of 𝐑𝑛 and let 𝜔 = 𝑑𝑥1 ∧ ⋯ ∧ 𝑑𝑥𝑛 . Then by (2.6.10) we have
𝑓⋆ 𝜔𝑝 = (𝑑𝑓1 )𝑝 ∧ ⋯ ∧ (𝑑𝑓𝑛 )𝑝
for all 𝑝 ∈ 𝑈. However,
𝜕𝑓𝑖
(𝑑𝑓𝑖 )𝑝 = ∑ (𝑝)(𝑑𝑥𝑗 )𝑝
𝜕𝑥𝑗
and hence by formula (1.7.10)
𝜕𝑓𝑖
𝑓⋆ 𝜔𝑝 = det [ (𝑝)] (𝑑𝑥1 ∧ ⋯ ∧ 𝑑𝑥𝑛 )𝑝 .
𝜕𝑥𝑗
In other words
𝜕𝑓𝑖
(2.6.13) 𝑓⋆ (𝑑𝑥1 ∧ ⋯ ∧ 𝑑𝑥𝑛 ) = det [ ] 𝑑𝑥1 ∧ ⋯ ∧ 𝑑𝑥𝑛 .
𝜕𝑥𝑗
In Exercise 2.6.iv and equation (2.6.6) we outline the proof of an important topological
property of the pullback operation.
Definition 2.6.14. Let 𝑈 be an open subset of 𝐑𝑛 , 𝑉 an open subset of 𝐑𝑚 , 𝐴 ⊂ 𝐑 an open
interval containing 0 and 1 and 𝑓0 , 𝑓1 ∶ 𝑈 → 𝑉 two 𝐶∞ maps. A 𝐶∞ map 𝐹 ∶ 𝑈 × 𝐴 → 𝑉 is
a 𝐶∞ homotopy between 𝑓0 and 𝑓1 if 𝐹(𝑥, 0) = 𝑓0 (𝑥) and 𝐹(𝑥, 1) = 𝑓1 (𝑥).
If there exists a homotopy between 𝑓0 and 𝑓1 , we say that 𝑓0 and 𝑓1 are homotopic and
write 𝑓0 ≃ 𝑓1 .
Intuitively, 𝑓0 and 𝑓1 are homotopic if there exists a family of 𝐶∞ maps 𝑓𝑡 ∶ 𝑈 → 𝑉,
where 𝑓𝑡 (𝑥) = 𝐹(𝑥, 𝑡), which “smoothly deform 𝑓0 into 𝑓1 ”. In Exercise 2.6.iv and equa-
tion (2.6.6) you will be asked to verify that for 𝑓0 and 𝑓1 to be homotopic they have to
satisfy the following criteria.
Theorem 2.6.15. Let 𝑈 be an open subset of 𝐑𝑛 , 𝑉 an open subset of 𝐑𝑚 , and 𝑓0 , 𝑓1 ∶ 𝑈 → 𝑉
two 𝐶∞ maps. If 𝑓0 and 𝑓1 are homotopic then for every closed form 𝜔 ∈ 𝛺𝑘 (𝑉) the form
𝑓1⋆ 𝜔 − 𝑓0⋆ 𝜔 is exact.
This theorem is closely related to the Poincaré lemma, and, in fact, one gets from it
a slightly stronger version of the Poincaré lemma than that described in Exercise 2.3.iv
and equation (2.3.7).
Definition 2.6.16. An open subset 𝑈 of 𝐑𝑛 is contractible if, for some point 𝑝0 ∈ 𝑈, the
identity map id𝑈 ∶ 𝑈 → 𝑈 is homotopic to the constant map
𝑓0 ∶ 𝑈 → 𝑈 , 𝑓0 (𝑝) = 𝑝0
at 𝑝0 .
Draft: March 28, 2018
From Theorem 2.6.15 it is easy to see that the Poincaré lemma holds for contractable
open subsets of 𝐑𝑛 . If 𝑈 is contractable every closed 𝑘-form on 𝑈 of degree 𝑘 > 0 is exact. To
see this, note that if 𝜔 ∈ 𝛺𝑘 (𝑈), then for the identity map id⋆𝑈 𝜔 = 𝜔 and if 𝑓 is the constant
map at a point of 𝑈, then 𝑓⋆ 𝜔 = 0.
Exercises for §2.6
3 3
Exercise 2.6.i. Let 𝑓 ∶ 𝐑 → 𝐑 be the map
𝑓(𝑥1 , 𝑥2 , 𝑥3 ) = (𝑥1 𝑥2 , 𝑥2 𝑥32 , 𝑥33 ) .
Compute the pullback, 𝑓⋆ 𝜔 for the following forms.
(1) 𝜔 = 𝑥2 𝑑𝑥3
(2) 𝜔 = 𝑥1 𝑑𝑥1 ∧ 𝑑𝑥3
(3) 𝜔 = 𝑥1 𝑑𝑥1 ∧ 𝑑𝑥2 ∧ 𝑑𝑥3
Exercise 2.6.ii. Let 𝑓 ∶ 𝐑2 → 𝐑3 be the map
𝑓(𝑥1 , 𝑥2 ) = (𝑥12 , 𝑥22 , 𝑥1 𝑥2 ) .
Complete the pullback, 𝑓⋆ 𝜔, for the following forms.
(1) 𝜔 = 𝑥2 𝑑𝑥2 + 𝑥3 𝑑𝑥3
(2) 𝜔 = 𝑥1 𝑑𝑥2 ∧ 𝑑𝑥3
(3) 𝜔 = 𝑑𝑥1 ∧ 𝑑𝑥2 ∧ 𝑑𝑥3
Exercise 2.6.iii. Let 𝑈 be an open subset of 𝐑𝑛 , 𝑉 an open subset of 𝐑𝑚 , 𝑓 ∶ 𝑈 → 𝑉 a 𝐶∞
map and 𝛾 ∶ [𝑎, 𝑏] → 𝑈 a 𝐶∞ curve. Show that for 𝜔 ∈ 𝛺1 (𝑉)
∫ 𝑓⋆ 𝜔 = ∫ 𝜔
𝛾 𝛾1
(i.e., none of the summands involve 𝑑𝑡). For a reduced form 𝜇, let 𝑄𝜇 ∈ 𝛺ℓ (𝑈) be the form
1
(2.6.17) 𝑄𝜇 ≔ (∑ ∫ 𝑓𝐼 (𝑥, 𝑡)𝑑𝑡) 𝑑𝑥𝐼
𝐼 0
ℓ
and let 𝜇1 , 𝜇2 ∈ 𝛺 (𝑈), be the forms
𝜇0 = ∑ 𝑓𝐼 (𝑥, 0)𝑑𝑥𝐼
𝐼
and
𝜇1 = ∑ 𝑓𝐼 (𝑥, 1)𝑑𝑥𝐼 .
Now recall from Exercise 2.4.v that every form 𝜔 ∈ 𝛺𝑘 (𝑈 × 𝐴) can be written uniquely as a
sum
(2.6.18) 𝜔 = 𝑑𝑡 ∧ 𝛼 + 𝛽
where 𝛼 and 𝛽 are reduced.
(1) Prove:
Draft: March 28, 2018
Even in three dimensions, however, there is a good reason for replacing the gradient,
divergence, and curl by the three operations (2.7.1). A problem that physicists spend a lot of
time worrying about is the problem of general covariance: formulating the laws of physics in
such a way that they admit as large a set of symmetries as possible, and frequently these for-
mulations involve differential forms. An example is Maxwell’s equations, the fundamental
laws of electromagnetism. These are usually expressed as identities involving div and curl.
However, as we explain below, there is an alternative formulation of Maxwell’s equations
based on the operations (2.7.1), and from the point of view of general covariance, this for-
mulation is much more satisfactory: the only symmetries of 𝐑3 which preserve div and curl
are translations and otations, whereas the operations (2.7.1) admit all diffeomorphisms of
𝐑3 as symmetries.
To describe how the gradient, divergence, and curl are related to the operations (2.7.1)
we first note that there are two ways of converting vector fields into forms.
The first makes use of the natural inner product 𝐵(𝑣, 𝑤) = ∑ 𝑣𝑖 𝑤𝑖 on 𝐑𝑛 . From this
inner product one gets by Exercise 1.2.ix a bijective linear map
𝐿 ∶ 𝐑𝑛 ⥲ (𝐑𝑛 )⋆
with the defining property: 𝐿(𝑣) = ℓ if and only if ℓ(𝑤) = 𝐵(𝑣, 𝑤). Via the identification
(2.1.2) 𝐵 and 𝐿 can be transferred to 𝑇𝑝 𝐑𝑛 , giving one an inner product 𝐵𝑝 on 𝑇𝑝 𝐑𝑛 and a
bijective linear map
𝐿𝑝 ∶ 𝑇𝑝 𝐑𝑛 ⥲ 𝑇𝑝⋆ 𝐑𝑛 .
Hence if we are given a vector field 𝒗 on 𝑈 we can convert it into a 1-form 𝒗♯ by setting
(2.7.2) 𝒗♯ (𝑝) ≔ 𝐿𝑝 𝒗(𝑝)
and this sets up a bijective correspondence between vector fields and 1-forms.
For instance
𝜕
𝒗= ⟺ 𝒗♯ = 𝑑𝑥𝑖 ,
𝜕𝑥𝑖
(see Exercise 2.7.iii) and, more generally,
𝑛 𝑛
𝜕
(2.7.3) 𝒗 = ∑ 𝑓𝑖 ⟺ 𝒗♯ = ∑ 𝑓𝑖 𝑑𝑥𝑖 .
𝑖=1 𝜕𝑥𝑖 𝑖=1
Moreover, every (𝑛 − 1)-form can be written uniquely as such a sum, so (2.7.5) and (2.7.6)
set up a bijective correspondence between vector fields and (𝑛 − 1)-forms. Under this corre-
spondence the 𝑑-operation gets converted into an operation on vector fields
𝒗 ↦ 𝑑𝜄𝒗 𝛺 .
Moreover, by (2.5.9)
𝑑𝜄𝒗 𝛺 = 𝐿𝒗 𝛺
and by (2.5.14)
𝐿𝒗 𝛺 = div(𝒗)𝛺
where
𝑛
𝜕𝑓𝑖
(2.7.7) div(𝒗) = ∑ .
𝑖=1 𝜕𝑥𝑖
In other words, this correspondence between (𝑛 − 1)-forms and vector fields converts the
𝑑-operation into the divergence operation (2.7.7) on vector fields.
Notice that divergence and gradient are well-defined as vector calculus operations in 𝑛
dimensions, even though one usually thinks of them as operations in 3-dimensional vector
calculus. The curl operation, however, is intrinsically a 3-dimensional vector calculus oper-
ation. To define it we note that by (2.7.6) every 2-form 𝜇 on an open subset 𝑈 ⊂ 𝐑3 can be
written uniquely as an interior product,
(2.7.8) 𝜇 = 𝜄𝒘 𝑑𝑥1 ∧ 𝑑𝑥2 ∧ 𝑑𝑥3 ,
for some vector field 𝒘, and the left-hand side of this formula determines 𝒘 uniquely.
Definition 2.7.9. Let 𝑈 be an open subset of 𝐑3 and 𝒗 a vector field on 𝑈. From 𝒗 we get by
(2.7.3) a 1-form 𝒗♯ , and hence by (2.7.8) a vector field 𝒘 satisfying
𝑑𝒗♯ = 𝜄𝒘 𝑑𝑥1 ∧ 𝑑𝑥2 ∧ 𝑑𝑥3 .
The curl of 𝒗 is the vector field
curl(𝒗) ≔ 𝒘 .
We leave for you to check that this definition coincides with the definition one finds in
calculus books. More explicitly we leave for you to check that if 𝒗 is the vector field
𝜕 𝜕 𝜕
𝒗 = 𝑓1 + 𝑓2 + 𝑓3
𝜕𝑥1 𝜕𝑥2 𝜕𝑥3
then
𝜕 𝜕 𝜕
curl(𝒗) = 𝑔1 + 𝑔2 + 𝑔3
𝜕𝑥1 𝜕𝑥2 𝜕𝑥3
where
𝜕𝑓2 𝜕𝑓3
𝑔1 = −
𝜕𝑥3 𝜕𝑥2
𝜕𝑓 𝜕𝑓
(2.7.10) 𝑔2 = 3 − 1
𝜕𝑥1 𝜕𝑥3
𝜕𝑓 𝜕𝑓
𝑔3 = 1 − 2 .
𝜕𝑥2 𝜕𝑥1
To summarize: the gradient, curl, and divergence operations in 3-dimensions are basi-
cally just the three operations (2.7.1). The gradient operation is the operation (2.7.1) in de-
gree zero, curl is the operation (2.7.1) in degree one, and divergence is the operation (2.7.1)
Draft: March 28, 2018
in degree two. However, to define gradient we had to assign an inner product 𝐵𝑝 to the tan-
gent space 𝑇𝑝 𝐑𝑛 for each 𝑝 ∈ 𝑈; to define divergence we had to equip 𝑈 with the 3-form
𝛺 and to define curl, the most complicated of these three operations, we needed the inner
products 𝐵𝑝 and the form 𝛺. This is why diffeomorphisms preserve the three operations
(2.7.1) but do not preserve gradient, curl, and divergence. The additional structures which
one needs to define grad, curl and div are only preserved by translations and rotations.
Maxwell’s equations and differential forms
We conclude this section by showing how Maxwell’s equations, which are usually for-
mulated in terms of divergence and curl, can be reset into “form” language. The discussion
below is an abbreviated version of [5, §1.20].
Maxwell’s equations assert:
(2.7.11) div(𝒗𝐸 ) = 𝑞
𝜕
(2.7.12) curl(𝒗𝐸 ) = − 𝒗
𝜕𝑡 𝑀
(2.7.13) div(𝒗𝑀 ) = 0
𝜕
(2.7.14) 𝑐2 curl(𝒗𝑀 ) = 𝒘 + 𝒗
𝜕𝑡 𝐸
where 𝒗𝐸 and 𝒗𝑀 are the electric and magnetic fields, 𝑞 is the scalar charge density, 𝒘 is the
current density and 𝑐 is the velocity of light. (To simplify (2.7.15) slightly we assume that
our units of space–time are chosen so that 𝑐 = 1.) As above let 𝛺 = 𝑑𝑥1 ∧ 𝑑𝑥2 ∧ 𝑑𝑥3 and let
𝜇𝐸 = 𝜄(𝒗𝐸 )𝛺
and
𝜇𝑀 = 𝜄(𝒗𝑀 )𝛺 .
We can then rewrite equations (2.7.11) and (2.7.13) in the form
(2.7.11′ ) 𝑑𝜇𝐸 = 𝑞𝛺
and
(2.7.13′ ) 𝑑𝜇𝑀 = 0 .
What about (2.7.12) and (2.7.14)? We leave the following “form” versions of these equa-
tions as an exercise:
♯ 𝜕
(2.7.12′ ) 𝑑𝒗𝐸 = − 𝜇𝑀
𝜕𝑡
and
♯ 𝜕
(2.7.14′ ) 𝑑𝒗𝑀 = 𝜄𝒘 𝛺 + 𝜇𝐸
𝜕𝑡
♯ ♯
where the 1-forms, 𝒗𝐸 and 𝒗𝑀 , are obtained from 𝒗𝐸 and 𝒗𝑀 by the operation (2.7.2).
These equations can be written more compactly as differential form identities in 3 + 1
dimensions. Let 𝜔𝑀 and 𝜔𝐸 be the 2-forms
♯
𝜔𝑀 = 𝜇𝑀 − 𝒗𝐸 ∧ 𝑑𝑡
and
♯
(2.7.15) 𝜔𝐸 = 𝜇𝐸 − 𝒗𝑀 ∧ 𝑑𝑡
and let 𝛬 be the 3-form
𝛬 ≔ 𝑞𝛺 + 𝜄𝒘 𝛺 ∧ 𝑑𝑡 .
Draft: March 28, 2018
We will leave for you to show that the four equations (2.7.11)–(2.7.14) are equivalent
to two elegant and compact (3 + 1)-dimensional identities
𝑑𝜔𝑀 = 0
and
𝑑𝜔𝐸 = 𝛬 .
Exercises for §2.7
Exercise 2.7.i. Verify that the curl operation is given in coordinates by equation (2.7.10).
Exercise 2.7.ii. Verify that Maxwell’s equations (2.7.11) and (2.7.12) become the equations (2.7.13)
and (2.7.14) when rewritten in differential form notation.
Exercise 2.7.iii. Show that in (3 + 1)-dimensions Maxwell’s equations take the form of equa-
tions (2.7.10) and (2.7.11).
Exercise 2.7.iv. Let 𝑈 be an open subset of 𝐑3 and 𝒗 a vector field on 𝑈. Show that if 𝒗 is
the gradient of a function, its curl has to be zero.
Exercise 2.7.v. If 𝑈 is simply connected prove the converse: If the curl of 𝒗 vanishes, 𝒗 is
the gradient of a function.
Exercise 2.7.vi. Let 𝒘 = curl(𝒗). Show that the divergence of 𝒘 is zero.
Exercise 2.7.vii. Is the converse to Exercise 2.7.vi true? Suppose the divergence of 𝒘 is zero.
Is 𝒘 = curl(𝒗) for some vector field 𝒗?
We can simplify this sum by noting that if the multi-index 𝐼 = (𝑖1 , …, 𝑖𝑛 ) is repeating, the
wedge product
(2.8.3) 𝑑𝑥𝑖1 ∧ 𝑑𝑦𝑖1 ∧ ⋯ ∧ 𝑑𝑥𝑖𝑛 ∧ 𝑑𝑥𝑖𝑛
Draft: March 28, 2018
involves two repeating 𝑑𝑥𝑖𝑗 and hence is zero, and if 𝐼 is non-repeating we can permute the
factors and rewrite (2.8.3) in the form
𝑑𝑥1 ∧ 𝑑𝑦1 ∧ ⋯ ∧ 𝑑𝑥𝑛 ∧ 𝑑𝑦𝑛 .
(See Exercise 1.6.v.)
Hence since these are exactly 𝑛! non-repeating multi-indices
𝜔𝑛 = 𝑛!𝑑𝑥1 ∧ 𝑑𝑦1 ∧ ⋯ ∧ 𝑑𝑥𝑛 ∧ 𝑑𝑦𝑛 ,
i.e.,
1 𝑛
(2.8.4) 𝜔 =𝛺
𝑛!
where
(2.8.5) 𝛺 = 𝑑𝑥1 ∧ 𝑑𝑦1 ∧ ⋯ ∧ 𝑑𝑥𝑛 ∧ 𝑑𝑦𝑛
is the symplectic volume form on 𝐑2𝑛 .
Definition 2.8.6. Let 𝑈 and 𝑉 be open subsets of 𝐑2𝑛 . A diffeomorphism 𝑓 ∶ 𝑈 ⥲ 𝑉 is a
symplectic diffeomorphism or symplectomorphism if 𝑓⋆ 𝜔 = 𝜔, where 𝜔 denotes the Dar-
boux form on 𝐑2𝑛 .
Definition 2.8.7. Let 𝑈 be an open subset of 𝐑2𝑛 , let
(2.8.8) 𝑓𝑡 ∶ 𝑈 ⥲ 𝑈 , − ∞ < 𝑡 < ∞
be a one-parameter group of diffeomorphisms of 𝑈, and 𝒗 be the vector field generating
(2.8.8) We say that 𝒗 is a symplectic vector field if the diffeomorphisms (2.8.8) are symplec-
tomorphisms, i.e., for all 𝑡 we have 𝑓𝑡⋆ 𝜔 = 𝜔.
Let us see what such vector fields have to look like. Note that by (2.6.23)
𝑑 ⋆
(2.8.9) 𝑓 𝜔 = 𝑓𝑡⋆ 𝐿𝒗 𝜔 ,
𝑑𝑡 𝑡
hence if 𝑓𝑡⋆ 𝜔 = 𝜔 for all 𝑡, the left hand side of (2.8.9) is zero, so
𝑓𝑡⋆ 𝐿𝒗 𝜔 = 0 .
In particular, for 𝑡 = 0, 𝑓𝑡 is the identity map so 𝑓𝑡⋆ 𝐿𝒗 𝜔 = 𝐿𝒗 𝜔 = 0. Conversely, if 𝐿𝒗 𝜔 = 0,
then 𝑓𝑡⋆ 𝐿𝒗 𝜔 = 0 so by (2.8.9) 𝑓𝑡⋆ 𝜔 does not depend on 𝑡. However, since 𝑓𝑡⋆ 𝜔 = 𝜔 for 𝑡 = 0
we conclude that 𝑓𝑡⋆ 𝜔 = 𝜔 for all 𝑡. To summarize, we have proved:
Theorem 2.8.10. Let 𝑓𝑡 ∶ 𝑈 → 𝑈 be a one-parameter group of diffeomorphisms and 𝒗 the
infinitesmal generator of this group. Then 𝒗 is symplectic of and only if 𝐿𝒗 𝜔 = 0.
There is an equivalent formulation of Theorem 2.8.10 in terms of the interior product
𝜄𝒗 𝜔. By (2.5.9)
𝐿𝒗 𝜔 = 𝑑𝜄𝒗 𝜔 + 𝜄𝒗 𝑑𝜔 .
But by (2.8.2) we have 𝑑𝜔 = 0, so
𝐿𝒗 𝜔 = 𝑑𝜄𝒗 𝜔 .
Thus we have shown:
Theorem 2.8.11. A vector field 𝒗 on an open subset of 𝐑2𝑛 is symplectic if and only if the form
𝜄𝒗 𝜔 is closed.
Draft: March 28, 2018
Let 𝑈 be an open subset of 𝐑2𝑛 and 𝒗 a vector field on 𝑈. If 𝜄𝒗 𝜔 is not only closed but is
exact we say that 𝒗 is a Hamiltonian vector field. In other words, 𝒗 is Hamiltonian if
(2.8.12) 𝜄𝒗 𝜔 = 𝑑𝐻
∞ ∞
for some 𝐶 function 𝐻 ∈ 𝐶 (𝑈).
Let is see what this condition looks like in coordinates. Let
𝑛
𝜕 𝜕
(2.8.13) 𝒗 = ∑ (𝑓𝑖 + 𝑔𝑖 ).
𝑖=1 𝜕𝑥𝑖 𝜕𝑦𝑖
Then
𝜄𝒗 𝜔 = ∑ 𝑓𝑖 𝜄𝜕/𝜕𝑥𝑖 (𝑑𝑥𝑗 ∧ 𝑑𝑦𝑗 ) + ∑ 𝑔𝑖 𝜄𝜕/𝜕𝑦𝑖 (𝑑𝑥𝑗 ∧ 𝑑𝑦𝑖 ) .
1≤𝑖,𝑗≤𝑛 1≤𝑖,𝑗≤𝑛
But
1, 𝑖 = 𝑗
𝜄𝜕/𝜕𝑥𝑖 𝑑𝑥𝑗 = {
0, 𝑖 ≠ 𝑗
and
𝜄𝜕/𝜕𝑥𝑖 𝑑𝑦𝑗 = 0
𝑛
so the first summand above is ∑𝑖=1 𝑓𝑖
𝑑𝑦𝑖 , and a similar argument shows that the second
𝑛
summand is − ∑𝑖=1 𝑔𝑖 𝑑𝑥𝑖 .
Hence if 𝒗 is the vector field (2.8.13), then
𝑛
(2.8.14) 𝜄𝒗 𝜔 = ∑(𝑓𝑖 𝑑𝑦𝑖 − 𝑔𝑖 𝑑𝑥𝑖 ) .
𝑖=1
Thus since
𝑛
𝜕𝐻 𝜕𝐻
𝑑𝐻 = ∑ ( 𝑑𝑥 + 𝑑𝑦 )
𝑖=1 𝜕𝑥𝑖 𝑖 𝜕𝑦𝑖 𝑖
we get from (2.8.12)–(2.8.14)
𝜕𝐻 𝜕𝐻
𝑓𝑖 = and 𝑔𝑖 = −
𝜕𝑦𝑖 𝜕𝑥𝑖
so 𝒗 has the form:
𝑛
𝜕𝐻 𝜕 𝜕𝐻 𝜕
(2.8.15) 𝒗 = ∑( − ).
𝑖=1 𝜕𝑦𝑖 𝜕𝑥𝑖 𝜕𝑥𝑖 𝜕𝑦𝑖
In particular if 𝛾(𝑡) = (𝑥(𝑡), 𝑦(𝑡)) is an integral curve of 𝒗 it has to satisfy the system of
differential equations
𝑑𝑥𝑖 𝜕𝐻
{
{ = (𝑥(𝑡), 𝑦(𝑡))
{ 𝑑𝑡 𝜕𝑦𝑖
(2.8.16) {
{ 𝑑𝑦𝑖 = − 𝜕𝐻 (𝑥(𝑡), 𝑦(𝑡)) .
{
{ 𝑑𝑡 𝜕𝑥𝑖
The formulas (2.8.13) and (2.8.14) exhibit an important property of the Darboux form 𝜔.
Every one-form on 𝑈 can be written uniquely as a sum
𝑛
∑ (𝑓𝑖 𝑑𝑦𝑖 − 𝑔𝑖 𝑑𝑥𝑖 )
𝑖=1
with 𝑓𝑖 and 𝑔𝑖 in 𝐶∞ (𝑈) and hence (2.8.13) and (2.8.14) imply:
Theorem 2.8.17. The map 𝒗 ↦ 𝜄𝒗 𝜔 defines a bijective between vector field and one-forms.
Draft: March 28, 2018
and interpreting the term on the right as the force exerted on the 𝑖th point-mass and the
term on the left as mass times acceleration this equation becomes Newton’s second law. □
In classical mechanics the equations (2.8.16) are known as the Hamilton–Jacobi equa-
tions. For a more detailed account of their role in classical mechanics we highly recommend
[1]. Historically these equations came up for the first time, not in Newtonian mechanics,
but in geometric optics and a brief description of their origins there and of their relation to
Maxwell’s equations can be found in [5].
We conclude this chapter by mentioning a few implications of the Hamiltonian descrip-
tion (2.8.16) of Newton’s equations (2.8.24).
(1) Conservation of energy: By (2.8.18) the energy function (2.8.22) is constant along the
integral curves of 𝒗, hence the energy of the system (2.8.16) does notchange in time.
(2) Noether’s principle: Let 𝛾𝑡 ∶ 𝐑2𝑛 → 𝐑2𝑛 be a one-parameter group of diffeomorphisms of
phase space and 𝒘 its infinitesimal generator. Then (𝛾𝑡 )𝑡∈𝐑 is a symmetry of the system
above if each 𝛾𝑡 preserves the function (2.8.22) and the vector field 𝒘 is Hamiltonian.
The Hamiltonian condition means that
𝜄𝒘 𝜔 = 𝑑𝐺
∞
for some 𝐶 function 𝐺, and what Noether’s principle asserts is that this function is an
integral of motion of the system (2.8.16), i.e., satisfies 𝐿𝒗 𝐺 = 0. In other words stated
more succinctly: symmetries of the system (2.8.16) give rise to integrals of motion.
(3) Poincaré recurrence: An important theorem of Poincaré asserts that if the function
𝐻 ∶ 𝐑2𝑛 → 𝐑
defined by (2.8.22) is proper then every trajectory of the system (2.8.16) returns arbi-
trarily close to its initial position at some positive time 𝑡0 , and, in fact, does this not just
once but does so infinitely often. We sketch a proof of this theorem, using (2.8.20), in
the next chapter.
Exercise 2.8.v. The expression (2.8.25) is known as the Poisson bracket of 𝐻1 and 𝐻2 , de-
noted by {𝐻1 , 𝐻2 }. Show that it is anti-symmetric
{𝐻1 , 𝐻2 } = −{𝐻2 , 𝐻1 }
and satisfies the Jacobi identity
0 = {𝐻1 , {𝐻2 , 𝐻3 }} + {𝐻2 , {𝐻3 , 𝐻1 }} + {𝐻3 , {𝐻1 , 𝐻2 }} .
Exercise 2.8.vi. Show that
{𝐻1 , 𝐻2 } = 𝐿𝒗𝐻 𝐻2 = −𝐿𝒗𝐻 𝐻1 .
1 2
Exercise 2.8.vii. Prove that the following three properties are equivalent.
(1) {𝐻1 , 𝐻2 } = 0.
(2) 𝐻1 is an integral of motion of 𝒗2 .
(3) 𝐻2 is an integral of motion of 𝒗1 .
Exercise 2.8.viii. Verify Noether’s principle.
Exercise 2.8.ix (conservation of linear momentum). Suppose the potential 𝑉 in equation (2.8.22)
is invariant under the one-parameter group of translations
𝑇𝑡 (𝑥1 , …, 𝑥𝑛 ) = (𝑥1 + 𝑡, …, 𝑥𝑛 + 𝑡) .
(1) Show that the function (2.8.22) is invariant under the group of diffeomorphisms
𝛾𝑡 (𝑥, 𝑦) = (𝑇𝑡 𝑥, 𝑦) .
(2) Show that the infinitesmal generator of this group is the Hamiltonian vector field 𝒗𝐺
𝑛
where 𝐺 = ∑𝑖=1 𝑦𝑖 .
(3) Conclude from Noether’s principle that this function is an integral of the vector field
𝒗𝐻 , i.e., that “total linear moment” is conserved.
(4) Show that “total linear momentum” is conserved if 𝑉 is the Coulomb potential
𝑚𝑖
∑ .
𝑖≠𝑗 |𝑥𝑖 − 𝑥𝑗 |
Exercise 2.8.x. Let 𝑅𝑖𝑡 ∶ 𝐑2𝑛 → 𝐑2𝑛 be the rotation which fixes the variables, (𝑥𝑘 , 𝑦𝑘 ), 𝑘 ≠ 𝑖
and rotates (𝑥𝑖 , 𝑦𝑖 ) by the angle, 𝑡:
𝑅𝑖𝑡 (𝑥𝑖 , 𝑦𝑖 ) = (cos 𝑡 𝑥𝑖 + sin 𝑡 𝑦𝑖 ,− sin 𝑡 𝑥𝑖 + cos 𝑡 𝑦𝑖 ) .
(1) Show that 𝑅𝑖𝑡 , −∞ < 𝑡 < ∞, is a one-parameter group of symplectomorphisms.
(2) Show that its generator is the Hamiltonian vector field, 𝒗𝐻𝑖 , where 𝐻𝑖 = (𝑥𝑖2 + 𝑦𝑖2 )/2.
𝑗
(3) Let 𝐻 be the harmonic oscillator Hamiltonian from Exercise 2.8.iii. Show that the 𝑅𝑡 ’s
preserve 𝐻.
(4) What does Noether’s principle tell one about the classical mechanical system with en-
ergy function 𝐻?
Exercise 2.8.xi. Show that if 𝑈 is an open subset of 𝐑2𝑛 and 𝒗 is a symplectic vector field on 𝑈
then for every point 𝑝0 ∈ 𝑈 there exists a neighborhood 𝑈0 of 𝑝0 on which 𝒗 is Hamiltonian.
Exercise 2.8.xii. Deduce from Exercises 2.8.iv and 2.8.xi that if 𝒗1 and 𝒗2 are symplectic
vector fields on an open subset 𝑈 of 𝐑2𝑛 their Lie bracket [𝒗1 , 𝒗2 ] is a Hamiltonian vector
field.
𝑛
Exercise 2.8.xiii. Let 𝛼 = ∑𝑖=1 𝑦𝑖 𝑑𝑥𝑖 .
(1) Show that 𝜔 = −𝑑𝛼.
Draft: March 28, 2018
(2) Show that if 𝛼1 is any one-form on 𝐑2𝑛 with the property 𝜔 = −𝑑𝛼1 , then
𝛼 = 𝛼1 + 𝑑𝐹
∞
for some 𝐶 function 𝐹.
(3) Show that 𝛼 = 𝜄𝒘 𝜔 where 𝒘 is the vector field
𝑛
𝜕
𝒘 ≔ − ∑ 𝑦𝑖 .
𝑖=1 𝜕𝑦𝑖
Exercise 2.8.xiv. Let 𝑈 be an open subset of 𝐑2𝑛 and 𝒗 a vector field on 𝑈. Show that 𝒗 has
the property, 𝐿𝒗 𝛼 = 0, if and only if
𝜄𝒗 𝜔 = 𝑑𝜄𝒗 𝛼 .
In particular conclude that if 𝐿𝒗 𝛼 = 0 then 𝑣 is Hamiltonian.
Hint: Equation (2.8.2).
Exercise 2.8.xv. Let 𝐻 be the function
𝑛
(2.8.26) 𝐻(𝑥, 𝑦) = ∑ 𝑓𝑖 (𝑥)𝑦𝑖 ,
𝑖=1
where the 𝑓𝑖 ’s are 𝐶∞ functions on 𝐑𝑛 . Show that
(2.8.27) 𝐿𝒗𝐻 𝛼 = 0 .
Exercise 2.8.xvi. Conversely show that if 𝐻 is any 𝐶∞ function on 𝐑2𝑛 satisfying equa-
tion (2.8.27) it has to be a function of the form (2.8.26).
Hints:
(1) Let 𝒗 be a vector field on 𝐑2𝑛 satisfying 𝐿𝒗 𝛼 = 0. By equation (2.8.16) we have 𝒗 = 𝒗𝐻 ,
where 𝐻 = 𝜄𝒗 𝛼.
(2) Show that 𝐻 has to satisfy the equation
𝑛
𝜕𝐻
∑ 𝑦𝑖 =𝐻.
𝑖=1 𝜕𝑦𝑖
𝜕𝐻
(3) Conclude that if 𝐻𝑟 = 𝜕𝑦𝑟
then 𝐻𝑟 has to satisfy the equation
𝑛
𝜕
∑ 𝑦𝑖 𝐻 =0.
𝑖=1 𝜕𝑦𝑖 𝑟
(4) Conclude that 𝐻𝑟 has to be constant along the rays (𝑥, 𝑡𝑦), for 0 ≤ 𝑡 < ∞.
(5) Conclude finally that 𝐻𝑟 has to be a function of 𝑥 alone, i.e., does not depend on 𝑦.
Exercise 2.8.xvii. Show that if 𝒗𝐑𝑛 is a vector field
𝑛
𝜕
∑ 𝑓𝑖 (𝑥)
𝑖=1 𝜕𝑥𝑖
on configuration space there is a unique lift of 𝒗𝐑𝑛 to phase space
𝑛
𝜕 𝜕
𝒗 = ∑ 𝑓𝑖 (𝑥) + 𝑔𝑖 (𝑥, 𝑦)
𝑖=1 𝜕𝑥𝑖 𝜕𝑦𝑖
satisfying 𝐿𝒗 𝛼 = 0.
Draft: March 28, 2018
Draft: March 28, 2018
CHAPTER 3
Integration of Forms
3.1. Introduction
The change of variables formula asserts that if 𝑈 and 𝑉 are open subsets of 𝐑𝑛 and
𝑓 ∶ 𝑈 → 𝑉 a 𝐶1 diffeomorphism then, for every continuous function 𝜙 ∶ 𝑉 → 𝐑 the integral
∫ 𝜙(𝑦)𝑑𝑦
𝑉
exists if and only if the integral
(the “hat” over the 𝑑𝑥𝑖 meaning that 𝑑𝑥𝑖 has to be omitted from the wedge product). Then
𝑛
𝜕𝑓𝑖
𝑑𝜇 = ∑(−1)𝑖−1 𝑑𝑥 ∧ ⋯ ∧ 𝑑𝑥𝑛 ,
𝑖=1 𝜕𝑥𝑖 1
and to show that the integral of 𝑑𝜇 is zero it suffices to show that each of the integrals
𝜕𝑓𝑖
(3.2.2)𝑖 ∫ 𝑑𝑥
𝐑𝑛 𝜕𝑥𝑖
is zero. By Fubini we can compute (3.2.2)𝑖 by first integrating with respect to the variable,
𝑥𝑖 , and then with respect to the remaining variables. But
𝜕𝑓𝑖 𝑥𝑖 =𝑏𝑖
∫ 𝑑𝑥𝑖 = 𝑓(𝑥)| =0
𝜕𝑥𝑖 𝑥𝑖 =𝑎𝑖
since 𝑓𝑖 is supported on 𝑈. □
We will prove that (1)⇒(2) by proving a somewhat stronger result. Let 𝑈 be an open
subset of 𝐑𝑚 . We’ll say that 𝑈 has property 𝑃 if every form 𝜔 ∈ 𝛺𝑚
𝑐 (𝑈) such that ∫𝑈 𝜔 = 0
we have 𝜔 ∈ 𝑑 𝛺𝑚−1
𝑐 (𝑈). We will prove:
Theorem 3.2.3. Let 𝑈 be an open subset of 𝐑𝑛−1 and 𝐴 ⊂ 𝐑 an open interval. Then if 𝑈 has
property 𝑃, 𝑈 × 𝐴 does as well.
Remark 3.2.4. It is very easy to see that the open interval 𝐴 itself has property 𝑃. (See
Exercise 3.2.i below.) Hence it follows by induction from Theorem 3.2.3 that
int 𝑄 = 𝐴 1 × ⋯ × 𝐴 𝑛 , 𝐴 𝑖 = (𝑎𝑖 , 𝑏𝑖 )
has property 𝑃, and this proves “(1)⇒(2)”.
Proof of Theorem 3.2.3. Let (𝑥, 𝑡) = (𝑥1 , …, 𝑥𝑛−1 , 𝑡) be product coordinates on 𝑈 × 𝐴. Given
𝜔 ∈ 𝛺𝑛𝑐 (𝑈 × 𝐴) we can express 𝜔 as a wedge product, 𝑑𝑡 ∧ 𝛼 with 𝛼 = 𝑓(𝑥, 𝑡) 𝑑𝑥1 ∧ ⋯ ∧ 𝑑𝑥𝑛−1
and 𝑓 ∈ 𝐶0∞ (𝑈 × 𝐴). Let 𝜃 ∈ 𝛺𝑛−1
𝑐 (𝑈) be the form
Then
∫ 𝜃 = ∫ 𝑓(𝑥, 𝑡)𝑑𝑥𝑑𝑡 = ∫ 𝜔
𝐑𝑛−1 𝐑𝑛 𝐑𝑛
so if the integral of 𝜔 is zero, the integral of 𝜃 is zero. Hence since 𝑈 has property 𝑃, 𝜃 = 𝑑𝜈
for some 𝜈 ∈ 𝛺𝑛−2 ∞
𝑐 (𝑈). Let 𝜌 ∈ 𝐶 (𝐑) be a bump function which is supported on 𝐴 and
whose integral over 𝐴 is 1. Setting
𝜅 = −𝜌(𝑡)𝑑𝑡 ∧ 𝜈
we have
𝑑𝜅 = 𝜌(𝑡)𝑑𝑡 ∧ 𝑑𝜈 = 𝜌(𝑡)𝑑𝑡 ∧ 𝜃 ,
and hence
𝜔 − 𝑑𝜅 = 𝑑𝑡 ∧ (𝛼 − 𝜌(𝑡)𝜃)
= 𝑑𝑡 ∧ 𝑢(𝑥, 𝑡)𝑑𝑥1 ∧ ⋯ ∧ 𝑑𝑥𝑛−1
where
𝑢(𝑥, 𝑡) = 𝑓(𝑥, 𝑡) − 𝜌(𝑡) ∫ 𝑓(𝑥, 𝑡)𝑑𝑡
𝐴
by (3.2.5). Thus
(3.2.6) ∫ 𝑢(𝑥, 𝑡) 𝑑𝑡 = 0 .
exist and are continuous as functions of 𝑥 and 𝑦. Prove the following “differentiation under
the integral sign” theorem (which we implicitly used in our proof of Theorem 3.2.3).
Theorem 3.2.8. The function 𝑔(𝑥) ≔ ∫𝐑ℓ 𝑓(𝑥, 𝑦)𝑑𝑦 is of class 𝐶1 and
𝜕𝑔 𝜕𝑓
(𝑥) = ∫ (𝑥, 𝑦)𝑑𝑦 .
𝜕𝑥𝑖 𝜕𝑥𝑖
Hints: For 𝑦 fixed and ℎ ∈ 𝐑𝑘 ,
𝑓(𝑥 + ℎ, 𝑦) − 𝑓(𝑥, 𝑦) = 𝐷𝑥 𝑓(𝑐, 𝑦)ℎ
for some point 𝑐 on the line segment joining 𝑥 to 𝑥+ℎ. Using the fact that 𝐷𝑥 𝑓 is continuous
as a function of 𝑥 and 𝑦 and compactly supported, conclude the following.
Lemma 3.2.9. Given 𝜀 > 0 there exists a 𝛿 > 0 such that for |ℎ| ≤ 𝛿
|𝑓(𝑥 + ℎ, 𝑦) − 𝑓(𝑥, 𝑦) − 𝐷𝑥 𝑓(𝑥, 𝑦)ℎ| ≤ 𝜀|ℎ| .
Now let 𝑄 ⊂ 𝐑ℓ be a rectangle with supp(𝑓) ⊂ 𝐑𝑘 × 𝑄 and show that
∫ 𝐷𝑥 𝑓(𝑥, 𝑦)𝑑𝑦 .
is of class 𝐶𝑟 .
Exercise 3.2.iv. Let 𝑈 be an open subset of 𝐑𝑛−1 , 𝐴 ⊂ 𝐑 an open interval and (𝑥, 𝑡) product
coordinates on 𝑈 × 𝐴. Recall from Exercise 2.3.v that every form 𝜔 ∈ 𝛺𝑘 (𝑈 × 𝐴) can be
written uniquely as a sum 𝜔 = 𝑑𝑡 ∧ 𝛼 + 𝛽 where 𝛼 and 𝛽 are reduced, i.e., do not contain a
factor of 𝑑𝑡.
(1) Show that if 𝜔 is compactly supported on 𝑈 × 𝐴 then so are 𝛼 and 𝛽.
(2) Let 𝛼 = ∑𝐼 𝑓𝐼 (𝑥, 𝑡)𝑑𝑥𝐼 . Show that the form
is in 𝛺𝑘−1
𝑐 (𝑈).
(3) Show that if 𝑑𝜔 = 0, then 𝑑𝜃 = 0.
Hint: By equation (3.2.11),
𝜕𝑓𝐼
𝑑𝜃 = ∑ (∫ (𝑥, 𝑡) 𝑑𝑡) 𝑑𝑥𝑖 ∧ 𝑑𝑥𝐼 = ∫ (𝑑𝑈 𝛼) 𝑑𝑡
𝐼,𝑖 𝐴 𝜕𝑥𝑖 𝐴
𝑑𝛽
and by equation (2.4.19) we have 𝑑𝑈 𝛼 = 𝑑𝑡
.
Exercise 3.2.v. In Exercise 3.2.iv, show that if 𝜃 is in 𝑑𝛺𝑘−2 𝑘−1
𝑐 (𝑈) then 𝜔 is in 𝑑𝛺𝑐 (𝑈 × 𝐴)
as follows.
Draft: March 28, 2018
𝜔 − 𝑑𝜅 = 𝑑𝑡 ∧ (𝛼 − 𝜌(𝑡)𝜃) + 𝛽
= 𝑑𝑡 ∧ (∑𝐼 𝑢𝐼 (𝑥, 𝑡)𝑑𝑥𝐼 ) + 𝛽 ,
where
𝑢𝐼 (𝑥, 𝑡) = 𝑓𝐼 (𝑥, 𝑡) − 𝜌(𝑡) ∫ 𝑓𝐼 (𝑥, 𝑡)𝑑𝑡 .
𝐴
(2) Let 𝑎 and 𝑏 be the end points of 𝐴 and let
𝑡
𝑣𝐼 (𝑥, 𝑡) = ∫ 𝑢𝐼 (𝑥, 𝑡) 𝑑𝑡 .
𝑎
𝑑𝛾 = 𝜔 − 𝑑𝜅 − 𝛽 + 𝑑𝑈 𝛾 .
Exercise 3.2.vi. Let 𝑈 be an open subset of 𝐑𝑚 . We say that 𝑈 has property 𝑃𝑘 , for 1 ≤ 𝑘 < 𝑚,
if every closed 𝑘-form 𝜔 ∈ 𝛺𝑘𝑐 (𝑈) is in 𝑑 𝛺𝑘−1
𝑐 (𝑈). Prove that if the open set 𝑈 ⊂ 𝐑
𝑛−1
in
Exercise 3.2.iv has property 𝑃𝑘−1 then 𝑈 × 𝐴 has property 𝑃𝑘 .
Exercise 3.2.vii. Show that if 𝑄 is the rectangle [𝑎1 , 𝑏1 ] × ⋯ × [𝑎𝑛 , 𝑏𝑛 ] and 𝑈 = int 𝑄 then
𝑈 has property 𝑃𝑘 .
(3.2.12) 𝐇𝑛 ≔ { (𝑥1 , …, 𝑥𝑛 ) ∈ 𝐑𝑛 | 𝑥1 ≤ 0 }
and let 𝜔 ∈ 𝛺𝑛𝑐 (𝐑𝑛 ) be the 𝑛-form 𝜔 ≔ 𝑓𝑑𝑥1 ∧ ⋯ ∧ 𝑑𝑥𝑛 with 𝑓 ∈ 𝐶0∞ (𝐑𝑛 ). Define
where the right hand side is the usual Riemann integral of 𝑓 over 𝐇𝑛 . (This integral makes
sense since 𝑓 is compactly supported.) Show that if 𝜔 = 𝑑𝜇 for some 𝜇 ∈ 𝛺𝑛−1 𝑛
𝑐 (𝐑 ) then
(3.2.14) ∫ 𝜔=∫ 𝜄⋆ 𝜇
𝐇𝑛 𝐑𝑛−1
(𝑥2 , …, 𝑥𝑛 ) ↦ (0, 𝑥2 , …, 𝑥𝑛 ) .
̂ 𝑖 ⋯ ∧ 𝑑𝑥𝑛 . Mimicking the “(2)⇒(1)” part of the proof
Hint: Let 𝜇 = ∑𝑖 𝑓𝑖 𝑑𝑥1 ∧ ⋯ 𝑑𝑥
of Theorem 3.2.2 show that the integral (3.2.13) is the integral over 𝐑𝑛−1 of the function
0
𝜕𝑓1
∫ (𝑥 , 𝑥 , …, 𝑥𝑛 )𝑑𝑥1 .
−∞ 𝜕𝑥1 1 2
Draft: March 28, 2018
3.3. The Poincaré lemma for compactly supported forms on open subsets of 𝐑𝑛
In this section we will generalize Theorem 3.2.2 to arbitrary connected open subsets of
𝐑𝑛 .
Theorem 3.3.1 (Poincaré lemma for compactly supported forms). Let 𝑈 be a connected open
subset of 𝐑𝑛 and let 𝜔 be a compactly supported 𝑛-form with supp(𝜔) ⊂ 𝑈. The the following
assertions are equivalent:
(1) ∫𝐑𝑛 𝜔 = 0.
(2) There exists a compactly supported (𝑛 − 1)-form 𝜇 with supp 𝜇 ⊂ 𝑈 and 𝜔 = 𝑑𝜇.
Proof that (2)⇒(1). The support of 𝜇 is contained in a large rectangle, so the integral of 𝑑𝜇
is zero by Theorem 3.2.2. □
Proof that (1)⇒(2). Let 𝜔1 and 𝜔2 be compactly supported 𝑛-forms with support in 𝑈. We
will write
𝜔1 ∼ 𝜔2
as shorthand notation for the statement: There exists a compactly supported (𝑛 − 1)-form,
𝜇, with support in 𝑈 and with 𝜔1 − 𝜔2 = 𝑑𝜇. We will prove that (1)⇒(2) by proving an
equivalent statement: Fix a rectangle, 𝑄0 ⊂ 𝑈 and an 𝑛-form, 𝜔0 , with supp 𝜔0 ⊂ 𝑄0 and
integral equal to one.
Theorem 3.3.2. If 𝜔 is a compactly supported 𝑛-form with supp(𝜔) ⊂ 𝑈 and 𝑐 = ∫ 𝜔 then
𝜔 ∼ 𝑐𝜔0 .
Thus in particular if 𝑐 = 0, Theorem 3.3.2 says that 𝜔 ∼ 0 proving that (1)⇒(2). □
To prove Theorem 3.3.2 let 𝑄𝑖 ⊂ 𝑈, 𝑖 = 1, 2, 3, …, be a collection of rectangles with
∞
𝑈 = ⋃𝑖=1 int(𝑄𝑖 ) and let 𝜙𝑖 be a partition of unity with supp(𝜙𝑖 ) ⊂ int(𝑄𝑖 ). Replacing 𝜔
𝑚
by the finite sum ∑𝑖=1 𝜙𝑖 𝜔 for 𝑚 large, it suffices to prove Theorem 3.3.2 for each of the
summands 𝜙𝑖 𝜔. In other words we can assume that supp(𝜔) is contained in one of the open
rectangles int(𝑄𝑖 ). Denote this rectangle by 𝑄. We claim that one can join 𝑄0 to 𝑄 by a
sequence of rectangles as in the figure below.
𝑄0
Lemma 3.3.3. There exists a sequence of rectangles 𝑅0 , …, 𝑅𝑁+1 such that 𝑅0 = 𝑄0 , 𝑅𝑁+1 = 𝑄
and int(𝑅𝑖 ) ∩ int(𝑅𝑖+1 ) is nonempty.
Proof. Denote by 𝐴 the set of points, 𝑥 ∈ 𝑈, for which there exists a sequence of rectangles,
𝑅𝑖 , 𝑖 = 0, …, 𝑁 + 1 with 𝑅0 = 𝑄0 , with 𝑥 ∈ int 𝑅𝑁+1 and with int 𝑅𝑖 ∩ int 𝑅𝑖+1 nonempty.
It is clear that this set is open and that its complement is open; so, by the connectivity of 𝑈,
𝑈 = 𝐴. □
Draft: March 28, 2018
To prove Theorem 3.3.2 with supp 𝜔 ⊂ 𝑄, select, for each 𝑖, a compactly supported
𝑛-form 𝜈𝑖 with supp(𝜈𝑖 ) ⊂ int(𝑅𝑖 ) ∩ int(𝑅𝑖+1 ) and with ∫ 𝜈𝑖 = 1. The difference, 𝜈𝑖 − 𝜈𝑖+1
is supported in int 𝑅𝑖+1 , and its integral is zero. So by Theorem 3.2.2, 𝜈𝑖 ∼ 𝜈𝑖+1 . Similarly,
𝜔0 ∼ 𝜈0 and, setting 𝑐 ≔ ∫ 𝜔, we have 𝜔 ∼ 𝑐𝜈𝑁 . Thus
𝑐𝜔0 ∼ 𝑐𝜈0 ∼ ⋯ ∼ 𝑐𝜈𝑁 = 𝜔
proving the theorem.
(3.4.2) ∫ 𝑓⋆ 𝜔 = deg(𝑓) ∫ 𝜔
𝑈 𝑉
holds for all 𝜔 ∈ 𝛺𝑛𝑐 (𝑉).
Before we prove this assertion let’s see what this formula says in coordinates. If
𝜔 = 𝜙(𝑦)𝑑𝑦1 ∧ ⋯ ∧ 𝑑𝑦𝑛
then at 𝑥 ∈ 𝑈
𝑓⋆ 𝜔 = (𝜙 ∘ 𝑓)(𝑥) det(𝐷𝑓(𝑥))𝑑𝑥1 ∧ ⋯ ∧ 𝑑𝑥𝑛 .
Hence, in coordinates, equation (3.4.2) takes the form
∫ 𝑓⋆ 𝜔 = 𝑐 ∫ 𝑓⋆ 𝜔0 = deg(𝑓) ∫ 𝜔 . □
𝑈 𝑉
We will show in § 3.6 that the degree of 𝑓 is always an integer and explain why it is a
“topological” invariant of 𝑓.
Proposition 3.4.4. For the moment, however, we’ll content ourselves with pointing out a sim-
ple but useful property of this invariant. Let 𝑈, 𝑉 and 𝑊 be connected open subsets of 𝐑𝑛 and
𝑓 ∶ 𝑈 → 𝑉 and 𝑔 ∶ 𝑉 → 𝑊 proper 𝐶∞ maps. Then
(3.4.5) deg(𝑔 ∘ 𝑓) = deg(𝑔) deg(𝑓) .
Draft: March 28, 2018
= deg(𝑔) deg(𝑓) ∫ 𝜔 . □
𝑊
From this multiplicative property it is easy to deduce the following result (which we
will need in the next section).
Theorem 3.4.6. Let 𝐴 be a non-singular 𝑛 × 𝑛 matrix and 𝑓𝐴 ∶ 𝐑𝑛 → 𝐑𝑛 the linear mapping
associated with 𝐴. Then deg(𝑓𝐴 ) = +1 if det 𝐴 is positive and −1 if det 𝐴 is negative.
A proof of this result is outlined in Exercises 3.4.v to 3.4.ix below.
and evaluate the integral on the right by Fubini’s theorem; i.e., by first integrating with re-
spect to the 𝑥1 variable and then with respect to the remaining variables. Note that by equa-
tion (3.4.8)
∫ 𝑓(𝑥1 + 𝜆𝑥2 , 𝑥2 , …, 𝑥𝑛 ) 𝑑𝑥1 = ∫ 𝑓(𝑥1 , 𝑥2 , …, 𝑥𝑛 ) 𝑑𝑥1 .
Show that if ℓ1,1 ≠ 0 one can write 𝐿 as a product, 𝐿 = 𝐵𝐴𝐶, where 𝐴, 𝐵 and 𝐶 are linear
mappings of the form (3.4.10).
Hint: First solve the equations
ℓ𝑗,1 = 𝑏𝑗
Draft: March 28, 2018
Proof of Theorem 3.5.1. Given a point 𝑎1 ∈ 𝑈, let 𝑎2 = −𝑓(𝑎1 ) and for 𝑖 = 1, 2 let 𝑔𝑖 ∶ 𝐑𝑛 →
𝐑𝑛 be the translation, 𝑔𝑖 (𝑥) = 𝑥 + 𝑎𝑖 . By equation (3.4.2) and Exercise 3.4.iv the composite
diffeomorphism
(3.5.4) 𝑔2 ∘ 𝑓 ∘ 𝑔1
has the same degree as 𝑓, so it suffices to prove the theorem for this mapping. Notice however
that this mapping maps the origin onto the origin. Hence, replacing 𝑓 by this mapping, we
can, without loss of generality, assume that 0 is in the domain of 𝑓 and that 𝑓(0) = 0.
Next notice that if 𝐴 ∶ 𝐑𝑛 ⥲ 𝐑𝑛 is a bijective linear mapping the theorem is true for 𝐴
(by Exercise 3.4.ix), and hence if we can prove the theorem for 𝐴−1 ∘ 𝑓, equation (3.4.2) will
tell us that the theorem is true for 𝑓. In particular, letting 𝐴 = 𝐷𝑓(0), we have
𝐷(𝐴−1 ∘ 𝑓)(0) = 𝐴−1 𝐷𝑓(0) = id𝑛
where id𝑛 is the identity mapping. Therefore, replacing 𝑓 by 𝐴−1 ∘ 𝑓, we can assume that
the mapping 𝑓 (for which we are attempting to prove Theorem 3.5.1) has the properties:
𝑓(0) = 0 and 𝐷𝑓(0) = id𝑛 . Let 𝑔(𝑥) = 𝑓(𝑥) − 𝑥. Then these properties imply that 𝑔(0) = 0
and 𝐷𝑔(0) = 0. □
Lemma 3.5.5. There exists a 𝛿 > 0 such that |𝑔(𝑥)| ≤ 12 |𝑥| for |𝑥| ≤ 𝛿.
Proof. Let 𝑔(𝑥) = (𝑔1 (𝑥), …, 𝑔𝑛 (𝑥)). Then
𝜕𝑔𝑖
(0) = 0 ;
𝜕𝑥𝑗
so there exists a 𝛿 > 0 such that
𝜕𝑔𝑖 1
|
(𝑥)| ≤
𝜕𝑥𝑗 2
for |𝑥| ≤ 𝛿. However, by the mean value theorem,
𝑛
𝜕𝑔𝑖
𝑔𝑖 (𝑥) = ∑ (𝑐)𝑥𝑗
𝑗=1 𝜕𝑥𝑗
It is clear that
(3.5.6) ̃ =𝑥
𝑓(𝑥) for |𝑥| ≥ 𝛿
and, since 𝑓(𝑥) = 𝑥 + 𝑔(𝑥),
̃ = 𝑓(𝑥) 𝛿
(3.5.7) 𝑓(𝑥) for |𝑥| ≤ .
2
In addition, for all 𝑥 ∈ 𝐑𝑛 :
̃ 1
(3.5.8) |𝑓(𝑥)| ≥ |𝑥| .
2
̃
Indeed, by (3.5.6), |𝑓(𝑥)| ≥ |𝑥| for |𝑥| ≥ 𝛿, and for |𝑥| ≤ 𝛿
̃
|𝑓(𝑥)| ≥ |𝑥| − 𝜌(𝑥)|𝑔(𝑥)|
1
≥ |𝑥| − |𝑔(𝑥)| ≥ |𝑥| − |𝑥|
2
1
= |𝑥|
2
by Lemma 3.5.5.
Now let 𝑄𝑟 be the cube 𝑄𝑟 ≔ { 𝑥 ∈ 𝐑𝑛 | |𝑥| ≤ 𝑟 }, and let 𝑄𝑐𝑟 ≔ 𝐑𝑛 ∖ 𝑄𝑟 . From (3.5.8)
we easily deduce that
(3.5.9) ̃ (𝑄𝑟 ) ⊂ 𝑄2𝑟
𝑓−1
for all 𝑟, and hence that 𝑓 ̃ is proper. Also notice that for 𝑥 ∈ 𝑄𝛿 ,
̃ 3
|𝑓(𝑥)| ≤ |𝑥| + |𝑔(𝑥)| ≤ |𝑥|
2
by Lemma 3.5.5 and hence
(3.5.10) ̃ (𝑄𝑐3 ) ⊂ 𝑄𝑐 .
𝑓−1 𝛿 𝛿
2
Next let 𝜔 be a compactly supported 𝑛-form with support in 𝑄𝑐3𝛿/2 and with integral
equal to one. Then 𝑓⋆̃ 𝜔 is supported in 𝑄𝛿𝑐 by (3.5.10), and hence since 𝑓(𝑥) = 𝑥 on 𝑄𝛿𝑐 , we
have 𝑓⋆̃ 𝜔 = 𝜔. Thus
deg(𝑓)̃ = ∫ 𝑓⋆̃ 𝜔 = ∫ 𝜔 = 1 .
Putting these two identities together we conclude that deg(𝑓) = 1. □
If the function, 𝜙, in equation (3.5.4) is a 𝐶∞ function, the identity (3.5.3) is an im-
mediate consequence of the result above and the identity (3.4.3). If 𝜙 is not 𝐶∞ , but is just
continuous, we will deduce equation (3.5.4) from the following result.
Draft: March 28, 2018
(3.5.12) ∫ 𝜌(𝑦)𝑑𝑦 = 1 .
Set
𝜓(𝑥) = ∫ 𝜌(𝑦 − 𝑥)𝜙(𝑦)𝑑𝑦 .
and hence
𝜙(𝑥) = ∫ 𝜙(𝑥)𝜌(𝑦 − 𝑥)𝑑𝑦
so
𝜙(𝑥) − 𝜓(𝑥) = ∫(𝜙(𝑥) − 𝜙(𝑦))𝜌(𝑦 − 𝑥)𝑑𝑦
and
|𝜙(𝑥) − 𝜓(𝑥)| ≤ ∫ |𝜙(𝑥) − 𝜙(𝑦)|𝜌(𝑦 − 𝑥)𝑑𝑦 .
But 𝜌(𝑦 − 𝑥) = 0 for |𝑥 − 𝑦| ≥ 𝛿; and |𝜙(𝑥) − 𝜙(𝑦)| < 𝜀 for |𝑥 − 𝑦| ≤ 𝛿, so the integrand on
the right is less than
𝜀 ∫ 𝜌(𝑦 − 𝑥) 𝑑𝑦 ,
and hence by equation (3.5.13) we have
|𝜙(𝑥) − 𝜓(𝑥)| ≤ 𝜀 . □
To prove the identity (3.5.3), let 𝛾 ∶ 𝐑𝑛 → 𝐑 be a 𝐶∞ cut-off function which is one on
a neighborhood 𝑉1 of the support of 𝜙 is non-negative, and is compactly supported with
supp 𝛾 ⊂ 𝑉, and let
𝑐 = ∫ 𝛾(𝑦) 𝑑𝑦 .
By Theorem 3.5.11 there exists, for every 𝜀 > 0, a 𝐶∞ function 𝜓, with support on 𝑉1 satis-
fying
𝜀
(3.5.14) |𝜙 − 𝜓| ≤ .
2𝑐
Draft: March 28, 2018
Thus
|∫ (𝜙 − 𝜓)(𝑦)𝑑𝑦| ≤ ∫ |𝜙 − 𝜓|(𝑦)𝑑𝑦
𝑉 𝑉
≤ ∫ 𝛾|𝜙 − 𝜓|(𝑥𝑦)𝑑𝑦
𝑉
𝜀 𝜀
≤ ∫ 𝛾(𝑦)𝑑𝑦 ≤
2𝑐 2
so
𝜀
(3.5.15) |∫ 𝜙(𝑦) 𝑑𝑦 − ∫ 𝜓(𝑦) 𝑑𝑦| ≤ .
𝑉 𝑉 2
Similarly, the expression
∫ (ℎ ∘ 𝑓)(𝑥)|det 𝐷𝑓(𝑥)|𝑑𝑥
𝑈
is well-defined and these two integrals are equal.
Hint: If (𝜙𝑖 )𝑖≥1 is a partition of unity on 𝑉 then 𝜓𝑖 = 𝜙𝑖 ∘ 𝑓 is a partition of unity on 𝑈
and
∫ 𝜙𝑖 ℎ𝑑𝑦 = ∫ 𝜓𝑖 (ℎ ∘ 𝑓(𝑥))|det 𝐷𝑓(𝑥)|𝑑𝑥 .
Draft: March 28, 2018
(3.5.17) ∫ 𝑓⋆ 𝜔 = ∫ 𝜔.
𝑈∩𝐇𝑛 𝑉∩𝐇𝑛
Hint: Interpret the left and right hand sides of this formula as improper integrals over
𝑈 ∩ int(𝐇𝑛 ) and 𝑉 ∩ int(𝐇𝑛 ).
Exercise 3.5.v. The boundary of 𝐇𝑛 is the set
𝜕𝐇𝑛 ≔ { (0, 𝑥2 , …, 𝑥𝑛 ) | (𝑥2 , …, 𝑥𝑛 ) ∈ 𝐑𝑛−1 }
so the map
𝜄 ∶ 𝐑𝑛−1 → 𝐇𝑛 , (𝑥2 , …, 𝑥𝑛 ) ↦ (0, 𝑥2 , …, 𝑥𝑛 )
in Exercise 3.2.viii maps 𝐑𝑛−1 bijectively onto 𝜕𝐇𝑛 .
(1) Show that the map 𝑓 ∶ 𝑈 → 𝑉 in Exercise 3.5.iv maps 𝑈 ∩ 𝜕𝐇𝑛 onto 𝑉 ∩ 𝜕𝐇𝑛 .
(2) Let 𝑈′ = 𝜄−1 (𝑈) and 𝑉′ = 𝜄−1 (𝑉). Conclude from (1) that the restriction of 𝑓 to 𝑈∩𝜕𝐇𝑛
gives one a diffeomorphism
𝑔 ∶ 𝑈′ → 𝑉 ′
satisfying:
𝜄∘𝑔=𝑓∘𝜄.
(3) Let 𝜇 be in 𝛺𝑛−1
𝑐 (𝑉). Conclude from equations (3.2.14) and (3.5.17):
∫ 𝑔 ⋆ 𝜄⋆ 𝜇 = ∫ 𝜄⋆ 𝜇
𝑈′ 𝑉′
and in particular show that the diffeomorphism 𝑔 ∶ 𝑈′ → 𝑉′ is orientation-preserving.
𝑈 in the pre-image and hence, a fortiori, contains no critical points in its pre-image. Notice
also that 𝐶𝑓 can be quite large. For instance, if 𝑐 ∈ 𝑉 and 𝑓 ∶ 𝑈 → 𝑉 is the constant map
which maps all of 𝑈 onto 𝑐, then 𝐶𝑓 = 𝑈. However, in this example, 𝑓(𝐶𝑓 ) = {𝑐}, so the set
of regular values of 𝑓 is 𝑉 ∖ {𝑐}, and hence (in this example) is an open dense subset of 𝑉.
We will show that this is true in general.
Theorem 3.6.2 (Sard). If 𝑈 and 𝑉 are open subsets of 𝐑𝑛 and 𝑓 ∶ 𝑈 → 𝑉 a proper 𝐶∞ map,
the set of regular values of 𝑓 is an open dense subset of 𝑉.
We will defer the proof of this to Section 3.7 and in this section explore some of its
implications. Picking a regular value 𝑞 of 𝑓 we will prove:
Theorem 3.6.3. The set 𝑓−1 (𝑞) is a finite set. Moreover, if 𝑓−1 (𝑞) = {𝑝1 , …, 𝑝𝑛 } there exist
connected open neighborhoods 𝑈𝑖 of 𝑝𝑖 in 𝑌 and an open neighborhood 𝑊 of 𝑞 in 𝑉 such that:
(1) for 𝑖 ≠ 𝑗 the sets 𝑈𝑖 and 𝑈𝑗 are disjoint;
(2) 𝑓−1 (𝑊) = 𝑈1 ∪ ⋯ ∪ 𝑈𝑛 ,
(3) 𝑓 maps 𝑈𝑖 diffeomorphically onto 𝑊.
Proof. If 𝑝 ∈ 𝑓−1 (𝑞) then, since 𝑞 is a regular value, 𝑝 ∉ 𝐶𝑓 ; so
𝐷𝑓(𝑝) ∶ 𝐑𝑛 → 𝐑𝑛
is bijective. Hence by the inverse function theorem, 𝑓 maps a neighborhood, 𝑈𝑝 of 𝑝 diffeo-
morphically onto a neighborhood of 𝑞. The open sets
{ 𝑈𝑝 | 𝑝 ∈ 𝑓−1 (𝑞) }
are a covering of 𝑓−1 (𝑞); and, since 𝑓 is proper, 𝑓−1 (𝑞) is compact. Thus we can extract a
finite subcovering
{𝑈𝑝1 , …, 𝑈𝑝𝑁 }
and since 𝑝𝑖 is the only point in 𝑈𝑝𝑖 which maps onto 𝑞, we have that 𝑓−1 (𝑞) = {𝑝1 , …, 𝑝𝑁 }.
Without loss of generality we can assume that the 𝑈𝑝𝑖 ’s are disjoint from each other; for,
if not, we can replace them by smaller neighborhoods of the 𝑝𝑖 ’s which have this property.
By Theorem 3.4.7 there exists a connected open neighborhood 𝑊 of 𝑞 in 𝑉 for which
𝑓−1 (𝑊) ⊂ 𝑈𝑝1 ∪ ⋯ ∪ 𝑈𝑝𝑁 .
To conclude the proof let 𝑈𝑖 ≔ 𝑓−1 (𝑊) ∩ 𝑈𝑝𝑖 . □
The main result of this section is a recipe for computing the degree of 𝑓 by counting
the number of 𝑝𝑖 ’s above, keeping track of orientation.
Theorem 3.6.4. For each 𝑝𝑖 ∈ 𝑓−1 (𝑞) let 𝜎𝑝𝑖 = +1 if 𝑓 ∶ 𝑈𝑖 → 𝑊 is orientation-preserving
and −1 if 𝑓 ∶ 𝑈𝑖 → 𝑊 is orientation reversing. Then
𝑁
(3.6.5) deg(𝑓) = ∑ 𝜎𝑝𝑖 .
𝑖=1
Proof. Let 𝜔 be a compactly supported 𝑛-form on 𝑊 whose integral is one. Then
𝑁
deg(𝑓) = ∫ 𝑓⋆ 𝜔 = ∑ ∫ 𝑓⋆ 𝜔 .
𝑈 𝑖=1 𝑈𝑖
Since 𝑓 ∶ 𝑈𝑖 → 𝑊 is a diffeomorphism
1, 𝑓 is orientation-preserving
∫ 𝑓⋆ 𝜔 = ± ∫ 𝜔 = {
𝑈𝑖 𝑊 −1, 𝑓 is not orientation-preserving .
Draft: March 28, 2018
0 = ∫ 𝑓⋆ 𝜔 = deg(𝑓) ∫ 𝜔 = deg(𝑓) . □
𝑈 𝑉
Remark 3.6.7. In applications the contrapositive of Theorem 3.6.6 is much more useful
than the theorem itself.
Theorem 3.6.8. If deg(𝑓) ≠ 0, then 𝑓 maps 𝑈 surjectively onto 𝑉.
In other words if deg(𝑓) ≠ 0 the equation
𝑓(𝑥) = 𝑦
has a solution, 𝑥 ∈ 𝑈 for every 𝑦 ∈ 𝑉.
We will now show that the degree of 𝑓 is a topological invariant of 𝑓: if we deform 𝑓 by
a “homotopy” we do not change its degree. To make this assertion precise, let’s recall what
we mean by a homotopy between a pair of 𝐶∞ maps. Let 𝑈 be an open subset of 𝐑𝑚 , 𝑉
an open subset of 𝐑𝑛 , 𝐴 an open subinterval of 𝐑 containing 0 and 1, and 𝑓1 , 𝑓2 ∶ 𝑈 → 𝑉
a pair of 𝐶∞ maps. Then a 𝐶∞ map 𝐹 ∶ 𝑈 × 𝐴 → 𝑉 is a homotopy between 𝑓0 and 𝑓1 if
𝐹(𝑥, 0) = 𝑓0 (𝑥) and 𝐹(𝑥, 1) = 𝑓1 (𝑥). (See Definition 2.6.14.) Suppose now that 𝑓0 and 𝑓1 are
proper.
Definition 3.6.9. A homotopy 𝐹 between 𝑓0 and 𝑓1 is a proper homotopy if the map
𝐹♯ ∶ 𝑈 × 𝐴 → 𝑉 × 𝐴
defined by (𝑥, 𝑡) ↦ (𝐹(𝑥, 𝑡), 𝑡) is proper.
Note that if 𝐹 is a proper homotopy between 𝑓0 and 𝑓1 , then for every 𝑡 between 0 and
1, the map
𝑓𝑡 ∶ 𝑈 → 𝑉 , 𝑓𝑡 (𝑥) ≔ 𝐹(𝑥, 𝑡)
is proper.
Now let 𝑈 and 𝑉 be open subsets of 𝐑𝑛 .
Theorem 3.6.10. If 𝑓0 and 𝑓1 are properly homotopic, then deg(𝑓0 ) = deg(𝑓1 ).
Proof. Let
𝜔 = 𝜙(𝑦)𝑑𝑦1 ∧ ⋯ ∧ 𝑑𝑦𝑛
be a compactly supported 𝑛-form on 𝑉 whose integral over 𝑉 is 1. The the degree of 𝑓𝑡 is
equal to
Applications
We’ll conclude this account of degree theory by describing a couple applications.
Application 3.6.12 (The Brouwer fixed point theorem). Let 𝐵𝑛 be the closed unit ball in 𝐑𝑛 :
𝐵𝑛 ≔ { 𝑥 ∈ 𝐑𝑛 | ‖𝑥‖ ≤ 1 } .
Theorem 3.6.13. If 𝑓 ∶ 𝐵𝑛 → 𝐵𝑛 is a continuous mapping then 𝑓 has a fixed point, i.e., maps
some point, 𝑥0 ∈ 𝐵𝑛 onto itself.
The idea of the proof will be to assume that there isn’t a fixed point and show that this
leads to a contradiction. Suppose that for every point 𝑥 ∈ 𝐵𝑛 we have 𝑓(𝑥) ≠ 𝑥. Consider
the ray through 𝑓(𝑥) in the direction of 𝑥:
𝑓(𝑥) + 𝑠(𝑥 − 𝑓(𝑥)) , 𝑠 ∈ [0, ∞) .
This ray intersects the boundary 𝑆𝑛−1 ≔ 𝜕𝐵𝑛 in a unique point 𝛾(𝑥) (see Figure 3.6.1 below);
and one of the exercises at the end of this section will be to show that the mapping 𝛾 ∶ 𝐵𝑛 →
𝑆𝑛−1 given by 𝑥 ↦ 𝛾(𝑥), is a continuous mapping. Also it is clear from Figure 3.6.1 that
𝛾(𝑥) = 𝑥 if 𝑥 ∈ 𝑆𝑛−1 , so we can extend 𝛾 to a continuous mapping of 𝐑𝑛 into 𝐑𝑛 by letting 𝛾
be the identity for ‖𝑥‖ ≥ 1. Note that this extended mapping has the property
‖𝛾(𝑥)‖ ≥ 1
𝑛
for all 𝑥 ∈ 𝐑 and
(3.6.14) 𝛾(𝑥) = 𝑥
for all ‖𝑥‖ ≥ 1. To get a contradiction we’ll show that 𝛾 can be approximated by a 𝐶∞ map
which has similar properties. For this we will need the following corollary of Theorem 3.5.11.
Lemma 3.6.15. Let 𝑈 be an open subset of 𝐑𝑛 , 𝐶 a compact subset of 𝑈 and 𝜙 ∶ 𝑈 → 𝐑 a
continuous function which is 𝐶∞ on the complement of 𝐶. Then for every 𝜀 > 0, there exists a
𝐶∞ function 𝜓 ∶ 𝑈 → 𝐑, such that 𝜙 − 𝜓 has compact support and |𝜙 − 𝜓| < 𝜀.
Proof. Let 𝜌 be a bump function which is in 𝐶0∞ (𝑈) and is equal to 1 on a neighborhood
of 𝐶. By Theorem 3.5.11 there exists a function 𝜓0 ∈ 𝐶0∞ (𝑈) such that |𝜌𝜙 − 𝜓0 | < 𝜀. To
complete the proof, let 𝜓(1 − 𝜌)𝜙 + 𝜓0 , and note that
𝜙 − 𝜓 = (1 − 𝜌)𝜙 + 𝜌𝜙 − (1 − 𝜌)𝜙 − 𝜓0
= 𝜌𝜙 − 𝜓0 . □
∞
By applying Lemma 3.6.15 to each of the coordinates of the map 𝛾, one obtains a 𝐶
map 𝑔 ∶ 𝐑𝑛 → 𝐑𝑛 such that
‖𝑔 − 𝛾‖ < 𝜀 < 1
and such that 𝑔 = 𝛾 on the complement of a compact set. However, by (3.6.14), this means
that 𝑔 is equal to the identity on the complement of a compact set and hence (see Exer-
cise 3.6.ix) that 𝑔 is proper and deg(𝑔) = 1. On the other hand by (3.6.19) and (3.6.14) we
have ‖𝑔(𝑥)‖ > 1 − 𝜀 for all 𝑥 ∈ 𝐑𝑛 , so 0 ∉ im(𝑔) and hence by Theorem 3.6.4 we have
deg(𝑔) = 0, which provides the desired contradiction.
𝑓(𝑥)
𝛾(𝑥)
with 𝐑2 via the map 𝐑2 → 𝐂 given by (𝑥, 𝑦) ↦ 𝑧 = 𝑥 + 𝑖𝑦, we can think of 𝑝 as defining a
mapping
𝑝 ∶ 𝐑2 → 𝐑2 , 𝑧 ↦ 𝑝(𝑧) .
We will prove:
Theorem 3.6.17. The mapping 𝑝 ∶ 𝐑2 → 𝐑2 is proper and deg(𝑝) = 𝑛.
Proof. For 𝑡 ∈ 𝐑 let
𝑛−1
𝑝𝑡 (𝑧) ≔ (1 − 𝑡)𝑧𝑛 + 𝑡𝑝(𝑧) = 𝑧𝑛 + 𝑡 ∑ 𝑎𝑖 𝑧𝑖 .
𝑖=0
We will show that the mapping
𝑔 ∶ 𝐑2 × 𝐑 → 𝐑2 , (𝑧, 𝑡) ↦ 𝑝𝑡 (𝑧)
is a proper homotopy. Let
𝐶 = sup |𝑎𝑖 | .
0≤𝑖≤𝑛−1
Then for |𝑧| ≥ 1 we have
|𝑎0 + ⋯ + 𝑎𝑛−1 𝑧𝑛−1 | ≤ |𝑎0 | + |𝑎1 ||𝑧| + ⋯ + |𝑎𝑛−1 | |𝑧|𝑛−1
≤ 𝐶𝑛|𝑧|𝑛−1 ,
and hence, for |𝑡| ≤ 𝑎 and |𝑧| ≥ 2𝑎𝐶𝑛,
|𝑝𝑡 (𝑧)| ≥ |𝑧|𝑛 − 𝑎𝐶𝑛|𝑧|𝑛−1
≥ 𝑎𝐶𝑛|𝑧|𝑛−1 .
If 𝐴 ⊂ 𝐂 is compact, then for some 𝑅 > 0, 𝐴 is contained in the disk defined by |𝑤| ≤ 𝑅,
and hence the set
{ 𝑧 ∈ 𝐂 | (𝑡, 𝑝𝑡 (𝑧)) ∈ [−𝑎, 𝑎] × 𝐴 }
is contained in the compact set
{ 𝑧 ∈ 𝐂 | 𝑎𝐶|𝑧|𝑛−1 ≤ 𝑅 } .
This shows that 𝑔 is a proper homotopy.
Thus for each 𝑡 ∈ 𝐑, the map 𝑝𝑡 ∶ 𝐂 → 𝐂 is proper and
deg(𝑝𝑡 ) = deg(𝑝1 ) = deg(𝑝) = deg(𝑝0 )
However, 𝑝0 ∶ 𝐂 → 𝐂 is just the mapping 𝑧 ↦ 𝑧𝑛 and an elementary computation (see
Exercises 3.6.v and 3.6.vi below) shows that the degree of this mapping is 𝑛. □
Draft: March 28, 2018
In particular for 𝑛 > 0 the degree of 𝑝 is nonzero; so by Theorem 3.6.4 we conclude that
𝑝 ∶ 𝐂 → 𝐂 is surjective and hence has zero in its image.
Theorem 3.6.18 (Fundamental theorem of algebra). Every positive-degree polynomial
𝑝(𝑧) = 𝑧𝑛 + 𝑎𝑛−1 𝑧𝑛−1 + ⋯ + 𝑎0
with complex coefficients has a complex root: 𝑝(𝑧0 ) = 0 for some 𝑧0 ∈ 𝐂.
Exercises for §3.6
Exercise 3.6.i. Let 𝑊 be a subset of 𝐑𝑛 and let 𝑎(𝑥), 𝑏(𝑥) and 𝑐(𝑥) be real-valued functions
on 𝑊 of class 𝐶𝑟 . Suppose that for every 𝑥 ∈ 𝑊 the quadratic polynomial
𝑎(𝑥)𝑠2 + 𝑏(𝑥)𝑠 + 𝑐(𝑥)
has two distinct real roots, 𝑠+ (𝑥) and 𝑠− (𝑥), with 𝑠+ (𝑥) > 𝑠− (𝑥). Prove that 𝑠+ and 𝑠− are
functions of class 𝐶𝑟 .
Hint: What are the roots of the quadratic polynomial: 𝑎𝑠2 + 𝑏𝑠 + 𝑐?
Exercise 3.6.ii. Show that the function 𝛾(𝑥) defined in Figure 3.6.1 is a continuous surjec-
tion 𝐵𝑛 → 𝑆2𝑛−1 .
Hint: 𝛾(𝑥) lies on the ray,
𝑓(𝑥) + 𝑠(𝑥 − 𝑓(𝑥)) , 𝑠 ∈ [0, ∞)
and satisfies ‖𝛾(𝑥)‖ = 1. Thus
𝛾(𝑥) = 𝑓(𝑥) + 𝑠0 (𝑥 − 𝑓(𝑥)) ,
where 𝑠0 is a non-negative root of the quadratic polynomial
‖𝑓(𝑥) + 𝑠(𝑥 − 𝑓(𝑥))‖2 − 1 .
Argue from Figure 3.6.1 that this polynomial has to have two distinct real roots.
Exercise 3.6.iii. Show that the Brouwer fixed point theorem isn’t true if one replaces the
closed unit ball by the open unit ball.
Hint: Let 𝑈 be the open unit ball (i.e., the interior of 𝐵𝑛 ). Show that the map
𝑥
ℎ ∶ 𝑈 → 𝐑𝑛 , ℎ(𝑥) ≔
1 − ‖𝑥‖2
is a diffeomorphism of 𝑈 onto 𝐑𝑛 , and show that there are lots of mappings of 𝐑𝑛 onto 𝐑𝑛
which do not have fixed points.
Exercise 3.6.iv. Show that the fixed point in the Brouwer theorem doesn’t have to be an
interior point of 𝐵𝑛 , i.e., show that it can lie on the boundary.
Exercise 3.6.v. If we identify 𝐂 with 𝐑2 via the mapping (𝑥, 𝑦) ↦ 𝑥 + 𝑖𝑦, we can think of a
𝐂-linear mapping of 𝐂 into itself, i.e., a mapping of the form
𝑧 ↦ 𝑐𝑧
for a fixed 𝑐 ∈ 𝐂, as an 𝐑-linear mapping of 𝐑2 into itself. Show that the determinant of this
mapping is |𝑐|2 .
Exercise 3.6.vi.
Draft: March 28, 2018
(1) Let 𝑓 ∶ 𝐂 → 𝐂 be the mapping 𝑓(𝑧) ≔ 𝑧𝑛 . Show that 𝐷𝑓(𝑧) is the linear map
𝐷𝑓(𝑧) = 𝑛𝑧𝑛−1
given by multiplication by 𝑛𝑧𝑛−1 .
Hint: Argue from first principles. Show that for ℎ ∈ 𝐂 = 𝐑2
(𝑧 + ℎ)𝑛 − 𝑧𝑛 − 𝑛𝑧𝑛−1 ℎ
|ℎ|
tends to zero as |ℎ| → 0.
(2) Conclude from Exercise 3.6.v that
det(𝐷𝑓(𝑧)) = 𝑛2 |𝑧|2𝑛−2 .
(3) Show that at every point 𝑧 ∈ 𝐂 ∖ {0}, 𝑓 is orientation preserving.
(4) Show that every point, 𝑤 ∈ 𝐂 ∖ {0} is a regular value of 𝑓 and that
𝑓−1 (𝑤) = {𝑧1 , …, 𝑧𝑛 }
with 𝜎𝑧𝑖 = +1.
(5) Conclude that the degree of 𝑓 is 𝑛.
Exercise 3.6.vii. Prove that the map 𝑓 from Exercise 3.6.vi has degree 𝑛 by deducing this
directly from the definition of degree.
Hints:
• Show that in polar coordinates, 𝑓 is the map (𝑟, 𝜃) ↦ (𝑟𝑛 , 𝑛𝜃).
• Let 𝜔 be the 2-form 𝜔 ≔ 𝑔(𝑥2 + 𝑦2 )𝑑𝑥 ∧ 𝑑𝑦, where 𝑔(𝑡) is a compactly supported
𝐶∞ function of 𝑡. Show that in polar coordinates 𝜔 = 𝑔(𝑟2 )𝑟𝑑𝑟 ∧ 𝑑𝜃, and compute
the degree of 𝑓 by computing the integrals of 𝜔 and 𝑓⋆ 𝜔 in polar coordinates and
comparing them.
Exercise 3.6.viii. Let 𝑈 be an open subset of 𝐑𝑛 , 𝑉 an open subset of 𝐑𝑚 , 𝐴 an open subin-
terval of 𝐑 containing 0 and 1, 𝑓0 , 𝑓1 ∶ 𝑈 → 𝑉 a pair of 𝐶∞ mappings, and 𝐹 ∶ 𝑈 × 𝐴 → 𝑉
a homotopy between 𝑓0 and 𝑓1 .
(1) In Exercise 2.4.iv you proved that if 𝜇 ∈ 𝛺𝑘 (𝑉) and 𝑑𝜇 = 0, then
(3.6.19) 𝑓0⋆ 𝜇 − 𝑓1⋆ 𝜇 = 𝑑𝜈
where 𝜈 is the (𝑘 − 1)-form 𝑄𝛼 in equation (2.6.17). Show (by careful inspection of the
definition of 𝑄𝛼) that if 𝐹 is a proper homotopy and 𝜇 ∈ 𝛺𝑘𝑐 (𝑉) then 𝜈 ∈ 𝛺𝑘−1
𝑐 (𝑈).
(2) Suppose in particular that 𝑈 and 𝑉 are open subsets of 𝐑𝑛 and 𝜇 is in 𝛺𝑛𝑐 (𝑉). Deduce
from equation (3.6.19) that
∫ 𝑓0⋆ 𝜇 = ∫ 𝑓1⋆ 𝜇
and deduce directly from the definition of degree that degree is a proper homotopy
invariant.
Exercise 3.6.ix. Let 𝑈 be an open connected subset of 𝐑𝑛 and 𝑓 ∶ 𝑈 → 𝑈 a proper 𝐶∞
map. Prove that if 𝑓 is equal to the identity on the complement of a compact set 𝐶, then 𝑓
is proper and deg(𝑓) = 1.
Hints:
• Show that for every subset 𝐴 ⊂ 𝑈, we have 𝑓−1 (𝐴) ⊂ 𝐴 ∪ 𝐶, and conclude from
this that 𝑓 is proper.
• Use the recipe (1.6.2) to compute deg(𝑓) with 𝑞 ∈ 𝑈 ∖ 𝑓(𝐶).
Draft: March 28, 2018
Exercise 3.6.x. Let (𝑎𝑖,𝑗 ) be an 𝑛 × 𝑛 matrix and 𝐴 ∶ 𝐑𝑛 → 𝐑𝑛 the linear mapping associated
with this matrix. Frobenius’ Theorem asserts: If the 𝑎𝑖,𝑗 are non-negative then 𝐴 has a non-
negative eigenvalue. In other words there exists a 𝑣 ∈ 𝐑𝑛 and a 𝜆 ∈ 𝐑, 𝜆 ≥ 0, such that
𝐴𝑣 = 𝜆𝑣. Deduce this linear algebra result from the Brouwer fixed point theorem.
Hints:
• We can assume that 𝐴 is bijective, otherwise 0 is an eigenvalue. Let 𝑆𝑛−1 be the
(𝑛 − 1)-sphere, defined by |𝑥| = 1, and 𝑓 ∶ 𝑆𝑛−1 → 𝑆𝑛−1 the map,
𝐴𝑥
𝑓(𝑥) = .
‖𝐴𝑥‖
Show that 𝑓 maps the set
𝑄 ≔ { (𝑥1 , …, 𝑥𝑛 ) ∈ 𝑆𝑛−1 | 𝑥1 , …, 𝑥𝑛 ≥ 0 }
into itself.
• It is easy to prove that 𝑄 is homeomorphic to the unit ball 𝐵𝑛−1 , i.e., that there
exists a continuous map 𝑔 ∶ 𝑄 → 𝐵𝑛−1 which is invertible and has a continuous
inverse. Without bothering to prove this fact deduce from it Frobenius’ theorem.
Proof. Let the length of each of the sides of 𝐴 be ℓ. Given 𝛿 > 0 one can subdivide 𝐴 into 𝑁𝑛
cubes, each of volume (ℓ/𝑁)𝑛 , such that if 𝑥 and 𝑦 are points of any one of these subcubes
𝜕𝑓𝑖 𝜕𝑓
(3.7.8) | (𝑥) − 𝑖 (𝑦)| < 𝛿 .
𝜕𝑥𝑗 𝜕𝑥𝑗
♯ ♯
Let 𝐴 1 , …, 𝐴 𝑚 be the cubes in this collection which intersect 𝐶𝑓 . Then for 𝑧0 ∈ 𝐴 𝑖 ∩ 𝐶𝑓 ,
𝜕𝑓𝑖
(𝑧 )
𝜕𝑥𝑗 0
= 0, so for 𝑧 ∈ 𝐴 𝑖 we have
𝜕𝑓𝑖
(3.7.9) | (𝑧)| < 𝛿
𝜕𝑥𝑗
by equation (3.7.8). If 𝑥 and 𝑦 are points of 𝐴 𝑖 then by the mean value theorem there exists
a point 𝑧 on the line segment joining 𝑥 to 𝑦 such that
𝑛
𝜕𝑓𝑖
𝑓𝑖 (𝑥) − 𝑓𝑖 (𝑦) = ∑ (𝑧)(𝑥𝑗 − 𝑦𝑗 )
𝑗=1 𝜕𝑥𝑗
for all 𝑝 ∈ 𝑈 so we’re reduced to proving Theorem 3.7.1 for maps 𝑓 ∶ 𝑈 → 𝐑𝑛 having the
property (3.7.11). Let 𝑔 ∶ 𝑈 → 𝐑𝑛 be defined by
(3.7.12) 𝑔(𝑥1 , …, 𝑥𝑛 ) = (𝑓1 (𝑥), 𝑥2 , …, 𝑥𝑛 ) .
Then
𝑔⋆ 𝑥1 = 𝑓⋆ 𝑥1 = 𝑓1 (𝑥1 , …, 𝑥𝑛 )
and
𝜕𝑓1
det(𝐷𝑔) = ≠0.
𝜕𝑥1
Thus, by the inverse function theorem, 𝑔 is locally a diffeomorphism at every point 𝑝 ∈ 𝑈.
This means that if 𝐴 is a compact subset of 𝑈 we can cover 𝐴 by a finite number of open
subsets 𝑈𝑖 ⊂ 𝑈 such that 𝑔 maps 𝑈𝑖 diffeomorphically onto an open subset 𝑊𝑖 in 𝐑𝑛 . To
conclude the proof of the theorem we’ll show that 𝐑𝑛 ∖ 𝑓(𝐶𝑓 ∩ 𝑈𝑖 ∩ 𝐴) is a dense subset of
𝐑𝑛 . Let ℎ ∶ 𝑊𝑖 → 𝐑𝑛 be the map ℎ = 𝑓∘𝑔−1 . To prove this assertion it suffices by Remark 3.7.4
to prove that the set 𝐑𝑛 ∖ ℎ(𝐶ℎ ) is dense in 𝐑𝑛 . This we will do by induction on 𝑛. First note
♯
that for 𝑛 = 1, we have 𝐶𝑓 = 𝐶𝑓 , so we’ve already proved Theorem 3.7.1 in dimension one.
Now note that by (3.7.12) we have ℎ⋆ 𝑥1 = 𝑥1 , i.e., ℎ is a mapping of the form
ℎ(𝑥1 , …, 𝑥𝑛 ) = (𝑥1 , ℎ2 (𝑥), …, ℎ𝑛 (𝑥)) .
Thus if we let
𝑊𝑐 ≔ { (𝑥2 , …, 𝑥𝑛 ) ∈ 𝐑𝑛−1 | (𝑐, 𝑥2 , …, 𝑥𝑛 ) ∈ 𝑊𝑖 }
and let ℎ𝑐 ∶ 𝑊𝑐 → 𝐑𝑛−1 be the map
ℎ𝑐 (𝑥2 , …, 𝑥𝑛 ) = (ℎ2 (𝑐, 𝑥2 , …, 𝑥𝑛 ), …, ℎ𝑛 (𝑐, 𝑥2 , …, 𝑥𝑛 )) .
Then
det(𝐷ℎ𝑐 )(𝑥2 , …, 𝑥𝑛 ) = det(𝐷ℎ)(𝑐, 𝑥2 , …, 𝑥𝑛 ) ,
and hence
(3.7.13) (𝑐, 𝑥) ∈ 𝑊𝑖 ∩ 𝐶ℎ ⟺ 𝑥 ∈ 𝐶ℎ𝑐 .
Now let 𝑝0 = (𝑐, 𝑥0 ) be a point in 𝐑𝑛 . We have to show that every neighborhood 𝑉 of 𝑝0
contains a point 𝑝 ∈ 𝐑𝑛 ∖ ℎ(𝐶ℎ ). Let 𝑉𝑐 ⊂ 𝐑𝑛−1 be the set of points 𝑥 for which (𝑐, 𝑥) ∈ 𝑉. By
induction 𝑉𝑐 contains a point 𝑥 ∈ 𝐑𝑛−1 ∖ ℎ𝑐 (𝐶ℎ𝑐 ) and hence 𝑝 = (𝑐, 𝑥) is in 𝑉 by definition
and in 𝐑𝑛 ∖ ℎ(𝐶ℎ ) by (3.7.13). □
Exercises for §3.7
Exercise 3.7.i. What are the set of critical points and the image of the set of critical points
for the following maps 𝐑 → 𝐑?
(1) The map 𝑓1 (𝑥) = (𝑥2 − 1)2 .
(2) The map 𝑓2 (𝑥) = sin(𝑥) + 𝑥.
(3) The map
0, 𝑥≤0
𝑓3 (𝑥) = { − 1
𝑒 𝑥, 𝑥 > 0 .
Exercise 3.7.ii (Sard’s theorem for affine maps). Let 𝑓 ∶ 𝐑𝑛 → 𝐑𝑛 be an affine map, i.e., a
map of the form
𝑓(𝑥) = 𝐴(𝑥) + 𝑥0
where 𝐴 ∶ 𝐑𝑛 → 𝐑𝑛 is a linear map and 𝑥0 ∈ 𝐑𝑛 . Prove Sard’s theorem for 𝑓.
Draft: March 28, 2018
Exercise 3.7.iii. Let 𝜌 ∶ 𝐑 → 𝐑 be a 𝐶∞ function which is supported in the interval (−1/2, 1/2)
and has a maximum at the origin. Let (𝑟𝑖 )𝑖≥1 be an enumeration of the rational numbers, and
let 𝑓 ∶ 𝐑 → 𝐑 be the map
∞
𝑓(𝑥) = ∑ 𝑟𝑖 𝜌(𝑥 − 𝑖) .
𝑖=1
Show that 𝑓 is a 𝐶∞ map and show that the image of 𝐶𝑓 is dense in 𝐑.
The moral of this example: Sard’s theorem says that the complement of 𝐶𝑓 is dense in 𝐑,
but 𝐶𝑓 can be dense as well.
Exercise 3.7.iv. Prove the assertion made in Remark 3.7.4.
Hint: You need to show that for every point 𝑝 ∈ 𝐑𝑛 and every neighborhood 𝑉 of 𝑝,
𝑉 ∩ ⋂𝑛≥1 𝑈𝑛 is nonempty. Construct, by induction, a family of closed balls (𝐵𝑘 )𝑘≥1 such that
• 𝐵𝑘 ⊂ 𝑉,
• 𝐵𝑘+1 ⊂ 𝐵𝑘 ,
• 𝐵𝑘 ⊂ ⋂𝑛≤𝑘 𝑈𝑛 ,
• The radius of 𝐵𝑘 is less than 𝑘1 ,
and show that ⋂𝑘≥1 𝐵𝑘 ≠ ∅.
Exercise 3.7.v. Verify equation (3.7.5).
Draft: March 28, 2018
Draft: March 28, 2018
CHAPTER 4
4.1. Manifolds
Our agenda in this chapter is to extend to manifolds the results of Chapters 2 and 3 and
to formulate and prove manifold versions of two of the fundamental theorems of integral
calculus: Stokes’ theorem and the divergence theorem. In this section we’ll define what we
mean by the term manifold, however, before we do so, a word of encouragement. Having
had a course in multivariable calculus, you are already familiar with manifolds, at least in
their one and two-dimensional emanations, as curves and surfaces in 𝐑3 , i.e., a manifold is
basically just an 𝑛-dimensional surface in some high dimensional Euclidean space. To make
this definition precise let 𝑋 be a subset of 𝐑𝑁 , 𝑌 a subset of 𝐑𝑛 and 𝑓 ∶ 𝑋 → 𝑌 a continuous
map. We recall:
Definition 4.1.1. The map 𝑓 is a 𝐶∞ map if for every 𝑝 ∈ 𝑋, there exists a neighborhood
𝑈𝑝 of 𝑝 in 𝐑𝑁 and a 𝐶∞ map 𝑔𝑝 ∶ 𝑈𝑝 → 𝐑𝑛 which coincides with 𝑓 on 𝑈𝑝 ∩ 𝑋.
We also recall:
Theorem 4.1.2. If 𝑓 ∶ 𝑋 → 𝑌 is a 𝐶∞ map, there exists a neighborhood 𝑈, of 𝑋 in 𝐑𝑁 and a
𝐶∞ map 𝑔 ∶ 𝑈 → 𝐑𝑛 such that 𝑔 coincides with 𝑓 on 𝑋.
(A proof of this can be found in Appendix a.)
We will say that 𝑓 is a diffeomorphism if 𝑓 is a bijection and 𝑓 and 𝑓−1 are both 𝐶∞
maps. In particular if 𝑌 is an open subset of 𝐑𝑛 , 𝑋 is an example of an object which we will
call a manifold. More generally:
Definition 4.1.3. Let 𝑁 and 𝑛 be nonnegative integers with 𝑛 ≤ 𝑁. A subset 𝑋 ⊂ 𝐑𝑁 is an
𝑛-manifold if for every 𝑝 ∈ 𝑋 there exists a neighborhood 𝑉 of 𝑝 in 𝐑𝑁 , an open subset 𝑈
in 𝐑𝑛 , and a diffeomorphism 𝜙 ∶ 𝑈 ⥲ 𝑋 ∩ 𝑉.
Thus 𝑋 is an 𝑛-manifold if, locally near every point 𝑝, 𝑋 “looks like” an open subset of
𝐑𝑛 .
Examples 4.1.4.
(1) Graphs of functions: Let 𝑈 be an open subset of 𝐑𝑛 and 𝑓 ∶ 𝑈 → 𝐑 a 𝐶∞ function. Its
graph
𝛤𝑓 = { (𝑥, 𝑓(𝑥)) ∈ 𝐑𝑛+1 | 𝑥 ∈ 𝑈 }
is an 𝑛-manifold in 𝐑𝑛+1 . In fact the map
𝜙 ∶ 𝑈 → 𝐑𝑛+1 , 𝑥 ↦ (𝑥, 𝑓(𝑥))
is a diffeomorphism of 𝑈 onto 𝛤𝑓 . (It’s clear that 𝜙 is a 𝐶∞ map, and it is a diffeomor-
phism since its inverse is the map 𝜋 ∶ 𝛤𝑓 → 𝑈 given by 𝜋(𝑥, 𝑡) ≔ 𝑥, which is also clearly
𝐶∞ .)
97
Draft: March 28, 2018
§4.1 Manifolds 99
(7) The 2-torus: In calculus books this is usually described as the surface of rotation in 𝐑3
obtained by taking the unit circle centered at the point, (2, 0), in the (𝑥1 , 𝑥3 ) plane and
rotating it about the 𝑥3 -axis. However, a slightly nicer description of it is as the product
manifold 𝑆1 × 𝑆1 in 𝐑4 . (As an exercise, reconcile these two descriptions.)
We’ll now turn to an alternative way of looking at manifolds: as solutions of systems of
equations. Let 𝑈 be an open subset of 𝐑𝑁 and 𝑓 ∶ 𝑈 → 𝐑𝑘 a 𝐶∞ map.
Definition 4.1.6. A point 𝑎 ∈ 𝐑𝑘 is a regular value of 𝑓 if for every point 𝑝 ∈ 𝑓−1 (𝑎), the
map 𝑓 is a submersion at 𝑝.
Some examples:
Examples 4.1.9.
(1) The 𝑛-sphere: Let
𝑓 ∶ 𝐑𝑛+1 → 𝐑
be the map,
(𝑥1 , …, 𝑥𝑛+1 ) ↦ 𝑥12 + ⋯ + 𝑥𝑛+1
2
−1.
Then
𝐷𝑓(𝑥) = 2(𝑥1 , …, 𝑥𝑛+1 )
so, if 𝑥 ≠ 0, then 𝑓 is a submersion at 𝑥. In particular, 𝑓 is a submersion at all points 𝑥
on the 𝑛-sphere
𝑆𝑛 = 𝑓−1 (0)
so the 𝑛-sphere is an 𝑛-dimensional submanifold of 𝐑𝑛+1 .
(2) Graphs: Let 𝑔 ∶ 𝐑𝑛 → 𝐑𝑘 be a 𝐶∞ map and, as in (2), let
𝛤𝑔 ≔ { (𝑥, 𝑦) ∈ 𝐑𝑛 × 𝐑𝑘 | 𝑦 = 𝑔(𝑥) } .
Proof. Let
𝑓 ∶ 𝐑 𝑛 × 𝐑𝑘 → 𝐑 𝑘
be the map 𝑓(𝑥, 𝑦) ≔ 𝑦 − 𝑔(𝑥). Then
𝐷𝑓(𝑥, 𝑦) = (−𝐷𝑔(𝑥) id𝑘 )
where id𝑘 is the identity map of 𝐑𝑘 onto itself. This map is always of rank 𝑘. Hence
𝛤𝑓 = 𝑓−1 (0) is an 𝑛-dimensional submanifold of 𝐑𝑛+𝑘 . □
(3) Let M 𝑛 be the set of all 𝑛 × 𝑛 matrices and let 𝒮𝑛 be the set of all symmetric 𝑛 × 𝑛
matrices, i.e., the set
𝒮𝑛 = { 𝐴 ∈ M 𝑛 | 𝐴 = 𝐴⊺ } .
The map
𝐴 = (𝑎𝑖,𝑗 ) ↦ (𝑎1,1 , 𝑎1,2 , …, 𝑎1,𝑛 , 𝑎2,1 , …, 𝑎2,𝑛 , …)
gives us an identification
2
M 𝑛 ⥲ 𝐑𝑛
and the map
𝐴 = (𝑎𝑖,𝑗 ) ↦ (𝑎1,1 , …, 𝑎1,𝑛 , 𝑎2,2 , …, 𝑎2,𝑛 , 𝑎3,3 , …, 𝑎3,𝑛 , …)
gives us an identification
𝑛(𝑛+1)
𝒮𝑛 ⥲ 𝐑 2 .
(Note that if 𝐴 is a symmetric matrix, then 𝑎𝑖,𝑗 = 𝑎𝑗,𝑖 , so this map avoids redundancies.)
Let
𝑂(𝑛) = { 𝐴 ∈ M 𝑛 | 𝐴⊺ 𝐴 = id𝑛 } .
This is the set of orthogonal 𝑛 × 𝑛 matrices, and we will leave for you as an exercise to
show that it’s an 𝑛(𝑛 − 1)/2-manifold.
Hint: Let 𝑓 ∶ M 𝑛 → 𝒮𝑛 be the map 𝑓(𝐴) = 𝐴⊺ 𝐴 − id𝑛 . Then
𝑂(𝑛) = 𝑓−1 (0) .
These examples show that lots of interesting manifolds arise as zero sets of submersions
𝑓 ∶ 𝑈 → 𝐑𝑘 . This is, in fact, not just an accident. We will show that locally every manifold
arises this way. More explicitly let 𝑋 ⊂ 𝐑𝑁 be an 𝑛-manifold, 𝑝 a point of 𝑋, 𝑈 a neighbor-
hood of 0 in 𝐑𝑛 , 𝑉 a neighborhood of 𝑝 in 𝐑𝑁 and 𝜙 ∶ (𝑈, 0) ⥲ (𝑉 ∩ 𝑋, 𝑝) a diffeomor-
phism. We will for the moment think of 𝜙 as a 𝐶∞ map 𝜙 ∶ 𝑈 → 𝐑𝑁 whose image happens
to lie in 𝑋.
Lemma 4.1.10. The linear map 𝐷𝜙(0) ∶ 𝐑𝑛 → 𝐑𝑁 is injective.
Proof. Since 𝜙−1 ∶ 𝑉 ∩ 𝑋 ⥲ 𝑈 is a diffeomorphism, shrinking 𝑉 if necessary, we can assume
that there exists a 𝐶∞ map 𝜓 ∶ 𝑉 → 𝑈 which coincides with 𝜙−1 on 𝑉 ∩ 𝑋. Since 𝜙 maps 𝑈
onto 𝑉 ∩ 𝑋, 𝜓 ∘ 𝜙 = 𝜙−1 ∘ 𝜙 is the identity map on 𝑈. Therefore,
𝐷(𝜓 ∘ 𝜙)(0) = (𝐷𝜓)(𝑝)𝐷𝜙(0) = id𝑛
by the chain rule, and hence if 𝐷𝜙(0)𝑣 = 0, it follows from this identity that 𝑣 = 0. □
Lemma 4.1.10 says that 𝜙 is an immersion at 0, so by the canonical immersion theorem
(see Theorem b.18) there exists a neighborhood 𝑈0 of 0 in 𝑈, a neighborhood 𝑉𝑝 of 𝑝 in 𝑉,
and a diffeomorphism
(4.1.11) 𝑔 ∶ (𝑉𝑝 , 𝑝) ⥲ (𝑈0 × 𝐑𝑁−𝑛 , 0)
Draft: March 28, 2018
in 𝑇𝑝⋆ 𝐑𝑛 are linearly independent, i.e., the system (4.1.14) is an “independent system of
defining equations” for 𝑋.
where ℓ = 𝑛 − 𝑘.
(2) Let 𝐺𝑘 be the set of 𝑘-dimensional subspaces of 𝐑𝑛 having the property (4.1.17) and
let M 𝑘,ℓ be the vector space of 𝑘 × ℓ matrices. Show that one gets from the identities
(4.1.18) a bijective map:
(4.1.19) 𝛾 ∶ M 𝑘,ℓ → 𝐺𝑘 .
Exercise 4.1.viii. Let 𝑆𝑛 be the vector space of linear mappings of 𝐑𝑛 into itself which are
self-adjoint, i.e., have the property 𝐴 = 𝐴⊺ .
Draft: March 28, 2018
Exercise 4.1.x. Let 𝑉0 = span(𝑒1 , …, 𝑒𝑘 ) and let 𝐺̃𝑘 = 𝜌(𝐺𝑘 ). Show that 𝜙 maps a neighbor-
hood of 𝑉0 in M 𝑘,ℓ diffeomorphically onto a neighborhood of 𝜋𝑉0 in 𝐺 ̃𝑘 .
̃𝑘 if and only if 𝑉 satisfies (4.1.17). For 1 ≤ 𝑖 ≤ 𝑘 let
Hints: 𝜋𝑉 is in 𝐺
𝑘 ℓ
𝑤𝑖 = 𝜋𝑉 (𝑒𝑖 ) = ∑ 𝑎𝑖,𝑗 𝑒𝑗 + ∑ 𝑐𝑖,𝑟 𝑒𝑘+𝑟 .
𝑗=1 𝑟=1
̃𝑘 .
• Show that if the matrix 𝐴 = (𝑎𝑖,𝑗 ) is invertible, 𝜋𝑉 is in 𝐺
̃
• Let 𝑂 ⊂ 𝐺𝑘 be the set of all 𝜋𝑉 ’s for which 𝐴 is invertible. Show that 𝜙−1 ∶ 𝑂 →
M 𝑘,ℓ is the map
𝜙−1 (𝜋𝑉 ) = 𝐵 = 𝐴−1 𝐶
where 𝐶 = [𝑐𝑖,𝑗 ].
Exercise 4.1.xi. Let 𝐺(𝑘, 𝑛) ⊂ 𝑆𝑛 be the set of rank 𝑘 projection operators. Prove that 𝐺(𝑘, 𝑛)
𝑛(𝑛+1)
is a 𝑘ℓ-dimensional submanifold of the Euclidean space 𝑆𝑛 ≅ 𝐑 2 .
Hints:
• Show that if 𝑉 is any 𝑘-dimensional subspace of 𝐑𝑛 there exists a linear mapping,
𝐴 ∈ 𝑂(𝑛) mapping 𝑉0 to 𝑉.
• Show that 𝜋𝑉 = 𝐴𝜋𝑉0 𝐴−1 .
• Let 𝐾𝐴 ∶ 𝑆𝑛 → 𝑆𝑛 be the linear mapping,
𝐾𝐴 (𝐵) = 𝐴𝐵𝐴−1 .
Show that
𝐾𝐴 ∘ 𝜙 ∶ M 𝑘,ℓ → 𝑆𝑛
maps a neighborhood of 𝑉0 in M 𝑘,ℓ diffeomorphically onto a neighborhood of
𝜋𝑉 in 𝐺(𝑘, 𝑛).
Draft: March 28, 2018
Remark 4.1.21. Let Gr(𝑘, 𝑛) be the set of all 𝑘-dimensional subspaces of 𝐑𝑛 . The identifi-
cation of Gr(𝑘, 𝑛) with 𝐺(𝑘, 𝑛) given by 𝑉 ↦ 𝜋𝑉 allows us to restate the result above in the
form.
Theorem 4.1.22. The Grassmannian Gr(𝑘, 𝑛) of 𝑘-dimensional subspaces of 𝐑𝑛 is a 𝑘ℓ-dimens-
𝑛(𝑛+1)
ional submanifold of 𝑆𝑛 ≅ 𝐑 2 .
Exercise 4.1.xii. Show that Gr(𝑘, 𝑛) is a compact submanifold of 𝑆𝑛 .
Hint: Show that it’s closed and bounded.
𝑦
ℓ
𝛤𝑓
𝑝0
𝑥
Then in this figure above the tangent line ℓ to 𝑋 at 𝑝0 = (𝑥0 , 𝑦0 ) is defined by the
equation
𝑦 − 𝑦0 = 𝑎(𝑥 − 𝑥0 )
where 𝑎 = 𝑓′ (𝑥0 ) In other words if 𝑝 is a point on ℓ then 𝑝 = 𝑝0 + 𝜆𝑣0 where 𝑣0 = (1, 𝑎)
and 𝜆 ∈ 𝐑. We would, however, like the tangent space to 𝑋 at 𝑝0 to be a subspace of the
tangent space to 𝐑2 at 𝑝0 , i.e., to be the subspace of the space: 𝑇𝑝0 𝐑2 = {𝑝0 } × 𝐑2 , and this
we’ll achieve by defining
𝑇𝑝0 𝑋 ≔ { (𝑝0 , 𝜆𝑣0 ) | 𝜆 ∈ 𝐑 } .
Draft: March 28, 2018
Example 4.2.3. Let 𝑆2 be the unit 2-sphere in 𝐑3 . The tangent plane to 𝑆2 at 𝑝0 is usually
defined to be the plane
{ 𝑝0 + 𝑣 | 𝑣 ∈ 𝐑3 and 𝑣 ⟂ 𝑝0 } .
However, this tangent plane is easily converted into a subspace of 𝑇𝑝 𝐑3 via the map 𝑝0 +𝑣 ↦
(𝑝0 , 𝑣) and the image of this map
{ (𝑝0 , 𝑣) | 𝑣 ∈ 𝐑3 and 𝑣 ⟂ 𝑝0 } .
will be our definition of 𝑇𝑝0 𝑆2 .
Let’s now turn to the general definition. As above let 𝑋 be an 𝑛-dimensional submani-
fold of 𝐑𝑁 , 𝑝 a point of 𝑋, 𝑉 a neighborhood of 𝑝 in 𝐑𝑁 , 𝑈 an open set in 𝐑𝑛 and
𝜙 ∶ (𝑈, 𝑞) ⥲ (𝑋 ∩ 𝑉, 𝑝)
a parameterization of 𝑋. We can think of 𝜙 as a 𝐶∞ map
𝜙 ∶ (𝑈, 𝑞) → (𝑉, 𝑝)
whose image happens to lie in 𝑋 ∩ 𝑉 and we proved in §4.1 that its derivative at 𝑞
(𝑑𝜙)𝑞 ∶ 𝑇𝑞 𝐑𝑛 → 𝑇𝑝 𝐑𝑁
is injective.
Definition 4.2.4. The tangent space 𝑇𝑝 𝑋 to 𝑋 at 𝑝 is the image of the linear map
(𝑑𝜙)𝑞 ∶ 𝑇𝑞 𝐑𝑛 → 𝑇𝑝 𝐑𝑁 .
In other words, 𝑤 ∈ 𝑇𝑝 𝐑𝑁 is in 𝑇𝑝 𝑋 if and only if 𝑤 = 𝑑𝜙𝑞 (𝑣) for some 𝑣 ∈ 𝑇𝑞 𝐑𝑛 . More
succinctly,
(4.2.5) 𝑇𝑝 𝑋 ≔ (𝑑𝜙𝑞 )(𝑇𝑞 𝐑𝑛 ) .
(Since 𝑑𝜙𝑞 is injective, 𝑇𝑝 𝑋 is an 𝑛-dimensional vector subspace of 𝑇𝑝 𝐑𝑁 .)
One problem with this definition is that it appears to depend on the choice of 𝜙. To
get around this problem, we’ll give an alternative definition of 𝑇𝑝 𝑋. In §4.1 we showed that
there exists a neighborhood 𝑉 of 𝑝 in 𝐑𝑁 (which we can without loss of generality take to
be the same as 𝑉 above) and a 𝐶∞ map
(4.2.6) 𝑓 ∶ (𝑉, 𝑝) → (𝐑𝑘 , 0) , where 𝑘 = 𝑁 − 𝑛 ,
such that 𝑋 ∩ 𝑉 = 𝑓−1 (0) and such that 𝑓 is a submersion at all points of 𝑋 ∩ 𝑉, and in
particular at 𝑝. Thus
𝑑𝑓𝑝 ∶ 𝑇𝑝 𝐑𝑁 → 𝑇0 𝐑𝑘
is surjective, and hence the kernel of 𝑑𝑓𝑝 has dimension 𝑛. Our alternative definition of 𝑇𝑝 𝑋
is
(4.2.7) 𝑇𝑝 𝑋 ≔ ker(𝑑𝑓𝑝 ) .
Lemma 4.2.8. The spaces (4.2.5) and (4.2.7) are both 𝑛-dimensional subspaces of 𝑇𝑝 𝐑𝑁 , and
we claim that these spaces are the same.
(Notice that the definition (4.2.7) of 𝑇𝑝 𝑋 does not depend on 𝜙, so if we can show that
these spaces are the same, the definitions (4.2.5) and (4.2.7)) will depend neither on 𝜙 nor
on 𝑓.)
Draft: March 28, 2018
Theorem 4.2.13 (Inverse function theorem for manifolds). Let 𝑋 and 𝑌 be 𝑛-manifolds and
𝑓 ∶ 𝑋 → 𝑌 a 𝐶∞ map. Suppose that at 𝑝 ∈ 𝑋 the map
𝑑𝑓𝑝 ∶ 𝑇𝑝 𝑋 → 𝑇𝑞 𝑌 , 𝑞 ≔ 𝑓(𝑝) ,
is bijective. Then 𝑓 maps a neighborhood 𝑈 of 𝑝 in 𝑋 diffeomorphically onto a neighborhood
𝑉 of 𝑞 in 𝑌.
Proof. Let 𝑈 and 𝑉 be open neighborhoods of 𝑝 in 𝑋 and 𝑞 in 𝑌 and let 𝜙0 ∶ (𝑈0 , 0) → (𝑈, 𝑝)
and 𝜓0 ∶ (𝑉0 , 0) → (𝑉, 𝑞) be parametrizations of these neighborhoods. Shrinking 𝑈0 and 𝑈
we can assume that 𝑓(𝑈) ⊂ 𝑉. Let
𝑔 ∶ (𝑈0 , 𝑝0 ) → (𝑉0 , 𝑞0 )
be the map 𝜓0−1 ∘ 𝑓 ∘ 𝜙0 . Then 𝜓0 ∘ 𝑔 = 𝑓 ∘ 𝜙, so by the chain rule
(𝑑𝜓0 )𝑞0 ∘ (𝑑𝑔)𝑝0 = (𝑑𝑓)𝑝 ∘ (𝑑𝜙0 )𝑝0 .
Since (𝑑𝜓0 )𝑞0 and (𝑑𝜙0 )𝑝0 are bijective it’s clear from this identity that if 𝑑𝑓𝑝 is bijective
the same is true for (𝑑𝑔)𝑝0 . Hence by the inverse function theorem for open subsets of 𝐑𝑛 ,
𝑔 maps a neighborhood of 𝑝0 in 𝑈0 diffeomorphically onto a neighborhood of 𝑞0 in 𝑉0 .
Shrinking 𝑈0 and 𝑉0 we assume that these neighborhoods are 𝑈0 and 𝑉0 and hence that 𝑔 is
a diffeomorphism. Thus since 𝑓 ∶ 𝑈 → 𝑉 is the map 𝜓0 ∘ 𝑔 ∘ 𝜙−1 0 , it is a diffeomorphism as
well. □
Theorem 4.2.14 (The canonical submersion theorem for manifolds). Let 𝑋 and 𝑌 be man-
ifolds of dimension 𝑛 and 𝑚, where 𝑚 < 𝑛, and let 𝑓 ∶ 𝑋 → 𝑌 be a 𝐶∞ map. Suppose that at
𝑝 ∈ 𝑋 the map
𝑑𝑓𝑝 ∶ 𝑇𝑝 𝑋 → 𝑇𝑞 𝑌 , 𝑞 ≔ 𝑓(𝑝) ,
is surjective. Then there exists an open neighborhood 𝑈 of 𝑝 in 𝑋, and open neighborhood, 𝑉
of 𝑓(𝑈) in 𝑌 and parametrizations 𝜙0 ∶ (𝑈0 , 0) → (𝑈, 𝑝) and 𝜓0 ∶ (𝑉0 , 0) → (𝑉, 𝑞) such that
in the square
𝜓0−1 𝑓𝜙0
𝑈0 𝑉0
𝜙0 𝜓0
𝑈 𝑉
𝑓
Theorem 4.2.15 (The canonical immersion theorem for manifolds). Let 𝑋 and 𝑌 be mani-
folds of dimension 𝑛 and 𝑚, where 𝑛 < 𝑚, and 𝑓 ∶ 𝑋 → 𝑌 a 𝐶∞ map. Suppose that at 𝑝 ∈ 𝑋
the map
𝑑𝑓𝑝 ∶ 𝑇𝑝 𝑋 → 𝑇𝑞 𝑌 , 𝑞 = 𝑓(𝑝)
is injective. Then there exists an open neighborhood 𝑈 of 𝑝 in 𝑋, an open neighborhood 𝑉 of
𝑓(𝑈) in 𝑌 and parametrizations 𝜙0 ∶ (𝑈0 , 0) → (𝑈, 𝑝) and 𝜓0 ∶ (𝑉0 , 0) → (𝑉, 𝑞) such that in
the square
𝜓0−1 𝑓𝜙0
𝑈0 𝑉0
𝜙0 𝜓0
𝑈 𝑉
𝑓
Proof. The proof is identical with the proof of Theorem 4.2.14 except for the last step. In the
last step one converts 𝑔 into the canonical immersion via a map 𝜓1 ∶ (𝑉1 , 0) → (𝑉0 , 0) with
the property 𝑔 ∘ 𝜓1 = 𝜄 and then replaces 𝜓0 by 𝜓0 ∘ 𝜓1 . □
Exercise 4.2.ii. Show that the tangent space to the (𝑛 − 1)-sphere 𝑆𝑛−1 at 𝑝 is the space of
vectors (𝑝, 𝑣) ∈ 𝑇𝑝 𝐑𝑛 satisfying 𝑝 ⋅ 𝑣 = 0.
Exercise 4.2.iii. Let 𝑓 ∶ 𝐑𝑛 → 𝐑𝑘 be a 𝐶∞ map and let 𝑋 = 𝛤𝑓 . What is the tangent space
to 𝑋 at (𝑎, 𝑓(𝑎))?
Exercise 4.2.iv. Let 𝜎 ∶ 𝑆𝑛−1 → 𝑆𝑛−1 be the antipodal map 𝜎(𝑥) = −𝑥. What is the derivative
of 𝜎 at 𝑝 ∈ 𝑆𝑛−1 ?
𝜙−1
2 ∘𝜙1
𝜙−1
1 (𝑋 ∩ 𝑉1 ∩ 𝑉2 ) 𝜙−1
2 (𝑋 ∩ 𝑉1 ∩ 𝑉2 )
(4.2.16)
𝜙1 𝜙2
𝑋 ∩ 𝑉1 ∩ 𝑉2 .
Draft: March 28, 2018
(2) Use overlap diagrams to give another proof that 𝑇𝑝 𝑋 is intrinsically defined.
𝜙⋆0 𝜔 = 𝜓⋆ 𝜙⋆1 𝜔 ,
so by (2.5.11) 𝜙⋆0 𝜔 is 𝐶∞ if 𝜙⋆1 𝜔 is 𝐶∞ . The same argument applied to 𝜓−1 shows that 𝜙⋆1 𝜔
is 𝐶∞ if 𝜙⋆0 𝜔 is 𝐶∞ .
The notion of “𝐶∞ ” for vector fields is defined similarly:
Definition 4.3.8. A vector field 𝒗 on 𝑈 is smooth, or simply 𝐶∞ , if 𝜙⋆0 𝒗 is 𝐶∞ .
By Proposition 4.3.4 𝜙⋆0 𝒗 = 𝜓⋆ 𝜙⋆1 𝒗, so, as above, this definition is independent of the
choice of parametrization.
We now globalize these definitions.
Definition 4.3.9. A 𝑘-form 𝜔 on 𝑋 is 𝐶∞ if, for every point 𝑝 ∈ 𝑋, 𝜔 is 𝐶∞ on a neigh-
borhood of 𝑝. Similarly, a vector field 𝒗 on 𝑋 is 𝐶∞ if, for every point 𝑝 ∈ 𝑋, 𝒗 is 𝐶∞ on a
neighborhood of𝑝.
We will also use the identities (4.3.5) and (4.3.18) to prove the following two results.
Proposition 4.3.10. Let 𝑋 and 𝑌 be manifolds and 𝑓 ∶ 𝑋 → 𝑌 a 𝐶∞ map. Then if 𝜔 is a 𝐶∞
𝑘-form on 𝑌, the pullback 𝑓⋆ 𝜔 is a 𝐶∞ 𝑘-form on 𝑋.
Proof. For 𝑝 ∈ 𝑋 and 𝑞 = 𝑓(𝑝) let 𝜙0 ∶ 𝑈0 → 𝑈 and 𝜓0 ∶ 𝑉0 → 𝑉 be parametrizations with
𝑝 ∈ 𝑈 and 𝑞 ∈ 𝑉. Shrinking 𝑈 if necessary we can assume that 𝑓(𝑈) ⊂ 𝑉. Let 𝑔 ∶ 𝑈0 → 𝑉0
be the map 𝑔 = 𝜓0−1 ∘ 𝑓 ∘ 𝜙0 . Then 𝜓0 ∘ 𝑔 = 𝑓 ∘ 𝜙0 , so 𝑔⋆ 𝜓0⋆ 𝜔 = 𝜙⋆0 𝑓⋆ 𝜔. Since 𝜔 is 𝐶∞ , we
see that 𝜓0⋆ 𝜔 is 𝐶∞ , so by equation (2.6.11) we have 𝑔⋆ 𝜓0⋆ 𝜔 is 𝐶∞ , and hence 𝜙⋆0 𝑓⋆ 𝜔 is 𝐶∞ .
Thus by definition 𝑓⋆ 𝜔 is 𝐶∞ on 𝑈. □
By exactly the same argument one proves.
Proposition 4.3.11. If 𝒘 is a 𝐶∞ vector field on 𝑌 and 𝑓 is a diffeomorphism, then 𝑓⋆ 𝒘 is a
𝐶∞ vector field on 𝑋.
Notation 4.3.12.
(1) We’ll denote the space of 𝐶∞ 𝑘-forms on 𝑋 by 𝛺𝑘 (𝑋) .
Draft: March 28, 2018
Proof. It’s not hard to prove this by the same argument that we used to prove this theorem
for vector fields on 𝐑𝑛 , but we’ll give a simpler proof that derives this directly from the 𝐑𝑛
result. Suppose 𝑋 is a submanifold of 𝐑𝑁 . Then for 𝑝 ∈ 𝑋,
𝑇𝑝 𝑋 ⊂ 𝑇𝑝 𝐑𝑁 = { (𝑝, 𝑣) | 𝑣 ∈ 𝐑𝑁 } ,
To conclude the proof of Theorem 4.3.17 we note that since 𝐑 is connected it follows
that if 𝛾(𝑡0 ) ∈ 𝑋 for some 𝑡0 ∈ 𝐑 then 𝛾(𝑡) ∈ 𝑋 for all 𝑡 ∈ 𝐑, and hence 𝛾 is an integral curve
of 𝒗. Thus in particular every integral curve of 𝒗 exists for all time, so 𝒗 is complete. □
Theorem 4.3.22. Let 𝑋 and 𝑌 be manifolds and 𝑓 ∶ 𝑋 → 𝑌 a 𝐶∞ map. Then for 𝜔 ∈ 𝛺𝑘 (𝑌)
we have
𝑓⋆ (𝑑𝜔) = 𝑑(𝑓⋆ 𝜔) .
Proof. For every 𝑝 ∈ 𝑋 we’ll check that this equality holds in a neighborhood of 𝑝. Let
𝑞 = 𝑓(𝑝) and let 𝑈 and 𝑉 be parametrizable neighborhoods of 𝑝 and 𝑞. Shrinking 𝑈 if
necessary we can assume 𝑓(𝑈) ⊂ 𝑉. Given parametrizations
𝜙 ∶ 𝑈0 → 𝑈
and
𝜓 ∶ 𝑉0 → 𝑉
we get by composition a map
𝑔 ∶ 𝑈0 → 𝑉0 , 𝑔 = 𝜓−1 ∘ 𝑓 ∘ 𝜙
with the property 𝜓 ∘ 𝑔 = 𝑓 ∘ 𝜙. Thus
𝜙⋆ 𝑑(𝑓⋆ 𝜔) = 𝑑𝜙⋆ 𝑓⋆ 𝜔 (by definition of 𝑑)
⋆
= 𝑑(𝑓 ∘ 𝜙) 𝜔
= 𝑑(𝜓 ∘ 𝑔)⋆ 𝜔
= 𝑑𝑔⋆ (𝜓⋆ 𝜔)
= 𝑔⋆ 𝑑𝜙⋆ 𝜔 (by (2.6.12))
⋆ ⋆
= 𝑔 𝜓 𝑑𝜔 (by definition of 𝑑)
⋆ ⋆
= 𝜙 𝑓 𝑑𝜔 .
⋆
Hence 𝑑𝑓 𝜔 = 𝑓 𝑑𝜔. ⋆
□
Draft: March 28, 2018
field
𝜕 𝜕 𝜕
𝒘 ≔ 𝑥1 + 𝑥2 + 2𝑥3 .
𝜕𝑥1 𝜕𝑥2 𝜕𝑥3
(1) Show that 𝒘 is tangent to 𝑋 and hence defines by restriction a vector field 𝒗 on 𝑋.
(2) What are the integral curves of 𝒗?
Exercise 4.3.ii. Let 𝑆2 be the unit 2-sphere, 𝑥12 + 𝑥22 + 𝑥32 = 1, in 𝐑3 and let 𝒘 be the vector
field
𝜕 𝜕
𝒘 ≔ 𝑥1 − 𝑥2 .
𝜕𝑥2 𝜕𝑥1
(1) Show that 𝒘 is tangent to 𝑆2 , and hence by restriction defines a vector field 𝒗 on 𝑆2 .
(2) What are the integral curves of 𝒗?
Exercise 4.3.iii. As in Exercise 4.3.ii let 𝑆2 be the unit 2-sphere in 𝐑3 and let 𝒘 be the vector
field
𝜕 𝜕 𝜕 𝜕
𝒘≔ − 𝑥3 (𝑥1 + 𝑥2 + 𝑥3 )
𝜕𝑥3 𝜕𝑥1 𝜕𝑥2 𝜕𝑥3
(1) Show that 𝒘 is tangent to 𝑆2 and hence by restriction defines a vector field 𝒗 on 𝑆2 .
(2) What do its integral curves look like?
Exercise 4.3.iv. Let 𝑆1 be the unit circle, 𝑥12 + 𝑥22 = 1, in 𝐑2 and let 𝑋 = 𝑆1 × 𝑆1 in 𝐑4 with
defining equations
𝑓1 = 𝑥12 + 𝑥22 − 1 = 0
𝑓2 = 𝑥32 + 𝑥42 − 1 = 0 .
Draft: March 28, 2018
4.4. Orientations
The last part of Chapter 4 will be devoted to the “integral calculus” of forms on mani-
folds. In particular we will prove manifold versions of two basic theorems of integral calculus
on 𝐑𝑛 , Stokes’ theorem and the divergence theorem, and also develop a manifold version
of degree theory. However, to extend the integral calculus to manifolds without getting in-
volved in horrendously technical “orientation” issues we will confine ourselves to a special
class of manifolds: orientable manifolds. The goal of this section will be to explain what this
term means.
Definition 4.4.1. Let 𝑋 be an 𝑛-manifold. An orientation of 𝑋 is a rule for assigning to each
𝑝 ∈ 𝑋 an orientation of 𝑇𝑝 𝑋.
Thus by Definition 1.9.3 one can think of an orientation as a “labeling” rule which, for
every 𝑝 ∈ 𝑋, labels one of the two components of the set 𝛬𝑛 (𝑇𝑝⋆ 𝑋)∖{0} by 𝛬𝑛 (𝑇𝑝⋆ 𝑋)+ , which
we’ll henceforth call the “plus” part of 𝛬𝑛 (𝑇𝑝⋆ 𝑋), and the other component by 𝛬𝑛 (𝑇𝑝⋆ 𝑋)− ,
which we’ll henceforth call the “minus” part of 𝛬𝑛 (𝑇𝑝⋆ 𝑋).
Definition 4.4.2. An orientation of a manifold 𝑋 is smooth, or simply 𝐶∞ , if for every
𝑝 ∈ 𝑋, there exists a neighborhood 𝑈 of 𝑝 and a non-vanishing 𝑛-form 𝜔 ∈ 𝛺𝑛 (𝑈) with the
property
𝜔𝑞 ∈ 𝛬𝑛 (𝑇𝑞⋆ 𝑋)+
for every 𝑞 ∈ 𝑈.
Remark 4.4.3. If we’re given an orientation of 𝑋 we can define another orientation by as-
signing to each 𝑝 ∈ 𝑋 the opposite orientation to the orientation we already assigned, i.e.,
by switching the labels on 𝛬𝑛 (𝑇𝑝⋆ )+ and 𝛬𝑛 (𝑇𝑝⋆ )− . We will call this the reversed orientation
of 𝑋. We will leave for you to check as an exercise that if 𝑋 is connected and equipped with
a smooth orientation, the only smooth orientations of 𝑋 are the given orientation and its
reversed orientation.
Hint: Given any smooth orientation of 𝑋 the set of points where it agrees with the given
orientation is open, and the set of points where it does not is also open. Therefore one of
these two sets has to be empty.
Note that if 𝜔 ∈ 𝛺𝑛 (𝑋) is a non-vanishing 𝑛-form one gets from 𝜔 a smooth orientation
of 𝑋 by requiring that the “labeling rule” above satisfy
𝜔𝑝 ∈ 𝛬𝑛 (𝑇𝑝⋆ 𝑋)+
for every 𝑝 ∈ 𝑋. If 𝜔 has this property we will call 𝜔 a volume form. It’s clear from this defi-
nition that if 𝜔1 and 𝜔2 are volume forms on 𝑋 then 𝜔2 = 𝑓2,1 𝜔1 where 𝑓2,1 is an everywhere
positive 𝐶∞ function.
Example 4.4.4. Let 𝑈 be an open subset 𝐑𝑛 . We will usually assign to 𝑈 its standard orien-
tation, by which we will mean the orientation defined by the 𝑛-form 𝑑𝑥1 ∧ ⋯ ∧ 𝑑𝑥𝑛 .
Draft: March 28, 2018
𝑥12 + ⋯ + 𝑥𝑛2 = 1
{
𝑥12 + ⋯ + 𝑥𝑛2 = 2 .
(3) Let 𝑆𝑛−1 be the unit sphere, 𝑥12 + ⋯ + 𝑥22 = 1 and let 𝐷 = 𝐑𝑛 ∖ 𝑆𝑛−1 . Then the boundary
of 𝐷 is 𝑆𝑛−1 but 𝐷 is not a smooth domain since the boundary of its closure is empty.
Draft: March 28, 2018
Definition 4.4.23. We will call an open set 𝑈 with the properties above a 𝐷-adapted parame-
trizable open set.
We will now show that if 𝑋 is oriented and 𝐷 ⊂ 𝑋 is a smooth domain then the bound-
ary 𝑍 ≔ 𝜕𝐷 of 𝐷 acquires from 𝑋 a natural orientation. To see this we first observe:
Lemma 4.4.24. The diffeomorphism 𝜓 ∶ 𝑈0 ⥲ 𝑈 in Theorem 4.4.17 can be chosen to be
orientation preserving.
Proof. If it is not, then by replacing 𝜓 with the diffeomorphism
𝜓♯ (𝑥1 , …, 𝑥𝑛 ) ≔ 𝜓(𝑥1 , …, 𝑥𝑛−1 , −𝑥𝑛 ) ,
we get a 𝐷-adapted parametrization of 𝑈 which is orientation preserving. (See (4.4.12)–
(4.4.13).) □
Let 𝑉0 = 𝑈0 ∩ 𝐑𝑛−1 be the boundary of 𝑈0 ∩ 𝐇𝑛 . The restriction of 𝜓 to 𝑉0 is a diffeomor-
phism of 𝑉0 onto 𝑈 ∩ 𝑍, and we will orient 𝑈 ∩ 𝑍 by requiring that this map be an oriented
parametrization. To show that this is an “intrinsic” definition, i.e., does not depend on the
choice of 𝜓. We’ll prove:
Theorem 4.4.25. If 𝜓𝑖 ∶ 𝑈𝑖 → 𝑈, 𝑖 = 0, 1, are oriented parametrizations of 𝑈 with the property
𝜓𝑖 ∶ 𝑈𝑖 ∩ 𝐇𝑛 → 𝑈 ∩ 𝐷
the restrictions of 𝜓𝑖 to 𝑈𝑖 ∩ 𝐑𝑛−1 induce compatible orientations on 𝑈 ∩ 𝑋.
Proof. To prove this we have to prove that the map 𝜙−1 𝑛
1 ∘ 𝜙0 , restricted to 𝑈 ∩ 𝜕𝐇 is an
𝑛−1 𝑛−1
orientation preserving diffeomorphism of 𝑈0 ∩ 𝐑 onto 𝑈1 ∩ 𝐑 . Thus we have to prove
the following:
Proposition 4.4.26. Let 𝑈0 and 𝑈1 be open subsets of 𝐑𝑛 and 𝑓 ∶ 𝑈0 → 𝑈1 an orientation
preserving diffeomorphism which maps 𝑈0 ∩ 𝐇𝑛 onto 𝑈1 ∩ 𝐇𝑛 . Then the restriction 𝑔 of 𝑓 to
the boundary 𝑈0 ∩ 𝐑𝑛−1 of 𝑈0 ∩ 𝐇𝑛 is an orientation preserving diffeomorphism
𝑔 ∶ 𝑈0 ∩ 𝐑𝑛−1 ⥲ 𝑈1 ∩ 𝐑𝑛−1 .
Let 𝑓(𝑥) = (𝑓1 (𝑥), …, 𝑓𝑛 (𝑥)). By assumption 𝑓1 (𝑥1 , …, 𝑥𝑛 ) is less than zero if 𝑥1 is less
than zero and equal to zero if 𝑥1 is equal to zero, hence
𝜕𝑓1
(4.4.27) (0, 𝑥2 , …, 𝑥𝑛 ) ≥ 0
𝜕𝑥1
and
𝜕𝑓1
(0, 𝑥2 , …, 𝑥𝑛 ) = 0
𝜕𝑥𝑖
for 𝑖 > 1 Moreover, since 𝑔 is the restriction of 𝑓 to the set 𝑥1 = 0
𝜕𝑓𝑖 𝜕𝑔
(0, 𝑥2 , …, 𝑥𝑛 ) = 𝑖 (𝑥2 , …, 𝑥1 )
𝜕𝑥𝑗 𝜕𝑥𝑗
for 𝑖, 𝑗 ≥ 2. Thus on the set defined by 𝑥1 = 0 we have
𝜕𝑓𝑖 𝜕𝑓 𝜕𝑔
(4.4.28) det ( ) = 1 det ( 𝑖 ) .
𝜕𝑥𝑗 𝜕𝑥1 𝜕𝑥𝑗
Since 𝑓 is orientation preserving the left hand side of equation (4.4.28) is positive at all
𝜕𝑓1 𝜕𝑔𝑖
points (0, 𝑥2 , …, 𝑥𝑛 ) ∈ 𝑈0 ∩ 𝐑𝑛−1 hence by (4.4.27) the same is true for 𝜕𝑥 and det ( 𝜕𝑥 ).
1 𝑗
Thus 𝑔 is orientation preserving. □
Draft: March 28, 2018
Remark 4.4.29. For an alternative proof of this result see Exercise 3.2.viii and Exercise 3.6.iv.
We will now orient the boundary of 𝐷 by requiring that for every 𝐷-adapted parametriz-
able open set 𝑈 the orientation of 𝑍 coincides with the orientation of 𝑈∩𝑍 that we described
above. We will conclude this discussion of orientations by proving a global version of Propo-
sition 4.4.26.
Proposition 4.4.30. For 𝑖 = 1, 2, let 𝑋𝑖 be an oriented manifold, 𝐷𝑖 ⊂ 𝑋𝑖 a smooth domain
and 𝑍𝑖 ≔ 𝜕𝐷𝑖 its boundary. Then if 𝑓 is an orientation preserving diffeomorphism of (𝑋1 , 𝐷1 )
onto (𝑋2 , 𝐷2 ) the restriction 𝑔 of 𝑓 to 𝑍1 is an orientation preserving diffeomorphism of 𝑍1
onto 𝑍2 .
Proof. Let 𝑈 be an open subset of 𝑋1 and 𝜙 ∶ 𝑈0 → 𝑈 an oriented 𝐷1 -compatible parametriza-
tion of 𝑈. Then if 𝑉 = 𝑓(𝑈) the map 𝑓𝜙 ∶ 𝑈 → 𝑉 is an oriented 𝐷2 -compatible parametriza-
tion of 𝑉 and hence 𝑔 ∶ 𝑈 ∩ 𝑍1 → 𝑉 ∩ 𝑍2 is orientation preserving. □
Exercises for §4.4
Exercise 4.4.i. Let 𝑉 be an oriented 𝑛-dimensional vector space, 𝐵 an inner product on 𝑉
and 𝑒1 , …, 𝑒𝑛 ∈ 𝑉 an oriented orthonormal basis. Given vectors 𝑣1 , …, 𝑣𝑛 ∈ 𝑉, show that if
(4.4.31) 𝑏𝑖,𝑗 = 𝐵(𝑣𝑖 , 𝑣𝑗 )
and
𝑛
𝑣𝑖 = ∑ 𝑎𝑗,𝑖 𝑒𝑗 ,
𝑗=1
(3) Show that the boundary orientation of 𝑆𝑛 agrees with the orientation of 𝑆𝑛 in Exer-
cise 4.4.viii.
(4.5.1) ∫ 𝜔
𝑊
where 𝜔 is a compactly supported 𝑛-form. We’ll begin by showing how to define this integral
when the support of 𝜔 is contained in a parametrizable open set 𝑈. Let 𝑈0 be an open subset
of 𝐑𝑛 and 𝜙0 ∶ 𝑈0 → 𝑈 a parametrization. As we noted in §4.4 we can assume without loss
of generality that this parametrization is oriented. Making this assumption, we’ll define
(4.5.2) ∫ 𝜔 = ∫ 𝜙⋆0 𝜔
𝑊 𝑊0
where 𝑊0 = 𝜙−1 ⋆
0 (𝑈 ∩ 𝑊). Notice that if 𝜙 𝜔 = 𝑓𝑑𝑥1 ∧ ⋯ ∧ 𝑑𝑥𝑛 , then, by assumption, 𝑓 is
∞
in 𝐶0 (𝑈0 ). Hence since
∫ 𝜙⋆0 𝜔 = ∫ 𝑓𝑑𝑥1 …𝑑𝑥𝑛
𝑊0 𝑊0
and since 𝑓 is a bounded continuous function and is compactly supported the Riemann
integral on the right is well-defined. (See Appendix b.) Moreover, if 𝜙1 ∶ 𝑈1 → 𝑈 is another
oriented parametrization of 𝑈 and 𝜓 ∶ 𝑈0 → 𝑈1 is the map 𝜓 = 𝜙−1 1 ∘ 𝜙0 then 𝜙0 = 𝜙1 ∘ 𝜓,
so by Proposition 4.3.4
𝜙⋆0 𝜔 = 𝜓⋆ 𝜙⋆1 𝜔 .
Moreover, by (4.3.18) 𝜓 is orientation preserving. Therefore since
𝑊1 = 𝜓(𝑊0 ) = 𝜙−1
1 (𝑈 ∩ 𝑊)
Thus the definition (4.5.2) is a legitimate definition. It does not depend on the parametriza-
tion that we use to define the integral on the right. From the usual additivity properties
of the Riemann integral one gets analogous properties for the integral (4.5.2). Namely for
𝜔1 , 𝜔2 ∈ 𝛺𝑛𝑐 (𝑈),
(4.5.4) ∫ (𝜔1 + 𝜔2 ) = ∫ 𝜔1 + ∫ 𝜔2
𝑊 𝑊 𝑊
by equation (4.5.4). Interchanging the orders of summation and resuming with respect to
the 𝑗’s this sum becomes
∞ ∞
∑ ∫ ∑ 𝜌𝑗′ 𝜌𝑖 𝜔
𝑖=1 𝑊 𝑗=1
or
∞
∑ ∫ 𝜌𝑖 𝜔 .
𝑖=1 𝑊
Hence
∞ ∞
∑ ∫ 𝜌𝑗′ 𝜔 = ∑ ∫ 𝜌𝑖 𝜔 ,
𝑖=1 𝑊 𝑖=1 𝑊
so the two sums are the same.
From equations (4.5.4) and (4.5.9) one easily deduces:
Draft: March 28, 2018
∫ 𝜔1 + 𝜔2 = ∫ 𝜔1 + ∫ 𝜔2
𝑊 𝑊 𝑊
and for 𝜔 ∈ 𝛺𝑛𝑐 (𝑋) and 𝑐 ∈ 𝐑
∫ 𝑐𝜔 = 𝑐 ∫ 𝜔 .
𝑊 𝑊
The definition of the integral (4.5.1) depends on the choice of an orientation of 𝑋, but
it’s easy to see how it depends on this choice. We pointed out in Section 4.4 that if 𝑋 is
connected, there is just one way to orient it smoothly other than by its given orientation,
namely by reversing the orientation of 𝑇𝑝 𝑋 at each point 𝑝 and it’s clear from the definitions
(4.5.2) and (4.5.9) that the effect of doing this is to change the sign of the integral, i.e., to
change ∫𝑋 𝜔 to − ∫𝑋 𝜔.
In the definition of the integral (4.5.1) we have allowed 𝑊 to be an arbitrary open subset
of 𝑋 but required 𝜔 to be compactly supported. This integral is also well-defined if we allow
𝜔 to be an arbitrary element of 𝛺𝑛 (𝑋) but require the closure of 𝑊 in 𝑋 to be compact. To
see this, note that under this assumption the sum (4.5.8) is still a finite sum, so the definition
of the integral still makes sense, and the double sum on the right side of (4.5.10) is still a
finite sum so it’s still true that the definition of the integral does not depend on the choice
of partitions of unity. In particular if the closure of 𝑊 in 𝑋 is compact we will define the
volume of 𝑊 to be the integral,
vol(𝑊) = ∫ 𝜎vol ,
𝑊
where 𝜎vol is the Riemannian volume form and if 𝑋 itself is compact we’ll define its volume
to be the integral
vol(𝑋) = ∫ 𝜎vol .
𝑋
We’ll next prove a manifold version of the change of variables formula (3.5.3).
Theorem 4.5.12 (change of variables formula). Let 𝑋′ and 𝑋 be oriented 𝑛-manifolds and
𝑓 ∶ 𝑋′ → 𝑋 an orientation preserving diffeomorphism. If 𝑊 is an open subset of 𝑋 and
𝑊′ ≔ 𝑓−1 (𝑊), then
(4.5.13) ∫ 𝑓⋆ 𝜔 = ∫ 𝜔
𝑊′ 𝑊
for all 𝜔 ∈ 𝛺𝑛𝑐 (𝑋).
Proof. By (4.5.9) the integrand of the integral above is a finite sum of 𝐶∞ forms, each of
which is supported on a parametrizable open subset, so we can assume that 𝜔 itself as this
property. Let 𝑉 be a parametrizable open set containing the support of 𝜔 and let 𝜙0 ∶ 𝑈 ⥲ 𝑉
be an oriented parameterization of 𝑉. Since 𝑓 is a diffeomorphism its inverse exists and is
a diffeomorphism of 𝑋 onto 𝑋1 . Let 𝑉′ ≔ 𝑓−1 (𝑉) and 𝜙′0 ≔ 𝑓−1 ∘ 𝜙0 . Then 𝜙′0 ∶ 𝑈 ⥲ 𝑉′ is
an oriented parameterization of 𝑉′ . Moreover, 𝑓 ∘ 𝜙′0 = 𝜙0 so if 𝑊0 = 𝜙−1
0 (𝑊) we have
𝑊0 = (𝜙′0 )−1 (𝑓−1 (𝑊)) = (𝜙′0 )−1 (𝑊′ )
and by the chain rule we have
𝜙⋆0 𝜔 = (𝑓 ∘ 𝜙′0 )⋆ 𝜔 = (𝜙′0 )⋆ 𝑓⋆ 𝜔
hence
∫ 𝜔 = ∫ 𝜙⋆0 𝜔 = ∫ (𝜙′0 )⋆ (𝑓⋆ 𝜔) = ∫ 𝑓⋆ 𝜔 . □
𝑊 𝑊0 𝑊0 𝑊′
Draft: March 28, 2018
(4.5.14) ∫ 𝑓⋆ 𝜔 = − ∫ 𝜔 .
𝑊′ 𝑊
We’ll conclude this discussion of “integral calculus on manifolds” by proving a prelim-
inary version of Stokes’ theorem.
Theorem 4.5.15. If 𝜇 is in 𝛺𝑛−1
𝑐 (𝑋) then
∫ 𝑑𝜇 = 0 .
𝑋
Proof. Let (𝜌𝑖 )𝑖≥1 be a partition of unity with the property that each 𝜌𝑖 is supported in a
parametrizable open set 𝑈𝑖 = 𝑈. Replacing 𝜇 by 𝜌𝑖 𝜇 it suffices to prove the theorem for
𝜇 ∈ 𝛺𝑛−1
𝑐 (𝑈). Let 𝜙 ∶ 𝑈0 → 𝑈 be an oriented parametrization of 𝑈. Then
∫ 𝑑𝜇 = ∫ 𝜙⋆ 𝑑𝜇 = ∫ 𝑑(𝜙⋆ 𝜇) = 0
𝑈 𝑈0 𝑈0
by Theorem 3.3.1. □
Exercises for §4.5
Exercise 4.5.i. Let 𝑓 ∶ 𝐑 → 𝐑 be a 𝐶∞ function. Orient the graph 𝑋 ≔ 𝛤𝑓 of 𝑓 by requiring
𝑛
Hint: An easier problem: Using polar coordinates integrate 𝜔 = 𝑥3 𝑑𝑥1 ∧ 𝑑𝑥2 over the
hemisphere defined by 𝑥3 = √1 − 𝑥12 − 𝑥22 , where 𝑥12 + 𝑥22 < 1.
𝑛
Exercise 4.5.vi. Let 𝛼 be the one-form ∑𝑖=1 𝑦𝑖 𝑑𝑥𝑖 in equation (2.8.2) and let 𝛾(𝑡), 0 ≤ 𝑡 ≤ 1,
be a trajectory of the Hamiltonian vector field (2.8.2). What is the integral of 𝛼 over 𝛾(𝑡)?
∫ 𝜄⋆ 𝜇
𝜕𝐷
is well-defined. We will prove:
Theorem 4.6.1 (Stokes’ theorem). For 𝜇 ∈ 𝛺𝑘−1
𝑐 (𝑋) we have
(4.6.2) ∫ 𝜄⋆ 𝜇 = ∫ 𝑑𝜇 .
𝜕𝐷 𝐷
Proof. Let (𝜌𝑖 )𝑖≥1 be a partition of unity such that for each 𝑖, the support of 𝜌𝑖 is contained
in a parametrizable open set 𝑈𝑖 = 𝑈 of one of the following three types:
(a) 𝑈 ⊂ int(𝐷).
(b) 𝑈 ⊂ ext(𝐷).
(c) There exists an open subset 𝑈0 of 𝐑𝑛 and an oriented 𝐷-adapted parametrization
𝜙 ∶ 𝑈0 ⥲ 𝑈 .
∞
Replacing 𝜇 by the finite sum ∑𝑖=1 𝜌𝑖 𝜇 it suffices to prove (4.6.2) for each 𝜌𝑖 𝜇 separately. In
other words we can assume that the support of 𝜇 itself is contained in a parametrizable open
set 𝑈 of type (a), (b), or (c). But if 𝑈 is of type (a)
∫ 𝑑𝜇 = ∫ 𝑑𝜇 = ∫ 𝑑𝜇
𝐷 𝑈 𝑋
and 𝜄⋆ 𝜇 = 0. Hence the left hand side of equation (4.6.2) is zero and, by Theorem 4.5.15,
the right hand side is as well. If 𝑈 is of type (b) the situation is even simpler: 𝜄⋆ 𝜇 is zero
and the restriction of 𝜇 to 𝐷 is zero, so both sides of equation (4.6.2) are automatically zero.
Thus one is reduced to proving (4.6.2) when 𝑈 is an open subset of type (c). In this case the
restriction of the map (4.6.2) to 𝑈0 ∩ 𝜕𝐇𝑛 is an orientation preserving diffeomorphism
(4.6.3) 𝜓 ∶ 𝑈0 ∩ 𝜕𝐇𝑛 → 𝑈 ∩ 𝑍
and
(4.6.4) 𝜄𝑍 ∘ 𝜓 = 𝜙 ∘ 𝜄𝐑𝑛−1
where the maps 𝜄 = 𝜄𝑍 and
𝜄𝐑𝑛−1 ∶ 𝐑𝑛−1 ↪ 𝐑𝑛
are the inclusion maps of 𝑍 into 𝑋 and 𝜕𝐇𝑛 into 𝐑𝑛 . (Here we’re identifying 𝜕𝐇𝑛 with 𝐑𝑛−1 .)
Thus
∫ 𝑑𝜇 = ∫ 𝜙⋆ 𝑑𝜇 = ∫ 𝑑𝜙⋆ 𝜇
𝐷 𝐇𝑛 𝐇𝑛
Draft: March 28, 2018
and by (4.6.4)
∫ 𝜄𝑍⋆ 𝜇 = ∫ 𝜓⋆ 𝜄𝑍⋆ 𝜇
𝑍 𝐑𝑛−1
=∫ 𝜄⋆𝐑𝑛−1 𝜙⋆ 𝜇
𝐑𝑛−1
=∫ 𝜄⋆𝐑𝑛−1 𝜙⋆ 𝜇 .
𝜕𝐇𝑛
Thus it suffices to prove Stokes’ theorem with 𝜇 replaced by 𝜙⋆ 𝜇, or, in other words, to prove
Stokes’ theorem for 𝐇𝑛 ; and this we will now do.
Theorem 4.6.5 (Stokes’ theorem for 𝐇𝑛 ). Let
𝑛
̂ 𝑖 ∧ ⋯ ∧ 𝑑𝑥𝑛 .
𝜇 = ∑(−1)𝑖−1 𝑓𝑖 𝑑𝑥1 ∧ ⋯ ∧ 𝑑𝑥
𝑖=1
Then
𝑛
𝜕𝑓𝑖
𝑑𝜇 = ∑ 𝑑𝑥1 ∧ ⋯ ∧ 𝑑𝑥𝑛
𝑖=1 𝜕𝑥𝑖
and
𝑛
𝜕𝑓𝑖
∫ 𝑑𝜇 = ∑ ∫ 𝑑𝑥 ⋯ 𝑑𝑥𝑛 .
𝐇𝑛 𝑖=1 𝐇𝑛 𝜕𝑥𝑖 1
We will compute each of these summands as an iterated integral doing the integration
with respect to 𝑑𝑥𝑖 first. For 𝑖 > 1 the 𝑑𝑥𝑖 integration ranges over the interval −∞ < 𝑥𝑖 < ∞
and hence since 𝑓𝑖 is compactly supported
∞ 𝑥𝑖 =+∞
𝜕𝑓𝑖
∫ 𝑑𝑥𝑖 = 𝑓𝑖 (𝑥1 , …, 𝑥𝑖 , …, 𝑥𝑛 )| =0.
−∞ 𝜕𝑥𝑖 𝑥𝑖 =−∞
On the other hand the 𝑑𝑥1 integration ranges over the interval −∞ < 𝑥1 < 0 and
0
𝜕𝑓1
∫ 𝑑𝑥 = 𝑓(0, 𝑥2 , …, 𝑥𝑛 ) .
−∞ 𝜕𝑥1 1
Thus integrating with respect to the remaining variables we get
∫ 𝐿𝒗 𝜔 = ∫ 𝜄𝑍⋆ (𝜄𝒗 𝜔) .
𝐷 𝑍
where the term on the right is by definition the flux of 𝒗 through 𝑍. In calculus books this
is written in a slightly different form. Letting
1
𝑛 2 2
𝜕𝑓
𝜎𝑍 = (∑ ( ) ) 𝜈𝑍
𝑖=1 𝜕𝑥𝑖
and letting
𝑛 2 − 21
𝜕𝑓 𝜕𝑓 𝜕𝑓
𝑛⃗ ≔ (∑ ( ) ) ( , …, )
𝑖=1 𝜕𝑥𝑖 𝜕𝑥1 𝜕𝑥𝑛
and 𝑣 ⃗ ≔ (𝑣1 , …, 𝑣𝑛 ) we have
𝐿𝒗 𝜈𝑍 = (𝑛⃗ ⋅ 𝑣)𝜎
⃗ 𝑍
and hence
In three dimensions 𝜎𝑍 is just the standard “infinitesimal element of area” on the surface 𝑍
and 𝑛𝑝 the unit outward normal to 𝑍 at 𝑝, so this version of the divergence theorem is the
version one finds in most calculus books.
As an application of Stokes’ theorem, we’ll give a very short alternative proof of the
Brouwer fixed point theorem. As we explained in § 3.6 the proof of this theorem basically
comes down to proving the following theorem.
Theorem 4.6.15. Let 𝐵𝑛 be the closed unit ball in 𝐑𝑛 and 𝑆𝑛−1 its boundary. Then the identity
map id𝑆𝑛−1 on 𝑆𝑛−1 cannot be extended to a 𝐶∞ map 𝑓 ∶ 𝐵𝑛 → 𝑆𝑛−1 .
Proof. Suppose that 𝑓 is such a map. Then for every (𝑛 − 1)-form 𝜇 ∈ 𝛺𝑛−1 (𝑆𝑛−1 ),
But 𝑑(𝑓⋆ 𝜇) = 𝑓⋆ (𝑑𝜇) = 0 since 𝜇 is an (𝑛−1)-form and 𝑆𝑛−1 is an (𝑛−1)-manifold, and since
𝑓 is the identity map on 𝑆𝑛−1 , (𝜄𝑆𝑛−1 )⋆ 𝑓⋆ 𝜇 = (𝑓 ∘ 𝜄𝑆𝑛−1 )⋆ 𝜇 = 𝜇. Thus for every 𝜇 ∈ 𝛺𝑛−1 (𝑆𝑛−1 ),
equation (4.6.16) says that the integral of 𝜇 over 𝑆𝑛−1 is zero. Since there are lots of (𝑛 − 1)-
forms for which this is not true, this shows that a map 𝑓 with the property above cannot
exist. □
and note by Exercise 4.4.ix that 𝜇 is the Riemannian volume form of 𝑆𝑛−1 .
Exercise 4.6.ii. Let 𝐷 ⊂ 𝐑𝑛 be a smooth domain. Show that there exists a neighborhood 𝑈
of 𝜕𝐷 in 𝐑𝑛 and a 𝐶∞ defining function 𝑔 ∶ 𝑈 → 𝐑 for 𝐷 with the properties:
(1) 𝑝 ∈ 𝑈 ∩ 𝐷 if and only if 𝑔(𝑝) < 0,
(2) and 𝑑𝑔𝑝 ≠ 0 if 𝑝 ∈ 𝑍
Draft: March 28, 2018
Hint: Deduce from Theorem 4.4.9 that a local version of this result is true. Show that
you can cover 𝑍 by a family U = {𝑈𝛼 }𝛼∈𝐼 of open subsets of 𝐑𝑛 such that for each there
exists a function 𝑔𝛼 ∶ 𝑈𝛼 → 𝐑 with the properties (1) and (2). Now let (𝜌𝑖 )𝑖≥1 be a partition
∞
of unity subordinate to U and let 𝑔 = ∑𝑖=1 𝜌𝑖 𝑔𝛼𝑖 where supp(𝜌𝑖 ) ⊂ 𝑈𝛼𝑖 .
Exercise 4.6.iii. In Exercise 4.6.ii suppose 𝑍 is compact. Show that there exists a global
defining function 𝑓 ∶ 𝐑𝑛 → 𝐑 for 𝐷 with properties (1) and (2).
Hint: Let 𝜌 ∈ 𝐶0∞ (𝑈) be a function such that 0 ≤ 𝜌(𝑥) ≤ 1 for all 𝑥 ∈ 𝑈 and 𝜌 is
identically 1 on a neighborhood of 𝑍, and replace 𝑔 by the function
𝑛
{𝜌(𝑥)𝑔(𝑥) + (1 − 𝜌(𝑥)), 𝑥 ∈ 𝐑 ∖ 𝐷
{
𝑓(𝑥) = {𝑔(𝑥), 𝑥 ∈ 𝜕𝐷
{
{𝜌(𝑥) − (1 − 𝜌(𝑥))𝑔(𝑥), 𝑥 ∈ int(𝐷) .
Exercise 4.6.iv. Show that the form 𝐿𝒗 𝑓𝜈𝑍 in equation (4.6.13) does not depend on what
choice we make of a defining function 𝑓 for 𝐷.
Hints:
• Show that if 𝑔 is another defining function then, at 𝑝 ∈ 𝑍, 𝑑𝑓𝑝 = 𝜆𝑑𝑔𝑝 , where 𝜆
is a positive constant.
• Show that if one replaces 𝑑𝑓𝑝 by (𝑑𝑔)𝑝 the first term in the product (𝐿𝒗 𝑓)(𝑝)(𝜈𝑍 )𝑝
changes by a factor 𝜆 and the second term by a factor 1/𝜆.
Exercise 4.6.v. Show that the form 𝜈𝑍 is intrinsically defined in the sense that if 𝜈 is any
(𝑛 − 1)-form satisfying equation (4.6.11), then 𝜈𝑍 = 𝜄𝑍⋆ 𝜈.
Hint: Exercise 4.5.vi.
Exercise 4.6.vi. Show that the form 𝜎𝑍 in equation (4.6.14) is the Riemannian volume form
on 𝑍.
Exercise 4.6.vii. Show that the (𝑛 − 1)-form
𝑛
̂ 𝑟 ∧ ⋯ 𝑑𝑥𝑛
𝜇 = (𝑥12 + ⋯ + 𝑥𝑛2 )−𝑛 ∑ (−1)𝑟−1 𝑥𝑟 𝑑𝑥1 ∧ ⋯ ∧ 𝑑𝑥
𝑟=1
is closed and prove directly that Stokes’ theorem holds for the annulus 𝑎 < 𝑥12 + ⋯ + 𝑥𝑛2 < 𝑏
by showing that the integral of 𝜇 over the sphere, 𝑥12 + ⋯ + 𝑥𝑛2 = 𝑎, is equal to the integral
over the sphere, 𝑥12 + ⋯ + 𝑥𝑛2 = 𝑏.
Exercise 4.6.viii. Let 𝑓 ∶ 𝐑𝑛−1 → 𝐑 be an everywhere positive 𝐶∞ function and let 𝑈 be a
bounded open subset of 𝐑𝑛−1 . Verify directly that Stokes’ theorem is true if 𝐷 is the domain
defined by
0 < 𝑥𝑛 < 𝑓(𝑥1 , …, 𝑥𝑛−1 ) , (𝑥1 , …, 𝑥𝑛−1 ) ∈ 𝑈
and 𝜇 an (𝑛 − 1)-form of the form
𝜙(𝑥1 , …, 𝑥𝑛 )𝑑𝑥1 ∧ ⋯ ∧ 𝑑𝑥𝑛−1
where 𝜙 is in 𝐶0∞ (𝐑𝑛 ).
Exercise 4.6.ix. Let 𝑋 be an oriented 𝑛-manifold and 𝒗 a vector field on 𝑋 which is complete.
Verify that for 𝜔 ∈ 𝛺𝑛𝑐 (𝑋)
∫ 𝐿𝒗 𝜔 = 0
𝑋
in the following ways:
(1) Directly by using the divergence theorem.
Draft: March 28, 2018
Proof. We have already verified the assertion (2) ⇒(1) (see Theorem 4.5.15), so what is left
to prove is the converse assertion. The proof of this is more or less identical with the proof
of the “(1) ⇒(2)” part of Theorem 3.2.2:
Step 1. Let 𝑈 be a connected parametrizable open subset of 𝑋. If 𝜔 ∈ 𝛺𝑛𝑐 (𝑈) has property (1),
then 𝜔 = 𝑑𝜇 for some 𝜇 ∈ 𝛺𝑛−1
𝑐 (𝑈).
∫ 𝜙⋆ 𝜔 = ∫ 𝜔 = 0
𝑈0 𝑋
Step 2. Fix a base point 𝑝0 ∈ 𝑋 and let 𝑝 be any point of 𝑋. Then there exists a collection
of connected parametrizable open sets 𝑊1 , …, 𝑊𝑁 with 𝑝0 ∈ 𝑊1 and 𝑝 ∈ 𝑊𝑁 such that for
1 ≤ 𝑖 ≤ 𝑁 − 1, the intersection 𝑊𝑖 ∩ 𝑊𝑖+1 is non-empty.
Proof of Step 2. The set of points, 𝑝 ∈ 𝑋, for which this assertion is true is open and the set
for which it is not true is open. Moreover, this assertion is true for 𝑝 = 𝑝0 . □
Step 3. We deduce Theorem 4.7.1 from a slightly stronger result. Introduce an equivalence
relation on 𝛺𝑛𝑐 (𝑋) by declaring that two 𝑛-forms 𝜔1 , 𝜔2 ∈ 𝛺𝑛𝑐 (𝑋) are equivalent if 𝜔1 − 𝜔2 ∈
𝑑𝛺𝑥𝑛−1 (𝑋). Denote this equivalence relation by a 𝜔1 ∼ 𝜔2 .
We will prove:
Theorem 4.7.2. For 𝜔1 and 𝜔2 ∈ 𝛺𝑛𝑐 (𝑋) the following are equivalent:
(1) ∫𝑋 𝜔1 = ∫𝑋 𝜔2
(2) 𝜔1 ∼ 𝜔2 .
Draft: March 28, 2018
Applying this result to a form 𝜔 ∈ 𝛺𝑛𝑐 (𝑋) whose integral is zero, we conclude that
𝜔 ∼ 0, which means that 𝜔 = 𝑑𝜇 for some 𝜇 ∈ 𝛺𝑛−1 𝑐 (𝑋). Hence Theorem 4.7.2 implies
Theorem 4.7.1. Conversely, if ∫𝑋 𝜔1 = ∫𝑋 𝜔2 , then ∫𝑋 (𝜔1 − 𝜔2 ) = 0, so 𝜔1 − 𝜔2 = 𝑑𝜇 for
some 𝜇 ∈ 𝛺𝑛𝑐 (𝑋). Hence Theorem 4.7.2 implies Theorem 4.7.1.
Step 4. By a partition of unity argument it suffices to prove Theorem 4.7.2 for 𝜔1 ∈ 𝛺𝑛𝑐 (𝑈1 )
and 𝜔2 ∈ 𝛺𝑛𝑐 (𝑈2 ) where 𝑈1 and 𝑈2 are connected parametrizable open sets. Moreover, if the
integrals of 𝜔1 and 𝜔2 are zero then 𝜔𝑖 = 𝑑𝜇𝑖 for some 𝜇𝑖 ∈ 𝛺𝑛𝑐 (𝑈𝑖 ) by Step 1, so in this case,
the theorem is true. Suppose on the other hand that
∫ 𝜔1 = ∫ 𝜔2 = 𝑐 ≠ 0 .
𝑋 𝑋
Then dividing by 𝑐, we can assume that the integrals of 𝜔1 and 𝜔2 are both equal to 1.
Step 5. Let 𝑊1 , …, 𝑊𝑁 be, as in Step 2, a sequence of connected parametrizable open sets with
the property that the intersections, 𝑊1 ∩ 𝑈1 , 𝑊𝑁 ∩ 𝑈2 and 𝑊𝑖 ∩ 𝑊𝑖+1 , for 𝑖 = 1, …, 𝑁 − 1, are all
non-empty. Select 𝑛-forms, 𝛼0 ∈ 𝛺𝑛𝑐 (𝑈1 ∩ 𝑊1 ), 𝛼𝑁 ∈ 𝛺𝑛𝑐 (𝑊𝑁 ∩ 𝑈2 ) and 𝛼𝑖 ∈ 𝛺𝑛𝑐 (𝑊𝑖 ∩ 𝑊𝑖+1 ),
for 𝑖 = 1, …, 𝑁 − 1, such that the integral of each 𝛼𝑖 over 𝑋 is equal to 1 By Step 1 we see that
Theorem 4.7.1 is true for 𝑈1 , 𝑈2 and the 𝑊1 , …, 𝑊𝑁 , hence Theorem 4.7.2 is true for 𝑈1 , 𝑈2
and the 𝑊1 , …, 𝑊𝑁 , so
𝜔1 ∼ 𝛼 0 ∼ 𝛼 1 ∼ ⋯ ∼ 𝛼 𝑁 ∼ 𝜔 2
and thus 𝜔1 ∼ 𝜔2 . □
Just as in (3.4.2) we get as a corollary of the theorem above the following “definition–
theorem” of the degree of a differentiable mapping:
Theorem 4.7.3. Let 𝑋 and 𝑌 be compact oriented 𝑛-manifolds and let 𝑌 be connected. Given a
proper 𝐶∞ mapping, 𝑓 ∶ 𝑋 → 𝑌, there exists a topological invariant deg(𝑓) with the defining
property:
(4.7.4) ∫ 𝑓⋆ 𝜔 = deg(𝑓) ∫ 𝜔 .
𝑋 𝑌
Proof. As in the proof of Theorem 3.4.6 pick an 𝑛-form 𝜔0 ∈ 𝛺𝑛𝑐 (𝑌) whose integral over 𝑌
is one and define the degree of 𝑓 to be the integral over 𝑋 of 𝑓⋆ 𝜔0 , i.e., set
(4.7.5) deg(𝑓) ≔ ∫ 𝑓⋆ 𝜔0 .
𝑋
Now let 𝜔 be any 𝑛-form in 𝛺𝑛𝑐 (𝑌) and let
(4.7.6) 𝑐≔∫ 𝜔.
𝑌
Then the integral of 𝜔 − 𝑐𝜔0 over 𝑌 is zero so there exists an (𝑛 − 1)-form 𝜇 ∈ 𝛺𝑛−1
𝑐 (𝑌) for
which 𝜔 − 𝑐𝜔0 = 𝑑𝜇. Hence 𝑓⋆ 𝜔 = 𝑐𝑓⋆ 𝜔0 + 𝑑𝑓⋆ 𝜇, so
∫ 𝑓⋆ 𝜔 = 𝑐 ∫ 𝑓⋆ 𝜔0 = deg(𝑓) ∫ 𝜔
𝑋 𝑋 𝑌
by equations (4.7.5) and (4.7.6). □
It’s clear from the formula (4.7.4) that the degree of 𝑓 is independent of the choice of
𝜔0 . (Just apply this formula to any 𝜔 ∈ 𝛺𝑛𝑐 (𝑌) having integral over 𝑌 equal to one.) It’s also
clear from (4.7.4) that “degree” behaves well with respect to composition of mappings:
Draft: March 28, 2018
∫ 𝑓⋆ 𝜔 = ∫ 𝜔
𝑈𝑝 𝑉
∫ 𝑓⋆ 𝜔 = − ∫ 𝜔
𝑈𝑝 𝑉
where 𝜎𝑝 = +1 if the map (4.7.10) is orientation preserving and 𝜎𝑝 = −1 if the map (4.7.10)
is orientation reversing.
We will next show that Sard’s Theorem is true for maps between manifolds and hence
that there exist lots of regular values. We first observe that if 𝑈 is a parametrizable open
subset of 𝑋 and 𝑉 a parametrizable open neighborhood of 𝑓(𝑈) in 𝑌, then Sard’s Theorem
is true for the map 𝑓 ∶ 𝑈 → 𝑉 since, up to diffeomorphism, 𝑈 and 𝑉 are just open subsets
of 𝐑𝑛 . Now let 𝑞 be any point in 𝑌, let 𝐵 be a compact neighborhood of 𝑞, and let 𝑉 be a
parametrizable open set containing 𝐵. Then if 𝐴 = 𝑓−1 (𝐵) it follows from Theorem 3.4.7
that 𝐴 can be covered by a finite collection of parametrizable open sets, 𝑈1 , …, 𝑈𝑁 such that
𝑓(𝑈𝑖 ) ⊂ 𝑉. Hence since Sard’s Theorem is true for each of the maps 𝑓 ∶ 𝑈𝑖 → 𝑉 and 𝑓−1 (𝐵)
is contained in the union of the 𝑈𝑖 ’s we conclude that the set of regular values of 𝑓 intersects
the interior of 𝐵 in an open dense set. Thus, since 𝑞 is an arbitrary point of 𝑌, we have proved:
Theorem 4.7.13. If 𝑋 and 𝑌 are 𝑛-manifolds and 𝑓 ∶ 𝑋 → 𝑌 is a proper 𝐶∞ map the set of
regular values of 𝑓 is an open dense subset of 𝑌.
Since there exist lots of regular values the formula (4.7.12) gives us an effective way
of computing the degree of 𝑓. We’ll next justify our assertion that deg(𝑓) is a topological
invariant of 𝑓. To do so, let’s generalize Definition 2.6.14 to manifolds.
Definition 4.7.14. Let 𝑋 and 𝑌 be manifolds and 𝑓0 , 𝑓1 ∶ 𝑋 → 𝑌 be 𝐶∞ maps. A 𝐶∞ map
(4.7.15) 𝐹 ∶ 𝑋 × [0, 1] → 𝑌
is a homotopy between 𝑓0 and 𝑓1 if for all 𝑥 ∈ 𝑋 we have 𝐹(𝑥, 0) = 𝑓0 (𝑥) and 𝐹(𝑥, 1) = 𝑓1 (𝑥).
Moreover, if 𝑓0 and 𝑓1 are proper maps, the homotopy, 𝐹, is a proper homotopy if 𝐹 is proper
as a 𝐶∞ map, i.e., for every compact set 𝐶 of 𝑌, 𝐹−1 (𝐶) is compact.
Let’s now prove the manifold analogue of Theorem 3.6.10.
Theorem 4.7.16. Let 𝑋 and 𝑌 be oriented 𝑛-manifolds and assume that 𝑌 is connected. If
𝑓0 , 𝑓2 ∶ 𝑋 → 𝑌 are proper maps and the map (4.7.8) is a proper homotopy, then deg(𝑓1 ) =
deg(𝑓2 ).
Proof. Let 𝜔 be an 𝑛-form in 𝛺𝑛𝑐 (𝑌) whose integral over 𝑌 is equal to 1, and let 𝐶 ≔ supp(𝜔)
be the support of 𝜔. Then if 𝐹 is a proper homotopy between 𝑓0 and 𝑓1 , the set, 𝐹−1 (𝐶), is
compact and its projection on 𝑋
(4.7.17) { 𝑥 ∈ 𝑋 | (𝑥, 𝑡) ∈ 𝐹−1 (𝐶) for some 𝑡 ∈ [0, 1] }
is compact. Let 𝑓𝑡 ∶ 𝑋 → 𝑌 be the map: 𝑓𝑡 (𝑥) = 𝐹(𝑥, 𝑡). By our assumptions on 𝐹, 𝑓𝑡 is a
proper 𝐶∞ map. Moreover, for all 𝑡 the 𝑛-form 𝑓𝑡⋆ 𝜔 is a 𝐶∞ function of 𝑡 and is supported
on the fixed compact set (4.7.17). Hence it’s clear from the definitions of the integral of a
form and 𝑓𝑡 that the integral
∫ 𝑓𝑡⋆ 𝜔
𝑋
is a 𝐶∞ function of 𝑡. On the other hand this integral is by definition the degree of 𝑓𝑡 and
hence by Theorem 4.7.3, deg(𝑓𝑡 ) is an integer, so it does not depend on 𝑡. In particular,
deg(𝑓0 ) = deg(𝑓1 ). □
Exercises for §4.7
Exercise 4.7.i. Let 𝑓 ∶ 𝐑 → 𝐑 be the map 𝑥 ↦ 𝑥𝑛 . Show that deg(𝑓) = 0 if 𝑛 is even and 1
if 𝑛 is odd.
Draft: March 28, 2018
∫ 𝑑𝐹⋆ 𝜇 = ∫ 𝜄⋆ 𝐹⋆ 𝜇 = ∫ 𝑓⋆ 𝜇 = deg(𝑓) ∫ 𝜇
𝐷 𝑍 𝑍 𝑌
by Stokes’ theorem since 𝐹 ∘ 𝜄 = 𝑓. Hence deg(𝑓) has to be zero. □
Draft: March 28, 2018
𝑆𝑛−1 , so its degree is −1, and this contradicts what we just deduced from the existence of the
homotopy (4.8.8). □
From this argument we can deduce another interesting fact about the sphere 𝑆𝑛−1 when
𝑛 − 1 is even. For 𝑣 ∈ 𝑆𝑛−1 the tangent space to 𝑆𝑛−1 at 𝑣 is just the space,
𝑇𝑣 𝑆𝑛−1 = { (𝑣, 𝑤) | 𝑤 ∈ 𝐑𝑛 and 𝑣 ⋅ 𝑤 = 0 } ,
so a vector field on 𝑆𝑛−1 can be viewed as a function 𝑔 ∶ 𝑆𝑛−1 → 𝐑𝑛 with the property
𝑔(𝑣) ⋅ 𝑣 = 0
𝑛−1
for all 𝑣 ∈ 𝑆 . If this function is nonzero at all points, then, letting ℎ be the function (4.8.6),
and arguing as above, we’re led to a contradiction. Hence we conclude:
Theorem 4.8.10. If 𝑛 − 1 is even and 𝒗 is a vector field on the sphere 𝑆𝑛−1 , then there exists a
point 𝑝 ∈ 𝑆𝑛−1 at which 𝒗(𝑝) = 0.
Draft: March 28, 2018
inside
𝑞
outside
Figure 4.8.1. A Jordan curve with 𝑝 inside of the curve and 𝑞 outside
and 4.8.3 below. The surface in Figure 4.8.2 is convex at 𝑝, and the surface in Figure 4.8.3 is
convex–concave.)
𝑝 𝛾
𝛾(𝑝)
𝛾
𝑝
𝛾(𝑝)
𝛾 𝑣
𝑝1 𝑝2 … 𝑝𝑔 𝑝0 0 𝑞
Figure 4.8.4. The Gauss map from a surface of genus 𝑔 to the sphere
Draft: March 28, 2018
𝑛 = 3 the Gauss–Bonnet theorem asserts that this topological invariant is just 1 − 𝑔 where
𝑔 is the genus of 𝑋 or, in other words, the “number of holes”. Figure 4.8.4 gives a pictorial
proof of this result. (Notice that at the points 𝑝1 , …, 𝑝𝑔 the surface 𝑋 is convex–concave so
the scalar curvature at these points is negative, i.e., the Gauss map is orientation reversing.
On the other hand, at the point 𝑝0 the surface is convex, so the Gauss map at this point is
orientation preserving.)
0 = ∫ 𝐹⋆ 𝑑𝜔 = ∫ 𝑑𝐹⋆ 𝜔
𝑊 𝑊
Hence
ind(𝒗, 𝐷) = deg(𝑓) = deg(𝑓1 ) = ind(𝒗, 𝐷) . □
Draft: March 28, 2018
Suppose now that 𝒗 has a finite number of isolated zeros 𝑝1 , …, 𝑝𝑘 in 𝐷. Let 𝐵𝜀 (𝑝𝑖 ) be
the open ball of radius 𝜀 with center at 𝑝𝑖 . By making 𝜀 small enough we can assume that
each of these balls is contained in 𝐷 and that they’re mutually disjoint. We will define the
local index ind(𝒗, 𝑝𝑖 ) of 𝒗 at 𝑝𝑖 to be the index of 𝒗 with respect to 𝐵𝜀 (𝑝𝑖 ). By Theorem 4.9.2
these local indices are unchanged if we replace 𝜀 by a smaller 𝜀, and by applying this theorem
to the domain
𝑘
𝐷1 = ⋃ 𝐵𝑝𝑖 (𝜀)
𝑖=1
we get, as a corollary of the theorem, the formula
𝑘
(4.9.3) ind(𝒗, 𝐷) = ∑ ind(𝒗, 𝑝𝑖 )
𝑖=1
which computes the global index of 𝒗 with respect to 𝐷 in terms of these local indices.
Let’s now see how to compute these local indices. Let’s first suppose that the point 𝑝 = 𝑝𝑖
is at the origin of 𝐑𝑛 . Then near 𝑝 = 0
𝒗 = 𝒗lin + 𝒗′
where
𝜕
𝒗 = 𝒗lin = ∑ 𝑎𝑖,𝑗 𝑥𝑖
1≤𝑖,𝑗≤𝑛 𝜕𝑥𝑗
the 𝑎𝑖,𝑗 ’s being constants and
𝑛
𝜕
𝒗𝑖 = ∑ 𝑓𝑖𝑗
𝑗=1 𝜕𝑥𝑗
where the 𝑓𝑖𝑗 ’s vanish to second order near zero, i.e., satisfy
(4.9.4) |𝑓𝑖,𝑗 (𝑥)| ≤ 𝐶|𝑥|2
for some constant 𝐶 > 0.
Definition 4.9.5. We’ll say that the point 𝑝 = 0 is a non-degenerate zero of 𝒗 is the matrix
𝐴 = (𝑎𝑖,𝑗 ) is non-singular.
This means in particular that
𝑛 𝑛
(4.9.6) ∑ |∑ 𝑎𝑖,𝑗 𝑥𝑖 | ≥ 𝐶1 |𝑥|
𝑗=1 𝑖=1
form some constant 𝐶1 > 0. Thus by (4.9.4) and (4.9.6) the vector field
𝒗𝑡 = 𝒗lin + 𝑡𝒗′ , 0 ≤ 𝑡 ≤ 1
has no zeros in the ball 𝐵𝜀 (0), other than the point 𝑝 = 0 itself, provided that 𝜀 is small
enough. Therefore if we let 𝑋𝜀 be the boundary of 𝐵𝜀 (0) we get a homotopy
𝐹 ∶ 𝑋𝜀 × [0, 1] → 𝑆𝑛−1 , (𝑥, 𝑡) ↦ 𝑓𝒗𝑡 (𝑥)
between the maps 𝑓𝒗lin ∶ 𝑋𝜀 → 𝑆𝑛−1 and 𝑓𝒗 ∶ 𝑋𝜀 → 𝑆𝑛−1 , thus by Theorem 4.7.16 we see that
deg(𝑓𝒗 ) = deg(𝑓𝒗lin ). Therefore,
(4.9.7) ind(𝒗, 𝑝) = ind(𝒗lin , 𝑝) .
Draft: March 28, 2018
and 𝒗′ vanishes to second order at 𝑝, and if 𝑝 is a non-degenerate zero of 𝑞, i.e., if the matrix
𝐴 = (𝑎𝑖,𝑗 ) is non-singular, then exactly the same argument as before shows that
ind(𝒗, 𝑝) = ind(𝒗lin , 𝑝) .
We will next prove:
Theorem 4.9.8. If 𝑝 is a non-degenerate zero of 𝒗, the local index ind(𝒗, 𝑝) is +1 or −1 de-
pending on whether the determinant of the matrix, 𝐴 = (𝑎𝑖,𝑗 ) is positive or negative.
Proof. As above we can, without loss of generality, assume that 𝑝 = 0. By (4.9.7) we can
assume 𝒗 = 𝒗lin . Let 𝐷 be the domain defined by
𝑛 𝑛 2
(4.9.9) ∑ (∑ 𝑎𝑖,𝑗 𝑥𝑖 ) < 1 ,
𝑗=1 𝑖=1
and let 𝑋 ≔ 𝜕𝐷 be its boundary. Since the only zero of 𝒗lin inside this domain is 𝑝 = 0 we
get from (4.9.3) and (4.9.4) that
ind(𝒗, 𝑝) = ind(𝒗lin , 𝑝) = ind(𝒗lin , 𝐷) .
Moreover, the map
𝑓𝒗lin ∶ 𝑋 → 𝑆𝑛−1
is, in view of (4.9.9), just the linear map 𝑣 ↦ 𝐴𝑣, restricted to 𝑋. In particular, since 𝐴 is
a diffeomorphism, this mapping is a diffeomorphism as well, so the degree of this map is
+1 or −1 depending on whether this map is orientation preserving or not. To decide which
𝑛 ̂ 𝑖 ∧ ⋯ 𝑑𝑥𝑛 be the Riemannian
of these alternatives is true let 𝜔 = ∑𝑖=1 (−1)𝑖 𝑥𝑖 𝑑𝑥1 ∧ ⋯ ∧ 𝑑𝑥
𝑛−1
volume form on 𝑆 then
∫ 𝐴⋆ 𝑑𝜔 = 𝑛 ∫ 𝐴⋆ 𝑑𝑥1 ∧ ⋯ ∧ 𝑑𝑥𝑛
𝐷 𝐷
We’ll briefly describe the various types of non-degenerate zeros that can occur for vector
fields in two dimensions. To simplify this description a bit we’ll assume that the matrix
𝑎11 𝑎12
𝐴=( )
𝑎21 𝑎22
is diagonalizable and that its eigenvalues are not purely imaginary. Then the following five
scenarios can occur:
(1) The eigenvalues of 𝐴 are real and positive. In this case the integral curves of 𝒗lin in a neigh-
borhood of 𝑝 look like the curves in Figure 4.9.1 and hence, in a small neighborhood
of 𝑝, the integral sums of 𝒗 itself look approximately like the curve in Figure 4.9.1.
(2) The eigenvalues of 𝐴 are real and negative. In this case the integral curves of 𝒗lin look like
the curves in Figure 4.9.1, but the arrows are pointing into 𝑝, rather than out of 𝑝, i.e.,
(3) The eigenvalues of 𝐴 are real, but one is positive and one is negative. In this case the
integral curves of 𝒗lin look like the curves in Figure 4.9.3.
(4) The eigenvalues of 𝐴 are complex and of the form 𝑎 ± √−𝑏, with 𝑎 positive. In this case
the integral curves of 𝒗lin are spirals going out of 𝑝 as in Figure 4.9.4.
(5) The eigenvalues of 𝐴 are complex and of the form 𝑎 ± √−𝑏, with 𝑎 negative. In this case
the integral curves are as in Figure 4.9.4 but are spiraling into 𝑝 rather than out of 𝑝.
Definition 4.9.10. Zeros of 𝒗 of types (1) and (4) are called sources; those of types (2) and
(5) are called sinks and those of type (3) are called saddles.
Draft: March 28, 2018
CHAPTER 5
(2) If 𝑋 is connected and non-compact 𝐻𝑐0 (𝑋) = 0. To see this, note that if 𝑓 is in 𝐶0∞ (𝑋)
and 𝑋 is non-compact, 𝑓 has to be zero at some point, and hence if 𝑑𝑓 = 0, then 𝑓 has
to be identically zero.
(3) If 𝑋 is 𝑛-dimensional,
𝛺𝑘 (𝑋) = 𝛺𝑘𝑐 (𝑋) = 0
for 𝑘 < 0 or 𝑘 > 𝑛, hence
𝐻𝑘 (𝑋) = 𝐻𝑐𝑘 (𝑋) = 0
for 𝑘 < 0 or 𝑘 > 𝑛.
(4) If 𝑋 is an oriented, connected 𝑛-manifold, the integration operation is a linear map
Theorem 5.1.10. If the maps (5.1.9) are homotopic then, for the maps they induce on coho-
mology
♯ ♯
(5.1.11) 𝑓0 = 𝑓1 .
Our proof of this will consist of proving this for an important special class of homo-
topies, and then by “pullback” tricks deducing this result for homotopies in general. Let 𝒗
be a complete vector field on 𝑋 and let
𝑓𝑡 ∶ 𝑋 → 𝑋 , − ∞ < 𝑡 < ∞
be the one-parameter group of diffeomorphisms that 𝒗 generates. Then
𝐹 ∶ 𝑋 × [0, 1] → 𝑋 , 𝐹(𝑝, 𝑡) ≔ 𝑓𝑡 (𝑝) ,
is a homotopy between 𝑓0 and 𝑓1 , and we’ll show that for this homotopic pair (5.1.11) is
true.
We’ll now describe how to extract from this result a proof of Theorem 5.1.10 for any
pair of homotopic maps. We’ll begin with the following useful observation.
Draft: March 28, 2018
i.e.,
𝐹 ∘ 𝛾1 ∘ 𝜄 = 𝑓1
and
𝐹 ∘ 𝛾0 ∘ 𝜄 = 𝑓0 .
Thus
𝜄⋆ (𝛾1⋆ 𝐹⋆ 𝜔 − 𝛾0⋆ 𝐹⋆ 𝜔) = 𝑓1⋆ 𝜔 − 𝑓0⋆ 𝜔
and on the other hand by (5.1.15)
𝜄⋆ (𝛾1⋆ 𝐹⋆ 𝜔 − 𝛾0⋆ 𝐹⋆ 𝜔) = 𝑑𝜄⋆ 𝛤𝐹⋆ 𝜔 + 𝜄⋆ 𝛤𝐹⋆ 𝑑𝜔 .
Letting
𝑄 ∶ 𝛺𝑘 (𝑌) → 𝛺𝑘−1 (𝑋)
be the “chain homotopy” operator
(5.1.16) 𝑄𝜔 ≔ 𝜄⋆ 𝛤𝐹⋆ 𝜔
we can write the identity above more succinctly in the form
(5.1.17) 𝑓1⋆ 𝜔 − 𝑓0⋆ 𝜔 = 𝑑𝑄𝜔 + 𝑄𝑑𝜔
and from this deduce, exactly as we did earlier, the identity (5.1.11).
This proof can easily be adapted to the compactly supported setting. Namely the oper-
ator (5.1.16) is defined by the integral
1
(5.1.18) 𝑄𝜔 = ∫ 𝜄⋆ 𝛾𝑡⋆ (𝜄𝜕/𝜕𝑡 𝐹⋆ 𝜔)𝑑𝑡 .
0
We’ll conclude this section by noting that the cohomology groups 𝐻𝑘 (𝑋) are equipped
with a natural product operation. Namely, suppose 𝜔𝑖 ∈ 𝛺𝑘𝑖 (𝑋), 𝑖 = 1, 2, is a closed form
and that 𝑐𝑖 = [𝜔𝑖 ] is the cohomology class represented by 𝜔𝑖 . We can then define a product
cohomology class 𝑐1 ⋅ 𝑐2 in 𝐻𝑘1 +𝑘2 (𝑋) by the recipe
(5.1.21) 𝑐1 ⋅ 𝑐2 ≔ [𝜔1 ∧ 𝜔2 ] .
To show that this is a legitimate definition we first note that since 𝜔2 is closed
𝑑(𝜔1 ∧ 𝜔2 ) = 𝑑𝜔1 ∧ 𝜔2 + (−1)𝑘1 𝜔1 ∧ 𝑑𝜔2 = 0 ,
so 𝜔1 ∧ 𝜔2 is closed and hence does represent a cohomology class. Moreover if we replace
𝜔1 by another representative 𝜔1 + 𝑑𝜇1 = 𝜔′ of the cohomology class 𝑐1 , then
𝜔′1 ∧ 𝜔2 = 𝜔1 ∧ 𝜔2 + 𝑑𝜇1 ∧ 𝜔2 .
But since 𝜔2 is closed,
𝑑𝜇1 ∧ 𝜔2 = 𝑑(𝜇1 ∧ 𝜔2 ) + (−1)𝑘1 𝜇1 ∧ 𝑑𝜔2
= 𝑑(𝜇1 ∧ 𝜔2 )
so
𝜔′1 ∧ 𝜔2 = 𝜔1 ∧ 𝜔2 + 𝑑(𝜇1 ∧ 𝜔2 )
and [𝜔′1∧ 𝜔2 ] = [𝜔1 ∧ 𝜔2 ]. Similarly (5.1.21) is unchanged if we replace 𝜔2 by 𝜔2 + 𝑑𝜇2 ,
so the definition of (5.1.21) depends neither on the choice of 𝜔1 nor 𝜔2 and hence is an
intrinsic definition as claimed.
There is a variant of this product operation for compactly supported cohomology classes,
𝑘
and we’ll leave for you to check that it’s also well defined. Suppose 𝑐1 is in 𝐻𝑐 1 (𝑋) and 𝑐2 is in
𝐻𝑘2 (𝑋) (i.e., 𝑐1 is a compactly supported class and 𝑐2 is an ordinary cohomology class). Let
𝑘
𝜔1 be a representative of 𝑐1 in 𝛺𝑐 1 (𝑋) and 𝜔2 a representative of 𝑐2 in 𝛺𝑘2 (𝑋). Then 𝜔1 ∧ 𝜔2
𝑘1 +𝑘2
is a closed form in 𝛺𝑐 (𝑋) and hence defines a cohomology class
(5.1.22) 𝑐1 ⋅ 𝑐2 ≔ [𝜔1 ∧ 𝜔2 ]
𝑘 +𝑘
in 𝐻𝑐 1 2 (𝑋). We’ll leave for you to check that this is intrinsically defined. We’ll also leave
for you to check that (5.1.22) is intrinsically defined if the roles of 𝑐1 and 𝑐2 are reversed, i.e.,
𝑘
if 𝑐1 is in 𝐻𝑘1 (𝑋) and 𝑐2 in 𝐻𝑐 2 (𝑋) and that the products (5.1.21) and (5.1.22) both satisfy
𝑐1 ⋅ 𝑐2 = (−1)𝑘1 𝑘2 𝑐2 ⋅ 𝑐1 .
Finally we note that if 𝑌 is another manifold and 𝑓 ∶ 𝑋 → 𝑌 a 𝐶∞ map then for 𝜔1 ∈ 𝛺𝑘1 (𝑌)
and 𝜔2 ∈ 𝛺𝑘2 (𝑌)
𝑓⋆ (𝜔1 ∧ 𝜔2 ) = 𝑓⋆ 𝜔1 ∧ 𝑓⋆ 𝜔2
by (2.6.8) and hence if 𝜔1 and 𝜔2 are closed and 𝑐𝑖 = [𝜔𝑖 ], then
(5.1.23) 𝑓♯ (𝑐1 ⋅ 𝑐2 ) = 𝑓♯ 𝑐1 ⋅ 𝑓♯ 𝑐2 .
Exercises for §5.1
Exercise 5.1.i (stereographic projection). Let 𝑝 ∈ 𝑆𝑛 be the point 𝑝 = (0, …, 0, 1). Show
that for every point 𝑥 = (𝑥1 , …, 𝑥𝑛+1 ) of 𝑆𝑛 ∖ {𝑝} the ray
𝑡𝑥 + (1 − 𝑡)𝑝 , 𝑡 > 0
intersects the plane 𝑥𝑛+1 = 0 in the point
1
𝛾(𝑥) = (𝑥 , …, 𝑥𝑛 )
1 − 𝑥𝑛+1 1
Draft: March 28, 2018
̂ 𝑖 ∧ ⋯ 𝑑𝑥𝑖
where 𝑑𝑥𝐼𝑟 ≔ 𝑑𝑥𝑖1 ∧ ⋯ ∧ 𝑑𝑥 𝑟 𝑘
𝐑 𝐿
𝐑
is multiplication by deg(𝑓).
Exercise 5.1.v. A homotopy equivalence between 𝑋 and 𝑌 is a pair of maps 𝑓 ∶ 𝑋 → 𝑌 and
𝑔 ∶ 𝑌 → 𝑋 such that 𝑔 ∘ 𝑓 ≃ id𝑋 and 𝑓 ∘ 𝑔 ≃ id𝑌 . Show that if 𝑋 and 𝑌 are homotopy
equivalent their cohomology groups are the same “up to isomorphism”, i.e., 𝑓 and 𝑔 induce
inverse isomorphisms 𝑓♯ ∶ 𝐻𝑘 (𝑌) → 𝐻𝑘 (𝑋) and 𝑔♯ ∶ 𝐻𝑘 (𝑋) → 𝐻𝑘 (𝑌).
Exercise 5.1.vi. Show that 𝐑𝑛 ∖ {0} and 𝑆𝑛−1 are homotopy equivalent.
Exercise 5.1.vii. What are the cohomology groups of the 𝑛-sphere with two points deleted?
Hint: The 𝑛-sphere with one point deleted is homeomorphic to 𝐑𝑛 .
Draft: March 28, 2018
For example the sequence (5.2.1) is exact if 𝑍𝑖 = 𝐵𝑖 for all 𝑖, or, in other words, if
𝐻𝑖 (𝐶) = 0 for all 𝑖. A simple example of an exact sequence that we’ll encounter a lot below
is a sequence of the form
𝛼1 𝛼2
0 𝑉1 𝑉2 𝑉3 0,
i.e., a five term exact sequence whose first and last terms are the vector space, 𝑉0 = 𝑉4 = 0,
and hence 𝛼0 = 𝛼3 = 0. This sequence is exact if and only if the following conditions hold:
(1) 𝛼1 is injective,
(2) the kernel of 𝛼2 equals the image of 𝛼1 , and
(3) 𝛼2 is surjective.
We will call an exact sequence of this form a short exact sequence. (We’ll also encounter
a lot below an even shorter example of an exact sequence, namely a sequence of the form
𝛼1
0 𝑉1 𝑉2 0.
𝑓
𝐴 𝐵
𝑖 𝑗
𝐶 𝑔 𝐷
commutes if 𝑗 ∘ 𝑓 = 𝑔 ∘ 𝑖, and a more complicated diagram of vector spaces and linear maps
like the diagram
𝐴1 𝐴2 𝐴3
𝐵1 𝐵2 𝐵3
𝐶1 𝐶2 𝐶3
𝑑 𝑑 𝑑
(5.2.6) 𝐶𝑟0 𝐶𝑟1 𝐶𝑟2 ⋯.
𝑖 𝑗
(5.2.7) 0 𝐶1𝑘 𝐶2𝑘 𝐶3𝑘 0
Draft: March 28, 2018
𝑖 𝑗
0 𝐶1𝑘−1 𝐶2𝑘−1 𝐶3𝑘−1 0
𝑑 𝑑 𝑑
𝑑 𝑑 𝑑
0 𝐶1𝑘+1 𝑖
𝐶2𝑘+1 𝑗
𝐶3𝑘+1 0
⋮ ⋮ ⋮ .
That is, assume that in the left hand squares, 𝑑𝑖 = 𝑖𝑑, and in the right hand squares, 𝑑𝑗 = 𝑗𝑑.
The Mayer–Vietoris theorem addresses the following question: If one has information
about the cohomology groups of two of the three complexes (5.2.6), what information about
the cohomology groups of the third can be extracted from this diagram? Let us first observe
that the maps 𝑖 and 𝑗 give rise to mappings between these cohomology groups. Namely, for
𝑟 = 1, 2, 3 let 𝑍𝑟𝑘 be the kernel of the map 𝑑 ∶ 𝐶𝑟𝑘 → 𝐶𝑟𝑘+1 , and 𝐵𝑟𝑘 the image of the map
𝑑 ∶ 𝐶𝑟𝑘−1 → 𝐶𝑟𝑘 . Since 𝑖𝑑 = 𝑑𝑖, 𝑖 maps 𝐵1𝑘 into 𝐵2𝑘 and 𝑍1𝑘 into 𝑍2𝑘 , therefore by (5.2.4) it gives
rise to a linear mapping
𝑖♯ ∶ 𝐻𝑘 (𝐶1 ) → 𝐻𝑘 (𝐶2 ) .
Similarly since 𝑗𝑑 = 𝑑𝑗, 𝑗 maps 𝐵2𝑘 into 𝐵3𝑘 and 𝑍2𝑘 into 𝑍3𝑘 , and so by (5.2.4) gives rise to a
linear mapping
𝑗♯ ∶ 𝐻𝑘 (𝐶2 ) → 𝐻𝑘 (𝐶3 ) .
Moreover, since 𝑗 ∘ 𝑖 = 0 the image of 𝑖♯ is contained in the kernel of 𝑗♯ . We’ll leave as an
exercise the following sharpened version of this observation:
Proposition 5.2.8. The kernel of 𝑗♯ equals the image of 𝑖♯ , i.e., the three term sequence
𝑖♯ 𝑗♯
(5.2.9) 𝐻𝑘 (𝐶1 ) 𝐻𝑘 (𝐶2 ) 𝐻𝑘 (𝐶3 )
is exact.
Since (5.2.7) is a short exact sequence one is tempted to conjecture that (5.2.9) is also
a short exact sequence (which, if it were true, would tell us that the cohomology groups
of any two of the complexes (5.2.6) completely determine the cohomology groups of the
third). Unfortunately, this is not the case. To see how this conjecture can be violated Let us
try to show that the mapping 𝑗♯ is surjective. Let 𝑐3𝑘 be an element of 𝑍3𝑘 representing the
cohomology class [𝑐3𝑘 ] in 𝐻3 (𝐶3 ). Since (5.2.7) is exact there exists a 𝑐2𝑘 in 𝐶2𝑘 which gets
mapped by 𝑗 onto 𝑐3𝑘 , and if 𝑐3𝑘 were in 𝑍2𝑘 this would imply
𝑗♯ [𝑐2𝑘 ] = [𝑗𝑐2𝑘 ] = [𝑐3𝑘 ] ,
Draft: March 28, 2018
i.e., the cohomology class [𝑐3𝑘 ] would be in the image of 𝑗♯ . However, since there’s no reason
for 𝑐2𝑘 to be in 𝑍2𝑘 , there’s also no reason for [𝑐3𝑘 ] to be in the image of 𝑗♯ . What we can say,
however, is that 𝑗𝑑𝑐2𝑘 = 𝑑𝑗𝑐2𝑘 = 𝑑𝑐3𝑘 = 0 since 𝑐3𝑘 is in 𝑍3𝑘 . Therefore by the exactness of
(5.2.7) in degree 𝑘 + 1 there exists a unique element 𝑐1𝑘+1 in 𝐶1𝑘+1 with property
(5.2.10) 𝑑𝑐2𝑘 = 𝑖𝑐1𝑘+1 .
Moreover, since
0 = 𝑑(𝑑𝑐2𝑘 ) = 𝑑𝑖(𝑐1𝑘+1 ) = 𝑖 𝑑𝑐1𝑘+1
and 𝑖 is injective, 𝑑𝑐1𝑘+1 = 0, i.e.,
(5.2.11) 𝑐1𝑘+1 ∈ 𝑍1𝑘+1 .
Thus via (5.2.10) and (5.2.11) we’ve converted an element 𝑐3𝑘 of 𝑍3𝑘 into an element 𝑐1𝑘+1 of
𝑍1𝑘+1 and hence set up a correspondence
(5.2.12) 𝑐3𝑘 ∈ 𝑍3𝑘 → 𝑐1𝑘+1 ∈ 𝑍1𝑘+1 .
Unfortunately this correspondence isn’t, strictly speaking, a map of 𝑍3𝑘 into 𝑍1𝑘+1 ; the 𝑐1𝑘 in
(5.2.12) isn’t determined by 𝑐3𝑘 alone but also by the choice we made of 𝑐2𝑘 . Suppose, how-
ever, that we make another choice of a 𝑐2𝑘 with the property 𝑗(𝑐2𝑘 ) = 𝑐3𝑘 . Then the difference
between our two choices is in the kernel of 𝑗 and hence, by the exactness of (5.2.6) at level
𝑘, is in the image of 𝑖. In other words, our two choices are related by
(𝑐2𝑘 )new = (𝑐2𝑘 )old + 𝑖(𝑐1𝑘 )
for some 𝑐1𝑘 in 𝐶1𝑘 , and hence by (5.2.10)
(𝑐1𝑘+1 )new = (𝑐1𝑘+1 )old + 𝑑𝑐1𝑘 .
Therefore, even though the correspondence (5.2.12) isn’t strictly speaking a map it does give
rise to a well-defined map
𝑍3𝑘 → 𝐻𝑘+1 (𝐶1 ) , 𝑐3𝑘 ↦ [𝑐3𝑘+1 ] .
Moreover, if 𝑐3𝑘 is in 𝐵3𝑘 , i.e., 𝑐3𝑘 = 𝑑𝑐3𝑘−1 for some 𝑐3𝑘−1 ∈ 𝐶3𝑘−1 , then by the exactness of (5.2.6)
at level 𝑘−1, 𝑐3𝑘−1 = 𝑗(𝑐2𝑘−1 ) for some 𝑐2𝑘−1 ∈ 𝐶2𝑘−1 and hence 𝑐3𝑘 = 𝑗(𝑑𝑐2𝑘−2 ). In other words we
can take the 𝑐2𝑘 above to be 𝑑𝑐2𝑘−1 in which case the 𝑐1𝑘+1 in equation (5.2.10) is just zero. Thus
the map (5.2.12) maps 𝐵3𝑘 to zero and hence by Proposition 1.2.9 gives rise to a well-defined
map
𝛿 ∶ 𝐻𝑘 (𝐶3 ) → 𝐻𝑘+1 (𝐶1 )
mapping [𝑐3𝑘 ] → [𝑐1𝑘+1 ]. We will leave it as an exercise to show that this mapping measures
the failure of the arrow 𝑗♯ in the exact sequence (5.2.9) to be surjective (and hence the failure
of this sequence to be a short exact sequence at its right end).
Proposition 5.2.13. The image of the map 𝑗♯ ∶ 𝐻𝑘 (𝐶2 ) → 𝐻𝑘 (𝐶3 ) is equal to the kernel of
the map 𝛿 ∶ 𝐻𝑘 (𝐶3 ) → 𝐻𝑘+1 (𝐶1 ).
Hint: Suppose that in the correspondence (5.2.12), 𝑐1𝑘+1 is in 𝐵1𝑘+1 . Then 𝑐1𝑘+1 = 𝑑𝑐1𝑘 for
some 𝑐1𝑘 in 𝐶1𝑘 . Show that 𝑗(𝑐2𝑘 − 𝑖(𝑐1𝑘 )) = 𝑐3𝑘 and 𝑑(𝑐2𝑘 − 𝑖(𝑐1𝑘 )) = 0, i.e., 𝑐2𝑘 − 𝑖(𝑐1𝑘 ) is in 𝑍2𝑘 and
hence 𝑗♯ [𝑐2𝑘 − 𝑖(𝑐1𝑘 )] = [𝑐3𝑘 ].
Let us next explore the failure of the map 𝑖♯ ∶ 𝐻𝑘+1 (𝐶1 ) → 𝐻𝑘+1 (𝐶2 ), to be injective. Let
𝑐1𝑘+1 be in 𝑍1𝑘+1 and suppose that its cohomology class, [𝑐1𝑘+1 ], gets mapped by 𝑖♯ into zero.
This translates into the statement
(5.2.14) 𝑖(𝑐1𝑘+1 ) = 𝑑𝑐2𝑘
Draft: March 28, 2018
for some 𝑐2𝑘 ∈ 𝐶2𝑘 . Moreover since 𝑑𝑐2𝑘 = 𝑖(𝑐1𝑘+1 ), 𝑗(𝑑𝑐2𝑘 ) = 0. But if
(5.2.15) 𝑐3𝑘 ≔ 𝑗(𝑐2𝑘 )
then 𝑑𝑐3𝑘 = 𝑑𝑗(𝑐2𝑘 ) = 𝑗(𝑑𝑐2𝑘 ) = 𝑗(𝑖(𝑐1𝑘+1 )) = 0, so 𝑐3𝑘 is in 𝑍3𝑘 , and by (5.2.14), (5.2.15) and the
definition of 𝛿
[𝑐1𝑘+1 ] = 𝛿[𝑐3𝑘 ] .
In other words the kernel of the map 𝑖♯ ∶ 𝐻𝑘+1 (𝐶1 ) → 𝐻𝑘+1 (𝐶2 ) is contained in the image of
the map 𝛿 ∶ 𝐻𝑘 (𝐶3 ) → 𝐻𝑘+1 (𝐶1 ). We will leave it as an exercise to show that this argument
can be reversed to prove the converse assertion and hence to prove
Proposition 5.2.16. The image of the map 𝛿 ∶ 𝐻𝑘 (𝐶1 ) → 𝐻𝑘+1 (𝐶1 ) is equal to the kernel of
the map 𝑖♯ ∶ 𝐻𝑘+1 (𝐶1 ) → 𝐻𝑘+1 (𝐶2 ).
Putting together the Propositions 5.2.8, 5.2.13 and 5.2.16 we obtain the main result of
this section: the Mayer–Vietoris theorem. The sequence of cohomology groups and linear
maps
𝛿 𝑖♯ 𝑗♯ 𝛿 𝑖♯
(5.2.17) ⋯ 𝐻𝑘 (𝐶1 ) 𝐻𝑘 (𝐶2 ) 𝐻𝑘 (𝐶3 ) 𝐻𝑘+1 (𝐶1 ) ⋯
is exact.
Remark 5.2.18. To see an illuminating real world example of an application of these ideas,
we strongly recommend that our readers stare at Exercise 5.2.v with gives a method for com-
puting the de Rham cohomology of the 𝑛-sphere 𝑆𝑛 in terms of the de Rham cohomologies
of 𝑆𝑛 ∖ {(0, …, 0, 1)} and 𝑆𝑛 ∖ {(0, …, 0, −1)} via an exact sequence of the form (5.2.17).
Remark 5.2.19. In view of the “⋯”’s this sequence can be a very long sequence and is com-
monly referred to as the long exact sequence in cohomology associated to the short exact
sequence of complexes (5.2.7).
Before we discuss the applications of this result, we will introduce some vector space
notation. Given vector spaces 𝑉1 and 𝑉2 , we’ll denote by 𝑉1 ⊕ 𝑉2 the vector space sum of 𝑉1
and 𝑉2 , i.e., the set of all pairs
(𝑢1 , 𝑢2 ) , 𝑢𝑖 ∈ 𝑉𝑖
with the addition operation
(𝑢1 , 𝑢2 ) + (𝑣1 + 𝑣2 ) ≔ (𝑢1 + 𝑣1 , 𝑢2 + 𝑣2 )
and the scalar multiplication operation
𝜆(𝑢1 , 𝑢2 ) ≔ (𝜆𝑢1 , 𝜆𝑢2 ) .
Now let 𝑋 be a manifold and let 𝑈1 and 𝑈2 be open subsets of 𝑋. Then one has a linear map
𝑖 ∶ 𝛺𝑘 (𝑈1 ∪ 𝑈2 ) → 𝛺𝑘 (𝑈1 ) ⊕ 𝛺𝑘 (𝑈2 )
defined by
(5.2.20) 𝜔 ↦ (𝜔|𝑈1 , 𝜔|𝑈2 )
where 𝜔|𝑈𝑖 is the restriction of 𝜔 to 𝑈𝑖 . Similarly one has a linear map
𝑗 ∶ 𝛺𝑘 (𝑈1 ) ⊕ 𝛺𝑘 (𝑈2 ) → 𝛺𝑘 (𝑈1 ∩ 𝑈2 )
defined by
(𝜔1 , 𝜔2 ) ↦ 𝜔1 |𝑈1 ∩𝑈2 − 𝜔2 |𝑈1 ∩𝑈2 .
We claim:
Draft: March 28, 2018
𝜙2 𝜔, on 𝑈1 ∩ 𝑈2
(5.2.23) 𝜔1 ≔ {
0, on 𝑈1 ∖ 𝑈1 ∩ 𝑈2
and let
−𝜙1 𝜔, on 𝑈1 ∩ 𝑈2
(5.2.24) 𝜔2 ≔ {
0, on 𝑈2 ∖ 𝑈1 ∩ 𝑈2 .
Since 𝜙2 is supported on 𝑈2 the form defined by (5.2.23) is 𝐶∞ on 𝑈1 and since 𝜙1 is sup-
ported on 𝑈1 the form defined by (5.2.24) is 𝐶∞ on 𝑈2 and since 𝜙1 + 𝜙2 = 1, 𝜔1 − 𝜔2 =
(𝜙1 + 𝜙2 )𝜔 = 𝜔 on 𝑈1 ∩ 𝑈2 . □
Now let
𝑑 𝑑 𝑑
(5.2.25) 0 𝐶10 𝐶11 𝐶12 ⋯.
be the vector space direct sum of the de Rham complexes of 𝑈1 and 𝑈2 , i.e., the complex
whose 𝑘th term is
𝐶2𝑘 = 𝛺𝑘 (𝑈1 ) ⊕ 𝛺𝑘 (𝑈2 )
Draft: March 28, 2018
with 𝑑 ∶ 𝐶2𝑘 → 𝐶2𝑘+1 defined to be the map 𝑑(𝜇1 , 𝜇2 ) = (𝑑𝜇1 , 𝑑𝜇2 ). Since 𝐶1𝑘 = 𝛺𝑘 (𝑈1 ∪ 𝑈2 )
and 𝐶3𝑘 = 𝛺𝑘 (𝑈1 ∩ 𝑈2 ) we have, by Theorem 5.2.21, a short exact sequence
𝑖 𝑗
0 𝐶1𝑘 𝐶2𝑘 𝐶3𝑘 0,
and it’s easy to see that 𝑖 and 𝑗 commute with the 𝑑’s:
𝑑𝑖 = 𝑖𝑑 and 𝑑𝑗 = 𝑗𝑑 .
Hence we’re exactly in the situation to which Mayer–Vietoris applies. Since the cohomology
groups of the complexes (5.2.25) and (5.2.26) are the de Rham cohomology groups 𝐻𝑘 (𝑈1 ∪
𝑈2 ) and 𝐻𝑘 (𝑈1 ∩ 𝑈2 ), and the cohomology groups of the complex (5.2.27) are the vector
space direct sums, 𝐻𝑘 (𝑈1 ) ⊕ 𝐻𝑘 (𝑈2 ), we obtain from the abstract Mayer–Vietoris theorem,
the following de Rham theoretic version of Mayer–Vietoris.
Theorem 5.2.28. Letting 𝑈 = 𝑈1 ∪ 𝑈2 and 𝑉 = 𝑈1 ∩ 𝑈2 one has a long exact sequence in de
Rham cohomology:
𝛿 𝑖♯ 𝑗♯ 𝛿 𝑖♯
⋯ 𝐻𝑘 (𝑈) 𝐻𝑐𝑘 (𝑈1 ) ⊕ 𝐻𝑐𝑘 (𝑈2 ) 𝐻𝑐𝑘 (𝑉) 𝐻𝑐𝑘+1 (𝑈) ⋯.
This result also has an analogue for compactly supported de Rham cohomology. Let
(5.2.29) 𝑖 ∶ 𝛺𝑘𝑐 (𝑈1 ∩ 𝑈2 ) → 𝐻𝑐𝑘 (𝑈1 ) ⊕ 𝛺𝑘𝑐 (𝑈2 )
be the map
𝑖(𝜔) ≔ (𝜔1 , 𝜔2 )
where
𝜔, on 𝑈1 ∩ 𝑈2
(5.2.30) 𝜔𝑖 ≔ {
0, on 𝑈𝑖 ∖ 𝑈1 ∩ 𝑈2 .
(Since 𝜔 is compactly supported on 𝑈1 ∩ 𝑈2 the form defined by equation (5.2.29) is a 𝐶∞
form and is compactly supported on 𝑈𝑖 .) Similarly, let
𝑗 ∶ 𝛺𝑘𝑐 (𝑈1 ) ⊕ 𝛺𝑘𝑐 (𝑈2 ) → 𝛺𝑘𝑐 (𝑈1 ∪ 𝑈2 )
be the map
𝑗(𝜔1 , 𝜔2 ) ≔ 𝜔̃1 − 𝜔̃2
where:
𝜔𝑖 , on 𝑈𝑖
𝜔̃𝑖 ≔ {
0, on (𝑈1 ∪ 𝑈2 ) − 𝑈𝑖 .
As above it’s easy to see that 𝑖 is injective and that the kernel of 𝑗 is equal to the image of 𝑖.
Thus if we can prove that 𝑗 is surjective we’ll have proved
Theorem 5.2.31. The sequence
𝑖 𝑗
0 𝛺𝑘𝑐 (𝑈1 ∩ 𝑈2 ) 𝛺𝑘𝑐 (𝑈1 ) ⊕ 𝛺𝑘𝑐 (𝑈2 ) 𝛺𝑘𝑐 (𝑈1 ∩ 𝑈2 ) 0.
is a short exact sequence.
Proof. To prove the surjectivity of 𝑗 we mimic the proof above. Given 𝜔 ∈ 𝛺𝑘𝑐 (𝑈1 ∪ 𝑈2 ), let
𝜔1 ≔ 𝜙1 𝜔|𝑈1
and
𝜔2 ≔ −𝜙2 𝜔|𝑈2 .
Then by (5.2.30) we have that𝜔 = 𝑗(𝜔1 , 𝜔2 ). □
Draft: March 28, 2018
Exercise 5.2.v. Let 𝑋 = 𝑆𝑛 and let 𝑈1 and 𝑈2 be the open subsets of 𝑆𝑛 obtained by removing
from 𝑆𝑛 the points, 𝑝1 = (0, …, 0, 1) and 𝑝2 = (0, …, 0, −1).
(1) Using stereographic projection show that 𝑈1 and 𝑈2 are diffeomorphic to 𝐑𝑛 .
(2) Show that 𝑈1 ∪ 𝑈2 = 𝑆𝑛 and 𝑈1 ∩ 𝑈2 is homotopy equivalent to 𝑆𝑛−1 . (See Exercise 5.1.v.)
Hint: 𝑈1 ∩ 𝑈2 is diffeomorphic to 𝐑𝑛 ∖ {0}.
(3) Deduce from the Mayer–Vietoris sequence that 𝐻𝑖+1 (𝑆𝑛 ) = 𝐻𝑖 (𝑆𝑛−1 ) for 𝑖 ≥ 1.
(4) Using part (3) give an inductive proof of a result that we proved by other means in
Section 5.1: 𝐻𝑘 (𝑆𝑛 ) = 0 for 1 ≤ 𝑘 < 𝑛.
Exercise 5.2.vi. Using the Mayer–Vietoris sequence of Exercise 5.2.v with cohomology re-
placed by compactly supported cohomology show that
𝐑, 𝑘 = 1, 𝑛
𝐻𝑐𝑘 (𝐑𝑛 ∖ {0}) ≅ {
0, otherwise .
Exercise 5.2.vii. Let 𝑛 a positive integer and let
𝑓1 𝑓𝑛−1
0 𝑉1 𝑉2 ⋯ 𝑉𝑛−1 𝑉𝑛 0
𝑛
be an exact sequence of finite dimensional vector spaces. Prove that ∑𝑖=1 dim(𝑉𝑖 ) = 0.
For manifolds the proof of Theorem 5.3.2 is somewhat trickier. The proof requires a
manifold analogue of the notion of convexity and there are several serviceable candidates.
The one we will use is the following. Let 𝑋 ⊂ 𝐑𝑁 be an 𝑛-manifold and for 𝑝 ∈ 𝑋 let 𝑇𝑝 𝑋
be the tangent space to 𝑋 at 𝑝. Recalling that 𝑇𝑝 𝑋 sits inside 𝑇𝑝 𝐑𝑁 and that
𝑇𝑝 𝐑𝑁 = { (𝑝, 𝑣) | 𝑣 ∈ 𝐑𝑁 }
we get a map
𝑇𝑝 𝑋 ↪ 𝑇𝑝 𝐑𝑁 → 𝐑𝑁 , (𝑝, 𝑥) ↦ 𝑝 + 𝑥 ,
and this map maps 𝑇𝑝 𝑋 bijectively onto an 𝑛-dimensional “affine” subspace 𝐿𝑝 of 𝐑𝑁 which
is tangent to 𝑋 at 𝑝. Let 𝜋𝑝 ∶ 𝑋 → 𝐿𝑝 be, as in the figure below, the orthogonal projection
of 𝑋 onto 𝐿𝑝 .
𝐿𝑝
𝑝 𝜋𝑝(𝑥)
Intuitively this assertion is pretty obvious: if 𝑞 is in 𝑈𝑝𝜀 and 𝜀 is small the map
𝜋𝑝−1 𝜋𝑞
𝐵𝑝𝜀 𝑈𝑝𝜀 𝐿𝑞
is to order 𝜀2 equal to the identity map, so it’s intuitively clear that its image is a slightly
warped, but still convex, copy of 𝐵𝜀 (𝑝). We won’t, however, bother to write out the details
that are required to make this proof rigorous.
A good cover is a particularly good “good cover” if it is a finite cover. We’ll codify this
property in the definition below.
Definition 5.3.7. An 𝑛-manifold 𝑋 is said to have finite topology if 𝑋 admits a finite cov-
ering by open sets 𝑈1 , …, 𝑈𝑁 with the property that for every multi-index, 𝐼 = (𝑖1 , …, 𝑖𝑘 ),
1 ≤ 𝑖1 ≤ 𝑖2 ⋯ < 𝑖𝐾 ≤ 𝑁, the set
(5.3.8) 𝑈𝐼 ≔ 𝑈𝑖1 ∩ ⋯ ∩ 𝑈𝑖𝑘
is either empty or is diffeomorphic to 𝐑𝑛 .
If 𝑋 is a compact manifold and U = {𝑈𝛼 }𝛼∈𝐼 is a good cover of 𝑋 then by the Heine–
Borel theorem we can extract from U a finite subcover 𝑈1 , …, 𝑈𝑁 , where
𝑈𝑖 ≔ 𝑈𝛼𝑖 , for 𝛼1 , …, 𝛼𝑁 ∈ 𝐼 ,
hence we conclude:
Theorem 5.3.9. Every compact manifold has finite topology.
More generally, for any manifold 𝑋, let 𝐶 be a compact subset of 𝑋. Then by Heine–
Borel we can extract from the cover U a finite subcollection 𝑈1 , …, 𝑈𝑁 , where
𝑈𝑖 ≔ 𝑈𝛼𝑖 , for 𝛼1 , …, 𝛼𝑁 ∈ 𝐼 ,
that covers 𝐶, hence letting 𝑈 ≔ 𝑈1 ∪ ⋯ ∪ 𝑈𝑁 , we’ve proved:
Theorem 5.3.10. If 𝑋 is an 𝑛-manifold and 𝐶 a compact subset of 𝑋, then there exists an open
neighborhood 𝑈 of 𝐶 in 𝑋 with finite topology.
We can in fact even strengthen this further. Let 𝑈0 be any open neighborhood of 𝐶 in
𝑋. Then in the theorem above we can replace 𝑋 by 𝑈0 to conclude:
Theorem 5.3.11. Let 𝑋 be a manifold, 𝐶 a compact subset of 𝑋, and 𝑈0 an open neighborhood
of 𝐶 in 𝑋. Then there exists an open neighborhood 𝑈 of 𝐶 in 𝑋, 𝑈 contained in 𝑈0 , so that 𝑈
has finite topology.
We will justify the term “finite topology” by devoting the rest of this section to proving.
Theorem 5.3.12. Let 𝑋 be an 𝑛-manifold. If 𝑋 has finite topology the de Rham cohomology
groups 𝐻𝑘 (𝑋), for 𝑘 = 0, …, 𝑛, and the compactly supported de Rham cohomology groups
𝐻𝑐𝑘 (𝑋), for 𝑘 = 0, …, 𝑛 are finite dimensional vector spaces.
The basic ingredients in the proof of this will be the Mayer–Vietoris techniques that we
developed in §5.2 and the following elementary result about vector spaces.
Lemma 5.3.13. Let 𝑉1 , 𝑉2 , and 𝑉3 be vector spaces and
𝛼 𝛽
(5.3.14) 𝑉1 𝑉2 𝑉3
an exact sequence of linear maps. Then if 𝑉1 and 𝑉3 are finite dimensional, so is 𝑉2 .
Draft: March 28, 2018
Proof. Since 𝑉3 is finite dimensional, the im(𝛽) is of finite dimension 𝑘. Hence there exist
vectors 𝑣1 , …, 𝑣𝑘 ∈ 𝑉2 such that
im(𝛽) = span(𝛽(𝑣1 ), …, 𝛽(𝑣𝑘 )) .
𝑘
Now let 𝑣 ∈ 𝑉2 . Then 𝛽(𝑣) is a linear combination 𝛽(𝑣) = ∑𝑖=1 𝑐𝑖 𝛽(𝑣𝑖 ), where 𝑐1 , …, 𝑐𝑘 ∈
𝐑. So
𝑘
(5.3.15) 𝑣′ ≔ 𝑣 − ∑ 𝑐𝑖 𝑣𝑖
𝑖=1
is in the kernel of 𝛽 and hence, by the exactness of (5.3.14), in the image of 𝛼. But 𝑉1 is finite
dimensional, so im(𝛼) is finite dimensional. Letting 𝑣𝑘+1 , …, 𝑣𝑚 be a basis of im(𝛼) we can
𝑚
by (5.3.15)) write 𝑣 as a sum 𝑣 = ∑𝑖=1 𝑐𝑖 𝑣𝑖 . In other words 𝑣1 , …, 𝑣𝑚 is a basis of 𝑉2 . □
Proof of Theorem 5.3.4. Our proof will be by induction on the number of open sets in a good
cover of 𝑋. More specifically, let
U = {𝑈1 , …, 𝑈𝑁 }
be a good cover of 𝑋. If 𝑁 = 1, 𝑋 = 𝑈1 and hence 𝑋 is diffeomorphic to 𝐑𝑛 , so 𝐻𝑘 (𝑋) = 0
for 𝑘 > 0, and 𝐻0 (𝑋) ≅ 𝐑, so the theorem is certainly true in this case. Let us now prove
it’s true for arbitrary 𝑁 by induction. Let 𝑈 ≔ 𝑈2 ∪ ⋯ ∪ 𝑈𝑁 . Then 𝑈 is a submanifold of 𝑋,
and, thinking of 𝑈 as a manifold in its own right, {𝑈2 , …, 𝑈𝑁 } is a good cover of 𝑈 involving
only 𝑁 − 1 sets. Hence the cohomology groups of 𝑈 are finite dimensional by the induction
hypothesis. The manifold 𝑈 ∩ 𝑈1 has a good cover given by {𝑈 ∩ 𝑈2 , …, 𝑈 ∩ 𝑈𝑁 }, so by the
induction hypothesis the cohomology groups of 𝑈∩𝑈1 are finite-dimensional. To prove that
the theorem is true for 𝑋 we note that 𝑋 = 𝑈1 ∪ 𝑈, and and the Mayer–Vietoris sequence
gives an exact sequence
𝛿 𝑖♯
𝐻𝑘−1 (𝑈1 ∩ 𝑈) 𝐻𝑘 (𝑋) 𝐻𝑘 (𝑈1 ) ⊕ 𝐻𝑘 (𝑈) .
Since the right-hand and left-hand terms are finite dimensional it follows from Lemma 5.3.13
that the middle term is also finite dimensional. □
The proof works practically verbatim for compactly supported cohomology. For 𝑁 = 1
𝐻𝑐𝑘 (𝑋) = 𝐻𝑐𝑘 (𝑈1 ) = 𝐻𝑐𝑘 (𝐑𝑛 )
so all the cohomology groups of 𝐻𝑘 (𝑋) are finite in this case, and the induction “𝑁 − 1” ⇒
“𝑁” follows from the exact sequence
𝑗♯ 𝛿
𝐻𝑐𝑘 (𝑈1 ) ⊕ 𝐻𝑐𝑘 (𝑈) 𝐻𝑐𝑘 (𝑋) 𝐻𝑐𝑘+1 (𝑈1 ∩ 𝑈) .
Remark 5.3.16. A careful analysis of the proof above shows that the dimensions of the
vector spaces 𝐻𝑘 (𝑋) are determined by the intersection properties of the open sets 𝑈𝑖 , i.e.,
by the list of multi-indices 𝐼 for which th intersections (5.3.8) are non-empty.
This collection of multi-indices is called the nerve of the cover U, and this remark
suggests that there should be a cohomology theory which has as input the nerve of U and
as output cohomology groups which are isomorphic to the de Rham cohomology groups.
Such a theory does exist, and we further address it in §5.8. (A nice account of it can also be
found in [13, Ch. 5]).
Draft: March 28, 2018
Exercise 5.3.iv. Show that every bounded, open convex subset, 𝑈, of 𝐑𝑛 is diffeomorphic
to 𝐑𝑛 .
Hints:
• Let 𝜓(𝑥) be the exhaustion function constructed in Exercise 5.3.iii and let
𝑓 ∶ 𝑈 → 𝐑𝑛
be the map: 𝑓(𝑥) = 𝜓(𝑥)𝑥. Show that this map is a bijective map of 𝑈 onto 𝐑𝑛 .
• Show that for 𝑥 ∈ 𝑈 and 𝑣 ∈ 𝐑𝑛
(𝑑𝑓)𝑥 𝑣 = 𝜓(𝑥)𝑣 + 𝑑𝜓𝑥 (𝑣)𝑥
and conclude that 𝑑𝑓𝑥 is bijective at 𝑥, i.e., that 𝑓 is locally a diffeomorphism of a
neighborhood of 𝑥 in 𝑈 onto a neighborhood of 𝑓(𝑥) in 𝐑𝑛 .
• Putting these together show that 𝑓 is a diffeomorphism of 𝑈 onto 𝐑𝑛 .
Exercise 5.3.v. Let 𝑈 ⊂ 𝐑 be the union of the open intervals, 𝑘 < 𝑥 < 𝑘 + 1, 𝑘 an integer.
Show that 𝑈 does not have finite topology.
Exercise 5.3.vi. Let 𝑉 ⊂ 𝐑2 be the open set obtained by deleting from 𝐑2 the points, 𝑝𝑛 =
(0, 𝑛), for every integer 𝑛. Show that 𝑉 does not have finite topology.
Hint: Let 𝛾𝑛 be a circle of radius 21 centered about the point 𝑝𝑛 . Using Exercises 2.1.viii
and 2.1.ix show that there exists a closed smooth one-form, 𝜔𝑛 on 𝑉 with the property that
∫𝛾 𝜔𝑛 = 1 and ∫𝛾 𝜔𝑛 = 0 for 𝑚 ≠ 𝑛.
𝑛 𝑚
Exercise 5.3.vii. Let 𝑋 be an 𝑛-manifold and U = {𝑈1 , 𝑈2 } a good cover of 𝑋. What are the
cohomology groups of 𝑋 if the nerve of this cover is:
(1) {1}, {2}.
(2) {1}, {2}, {1, 2}.
Exercise 5.3.viii. Let 𝑋 be an 𝑛-manifold and U = {𝑈1 , 𝑈2 , 𝑈3 } a good cover of 𝑋. What
are the cohomology groups of 𝑋 if the nerve of this cover is:
(1) {1}, {2}, {3}.
(2) {1}, {2}, {3}, {1, 2}.
(3) {1}, {2}, {3}, {1, 2}, {1, 3}
(4) {1}, {2}, {3}, {1, 2}, {1, 3}, {2, 3}.
(5) {1}, {2}, {3}, {1, 2}, {1, 3}, {2, 3}, {1, 2, 3}.
Exercise 5.3.ix. Let 𝑆1 be the unit circle in 𝐑3 parametrized by arc length: (𝑥, 𝑦) = (cos 𝜃, sin 𝜃).
Let 𝑈1 be the set: 0 < 𝜃 < 2𝜋
3
, 𝑈2 the set: 𝜋2 < 𝜃 < 3𝜋
2
, and 𝑈3 the set: − 2𝜋
3
< 𝜃 < 𝜋3 .
(1) Show that the 𝑈𝑖 ’s are a good cover of 𝑆1 .
(2) Using the previous exercise compute the cohomology groups of 𝑆1 .
Exercise 5.3.x. Let 𝑆2 be the unit 2-sphere in 𝐑3 . Show that the sets
𝑈𝑖 = { (𝑥1 , 𝑥2 , 𝑥3 ) ∈ 𝑆2 | 𝑥𝑖 > 0 }
𝑖 = 1, 2, 3 and
𝑈𝑖 = { (𝑥1 , 𝑥2 , 𝑥3 ) ∈ 𝑆2 | 𝑥𝑖−3 < 0 } ,
𝑖 = 4, 5, 6, are a good cover of 𝑆2 . What is the nerve of this cover?
Exercise 5.3.xi. Let 𝑋 and 𝑌 be manifolds. Show that if they both have finite topology, their
product 𝑋 × 𝑌 does as well.
Exercise 5.3.xii.
Draft: March 28, 2018
(1) Let 𝑋 be a manifold and let 𝑈1 , …, 𝑈𝑁 , be a good cover of 𝑋. Show that 𝑈1 ×𝐑, …, 𝑈𝑁 ×𝐑
is a good cover of 𝑋 × 𝐑 and that the nerves of these two covers are the same.
(2) By Remark 5.3.16,
𝐻𝑘 (𝑋 × 𝐑) ≅ 𝐻𝑘 (𝑋) .
Verify this directly using homotopy techniques.
(3) More generally, show that for all ℓ ≥ 0
𝐻𝑘 (𝑋 × 𝐑ℓ ) ≅ 𝐻𝑘 (𝑋)
(a) by concluding that this has to be the case in view of the Remark 5.3.16,
(b) and by proving this directly using homotopy techniques.
Let us now apply these remarks to de Rham theory. Let 𝑋 be a connected, oriented
𝑛-manifold. If 𝑋 has finite topology the vector spaces, 𝐻𝑐𝑛−𝑘 (𝑋) and 𝐻𝑘 (𝑋) are both finite
dimensional. We will show that there is a natural bilinear pairing between these spaces, and
hence by the discussion above, a natural linear mapping of 𝐻𝑘 (𝑋) into the vector space dual
of 𝐻𝑐𝑛−1 (𝑋). To see this let 𝑐1 be a cohomology class in 𝐻𝑐𝑛−𝑘 (𝑋) and 𝑐2 a cohomology class
in 𝐻𝑘 (𝑋). Then by (5.1.22) their product 𝑐1 ⋅ 𝑐2 is an element of 𝐻𝑐𝑛 (𝑋), and so by (5.1.2) we
can define a pairing between 𝑐1 and 𝑐2 by setting
(5.4.5) 𝐵(𝑐1 , 𝑐2 ) ≔ 𝐼𝑋 (𝑐1 ⋅ 𝑐2 ) .
Notice that if 𝜔1 ∈ 𝛺𝑛−𝑘
𝑐 (𝑋) and 𝜔2 ∈ 𝛺𝑘 (𝑋) are closed forms representing the cohomology
classes 𝑐1 and 𝑐2 , respectively, then by (5.1.22) this pairing is given by the integral
𝐵(𝑐1 , 𝑐2 ) ≔ ∫ 𝜔1 ∧ 𝜔2 .
𝑋
We’ll next show that this bilinear pairing is non-singular in one important special case:
Proposition 5.4.6. If 𝑋 is diffeomorphic to 𝐑𝑛 the pairing defined by (5.4.5) is non-singular.
Proof. To verify this there is very little to check. The vector spaces 𝐻𝑘 (𝐑𝑛 ) and 𝐻𝑐𝑛−𝑘 (𝐑𝑛 )
are zero except for 𝑘 = 0, so all we have to check is that the pairing
𝐻𝑐𝑛 (𝑋) × 𝐻0 (𝑋) → 𝐑
is non-singular. To see this recall that every compactly supported 𝑛-form is closed and that
the only closed zero-forms are the constant functions, so at the level of forms, the pairing
(5.4.5) is just the pairing
𝛺𝑛 (𝑋) × 𝐑 → 𝐑 , (𝜔, 𝑐) ↦ 𝑐 ∫ 𝜔 ,
𝑋
and this is zero if and only if 𝑐 is zero or 𝜔 is in 𝑑𝛺𝑛−1
𝑐 (𝑋). Thus at the level of cohomology
this pairing is non-singular. □
We will now show how to prove this result in general.
Theorem 5.4.7 (Poincaré duality). Let 𝑋 be an oriented, connected 𝑛-manifold having finite
topology. Then the pairing (5.4.5) is non-singular.
The proof of this will be very similar in spirit to the proof that we gave in the last sec-
tion to show that if 𝑋 has finite topology its de Rham cohomology groups are finite dimen-
sional. Like that proof, it involves Mayer–Vietoris plus some elementary diagram-chasing.
The “diagram-chasing” part of the proof consists of the following two lemmas.
Lemma 5.4.8. Let 𝑉1 , 𝑉2 and 𝑉3 be finite dimensional vector spaces, and let
𝛼 𝛽
𝑉1 𝑉2 𝑉3
be an exact sequence of linear mappings. Then the sequence of transpose maps
𝛽⋆ 𝛼⋆
𝑉3⋆ 𝑉2⋆ 𝑉1⋆
is exact.
Proof. Given a vector subspace, 𝑊2 , of 𝑉2 , let
𝑊2⟂ = { ℓ ∈ 𝑉2⋆ | ℓ(𝑤) = 0 for all 𝑤 ∈ 𝑊 } .
We’ll leave for you to check that if 𝑊2 is the kernel of 𝛽, then 𝑊2⟂ is the image of 𝛽⋆ and that
if 𝑊2 is the image of 𝛼, 𝑊2⟂ is the kernel of 𝛼⋆ . Hence if 𝛽 = im(𝛼), im(𝛽⋆ ) = ker(𝛼⋆ ). □
Draft: March 28, 2018
𝐴1 𝛼1 𝐴2 𝛼2 𝐴3 𝛼3 𝐴4 𝛼4 𝐴5 .
⋯ 𝐻𝑐𝑘−1 (𝑀) 𝐻𝑐𝑘 (𝑈1,2 ) 𝐻𝑐𝑘 (𝑈1 ) ⊕ 𝐻𝑐𝑘 (𝑈2 ) 𝐻𝑐𝑘 (𝑀) ⋯
⋯ 𝐻𝑛−(𝑘−1) (𝑀)⋆ 𝐻𝑛−𝑘 (𝑈1,2 )⋆ 𝐻𝑛−𝑘 (𝑈1 )⋆ ⊕ 𝐻𝑛−𝑘 (𝑈2 )⋆ 𝐻𝑛−𝑘 (𝑀)⋆ ⋯
By Mayer–Vietoris the bottom row of the diagram is exact and by Mayer–Vietoris and
Lemma 5.4.8 the top row of the diagram is exact. Hence we can apply the “five lemma” to
the diagram to conclude:
Draft: March 28, 2018
(5.4.14) ∫ 𝑓⋆ 𝜇 ∧ 𝜈 = ∫ 𝜇 ∧ 𝑓⋆ 𝜈
𝑈 𝑈×𝐑
Exercise 5.4.iv. Show that the mapping 𝑓⋆ in Exercise 5.4.iii satisfies 𝑓⋆ 𝑑𝜔 = 𝑑𝑓⋆ 𝜔.
Exercise 5.4.v. Show that if 𝜔 is a closed compactly supported (𝑘 + 1)-form on 𝑈 × 𝐑 then
[𝑓⋆ 𝜔] = 𝑓! [𝜔]
where 𝑓! is the mapping (5.4.13) and 𝑓⋆ the mapping (5.4.14).
Exercise 5.4.vi.
(1) Let 𝑈 be an open subset of 𝐑𝑛 and let 𝑓 ∶ 𝑈 × 𝐑ℓ → 𝑈 be the projection, 𝑓(𝑥, 𝑡) = 𝑥.
Show that there is a unique linear mapping
𝑓⋆ ∶ 𝛺𝑘+ℓ ℓ 𝑘
𝑐 (𝑈 × 𝐑 ) → 𝛺𝑐 (𝑈)
with the property
∫ 𝑓⋆ 𝜇 ∧ 𝜈 = ∫ 𝜇 ∧ 𝑓⋆ 𝜈
𝑈 𝑈×𝐑ℓ
for all 𝜇 ∈ 𝛺𝑘+ℓ ℓ
𝑐 (𝑈 × 𝐑 ) and 𝜈 ∈ 𝛺
𝑛−𝑘
(𝑈).
Hint: Exercise 5.4.iii plus induction on ℓ.
(2) Show that for 𝜔 ∈ 𝛺𝑘+ℓ
𝑐 (𝑈 × 𝐑 )
ℓ
𝑑𝑓⋆ 𝜔 = 𝑓⋆ 𝑑𝜔 .
(3) Show that if 𝜔 is a closed, compactly supported (𝑘 + ℓ)-form on 𝑋 × 𝐑ℓ
[𝑓⋆ 𝜔] = 𝑓! [𝜔]
where 𝑓! ∶ 𝐻𝑐𝑘+ℓ (𝑈 × 𝐑ℓ ) → 𝐻𝑐𝑘 (𝑈) is the map (5.4.13).
Exercise 5.4.vii. Let 𝑋 be an 𝑛-manifold and 𝑌 an 𝑚-manifold. Assume 𝑋 and 𝑌 are com-
pact, oriented and connected, and orient 𝑋 × 𝑌 by giving it its natural product orientation.
Let
𝑓∶ 𝑋 × 𝑌 → 𝑌
be the projection map 𝑓(𝑥, 𝑦) = 𝑦. Given
𝜔 ∈ 𝛺𝑚 (𝑋 × 𝑌)
and 𝑝 ∈ 𝑌, let
(5.5.5) ∫ 𝜏𝑌 ∧ 𝜔 = ∫ 𝜄⋆𝑌 𝜔 .
𝑋 𝑌
𝑛−𝑘
In other words, for every closed form 𝜔 ∈ 𝛺𝑐 (𝑋),
the integral of 𝜔 over 𝑌 is equal to
the integral over 𝑋 of 𝜏𝑌 ∧ 𝜔. A closed form 𝜏𝑌 with this “reproducing” property is called a
Thom form for 𝑌. Note that if we add to 𝜏𝑌 an exact (𝑛 − 𝑘)-form 𝜇 ∈ 𝑑𝛺𝑛−𝑘−1 (𝑋), we get
Draft: March 28, 2018
another representative 𝜏𝑌 + 𝜇 of the cohomology class 𝑇𝑌 , and hence another form with this
reproducing property. Also, since equation (5.5.5) is a direct translation into form language
of equation (5.5.4), any closed (𝑛 − 𝑘)-form 𝜏𝑌 with the reproducing property (5.5.5) is a
representative of the cohomology class 𝑇𝑌 .
These remarks make sense as well for compactly supported cohomology. Suppose 𝑌 is
compact. Then from the inclusion map we get a pullback map
♯
𝜄𝑌 ∶ 𝐻𝑘 (𝑋) → 𝐻𝑘 (𝑌)
and since 𝑌 is compact, the integration operation 𝐼𝑌 is a map 𝐻𝑘 (𝑌) → 𝐑, so the composi-
tion of these two operations is a map,
ℓ𝑌 ∶ 𝐻𝑘 (𝑋) → 𝐑
which by (5.5.3) gets converted into a cohomology class
𝑇𝑌 = 𝐿−1 𝑛−𝑘
𝐵 (ℓ𝑌 ) ∈ 𝐻𝑐 (𝑋) .
Moreover, if 𝜏𝑌 ∈ 𝛺𝑛−𝑘
𝑐 (𝑋) is a closed form, it represents this cohomology class if and only
if it has the reproducing property
(5.5.6) ∫ 𝜏𝑌 ∧ 𝜔 = ∫ 𝜄⋆𝑌 𝜔
𝑋 𝑌
𝑛−𝑘
for closed forms 𝜔 ∈ 𝛺 (𝑋). (There’s a subtle difference, however, between formula (5.5.5)
and formula (5.5.6). In (5.5.5) 𝜔 has to be closed and compactly supported and in (5.5.6) it
just has to be closed.)
As above we have a lot of latitude in our choice of 𝜏𝑌 : we can add to it any element of
𝑑𝛺𝑛−𝑘−1
𝑐 (𝑋). One consequence of this is the following.
Theorem 5.5.7. Given a neighborhood 𝑈 of 𝑌 in 𝑋, there exists a closed form 𝜏𝑌 ∈ 𝛺𝑛−𝑘
𝑐 (𝑈)
with the reproducing property
(5.5.8) ∫ 𝜏𝑌 ∧ 𝜔 = ∫ 𝜄⋆𝑌 𝜔
𝑈 𝑌
𝑘
for all closed forms 𝜔 ∈ 𝛺 (𝑈).
Hence, in particular, 𝜏𝑌 has the reproducing property (5.5.6) for all closed forms 𝜔 ∈
𝛺𝑛−𝑘 (𝑋). This result shows that the Thom form 𝜏𝑌 can be chosen to have support in an
arbitrarily small neighborhood of 𝑌.
Proof of Theorem 5.5.7. By Theorem 5.3.9 we can assume that 𝑈 has finite topology and
hence, in our definition of 𝜏𝑌 , we can replace the manifold 𝑋 by the open submanifold 𝑈.
This gives us a Thom form 𝜏𝑌 with support in 𝑈 and with the reproducing property (5.5.8)
for closed forms 𝜔 ∈ 𝛺𝑛−𝑘 (𝑈). □
Let us see what Thom forms actually look like in concrete examples. Suppose 𝑌 is de-
fined globally by a system of ℓ independent equations, i.e., suppose there exists an open
neighborhood 𝑂 of 𝑌 in 𝑋, a 𝐶∞ map 𝑓 ∶ 𝑂 → 𝐑ℓ , and a bounded open convex neighbor-
hood 𝑉 of the origin in 𝐑𝑛 satisfying the following properties.
Properties 5.5.9.
(1) The origin is a regular value of 𝑓.
(2) 𝑓−1 (𝑉) is closed in 𝑋.
(3) 𝑌 = 𝑓−1 (0).
Draft: March 28, 2018
Then by (1) and (3) 𝑌 is a closed submanifold of 𝑂 and by (2) it’s a closed submanifold
of 𝑋. Moreover, it has a natural orientation: For every 𝑝 ∈ 𝑌 the map
𝑑𝑓𝑝 ∶ 𝑇𝑝 𝑋 → 𝑇0 𝐑ℓ
is surjective, and its kernel is 𝑇𝑝 𝑌, so from the standard orientation of 𝑇0 𝐑ℓ one gets an
orientation of the quotient space
𝑇𝑝 𝑋/𝑇𝑝 𝑌 ,
and hence since 𝑇𝑝 𝑋 is oriented, one gets by Theorem 1.9.9 an orientation on 𝑇𝑝 𝑌. (See
Example 4.4.5.) Now let 𝜇 be an element of 𝛺ℓ𝑐 (𝑋). Then 𝑓⋆ 𝜇 is supported in 𝑓−1 (𝑉) and
hence by property (2) we can extend it to 𝑋 by setting it equal to zero outside 𝑂. We will
prove:
Theorem 5.5.10. If ∫𝑉 𝜇 = 1, then 𝑓⋆ 𝜇 is a Thom form for 𝑌.
To prove this we’ll first prove that if 𝑓⋆ 𝜇 has property (5.5.5) for some choice of 𝜇 it has
this property for every choice of 𝜇.
Lemma 5.5.11. Let 𝜇1 and 𝜇2 be forms in 𝛺ℓ𝑐 (𝑉) such that ∫𝑉 𝜇1 = ∫𝑉 𝜇2 = 1. Then for every
closed 𝑘-form 𝜈 ∈ 𝛺𝑘𝑐 (𝑋) we have
∫ 𝑓⋆ 𝜇1 ∧ 𝜈 = ∫ 𝑓⋆ 𝜇2 ∧ 𝜈 .
𝑋 𝑋
∫ 𝑓⋆ 𝜇𝑡 ∧ 𝜈 = ∫ 𝑓⋆ 𝜇 ∧ 𝜈
𝑋 𝑋
for all closed forms 𝜈 ∈ 𝛺𝑘𝑐 (𝑋). Hence to prove that 𝑓⋆ 𝜇 has the property (5.5.5), it suffices
to prove that
small 𝑓(supp 𝜈) is disjoint from 𝑡𝑉, and both sides of (5.5.12) are zero. Next suppose that 𝜈
is in 𝛺𝑘𝑐 (𝑈𝑝 ). Then 𝜓⋆ 𝜈 is a compactly supported 𝑘-form on 𝑊 so we can write it as a sum
𝜓⋆ 𝜈 = ∑ ℎ𝐼 (𝑥)𝑑𝑥𝐼 , ℎ𝐼 ∈ 𝐶0∞ (𝑊)
𝐼
the 𝐼’s being strictly increasing multi-indices of length 𝑘. Let 𝐼0 = (ℓ + 1, ℓ2 + 2, …, 𝑛). Then
𝑥 𝑥
(5.5.15) 𝜋⋆ 𝜇𝑡 ∧ 𝜓⋆ 𝜈 = 𝑡ℓ 𝜌 ( 1 , …, ℓ ) ℎ𝐼0 (𝑥1 , …, 𝑥𝑛 )𝑑𝑥𝑟 ∧ ⋯ 𝑑𝑥𝑛
𝑡 𝑡
and by (5.5.14)
𝜓⋆ (𝑓⋆ 𝜇𝑡 ∧ 𝜈) = 𝜋⋆ 𝜇𝑡 ∧ 𝜓⋆ 𝜈
and hence since 𝜓 is orientation preserving
𝑥 𝑥
∫ 𝑓⋆ 𝜇𝑡 ∧ 𝜈 = ∫ 𝑓⋆ 𝜇𝑡 ∧ 𝜈 = 𝑡ℓ ∫ 𝜌 ( 1 , …, ℓ ) ℎ𝐼0 (𝑥1 , …, 𝑥𝑛 )𝑑𝑥
𝑋 𝑈𝑝 𝐑𝑛 𝑡 𝑡
This, however, is just the integral of 𝜓⋆ 𝜈 over the set 𝜋−1 (0)∩𝑊. By (5.5.12), 𝜓 maps 𝜋−1 (0)∩
𝑊 diffeomorphically onto 𝑌 ∩ 𝑈𝑝 and by our recipe for orienting 𝑌 this diffeomorphism is
an orientation-preserving diffeomorphism, so the integral (5.5.16) is equal to ∫𝑌 𝜈. □
We’ll now describe some applications of Thom forms to topological intersection theory.
Let 𝑌 and 𝑍 be closed, oriented submanifolds of 𝑋 of dimensions 𝑘 and ℓ where 𝑘 + ℓ = 𝑛,
and Let us assume one of them (say 𝑍) is compact. We will show below how to define an
intersection number 𝐼(𝑌, 𝑍), which on the one hand will be a topological invariant of 𝑌
and 𝑍 and on the other hand will actually count, with appropriate ±-signs, the number of
points of intersection of 𝑌 and 𝑍 when they intersect non-tangentially. (Thus this notion is
similar to the notion of the degree deg(𝑓) for a 𝐶∞ mapping 𝑓. On the one hand deg(𝑓)
is a topological invariant of 𝑓. It’s unchanged if we deform 𝑓 by a homotopy. On the other
hand if 𝑞 is a regular value of𝑓, then deg(𝑓) counts with appropriate ±-signs the number of
points in the set 𝑓−1 (𝑞).)
We’ll first give the topological definition of this intersection number. This is by the for-
mula
(5.5.17) 𝐼(𝑌, 𝑍) = 𝐵(𝑇𝑌 , 𝑇𝑍 )
ℓ
where 𝑇𝑌 ∈ 𝐻 (𝑋) and 𝑇𝑍 ∈ and 𝐵 is the bilinear pairing (5.4.5). If 𝜏𝑌 ∈ 𝛺ℓ (𝑋)
𝐻𝑐𝑘 (𝑋)
and 𝜏𝑍 ∈ 𝛺𝑘𝑐 (𝑋) are Thom forms representing 𝑇𝑌 and 𝑇𝑍 , (5.5.17) can also be defined as the
integral
𝐼(𝑌, 𝑍) = ∫ 𝜏𝑌 ∧ 𝜏𝑍
𝑋
or by (5.5.8), as the integral over 𝑌,
To prove this we first need to show that transverse intersections look nice locally.
Theorem 5.5.23. If 𝑌 and 𝑍 intersect transversally, then for every 𝑝 ∈ 𝑌 ∩ 𝑍, there exists
an open neighborhood 𝑉𝑝 of 𝑝 in 𝑋, an open neighborhood 𝑈𝑝 of the origin in 𝐑𝑛 , and an
orientation-preserving diffeomorphism 𝜓𝑝 ∶ 𝑉𝑝 ⥲ 𝑈𝑝 which maps 𝑉𝑝 ∩ 𝑌 diffeomorphically
onto the subset of 𝑈𝑝 defined by the equations: 𝑥1 = ⋯ = 𝑥ℓ = 0, and maps 𝑉 ∩ 𝑍 onto the
subset of 𝑈𝑝 defined by the equations: 𝑥ℓ+1 = ⋯ = 𝑥𝑛 = 0.
Proof. Since this result is a local result, we can assume that 𝑋 = 𝐑𝑛 and hence by Theo-
rem 4.2.15 that there exists a neighborhood 𝑉𝑝 of 𝑝 in 𝐑𝑛 and submersions 𝑓 ∶ (𝑉𝑝 , 𝑝) →
(𝐑ℓ , 0) and 𝑔 ∶ (𝑉𝑝 , 𝑝) → (𝐑𝑘 , 0) with the properties
(5.5.24) 𝑉𝑝 ∩ 𝑌 = 𝑓−1 (0)
and
(5.5.25) 𝑉𝑝 ∩ 𝑍 = 𝑔−1 (0) .
Draft: March 28, 2018
Moreover, by (4.3.5) 𝑇𝑝 𝑌 = (𝑑𝑓𝑝 )−1 (0) and 𝑇𝑝 𝑍 = (𝑑𝑔𝑝 )−1 (0). Hence by (5.5.21), the equa-
tions
(5.5.26) 𝑑𝑓𝑝 (𝑣) = 𝑑𝑔𝑝 (𝑣) = 0
for 𝑣 ∈ 𝑇𝑝 𝑋 imply that 𝑣 = 0. Now let 𝜓𝑝 ∶ 𝑉𝑝 → 𝐑𝑛 be the map
(𝑓, 𝑔) ∶ 𝑉𝑝 → 𝐑ℓ × 𝐑𝑘 = 𝐑𝑛 .
Then by (5.5.26), 𝑑𝜓𝑝 is bijective, therefore, shrinking 𝑉𝑝 if necessary, we can assume that
𝜓𝑝 maps 𝑉𝑝 diffeomorphically onto a neighborhood 𝑈𝑝 of the origin in 𝐑𝑛 , and hence by
(5.5.24) and (5.5.25)), 𝜓𝑝 maps 𝑉𝑝 ∩ 𝑌 onto the set: 𝑥1 = ⋯ = 𝑥ℓ = 0 and maps 𝑉𝑝 ∩ 𝑍 onto
the set: 𝑥ℓ+1 = ⋯ = 𝑥𝑛 = 0. Finally, if 𝜓 isn’t orientation preserving, we can make it so by
composing it with the involution, (𝑥1 , …, 𝑥𝑛 ) ↦ (𝑥1 , 𝑥2 , …, 𝑥𝑛−1 , −𝑥𝑛 ). □
From this result we deduce:
Theorem 5.5.27. If 𝑌 and 𝑍 intersect transversally, their intersection is a finite set.
Proof. By Theorem 5.5.23 the only point of intersection in 𝑉𝑝 is 𝑝 itself. Moreover, since 𝑌
is closed and 𝑍 is compact, 𝑌 ∩ 𝑍 is compact. Therefore, since the 𝑉𝑝 ’s cover 𝑌 ∩ 𝑍 we can
extract a finite subcover by the Heine–Borel theorem. However, since no two 𝑉𝑝 ’s cover the
same point of 𝑌 ∩ 𝑍, this cover must already be a finite subcover. □
We will now prove Theorem 5.5.22.
Proof of Theorem 5.5.22. Since 𝑌 is closed, the map 𝜄𝑌 ∶ 𝑌 ↪ 𝑋 is proper, so by Theo-
rem 3.4.7 there exists a neighborhood 𝑈 of 𝑍 in 𝑋 such that 𝑈 ∩ 𝑌 is contained in the union
of the open sets 𝑉𝑝 above. Moreover by Theorem 5.5.7 we can choose 𝜏𝑍 to be supported
in 𝑈 and by Theorem 5.3.2 we can assume that 𝑈 has finite topology, so we’re reduced to
proving the theorem with 𝑋 replaced by 𝑈 and 𝑌 replaced by 𝑌 ∩ 𝑈. Let
𝑂 ≔ 𝑈 ∩ ⋃ 𝑉𝑝 ,
𝑝∈𝑋
ℓ
let 𝑓 ∶ 𝑂 → 𝐑 be the map whose restriction to 𝑉𝑝 ∩ 𝑈 is 𝜋 ∘ 𝜓𝑝 where 𝜋 is, as in equa-
tion (5.5.14), the canonical submersion of 𝐑𝑛 onto 𝐑ℓ , and finally let 𝑉 be a bounded con-
vex neighborhood of 𝐑ℓ , whose closure is contained in the intersection of the open sets,
𝜋 ∘ 𝜓𝑝 (𝑉𝑝 ∩ 𝑈). Then 𝑓−1 (𝑉) is a closed subset of 𝑈, so if we replace 𝑋 by 𝑈 and 𝑌 by
𝑌 ∩ 𝑈, the data (𝑓, 𝑂, 𝑉) satisfy Properties 5.5.9. Thus to prove Theorem 5.5.22 it suffices by
Theorem 5.5.10 to prove this theorem with
𝜏𝑌 = 𝜎𝑝 (𝑌)𝑓⋆ 𝜇
on 𝑉𝑝 ∩ 𝑂 where 𝜎𝑝 (𝑌) = +1 or −1 depending on whether the orientation of 𝑌 ∩ 𝑉𝑝 in
Theorem 5.5.10 coincides with the given orientation of 𝑌 or not. Thus
𝐼(𝑌, 𝑍) = (−1)𝑘ℓ 𝐼(𝑍, 𝑌)
= ∑ (−1)𝑘ℓ 𝜎𝑝 (𝑌) ∫ (𝜋 ∘ 𝜓𝑝 ∘ 𝜄𝑍 )⋆ 𝜇 .
𝑝∈𝑋 𝑍∩𝑉𝑝
Draft: March 28, 2018
∫ (𝜋 ∘ 𝜓𝑝 ∘ 𝜄𝑍 )⋆ 𝜇 = 𝜎𝑝 (𝑍) ∫ 𝜇 = 𝜎𝑝 (𝑍) ,
𝑍∩𝑈𝑝 𝑉
We will leave as an exercise the task of unraveling these orientations and showing that
(−1)𝑘ℓ 𝜎𝑝 (𝑌)𝜎𝑝 (𝑍) = 𝐼𝑝 (𝑌, 𝑍)
and hence that 𝐼(𝑌, 𝑍) = ∑𝑝∈𝑋 𝐼𝑝 (𝑌, 𝑍). □
Exercise 5.5.vii. Show that in every neighborhood 𝑉1 of 𝑞 in 𝑉 there exists a point 𝑞1 whose
preimage
𝑌1 ≔ 𝑓−1 (𝑞1 )
intersects 𝑍 transversally. Conclude that one can “deform 𝑌 an arbitrarily small amount so
that it intersects 𝑍 transversally”.
Hint: Exercise 5.5.iv plus Sard’s theorem.
Exercise 5.5.viii (Intersection theory for mappings). Let 𝑋 be an oriented, connected 𝑛-
manifold, 𝑍 a compact, oriented ℓ-dimensional submanifold, 𝑌 an oriented manifold of
dimension 𝑘 ≔ 𝑛 − ℓ and 𝑓 ∶ 𝑌 → 𝑋 a proper 𝐶∞ map. Define the intersection number of
𝑓 with 𝑍 to be the integral
𝐼(𝑓, 𝑍) ≔ ∫ 𝑓⋆ 𝜏𝑍 .
𝑌
(1) Show that 𝐼(𝑓, 𝑍) is a homotopy invariant of 𝑓, i.e., show that if 𝑓0 , 𝑓1 ∶ 𝑌 → 𝑋 are
proper 𝐶∞ maps and are properly homotopic, then
𝐼(𝑓0 , 𝑍) = 𝐼(𝑓1 , 𝑍) .
(2) Show that if 𝑌 is a closed submanifold of 𝑋 of dimension 𝑘 = 𝑛 − ℓ and 𝜄𝑌 ∶ 𝑌 → 𝑋 is
the inclusion map, then 𝐼(𝜄𝑌 , 𝑍) = 𝐼(𝑌, 𝑍).
Exercise 5.5.ix.
(1) Let 𝑋 be an oriented connected 𝑛-manifold and let 𝑍 be a compact zero-dimensional
submanifold consisting of a single point 𝑧0 ∈ 𝑋. Show that if 𝜇 is in 𝛺𝑛𝑐 (𝑋) then 𝜇 is a
Thom form for 𝑍 if and only if its integral is 1.
(2) Let 𝑌 be an oriented 𝑛-manifold and 𝑓 ∶ 𝑌 → 𝑋 a 𝐶∞ map. Show that for 𝑍 = {𝑧0 } as
in part (1) we have 𝐼(𝑓, 𝑍) = deg(𝑓).
An example of a graph is the graph of the identity map of 𝑋 onto itself. This is the
diagonal in 𝑋 × 𝑋
𝛥 ≔ { (𝑥, 𝑥) | 𝑥 ∈ 𝑋 } ⊂ 𝑋 × 𝑋
and its intersection with 𝛤𝑓 is the set
𝛤𝑓 ∩ 𝛥 = { (𝑥, 𝑥) | 𝑓(𝑥) = 𝑥 } ,
which is just the set of fixed points of 𝑓. Hence a natural way to count the fixed points of 𝑓 is
as the intersection number of 𝛤𝑓 and 𝛥 in 𝑋 × 𝑋. To do so we need these three manifolds to
be oriented, but, as we noted above, 𝛤𝑓 and 𝛥 acquire orientations from the identifications
(5.6.1) and, as for 𝑋×𝑋, we’ll give it its natural orientation as a product of oriented manifolds.
(See §4.5.)
Definition 5.6.2. The global Lefschetz number of 𝑓 is the intersection number
𝐿(𝑓) ≔ 𝐼(𝛤𝑓 , 𝛥) .
In this section we’ll give two recipes for computing this number: one by topological
methods and the other by making transversality assumptions and computing this number
as a sum of local intersection numbers via Theorem 5.5.22. We first show what one gets from
the transversality approach.
Definition 5.6.3. The map 𝑓 ∶ 𝑋 → 𝑋 is a Lefschetz map, or simply Lefschetz, if 𝛤𝑓 and 𝛥
intersect transversally.
Let us see what being Lefschetz entails. Suppose 𝑝 is a fixed point of 𝑓. Then at the point
𝑞 = (𝑝, 𝑝) of 𝛤𝑓
(5.6.4) 𝑇𝑞 (𝛤𝑓 ) = (𝑑𝛾𝑓 )𝑝 𝑇𝑝 𝑋 = { (𝑣, 𝑑𝑓𝑝 (𝑣)) | 𝑣 ∈ 𝑇𝑝 𝑋 }
and, in particular, for the identity map,
𝑇𝑞 (𝛥) = { (𝑣, 𝑣) | 𝑣 ∈ 𝑇𝑝 𝑋 } .
Therefore, if 𝛥 and 𝛤𝑓 are to intersect transversally, the intersection of 𝑇𝑞 (𝛤𝑓 ) ∩ 𝑇𝑞 (𝛥) inside
𝑇𝑞 (𝑋 × 𝑋) has to be the zero space. In other words if
(5.6.5) (𝑣, 𝑑𝑓𝑝 (𝑣)) = (𝑣, 𝑣)
then 𝑣 = 0. But the identity (5.6.5) says that 𝑣 is a fixed point of 𝑑𝑓𝑝 , so transversality at 𝑝
amounts to the assertion
𝑑𝑓𝑝 (𝑣) = 𝑣 ⟺ 𝑣 = 0 ,
or in other words the assertion that the map
(5.6.6) (id𝑇𝑝 𝑋 −𝑑𝑓𝑝 ) ∶ 𝑇𝑝 𝑋 → 𝑇𝑝 𝑋
is bijective.
Proposition 5.6.7. The local intersection number 𝐼𝑝 (𝛤𝑓 , 𝛥) is 1 if (5.6.6) is orientation-preserving
and −1 otherwise.
In other words 𝐼𝑝 (𝛤𝑓 , 𝛥) is the sign of det(id𝑇𝑝 𝑋 −𝑑𝑓𝑝 ).
Let 𝜋1 and 𝜋2 be the projection of 𝑋 × 𝑋 onto its first and second factors, i.e., for 𝑖 = 1, 2
let
𝜋𝑖 ∶ 𝑋 × 𝑋 → 𝑋
be the projection 𝜋𝑖 (𝑥1 , 𝑥2 ) ≔ 𝑥𝑖 . Then by equation (5.6.1) we have
(5.6.11) 𝜋1 ⋅ 𝛾𝑓 = id𝑋
and
(5.6.12) 𝜋2 ⋅ 𝛾𝑓 = 𝑓 .
𝑛
Lemma 5.6.13. If 𝜔1 and 𝜔2 are in 𝛺 (𝑋) then
Proof. By a partition of unity argument we can assume that 𝜔𝑖 has compact support in a
parametrizable open set, 𝑉𝑖 . Let 𝑈𝑖 be an open subset of 𝐑𝑛 and 𝜙𝑖 ∶ 𝑈𝑖 → 𝑉𝑖 an orientation-
preserving diffeomorphism. Then
𝜙⋆𝑖 𝜔 = 𝜌𝑖 𝑑𝑥1 ∧ ⋯ ∧ 𝑑𝑥𝑛
with 𝜌𝑖 ∈ 𝐶0∞ (𝑈𝑖 ), so the right hand side of (5.6.14) is the product of integrals over 𝐑𝑛 :
This is the corollary of the Künneth theorem that we alluded to above (and whose proof
we’ll give in §5.7). Using these results we prove the following.
Theorem 5.6.19. The Thom class 𝑇𝛥 ∈ 𝐻𝑛 (𝑋 × 𝑋) is given explicitly by the formula
𝑑𝑘
♯ ♯
(5.6.20) 𝑇𝛥 = ∑ (−1)ℓ ∑ 𝜋1 𝜇𝑘𝑖 ⋅ 𝜋2 𝜈𝑖ℓ .
𝑘+ℓ=𝑛 𝑖=1
Proof. We have to check that for every cohomology class 𝑐 ∈ 𝐻𝑛 (𝑋×𝑋), the class 𝑇𝛥 defined
by (5.6.20) has the reproducing property
♯
(5.6.21) 𝐼𝑋×𝑋 (𝑇𝛥 ⋅ 𝑐) = 𝐼𝛥 (𝜄𝛥 𝑐)
where 𝜄𝛥 is the inclusion map of 𝛥 into 𝑋 × 𝑋. However the map
𝛾𝛥 ∶ 𝑋 → 𝑋 × 𝑋 , 𝑥 ↦ (𝑥, 𝑥)
is an orientation-preserving diffeomorphism of 𝑋 onto 𝛥, so it suffices to show that
♯
(5.6.22) 𝐼𝑋×𝑋 (𝑇𝛥 ⋅ 𝑐) = 𝐼𝑋 (𝛾𝛥 𝑐)
and by Lemma 5.6.18 it suffices to verify (5.6.22) for 𝑐’s of the form
♯ ♯
𝑐 = 𝜋1 𝜈𝑟ℓ ⋅ 𝜋2 𝜇𝑘𝑠 .
The product of this class with a typical summand of (5.6.20), for instance, the summand
′ ♯ ′ ♯ ′
(5.6.23) (−1)ℓ 𝜋1 𝜇𝑘𝑖 ⋅ 𝜋2 𝜈𝑖ℓ , 𝑘′ + ℓ′ = 𝑛 ,
is equal, up to sign to,
♯ ′ ♯ ′
𝜋1 𝜇𝑘𝑖 ⋅ 𝜈𝑟ℓ ⋅ 𝜋2 𝜇𝑘𝑠 ⋅ 𝜈𝑖ℓ .
Notice, however, that if 𝑘 ≠ 𝑘′ this product is zero: For 𝑘 < 𝑘′ , 𝑘′ + ℓ is greater than 𝑘 + ℓ
and hence greater than 𝑛. Therefore
′ ′
𝜇𝑘𝑖 ⋅ 𝜈𝑟ℓ ∈ 𝐻𝑘 +ℓ (𝑋)
′
is zero since 𝑋 is of dimension 𝑛, and for 𝑘 > 𝑘′ , ℓ′ is greater than ℓ and 𝜇𝑘𝑠 ⋅ 𝜈𝑖ℓ is zero for
the same reason. Thus in taking the product of 𝑇𝛥 with 𝑐 we can ignore all terms in the sum
except for the terms, 𝑘′ = 𝑘 and ℓ′ = ℓ. For these terms, the product of (5.6.23) with 𝑐 is
♯ ♯
(−1)𝑘ℓ 𝜋1 𝜇𝑘𝑖 ⋅ 𝜈𝑟ℓ ⋅ 𝜋2 𝜇𝑘3 ⋅ 𝜈𝑖ℓ .
2
(Exercise: Check this. Hint: (−1)ℓ (−1)ℓ = 1.) Thus
𝑑𝑘
♯ ♯
𝑇𝛥 ⋅ 𝑐 = (−1)𝑘ℓ ∑ 𝜋1 𝜇𝑘𝑖 ⋅ 𝜈𝑟ℓ ⋅ 𝜋2 𝜇𝑘𝑠 ⋅ 𝜈𝑖ℓ
𝑖=1
♯ ♯
On the other hand for 𝑐 = 𝜋1 𝜈𝑟ℓ ⋅ 𝜋2 𝜇𝑘𝑠
♯ ♯ ♯ ♯ ♯
𝛾𝛥 𝑐 = 𝛾𝛥 𝜋1 𝜈𝑟ℓ ⋅ 𝛾𝛥 𝜋2 𝜇𝑘𝑠
= (𝜋1 ⋅ 𝛾𝛥 )♯ 𝜈𝑟ℓ (𝜋2 ⋅ 𝛾𝛥 )♯ 𝜇𝑘𝑠
= 𝜈𝑟ℓ ⋅ 𝜇𝑘𝑠
since
𝜋1 ⋅ 𝜈𝛥 = 𝜋2 ⋅ 𝛾𝛥 = id𝑋 .
So
♯
𝐼𝑋 (𝛾𝛥 𝑐) = 𝐼𝑋 (𝜈𝑟ℓ ⋅ 𝜇𝑘𝑠 ) = (−1)𝑘ℓ 𝛿𝑟,𝑠
by (5.6.17). Thus the two sides of (5.6.21) are equal. □
We’re now in position to compute 𝐿(𝑓) , i.e., to compute the expression on 𝐼𝑋 (𝛾𝑓⋆ 𝑇𝛥 )
ℓ ℓ ℓ
the right hand side of (5.6.10). Since 𝜈1 , …, 𝜈𝑑ℓ is a basis of 𝐻 (𝑋) the linear mapping
(5.6.24) 𝑓♯ ∶ 𝐻ℓ (𝑋) → 𝐻ℓ (𝑋)
ℓ
can be described in terms of this basis by a matrix [𝑓𝑖,𝑗 ] with the defining property
𝑑ℓ
𝑓♯ 𝜈𝑖ℓ = ∑ 𝑓𝑗,𝑖
ℓ ℓ
𝜈𝑗 .
𝑗=1
(2) Show that if id𝑉 −𝐹(𝑡0 ) ∶ 𝑉 → 𝑉 is bijective for some 𝑡0 , then 𝐴 ∶ 𝑉 → 𝑉 is bijective.
Hint: Show that if 𝐴𝑣 = 0 for some 𝑣 ∈ 𝑉 − {0}, 𝐹(𝑡)𝑣 = 𝑣.
Exercise 5.6.viii. Let 𝒗 be a vector field on a compact manifold 𝑋 and let (5.6.26) be the
one-parameter group of diffeomorphisms generated by 𝒗. If 𝒗(𝑝) = 0 then by part (1) of
Exercise 5.6.vi, 𝑝 is a fixed point of 𝑓𝑡 for all 𝑡.
(1) Show that
(𝑑𝑓𝑡 ) ∶ 𝑇𝑝 𝑋 → 𝑇𝑝 𝑋
is a one-parameter group of linear mappings of 𝑇𝑝 𝑋 onto itself.
(2) Conclude from Exercise 5.6.vii that there exists a linear map
(5.6.27) 𝐿𝒗 (𝑝) ∶ 𝑇𝑝 𝑋 → 𝑇𝑝 𝑋
with the property
exp 𝑡𝐿𝒗 (𝑝) = (𝑑𝑓𝑡 )𝑝 .
Exercise 5.6.ix. Suppose 𝑓𝑡0 is a Lefschetz map for some 𝑡0 . Let 𝑎 = 𝑡0 /𝑁 where 𝑁 is a
positive integer. Show that 𝑓𝑎 is a Lefschetz map.
Hints:
• Show that
𝑓𝑡0 = 𝑓𝑎 ∘ ⋯ ∘ 𝑓𝑎 = 𝑓𝑎𝑁
(i.e., 𝑓𝑎 composed with itself 𝑁 times).
• Show that if 𝑝 is a fixed point of 𝑓𝑎 , it is a fixed pointof 𝑓𝑡0 .
• Conclude from Exercise 5.6.vi that the fixed points of 𝑓𝑎 are the zeros of 𝑣.
• Show that if 𝑝 is a fixed point of 𝑓𝑎 ,
(𝑑𝑓𝑡0 )𝑝 = (𝑑𝑓𝑎 )𝑝𝑁 .
• Conclude that if (𝑑𝑓𝑎 )𝑝 𝑣 = 𝑣 for some 𝑣 ∈ 𝑇𝑝 𝑋 ∖ {0}, then (𝑑𝑓𝑡0 )𝑝 𝑣 = 𝑣.
Exercise 5.6.x. Show that for all 𝑡, 𝐿(𝑓𝑡 ) = 𝜒(𝑋).
Hint: Exercise 5.6.ii.
Exercise 5.6.xi (Hopf Theorem). A vector field 𝒗 on a compact manifold 𝑋 is a Lefschetz
vector field if for some 𝑡0 ∈ 𝐑 the map 𝑓𝑡0 is a Lefschetz map.
(1) Show that if 𝒗 is a Lefschetz vector field then it has a finite number of zeros and for each
zero 𝑝 the linear map (5.6.27) is bijective.
(2) For a zero 𝑝 of 𝒗 let 𝜎𝑝 (𝒗) = +1 if the map (5.6.27) is orientation-preserving and let
𝜎𝑝 (𝒗) = −1 if it’s orientation-reversing. Show that
𝜒(𝑋) = ∑ 𝜎𝑝 (𝒗) .
𝒗(𝑝)=0
(3) Let 𝑉 be and 𝑛-dimensional vector space and 𝐿 ∶ 𝑉 → 𝑉 a linear map. Fix a basis
𝑣1 , …, 𝑣𝑛 of 𝑉 and define the trace of 𝐿 to be the trace of 𝐴 where 𝐴 is the defining
matrix for 𝐿 in this basis, i.e.,
𝑛
𝐿𝑣𝑖 = ∑ 𝑎𝑗,𝑖 𝑣𝑗 .
𝑗=1
be a basis of 𝐻ℓ (𝑌). Then for 𝑘 + ℓ = 𝑚 the product, 𝜋♯ 𝜇𝑘𝑖 ⋅ 𝜌♯ 𝜈𝑗ℓ , is in 𝐻𝑚 (𝑋 × 𝑌). The
Künneth theorem asserts
Theorem 5.7.1. The product manifold 𝑋 × 𝑌 has finite topology and hence the cohomology
groups 𝐻𝑚 (𝑋 × 𝑌) are finite dimensional. Moreover, the products over 𝑘 + ℓ = 𝑚
(5.7.2) 𝜋♯ 𝜇𝑘𝑖 ⋅ 𝜌♯ 𝜈𝑗ℓ , 0 ≤ 𝑖 ≤ dim 𝐻𝑘 (𝑋) , 0 ≤ 𝑗 ≤ dim 𝐻ℓ (𝑌) ,
are a basis for the vector space 𝐻𝑚 (𝑋 × 𝑌).
The fact that 𝑋 × 𝑌 has finite topology is easy to verify. If 𝑈1 , …, 𝑈𝑀 , is a good cover of
𝑋 and 𝑉1 , …, 𝑉𝑁 , is a good cover of 𝑌 the products of these open sets 𝑈𝑖 × 𝑉𝑗 , for 1 ≤ 𝑖 ≤ 𝑀
and 1 ≤ 𝑗 ≤ 𝑁 is a good cover of 𝑋 × 𝑌: For every multi-index 𝐼, 𝑈𝐼 is either empty or
diffeomorphic to 𝐑𝑛 , and for every multi-index 𝐽, 𝑉𝐽 is either empty or diffeomorphic to 𝐑𝑟 ,
hence for any product multi-index (𝐼, 𝐽), the product 𝑈𝐼 ×𝑉𝐽 is either empty or diffeomorphic
to 𝐑𝑛 × 𝐑𝑟 . The tricky part of the proof is verifying that the products from (5.7.2) are a basis
of 𝐻𝑚 (𝑋 × 𝑌), and to do this it will be helpful to state the theorem above in a form that
avoids our choosing specified bases for 𝐻𝑘 (𝑋) and 𝐻ℓ (𝑌). To do so we’ll need to generalize
slightly the notion of a bilinear pairing between two vector space.
Definition 5.7.3. Let 𝑉1 , 𝑉2 and 𝑊 be finite dimensional vector spaces. A map 𝐵 ∶ 𝑉1 ×𝑉2 →
𝑊 of sets is a bilinear map if it is linear in each of its factors, i.e., for 𝑣2 ∈ 𝑉2 the map
𝑣1 ∈ 𝑉1 ↦ 𝐵(𝑣1 , 𝑣2 )
is a linear map of 𝑉1 into 𝑊 and for 𝑣1 ∈ 𝑉1 so is the map
𝑣2 ∈ 𝑉2 ↦ 𝐵(𝑣1 , 𝑣2 ) .
Draft: March 28, 2018
It’s clear that if 𝐵1 and 𝐵2 are bilinear maps of 𝑉1 × 𝑉2 into 𝑊 and 𝜆1 and 𝜆2 are real
numbers the function
𝜆1 𝐵1 + 𝜆2 𝐵2 ∶ 𝑉1 × 𝑉2 → 𝑊
is also a bilinear map of 𝑉1 × 𝑉2 into 𝑊, so the set of all bilinear maps of 𝑉1 × 𝑉2 into 𝑊
forms a vector space. In particular the set of all bilinear maps of 𝑉1 × 𝑉2 into 𝐑 is a vector
space, and since this vector space will play an essential role in our intrinsic formulation of
the Künneth theorem, we’ll give it a name. We’ll call it the tensor product of 𝑉1⋆ and 𝑉2⋆ and
denote it by 𝑉1⋆ ⊗ 𝑉2⋆ . To explain where this terminology comes from we note that if ℓ1 and
ℓ2 are vectors in 𝑉1⋆ and 𝑉2⋆ then one can define a bilinear map
ℓ1 ⊗ ℓ2 ∶ 𝑉1 × 𝑉2 → 𝐑
by setting (ℓ1 ⊗ ℓ2 )(𝑣1 , 𝑣2 ) ≔ ℓ1 (𝑣1 )ℓ2 (𝑣2 ). In other words one has a tensor product map:
(5.7.4) 𝑉1⋆ × 𝑉2⋆ → 𝑉1⋆ ⊗ 𝑉2⋆
mapping (ℓ1 , ℓ2 ) to ℓ1 ⊗ ℓ2 . We leave for you to check that this is a bilinear map of 𝑉1⋆ × 𝑉2⋆
into 𝑉1⋆ ⊗ 𝑉2⋆ and to check as well
Proposition 5.7.5. If ℓ11 , …, ℓ𝑚
1
is a basis of 𝑉1⋆ and ℓ12 , …, ℓ𝑛2 is a basis of 𝑉2⋆ then ℓ𝑖1 ⊗ ℓ𝑗2 , for
1 ≤ 𝑖 ≤ 𝑚 and 1 ≤ 𝑗 ≤ 𝑛, is a basis of 𝑉1⋆ ⊗ 𝑉2⋆ .
Hint: If 𝑉1 and 𝑉2 are the same vector space you can find a proof of this in §1.3 and the
proof is basically the same if they’re different vector spaces.
Corollary 5.7.6. Let 𝑉1 and 𝑉2 be finite-dimensional vector spaces. Then
dim(𝑉1⋆ ⊗ 𝑉2⋆ ) = dim(𝑉1⋆ ) dim(𝑉2⋆ ) = dim(𝑉1 ) dim(𝑉2 ) .
We’ll now perform some slightly devious maneuvers with “duality” operations. First
note that for any finite dimensional vector space 𝑉, the evaluation pairing
𝑉 × 𝑉⋆ → 𝐑 , (𝑣, ℓ) ↦ ℓ(𝑣)
is a non-singular bilinear pairing, so, as we explained in §5.4 it gives rise to a bijective linear
mapping
(5.7.7) 𝑉 → (𝑉⋆ )⋆ .
Next note that if
(5.7.8) 𝐿 ∶ 𝑉1 × 𝑉2 → 𝑊
is a bilinear mapping and ℓ ∶ 𝑊 → 𝐑 a linear mapping (i.e., an element of 𝑊⋆ ), then the
composition of ℓ and 𝐿 is a bilinear mapping
ℓ ∘ 𝐿 ∶ 𝑉1 × 𝑉2 → 𝐑
and hence by definition an element of 𝑉1⋆ ⊗ 𝑉2⋆ . Thus from the bilinear mapping (5.7.8) we
get a linear mapping
(5.7.9) 𝐿♯ ∶ 𝑊⋆ → 𝑉1⋆ ⊗ 𝑉2⋆ .
We’ll now define a notion of tensor product for the vector spaces 𝑉1 and 𝑉2 themselves.
Definition 5.7.10. The vector space 𝑉1 ⊗ 𝑉2 is the vector space dual of 𝑉1⋆ ⊗ 𝑉2⋆ , i.e., is the
space
(5.7.11) 𝑉1 ⊗ 𝑉2 ≔ (𝑉1⋆ ⊗ 𝑉2⋆ )⋆ .
Draft: March 28, 2018
Notice that by Proposition 5.7.13 the property (5.7.14) is the defining property of 𝐿# ;
it uniquely determines this map. (This is in fact the whole point of the tensor product con-
struction. Its purpose is to convert bilinear maps, which are not linear, into linear maps.)
After this brief digression (into an area of mathematics which some mathematicians
unkindly refer to as “abstract nonsense”), let us come back to our motive for this digression:
an intrinsic formulation of the Künneth theorem. As above let 𝑋 and 𝑌 be manifolds of
dimension 𝑛 and 𝑟, respectively, both having finite topology. For 𝑘+ℓ = 𝑚 one has a bilinear
map
𝐻𝑘 (𝑋) × 𝐻ℓ (𝑌) → 𝐻𝑚 (𝑋 × 𝑌)
mapping (𝑐1 , 𝑐2 ) to 𝜋⋆ 𝑐1 ⋅ 𝜌⋆ 𝑐2 , and hence by Proposition 5.7.13 a linear map
(5.7.15) 𝐻𝑘 (𝑋) ⊗ 𝐻ℓ (𝑌) → 𝐻𝑚 (𝑋 × 𝑌) .
Let
𝐻1𝑚 (𝑋 × 𝑌) ≔ ⨁ 𝐻𝑘 (𝑋) ⊗ 𝐻ℓ (𝑌) .
𝑘+ℓ=𝑚
The maps (5.7.15) can be combined into a single linear map
(5.7.16) 𝐻1𝑚 (𝑋 × 𝑌) → 𝐻𝑚 (𝑋 × 𝑌)
and our intrinsic version of the Künneth theorem asserts the following.
Theorem 5.7.17. The map (5.7.16) is bijective.
Here is a sketch of how to prove this. (Filling in the details will be left as a series of
exercises.) Let 𝑈 be an open subset of 𝑋 which has finite topology and let
𝑚
H 1 (𝑈) ≔ ⨁ 𝐻𝑘 (𝑈) ⊗ 𝐻ℓ (𝑌)
𝑘+ℓ=𝑚
and
𝑚
H 2 (𝑈) ≔ 𝐻𝑚 (𝑈 × 𝑌) .
As we’ve just seen there’s a Künneth map
𝑚 𝑚
𝜅 ∶ H 1 (𝑈) → H 2 (𝑈) .
Draft: March 28, 2018
tensor each term in this sequence with 𝐻ℓ (𝑌), and sum over 𝑘 + ℓ = 𝑚.
Exercise 5.7.ii. Show that for H 2 there is a similar sequence.
Hint: Apply Mayer–Vietoris to the open subsets 𝑈1 × 𝑌 and 𝑈2 × 𝑌 of 𝑀.
Exercise 5.7.iii. Show that the diagram below commutes:
𝛿 𝑚 𝑚 𝑚 𝑚 𝛿 𝑚+1
⋯ H 1 (𝑈) H 1 (𝑈1 ) ⊕ H 1 (𝑈2 ) H 1 (𝑈1 ∩ 𝑈2 ) H1 (𝑈) ⋯
𝜅 𝜅 𝜅 𝜅
𝛿 𝑚 𝑚 𝑚 𝑚 𝛿 𝑚+1
⋯ H 2 (𝑈) H 2 (𝑈1 ) ⊕ H 2 (𝑈2 ) H 2 (𝑈1 ∩ 𝑈2 ) H2 (𝑈) ⋯
(This looks hard but is actually very easy: just write down the definition of each arrow in the
language of forms.)
Exercise 5.7.iv. Conclude from Exercise 5.7.iii that if the Künneth map is bijective for 𝑈1 ,
𝑈2 and 𝑈1 ∩ 𝑈2 it is bijective for 𝑈.
Exercise 5.7.v. Prove the Künneth theorem by induction on the number of open sets in a
good cover of 𝑋. To get the induction started, note that
𝐻𝑘 (𝐑𝑛 × 𝑌) ≅ 𝐻𝑘 (𝑌) .
(See Exercise 5.3.xi.)
We claim that 𝛿 actually is a coboundary operator, i.e., 𝛿 ∘ 𝛿 = 0. To see this, we note that
for 𝑐 ∈ 𝐶𝑘−1 (U; 𝐑) and 𝐼 ∈ 𝑁𝑘 (U), by (5.8.1) we have
𝑘+1
𝛿(𝛿(𝑐))(𝐼) = ∑ (−1)𝑟 𝛿(𝑐)(𝐼𝑟 )
𝑟=0
𝑘+1
= ∑ (−1)𝑟 (∑(−1)𝑠 𝑐(𝐼𝑟,𝑠 ) + ∑(−1)𝑠−1 𝑐(𝐼𝑟,𝑠 )) .
𝑟=0 𝑠<𝑟 𝑠>𝑟
Thus the term 𝑐(𝐼𝑟,𝑠 ) occurs twice in the sum, but occurs once with opposite signs, and hence
𝛿(𝛿(𝑐))(𝐼) = 0.
The cochain complex
𝛿 𝛿 𝛿 𝛿
(5.8.2) 0 𝐶0 (U; 𝐑) 𝐶1 (U; 𝐑) ⋯ 𝐶𝑘 (U; 𝐑) ⋯
is called the Čech cochain complex of the cover U. The Čech cohomology groups of the cover
U are the cohomology groups of the Čech cochain complex:
ker(𝛿 ∶ 𝐶𝑘 (U; 𝐑) → 𝐶𝑘+1 (U; 𝐑))
𝐻𝑘 (U; 𝐑) ≔ .
im(𝛿 ∶ 𝐶𝑘−1 (U; 𝐑) → 𝐶𝑘 (U; 𝐑))
The rest of the section is devoted to proving the following theorem:
Theorem 5.8.3. Let 𝑋 be a manifold and U a finite good cover of 𝑋. Then for all 𝑘 ≥ 0 we
have isomorphisms
𝐻𝑘 (U; 𝐑) ≅ 𝐻𝑘 (𝑋) .
Remarks 5.8.4.
(1) The definition of 𝐻𝑘 (U; 𝐑) only involves the nerve 𝑁(U) of this cover so Theorem 5.8.3
is in effect a proof of the claim we made above: the cohomology of 𝑋 is in principle deter-
mined by the intersection properties of the 𝑈𝑖 ’s.
(2) Theorem 5.8.3 gives us another proof of an assertion we proved earlier: if 𝑋 admits a
finite good cover the cohomology groups of 𝑋 are finite-dimensional.
Draft: March 28, 2018
The proof of Theorem 5.8.3 will involve an interesting de Rham theoretic generalization
of the Čech cochain complex (5.8.2). Namely, for 𝑘 and ℓ nonnegative integers we define a
Čech cochain of degree 𝑘 with values in 𝛺ℓ to be a map 𝑐 which assigns each 𝐼 ∈ 𝑁𝑘 (U) an
ℓ-form 𝑐(𝐼) ∈ 𝛺ℓ (𝑈𝐼 ).
The set of these cochains forms an infinite dimensional vector space 𝐶𝑘 (U; 𝛺ℓ ). We
will show how to define a coboundary operator
𝛿 ∶ 𝐶𝑘−1 (U; 𝛺ℓ ) → 𝐶𝑘 (U; 𝛺ℓ )
similar to the Čech coboundary operator (5.8.1). To define this operator, let 𝐼 ∈ 𝑁𝑘 (U) and
for 0 ≤ 𝑟 ≤ 𝑘 let 𝛾𝑟 ∶ 𝛺ℓ (𝑈𝐼𝑟 ) → 𝛺ℓ (𝑈𝐼 ) be the restriction map defined by
𝜔 ↦ 𝜔|𝑈𝐼 .
(Note that the restiction 𝜔|𝑈𝐼 is well-defined since 𝑈𝐼 ⊂ 𝑈𝐼𝑟 .) Thus given a Čech cochain
𝑐 ∈ 𝐶𝑘−1 (U; 𝛺ℓ ) we can, mimicking (5.8.1), define a Čech cochain 𝛿(𝑐) ∈ 𝐶𝑘 (U; 𝛺ℓ ) by
setting
𝑘 𝑘
(5.8.5) 𝛿(𝑐)(𝐼) ≔ ∑ 𝛾𝑟 (𝑐(𝐼𝑟 )) = ∑ 𝑐(𝐼𝑟 )|𝑈𝐼 .
𝑟=0 𝑟=0
In other words, except for the 𝛾𝑟 ’s the definition of this cochain is formally identical with
the definition (5.8.1). We leave it as an exercise that the operators
𝛿 ∶ 𝐶𝑘−1 (U; 𝛺ℓ ) → 𝐶𝑘 (U; 𝛺ℓ )
satisfy 𝛿 ∘ 𝛿 = 0.1
Thus, to summarize, for every nonnegative integer ℓ we have constructed a Čech cochain
complex with values in 𝛺ℓ
𝛿 𝛿 𝛿 𝛿
(5.8.6) 0 𝐶0 (U; 𝛺ℓ ) 𝐶1 (U; 𝛺ℓ ) ⋯ 𝐶𝑘 (U; 𝛺ℓ ) ⋯.
Our next task in this section will be to compute the cohomology of the complex (5.8.6). We’ll
begin with the 0th cohomology group of (5.8.6), i.e., the kernel of the coboundary operator
𝛿 ∶ 𝐶0 (U; 𝛺ℓ ) → 𝐶1 (U; 𝛺ℓ ) .
An element 𝑐 ∈ 𝐶0 (U; 𝛺ℓ ) is, by definition, a map which assigns to each 𝑖 ∈ {1, …, 𝑑} an
element 𝑐(𝑖) ≕ 𝜔𝑖 of 𝛺ℓ (𝑈𝑖 ).
Now let 𝐼 = (𝑖0 , 𝑖1 ) be an element of 𝑁1 (U). Then, by definition, 𝑈𝑖0 ∩ 𝑈𝑖1 is nonempty
and
𝛿(𝑐)(𝐼) = 𝛾𝑖0 𝜔𝑖1 − 𝛾𝑖1 𝜔𝑖0 .
Thus 𝛿(𝑐) = 0 if and only if
𝜔𝑖0 |𝑈𝑖 ∩𝑈𝑖 = 𝜔𝑖1 |𝑈𝑖 ∩𝑈𝑖
0 1 0 1
for all 𝐼 ∈ 𝑁1 (U). This is true if and only if each ℓ-form 𝜔𝑖 ∈ 𝛺ℓ (𝑈𝑖 ) is the restriction to 𝑈𝑖
of a globally defined ℓ-form 𝜔 on 𝑈1 ∪ ⋯ ∪ 𝑈𝑑 = 𝑋. Thus
ker(𝛿 ∶ 𝐶0 (U; 𝛺ℓ ) → 𝐶1 (U; 𝛺ℓ )) = 𝛺ℓ (𝑋) .
Inserting 𝛺ℓ (𝑋) into the sequence (5.8.6) we get a new sequence
𝛿 𝛿 𝛿
(5.8.7) 0 𝛺ℓ (𝑋) 𝐶0 (U; 𝛺ℓ ) ⋯ 𝐶𝑘 (U; 𝛺ℓ ) ⋯,
and we prove:
1Hint: Except for keeping track of the 𝛾 ’s the proof is identical to the proof of the analogous result for the
𝑟
coboundary operator (5.8.1).
Draft: March 28, 2018
Theorem 5.8.8. Let 𝑋 be a manifold and U a finite good cover of 𝑋. Then for each nonnegative
integer ℓ, the sequence (5.8.7) is exact.
We’ve just proved that the sequence (5.8.7) is exact in the first spot. To prove that (5.8.7)
is exact in its 𝑘th spot, we will construct a chain homotopy operator
𝑄 ∶ 𝐶𝑘 (U; 𝛺ℓ ) → 𝐶𝑘+1 (U; 𝛺ℓ )
with the property that
(5.8.9) 𝛿𝑄(𝑐) + 𝑄𝛿(𝑐) = 𝑐
𝑘 ℓ
for all 𝑐 ∈ 𝐶 (U; 𝛺 ). To define this operator we’ll need a slight generalization of Theo-
rem 5.2.22.
Theorem 5.8.10. Let 𝑋 be a manifold with a finite good cover {𝑈1 , …, 𝑈𝑁 }. Then there exist
𝑁
functions 𝜙1 , …, 𝜙𝑁 ∈ 𝐶∞ (𝑋) such that supp(𝜙𝑖 ) ⊂ 𝑈𝑖 for 𝑖 = 1, …, 𝑁 and ∑𝑖=1 𝜙𝑖 = 1.
Proof of Theorem 5.8.10. Let (𝜌𝑖 )𝑖≥1 , where 𝜌𝑖 ∈ 𝐶0∞ (𝑋), be a partition of unity subordinate
to the cover U and let 𝜙1 be the sum of the 𝜌𝑗 ’s with support in 𝑈1 :
𝜙1 ≔ ∑ 𝜌𝑗 .
𝑗≥1
supp(𝜌𝑗 )⊂𝑈1
Deleting those 𝜌𝑗 such that supp(𝜌𝑗 ) ⊂ 𝑈1 from the sequence (𝜌𝑖 )𝑖≥1 , we get a new sequence
(𝜌′𝑖 )𝑖≥1 . Let 𝜙2 be the sum of the 𝜌𝑗′ ’s with support in 𝑈1 :
𝜙2 ≔ ∑ 𝜌𝑗′ .
𝑗≥1
supp(𝜌𝑗′ )⊂𝑈2
Now we delete those 𝜌𝑗′ such that supp(𝜌𝑗′ ) ⊂ 𝑈2 from the sequence (𝜌′𝑖 )𝑖≥1 and construct
𝜙3 by the same method we used to construct 𝜙1 and 𝜙2 , and so on. □
Now we define the operator 𝑄 and prove Theorem 5.8.8.
Proof of Theorem 5.8.8. To define the operator (5.8.9) let 𝑐 ∈ 𝐶𝑘 (U; 𝛺ℓ ). Then we define
𝑄𝑐 ∈ 𝐶𝑘−1 (U; 𝛺ℓ ) by defining its value at 𝐼 ∈ 𝑁𝑘−1 (𝐼) to be
𝑑
𝑄𝑐(𝐼) ≔ ∑ 𝜙𝑖 𝑐(𝑖, 𝑖0 , …, 𝑖𝑘−1 ) ,
𝑖=0
th
where 𝐼 = (𝑖0 , …, 𝑖𝑘−1 ) and the 𝑖 summand on the right is defined by be equal to 0 when 𝑈𝑖 ∩
𝑈𝐼 is empty and defined to be the product of the function 𝜙𝑖 and the ℓ-form 𝑐(𝑖, 𝑖0 , …, 𝑖𝑘−1 )
when 𝑈𝑖 ∩ 𝑈𝐼 is nonempty. (Note that in this case, the multi-index (𝑖, 𝑖0 , …, 𝑖𝑘−1 ) is in 𝑁𝑘 (U)
and so by definition 𝑐(𝑖, 𝑖0 , …, 𝑖𝑘−1 ) is an element of 𝛺ℓ (𝑈𝑖 ∩ 𝑈𝐼 ). However, since 𝜙𝑖 is sup-
ported on 𝑈𝑖 , we can extend the ℓ-form 𝜙𝑖 𝑐(𝑖, 𝑖0 , …, 𝑖𝑘−1 ) ti 𝑈𝐼 by setting it equal to zero
outside of 𝑈𝑖 ∩ 𝑈𝐼 .)
To prove (5.8.9) we note that for 𝑐 ∈ 𝐶𝑘 (U; 𝛺ℓ ) we have
𝑑
𝑄𝛿𝑐(𝑖0 , …, 𝑖𝑘 ) = ∑ 𝜙𝑖 𝛿𝑐(𝑖, 𝑖0 , …, 𝑖𝑘 ) ,
𝑖=1
so by (5.8.5) the right hand side is equal to
𝑑 𝑑 𝑘
(5.8.11) ∑ 𝜙𝑖 𝛾𝑖 𝑐(𝑖0 , …, 𝑖𝑘 ) + ∑ ∑(−1)𝑠+1 𝜙𝑖 𝛾𝑖𝑠 𝑐(𝑖, 𝑖0 , …, 𝚤𝑠̂ , …, 𝑖𝑘 ) ,
𝑖=1 𝑖=1 𝑠=0
Draft: March 28, 2018
However, this is the negative of the second summand of (5.8.11); so, by adding these two
summands we get
𝑑
(𝑄𝛿𝑐 + 𝛿𝑄𝑐)(𝑖0 , …, 𝑖𝑘 ) = ∑ 𝜙𝑖 𝛾𝑖 𝑐(𝑖0 , …, 𝑖𝑘 )
𝑖=1
𝑑
and since ∑𝑖=1 𝜙𝑖 = 1, this identity reduced to
(𝑄𝛿𝑐 + 𝛿𝑄𝑐)(𝐼) = 𝑐(𝐼) ,
𝑘
or, since 𝐼 ∈ 𝑁 (U) is any element, 𝑄𝛿𝑐 + 𝛿𝑄𝑐 = 𝑐.
Finally, note the Theorem 5.8.8 is an immediate consequence of the existence of this
chain homotopy operator. Namely, if 𝑐 ∈ 𝐶𝑘 (U; 𝛺ℓ ) is in the kernel of 𝛿, then
𝑐 = 𝑄𝛿(𝑐) + 𝛿𝑄(𝑐) = 𝛿𝑄(𝑐) ,
so 𝑐 is in the image of 𝛿 ∶ 𝐶 𝑘−1
(U; 𝛺ℓ ) → 𝐶𝑘 (U; 𝛺ℓ ). □
To prove Theorem 5.8.3, we note that in addition to the exact sequence
𝛿 𝛿 𝛿
0 𝛺ℓ (𝑋) 𝐶0 (U; 𝛺ℓ ) ⋯ 𝐶𝑘 (U; 𝛺ℓ ) ⋯,
we have another exact sequence
(5.8.12)
𝑑 𝑑 𝑑 𝑑 𝑑
0 𝐶𝑘 (U; 𝐑) 𝐶𝑘 (U; 𝛺0 ) 𝐶𝑘 (U; 𝛺1 ) ⋯ 𝐶𝑘 (U; 𝛺ℓ ) ⋯,
where the 𝑑’s are defined as follows: given 𝑐 ∈ 𝐶𝑘 (U; 𝛺ℓ ) and 𝐼 ∈ 𝑁𝑘 (U), we have that
𝑐(𝐼) ∈ 𝛺ℓ (𝑈𝐼 ) and 𝑑(𝑐(𝐼)) ∈ 𝛺ℓ+1 (𝑈𝐼 ), and we define 𝑑(𝑐) ∈ 𝐶𝑘 (U; 𝛺ℓ+1 ) to be the map
given by
𝐼 ↦ 𝑑(𝑐(𝐼)) .
It is clear from this definition that 𝑑(𝑑(𝑐)) = 0 so the sequence
𝑑 𝑑 𝑑 𝑑
(5.8.13) 𝐶𝑘 (U; 𝛺0 ) 𝐶𝑘 (U; 𝛺1 ) ⋯ 𝐶𝑘 (U; 𝛺ℓ ) ⋯
is a complex and the exactness of this sequence follows from the fact that, since U is a good
cover, the sequence
𝑑 𝑑 𝑑 𝑑
𝛺0 (𝑈𝐼 ) 𝛺1 (𝑈𝐼 ) ⋯ 𝛺ℓ (𝑈𝐼 ) ⋯
and the 𝛿 operation 𝛿 ∶ 𝐶𝑘 (U; 𝛺ℓ ) → 𝐶𝑘+1 (U; 𝛺ℓ ), commute. i.e., for 𝑐 ∈ 𝐶𝑘 (U; 𝛺ℓ ) we
have
(5.8.14) 𝛿𝑑𝑐 = 𝑑𝛿𝑐 .
and
𝑘
𝑑𝛿(𝐼) = ∑ 𝑑𝜔𝐼𝑟 |𝑈𝐼 = 𝛿𝑑𝑐(𝐼) . □
𝑟=0
⋮ ⋮ ⋮ ⋮
𝛿 𝛿
0 𝛺2 (𝑋) 𝐶0 (U; 𝛺2 ) 𝐶1 (U; 𝛺2 ) 𝐶2 (U; 𝛺2 ) ⋯
𝑑 𝑑 𝑑 𝑑
𝑑 𝑑 𝑑 𝑑
0 0 0 .
In this diagram all columns are exact except for the extreme left and column, which is the
usual de Rham complex, and all rows are exact except for the bottom row which is the usual
Čech cochain complex.
𝑐 𝑐𝑘+1,0
𝑑
𝑐𝑘,0 𝑐𝑘,1
𝛿
𝑑
𝑐𝑘−1,1 𝑐𝑘−1,2
(5.8.15) 𝛿
𝑑
𝑐𝑘−2,2 𝑐𝑘−2,3
𝛿
𝑑
𝑐𝑘−3,3 ⋯
𝛿
Draft: March 28, 2018
APPENDIX a
1
𝑒− 𝑥 , 𝑥 > 0
(a.1) 𝜌(𝑥) ≔ {
0, 𝑥≤0,
which is positive for 𝑥 positive, zero for 𝑥 negative and everywhere 𝐶∞ . (We will sketch a
proof of this last assertion at the end of this appendix). From 𝜌 one can construct a number
of other interesting 𝐶∞ functions.
Example a.2. For 𝑎 > 0 the function 𝜌(𝑥) + 𝜌(𝑎 − 𝑥) is positive for all 𝑥 so the quotient
𝜌𝑎 (𝑥) of 𝜌(𝑥) by this function is a well-defined 𝐶∞ function with the properties:
{𝜌 (𝑥) = 0, for 𝑥 ≤ 0
{ 𝑎
{0 ≤ 𝜌𝑎 (𝑥) ≤ 1
{
{𝜌𝑎 (𝑥) = 1, for 𝑥 ≥ 𝑎 .
Example a.3. Let 𝐼 be the open interval (𝑎, 𝑏), and 𝜌𝐼 the product of 𝜌(𝑥 − 𝑎) and 𝜌(𝑏 − 𝑥).
Then 𝜌𝐼 (𝑥) is positive for 𝑥 ∈ 𝐼 and zero on the complement of 𝐼.
Example a.4. More generally let 𝐼1 , …, 𝐼𝑛 be open intervals and let 𝑄 = 𝐼1 × ⋯ × 𝐼𝑛 be the
open rectangle in 𝐑𝑛 having these intervals as sides. Then the function
𝜌𝑄 (𝑥) ≔ 𝜌𝐼1 (𝑥1 ) ⋯ 𝜌𝐼𝑛 (𝑥𝑛 )
is a 𝐶∞ function on 𝐑𝑛 that’s positive on 𝑄 and zero on the complement of 𝑄.
Using these functions we will prove:
Lemma a.5. Let 𝐶 be a compact subset of 𝐑𝑛 and 𝑈 an open set containing 𝐶. Then there
exists function 𝜙 ∈ 𝐶0∞ (𝑈) such that
𝜙(𝑥) ≥ 0, for all 𝑥 ∈ 𝑈
{
𝜙(𝑥) > 0, for 𝑥 ∈ 𝐶 .
Proof. For each 𝑝 ∈ 𝐶 let 𝑄𝑝 be an open rectangle with 𝑝 ∈ 𝑄𝑝 and 𝑄𝑝 ⊂ 𝑈. The 𝑄𝑝 ’s cover
𝑁
𝐶𝑖 so, by Heine–Borel there exists a finite subcover 𝑄𝑝1 , …, 𝑄𝑝𝑁 . Now let 𝜙 ≔ ∑𝑖=1 𝜌𝑄𝑝 □
𝑖
Proof. Let 𝜙 be as in Lemma a.5 and let 𝑎 > 0 be the greatest lower bound of the restriction
of 𝜙 to 𝐶. Then if 𝜌𝑎 is the function in Example a.2 the function 𝜌𝑎 ∘ 𝜙 has the properties
indicated above. □
Remark a.7. The function 𝜓 in this theorem is an example of a bump function. If one wants
to study the behavior of a vector field 𝒗 or a 𝑘-form 𝜔 on the set 𝐶, then by multiplying 𝒗
(or 𝜔) by 𝜓 one can, without loss or generality, assume that 𝒗 (or 𝜔) is compactly supported
on a small neighborhood of 𝐶.
Bump functions are one of the standard tools in calculus for converting global problems
to local problems. Another such tool is partitions of unity.
Let 𝑈 be an open subset of 𝐑𝑛 and U = {𝑈𝛼 }𝛼∈𝐼 a covering of 𝑈 by open subsets (in-
dexed by the elements of the index set 𝐼). Then the partition of unity theorem asserts:
Theorem a.8. There exists a sequence of functions 𝜌1 , 𝜌2 , … ∈ 𝐶0∞ (𝑈) such that
(1) 𝜌𝑖 ≥ 0
(2) For every 𝑖 there is an 𝛼 ∈ 𝐼 with 𝜌𝑖 ∈ 𝐶0∞ (𝑈𝛼 )
(3) For every 𝑝 ∈ 𝑈 there exists a neighborhood 𝑈𝑝 of 𝑝 in 𝑈 and an 𝑁𝑝 > 0 such that
𝜌𝑖 |𝑈𝑝 = 0 for 𝑢 > 𝑁𝑝 .
(4) ∑ 𝜌𝑖 = 1
Remark a.9. Because of item (3) the sum in item (4) is well defined. We will derive this
result from a somewhat simpler set theoretical result.
Theorem a.10. There exists a countable covering of 𝑈 by open rectangles (𝑄𝑖 )𝑖≥1 such that
(1) 𝑄𝑖 ⊂ 𝑈
(2) For each 𝑖 there is an 𝛼 ∈ 𝐼 with 𝑄𝑖 ⊂ 𝑈𝛼
(3) For every 𝑝 ∈ 𝑈 there exists a neighborhood 𝑈𝑝 of 𝑝 in 𝑈 and 𝑁𝑝 > 0 such that 𝑄𝑖 ∩ 𝑈𝑝
is empty for 𝑖 > 𝑁𝑝
We first note that Theorem a.10 implies Theorem a.8. To see this note that the func-
tions 𝜌𝑄𝑖 in Example a.4 above have all the properties indicated in Theorem a.8 except for
∞
property (4). Moreover since the 𝑄𝑖 ’s are a covering of 𝑈 the sum ∑𝑖=1 𝜌𝑄𝑖 is everywhere
positive. Thus we get a sequence of 𝜌𝑖 ’s satisfying (1)–(4) by taking 𝜌𝑖 to be the quotient of
𝜌𝑄𝑖 by this sum.
Proof of Theorem a.10. Let 𝑑(𝑥, 𝑈𝑐 ) be the distance of a point 𝑥 ∈ 𝑈 to the complement
𝑈𝑐 ≔ 𝐑𝑛 ∖ 𝑈 of 𝑈 in 𝐑𝑛 , and let 𝐴 𝑟 be the compact subset of 𝑈 consisting of points, 𝑥 ∈ 𝑈,
satisfying 𝑑(𝑥, 𝑈𝑐 ) ≥ 1/𝑟 and |𝑥| ≤ 𝑟. By Heine–Borel we can find, for each 𝑟, a collection of
open rectangles, 𝑄𝑟,𝑖 , 𝑖 = 1, …, 𝑁𝑟 , such that 𝑄𝑟,𝑖 is contained in int(𝐴 𝑟+1 ∖𝐴 𝑟−2 ) and in some
𝑈𝛼 and such that the 𝑄𝑖 , 𝑟’s are a covering of 𝐴 𝑟 ∖int(𝐴 𝑟−1 ). Thus the 𝑄𝑟,𝑖 ’s have the properties
listed in Theorem a.10, and by relabelling, i.e., setting 𝑄𝑖 = 𝑄1,𝑖 for 1 ≤ 𝑖 ≤ 𝑁𝑖 , 𝑄𝑁1 +𝑖 ≔ 𝑄2,𝑖 ,
for 1 ≤ 𝑖 ≤ 𝑁2 , etc., we get a sequence, 𝑄1 , 𝑄2 , … with the desired properties. □
Applications
We will next describe a couple of applications of Theorem a.10.
Application a.11 (Improper integrals). Let 𝑓 ∶ 𝑈 → 𝐑 be a continuous function. We will
say that 𝑓 is integrable over 𝑈 if the infinite sum
∞
(a.12) ∑ ∫ 𝜌𝑖 (𝑥) |𝑓(𝑥)| 𝑑𝑥
𝑖=1 𝑈
Draft: March 28, 2018
203
converges and if so we will define the improper integral of 𝑓 over 𝑈 to be the sum
∞
(a.13) ∑ ∫ 𝜌𝑖 (𝑥)𝑓(𝑥)𝑑𝑥
𝑖=1 𝑈
(Notice that each of the summands in this series is the integral of a compactly supported
continuous function over 𝐑𝑛 so the individual summands are well-defined. Also it’s clear
that if 𝑓 itself is a compactly supported function on 𝐑𝑛 , (a.12) is bounded by the integral of
|𝑓| over 𝐑𝑛 , so for every 𝑓 ∈ 𝐶0 (𝐑𝑛 ) the improper integral of 𝑓 over 𝑈 is well-defined.)
Application a.14 (An extension theorem for 𝐶∞ maps). Let 𝑋 be a subset of 𝐑𝑛 and 𝑓 ∶
𝑋 → 𝐑𝑚 a continuous map. We will say that 𝑓 is 𝐶∞ if, for every 𝑝 ∈ 𝑋, there exists an
open neighborhood, 𝑈𝑝 , of 𝑝 in 𝐑𝑛 and a 𝐶∞ map, 𝑔𝑝 ∶ 𝑈𝑝 → 𝐑𝑚 such that 𝑔𝑝 = 𝑓 on
𝑈𝑝 ∩ 𝑋.
Theorem a.15 (Extension Theorem). If 𝑓 ∶ 𝑋 → 𝐑𝑚 is 𝐶∞ there exists an open neighbor-
hood 𝑈 of 𝑋 in 𝐑𝑛 and a 𝐶∞ map 𝑔 ∶ 𝑈 → 𝐑𝑚 , such that 𝑔 = 𝑓 on 𝑋.
Proof. Let 𝑈 = ⋃𝑝∈𝑋 𝑈𝑝 and let (𝜌𝑖 )𝑖≥1 be a partition of unity subordinate to the covering
U = {𝑈𝑝 }𝑝∈𝑋 of 𝑈. Then, for each 𝑖, there exists a 𝑝 such that the support of 𝜌𝑖 is contained
in 𝑈𝑝 . Let
𝜌𝑖 𝑔𝑝 (𝑥), 𝑥 ∈ 𝑈𝑝
𝑔𝑖 (𝑥) = {
0, 𝑥 ∈ 𝑈𝑝𝑐 .
∞
Then 𝑔 = ∑𝑖=1 𝑔𝑖 is well defined by item (3) of Theorem a.8 and the restriction of 𝑔 to 𝑋 is
∞
given by ∑𝑖=1 𝜌𝑖 𝑓, which is equal to 𝑓. □
Exercises for Appendix a
Exercise a.i. Show that the function (a.1) is 𝐶∞ .
Hints:
(1) From the Taylor series expansion
∞
𝑥𝑘
𝑒𝑥 = ∑
𝑘=0 𝑘!
conclude that for 𝑥 > 0
𝑥𝑘+𝑛
𝑒𝑥 ≥ .
(𝑘 + 𝑛)!
(2) Replacing 𝑥 by 1/𝑥 conclude that for 𝑥 > 0,
1 1
𝑒1/𝑥 ≥ .
(𝑛 + 𝑘)! 𝑥𝑛+𝑘
(3) From this inequality conclude that for 𝑥 > 0,
1
𝑒− 𝑥 ≤ (𝑛 + 𝑘)!𝑥𝑛+𝑘 .
(4) Let 𝑓𝑛 (𝑥) be the function
1
𝑒− 𝑥 𝑥−𝑛 , 𝑥 > 0
𝑓𝑛 (𝑥) ≔ {
0, 𝑥≤0.
Conclude from (3) that for 𝑥 > 0,
𝑓𝑛 (𝑥) ≤ (𝑛 + 𝑘)!𝑥𝑘
for all 𝑘
Draft: March 28, 2018
APPENDIX b
than 1/2, so 𝑓 is strictly increasing on this interval and its graph looks like the curve in the
figure below with 𝑐 = 𝑓(𝑎) ≤ − 12 𝑎 and 𝑑 = 𝑓(𝑏) ≥ 12 𝑏.
𝑎 𝑏
𝑥
The graph of the inverse function 𝑔 ∶ [𝑐, 𝑑] → [𝑎, 𝑏] is obtained from this graph by just
rotating it through ninety degrees, i.e., making the 𝑦-axis the horizontal axis and the 𝑥-axis
the vertical axis. (From the picture it’s clear that 𝑦 = 𝑓(𝑥) ⟺ 𝑥 = 𝑔(𝑦) ⟺ 𝑓(𝑔(𝑦)) =
𝑦.)
Most elementary text books regard this intuitive argument as being an adequate proof
of the inverse function theorem; however, a slightly beefed-up version of this proof (which is
completely rigorous) can be found in [12, Ch. 12]. Moreover, as Spivak points out in [12, Ch.
12], if the slope of the curve in the figure above at the point (𝑥, 𝑦) is equal to 𝜆 the slope of
the rotated curve at (𝑦, 𝑥) is 1/𝜆, so from this proof one concludes that if 𝑦 = 𝑓(𝑥)
−1 −1
𝑑𝑔 𝑑𝑓 𝑑𝑓
(b.7) (𝑦) = ( (𝑥)) =( (𝑔(𝑦))) .
𝑑𝑦 𝑑𝑥 𝑑𝑥
Since 𝑓 is a continuous function, its graph is a continuous curve and, therefore, since the
graph of 𝑔 is the same curve rotated by ninety degrees, 𝑔 is also a continuous functions.
Hence by (b.7), 𝑔 is also a 𝐶1 function and hence by (b.7), 𝑔 is a 𝐶2 function, etc. In other
words 𝑔 is in 𝐶∞ ([𝑐, 𝑑]).
Let’s now prove that the implicit function theorem with 𝑘 = 1 and 𝑛 arbitrary. This
amounts to showing that if the function 𝑓 in the discussion above depends on the param-
eters, 𝑥2 , …𝑥𝑛 in a 𝐶∞ fashion, then so does its inverse 𝑔. More explicitly let’s suppose
𝑓(𝑥1 , …, 𝑥𝑛 ) is a 𝐶∞ function on a subset of 𝐑𝑛 of the form [𝑎, 𝑏] × 𝑉 where 𝑉 is a compact,
convex neighborhood of the origin in 𝐑𝑛 and satisfies 𝜕𝑓/𝜕𝑥1 ≥ 21 on this set. Then by the
argument above there exists a function, 𝑔 = 𝑔(𝑦, 𝑥2 , …, 𝑥𝑛 ) defined on the set [𝑎/2, 𝑏/2] × 𝑉
with the property
(b.8) 𝑓(𝑥1 , 𝑥2 , …, 𝑥𝑛 ) = 𝑦 ⟺ 𝑔(𝑦, 𝑥2 , …, 𝑥𝑛 ) = 𝑥1 .
Moreover by (b.7) 𝑔 is a 𝐶∞ function of 𝑦 and
𝜕𝑔 𝜕𝑓
(b.9) (𝑦, 𝑥2 , …, 𝑥𝑛 ) = (𝑥 , 𝑥 , …, 𝑥𝑛 )−1
𝜕𝑦 𝜕𝑥1 1 2
Draft: March 28, 2018
207
𝜕𝑓 1
at 𝑥1 = 𝑔(𝑦). In particular, since 𝜕𝑥1
≥ 2
𝜕𝑔
0< <2
𝜕𝑦
and hence
(b.10) |𝑔(𝑦′ , 𝑥2 , …, 𝑥𝑛 ) − 𝑔(𝑦, 𝑥2 , …, 𝑥𝑛 )| < 2|𝑦′ − 𝑦|
for points 𝑦 and 𝑦′ in the interval [𝑎/2, 𝑏/2].
The 𝑘 = 1 case of Theorem b.1 is almost implied by (b.8) and (b.9) except that we must
still show that 𝑔 is a 𝐶∞ function, not just of 𝑦, but of all the variables, 𝑦, 𝑥2 , …, 𝑥𝑛 , and this
we’ll do by quoting another theorem from freshman calculus (this time a theorem from the
second semester of freshman calculus).
Theorem b.11 (Mean Value Theorem in 𝑛 variables). Let 𝑈 be a convex open set in 𝐑𝑛 and
𝑓 ∶ 𝑈 → 𝐑 a 𝐶∞ function. Then for 𝑎, 𝑏 ∈ 𝑈 there exists a point 𝑐 on the line interval joining
𝑎 to 𝑏 such that
𝑛
𝜕𝑓
𝑓(𝑏) − 𝑓(𝑎) = ∑ (𝑐)(𝑏𝑖 − 𝑎𝑖 ) .
𝑖=1 𝜕𝑥 𝑖
209
A second result which we’ll extract from Theorem b.1 is the canonical submersion the-
orem.
Theorem b.17. Let 𝑈 be an open subset of 𝐑𝑛 and 𝜙 ∶ (𝑈, 𝑝) → (𝐑𝑘 , 0) a 𝐶∞ map. Suppose
𝜙 is a submersion at 𝑝, i.e., suppose its derivative
𝐷𝜙(𝑝) ∶ 𝐑𝑛 → 𝐑𝑘
is onto. Then there exists a neighborhood 𝑉 of 𝑝 in 𝑈, a neighborhood 𝑈0 of the origin in 𝐑𝑛 ,
and a diffeomorphism 𝑔 ∶ (𝑈0 , 0) → (𝑉, 𝑝) such that the map 𝜙 ∘ 𝑔 ∶ (𝑈0 , 0) → (𝐑𝑛 , 0) is the
restriction to 𝑈0 of the canonical submersion:
𝜋 ∶ 𝐑𝑛 → 𝐑 𝑘 , 𝜋(𝑥1 , …, 𝑥𝑛 ) ≔ (𝑥1 , …, 𝑥𝑘 ) .
Proof. Let 𝜙 = (𝑓1 , …, 𝑓𝑘 ). Composing 𝜙 with a translation we can assume 𝑝 = 0 and by a
permutation of the variables 𝑥1 , …, 𝑥𝑛 we can assume that the matrix (b.2) is non-singular.
By Theorem b.1 we conclude that there exists a diffeomorphism 𝑔 ∶ (𝑈0 , 0) ↪ (𝑈, 𝑝) with
the properties 𝑔⋆ 𝑓𝑖 = 𝑥𝑖 for 𝑖 = 1, …, 𝑘, and hence,
𝜙 ∘ 𝑔(𝑥1 , …, 𝑥𝑛 ) = (𝑥1 , …, 𝑥𝑘 ) . □
As a third application of Theorem b.1 we’ll prove a theorem which is similar in spirit to
Theorem b.17, the canonical immersion theorem.
Theorem b.18. Let 𝑈 be an open neighborhood of the origin in 𝐑𝑘 and 𝜙 ∶ (𝑈, 0) → (𝐑𝑛 , 𝑝)
a 𝐶∞ map. Suppose that the derivative of 𝜙 at 0
𝐷𝜙(0) ∶ 𝐑𝑘 → 𝐑𝑛
is injective. Then there exists a neighborhood 𝑈0 of 0 in 𝑈, a neighborhood 𝑉 of 𝑝 in 𝐑𝑛 , and
a diffeomorphism.
𝜓 ∶ 𝑉 → 𝑈0 × 𝐑𝑛−𝑘
such that the map 𝜓 ∘ 𝜙 ∶ 𝑈0 → 𝑈0 × 𝐑𝑛−𝑘 is the restriction to 𝑈0 of the canonical immersion
𝜄 ∶ 𝐑𝑘 → 𝐑𝑘 × 𝐑𝑛−𝑘 , 𝜄(𝑥1 , …, 𝑥𝑘 ) = (𝑥1 , …, 𝑥𝑘 , 0, …, 0) .
Proof. Let 𝜙 = (𝑓1 , …, 𝑓𝑛 ). By a permutation of the 𝑓𝑖 ’s we can arrange that the matrix
𝜕𝑓
[ 𝜕𝑥𝑗𝑖 (0)] , 1 ≤ 𝑖, 𝑗 ≤ 𝑘
is non-singular and by composing 𝜙 with a translation we can arrange that 𝑝 = 0. Hence by
Theorem b.16 the map
𝜒 ∶ (𝑈, 0) → (𝐑𝑘 , 0) 𝑥 ↦ (𝑓1 (𝑥), …, 𝑓𝑘 (𝑥))
maps a neighborhood 𝑈0 of 0 diffeomorphically onto a neighborhood 𝑉0 of 0 in 𝐑𝑘 . Let
𝜓 ∶ (𝑉0 , 0) → (𝑈0 , 0) be its inverse and let
𝛾 ∶ 𝑉0 × 𝐑𝑛−𝑘 → 𝑈0 × 𝐑𝑛−𝑘
be the map
𝛾(𝑥1 , …, 𝑥𝑛 ) ≔ (𝜓(𝑥1 , …, 𝑥𝑘 ), 𝑥𝑘+1 , …, 𝑥𝑛 ) .
Then
(b.19) (𝛾 ∘ 𝜙)(𝑥1 , …, 𝑥𝑘 ) = 𝛾(𝜒(𝑥1 , …, 𝑥𝑘 ), 𝑓𝑘+1 (𝑥), …, 𝑓𝑛 (𝑥))
= (𝑥1 , …, 𝑥𝑘 , 𝑓𝑘+1 (𝑥), …, 𝑓𝑛 (𝑥)) .
Now note that the map
ℎ ∶ 𝑈0 × 𝐑𝑛−𝑘 → 𝑈0 × 𝐑𝑛−𝑘
Draft: March 28, 2018
defined by
ℎ(𝑥1 , …, 𝑥𝑛 ) ≔ (𝑥1 , …, 𝑥𝑘 , 𝑥𝑘+1 − 𝑓𝑘+1 (𝑥), …, 𝑥𝑛 − 𝑓𝑛 (𝑥))
is a diffeomorphism. (Why? What is its inverse?) and, by (b.19),
(ℎ ∘ 𝛾 ∘ 𝜙)(𝑥1 , …, 𝑥𝑘 ) = (𝑥1 , …, 𝑥𝑘 , 0, …, 0) ,
i.e., (ℎ ∘ 𝛾) ∘ 𝜙 = 𝜄. Thus we can take the 𝑉 in Theorem b.18 to be 𝑉0 × 𝐑𝑛−𝑘 and the 𝜓 to be
ℎ ∘ 𝛾. □
Remark b.20. The canonical submersion and immersion theorems can be succinctly sum-
marized as saying that every submersion “looks locally like the canonical submersion” and
every immersion “looks locally like the canonical immersion”.
Draft: March 28, 2018
APPENDIX c
and that the right-hand side is positive because of the positive-definiteness of (c.2).
To prove (c.5) we note that 𝑑𝑓 𝑑𝑡
is strictly increasing by (c.4). Hence 𝑑𝑓
𝑑𝑡
(0) > 0 would
𝑑𝑓
imply that 𝑑𝑡 (𝑡) > 0 for all 𝑡 ∈ [0, 1] and hence would imply that 𝑓 is strictly increasing.
But
𝑓(0) = 𝑓(1) = 𝜙(𝑝) = 𝜙(𝑞) = 𝑐 ,
so this would be a contradiction. We conclude that 𝑑𝑓 𝑑𝑡
(0) < 0. A similar argument shows
𝑑𝑓
that 𝑑𝑡 (1) > 0.
Now we use equations (c.4) and (c.5) to show that 𝑈𝑐 is convex. Since 𝑑𝑓 𝑑𝑡
is strictly
𝑑𝑓 𝑑𝑓
increasing it follows from (c.5) that 𝑑𝑡 (𝑡0 ) = 0 for some 𝑡0 ∈ (0, 1) and that 𝑑𝑡 (𝑡) < 0 for
211
Draft: March 28, 2018
𝑡 ∈ [0, 𝑡0 ) and 𝑑𝑓
𝑑𝑡
(𝑡) > 0 for 𝑡 ∈ (𝑡0 , 1]. Therefore, since 𝑓(0) = 𝑓(1) = 𝑐, we have 𝑓(𝑡) < 𝑐
for all 𝑡 ∈ (0, 1); so 𝜙(𝑥) < 𝑐 on the line segment defined by (1 − 𝑡)𝑝 + 𝑡𝑞 for 𝑡 ∈ (0, 1). Hence
𝑈𝑐 is convex. □
We will now use Theorem c.6 to prove that 𝑋 admits a good cover: for every 𝑝 ∈ 𝑋, let
√𝑐(𝑝)
𝜀(𝑝) ≔
3
and let
𝑈𝑝 ≔ { 𝑞 ∈ 𝑋 | |𝑝 − 𝑞| < 𝜀(𝑝) } .
Theorem c.10. The 𝑈𝑝 ’s are a good cover of 𝑋.
Draft: March 28, 2018
213
Bibliography
1. Vladimir Igorevich Arnol’d, Mathematical methods of classical mechanics, Vol. 60, Springer Science & Business
Media, 2013.
2. Garrett Birkhoff and Gian-Carlo Rota, Ordinary differential equations, 4th ed., John Wiley & Sons, 1989.
3. Raoul Bott and Loring W. Tu, Differential forms in algebraic topology, Vol. 82, Springer Science & Business
Media, 2013.
4. Victor Guillemin and Alan Pollack, Differential topology, Vol. 370, American Mathematical Society, 2010.
5. Victor Guillemin and Shlomo Sternberg, Symplectic techniques in physics, Cambridge University Press, 1990.
6. Nikolai V. Ivanov, A differential forms perspective on the Lax proof of the change of variables formula, The Amer-
ican Mathematical Monthly 112 (2005), no. 9, 799–806.
7. John L. Kelley, General topology, Graduate Texts in Mathematics, vol. 27, Springer, 1975.
8. Peter D. Lax, Change of variables in multiple integrals, The American Mathematical Monthly 106 (1999), no. 6,
497–501.
9. , Change of variables in multiple integrals II, The American Mathematical Monthly 108 (2001), no. 2,
115–119.
10. James R. Munkres, Analysis on manifolds, Westview Press, 1997.
11. , Topology, 2nd ed., Prentice Hall, 2000 (English).
12. Michael Spivak, Calculus on manifolds: a modern approach to classical theorems of advanced calculus, Mathe-
matics Monographs, Westview Press, 1971.
13. Frank W. Warner, Foundations of differentiable manifolds and Lie groups, Graduate Texts in Mathematics,
vol. 94, Springer-Verlag New York, 1983.
215
Draft: March 28, 2018
Draft: March 28, 2018
Index of Notation
A 𝑘 (𝑉), 14 𝛬𝑘 (𝑉⋆ ), 19
Alt, 14 𝐿𝒗 , 34
𝐵𝑘 (𝑋), 149
𝐵𝑐𝑘 (𝑋), 149 𝑁(U), 195
𝑁𝑘 (U), 194
( 𝑛𝑘 ), 16
𝐻𝑘 (𝑋), 149
𝑇𝑝 𝐑𝑛 , 33
𝐻𝑘 (U; 𝐑), 195
𝑇𝑝 (𝑋), 104, 105
𝐻𝑐𝑘 (𝑋), 149
⊗, 9
≃, 56
𝑇𝑌 , 176
I𝑘 (𝑉), 17 𝑇𝜎 , 13
im, 2 𝜏𝑌 , 176
𝜄𝑣 , 23
𝜄𝒗 , 36 𝑊(𝑋, 𝑝), 139
∧, 20
ker, 2
217
Draft: March 28, 2018
Draft: March 28, 2018
Glossary of Terminology
221