Lecturenotes4 3
Lecturenotes4 3
∗
Department of Mathematics, Iowa State University, Ames, Iowa, U.S.A. Electronic address: da-
[email protected]
1
Contents
I Calculus On Manifolds 5
1 Manifolds 6
1.1 Examples and further definitions . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
1.1.1 Manifolds of dimension m=1 . . . . . . . . . . . . . . . . . . . . . . . . . 7
1.1.2 Surfaces . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
1.1.3 n-dimensional Spheres . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
1.1.4 Product manifolds . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
1.1.5 Projective spaces . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
1.1.6 Grassmann Manifolds . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11
1.2 Maps between manifolds . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14
1.3 Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
5 Lie Derivative 44
5.1 Lie derivative of a vector field . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 44
5.2 Lie derivatives of co-vector fields and general tensor fields . . . . . . . . . . . . . 50
5.3 Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52
2
7 Differential Forms Part II: Fields and the Exterior Derivative 61
7.1 Fields . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 61
7.2 The exterior derivative . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 61
7.2.1 Independence of coordinates . . . . . . . . . . . . . . . . . . . . . . . . . . 62
7.3 Properties of the exterior derivative . . . . . . . . . . . . . . . . . . . . . . . . . . 63
7.3.1 Examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 64
7.3.2 Closed and Exact Forms . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65
7.4 Interior product . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66
7.4.1 Properties of the interior product . . . . . . . . . . . . . . . . . . . . . . . 67
7.5 Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 69
11 Exercises 110
3
Other resources:
2. J.M. Lee, Introduction to Smooth Manifolds, 2-nd Edition, Graduate Texts in Mathematics,
218, Springer, 2012.
4. M. Nakahara, Geometry, Topology and Physics (Graduate Student Series in Physics) 2nd
Edition, Taylor and Francis Group, New York, 2003.
4
Part I
Calculus On Manifolds
5
1 Manifolds
I m . More in detail:
A manifold is an object that locally looks like R
Additionally, the maps φj are assumed to satisfy the smooth compatibility condi-
j, k such that Uj Uk 6= ∅, the map φk ◦ φ−1
T
tion, T
which means that,
T for any pair j :
∞
φj (Uj Uk ) → φk (Uj Uk ) is in C in the usual sense of calculus for maps R Im→ R I m.
The maps φk ◦ φ−1j are referred to as transition functions. (cf. Figure 1) Two more
conditions are usually assumed and we will do so as well:
1) M is assumed to be second countable, i.e., there exists a countable base in its topol-
ogy.a
2) M is Hausdorff, that is, every two distinct points in M have disjoint neighborhoods.
a
A countable family of open sets such that every open set can be written as the union of sets in this
family.
6
Figure 1: Definition of Manifold and the Smooth Compatibility Condition
.
7
Figure 2: Manifold structure and transition function for S 1 .
.
1.1.2 Surfaces
I 3 , which we studied
Much of our intuition about manifolds comes from curves and surfaces in R
in multivariable calculus. There,
Definition 1.1.1: Surface
Closed surfaces such as cylinders and spheres can be still described by parametric surfaces
locally. Therefore, they still can be given the structure of a manifold, but the corresponding atlas
contains more than one chart. The situation is similar to the one for the circle S m described
above.
8
1.1.3 n-dimensional Spheres
Definition 1.1.2: S n
with the subset topology induced from R I n+1 . Consider the charts (Uj+ , φj+ ) and
(Uj− , φj− ), for j = 0, 1, . . . , n, where the coordinate neighborhoods Uj+ ’s and Uj− ’s are
defined as
The set of coordinate charts (Uj± , φj± ) is an atlas since the Uj± , j = 0, 1, 2, . . . , n form a
cover for S n (if (x0 , x1 , . . . , xn ) does not belong to any of the Uj± , then all the components,
x0 , x1 , . . . , xn , must be zero which contradicts (1.1.1)). Assume j < k. On Uj+ ∩ Uk+ , both
xj and xk are strictly positive. φ−1 0 1
j+ maps (x , x , . . . , x I n to the point
j−1 , xj+1 , . . . , xn ) ∈ R
q
(x0 , x1 , . . . , xj−1 , 1 − l6=j (xl )2 , xj+1 , . . . , xn ) ∈ S n which is mapped to
P
s X
0 1 j−1
(x , x , . . . , x , 1− (xl )2 , xj+1 , . . . , xk−1 , xk+1 . . . , xn ) ∈ R
In
l6=j
9
1.1.4 Product manifolds
Definition 1.1.3: Product manifold
Suppose M and N are manifolds. Then M × N has a natural manifold structure under
the product topology and charts (Um × Un , φm × φn ) with (Um , φm ) and (Un , φn ) the
respective charts of M and N . We define this to be the product maifold.
It is a known fact from topology that the product of second countable Hausdorff spaces is a
second countable Hausdorff space in the product topology. Then under the product topology we
now naturally consider (Um × Un , φm × φn ). This function is still a homeomorphism to its image.
Finally, for overlapping charts, (φmi × φni ) ◦ ((φmj × φnj )−1 ) is still smooth. This is verified by
showing the product of continuous functions is continuous, the product of open maps is an open
map, and the product of differentiable functions is differentiable.
The real projective space RP n is the quotient space ( R I n+1 − {0})/ ∼, where ∼ is the
I n+1 .
equivalence relation which associates P with λP , for arbitrary λ 6= 0 and P ∈ R
10
Figure 3: Hausdorff property in RP 1 . U0 (U1 ) is the set of lines whose points have nonzero
components along the x0 (x1 ) axis. All pairs of lines such as ([B], [A1 ]) or ([B], [A2 ]) are such
that both elements are in U0 or U1 . The only exception in ([A1 ], [A2 ]). To handle this case we
introduce a new Uθ by rotating the x0 axis by an angle θ and obtaining the axis represented with
thicker pointed line. The elements in Uθ are all the lines whose point have non zero component
along the θ-rotated axis. This way, both [A1 ] and [A2 ] are in Uθ .
A = gB.
The space of equivalence classes under this equivalence relation is called the Grassman-
nian, or Grassmann Manifold, Grk,n ( R I ), that is Grk,n ( R
I ) := Stk,n ( RI )/ ∼. This
can be thought of as the space of k-dimensional subspaces of R I n , each subspace being
11
spanned by the rows of a representative in the given equivalence class. Since Stk,n ( R I)
has a given (natural) topology, Grk,n ( R I ) is equipped with the quotient topology, that is,
I ) → Grk,n ( R
if π : Stk,n ( R I ) is the natural projection mapping an element A ∈ Stk,n ( R I)
to its equivalence class [A] in Grk,n ( R I ) iff π −1 (U ) is open in
I ), then U is open in Grk,n ( R
Stk,n ( R
I ).
Then φl ([A]) := Q−1 N ∈ Mk,n−k ( R I)' R I k(n−k) . This map is well defined since its definition
does not depend on the representative A of the equivalence class [A]. In fact, if P is a k × k
nonsingular matrix, φl ([P A]) = Q−1 P −1 P N = Q−1 N . The map φl is a homeomorphism between
Ul and Rk(n−k) . To show that {Ul , φl } give a differentiable structure on Grk,n ( R
I ) we need to show
−1 k(n−k) k(n−k) n
that the transition functions φl ◦φj : R I → R I are smooth for every l, j = 1, . . . , .
k
Assume for simplicity of notation j = 1. We have
φ−1 φ
I k(n−k) −−
W ∈ R 1 l
→ [(1 W )] −→ Q−1 N,
where Q is the l-th submatrix of the matrix [1 W ] and N is the remaining k × (n − k) matrix.
This is clearly a C ∞ map R
I k(n−k) → RI k(n−k) .
We now turn to the proof that Grk,n ( R I ) is Hausdorff. Given two points [A1 ] and [A2 ] in
Grk,n ( RI ) we need to exhibit two disjoint neighborhoods of [A1 ] and [A2 ]. If [A1 ] and [A2 ] are
both in the same Ul , for some l, this is easily obtained since the coordinate map φl gives a
homeomorphism that establishes a one to one and onto correspondence between Ul and R I k(n−k) .
k(n−k)
Since R I is Hausdorff we can choose two disjoint neighborhoods N1 and N2 of φl ([A1 ]) and
φl ([A2 ]), respectively and φ−1 −1
l (N1 ) and φl (N2 ) will give two disjoint neighborhoods of [A1 ] and
[A2 ] in Grk,n ( R
I ). This is the generic case since the set of pairs ([A1 ], [A2 ]) in Grk,n ( R
I )×Grk,n ( R
I)
that do not share any coordinate neighborhood Ul is a closed set. (it corresponds to the closed
12
I ) given by the zeros of the function l (det(D1,l ))2 (det(D2,l )))2 , where
P
I ) × Stk,n ( R
set in Stk,n ( R
D1,l (D2,l ) is the l−th submatrix f the k × n matrix in the first (second) factor of the pair
(A, B) ∈ Stk,n ( R I ) × Stk,n ( R
I )) In order to extend this proof in general, we define a richer set of
coordinate charts.
Consider an action of Gl(n, R I ) on1 Grk,n ( RI ), that is, for every g ∈ Gl(n, R I ) a map σg :
I ) → Grk,n ( R
Grk,n ( R I ), defined by σg ([A]) = [Ag]. It is easily seen that σg is well defined and
continuous map. Moreover Gl(n, R I ) acts transitively2 on Grk,n ( R I ), that is, for every pair [A]
and [B] there exists a g ∈ Gl(n, R I ) such that σg ([A]) = [B].3
Consider now the chart {U1 , φ1 } defined above. From this we can define a new neighborhood
σg (U1 ) and a new coordinate function φg := φ1 ◦ σg−1 on σg (U1 ) for any g ∈ GL(n, R). In
particular the above described coordinate neighborhoods Ul and coordinate maps φl can be
obtained as a special case of this construction by choosing for g a matrix which moves the first
k columns to the positions identified by the l−th permutation. In general, we can denote by Ug
and φg the neighborhood and coordinate map, respectively, corresponding to a certain g
(U1 and φ1 correspond to the identity matrix). If [A1 ] and [A2 ] belong to the same Ug , we
can repeat the argument we have done for Ul and show that they have disjoint neighborhoods.
Therefore the result is proven if we show that there exists such a g. Summarizing, we need to prove
that for any pair [A1 ], [A2 ] in Grk,n ( R I ), there exists a g ∈ Gl(n, RI ) such that [A1 ] ∈ Ug = σg (U1 )
and [A2 ] ∈ Ug = σg (U1 ).
We first give an alternative description of Ug , for a given g (Step 1 below) and then (Step 2
below) we use this characterization to show that there exists a g so that [A1 ] and [A2 ] are both
in Ug .
Step 1: Define the (n − k) × n matrix W0 := [0n−k,k 1n−k,n−k ], where 1n−k,n−k is the
(n − k) × (n − k) identity. Given g ∈ Gl(n, R
I)
or, equivalently,
vkT Ag −1 = wn−k
T
W0 .
the first k entries of the right hand side are equal to zero. Moreover since [Ag −1 ] ∈ U1 ,
Ag −1 = [K M ] with the k × k matrix K nonsingular. This implies, since vkT K = 0, that
1
The notation Gl(n, R I ) stands for the ‘linear group’ of n × n nonsingular matrices with entries in RI.
2
We shall elaborate more on ‘actions’ and ‘transitivity’ when we will talk about Lie transformation groups
later in the course.
3 A
To see this, take A and complete it with an (n − k) × n matrix à so that the n × n matrix is nonsingular;
Ã
B
Analogously, take B and complete it with an (n − k) × n matrix B̃ so that the n × n matrix is nonsingular.
B̃
−1
A B
The matrix g = is such that σg ([A]) = [B].
à B̃
13
vk = 0, which is a contradiction.
To show the converse implication (⇐ in (1.1.2)), assume span(A) ∩ span(W0 g) = {0} or, equiv-
alently, span(Ag −1 ) ∩ span(W0 ) = {0}, the matrix Ag −1 has the form Ag −1 = [K M ]. The
submatrix matrix K is not singular. If it was singular, there would be a nonzero k−vector vk ,
such that vkT K = 0 which would still give vkT Ag −1 6= 0 since Ag −1 has rank k. The (row) vector
vkT Ag −1 6= 0 would therefore be of the form [0 0 . . . 0 a1 , a2 , . . . an−k ] and therefore in span(W0 )
and different from zero which contradicts the right hand side of (1.1.2). Since Ag −1 = [K M ]
with K (k × k) nonsingular, [Ag −1 ] ∈ U1 .
Step 2 The key observation is the separation property, that is, given two k-dimensional
subspaces V1 and V2 of R I n there exists an (n − k)−dimensional subspace W , such that V1 ∩ W =
V2 ∩ W = {0} (Geometrically this is clear by doing examples in R I 2 or R
I 3 ). We construct such
a subspace as the span of vectors not included in V1 ∪ V2 . If V1 = V2 , then we clearly can take a
vector w 6= 0 with w ∈ / V1 and w ∈ / V2 . We can do this even if V1 and V2 are different. In fact,
if V1 and V2 are different spaces with the same dimension, there exist nonzero v2 ∈ V2 , v2 ∈ / V1 ,
and nonzero v1 ∈ V1 , v1 ∈ / V2 . The vector v := av1 + bv2 with a and b different from zero is
not in V1 ∪ V2 . Consider now the spaces V1 ⊕ span(v) and V2 ⊕ span(v). To these two spaces
we can re-apply the same argument and find one vector which lies outside the union of the two
spaces. Continuing this way, we find n − k vectors w1 = v, w2 ,...,wn−k . These vector are linearly
independent (because at each step the new vector is not a linear combination of the previous
ones), and therefore form a a basis of an (n − k)-dimensional vector space W whose intersection
with V1 and V2 , by construction, is {0}. Given this fact, we are now ready to conclude the proof.
Given [A1 ] and [A2 ] apply the above fact with V1 = span(A1 ) and V2 = span(A2 ) and find
the corresponding W , then select g so that W = span(W0 g). This is always possible since we
know that Gl(n, R I ) is transitive on Grn−k,n ( RI ). Since span(A1 ) ∩ W = span(A2 ) ∩ W = {0}
with W = span(W0 g), condition (1.1.2) applies for both A1 and A2 . Therefore both [A1 ] and
[A2 ] are in Ug and this completes the proof. The geometric idea of the proof is illustrated in
Figure 3
The proof of the Hausdorff property follows mainly the argument of Prof. Schlichtkrull
(University of Copenhagen) http://www.math.ku.dk/∼schlicht/Geom2/Grassmann.pdf, which
was somehow streamlined and adapted to our notations.
Remark 1.1.1: Notation
We shall use the Einstein summation convention in that upper and lowerPcorrespond-
ing indexes in an expression indicate a sum so that for instance X i ωi := i X i ωi , and
X i1 ,i2 ,...,in ,k ωi1 ,i2 ,...,in := i1 ,i2 ,...,in X i1 ,i2 ,...,in ,k ωi1 ,i2 ,...,in , which depends on k.
P
Given a chart (U, φ) for Mm and a chart (V, ψ)) for Nn such that f (U ) ⊆ V 6= ∅, it is
possible to define the map ψ ◦ f ◦ φ−1 : R
Im → R I n which is the (local) coordinate
14
representation of f . The map f is called smooth at p ∈ M if ψ ◦ f ◦ φ−1 is smooth
as a function RIm → R I n at φ(p) ∈ R I m . Such a property is independent of the system
of coordinates chosen on M or on N . For instance, consider two charts (U1 , φ1 ), (U2 , φ2 )
with p ∈ U1 ∩ U2 . Then if ψ ◦ f ◦ φ−1
1 is smooth, so is
ψ ◦ f ◦ φ−1 −1 −1
2 = ψ ◦ f ◦ φ1 ◦ (φ1 ◦ φ2 ),
from the compatibility condition. The same argument holds for coordinate charts (V1 , ψ1 ),
(V2 , ψ2 ), on N , such that f (p) ∈ V1 ∩ V2 .
15
a coordinate map. Open curves and functions on a manifolds form the basic building blocks of
many of the constructions in differential geometry. The space of smooth functions on a manifold
is denoted by F(M ).
16
1.3 Exercises
Exercise 1.1: Consider the self intersecting curve of part a) of Figure 5 and the open square
I 2.
of part b) of the same figure, with the subset topology derived by the standard topology on R
∞
Show that a) cannot be given the structure of a C manifold (of dimension 1) and b) can be
given the structure of a C ∞ manifold.
Exercise 1.3: Derive the atlas, i.e., the differentiable structure for R I P n as a special case
of what described for Grk,n , that is, describe the coordinate neighborhood U0 , . . . , Un the corre-
sponding coordinate maps φ0 , φ1 , . . . , φn , and the transition functions φk ◦ φ−1
j , and show that
the transition functions are in C ∞ .
I P 2.
Exercise 1.4: Reconsider the differentiable structure of Grassmann manifold for Gr1,3 = R
Consider g1 and g2 in Gl(3, R
I ),
1 1 0 1 1 1
g1 := −1 1 0 , g2 := 0 1 0 .
0 0 1 0 0 1
Describe Ug1 and Ug2 and calculate the transition function φg1 ◦ φ−1 I2→ R
g2 : R I 2.
Exercise 1.6: Consider the map S 1 → R I which associates p ∈ S with 1 if p is to the right of
the y-axis, f (p) = −1 if p is to the left of the y axis and f (p) = 0 if p is on the y axis. Write the
coordinate representations of this map ( R I → R I ) with respect to the atlas defined in subsection
1.1. Use these coordinate representations to show that this map is not smooth.
17
2 Tangent and cotangent spaces
2.1 Tangent vector and tangent spaces
I 3,
In multivariable calculus we have learned the definition of a tangent vector to a curve c in R
x = x(t), y = y(t), z = z(t) at a point c(0) := (x(0), y(0), z(0)). This is given by
If c is a curve with image belonging to a surface M , we can think of the tangent vector ~v as a
tangent vector to the surface (manifold) M (cf. Figure 6)
I3 → R
If f = f (x, y, z) is a function f : R I the tangent vector ~v determines a directional
derivative for f at the point p = c(0), given by
d
∇f |p · ~v = fx (p)ẋ(0) + fy (p)ẏ(0) + fz (p)ż(0) = f (x(t), y(t), z(t))|t=0 . (2.1.1)
dt
We remark that to be defined, the right hand side of (2.1.1) does not require that f is defined
on all of RI 3 but only that f is defined (and smooth) on M , in fact in a neighborhood of p ∈ M .
Generalizing this picture to a general manifold M , a curve c : (−a, a) → M determines a
tangent vector at p = c(0), in that, for every function f : M → R I , it gives a value in RI , given by
d d
dt f ◦c(t)|t=0 . Notice that there may be several curves which give the same value for dt f ◦c(t)|t=0 .
Therefore,
Definition 2.1.1: Tangent vector
18
1. c1 (0) = c2 (0) = p.
d d
2. dt (f ◦ c1 (t))|t=0 = dt (f ◦ c2 (t))|t=0 , for every f ∈ F(M ).
Now consider a coordinate chart (U, φ) with p ∈ U , and calculate for a curve c
d d ∂f dxµ
(f ◦ c(t))|t=0 = (f ◦ φ−1 ◦ φ ◦ c(t))|t=0 = |φ(p) |t=0 . (2.1.2)
dt dt ∂xµ dt
Here we use (for the first time) the Einstein summation convention. See that xµ is the µ
d
component of the coordinate map φ := (x1 , x2 , . . . , xm ). Call X µ := dt (xµ ◦ c(t))|t=0 , which
alone in (2.1.2) contains the information about the equivalence class of c. Then we can look at
a tangent vector as a linear operator acting on F(M ) written as
∂
X = Xµ , (2.1.3)
∂xµ
∂
where ∂xµ is the operator F(M ) → R
I,
∂ ∂
µ
f := (f ◦ φ−1 )|φ(p) . (2.1.4)
∂x ∂xµ
Notice there is some ambiguity in the notation ∂x∂ µ used in (2.1.4). This is a usual issue in
differential geometry and we shall keep the ambiguity as the meaning of notation should be clear
from the context. Notice ∂x∂ µ on the left hand side denotes a linear operator F(M ) → R I . Here
the reference to the point p is (sometimes) omitted. This operator is defined as in the right hand
side where now ∂x∂ µ is the standard partial derivative of multivariable calculus.
Definition 2.1.2: Tangent Space
From formula (2.1.3), all tangent vectors can be written as linear combinations of the
operators ∂x∂ µ , with µ = 1, 2, . . . , m, which shows that the space of tangent vectors at
p ∈ M is a vector space. It is called the tangent space at p and denoted by Tp M .
The basis of the vector space { ∂x∂ µ }|p at p, of course depends on the coordinate map φ in the
coordinate chart (U, φ) considered. If two overlapping charts are given, (U, φ), (V, ψ), such that
p ∈ U ∩ V and denote by xµ (y ν ) the coordinates associated with φ (ψ), then the same tangent
vector X can be expanded in terms of { ∂x∂ µ } or in terms of { ∂y∂ ν }, that is,
∂ ∂
X = Xµ = Yν ν. (2.1.5)
∂xµ ∂y
To understand the relation between the components X µ and Y ν , we apply X to the coordinate
function xk which gives the k−th component of φ. We have, using (2.1.5)
∂ ∂
X(xk ) = X µ µ
(xk ) = Y ν ν (xk ). (2.1.6)
∂x ∂y
19
∂ k
Since ∂xµ (x ) = δk,µ , where δk,µ is the Kronecker delta, we have
X(xk ) = X µ δk,µ = X k ,
and, using this in (2.1.6), we obtain the relation between the components
∂ k
Xk = Y µ x . (2.1.7)
∂y µ
Recall that ∂y∂µ xk is the partial derivative (in the usual sense of multivariable calculus)
with respect to y µ of the k−th component of the function φ ◦ ψ −1 calculated at ψ(p) (cf.
Figure 1). The matrix ∂y∂µ xk |ψ(p) is called the Jacobian associated with the given change
of coordinates at p.
In the definition (2.1.8), Vp is assumed to give the same value on each function f which
belongs to the same germ at p. A germ is a subset of F(M ) of functions which are equal in a
neighborhood of p. Belonging to the same germ is an equivalence relation ‘ ∼0 and Vp can be
seen as acting on equivalence classes of functions that have the same value in a neighborhood of
p ∈ M . A tangent vector Vp is a derivation defined on elements of the same germ, that is, it can
be seen as acting on F(M )/ ∼.
The definition in terms of derivation is equivalent to the one in terms of equivalence classes
of curves, that is, there exists a one to one and onto correspondence between equivalence classes
of curves and derivations defined on F(M )/ ∼. The proof (not difficult) can be found in Spivak
pg. 79, ff. The lengthier part is where one wants to show that given a derivation Vp , it is possible
to find the corresponding equivalence class of curves. The crucial step is to show that every
derivation Vp can be written as
∂
Vp = X j j |p , (2.1.9)
∂x
for some coefficients X j (which are necessarily given by X j = Vp (xj )). Then one chooses a
d
(equivalence class of) curve c such that dt [xj ◦ c(t)]|t=0 = X j , and this will give the curve
corresponding to the derivation Vp .4
4
Just choose a curve in RI m whose tangent vector has components X j and then map it back to the manifold
M via the inverse coordinate transformation φ−1 .
20
2.2 Co-tangent vectors and co-tangent space
Definition 2.2.1: Contangent Vector
Like any vector space Tp M has a dual space defined as the space of linear maps ω : Tp M →
I . An element ω ∈ Tp∗ M is called a cotangent vector or a one form.
R
Let dxj denote the differential of the function xj as defined above. If we calculate, using the
definition of a differential of a function,
∂xj
∂
dx j
:= = δkj , (2.2.3)
∂xk ∂xk
we find that {dx1 , . . . , dxm } is the dual basis of the basis of Tp M in (2.2.2). Every element
ω ∈ Tp∗ M can be written as
ω := ωµ dxµ , (2.2.4)
and, in general, if ω can be written as in (2.2.4) and V ∈ Tp M is given by V = V j ∂x∂ j , we have
j ∂ ∂
µ
ω(V ) = ωµ dx V j
= ωµ V dx µ
= ωµ V j δjµ = ωµ V µ , (2.2.5)
∂xj ∂xj
which can be interpreted as a product between elements in Tp∗ M and elements in Tp M . If there
are two overlapping coordinate systems x and y, a one form can be written as ω = ωi dxi or
ω = ω̃j dy j . To see the relation between the ωi and the ω̃j , fix an index k and apply ω to ∂y∂ k .
We obtain
∂xi
∂ i ∂
ω = ωi dx = ω i = (2.2.6)
∂y k ∂y k ∂y k
∂y j
∂
ω̃j dy j
= ω̃ j = ω̃j δkj = ω̃k .
∂y k ∂y k
21
Therefore, the Jacobian of the transition map gives the desired transformation between the
expression of a one form in one coordinate system and the other, that is,
∂xi
ω̃k = ωi . (2.2.7)
∂y k
In order to remember formula (2.2.7) one may start from the equality ω̃k dy k = ωi dxi and
divide ‘formally’ by dy k , and then change the d in ∂. Notice also the Jacobian of the
∂xi
transition function φ ◦ ψ −1 , J = ∂y k in (2.2.7) is multiplied as follows: If one thinks of
ω̃ and ω as rows, equation (2.2.7) reads as ω̃ = ωJ. Analogously, formula (2.1.7) can
be remembered by using X k ∂x∂ k = Y µ ∂y∂µ , neglecting the ∂ signs on top and ‘formally’
∂x k
multiplying by ∂xk . If we think of X and Y as vectors and J := ∂y µ as a matrix, the
relation writes as X = JY ; consistently with this we get that ω̃Y = ωJ(J −1 X) = ωX,
independently of the coordinates.
Notice in order not to confuse this concept with the concept of differential of a function
defined when discussing one forms, the term ’push-forward’ is used instead of ‘differential’. This
term is justified by the fact that the map f∗ maps elements in Tp M ‘forward’ to elements in
Tf (p) N .
The definition (2.3.1) is given in terms of the ‘derivation definition’ of a tangent vector and
it is straightforward to verify that if V is a derivation at p ∈ M then f∗ V defined in (2.3.1) is a
derivation at f (p) ∈ N . Moreover, if we use the definition in terms of equivalence class of curves
d
and c is a curve corresponding to V , that is, for h ∈ F(M ), V (h) = dt (h ◦ c)|t=0 , then we have,
for g ∈ F(N )
d
f∗ V (g) := V (g ◦ f ) = (g ◦ f ◦ c)|t=0 . (2.3.2)
dt
Therefore, f ◦ c is the curve associated with f∗ V (cf. Figure (7)).
22
Remark 2.3.1: Linearity of Push-forward
From our real vector space Tp M , we also see that the pushforward of a smooth map f is
linear. Take V1 , V2 in Tp M and real numbers a, b. Remembering that these vectors are
derivations, which are linear operators, we have f∗ (aV1 + bV2 )(g) = (aV1 + bV2 )(g(f )) =
(aV1 )(g(f )) + (bV2 )(g(f )) = aV1 (g(f )) + bV2 (g(f )) = af∗ (V1 ) + bf∗ (V2 ).
We now investigate how to express the push-forward in terms of coordinates. Let x denote
23
a coordinate system at p ∈ M and y a coordinate system at f (p) ∈ N . Then we can write
V = V µ ∂x∂ µ and f∗ V = Y ν ∂y∂ ν . To discover the relation between the Y ν coefficients and the V µ
coefficients, we recall that Y k = f∗ V (y k ). Therefore, using the definition of f∗ V , we get
∂(y k ◦ f )
Y k = (f∗ V )(y k ) = V (y k ◦ f ) = V µ . (2.3.3)
∂xµ
Recall that ∂
∂xµ (g) is really the partial derivative with respect to xµ of the function g ◦ φ−1 :=
k
In → R
R I . Therefore J := ∂(y∂xµ◦f ) is really the {ν, µ} entry of the Jacobian associated with the
map ψ ◦ f ◦ φ−1 : R Im → R I n (cf. Figure 4). Also notice that thinking of Y and V as vectors,
they (2.3.3) can be related as follows
Y = JV . (2.3.4)
1. Express this tangent vector in terms of the coordinates associated with the chart (U0 , φ0+ ).
2. Express such a tangent vector in terms of the spherical coordinates, (θ, φ) ∈ (−π, π)×(0, π),
which are (recall)
I 3 at (1, 0, 0).
in R
As for 1) we use the coordinates (x1 , x2 ) and the corresponding basis in the tangent space
{ ∂x∂ 1 , ∂x∂ 2 },
so that the tangent vector can be written as
∂ ∂
Vp = X 1 1
+ X2 2 .
∂x ∂x
24
Recall the X 1 (X 2 ) is the result of applying the tangent vector to the coordinate function. If we
do that for x1 , we get
d d p
X 1 = Vp x1 := |t=0 (x1 ◦ c(t)) = |t=0 ( 1 − t2 sin(10πt)) = 10π.
dt dt
For the same reason, we get X 2 = 1, so that
∂ ∂
Vp = 10π 1
+ 2. (2.3.6)
∂x ∂x
∂ ∂
To do 2) now we want to express Vp in terms of { ∂θ , ∂φ }. We know that this change of coordinates
involves the Jacobian as in (2.1.7). However, instead of remembering and applying the formula,
per se, it is more common among differential geometers to re-act the calculation the brought to
the formula. The transition functions (θ, φ) → (x1 , x2 ) are the last two of (2.3.5). Calculating
∂ 1 ∂
X 1 = Vp x1 (θ, φ) = Y 1 x (θ, φ) + Y 2 x1 (θ, φ),
∂θ ∂θ
we get:
10π = Y 1 (sin(φ) cos(θ))|φ= π2 ,θ=0 + Y 2 (cos(φ) sin(θ))|φ= π2 ,θ=0 = Y 1 .
Analogously, calculating
∂ 2 ∂
X 2 = Vp x2 (θ, φ) = Y 1 x (θ, φ) + Y 2 x2 (θ, φ),
∂φ ∂φ
we get
1 = Y 1 (0) + Y 2 (−sin(φ))φ= π2 ,θ=0 = −Y 2 .
So we have
∂ ∂
Vp = 10π − .
∂θ ∂φ
∂ ∂ ∂
To do 3) we write i∗ Vp as i∗ Vp = X̃ 1 ∂x + X̃ 2 ∂y + X̃ 3 ∂z (we have switched to notation (x, y, z)
I 3 and coordinates
instead of (x0 , x1 , x2 ) to emphasize the distinction between coordinates on R
for S 2 ). We have
d d p
X̃1 = (i∗ Vp )(x) := Vp (x ◦ i) := |t=0 (x ◦ i ◦ c(t)) = |t=0 ( 1 − t2 cos(10πt)) = 0.
dt dt
and analogously we get X̃2 = 10π and X̃3 = −1, so that
∂ ∂
i∗ Vp = 10π − .
∂y ∂z
Finally for 4) we can calculate f∗ Vp as described in part 4) with coordinates x and y. By writing
∂ ∂
f∗ Vp := Ỹ1 ∂x + Ỹ2 ∂y , we know that Ỹ1 = f∗ Vp (x) = Vp (x ◦ f ), Ỹ2 = f∗ Vp (y) = Vp (y ◦ f ).
Moreover, in the coordinates (x, y), f ◦ c(t) is
0 √
x (t) − x1 (t)
1 −√t2 (cos(10πt) − sin(10πt))
f ◦ c(t) := = .
x1 (t) − x2 (t) 1 − t2 sin(10πt) + t
25
Therefore,
d d p
Vp (x ◦ f ) := |t=0 (x ◦ f ◦ c(t)) = |t=0 ( 1 − t2 (cos(10πt) − sin(10πt)) = −10π.
dt dt
Analogously,
d p
Ỹ2 = f∗ Vp (y) = Vp (y ◦ f ) = |t=0 1 − t2 sin(10πt) + t = 10π + 1.
dt
So we get
∂ ∂
f∗ Vp = −10π + (1 + 10π) . (2.3.7)
∂x ∂y
I 2 (which is the
We could have also used formula (2.3.4). Here, with ψ the coordinate map on R
identity map) and φ0+ the coordinate map on S 2 , we have
1 − (x1 )2 − (x2 )2 − x1
p
−1
ψ◦f ◦φ = ,
x1 − x2
which multiplied by the components of Vp , i.e., (10π, −1)T gives the components in (2.3.7), that
is, (−10π, 10π + 1)T .
for any V ∈ Tp M .a
a
In general let V and V ∗ a vector space and its dual, respectively, and W and W ∗ another vector
space and its dual, let A be a linear map A := V → W , then the dual map A∗ is the defined as the map
A∗ : W ∗ → V ∗ such that for every w∗ ∈ W ∗ and v ∈ V , w∗ (Av) = (A∗ w∗ )(v). Definition (2.4.1) is a
special case with A = f∗ , v = V , f ∗ = A∗ , ω = w∗ and the vector spaces identified accordingly.
26
To see how f ∗ transform ω in local coordinates, let us choose coordinates x in a neighborhood
of p ∈ M and coordinates y in a neighborhood of f (p) ∈ N . Write ω := ωi dy i and f ∗ ω := λj dxj .
We know that, for fixed k, λk is the result of applying f ∗ ω to ∂x∂ k . Therefore we have
∂ ∂ ∂
λk = f ∗ ω( k
) = ωi dyi (f∗ k ) = ωi k (y i ◦ f ), (2.4.2)
∂x ∂x ∂x
which gives the transformation rule among the coefficients in terms of the Jacobian of f . Once
again, it is convenient to think of ω and λ as rows and given the Jacobian J := ∂x∂ k (y i ◦ f ) on
the right hand side of (2.4.2), the formula simply says λ = ωJ.
Remark 2.4.1: Linearity of Pull-back
Linearity of the pullback similarly holds using the fact that our contangent vectors are
functionals. We will use this fact in the next example.
Therefore, f ∗ ω = −dx2 .
To write f ∗ ω in the spherical coordinates (θ, φ) we use f ∗ ω = −dx2 and (2.3.5) which gives
27
so that
f ∗ ω = dφ,
where in the last one we used φ = π2 . Finally for 3), from the expression of Vp in (2.3.6), we get
∗ 2 ∂ ∂
(f ω)(Vp ) = −dx 10π 1 + 2 = −1.
∂x ∂x
Since we have already discussed linearity of the pullback, showing it is a bijection is sufficent
for the forward direction specifically, f∗−1 = (f −1 )∗ . For the reverse direction, we consider
f = ψf φ−1 with coordinate charts of p and f (p) respectively. Now we use the inverse function
theorem from calculus on f to get the desired result. For details see ?? page 79
Definition 2.5.1: Immersions
Let f be a smooth map f : M → N , with dim M ≤ dim N . f is an immersion if, at every
point p ∈ M , f∗ is an injection, that is rank(f∗ ) = dim(M ). Consider now an immersion
f which is also injective, the image f (M ) can be made a manifold by ‘borrowing’ the
topological and differentiable structure of M , that is open sets in f (M ) are the images of
open sets in M , and charts on f (M ) can be defined using the charts of M . The manifold
f (M ) is diffeomorphic to M and it is called an immersed submanifold of N .
If the injective immersion f has the additional property that the images of coordinate
neighborhoods in M f (V ) are equal to f (M )∩U for some U coordinate neighborhood of N
(subset topology), then f is called an embedding and f (M ) an embedded submanifold
or simply a submanifold.
Often f is taken as the inclusion map so that for example S 1 is a submanifold of R2 , etc.
Figure 8 shows examples of immersed and embedded submanifolds.
28
Figure 8: The image of an immersion a); An immersed submanifold b); An embedded submanifold
c)
29
2.6 Exercises
Exercise 2.1 Show that two diffeomorphic manifolds have the same dimension.
Exercise 2.2 The tangent space at p, Tp M is a subspace of the space of linear operators on
F(M ). Show that it has dimension m = dim(M ) and ∂x∂ µ , µ = 1, 2, . . . , m, form a basis for
Tp M .
Exercise 2.3 Consider the sphere S 2 as defined in subsection 1.1, with the given atlas, and
the point p ≡ ( √13 , √13 , √13 ) ∈ S 2 . Find two overlapping coordinate charts (U, φ), (V, ψ), with
p ∈ U ∩ V , and the Jacobian giving the change of coordinates from the φ coordinates to the ψ
coordinates.
30
3 Tensors and Tensor Fields
3.1 Tensors
Definition 3.1.1: Tensor
Given vector spaces over R I , V1 , . . . , Vn , we consider the vector space generated by elements
in V1 × V2 × · · · × Vn , and denote it by V1 ⊗ V2 ⊗ · · · ⊗ Vn . A tensor T on V1 ⊗ V2 ⊗ · · · ⊗ Vn
is a multilinear map (linear is each variable separately) T : V1 ⊗ V2 ⊗ · · · ⊗ Vn → R I . If
T1 is a tensor on V1 and T2 is a tensor on V2 , we can define the product tensor, T1 ⊗ T2 ,
as (for v1 ∈ V1 , v2 ∈ V2 ) a map that is multilinear and satisfies
This construction inductively extends to tensor products of more than two spaces.
Proposition 3.1.1: Basis of Linear Operators
(See [3] for proof) For j = 1, . . . , n, let {Tj,kj }, kj = 1, . . . , nj , a basis in the space (of
dimension nj = dim (Vj )) of linear operators on Vj . Then a basis in the space of linear
operators on V1 ⊗ V2 ⊗ · · · ⊗ Vn is given by {T1,k1 ⊗ T2,k2 ⊗ · · · ⊗ Tn,kn }, with kj = 1, . . . , nj .
The space of tensors on V1 ⊗ V2 ⊗ · · · ⊗ Vn has therefore dimension n1 n2 · · · nn .
The vector spaces we are mostly interested in are Tp M and Tp∗ M and we are interested in
tensors of the type (q, r) defined as tensors on ⊗q Tp∗ M ⊗r Tp M . We can use a basis of Tp M
as a basis of (Tp∗ M )∗ (they are naturally isomorphic), by identifying a vector V ∈ Tp M with
its action on Tp∗ M , i.e., for ω ∈ Tp∗ M , V (ω) := ω(V ). If we do that, a basis of Tp M gives us a
basis of (Tp∗ M )∗ . Using this identification, our privileged bases in Tp M and Tp∗ M (for a given
coordinate system) and Proposition 3.3.1, every (q, r) tensor, T , can be written as
µ ···µ ∂ ∂
T = Tν11···νrq ⊗ · · · ⊗ µq ⊗ dxν1 ⊗ · · · ⊗ dxνr .
∂xµ1 ∂x
k
The action on vectors (ω1 , . . . , ωq , X1 , . . . , Xr ) defined as ωj = ωjlj dxlj and Xj := Xj j ∂kj , gives
∂x
(See Homework)
µ ···µ
T (ω1 , . . . , ωq , X1 , . . . , Xr ) = Tν11···νrq ω1µ1 · · · ωqµq X1ν1 . . . Xrνr . (3.1.2)
q
The space of (q, r) tensors at p ∈ M is denoted by Tr,p (M ) and it has dimension mq+r
(where m is the dimension of M ). In particular elements of T0,p1 (M ) are tangent vectors and
1 (M ) ' T M , while elements of T 0 (M ) are one forms and T 0 (M ) ' T ∗ M .
T0,p p 1,p 1,p p
31
q q
defined as F∗ : T0,p (M ) → T0,f (p) (N ) and it transforms as
µ1 ···µq ∂ ∂ ∂ ∂
F∗ T := F∗ T µ
⊗ · · · ⊗ µq := T µ1 ···µq (f∗ µ1 ) ⊗ · · · ⊗ (f∗ µq ). (3.1.3)
∂x 1 ∂x ∂x ∂x
q
See F∗ T is an element of T0,f (p) (N ).
The word ‘smooth’ here means that the value of Vp f , depends smoothly on p for every smooth
function f ∈ F(M ). If X is a vector field, its value at p, X(p) is a tangent vector and X(p)f is
a smooth function of p for every f ∈ F(M ).
Definition 3.2.2: Lie Derivative of a Function
A vector field X defines a map F(M ) → F(M ), which is also denoted by X and it is
defined as
(Xf )(p) = X(p)f. (3.2.1)
The value Xf is also called the Lie derivative of the function f along X at p.
Consider now a coordinate chart given. The simplest example of a vector field (described
locally in terms of the local coordinates) is the association to every p ∈ M of ∂x∂ µ |p , for a certain
fixed µ. We will denote such vector field by ∂x∂ µ (we have used this symbol before to indicate
a tangent vector at p omitting the |p symbol, and also for denoting the usual partial derivative
with respect to xµ , the context will resolve ambiguity among these three different meanings of
this notation).
32
Proposition 3.2.1
This vector field is smooth in that it transforms smooth functions into smooth functions.
Proof: The proof is a simple but useful exercise to recall some of the definitions we have
given. If f is a smooth function, then the function g := ∂x∂ µ (f ) is defined as g(p) =
∂ −1 ∂
I m and φ is
∂xµ (f ◦ φ )|φ(p) where now ∂xµ denotes the standard partial derivative in R
the coordinate function. Recall that g smooth means that the function g ◦ φ−1 : R Im→ R I
∞
is in C in the usual calculus sense. This function is given by
∂(f ◦ φ−1 )
g ◦ φ−1 = , (3.2.2)
∂xµ
which is smooth since f is smooth.
Extending this, if we multiply ∂x∂ µ by a smooth function g µ and sum over µ, we still obtain
a smooth vector field, and, in fact, every vector field can be written locally like this. Therefore,
we shall write a vector field in local coordinates as
∂
X = gµ . (3.2.3)
∂xµ
Smooth in this case means that it maps every smooth vector field to a smooth function.
Special differential forms are the ones that associate with every p the one form dxµ |p and every
dual vector field can be written as (cf. (3.2.3))
ω = ωµ dxµ , (3.2.4)
µ ,...,µ ∂ ∂
p → Tν11,...,νrq (p) | p ⊗ · · · ⊗ |p ⊗ dxν1 |p ⊗ · · · ⊗ dxνr |p , (3.2.5)
µ
∂x 1 ∂xµq
µ ,...,µ
where the functions Tν11,...,νrq are smooth.
We remark that the definitions of smooth vector field and co-vector field were given in a
coordinate free fashion while the definition (3.2.5) seems to be dependent of the coordinates
chosen (see exercise 3.3.6).
The space of tensor fields of type (q, r) is denoted by Trq (M ). Special notations are reserved
to the space of vector fields, T01 (M ), which is also denoted by X (M ), and the space of dual
33
vector fields T10 (M ), which is also denoted by Ω1 especially, in the context of differential forms
as we shall see later. The space of smooth functions F(M ) is also denoted by T00 (M ) or Ω0 (M )
in the context of differential forms.
In general, we shall say that two vector fields X on M and Y on N are f −related if for
every p ∈ M
f∗ (X(p)) = Y (f (p)), (3.2.7)
that is the following commutative diagram holds.
34
f
M N
X Y
f∗
Tp M Tf (p) N
ω1 (p) = F ∗ ω2 (f (p)).
35
3.3 Exercises
Exercise 3.1 Consider R I n and R I m with the standard bases {~ej1 }, j = 1, 2, . . . , n and {~ek2 },
I n⊗ R
k = 1, 2, . . . , m. An ordered basis for R I m is given by (~ej1 , ~ek2 ), j = 1, 2, . . . , n, k = 1, 2, . . . , m
and
(~ej11 , ~ek21 ) < (~ej12 , ~ek22 ),
Exercise 3.2 The definition of F∗ T in (3.1.3) is given in terms of coordinates. Show that such
a definition does not depend on coordinates, i.e., the action of F∗ T on ⊗q Tf∗(p) N is uniquely
determined by the definition (3.1.3) independently of the coordinates used.
q
Exercise 3.3 For an element T̃ of T0,f (p) (M ) in a given system of coordinates at f (p) write
∂ ∂ ∂
T̃ := T̃ ν1 ,ν2 ,...,νq ν
⊗ ν2 ⊗ νq .
∂y 1 ∂y ∂y
Exercise 3.4 The definition of F ∗ T in (3.1.4) is given for given in terms of coordinates. Show
that such a definition does not depend on coordinates, i.e., the action of F ∗ T on ⊗r Tp M is
uniquely determined by the definition (3.1.4) independently of the coordinates used.
36
Exercise 3.6 Prove the smoothness in definition (3.2.5) does not depend on the coordinate
system used. That is, if we replace the coordinates xµ with coordinates y µ , the corresponding
µ ,...,µ
functions T̃ν11,...,νrq in
µ ,...,µ ∂ ∂
p → T̃ν11,...,νrq (p) |p ⊗ · · · ⊗ µq |p ⊗ dy ν1 |p ⊗ · · · ⊗ dy νr , (3.3.2)
∂y µ1 ∂y
are still smooth.
Exercise 3.7 Verify that every vector field X can be written locally as X = g µ ∂x∂ µ .
37
4 Integral curves and flows
Consider a vector field X on M and a point p ∈ M . Recall that we have denoted the tangent
d
I → M , using the push-forward, as c∗ dt
vector corresponding to a curve c : R .
Definition 4.0.1: Integral Curve
1. c(0) = p,
2.
d
c∗
= X(c(t)), (4.0.1)
dt
for every t ∈ (−a, b). This means that for every point c(t), with t ∈ (−a, b), the
tangent vector associated with c coincides with the value of the vector field at that
point.
Proposition 4.1.1
Let c : (−a, b) → M a curve such that c((−a, b)) ⊆ U , for the coordinate neighborhood
U , in the coordinate chart (U, φ), with p ∈ U . Then c is an integral curve for X = X ν ∂x∂ ν
at p if and only if cφ in (4.1.1) satisfies the Initial Value Problem
The proposition shows that the study of integral curves (upstairs) on a manifold can be
I m.
locally reduced to the study of solutions of differential equations (downstairs) in R
Proof. Assume c is an integral curve. Then applying both sides of (4.0.1) to the coordinate
functions xν , we obtain
d ∂
c∗ xν = X µ (c(t)) µ xν ,
dt ∂x
ċνφ (t) = X µ (c(t))δµν = X ν (c(t)).
That is, the differential equation in (4.1.2). Viceversa, if cφ = cφ (t) satisfies (4.1.2) consider
c := φ−1 ◦ cφ . Let us apply the left hand side and right hand side of (4.0.1) to a function
f ∈ F(M ), with this c, and for a given t ∈ (−a, b). The left hand side gives
d d d ∂f ◦ φ−1 ν
c∗ f = (f ◦ c(t)) = (f ◦ φ−1 ◦ cφ (t)) = ċφ . (4.1.3)
dt dt dt ∂xν
38
The right hand side gives
∂ ∂f ◦ φ−1
X(c(t))f = X(φ−1 ◦ cφ (t))(f ) = X µ (φ−1 ◦ cφ (t)) |φ
µ −1
−1 ◦c (t) f := X (φ (cφ (t))) ,
∂xµ φ
∂xν
(4.1.4)
which give the same result because of (4.1.2).
In particular, there is a one to one correspondence locally between what happens upstairs
and downstairs, and properties of integral curves can be obtained from properties of solutions of
differential equations. In particular, given a vector field X, at a point p ∈ M , we obtain that an
integral curve exists and is unique.
dcφ
= c2φ (t), cφ (0) = p,
dt
39
p
to be σ(t, p) = 1−pt , and it is defined only for the points (t, p) between the two curves in Figure
11.
∂
Figure 11: Domain of the flow σ = σ(t, p) for X = x2 ∂x I × R
, in R I
Proposition 4.2.1
1.
σ(0, p) = p, ∀p ∈ M, (4.2.1)
2.
σ(t, (σ(s, p)) = σ(t + s, p), (4.2.2)
for every t and s (small enough) so that both left hand side and right hand side
exist.
Proof. In the following proof, we use the notation σ to indicate a curve in M and σ µ to
indicate the µ-th component of its coordinate representation. Analogously, X represents
a vector field, X µ represents the µ-th component of its coordinate representation. See
that 4.2.1 follows from definition of integral curve of X at p.
40
Fix s and map left hand side and right hand side using a coordinate map φ (downstairs)
I m . If they map to the same curve, they are the same curve in M . For t = 0, they
to R
both map to φ(σ(s, p)). Moreover, taking the derivative of the left hand side, we have
d µ
(σ (t, σ(s, p))) = X µ (σ µ (t, σ(s, p))), (4.2.3)
dt
and taking the derivative of the right hand side,
d µ
(σ (t + s, p)) = X µ (σ µ (t + s, p)). (4.2.4)
dt
The two functions satisfy the same O.D.E with the same initial condition, so they are the
same by the uniqueness theorem. If σ(t + s, p) exits the coordinate neighborhood U , we
can inductively repeat the same argument on every other coordinate neighborhood.
A vector field X is called complete if the corresponding flow σ = σ(t, p) is defined on all
I × M.
of R
Theorem 4.2.1
If M is compact, every vector field X on M is complete. (see Exercise 4.2)
We have seen that, by definition σ(t, p) is, for fixed p, the integral curve at p associated with
the vector field X. We now fix t and look at σ(t, p) as a function on p. By varying the values
of t, we get a family of maps σ(t, ·) : M × M . These maps form a one parameter commutative
group of transformations which is local, i.e., its elements are defined for small t and the interval
of values of t depending on p where σ(t, p) is defined. It is a global one parameter group of
transformations if M is compact (cf. Theorem 4.2.1). We denote the transformation σ(t, ·) by
σt , and we have
1. σ0 = identity
2. σt+s = σt ◦ σs = σs ◦ σt ,
3. σ−s = (σs )−1 .
The following three lemmas give some more properties of the flow. Although these properties
are almost obvious consequences of the definition, it is useful to state them formally for future
reference. Let X a vector field and σt the associated one parameter group of transformations.
Lemma 4.2.1
For every q ∈ M and every function f ∈ F(M )
d
Xσt (q) f = (f (σt (q))). (4.2.5)
dt
41
Proof. Using the definition of the flow ,
d
Xσt (q) = σ∗ , σ0 (q) = q, (4.2.6)
dt
apply left hand side and right hand side to a function f ∈ F(M ),
d d
Xσt (q) f = σ∗ f= (f ◦ σt (q)) . (4.2.7)
dt dt
We can rewrite formula (4.2.7) in the form in which it is usually used, that is, for t = 0 and
q fixed
d
Xq f := Xf (q) = |t=0 (f ◦ σt (q)) , (4.2.8)
dt
d
which can be seen, as q varies in M , as an equality between the two functions Xf and dt |t=0 (f ◦ σt ).
Lemma 4.2.2
If σt is the flow associated with X, then σ−t is the flow associated with −X.
Proof. This fact follows from the upstairs-downstairs correspondence of equation (4.1.1),
I m : If cφ = cφ (t) satisfies
since it is easily verified in R
d ν
c = X ν (cφ (t)), (4.2.9)
dt φ
then cφ (−t) satisfies
d ν dcφ (−t)
(cφ (−t)) = − = −X ν (cφ (−t)). (4.2.10)
dt d(−t)
d
−Xσ−t (q) f = (f ◦ σ−t (q)), (4.2.11)
dt
and specializing at t = 0,
1
−Xq f = lim (f ◦ σ− (q) − f (q)) . (4.2.12)
→0
42
4.3 Exercises
Exercise 4.1 The notion of f −related vector fields introduced in subsection 3.2.1 is a gener-
alization of the familiar notion of change of coordinates for a system of differential equations.
Assume the vector field Y on N is f −related to the vector field X on M (cf. (3.2.7)). Prove
that if c = c(t) is an integral curve of X on M at p then f ◦ c is an integral curve of Y on N at
f (p).
Exercise 4.3 Calculate the flow of the vector field X = x1 ∂x∂ 1 + (x1 + x2 ) ∂x∂ 2 on S 2 in the
coordinate neighborhood U0+ . Recall that φ0+ ((x, y, z)) = (y, z).
Exercise 4.4 Recall that a smooth structure on RP n is defined as {Ul , φl } where Ul are classes
of representatives whose l − th component a in Rn+1 is non-zero and φl ([A]) = a−1 B with B
the vector in Rn with the l − th component removed. Calculate the flow of the vector field
X = ∂x∂ 0 + ... ∂x∂n−1 in neighborhood Un .
43
5 Lie Derivative
5.1 Lie derivative of a vector field
Consider a vector field X and its associated flow σt . Given another vector field Y we want to
analyze how the vector field Y varies along the integral curve σt (p), for a certain p ∈ M . So
we want to compare Yp and Yσ (p) . However, we cannot simply take the difference Yp − Yσ (p)
because these two vectors belong to different tangent spaces. The idea then is to take Yσ (p) and
bring it to Tp M by applying σ− ∗ . It makes sense to look at the difference ∆Y of two vectors in
Tp M (cf. Figure 12)
∆Y := (σ− )∗ Yσ (p) − Yp . (5.1.1)
Notice that this difference is itself a tangent vector and so is the limit as → 0.
Figure 12: Construction for the Lie derivative. Remark: In the picture is not necessarily small,
as a matter of fact Yσ (p) is not similar to Yp .
44
1. By changing to −k in (5.1.2) we obtain
1
LX Y = lim (σk )∗ Yσ−k (p) − Yp = (5.1.3)
k→0 −k
1
lim Yp − (σk )∗ Yσ−k (p) .
k→0 k
2. By collecting (σ− )∗ in (5.1.2) and using the fact that lim→0 (σ− )∗ is equal to the identity
operator, we obtain
1
LX Y = lim Yσ (p) − (σ )∗ Yp . (5.1.4)
→0
In this last definition, we transport the tangent vector Yp ∈ Tp M to Tσ (p) M before we
compare it with Yσ (p) .
[X, Y ] := XY − Y X, (5.1.5)
the product on the left hand side means composition, so that for any f ∈ F(M )
The following theorem states that the Lie derivative can be expressed as a commutator.
Theorem 5.1.1: Lie Derivative with Commuator
At a point p,
LX Y = [X, Y ]p . (5.1.7)
Proof. Take a function f ∈ F(M ) and calculate (LX Y )(f ). By definition, we have
1
(LX Y )(f ) = lim ((σ− )∗ Yσ (p) )f − Yp (f ) = (5.1.8)
→0
1
lim Yσ (p) (f ◦ σ− ) − Yp (f ) ;
→0
1 1
LX Y (f ) = lim Yσ (p) [(f ◦ σ− ) − f ] + lim Yσ (p) f − Yp f . (5.1.9)
→0 →0
45
where, in the last equality, we used (4.2.12). For the second limit, we have,
1 1
lim (Yσ (p) f − Yp f ) = lim ((Y f )(σ (p)) − (Y f )(p)) = Xp (Y f ), (5.1.11)
→0 →0
where in the first equality we used the definition of the action of a vector field on a function
(3.2.1) and in the last equality we used formula (4.2.5) of Lemma 4.2.1 with t = 0, q = p
and f replaced by Y f . Therefore, the theorem is proven.
The following propositions collect some of the main properties of the Lie derivative.
Proposition 5.1.1: Properties of Lie Derivative
3. (skew-symmetry) For X, Y ∈ X (M )
Properties 2),3) and 4) above qualify the vector space X (M ) with the operation (X, Y ) →
[X, Y ] as a Lie algebra. The commutator is also sometimes called a Lie bracket.
Proposition 5.1.3 is important for practical calculations especially when the map f is the
inclusion map. It is often convenient to work in a larger manifold (typically R I n ) than
the original manifold M (where we have to choose various systems of coordinates to be
patched together). The proposition ensures us that the algebra we do for the vector fields
46
on the larger space coincide with the algebra we would do on the smaller space. Consider
for instance the sphere in Figure 13 and the vector field Y in part a) whose integral curves
are rotations on the sphere about the y axis and the vector field Z in part b) whose integral
curves are rotations about the z axis.
Then we have
∂ ∂ ∂ ∂
i∗ Y = z −x , i∗ Z = −y +x .
∂x ∂z ∂x ∂y
From this we can calculate
∂ ∂ ∂ ∂
[i∗ Y, i∗ Z] = i∗ [Y, Z] = z − x , −y +x
∂x ∂z ∂x ∂y
47
Figure 13: Figure for Example 5.1.1
∂Y k µ ∂X
k
Z k = [X, Y ]xk = X(Y xk ) − Y (Xxk ) = X(Y k ) − Y (X k ) = X µ − Y . (5.1.13)
∂xµ ∂xµ
Im→ R
Notice we can represent X, Y and Z with a column vector of functions R I m . Now denote
JX (JY ) the Jacobian matrix corresponding to X (Y ). Then formula (5.1.13) can be written in
matrix-vector notation as
Z = JY X − JX Y.
The equality between the Lie derivative and the commutator suggests a geometric interpre-
tation of the Lie derivative as a measure of noncommutativity of two flows. Consider σt the flow
associated with X and τs the flow associated with Y . Starting from a point p, we move following
Y for time s first and then following X for time t. We arrive in σ(t, τs (p)). Now, starting from p
follow X first for time t and then follow Y for time s. We arrive in τ (s, σ(t, p)) (cf. Figure 14).
If the two flows commuted, τs (σt (p)) = σt (τs (p)). However, in general, this is not the case.
Let us compare τ (s, σ(t, p)) and σ(t, τ (s, p)). We cannot simply take the difference between these
two points as we are on a manifold (where in general we have not defined a distance function).
48
Figure 14: Noncommuting flows
We can however, take a (arbitrary) function f in F(M ) and consider the difference
Expand d(t, s) in a Taylor series about the point (0, 0) and calculate the various terms. We are
going to routinely use formula (4.2.7) in various forms. Clearly d(0, 0) = f (p) − f (p) = 0.
∂d ∂ ∂
= f (τs (σt (p))) − f (σt (τs (p))) =
∂s ∂s ∂s
and
Yτs (p) (f ◦ σt ) = (Y (f ◦ σt ))(τs (p)).
Taking the derivative with respect to s in (5.1.14), we get
This at s = 0, t = 0 gives
Yp (Y f ) − Yp (Y f ) = 0.
49
Analogously, one shows that the second derivative with respect to t is zero. We now calculate
the mixed derivative taking
∂
[Y f (τs ◦ σt (p)) − (Y (f ◦ σt ))(τs (p))] =
∂t
∂
Xσt (p) ((Y f ) ◦ τs ) − (Y f ◦ σt )(τs (p)),
∂t
∂
where in the second term we have switched the order of Y and ∂t . This is equal to
Xσt (p) ((Y f ) ◦ τs ) − (Y (Xf ))(τs (p)) := Xσt (p) ((Y f ) ◦ τs ) − Yτs (p) (Xf ),
which at t = 0, s = 0, gives
Xp (Y f ) − Yp (Xf ) = [X, Y ]p f.
So the commutator gives the first nonzero term in the Taylor expansion measuring the distance
between the two values of the function f .
The motivation for this definition is similarly to (5.1.2), that is, we want to compare the
values of the co-vector field at two different points, but since these values belong to two different
co-tangent spaces, we use (this time) the pull-back to bring both co-tangent vectors to Tp∗ M .
The expression in local coordinates for LX ω can be found to be
∂ωµ ν ∂X ν
LX ω = X + ων dxµ . (5.2.2)
∂xν ∂xµ
To extend the definition of Lie derivative to general tensors, we need to extend the definition
of push-forward and pull-back to tensors of mixed type. This can be done if the map f : M → N
used in the definition is a diffeomorphism (like for example σ ).
Definition 5.2.2: Push-forward and Pull-back of Tensor Extension
q q q
For a tensor in Tr,p (M ), the push-forward F∗ : Tr,p (M ) → Tr,f (p) (N ) is defined as
F ∗ (V1 ⊗ · · · ⊗ Vq ⊗ ω1 ⊗ · · · ⊗ ωr ) = f −1 ∗ V1 ⊗ · · · ⊗ f −1 ∗ Vq ⊗ f ∗ ω1 ⊗ · · · ⊗ f ∗ ωr . (5.2.4)
These defintions are extensions of the coordinate definitions (3.1.3 and 3.1.4) (See Exercise
50
3.2 and 3.4).
We remark, in particular, that the flow σ is a diffeomorphism. Let σ− ∗ denote the pushfor-
ward associated with the flow σ− of a vector field X.
Definition 5.2.3: Lie Deriviative of Tensor field
The Lie derivative of a general tensor field T ∈ Trq (M ) is defined as
1
LX T := lim σ− ∗ Tσ (p) − Tp . (5.2.5)
→0
The following proposition (see Proposition 5.1 [1] and proof there) allows us to reduce the
study of Lie derivatives of general tensors to Lie derivatives of vector fields and co-vector fields
only.
Proposition 5.2.1: Lie Derivative of Tensors to Vector fields
51
5.3 Exercises
Exercise 5.1 Prove Proposition 5.1.1
Exercise 5.4 Give an alternative proof of Theorem 5.1.1 by proving that LX Y xk is equal to the
right hand side of (5.1.13).
∂
Exercise 5.5 Prove formula (5.2.2). (Hint: Apply (5.2.1) to the tangent vector ∂xµ |p .)
Exercise 5.6 Prove that the pushforward distributes with the Lie derivative that is
f∗ [X, Y ] = [f∗ X, f∗ Y ] .
52
6 Differential Forms Part I: Algebra on Tensors
6.1 Preliminaries: Permutations acting on tensors
Definition 6.1.1: Permutation on a Tensor
0 (M ) and P a permutation in the symmetric
Consider a tensor of the type (0, r), i.e., ω ∈ Tr,p
a
group on r elements, Sr . A new tensor, P ω, is defined as
Notice here there is a little abuse of notation since P denotes both the operation on the tensor
ω and the permutation on 1, . . . , r. Also, P (j) denotes the transformation of j in the permuted
r−ple P (1, 2, . . . , r). For instance, if r = 3 and P is given by P (1, 2, 3) := (3, 2, 1), then
P ω(V1 , V2 , V3 ) = ω(V3 , V2 , V1 ).
Question: If ω is described in a given basis as ωµ1 ,...,µr dxµ1 ⊗ · · · ⊗ dxµr , how is P ω written
in the same basis?
We have
∂ ∂
(P ω)µ1 ,...,µr = (P ω)( µ1 , . . . , µr ) := (6.1.2)
∂x ∂x
∂ ∂
ω , . . . , = ωµP (1) ,...,µP (r) ,
∂xµP (1) ∂xµP (r)
that is the indexes are permuted according to the permutation P . For instance, in the above
example where P (1, 2, 3) = (3, 2, 1), we have
Recall that a tensor of type (0, r) is specified by mr elements. As a special (familiar) case,
tensors of the type (0, 2) are specified by m × m matrices ωµ1 ,µ2 . The are only two possible
permutations for (1, 2), the identity and P (1, 2) = (2, 1). Since (P ω)µ1 ,µ2 = ωµ2 ,µ1 the P
operation corresponds to exchanging the first (row) index with the second (column) index
53
and therefore corresponds to matrix transposition.
Using the definition (6.1.1), we define the symmetrizer operation, S, for elements
0 (M ),
ω ∈ Tr,p
1 X
S(ω) := P ω, (6.1.4)
r!
P ∈Sr
P ω = ω, ∀P ∈ Sr . (6.1.6)
Sω = ω, Aω = 0, (6.1.8)
Sω = 0, Aω = ω. (6.1.9)
Notice in particular that, for every tensor ω, Sω (Aω) is totally symmetric (totally anti-symmetric),
so that we have, using (6.1.8) and (6.1.9),
54
look more closely at totally anti-symmetric tensors of type (0, r), written in a coordinate basis
as
ω = ωµ1 ,...,µr dxµ1 ⊗ · · · ⊗ dxµr .
Given a permutation P , we have from (6.1.2) and using the antisymmetry property
(P ω)µ1 ,...,µr = ωµP (1) ,...,µP (r) = sign(P )ωµ1 ,...,µr . (6.1.10)
So, totally anti-symmetric tensors have the property that if we permute the indexes according
to a certain permutation P they change or do not change the sign according to the sign of P .
Example 6.1.2: Anti-symmetric Tensor Examples
For example, for a tensor of type (0, 2), totally anti-symmetric, ωµ,ν , µ, ν = 1, . . . , m,
is such that ωµ,ν = −ων,µ as for an antisymmetric matrix, as we have seen. Another
example is the Levi-Civita symbol µ,ν,λ , µ, ν, λ = 1, 2, 3, which is zero any time there are
repeated indices and it is otherwise defined by P (1),P (2),P (3) := sign(P ). So, for example,
1,2,3 = 1, and 2,1,3 = −1.
Differential forms of order r form a vector space denoted by Ωrp (M ), and we have Ω0p (M ) := R
I,
Ω1p (M )= Tp∗ M .
Consider now a tensor ω ∈ Ωrp (M ) and a tensor ξ ∈ Ωqp (M ). In (3.1.1) we have defined the
product of ω and ξ as
(r + q)!
ω∧ξ = A(ω ⊗ ξ). (6.2.2)
r!q!
55
1 X
sign(P )(P ω ⊗ ξ)(V1 , . . . , Vr , Vr+1 , . . . , Vr+q ) =
r!q!
P ∈Sr+q
1 X
sign(P )ω(VP (1) , . . . , VP (r) )ξ(VP (r+1) , . . . , VP (r+q) ).
r!q!
P ∈Sr+q
The following theorem gives some of the properties of the exterior product, which are used in
most of the calculation with r−forms
Theorem 6.2.1: Properties of the Exterior Product
ω ∧ ξ = (−1)qr ξ ∧ ω.
3. Associative
(ω ∧ η) ∧ ξ = ω ∧ (η ∧ ξ).
4. Such that
F ∗ (ω ∧ ξ) = (F ∗ ω) ∧ (F ∗ ξ), (6.2.3)
where F ∗ is the pull back associated with a smooth map f : M → N .
We give a sketch of the proof of property 3 and postpone the proof of the properties 1,2, and
4 to Exercise 6.1, below. The proof uses the following Lemma.
Lemma 6.2.1: Property of Anti-symmetrizer Operation
(Theorem 2 part (1) in Spivak, pg. 203) If A(ω) = 0, for a tensor ω, A(ξ ⊗ω) = A(ω ⊗ξ) =
0 for every tensor ξ.
56
We calculate, using the definition (assume ω ∈ Ωrp (M ), η ∈ Ωqp (M ), ξ ∈ Ωsp (M ))
(r + q + s)! (r + q + s)! (r + q)
(ω ∧ η) ∧ ξ := A((ω ∧ η) ⊗ ξ) = A A(ω ⊗ η) ⊗ ξ =
(r + q)!s! (r + q)!s! r!q!
(r + q + s)! (r + q + s)!
A((A(ω ⊗ η)) ⊗ ξ) = A((A(ω ⊗ η) + ω ⊗ η − ω ⊗ η) ⊗ ξ) =
r!q!s! r!q!s!
(r + q + s)!
[A ((A(ω ⊗ η) − ω ⊗ η) ⊗ ξ) + A(ω ⊗ η ⊗ ξ)] .
r!q!s!
0
The first term in the square bracket is zero because of Lemma 6.2.1, since with ω := A(ω ⊗ η) −
0
ω ⊗ η we have Aω = 0. Therefore, we have
(r + q + s)!
(ω ∧ η) ∧ ξ = A(ω ⊗ η ⊗ ξ). (6.2.4)
r!q!s!
The same result is obtained if we start with ω ∧ (η ∧ ξ), which proves the claim.5
In view of part 2 of Proposition 6.2.1, we can simply write ω ∧η ∧ξ for (ω ∧η)∧ξ or ω ∧(η ∧ξ).
Moreover extending inductively formula (6.2.4), we have
Corollary 6.2.1: n Exterior Products
For ωi ∈ Ωrpi (M ), i = 1, . . . , k
(r1 + r2 + · · · + rk )!
ω1 ∧ ω2 ∧ · · · ∧ ωk = A(ω1 ⊗ ω2 ⊗ · · · ⊗ ωk ), (6.2.5)
r1 !r2 ! · · · rk !
This is the oriented area of the parallelogram with sides V1 and V2 , which justifies the fact
that such form is called the area element. If we use polar coordinates r and θ instead,
from dx = cos(θ)dr − r sin(θ)dθ, and dy = sin(θ)dr + r cos(θ)dθ, we get, that the area
element is
5
For more details see the proof of part (3) of the Theorem 2 in Spivak, pg. 203.
57
6.3 Characterization of the vector spaces Ωrp (M )
We now want to find a suitable basis for Ωrp (M ). An element ω ∈ Ωrp (M ) can be written as
and unless we have information on the coefficients ωµ1 ,...,µr , there is no sign that this tensor is
totally antisymmetric. We now show that the exterior product forms a basis.
Proposition 6.3.1: dxµ1 ∧ · · · ∧ dxµr is a basis of Ωrp (M )
In order to highlight this fact, we apply A to both left hand side and right hand side of
(6.3.1) using the fact that Aω = ω, and we have
1
ωµ ,...,µr dxµ1 ∧ · · · ∧ dxµr .
r! 1
This shows that dxµ1 ∧· · ·∧dxµr , µ1 , . . . , µr ∈ {1, . . . , m}, span Ωrp (M ). Moreover, because
of the properties of Theorem 6.2.1, we can assume that µ1 < µ2 < · · · < µr . The m r
tensors dxµ1 ∧ · · · ∧ dxµr , with µ1 < µ1 < · · · < µr , therefore span Ωrp (M ). To show that
this is a basis we have to show that these tensors are linearly independent. Take a linear
combination of them which gives zero:
X
0= aµ1 ,...,µr dxµ1 ∧ · · · ∧ dxµr , (6.3.3)
µ1 <µ2 <···<µr
∂
and apply it to ∂xk1
, . . . , ∂x∂kr for fixed k1 , . . . , kr , with k1 < k2 < . . . < kr . This gives
X
µ1 µr ∂ ∂
0= aµ1 ,...,µr r!A(dx ⊗ · · · ⊗ dx ) k
, . . . , kr = (6.3.4)
µ1 <µ2 <···µr
∂x 1 ∂x
" #
X X ∂ ∂
aµ1 ,...,µr sign(P ) dxµ1 · · · dxµr .
µ1 <µ2 <···µr P ∈Sr
∂xkP (1) ∂xkP (r)
Consider one of the terms inside the square brackets corresponding to a fixed value µ1 <
µ2 < · · · µr . It is easily seen that the only term in the sum which is possibly different from
zero is the one corresponding to the basic permutation, P : (1, 2, . . . , r) → (1, 2, . . . , r). If
this was not the case we would have at least an inversion, that is i < j and kP (j) < kP (i) .
µ
Then we would have terms in the products δkµPi (i) and δkPj (j) , both equal to 1, that is
µi = kP (i) and µj = kP (j) . However, kP (j) < kP (i) contradicts µi < µj . Moreover,
necessarily we must have kj = µj for j = 1, . . . , r. This therefore shows ak1 ,...,kr = 0.
We have therefore found a basis. We notice that, from Part 2 of Theorem 6.2.1, Ωqp (M ) is
zero if q > m. Moreover, by the equality of the dimensions, Ωrp (M ) is isomorphic to Ωm−r
p (M ).
The space
Ω0p (M ) ⊕ Ω1p (M ) ⊕ · · · ⊕ Ωm
p (M ), (6.3.5)
58
with the wedge product has the structure of a graded algebra, the grading being given by the j
in Ωjp (M ).
59
6.4 Exercises
Exercise 6.1 Prove parts 1,2,4 of Theorem 6.2.1.
Exercise 6.2 Prove or disprove that properties 1,2,3 hold if the exterior product ∧ is replaced
by the tensor product ⊗. (For property 4 see (3.1.4)).
Exercise 6.3 Use Part 2 of Theorem 6.2.1 to prove Ωqp (M ) is zero if q > m.
60
7 Differential Forms Part II: Fields and the Exterior Derivative
7.1 Fields
Definition 7.1.1: r-form Field
Analogously to what we have done for general tensor fields, we define a r- form field
as a smooth assignment to every point p of an r-form. The space of r-form fields will be
denoted by Ωr (M ) (without reference to the point p).
where the arbitrary functions of ωµ1 ,µ2 ,...,µr for µ1 < µ2 < · · · < µr automatically by anti-symmetry
determine the other ωµ1 ,µ2 ,...,µr .
61
7.2.1 Independence of coordinates
Consider now each of the r terms in the above sum, which is itself a sum over ν, ν1 , . . . , νr . Fix a
term in the sum corresponding to certain values ν = ν̄, ν1 = ν̄1 ,..., νr = ν̄r , i.e., the (r + 1)-ple,
ν̄, ν̄1 , ..., ν̄j , ..., ν̄r . The term corresponding to the (r + 1)-ple, ν̄j , n̄u1 , ..., n̄u, ..., n̄ur is equal and
opposite because of the equality of mixed derivatives and property 2. of Theorem 6.2.1. These
terms all cancel, and all the r terms in the sum are zero. We are left with
µ1 µ2
∂xµr
0 1 ∂ωµ1 ,...,µr ∂x ∂x
dr ω = · · · νr dy ν ∧ dy ν1 ∧ · · · ∧ dy νr = (7.2.6)
r! ∂y ν ∂y ν1 ∂y ν2 ∂y
µ1 µ2 µr
1 ∂ωµ1 ,...,µr ν ∂x ν1 ∂x ν2 ∂x νr
dy ∧ dy ∧ dy ∧ ··· ∧ dy =
r! ∂y ν ∂y ν1 ∂y ν2 ∂y νr
1 ∂ωµ1 ,...,µr
dy ν ∧ dxµ1 ∧ dxµ2 ∧ · · · ∧ dxµr .
r! ∂y ν
The first factor in the expression above is the differential of the function ωµ1 ,...,µr . If we write
this differential in the x coordinates, we find the expression of dr ω as desired.
62
7.3 Properties of the exterior derivative
The following Theorem describes some of the main proprties of exterior differentiation. We drop
in the following the index r in dr and denote the exterior derivative simply by d.
Theorem 7.3.1: Properties of the Exterior Derivative
1. is linear
d(aω1 + bω2 ) = adω1 + bdω2 ; (7.3.1)
2. satisfies for ξ ∈ Ωq (M ), ω ∈ Ωr (M ),
3. satisfies
d2 = 0. (7.3.3)
Proof. 1. follows directly from the definition while 2. is left as exercise. We only prove 3.
1 ∂ωµ1 ,...,µr µ µ1 µr
d(dω) = d dx ∧ dx ∧ · · · ∧ dx =
r! ∂xµ
1 ∂ ∂ωµ1 ,...,µr k
dx ∧ dxµ ∧ dxµ1 ∧ · · · ∧ dxµr .
(r!)(r + 1)! ∂xk ∂xµ
Inside the round parentheses, the terms with k = µ are zero because dxk ∧ dxk = 0. The
terms with k 6= µ pairwise cancel because of the equality of mixed derivatives.
The following theorem says that properties (7.3.1), (7.3.2) and (7.3.3) uniquely define in a
coordinate independent fashion, the exterior derivative.
Theorem 7.3.2: Uniqueness of the Exterior Derivative
0
Consider an operator d : Ωr (M ) → Ωr+1 (M ), which satisfies (7.3.1), (7.3.2) and (7.3.3)
0 0
(with d replaced by d ) and agrees with d on functions, then d = d.
Notice this theorem is another way to show that the definition of d does not depend on
coordinates. We could have simply defined d as the unique operator that satisfies (7.3.1),
(7.3.2)and (7.3.3) and it is equal to the differential on functions in F(M ).
Proof. (the proof follows Spivak pg. 211-212.) By linearity, it is enough to show that
0
d (f dxµ1 ∧ · · · ∧ dxµr ) = d(f dxµ1 ∧ · · · ∧ dxµr ) := df ∧ dxµ1 ∧ · · · ∧ dxµr . (7.3.4)
63
0
because of equality with d on functions. It is therefore enough to show that d (dxµ1 ∧
0
· · · ∧ dxµr ) = 0, or equivalently since dxµj = d xµj (because of equality on functions), that
0 0 0
d (d xµ1 ∧ · · · ∧ d xµr ) = 0. This is shown by induction on r. For r = 1, it is true because
of (7.3.3), and using property (7.3.2),
0 0 0 0 0 0 0 0 0 0 0
d (d xµ1 ∧· · ·∧d xµr ) = d (d xµ1 )∧d xµ2 ∧· · ·∧d xµr −d xµ1 ∧d (d xµ2 ∧· · ·∧d xµr ) = 0−0 = 0,
(7.3.6)
where we have used the inductive assumption.
ω : X (M ) × X (M ) × · · · × X (M ) → F(M ).
One more property, which can also be taken as coordinate independent definition of the exterior
derivative, is given in terms of how, for ω ∈ Ωr (M ), dω ∈ Ωr+1 (M ) acts on (r + 1)-tuples of
vector fields to give functions in F(M ). We have (the proof that this definition is equivalent to
the one previously given is omitted; it can be found for instance in Spivak Theorem 13, Chapter
7) Xj ∈ X (M ), j = 1, . . . , r + 1,
r+1
X
dω(X1 , . . . , Xr+1 ) = (−1)i+1 Xi (ω(X1 , . . . , Xi−1 , Xi+1 , . . . , Xr+1 ))+ (7.3.7)
i=1
X
(−1)i+j ω([Xi , Xj ], X1 , . . . , Xi−1 , Xi+1 , . . . , Xj−1 , Xj+1 , . . . , Xr+1 ).
i<j
Special cases of this formula are important and often used in calculations. For r = 0, ω is just
a smooth function and the formula (7.3.7) gives dω(X1 ) = X1 (ω), that is the definition of the
differential of a function (cf. (2.2.1)). If ω is a 1− form, we get, with X, Y ∈ X (M ),
7.3.1 Examples
Example 7.3.1: M = R3
Consider M = R I 3 . Since Ω0p (M ) and Ω3p (M ) are both one-dimensional, both the elements
of Ω (M ) and Ω3 (M ) can be identified with functions. That is, an element in Ω0 (M )
0
64
well known operations of vector calculus. In particular,
∂ω0 ∂ω0 ∂ω0
dω0 = dx + dy + dz,
∂x ∂y ∂z
~
∂B
~ = 0,
∇·B ~ = 0,
+∇×E (7.3.10)
∂t
follow from the fact that d2 A = 0, i.e., by setting each of the four components of this
3-tensor equal to zero.
H r (M ) := Z r (M )/B r (M ), (7.3.11)
65
is called the r−th co-homology group.
My = Nx .
fx = M
and
fy = N,
giving
f (x, y) = c
as the solution. This is not precisely the case since this says that a covector field being
closed implies that it is exact. This is why in these exact equations we require that the
open set in R2 where the covector field is defined is simply connected. This means it is
path connected, and any two smooth paths on the open set with the same endpoints can
be continuously deformed into each other, or simply the set has no holes. This condition
lets closed imply exact for our covector field in R2 .
The interior product is a form of ‘contraction’ (cf. subsection ??) between the vector
field X and the r−form ω. In order to see this, let us express iX ω in given coordinates.
We write X = X µ ∂x∂ µ and ω := r!1 ωµ1 ,µ2 ,...,µr dxµ1 ∧ · · · ∧ dxµr , and iX ω in coordinates as
1
iX ω = (r−1)! (iX ω)ν1 ,...,νr−1 dxν1 ∧ · · · ∧ dxνr−1 . We know that
∂ ∂
(iX ω)ν1 ,...,νr−1 = iX ω , . . . , νr−1 := (7.4.2)
∂xν1 ∂x
∂ ∂ µ ∂ ∂ ∂
ω X, ν1 , . . . , νr−1 = ω X , , . . . , νr−1 = X µ ωµ,ν1 ,...,νr−1 .
∂x ∂x ∂xµ ∂xν1 ∂x
66
Therefore,
1
iX ω = X µ ωµ,ν1 ,...,νr−1 dxν1 ∧ · · · ∧ dxνr−1 .
(r − 1)!
3.
iX (ω ∧ ξ) = (iX ω) ∧ ξ + (−1)r ω ∧ iX ξ,
if ω ∈ Ωr (M ).
5. i[X,Y ] = LX ◦ iY − iY ◦ LX
Proof. Property 1. follows directly from the definition. Property 2. also follows from the
definition of Lie derivative of a function f and the definition of differential of a function.
Properties 3., 4. and 5. are left as exercise.
A lot of algebra can be done with differential forms and the operations d, ∧ and iX and
Lie derivatives of vector fields to obtain identities relating these operations. One important
formula is the so-called Cartan magic formula, or, simply, Cartan identity, which relates the Lie
derivative along a vector field X, with the interior product with respect to X. It says that the
Lie derivative of a differential form along X is the anticommutator of the exterior derivative d
and the interior product iX . In particular, we have
Theorem 7.4.2
For every ω ∈ Ωr (M ),
LX ω = (iX ◦ d + d ◦ iX )ω. (7.4.3)
67
Remark 7.4.1: Another Proof of Theorem 7.4.2
The proof we present is different from Nakahara’s one and it does not use the
representation of the various operations in given coordinates. It follows Dr. Zuoqin
Wang’s online notes from the University of Michigan (cf. Lecture 28 in
http://www-personal.umich.edu/ wangzuoq/437W13/.) In principle, there is noth-
ing wrong with proofs which are done in specific coordinates. However, they tend
to be a little messier with the indices and it might be difficult to arrange the var-
ious terms and indices to get exactly the equality we want. Some tricks are found
repeatedly in proofs of equalities that involve iX , LX and d. One of them is to use
induction on the order r of a differential form. Another trick is to use property 2.
of Proposition 5.2.1., which extends to differential forms as
Often proofs for low degree tensors can be done using the expression in local coordi-
nates without too much confusion with indexes. One tool often used is the equality
of mixed derivatives. These elements are present in the following proof.
Proof. By linearity, it is enough to show (7.4.3) for an r−form of the type ω = f dx1 ∧
dx2 ∧ · · · ∧ dxr for f ∈ F(M ). If r = 0, (7.4.3) is true since iX ω = 0 and property 2. of
Theorem 7.4.1 holds. To prove it for general r ≥ 1, assume it true for r − 1 and write
ω1 := f dx2 ∧ · · · ∧ dxr , so that
ω = dx1 ∧ ω1 . (7.4.5)
Calculate the left hand side of (7.4.3), using (7.4.4) and (7.4.5). We have
LX ω = (LX dx1 ) ∧ ω1 + dx1 ∧ (LX ω1 ). (7.4.6)
The right hand side of (7.4.3) is
(iX ◦ d + d ◦ iX )(dx1 ∧ ω1 ) = iX ◦ d(dx1 ∧ ω1 ) + d ◦ iX (dx1 ∧ ω1 ) =
−iX (dx1 ∧ dω1 ) + d[(iX dx1 )ω1 − dx1 ∧ iX ω1 ],
using d2 = 0, and property 3. of Theorem 7.4.1, and (7.3.2). Again using these properties,
we get:
= −(iX dx1 )dω1 + dx1 ∧ iX ◦ dω1 + (diX dx1 ) ∧ ω1 + (iX dx1 )dω1 + dx1 ∧ d(iX ω1 ) =
68
7.5 Exercises
Exercise 7.1 Prove part 2 of the above theorem.
69
8 Integration of differential forms on manifolds part I: Prelimi-
nary Concepts
Consider a differential r−form ω defined on a (in some sense measurable) set S ⊆ M , for
example a submanifold of dimension r. Just like we did in the previous treatment, we may
define integration of an r−form on an (r−dimensional) sub-manifold in terms of integration on
I r . Since S is (locally) Euclidean, there exists a coordinate chart φ (locally) mapping S to a set
R
φ(S) in Rr , with coordinates x1 , . . . , xr . If ω := f dx1 ∧ dx2 ∧ · · · ∧ dxr , we can (roughly) define
Z Z
ω= f (φ−1 (x1 , x2 , . . . , xr ))dx1 dx2 · · · dxr , (8.0.1)
S φ(S)
where the integral on the right hand side is a standard Riemann integral in R I r . Beside the easily
settled issue of independence of coordinates in this definition, there are few issues to be carefully
analyzed. First of all, for domains of integration of RRiemannian integrals, we have a concept of
b Ra
orientation, which leads for instance to the fact that a f (x)dx = − b f (x)dx. We need to define
a notion of orientation on manifolds, which generalizes the one on R I . Then, for most manifolds
and sets, it is not possible to find a single coordinate map φ mapping S to φ(S) since S might
be covered by more than one coordinate neighborhood. The device used to patch together these
various coordinate charts in the definition and calculation of the integral is called a partition of
unity. These topics will be treated in the next two subsections. Then we shall study integration
on special sets called chains which will lead us to the first version of Stokes’ theorem. We will
then generalize to integration over more general sets which we call regular domains and, on those,
we shall state a second version of Stokes’ theorem. This theorem is the most important one for
integration on manifolds, generalizing several theorems of multivariable calculus such as the
fundamental theorem of calculus, Green’s theorem, and the divergence theorem. Our treatment
mostly follows F. Warner’s book (Chapter 4).
∂ ∂xi ∂
| p = |p |p . (8.1.1)
∂y j ∂y j ∂xi
If
∂xi
det |p > 0, (8.1.2)
∂y j
∂ ∂
then the two bases { ∂x i |p }, and { ∂y j |p }, are said to have the same orientation. Other-
wise, if this determinant is < 0, they are said to have opposite orientation.
70
Definition 8.1.2: Orientable Manifold
A connected manifold M with atlas {Uj , φj } is said to be orientable if there exists a
subcover of {Uj }, such that for every pair of overlapping charts in this subcover and at
any point p in the intersection of the two coordinate neighborhoods, the corresponding
∂ ∂
bases of Tp M , { ∂x i |p } and { ∂y j |p } have the same orientation. The atlas (or the subcover)
given this way is said to be orientation compatible. In other terms, a manifold is
orientable if and only if there exists an orientation compatible atlas.
Example 8.1.1: S 1
All of the manifolds introduced in the examples in subsection 1.1 are orientable. Consider,
for instance, the circle S 1 with the two charts
71
Example 8.1.2: Möbius strip
The Möbius strip in Figure 16 is the prototipical example of non orientable manifold. It
consists of a strip depicted in the upper part of the figure where the two ends are glued
together. However, a half twist is performed so that the points A and B on the left and
right hand of the strip match.
This manifold can be covered by two overlapping charts depicted in Figure 16. To give the
resulting manifold the structure of a differentiable manifold, we cover it with two charts.
The first chart with coordinates (x1 , y1 ) is depicted in the upper part of the figure and
associated coordinate neighborhood consists of the portion of the strip from −3 3
2 to 2 . It
2
is mapped via the identity to R I . The second coordinate neighborhood in the lower part
of the figure contains the segment A − B, which is mapped to a segment on the y axis.
The boxes between −2 and −1 and between 1 and 2 are mapped to the two boxes on the
left and right side of the y axis. However, the left box is twisted so as to make the points
A and B match. The coordinates are x2 and y2 . The regions U and V are where the two
coordinate neighborhoods overlap. On V , we have for the transition functions x1 = 2−x2 ,
y1 = −y2 , so that the Jacobian is
−1 0
JV = ,
0 −1
72
Figure 16: Differentiable structure for a Möbius strip
We need to assume that the manifold M is paracompact, that is, every open cover has an
open cover that is a locally finite refinement. Then we can say it has an atlas (compatible
with the assigned differentiable structure) {Uj , φj } such that every point p, is covered by
only a finite subset of the {Uj }’s.
This hypothesis is not really a hypothesis since we have assumed that M is second countable
and Hausdorff, and it can be proven that this is sufficient to conclude that M is paracompact if M
is locally compact (See 1.3) (cf. for instance Lemma 1.9 in F. Warner). Clearly compact spaces
(manifolds) are automatically paracompact and, for example, all manifolds we have discussed in
the first lecture, which have atlases with a finite number of coordinate neighborhoods, are also
automatically paracompact. We assume therefore in the following that the atlas we use on our
manifold M has a finite set of coordinate neighborhoods at each point.
73
Definition 8.2.2: Partition of unity
Given such a cover {Uj }, a partition of unity subordinate to the cover {Uj } is a
family of smooth functions j : M → R
I such that
1. j (p) ≥ 0, ∀p ∈ M ,
2. j (p) = 0, if p ∈
/ Uj ,
Notice that in the last condition, because of the choice of the cover and condition 2, only
a finite number of terms has to be considered. However, such a number depends on the point
p ∈ M.
The existence of a partition of unity is not a completely trivial matter, and a construction can
be found in Warner or Spivak. We summarize here the main ideas without going into technical
details. The main tool is the function
1
h1 (x) = e− x2 , x ∈ (0, +∞), (8.2.1)
is equal to zero outside the interval (−2, 2), equal to 1 in the interval [−1, 1], and continuously
grows from zero to 1 in [−2, −1] and decreases from 1 to zero in [1, 2]. The bump function
(8.2.4) can be extended to higher dimensions by simply multiplying several bump functions with
each-other, one for each dimension, i.e., b(x)b(y)b(z) · · · , and can be made asymmetric or can be
shrunk or shifted by simply scaling or shifting the variables. Consider now the given open cover
on M , {Uj }. According to the shrinking Lemma, (Theorem 14 pg 51 in Spivak), it is possible to
0 0 0 0
obtain another cover {Uj }, such that, for every j, Ūj ⊂ Uj , where Ūj is the closure of Uj . Now,
0
let us consider one of the neighborhoods Uj and corresponding Uj , and let us map it using the
0
coordinate map φj to R I m . Let us assume for simplicity that φj (Uj ) is bounded so that φj (Ūj )
0
is compact. We can cover φj (Ūj ) with a finite number (hyper)-boxes so that the union of these
boxes is entirely contained in φj (Uj ). This is shown in Figure 17.
74
Figure 17: Construction for Partition of Unity
For each box, construct a (generalized) bump function, which is 1 inside the box and zero
outside φj (Uj ). By summing all these (finite number of) functions, we obtain a function, which
0
is strictly positive on φj (Uj ) and zero outside φj (Uj ). Call this function fj . The function
0
Fj = fj ◦ φj is a smooth function (back) on the manifold M , which is strictly positive on Uj and
zero outside Uj . By the way it was constructed, this function is smooth. The functions
Fj (p)
j (p) := P , (8.2.5)
k Fk (p)
where the sum is taken over the finite number of coordinate neighborhoods containing p, give a
partition of unity for the given manifold, subordinate to the cover {Uj }. Notice in particular that
0
the fact that the Uj forms a cover guarantees that the denominator in (8.2.5) is always different
from zero.
75
8.3 Orientation and existence of a nowhere vanishing form
Definition 8.3.1: Volume form
One of the main features of an orientable manifold is that it admits an m-form (recall
m = dim M ) called a volume form, which is nowhere vanishing and plays the role of a
measure on the manifold.
Proof. (Proof of Theorem 8.3.1) Assume there exists a nowhere vanishing m−form ω for
the m-dimensional manifold M , which we write locally for a chart (U, φ) as
for f (x) > 0, on φ(U ). We remark that f is always positive since it is continuous, and in
order to change sign, it has to be equal to zero at some point, which contradicts the fact
that it is nowhere vanishing. Consider now the form ω on another chart (V, ψ), overlapping
with (U, φ), and write it there as
ω = g(y)dy 1 ∧ dy 2 ∧ · · · ∧ dy m , (8.3.2)
with g(y) 6= 0 defined in ψ(V ). As we have done for f , since g(y) is continuous and
never zero, we can assume that g(y) is always positive or always negative. If it is always
negative, we can change, in the chart, the coordinate y 1 to −y 1 which, changes the sign of
g(y). Therefore, we can assume that the two coordinate charts are such that both f and
g are positive on the domain where they are defined. On U ∩ V ,
∂ ∂
f (φ(p)) = g(ψ(p))(dy 1 ∧ · · · ∧ dy m ) |p , . . . , | p = (8.3.4)
∂x1 ∂xm
g(ψ(p)) X 1 m ∂ ∂
sign(P )(dy ⊗ · · · ⊗ dy ) P (1) p
| , . . . , P (m) |p =
r! ∂x ∂x
P ∈Sm
∂y 1 ∂y 2 ∂y m
g(ψ(p)) X
sign(P ) |p |p · · · |p .
r!
P ∈S
∂xP (1) ∂xP (2) ∂xP (m)
m
76
∂y i
The sum on the right hand side is the determinant of the Jacobian matrix ∂x j at p [Recall
that one of the definitions of the determinant of a m × m matrix {ai,j } is
X
det({ai,j }) = sign(P )a1,P (1) a2,P (2) · · · ar,P (m) ].
P ∈Sm
Therefore, given the signs of f and g this determinant is positive, which implies that the
two coordinate systems share the same orientation. Since this can be repeated for any
two overlapping coordinate systems, the manifold is orientable.
To show the converse in the theorem statement, we assume that a cover exists such that
overlapping charts have the same orientation. Then taking one of these charts (U, φ) a
nowhere vanishing form defined on U certainly exists. It is
The form ω is then extended to the other charts. In particular, let ωj be the form associated
with the coordinate chart (Uj , φj ). The nowhere vanishing form is then defined using a
partition of unity {j } subordinate to the cover {Uj } as
X
ω := j ωj . (8.3.6)
j
To show that this form is nowhere vanishing, consider a point p ∈ M and according to the
paracompactness assumption, there are only a finite number of coordinate charts covering
p. Take the (finite) intersection of the corresponding coordinate neighborhoods. Consider
one of the coordinate neightborhoods and the corresponding basis of the tangent space
{ ∂x∂ 1 |p , . . . , ∂x∂m |p }. For every j, let ωj be written as ωj := dy 1 ∧ · · · dy m , and we have (cf.
calculation (8.3.4))
j
∂ ∂ ∂y
ωj 1
|p , . . . , m |p = det |p > 0. (8.3.7)
∂x ∂x ∂xk
Therefore, ω applied to ∂x∂ 1 |p , . . . , ∂x∂m |p gives a finite sum weighed by the j (p) of strictly
positive numbers. Since the j (p) are not all zero (their sum gives 1), this value is strictly
positive.
The first type of regions where we shall integrate differential forms on M will be singular
simplexes and singular chains, which we define in the following two subsections.
77
8.4 Simplexes
Definition 8.4.1: Geometrically independent points
Figure 18: Examples of geometrically independent points (p0 , p1 , p2 in part a)) and geometrically
dependent points (q0 , q1 , q2 in part b))
The numbers Pcj ’s are called barycentric coordinates of the point p ∈ σr , since the point
p, given by rj=0 cj pj represents the baricenter of a system of masses c0 , c1 , . . . , cr placed
in the points p0 , p1 , . . . , pr .
78
Definition 8.4.3: Standard r-simplex
A special case is the standard n-simplex in R I n , σ̄n , which is the simplex σ̄n :=
(p0 , p1 , . . . , pn ), with p0 := (0, 0, . . . , 0) and pi := (0, 0, . . . , 0, 1, 0, . . . , 0) where 1 appears
in the i−th position. In this case, formula (8.4.1) specializes as
n
X
σ̄n := p := (x1 , . . . , xn ) ∈ R
I n| xj ≤ 1, xj ≥ 0 . (8.4.2)
j=1
The standard 0− simplex is a point, while the standard 1− simplex is the segment [0, 1] in R
I.
I 2 , while the tetrahedron
The triangle in part c) of Figure 6.1 19 is the standard 2−simplex in R
in part d) in the same figure is the standard 3−simplex in RI 3.
Figure 19: Standard 0, 1, 2, and 3 simplexes in parts a), b), c) and d), respectively.
Consider now a subset of q + 1 points of the r + 1 points {p0 , p1 , . . . , pr }, and notice (using
the criterion on the linear independence of the position vectors) that since {p0 , p1 , . . . , pr } are
geometrically independent, they also are geometrically independent.
Definition 8.4.4: q-face
Therefore, they identify a q-simplex σq , which is called a q− face of the simplex σr , and
we denote σq ≤ σr . A q-face σq , is the subset of σr where r − q baricentric coordinates
are set to zero.
79
Definition 8.4.5: Orientation
An orientation can be defined on a simplex by choosing an order for the points
{p0 , p1 , . . . , pr }. For example, for a 1-simplex, the simplex (p0 , p1 ) as an oriented sim-
plex is different from (p1 , p0 ).a A direction is defined on the segment [p0 , p1 ] going from p0
to p1 or viceversa. Formally, consider ordered (r + 1)−ples of the points {p0 , p1 , . . . , pr }.
We say that two ordered (r + 1)−ples have the same orientation if it is possible to go
from one to the other via an even permutation. Otherwise, we say that they have oppo-
site orientation. An oriented r-simplex is defined as an equivalence class (r + 1)-ples
according to this equivalence relation. Two oriented r-simplexes σr1 and σr2 , which only
differ by the orientation are formally in the relation σr2 = −σr1 .
a
Following standard notation we use round brackets as in (p0 , p1 , . . . , pr ) to denote oriented simplexes.
where p̂i means that pi is omitted. Notice that each element in the sum in the right-hand
side of (8.4.5) is an (oriented) r − 1 face of the simplex σr , hence the name ‘boundary’
operator. We shall sometimes denote for brevity (p0 , p1 , . . . , p̂i , . . . , pr ) := σri , i.e., the
i−th (r − 1)-face of σr .
∂r ◦ ∂r+1 = 0. (8.4.6)
80
Remark 8.4.1: Parallel with Exterior Derivative
Notice the similarity of property (8.4.6) with the property (7.3.3). We shall take this
parallel much further in the following.
Let σ̄r be the standard r−simplex, and sr : σ̄r → M be a smooth map with image in
a manifold M . sr is assumed to be smooth on an open set containing σ̄r in R I r . sr is
called a singular r−simplex in M . For r = 0, sr maps the 0−simplex to a point in M .
In general, notice there is no requirement on sr beside smoothness. In particular it does
not have to be one to one, hence the word ‘singular’. Let {sr,i } be a set of such singular
r-simplexes. A (singular) r−chain c is a formal linear combination of the sr,i ’s, with
coefficients in R
I , i.e., X
c := ai sr,i . (8.5.1)
i
These are formal linear combinations of smooth maps. They form a vector space under
the operation, which associates to c1 + c2 the chain with coefficients, which are sums of the
coefficients corresponding to the same simplex. This vector space is denoted by Cr (M )
and called the r-chain group.
More precisely, we define, for i = 0, 1, . . . , r, functions fri , that are maps σ̄r−1 → σ̄r , mapping
σ̄r−1 to the i−th face of σ̄r , the one obtained by omitting pi . They are defined as
r−1
X
fr0 (x1 , x2 , . . . , xr−1 ) := (1 − xj , x1 , x2 , . . . , xr−1 ), (8.5.3)
j=1
For example, for the standard simplexes σ̄1 and σ̄2 (cf. b) and c) in Figure 19) with t ∈ [0, 1],
we have
f20 (t) = (1 − t, t), f21 (t) = (0, t), f22 (t) = (t, 0).
81
Then sir is defined as
sir := sr ◦ fri , (8.5.6)
and it is an (r − 1)-singular simplex.
Definition 8.5.2: Boundary Operator
which is a chain of singular (r−1)−simplexes with the induced orientation. This operation
extends to r-chains by linearity, i.e., as
X X
∂ aj sr,j := aj ∂sr,j , (8.5.8)
j j
Figure 20: An example of a singular complex s2 and its boundary ∂s2 = s02 − s12 + s22 and its
image. Notice we have a consistent direction on the boundary.
82
Theorem 8.5.1: Boundary Operator Property
∂ 2 = 0. (8.5.9)
cr = ∂cr+1 , (8.5.10)
is called an r-boundary. An element cr such that ∂cr = 0 is called an r-cycle. That is,
r-cycles span the kernel of the linear operator ∂ on Cr (M ).
Both boundaries and cycles form vector subspaces of Cr (M ), which are denoted respec-
tively by Br (M ) and Zr (M ). Moreover, from the property (8.5.9) it follows that if an
r-chain is a boundary, then it is also an r-cycle, that is,
Br (M ) ⊆ Zr (M ). (8.5.11)
Hr (M ) = Zr (M )/Br (M ). (8.5.12)
The situation is exactly parallel to the one discussed in subsection 7.3.2. The spaces C0 (M ),
C1 (M ),...,Cm (M ) form a ‘chain complex’ of vector spaces with Cl (M ) = 0 for l > m, i.e.,
C0 (M ) ← C1 (M ) ← · · · ← Cm (M ), (8.5.13)
with the boundary operator ∂ mapping in the direction of the arrow, and ∂ 2 = 0. Analogously,
we have the (co-)chain complex
Ω0 (M ) → Ω1 (M ) → · · · → Ωm (M ) (8.5.14)
where Ωl = 0 for l > m, and the exterior derivative d, which maps in the direction of the
arrows and d2 = 0. Notice the correspondence between boundaries and exact forms and the
correspondence between cycles and closed forms, which is reflected in the notations Zr (M ) ↔
Z r (M ), Br (M ) ↔ B r (M ), Hr (M ) ↔ H r (M ). The result that links homology and co-homology
is Stokes theorem.
83
8.6 Exercises
Exercise 8.1 Prove that the real projective space RP n in subsection 1.1 is orientable for every
n.
84
9 Integration of differential forms on manifolds part II: Stokes
theorem
9.1 Integration of differential r-forms over r−chains; Stokes theorem
Consider now a differential r-form ω defined on M , and a chain c ∈ Cr (M ). We want to define
the integral of ω on c.
Definition 9.1.1: Integral of r-form on a singular r-simplex
where σ̄r s∗r ω is the standard Riemann integral on R I r calculated on σ̄r , and s∗r ω is an
R
r
r − f orm in R I . In general, for r ≥ 1, for an r−form ξ := adx1 ∧ · · · ∧ dxr on A ⊆ R I r,
we define Z Z
1 r
adx ∧ · · · ∧ dx := adx1 · · · dxr , (9.1.2)
A A
r
which is the standard Riemann integral in R
I . For r = 0 and for a 0−form, i.e., a function
at a point p, the definition is Z
ω := ω(p). (9.1.3)
p
Notice that the integral in (9.1.1) depends, in general, not just on the image of sr , but on sr
itself. The following simple examples clarify this fact as well as the meaning of the definition.
Example 9.1.1: Multivariable Calculus
However, the integrand on the right hand side is nothing but c∗ ω as it can be easily
verified (just calculate c∗ ω( dt
d
) using the definition).
85
and Z Z 1 Z Z 1
ω := 2dt = 2 6= ω := −2dt = −2
s1 0 f1 0
Notice that the integral can be seen as an inner product between Ωr (M ) and C r (M ),
hω, ci. Stokes theorem says that the boundary operator ∂ is the adjoint of the exterior
derivative, i.e.,
h∂c, ωi = hc, dωi. (9.1.7)
Proof. (We mostly follow the proof of Theorem 6.1 in Nakahara and Theorem 4.7 in
Warner ) By linearity, it is enough to prove (9.1.6) on simplexes, i.e.,
Z Z
dω = ω, (9.1.8)
sr ∂sr
with ω ∈ Ωr−1 (M ).
Moreover, we observe the following general Lemma .
86
Lemma 9.1.1: Commutativity of Exterior Derivative and Pull-back
f ∗ dω = df ∗ ω, (9.1.9)
Using the fact that the property is true on functions, we can write the right hand
side as
The second term of this sum is zero if we distribute the d according to (7.3.2) and
use recursively d2 = 0. So we are left with
df ∗ ω = f ∗ dg ∧ f ∗ dx1 ∧ · · · ∧ f ∗ dxr−1 ,
So, we are only left with proving the result for zero forms ω. Let’s examine how
f ∗ dω acts on a vector field X on M , or, if we consider at a point p on a tangent
vector at that point. We have
In all the previous steps, we used the definitions. Since X is arbitrary, we have
equality (9.1.9) for zero forms, and the Lemma is proved.
87
Using this Lemma, the left hand side of (9.1.8) is
Z Z
dω = d(s∗r ω). (9.1.12)
sr σ̄r
Since Z Z Z Z
(sir )∗ ω = (sr ◦ fri )∗ ω = (fri )∗ (s∗r ω) := s∗r ω
σ̄r−1 σ̄r−1 σ̄r−1 fri
from the definition (9.1.3). Thus, the theorem follows from the fundamental theorem of
calculus. Therefore, we now prove (9.1.15) for r ≥ 2.
By linearity, again we can assume ψ = a(x1 , ...., xr )dx1 ∧·∧dxr−1 . Calculate the integrand
of the left hand side of (9.1.15)
∂a µ ∂a
dψ = µ
dx ∧ dx1 ∧ · · · dxr−1 = (−1)r−1 r dx1 ∧ · · · ∧ dxr . (9.1.17)
∂x ∂x
Therefore, the left hand side of (9.1.15) is (from Calculus)
Z Z
r−1 ∂a 1 2
dψ = (−1) r
dx dx · · · dxr = (9.1.18)
σ̄r σ̄r ∂x
Pr−1
Z Z 1− xj
r−1 1 2 r−1
j=1 ∂a r
(−1) dx dx · · · dx dx ,
Ar 0 ∂xr
88
where Ar is the set
r−1
X
1 r−1 r−1 j
Ar := {(x , . . . , x )∈ R
I | x ≥ 0, xj ≤ 1},
j=1
which is the standard (r − 1)−simplex, i.e., Ar := σ̄r−1 . Using the fundamental theorem
of calculus on the inside integral in (9.1.18), we get
Z
dψ = (9.1.19)
σ̄r
Z r−1
X
(−1)r−1 a(x1 , x2 , . . . , xr−1 , 1 − xj ) − a(x1 , x2 , . . . , xr−1 , 0) dx1 dx2 · · · dxr−1 =
σ̄r−1 j=1
(9.1.20)
Z r−1
X
r−1 1 2 r−1
(−1) a(x , x , . . . , x ,1 − xj )dx1 dx2 · · · dxr−1 −
σ̄r−1 j=1
Z
(−1)r−1 a(x1 , x2 , . . . , xr−1 , 0)dx1 dx2 · · · dxr−1 .
σ̄r−1
(fri )∗ (adx1 ∧ · · · dxr−1 ) = ((fri )∗ a)((fri )∗ dx1 ) ∧ · · · ∧ ((fri )∗ dxi ) ∧ · · · ∧ ((fri )∗ dxr−1 ).
89
As for the first term on the right hand side of (9.1.23), we have
Z Z
1 r 1
a(x , ...., x )dx ∧ · ∧ dx r−1
= (fr0 )∗ a((fr0 )∗ dx1 ) ∧ ((fr0 )∗ dx2 ) ∧ · · · ((fr0 )∗ dxr−1 ).
fr0 σ̄r−1
(9.1.24)
Using the definition (8.5.3), for j = 1, 2, ..., r − 1, we have
r−1
X
(fr0 )∗ dx1 = d(x1 ◦ fr0 ) = − dxk ,
k=1
and
(fr0 )∗ dxj = d(xj ◦ fr0 ) = dxj−1 , for j = 2, ..., r − 1.
Therefore,
r−1
!
X
((fr0 )∗ dx1 ) ∧ ((fr0 )∗ dx2 ) ∧ · · · ((fr0 )∗ dxr−1 ) = − dxk ∧ dx1 ∧ dx2 ∧ · · · ∧ dxr−2 =
k=1
(9.1.25)
−dxr−1 ∧ dx1 ∧ dx2 ∧ · · · ∧ dxr−2 = (−1)r−1 dx1 ∧ dx2 ∧ · · · ∧ dxr−2 ∧ dxr−1 .
Since
r−1
X
((fr0 )∗ a)(x1 , x2 , ..., xr−1 ) = a 1 − xj , x1 , ..., xr−1 ,
j=1
from (8.5.3), we have that the first term on the right hand side of (9.1.23) is equal to
Z r−1
X
(−1)r−1 a 1 − xj , x1 , ..., xr−1 dx1 dx2 · · · dxr−1 . (9.1.26)
σ̄r−1 j=1
90
Z
r−1
−(−1) a(x1 , x2 , . . . , xr−1 , xr )dx1 · · · dxr−1 ,
σ̄r−1
9.2 Integration of Differential forms on regular domains and the second ver-
sion of Stokes’ Theorem
9.2.1 Regular Domains
I m |xm ≥ 0}.
The hyper-plane H m is defined as H m := {{x1 , x2 , ..., xm } ∈ R
A regular domain D ⊆ M is a closed subset of M , with non empty interior, such that
for every point p ∈ ∂D there exists a chart (U, φ), with p ∈ U and φ(U ∩ D) = φ(U ) ∩ H m
(cf. Figura 21).
91
Now we want to give ∂D the structure of a differentiable manifold by displaying an atlas of
compatible charts.
The set Ũ := U ∩ ∂D is open in ∂D since ∂D is endowed with the subset topology, and it is
mapped homeomorphically by φ̃, the restriction of φ to U ∩ ∂D to (φ(U ) ∩ {x ∈ R I m |xm = 0}) ⊆
I m−1 . Therefore, (Ũ , φ̃) is a coordinate chart for ∂D at the point p. Since, by definition, such
R
a coordinate chart exists for every p ∈ ∂D, we have an atlas for ∂D as long as the compatibility
of two overlapping charts is verified. We have:
Proposition 9.2.1: Structure on ∂M
1. The above coordinate charts (Ũ , φ̃) defined as Ũ := U ∩∂D, φ̃ := φŨ are compatible.
2. If two overlapping charts (U, φ), (V, ψ) have compatible orientations on M and D,
then the induced charts (Ũ , φ̃), (Ṽ , ψ̃) have compatible orientations.
Proof. Consider two overlapping charts (Ũ , φ̃) and (Ṽ , ψ̃) and the corresponding charts on
M (U, φ) and (V, ψ) at a point p, so that φ(p) = (φ̃(p), 0) := (a1 , a2 , . . . , am−1 , 0). Denote
also by ỹ j , x̃j , j = 1, . . . , m − 1, the coordinate functions associated with the maps ψ̃
and φ̃, respectively and by y j , xj , j = 1, . . . , m the coordinate functions associated with
the map ψ and φ, respectively. Consider ψ̃ ◦ φ̃−1 : (x̃1 , . . . , x̃m−1 ) → (ỹ 1 , . . . , ỹ m−1 ). For
j = 1, 2, . . . , m − 1, we have,
ỹ j (x̃1 , x̃2 , . . . , x̃m−1 ) = y j (x̃1 , x̃2 , . . . , x̃m−1 , 0). (9.2.1)
Therefore, smoothness of the y j functions implies smoothness of the ỹ j functions, and this
gives the compatibility of the two charts (Ũ , φ̃) and (Ṽ , ψ̃). Moreover from (9.2.1), we
have.
∂ ỹ j ∂y j
|φ̃(p) = | , j = 1, 2, . . . , m − 1, k = 1, 2, . . . , m − 1. (9.2.2)
∂ x̃ k ∂xk φ(p)
m
Examine now ∂y ∂xk
, for k = 1, 2, . . . , m − 1. Since (locally) φ−1 maps points in {x ∈
m m
I |x = 0} to ∂D, and ψ maps points on ∂D to {y ∈ R
R I m |y m = 0}, we have (in a
neighborhood of φ(p)) that y m as a funcion of x1 , . . . , xm−1 is constant and equal to zero
which gives
∂y m
| = 0, (9.2.3)
∂xk φ(p)
for k = 1, 2, . . . , m − 1. Moreover
∂y m
| > 0, (9.2.4)
∂xm φ(p)
because an increase of xm , that is the point moves to the upper hyperspace induces an
increase in y m as the image of the point also moves in the upper hemisphere.
∂y j
The Jacobian ∂x k |φ(p) , j = 1, 2, . . . , m, k = 1, 2, . . . , m has the form
∂ ỹ r
!
∂y j ∂x s |φ̃(p) ∗
| = ∂y m , (9.2.5)
∂xk φ(p) 0 ∂xm |φ(p)
92
and it has positive determinant because the two charts on M (U, φ) and (V, ψ) have the
∂ ỹ j
same orientation. Since (9.2.4) holds, the Jacobian ∂x k |φ̃(p) also has positive determinant.
Therefore, the two charts (Ũ , φ̃) and (Ṽ , ψ̃) also have compatible orientation. Then ∂D is
orientable.
∂ ∂
ω 1
, ..., m > 0,
∂x ∂x
then the orientation at q, for Tq M , is chosen so that the ordered basis { ∂y∂ 1 , . . . , ∂y∂m } for coor-
dinates y 1 , ..., y m at q, is6
∂ ∂
ω , ..., m > 0.
∂y 1 ∂y
Assume now, an orientation on M , and therefore on a regular domain D in M , is given,
and consider a point p ∈ ∂D. Let the oriented basis of Tp M be given as [~n, ~e1 , . . . , ~em−1 ] so
that [~e1 , . . . , ~em−1 ] is an ordered basis of Tp ∂D and ~n is an outward vector, that is a vector
corresponding to a curve c = c(t) such that c(t) ∈ M − D for 0 < t < for some > 0.7 The
orientation on ∂D is by convention chosen as [~e1 , . . . , ~em−1 ]. So, in summary, the orientation
of ∂D is chosen by taking (at a point p ∈ ∂D) a basis of Tp (∂D), [~e1 , . . . , ~em−1 ], so that the
oriented basis [~n, ~e1 , . . . , ~em−1 ] has the original orientation in Tp M .
6
Notice that since we have an orientation compatible atlas this statement is independent of the coordinate
chosen. If we consider coordinates (z 1 , z 2 , ..., z m ) and an ordered basis of Tq M , { ∂z∂ 1 , ..., ∂z∂m }, we have with
g(q) : ω ∂y∂ 1 , ..., ∂y∂m > 0
∂y j
∂ ∂ ∂ ∂
ω( , ..., m ) = g(q)dy 1 ∧ · · · ∧ dy m ( 1 , ..., m ) = g(q) det > 0.
∂z 1 ∂z ∂z ∂z ∂z k
Analogously, it does not depend depend on the choice of coordinates in p.
7
Usually the vector ~n is chosen perpendicular to ~e1 , . . . , ~em in cases where we have an inner product defined
on the tangent space as for Riemannian manifolds.
93
Figure 22: Induced orientation on the boundary of a regular domain. [~n, e1 ] is the same orienta-
tion class as [~e1 , ~e2 ].
Let us carry out this program for H n , where the boundary is R I n−1 . The orientation for
H n is always chosen as [~e1 , ~e2 , . . . , ~en ], given by the standard basis in the standard order.
For n = 2, ~n = −~j, and a basis for ∂H n = R I 1 could be chosen as ±~i. We choose the sign
+ because [−~j,~i] = [~i, ~j] coincides with the standard basis in H 2 . So the induced basis on
RI 1 coincides in this case with the usual basis ~i. Consider now n = 3. In this case ~n = −~k.
We could choose as orientation of ∂H 3 = R I 2 , [~i, ~j] or [~j,~i]. We choose [~j,~i], because
[−~k, ~j,~i] = [~j, ~k,~i] = −[~j,~i, ~k] = [~i, ~j, ~k]. (9.2.6)
Therefore, in this case, the induced orientation is the opposite of the standard orientation
on RI n . Extending this reasoning, one can prove that in general the induced orientation
on ∂H n coincides with the standard orientation for R I n−1 if n is even, and it is opposite
if n is odd.
To avoid issues concerning convergence and improper integrals, we shall assume that a
differential form ω has compact support, where we recall that the support of ω, supp(ω) is
defined as
supp(ω) = cl{p ∈ M |ω(p) 6= 0}, (9.2.7)
where clA denotes the closure of a set A.
94
Definition 9.2.3: Regular Simplex
Recall that the standard simplex σ̄m := (p0 , p1 , . . . , pm ) has an orientation given by the
standard ordering of the points p0 := (0, 0, ..., 0), p1 = (1, 0, 0, . . . , 0), etc. This coincides with the
standard orientation of R I m in that it determines the standard ordering of the vectors p0~p1 := ~e1 ,
p0~p2 := ~e2 ,...,p0~pm := ~em , which is also an ordering for the basis of the tangent space in R I m at
any point.
Definition 9.2.4: Orientation preserving diffeomorphism
Consider a coordinate cover {Uj } of M , which is also a cover of supp(ω). Since supp(ω)
is compact, we can choose a finite subcover of (supp(ω)) ∩ D, {U1 , U2 , . . . , Uk }, and if U0 :=
M −((suppω)∩D), then {U0 , U1 , U2 , . . . , Uk } is a cover of M . Let {0 , 1 , . . . , k } be a partition of
unity subordinate to the cover {U0 , U1 , U2 , . . . , Uk }. Now, associate to each of the above elements
U1 , U2 , . . . , Uk of the cover {U0 , U1 , U2 , . . . , Uk } a regular m−simplex s1 ,...,sk 8 corresponding to
diffeomorphisms with images s1 (σ̄m ),...,sk (σ̄m ) in D. We assume (without loss of generality as
we can always get smaller open sets) that the open sets U1 , . . . , Uk are small enough so that
either
1)
that is, the integral is defined in terms of the integrals over simplexes as defined in (9.1.1).
8
For simplicity of notation we omit the subscript m, in sm . It is understood that we are dealing with
m−simplexes.
95
Figure 23: Two types of open sets U and V in the cover of supp(ω) ∩ D and the associated
regular simplexes s1 , s2 , respectively.
We now show that this definition is independent of the cover used, simplexes used, and the
partition of unity used. It only depends on ω and D. Let {U0 , U1 , ..., Uk } be the above cover with
associated regular simplexes s1 , ..., sk and partition of unity {0 , 1 , ..., k } as for the definition
(9.2.10) and {V0 , V1 , . . . , Vl } be another cover with associated simplexes f1 , f2 , ..., fl and partition
of unity δ0 , δ1 , ...., δl . Using the latter set-up, the definition (9.2.10) gives
Z l Z
X
ω := δj ω. (9.2.11)
D j=1 fj
To show that (9.2.12) and (9.2.13) are equal, fix i and j in the right hand side of (9.2.12). By
definition we have Z Z
(δj i ω) := s∗i (δj i ω), (9.2.14)
si σ̄m
m
where the right hand side is an integral in R I m , we consider a variable transformation
I . On R
96
s−1
i ◦ fj , which is orientation preserving. We have the following:
Z Z Z
∗
(δj i ω) := si (δj i ω) = s∗i (δj i ω) = (9.2.15)
si σ̄m s−1
i (Vj ∩Ui )
Z Z
(s−1
i ◦ fj )∗ s∗i (δj i ω) = fj∗ (δj i ω) =
fj−1 (Vj ∩Ui ) fj−1 (Vj ∩Ui )
Z Z
fj∗ (δj i ω) := (δj i ω).
σ̄m fj
The second equality is due to the fact that δj i ω is possibly different from zero only on Vj ∩ Ui .
The third equality is an application of the transformation of variables formula for integrals in
I m together with the fact that s−1
R i ◦ fj is assumed to be orientation preserving. The fourth
equality follows directly from properties of the pullback. The fifth equality follows again from
the fact that δj i ω is possibly different from zero only on Vj ∩ Ui .
This definition of the integral on a regular domain allows us to relate the integral of dω on
D with the integral of ω on ∂D, for ω ∈ Ωm−1 (M ), which is the content of the second version of
Stokes theorem.
Proof. We have
k k k k
!
X X X X
d(i ω) = di ∧ ω + i dω = di ∧ω+ i dω.
i=1 i=1 i=1 i=1
Pk
The first term of this expression is zero since i=1 j ≡ 1 and therefore constant on
supp(ω) ∩ D. Moreover,
k k
!
X X
i dω = i dω = dω. (9.2.17)
i=1 i=1
97
Proof. Using Lemma 9.2.1, we write
Z k Z
X k X
X k Z k Z
X
dω = d(i ω) := j d(i ω) = i ω. (9.2.19)
D i=1 D i=1 j=1 sj i=1 ∂si
R
Here the second equality is due to the fact that, as we have seen in the definition, D η for
every m-form η does not depend on the cover of supp(η) ∩ D or on the partition of unity
and simplexes chosen. In our case η := d(i ω) is zero outside of Ui , and we can take as
a cover of supp(η) ∩ D all the Uj ∩ Ui ’s as j = 0, 1, ..., k. We can also take all simplexes
equal to si . Therefore for the independence of the choice of the simplexes, we have
k Z
X k Z
X Z k
X Z
j d(i ω) = j d(i ω) = ( j )d(i ω) = d(i ω).
j=1 sj j=1 si si j=1 si
The third equality in (9.2.19) follows from the first version of Stokes’ theorem.
We now look at Z
i ω,
∂si
in (9.2.19), for a given i. If Ui ⊂ int(si (σ̄m )), that is we are in the situation described by
(9.2.8), we have Z
i ω = 0, (9.2.20)
∂si
as desired.
98
9.3 De Rham Theorem and Poincare’ Lemma
Definition 9.3.1: Inner product on Cr (M ) × Ωr (M )
The integral of a differential form over an r−chain, determines an inner product (·, ·) :
Cr (M ) × Ωr (M ) → RI , defined as
Z
(c, ω) := ω, c ∈ Cr (M ), ω ∈ Ωr (M ). (9.3.1)
c
Elementary properties of the integral prove that this is indeed an inner product, i.e., linear
in both c and ω. Stokes theorem is a statement about the fact that the operators ∂ and d are
adjoint of each other with respect to this inner product. It reads for c ∈ Cr (M ), ω ∈ Ωr−1 (M )
as Z Z
(∂c, ω) := ω = dω := (c, dω). (9.3.2)
∂c c
We have to check that this definition is well posed, i.e., independent of the representatives c
and ω. Let us choose a different representative for [c], c + ∂c1 . We have
Z Z Z Z Z Z
ω= ω+ ω= ω+ dω = ω, (9.3.4)
c+∂c1 c ∂c1 c c1 c
where we used Stokes’ theorem and in the last equality the fact that ω is a closed form so that
dω = 0. Also we have, for ψ ∈ Ωr−1 (M )
Z Z Z Z Z Z
ω + dψ = ω + dψ = ω + ψ = ω, (9.3.5)
c c c c ∂c c
where again we used Stokes’ theorem and the fact that c is a cycle so that ∂c = 0.
If we fix [ω] ∈ H r (M ), Λ(·, [ω]) is a linear map Hr (M ) → R
I and therefore, an element of
the dual space (Hr (m))∗ . Therefore, we have a linear map λ : H r (M ) → (Hr (M ))∗ , which
associates to [ω], the element Λ(·, [ω]). Is this map an isomorphism? This is the content of De
Rham Theorem:
Theorem 9.3.1: De Rham Theorem
If M is a compact manifold, then H r (M ) and Hr (M ) are finite dimensional and the above
map λ : [ω] → Λ(·, [ω]) is an isomorphism λ : H r (M ) → (Hr (M ))∗ .
Therefore, under the assumptions of the theorem, Hr (M ) and H r (M ) are dual of each other,
and information on one can be obtained from information on the other. Moreover, the inner
product Λ under the assumptions of the theorem is nondegenerate (see Exercise 9.3).
99
9.3.1 Consequences of the De Rham Theorem
As a consequence of the De Rham theorem the Betti numbers, br (M ) := dim Hr (M ) are equal
to dim(H r (M )). Moreover, if [c1 ], [c2 ],...,[ck ], is a basis of Hr (M ), then ω ∈ Z r (M ) is exact if
and only if Z
(cj , ω) := ω = 0, ∀j = 1, 2, ..., k. (9.3.6)
cj
In fact, this would mean that λ([ω]) is the zero map in (Hr (M ))∗ , but since λ is an isomorphism,
[ω] must be zero, that is ω is an exact form. This means that
we can check that a form is exact by calculating its integral over a finite number of chains.
Given a basis {[c1 ], [c2 ], ..., [ck ]} of Hr (M ) we can use the non-degenerate inner product Λ to
find a dual basis {[ω1 ], ..., [ωk ]} for H r (M ), which satisfies the requirement
To see this, just take any basis {[ω̃l ]} of H r (M ) and for fixed j, solve the equations
br
X br
X
Λ([cs ], xl [ω̃l ]) = xl Λ([cs ], [ω̃l ]) = δs,j , s = 1, 2, . . . br , (9.3.8)
l=1 l=1
100
9.4 Exercises
Exercise 9.1 Using the linearity of the Riemann integral, prove that the integral defined in 9.1.1
is also linear.
Exercise 9.3 Prove the inner product Λ is nondegenerate if the manifold is compact.
101
10 Lie groups Part I; Basic Concepts
10.1 Basic Definitions and Examples
Definition 10.1.1: Lie Groups
Lie group G is a differentiable manifold which is also group and such that the group
operations, product, · : G × G → G, (g1 , g2 ) → g1 · g2 , and inversion −1 , G → G, g → g −1 ,
are C ∞ (here G × G is the product manifold of G with itself). The dimension of the Lie
group is its dimension as a manifold.
The simplest example of a Lie group is R I ∗ := R I − {0} with the multiplication operation,
which is a manifold with two connected components and a Lie group since the operations of
multiplication and inversion are C ∞ . The connected component R I + of positive real numbers
∗
is also a Lie group, with the same operations of R I . It has obviously dimension 1. A general-
∗
ization of RI is the general linear group Gl(n, R I ) of n × n real matrices with determinant
different from zero, where the product and the inversions are the standard product of matri-
ces and inversion of a matrix. As a manifold, Gl(n, R I ) has the subset topology derived from
2 2
the topology of Mn,n ( RI) ≡ R I n and in fact Gl(n, R I ) is the same as R I n with the closed set
{A ∈ Mn,n ( R I ) | det(A) = 0} removed. Analogously to R I + , Gl+ (n, RI ) is the Lie group of
nonsingular matrices with positive determinant. The general linear group Gl(n, CI ) is the
Lie group of nonsingular n × n matrices with complex entries, with the product and inversion
operation of a matrix. This is a 2n2 dimensional Lie group, the 2n2 real parameters give the
coordinate functions of a (global) chart.
All Lie groups which are subgroups (in the algebraic sense) of Gl(n, R I ) or Gl(n, CI ) are
called matrix Lie groups or linear Lie groups. They give some of the most important
examples of Lie groups. Among these, SL(n, R) (SL(n, C)), the special Linear group, is
the Lie group of matrices in Gl(n, R I ) (Gl(n, CI )) with determinant equal to 1. The Lie group
O(n), the orthogonal group, is the Lie group of n × n orthogonal matrices in Gl(n, R I ), i.e.,
A ∈ O(n) ⇔ AT A = AAT = 1; SO(n), the special orthogonal group is the Lie group of
matrices in O(n) with determinant equal to 1. The Lie group U (n), the unitary group, is the
Lie group (subgroup of Gl(n, C I )) of n × n unitary matrices, i.e., U ∈ U (n) ⇔ U † U = U U † = 1.
The Lie group SU (n), the special unitary group, is the Lie group of matrices in U (n) with
determinant equal to one.
O(n) and U (n) are Lie groups defined by the fact that they preserve some inner product; For
example if and only if A ∈ O(n), (A~x)T (A~x) = ~xT AT A~x = ~xT ~x, for every ~x ∈ Rn . More in
general, one can consider Lie groups which preserve a quadratic form, not necessarily positive
definite. Consider the matrix 1p,q := diag(1p , −1q ), where 1s is the s × s identity matrices. This
defines a quadratic form ~xT 1p,q ~x := x21 + x22 + · · · + x2p − x2p+1 − x2p+2 − · · · − x2p+q . Then the Lie
O(p, q) is the Lie group of all matrices A such that
AT 1p,q A = 1p,q .
A particularly important case is O(1, 3), the Lorentz group where 11,3 represents the Minkowski
I 4.
metric in time-space R
102
10.2 Lie subgroups and coset spaces
Definition 10.2.1: Lie subgroups
A Lie subgroup H of a Lie group G is a subgroup of G (in the algebraic sense) which is
also a Lie group and a submanifold of G
Notice in particular that since the inclusion map i : H → G is supposed to be smooth and
therefore continuous and open, the topology on H has to coincide with the subset topology of
H as a subset of G. This observation gives rise to examples of Lie groups which are subgroups
but not Lie subgroups. The following example is an instance of this fact.
Example Let T 2 be the 2-dimensional torus defined as T 2 := S 1 × S 1 , which is a Lie group
with the group operation (eiθ1 , eiθ2 ) · (eiθ3 , eiθ4 ) := (ei(θ1 +θ3 ) , ei(θ2 +θ4 ) ). Consider now the Lie
group Hλ := {P ∈ T 2 |P = (eit , eiλt ), t ∈ R
I } which is a subgroup of T 2 and it is a Lie group with
2
the operation inherited from T . The topology on H is such that H is locally homeomoprhic to R I.
Assume however that λ is an irrational number. Then this topology does not coincide with the
subset topology. In particular, consider the open set N in H, N := {(eit , eiλt ) ∈ H|t ∈ (−, )}.
If we try to realize N as H ∩ V where V is an open set in T 2 we should be able to find a δ
sufficiently small so that H ∩ Iδ ⊆ N , where Iδ : {(eiγ , eiψ ) ∈ G | γ, ψ ∈ (−δ, δ)}. In fact, any
open set V in G is the union of Iδ ’s, therefore N = H ∩ V = H ∩ (∪δIδ ) implies that there exists
a δ such that
H ∩ Iδ ⊆ N . (10.2.1)
Let us have a closer look at H ∩ Iδ . (eiγ , eiψ ) is in H, if and only if, for some real x there
exist integers k and m such that
Consider now the coordinate chart φ on T 2 which maps Iδ to the box (−δ, δ) × (−δ, δ) ⊆ R2 .
We should have, assuming sufficiently small, from (10.2.1)
However φ(N ) is a single segment in (−δ, δ) × (−δ, δ) while φ(H ∩ Iδ ) is made up of segments
of lines (obtained by solving for x in (10.2.2)) ψ = λ(γ − 2kπ) + 2mπ all with the same slope λ
and with intersection with the ψ axis
δ
|m − λk| < ? (10.2.5)
2π
There are in fact an infinite numbers of them because λ is irrational. This is a consequence
of Dirichlet’s approximation theorem [2] .This says that for every real number λ and positive
103
integer N , there exists two integers k and m, with 1 ≤ k ≤ N such that |kλ − m| < N1 . Now
assume that there was only a finite number of pairs and take the pair (m, k) which gives the
minimum of |m − λk|. Since λ is irrational such a minimum, call it min cannot be zero. Then
choose an N such that N1 < min and we can choose (k, m) so that |m − λk| < min which is
a contradiction. Therefore no matter what δ is there are an infinite number of values of ψm,k
which is not compatible with (10.2.3) .
The following important theorem gives a condition for a subgroup H of a Lie group G to be
a Lie subgroup.
Theorem 10.2.1: Closed subgroup theorem
A Lie group H which is a subgroup of a Lie group G and it is closed in the subset topology,
is a Lie subgroup.
This theorem in particular shows that the matrix Lie groups SL(n, R I ), SO(n) and so on
are Lie subgroups of Gl(n, R I ) as they are defined as the inverse images of a smooth map from
Gl(n, RI ) to a closed set of RI m , for appropriate m. For example, SL(n, R I ) is the inverse immage
of the set {1} ∈ R I under the function det; SO(n) is defined as the inverse image of the identity
under the function that maps elements A in Gl(n, R I ) to the n2 entries of AT A and det(A) = 1,
n2 +1
a closed set in R I . Analogous discussion can be applied to SU (n), etc, as Lie subgroups of
Gl(n, CI ).
If H is a Lie subgroup of G, we can define an equivalence relation on elements of G, by saying
0
the g ∼ g 0 if and only if there exists h ∈ H with g = gh. The space of all equivalence classes
is called the (left)9 coset space, and any equivalence class is called a (left) coset. The coset
space G/H is a manifold of dimension dim(G) − dim(H) (see subsection ?? below). It is also a
group if H is a normal subgroup, that is, ghg −1 ∈ H for every h ∈ H and g ∈ G. In this case
the product operation can be defined between two equivalence classes [g1 ][g2 ] := [g1 g2 ] which is
independent of the representative chosen since
where h̃ := gh−1 g −1 ∈ H if h ∈ H.
104
Definition 10.3.1: (Invariant vector field)
A vector field on G is said to be left invariant (right invariant) if, for every g ∈ G
La∗ Xg = XLa (g) = XLa (g) , (Ra∗ Xg = XRa (g) = Xga ), (10.3.1)
From now on we shall deal only with left invariant vector fields as the theory for right invariant
vector fields can be obtained analogously.
Example 10.1. In a quest to understand how left invariant vector fields look like , we first
I ∗ with the multiplication as the group operation. We know this is a
consider the Lie group R
one dimensional manifold so in the standard coordinate, x, a vector field will look like
d
X=f , (10.3.3)
dx
d
where f is a function of the point x. Now the vector field La∗ X at ax is written as h dx where h
is La∗ X applied to the coordinate function φ. That is,
d d
h = [La∗ X(φ)] ◦ La := f (x) (φ ◦ La ) = f (x) (ax) = f (x)a.
dx dx
However, since La∗ X = X, we must have f (ax) = h. Therefore f (ax) = af (x), i.e., f is a linear
function and the left invariant vector field (10.3.3) are of the form
d
X = ax . (10.3.4)
dx
We shall generalize this example to Gl(n, R
I ) later.
Example 10.2. As another simple example consider R I 2 but with the Lie group operation +,
(a1 , a2 ) + (b1 , b2 ) = (a1 + b1 , a2 + b2 ) (Notice this group is commutative so left translation and
right translation coincide). Let a := (a1 , a2 ). A general vector field, in the standard (Cartesian)
coordinates will have the form
∂ ∂
X=f +g , (10.3.5)
∂x ∂y
∂ ∂
for smooth functions f = f (x, y) and g = g(x, y). Write La∗ X as La∗ X := h1 (x, y) ∂x +h2 (x, y) ∂y
and recall that h1 is La∗ X applied to the first coordinate function (say φ) and h2 is La∗ X applied
to the second coordinate function (say ψ). That is, h1 (x + a1 , y + a2 ) = (La∗ X)|(x+a1 ,y+a2 ) φ,
h2 (x + a1 , y + a2 ) = (La∗ X)|(x+a1 ,y+a2 ) ψ.
In particular
∂ ∂
(La∗ X)x+a1 ,y+a2 φ = Xx,y (φ ◦ La ) = f (x, y) (x + a1 ) + g(x, y) (x + a1 ) = f (x, y). (10.3.6)
∂x ∂y
105
Recall the fact that φ ◦ La = x + a1 so that the derivative with respect to x gives 1. Analogously,
applying La∗ X to the second component, ψ, we get h2 (x + a1 , y + a2 ) = g(x, y). Therefore
∂ ∂
(La∗ X)|(x+a1 ,y+a2 ) = f (x, y) |(x+a1 ,y+a2 ) + g(x, y) |(x+a1 ,y+a2 ) . (10.3.7)
∂x ∂y
On the other hand
∂ ∂
X(x+a1 ,y+a2 ) = f (x + a1 , y + a2 ) |(x+a1 ,y+a2 ) + g(x + a1 , y + a2 ) |(x+a1 ,y+a2 ) . (10.3.8)
∂x ∂y
Since (10.3.7) and (10.3.8) must be equal because of invariance, we have that f and g must be
constants in (10.3.5)
∂ ∂2 ∂ ∂2
axb + abx2 2 − bxa − bax2 2 = 0.
∂x ∂x ∂x ∂x
We are therefore left wondering whether these features are true in general.
The fact that the set of left invariant vector fields is a vector space follows immediately from
the definition and the fact that it is closed under Lie bracket also follows from the definition.
In fact take two left invariant vector fields X and Y and check the definition of left invariance
for [X, Y ], we have using the distributivity property of the pushforward with respect to the Lie
derivative (see, Exercise 5.6)
We shall denote by g the Lie algebra of left invariant vector fields on G. The fact that the g is
a vector space of dimension equal to the dimension of the manifold G follows from the following
proposition.
Proposition 10.3.1
There exists an isomorphism between the tangent space at the identity Te G and the
space of left invariant vector fields of G. This isomorphism associates to V ∈ Te G the left
106
invariant vector field XV defined by
XV |a = La∗ V. (10.3.10)
Proof. It is clear that the map (10.3.10) is linear. Moreover it gives indeed left invariant vector
fields since for b ∈ G,
Moreover, the map is onto since for any left invariant vector field X the vector V = Xe ∈ Te G
produces the vector field X according to (10.3.10). Moreover, it is one to one since different
vectors in Te G give different left invariant vector fields. In fact, the map XV → V = XV |e
provides the inverse of the isomorphism.
The Lie algebra associated with a Lie group is by definition the Lie algebra of left
invariant vector fields on the Lie group. As we have seen, it is a vector space isomorphic to
the tangent space at the identity. In the following we shall follow the standard notation to
denote by lower case gothic letter(s), g the Lie algebra corresponding to a Lie group G denoted
by the same upper case letter(s). For instance, gl(n, RI ) denotes the Lie algebra associated with
Gl(n, RI ) and so(n) denotes the Lie algebra associated with SO(n).
107
corresponding to [Xi∗V , Xi∗W ] := [Φ(XV ), Φ(XW )] in Te G as for the isomorphism (10.3.10),
that is [Xi∗V , Xi∗W ](e), is exactly the same as the one corresponding to Φ([XV , XW ]), that is,
i∗ ([XV , XW ](e)) as for the isomorphism of (10.3.10) Therefore we have to show that
vector field is left invariant? We determine this here generalizing what we have done in Example
10.1 for R∗ . It is useful, although not necessary, to use the isomorphism (10.3.10) and start with
a general tangent vector at 1n ∈ Gl(n, R I ),
∂
V = V jk |1 . (10.4.1)
∂xjk n
The corresponding element of gl(n, R I ), XV , is given at p ∈ Gl(n, R
I ) by Lp∗ V . The corresponding
functions f jk (p) are given by this vector fields applied to the jk-th component of the coordinate
functions, which we denote by φjk . So we have
∂
f jk (p) = Lp∗ V (φjk ) := V (φjk ◦ Lp ) = V lm |1 (φjk ◦ Lp ) = (10.4.2)
∂xlm n
∂ X X ∂ X X
V lm |1 n pjb x bk
= V lm
|1 n pjb x bk
= V lm
p jb δl
b k
δm = pjl V lk .
∂xlm ∂xlm
b b b l
108
To discover the expression of [XV , XW ] once again we apply [XV , XW ] := XV XW − XW XV to
the coordinate function φrq using (10.4.3) and the result of (10.4.2). We have
∂
XW |p φrq = pik W kj (φrq ) = pik W kj δri δjq = prk W kq , (10.4.4)
∂xij
and analogously
∂
XV |p φrq = pik V kj (φrq ) = pik V kj δri δjq = prk V kq . (10.4.5)
∂xij
Now
∂
XV XW φrq = XV |p (xrk W kq ) = pjl V lm (xrk W kq ) = (10.4.6)
∂xjm
pjl V lm δjr δm
k
W kq = prl V lm W mq ,
which in matrix form gives P Ṽ W̃ . Analogously
∂
XW XV φrq = XW |p (xrk V kq ) = pjl W lm (xrk V kq ) = (10.4.7)
∂xjm
pjl W lm δjr δm
k kq
V = prl W lm V mq ,
which, in matrix form, is P W̃ Ṽ . We obtain the result that [XV , XW ] at the point P ∈ Gl(n, R
I)
can be represented by P Ṽ W̃ − P W̃ Ṽ = P [Ṽ , W̃ ], i.e., by the commutator, in the sense of
matrices.
109
11 Exercises
Exercise 10.1 Find the dimensions of O(n), SO(n), and O(p, q),
110
12 Lie groups Part II
111
13 Fiber bundles; Part I
112
14 Fiber bundles; Part II
113
References
[1] M. Nakahara, Geometry, Topology and Physics (Graduate Student Series in Physics) 2nd
Edition, Taylor and Francis Group, New York, 2003.
[2] W.M., Schmidt, Diophantine approximation, Lecture Notes in Mathematics. 785. Springer,
(1980).
114