52H Text 2014
52H Text 2014
52H Text 2014
and V . . . . . . . . . . . . . . . . . . . . . . 11
1.3 The map /
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11
1.4 Multilinear functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14
1.5 Symmetric bilinear functions and quadratic forms . . . . . . . . . . . . . . . . . . . . 15
2 Tensor and exterior products 17
2.1 Tensor product . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
2.2 Spaces of multilinear functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
2.3 Symmetric and skew-symmetric tensors . . . . . . . . . . . . . . . . . . . . . . . . . 19
2.4 Symmetrization and anti-symmetrization . . . . . . . . . . . . . . . . . . . . . . . . 20
2.5 Exterior product . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21
2.6 Spaces of symmetric and skew-symmetric tensors . . . . . . . . . . . . . . . . . . . . 24
2.7 Operator /
on spaces of tensors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24
3 Orientation and Volume 29
3.1 Orientation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29
3.2 Orthogonal transformations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30
3.3 Determinant and Volume . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
3.4 Volume and Gram matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33
3
4 Dualities 35
4.1 Duality between k-forms and (n k)-forms on a n-dimensional Euclidean space V . 35
4.2 Euclidean structure on the space of exterior forms . . . . . . . . . . . . . . . . . . . 41
4.3 Contraction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
5 Complex vector spaces 49
5.1 Complex numbers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 49
5.2 Complex vector space . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52
5.3 Complex linear maps . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54
II Calculus of dierential forms 57
6 Topological preliminaries 59
6.1 Elements of topology in a vector space . . . . . . . . . . . . . . . . . . . . . . . . . . 59
6.2 Everywhere and nowhere dense sets . . . . . . . . . . . . . . . . . . . . . . . . . . . 61
6.3 Compactness and connectedness . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 61
6.4 Connected and path-connected components . . . . . . . . . . . . . . . . . . . . . . . 64
6.5 Continuous maps and functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65
7 Vector elds and dierential forms 69
7.1 Dierential and gradient . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 69
7.2 Smooth functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 71
7.3 Gradient vector eld . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 71
7.4 Vector elds . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 72
7.5 Dierential forms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 75
7.6 Coordinate description of dierential forms . . . . . . . . . . . . . . . . . . . . . . . 75
7.7 Smooth maps and their dierentials . . . . . . . . . . . . . . . . . . . . . . . . . . . 76
7.8 Operator f
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 77
7.9 Coordinate description of the operator f
. . . . . . . . . . . . . . . . . . . . . . . . 79
7.10 Examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 80
7.11 Pfaan equations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 81
4
8 Exterior dierential 83
8.1 Coordinate denition of the exterior dierential . . . . . . . . . . . . . . . . . . . . . 83
8.2 Properties of the operator d . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 85
8.3 Curvilinear coordinate systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 89
8.4 Geometric denition of the exterior dierential . . . . . . . . . . . . . . . . . . . . . 89
8.5 More about vector elds . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 91
8.6 Case n = 3. Summary of isomorphisms . . . . . . . . . . . . . . . . . . . . . . . . . . 93
8.7 Gradient, curl and divergence of a vector eld . . . . . . . . . . . . . . . . . . . . . . 94
8.8 Example: expressing vector analysis operations in spherical coordinates . . . . . . . 95
8.9 Complex-valued dierential k-forms . . . . . . . . . . . . . . . . . . . . . . . . . . . . 98
9 Integration of dierential forms and functions 101
9.1 Useful technical tools: partition of unity and cut-o functions . . . . . . . . . . . . . 101
9.2 One-dimensional Riemann integral for functions and dierential 1-forms . . . . . . . 104
9.3 Integration of dierential 1-forms along curves . . . . . . . . . . . . . . . . . . . . . . 107
9.4 Integrals of closed and exact dierential 1-forms . . . . . . . . . . . . . . . . . . . . . 112
9.5 Integration of functions over domains in high-dimensional spaces . . . . . . . . . . . 113
9.6 Fubinis Theorem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 125
9.7 Integration of n-forms over domains in n-dimensional space . . . . . . . . . . . . . . 128
9.8 Manifolds and submanifolds . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 135
9.8.1 Manifolds . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 135
9.8.2 Gluing construction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 138
9.8.3 Examples of manifolds . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 144
9.8.4 Submanifolds of an n-dimensional vector space . . . . . . . . . . . . . . . . . 146
9.8.5 Submanifolds with boundary . . . . . . . . . . . . . . . . . . . . . . . . . . . 148
9.9 Tangent spaces and dierential . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 150
9.10 Vector bundles and their homomorphisms . . . . . . . . . . . . . . . . . . . . . . . . 153
9.11 Orientation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 154
9.12 Integration of dierential k-forms over k-dimensional submanifolds . . . . . . . . . . 154
5
III Stokes theorem and its applications 161
10 Stokes theorem 163
10.1 Statement of Stokes theorem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 163
10.2 Proof of Stokes theorem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 167
10.3 Integration of functions over submanifolds . . . . . . . . . . . . . . . . . . . . . . . . 169
10.4 Work and Flux . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 175
10.5 Integral formulas of vector analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . 178
10.6 Expressing div and curl in curvilinear coordinates . . . . . . . . . . . . . . . . . . . . 179
11 Applications of Stokes formula 183
11.1 Integration of closed and exact forms . . . . . . . . . . . . . . . . . . . . . . . . . . . 183
11.2 Approximation of continuous functions by smooth ones . . . . . . . . . . . . . . . . . 184
11.3 Homotopy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 186
11.4 Winding and linking numbers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 192
11.5 Properties of k-forms on k-dimensional manifolds . . . . . . . . . . . . . . . . . . . . 195
6
Part I
Multilinear Algebra
7
Chapter 1
Linear and multilinear functions
1.1 Dual space
Let V be a nite-dimensional real vector space. The set of all linear functions on V will be denoted
by V
.
Proposition 1.1. V
, x V, R
It is straightforward to check that all axioms of a vector space are satised for V
. Let us now
check that dimV = dimV
.
Choose a basis v
1
. . . v
n
of V . For any x V let
_
_
_
_
_
x
1
.
.
.
x
n
_
_
_
_
_
be its coordinates in the basis v
1
. . . v
n
.
Notice that each coordinate x
1
, . . . , x
n
can be viewed as a linear function on V . Indeed,
1) the coordinates of the sum of two vectors are equal to the sum of the corresponding coordinates;
9
2) when a vector is multiplied by a number, its coordinates are being multiplied by the same
number.
Thus x
1
, . . . , x
n
are vectors from the space V
.
Indeed, any linear function l V
.
On the other hand, if a
1
x+. . . +a
n
x
n
is the 0-function, then all the coecients must be equal to 0;
i.e. functions x
1
, . . . , x
n
are linearly independent. Hence x
1
, . . . , x
n
form a basis of V and therefore
dimV
= n = dimV.
The space V
we can
establish an isomorphism i
S
: V V
and V
The space (V
by the formula
x I(x) (V
.
Exercise 1.6. Prove that if V is nite-dimensional then I is an isomorphism. What can go wrong
in the innite-dimensional case?
1.3 The map /
: W
(l) V
by the formula /
(l)(x) = l(A(x)), x V . In
other words, /
(l) = l /.
2
2
In fact, the above formula makes sense in much more general situation. Given any map : X Y between two
sets X and Y the formula
and W
and Y = y
1
, . . . , y
k
which consist of coordinate functions corresponding to the basis B
v
and B
w
.
Let us denote by A
= M
Y X
(/
).
Proposition 1.7. The matrices A and A
= A
T
.
Proof. By the denition of the matrix of a linear map we should take vectors of the basis Y =
y
1
, . . . , y
k
, apply to them the map /
. Set y
i
= /
(y
i
), i = 1, . . . , k. For
any vector u =
n
j=i
u
j
v
j
V, we have y
i
(u) = y
i
(/(u)). The coordinates of the vector /(u) in the
basis w
1
, . . . , w
k
may be obtained by multiplying the matrix A by the column
_
_
_
_
_
u
1
.
.
.
u
n
_
_
_
_
_
. Hence,
y
i
(u) = y
i
(/(u)) =
n
j=1
a
ij
u
j
.
But we also have
n
j=1
a
ij
x
j
(u) =
n
j=1
a
ij
u
j
.
Hence, the linear function y
i
V
has an expansion
n
j=1
a
ij
x
j
in the basis X = x
1
, . . . x
n
of the
space V
equals
_
_
_
_
_
a
i1
.
.
.
a
in
_
_
_
_
_
, so that the whole matrix A
=
_
_
_
_
_
a
11
a
k1
. . .
a
1n
a
kn
_
_
_
_
_
= A
T
.
12
Exercise 1.8. Given a linear map / : V W with a matrix A, nd /
(y
i
).
Answer. The map /
j=1
a
ij
x
j
on V , i.e. to
its expression in coordinates x
i
.
Proposition 1.9. Consider linear maps
U
A
V
B
W .
Then (B /)
= /
.
Proof. For any linear function l W
we have
(B /)
(l)(x) = l(B(/(x)) = /
(B
(l))(x)
for any x U.
Exercise 1.10. Suppose that V is a Euclidean space and / is a linear map V V . Prove that
for any two vectors X, Y V we have
/(X), Y ) = X, T
1
/
. Applying this to Z = /
T(Y )(X). On the other hand, the left-hand side can be rewritten as T(Y )(/(X)).
But /
. Clearly, (/
= /. In any
orthonormal basis the matrices of an operator and its adjoint are transpose to each other. An
13
operator / : V V is called self-adjoint if /
is equal to T
1
/
T = T /.
Remark 1.11. As it follows from Proposition 1.7 and Exercise 1.10 the matrix of a self-adjoint
operator in any orthonormal basis is symmetric. This is not true in an arbitrrary basis.
1.4 Multilinear functions
A function l(X
1
, X
2
, . . . , X
k
) of k vector arguments X
1
, . . . , X
k
V is called k-linear (or multi-
linear) if it is linear with respect to each argument when all other arguments are xed. We say
bilinear instead of 2-linear. Multilinear functions are also called tensors. Sometimes, one may also
say a k-linear form, or simply k-form instead of a k-linear functions. However, we will reserve
the term k-form for a skew-symmetric tensors which will be dened in Section 2.3 below.
If one xes a basis v
1
. . . v
n
in the space V then with each bilinear function f(X, Y ) one can
associate a square n n matrix as follows. Set a
ij
= f(v
i
, v
j
). Then A = (a
ij
)
i,j=1,...,n
is called the
matrix of the function f in the basis v
1
, . . . , v
n
. For any 2 vectors
X =
n
1
x
i
v
i
, Y =
n
1
y
j
v
j
we have
f(X, Y ) = f
_
_
n
i=1
x
i
v
i
,
n
j=1
y
j
v
j
_
_
=
n
i,j=1
x
i
y
j
f(v
i
, v
j
) =
n
i,j=1
a
ij
x
i
y
j
= X
T
AY .
Exercise 1.12. How does the matrix of a bilinear function depend on the choice of a basis?
Answer. The matrices A and
A of the bilinear form f(x, y) in the bases v
1
, . . . , v
n
and v
1
, . . . , v
n
are related by the formula
A = C
T
AC, where C is the matrix of transition from the basis v
1
. . . v
n
14
to the basis v
1
. . . v
n
, i.e the matrix whose columns are the coordinates of the basis v
1
. . . v
n
in the
basis v
1
. . . v
n
.
Similarly, with a k-linear function f(X
1
, . . . , X
k
) on V and a basis v
1
, . . . , v
n
one can associate
a k-dimensional matrix
A = a
i
1
i
2
...i
k
; 1 i
1
, . . . , i
k
n,
where
a
i
1
i
2
...i
k
= f(v
i
1
, . . . , v
i
k
) .
If X
i
=
n
j=1
x
ij
v
j
, i = 1, . . . , k , then
f(X
1
, . . . , X
k
) =
n
i
1
,i
2
,...i
k
=1
a
i
1
i
2
...i
k
x
1i
1
x
2i
2
. . . x
ki
k
,
see Proposition 2.1 below.
1.5 Symmetric bilinear functions and quadratic forms
A function Q : V R on a vector space V is called quadratic if there exists a bilinear function
f(X, Y ) such that
Q(X) = f(X, X), X V. (1.5.1)
One also uses the term quadratic form. The bilinear function f(X, Y ) is not uniquely determined
by the equation (1.5.1). For instance, all the bilinear functions x
1
y
2
, x
2
y
1
and
1
2
(x
1
y
2
+ x
2
y
1
) on
R
2
dene the same quadratic form x
1
x
2
.
On the other hand, there is a 1-1 corerspondence between quadratic form and symmetric bilinear
functions. A bilinear function f(X, Y ) is called symmetric if f(X, Y ) = f(Y, X) for all X, Y V .
Lemma 1.13. Given a quadratic form Q : V R there exists a unique symmetric bilinear form
f(X, Y ) such that Q(X) = f(X, X), X V .
Proof. If Q(X) = f(X, X) for a symmetric f then
Q(X +Y ) = f(X +Y, X +Y ) = f(X, X) +f(X, Y ) +f(Y, X) +f(Y, Y )
= Q(X) + 2f(X, Y ) +Q(Y ),
15
and hence
f(X, Y ) =
1
2
_
Q(X +Y ) Q(X) Q(Y )
_
. (1.5.2)
I leave it as an exercise to check that the formula (1.5.2) always denes a symmetric bilinear
function.
Let S = v
1
, . . . , v
n
is a basis of V . The matrix A = (a
ij
) of a symmetric bilinear form f(X, Y )
in the basis S is called also the matrix of the corresponding quadratic form Q(X) = f(X, X). This
matrix is symmetric, and
Q(X) =
ij
a
ij
x
i
x
j
= a
11
x
2
1
+ +a
nn
x
2
n
+ 2
i<j
a
ij
x
i
x
j
.
Thus the matrix A is diagonal if and only if the quadratic form Q is the sum of squares (with
coecients). Let us recall that if one changes the basis S to a basis
S = v
1
, . . . , v
n
then the
matrix of a bilinear form f changes to
C = C
T
AC.
Exercise 1.14. (Sylvesters inertia law) Prove that there is always exists a basis
S = v
1
, . . . , v
n
in which a quadratic form Q is reduced to a sum of squares. The number of positive, negative and
zero coecients with the squares is independent of the choice of the basis.
Thus, in some coordinate system a quadratic form can always be written as
1
x
2
i
+
k+l
k+1
x
2
j
, k +l n.
The number k of negative squares is called the negative index or simply the index of the quadratic
form Q, the total number of k + l of non-zero squares is called the rank of the form. It coincides
with the rank of the matrix of the form in any basis. A bilinear (and quadratic) form is called
non-degenerate if its rank is maximal possible, i.e. equal to n. For a non-degenerate quadratic for
Q the dierence l k between the number of positive and negative squares is called the signature.
A quadratic form Q is called positive denite if Q(X) 0 and if Q(X) = 0 then X = 0.
Equivalently, one can say that a form is positive denite if it is non-degenerate and its negative
index is equal to 0.
16
Chapter 2
Tensor and exterior products
2.1 Tensor product
Given a k-linear function and a l-linear function , one can form a (k +l)-linear function, which
will be denoted by and called the tensor product of the functions and . By denition
(X
1
, . . . , X
k
, X
k+1
, . . . , X
k+l
) := (X
1
, . . . , X
k
) (X
k+1
, . . . , X
k+l
).
For instance, the tensor product two linear functions l
1
and l
2
is a bilinear function l
1
l
2
dened
by the formula
l
1
l
2
(U, V ) = l
1
(U)l
2
(V ).
Let v
1
. . . v
n
be a basis in V and x
1
, . . . , x
n
a dual basis in V
, i.e. x
1
, . . . , x
n
are coordinates of a
vector with respect to the basis v
1
, . . . , v
n
.
The tensor product x
i
x
j
is a bilinear function x
i
x
j
(Y, Z) = y
i
z
j
. Thus a bilinear function
f with a matrix A can be written as a linear combination of the functions x
i
x
j
as follows:
f =
n
i,j=1
a
ij
x
i
x
j
,
where a
ij
is the matrix of the form f in the basis v
1
. . . v
n
. Similarly any k-linear function f with
a k-dimensional matrix A = a
i
1
i
2
...i
k
can be written (see 2.1 below) as a linear combination of
functions
x
i
1
x
i
2
x
i
k
, 1 i
1
, i
2
, . . . , i
k
n.
17
Namely, we have
f =
n
i
1
,i
2
,...i
k
=1
a
i
1
i
2
...i
k
x
i
1
x
i
2
x
i
k
.
2.2 Spaces of multilinear functions
All k-linear functions, or k-tensors, on a given n-dimensional vector space V themselves form a
vector space, which will be denoted by V
k
. The space V
1
is, of course, just the dual space V
.
Proposition 2.1. Let v
1
, . . . v
n
be a basis of V , and x
1
, . . . , x
k
be the dual basis of V
formed
by coordinate functions with respect to the basis V . Then n
k
k-linear functions x
i
1
x
i
k
,
1 i
1
, . . . , i
k
n, form a basis of the space V
k
.
Proof. Take a k-linear function F from V
k
and evaluate it on vectors v
i
1
, . . . , v
i
k
:
F(v
i
1
, . . . , v
i
k
) = a
i
1
...i
k
.
We claim that we have
F =
1i
1
,...,i
k
n
a
i
1
...i
k
x
i
1
x
i
k
.
Indeed, the functions on the both sides of this equality being evaluated on any set of k basic vectors
v
i
1
, . . . , v
i
k
, give the same value a
i
1
...i
k
. The same argument shows that if
1i
1
,...,i
k
n
a
i
1
...i
k
x
i
1
x
i
k
= 0 , then all coecients a
i
1
...i
k
should be equal to 0. Hence the functions x
i
1
x
i
k
,
1 i
1
, . . . , i
k
n, are linearly independent, and therefore form a basis of the space V
k
Similar to the case of spaces of linear functions, a linear map / : V W induces a linear map
/
: W
k
V
k
, which sends a k-linear function F W
k
to a k-linear function /
(F) V
k
,
dened by the formula
/
(F)(X
1
, . . . , X
k
) = F(/(X
1
), . . . , /(X
k
)) for any vectors X
1
, . . . X
k
V .
Exercise 2.2. Suppose V is provided with a basis v
1
, . . . , v
n
and x
i
1
x
i
k
, 1 i
1
, . . . , i
k
n, is
the corresponding basis of the space V
k
. Suppose that the map / : V V has a matrix A = (a
ij
)
in the basis v
1
, . . . , v
n
. Find the matrix of the map /
: V
k
V
k
in the basis x
i
1
x
i
k
.
18
2.3 Symmetric and skew-symmetric tensors
A multilinear function (tensor) is called symmetric if it remains unchanged under the transposition
of any two of its arguments:
f(X
1
, . . . , X
i
, . . . , X
j
, . . . , X
k
) = f(X
1
, . . . , X
j
, . . . , X
i
, . . . , X
k
)
Equivalently, one can say that a k-tensor f is symmetric if
f(X
i
1
, . . . , X
i
k
) = f(X
1
, . . . , X
k
)
for any permutation i
1
, . . . , i
k
of indices 1, . . . , k.
Exercise 2.3. Show that a bilinear function f(X, Y ) is symmetric if and only if its matrix (in any
basis) is symmetric.
Notice that the tensor product of two symmetric tensors usually is not symmetric.
Example 2.4. Any linear function is (trivially) symmetric. However, the tensor product of two
functions l
1
l
2
is not a symmetric bilinear function unless l
1
= l
2
. On the other hand, the function
l
1
l
2
+l
2
l
1
, is symmetric.
A tensor is called skew-symmetric (or anti-symmetric) if it changes its sign when one transposes
any two of its arguments:
f(X
1
, . . . , X
i
, . . . , X
j
, . . . , X
k
) = f(X
1
, . . . , X
j
, . . . , X
i
, . . . , X
k
).
Equivalently, one can say that a k-tensor f is anti-symmetric if
f(X
i
1
, . . . , X
i
k
) = (1)
inv(i
1
...i
k
)
f(X
1
, . . . , X
k
)
for any permutation i
1
, . . . , i
k
of indices 1, . . . , k, where inv(i
1
. . . i
k
) is the number of inversions in
the permutation i
1
, . . . , i
k
. Recall that two indices i
k
, i
l
form an inversion if k < l but i
k
> i
l
.
The matrix A of a bilinear skew-symmetric function is skew-symmetric, i.e.
A
T
= A.
19
Example 2.5. The determinant det(X
1
, . . . , X
n
) (considered as a function of columns X
1
, . . . , X
n
of a matrix) is a skew-symmetric n-linear function.
Exercise 2.6. Prove that any n-linear skew-symmetric function on R
n
is proportional to the de-
terminant.
Linear functions are trivially anti-symmetric (as well as symmetric).
As in the symmetric case, the tensor product of two skew-symmetric functions is not skew-
symmetric. We will dene below in Section 2.5 a new product, called an exterior product of skew-
symmetric functions, which will again be a skew-symmetric function.
2.4 Symmetrization and anti-symmetrization
The following constructions allow us to create symmetric or anti-symmetric tensors from arbitrary
tensors. Let f(X
1
, . . . , X
k
) be a k-tensor. Set
f
sym
(X
1
, . . . , X
k
) :=
(i
1
...i
k
)
f(X
i
1
, . . . , X
i
k
)
and
f
asym
(X
1
, . . . , X
k
) :=
(i
1
...i
k
)
(1)
inv(i
1
,...,i
k
)
f(X
i
1
, . . . , X
i
k
)
where the sums are taken over all permutations i
1
, . . . , i
k
of indices 1, . . . , k. The tensors f
sym
and f
asym
are called, respectively, symmetrization and anti-symmetrization of the tensor f. It is
now easy to see that
Proposition 2.7. The function f
sym
is symmetric. The function f
asym
is skew-symmetric. If f
is symmetric then f
sym
= k!f and f
asym
= 0. Similarly, if f is anti-symmetric then f
asym
= k!f,
f
sym
= 0.
Exercise 2.8. Let x
1
, . . . , x
n
be coordinate function in R
n
. Find (x
1
x
2
. . . x
n
)
asym
.
Answer. The determinant.
20
2.5 Exterior product
For our purposes skew-symmetric functions will be more important. Thus we will concentrate on
studying operations on them.
Skew-symmetric k-linear functions are also called exterior k-forms. Let be an exterior k-form
and an exterior l-form. We dene an exterior (k + l)-form , the exterior product of and
, as
:=
1
k!l!
( )
asym
.
In other words,
(X
1
, . . . , X
k
, X
k+1
, . . . , X
k+l
) =
1
k!l!
i
1
,...i
k+l
(1)
inv(i
1
,...,i
k+l
)
(X
i
1
. . . , X
i
k
)(X
i
k+1
, . . . , X
i
k+l
),
where the sum is taken over all permutations of indices 1, . . . , k +l.
Note that because the anti-symmetrization of an anti-symmetric k-tensor amounts to its mul-
tiplication by k!, we can also write
(X
1
, . . . , X
k+l
) =
i
1
<...<i
k
,i
k+1
<...<i
k+l
(1)
inv(i
1
,...,i
k+l
)
(X
i
1
, . . . , X
i
k
)(X
i
k+1
, . . . , X
i
k+l
),
where the sum is taken over all permutations i
1
, . . . , i
k+l
of indices 1, . . . , k +l.
Exercise 2.9. The exterior product operation has the following properties:
For any exterior k-form and exterior l-form we have = (1)
kl
.
Exterior product is linear with respect to each factor:
(
1
+
2
) =
1
+
2
() = ( )
for k-forms ,
1
,
2
, l-form and a real number .
Exterior product is associative: ( ) = ( ).
21
First two properties are fairly obvious. To prove associativity one can check that both sides of
the equality ( ) = ( ) are equal to
1
k!l!m!
( )
asym
.
In particular, if , and are 1-forms, i.e. if k = l = m = 1 then
= ( )
asym
.
This formula can be generalized for computing the exterior product of any number of 1-forms:
1
k
= (
1
k
)
asym
. (2.5.1)
Example 2.10. x
1
x
2
= x
1
x
2
x
2
x
1
. For 2 vectors, U =
_
_
_
_
_
u
1
.
.
.
u
n
_
_
_
_
_
, V =
_
_
_
_
_
v
1
.
.
.
v
n
_
_
_
_
_
, we have
x
1
x
2
(U, V ) = u
1
v
2
u
2
v
1
=
u
1
v
1
u
2
v
2
.
For 3 vectors U, V, W we have
x
1
x
2
x
3
(U, V, W) =
= x
1
x
2
(U, V )x
3
(W) +x
1
x
2
(V, W)x
3
(U) +x
1
x
2
(W, U)x
3
(V ) =
(u
1
v
2
u
2
v
1
)w
3
+ (v
1
w
2
v
2
w
1
)u
3
+ (w
1
u
2
w
2
u
1
)v
3
=
u
1
v
1
w
1
u
2
v
2
w
2
u
3
v
3
w
3
.
The last equality is just the expansion formula of the determinant according to the last row.
Proposition 2.11. Any exterior 2-form f can be written as
f =
1i<jn
a
ij
x
i
x
j
22
Proof. We had seen above that any bilinear form can be written as f =
ij
a
i,j
x
i
x
j
. If f is
skew-symmetric then the matrix A = (a
ij
) is skew-symmetric, i.e. a
ii
= 0, a
ij
= a
ji
for i ,= j.
Thus, f =
1i<jn
a
ij
(x
i
x
j
x
j
x
i
) =
1i<jn
a
ij
x
i
x
j
.
Exercise 2.12. Prove that any exterior k-form f can be written as
f =
1i,<...<i
k
n
a
i
1
...i
k
x
i
1
x
i
2
. . . x
i
k
.
The following proposition can be proven by induction over k, similar to what has been done in
Example 2.10 for the case k = 3.
Proposition 2.13. For any k 1-forms l
1
, . . . , l
k
and k vectors X
1
, . . . , X
k
we have
l
1
l
k
(X
1
, . . . , X
k
) =
l
1
(X
1
) . . . l
1
(X
k
)
. . . . . . . . .
l
k
(X
1
) . . . l
k
(X
k
)
. (2.5.2)
Corollary 2.14. The 1-forms l
1
, . . . , l
k
are linearly dependent as vectors of V
if and only if
l
1
. . . l
k
= 0. In particular, l
1
. . . l
k
= 0 if k > n = dimV .
Proof. If l
1
, . . . , l
k
are dependent then for any vectors X
1
, . . . , X
k
V the rows of the determinant
in the equation (2.13) are linearly dependent. Therefore, this determinant is equal to 0, and hence
l
1
. . . l
k
= 0. In particular, when k > n then the forms l
1
, . . . , l
k
are dependent (because
dimV
= dimV = n).
On the other hand, if l
1
, . . . , l
k
are linearly independent, then the vectors l
1
, . . . , l
k
V
can be
completed to form a basis l
1
, . . . , l
k
, l
k+1
, . . . , l
n
of V
. In other words, l
1
, . . . , l
n
can be viewed as
coordinate functions with respect to the basis w
1
, . . . , w
n
. In particular, we have l
i
(w
j
) = 0 if i ,= j
and l
i
(w
i
) = 1 for all i, j = 1, . . . , n. Hence we have
l
1
l
k
(w
1
, . . . , w
k
) =
l
1
(w
1
) . . . l
1
(w
k
)
. . . . . . . . .
l
k
(w
1
) . . . l
k
(w
k
)
1 . . . 0
. . . 1 . . .
0 . . . 1
= 1 ,
23
i.e. l
1
l
k
,= 0.
Proposition 2.13 can be also deduced from formula (2.5.1).
Corollary 2.14 and Exercise 2.12 imply that there are no non-zero k-forms on an n-dimensional
space for k > n.
2.6 Spaces of symmetric and skew-symmetric tensors
As was mentioned above, k-tensors on a vector space V form a vector space under the operation of
addition of functions and multiplication by real numbers. We denoted this space by V
k
. Symmet-
ric and skew-symmetric tensors form subspaces of this space V
k
, which we denote, respectively,
by S
k
(V
) and
k
(V
). In particular, we have
V
= S
1
(V
) =
1
(V
)
.
Exercise 2.15. What is the dimension of the spaces S
k
(V
) and
k
(V
)?
Answer.
dim
k
(V
) =
_
_
n
k
_
_
=
n!
k!(n k)!
dimS
k
(V
) =
(n +k 1)!
k!(n 1)!
.
The basis of
k
(V
on spaces of tensors
For any linear operator / : V W we introduced above in Section 1.3 the notion of a dual linear
operator /
: W
. Namely /
: W
k
V
k
.
24
Given a k-tensor W
k
and k vectors X
1
, . . . , X
k
V we dene
/
()(X
1
, . . . , X
k
) = (/(X
1
), . . . , /(X
k
)) .
Note that if is symmetric, or anti-symmetric, so is /
) S
k
(V
) and
k
(W
)
k
(V
for both of
these maps as well.
Proposition 2.16. Let / : V W be a linear map. Then
1. /
( ) = /
() /
() for any W
k
, W
l
;
2. /
(
asym
) = (/
())
asym
, /
(
sym
) = (/
())
sym
;
3. /
( ) = /
() /
() for any
k
(W
),
l
(W
).
If B : W U is another linear map then (B /)
= /
.
Proof.
1. Take any k +l vectors X
1
, . . . , X
k+l
V . Then by denition of the operator /
we have
/
( )(X
1
, . . . , X
k+l
) = (/(X
1
), . . . , /(X
n+k
) =
(/(X
1
), . . . , /(X
k
)(/(X
k+1
), . . . , /(X
n+k
)) =
/
()(X
1
, . . . , X
k
)/
()(X
k+1
, . . . , X
k+n
) =
/
() /
()(X
1
, . . . , X
k+l
).
2. Given k vectors X
1
, . . . , X
k
V we get
/
(
asym
)(X
1
, . . . , X
k
) =
asym
(/(X
1
), . . . , /(X
k
)) =
(i
1
...i
k
)
(1)
inv(i
1
,...,i
k
)
(/(X
i
1
), . . . , /(X
i
k
)) =
(i
1
...i
k
)
(1)
inv(i
1
,...,i
k
)
/
()(X
i
1
, . . . , X
i
k
) = (/
())
asym
(X
1
, . . . , X
k
),
where the sum is taken over all permutations i
1
, . . . , i
k
of indices 1, . . . , k. Similarly one proves that
/
(
sym
) = (/
())
sym
.
3. /
( ) =
1
k!l!
/
(( )
asym
)) =
1
k!l!
(/
( ))
asym
= /
() /
().
The last statement of Proposition 2.16 is straightforward and its proof is left to the reader.
25
Let us now discuss how to compute /
(y
k
) = l
k
=
m
j=1
a
kj
x
j
, k = 1, . . . , n. Indeed, the
coecients of the function l
k
= /
(y
k
) form the k-th column of the transpose matrix A
T
. Hence,
using Proposition 2.16 we compute:
/
(y
j
1
y
j
k
) = l
j
1
l
j
k
and
/
(y
j
1
y
j
k
) = l
j
1
l
j
k
.
Consider now the case when V = W, n = m, and we use the same basis v
1
, . . . , v
n
in the source
and target spaces.
Proposition 2.17.
/
(x
1
x
n
) = det Ax
1
x
n
.
Note that the determinant det A is independent of the choice of the basis. Indeed, the matrix
of a linear map changes to a similar matrix C
1
AC in a dierent basis, and det C
1
AC = det A.
Hence, we can write det / instead of det A, i.e. attribute the determinant to the linear operator /
rather than to its matrix A.
26
Proof. We have
/
(x
1
x
n
) = l
1
l
n
=
n
i
1
=1
a
1i
1
x
i
1
n
in=1
a
nin
x
in
=
n
i
1
,...,in=1
a
1i
1
. . . a
nin
x
i
1
x
in
.
Note that the in the latter sum all terms with repeating indices vanish, and hence we can replace
this sum by a sum over all permutations of indices 1, . . . , n. Thus, we can continue
/
(x
1
x
n
) =
i
1
,...,in
a
1i
1
. . . a
nin
x
i
1
x
in
=
_
_
i
1
,...,in
(1)
inv(i
1
,...,in)
a
1i
1
. . . a
nin
_
_
x
1
x
n
= det Ax
1
x
n
.
Exercise 2.18. Apply the equality
/
(x
1
x
k
x
k+1
x
n
) = /
(x
1
x
k
) /
(x
k+1
x
n
)
for a map / : R
n
R
n
to deduce the formula for expansion of a determinant according to its rst
k rows:
det A =
i
1
<<i
k
, j
1
<<j
nk
; im=j
l
(1)
inv(i
1
,...,j
nk
)
a
1,i
1
. . . a
1,i
k
.
.
.
.
.
.
.
.
.
a
k,i
1
. . . a
k,i
k
a
k+1,j
1
. . . a
k+1,j
nk
.
.
.
.
.
.
.
.
.
a
n,j
1
. . . a
n,j
nk
.
27
28
Chapter 3
Orientation and Volume
3.1 Orientation
We say that two bases v
1
, . . . , v
k
and w
1
, . . . , w
k
of a vector space V dene the same orientation of
V if the matrix of transition from one of these bases to the other has a positive determinant. Clearly,
if we have 3 bases, and the rst and the second dene the same orientation, and the second and the
third dene the same orientation then the rst and the third also dene the same orientation. Thus,
one can subdivide the set of all bases of V into the two classes. All bases in each of these classes
dene the same orientation; two bases chosen from dierent classes dene opposite orientation of
the space. To choose an orientation of the space simply means to choose one of these two classes of
bases.
There is no way to say which orientation is positive or which is negativeit is a question
of convention. For instance, the so-called counter-clockwise orientation of the plane depends from
which side we look at the plane. The positive orientation of our physical 3-space is a physical, not
mathematical, notion.
Suppose we are given two oriented spaces V, W of the same dimension. An invertible linear map
(= isomorphism) / : V W is called orientation preserving if it maps a basis which denes the
given orientation of V to a basis which denes the given orientation of W.
Any non-zero exterior n-form on V induces an orientation of the space V . Indeed, the preferred
set of bases is characterized by the property (v
1
, . . . , v
n
) > 0.
29
3.2 Orthogonal transformations
Let V be a Euclidean vector space. Recall that a linear operator | : V V is called orthogonal if
it preserves the scalar product, i.e. if
|(X), |(Y )) = X, Y ), (3.2.1)
for any vectors X, Y V . Recall that we have
|(X), |(Y )) = X, |
(|(Y ))),
where |
| = Id, or |
= |
1
.
Here we denoted by Id the identity operator, i.e. Id(X) = X for any X V .
Let us recall, see Exercise 1.10, that the adjoint operator U
: V
by the formula
U
= T
1
U
T.
Hence, for an orthogonal operator |, we have T
1
U
T = |
1
, i.e.
|
T = T |
1
. (3.2.2)
Let v
1
, . . . , v
n
be an orthonormal basis in V and U be the matrix of | in this basis. The matrix
of the adjoint operator in an orthonormal basis is the transpose of the matrix of this operator.
Hence, the equation |
i
u
ki
u
ji
=
_
_
1, if k = j,
0, if k ,= j,
.
30
Similarly, the equation UU
T
= E can be rewritten as
i
u
ik
u
ij
=
_
_
1, if k = j,
0, if k ,= j,
.
The above identities mean that columns (and rows) of an orthogonal matrix U form an or-
thonormal basis of R
n
with respect to the dot-product.
In particular, we have
1 = det(U
T
U) = det(U
T
) det U = (det U)
2
,
and hence det U = 1. In other words, the determinant of any orthogonal matrix is equal 1. We
can also say that the determinant of an orthogonal operator is equal to 1 because the determinant
of the matrix of an operator is independent of the choice of a basis. Orthogonal transformations
with det = 1 preserve the orientation of the space, while those with det = 1 reverse it.
Composition of two orthogonal transformations, or the inverse of an orthogonal transformation
is again an orthogonal transformation. The set of all orthogonal transformations of an n-dimensional
Euclidean space is denoted by O(n). Orientation preserving orthogonal transformations sometimes
called special, and the set of special orthogonal transformations is denoted by SO(n). For instance
O(1) consists of two elements and SO(1) of one: O(1) = 1, 1, SO(1) = 1. SO(2) consists of
rotations of the plane, while O(2) consists of rotations and reections with respect to lines.
3.3 Determinant and Volume
We begin by recalling some facts from Linear Algebra. Let V be an n-dimensional Euclidean space
with an inner product , ). Given a linear subspace L V and a point x V , the projection
proj
L
(x) is a vector y L which is uniquely characterized by the property x y L, i.e.
x y, z) = 0 for any z L. The length [[x proj
L
(x)[[ is called the distance from x to L; we
denote it by dist(x, L).
Let U
1
, . . . , U
k
V be linearly independent vectors. The k-dimensional parallelepiped spanned
by vectors U
1
, . . . , U
k
is, by denition, the set
P(U
1
, . . . , U
k
) =
_
k
j
U
j
; 0
1
, . . . ,
k
1
_
Span(U
1
, . . . , U
k
).
31
Given a k-dimensional parallelepiped P = P(U
1
, . . . , U
k
) we will dene its k-dimensional volume
by the formula
VolP = [[U
1
[[dist(U
2
, Span(U
1
))dist(U
3
, Span(U
1
, U
2
)) . . . dist(U
k
, Span(U
1
, . . . , U
k1
)). (3.3.1)
Of course we can write dist(U
1
, 0) instead of [[U
1
[[. This denition agrees with the denition of the
area of a parallelogram, or the volume of a 3-dimensional parallelepiped in the elementary geometry.
Proposition 3.1. Let v
1
, . . . , v
n
be an orthonormal basis in V . Given n vectors U
1
, . . . , U
n
let us
denote by U the matrix whose columns are coordinates of these vectors in the basis v
1
, . . . , v
n
:
U :=
_
_
_
_
_
u
11
. . . u
1n
.
.
.
u
n1
. . . u
nn
_
_
_
_
_
Then
Vol P(U
1
, . . . , U
n
) = [ det U[.
Proof. If the vectors U
1
, . . . , U
n
are linearly dependent then Vol P(U
1
, . . . , U
n
) = det U = 0.
Suppose now that the vectors U
1
, . . . , U
n
are linearly independent, i.e. form a basis. Consider rst
the case where this basis is orthonormal. Then the matrix U is orthogonal. i.e. UU
T
= E, and
hence det U = 1. But in this case Vol P(U
1
, . . . , U
n
) = 1, and hence Vol P(U
1
, . . . , U
n
) = [ det U[.
Now let the basis U
1
, . . . , U
n
be arbitrary. Let us apply to it the Gram-Schmidt orthonormaliza-
tion process. Recall that this process consists of the following steps. First, we normalize the vector
U
1
, then subtract from U
2
its projection to Span(U
1
), Next, we normalize the new vector U
2
, then
subtract from U
3
its projection to Span(U
1
, U
2
), and so on. At the end of this process we obtain
an orthonormal basis. It remains to notice that each of these steps aected Vol P(U
1
, . . . , U
n
) and
[ det U[ in a similar way. Indeed, when we multiplied the vectors by a positive number, both the
volume and the determinant were multiplied by the same number. When we subtracted from a vec-
tor U
k
its projection to Span(U
1
, . . . , U
k1
), this aected neither the volume nor the determinant.
(x
1
x
n
)(U
1
, . . . , U
n
) = det /x
1
. . . x
n
(U
1
, . . . , U
n
).
In view of Proposition 3.1 and the rst part of Corollary 3.2 the value
x
1
. . . x
n
(U
1
, . . . , U
n
) = det U
is called sometimes the signed volume of the parallelepiped P(U
1
, . . . , U
n
). It is positive when the
basis U
1
, . . . , U
n
denes the given orientation of the space V , and it is negative otherwise.
Note that x
1
. . . x
k
(U
1
, . . . , U
k
) for 0 k n is the signed k-dimensional volume of the
orthogonal projection of the parallelepiped P(U
1
, . . . , U
k
) to the coordinate subspace x
k+1
=
= x
n
= 0.
For instance, let be the 2-form x
1
x
2
+x
3
x
4
on R
4
. Then for any two vectors U
1
, U
2
R
4
the value (U
1
, U
2
) is the sum of signed areas of projections of the parallelogram P(U
1
, U
2
) to the
coordinate planes spanned by the two rst and two last basic vectors.
3.4 Volume and Gram matrix
In this section we will compute the Vol
k
P(v
1
, . . . , v
k
) in the case when the number k of vectors is
less than the dimension n of the space.
Let V be an Euclidean space. Given vectors v
1
, . . . , v
k
V we can form a k k-matrix
G(v
1
, . . . , v
k
) =
_
_
_
_
_
v
1
, v
1
) . . . v
1
, v
k
)
. . . . . . . . .
v
k
, v
1
) . . . v
k
, v
k
)
_
_
_
_
_
, (3.4.1)
which is called the Gram matrix of vectors v
1
, . . . , v
k
.
Suppose we are given Cartesian coordinate system in V and let us form a matrix C whose
columns are coordinates of vectors v
1
, . . . , v
k
. Thus the matrix C has n rows and k columns. Then
G(v
1
, . . . , v
k
) = C
T
C,
33
because in Cartesian coordinates the scalar product looks like the dot-product.
We also point out that if k = n and vectors v
1
, . . . , v
n
form a basis of V , then G(v
1
, . . . , v
k
) is
just the matrix of the bilinear function X, Y ) in the basis v
1
, . . . , v
n
. It is important to point out
that while the matrix C depends on the choice of the basis, the matrix G does not.
Proposition 3.3. Given any k vectors v
1
, . . . , v
k
in an Euclidean space V the volume Vol
k
P(v
1
, . . . , v
k
)
can be computed by the formula
Vol
k
P(v
1
, . . . , v
k
)
2
= det G(v
1
, . . . , v
k
) = det C
T
C, (3.4.2)
where G(v
1
, . . . , v
k
) is the Gram matrix and C is the matrix whose columns are coordinates of
vectors v
1
, . . . , v
k
in some orthonormal basis.
Proof. Suppose rst that k = n. Then according to Proposition 3.1 we have Vol
k
P(v
1
, . . . , v
k
) =
[ det C[. But det C
T
C = det C
2
, and the claim follows.
Let us denote vectors of our orthonormal basis by w
1
, . . . , w
n
. Consider now the case when
Span(v
1
, . . . , v
k
) Span(w
1
, . . . , w
k
). (3.4.3)
In this case the elements in the j-th row of the matrix C are zero if j > k. Hence, if we denote
by
C the square k k matrix formed by the rst k rows of the matrix C, then C
T
C =
C
T
C and
thus det C
T
C = det
C
T
C. But det
C
T
C = Vol
k
P(v
1
, . . . , v
k
) in view of our above argument in the
equi-dimensional case applied to the subspace Span(w
1
, . . . , w
k
) V , and hence
Vol
2
k
P(v
1
, . . . , v
k
) = det C
T
C = det G(v
1
, . . . , v
k
).
But neither Vol
k
P(v
1
, . . . , v
k
), nor the Gram matrix G(v
1
, . . . , v
k
) depends on the choice of an
orthonormal basis. On the other hand, using Gram-Schmidt process one can always nd an or-
thonormal basis which satises condition (3.4.3).
Remark 3.4. Note that det G(v
1
, . . . , v
k
) 0 and det G(v
1
, . . . , v
k
) = 0 if an and only if the
vectors v
1
, . . . , v
k
are linearly dependent.
34
Chapter 4
Dualities
4.1 Duality between k-forms and (nk)-forms on a n-dimensional
Euclidean space V
Let V be an n-dimensional vector space. As we have seen above, the space
k
(V
) of k-forms, and
the space
nk
(V
)
nk
(V
) .
Denition 4.1. Let be a k-form. Then given any vectors U
1
, . . . , U
nk
, the value (U
1
, . . . , U
nk
)
can be computed as follows. If U
1
, . . . , U
nk
are linearly dependent then (U
1
, . . . , U
nk
) = 0. Oth-
erwise, let S
such that
Vol
k
(Z
1
, . . . , Z
k
) = Vol
nk
(U
1
, . . . , U
nk
)
and the basis Z
1
, . . . , Z
k
, U
1
, . . . , U
nk
denes the given orientation of the space V . Then
(U
1
, . . . , U
nk
) = (Z
1
, . . . , Z
k
). (4.1.1)
35
Let us rst show that
Lemma 4.2. is a (n k)-form, i.e. is skew-symmetric and multilinear.
Proof. To verify that is skew-symmetric we note that for any 1 i < j n q the bases
Z
1
, Z
2
, . . . , Z
k
, U
1
, . . . , U
i
, . . . , U
j
, . . . , U
nk
and
Z
1
, Z
2
, . . . , Z
k
, U
1
, . . . , U
j
, . . . , U
i
, . . . , U
nk
dene the same orientation of the space V , and hence
(U
1
, . . . , U
j
, . . . , U
i
, . . . , U
nk
) = (Z
1
, Z
2
, . . . , Z
k
)
=(Z
1
, Z
2
, . . . , Z
k
) = (U
1
, . . . , U
i
, . . . , U
j
, . . . , U
nk
).
Hence, in order to check the multi-linearity it is sucient to prove the linearity of with respect
to the rst argument only. It is also clear that
(U
1
, . . . , U
nk
) = (U
1
, . . . , U
nk
). (4.1.2)
Indeed, multiplication by ,= 0 does not change the span of the vectors U
1
, . . . , U
nq
, and hence
if (U
1
, . . . , U
nk
) = (Z
1
, . . . , Z
k
) then (U
1
, . . . , U
nk
) = (Z
1
, . . . , Z
k
) = (Z
1
, . . . , Z
k
).
Thus it remains to check that
(U
1
+
U
1
, U
2
, . . . , U
nk
) = (U
1
, U
2
, . . . , U
nk
) +(
U
1
, U
2
, . . . , U
nk
)).
Let us denote L := Span(U
2
, . . . , U
nk
) and observe that proj
L
(U
1
+
U
1
) = proj
L
(U
1
) +
proj
L
(
U
1
). Denote N := U
1
proj
L
(U
1
) and
N :=
U
1
proj
L
(
U
1
). The vectors N and
N are
normal components of U
1
and
U
1
with respect to the subspace L, and the vector N +
N is the
normal component of U
1
+
U
1
with respect to L. Hence, we have
(U
1
, . . . , U
nk
) = (N, . . . , U
nk
), (
U
1
, . . . , U
nk
) = (
N, . . . , U
nk
),
and
(U
1
+
U
1
, . . . , U
nk
) = (N +
N, . . . , U
nk
).
Indeed, in each of these three cases,
36
- vectors on both side of the equality span the same space;
- the parallelepiped which they generate have the same volume, and
- the orientation which these vectors dene together with a basis of the complementary space
remains unchanged.
Hence, it is sucient to prove that
(N +
N, U
2
, . . . , U
nk
) = (N, U
2
, . . . , U
nk
) +(
N, U
2
, . . . , U
nk
). (4.1.3)
If the vectors N and
N are linearly dependent, i.e. one of them is a multiple of the other, then
(4.1.3) follows from (4.1.2).
Suppose now that N and
N are linearly independent. Let L
= k + 1 and we have N,
N L
. Let us denote by M
the plane in L
.
Note that dimM
= k 1.
Choose any orientation of M so that we can talk about counter-clockwise rotation of this plane.
Let Y,
a basis Z
2
, . . . , Z
k
such that
Vol
k1
P(Z
2
, . . . , Z
k
) = Vol
nk1
P(U
2
, . . . , U
nk
).
Note that the orthogonal complements to Span(N, U
2
, . . . , U
nk
), Span(
N, U
2
, . . . , U
nk
), and to
Span(N+
N, U
2
, . . . , U
nk
) in V coincide, respectively, with the orthogonal complements to the the
vectors N,
N and to N +
N in L
V
= Span(Y, Z
2
, . . . , Z
k
),
_
Span(
N, U
2
, . . . , U
nk
)
_
V
= Span(
Y , Z
2
, . . . , Z
k
) and
_
Span(N +
N, U
2
, . . . , U
nk
)
_
V
= Span(Y +
Y , Z
2
, . . . , Z
k
).
Next, we observe that
Vol
nk
P(N, U
2
, . . . , U
nk
) = Vol
k
P(Y, Z
2
, . . . , Z
k
),
37
Vol
nk
P(
N, U
2
, . . . , U
nk
) = Vol
k
P(
Y , Z
2
, . . . , Z
k
) and
Vol
nk
P(N +
N, U
2
, . . . , U
nk
) = Vol
k
P(Y +
Y , Z
2
, . . . , Z
k
).
Consider the following 3 bases of V :
Y, Z
2
, . . . , Z
k
, N, U
2
, . . . , U
nk
,
Y , Z
2
, . . . , Z
k
,
N, U
2
, . . . , U
nk
,
Y +
Y , Z
2
, . . . , Z
k
, N +
N, U
2
, . . . , U
nk
,
and observe that all three of them dene the same a priori given orientation of V . Thus, by denition
of the operator we have:
(N +
N, U
2
, . . . , U
nk
) = (Y +
Y , Z
2
, . . . , Z
k
)
= (Y, Z
2
, . . . , Z
k
) +(
Y , Z
2
, . . . , Z
k
) = (N, U
2
, . . . , U
nk
) +(
N, U
2
, . . . , U
nk
).
This completes the proof that is an (n k)-form.
Thus the map denes a map :
k
(V
)
nk
(V
).
Lemma 4.3.
x
i
1
x
i
2
x
i
k
= (1)
inv(i
1
,...,i
k
,j
1
,...,j
nk
)
x
j
1
x
j
2
x
j
nk
, (4.1.4)
where j
1
< < j
nk
is the set of indices, complementary to i
1
, . . . , i
k
. In other words, i
1
, . . . , i
k
, j
1
, . . . , j
nk
is a permutation of indices 1, . . . , n.
Proof. Evaluating (x
i
1
x
i
k
) on basic vectors v
j
1
, . . . , v
j
nk
, 1 j
1
< < j
nq
n, we get
0 unless all the indices j
1
, . . . , j
nk
are all dierent from i
1
, . . . , i
k
, while in the latter case we get
(x
i
1
x
i
k
)(v
j
1
, . . . , v
j
nk
) = (1)
inv(i
1
,...,i
k
,j
1
,...,j
nk
)
.
38
Hence,
x
i
1
x
i
2
x
i
k
= (1)
inv(i
1
,...,i
k
,j
1
,...,j
nk
)
x
j
1
x
j
2
x
j
nk
.
) and
the space
nk
(V
1i
1
<<i
k
n
a
i
1
...i
k
x
i
1
x
i
k
) we have
=
1i
1
<<i
k
n
a
i
1
...i
k
(x
i
1
x
i
k
) .
Examples.
1. C = Cx
1
x
n
; in other words the isomorphism acts on constants (= 0-forms) by
multiplying them by the volume form.
2. In R
3
we have
x
1
= x
2
x
3
, x
2
= x
1
x
3
= x
3
x
1
, x
3
= x
1
x
2
,
(x
1
x
2
) = x
3
, (x
3
x
1
) = x
2
, (x
2
x
3
) = x
1
.
3. More generally, given a 1-form l = a
1
x
1
+ +a
n
x
n
we have
l = a
1
x
2
x
n
a
2
x
1
x
3
x
n
+ + (1)
n1
a
n
x
1
x
n1
.
In particular for n = 3 we have
(a
1
x
1
+a
2
x
2
+a
3
x
3
) = a
1
x
2
x
3
+a
2
x
3
x
1
+a
3
x
1
x
2
.
Proposition 4.4.
2
= (1)
k(nk)
Id, i.e. () = (1)
k(nk)
for any k-form .
In particular, if dimension n = dimV is odd then
2
= Id. If n is even and is a k-form then
() = if k is even, and () = if k is odd.
39
Proof. It is sucient to verify the equality
() = (1)
k(nk)
_
(x
i
1
x
i
k
)
_
=
_
(1)
inv(i
1
,...,i
k
,j
1
,...,j
nk
)
x
j
1
x
j
2
x
j
nk
_
= (1)
inv(i
1
,...,i
k
,j
1
,...,j
nk
)
(x
j
1
x
j
2
x
j
nk
)
= (1)
inv(i
1
,...,i
k
,j
1
,...,j
nk
)+inv(j
1
,...,j
nk
,i
1
,...,i
k
)
x
i
1
x
i
k
= (1)
k(nk)
x
i
1
x
i
k
.
Exercise 4.5. (a) For any special orthogonal operator / the operators /
= /
.
(b) Let A be an orthogonal matrix of order n with det A = 1. Prove that for any k 1, . . . , n
the absolute value of each k-minor M of A is equal to the absolute value of its complementary
minor of order (n k). (Hint: Apply (a) to the form x
i
1
x
i
k
).
40
(c) Let V be an oriented 3-dimensional Euclidean space. Prove that for any two vectors X, Y V ,
their cross-product can be written in the form
X Y = T
1
((T(X) T(Y ))) .
4.2 Euclidean structure on the space of exterior forms
Suppose that the space V is oriented and Euclidean, i.e. it is endowed with an inner product , )
and an orientation.
Given two forms ,
k
(V
), k = 0, . . . , n, let us dene
, )) = ( ).
Note that is an n-form for every k, and hence, , )) is a 0-form, i.e. a real number.
Proposition 4.6. 1. The operation , )) denes an inner product on
k
(V
) for each k =
0, . . . , n.
2. If / : V V is a special orthogonal operator then the operator /
:
k
(V
)
k
(V
) is
orthogonal with respect to the inner product , )).
Proof. 1. We need to check that , )) is a symmetric bilinear function on
k
(V
) and , )) > 0
unless = 0. Bilinearity is straightforward. Hence, it is sucient to verify the remaining properties
for basic vectors = x
i
1
x
i
k
, = x
j
1
x
j
k
, where 1 i
1
< < i
k
n, 1 j
1
< <
j
k
n. Here (x
1
, . . . , x
n
) is any Cartersian coordinates in V which dene its given orientation.
Note that , )) = 0 = , )) unless i
m
= j
m
for all m = 1, . . . , k, and in the the latter case
we have = . Furthermore, we have
, )) = ( ) = (x
1
x
n
) = 1 > 0.
2. The inner product , )) is dened only in terms of the Euclidean structure and the orientation
of V . Hence, for any special orthogonal operator / (which preserves these structures) the induced
operator /
:
k
(V
)
k
(V
1i
1
<<i
k
n
a
i
1
...i
k
x
i
1
x
i
k
.
Then
[[ [[
2
= , )) =
1i
1
<<i
k
n
a
2
i
1
...i
k
.
Exercise 4.8. Show that if , are 1-forms on an Euclidean space V . Then
, )) = T
1
(), T
1
()),
i.e the scalar product , )) on V
1i
1
<<i
k
n
Z
2
i
1
,...,i
k
, (4.2.1)
where
Z
i
1
,...,i
k
=
z
i
1
1
. . . z
i
1
k
. . . . . . . . .
z
i
k
1
. . . z
i
k
k
.
Proof. Consider linear functions l
j
= T(Z
j
) =
n
i=1
z
ij
x
i
V
, j = 1, . . . , k. Then
l
1
l
k
=
n
i
1
=1
z
i
1
j
x
i
1
n
i
k
=1
z
i
k
j
x
i
k
=
i
1
,...,i
k
z
i
1
. . . z
i
k
x
i
1
x
i
k
=
1i
1
<<i
k
n
Z
i
1
,...,i
k
x
i
1
. . . x
i
k
. (4.2.2)
In particular, if one has Z
1
, . . . Z
k
Span(e
1
, . . . , e
k
) then Z
1...k
= Vol P(Z
1
, . . . , Z
k
) and hence
l
1
l
k
= Z
1...k
x
1
x
k
= Vol P(Z
1
, . . . , Z
k
)x
1
x
k
,
which yields the claim in this case.
42
In the general case, according to Proposition 4.7 we have
[[ l
1
l
k
[[
2
=
1i
1
<<i
k
n
Z
2
i
1
,...,i
k
, (4.2.3)
which coincides with the right-hand side of (4.2.1). Thus it remains to check to that
[[ l
1
l
k
[[= Vol
k
P(Z
1
, . . . , Z
k
).
Given any orthogonal transformation / : V V we have, according to Proposition 4.6, the equality
[[ l
1
l
k
[[=[[ /
l
1
/
l
k
[[ . (4.2.4)
We also note that any orthogonal transformation B : V V preserves k-dimensional volume of all
k-dimensional parallelepipeds:
[Vol
k
P(Z
1
, . . . , Z
k
)[ = [Vol
k
P(B(Z
1
), . . . , B(Z
k
))[. (4.2.5)
On the other hand, there exists an orthogonal transformation / : V V such that /
1
(Z
1
), . . . , /
1
(Z
k
)
Span(e
1
, . . . , e
k
). Denote
Z
j
:= /
1
(Z
j
), j = 1, . . . , k. Then, according to (3.2.2) we have
l
j
:= T(
Z
j
) = T(/
1
(Z
j
)) = /
(T(Z
j
)) = /
l
j
.
As was pointed out above we then have
[Vol
k
P(
Z
1
, . . . ,
Z
k
)[ =[[
l
1
l
k
[[=[[ /
l
1
/
l
k
[[, (4.2.6)
and hence, the claim follows from (4.2.4) and (4.2.5) applied to B = /
1
.
We recall that an alternative formula for computing Vol
k
P(Z
1
, . . . , Z
k
) was given earlier in
Proposition 3.3.
4.3 Contraction
Let V be a vector space and
k
(V
); R.
The proof is straightforward.
Let be a non-zero n-form. Then we have
Proposition 4.11. The map : V
n1
(V
).
Proof. Take a basis v
1
, . . . , v
n
. Let x
1
, . . . , x
n
V
1
a
j
x
1
x
j1
x
j+1
x
n
.
Then
v
i
(v
1
, . . . , v
l1
, v
l+1
, . . . , v
n
)
=
n
1
a
j
x
1
x
j1
x
j+1
x
n
(v
1
, . . . , v
l1
, v
l+1
, . . . , v
n
) = a
l
, (4.3.1)
44
but on the other hand,
v
i
(v
1
, . . . , v
l1
, v
l+1
, . . . , v
n
) = (v
i
, v
1
, . . . , v
l1
, v
l+1
, . . . , v
n
)
= (1)
i1
(v
1
, . . . , v
l1
, v
i
, v
l+1
, . . . , v
n
) =
_
_
(1)
i1
, l = i ;
0 , otherwise
(4.3.2)
Thus,
v
i
= (1)
i1
x
1
x
i1
x
i+1
x
n
.
Hence, the map sends a basis of V
into a basis of
n1
(V
Take a vector v =
n
1
a
j
v
j
. Then we have
v (x
1
x
n
) =
n
1
(1)
i1
a
i
x
1
x
i1
x
i+1
x
n
. (4.3.3)
This formula can be interpreted as the formula of expansion of a determinant according to the rst
column (or the rst row). Indeed, for any vectors U
1
, . . . , U
n1
we have
v (U
1
, . . . , U
n1
) = det(v, U
1
, . . . , U
n1
) =
a
1
u
1,1
. . . u
1,n1
. . . . . . . . . . . .
a
n
u
n,1
. . . u
n,n1
,
where
_
_
_
_
_
u
1,i
.
.
.
u
n,i
_
_
_
_
_
are coordinates of the vector U
i
V in the basis v
1
, . . . , v
n
. On the other hand,
v (U
1
, . . . , U
n1
) =
n
1
(1)
i1
a
i
x
1
x
i1
x
i+1
x
n
(U
1
, . . . , U
n1
)
= a
1
u
2,1
. . . u
2,n1
u
3,1
. . . u
3,n1
. . . . . . . . .
u
n,1
. . . u
n,n1
+ + (1)
n1
a
n
u
1,1
. . . u
1,n1
u
2,1
. . . u
2,n1
. . . . . . . . .
u
n1,1
. . . u
n1,n1
. (4.3.4)
45
Suppose that dimV = 3. Then the formula (4.3.3) can be rewritten as
v (x
1
x
2
x
3
) = a
1
x
2
x
3
+a
2
x
3
x
1
+a
3
x
1
x
2
,
where
_
_
_
_
_
a
1
a
2
a
3
_
_
_
_
_
are coordinates of the vector V . Let us describe the geometric meaning of the
operation . Set = v (x
1
x
2
x
3
). Then (U
1
, U
2
) is the volume of the parallelogram dened
by the vectors U
1
, U
2
and v. Let be the unit normal vector to the plane L(U
1
, U
2
) V . Then we
have
(U
1
, U
2
) = Area P(U
1
, U
2
) v, ) .
If we interpret v as the velocity of a uid ow in the space V then (U
1
, U
2
) is just an amount
of uid own through the parallelogram generated by vectors U
1
and U
2
for the unit time. It is
called the ux of v through the parallelogram .
Let us return back to the case dimV = n.
Exercise 4.12. Let
= x
i
1
x
i
k
, 1 i
1
< < i
k
n
and v = (a
1
, . . . , a
n
). Show that
v =
k
j=1
(1)
j+1
a
i
j
x
i
1
. . . x
i
j1
x
i
j+1
. . . x
i
k
.
The next proposition establishes a relation between the isomorphisms , and T.
Proposition 4.13. Let V be a Euclidean space, and x
1
, . . . , x
n
be coordinates in an orthonormal
basis. Then for any vector v V we have
Tv = v (x
1
x
n
) .
Proof. Let v = (a
1
, . . . , a
n
). Then Tv = a
1
x
1
+ +a
n
x
n
and
Tv = a
1
x
2
x
n
a
2
x
1
x
3
x
n
x
1
+ + (1)
n1
a
n
x
1
x
n1
.
But according to Proposition 4.11 the (n1)-form v (x
1
x
n
is dened by the same formula.
46
We nish this section by the proposition which shows how the contraction operation interacts
with the exterior product.
Proposition 4.14. Let , be forms of order k and l, respectively, and v a vector. Then
v ( ) = (v ) + (1)
k
(v ).
Proof. Note that given any indices k
1
, . . . k
m
(not necessarily ordered) we have e
i
x
k
1
x
km
= 0
if i / k
1
, . . . , k
m
and e
i
x
k
1
x
km
= (1)
J
x
k
1
. . .
i
x
km
, where J = J(i; k
1
, . . . , k
m
)
is the number of variables ahead of x
i
.
By linearity it is sucient to consider the case when v, , are basic vector and forms, i.e.
v = e
i
, = x
i
1
x
i
k
, = x
j
1
. . . x
j
l
.
We have v ,= 0 if and only if the index i is among the indices i
1
, . . . , i
k
. In that case
v = (1)
J
x
i
1
. . .
i
x
i
k
and if v ,= 0 then
v = (1)
J
x
j
1
. . .
i
x
j
l
,
where J = J(i; i
1
, . . . , i
k
), J
= J(i; j
1
, . . . , j
l
).
If it enters bot not then
v ( ) = (1)
J
x
i
1
. . .
i
x
i
k
x
j
1
x
j
l
= (v ) ,
while (v ) = 0.
Similarly, if it enters but not then
v ( ) = (1)
J
+m
x
i
1
x
i
k
x
j
1
. . .
i
x
j
l
= (1)
k
(v ),
while v = 0. Hence, in both these cases the formula holds.
If x
i
enters both products then = 0, and hence v ( = 0.
On the other hand,
(v ) + (1)
k
(v ) = (1)
J
x
i
1
. . .
i
x
i
k
x
j
1
x
j
l
+ (1)
k+J
x
i
1
x
i
k
x
j
1
. . .
i
x
j
l
= 0,
47
because the products x
i
1
. . .
i
x
i
k
x
j
1
x
j
l
and x
i
1
x
i
k
x
j
1
. . .
i
x
j
l
dier
only in the position of x
i
. In the rst product it is at the (k + J
+k+1
, which leads to the required
cancellation.
48
Chapter 5
Complex vector spaces
5.1 Complex numbers
The space R
2
can be endowed with an associative and commutative multiplication operation. This
operation is uniquely determined by three properties:
it is a bilinear operation;
the vector (1, 0) is the unit;
the vector (0, 1) satises (0, 1)
2
= (0, 1).
The vector (0, 1) is usually denoted by i, and we will simply write 1 instead of the vector (1, 0).
Hence, any point (a, b) R
2
can be written as a +bi, where a, b R, and the product of a +bi and
c +di is given by the formula
(a +bi)(c +di) = ac bd + (ad +bc)i.
The plane R
2
endowed with this multiplication is denoted by C and called the set of complex
numbers. The real line generated by 1 is called the real axis, the line generated by i is called the
imaginary axis. The set of real numbers R can be viewed as embedded into C as the real axis.
Given a complex number z = x +iy, the numbers x and y are called its real and imaginary parts,
respectively, and denoted by Rez and Imz, so that z = Rez +iImz.
49
For any non-zero complex number z = a + bi there exists an inverse z
1
such that z
1
z = 1.
Indeed, we can set
z
1
=
a
a
2
+b
2
b
a
2
+b
2
i.
The commutativity, associativity and existence of the inverse is easy to check, but it should not
be taken for granted: it is impossible to dene a similar operation any R
n
for n > 2.
Given z = a +bi C its conjugate is dened as z = a bi. The conjugation operation z z is
the reection of C with respect to the real axis R C. Note that
Rez =
1
2
(z + z), Imz =
1
2i
(z z).
Let us introduce the polar coordinates (r, ) in R
2
= C. Then a complex number z = x + yi
can be written as r cos + ir sin = r(cos + i sin). This form of writing a complex number
is called, sometimes, ttrigonometric. The number r =
_
x
2
+y
2
is called the modulus of z and
denoted by [z[ and is called the argument of and denoted by arg z. Note that the argument is
dened only mod 2. The value of the argument in [0, 2) is sometimes called the principal value
of the argument. When z is real than its modulus [z[ is just the absolute value. We also not that
[z[ =
z z.
An important role plays the triangle inequality
[z
1
[ [z
2
[
[z
1
+z
2
[ [z
1
[ +[z
2
[.
Exponential function of a complex variable
Recall that the exponential function e
x
has a Taylor expansion
e
x
=
0
x
n
n!
= 1 +x +
x
2
2
+
x
3
6
+. . . .
We then dene for a complex the exponential function by the same formula
e
z
:= 1 +z +
z
2
2!
+ +
z
n
n!
+. . . .
One can check that this power series absolutely converging for all z and satises the formula
e
z
1
+z+2
= e
z
1
e
z
2
.
50
Figure 5.1: Leonhard Euler (1707-1783)
In particular, we have
e
iy
= 1 +iy
y
2
2!
i
y
3
3!
+
y
4
4!
+ +. . . (5.1.1)
=
k=0
(1)
k
y
2k
2k!
+i
k=0
(1)
k
y
2k+1
(2k + 1)!
. (5.1.2)
But
k=0
(1)
k y
2k
2k!
= cos y and
k=0
(1)
k y
2k+1
(2k+1)!
= siny, and hence we get Eulers formula
e
iy
= cos y +i siny,
and furthermore,
e
x+iy
= e
x
e
iy
= e
x
(cos y +i siny),
i.e. [e
x+iy
[ = e
x
, arg(e
z
) = y. In particular, any complex number z = r(cos + i sin) can be
rewritten in the form z = re
i
. This is called the exponential form of the complex number z.
Note that
_
e
i
_
n
= e
in
,
and hence if z = re
i
then z
n
= r
n
e
in
= r
n
(cos n +i sinn).
Note that the operation z iz is the rotation of C counterclockwise by the angle
2
. More
generally a multiplication operation z zw, where w = e
i
is the composition of a rotation by
the angle and a radial dilatation (homothety) in times.
51
Exercise 5.1. 1. Compute
n
0
cos k and
n
1
sink.
2. Compute 1 +
_
n
4
_
+
_
n
8
_
+
_
n
12
_
+. . . .
5.2 Complex vector space
In a real vector space one knows how to multiply a vector by a real number. In a complex vector
space there is dened an operation of multiplication by a complex number. Example is the space
C
n
whose vectors are n-tuples z = (z
1
, . . . , z
n
) of complex numbers, and multiplication by any
complex number = + i is dened component-wise: (z
1
, . . . , z
n
) = (z
1
, . . . , z
n
). Complex
vector space can be viewed as an upgrade of a real vector space, or better to say as a real vector
space with an additional structure.
In order to make a real vector space V into a complex vector space, one just needs to dene how
to multiply a vector by i. This operation must be a linear map : V V which should satisfy
the condition
2
= Id, i.e ((v)) = i(iv) = v.
Example 5.2. Consider R
2n
with coordinates (x
1
, y
1
, . . . , x
n
, y
n
). Consider a 2n 2n-matrix
J =
_
_
_
_
_
_
_
_
_
_
_
_
_
_
_
_
_
0 1 0 0
1 0 0 0
0 0 0 1
0 0 1 0
. . .
0 1
1 0
_
_
_
_
_
_
_
_
_
_
_
_
_
_
_
_
_
Then J
2
= I. Consider a linear operator : R
2n
R
2n
with this matrix, i.e. (Z) = JZ for
any vector Z R
2n
which we view as a column-vector. Then
2
= Id, and hence we can dene
on R
2n
a complex structure (i.e.the multiplication by i by the formula
iZ = (Z), Z R
2n
.
This complex vector space is canonically isomorphic to C
n
, where we identify the real vector
(x
1
, y
1
, . . . , x
n
, y
n
) R
2n
with a complex vector (z
1
= x
1
+iy
1
, . . . , z
n
x
n
+iy
n
).
52
On the other hand, any complex vector space can be viewed as a real vector space. In order to
do that we just need to forget how to multiply by i. This procedure is called the realication of
a complex vector space. For example, the realication of C
n
is R
2n
. Sometimes to emphasize the
realication operation we will denote the realication of a complex vector space V by V
R
. As the
sets these to objects coincide.
Given a complex vector space V we can dene linear combinations
i
v
i
, where
i
C are
complex numbers, and thus similarly to the real case talk about really dependent, really independent
vectors. Given vectors v
1
, . . . v
n
V we dene its complex span Span
C
(V
1
, . . . , v
n
) by the formula
Span
C
(v
1
, . . . , v
n
) =
_
n
i
v
i
,
i
C.
_
. A basis of a complex vector space is a system of complex linear independent vectors v
1
, . . . , v
n
such that Span
C
(v
1
, . . . , v
n
) = V . The number of vectors in a complex basis is called the complex
dimension of V and denoted dim
C
V .
For instance dimC
n
= n. On the other hand, its realication R
2n
has real dimension 2n. In
particular, C is a complex vector space of dimension 1, and therefore it is called a complex line
rather than a plane.
Exercise 5.3. Let v
1
, . . . , v
n
be a complex basis of a complex vector space V . Find the real basis of
its realication V
R
.
Answer. v
1
, iv
1
, v
2
, iv
2
, . . . , v
n
, iv
n
.
There is another important operation which associates with a real vector space V of real dimen-
sion n a complex vector space V
C
of complex dimension n. It is done in a way similar to how we
made complex numbers out of real numbers. As a real vector space the space V
C
is just the direct
sum V V = (v, w); v, w V . This is a real space of dimension 2n. We then make V V into
a complex vector space by dening the multiplication by i by the formula:
i(v, w) = (w, v).
We will write vector (v, 0) simply by v and (0, v) = i(v, 0) by iv. Hence, every vector of V
C
can be
written as v + iw, where v, w V . If v
1
, . . . , v
n
is a real basis of V , then the same vectors form a
complex basis of V
C
.
53
5.3 Complex linear maps
Complex linear maps and their realications
Given two complex vector spaces V, W a map / : V W is called complex linear (or C-linear)
if /(X + Y ) = /(X) + /(Y ) and /(X) = /(X) for any vectors X, Y V and any complex
number C. Thus complex linearity is stronger condition than the real linearity. The dierence is
in the additional requirement that /(iX) = i/(X). In other words, the operator / must commute
with the operation of multiplication by i.
Any linear map / : C C is a multiplication by a complex number a = c + id. If we view C
as R
2
and right the real matrix of this map in the standard basis 1, i we get the matrix
_
_
c d
d c
_
_
.
Indeed, /(1) = a = c + di and /(i) = ai = d + ci, so the rst column of the matrix is equal to
_
_
c
d
_
_
and the second one is equal to
_
_
d
c
_
_
.
If we have bases v
1
, . . . , v
n
of V and w
1
, . . . w
m
of W then one can associate with / an m n
complex matrix A by the same rule as in the real case.
Recall (see Exercise 5.3) that vectors v
1
, v
1
= iv
1
, v
2
, v
2
= iv
2
, . . . , v
n
, v
n
= iv
n
and w
1
, w
1
=
iw
1
, w
2
, w
2
= iw
2
, . . . w
m
, w
m
= iw
m
) form real bases of the realications V
R
and W
R
of the spaces
V and W. if
A =
_
_
_
_
_
a
11
. . . a
1n
. . .
a
m1
. . . a
mn
_
_
_
_
_
is the complex matrix of / then the real matrix A
R
of the map / is the real basis has order 2n2n
and is obtained from A by replacing each complex element a
kl
= c
kl
+ id
kl
by a 2 2 matrix
_
_
c
kl
d
kl
d
kl
c
kl
_
_
.
Exercise 5.4. Prove that
det A
R
= [ det A[
2
.
54
Complexication of real linear maps
Given a real linear map / : V W one can dene a complex linear map /
C
: V
C
W
C
by the
formula
/
C
(v +iw) = /(v) +i/(w).
. If A is the matrix of / in a basis v
1
, . . . , v
n
then /
C
has the same matrix in the same basis viewed
as a complex basis of V
C
. The operator /
C
is called the complexication of the operator /.
In particular, one can consider C-linear functions V C on a complex vector space V . Complex
coordinates z
1
, . . . , z
n
in a complex basis are examples of C-linear functions, and any other C-linear
function on V has a form c
1
z
1
+. . . c
n
z
n
, where c
1
, . . . , c
n
C are complex numbers.
Complex-valued R-linear functions
It is sometimes useful to consider also C-valued R-linear functions on a complex vector space V ,
i.e. R-linear maps V C (i.e. a linear map V
R
R
2
). Such a C-valued function has the form
= + i, where , are usual real linear functions. For instance the function z on C is a
C-valued R-linear function which is not C-linear.
If z
1
, . . . , z
n
are complex coordinates on a complex vector space V then any R-linear complex-
valued function can be written as
n
1
a
i
z
i
+b
i
z
i
, where a
i
, b
i
C are complex numbers.
We can furthermore consider complex-valued tensors and, in particular complex-valued exterior
forms. A C-valued k-form can be written as + i where and are usual R-valued k-forms.
For instance, we can consider on C
n
the 2-form =
i
2
n
1
z
k
z
k
. It can be rewritten as =
i
2
n
1
(x
k
+iy
k
) (x
k
iy
k
) =
n
1
x
k
y
k
.
55
56
Part II
Calculus of dierential forms
57
Chapter 6
Topological preliminaries
6.1 Elements of topology in a vector space
We recall in this section some basic topological notions in a nite-dimensional vector space and
elements of the theory of continuous functions. The proofs of most statements are straightforward
and we omit them.
Let us choose in V a scalar product.
Notation B
r
(p) := x V ; [[x p[[ < r, D
r
(p) := x V ; [[x p[[ r and S
r
(p) := x
V ; [[x p[[ = r stand for open, closed balls and the sphere of radius r centered at a point p V .
Open and closed sets
A set U V is called open if for any x U there exists > 0 such that B
(x) U.
A set A V is called closed if its complement V A is open. Equivalently,
Lemma 6.1. The set A is closed if and only if for any sequence x
n
A, n = 1, 2, . . . which
converges to a point a V , the limit point a belongs to A.
Remark 6.2. It is important to note that the notion of open and closed sets are independent of
the choice of the auxiliary Euclidean structure in the space V .
Points which appear as limits of sequences of points x
n
A are called limit points of A.
There are only two subsets of V which are simultaneously open and closed: V and .
59
Figure 6.1: Bernard Bolzano (1781-1848)
Lemma 6.3. 1. For any family U
is open.
2. For any family A
is closed.
3. The union
n
1
A
i
of a nite family of closed sets is closed.
4. The intersection
n
1
U
i
of a nite family of open sets is open.
By a neighborhood of a point a V we understand any open set U p.
Given a subset X V
- a subset Y X is called relatively open in Y if there exists an open set U V such that
Y = X U.
- a subset Y X is called relatively closed in Y if there exists a closed set A V such that
Y = X A.
One also call relatively and open and closed subsets of X just open and closed in X.
Given any subset X V a point a V is called
interior point for A if there is a neighborhood U p such that U A;
boundary point if it is not an interior point neither for A nor for its complement V A.
60
We emphasize that a boundary point of A may or may not belong to A. Equivalently, a point a V
is a boundary point of A if it is a limit point both for A and V A.
The set of all interior points of A is called the interior of A and denoted Int A. The set of all
boundary points of A is called the boundary of A and denoted A. The union of all limit points of
A is called the closure of A and denoted A.
Lemma 6.4. 1. We have A = A A, Int A = A A.
2. A is equal to the intersection of all closed sets containg A
3. Int U is the union of all open sets contained in A.
6.2 Everywhere and nowhere dense sets
A closed set A is called nowhere dense if Int A = . For instance any nite set is nowhere dense.
Any linear subspace L V is nowhere dense in V if dimL < dimV . Here is a more interesting
example of a nowhere dense set.
Fix some number
1
3
. For any interval = [a, b] we denote by
. Thus C() consists of two disjoint smaller closed intervals. Let I = [0, 1]. Take
C(I) = I
1
I
2
. Take again C(I
1
) C(I
2
) then again apply the operation C to four new closed
intervals. Continue the process, and take the intersection of all sets arising on all steps of this
construction. The resulted closed set K
1
, . . . , U
k
which cover A, i.e.
k
1
U
k
A.
The equivalence of these denitions is a combination of theorems of Bolzano-Weierstrass and
Emile Borel.
A set A is called path-connected if for any two points a
0
, a
1
A there is a continuous path
: [0, 1] A such that (0) = a
0
and (1) = a
1
.
A set A is called connected if one cannot present A as a union A = A
1
A
2
such that A
1
A
2
= ,
A
1
, A
2
,= and both A
1
and A
2
are simultaneously relatively closed and open in A.
Lemma 6.5. Any path-connected set is connected. If a connected set A is open then it is path
connected.
Proof. Suppose that A is disconnected. Then it can be presented as a union A = A
0
A
1
of two
non-empty relatively open (and hence relatively closed) subsets. Consider the function : A R
dened by the formula
(x) =
_
_
0, x A
0
,
1, x A
1
.
We claim that the function is is continuous. Indeed, For each i = 0, 1 and any point a A
i
there exists > 0 such that B
(x) A A
i
. Hence the function is constant on B
(x) A,
62
Figure 6.3:
Emile Borel (1871-1956)
and hence continuous at the point x. Now take points x
0
A
0
and x
1
A
1
and connect them
by a path : [0, 1] A (this path exists because A is path-connected). Consider the function
:= : [0, 1] R. This function is continuous (as a composition of two continuous maps).
Furthermore, (0) = 0, (1) = 1. Hence, by an intermediate value theorem of Cauchy the function
must take all values in the interval [0, 1]. But this is a contradiction because by construction the
function takes no other values except 0 and 1.
Lemma 6.6. Any open connected subset U R
n
is path connected.
Proof. Take any point a U. Denote by C
a
the set of all points in U which can be connected with
a by a path. We need to prove that C
a
= U.
First, we note that C
a
is open. Indeed, if b C
a
then using openness of U we can nd > 0
such the ball B
(b)
with b, and hence it can be connected by a path with a, i.e. c C
a
. Thus B
(b) C
1
, and hence
C
a
is open. Similarly we prove that the complement U C
a
is open. Indeed, take b / C
a
. As above,
there exists an open ball B
(b) U. Then B
(b) U C
a
. Indeed, if it were possible to connect a
point c B
(b) with a by a path, then the same would be true for b, because b and c are connected
by the interval I
bc
. Thus, we have U = C
a
(U C
a
), both sets C
a
and U C
a
are open and C
a
is
non-empty. Hence, U C
a
is to be empty in view of connectedness of U. Thus, C
a
= U, i.e. U is
path-connected.
63
Exercise 6.7. Prove that any non-empty connected (= path-connected) open subset of R is equal
to an interval (a, b) (we allow here a = and b = ). If one drops the condition of openness,
then one needs to add a closed and semi-closed intervals and a point.
Remark 6.8. One of the corollaries of this exercise is that in R any connected set is path-connected.
Solution. Let A R be a non-empty connected subset. Let a < b be two points of A. Suppose
that a point c (a, b) does not belong to A. Then we can write A = A
0
A
1
, where A
0
=
A (, 0), A
1
= A (0, ). Both sets A
0
and A
1
are relatively open and non-empty, which
contradicts connectedness of A. Hence if two points a and b, a < b, are in A, then the whole
interval [a, b] is also contained in A. Denote m := inf A and M := supA (we assume that m =
if A is unbounded from below and M = + if A is unbounded from above). Then the above
argument shows that the open interval (m, M) is contained in A. Thus, there could be 5 cases:
m, M / A; in this case A = (m, M);
m A, M / A; in this case A = [m, M);
m / A, M A; in this case A = (m, M];
m, M A and m < M; in this case A = [m, M];
m, M A and m = M; in this case A consists of one point.
,= . Then
and A
.
Suppose now that A
of disjoint relatively open subsets, where we denoted by U the set which contains
64
the point a
:= U A
and U
:= U
are
relatively open in A
. We have A
= U
. By assumption, U
a, and hence U
,= . Hence,
connectedness of A
implies that U
= . But then U
Given any set A V and a point a A the connected component (resp. path-connected com-
ponent C
a
A of the point a A is the union of all connected (resp. path-connected) subsets
of A which contains the point a. Due to Lemma 6.9 the (path-)connected component C
a
is itself
(path-)connected, and hence it is the biggest (path-)connected subset of A which contains the point
a. The path-connected component of a can be equivalently dened as the set of all points of A one
can connect with a by a path in A.
Note that (path-)connected components of dierent points either coincide or do not intersect,
and hence the set A can be presented as a disjoint union of (path-)-connected components.
Lemma 6.6 shows that for open sets in a vector space V the notions of connected and path-
connected components coincide, and due to Exercise 6.7 the same is true for any subsets in R. In
particular, any open set U R can be presented as a union of disjoint open intervals, which are its
connected (= path-connected) components. Note that the number of these intervals can be innite,
but always countable.
6.5 Continuous maps and functions
Let V, W be two Euclidean spaces and A is a subset of V . A map f : A W is called continuous
if one of the three equivalent properties hold:
1. For any > 0 and any point x A there exists > 0 such that f(B
(x) A) B
(f(x)).
2. If for a sequence x
n
A there exists limx
n
= x A then the sequence f(x
n
) W converges
to f(x).
3. For any open set U W the pre-image f
1
(U) is relatively open in A.
4. For any closed set B W the pre-image f
1
(B) is relatively closed in A.
65
Let us verify equivalence of 3 and 4. For any open set U W its complement B = W U is closed
and we have f
1
(U) = A f
1
(B). Hence, if f
1
(U) is relatively open, i.e. f
1
(U) = U
A for
an open set U
V , then f
1
(B) = A (V U
), i.e. f
1
(B) is relatively closed. The converse is
similar.
Let us deduce 1 from 3. The ball B
(x) A f
1
(B
(x) A) B
(f(x)).
We leave the converse and the equivalence of denition 2 to the reader.
Remark 6.10. Consider a map f : A W and denote B := f(A). Then denition 3 can be
equivalently stated as follows:
3
W, while f
1
(U) = f
1
(U
).
The following theorem summarize properties of continuous maps.
Theorem 6.11. Let f : A W be a continuous map. Then
1. if A is compact then f(A) is compact;
2. if A is connected then f(A) is connected;
3. if A is path connected then f(A) is path-connected.
Proof. 1. Take any innite sequence y
n
f(A). Then there exist points x
n
A such that y
n
=
f(x
n
), n = 1, . . . . Then there exists a converging subsequence x
n
k
a A. Then by continuity
limk f(x
n
k
) = f(a) f(A), i.e. f(A) is compact.
2. Suppose that f(A) can be presented as a union B
1
B
2
of simultaneously relatively open
and closed disjoint non-empty sets. Then f
1
(B
1
), f
1
(B
2
) A are simultaneously relatively open
and closed in A, disjoint and non-empty. We also have f
1
(B
1
) f
1
(B
2
) = f
1
(B
1
B
2
) =
f
1
(f(A)) = A. Hence A is disconnected which is a contradiction.
3. Take any two points y
0
, y
1
f(A). Then there exist x
0
, x
1
A such that f(x
0
) = y
0
, f(x
1
) =
y
1
. But A is path-connected. Hence the points x
0
, x
1
can be connected by a path : [0, 1] A.
Then the path f : [0, 1] f(A) connects y
0
and y
1
, i.e. f(A) is path-connected.
66
Figure 6.4: George Cantor (1845-1918)
Note that in the case W = R Theorem 6.11.1 is just the Weierstrass theorem: a continuos
function on a compact set is bounded and achieves its maximal and minimal values.
We nish this section by a theorem of George Cantor about uniform continuity.
Theorem 6.12. Let A be compact and f : A W a continuous map. Then for any > 0 there
exists > 0 such that for any x A we have f(B
(x)) B
(f(x)).
Proof. Choose > 0. By continuity of f for every point x A there exists (x) > 0 such that
f(B
(x)
(x)) B
4
(f(x)).
We need to prove that inf
xA
(x) > 0. Note that for any point in y B(x)
2
(x) we have B(x)
2
(y)
B
(x)
(x), and hence f(B(x)
2
(y)) B
(x)) B
x
for each x V . Set
f(x) = T
1
(d
x
f).
71
The vector f(x) is called the gradient of the function f at the point x U. We will also use
the notation gradf(x).
By denition we have
f(x), h) = d
x
f(h) for any vector h V .
If [[h[[ = 1 then d
x
f(h) = [[f(x)[[ cos , where is the angle between the vectors f(x) and h. In
particular, the directional derivative d
x
f(h) has its maximal value when = 0. Thus the direction
of the gradient is the direction of the maximal growth of the function and the length of the gradient
equals this maximal value.
As in the case of a dierential, the gradient varies from point to point, and the family of vectors
f(x)
xU
is called the gradient vector eld f.
We discuss the general notion of a vector eld in Section 7.4 below.
7.4 Vector elds
A vector eld v on a domain U V is a function which associates to each point x U a vector
v(x) V
x
, i.e. a vector originated at the point x.
A gradient vector eld f of a function f provides us with an example of a vector eld, but as
we shall see, gradient vector elds form only a small very special class of vector elds.
Let v be a vector eld on a domain U V . If we x a basis in V , and parallel transport this
basis to all spaces V
x
, x V , then for any point x V the vector v(x) V
x
is described by its
coordinates (v
1
(x), v
2
(x), . . . , v
n
(x)). Thus to dene a vector eld on U is the same as to dene n
functions v
1
, . . . , v
n
on U, i.e. to dene a map (v
1
, . . . , v
n
) : U R
n
.
Thus, if a basis of V is xed, then the dierence between the maps U R
n
and vector elds
on U is just a matter of geometric interpretation. When we speak about a vector eld v we view
v(x) as a vector in V
x
, i.e. originated at the point x U. When we speak about a map v : U R
n
we view v(x) as a point of the space V , or as a vector, with its origin at 0 V .
Vector elds naturally arise in a context of Physics, Mechanics, Hydrodynamics, etc. as force,
velocity and other physical elds.
72
There is another very important interpretation of vector elds as rst order dierential opera-
tors.
Let C
(U) C
(U),
given by the formula
D
v
(f) = df(v), f C
(U).
In other words, we compute at any point x U the directional derivative of f in the direction of
the vector v(x). Clearly, the operator D
v
is linear: D
v
(af +bg) = aD
v
(f)+bD
v
(g) for any functions
f, g C
(U) and any real numbers a, b R. It also satises the Leibniz rule:
D
v
(fg) = D
v
(f)g +fD
v
(g).
In view of the above correspondence between vector elds and rst order dierential operatopts
, it is sometimes convenient just to view a vector eld as a dierential operator. Hence, when it will
not be confusing we may drop the notation D
v
and just directly apply the vector v to a function
f.
Let v
1
, . . . , v
n
be a basis of V , and x
1
, . . . , x
n
be the coordinate functions in this basis. We
would like to introduce the notation for the vector eld obtained from vectors v
1
, . . . , v
n
by parallel
transporting them to all points of the domain U. To motivate the notation which we are going
to introduce, let us temporarily denote these vector elds by v
1
, . . . , v
n
. Observe that D
v
i
(f) =
f
x
i
, i = 1, . . . , n. Thus the operator D
v
i
is just the operator
x
i
of taking i-th partial derivative.
Hence, viewing the vector eld v
i
as a dierential operator we will just use the notation
x
i
instead
of v
i
. Given any vector eld v with coordinate functions a
1
, a
2
, . . . , a
n
: U R we have
D
v
(f)(x) =
n
i=1
a
i
(x)
f
x
i
(x), for any f C
(U),
and hence we can write v =
n
i=1
a
i
x
i
. Note that the coecients a
i
here are functions and not
constants.
73
Suppose that V, , ) is a Euclidean vector space. Choose a (not necessarily orthonormal) basis
v
1
, . . . , v
n
. Let us nd the coordinate description of the gradient vector eld f, i.e. nd the
coecients a
j
in the expansion f(x) =
n
1
a
i
(x)
x
i
. By denition we have
f(x), h) = d
x
f(h) =
n
1
f
x
j
(x)h
j
(7.4.1)
for any vector h V
x
with coordinates (h
1
, . . . , h
n
) in the basis v
1
, . . . , v
n
parallel transported to
V
x
. Let us denote g
ij
= v
i
, v
j
). Thus G = (g
ij
) is a symmetric n n matrix, which is called the
Gram matrix of the basis v
1
, . . . , v
n
. Then the equation (7.4.1) can be rewritten as
n
i,j=1
g
ji
a
i
h
j
=
n
1
f
x
j
(x)h
j
.
Because h
j
are arbitrarily numbers it implies that the coecients with h
j
in the right and left sides
should coincide for all j = 1, . . . , n. Hence we get the following system of linear equations:
n
i=1
g
ij
a
i
=
f
x
j
(x), j = 1, . . . , n, (7.4.2)
or in matrix form
G
_
_
_
_
_
a
1
.
.
.
a
n
_
_
_
_
_
=
_
_
_
_
_
f
x
1
(x)
.
.
.
f
xn
(x)
_
_
_
_
_
,
and thus
_
_
_
_
_
a
1
.
.
.
a
n
_
_
_
_
_
= G
1
_
_
_
_
_
f
x
1
(x)
.
.
.
f
xn
(x)
_
_
_
_
_
, (7.4.3)
i.e.
f =
n
i,j=1
g
ij
f
x
i
(x)
x
j
, (7.4.4)
where we denote by g
ij
the entries of the inverse matrix G
1
= (g
ij
)
1
If the basis v
1
, . . . , v
n
is orthonormal then G is the unit matrix, and thus in this case
f =
n
1
f
x
j
(x)
x
j
, (7.4.5)
74
i.e. f has coordinates (
f
x
1
, . . . ,
f
xn
). However, simple expression (7.4.5) for the gradient holds
only in the orthonormal basis. In the general case one has a more complicated expression
(7.4.4).
7.5 Dierential forms
Similarly to vector elds, we can consider elds of exterior forms, i.e. functions on U V which
associate to each point x U a k-form from
k
(V
x
). These elds of exterior k-forms are called
dierential k-forms.
Thus the relation between k-forms and dierential k-forms is exactly the same as the relation
between vectors and vector-elds. For instance, a dierential 1-form associates with each point
x U a linear function (x) on the space V
x
. Sometimes we will write
x
instead of (x) to leave
space for the arguments of the function (x).
Example 7.2. 1. Let f : V R be a smooth function. Then the dierential df is a dierential
1-form. Indeed, with each point x V it associates a linear function d
x
f on the space V
x
.
As we shall see, most dierential 1-form are not dierentials of functions (just as most vector
elds are not gradient vector elds).
2. A dierential 0-form f on U associates with each point x U a 0-form on V
x
, i. e. a number
f(x) R. Thus dierential 0-forms on U are just functions U R.
7.6 Coordinate description of dierential forms
Let x
1
, . . . , x
n
be coordinate linear functions on V , which form the basis of V
dual to a chosen
basis v
1
, . . . , v
n
of V . For each i = 1, . . . , n the dierential dx
i
denes a linear function on each
space V
x
, x V . Namely, if h = (h
1
, . . . , h
n
) V
x
then dx
i
(h) = h
i
. Indeed
d
x
x
i
(h) = lim
t0
x
i
+th
i
x
i
t
= h
i
,
independently of the base point x V . Thus dierentials dx
1
, . . . , dx
n
form a basis of the space V
x
for each x V . In particular, any dierential 1-form on v can be written as
75
= f
1
dx
1
+. . . +f
n
dx
n
,
where f
1
, . . . , f
n
are functions on V . In particular,
df =
f
x
1
dx
1
+. . . +
f
x
n
dx
n
. (7.6.1)
Let us point out that this simple expression of the dierential of a function holds in an arbitrary
coordinate system, while an analogous simple expression (7.4.5) for the gradient vector eld
is valid only in the case of Cartesian coordinates. This reects the fact that while the notion of
dierential is intrinsic and independent of any extra choices, one needs to have a background inner
product to dene the gradient.
Similarly, any dierential 2-form w on a 3-dimensional space can be written as
= b
1
(x)dx
2
dx
3
+b
2
(x)dx
3
dx
1
+b
3
(x)dx
1
dx
2
where b
1
, b
2
, and b
3
are functions on V . Any dierential 3-form on a 3-dimensional space V has
the form
= c(x)dx
1
dx
2
dx
3
for a function c on V .
More generally, any dierential k-form can be expressed as
=
1i
1
<i
2
<i
k
n
a
i
1
...i
k
dx
i
1
dx
i
k
for some functions a
i
1
...i
k
on V .
7.7 Smooth maps and their dierentials
Let V, W be two vector spaces of arbitary (not, necessarily, equal) dimensions and U V be an
open domain in V .
Recall that a map f : U W is called dierentiable if for each x U there exists a linear map
l : V
x
W
f(x)
76
such that
l(h) = lim
t0
f(x +th) f(x)
t
for any h V
x
. In other words,
f(x +th) f(x) = tl(h) +o(t), where
o(t)
t
t0
0.
The map l is denoted by d
x
f and is called the dierential of the map f at the point x U. Thus,
d
x
f is a linear map V
x
W
f(x)
.
The space W
f(x)
can be identied with W via a parallel transport, and hence sometimes it is
convenient to think about the dierential as a map V
x
W, In particular, in the case of a linear
function. i.e. when W = R it is customary to do that, and hence we dened earlier in Section 7.1
the dierential of a function f : U R at a point x U as a linear function V
x
R, i.e. an
element of V
x
, rather than a linear map V
x
W
f(x)
.
Let us pick bases in V and W and let (x
1
, . . . , x
k
) and (y
1
, . . . , y
n
) be the corresponding coor-
dinate functions. Then each of the spaces V
x
and W
y
, x V, y W inherits a basis obtained by
parallel transport of the bases of V and W. In terms of these bases, the dierential d
x
f is given by
the Jacobi matrix
_
_
_
_
_
f
1
x
1
. . .
f
1
x
k
. . . . . . . . .
fn
x
1
. . .
fn
x
k
_
_
_
_
_
In what follows we will consider only suciently smooth maps, i.e. we assume that all maps
and their coordinate functions are dierentiable as many times as we need it.
7.8 Operator f
Let U be a domain in a vector space V and f : U W a smooth map. Then the dierential df
denes a linear map
d
x
f : V
x
W
f(x)
77
for each x V .
Let be a dierential k-form on W. Thus denes an exterior k-form on the space W
y
for
each y W.
Let us dene the dierential k-form f
on U by the formula
(f
)[
Vx
= (d
x
f)
([
W
f(x)
).
Here the notation [
Wy
stands for the exterior k-form dened by the dierential form on the
space W
y
.
In other words, for any k vectors, H
1
, . . . , H
k
V
x
we have
f
(H
1
, . . . , H
k
) = (d
x
f(H
1
), . . . , d
x
f(H
k
)).
We say that the dierential form f
is the pull-back
of by f.
Example 7.3. Let = h(x)dx
1
dx
n
. Then formula (3.2) implies
f
= h f det Dfdx
1
dx
n
.
Here
det Df =
f
1
x
1
. . .
f
1
xn
. . . . . . . . .
fn
x
1
. . .
fn
xn
() = f
(g
) .
78
An important special case of the pull-back operator f
on the domain U
U
. Thus the restricted form [
U
is the same form but viewed as function of a point a U
and vectors T
1
, . . . , T
k
L
a
.
7.9 Coordinate description of the operator f
). Let us x coordinate
systems x
1
, . . . , x
k
in V and y
1
, . . . , y
n
in W. If A is the matrix of the map / then we already have
seen in Section 2.7 that
/
y
j
= l
j
(x
1
, . . . , x
k
) = a
j1
x
1
+a
j2
x
2
+. . . +a
jk
x
k
, j = 1, . . . , n,
and that for any exterior k-form
=
1i
1
<...<ipn
A
i
1
,...,ip
y
i
1
. . . y
ip
we have
/
1i,<...<ipn
A
i
1
...ip
l
i
1
. . . l
ip
.
Now consider the non-linear situation. Let be a dierential p-form on W. Thus it can be
written in the form
=
A
i
1
...ip
(y)dy
i
1
. . . dy
ip
for some functions A
i
1
...ip
on W .
Let U be a domain in V and f : U W a smooth map.
Proposition 7.5. f
A
i
1
...,ip
(f(x))df
i
1
. . . df
ip
, where f
1
, . . . , f
n
are coordinate functions
of the map f.
79
Proof. For each point x U we have, by denition,
f
[
Vx
= (d
x
f)
([
W
fx
)
But the coordinate functions of the linear map d
x
f are just the dierentials d
x
f
i
of the coordinate
functions of the map f. Hence the desired formula follows from the linear case proven in the previous
proposition.
7.10 Examples
1. Consider the domain U = r > 0, 0 < 2 on the plane V = R
2
with cartesian coordinates
(r, ). Let W = R
2
be another copy of R
2
with cartesian coordinates (x, y). Consider a map
P : V W given by the formula
P(r, ) = (r cos , r sin).
This map introduces (r, ) as polar coordinates on the plane W. Set = dx dy. It is called
the area form on W. Then
P
i
1
<...<i
k
da
i
1
...i
k
dx
i
1
. . . dx
i
k
.
Examples. 1. Let w
1
(U), i.e. w =
n
i=1
a
i
dx
i
. Then
83
Figure 8.1:
Elie Cartan (18691951)
dw =
n
i=1
da
i
dx
i
=
n
i=1
_
_
n
j=1
a
i
x
j
dx
j
_
_
dx
i
=
1i<jn
_
a
j
x
i
a
i
x
j
_
dx
i
dx
j
.
For instance, when n = 2 we have
d(a
1
dx
1
+a
2
dx
2
) =
_
a
2
x
1
a
1
x
2
_
dx
1
dx
2
For n = 3, we get
d(a
1
dx
1
+a
2
dx
2
+a
3
dx
3
) =
_
a
3
x
2
a
2
x
3
_
dx
2
dx
3
+
_
a
1
x
3
a
3
x
1
_
dx
3
dx
1
+
_
a
2
x
1
a
1
x
2
_
dx
1
dx
2
.
2. Let n = 3 and w
2
(U). Then
w = a
1
dx
2
dx
3
+a
2
dx
3
dx
1
+a
3
dx
1
dx
2
and
dw = da
1
dx
2
dx
3
+da
2
dx
3
dx
1
+da
3
dx
1
dx
2
=
_
a
1
x
1
+
a
2
x
2
+
a
3
x
3
_
dx
1
dx
2
dx
3
84
Figure 8.2: Henri Poincare (18541912)
3. For 0-forms, i.e. functions the exterior dierential coincides with the usual dierential of a
function.
8.2 Properties of the operator d
Proposition 8.1. For any two dierential forms,
k
(U),
l
(U) we have
d( ) = d + (1)
k
d.
Proof. We have
=
i
1
<...<i
k
a
i
1
...i
k
dx
i
1
. . . dx
i
k
=
j
1
<...<j
l
b
j
i
...j
l
dx
j
1
. . . dx
j
l
=
_
_
i
1
<...<i
k
a
i
1
...i
k
dx
i
1
. . . dx
i
k
_
_
_
_
j
i
<...<j
l
b
j
1
...j
l
dx
j
1
. . . dx
j
l
_
_
=
i
1
<...<i
k
j
1
<...<j
l
a
i
1
...i
k
b
j
1
...j
l
dx
i
l
. . . dx
i
k
dx
j
l
. . . dx
j
l
85
d( ) =
i
1
<...<i
k
j
1
<...<j
l
(b
j
1
...j
l
da
i
1
...i
k
+a
i
1
...i
k
db
j
1
...j
l
) dx
i
1
. . . dx
i
k
dx
j
1
. . . dx
j
l
=
i
1
<...<i
k
j
1
<...<j
l
b
j
1
...j
l
da
i
1
...i
k
dx
i
1
. . . dx
i
k
dx
j
l
. . . dx
j
l
+
i
1
<...<i
k
j
1
<...<j
l
a
i
1
...i
k
db
j
1
...j
l
dx
i
1
. . . dx
i
k
dx
j
1
. . . dx
j
l
=
_
_
i
1
<...<i
k
da
i
1
...i
k
dx
i
1
...
dx
i
k
_
_
b
j
1
...j
l
dx
j
1
. . . dx
j
l
_
+ (1)
k
_
_
i
1
<...<i
k
a
i
1
...i
k
dx
i
1
. . . dx
i
k
_
_
_
_
j
i
<...<j
l
db
j
1
...j
l
dx
j
1
. . . dx
j
l
_
_
= d + (1)
k
d.
Notice that the sign (1)
k
appeared because we had to make k transposition to move db
j
1
...j
l
to its
place.
Proposition 8.2. For any dierential k-form w we have
ddw = 0.
Proof. Let w =
i
1
<...<i
k
a
i
1
...i
k
dx
i
1
. . . dx
i
k
. Then we have
dw =
i
1
<...<i
k
da
i
1
...i
k
dx
i
1
. . . dx
i
k
.
Applying Proposition 8.1 we get
ddw =
i
1
<...<i
k
dda
i
1
...i
k
dx
i
1
. . . dx
i
k
da
i
1
...i
k
ddx
i
1
. . . dx
i
k
+. . .
+ (1)
k
da
i
1
...i
k
dx
i
1
. . . ddx
i
k
.
86
But ddf = 0 for any function f as was shown above. Hence all terms in this sum are equal to 0, i.e.
ddw = 0.
Denition. A k-form is called closed if d = 0. It is called exact if there exists a (k 1)-form
such that d = . The form is called the primitive of the form . The previous theorem can be
reformulated as follows:
Corollary 8.3. Every exact form is closed.
The converse is not true in general. For instance, take a dierential 1-form
=
xdy ydx
x
2
+y
2
on the punctured plane U = R
2
0 (i.e the plane R
2
with the deleted origin). It is easy to calculate
that d = 0, i.e is closed. On the other hand it is not exact. Indeed, let us write down this form
in polar coordinates (r, ). We have
x = r cos , y = r sin.
Hence,
=
1
r
2
(r cos (sindr +r cos d) r sin(cosdr r sind)) = d .
If there were a function H on U such that dH = , then we would have to have H = + const,
but this is impossible because the polar coordinate is not a continuous univalent function on U.
Hence is not exact.
However, a closed form is exact if it is dened on the whole vector space V .
Proposition 8.4. Operators f
w = f
dw
.
Proof. Suppose rst that k = 0, i.e. w is a function : W R. Then f
= f. Then
d( f) = f
(d(X)).
Consider now the case of arbitrary k-form w,
w =
i
1
<...<i
k
a
i
1
...i
k
dx
i
1
. . . dx
i
k
.
Then
f
w =
i
1
<...<i
k
a
i
1
...i
k
fdf
i
1
. . . df
i
k
where f
1
, . . . , f
n
are coordinate functions of the map f. Using the previous theorem and taking into
account that d(df
i
) = 0, we get
d(f
w) =
i
1
<...<i
k
d(a
i
1
...i
k
f) df
i
1
. . . df
i
k
.
On the other hand
dw =
i
1
<...<i
k
da
i
1
...i
k
dx
i
1
. . . dx
i
k
and therefore
f
dw =
i
1
<...<i
k
f
(da
i
1
...i
k
) df
i
1
. . . df
i
k
.
But according to what is proven above, we have
f
da
i
1
...i
k
= d(a
i
1
...i
k
f)
Thus,
f
dw =
i
1
<...<i
k
d(a
i
1
...i
k
f) df
i
1
. . . df
i
k
= df
The above theorem shows, in particular, that the denition of the exterior dierential is in-
dependent of the choice of the coordinate. Moreover, one can even use non-linear (curvilinear)
coordinate systems, like polar coordinates on the plane.
88
8.3 Curvilinear coordinate systems
A (non-linear) coordinate system on a domain U in an n-dimensional space V is a smooth map
f = (f
1
, . . . , f
n
) : U R
n
such that
1. For each point x U the dierentials d
x
f
1
, . . . , d
x
f
n
(V
x
)
= f(U) R
n
is the image of U under the map f. If one already has another set of coordinates x
1
. . . x
n
on
U, then the coordinate map f expresses new coordinates y
1
. . . y
n
through the old one, while the
parametrization map expresses the old coordinate through the new one. Thus the statement
g
dw = dg
w
applied to the parametrization map g just tells us that the formula for the exterior dierential is
the same in the new coordinates and in the old one.
Consider a space R
n
with coordinates (u
1
, . . . , u
n
). The j-th coordinate line is given by equations
u
i
= c
i
, i = 1, . . . , n; i ,= j. Given a domain U
R
n
consider a parameterization map g : U
U V . The images gu
i
= c
i
, i ,= j) U of coordinates lines u
i
= c
i
, i ,= j U
are called
coordinate lines in U with respect to the curvilinear coordinate system (u
1
, . . . , u
n
). For instance,
coordinate lines for polar coordinates in R
2
are concentric circles and rays, while coordinate lines
for spherical coordinates in R
3
are rays from the origin, and latitudes and meridians on concentric
spheres.
8.4 Geometric denition of the exterior dierential
We will show later (see Lemma 10.2) that one can give another equivalent denition of the operator
d) without using any coordinates at all. But as a rst step we give below an equivalent denition
of the exterior dierential which is manifestly invariant of a choice of ane coordinates.
89
Given a point a V and vectors h
1
, . . . , h
k
V
a
we will denote by P
a
(h
1
, . . . , h
k
) the k-
dimensional parallelepiped
_
a +
1
2
k
1
u
j
h
j
; [u
1
[, . . . , [u
k
[ 1
_
centered at the point a with its sides parallel to the vectors h
1
, . . . , h
k
, see Fig. ??. For instance,
for k = 1 the parallelepiped P
a
(h) is an interval centered at a of length [a[.
As it was explained above in Section 8.1 for a 0-form f
0
(U), i.e. a function f : U R its
exterior dierential is just its usual dierential, and hence it can be dened by the formula
(df)
a
(h) = d
a
f(h) = lim
t0
1
t
_
f
_
a +
th
2
_
f
_
a
th
2
__
, t R, h V
a
, a U
Proposition 8.5. Let
k1
1
(U) be a dierential (k 1)-form on a domain U R
n
. Then for
any point a U and any vectors h
1
, . . . , h
k
V
a
we have
(d)
a
(h
1
, . . . , h
k
) = lim
t0
1
t
k
k
j=1
(1)
k1
_
a+
th
j
2
(th
1
, . . . ,
j
, . . . , th
k
)
a
th
j
2
(th
1
, . . . ,
j
, . . . , th
k
)
_
= lim
t0
1
t
k
j=1
(1)
k1
_
a+
th
j
2
(h
1
, . . . ,
j
, . . . , h
k
)
a
th
j
2
(h
1
, . . . ,
j
, . . . , h
k
)
_
(8.4.1)
For instance, for a 1-form
1
(U), we have
(d)
a
(h
1
, h
2
) := lim
t0
1
t
2
_
a+
th
1
2
(th
2
) +
a
th
1
2
(th
2
) +
a+
th
2
2
(th
1
) +
a
th
2
2
(th
1
)
_
= lim
t0
1
t
_
a+
th
1
2
(h
2
)
a
th
1
2
(h
2
)
a+
th
2
2
(h
1
) +
a
th
2
2
(h
1
)
_
Proof. First we observe that for a C
1
-smooth form the limit in the right-hand side of formula
(8.4.1) exists and dene a k-form (i.e. for each a it is a multilinear skewsymmetric function of
vectors h
1
, . . . , h
k
R
n
a
). We can also assume that the vectors h
1
, . . . , h
k
are linearly independent.
Otherwise, both denitions give 0. Denote L
a
= Span(h
1
, . . . , h
k
) V
a
and consider the restriction
:= [
LU
. Let (y
1
, . . . , y
k
) be coordinates in L
a
corresponding to the basis h
1
, . . . , h
k
. Then we
90
can write
:=
P
i
dy
1
. . .
i
dy
k
and thus
lim
t0
1
t
_
_
k
j=1
a+
th
j
2
(h
1
, . . . ,
i
, . . . , h
k
)
a+
th
j
2
(h
1
, . . . ,
i
, . . . , h
k
)
_
_
=
k
j=1
k
i=1
lim
t0
P
j
(a +
th
j
2
) P
j
(a
th
j
2
)
t
dy
1
. . .
i
dy
k
(h
1
, . . . ,
j
, . . . , h
k
)
=
k
j=1
k
i=1
P
j
y
j
(a)dy
1
. . .
i
dy
k
(h
1
, . . . ,
i
, . . . , h
k
).
(8.4.2)
But
dy
1
. . .
i
dy
k
(h
1
, . . . ,
j
, . . . , h
k
) =
_
_
(1)
i1
, if i = j
0, if i ,= j.
Hence, the expression in (8.4.2) is equal to
k
1
(1)
i1 P
i
y
i
(a). But
d
=
_
k
1
(1)
i1
P
i
y
i
_
dy
1
dy
k
,
and therefore (d
)
a
(h
1
, . . . , h
k
) =
k
1
(1)
i1 P
i
y
i
(a), i.e. the two denitions of the exterior dieren-
tial coincide.
8.5 More about vector elds
Similarly to the case of linear coordinates, given any curvilinear coordinate system (u
1
, . . . , u
n
) in
U, one denotes by
u
1
, . . . ,
u
n
the vector elds which correspond to the partial derivatives with respect to the coordinates u
1
, . . . , u
n
.
In other words, the vector eld
u
i
is tangent to the u
i
-coordinate lines and represents the the ve-
locity vector of the curves u
1
= const
1
, . . . , u
i1
= const
i1
, u
i+1
= const
i+1
, . . . , u
n
= const
n
,
parameterized by the coordinate u
i
.
For instance, suppose we are given spherical coordinates (r, , ), r 0, [0, ], [0, 2) in
91
R
3
. The spherical coordinates are related to the cartesian coordinates (x, y, z) by the formulas
x = r sincos , (8.5.1)
y = r sinsin, (8.5.2)
z = r cos . (8.5.3)
Then the vector elds
r
,
, and
are mutually orthogonal and dene the same orientation of the space as the standard Cartesian
coordinates in R
3
. We also have [[
r
[[ = 1. However the length of vector elds
and
vary. When
r and are xed and varies, then the corresponding point (r, , ) is moving along a meridian of
radius r with a constant angular speed 1. Hence,
[[
[[ = r .
When r and are xed and varies, then the point (r, , ) is moving along a latitude of radius
r sin with a constant angular speed 1. Hence,
[[
[[ = r sin.
Note that it is customary to introduce unit vector elds in the direction of
r
,
and
:
e
r
:=
r
, e
:=
1
r
, e
=
1
r sin
,
which form an orthonormal basis at every point. The vector elds e
r
, e
and e
x
1
, . . . ,
xn
. Indeed, for any function f : U R we have
f
u
i
=
n
j=1
f
x
j
x
j
u
i
,
and, therefore,
u
i
=
n
j=1
x
j
u
i
x
j
,
92
For instance, spherical coordinates (r, , ) are related to the cartesian coordinates (x, y, z) by
the formulas (8.5.1), and hence we derive the following expression of the vector elds
r
,
r
= sincos
x
+ sinsin
y
+ cos
z
,
= r sinsin
x
+r sincos
y
,
= r cos cos
x
+r cos sin
y
r sin
z
.
8.6 Case n = 3. Summary of isomorphisms
Let U be a domain in the 3-dimensional space V . We will consider 5 spaces associated with U.
0
(U) = C
k
(U) for k = 1, 2, 3the spaces of dierential k-forms on U;
Vect(U)the space of vector elds on U.
Let us x a volume form w
3
(U) that is any nowhere vanishing dierential 3-form. In coordinates
w can be written as
w = f(x)dx
1
dx
2
dx
3
where the function f : U R is never equal to 0. The choice of the form w allows us to dene the
following isomorphisms.
1.
w
: C
(U)
3
(U),
w
(h) = hw for any function h C
(U).
2.
w
: Vect(U)
2
(U)
w
(v) = v w.
Sometimes we will omit the subscript w an write just and .
Our third isomorphism depends on a choice of a scalar product <, > in V . Let us x a scalar
product. This enables us to dene an isomorphism
T = T
<,>
: Vect(U)
1
(U)
93
which associates with a vector eld v on U a dierential 1-form T(v) = v, .). Let us write down the
coordinate expressions for all these isomorphisms. Fix a cartesian coordinate system (x
1
, x
2
, x
3
) in
V so that the scalar product x, y) in these coordinates equals x
1
y
1
+ x
2
y
2
+ x
3
y
3
. Suppose also
that w = dx
1
dx
2
dx
3
. Then (h) = hdx
1
dx
2
dx
3
.
(v) = v
1
dx
2
dx
3
+v
2
dx
3
dx
1
+v
3
dx
1
dx
2
where v
1
, v
2
, v
3
are coordinate functions of the vector eld v.
T(v) = v
1
dx
1
+v
2
dx
2
+v
3
dx
3
.
If V is an oriented Euclidean space then one also has isomorphisms
:
k
(V )
3k
(V ), k = 0, 1, 2, 3.
If w is the volume form on V for which the unit cube has volume 1 and which dene the given
orientation of V (equivalently, if w = x
1
x
2
x
3
for any Cartesian positive coordinate system on
V ), then
w
(v) = T(v), and
w
= :
0
(V )
3
(V ) .
8.7 Gradient, curl and divergence of a vector eld
The above isomorphism, combined with the operation of exterior dierentiation, allows us to dene
the following operations on the vector elds. First recall that for a function f C
(U),
gradf = T
1
(df).
Now let v Vect (U) be a vector eld. Then its divergence div v is the function dened by the
formula
div v =
1
(d( v))
In other words, we take the 2-form v w (w is the volume form) and compute its exterior dierential
d(v w). The result is a 3-form, and, therefore is proportional to the volume form w, i.e. d(v w) = hw.
94
This proportionality coecient (which is a function; it varies from point to point) is simply the
divergence: div v = h.
Given a vector eld v, its curl is as another vector eld curl v dened by the formula
curl v :=
1
d(Tv) = T
1
d(Tv).
If one xes a cartesian coordinate system in V such that w = dx
1
dx
2
dx
3
and x, y) =
x
1
y
1
+x
2
y
2
+x
3
y
3
then we get the following formulas
grad f =
_
f
x
1
,
f
x
2
,
f
x
3
_
div v =
v
1
x
1
+
v
2
x
2
+
v
3
x
3
curl v =
_
v
3
x
2
v
2
x
3
,
v
1
x
3
v
3
x
1
,
v
2
x
1
v
1
x
2
_
where v = (v
1
, v
2
, v
3
).
We will discuss the geometric meaning of these operations later in Section ??.
8.8 Example: expressing vector analysis operations in spherical
coordinates
Given spherical coordinates (r, , ) in R
3
, consider orthonormal vector elds
e
r
=
r
, e
=
1
r
and e
=
1
r sin
tangent to the coordinate lines and which form orthonormal basis at every point.
Operator T.
Given a vector eld v = a
1
e
r
+ a
2
e
+ a
3
e
+h
3
e
+h
3
e
)
= c
1
h
1
dr(e
r
) +c
2
h
2
d(e
) +c
3
h
3
d(e
) = c
1
h
1
+
c
2
r
h
2
+
c
3
r sin
h
3
.
Hence, c
1
= a
1
, c
2
= ra
2
, c
3
= ra
3
sin, i.e.
Tv = a
1
dr +ra
2
d +ra
3
sind. (8.8.1)
Operator . First we express the volume form = dx dy dz in spherical coordinates. One
way to do that is just to plug into the form the expression of x, y, z through spherical coordinates.
But we can also argue as follows. We have = cdr d d. Let us evaluate both sides of this
equality on vectors e
r
, e
, e
. Then (e
r
, e
, e
, e
)
=
1
r
2
sin
(dr d d)(
r
,
) =
1
r
2
sin
,
and therefore c = r
2
sin, i.e.
= r
2
sindr d d.
In particular, given a 3-form = adr d d we get =
a
r
2
sin
.
Let us now compute dr, d and d. We argue that
dr = Ad d, d = Bdr d, d = Cd dr.
Indeed, suppose that dr = Ad d +A
d dr +A
, e
r
:
dr(e
, e
r
) = dr(e
) = 0,
because the 2-dimensional volume (i.e the area) of the 2-dimensional parallelepiped P(e
, e
r
) and
the 1-dimensional volume (i.e. the length) of the 1-dimensional parallelepiped P(e) are both equal
to 1. On the other hand,
(Ad d +A
d dr +A
dr d)(e
, e
r
) =
1
r sin
A
,
96
and therefore A
= 0. Similarly, A
) = dr(e
r
) = 1 = Ad d(e, e
) =
A
r
2
sin
,
d(e
r
, e
) = d(e
) =
1
r sin
= Bdr d(e
r
, e
) =
B
r
,
d(e
, e
r
) = d(e
) =
1
r
= Cd dr(e
, e
r
) =
C
r sin
.
Thus, i.e.
A = r
2
sin, B =
1
sin
, C = sin,
dr = r
2
sind d, d =
1
sin
dr d, d = sind dr . (8.8.2)
Hence we also have
d d =
1
r
2
dr,
dr d = sind,
d dr =
1
sin
d.
(8.8.3)
Now we are ready to express the vector analysis operations in spherical coordinates.
Exercise 8.6. (Gradient) Given a function f(r, , ) expressed in cylindrical coordinates, com-
pute f.
Solution. We have
f = D
1
(f
r
dr +f
d +f
d) = f
r
e
r
+
f
r
e
+
f
r sin
e
.
Here we denote by f
r
, f
and f
+ce
+ce
) r
2
sindr d d
= ar
2
sind d +cr dr d +br sind dr
and
d( v ) =
_
(r
2
a
r
+ 2ra) sin +rc
+rb
r sin
+
b
r
+
b
r
cot ) .
Hence,
div v = a
r
+
2a
r
+
c
r sin
+
b
r
+
b
r
cot . (8.8.4)
b) Curl. We have
curl v = T
1
dTv = T
1
d(adr +rbd +rc sind)
= T
1
_
(a
b rb
r
)dr d + (c sin +rc
r
sin a
)d dr
+ (rb
rc
sin rc cos )d d
_
= T
1
_
rb
rc
sin rc cos
r
2
dr +
c sin +rc
r
sin a
sin
d
+ (a
b rb
r
) sind
_
=
rb
rc
sin rc cos
r
2
e
r
+
c sin +rc
r
sin a
r sin
e
+
a
b rb
r
r
e
.
(8.8.5)
( +i) = f
+if
, d( +i) = d +id.
We will be in particular interested in complex valued on C. Note that a complex-valued function
(or 0-form) is on a domain U C is just a map f = u+iv : U C. Its dierential df is the same as
the dierntial of this map, but it also can be viewed as a C-valued dierential 1-form df = du+idv.
Example 8.8.
dz = dx +idy, d z = dx idy, zdz = (x +iy)(dx +idy) = xdx ydy +i(xdy +ydx),
dz d z = (dx +idy) (dx idy) = 2idx dy.
Exercise 8.9. Prove that d(z
n
) = nz
n1
dz for any integer n ,= 0.
Solution. Let us do the computation in polar coordinates. Then z
n
= r
n
e
in
and assuming
that n ,= 0 we have
d(z
n
) = nr
n1
e
in
dr +inr
n
e
in
d = nr
n1
e
in
(dr +id).
On the other hand,
nz
n1
dz = nr
n1
e
i(n1)
d(re
i
) = nr
n1
e
i(n1)
_
e
i
dr +ie
i
d
_
= nr
n1
e
in
(dr +id).
Comparing the two expressions we conclude that d(z
n
) = nz
n1
dz.
It follows that the 1-form
dz
z
n
is exact on C 0 for n > 1. Indeed,
dz
z
n
= d
_
1
(1 n)z
n1
_
.
On the other hand the form
dz
z
is closed on C 0 but not exact. Indeed,
dz
z
=
d(re
i
)
re
i
=
e
i
dr +ire
i
d
re
i
=
dr
r
+id,
and hence d
_
dz
z
_
= 0.
On the other hand, we already had seen that the form d, and hence
dz
z
, is not exact.
99
100
Chapter 9
Integration of dierential forms and
functions
9.1 Useful technical tools: partition of unity and cut-o functions
Let us recall that the support of a function is the closure of the set of points where it is not
equal to 0. We denote the support by Supp(). We say that is supported in an open set U if
Supp() U.
Lemma 9.1. There exists a C
_
e
1
x
2
, x > 0
0 , x 0.
(9.1.1)
The function e
1
x
2
has the property that all its derivatives at 0 are equal to 0, and hence the
function h is C
-smooth. Then the function (x) := h(1 +x)h(1 x) has the required properties.
101
Lemma 9.2. Existence of cut-off functions Let C V be compact set and U C its open
neighborhood. Then there exists a C
-smooth function
C,U
: V [0, ) with its support in U
which is equal to 1 on C
Proof. Let us x a Euclidean structure in V and a Cartesian coordinate system. Thus we can
identify V with R
n
with the standard dot-product. Given a point a V and > 0 let us denote
by
a,
the bump function on V dened by
a,
(x) :=
_
[[x a[[
2
2
_
, (9.1.2)
where : R
-function
with Supp(
a,
) = D
:= B
(a).
Let us denote by U
(z
1
), . . . , B
(z
N
) U cover U
2
(C), i.e. U
2
(C)
N
1
B
(z
j
).
Consider a function
1
:=
N
z
i
,
2
: V R.
The function
1
is positive on U
2
(C) and has Supp(
1
) U.
The complement E = V U
2
(C) is a closed but unbounded set. Take a large R > 0 such that
B
R
(0) U. Then E
R
= D
R
(0) U
2
(C) is compact. Choose nitely many points x
1
, . . . , x
M
E
R
such that
M
1
B
4
(x
i
) E
R
. Notice that
M
1
B
4
(x
i
) C = . Denote
2
:=
M
x
i
,
4
.
Then the function
2
is positive on V
R
and vanishes on C. Note that the function
1
+
2
is positive
on B
R
(0) and it coincides with
1
on C. Finally, dene the function
C,U
by the formula
C,U
:=
1
1
+
2
102
on B
(
R)(0) and extend it to the whole space V as equal to 0 outside the ball B
R
(0). Then
C,U
= 1
on C and Supp(
C,U
) U, as required. .
Let C V be a compact set. Consider its nite covering by open sets U
1
, . . . , U
N
, i.e.
N
_
1
U
j
C.
We say that a nite sequence
1
, . . . ,
K
of C
j=1,...,N
if
j
(x) = 1 for all x C;
Each function
j
, j = 1, . . . , K is supported in one of the sets U
i
, i = 1, . . . , K.
Lemma 9.3. For any compact set C and its open covering U
j
j=1,...,N
there exists a partition of
unity over C subordinated to this covering.
Proof. In view of compactness of there exists > 0 and nitely many balls B
(z
j
) centered at
points z
j
C, j = 1, . . . , K, such that
K
1
B
(z
j
) C and each of these balls is contained in one of
the open sets U
j
, j = 1, . . . , N. Consider the functions
z
j
,
dened in (9.1.2). We have
K
z
j
,
> 0
on some neighborhood U C. Let
C,U
be the cut-o function constructed in Lemma 9.2. For
j = 1, . . . , K we dene
j
(x) =
_
_
z
j
,(x)
C,U
(x)
K
P
1
z
j
,(x)
, if x U,
0, otherwise
.
Each of the functions is supported in one of the open sets U
j
, j = 1, . . . , N, and we have for every
x C
K
j
(x) =
K
z
j
,
(x)
K
z
j
,
(x)
= 1.
103
Figure 9.1: Bernhard Riemann (1826-1866)
9.2 One-dimensional Riemann integral for functions and dieren-
tial 1-forms
A partition T of an interval [a, b] is a nite sequence a = t
0
< t
1
< < t
N
= b. We will
denote by T
j
, j = 0, . . . , N the vector t
j+1
t
j
R
t
j
and by
j
the interval [t
j
, t
j+1
]. The length
t
j+1
t
j
= [[T
j
[[ of the interval
j
will be denoted by
j
. The number max
j=1,...,N
j
is called the
neness or the size of the partition T and will be denoted by (T). Let us rst recall the denition
of (Riemann) integral of a function of one variable. Given a function f : [a, b] R we will form a
lower and upper integral sums corresponding to the partition T:
L(f; T) =
N1
0
( inf
[t
j
,t
j+1
]
f)(t
j+1
t
j
) ,
U(f; T) =
N1
0
( sup
[t
j
,t
j+1
]
f)(t
j+1
t
j
) , (9.2.1)
The function is called Riemann integrable if
sup
P
L(f; T) = inf
P
U(f; T),
and in this case this number is called the (Riemann) integral of the function f over the interval
[a, b]. The integrability of f can be equivalently reformulated as follows. Let us choose a set C =
104
c
1
, . . . , c
N1
, c
j
j
, and consider an integral sum
I(f; T, C) =
N1
0
f(c
j
)(t
j+1
t
j
), c
j
j
. (9.2.2)
Then the function f is integrable if there exists a limit lim
(P)0
I(f, T, C). In this case this limit is
equal to the integral of f over the interval [a, b]. Let us emphasize that if we already know that the
function is integrable, then to compute the integral one can choose any sequence of integral sum,
provided that their neness goes to 0. In particular, sometimes it is convenient to choose c
j
= t
j
,
and in this case we will write I(f; T) instead of I(f; T, C).
The integral has dierent notations. It can be denoted sometimes by
_
[a,b]
f, but the most common
notation for this integral is
b
_
a
f(x)dx. This notation hints that we are integrating here the dierential
form f(x)dx rather than a function f. Indeed, given a dierential form = f(x)dx we have
f(c
j
)(t
j+1
t
j
) =
c
j
(T
j
),
1
and hence
I(; T, C) = I(f; T, C) =
N1
c
j
(T
j
), c
j
j
. (9.2.3)
We say that a dierential 1-form is integrable if there exists a limit lim
(P)0
I(, T, C), which is
called in this case the integral of the dierential 1-form over the oriented interval [a, b] and will
be denoted by
_
[a,b]
, or simply
b
_
a
. By denition, we say that
_
[a,b]
=
_
[a,b]
. This agrees with the
denition
_
[a,b]
= lim
(P)0
N1
c
j
(T
j
), and with the standard calculus rule
a
_
b
f(x)dx =
b
_
a
f(x)dx.
Let us recall that a map : [a, b] [c, d] is called a dieomorphism if it is smooth and has a
smooth inverse map
1
: [c, d] [a, b]. This is equivalent to one of the following:
(a) = c; (b) = d and
[a,b]
=
_
[c,d]
, (9.2.4)
if preserves the orientation and
_
[a,b]
=
_
[c,d]
=
_
[c,d]
,
if reverses the orientation.
Remark 9.5. We will show later a stronger result:
_
[a,b]
=
_
[c,d]
for any : [a, b] [c, d] with (a) = c, (b) = d, which is not necessarily a dieomorphism.
Proof. We consider only the orientation preserving case, and leave the orientation reversing one
to the reader. Choose any partition T = a = t
0
< < t
N1
< t
N
= b of the interval [a, b] and
choose any set C = c
0
, . . . , c
N1
such that c
j
j
. Then the points
t
j
= (t
j
) [c, d], j = 0, . . . N
form a partition of [c, d]. Denote this partition by
T, and denote
j
:= [
t
j
,
t
j+1
] [c, d], c
j
= (c
j
),
C = (C) = c
0
, . . . , c
N1
. Then we have
I(
, T, C) =
N1
c
j
(T
j
) =
N1
ec
j
(d(T
j
)) =
N1
ec
j
(
(c
j
)
j
). (9.2.5)
Recall that according to the mean value theorem there exists a point d
j
j
, such that
T
j
=
t
j+1
t
j
= (t
j+1
) (t
j
) =
(d
j
)(t
j+1
t
j
).
Note also that the function
is uniformly continuous, i.e. for any > 0 there exists > 0 such
that for any t, t
[ < we have [
(t)
(t
106
is bounded above and below by some positive constants: m <
< M. Hence m
j
<
j
< M
j
for
all j = 1, . . . , N 1. Hence, if (T) < then we have
I(
, T, C) I(;
T,
C)
N1
ec
j
_
(
(c
j
)
(d
j
))
j
_
N1
1
f(c
j
)
=
m
I(;
T,
C)
. (9.2.6)
When (T) 0 we have
(T) = 0, and hence by assumption I(;
T,
C)
d
_
c
, but this implies
that I(
, T, C) I(;
T,
C) 0, and thus
= lim
(P)0
I(
, T, C) = lim
(
e
P)0
I(,
T,
C) =
d
_
c
.
= f((t))
(t)dt.
9.3 Integration of dierential 1-forms along curves
Curves as paths
A path, or parametrically given curve in a domain U in a vector space V is a map : [a, b] U. We
will assume in what follows that all considered paths are dierentiable. Given a dierential 1-form
in U we dene the integral of over by the formula
_
=
_
[a,b]
.
Example 9.6. Consider the form = dz ydx +xdy on R
3
. Let : [0, 2] R
3
be a helix given
by parametric equations x = Rcos t, y = Rsint, z = Ct. Then
_
=
2
_
0
(Cdt +R
2
(sin
2
tdt + cos
2
tdt)) =
2
_
0
(C +R
2
)dt = 2(C +R
2
).
107
Note that
_
= 0 when C = R
2
. One can observe that in this case the curve is tangent to the
plane eld given by the Pfaan equation = 0.
Proposition 9.7. Let a path be obtained from : [a, b] U by a reparameterization, i.e.
= , where : [c, d] [a, b] is an orientation preserving dieomorphism. Then
_
e
=
_
.
Indeed, applying Theorem 9.4 we get
_
e
=
d
_
c
=
d
_
c
)
b
_
a
=
_
.
A vector
(t) V
(t)
is called the velocity vector of the path .
Curves as 1-dimensional submanifolds
A subset U is called a 1-dimensional submanifold of U if for any point x there is a
neighborhood U
x
U and a dieomorphism
x
: U
x
x
R
n
, such that
x
(x) = 0 R
n
and
x
(U
x
) either coincides with x
2
= . . . x
n
= 0
x
, or with x
2
= . . . x
n
= 0, x
1
0
x
. In
the latter case the point x is called a boundary point of . In the former case it is called an interior
point of .
A 1-dimensional submanifold is called closed if it is compact and has no boundary. An example
of a closed 1-dimensional manifold is the circle S
1
= x
2
+y
2
= 1 R
2
.
WARNING. The word closed is used here in a dierent sense than when one speaks about closed
subsets. For instance, a circle in R
2
is both, a closed subset and a closed 1-dimensional submanifold,
while a closed interval is a closed subset but not a closed submanifold: it has 2 boundary points.
An open interval in R (or any R
n
) is a submanifold without boundary but it is not closed because
it is not compact. A line in a vector space is a 1-dimensional submanifold which is a closed subset
of the ambient vector space. However, it is not compact, and hence not a closed submanifold.
Proposition 9.8. 1. Suppose that a path : [a, b] U is an embedding. This means that
) if t ,= t
.
2
Then = ([a, b]) is 1-dimensional
compact submanifold with boundary.
2
If only the former property is satised that is called an immersion.
108
2. Suppose U is given by equations F
1
= 0, . . . , F
n1
= 0 where F
1
, . . . , F
n1
: U R are
smooth functions such that for each point x the dierential d
x
F
1
, . . . , d
x
F
n1
are linearly
independent. Then is a 1-dimensional submanifold of U.
3. Any compact connected 1-dimensional submanifold U can be parameterized either by an
embedding : [a, b] U if it has non-empty boundary, or by an embedding : S
1
U if it is closed.
Proof. 1. Take a point c [a, b]. By assumption
(c) V
C
coincide
with the rst basic vector. In these coordinates the map gamma can be written as (
1
, . . . ,
n
)
where
1
(c) = 1,
j
(c) = 0 for j > 1 and
j
(c) = 0 for all j = 1, . . . , n. By the inverse function
theorem the function
1
is a dieomorphism of a neighborhood of c onto a neighborhood of 0 in
R (if c is one of the end points of the interval, then it is a dieomorphism onto the corresponding
one-sided neighborhood of 0). Let be the inverse function dened on the interval equal to
(, ), [0, ) and (, 0], respectively, depending on whether c is an interior point, c = a or c = b].
so that
1
((u)) = u for any u . Denote
= () [a, b]. Then (
) U can be given by
the equations:
y
2
=
2
(u) :=
2
((y
1
)), . . . , y
n
=
n
(u) :=
n
((y
1
)); y
1
.
Let us denote
= () := max
j=2,...,n
max
u
[
j
(u)[.
Denote
P
:= [y
1
[ , [y
j
[ ().
We have () P
.
We will show now that for a suciently small we have ([a, b]) P
)
for t ,= t
. Hence d(t) > 0 for all t [a, b] Int . The function d(t) is continuous and hence
achieve the minimum value on the compact set [a, b] Int . Denote d := min
t[a,b]\Int
d(t) > 0. Chose
) = max
j=2,...,n
max
|uc|<
[
j
(u)[ < d. Let
= [u[
[. Then
([a, b]) P
= y
2
=
2
(u), . . . , y
n
=
n
(u); y
1
.
109
2. Take a point c . The linear independent 1-forms d
c
F
1
, . . . , d
c
F
n1
V
c
can be completed
by a 1-form l V
c
to a basis of V
c
. We can choose an ane coordinate system in V with c as its
origin and such that the function x
n
coincides with l. Then the Jacobian matrix of the functions
F
1
, . . . , F
n1
, x
n
is non-degenerate at c = 0, and hence by the inverse function theorem the map
F = (F
1
, . . . , F
n1
, x
n
) : V R
n
is invertible in the neighborhood of c = 0, and hence these
functions can be chosen as new curvilinear coordinates y
1
= F
1
, . . . , y
n1
= F
n1
, y
n
= x
n
near the
point c = 0. In these coordinates the curve is given near c by the equations y
1
= = y
n1
= 0.
3. See Exercise ??.
In the case where is closed we will usually parameterize it by a path : [a, b] U
with (a) = (b). For instance, we parameterize the circle S
1
= x
2
+ y
2
= 1 R
2
by a path
[0, 2] (cos t, sint). Such , of course, cannot be an embedding, but we will require that
(t) ,= 0
and that for t ,= t
we have (t) ,= (t
(t) for any local parameterization : [a, b] with (t) = x. If is given implicitly, as in
9.8.2, then the tangent line is dened in V
x
by the system of linear equations d
x
F
1
= 0, . . . , d
x
F
n1
=
0.
Orientation of a curve is the continuously depending on points orientation of all its tangent
lines. If the curve is given as a path : [a, b] U such that
(t) l
x
.
It turns out that one can dene an integral of a dierential form over an oriented compact
curve directly without referring to its parameterization. For simplicity we will restrict our discussion
to the case when the form is continuous.
Let be a compact connected oriented curve. A partition of is a sequence of points T =
z
0
, z
1
, . . . , z
N
ordered according to the orientation of the curve and such that the boundary
points of the curve (if they exist) are included into this sequence. If is closed we assume that
z
N
= z
0
. The neness (T) of T is by denition is max
j=0,...,N1
dist(z
j
, z
j+1
) (we assume here that V
a Euclidean space).
110
Denition 9.9. Let be a dierential 1-form and a compact connected oriented curve. Let
T = z
0
, . . . , z
N
be its partition. Then we dene
_
= lim
(P)0
I(, T),
where I(, T) =
N1
z
j
(Z
j
), Z
j
= z
j+1
z
j
V
z
j
.
When is a closed submanifold then one sometimes uses the notation
_
instead of
_
.
Proposition 9.10. If one chooses a parameterization : [a, b] which respects the given
orientation of then
_
=
b
_
a
=
_
.
Proof. Indeed, let
T = t
0
, . . . , t
N
be a partition of [a, b] such that (t
j
) = z
j
, j = 0, . . . , N.
I(
,
T) =
N1
t
j
(T
j
) =
N1
z
j
(U
j
),
where U
j
= d
t
j
(T
j
)) V
z
j
is a tangent vector to at the point z
j
. Let us evaluate the dierence
U
j
Z
j
. Choosing some Cartesian coordinates in V we denote by
1
, . . . ,
n
the coordinate functions
of the path . Then using the mean value theorem for each of the coordinate functions we get
i
(t
j+1
)
i
(t
j
) =
i
(c
i
j
)
j
for some c
i
j
j
, i = 1, . . . , n; j = 0, . . . , N 1. Thus
Z
j
= (t
j+1
) (t
j
) = (
1
(c
1
j
), . . . ,
i
(c
n
j
))
j
.
On the other hand, U
j
= d
t
j
(T
j
)) V
z
j
=
(t
j
)
j
. Hence,
[[Z
j
U
j
[[ =
j
_
n
1
(
i
(c
i
j
)
i
(t
j
))
2
.
Note that if (T) 0 then we also have (
z
j
(
T
j
)
N1
z
j
(Z
j
)
(P)
0,
111
and therefore
b
_
a
= lim
(
e
P)0
I(
,
T) = lim
(P)0
I(, T) =
_
= f(B) f(A).
In particular, if is a loop then
_
= 0.
Similarly for an oriented curve U with boundary = B A we have
_
= f(B) f(A).
Proof. We have
_
df =
b
_
a
df =
b
_
a
d(f ) = f((b)) f((a)) = f(B) f(A).
It turns out that closed forms are locally exact. A domain U V is called star-shaped with
respect to a point a V if with any point x U it contains the whole interval I
a,x
connecting a
and x, i.e. I
a,x
= a +t(x a); t [0, 1]. In particular, any convex domain is star-shaped.
Proposition 9.12. Let be a closed 1-form in a star-shaped domain U V . Then it is exact.
Proof. Dene a function F : U R by the formula
F(x) =
_
Ia,x
, x U,
where the intervals I
a,x
are oriented from 0 to x.
We claim that dF = . Let us identify V with the R
n
choosing a as the origin a = 0. Then
can be written as =
n
1
P
k
(x)dx
k
, and I
0,x
can be parameterized by
t tx, t [0, 1].
112
Hence,
F(x) =
_
I
0,x
=
1
_
0
n
1
P
k
(tx)x
k
dt. (9.4.1)
Dierentiating the integral over x
j
as parameters, we get
F
x
j
=
1
_
0
n
k=1
tx
k
P
k
x
j
(tx)dt +
1
_
0
P
j
(tx)dt.
But d = 0 implies that
P
k
x
j
=
P
j
x
k
, and using this we can further write
F
x
j
=
1
_
0
n
k=1
tx
k
P
j
x
k
(tx)dt +
1
_
0
P
j
(tx)dt =
1
_
0
t
dP
j
(tx)
dt
dt +
1
_
0
P
j
(tx)dt
= (tP
j
(tx))[
1
0
1
_
0
P
j
(tx)dt +
1
_
0
P
j
(tx)dt = P
j
(tx)
Thus
dF =
n
j=1
F
x
j
dx
j
=
n
j=1
P
j
(x)dx =
9.5 Integration of functions over domains in high-dimensional spaces
Riemann integral over a domain in R
n
.
In this section we will discuss integration of bounded functions over bounded sets in a vector space
V . We will x a basis e
1
, . . . , e
n
and the corresponding coordinate system x
1
, . . . , x
n
in the space
and thus will identify V with R
n
. Let denote the volume form x
1
. . . x
n
. As it will be clear
below, the denition of an integral will not depend on the choice of a coordinate system but only
on the background volume form, or rather its absolute value because the orientation of V will be
irrelevant.
We will need a special class of parallelepipeds in V , namely those which are generated by vectors
proportional to basic vectors, or in other words, parallelepipeds with edges parallel to the coordinate
axes. We will also allow these parallelepipeds to be parallel transported anywhere in the space. Let
us denote
P(a
1
, b
1
; a
2
, b
2
; . . . ; a
n
, b
n
) := a
i
x
i
b
i
; i = 1, . . . , n R
n
.
113
We will refer to P(a
1
, b
1
; a
2
, b
2
; . . . ; a
n
, b
n
) as a special parallelepiped, or rectangle.
Let us x one rectangle P := P(a
1
, b
1
; a
2
, b
2
; . . . ; a
n
, b
n
). Following the same scheme as we used
in the 1-dimensional case, we dene a partition T of P as a product of partitions a
1
= t
1
0
< <
t
1
N
1
= b
1
, . . . , a
n
= t
n
0
< < t
n
Nn
= b
n
, of intervals [a
1
, b
1
], . . . , [a
n
, b
n
]. For simplicity of notation
we will always assume that each of the coordinate intervals is partitioned into the same number
of intervals, i.e. N
1
= = N
n
= N. This denes a partition of P into N
n
smaller rectangles
P
j
= t
1
j
1
x
1
t
1
j
1
+1
, . . . , t
n
jn
x
n
t
n
jn+1
, where j = (j
1
, . . . , j
n
) and each index j
k
takes
values between 0 and N 1. Let us dene
Vol(P
j
) :=
n
k=1
(t
k
j
k
+1
t
k
j
k
). (9.5.1)
This agrees with the denition of the volume of a parallelepiped which we introduced earlier (see
formula (3.3.1) in Section 3.3). We will also denote
j
:= max
k=1,...,n
(t
k
j
k
+1
t
k
j
k
) and (T) := max
j
(
j
).
Let us x a point c
j
P
j
and denote by C the set of all such c
j
. Given a function f : P R we
form an integral sum
I(f; T, C) =
j
f(c
j
)Vol(P
j
) (9.5.2)
where the sum is taken over all elements of the partition. If there exists a limit lim
(P)0
I(f; T, C)
then the function f : P R is called integrable (in the sense of Riemann) over P, and this limit is
called the integral of f over P. There exist several dierent notations for this integral:
_
P
f,
_
P
fdV ,
_
P
fdVol, etc. In the particular case of n = 2 one often uses notation
_
P
fdA, or
__
P
fdA. Sometime, the
functions we integrate may depend on a parameter, and in these cases it is important to indicate with
respect to which variable we integrate. Hence, one also uses the notation like
_
P
f(x, y)dx
n
, where
the index n refers to the dimension of the space over which we integrate. One also use the notation
_
. . .
_
. .
P
f(x
1
, . . . x
n
)dx
1
. . . dx
n
, which is reminiscent both of the integral
_
P
f(x
1
, . . . x
n
)dx
1
dx
n
which will be dened later in Section 9.7 and the notation
bn
_
an
. . .
b
1
_
a
1
f(x
1
, . . . x
n
)dx
1
. . . dx
n
for
n interated integral which will be discussed in Section 9.6.
Alternatively and equivalently the integrability can be dened via upper and lower integral sum,
114
similar to the 1-dimensional case. Namely, we dene
U(f; T) =
j
M
j
(f)Vol(P
j
), L(f; T) =
j
m
j
(f)Vol(P
j
),
where M
j
(f) = sup
P
j
f, m
j
(f) = inf
P
j
f, and say that the function f is integrable over P if inf
P
U(f; T) =
sup
P
L(f, T).
Note that inf
P
U(f; T) and sup
P
L(f, T) are sometimes called upper and lower integrals, respec-
tively, and denoted by
_
P
f and
_
P
f. Thus a function f : P R is integrable i
_
P
f =
_
P
f.
Let us list some properties of Riemann integrable functions and integrals.
Proposition 9.13. Let f, g : P R be integrable functions. Then
1. af +bg, where a, b R, is integrable and
_
P
af +bg = a
_
P
f +b
_
P
g;
2. If f g then
_
P
f
_
P
g;
3. h = max(f, g) is integrable; in particular the functions f
+
:= max(f, 0) and f
:= max(f, 0)
and [f[ = f
+
+f
are integrable;
4. fg is integrable.
Proof. Parts 1 and 2 are straightforward and we leave them to the reader as an exercise. Let us
check properties 3 and 4.
3. Take any partition T of P. Note that
M
j
(h) m
j
(h) max(M
j
(f) m
j
(f), M
j
(g) m
j
(g)) . (9.5.3)
Indeed, we have M
j
(h) = max(M
j
(f), M
j
(g)) and m
j
(h) max(m
j
(f), m
j
(g)). Suppose for deter-
minacy that max(M
j
(f), M
j
(g)) = M
j
(f). We also have m
j
(h) m
j
(f). Thus
M
j
(h) m
j
(h) M
j
(f) m
j
(f) max(M
j
(f) m
j
(f), M
j
(g) m
j
(g)).
Then using (9.5.3) we have
U(h; T) L(h; T) =
j
(M
j
(h) m
j
(h))Vol(P
j
)
j
max (M
j
(f) m
j
(f), M
j
(g) m
j
(g)) Vol(P
j
) =
max (U(f; T) L(f; T), U(f; T) L(f; T)) .
115
By assumption the right-hand side can be made arbitrarily small for an appropriate choice of the
partition T, and hence h is integrable.
4. We have f = f
+
f
, g = g
+
g
and fg = f
+
g
+
+ f
f
+
g
g
+
. Hence, using 1
and 3 we can assume that the functions f, g are non-negative. Let us recall that the functions f, g
are by assumption bounded, i.e. there exists a constant C > 0 such that f, g C. We also have
M
j
(fg) M
j
(f)M
j
(g) and m
j
(fg) m
j
(f)m
j
(g). Hence
U(fg; T) L(fg; T) =
j
(M
j
(fg) m
j
(fg)) Vol(P
j
)
j
(M
j
(f)M
j
(g) m
j
(f)m
j
(g))Vol(P
j
) =
j
(M
j
(f)M
j
(g) m
j
(f)M
j
(g) +m
j
(f)M
j
(g) m
j
(f)m
j
(g)) Vol(P
j
)
j
((M
j
(f) m
j
(f))M
j
(g) +m
j
(f)(M
j
(g) m
j
(g))) Vol(P
j
)
C(U(f; T) L(f; T) +U(g; T) L(g; T)).
By assumption the right-hand side can be made arbitrarily small for an appropriate choice of the
partition T, and hence fg is integrable.
Consider now a bounded subset K R
n
and choose a rectangle P K. Given any function
f : K R one can always extend it to P as equal to 0. A function f : K R is called integrable
over K if this trivial extension f is integrable over P, and we dene
_
K
fdV :=
_
fdV . When this
will not be confusing we will usually keep the notation f for the above extension.
Volume
We further dene the volume
Vol(K) =
_
K
1dV =
_
P
K
dV,
provided that this integral exists. In this case we call the set K measurable in the sense of Riemann,
or just measurable.
3
Here
K
is the characteristic or indicator function of K, i.e. the function which
3
There exists a more general and more common notion of measurability in the sense of Lebesgue. Any Riemann
measurable set is also measurable in the sense of Lebesgue, but not the other way around. Historically an attribution
116
is equal to 1 on K and 0 elsewhere. In the 2-dimensional case the volume is called the area, and in
the 1-dimensional case the length.
Remark 9.14. For any bounded set A there is dened a lower and upper volumes,
Vol(A) =
_
A
dV Vol(A) =
_
A
dV.
The set is measurable i Vol(A) = Vol(A). If Vol(A) = 0 then Vol(A) = 0, and hence A is
measurable and Vol(A) = 0.
Exercise 9.15. Prove that for the rectangles this denition of the volume coincides with the one
given by the formula (9.5.1).
The next proposition lists some properties of the volume.
Proposition 9.16. 1. Volume is monotone, i.e. if A, B P are measurable and A B then
Vol(A) Vol(B).
2. If sets A, B P are measurable then AB, A B and AB are measurable as well and we
have
Vol(A B) = Vol(A) + Vol(B) Vol(A B).
3. If A can be covered by a measurable set of arbitrarily small total volume then Vol(A) = 0.
Conversely, if Vol(A) = 0 then for any > 0 there exists a > 0 such that for any partition
T with (T) < the elements of the partition which intersect A have arbitrarily small total
volume.
4. A is measurable i Vol (A) = 0.
Proof. The rst statement is obvious. To prove the second one, we observe that
AB
= max(
A
,
B
),
AB
=
A
B
, max(
A
,
B
) =
A
+
B
A
B
,
A\B
=
A
AB
and then apply Proposition
9.13. To prove 9.16.3 we rst observe that if a set B is measurable and VolB < then then for a
suciently ne partition T we have U(
B
; T) < VolB + < 2. Since A B then
A
B
, and
of this notion to Riemann is incorrect. It was dened by Camille Jordan and Giuseppe Peano before Riemann integral
was introduced. What we call in these notes volume is also known by the name Jordan content.
117
therefore U(
A
, T) U(
B
, T) < 2. Thus, inf
P
U(
A
, T) = 0 and therefore A is measurable and
Vol(A) = 0. Conversely, if Vol(A) = 0 then for any > 0 for a suciently ne partition T we have
U(
A
; T) < . But U(
A
; T) is equal to the sum of volumes of elements of the partition which have
non-empty intersection with A.
Finally, let us prove 9.16.4. Consider any partition T of P and form lower and upper integral
sums for
A
. Denote M
j
:= M
j
(
A
) and m
j
= m
j
(
A
). Then all numbers M
j
, m
j
are equal to either
0 or 1. We have M
j
= m
j
= 1 if P
j
A; M
j
= m
j
= 0 if P
j
A = and M
j
= 1, m
j
= 0 if P
j
has
non-empty intersection with both A and P A. In particular,
B(T) :=
_
j;M
j
m
j
=1
P
j
A.
Hence, we have
U(
A
; T) L(
A
; T) =
j
(M
j
m
j
)Vol(P
j
) = VolB(T).
Suppose that A is measurable. Then there exists a partition such that U(
A
; T) L(
A
; T) < ,
and hence A is can be covered by the set B(T) of volume < . Thus applying part 3 we conclude
that Vol(A) = 0. Conversely, we had seen below that if Vol(A) = 0 then there exists a partition
such that the total volume of the elements intersecting A is < . Hence, for this partition we have
L(
A
; T) U(
A
; T) < , which implies the integrability of
A
, and hence measurability of A.
Corollary 9.17. If a bounded set A V is measurable then its interior Int A and its closure A
are also measurable and we have in this case
VolA = Vol Int A = VolA.
Proof. 1. We have A A and (Int A) A. Therefore, Vol A = Vol Int A = 0, and
therefore the sets A and Int A are measurable. Also Int AA = A and Int AA = . Hence, the
additivity of the volume implies that VolA = Vol Int A+Vol A = Vol Int A. On the other hand,
Int A A A. and hence the monotonicity of the volume implies that Vol Int A Vol A Vol A.
Hence, Vol A = Vol Int A = Vol A.
Exercise 9.18. If Int A or A are measurable then this does not imply that A is measurable. For
instance, if A is the set of rational points in interval I = [0, 1] R then Int A = and A = I.
However, show that A is not Riemann measurable.
118
2. A set A is called nowhere dense if Int A = . Prove that if A is nowhere dense then either
VolA = 0, or A is not measurable in the sense of Riemann. Find an example of a non-measurable
nowhere dense set.
Volume and smooth maps
We will further need the following lemma. Let us recall that given a compact set C V , we say
that a map f : C W is smooth if it extends to a smooth map dened on an open neighborhood
U C. Here V, W are vector spaces.
Lemma 9.19. Let A V be a compact set of volume 0 and f : V W a C
1
-smooth map, where
dimW dimV . Then Volf(A) = 0.
Proof. The C
1
-smoothness of f and compactness of A imply that there exists a constant K such
that [[d
x
f(h)[[ K[[h[[ for any x A and h V
x
. In particular, the image d
x
f(P) of every
rectangle P of size in V
x
, x A, is contained in a cube of size K in W
f(x)
. C
1
-smoothness
of f also implies that for any > 0 there exists such > 0 that if x A and [[h[[ then
[[f(x+h) f(x) d
x
f(h)[[ [[h[[. This implies that if we view P as a subset of V , rather than V
x
,
then the image f(P) is contained in a cube in W of size 2K if is small enough, and if P A ,= .
Let us denote dimensions of V and W by n and m, respectively. By assumption A can be covered
by N cubes of size such that the total volume of these cubes is equal to N
n
. Hence, f(A)
can be covered by N cubes of size 2K of total volume
N(2K)
m
= N
n
(2K)
m
mn
= (2K)
m
mn
0
0,
because m n.
Corollary 9.20. Let A V be a compact domain and f : A W a C
1
-smooth map. Suppose that
n = dimV < m = dimW. Then Vol(f(A)) = 0.
Indeed, f can be extended to a smooth map dened on a neighborhood of A0 in V R (e.g.
as independent of the new coordinate t R). But Vol
n+1
(A 0) = 0 and m n + 1. Hence, the
required statement follows from Lemma 9.19.
119
Remark 9.21. The statement of Corollary 9.20 is wrong for continuous maps. For instance, there
exists a continuous map h : [0, 1] R
2
such that h([0, 1]) is the square 0 x
1
, x
1
1. (This is
a famous Peano curve passing through every point of the square.)
Corollary 9.20 is a simplest special case of Sards theorem which asserts that the set of critical
values of a suciently smooth map has volume 0. More precisely,
Proposition 9.22. (A. Sard, 1942) Given a C
k
-smooth map f : A W (where A is a compact
subset of V , dimV = n, dimW = m) let us denote by
(f) : x C; rank d
x
f < m.
Then if k max(n m+ 1, 1) then Vol
m
(f((f)) = 0.
If m > n then (f) = A, and hence the statement is equivalent to Corollary 9.20.
Proof. We prove the proposition only for the case m = n. The proof in this case is similar to the
proof of Corollary 9.19. C
1
-smoothness of f and compactness of (f) imply that there exists a
constant K such that
[[d
x
f(h)[[ K[[h[[ (9.5.4)
for any x A and h V
x
. C
1
-smoothness of f also implies that for any > 0 there exists such
> 0 that if x A and [[h[[ then
[[f(x +h) f(x) d
x
f(h)[[ [[h[[. (9.5.5)
Take a partition T of a rectangle P A by N
n
smaller rectangles of equal size. Let B be the
union of the rectangles intersecting (f). For any such rectangle P
j
B choose a point c
j
(f).
Viewing P
j
as a subset of V
c
j
we can take its image
P
j
= d
c
j
(P
bj
) W
f(c
j
)
. Then the parallelepiped
P
j
is contained in a subspace L W
f(c
j
)
of dimension r = rank(d
c
j
f) < m. In view of (9.5) it also
contained in a ball of radius
K
r
N
centered at the point f(c
j
). On the other hand, the inequality
(9.5.5) implies that if N is large enough then for any point u f(P
j
) there is a point in u
P
j
such [[u u[[
N
. This means that f(P
j
) is contained in a parallelepiped centered at c
j
and
generated by r orthogonal vectors of length
C
1
N
parallel to the subspace L and n r > 0 vectors
120
of length
C
2
N
which are orthogonal to L, where C
1
, C
2
are some positive constants. The volume of
this parallelepiped is equal to
C
r
1
C
nr
2
nr
N
n
=
C
3
nr
N
n
,
and hence Vol(f(P
j
) <
C
3
nr
N
n
. The set B contains no more that N
n
cubes P
j
, and hence
Vol(f((f)) Volf(C) N
n
C
3
nr
N
n
= C
3
nr
0
0.
of all rectangles of
the partition which have non-empty intersection with A has volume < . The functions f, g are
bounded, i.e. there exists C > 0 such C [f(x)[, [g(x)[ C for all x P. Due to integrability of
f we can choose small enough so that [U(f, T) L(f, T)[ when (T) . Then we have
[U(g, T) U(f, T)[ =
J: P
J
B
sup
P
J
g sup
P
J
f)
2CVolB
2C.
Similarly, [L(g, T) L(f, T)[ 2C, and hence
[U(g, T) L(g, T)[ [U(g, T) U(f, T)[ +[U(f, T) L(f, T)[ +[L(f, T) L(g, T)[
+ 4C
0
0,
121
and hence g is integrable and
_
P
g = lim
(P)0
U(g, T) = lim
(P)0
U(f, T) =
_
P
f.
B and [[x x
we have (T
) < . By assumption
the function f is bounded, i.e. there exists a constant K > 0 such that M
j
(f) m
j
(f) < K for all
indices j. Then we have
U(f; T
) L(f, T
) =
j
(M
j
(f) m
j
(f))Vol(P
j
) =
j;P
j
B
(M
j
(f) m
j
(f))Vol(P
j
) +
j;P
j
C
(M
j
(f) m
j
(f))Vol(P
j
) <
VolB +KVolC < (VolP +K).
Hence inf
P
U(f; T) = sup
P
L(f; T), i.e. the function f is integrable.
122
2. If x is an interior point of A and detDf(x) ,= 0 then the inverse function theorem implies
that f(x) Int f(A). Denote C = x A; det Df(x) = 0. Hence, f(A) f(A) f(C).
But Vol(A) = 0 because A is measurable and Vol f(C) = 0 by Sards theorem 9.22. Therefore,
Vol f(A) = 0 and thus f(A) is measurable.
Orthogonal invariance of the volume and volume of a parallelepiped
The following lemma provides a way of computing the volume via packing by balls rather then
cubes. An admissible set of balls in A is any nite set of disjoint balls B
1
, . . . , B
K
A
Lemma 9.25. Let A be a measurable set. Then VolA is the supremum of the total volume of
admissible sets of balls in A. Here the supremum is taken over all admissible sets of balls in A.
Proof. Let us denote this supremum by . The monotonicity of volume implies that VolA.
Suppose that < V olA. Let us denote by
n
the volume of an n-dimensional ball of radius 1 (we
will compute this number later on). This ball is contained in a cube of volume 2
n
. It follows then
that the ratio of the volume of any ball to the volume of the cube to which it is inscribed is equal
to
n
2
n
. Choose an <
n
2
n
(VolA). Then there exists a nite set of disjoint balls B
1
, . . . , B
K
A
such that Vol(
K
1
B
j
) > . The volume of the complement C = A
K
1
B
j
satises
VolC = VolAVol
_
K
_
1
B
j
_
> VolA.
Hence there exists a partition T of P by cubes such that the total volume of cubes Q
1
, . . . , Q
L
con-
tained in C is > VolA. Let us inscribe in each of the cubes Q
j
a ball
B
j
. Then B
1
, . . . , B
K
,
B
1
, . . . ,
B
L
is an admissible set of balls in A. Indeed, all these balls are disjoint and contained in A. The total
volume of this admissible set is equal to
K
1
VolB
j
+
L
1
Vol
B
i
+
n
2
n
(VolA) > ,
in view of our choice of , but this contradicts to our assumption < VolA. Hence, we have
= VolA.
123
Lemma 9.26. Let A V be any measurable set in a Euclidean space V . Then for any linear
orthogonal transformation F : V V the set F(A) is also measurable and we have Vol(F(A)) =
Vol(A).
Proof. First note that if VolA = 0 then the claim follows from Lemma 9.19. Indeed, an orthogonal
transformation is, of course a smooth map.
Let now A be an arbitrary measurable set. Note that F(A) = F(A). Measurability of A
implies Vol(A) = 0. Hence, as we just have explained, Vol(F(A)) = Vol(F(A)) = 0, and hence
F(A) is measurable. According to Lemma 9.25 the volume of a measurable set can be computed
as a supremum of the total volume of disjoint inscribed balls. But the orthogonal transformation
F moves disjoint balls to disjoint balls of the same size, and hence VolA = VolF(A).
Next proposition shows that the volume of a parallelepiped can be computed by formula (3.3.1)
from Section 3.3.
Proposition 9.27. Let v
1
, . . . , v
n
V be linearly independent vectors. Then
Vol P(v
1
, . . . , v
n
) = [x
1
x
n
(v
1
. . . , v
n
)[. (9.5.6)
Proof. The formula (9.5.6) holds for rectangles, i.e. when v
j
= c
j
e
j
for some non-zero numbers c
j
,
j = 1, . . . n. Using Lemma 9.26 we conclude that it also holds for any orthogonal basis. Indeed, any
such basis can be moved by an orthogonal transformation to a basis of the above form c
j
e
j
, j =
1, . . . n. Lemma 9.26 ensures that the volume does not change under the orthogonal transformation,
while Proposition 2.17 implies the same about [x
1
x
n
(v
1
. . . , v
n
)[.
The Gram-Schmidt orthogonalization process shows that one can pass from any basis to an
orthogonal basis by a sequence of following elementary operations:
- reordering of basic vectors, and
shears, i.e. an addition to the last vector a linear combination of the other ones:
v
1
, . . . , v
n1
, v
n
v
1
, . . . , v
n1
, v
n
+
n1
j
v
j
.
Note that the reordering of vectors v
1
, . . . , v
n
changes neither VolP(v
1
, . . . , v
n
), nor the absolute
value
124
[x
1
x
n
(v
1
. . . , v
n
)[. On the other hand, a shear does not change
x
1
x
n
(v
1
. . . , v
n
).
It remains to be shown that a shear does not change the volume of a parallelepiped. We will consider
here only the case n = 2 and will leave to the reader the extension of the argument to the general
case.
Let v
1
, v
2
be two orthogonal vectors in R
2
. We can assume that v
1
= (a, 0), v
2
= (0, b) for
a, b > 0, because we already proved the invariance of volume under orthogonal transformations.
Let v
2
= v
2
+v
1
= (a
, b), where a
2
) that is not covered
by rectangles
P
i,j
, i, j = 0, . . . , N 1 converges to 0.
i
m
j,i
(f)Vol
n
P
j,i
.
Here the rst sum is taken over all multi-indices j of the partition T
1
, and the second sum is taken
126
over all multi-indices i of the partition T
2
. On the other hand,
L(I, T
1
) =
j
inf
xP
j
1
_
_
_
_
P
2
f
x
dV
nk
_
_
_Vol
k
P
j
1
.
Note that for every x P
j
1
we have
_
P
2
f
x
dV
nk
L(f
x
; T
2
) =
i
m
i
(f
x
)Vol
nk
(P
i
2
)
i
m
i,j
(f)Vol
nk
(P
i
2
),
and hence
inf
xP
j
1
_
P
2
f
x
dV
nk
i
m
i,j
(f)Vol
nk
(P
i
2
).
Therefore,
L(I, T
1
)
i
m
i,j
(f)Vol
nk
(P
i
2
)Vol
k
(P
j
1
) =
i
m
j,i
(f)Vol
n
(P
j,i
) = L(f, T).
Similarly, one can check that U(I, T
1
) U(f, T). Together with an obvious inequality L(I, T
1
)
U(I, T
1
) this completes the proof of (9.6.1). Thus we have
max(U(I, T
1
) L(I, T
1
), U(I, T
1
) L(I, T
1
)) U(I, T
1
) L(I, T
1
) U(f, T) L(f, T).
By assumption for appropriate choices of partitions, the right-hand side can be made < for any
a priori given > 0. This implies the integrability of the function I(x) and I(x) over P
1
. But then
we can write
_
P
1
I(x)dV
nk
= lim
(P
1
)0
L(I; T
1
)
and
_
P
1
I(x)dV
nk
= lim
(P
1
)0
U(I; T
1
).
We also have
lim
(P)0
L(f; T) = lim
(P)0
U(f; T) =
_
P
fdV
n
.
Hence, the inequality (9.6.1) implies that
_
P
fdV
n
=
_
P
1
_
_
_
_
P
2
f
x
dV
nk
_
_
_dV
k
=
_
P
1
_
_
_
P
2
f
x
dV
nk
_
_
dV
k
.
127
Corollary 9.29. Suppose f : P R is a continuous function. Then
_
P
f =
_
P
1
_
P
2
f
x
=
_
P
2
_
P
1
f
y
.
Thus if we switch back to the notation x
1
, . . . , x
n
for coordinates in R
n
, and if P = a
1
x
1
b
1
, . . . , a
n
x
n
b
n
then we can write
_
P
f =
bn
_
an
_
_
. . .
_
_
b
1
_
a
1
f(x
1
, . . . , x
n
)dx
1
_
_
. . .
_
_
dx
n
. (9.6.2)
The integral in the right-hand side of (9.6.2) is called an iterated integral. Note that the order
of integration is irrelevant there. In particular, for continuous functions one can change the order
of integration in the iterated integrals.
j
(T
j
1
, T
j
2
, . . . , T
j,n
), (9.7.1)
where the sum is taken over all elements of the partition. We call an n-form integrable if there
exists a limit lim
(P)0
I(; T, C) which we denote by
_
P
and call the integral of over P. Note
that if = f(x)dx
1
dx
n
then the integral sum I(, T, C) from (9.7.1) coincides with the
integral sum I(f; T, C) from (9.5.2) for the function f. Thus the integrability of is the same as
integrability of f and we have
_
P
f(x)dx
1
dx
n
=
_
P
fdV. (9.7.2)
Note, however, that the equality (9.7.2) holds only if the coordinate system (x
1
, . . . , x
n
) denes the
given orientation of the space V . The integral
_
P
f(x)dx
1
dx
n
changes its sign with a change
of the orientation while the integral
_
P
fdV is not sensitive to the orientation of the space V .
It is not clear from the above denition whether the integral of a dierential form depends on
our choice of the coordinate system. It turns out that it does not, as the following theorem, which
is the main result of this section, shows. Moreover, we will see that one even can use arbitrarty
curvilinear coordinates.
In what follows we use the convention introduced at the end of Section 6.1. Namely by a
dieomorphism between two closed subsets of vector spaces we mean a dieomorphism between
their neighborhoods.
Theorem 9.30. Let A, B R
n
be two measurable compact subsets. Let f : A B be an orientation
preserving dieomorphism. Let be a dierential n-form dened on B. Then if is integrable over
B then f
=
_
B
. (9.7.3)
For an orientation reversing dieomorphism f we have
_
A
f
=
_
B
.
Let = g(x)dx
1
dx
n
. Then f
= g f det Dfdx
1
dx
n
, and hence the formula
(9.7.3) can be rewritten as
_
P
g(x
1
, . . . , x
n
)dx
1
dx
n
=
_
P
g f det Df dx
1
dx
n
.
129
Here
det Df =
f
1
x
1
. . .
f
1
xn
. . . . . . . . .
fn
x
1
. . .
fn
xn
n
=
n||F
1
||
. There exists > 0 such that if [[h[[ then
[[f(h) F(h)[[ [[h[[ .
Denote :=
n
. Then I B
ne
(F(z)) = B
(F(z)),
and therefore in view of (9.7.7)
f((I)) (1 +)F(I) (1 )F(I).
But this is equivalent to inclusion (9.7.6).
132
Lemma 9.35. Let F : R
n
R
n
be a non-degenerate orientation preserving linear map, P =
P(v
1
, . . . , v
n
) a parallelepiped, and = dx
1
dx
n
. Then
_
F(P)
=
_
P
F
=
det F, and hence
_
P
F
= det F
_
P
= (det F)VolP.
Proof of Proposition 9.32. We have f
= (det Df)dx
1
dx
n
, and hence the form f
is
integrable because f is C
1
-smooth, and hence det Df is continuous.
Choose a partition T of the cube I by N
n
small cubes I
K
, K = 1, . . . , N
n
, of the same size
1
N
.
Let c
K
I
K
be the center of the cube I
K
. Then
_
I
f
=
N
n
K=1
_
I
K
f
.
Note that in view of the uniform continuity of the function det Df, for any > 0 the number N
can be chosen so large that [ det Df(x) det Df(x
I
K
and any
K = 1, . . . , N
n
. Let
K
be the form det Df(c
k
)dx
1
dx
n
on I
K
. Then
_
I
K
f
_
I
K
_
I
K
[det Df(x) det Df(c
K
)[ dx
1
dx
n
Vol(I
K
) =
N
n
.
Thus
_
I
f
N
n
K=1
_
I
K
. (9.7.8)
Next, let us analyze the integral
_
I
K
K
. Denote F
K
:= d
c
K
(f). We can assume without loss of
generality that c
K
= 0, and f(c
K
) = 0, and hence F
K
can be viewed just as a linear map R
n
R
n
.
Using Lemma 9.34 we have for a suciently large N
(1 )F
K
(I
K
) f(I
K
) (1 +)F
K
(I
K
).
Again in view of compactness of I the number can be chosen the same for all cubes I
K
. Hence
(1 )
n
Vol(F
K
(I
K
)) Volf(I
K
) (1 +)
n
Vol(F
K
(I
K
)). (9.7.9)
133
Note that Volf(I
K
) =
_
f(I
K
)
, and hence summing up inequality (9.7.9) over K we get
(1 )
n
N
n
K=1
Vol(F
K
(I
K
))
N
n
K=1
Volf(I
K
) =
N
n
K=1
_
f(I
K
)
=
_
f(I)
(1 +)
n
N
n
K=1
Vol(F
K
(I
K
)).
(9.7.10)
Note that
K
= F
K
and by Lemma 9.35 we have
_
I
K
K
=
_
I
K
F
K
=
_
F
K
(I
K
)
= Vol(F
K
(I
K
)).
Hence, it follows from (9.7.10) that
(1 )
n
N
n
K=1
_
I
K
K
_
f(I)
(1 +)
n
N
n
K=1
_
I
K
K
. (9.7.11)
Recall that from (9.7.8) we have
_
I
f
N
n
K=1
_
I
K
_
I
f
+.
Combining with (9.7.11) we get
(1 )
n
_
_
_
I
f
_
_
_
f(I)
(1 +)
n
_
_
_
I
f
+
_
_
. (9.7.12)
Passing to the limit when 0 we get
_
I
f
_
f(I)
_
I
f
, (9.7.13)
i.e.
_
I
f
=
_
f(I)
.
Proof of Theorem 9.30. Let us recall that the dieomorphism f is dened as a dieomorphism
between open neighborhoods U A and U
by cubes I
j
of the same size . Consider forms
+
j
:= M
j
(h)dx
1
dx
n
and
j
:= m
j
(h)dx
1
dx
n
on I
j
, where m
j
(h) = inf
I
j
h, M
j
(h) = sup
I
j
(h). Let
be the
form on U
equal to
j
on each cube I
j
. The assumption of integrability of over B guarantees
that for any > 0 if is chosen small enough we have
_
B
+
_
B
. The forms f
are a.e.
continuous, and hence integrable over A and we have
_
A
f
_
A
f
_
A
f
+
. Hence, if we prove
that
_
A
f
=
_
B
j
_
I
j
j
and
_
A
f
j
_
B
j
f
j
, where B
j
= f
1
(I
j
). But according
to Proposition 9.32 we have
_
B
j
f
j
=
_
I
j
j
, and hence
_
A
f
=
_
B
.
9.8 Manifolds and submanifolds
9.8.1 Manifolds
Manifolds of dimension n are spaces which are locally look like open subsets of R
n
but globally
could be much more complicated. We give a precise denition below.
Let U, U
R
n
be open sets. A map f : U U
U.
A map f : U U
is called a C
k
-dieomorpism, k = 1, . . . , , if it is C
k
-smooth, one-to-one
map which has a C
k
-smooth inverse f
1
: U
: U
R
n
such that
M1. M =
.
M2. The image G
(U
) is an open set in R
n
.
M3. The map
viewed as a map U
is one-to-one.
M4. For any two sets U
, U
, , the images
(U
),
(U
) R
n
are open and
the map
h
:=
(U
(U
) R
n
is a C
k
-dieomorphism (resp. homeomorphism).
Sets U
: U
R
n
are called coordinate maps.
The pairs (U
=
1
: G
and A
= U
=
(U
), (U
)
,
is again an atlas on X. In other words, two atlases dene the same
smooth structure if transition maps from local coordinates in one of the atlases to the local coor-
dinates in the other one are given by smooth functions.
A subset G M is called open if for every the image
(G U
) R
n
is open. In
particular, coordinate charts U
) in M and
(
) in
M
, such that a U
, f(U
)
U
(U
R
n
is continuous.
Similarly, for k = 1, . . . , a map f : M
M is called C
k
-smooth if for every point a M
there exist local coordinate charts (U
) in M and (
) in
M
, such that a U
, f(U
)
U
(U
R
n
is C
k
-smooth. In other words, a map is continuous or smooth, if it is continuous or smooth when
expressed in local coordinates.
A map f : M N is called a dieomorphism if it is smooth, one-to-one, and the inverse map
is also smooth. One-to-one continuous maps with continuous inverses are called homeomorphisms.
Note that in view of the chain rule the C
k
-smoothness is independent of the choice of local
coordinate charts (U
) and (
(U
to points
in R
n1
= R
n
+
are called the boundary points of M. The set of boundary points is called the
boundary of M and denoted by M. It is itself a manifold of dimension n 1.
Note that any (interior) point a of an n-dimensional manifold M has a neighborhood B dieo-
morphic to an open ball B
1
(0) R
n
, while any boundary point has a neighborhood dieomorphic
to a semi-ball B
1
(0) x
1
0 R
n
.
Exercise 9.36. Prove that a boundary point does not have a neighborhood dieomorphic to an
open ball. In other words, the notion of boundary and interior point of a manifold with boundary
are well dened.
Next we want to introduce a notion of compactness for subsets in a manifold. Let us recall that
for subsets in a Euclidean vector space we introduced three equivalent denition of compactness,
see Section 6.1. The rst denition, COMP1 is unapplicable because we cannot talk about bounded
sets in a manifold. However, denitions COMP2 and COMP3 make perfect sense in an arbitrary
manifold. For instance, we can say that a subset A M is compact if from any innite sequence
of points in A one can choose a subsequence converging to a point in A.
A compact manifold (without boundary) is called closed. Note that the word closed is used
here in a dierent sense than a closed set. For instance, a closed interval is not a closed manifold
because it has a boundary. An open interval or a real line R is not a closed manifold because it is
not compact. On the other hand, a circle, or a sphere S
n
of any dimension n is a closed manifold.
The notions of connected and path connected subsets of a manifold are dened in the same way
as in an Euclidean space.
137
9.8.2 Gluing construction
The construction which is described in this section is called gluing or quotient construction. It
provides a rich source of examples of manifolds. We discuss here only very special cases of this
construction.
a) Let M be a manifold and U, U
.
4
Consider a dieomorphism f : U U
.
Let us denote by M/f(x) x the set obtained from M by identifying each point x U with
its image f(x) U
),
or a pair of points (x, f(x)), where x U. Note that there exists a canonical projection : M
M/f(x) x. Namely (x) = x if x / U U
= (3, 5).
Then M is a 1-dimensional manifold. Denote U := (0, 1) I, U
:= (3, 4) I
. Consider a
dieomorphism f : U U
.
138
Figure 9.4: Felix Hausdor (1868-1942)
In order to avoid such pathological examples one usually (but not always) requires that manifolds
satisfy an additional axiom, called Hausdor property:
M5. Any two distinct points x, y M have non-intersecting neighborhoods U x, G y.
In what follows we always assume that the manifolds satisfy the Hausdor property M5.
Let us make the following general remark about dieomorphisms f : (a, b) (c, d) between
two open intervals. Such dieomorphism is simply a dierentiable function whose derivative never
vanishes and whose range is equal to the interval (c, d). If derivative is positive then the dieomor-
phism is orientation preserving, and it is orientation reversing otherwise. The function f always
extends to a continuous (but necessarily dierentiable function f : [a, b] [c, d] such that f(a) = c
and f(b) = d in the orientation preserving case, and f(a) = d f(b) = c in the orientation reversing
case.
Lemma 9.38. Given a, b, a
< b
) and (b, 1) (b
b (b, 1) there
exists a dieomorphism F : (0, 1) (0, 1) which coincides with f on (0, a) and coincides with g on
(
b, 1).
Proof. Choose real numbers c, c,
-
function : (0, 1) (0, 1) which is equal to 1 on (0, a] [c,
d] [
_
(1 )(x)f
(x) on [
,C
is positive, and hence the function F
,C
is strictly increasing. It
coincides with f on (0, a] and coincides up to a constant with g on (
b) < b
< g(
b) = g(
: U
(0, 1) such that (U U
(U U
) = (0, a
) (b
1) for some a, a
, b, b
(0, 1)
with a < b, a
< b
)
1
, and set G := (U U
)
and G
:=
(U U
). Let h =
: G G
), h((b, 1)) = (b
, h(a) = 0,
h(b) = 1 and h(1) = b
:=
(a
) M are disjoint.
On the other hand, for any neighborhood A its image () I contains an interval (a , a),
140
and similarly for any neighborhood
its image
() I contains an interval (a
, a
). for
a suciently small . But h((a , a)) = (a
, a
) for some
((a
, a
)) =
h((a, a )) =
((a, a )) = ((a, a )) ,
i.e.
M intersect, which
violates axiom M5. Similarly we can check that h(b) = b
.
Now take the unit circle S
1
R
2
and consider the polar coordinate on S
1
. Let us dene a
map g
: U
S
1
by the formula =
(x). Thus g
is a dieomorphism of U
onto an arc
of S
1
given in polar coordinates by < < 0. The points A
(a
) and B
(b
) are
mapped to points with polar coordinates = a
and = b
. On the intersection U U
:
(0, a) (b, 1) S
1
. We have f(0) = g
(A
), f(a) = g
(0), f(1) = g
(B
), f(b) = g
(1). Here we
denoted by f the continuous extension of f to [0, a] [b, 1]. Thus f((0, a)) = a
< < 3 b
S
1
which coincides with f on (0, a) (
b, 1) for
any a (0, a) and
:= h(a),
= h(
((a,
F(x) =
_
_
g(x), x
U
;
F (x), x U.
: U
(0, 1) such that U U
i=1
U
x
i
= M. We can further assume that
none of these neighborhoods is completrely contained inside another one. Denote U
1
:= U
x
1
,
1
:=
x
1
. Note that U
1
k
2
U
x
k
,= . Indeed, if this were the case then due to connectedness of M we
would have
k
2
U
x
i
= and hence M = U
1
, but this is impossible because M is compact. Thus,
there exists i = 2, . . . , k such that U
x
i
U
1
,= . We set U
2
:= U
x
i
,
2
=
x
i
. Consider open sets
G
1,2
:=
1
1
(U
1
U
2
), G
2,1
=
1
2
(U
1
U
2
) I. The transition map h
1,2
:=
1
2
1
[
G
1,2
: G
1,2
G
2,1
is a dieomorphism.
Let us show that the set G
1,2
(and hence G
2,1
) cannot contain more that two connected com-
ponents. Indeed, in that case one of the components of G
1,2
has to be a subinterval I
= (a, b)
I = (0, 1) where 0 < a < b < 1. Denote I
:= h
1,2
(I
, has to
be an interior point c I = (0, 1). We will assume for determinacy that I
2
(c) in M we have
2
((c, c+))
))) =
2
(c, c+
),
where c +
= h
1,2
(a+). Hence
n+1
1
x
2
j
= 1 R
n+1
. Let
introduce on S
n
the structure of an n-dimensional manifold. Let N = (0, . . . , 1) and S = (0, . . . , 1)
be the North and South poles of S
n
, respectively.
Denote U
= S
n
S, U
+
= S
n
N and consider the maps p
: U
R
n
given by the formula
p
(x
1
, . . . , x
n+1
) =
1
1 x
n+1
(x
1
, . . . , x
n
). (9.8.1)
The maps p
+
: U
+
R
n
and p
: U
R
n
are called stereographic projections from the North
and South poles, respectively. It is easy to see that stereographic projections are one-to one maps.
Note that U
+
U
= S
n
S, N and both images, p
+
(U
+
U
) and p
(U
+
U
) coincide with
R
n
0. The map p
p
1
+
: R
n
0 R
n
0 is given by the formula
p
p
1
+
(x) =
x
[[x[[
2
, (9.8.2)
and therefore it is a dieomorphism R
n
0 R
n
0.
Thus, the atlas which consists of two coordinate charts (U
+
, p
+
) and (U
, p
) denes on S
n
a
structure of an n-dimensional manifold. One can equivalently denes the manifold S
n
as follows.
Take two disjoint copies of R
n
, let call them R
n
1
and R
n
2
. Denote M = R
n
1
R
n
2
, U = R
n
1
0 and
U
= R
n
2
0. Let f : U U
1
U
j
= RP
n
. There is a natural one-to one map
j
: U
j
j
144
which associates with each line U
j
the unique intersection point of with
j
. Furthermore,
each
j
can be identied with R
n
, and hence pairs (U
j
,
j
), j = 1, . . . , n + 1 can be chosen as an
atlas of coordinate charts. We leave it to the reader to check that this atlas indeed dene on RP
n
a structure of a smooth manifold, i.e. that the transition maps between dierent coordinate charts
are smooth.
Exercise 9.43. Let us view S
n
as the unit sphere in R
n+1
. Consider a map p : S
n
RP
n
which
associates to a point of S
n
the line passing through this point and the origin. Prove that this
two-to-one map is smooth, and moreover a local dieomorphism, i.e. that the restriction of p to
a suciently small neighborhood of each point is a dieomorphism. Use it to show that RP
n
is
dieomorphic to the quotient space S
n
/f, where f : S
n
S
n
is the antipodal map f(x) = x.
3. Products of manifolds and n-dimensional tori. Given two manifolds, M and N of
dimension m and n, respectively, one can naturally endow the direct product
M N = (x, y); x M, y N
with a structure of a manifold of dimension m+n. Let (U
and (V
be atlases
for M and N, so that
: U
R
m
,
: V
R
n
are dieomorphisms on open
subsets of R
m
and R
n
. Then the smooth structure on M N can be dened by an atlas
(U
)
,
,
where we denote by
: U
R
m
R
n
= R
m+n
are dieomorphisms dened
by the formula (x, y) (
(x)
(y)) for x U
and y V
.
One can similarly dene the direct product of any nite number of smooth manifolds. In par-
ticular the n-dimensional torus T
n
is dened as the product of n circles: T
n
= S
1
S
1
. .
n
. Let
us recall, that the circle S
1
is dieomorphic to R/T, i.e. a point of S
1
is a real number up to adding
any integer. Hence, the points of the torus T
n
can viewed as the points of R
n
up to adding any
vector with all integer coordinates.
145
9.8.4 Submanifolds of an n-dimensional vector space
Let V be an n-dimensional vector space. A subset A V is called a k-dimensional submanifold
of V , or simply a k-submanifold of V , 0 k n, if for any points a A there exists a local
coordinate chart (U
a
, u = (u
1
, . . . , u
n
) R
n
) such that u(a) = 0 (i.e. the point a is the origin in
this coordinate system) and
A U
a
= u = (u
1
, . . . , u
n
) U
a
; u
k+1
= = u
n
= 0. (9.8.3)
We will always assume the local coordinates at least as smooth as necessary for our purposes (but
at least C
1
-smooth), but more precisely, one can talk of C
m
- submanifolds if the implied coordinate
systems are at least C
m
-smooth.
Note that in the above we can replace the vector space V by any n-dimensional manifold, and
thus will get a notion of a k-dimesional submanifold of an n-dimensional manifold V .
Example 9.44. Suppose a subset A U V is given by equations F
1
= = F
nk
= 0 for
some C
m
-smooth functions F
1
, . . . , F
nk
on U. Suppose that for any point a A the dierentials
d
a
F
1
, . . . , d
a
F
nk
V
a
are linearly independent. Then A U is a C
m
-smooth submanifold.
Indeed, for each a A one can choose a linear functions l
1
, . . . , l
k
V
a
such that together
with d
a
F
1
, . . . , d
a
F
nk
V
a
they form a basis of V
. Consider functions L
1
, . . . , L
k
: V R,
dened by L
j
(x) = l
j
(x a) so that d
a
(L
j
) = l
j
, j = 1, . . . , k. Then the Jacobian det D
a
F of the
map F : (L
1
, . . . , L
k
, F
1
, . . . , F
nk
) : U R
n
does not vanish at a, and hence the inverse function
theorem implies that this map is invertible in a smaller neighborhood U
a
U of the point a A.
Hence, the functions u
1
= L
1
, . . . , u
k
= L
k
, u
k+1
= F
1
, . . . , u
n
= F
nk
can be chosen as a local
coordinate system in U
a
, and thus A U
a
= u
k+1
= = u
n
= 0.
Note that the map u
= (u
1
, . . . , u
k
) maps U
A
a
= U
a
A onto an open neighborhood
U =
u(U
a
) R
k
of the origin in R
k
R
n
, and therefore u
= (u
1
, . . . u
k
) denes a local coordinates,
so that the pair (U
A
a
, u
f
= (x, y) R
k
R
nk
= R
n
; x B, y = f(x).
A (C
m
)-submanifold A V is called graphical with respect to a splitting : R
k
R
nk
V ,
if there exist an open set U R
k
and a (C
m
)-smooth map f : U R
nk
such that
A = (
f
).
In other words, A is graphical if there exists a coordinate system in V such that
A = x = (x
1
, . . . , x
n
); (x
1
, . . . x
k
) U, x
j
= f
j
(x
1
, . . . , x
k
), j = k + 1, . . . , n.
for some open set U R
k
and smooth functions, f
k+1
, . . . , f
n
: U R.
For a graphical submanifold there is a global coordinate system given by the projection of the
submanifold to R
k
.
It turns out that that any submanifold locally is graphical.
Proposition 9.45. Let A V be a submanifold. Then for any point a A there is a neighborhood
U
a
a such that U
a
A is graphical with respect to a splitting of V . (The splitting may depend on
the point a A).
We leave it to the reader to prove this proposition using the implicit function theorem.
One can generalize the discussion in this section and dene submanifolds of any manifold M,
and not just the vector space V . In fact, the denition (9.8.3) can be used without any changes to
dene submanifolds of an arbitrary smooth manifold.
A map f : M Q is called an embedding of a manifold M into another manifold Q if it is a
dieomorphism of M onto a submanifold A Q. In other words, f is an embedding if the image
A = f(M) is a submanifold of Q and the map f viewed as a map M A is a dieomorphism.
147
One can prove that any n-dimensional manifold can be embedded into R
N
with a suciently large
N (in fact N = 2n + 1 is always sucient).
Hence, one can think of manifold as submanifold of some R
n
given up to a dieomorphism, i.e.
ignoring how this submanifold is embedded in the ambient space.
In the exposition below we mostly restrict our discussion to submanifolds of R
n
rather than
general abstract manifolds.
9.8.5 Submanifolds with boundary
A slightly dierent notion is of a submanifold with boundary. A subset A V is called a k-
dimensional submanifold with boundary, or simply a k-submanifold of V with boundary, 0 k < n,
if for any points a A there is a neighborhood U
a
a in V and local (curvi-linear) coordinates
(u
1
, . . . , u
n
) in U
a
with the origin at a if one of two conditions is satised: condition (9.8.3), or the
following condition
A U
a
= u = (u
1
, . . . , u
n
) U
a
; u
1
0, u
k+1
= = u
n
= 0. (9.8.4)
In the latter case the point a is called a boundary point of A, and the set of all boundary points is
called the boundary of A and is denoted by A.
As in the case of submanifolds without boundary, any submanifold with boundary has a struc-
ture of a manifold with boundary.
Exercise 9.46. Prove that if A is k-submanifold with boundary then A is a (k 1)-dimensional
submanifold (without boundary).
Remark 9.47. 1. As we already pointed out when we discussed manifolds with boundary, a
submanifold with boundary is not a submanifold!
2. As it was already pointed out when we discussed 1-dimensional submanifolds with boundary,
the boundary of a k-submanifold with boundary is not the same as its set-theoretic boundary,
though traditionally the same notation A is used. Usually this should be clear from the
context, what the notation A stands for in each concrete case. We will explicitly point this
dierence out when it could be confusing.
148
Figure 9.5: The parameterization introducing local coordinates near an interior point a and a
boundary point b.
A compact manifold (without boundary) is called closed. The boundary of any compact manifold
with boundary is closed, i.e. (A) = .
Example 9.48. An open ball B
n
r
= B
n
R
(0) =
n
1
x
2
j
< 1 R
n
is a codimension 0 submanifold,
A closed ball D
n
r
= D
n
R
(0) =
n
1
x
2
j
1 R
n
ia codimension 0 submanifold with boundary.
Its boundary D
n
R
is an (n 1)-dimensional sphere S
n1
R
=
n
1
x
2
j
= 1 R
n
. It is a closed
hypersurface. For k = 0, 1 . . . n1 let us denote by L
k
the subspace L
k
= x
k+1
= = x
n
= 0
R
n
. Then the intersections
B
k
R
= B
n
R
L
k
, D
k
R
= D
n
R
L
k
, and S
k1
R
= S
n1
R
L
k
R
n
are, respectively a k-dimensional submanifold, a k-dimensional submanifold with boundary and a
closed (k 1)-dimensional submanifold of R
n
. Among all above examples there is only one (which
one?) for which the manifold boundary is the same as the set-theoretic boundary.
A neighborhood of a boundary point a A can be always locally parameterized by the semi-
149
open upper-half ball
B
+
(0) = x = (x
1
, . . . , x
k
) R
k
; x
1
0,
n
1
x
2
j
< 1.
We will nish this section by dening submanifolds with piece-wise smooth boundary. A subset
A V is called a k-dimensional submanifold of V with piecewise smooth boundary or with boundary
with corners, 0 k < n, if for any points a A there is a neighborhood U
a
a in V and local
(curvi-linear) coordinates (u
1
, . . . , u
n
) in U
a
with the origin at a if one of three condittions satised:
conditions (9.8.3), (9.8.4) or the following condition
A U
a
= u = (u
1
, . . . , u
n
) U
a
; l
1
(u) 0, . . . , l
m
(u) 0, u
k+1
= = u
n
= 0, (9.8.5)
where m > 1 and l
1
, . . . , l
m
(R
k
)
G).
Denote G
1
:=
1
(U),
G
1
:=
1
(
[
e
G
1
_
1
: U
G
1
.
Denote h := u
[
e
G
1
=
G
1
G
1
. We have
= u
= h,
and hence the dierentials d
0
and d
0
of parameterizations and
at the origin map R
k
0
isomor-
phically onto the same k-dimensional linear subspace T V
a
. Indeed, d
0
= d
0
d
0
h. Thus the
space T = d
0
(R
k
0
) V
a
is independent of the choice of parameterization. It is called the tangent
150
space to the submanifold A at the point a A and will be denoted by T
a
A. If A is a submanifold
with boundary and a A then there are dened both the k-dimensional tangent space T
a
A and
its (k 1)-dimensional subspace T
a
(A) T
a
A tangent to the boundary.
Example 9.49. 1. Suppose a submanifold A V is globally parameterized by an embedding
: G A V , G R
k
. Suppose the coordinates in R
k
are denoted by (u
1
, . . . , u
k
). Then
the tangent space T
a
A at a point a = (b), b G is equal to the span
Span
_
u
1
(a), . . . ,
u
k
(a)
_
.
2. In particular, suppose a submanifold A is graphical and given by equations
x
k+1
= g
1
(x
1
, . . . , x
k
), . . . , x
n
= g
nk
(x
1
, . . . , x
k
), (x
1
, . . . , x
k
) G R
k
.
Take points b = (b
1
, . . . b
k
) G and a = (b
1
, . . . , b
k
, g
1
(b), . . . , g
nk
(b)) A. Then T
a
A =
Span(T
1
, . . . T
k
) , where
T
1
=
_
_
1, 0, . . . , 0
. .
k
,
g
1
x
1
(b), . . . ,
g
nk
x
1
(b)
_
_
,
T
1
=
_
_
0, 1, . . . , 0
. .
k
,
g
1
x
2
(b), . . . ,
g
nk
x
2
(b)
_
_
,
. . .
T
1
=
_
_
0, 0, . . . , 1
. .
k
,
g
1
x
k
(b), . . . ,
g
nk
x
k
(b)
_
_
.
3. Suppose a hypersurface R
n
is given by an equation = F = 0 for a smooth function
F dened on an neighborhood of and such that d
a
F ,= 0 for any a . In other words,
the function F has no critical points on . Take a point a . Then T
a
R
n
a
is given by a
linear equation
n
1
F
x
j
(a)h
j
= 0, h = (h
1
, . . . , h
n
) R
n
a
.
Note that sometimes one is interested to dene T
a
as an ane subspace of R
n
= R
n
0
and
not as a linear subspace of R
n
a
. We get the required equation by shifting the origin:
T
a
= x = (x
1
, . . . , x
n
) R
n
;
n
1
F
x
j
(a)(x
j
a
j
) = 0.
151
If for some parameterization : G A with (0) = a the composition f is dierentiable at
0, and the linear map
d
0
(f ) (d
0
)
1
: T
a
A W
f(a)
is called the dierential of f at the point a and denoted, as usual, by d
a
f. Similarly one can dene
C
m
-smooth maps A W.
Exercise 9.50. Show that a map f : A W is dierentiable at a point a A i for some
neighborhood U of a in V there exists a map F : U W that is dierentiable at a and such
that F[
UA
= f[
UA
, and we have dF[
TaA
= d
a
f. As it follows from the above discussion the map
dF[
TaA
is independent of this extension. Similarly, any C
m
-smooth map of A locally extends to a
C
m
-smooth map of a neighborhood of A in V .
Suppose that the image f(A) of a smooth map f : A W is contained in a submanifold B W.
In this case the image d
a
f(T
a
A) is contained in T
f(a)
B. Hence, given a smooth map f : A B
between two submanifolds A V and B W its dierential at a point a can be viewed as a linear
map d
a
f : T
a
A T
f(a)
B.
Let us recall, that given two submanifolds A V and B W (with or without boundary), a
smooth map f : A B is called a dieomorphism if there exists a smooth inverse map : B A,
i.e. f g : B B and g f : A A are both identity map. The submanifolds A and B are called
in this case dieomorphic.
Exercise 9.51. 1. Let A, B be two dieomorphic submanifolds. Prove that
(a) if A is path-connected then so is B;
(b) if A is compact then so is B;
(c) if A = then B = ;
(d) dimA = dimB;
5
2. Give an example of two dieomorphic submanifolds, such that one is bounded and the other
is not.
5
In fact, we will prove later that even homeomorphic manifolds should have the same dimension.
152
3. Prove that any closed connected 1-dimensional submanifold is dieomorphic to the unit circle
S
1
= x
2
1
+x
2
2
= 1 R
2
.
9.10 Vector bundles and their homomorphisms
Let us put the above discussion in a bit more global and general setup.
A collection of all tangent spaces T
a
A
aA
to a submanifold A is called its tangent bundle
and denoted by TA or T(A). This is an example of a more general notion of a vector bundle of
rank r over a set A V . One understands by this a family of r-dimensional vector subspaces
L
a
V
a
, parameterized by points of A and continuously (or C
m
-smoothly) depending on a. More
precisely one requires that each point a A has a neighborhood U A such that there exist linear
independent vector elds v
1
(a), . . . , v
r
(a) L
a
which continuously (smoothly, etc.) depend on a.
Besides the tangent bundle T(A) over a k-submanifold A an important example of a vector
bundle over a submanifold A is its normal bundle NA = N(A), which is a vector bundle of rank
n k formed by orthogonal complements N
a
A = T
a
A V
a
of the tangent spaces T
a
A of A. We
assume here that V is Euclidean space.
A vector bundle L of rank k over A is called trivial if one can nd k continuous linearly
independent vector elds v
1
(a), . . . , v
k
(a) L
a
dened for all a A. The set A is called the base of
the bundle L.
An important example of a trivial bundle is the bundle TV = V
a
aV
.
Exercise 9.52. Prove that the tangent bundle to the unit circle S
1
R
2
is trivial. Prove that the
tangent bundle to S
2
R
3
is not trivial, but the tangent bundle to the unit sphere S
3
R
4
is
trivial. (The case of S
1
is easy, of S
3
is a bit more dicult, and of S
2
even more dicult. It turns
out that the tangent bundle TS
n1
to the unit sphere S
n1
R
n
is trivial if and only if n = 2, 4
and 8. The only if part is a very deep topological fact which was proved by F. Adams in 1960.
Suppose we are given two vector bundles, L over A and
L over
A and a continuous (resp. smooth)
map : A
A. By a continuous (resp. smooth) homomorphism : L
L which covers the map
: A
A we understand a continuous (resp. smooth) family of linear maps
a
: L
a
L
(a)
. For
instance, a C
m
-smooth map f : A B denes a C
m1
-smooth homomorphism df : TA TB
153
which covers f : A B. Here df = d
a
f
aA
is the family of linear maps d
a
f : T
a
A T
f(a)
B,
a A.
9.11 Orientation
By an orientation of a vector bundle L = L
a
aA
over A we understand continuously depending
on a orientation of all vector spaces L
a
. An orientation of a submanifold k is the same as an
orientation of its tangent bundle T(A). A co-orientation of a k- submanifold A is an orientation of
its normal bundle N(A) = T
A in V . Note that not all bundles are orientable, i.e. some bundles
admit no orientation. But if L is orientable and the base A is connected, then L admits exactly two
orientations. Here is a simplest example of a non-orientable rank 1 bundle of the circle S
1
R
2
.
Let us identify a point a S
1
with a complex number a = e
i
, and consider a line l
a
R
2
a
directed
by the vector e
i
2
. Hence, when the point completes a turn around S
1
the line l
a
rotates by the
angle . We leave it to the reader to make a precise argument why this bundle is not orientable .
In fact, rank 1 bundles are orientable if and only if they are trivial.
If the ambient space V is oriented then co-orientation and orientation of a submanifold A deter-
mine each other according to the following rule. For each point a, let us choose any basis v
1
, . . . , v
k
of T
a
(A) and any basis w
1
, . . . , w
nk
of N
a
(A). Then w
1
, . . . , w
nk
, v
1
, . . . , v
k
is a basis of V
a
= V .
Suppose one of the bundles, say N(A), is oriented. Let us assume that the basis w
1
, . . . , w
nk
de-
nes this orientation. Then we orient T
a
A by the basis v
1
, . . . , v
k
if the basis w
1
, . . . , w
nk
, v
1
, . . . , v
k
denes the given orientation of V , and we pick the opposite orientation of T
a
A otherwise.
Example 9.53. (Induced orientation of the boundary of a submanifold.) Suppose A is an
oriented manifold with boundary. Let us co-orient the boundary A by orienting the rank 1 normal
bundle to T(A) in T(A) by the unit ourtward normal to T(A) in T(A) vector eld. Then the
above rule determine an orientation of T(A), and hence of A.
9.12 Integration of dierential k-forms over k-dimensional sub-
manifolds
Let be a dierential k-form dened on an open set U V .
154
Figure 9.6: The orientation of the surface is induced by its co-orientation by the normal vector n.
The orientation of the boundary us induced by the orientation of the surface.
155
Consider rst a k-dimensional compact submanifold with boundary A U dened parametri-
cally by an embedding : G A U, where G R
k
is possibly with boundary. Suppose that A
is oriented by this embedding. Then we dene
_
A
:=
_
G
.
Note that if we dene A by a dierent embedding
:
G A, then we have
= h, where
h =
1
:
G G is a dieomorphism. Hence, using Theorem 9.30 we get
_
e
G
=
_
e
G
h
) =
_
G
,
and hence
_
A
is independent of a choice of parameterization, provided that the orientation is
preserved.
Let now A be any compact oriented submanifold with boundary. Let us choose a partition of
unity 1 =
K
j
in a neighborhood of A such that each function is supported in some coordinate
neighborhood of A. Denote
j
=
j
. Then =
K
j
, where each form
j
is supported in one
of coordinate neighborhoods. Hence there exist orientation preserving embeddings
j
: G
j
A
of domains with boundary G
j
R
k
, such that
j
(G
j
) Supp(
j
), j = 1, . . . , K. Hence, we can
dene
_
A
j
:=
_
G
j
j
and
_
A
:=
K
1
_
A
j
.
Lemma 9.54. The above denition of
_
A
is independent of a choice of a partition of unity.
Proof. Consider two dierent partitions of unity 1 =
K
j
and 1 =
e
K
j
subordinated to coverings
U
1
, . . . , U
K
and
U
1
, . . . ,
U
e
K
, respectively. Taking the product of two partitions we get another
partition 1 =
K
i=1
e
K
j=1
ij
, where
ij
:=
i
j
, which is subordinated to the covering by intersections
U
i
U
j
, i = 1, . . . , K, j = 1, . . . ,
K. Denote
i
:=
i
,
j
:=
j
and
ij
=
ij
. Then
K
i=1
ij
=
j
,
156
e
K
j=1
ij
=
i
and =
K
i
=
e
K
1
j
. Then, using the linearity of the integral we get
K
1
_
A
i
=
K
i=1
e
K
j=1
_
A
ij
=
e
K
1
_
A
j
.
When k = 1 the above denition of the integral coincides with the denition of the integral of
a 1-form over an oriented curve which was given above in Section 9.2.
Let us extend the denition of integration of dierential forms to an important case of integration
of 0-form over oriented 0-dimensional submanifolds. Let us recall a compact oriented 0-dimensional
submanifold of V is just a nite set of points a
1
, . . . , a
m
V with assigned signs to every point.
So in view of the additivity of the integral it is sucient to dene integration over 1 point with a
sign. On the other hand, a 0-form is just a function f : V R. So we dene
_
a
f := f(a).
A partition of unity is a convenient tool for studying integrals, but not so convenient for practical
computations. The following proposition provides a more practical method for computations.
Proposition 9.55. Let A be a compact oriented submanifold of V and a dierential k-form given
on a neighborhood of A. Suppose that A presented as a union A =
N
1
A
j
, where A
j
are codimension
0 submanifolds of A with boundary with corners. Suppose that A
i
and A
j
for any i ,= j intersect
only along pieces of their boundaries. Then
_
A
=
N
1
_
A
j
.
In particular, if each A
j
is parameterized by an orientation preserving embedding
j
: G
j
A,
where G
j
R
k
is a domain with boundary with corners. Then
_
A
=
N
1
_
G
j
j
.
157
We leave the proof to the reader as an exercise.
Exercise 9.56. Compute the integral
_
S
1/3(x
1
dx
2
dx
3
+x
2
dx
3
dx
1
+x
3
dx
1
dx
2
),
where S is the sphere
x
2
1
+x
2
2
+x
2
3
= 1,
cooriented by its exterior normal vector.
Solution. Let us present the sphere as the union of northern and southern hemispheres:
S = S
S
+
, where S
= S x
3
0, S
+
= S x
3
0.
Then
_
S
=
_
S
+
+
_
S
R
2
u
2
v
2
), (u, v) u
2
+ v
2
R
2
=
T
R
. One can check that this parametrization agrees with the prescribed coorientation of S. Thus,
we have
_
S
+
= 1/3
_
D
R
_
udv d
_
R
2
u
2
v
2
+vd
_
R
2
u
2
v
2
du +
_
R
2
u
2
v
2
du dv
_
.
Passing to polar coordinates (r, ) in the plane (u, v) we get
_
S
+
= 1/3
_
P
r cos d(r sin) d
_
R
2
r
2
+r sind
_
R
2
r
2
d(r cos )
+
_
R
2
r
2
d(r cos ) d(r sin),
where P = 0 r R, 0 2. Computing this integral we get
158
_
S
+
=
1
3
_
P
r
3
cos
2
d dr
R
2
r
2
+
r
3
sin
2
dr d
R
2
r
2
+
_
R
2
r
2
dr d
=
1
3
_
P
_
r
3
R
2
r
2
+r
_
R
2
r
2
_
dr d
=
2
3
_
R
0
rR
2
R
2
r
2
dr =
2R
2
3
_
R
2
r
2
_
R
0
=
2R
3
3
Similarly, one can compute that
_
S
=
2R
3
3
.
Computing this last integral, one should notice the fact that the parametrization
(u, v) (u, v,
_
R
2
u
2
v
2
)
denes the wrong orientation of S
,
or
_
U
_
P
2
x
1
P
1
x
2
_
dx
1
dx
2
=
_
P
1
dx
1
+P
2
dx
2
.
163
Figure 10.1: George Stokes (1819-1903)
This is called Greens formula. In particular, when d = dx
1
dx
2
, e.g. = xdy or =
1
2
(xdy ydx), the integral
_
df =
_
B(A)
f = f(B) f(A),
which we already proved earlier, see Theorem 9.11 .
Consider now the case n = 3, k = 1.
Thus V = R
3
and = P
1
dx
1
+P
2
dx
2
+P
3
dx
3
. Let S R
3
be an oriented surface with boundary
165
Figure 10.3: Carl Friedrich Gauss (1777-1855) Mikhail Ostrogradski (1801-1862)
. We orient in the same way, as in Greens theorem. Then Stokes formula
_
S
d =
_
P
2
x
3
_
dx
2
dx
3
+
_
P
1
x
3
P
3
x
1
_
dx
3
dx
1
+
_
P
2
x
1
P
1
x
2
_
dx
1
dx
2
=
_
P
1
dx
1
+P
2
dx
2
+P
3
dx
3
.
This is the original Stokes theorem.
Stokes theorem allows one to clarify the geometric meaning of the exterior dierential.
Lemma 10.2. Let be a dierential k-form in a domain U V . Take any point a U and
vectors X
1
, . . . , X
k+1
V
a
. Given > 0 let us consider the parallelepiped P(X
1
, . . . , X
k+1
) as a
subset of V with vertices at points a
i
1
...i
k+1
= a +
k+1
1
i
j
X
j
, where each index i
j
takes values 0, 1.
Then
d
a
(X
1
, . . . X
k+1
) = lim
0
1
k+1
_
P(X
1
,...,X
k+1
)
.
166
Proof. First, it follows from the denition of integral of a dierential form that
d
a
(X
1
, . . . , X
k+1
) = lim
0
1
k+1
_
P(X
1
,...,X
k+1
)
d. (10.1.1)
Then we can continue using Stokes formula
d
a
(X
1
, . . . , X
k+1
) = lim
0
1
k+1
_
P(X
1
,...,X
k+1
)
d = lim
0
1
k+1
_
P(X
1
,...,X
k+1
)
. (10.1.2)
j
be a partition of unity subordinated to this covering. We
set
j
=
j
, so that =
N
j
, and each of
j
is supported in one of coordinate neighborhoods.
Hence, if formula 10.1 holds for each
j
it also holds for .
Let us now assume that is supported in one of coordinate neighborhoods. Consider the
corresponding parameterization : G U V , G R
n
, introducing coordinates u
1
, . . . , u
n
. Then
AU = (GL), where L is equal to the subspace R
k
= u
k+1
= . . . u
n
= 0 in the case (9.8.3) and
the upper-half space R
k
u
1
0. By denition, we have
_
A
d =
_
U
d =
_
GL
d =
_
GL
d
.
1
Though the form =
[
GL
is dened only on G L, it is supported in this neighborhood,
and hence we can extend it to a smooth form on the whole L by setting it equal to 0 outside the
neighborhood. With this extension we have
_
GL
d =
_
L
d . The (k 1)-form can be written in
coordinates u
1
, . . . , u
k
as
=
j
1
f
j
(u)du
1
. . .
j
du
k
.
1
We assume here that the coordinates u1, . . . , u
k
dene the given orientation of A.
167
Then
_
GL
d =
_
L
_
k
1
f
j
u
j
_
du
1
du
k
.
Let us choose a suciently R > 0 so that the cube I = [u
i
[ R, i = 1, . . . , k contains Supp( ).
Thus in the case (9.8.3) we have
_
GL
d =
k
1
_
R
k
f
j
u
j
dV =
k
1
R
_
R
. . .
R
_
R
f
j
u
j
du
1
. . . du
n
=
k
1
R
_
R
. . .
_
_
R
_
R
f
j
u
j
du
j
_
_
du
1
. . . du
j1
du
j+1
. . . du
n
= 0 (10.2.1)
because
R
_
R
f
j
u
j
du
j
= f
j
(u
1
, . . . , u
i1
, R, u
i
, . . . , u
n
) f
j
(u
1
, . . . , u
i1
, R, u
i
, . . . , u
n
)) = 0.
On the other hand, in this case
_
A
= 0, because the support of does not intersect the boundary
of A. Hence, Stokes formula holds in this case. In case (9.8.4) we similarly get
_
GL
d =
k
1
_
{u
1
0}
f
j
u
j
dV =
k
1
R
_
0
_
_
R
_
R
. . .
R
_
R
f
j
u
j
du
n
. . . du
2
_
_
du
1
=
R
_
R
_
_
R
_
R
. . .
R
_
0
f
1
u
1
du
1
. . . du
n1
_
_
du
n
=
R
_
R
. . .
R
_
R
f
1
(0, u
2
, . . . , u
n
)du
2
. . . du
n
. (10.2.2)
because all terms in the sum with j > 1 are equal to 0 by the same argument as in (10.2.1). On
the other hand, in this case
_
A
=
_
{u
1
=0}
=
_ _
{u
1
=0}
f
1
(0, u
2
, . . . , u
n
)du
2
du
n
=
R
_
R
. . .
R
_
R
f
1
(0, u
2
, . . . , u
n
)du
2
. . . du
n
. (10.2.3)
168
The sign minus appears in the last equality in front of the integral because the induced orien-
tation on the space u
1
= 0 as the boundary of the upper-half space u
1
0 is opposite to the
orientation dened by the volume form du
2
du
n
. Comparing the expressions (10.2.2) and
(10.2.3) we conclude that
_
A
d =
_
A
, as required.
10.3 Integration of functions over submanifolds
In order to integrate functions over a submanifold we need a notion of volume for subsets of the
submanifold.
Let A V be an oriented k-dimensional submanifold, 0 k n. By denition, the volume
form =
A
of A (or the area form if k = 2, or the length form if k = 1) is a dierential k-form
on A whose value on any k tangent vectors v
1
, . . . , v
k
T
x
A equals the oriented volume of the
parallelepiped generated by these vectors.
Given a function f : A R we dene its integral over A by the formula
_
A
fdV =
_
A
f
A
, (10.3.1)
and, in particular,
Vol A =
_
A
A
.
Notice that the integral
_
A
fdV is independent of the orientation of A. Indeed, changing the orienta-
tion we also change the sign of the form
A
, and hence the integral remains unchanged. This allows
us to dene the integral
_
A
fdV even for a non-orientable A. Indeed, we can cover A by coordinate
charts, nd a subordinated partition of unity and split correspondingly the function f =
N
1
f
j
in such a way that each function f
j
is supported in a coordinate neighborhood. By orienting in
arbitrary ways each of the coordinate neighborhoods we can compute each of the integrals
_
A
f
j
dV ,
j = 1, . . . , N. It is straightforward to see that the integral
_
A
fdV =
j
_
A
f
j
dV is independent of the
choice of the partition of unity.
Let us study in some examples how the form
A
can be eectively computed.
169
Example 10.3. Volume form of a hypersurface. Let us x a Cartesian coordiantes in V . Let A V
is given by the equation
A = F = 0
for some function F : V R which has no critical points on A. The vector eld F is orthogonal
to A, and
n =
F
[[F[[
is the unit normal vector eld to A. Assuming A to be co-oriented by n we can write down the
volume form of A as the contraction of n with the volume form = dx
1
dx
n
of R
n
, i.e.
A
= n =
1
[[F[[
n
1
(1)
i1
F
x
i
dx
1
i
. . . dx
n
.
In particular, if n = 3 we get the following formula for the area form of an implicitely given
2-dimensional surface A = F = 0 R
3
:
A
=
1
_
_
F
x
1
_
2
+
_
F
x
2
_
2
+
_
F
x
3
_
2
_
F
x
1
dx
2
dx
3
+
F
x
2
dx
3
dx
1
+
F
x
3
dx
1
dx
2
_
. (10.3.2)
Example 10.4. Length form of a curve.
Let R
n
be an oriented curve given parametrically by a map : [a, b] R
n
. Let =
be
the length form. Let us compute the form
= f(t)dt,
where
f(t) =
(e) =
(t)
_
= [[
(t)[[ .
In particular the length of is equal to
_
=
b
_
a
[[
(t)[[dt =
b
_
a
_
n
i=1
(x
i
(t))
2
dt ,
where
(t) = (x
1
(t), . . . , x
n
(t)) .
170
Similarly, given any function f : R we have
_
fds =
b
_
a
f ((t)) [[
(t)[[dt .
Example 10.5. Area form of a surface given parametrically.
Suppose a surface S R
n
is given parametrically by a map : U R
n
where U in the plane R
2
with coordinates (u, v).
Let us compute the pull-back form
S
. In other words, we want to express
S
in coordinates
u, v. We have
S
= f(u, v)du dv .
To determine f(u, v) take a point z = (u, v) R
2
and the standard basis e
1
, e
2
R
2
z
. Then
(
S
)
z
(e
1
, e
2
) = f(u, v)du dv(e
1
, e
2
) . (10.3.3)
On the other hand, by the denition of the pull-back form we have
(
S
)
z
(e
1
, e
2
) = (
S
)
(z)
(d
z
(e
1
), d
z
(e
2
)) . (10.3.4)
But d
z
(e
1
) =
u
(z) =
u
(z) and d
z
(e
2
) =
v
(z) =
v
(z). Hence from (10.3.3) and (10.3.4) we
get
f(u, v) =
S
(
u
,
v
) . (10.3.5)
The value of the form
S
on the vectors
u
,
v
is equal to the area of the parallelogram gener-
ated by these vectors, because the surface is assumed to be oriented by these vectors, and hence
S
(
u
,
v
) > 0. Denoting the angle between
u
and
v
by we get
2
S
(
u
,
v
) = [[
u
[[ [[
v
[[ sin.
Hence
S
(
u
,
v
)
2
= [[
u
[[
2
[[
v
[[
2
sin
2
= [[
u
[[
2
[[
v
[[
2
(1 cos
2
) = [[
u
[[
2
[[
v
[[
2
(
u
v
)
2
,
and therefore,
f(u, v) =
S
(
u
,
v
) =
_
[[
u
[[
2
[[
v
[[
2
(
u
v
)
2
.
2
See a computation in a more general case below in Example 10.6.
171
It is traditional to introduce the notation
E = [[
u
[[
2
, F =
u
v
, G = [[
v
[[
2
,
so that we get
S
=
_
EGF
2
du dv ,
and hence we get for any function f : S R
_
S
fdS =
_
S
f
S
=
_
U
f((u, v))
_
EGF
2
du dv =
_ _
U
f((u, v))
_
EGF
2
dudv . (10.3.6)
Consider a special case when the surface S dened as a graph of a function over a domain
D R
2
. Namely, suppose
S = z = (x, y), (x, y) D R
2
.
The surface S as parametrized by the map
(x, y)
(x, y, (x, y)).
Then
E = [[
x
[[
2
= 1 +
2
x
, G = [[
y
[[
2
= 1 +
2
y
, F =
x
y
=
x
y
,
and hence
EGF
2
= (1 +
2
x
)(1 +
2
y
)
2
x
2
y
= 1 +
2
x
+
2
y
.
Therefore, the formula (10.3.6) takes the form
_
S
fdS =
_ _
D
f((x, y))
_
EGF
2
dx dy =
_ _
D
f (x, y, (x, y))
_
1 +
2
x
+
2
y
dxdy . (10.3.7)
Note that the formula (10.3.7) can be also deduced from (10.3.2). Indeed, the surface
S = z = (x, y), (x, y) D R
2
,
can also be dened implicitly by the equation
F(x, y, z) = z (x, y) = 0, (x, y) D.
172
We have
F = (
x
,
y
, 1),
and, therefore,
_
S
fdS =
_
S
f(x, y, z)
[[F[[
_
F
x
dy dz +
F
y
dz dx +
F
z
dx dy
_
=
_ _
D
f(x, y, (x, y))
_
1 +
_
x
_
2
+
_
y
_
2
_
x
dy d
y
d dx +dx dy
_
=
_ _
D
f(x, y, (x, y))
_
1 +
_
x
_
2
+
_
y
_
2
_
1 +
_
x
_
2
+
_
y
_
2
_
dxdy
=
_ _
D
f(x, y, (x, y))
1 +
_
x
_
2
+
_
y
_
2
dxdy .
Example 10.6. Integration over a parametrically given k-dimensional submanfold.
Consider now a more general case of a parametrically given k-submanifold A in an n-dimensional
Euclidean space V . We x a Cartersian coordinate system in V and thus identify V with R
n
with
the standard dot-product.
Let U R
k
be a compact domain with boundary and : U R
n
be an embedding. We assume
that the submanifold with boundary A = (U) is oriented by this parameterization. Let
A
be the
volume form of A. We will nd an explicit expression for
A
. Namely. denoting coordinates in
R
k
by (u
1
, . . . , u
k
) we have
A
= f(u)du
1
du
k
, and our goal is to compute the function f.
By denition, we have
f(u) =
(
A
)
u
(e
1
, . . . , e
k
) = (
A
)
u
(d
u
(e
1
), . . . , d
u
(e
k
)) =
(
A
)
u
_
u
1
(u), . . . ,
u
k
(u)
_
= Vol
k
_
P
_
u
1
(u), . . . ,
u
k
(u)
__
(10.3.8)
In Section 4.2 we proved two formulas for the volume of a parallelepiped. Using formula (4.2.1)
we get
Vol
k
_
P
_
u
1
(u), . . . ,
u
k
(u)
__
=
1i
1
<<i
k
n
Z
2
i
1
...i
k
. (10.3.9)
173
where
Z
i
1
...i
k
=
i
1
u
1
(u) . . .
i
1
u
k
(u)
. . . . . . . . .
i
k
u
1
(u) . . .
i
k
u
k
(u)
. (10.3.10)
Thus
_
A
fdV =
_
A
f
A
=
_
U
f((u))
1i
1
<<i
k
n
Z
2
i
1
...i
k
du
1
du
k
.
Rewriting this formula for k = 2 we get
_
A
fdV =
_
U
f((u))
1i<jn
i
u
1
i
u
2
j
u
1
j
u
2
2
du
1
du
2
. (10.3.11)
Alternatively we can use formula (??). Then we get
Vol
k
_
P
_
u
1
(u), . . . ,
u
k
(u)
__
=
_
det
_
(D)
T
D
_
, (10.3.12)
where
D =
_
_
_
_
_
1
u
1
. . .
1
u
k
. . . . . . . . .
n
u
1
. . .
n
u
k
_
_
_
_
_
is the Jacobi matrix of .
3
Thus using this expression we get
_
A
fdV =
_
A
f
A
=
_
U
f((u))
_
det
_
(D)
T
D
_
dV. (10.3.13)
Exercise 10.7. In case n = 3, k = 2 show explicitely equivalence of formulas (10.3.6), (10.3.11)
and (10.3.13).
Exercise 10.8. Integration over an implicitly dened k-dimensional submanfold. Suppose that
A = F
1
= = F
nk
= 0 and the dierentials of dening functions are linearly independent at
points of A. Show that
A
=
(dF
1
dF
nk
)
[[ dF
1
dF
nk
[[
.
3
The symmetric matrix (D)
T
D is the Gram matrix of vectors
u
1
(u), . . . ,
u
k
(u), and its entrees are pairwise
scalar products of these vectors, see Remark ??.
174
Example 10.9. Let us compute the volume of the unit 3-sphere S
3
= x
2
1
+x
2
2
+x
2
3
+x
2
4
= 1.
By denition, Vol(S
3
) =
_
S
3
n , where = dx
1
dx
2
dx
3
dx
4
, and n is the outward unit
normal vector to the unit ball B
4
. Here S
3
should be co-oriented by the vector eld n. Then using
Stokes theorem we have
_
S
3
n =
_
S
3
(x
1
dx
2
dx
3
dx
4
x
2
dx
1
dx
3
dx
4
+x
3
dx
1
dx
2
dx
4
x
4
dx
1
dx
2
dx
3
) =
4
_
B
4
dx
1
dx
2
dx
3
dx
4
= 4
_
B
4
dV.
(10.3.14)
Introducing polar coordinates (r, ) and (, ) in the coordinate planes (x
1
, x
2
) and (x
3
, x
4
) and
using Fubinis theorem we get
_
B
4
=
2
_
0
2
_
0
1
_
0
1r
2
_
0
rddrdd =
2
2
1
_
0
(r r
3
)dr =
2
2
. (10.3.15)
Hence, Vol(S
3
) = 2
2
.
Exercise 10.10. Find the ratio
Voln(B
n
R
)
Vol
n1
(S
n1
R
)
.
10.4 Work and Flux
We introduce in this section two fundamental notions of vector analysis: a work of a vector eld
along a curve, and a ux of a vector eld through a surface. Let be an oriented smooth curve in
a Euclidean space V and T the unit tangent vector eld to . Let v be another vector eld, dened
along . The function v, T) equals the projection of the vector eld v to the tangent directions
to the curve. If the vector eld v is viewed as a force eld, then the integral
_
v, T)ds. As we
175
already indicated earlier, the sign
_
in this case has precisely the same meaning as
_
, and it is used
only to stress the point that we are integrating along a closed curve.
Consider now a co-oriented hypersurface V and denote by n the unit normal vector eld
to which determines the given co-orientation of . Given a vector eld v along we will view
it as the velocity vector eld of a ow of a uid in the space. Then we can interpret the integral
_
v, n)dV
as the ux Flux
= ( v )
,
where is the volume form in V .
2. For any oriented curve and a vector eld v near we have
v, T)
= T( v)[
.
Proof. 1. For any n 1 vectors T
1
, . . . , T
n1
T
x
we have
v (T
1
, . . . , T
n1
) = ( v, T
1
, . . . , T
n1
) = VolP( v, T
1
, . . . , T
n1
).
Using (3.3.1) we get
VolP( v, T
1
, . . . , T
n1
) = v, n)Vol
n1
P(T
1
, . . . , T
n1
) =
v, n)VolP( n, T
1
, . . . , T
n1
) = v, n)
(T
1
, . . . , T
n1
). (10.4.1)
2. The tangent space T
x
is generated by the vector T, and hence we just need to check that
v, T)
(T).
176
1
a
j
x
j
, then
v =
n1
1
(1)
i1
a
i
dx
1
i
. . . dx
n
, T( v) =
n
1
a
i
dx
i
.
Thus, we have
Corollary 10.12.
Flux
( v) =
_
v, n)dV =
_
( v ) =
_
1
(1)
i1
a
i
dx
1
i
. . . dx
n
;
Work
( v) =
_
v, T) =
_
T( v) =
_
1
a
i
dx
i
.
In particular if n = 3 we have
Flux
( v) =
_
a
1
dx
2
dx
3
+a
2
dx
3
dx
1
+a
3
dx
1
dx
2
.
Let us also recall that in a Euclidean space V we have v = T(v). Hence, the equation
= v is equivalent to the equation
v = T
1
_
_
= (1)
n1
T
1
().
In particular, when n = 3 we get v = T
1
(). Thus we get
Corollary 10.13. For any dierential (n1)-form and an oriented compact hypersurface we
have
_
= Flux
v,
where v = (1)
n1
T
1
().
Integration of functions along curves and surfaces can be interpreted as the work and the ux
of appropriate vector elds. Indeed, suppose we need to compute an integral
_
fds = Work
( v) =
_
T( v)
177
.
Note that we can also express v through by the formula v = T
1
, see Section 8.7.
Similarly, to compute an integral
_
fdS =
_
v, n)dS = Flux
( v) =
_
,
where = v .
10.5 Integral formulas of vector analysis
We interpret in this section Stokes formula in terms of integrals of functions and operations on
vector elds. Let us consider again dierential forms, which one can associate with a vector eld
v in an Euclidean 3-space. Namely, this is a dierential 1-form = T( v) and a dierential 2-form
= v , where = dx dy dz is the volume form.
Using Corollary 10.12 we can reformulate Stokes theorem for domains in a R
3
as follows.
Theorem 10.14. Let v be a smooth vector eld in a domain U R
3
with a smooth (or piece-wise)
smooth boundary . Suppose that is co-oriented by an outward normal vector eld. Then we have
Flux
v =
_ _ _
U
div vdxdydz .
Indeed, div v = d. Hence we have
_
U
div vdV =
_
U
(d)dx dy dz =
_
U
d =
_
=
_
v = Flux
v .
This theorem claries the meaning of div v:
Let B
r
(x) be the ball of radius r centered at a point x R
3
, and S
r
(x) = B
r
(x) be its
boundary sphere co-oriented by the outward normal vector eld. Then
div v(x) = lim
r0
Flux
Sr(x)
v
Vol(B
r
(x))
.
178
Theorem 10.15. Let be a piece-wise smooth compact oriented surface in R
3
with a piece-wise
smooth boundary = oriented respectively. Let v be a smooth vector eld dened near . Then
Flux
(curl v) =
_
curl v, n)dV =
_
v Tds = Work
v .
To prove the theorem we again use Stokes theorem and the connection between integrals of
functions and dierential forms. Set = T( v). Then curl v = T
1
d. We have
_
v Tds =
_
=
_
d = Flux
(T
1
(d)) = Flux
(curl v) .
Again, similar to the previous case, this theorem claries the meaning of curl. Indeed, let us
denote by D
r
(x, w) the 2-dimensional disc of radius r in R
3
centered at a point x R
3
and
orthogonal to a unit vector w R
3
x
. Set
c(x, w) = lim
r0
Work
Dr(x, w)
v
r
2
.
Then
c(x, w) = lim
r0
Flux
Dr(x, w)
(curl v)
r
2
= lim
r0
_
Dr(x, w)
curl v, w)
r
2
= curl v, w).
Hence, [[curl v(x)[[ = max
w
c(x, w) and direction of curl v(x) coincides with the direction of the
vector w for which the maximum value of c(x, w) is achieved.
10.6 Expressing div and curl in curvilinear coordinates
Let us show how to compute div v and curl v of a vector eld v in R
3
given in a curvilinear
coordinates u
1
, u
2
, u
3
, i.e. expressed through the coordinate vector elds
u
1
,
u
2
and
u
3
. Let
= f(u
1
, u
2
, u
3
)du
1
du
2
du
3
be the volume form dx
1
dx
2
dx
3
expressed in coordinates u
1
, u
2
, u
3
.
Let us rst compute (du
1
du
2
du
3
). We have
(du
1
du
2
du
3
) =
_
1
f
dx
1
dx
2
dx
3
_
=
1
f
.
179
Let
v = a
1
u
1
+a
2
u
2
+a
3
u
3
.
Then we have
div v = d( v )
= d
__
3
1
a
i
u
i
_
fdu
1
du
2
du
3
_
= d (fa
1
du
2
du
3
+fa
2
du
3
du
1
+fa
3
du
1
du
2
)
=
__
(fa
1
)
u
1
+
(fa
2
)
u
2
+
(fa
3
)
u
3
_
du
1
du
2
du
3
_
=
1
f
_
(fa
1
)
u
1
+
(fa
2
)
u
2
+
(fa
3
)
u
3
_
=
a
1
u
1
+
a
2
u
2
+
a
3
u
3
+
1
f
_
f
u
1
a
1
+
f
u
2
a
2
+
f
u
3
a
3
_
.
In particular, we see that the divergence of a vector eld is expressed by the same formulas as in
the cartesian case if and only if the volume form is proportional to the form du
1
du
2
du
3
with
a constant coecient.
For instance, in the spherical coordinates the volume form can be written as
= r
2
sindr d d ,
4
and hence the divergence of a vector eld
v = a
r
+b
+c
+
c
+
2a
r
+c cot .
The general formula for curl v in curvilinear coordinates looks pretty complicated. So instead
of deriving the formula we will explain here how it can be obtained in the general case, and then
illustrate this procedure for the spherical coordinates.
By the denition we have
4
Note that the spherical coordinates ordered as (r, , ) determine the same orientation of R
3
as the cartesian
coordinates (x, y, z).
180
curl v = D
1
( (d (D( v)))) .
Hence we rst need to compute D( v).
To do this we need to introduce a symmetric matrix
G =
_
_
_
_
_
g
11
g
12
g
13
g
21
g
22
g
23
g
31
g
32
g
33
_
_
_
_
_
,
where
g
ij
=
u
i
,
u
j
), i, j = 1, 2, 3 .
The matrix G is called the Gram matrix.
Notice, that if D( v) = A
1
du
1
+Bdu
2
+Cdu
3
then for any vector h = h
1
u
1
+h
2
u
2
+h
3
u
3
.
we have
D( v)(h) = A
1
h
1
+A
2
h
2
+A
3
h
3
=
_
A
1
A
2
A
3
_
_
_
_
_
_
h
1
h
2
h
3
_
_
_
_
_
= v, h) .
But
v, h) =
_
a
1
a
2
a
3
_
G
_
_
_
_
_
h
1
h
2
h
3
_
_
_
_
_
.
Hence
_
A
1
A
2
A
3
_
=
_
a
1
a
2
a
3
_
G,
or, equivalently, because the Gram matrix G is symmetric we can write
_
_
_
_
_
A
1
A
2
A
3
_
_
_
_
_
= G
_
_
_
_
_
a
1
a
2
a
3
_
_
_
_
_
,
and therefore,
A
i
=
n
i=1
g
ij
a
j
, i = 1, 2, 3 .
181
After computing
= d(D( v)) = B
1
du
2
du
3
+B
2
du
3
du
1
+B
3
du
1
du
2
we compute curl v by the formula = curl v = D
1
(). Let us recall (see Proposition 4.13 above)
that for any vector eld w the equality Dw = is equivalent to the equality w = , where
= fdu
1
du
2
du
3
is the volume form. Hence, if
curl v = c
1
u
1
+c
2
u
2
+c
3
u
3
then we have
w = fc
1
du
2
du
3
+fc
2
du
3
du
1
+fc
3
du
1
du
2
,
and therefore,
curl v =
B
1
f
u
1
+
B
2
f
u
2
+
B
3
f
u
3
.
Let us use the above procedure to compute curl v of the vector eld
v = a
r
+b
+c
given in the spherical coordinates. The Gram matrix in this case is the diagonal matrix
G =
_
_
_
_
_
1 0 0
0 r
2
0
0 0 r
2
sin
2
_
_
_
_
_
.
Hence,
Dv = adr +br
2
d +cr
2
sin
2
d ,
and
= d(Dv) = da dr +d(br
2
) d +d(cr
2
sin
2
) d
=
_
r
b
+r
2
c sin2 +r
2
sin
2
_
d d
+
_
+r
2
b
r
+ 2br
_
dr d +
_
2rc sin
2
r
2
sin
2
c
r
+
a
_
d dr .
Finally, we get the following expression for curl v:
curl v =
r
b
+r
2
c sin2 +r
2
sin
2
r
2
cos
r
+
+r
2 b
r
+ 2br
r
2
cos
+
2rc sin
2
r
2
sin
2
c
r
+
a
r
2
cos
.
182
Chapter 11
Applications of Stokes formula
11.1 Integration of closed and exact forms
Let us recall that a dierential k-form is called closed if d = 0, and that it is called exact if
there exists a (k 1)-form , called primitive of , such that = d.
Any exact form is closed, because d(d) = 0. Any n-form in a n-dimensional space is closed.
Proposition 11.1. a) For a closed k-form dened near a (k + 1)-dimensional submanifold
with boundary we have
_
= 0 .
b) If is exact k-form dened near a closed k-dimensional submanifold S then
_
S
= 0 .
The proof immediately follows from Stokes formula. Indeed, in case a) we have
_
=
_
d = 0 .
In case b) we have = d and S = . Thus
_
S
d =
_
= 0 .
Proposition 11.1b) gives a necessary condition for a closed form to be exact.
183
Example 11.2. The dierential 1-form =
1
x
2
+y
2
(xdy ydx) dened on the punctured plane
R
2
0 is closed but not exact.
Indeed, it is straightforward to check that is exact (one can simplify computations by passing
to polar coordinates and computing that = d). To check that it is not exact we compute the
integral
_
S
, where S in the unit circle x
2
+y
2
= 1. We have
_
S
=
2
_
0
d = 2 ,= 0 .
More generally, an (n 1)-form
n
=
n
i=1
(1)
i1
x
i
r
n
dx
1
i
. . . dx
n
(dx
i
is missing)
is closed in R
n
0. However, it is not exact. Indeed, let us show that
_
S
n1
n
,= 0, where S
n1
is the
unit sphere oriented as the boundary of the unit ball. Let us recall that the volume form
S
n1 on
the unit sphere is dened as
S
n1 = n =
n
i=1
(1)
i1
x
i
r
dx
1
i
. . . dx
n
.
Notice that
n
[
S
n1 =
S
n1, and hence
_
S
n1
n
=
_
S
n1
S
n1 =
_
S
n1
dV = Vol(S
n1
) > 0.
11.2 Approximation of continuous functions by smooth ones
Theorem 11.3. Let C V be a compact domain with smooth boundary. Then any continuous
function f : C R can be C
0
-approximated by C
-smooth function g : C R such that [f(x) g(x)[ < for any x C. Moreover, if
the function f is already C
_
f(x)
if t [0,
2
] and f(x, t) = 0 if t (
2
, ). Now we can extend f to U by the formula f(y) = F
1
(y)
if y U C, and setting it to 0 outside U.
Consider the function
=
1
_
D(0)
0,
dV
0,
,
where
0,
is a bump function dened above in (9.1.2). It is supported in the disc D
(0), non-
negative, and satises
_
D(0)
dV = 1.
. Given a continuous function f : V R we dene a function f
: V R by the formula
f
(x) =
_
f(x y)
(y)d
n
y. (11.2.1)
Then
Lemma 11.5. 1. The function f
is C
-smooth.
2. For any > 0 there exists > 0 such that for all x C we have [f(x) f
(x) =
_
D(0)
f(x y)
(y)d
n
y =
_
D(x)
f(u)
(x +u)d
n
u =
_
V
f(u)
(x +u)d
n
u.
185
But the expression under the latter integral depends on x C
in
C
-smooth.
2. Fix some
0
> 0. The function f is uniformly continuous in U
0
(C). Hence there exists > 0
such that x, x
0
(C) and [[x x
(x) f(x)[ = [
_
D(0)
f(x y)
(y)d
n
y
_
D(0)
f(x)
(y)d
n
y[
_
D(0)
[f(x y) f(x)[
(y)d
n
y
_
D(0)
(y)d
n
y = . (11.2.2)
Proof of Theorem 11.3. Lemma 11.5 implies that for a suciently small the function g = f
is the required C
+ (f f
)
B,U
.
Theorem 11.3 implies a similar theorem form continuous maps C R
n
by applying it to all
coordinate functions.
11.3 Homotopy
Let A, B be any 2 subsets of vector spaces V and W, respectively. Two continuous maps f
0
, f
1
: A
B are called homotopic if there exists a continuous map F : A[0, 1] B such that F(x, 0) = f
0
(x)
and F(x, 1) = f
1
(x) for all t [0, 1]. Notice that the family f
t
: A B, t [0, 1], dened by the
formula f
t
(x) = F(x, t) is a continuous deformation connecting f
0
and f
1
. Conversely, any such
continuous deformation f
t
t[0,1]
provides a homotopy between f
0
and f
1
.
Given a subset C A, we say that a homotopy f
t
t[0,1]
is xed over C if f
t
(x) = f
0
(x) for
all x C and all t [0, 1].
A set A is called contractible if there exists a point a A and a homotopy f
t
: A A, t [0, 1],
such that f
1
= Id and f
0
is a constant map, i.e. f
1
(x) = x for all x A and f
0
(x) = a A for all
x A.
186
Example 11.6. Any star-shaped domain A in V is contractible. Indeed, assuming that it is star-
shaped with respect to the origin, the required homotopy f
t
: A A, t [0, 1], can be dened by the
formula f
t
(x) = tx, x A.
Remark 11.7. In what follows we will always assume all homotopies to be smooth. According to
Theorem 11.3 this is not a serious constraint. Indeed, any continuous map can be C
0
-approximated
by smooth ones, and any homotopy between smooth maps can be C
0
-approximated by a smooth
homotopy between the same maps.
Lemma 11.8. Let U V be an open set, A a compact oriented manifold (possibly with boundary)
and a smooth closed dierential k-form on U. Let f
0
, f
1
: A U be two maps which are homotopic
relative to the boundary A. Then
_
A
f
0
=
_
A
f
1
.
Proof. Let F : A [0, 1] U be the homotopy map between f
0
and f
1
. By assumption d = 0,
and hence
_
A[0,1]
F
d =
_
A[0,1]
dF
=
_
(A[0,1])
F
=
_
A[0,1]
F
+
_
A1
F
+
_
A0
F
t
on A 0, where we denote by t
the coordinate corresponding to the factor [0, 1]. First, we notice that F
[
A[0,1]
= 0 because the
map F is independent of the coordinate t, when restricted to A [0, 1]. Hence
_
A[0,1]
F
= 0.
Consider the inclusion maps A A[0, 1] dened by the formulas j
0
(x) = (x, 0) and j
1
(x) = (x, 1).
Note that j
0
, j
1
are dieomorphisms A A 0 and A A 1, respectively. Note that the map
j
1
preserves the orientation while j
0
reverses it. We also have F j
0
= f
0
, F j
1
= f
1
. Hence,
_
A1
F
=
_
A
f
1
and
_
A0
F
=
_
A
f
0
. Thus,
0 =
_
(A[0,1])
F
=
_
A[0,1]
F
+
_
A1
F
+
_
A0
F
=
_
A
f
1
_
A
f
0
.
187
Lemma 11.9. Let A be an oriented m-dimensional manifold, possibly with boundary. Let (A)
denote the space of dierential forms on A and (A[0, 1]) denote the space of dierential forms
on the product A[0, 1]. Let j
0
, j
1
: A A[0, 1] be the inclusion maps j
0
(x) = (x, 0) A[0, 1]
and j
1
(x) = (x, 1) A[0, 1]. Then there exists a linear map K : (A[0, 1]) (A) such that
If is a k-form, k = 1, . . . , m then K() is a (k 1)-form;
d K +K d = j
1
j
0
, i.e. for each dierential k-form
k
(A [0, 1] one has dK() +
K(d) = j
1
j
0
.
Remark 11.10. Note that the rst d in the above formula denotes the exterior dierential
k
(A)
k
(A), while the second one is the exterior dierential
k
(A[0, 1])
k
(A[0, 1]).
Proof. Let us write a point in A [0, 1] as (x, t), x A, t [0, 1]. To construct K() for a given
k
(A[0, 1] we rst contract with the vector eld
t
and then integrate the resultant form
with respect to the t-coordinate. More precisely, note that any k-form on A[0, 1] can be written
as = (t) +dt (t), t [0, 1], where for each t [0, 1]
(t)
k
(A), (t)
k1
(A).
Then
t
= (t) and we dene K() =
1
_
0
(t)dt.
If we choose a local coordinate system (u
1
, . . . , u
m
) on A then (t) can be written as (t) =
1i
1
<<i
m
h
i
1
...i
k
(t)du
i
1
du
i
k
, and hence
K() =
1
_
0
(t)dt =
1i
1
<<i
m
_
_
1
_
0
h
i
1
...i
k
(t)dt
_
_
du
i
1
du
i
k
.
Clearly, K is a linear operator
k
(AI)
k1
(A).
Note that if = (t) +dt (t)
k
(A[0, 1]) then
j
0
= (0), j
1
= (1).
We further have
K() =
1
_
0
(t)dt;
d = d
UI
= d
U
(t) +dt
(t) dt d
U
(t) = d
U
(t) +dt (
(t) d
U
(t)),
188
where we denoted
(t) :=
(t)
t
and I = [0, 1]. Here the notation d
UI
stands for exterior dierential
on (U I) and d
U
denotes the exterior dierential on (U). In other words, when we write d
U
(t)
we view (t) as a form on U depending on t as a parameter. We do not write any subscript for d
when there could not be any misunderstanding.
Hence,
K(d) =
1
_
0
(
(t) d
U
(t))dt = (1) (0)
1
_
0
d
U
(t)dt;
d(K()) =
1
_
0
d
U
(t)dt.
Therefore,
d(K()) +K(d()) = (1) (0)
1
_
0
d
U
(t)dt +
1
_
0
d
U
(t)dt
= (1) (0) = j
1
() j
0
().
Theorem 11.11. (Poincar es lemma) Let U be a contractible domain in V . Then any closed form
in U is exact. More precisely, let F : U [0, 1] U be the contraction homotopy to a point a U,
i.e. F(x, 1) = x, F(x, 0) = a for all x U. Then if a closed k-form in U then
= dK(F
),
where K :
k+1
(U [0, 1])
k
(U) is an operator constructed in Lemma 11.9.
Proof. Consider a contraction homotopy F : U [0, 1] U. Then F j
0
(x) = a U and
F j
1
(x) = x for all x U. Consider an operator K : U) (U) constructed above. Thus
K d +d K = j
1
j
0
.
Let be a closed k-form on U. Denote := F
= F
d = 0, j
1
= (F j
1
)
= and j
0
= (F j
0
)
1
j
0
= , (11.3.1)
189
i.e. = dK(F
).
In particular, any closed form is locally exact.
Example 11.12. Let us work out explicitly the formula for a primitive of a closed 1-form in a
star-shaped domain U R
n
. We can assume that U is star-shaped with respect to the origin.
Let =
n
1
f
i
dx
i
be a closed 1-form. Then according to Theorem 11.11 we have = dF, where
F = K(
=
n
1
f
i
(tx)d(tx
i
) =
n
1
tf
i
(tx)dx
i
+
n
1
x
i
f
i
(tx)dt.
Hence,
K() =
1
_
0
=
1
_
0
_
n
1
x
i
f
i
(tx)
_
dt.
Note that this expression coincides with the expression in formula (9.4.1) in Section 9.4.
Exercise 11.13. Work out an explicit expression for a primitive of a closed 2-form = Pdy dz +
Qdz dx +Rdx dy on a star-shaped domain U R
3
.
Example 11.14. 1. R
n
+
= x
1
0 R
n
is not dieomorphic to R
n
. Indeed, suppose there exists
such a dieomorphism f : R
n
+
R
n
. Denote a := f(0). Without loss of generality we can assume
that a = 0. Then
f = f[
R
n
+
\0
is a dieomorphism R
n
+
0 R
n
0. But R
n
+
0 is star-shaped with
respect to any point with positive coordinate x
1
, and hence it is contractible. In particular any
closed form on R
n
+
0 is exact. On the other hand, we exhibited above in 11.2 a closed (n1)-form
on R
n
0 which is not exact.
2. Borsuks theorem: There is no continuous map D
n
D
n
which is the identity on D
n
. We
denote here by D
n
the unit disc in R
n
and by D
n
its boundary (n 1)-sphere.
Proof. Suppose that there is such a map f : D
n
D
n
. One can assume that f is smooth.
Indeed, according to Theorem 11.3 one can approximate f by a smooth map, keeping it xed on
the boundary where it is the identity map, and hence smooth. Take the closed non-exact form
n
from Example 11.2 on D
n
. Then
n
= f
n
is a closed (n 1)-form on D
n
which coincides with
190
n
on D
n
. D
n
is star-shaped, and therefore
n
is exact,
n
= d. But then
n
= d([
D
n) which
is a contradiction.
3. Brouwers fixed point theorem: Any continuous map f : D
n
D
n
has at least 1 xed point.
Proof. Suppose f : D
n
D
n
has no xed points. Let us dene a map F : D
n
D
n
as follows.
For each x D
n
take a ray r
x
from the point f(x) which goes through x till it intersects D
n
at a
point which we will denote F(x). The map is well dened because for any x the points x and f(x)
are distinct. Note also that if x D
n
then the ray r
x
intersects D
n
at the point x, and hence
F(x) = x in this case. But existence of such F is ruled out by Borsuk theorem.
k-connected manifolds
A subset A V is called k-connected, k = 0, 1, . . . , if for any m k any two continuous maps
of discs f
0
, f
1
: D
m
A which coincide along D
m
are homotopic relative to D
m
. Thus, 0-
connectedness is equivalent to path-connectedness. 1-connected submanifolds are also called simply
connected.
Exercise 11.15. Prove that k-connectedness can be equivalently dened as follows: A is k-connected
if any map f : S
m
A, m k is homotopic to a constant map.
Example 11.16. 1. If A is contractible then it is k-connected for any k. For some classes of
subsets, e.g. submanifolds, the converse is also true (J.H.C Whiteheads theorem) but this is
a quite deep and non-trivial fact.
2. The n-sphere S
n
is (n1)-connected but not n-connected. Indeed, to prove that S
n1
simply
connected we will use the second denition. Consider a map f : S
k
S
n
. We rst notice
that according to Theorem 11.3 we can assume that the map f is smooth. Hence, according
to Corollary 9.20 Vol
n
f(S
k
) = 0 provided that k < n. In particular, there exists a point
p S
n
f(S
k
). But the complement of a point p in S
n
is dieomorphic to S
n
vis the
stereographic projection from the point p. But R
n
is contractible, and hence f is homotopic
to a constant map. On the other hand, the identity map Id : S
n
S
n
is not homotopic to
a constant map. Indeed, we know that there exists a closed n-form on S
n
, say the form
n
191
from Example 11.2, such that
_
S
n
n
,= 0. Hence,
_
S
n
Id
n
,= 0. On the other hand if Id were
homotopic to a constant map this integral would vanish.
Exercise 11.17. Prove that R
n+1
0 is (n 1) connected but not n-connected.
Proposition 11.18. Let U V be a m-connected domain. Then for any k m any closed
dierential k-form in U is exact.
Proof. We will prove here only the case m = 1. Though the general case is not dicult, it requires
developing certain additional tools. Let be a closed dierential 1-form. Choose a reference point
b U. By assumption, U is path-connected. Hence, any other point x can be connected to b by
a path
x
: [0, 1] U, i.e.
x
(0) = b,
x
(1) = x. Let us dene the function F : U R by the
formula F(x) =
_
x
. Note that due to the simply-connectedness of the domain U, any : [0, 1] U
connecting b and x is homotopic to
x
relative its ends, and hence according to Lemma 11.8 we
have
_
x
=
_
. Thus the above denition of the function F is independent of the choice of paths
x
. We claim that the function F is dierentiable and dF = . Note that if the primitive of exists
than it has to be equal to F up to an additive constant. But we know that in a suciently small
ball B
(a) centered at any point a U there exists a primitive G of . Hence, G(x) = F(x) +const,
and the the dierentiability of G implies dierentiablity of F and we have dF = dG = .
11.4 Winding and linking numbers
Given a loop : S
1
R
2
0 we dene its winding number around 0 as the integral
w() =
1
2
1
_
S
2
=
1
2
_
S
1
xdy ydx
x
2
+y
2
,
where we orient S
1
as the boundary of the unit disc in R
2
. For instance, if j : S
1
R
2
is the
inclusion map then w(j) = 1. For the loop
n
parameterized by the map t (cos nt, sinnt), t
[0, 2] we have w(
n
) = n.
Proposition 11.19. 1. For any loop the number w() is an integer.
2. If loops
0
,
1
: S
1
R
2
0 are homotopic in R
2
0 then w(
0
) = w(
1
).
192
Figure 11.1: w() = 2.
3. w() = n then the loop is homotopic (as a loop in R
2
0) to the loop
n
: [0, 1] R
2
0
given by the formula
n
(t) = (cos 2nt, sin2nt).
Proof. 1. Let us dene the loop parametrically in polar coordinates:
r = r(s), = (s), s [0, 1],
where r(0) = r(1) and (1) = (0) + 2n. The form
2
in polar coordinates is equal to d, and
hence
_
=
1
2
1
_
0
(s)ds =
(1) (0)
2
= n.
2. This is an immediate corollary of Proposition 11.8.
3. Let us write both loops and
n
in polar coordinates. Respectively,we have r = r(t), = (s)
for and r = 1, = 2ns for
n
, s [0, 1]. The condition w() = n implies, in view of part 1, that
(1) = (0) + 2n. Then the required homotopy
t
, t [0, 1], connecting the loops
0
= and
1
=
n
can be dened by the parametric equations r = (1t)r(s)+t, =
t
(s) = (1t)(s)+2nst.
Note that for all t [0, 1] we have
t
(1) =
t
(0) + 2n. Therefore,
t
is a loop for all t [0, 1].
193
Figure 11.2: l(
1
,
2
) = 1.
Given two disjoint loops , : S
1
R
3
(i.e. (s) ,= (t) for any s, t S
1
) consider a map
F
,
: T
2
R
3
0, where T
2
= S
1
S
1
is the 2-torus, dened by the formula
F
,
(s, t) = (s) (t).
Then the number
l(, ) :=
1
4
2
_
T
F
3
=
1
4
2
_
T
F
,
_
xdy dz +ydz dx +zdx dy
(x
2
+y
2
+z
2
)
3
2
_
is called the linking number of loops , .
1
Exercise 11.20. Prove that
1. The number l(, ) remains unchanged if one continuously deforms the loops , keeping
them disjoint;
2. The number l(, ) is an integer for any disjoint loops , ;
1
This denition of the linking number is due to Carl Friedrich Gauss.
194
3. l(, ) = l(, );
4. Let (s) = (cos s, sins, 0), s [0, 2] and (t) = (1 +
1
2
cos t, 0,
1
2
sint), t [0, 2]. Then
l(, ) = 1.
11.5 Properties of k-forms on k-dimensional manifolds
A k-form on k-dimensional submanifold is always closed. Indeed, d is a (k +1)-form and hence
it is identically 0 on a k-dimensional manifold.
Remark 11.21. Given a k-dimensional submanifold A V , and a k-form on V , the dierential
d
x
does not need to vanish at a point x A. However, d
x
[
Tx(A)
does vanish.
The following theorem is the main result of this section.
Theorem 11.22. Let A V be an orientable compact connected k-dimensional submanifold,
possibly with boundary, and a dierential k-form on A.
1. Suppose that A ,= . Then is exact, i.e. there exists a (k 1)-form on A such that
d = .
2. Suppose that A is closed, i.e. A = . Then is exact if and only if
_
A
= 0.
To prove Theorem 11.22 we will need a few lemmas.
Lemma 11.23. Let I
k
be the k-dimensional cube 1 x
j
1, j = 1, . . . , k.
1. Let be a dierential k-form on I
k
such that
Supp()
_
0 I
k1
[0, 1] I
k1
_
= .
Then there exists a (k1)-form such that d = and such that Supp()
_
1 I
k1
[1, 1] I
k1
_
=
.
2. Let be a dierential k-form on I
k
such that Supp() Int I
k
and
_
I
k
= 0. Then there exists
a (k 1)-form such that d = and Supp() Int I
k
.
195
Proof. We have
= f(x
1
, . . . , x
k
)dx
1
dx
k
,
In the rst case of the lemma the function f vanishes on 0 I
k1
[1, 1] I
k1
. We will look
for in the form
= g(x
1
, . . . , x
k
)dx
2
dx
k
.
Then
d =
g
x
1
(x
1
, . . . , x
k
)dx
1
dx
2
dx
k
.
and hence the equation d = is equivalent to
g
x
1
(x
1
, . . . , x
k
) = f(x
1
, . . . , x
k
).
Hence, if we dene
g(x
1
, . . . , x
k
) :=
x
1
_
1
f(u, x
2
, . . . , x
k
)du,
then the form = g(x
1
, . . . , x
k
)dx
2
dx
k
has the required properties.
The second part of the lemma we will prove here only for the case k = 2. The general case can
be handled similarly by induction over k. We have in this case Supp(f) Int I
2
and
_
I
2
fdS = 0.
Let us denote h(x
2
) :=
1
_
1
f(x
1
, x
2
)dx
1
. Note that h(u) = 0 if u is suciently close to 1 or 1.
According to Fubinis theorem,
1
_
1
h(x
2
)dx
2
= 0. We can assume that f(x
1
, x
2
) = 0 for x
1
1 ,
and hence
u
_
1
f(x
1
, . . . , x
k1
, t)dt = h(x
1
, . . . , x
k1
) for u [1 , 1]. Consider any non-negative
C
3
, 1].
Dene a function g
1
: I
2
R by the formula
g
1
(x
1
x
2
) =
_
_
x
1
_
1
f(u, x
2
)du, x
1
[1, 1 ],
h(x
2
)(x
1
), x
1
(1 , 1].
Denote
1
= g
1
(x
1
, x
2
)dx
2
. Then d = on [1, 1 ] [0, 1] and d
1
= h(x
2
)
(x
1
)dx
1
dx
2
on
[1 , 1] [0, 1]. Note that Supp(
1
) Int I
2
.
196
Let us dene
g
2
(x
1
, x
2
) :=
_
_
0, x
1
[1, 1 ],
(x
1
)
x
2
_
1
h(u)du, x
1
(1 , 1]
and denote
2
= g
2
(x
1
, x
2
)dx
1
. Then d
2
= 0 on [1, 1 ] [1, 1] and
d
2
= h(x
2
)
(x
1
)dx
1
dx
2
on [1, 1][1, 1]. Note that g
2
(x
1
, 1) =
(x
1
)
1
_
1
h(u)du = 0. Taking into account that h(u) = 0
when u is suciently close to 1 or 1 we conclude that h(x
1
, x
2
) = 0 near I
2
, i.e. Supp(
2
) Int I
2
.
Finally, if we dene =
1
+
2
then we have d = and Supp() Int I
2
.
The following lemma is a special case of the, so-called, tubular neighborhood theorem.
Lemma 11.24. Let A V be a compact k-dimensional submanifold with boundary. Let :
[1, 1] A be an embedding such that (1) A,
(1) T
(1)
(A) and ([0, 1)) Int A.
Then the embedding can be extended to an embedding : [1, 1] I
k1
A such that
(t, 0) = (t), for t [1, 1], 0 I
k1
;
(1 I
k1
) A, ([1, 1) I
k1
) Int A;
t
(1, x) / T(A) for all x I
k1
.
There are many ways to prove this lemma. We will explain below one of the arguments.
Proof. Step 1. We rst construct k 1 ortonormal vector elds
1
, . . . ,
k
along = ([1, 1])
which are tangent to A and normal to . To do that let us denote by N
u
the normal (k 1)-
dimensional space N
u
to T
u
in T
u
A. Let us observe that in view of compactness of there is an
> 0 with the following property: for any two points u = (t), u
= (t
) , t, t
1
= ([1, t
1
]), project them orthogonally to the normal spaces N
u
in these points, and then
orthonormalize the resulted bases via the Gram-Schmidt process. Thus we constructed orthonormal
vector elds
1
(t), . . .
k
(t) N
(t)
, t [1, t
1
]. Now we repeat this procedure beginning with the
197
basis
1
(t
1
), . . .
k
(t
1
) N
(t
1
)
= N
u
1
and extend the vector elds
1
, . . . ,
k
to
2
= ([t
1
, t
2
]).
Continuing this process we will construct the orhonormal vector elds
1
, . . . ,
k
along the whole
curve .
2
Step 2. Consider a map : [1, 1] I
k1
V given by the formula
(t, x
1
, . . . , x
k1
) = (t) +
k1
1
x
j
j
(t), t, x
1
, . . . , x
k1
[1, 1],
where a small positive number will be chosen later. The map is an embedding if is chosen
small enough.
3
Unfortunately the image ([1, 1] I
k1
) is not contained in A. We will correct
this in the next step.
Step 3. Take any point a A and denote by
a
the orthogonal projection V T
a
A. Let us make
the following additional assumption (in the next step we will show how to get rid of it): there exists
a neighborhood U a = (1) in A such that
a
(U) N
a
T
a
A. Given > 0 let us denote by
B
T
a
A. In view of compactness of A
one can choose an > 0 such that for all points a there exists an embedding e
a
: B
(a) A
such that
a
e
a
= Id. Then for a suciently small <
k1
the map
: [1, 1] I
k1
A
dened by the formula
(t, x) = e
(t)
(t, x), t [1, 1], x I
k1
is an embedding with the required properties.
Step 4 It remains to show how to satisfy the additional condition at the boundary point (1)
A which were imposed above in Step 3. Take the point a = (1) A. Without loss of
generality we can assume that a = 0 V . Choose an orthonormal basis v
1
, . . . , v
n
of V such that
v
1
, . . . , v
k
N
a
and v
k
is tangent to and pointing inward . Let (y
1
, . . . , y
n
) be the corresponding
cartesian coordinates in V . Then there exists a neighborhood U a in A which is graphical in
these coordinates and can be given by
y
j
=
j
(y
1
, . . . , y
k
), j = k + 1, . . . , n, y
k
k
(y
1
, . . . , y
k1
), [y
i
[ , i = 1, . . . , k 1,
2
Strictly speaking, the constructed vector elds only piece-wise smooth, because we did not make any special
precautions to ensure smoothness at the points uj, j = 1, . . . , N1. This could be corrected via a standard smoothing
procedure.
3
Exercise: prove it!
198
where all the rst partial derivatives of the functions
k
, . . . ,
n
vanish at the origin. Take a C
_
, y
k+1
, . . . , y
n
).
For a suciently small > 0 this is a dieomorphism supported in an -ball in V centered in the
origin. On the other hand, the manifold
A = F(A) satises the extra condition of Step 3.
Lemma 11.25. Let A V be a (path)-connected submanifold with a non-empty boundary. Then
for any point a A there exists an embedding
a
: [1, 1] A such that
a
(0) = a,
a
(1) A
and
a
(1) T
a(1)
(A).
Sketch of the proof. Because A is path-connected with non-empty boundary, any interior point
can be connected by a path with a boundary point. However, this path need not be an embedding.
First, we perturb this path to make it an immersion : [1, 1] A, i.e. a map with non-vanising
derivative. This can be done as follows. As in the proof of the previous lemma we consider a su-
ciently small partition of the path, so that two neighboring subdivision points lie in a coordinate
neighborhood. Then we can connect these points by a straight segment in these coordinate neigh-
borhoods. Finally we can smooth the corners via the standard smoothing procedure. Unfortunately
the constructed immersed path may have self-intersection points. First, one can arrange that
there are only nitely many intersections, and then cut-out the loops, i.e. if (t
1
) = (t
2
) for
t
1
< t
2
we can consider a new piece-wise smooth path which consists of [
[1,t
1
]
and [
[t
2
,1]
The
new path has less self-intersection points, and thus continuing by induction we will end with a
piece-wise smooth embedding. It remains to smooth again the corners.
Proof of Theorem 11.22. 1. For every point a A choose an embedding
a
: [1, 1] A, as in
Lemma 11.25, and using Lemma 11.24 extend
a
to an embedding
a
: [1, 1] I
k1
A such
that
-
a
(t, 0) = (t), for t [1, 1], 0 I
k1
;
-
a
(1 I
k1
) A,
a
([1, 1) I
k1
) Int A;
-
a
t
(1, x) / T(A) for all x I
k1
.
199
Due to compactness of A we can choose nitely many such embeddings
j
=
a
j
, j = 1, . . . , N,
such that
N
j
((1, 1] Int(I
k1
) = A. Choose a partition of unity subordinated to this covering
and split the k-form as a sum =
K
j
, where each
i
is supported in
j
((1, 1] Int(I
k1
))
for some j = 1, . . . , N. To simplify the notation we will assume that N = K and each
j
is
supported in
j
((1, 1] Int(I
k1
)), j = 1, . . . , N. Consider the pull-back form
j
=
j
on
I
k
= [1, 1]I
k1
. According to Lemma 11.23.1 there exists a (k1)-form
j
such that Supp(
j
)
(1, 1] Int(I
k1
) and d
j
=
j
. Let us transport the form
j
back to A. Namely, set
j
equal to
(
1
j
)
j
on
j
((1, 1] Int(I
k1
)) A and extend it as 0 elsewhere on A. Then d
j
=
j
, and
hence d(
N
j
) =
N
j
= .
2. Choose a point a A and parameterize a coordinate neighborhood U A by an embedding
: I
k
A such that (0) = a. Take a small closed ball D
(0) I
k
R
k
and denote
D = (D
(0)).
Then
A = A Int
D is a submanifold with non-empty boundary, and
A =
= [
e
A
. Let us extent the form
in
any way to a form, still denoted by
on the whole submanifold A. Then d
= + where
Supp()
D Int (I
k
). Note that
_
(I
k
)
=
_
A
=
_
A
_
A
d
= 0
because
_
A
= 0 by our assumption, and
_
A
d
= 0, and
hence, we can apply Lemma 11.23.2 to the form
on I
k
and construct a (k 1)-form on
I
k1
such that d =
on (I
k
) and equal to 0 elsewhere. Finally, we have
d(
) = d
+ = , and hence =
+
=
_
A
.
200
Proof. According to Theorem 11.22.1 there exists a form such that = d. Then
_
A
f
=
_
A
f
d =
_
A
df
=
_
A
f
=
_
A
=
_
A
.
Degree of a map
Consider two closed connected oriented submanifolds A V , B W of the same dimension k. Let
be an n-form on B such that
_
B
= 1. Given a smooth map f : A B the integer deg(f) :=
_
A
f
=
_
A
f
,
for any smooth map f : A B, and thus deg(f) is independent of the choice of the form
on B with the property
_
A
= 1.
2. If the maps f, g : A B are homotopic then deg(f) = deg(g).
3. Let b B be a regular value of the map f. Let f
1
(b) = a
1
, . . . , a
d
. Then
deg(f) =
d
1
sign(det Df(a
j
)).
In particular, deg(f) is an integer number.
Proof. The second part follows from Lemma 11.8. To prove the rst part, let us write = +,
where
_
B
= 0. Using Theorem 11.22.2 we conclude that = d for some (k 1)-form on B.
Then
_
A
f
=
_
A
f
+
_
A
f
=
_
A
f
+
_
A
df
=
_
A
f
.
Let us prove the last statement of the theorem. By the inverse function theorem there exists a
neighborhood U b in B and neighborhoods U
1
a
1
, . . . , U
d
a
d
in A such that the restrictions
of the map f to the neighborhoods U
1
, . . . , U
d
are dieomorphisms f[
U
j
: U
j
U, j = 1, . . . , d. Let
us consider a form on B such that Supp U and
_
B
=
_
U
= 1. Then
deg(f) =
_
A
f
=
d
1
_
U
j
f
=
d
1
sign(det Df(a
j
)),
201
because according to Theorem ?? we have
_
U
j
f
= sign(det Df(a
j
))
_
U
= sign(det Df(a
j
)).
for each j = 1, . . . , d.
Remark 11.28. Any continuous map f : A B can be approximated by a homotopic to f smooth
map A B, and any two such smooth approximations of f are homotopic. Hence this allows us
to dene the degree of any continuous map f : A B.
Exercise 11.29. 1. Let us view R
2
as C. In particular, we view the unit sphere S
1
= S
1
1
(0) as the
set of complex numbers of modulus 1:
S
1
= z C; [z[ = 1.
Consider a map h
n
: S
1
S
1
given by the formula h
n
(z) = z
n
, z S
1
. Then deg(h
n
) = n.
2. Let f : S
n1
S
n1
be a map of degree d. Let p
1
x
2
j
(x
1
, . . . , x
n
) S
n1
and dene a map f : S
n
S
n
by the formula
f(x) =
_
_
p
, if x = p
,
_
n
1
x
2
j
f((x)), x
n+1
_
, if x ,= p
.
Prove that deg((f)) = d.
4
3. Prove that two maps f, g : S
n
S
n
are homotopic if and only if they have the same degree.
In particular, any map of degree n is homotopic to the map h
n
. (Hint: For n=1 this follows from
Proposition 11.19. For n > 1 rst prove that any map is homotopic to a suspension.)
4. Give an example of two non-homotopic orientation preserving dieomorphisms T
2
T
2
. Note
that the degree of both these maps is 1. Hence, for manifolds, other than spheres, having the same
degree is not sucient for their homotopy.
4
The map f is called the suspension of the map f.
202
5. Let , : S
1
R
3
be two disjoint loops in R
3
. Consider a map
F
,
: T
2
S
2
dened by the
formula
F
,
(s, t) =
(s) (t)
[[(s) (t)[[
, s, t S
1
.
Prove that l(, ) = deg(
F
,
). Use this to solve Exercise 11.20.4 above.
203