01 Csikos Differential Geometry
01 Csikos Differential Geometry
01 Csikos Differential Geometry
Algoritmuselmelet
Algoritmusok bonyolultsaga
Analitikus modszerek a penzugyekben
Bevezetes az analzisbe
Differential Geometry
Diszkret optimalizalas
Diszkret matematikai feladatok
Geometria
Igazsagos elosztasok
Interaktv analzis feladatgyujtemeny matematika BSc hallgatok szamara
Introductory Course in Analysis
Matematikai penzugy
Mathematical Analysis-Exercises 1-2
Mertekelmelet es dinamikus programozas
Numerikus funkcionalanalzis
Operaciokutatas
Operaciokutatasi peldatar
Optimalis iranytasok
Parcialis differencialegyenletek
Peldatar az analzishez
Szimmetrikus kombinatorikai strukturak
Tobbvaltozos adatelemzes
Balazs Csikos
DIFFERENTIAL
GEOMETRY
Typotex
2014
20142019, Balazs Csikos,
Eotvos Lorand University, Faculty of Science
1 Preliminaries 1
1.1 Categories and Functors . . . . . . . . . . . . . . . . . . . . . 1
1.2 Linear Algebra . . . . . . . . . . . . . . . . . . . . . . . . . . 3
1.2.1 Linear Spaces and Linear Maps . . . . . . . . . . . . . 3
1.2.2 Determinant of Matrices and Linear Endomorphisms . 7
1.2.3 Orientation of a Linear Space . . . . . . . . . . . . . . 11
1.2.4 Tensor Product . . . . . . . . . . . . . . . . . . . . . . 11
1.2.5 Exterior Powers . . . . . . . . . . . . . . . . . . . . . . 15
1.2.6 Euclidean Linear Spaces . . . . . . . . . . . . . . . . . 21
1.2.7 Hodge Star Operator . . . . . . . . . . . . . . . . . . . 30
1.3 Geometry . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33
1.3.1 Affine Geometry . . . . . . . . . . . . . . . . . . . . . 33
1.3.2 Euclidean Spaces . . . . . . . . . . . . . . . . . . . . . 40
1.4 Topology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 49
1.4.1 Separation and Countability Axioms . . . . . . . . . . 56
1.4.2 Compactness . . . . . . . . . . . . . . . . . . . . . . . 58
1.4.3 Fundamental Group and Covering Spaces . . . . . . . 60
1.5 Multivariable Calculus . . . . . . . . . . . . . . . . . . . . . . 62
1.6 Measure and Integration . . . . . . . . . . . . . . . . . . . . . 67
1.7 Ordinary Differential Equations . . . . . . . . . . . . . . . . . 75
2 Curves in En 81
2.1 The Notion of a Curve . . . . . . . . . . . . . . . . . . . . . . 81
2.2 The Length of a Curve . . . . . . . . . . . . . . . . . . . . . . 83
2.3 Croftons Formula . . . . . . . . . . . . . . . . . . . . . . . . 86
2.4 The Osculating k-planes . . . . . . . . . . . . . . . . . . . . . 92
2.5 Frenet Frames and Curvatures . . . . . . . . . . . . . . . . . 97
2.6 Limits of Some Geometrical Quantities . . . . . . . . . . . . . 107
2.7 Osculating Spheres . . . . . . . . . . . . . . . . . . . . . . . . 112
2.8 Plane Curves . . . . . . . . . . . . . . . . . . . . . . . . . . . 117
i
2.8.1 Evolute, Involute, Parallel Curves . . . . . . . . . . . 118
2.8.2 The Rotation Number Theorem . . . . . . . . . . . . 124
2.8.3 Convex Curves . . . . . . . . . . . . . . . . . . . . . . 127
2.8.4 The Four Vertex Theorem . . . . . . . . . . . . . . . . 129
2.9 Curves in R3 . . . . . . . . . . . . . . . . . . . . . . . . . . . 131
2.9.1 Orthogonal Projections onto Planes Spanned by Frenet
Vectors . . . . . . . . . . . . . . . . . . . . . . . . . . 133
2.9.2 Fenchels Theorem . . . . . . . . . . . . . . . . . . . . 135
2.9.3 The Fary-Milnor Theorem . . . . . . . . . . . . . . . . 139
3 Hypersurfaces in Rn 141
3.1 General Theory . . . . . . . . . . . . . . . . . . . . . . . . . . 141
3.1.1 Definition of a Parameterized Hypersurface . . . . . . 141
3.1.2 Curvature of Curves on a Hypersurface . . . . . . . . 143
3.1.3 The Weingarten Map and the Fundamental Forms . . 145
3.1.4 Umbilical Points . . . . . . . . . . . . . . . . . . . . . 151
3.1.5 The Fundamental Equations of Hypersurface Theory 152
3.1.6 Surface Volume . . . . . . . . . . . . . . . . . . . . . . 161
3.2 Surfaces in R3 . . . . . . . . . . . . . . . . . . . . . . . . . . . 162
3.2.1 Surfaces of Revolution . . . . . . . . . . . . . . . . . . 162
3.2.2 Lines of Curvature, Triply Orthogonal Systems . . . . 166
3.2.3 Ruled and Developable Surfaces . . . . . . . . . . . . 173
3.2.4 Asymptotic Curves on Negatively Curved Surfaces . . 179
3.2.5 Surfaces of Constant Negative Curvature . . . . . . . 183
3.2.6 Minimal Surfaces . . . . . . . . . . . . . . . . . . . . . 195
4 Manifolds 201
4.1 Topological and Differentiable Manifolds . . . . . . . . . . . . 201
4.1.1 Basic Definitions . . . . . . . . . . . . . . . . . . . . . 201
4.1.2 Configuration Spaces . . . . . . . . . . . . . . . . . . . 205
4.1.3 Submanifolds of Rn . . . . . . . . . . . . . . . . . . . 206
4.1.4 Remarks on the Classification of Manifolds . . . . . . 209
4.2 The Tangent Bundle . . . . . . . . . . . . . . . . . . . . . . . 212
4.2.1 The Tangent Bundle . . . . . . . . . . . . . . . . . . . 217
4.2.2 The Derivative of a Smooth Map . . . . . . . . . . . . 218
4.3 The Lie Algebra of Vector Fields . . . . . . . . . . . . . . . . 219
4.3.1 The Lie Algebra of Lie Groups . . . . . . . . . . . . . 227
4.3.2 Subspace Distributions and the Frobenius Theorem . 229
4.4 Tensor Bundles and Tensor Fields . . . . . . . . . . . . . . . 233
4.5 The Lie Derivative . . . . . . . . . . . . . . . . . . . . . . . . 238
4.6 Differential Forms . . . . . . . . . . . . . . . . . . . . . . . . 243
4.6.1 Interior Product by a Vector Field . . . . . . . . . . . 243
ii
4.6.2 Exterior Differentiation . . . . . . . . . . . . . . . . . 245
4.6.3 De Rham Cohomology . . . . . . . . . . . . . . . . . . 251
4.7 Integration of Differential Forms . . . . . . . . . . . . . . . . 254
4.7.1 Integration on Chains . . . . . . . . . . . . . . . . . . 255
4.7.2 Integration on Regular Domains . . . . . . . . . . . . 260
4.7.3 Integration on Riemannian Manifolds . . . . . . . . . 265
4.8 Differentiation of Vector Fields . . . . . . . . . . . . . . . . . 266
4.9 Curvature . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 275
4.10 Decomposition of Algebraic Curvature Tensors . . . . . . . . 290
4.11 Conformal Invariance of the Weyl Tensor . . . . . . . . . . . 298
4.12 Geodesics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 304
4.13 Applications to Hypersurface Theory . . . . . . . . . . . . . . 311
4.13.1 Geodesic Curves on Hypersurfaces . . . . . . . . . . . 311
4.13.2 Clairauts Theorem . . . . . . . . . . . . . . . . . . . . 315
4.13.3 Moving Orthonormal Frames Along a Hypersurface . . 317
4.13.4 Relation to Earlier Formulae for Parameterized Hyper-
surfaces . . . . . . . . . . . . . . . . . . . . . . . . . . 318
4.13.5 The GaussBonnet Formula . . . . . . . . . . . . . . . 320
4.13.6 Steiners Formula . . . . . . . . . . . . . . . . . . . . . 321
4.13.7 Minkowskis Formula . . . . . . . . . . . . . . . . . . . 325
4.13.8 Rigidity of Convex Surfaces . . . . . . . . . . . . . . . 330
Bibliography 336
Index 338
iv
Chapter 1
Preliminaries
In this chapter, we collect some definitions and facts that will be used later
in the text.
(f g) h = f (g h).
1
2 1. Preliminaries
Definition 1.1.6. Let F and G be two covariant functors from the category
C to the category D. Then a natural transformation from F to G assigns
to every X ObC a morphism X MorD (F (X), G(X)) so that for any
morphism f MorC (X, Y ), the diagram
F (f )
F (X) F (Y )
X y
y Y
G(f )
G(X) G(Y )
The definition of natural transformations can be extended also for the case,
when one of the functors or both are contravariant. Then the contravariant
functors should be substituted by their covariant counterparts going into the
opposite categories. Natural transformations and natural isomorphisms can
be defined by obvious modifications also for multivariable functors.
There is a principle in mathematics, that if there is an isomorphism between
two objects of a category, then the two objects can be identified, as they
behave in the same way and look like the same way from the view point of
the category.
The importance of naturality is that if there is a natural isomorphism between
two functors, then they can be identified with one another, or considered to
be essentially the same.
(i) (x + y) + z = x + (y + z) (associativity);
(iv) x + y = y + x (commutativity);
(v) (x + y) = x + y;
(vi) ( + )x = x + x;
(viii) 1x = x.
It is known that any two bases of a linear space have the same cardinality.
Definition 1.2.3. The dimension dim V of the linear space V is the cardi-
nality of a basis.
l11 . . . ln1
.. .. .
[L] = . . .. ,
l1m . . . lnm
we obtain the matrix of L with respect to the bases (e1 , . . . , en ) and (f1 , . . . , fm ).
If we arrange the coordinates of x with respect to the basis (e1 , . . . , en ) into
a column vector [x], then the column vector [L(x)] of the coordinates of L(x)
with respect to the basis (f1 , . . . , fm ) can be computed by the matrix mul-
tiplication [L(x)] = [L][x]. For an endomorphism of V , we usually use the
same basis for V and W = V .
Examples.
xy = (x1 y1 , . . . , xn yn ),
x = (x1 , . . . , xn ).
6 1. Preliminaries
is the Kronecker delta symbol, form a basis of V . This basis is called the
dual basis of the basis (e1 , . . . , en ).
We remark that the Kronecker delta symbol ij is also denoted by ij
and ij . In formulae involving the Kronecker delta symbol, we always
position the indices as dictated by the Einstein convention.
Though for a finite dimensional linear space V , the dual space V has the
same dimension as V , there is no natural isomorphism between these two
spaces. In other words, the identical functor on the category of finite dimen-
sional linear spaces is not naturally isomorphic to the dual space functor. On
the other hand, there is a natural transformation from the identical functor
to the double dual space functor, given by the embeddings V : V V ,
((v))(l) = l(v), where v V , l V . The restriction of this natural trans-
formation onto the category of finite dimensional linear spaces is a natural
isomorphism.
With the help of this natural isomorphism, elements of a finite dimensional
linear space V can be identified with elements of V . With this identification,
the dual basis of the dual basis of a basis of V will be equal to the original
basis.
Proposition 1.2.8. (1) If all but one columns of a square matrix are
fixed, then the determinant is a linear function of the varying column. This
means that if we denote by (a1 , . . . , an ) the square matrix with column vectors
a1 , . . . , an Rn , then
det(a1 , . . . , aj + aj , . . . , an ) =
det(a1 , . . . , aj , . . . , an ) + det(a1 , . . . , aj , . . . , an ).
(3) The value of the determinant does not change if an arbitrary multiple of
a column is added to another column.
(4) The determinant of a matrix vanishes if and only if its columns are lin-
early dependent.
(5) A square matrix A and its transposition A> , that is the reflection of A in
the main diagonal, have the same determinant
det([L]e ) = det([L]f ).
and
n X
X n n
X Xn n X
X n
lj sk ek = lj S(ej ) = L(fi ) = L sj
e = sji ljk ek .
i j i i j
j=1 k=1 j=1 j=1 j=1 k=1
which means that [L]f [S]e = [S]e [L]e . Taking the determinant of both sides
we obtain
det([L]f ) det([S]e ) = det([S]e ) det([L]e ).
Since the columns of [S]e are linearly independent, as they are the coordinate
vectors of the basis vectors fi with respect to the basis (e1 , . . . , en ), the deter-
minant of [S]e is nonzero. Thus, equation (1.3) implies the proposition.
Definition 1.2.11. The determinant of a linear endomorphism L : V V
is the determinant of the matrix of L with respect to an arbitrary basis of V .
The definition is correct according to the previous proposition.
10 1. Preliminaries
pA () = (1 ) (n ).
det A = pA (0) = 1 n .
which is also equal to the sum of the eigenvalues of A. The trace of a linear
endomorphism is the trace of its matrix with respect to an arbitrary basis.
Pn
If T = i1 ,...,ik ,j1 ,...,jl =1 Tij11...i
...jl i1 ...ik
e
k j1 ...jl
is a tensor of type (k, l), then the num-
j1 ...jl
bers Ti1 ...ik are called the coordinates or components of the tensor T with
respect to the basis e1 , . . . , en .
Exercise 1.2.24. The coordinates of a type (0, 2) or a type (1, 1) tensor can
always be arranged into an n n matrix. Do the trace and determinant of
this matrix depend on the choice of the basis?
(a) (V W )
= V W ;
(b) Hom(V, W )
= V W , in particular, End(V )
= T (1,1) V and
(2,0)
Hom(V, V ) = T V;
(c) (T (k,l) V )
= T (k,l) V
= T (l,k) V ;
Then
c11 ck1
v1 vk = det ... .. .. w w .
. . 1 k
c1k ckk
The next statement is a corollary of the previous two ones.
Proposition 1.2.32. The vectors v1 , . . . , vk are linearly independent if and
only if v1 vk 6= 0. Two linearly independent k-tuples of vectors
v1 , . . . , vk and w1 , . . . , wk span the same k-dimensional linear subspace if
and only if v1 vk = c w1 wk for some c R \ {0}. In addition,
these two collections are bases of the same orientation in the linear space they
span if and only if c > 0.
Corollary 1.2.33. The Grassmann manifold Grk (V ) can be embedded into
the projective space P(k V ) by assigning to the k-dimensional subspace span-
ned by the linearly independent vectors v1 , . . . , vk the 1-dimensional linear
space spanned by v1 vk . This embedding is called the Plucker embed-
ding.
Corollary 1.2.34. There is a natural one-to-one correspondence between
orientations of a k-dimensional linear subspace W of an n-dimensional lin-
ear space V and orientations of the 1-dimensional linear subspace of k (V )
assigned to W by the Plucker embedding.
Ldim V
The direct sum (V ) = k=0 k (V ) of all the exterior powers of V be-
defined uniquely by the
comes an associative algebra with multiplication
rule
(l (vk+1 , . . . , vk+l )) = k+l (v1 , . . . , vk+l )
(k (v1 , . . . , vk ))
k, l N; v1 , . . . , vk+l V.
As we have
...
k (v1 , . . . , vk ) = v1 vk
1.2. Linear Algebra 17
T (0,k) (V ) = Ak (V ) Wk ,
k k : V k Ak (V ), (v1 , . . . , vk ) 7 k (v1 vk )
k (v1 vk ) 7 k (v1 vk ).
Since
X
k (k (v1 vk ) = sgn v(1) v(k) = k! v1 vk ,
Sk
k1 = k /k!.
It is a matter of taste which isomorphism is used to identify wedge products
of vectors with alternating tensors, but to avoid confusion, we should choose
one of the identifications and then insist on that for the rest of the discussion.
Making this decision, from now on we use k to identify k (V ) and Ak (V ).
Nevertheless, one should be aware of the fact that some authors prefer the
identification by k .
1.2. Linear Algebra 19
Ak (V1 V2 )
M
= Ar (V1 ) As (V2 ).
r+s=k
If Sk is a permutation, then
n
X
((k )1 ( )) = (ei1 , . . . , eik )e(i1 )...(ik )
i1 ,...,ik =1
X n
= (e1 (i1 ) , . . . , e1 (ik ) )ei1 ...ik ,
i1 ,...,ik =1
h(x1 , . . . , xn ), (y1 , . . . , yn )i = x1 y1 + + xn yn
hv, wi
1 1.
kvk kwk
hv, wi
cos = .
kvk kwk
The number is called the angle enclosed by the vectors v and w (measured
in radians).
Definition 1.2.53. Two non-zero vectors are orthogonal if the angle enclosed
by them is /2. Orthogonality of two non-zero vectors is equivalent to the
condition hv, wi = 0. Since the latter equation is automatically fulfilled when
v = 0, we agree, that 0 is said to be orthogonal to every vector.
Definition 1.2.54. A collection of some vectors e1 , . . . , ek of a Euclidean
linear space V is said to be an orthonormal system if hei , ej i = ij for all
1 i, j k, i.e., if the vectors have unit length and are mutually orthogonal
to one another.
1.2. Linear Algebra 23
consequently,
ek = kk fk (hfk , e1 ie1 + + hfk , ek1 iek1 ) .
The parameter kk must be used to normalize the vector which stands on the
right of it. Thus,
kk = kfk (hfk , e1 ie1 + + hfk , ek1 iek1 )k1 .
Since kk > 0, the only possible choice of ek is
fk (hfk , e1 ie1 + + hfk , ek1 iek1 )
ek = .
kfk (hfk , e1 ie1 + + hfk , ek1 iek1 )k
It is not difficult to check that the vectors e1 , . . . , ek will satisfy the require-
ments.
Applying the Gram-Schmidt orthogonalization to a basis we obtain the fol-
lowing corollary.
Corollary 1.2.56. Every finite dimensional Euclidean linear space has an
orthonormal basis.
Definition 1.2.57. The Gram matrix of a system of vectors f1 , . . . , fk of a
Euclidean linear space V is the matrix
hf1 , f1 i . . . hf1 , fk i
G(f1 , . . . , fk ) = ... .. .. .
. .
hfk , f1 i . . . hfk , fk i
Corollary 1.2.58. Let G be the Gram matrix of the vectors f1 , . . . , fk of a
Euclidean linear space V . Then det G 0, and det G = 0 if and only if the
vectors f1 , . . . , fk are linearly dependent.
Pk
Proof. If there is a non-trivial linear relation i=1 i fi = 0, then the rows
of G are linearly dependent with the same coefficients i therefore det G = 0.
If f1 , . . . , fk are linearly independent, then applying the GramSchmidt or-
thogonalization to them we obtain an orthonormal system e1 , . . . , ek and we
can express the vectors fs as follows
f1 = 11 e1
..
.
fs = s1 e1 + + ss es
..
.
fk = k1 e1 + k2 e2 + + kk ek .
1.2. Linear Algebra 25
The lower triangular matrix T put together from the coefficients ij is the
inverse of the matrix (1.5) coming from the decompositions (1.4), in particular
s1
1
X
ss = = hf i
s i
> 0 for s = 1, . . . , k.
, e (1.7)
s
ss
f
i=1
It is clear that G = T T T , so
det G = det(T T T ) = det(T )2 = (11 kk )2 > 0.
Corollary 1.2.59. With the notation used in Corollary 1.2.58, the identity
s1
X
2
det G(f1 , . . . , fs ) =
fs hfs , ei iei
det G(f1 , . . . , fs1 )
i=1
holds.
Corollary 1.2.60. A symmetric bilinear function h , i on a linear space V is
positive definite if and only if there is a basis f1 , . . . , fn such that the Gram
matrices G(f1 , . . . , fs ) have positive determinants for s = 1, . . . , n.
Proof. The previous proposition shows that if h , i is positive definite, then
any basis will be good. Conversely, assume that f1 , . . . , fn is a basis such that
G(f1 , . . . , fs ) > 0 for s = 1, . . . , n. Then the recursive formula
Ps1
f1 fs i=1 hfs , ei iei
e1 = p , es = p
det G(f1 ) det G(f1 , . . . , fs )/ det G(f1 , . . . , fs1 )
defines an orthonormal basis for h , i, i.e. hei , ej i = ij . (Check this!) The
existence
Pnof an orthonormal basis Pimplies that h , i is positive definite, because
n
if v = i=1 v i ei , then hv, vi = i=1 (v i )2 0, and equality holds only when
v = 0.
1
{x, y} = Q{,} (x + y) Q{,} (x) Q{,} (y) . (1.8)
2
The following proposition establishes a bijection between bilinear functions
on a Euclidean linear space V and linear endomorphisms L : V V .
is a bilinear function on V .
(ii) For any bilinear function { , } on V , there is a unique linear map
L : V V such that { , } = { , }L .
(iii) The bilinear function { , }L is symmetric if and only if L satisfies the
identity hLx, yi = hx, Lyi.
Proof. We prove only (ii), the rest is obvious. Choose an orthonormal basis
(e1 , . . . , en ) in V . Given a bilinear function { , }, the only possible choice for
L(x) is
Xn n
X
L(x) = hLx, ei iei = {x, ei }ei ,
i=1 i=1
so L is a good choice.
Proof. It is enough to show the second part of the proposition since orthonor-
mal basis exists only for symmetric positive definite bilinear functions. This
follows from
hei1 ...ik , ej1 ...jk i = L(ei1 eik ej1 ejk )
= hei1 , ej1 i heik , ejk i = i1 ,j1 ik ,jk .
When the sequences are equal, then the permuted sequences i(1) , . . . , i(k)
and j(1) , . . . , j(k) coincide if and only if = , therefore
hei1 eik , ei1 eik i
1D X X E
= sgn ei(1) ...i(k) , sgn ei(1) ...i(k)
k!
Sk Sk
1 X k!
= kei(1) ...i(k) k2 = = 1.
k! k!
Sk
hv1 vk , w1 wk i
1D X X E
= sgn v(1) v(k) , sgn w(1) . . . w(k)
k!
Sk Sk
1 X X
= sgn sgn hv(1) , w(1) i hv(k) , w(k) i
k!
Sk Sk
hv(1) , w1 i hv(1) , wk i
1 X .. .. ..
= sgn det
k! . . .
Sk hv(k) , w1 i hv(k) , wk i
hv1 , w1 i hv1 , wk i
= det .. .. ..
.
. . .
hvk , w1 i hvk , wk i
Both sides of (1.10) are linear in , so the equation holds for all k (V ) if
and only if it is true for all vectors of a basis. Substituting the basis vector
= ei1 eik into (1.10) we obtain the equality
where j1 < < jnk are the elements of the complementary set {1, 2, . . . , n}\
{i1 , . . . , ik } arranged in increasing order. Thus, denoting by (i1 , . . . , ik ) the
sign of the permutation (i1 , . . . , ik , j1 , . . . , jnk ),
Corollary 1.2.72. Since the Hodge operator maps the orthonormal basis
(1.11) into an orthonormal basis of nk V it is an orthogonal transformation.
Corollary 1.2.73. As
Cross Product
Definition 1.2.74. Let V be a 3-dimensional oriented Euclidean vector field.
Then the binary operation : V V V , (a, b) 7 a b = (a b) is called
the cross product operation.
Proposition 1.2.75. The cross product has the following properties, which
give a geometrical way to construct it.
a b is orthogonal to both a and b.
The length of a b is equal to the area of the parallelogram spanned by
a and b. In particular, a b = 0 if and only if a and b are linearly
dependent. If a and b are nonzero and the angle between them is ,
then ka bk = kak kbk sin(). Geometrically, kak kbk sin() is the
area of the parallelogram spanned by the vectors a and b.
If a and b are linearly independent, then (a, b, a b) is a positively
oriented basis of V .
Proof. Let be the only positively oriented trivector of unit length in 3 V .
Then
kak2 ha, bi
2 2
ka bk = ka bk = det G(a, b) = det .
hb, ai |bk2
a b (a b) = ha b, (a b)i = ha b, a bi = ka bk2 .
ha, bci = hb, cai = hc, abi (permutation rule for mixed product);
ha, ci ha, di
ha b, c di = det (Lagrange identity);
hb, ci hb, di
a (b c) = ha, cib ha, bic (triple product expansion formula);
a (b c) + b (c a) + c (a b) = 0 (Jacobi identity).
Proof. Skew-commutativity and bilinearity follow directly from analogous
properties of the wedge product. The permutation rule is a consequence of
the identity a (b c) = ha, b ci and the alternating property a b c =
b c a = c a b of the wedge product.
As the Hodge star operator preserves the inner product, ha b, c di =
ha b, c di, and thus, Lagrange identity is a special case of equation (1.9).
Combining the permutation rule and the Lagrange identity, we obtain that
ha(bc), di = hda, bci = hd, biha, cihd, ciha, bi = hha, cibha, bic, di.
Since the first and last terms are equal for any d, a (b c) must be equal
to ha, cib ha, bic.
Expanding the summands on the left-hand side of the Jacobi identity we
obtain
(ha, cib ha, bic) + (hb, aic hb, cia) + (hc, bia hc, aib) = 0.
a b = (a1 e1 + a2 e2 + a3 e3 ) (b1 e1 + b2 e2 + b3 e3 )
= (a2 b3 a3 b2 )e1 + (a3 b1 a1 b3 )e2 + (a1 b2 a2 b1 )e3 .
Proposition 1.2.78.
The vector c = (a1 an1 ) is an alternating (n1)-linear function
of the vectors a1 , . . . , an1 .
c is orthogonal to ai for 1 i n 1.
c = 0 if and only if the vectors a1 , . . . , an1 are linearly dependent.
If a1 , . . . , an1 are linearly independent, then the length of c is the (n
1)-dimensional volume of the (n1)-dimensional p parallelepiped spanned
by the vectors a1 , . . . , an1 , which is equal to det G(a1 , . . . , an1 ).
(a1 , . . . , an1 , c) is a positively oriented basis of V .
If the coordinates of ai with respect to a positively oriented orthonormal
basis (e1 , . . . , en ) of V are (a1i , . . . , ani ), then
. . . a1n1
1
an1
a1
.. .. .. ..
c = det .
. . .
.
a1 n1 n
n1 . . . a n1 an1
e1 . . . en1 en
1.3 Geometry
1.3.1 Affine Geometry
Affine Spaces
In traditional axiomatic treatment of Euclidean geometry, vectors are defined
as equivalence classes of ordered pairs of points (also called directed seg-
ments). If X is the set of points of the space, then an ordered pair of points
is simply an element of the Cartesian product X X. If (A, B) X X is an
ordered pair of points, then A is called the initial point, and B is called the
endpoint of the ordered pair. The ordered pairs of points (A, B) and (C, D)
are said to be equivalent if the midpoint of the segment [A, D] coincides with
the midpoint of the segment [B, C]. It can be shown that this is indeed an
equivalence relation. The equivalence classes are called (free) vectors. The
set V of all vectors is equipped with a linear space structure. The sum of two
vectors is constructed by the triangle or parallelogram rule, multiplication by
scalars is defined in the usual way. The map : X X V, which assigns
to each ordered pair of points (A, B) the vector (A, B) = AB represented
by it, i.e., its equivalence class, satisfies the following properties.
34 1. Preliminaries
Figure 1.1: The construction of the sum of two vectors by the triangle and
the parallelogram rule.
Linear spaces and affine spaces are very similar objects. The main difference
is that in a linear space, we always have a distinguished point, the origin 0,
whereas in an affine space, none of the points of X is distinguished. This is
essentially the only difference, because if we choose any of the points of X
for the origin, we can turn X into a linear space isomorphic to V .
Indeed, choose a point O X. By the first axiom (A1) of an affine space,
the map O : X V , O (P ) = (O, P ) is a bijection between X and V ,
and we have
(A, B) = O (B) O (A)
by (A2). This identity means that O together with the linear map L = idV
is an isomorphism between the affine space A = (X, V, ) and the affine space
AV .
The vector O (P ) is called the position vector of A from the base point O.
Identification of points of an affine space with vectors with the help of a fixed
base point O is called vectorization of the affine space.
Affine Subspaces
Definition 1.3.3. A nonempty subset Y X of the point set of an affine
space A = (X, V, ) is an affine subspace of A if the image (Y Y ) = W
of Y Y under is a linear subspace of V and (Y, W, |Y Y ) is an affine
space. W is called the direction space of the affine subspace, its elements are
the direction vectors of Y , or the vectors parallel to Y . Since affine subspaces
are affine spaces themselves, their dimension is properly defined.
(y1 , y2 ) = y2 y1 = w2 w1 ,
1 p1 + + k pk = 1 (p1 a) + + k (pk a)
1 p1 + + k pk a = 1 (p1 a) + + k (pk a)
should hold for any choice of a. This condition is clearly equivalent to the
condition that 1 + + k = 1.
Definition 1.3.7. A linear combination 1 p1 + + k pk is an affine com-
bination if the sum 1 + + k of the coefficients is equal to 1.
The importance of affine combinations is summarized in the following propo-
sition.
Proposition 1.3.8. The point represented by an affine combination 1 p1 +
+ k pk of the position vectors of some points P1 , . . . , Pk does not depend
on the choice of the base point. This way, it makes sense to denote it by
1 P1 + + k Pk X.
Affine subspaces can be characterized in terms of affine combinations.
Proposition 1.3.9. A subset Y of an affine space is an affine subspace if
and only if it is nonempty and it contains all affine combinations of its points.
Proof. Identify the space with AV by vectorization.
If Y is an affine subspace, then it is a translation Ta W of its direction space.
Since 0 W , a Y is not empty. Furthermore, if yi = a + wi Y ,
(i = 1, . . . , k) are some points, then an affine combination of the has the form
1 y1 + + k yk = (1 + + k )a + (1 w1 + + k wk )
= a + (1 w1 + + k wk ).
Affine Independence
A direction space of the affine subspace spanned by a system of k + 1 points
P0 , . . . , Pk consists of 0-weight linear combinations of these points. This
linear space is generated by the differences P1 P0 , . . . , Pk P0 , therefore its
dimension is at most k.
Definition 1.3.15. The points P0 , . . . , Pk are called affinely independent if
any of the following equivalent conditions is fulfilled.
P0 , . . . , Pk span a k-dimensional affine subspace.
The vectors P1 P0 , . . . , Pk P0 are linearly independent.
A 0-weight linear combination of P0 , . . . , Pk equals 0 only if all the
coefficients are equal to 0.
1.3. Geometry 39
Figure 1.2: Pushing a body with equal forces at different base points.
40 1. Preliminaries
Definition 1.3.28. The angle (e, f ) between two 1-dimensional linear sub-
spaces e = lin[v] and f = lin[w] of a Euclidean linear space V , is the smaller
of and , where is the angle between v and w.
|hv, wi|
cos((e, f )) = .
kvk2 kwk2
Proof. We prove only the triangle inequality, the rest is trivial. Let e = lin[v],
f = lin[w], g = lin[z] be three 1-dimensional subspaces and let us show that
Assume that the vectors v, w and z have unit length. Denote by , and
the angles between (v, w), (w, z) and (z, v) respectively. Changing the
direction of the vectors w and z if necessary, we may assume that ,
[0, /2]. By Corollary 1.2.58, the Gram matrix of the vectors v, w, z has
nonnegative determinant, thus
1 cos() cos()
det cos() 1 cos()
cos() cos() 1
= 1 + 2 cos() cos() cos() cos2 () cos2 () cos2 () 0.
This is a quadratic inequality for cos(), which is fulfilled if and only if cos()
is in the closed interval the endpoints of which are the roots of the polynomial
Since the cosine function is strictly decreasing on the interval [0, ], this
implies
| | + ,
in particular,
We define the angle between k-dimensional linear subspaces using the Plucker
embedding Grk (V ) P(k V ) (see Corollary 1.2.33).
Since two different affine subspaces can have the same direction space, the
angle function is not a metric on the affine Grassmann manifold AGrk (A).
r : Rk V, r(x1 , . . . , xk ) = x1 a1 + + xk ak
This way, W contains all vectors orthogonal to (a1 an1 ). This last
vector is a normal vector of W .
Affine subspaces
Assume that we want to parameterize or write the equation of a k-dimensional
affine subspace W spanned by k + 1 affinely independent points a0 , . . . , ak .
It is not difficult to parameterize W , that is to obtain it as the image of a
bijection r : Rk W . We know that a point belongs to W if and only if it is
an affine combination of the points a0 , . . . , ak , consequently, the mapping
is a parameterization of W .
1.3. Geometry 45
x a0 ak = 0.
c0 + c1 x1 + + cn xn = 0. (1.14)
Equations of Spheres
Definition 1.3.33. Let be a (k+1)-dimensional subspace of an n-dimensional
Euclidean space, O be a point, r > 0 be a positive number. Then the
k-dimensional sphere or shortly the k-sphere in centered at O with radius
r is the set of those points in the distance of which from O is r. A hyper-
sphere is an (n 1)-dimensional sphere. The case r = 0 is considered to be
a degenerate case, when the sphere degenerates to a point.
This proves that every hypersphere can be defined by an equation of the form
where a, c R, b Rn .
Proof. The linear case is simple, the quadratic case follows from the equiva-
lent rearrangement
2 2
x b
= kbk 4ac
2a
4a2
of equation (1.15).
S = when r > d.
Because of the quadratic term in the last coordinate, this is not an affine
embedding. Its image V is a paraboloid. The advantage of this embedding
is that a linear equation on the coordinates of v is a hypersphere equation
on v.
Proposition 1.3.38. Let a0 , . . . , ak be (k + 1) 2 affinely independent
points. Then there is a unique (k 1)-sphere through these points and it can
be defined by the equation
x a0 ak = 0. (1.17)
ai a0 ak = 0.
1 = 0 + + k ,
x = 0 a0 + + k ak ,
kxk2 = 0 ka0 k2 + + k kak k2 .
The first two equations show that x must be an affine combination of the
points a0 , . . . , ak . Thus, S , and we have only two possibilities left.
Either S is a (k 1)-sphere, in which case we are done, or S = . The latter
case can be excluded by showing that the midpoint (a0 + a1 )/2 is not in S. If
it were in S, then by affine independence of the ai s, the only possible choice
for the i s would be 0 = 1 = 1/2 and i = 0 for 2 i k, however, the
third equation is not fulfilled with these coefficients as
a0 + a1
2
a0 a1
2
1 1
ka0 k2 + ka1 k2
=
> 0.
2 2 2
2
Denote by SPHk (V ) the set of all k-spheres in the Euclidean linear space
V . As a corollary of the previous proposition, we can construct an embed-
ding SPHk (V ) Gr1 (k+1 hhV ii) in the following way. Given a k-sphere
S, choose k affinely independent points a0 , . . . , ak from it, and assign to S
the 1-dimensional linear space spanned by the (k + 1)-vector a0 ak
k+1 hhV ii.
Exercise 1.3.39. Prove that the center of the sphere defined by (1.17) is the
point 0 a0 + + k ak , where the coefficients are obtained as the solution
of the following system of linear equations
0 + + k = 1
0 hai a0 , a0 i + + k hai a0 , ak i = kai k2 ka0 k2 for i = 1, . . . , k.
1.4 Topology
Definition 1.4.1. A pair (X, ) is said to be a topological space if X is a set,
is a collection of subsets of X, which we call the open subsets of X, such
that
50 1. Preliminaries
Metric Topology
The metric topology of a metric space (X, d) is introduced as follows.
Definition 1.4.6. The open ball in X with center x X and radius > 0 (or
an -ball centered at x) is the set B (x) = {y X | d(x, y) < }. Similarly,
the closed ball in X with center x X and radius > 0 (or a closed -ball
centered at x) is defined as the set B (x) = {y X | d(x, y) }.
Definition 1.4.7. A subset U of a metric space X is called open if for each
x U there is a positive such that the ball B (x) is contained in U .
1.4. Topology 51
U e
x
Proof. Obviously, the empty set and X are open. If U and V are open
subsets, and x is a common point of them, then there exist positive numbers
1 , 2 such that B1 (x) U and B2 (x) V . Let be the smaller of 1
and 2 . Then B (x) U V , showing that the intersection U V is open.
Finally, let {Ui : i I} be an arbitrary family of open sets, and x be an
element of their union. S S j I and a positive
Then we can find an index
such that B (x) Uj iI Ui , thus the union iI Ui is open.
Examples.
Subspace Topology
The topology of a topological space defines a topology on every of its subsets
by the following construction.
(i) Y = , X Y =Y;
(ii) (U Y ) (V Y ) = (U V ) Y ;
S S
(iii) iI (Ui Y ) = ( iI Ui ) Y .
Examples.
At the end of Section 1.3.2 the affine Grassmann manifolds and the
set of k-spheres of a Euclidean linear space were embedded into a pro-
jective space, which has a standard topology. Through these embed-
dings, affine Grassmann manifolds AGrk (V ) and the sets SPHk (V ) of
k-spheres also inherit a subspace topology.
1.4. Topology 53
Factor Topology
The topology of the Grassmann manifolds of all k-dimensional affine or linear
subspaces of an n-dimensional linear space can also be described with the help
of the factor space topology construction.
Assume that a topological space (X, ) is divided into a disjoint union of its
subsets. Such a subdivision can always be thought of as a splitting of X
into the equivalence classes of an equivalence relation on X. Denoting by
Y = X/ the set of equivalence classes we have a natural mapping : X Y
assigning to an element x X its equivalence class [x] Y .
Proposition 1.4.11. The set
0 = {U Y | 1 (U ) }
is a topology on Y .
Proof. The proof follows from the following set theoretical identities:
(i) 1 () = , 1 (Y ) = X;
(ii) 1 (U V ) = 1 (U ) 1 (V );
(iii) 1 ( iI Ui ) = iI 1 (Ui ).
S S
Product Topology
Definition 1.4.14. If (X1 , 1 ), . . . , (Xn , n ) are topological spaces, then the
product topology on the Cartesian product X1 Xn is defined as follows.
A subset U X1 Xn is open with respect to the product topology
if and only if for each point (x1 , . . . , xn ) U , we can find open subsets
U1 1 , . . . , Un n such that (x1 , . . . , xn ) U1 Un U .
Exercise 1.4.15. Show that the standard topology on Rn coincides with the
product topology on n copies of R.
Convergence, Continuity
Definition 1.4.16. We say that a sequence x1 , x2 , . . . of points of a topo-
logical space (X, ) converges to a point x X, if for any neighborhood U
of x there is a natural number N such that for n > N , xn U .
We say that two topological spaces are homeomorphic or have the same topo-
logical type, if there is a homeomorphism between them.
Homeomorphic topological spaces are considered to be the same from the
viewpoint of topology.
not
homeomorphic homeomorphic homeomorphic
The two point set X = {p, q} with topology = {X, {q}, } is normal and
T0 , but not T1 , hence not T4 . Every T4 space is T3 . The Sorgenfrey plane is an
example of a T3 space which is not normal. The Sorgenfrey plane is obtained
as the product space (R,S ) (R, ), where the (nonstandard) topology on
R consists of all unions iI [ai , bi ) of left closed right open intervals. (See
[9] Example 3. in 31, Ch. 4. for details.)
Countability axioms are of different character. They typically require the
existence of a countable family of subsets with some properties. Recall that
two sets have the same cardinality if and only if there is a bijection between
them. There is an ordering on the family of cardinalities. The cardinality
of a set A is less then or equal to the cardinality of the set B if there is an
injective map from A into B. A set is called countable if it is either finite
or has the same cardinality as the set of natural numbers. Equivalently, a
set is countable if there is a sequence p1 , p2 , . . . listing all elements of the
set. Every infinite set contains a countable infinite set, therefore a countable
infinite sets have the smallest cardinality among all infinite sets.
Definition 1.4.29. A topological space (X, ) is first-countable if for any
point p X, there is a countable family of open sets U1 , U2 , . . . such that
for any U containing p, there is an element Ui of the family for which
p Ui U .
Metric topologies are always first-countable. Choosing for Ui the open ball
of radius 1/i centered at p, we obtain a countable family of neighborhoods
of p satisfying the conditions.
Definition 1.4.30. A topological space (X, ) is second-countable if there is
a countable family of open sets U1 , U2 , . . . such that every openSset U is
the union of those elements Ui that are covered by U , i.e., U = Ui U Ui .
Every second-countable space is first-countable. The discrete topology on R
is metrizable but not second-countable.
Definition 1.4.31. A topological space (X, ) is said to be separable if con-
tains a countable dense subset.
Proposition 1.4.32. Every first-countable separable space is second-countable.
1.4.2 Compactness
Definition 1.4.33. A topological
S space (X, ) is said to be compact if from
any open covering X = iI Ui , Ui of X, we can choose a finite subcov-
ering X = Ui1 Uik , where i1 , . . . , ik I.
A related notion is sequential compactness.
1.4. Topology 59
A : Rm Rn if
kF (x) F (x0 ) A(x x0 )k
lim = 0.
xx0 kx x0 k
i F (x0 ) = ei F (x0 )
1 F 1 (x0 ) . . . m F 1 (x0 )
F 0 (x0 ) =
.. ..
. .
1 F n (x0 ) . . . m F n (x0 )
Definition 1.5.7. If the real valued function F : U R defined in a neigh-
borhood U of x0 Rm is differentiable at x0 , then its derivative matrix is a
row vector
grad F (x0 ) = (1 F (x0 ), . . . , m F (x0 )),
which we call the gradient vector of F at x0 .
According to this definition, for a vector valued function F , the rows of the
derivative matrix of F are the gradient vectors of the coordinates functions
of F .
1.5. Multivariable Calculus 65
Using the explicit form of the derivatives matrices and the multiplication
rule for matrices, we can reformulate the chain rule equating the matrix
coefficients.
(F i G)
Proposition 1.5.8 (Chain Rule | a reformulation). If U Rm is a neighbor-
hood of x0 U , and the function F = (F 1 , . . . , F n ) : U Rn is differentiable
at x0 , furthermore, the function G = (G1 , . . . , Gp ) : V Rp is defined on a
neighborhood of F (x0 ) and it is also differentiable at F (x0 ), then
n
X
i (G F )j (x0 ) = i (Gj F )(x0 ) = k Gj (F (x0 )) i F k (x0 ).
k=1
Bump Functions
Proposition 1.5.13. If K Rn is a compact subset of Rn and > 0, then
there exists a smooth function h : Rn [0, 1] such that h(x) is equal to 1 if
x K and h(x) = 0 if d(x, K) .
Proof. Define the function h1 : R [0, 1] by the formula
( 1
2n
xy
Z
h(x) = (y)hn dy
cn n Rn /2
Measure Spaces
Definition 1.6.1. A -algebra of subsets of a set X is a family of subsets
of X satisfying the following axioms.
X .
If A , then X \ A .
S
If A1 , A2 , . . . is a sequence of elements of , then i=1 Ai , i.e.,
is closed under countable unions.
Exercise 1.6.7. Show that given a measure space (X, , ), there is a unique
complete measure space (X, , ) such that A if and only if there exist
-measurable sets A, B such that (A \ A) (A \ A) B and (B) =
0, furthermore, if this is the case then (A) = (A). The measure space
(X, , ) is called the completion of the measure space (X, , ).
1.6. Measure and Integration 69
If compact sets of the space are Borel sets, for example, if the space is Haus-
dorff, then is said to be inner regular if for any Borel set B, we have
Lebesgue Measure
There are several methods to construct the Lebesgue measure n on Rn .
We can start with the construction of 1 as the completion of the unique
regular Borel measure 1 the value of which on open intervals is given by
1 ((a, b)) = b a. Uniqueness of 1 is clear since any open subset of R is a
countable union of disjoint open intervals, so the 1 measure of open subsets
is uniquely prescribed by -additivity. Then the measure of other Borel sets
is uniquely determined by outer regularity of 1 . The proof of existence
requires more work.
To proceed from 1 to n , we can use the product measure construction.
Theorem 1.6.9. Let (X1 , 1 , 1 ) and (X2 , 2 , 2 ) be two measure spaces.
Denote by 1 2 the smallest -algebras of subsets of X1 X2 containing
the family of subsets {A1 A2 | A1 1 , A2 2 }. Then there exists a
measure 1 2 on 2 such that
for all A1 1 and A2 2 . If both measure spaces are -finite, that is, if
they can be presented as a countable union of sets of finite measure, then the
measure 1 2 is unique.
Definition 1.6.10. If 1 and 2 are -finite, then the unique measure 1 2
is called the product measure of 1 and 2 .
Exercise 1.6.11. Prove that if X1 = Rk , X2 = Rl and i is the Borel
algebra of Xi , then 1 2 is the Borel algebra of Rk+l .
Starting from the Borel measure 1 on R, the product measure construction
gives a Borel measure n = 1 n1 on Rn = R Rn1 recursively for all
n. Completion of the Borel measure n is the Lebesgue measure n .
Exercise 1.6.12. Show that k l = k+l .
70 1. Preliminaries
where z = (x, y). However, one has to modify slightly the statement for
Lebesgue measures.
R For example, in the case of the Lebesgue measure, the
partial integral Rk f (x, y)dx may not exist for all x. Nevertheless, the set
of those points, for which the integral is not defined has Lebesgue measure
0. Since Lebesgue integrability and the Lebesgue integral of a function is
not affected by the modification of the function on a set of measure zero,
integration of functions whose values are not defined on a set of measure 0
makes sense.
bijection between U and V such that both h and h1 are of class C 1 . Then for
any Lebesgue measurable function f : V [ , ], the Lebesgue integral
of f over V exists if and only if the Lebesgue integral of (f h) | det h0 | over
U exists and if the integrals exist, then they are equal
Z Z
f (v)dv = f (h(u)) | det h0 (u)|du.
V U
Then Z Z
f d = n f c d. (1.18)
Sn1 Bn
Equation (1.18) can be checked easily for step functions. The general case
can be proved by approximating measurable functions with step functions.
A Lebesgue integral over a ball, can be obtained by integrating the function
over concentric spheres with respect to a suitably scaled measure and then
integrating these spherical integrals with respect to the radius.
n
Theorem 1.6.25 (Integration in Spherical Coordinates). Let f : BR R
n
be a Lebesgue measurable function on the n-dimensional ball BR of radius R
centered at the origin. Then
Z Z R Z
f d = rn1 f (ru)dudr,
n
BR 0 Sn1
where integrals over Sn1 are taken with respect to the measure , all the
other integrals are computed with respect to the Lebesgue measure.
Proof. Using equation (1.18), the right-hand side can be transferred to an
integral on B n [1, R]
Z R Z Z RZ
x
rn1 f (ru)dudr = n rn1 f r dxdr.
0 Sn1 0 B n \{0} kxk
Compute the last integral by the substitution
y
n
h : (BR \ {0}) (0, 1] (B n \ {0}) (0, R], h(y, s) = s , kyk .
kyk
The integrand composed with h is
y/kyk
kykn1 f kyk
y/kyk
= kyk
n1
f (y).
where In is the n n unit matrix. The second equality holds because sub-
tracting suitable multiples of the last column from the preceding ones we can
eliminate the term syyT /kyk3 from the upper left corner.
Exercise 1.6.26. Show that for any column vectors a, b Rn , we have
In a
= ha, bi.
det
T
b 0
Applying the result of the exercise,
sIn y
det
kyk kyk
yT /kyk 0
n+1 y n1
s In s
= det s = .
kyk kyk
yT /s 0
Therefore, the substitution gives
Z RZ Z 1Z n1
n1 x n1 s
n r f r dxdr = n kyk f (y) dyds
0 B n \{0} kxk 0 BR n kyk
Z 1 Z Z
=n sn1 ds f (y)dy = f (y)dy,
0 n
BR n
BR
Despite its more general form, every kth order differential equation is equiva-
lent to a first order one. The equivalent problem is to find a curve (, 1 , . . . ,
k1 ) : I U (Rn )k1 which satisfies the first order differential equation
78 1. Preliminaries
By Youngs Theorem, the right hand side of this equality should not change
if we flip the role of j and k.
Theorem 1.7.10 (Frobenius Theorem). The system of total differential
equations (1.22) is integrable if and only if
n
X n
X
k Fji + m+s Fji Fks = j Fki + m+s Fki Fjs
s=1 s=1
1.7. Ordinary Differential Equations 79
Curves in En
Probably anyone agrees that simple arcs are curves, but since open segments,
straight lines, conic sections and many other important examples of curves
are not simple arcs, this class of curves is too narrow.
We could define curves as finite unions of simple arcs. This wider class
includes circles, ellipses, but still excludes non-compact examples like straight
lines, hyperbolae, parabolas. Non-compact examples would be included if we
considered countable unions of simple arcs. This class of curves seems to
be wide enough, but maybe too wide. For example, it contains the set of
all those points in Rn which have at least one rational coordinate and it is
questionable whether we could call this set a curve.
81
82 2. Curves in En
Algebraic plane curves may have a finite number of singular points, for exam-
ple self intersections, so they are not necessarily 1-dimensional manifolds, but
removing the singular points, the remaining set is a 1-dimensional manifold,
maybe empty. On the other hand, algebraic curves are very specific curves.
For example, if a straight line intersects an algebraic curve in an infinite
number of points, then it is contained in the curve. In particular the graphs
periodic non-constant functions (like the sine function) are not an algebraic
curves.
One can also define curves as 1-dimensional topological or metric spaces.
For such a definition one must have a proper notion of dimension. Possible
definitions of dimension for a topological or metric space are discussed in
a branch of topology called dimension theory, and within the framework of
geometric measure theory. These theories are out of the main focus of this
textbook.
All the above definitions define curves as topological spaces or subsets of
topological spaces having a certain property satisfied by a sufficiently large
family of known examples of curves.
The second approach, which will be more suitable for our purposes, derives
curves from the motion of a point. This view is reflected in the definition of
a continuous curve:
By the Lagrange mean value theorem we can find real numbers ij such that
Fix a positive . By Proposition 1.4.43, for each > 0, we can find a positive
such that t, t [a, b] and |t t | < imply |x0j (t) x0j (t )| < for all
1 j n.
Suppose that the approximating broken line is fine enough in the sense that
|ti ti1 | < for all 1 i N . Then we have by the triangle inequality
v v v
u n
u n 0
u n
uX
uX uX
x 0 ( )2 t xj (ti )2 t (x0 (ij ) x0 (ti ))2 n.
ij
t
j j j
j=1 j=1 j=1
In this formula
v
N
u n 0 N
X uX X
(ti ti1 )t xj (ti )2 = (ti ti1 )| 0 (ti )|
i=1 j=1 i=1
Rb
is just an integral sum which converges to the integral a | 0 (t)|dt when
maxi |ti ti1 | tends to zero. Taking into account inequality (2.1) we see
that in this case the length of the inscribed broken lines also tends to this
integral.
2.2. The Length of a Curve 85
x = r cos(), y = r sin().
the surjective map , then E will be a Mobius band without its boundary
circle, and becomes a covering map, the universal covering map of E.
The Lebesgue measure on R2 defines a measure on E with the help
of the covering map as follows. Let a subset A E be -measurable if
and only if 1 (A) ([0, ] R) Lebesgue measurable and then let (A) =
(1 (A) ([0, ] R)).
Proposition 2.3.1. The measure is invariant under the isometry group
of the plane, that is, if Iso(R2 ) is an arbitrary isometry, A E is
a -measurable set of straight lines, then (A) is also -measurable, and
(A) = ((A)).
Proof. The isometry group of the plane is generated by the following three
types of transformations:
Rotations R by angle about the origin;
Translations Ta by a vector (a, 0) parallel to the x axis;
Reflection M in the x axis.
Hence it is enough to check the invariance of the measure under these trans-
formations.
It is clear that R (e,p ) = e+,p . Since the translation (, p) 7 ( + , p)
preserves Lebesgue measure, the statement is true for the rotations R .
The action of the translation Ta on the straight line parameters is not so sim-
ple as Ta (e,p ) = e,p+a cos() , nevertheless, as the determinant of the deriva-
tive of the transformation (, p) 7 (, p + a cos()) is 1, this transformation
also preserves Lebesgue measure.
Finally, the action of M on the line parameters is given by the map (, p) 7
(, p). The latter transformation is a reflection, therefore preserves the
Lebesgue measure.
as we wanted to show.
Crofton Formula in Rn
Theorem 2.3.3. Let : [a, b] Rn be a C 1 curve, and m : H N {}
be the map assigning to a hyperplane H the number m(H) = #{t [a, b] |
(t) H} of intersection points of the curve and the hyperplane H counted
with multiplicities. Then the length of is
Z
1
l = md.
n1 H
Proof. The proof is analogous to the planar case. Consider the C 1 -map
To compute the integral B n |hx, 0 (t)i|dx for a fixed t, write 0 (t) as v(t)u,
R
where v(t) = k 0 (t)k, u is a unit vector. Slice the ball with the hyperplanes
90 2. Curves in En
n
Hu, orthogonal to u. If [1, 1], then Hu, 0B is an (n 1)-dimensional
2
ball of radius 1 . Since the function hx, (t)i is equal to the constant
v(t) on this ball,
Z
|hx, 0 (t)i|dx = n1 v(t)| |(1 2 )(n1)/2 .
Hu, B n
Proof. For a fixed t [a, b], any vector w Rn can be decomposed uniquely
into the sum of two vectors so that the first vector is parallel to (t), the
second one is orthogonal to it. The decomposition is
Adding suitable multiples of the last column to the previous ones we can
eliminate the (t) T (t) term from the upper left corner. Then applying
Exercise 1.6.26 we get
In (t)
det(h0 (w, t)) = det
hw, (t)i 0 (t) hw, 0 (t)i
0 0
= hw, (t)i + hw, (t)ih(t), (t)i.
This reduces to det(h0 (w, t)) = hw, 0 (t)i as is on the unit sphere, therefore
(kk2 )0 = 2h, 0 i 0.
Applying Theorem 1.6.24 for h and the constant 1 function on B n+1 yields
Z bZ Z
0
|hw, (t)i|dwdt = m(H kxk
x )dxds
a Bn (x,s)B n+1
Z p
= 2 1 kxk2 m(H kxk x )dx.
Bn
The second equation comes from Fubinis theorem by slicing the ball B n+1
by straight lines orthogonal to the hyperplane Rn {0}. By equation (2.2),
we have
92 2. Curves in En
Z b Z
2n1
|hw, 0 (t)i|dwdt = l .
a Bn n+1
R1
According to Corollary 1.6.27, if we denote the integral 0 rn1 1 r2 dr by
an , then
Z p Z
x )
2m(H kxk 1 kxk2 dx = 2an m(Hu ) du.
Bn Sn1
where cn = (n+1)a n
n1 . The value of cn can be obtained by computing an , but
it can also be obtained by evaluating (2.3) for a great circle. The length of a
great circle on Sn1 is 2 and almost all hyperplanes through the origin cut
a great circle in exactly 2 points, that is the set of exceptional
R hyperplanes,
that contain the great circle has measure 0. This way, Sn1 m(Hu )du for a
great circle is twice the surface measure (Sn1 ) = nn of the sphere. This
gives that cn = /(nn ).
Exercise 2.3.5. Find a direct proof of the equations
Z 1 p n1 n+1
an = rn1 1 r2 dr = =
0 n(n + 1)n 2nn
2k1 (k1)!
(2k+1)!! , if n = 2k,
=
(2k1)!!
(2k+2)!! 2 , if n = 2k + 1.
We shall deal now with the problem of finding the k-plane which is tangent
to a curve at a given point with the highest possible order. The classical
approach to this problem is the following. A k-plane is determined uniquely
by k + 1 points of it not lying in a (k 1)-plane. Let us take k + 1 points
(t0 ), (t1 ), . . . , (tk ) on the curve. If is a curve of general type then
these points span a unique k-plane, which will be denoted by A(t0 , . . . , tk ).
The k-plane we look for is the limit position of the k-planes A(t0 , . . . , tk ) as
t0 , . . . , tk tend to t.
To properly understand the last sentence, we need a definition of convergence
of k-planes. In section 1.4, we defined a topology on the affine Grassmann
manifold AGrk (Rn ) in different but equivalent ways. Constructing the topol-
ogy as a factor topology gives the following notion of convergence of k-planes.
Proposition 2.4.1. The sequence of k-planes X1 , X2 , . . . tends to the k-
plane X with respect to the topology of the affine Grassmann manifold
AGrk (Rn ) if and only if one can find points pj Xj , p X and linearly
independent direction vectors v1j , . . . , vkj of Xj and v1 , . . . , vk of X such that
limj pj = p and limj vij = vi for i = 1, . . . , k.
2
v1
vv
1 1
1
p
p1 p2 v2
2
v21 v2
Figure 2.3: Convergence of k-planes
Proof. Suppose that the sequence X1 , X2 , . . . has two limits, say X and Y .
Then by the definition, one can find points pj , qj Xj , p X, q Y and
linearly independent direction vectors {v1j , . . . , vkj } and {w1j , . . . , wkj } of Xj ,
{v1 , . . . , vk } of X and {w1 , . . . , wk } of Y , such that lim pj = p, lim qj = q,
lim vij = vi and lim wij = wi for i = 1, . . . , k as j tends to infinity. Since
{v1j , . . . , vkj } and {w1j , . . . , wkj } span the same linear space, which contains
the direction vector pj qj , there exist a unique k k matrix (ajrs )1r,sk
and a vector (bj1 , . . . , bjk ) such that
k
X
vij = ajis wsj for i = 1, . . . , k;
s=1
(2.4)
k
X
pj qj = bjs wsj
s=1
The components of this matrix and the numbers bjs can be determined
ajrs
by solving the system (2.4) of linear equations, thus by Cramers rule, they
are rational functions (quotients of polynomials) of the components of the
vectors vij and wij . Using the fact that rational functions are continuous at
each point where they are defined one can show that the limits lim ajrs = ars ,
lim bjr = br exist as j . Taking j in (2.4) we obtain
k
X
vi = ais ws for i = 1, . . . , k;
s=1
k
X
pq= bs ws ;
s=1
from which follows that the vectors v1 , . . . , vk and w1 , . . . , wk span the same
linear space, i.e. the k-planes X and Y must be parallel and that the point
q is a common point of them, consequently X = Y .
f0 (t0 ) := f (t0 ),
f (t1 ) f (t0 )
f1 (t0 , t1 ) := ,
t1 t0
..
.
fk1 (t1 , . . . , tk ) fk1 (t0 , . . . , tk1 )
fk (t0 , t1 , . . . , tk ) := .
tk t0
Exercise 2.4.7. Show that the k-th order divided difference is a symmetric
function of the variables t0 , . . . , tk and has the following explicit form
k
X 1
fk (t0 , t1 , . . . , tk ) = f (ti ) ,
i=0
0 (ti )
f (k) ()
fk (t0 , t1 , . . . , tk ) = .
k!
Proof. Let P (t) be the polynomial of degree k for which f (ti ) = P (ti ) for
i = 0, 1, . . . , k. Such a polynomial exists and is unique. P is unique, since
if Q is also a polynomial of degree k such that f (ti ) = P (ti ) = Q(ti ) for
i = 0, . . . , k, then the polynomial P Q is of degree k, and P Q has k + 1
roots, which is possible only in the case when P Q = 0, i.e., P = Q. We
show the existence of P by an explicit construction. Set
Corollary 2.4.9. Since can be chosen from the interval spanned by the
points t0 , . . . , tk , if these points tend to t [a, b], then also tends to t,
f (k) () f (k) (t)
consequently fk (t0 , t1 , . . . , tk ) = tends to .
k! k!
Corollary 2.4.10. Applying the previous corollary to the components of a
vector valued function f : [a, b] Rn , we obtain that fk (t0 , t1 , . . . , tk ) tends
f (k) (t)
to as t0 , . . . , tk tend to t.
k!
Proof of Theorem 2.4.5. Let us recall that if p0 , . . . , pk are position vectors
of k + 1 points in Rn , then the affine plane spanned by them consists of linear
combinations the coefficients in which have sum equal to 1
A(p0 , . . . , pk ) = {0 p0 + + k pk | 0 + + k = 1}.
The direction vectors of this affine plane are linear combinations 0 p0 + +
k pk such that 0 + + k = 0.
We claim that if : [a, b] Rn is a curve in Rn , then k (t0 , . . . , tk ), (k 1) is
a direction vector of the affine subspace spanned by the points (t0 ), . . . (tk ).
Pk
To see this, it suffices to show that i=0 1/ 0 (ti ) = 0. Consider the function
f = 1 and construct the polynomial P of degree k which coincides with
f at t0 , . . . , tk using the general formulae. By the above proposition, there
exists a number such that
k
X 1
0 = f (k) () = P (k) () = ,
i=0
0 (ti )
as we wanted to show.
2.5. Frenet Frames and Curvatures 97
This way, (t0 ) is a point and 1 (t1 , t0 ), . . . , k (tk , . . . , t0 ) are direction vectors
of the affine subspace spanned by the points (t0 ), . . . , (tk ). If is smooth,
then (t0 ) tends to (t), 1 (t1 , t0 ) tends to 0 (t), and so on, k (tk , . . . , t0 )
tends to (k) (t)/k! as the points t0 , . . . , tk tend to t [a, b]. Since by our
assumption the first k derivatives of are linearly independent at t, so are
the vectors 1 (t1 , t0 ), . . . , k (tk , . . . , t0 ) if t0 , . . . , tk are in a sufficiently small
neighborhood of t, and in this case the k-plane A(t0 , . . . , tk ) tends to the
k-plane that passes through (t) with direction vectors 0 (t), 00 (t)/2, . . . ,
(k) (t)/k!.
Exercise 2.4.11. Prove by induction on k that k (t0 , . . . , tk ), (k 1) is a
direction vector of the affine subspace spanned by the points (t0 ), . . . , (tk ).
span the osculating k-plane at the point in question, then measuring the
speed of change of this basis.
Definition 2.5.1. A (smooth) vector field along a curve : I Rn is a
smooth mapping v : I Rn .
Remark. There is no formal difference between a curve and a vector field
along a curve. The difference is only in the interpretation. When we think of
a map v : I Rn as a vector field along the curve : I Rn we represent
(depict) v(t) by a directed segment starting from (t).
Definition 2.5.2. A moving (orthonormal) frame along a curve : I Rn
is a collection of n vector fields t1 , . . . , tn along such that hti (t), tj (t)i = ij
for all t I.
There are many moving frames along a curve and most of them have nothing
to do with the geometry of the curve. This is not the case for Frenet frames.
Definition 2.5.3. A moving frame t1 , . . . , tn along a curve is called a
Frenet frame if for all k, 1 k n, (k) (t) is contained in the linear span of
t1 (t), . . . , tk (t).
t2(t) t2(t)
v(t) t1(t)
g(t) g(t) g(t)
t1(t)
Figure 2.4: A vector field, a moving frame and the distinguished Frenet frame
along a curve.
Exercise 2.5.4. Construct a curve which has no Frenet frame and one with
infinitely many Frenet frames. Show that a curve of general type in Rn has
exactly 2n Frenet frames.
According to the exercise, a Frenet frame along a curve of general type is
almost unique. To select a distinguished Frenet frame from among all of
them, we use orientation.
Definition 2.5.5. A Frenet frame t1 , . . . , tn of a curve of general type in
Rn is called a distinguished Frenet frame if for all k, 1 k n 1, the
vectors t1 (t), . . . , tk (t) have the same orientation in their linear span as the
vectors 0 (t), . . . , (k) (t), and the basis t1 (t), . . . , tn (t) is positively oriented
with respect to the standard orientation of Rn .
2.5. Frenet Frames and Curvatures 99
1 0
ij + ji = (ht , tj i + hti , t0j i) = 0
v i
by differentiation. The proposition is proved.
1 = 12 , 2 = 23 , . . . ., n1 = n1,n ,
1 0
t = 1 t2
v 1
1 0
t = 1 t1 + 2 t3
v 2
..
.
1 0
t = n2 tn2 + n1 tn
v n1
1 0
t = n1 tn1 .
v n
These formulae are called the Frenet formulae for a curve of general type
in Rn . The functions 1 , . . . , n1 are called the curvature functions of the
curve.
We formulate two invariance theorems concerning the curvatures of a curve.
They are intuitively clear and their proof is straightforward.
The coefficients aji can be expressed recursively by the speed length function
v = k 0 k and the curvature functions. To illustrate this, let us compute the
first three decompositions assuming n 3.
The speed vector is simply
0 = vt1 . (2.7)
Differentiating and applying the first Frenet equation
00 = v 0 t1 + vt01 = v 0 t1 + v 2 1 t2 . (2.8)
Differentiating again and using the first two Frenet equations we get
000 = v 00 t1 + v 0 t01 + (v 2 1 )0 t2 + v 2 1 t02
= v 00 t1 + v 0 v1 t2 + (v 2 1 )0 t2 + v 2 1 v(2 t3 1 t1 ) (2.9)
00 0 0
= (v v 3 21 )t1 2 3
+ (v v1 + (v 1 ) )t2 + v 1 2 t3 .
102 2. Curves in En
In principle, iterating this method we can compute all the coefficients aji but
already the third derivative shows that the coefficients become complicated
and do not show any pattern except for the last coefficients.
Proposition 2.5.12.
akk = v k 1 k1 for 1 k n.
Proof. We prove the equation by induction on k. The base cases are verified
above for k 3. Assume now that the statement holds for k 1 < n. Then
differentiating
We see that tk appears only in the last summand, when j = k 1 and its
coefficient is
akk = ak1
k1 vk1 = (v
k1
1 k2 )vk1 = v k 1 k1 .
Consequently, the sign of the last curvature coincides with the the sign of
n , which is positive (or negative) if and only if ( 0 , . . . , (n) ) is a positively
(or negatively) oriented basis of Rn . It is 0 if and only if 0 , . . . , (n) are
linearly dependent. The remark at the end of the proposition follows from
the Fundamental Theorem of Curve Theory, which will be proved below.
This theorem that for any smooth function n1 on an interval, there is a
curve of general type in Rn the last curvature of which is n1 .
As a byproduct of the proof, we obtain the following expressions for the
curvatures
a2 2 /1 2
1 = 21 = = 3
va1 v1 v
and
ak+1
k+1 k+1 /k k+1 k1
k = = =
vakk vk /k1 v2k
for 2 k n 1. Thus, to find formulae for the curvatures it is enough
to express the determinants k or the diagonal elements akk in terms of the
derivatives of .
If we take the wedge product of the first k equations (2.6) and apply Propo-
sition 1.2.31, we get
0 (k) = k t1 tk . (2.11)
As t1 , (t) . . . , tn (t) an orthonormal basis for each parameter t,
k 0 (k) k = |k |. (2.12)
Since the determinants k are positive for k n 2, the absolute value can
be omitted for all but the last k. The length of the k-vector 0 (k) can
be expressed using (1.9) with the help of the Gram matrix G( 0 , . . . , (k) ).
Thus for k n 1,
q
k = k 0 (k) k = G( 0 , . . . , (k) )
v 0 0
h , i . . . h 0 , (k) i
u
u
=u
u .. .. ..
tdet
. . .
h (k) , 0 i . . . h (k) , (k) i
104 2. Curves in En
h (k) , ti i = aik ,
and
k1
X
(k)
fk = aik ti = akk tk .
i=1
This way, the length of fk , which we also have to compute for evaluating the
equation
fk
tk = ,
kfk k
is equal to
kfk k = |akk | = v k |1 k1 |,
and therefore
kfk+1 k
|k | = .
vkfk k
This formula allows us to compute also |n1 |, but to determine the sign
of n1 , one needs further consideration of the orientation of the first n
derivatives of .
Exercise 2.5.15. Compute the curvatures of the moment curve (t) =
(t, t2 , . . . , tn ) at t = 0.
T0 = C T (2.13)
Thus, prescribing the Frenet basis at a given point T(t0 ), we can obtain
the whole moving Frenet frame as a unique solution of (2.13). For each
t I, T(t) must be an orthogonal matrix, that is a matrix the rows of which
form an orthonormal basis of Rn . Orthogonality is equivalent to the equation
T TT = TT T = In , where TT is the transposition of T, In is the n n unit
matrix. Furthermore, since the Frenet frame must be a positively oriented
basis of Rn , T(t) must have positive determinant for all t I.
Recall that orthogonal matrices with positive determinant are called special
orthogonal matrices. The set (in fact group) of n n special orthogonal
matrices is usually denoted by SO(n).
We claim that if T is a solution of (2.13) and there is a number t0 , for which
T(t0 ) SO(n), then T(t) SO(n) for all t I. This means that if we take
care of the restrictions on T when we choose the initial matrix T0 , then we
do not have to worry about the other values of T.
To show the statement, set M = TT T. Then
M0 = (T0 )T T + TT T0 = (C T)T T + TT C T = TT (C T + C) T = 0,
(2.14)
as C is skew symmetric. (2.14) shows that M is constant, therefore M
M(t0 ) = In . This proves that T(t) is an orthogonal matrix for all t I.
As for the determinant of T(t), observe that the determinant of an orthogonal
matrix A Rnn satisfies
The derivative of is vt1 , thus we can obtain from the Frenet frame by
integrating vt1
Z t
(t) = v( )t1 ( )d. (2.15)
t0
Solving (2.13) we get the only possible Frenet frame, then formula (2.15)
defines the only curve which can satisfy the conditions. This proves unique-
ness. To show the existence part of the theorem we have to check that the
curve we have obtained has the prescribed curvature and length of speed
functions. The lengthy but straightforward verification of this fact is left to
the reader.
Exercise 2.5.17. Plot the curve, called astroid, given by the parameteriza-
tion : [0, 2] R2 , (t) = (cos3 (t), sin3 (t)). Is the astroid a smooth curve?
Is it regular? Is it a curve of general type? If the answer is no for a property,
characterize those arcs of the astroid which have the property. Compute the
length of the astroid. Show that the segment of a tangent lying between the
axis intercepts has the same length for all tangents.
Exercise 2.5.18. Find the distinguished Frenet frame and the equation of
the osculating 2-plane of the elliptical helix t 7 (a cos t, b sin t, ct) at the point
(a, 0, 0) (a, b and c are given positive numbers).
k (s)/2
lim = 1.
s0 sin(k (s)/2)
As a corollary, we get
k (s)
(t1 tk )(t0 + s) (t1 tk )(t0 )
lim = lim
s0 |s| s0 s
= k(t1 tk )0 (t0 )k.
Computing the derivative with the Leibniz rule and applying the Frenet for-
mulae
k
X
(t1 tk )0 = t1 t0i tk
i=1
k
X
= (1 t2 t2 tk + t1 (i ti+1 i1 ti1 ) tk .
i=2
As the wedge product of vectors vanishes if two of the vectors are equal, the
only nonzero summand on the right-hand side occurs at i = k
Proof. The volume of a simplex can be obtained from the wedge product of
the edge vectors running out from one of the vertices
1
k((t1 ) (t0 )) ((tk ) (t0 ))k.
Vk (t0 , . . . , tk ) =
k!
We know that the mth order divided difference of has the following explicit
form
m
X 1
m (t0 , . . . , tm ) = Qm (ti ).
j=0 (ti tj )
i=0 j6=i
also holds. Consider equation (2.16) for m = 1, . . . , k and take the wedge
product of these k equations. Applying Proposition 1.2.31 we get
1 (t0 , t1 ) k (t0 , . . . , tk )
1
= Qk Qm1 ((t1 ) (t0 )) ((tk ) (t0 )).
m=1 j=0 (tm tj )
(2.17)
110 2. Curves in En
Taking the length of both sides and rearranging the equation gives
As we know the limits of the higher order divided differences as all the ti s
tend to t, we can compute the limit of the right-hand side easily
k
!
Vk (t0 , . . . , tk ) 1 Y 1 k( 0 (k) )(t)k
lim Q = .
i:ti t 0i<jk k(tj ) (ti )k k! m=1
m! k 0 (t)k(k+1)k/2
Though for k > 2 the limit in Proposition 2.6.2 involves several curvatures,
it is possible to construct expressions of the volumes of faces of various di-
mensions of a simplex whose limit gives one single curvature. An example of
such a formula is Darbouxs formula for the second curvature.
Corollary 2.6.4 (Darboux). Using the notations of Proposition 2.6.2,
9 V3 (t0 , t1 , t2 , t3 )
lim p = |2 (t)|.
t0 ,t1 ,t2 ,t3 t 2 V2 (t1 , t2 , t3 )V2 (t0 , t2 , t3 )V2 (t0 , t1 , t3 )V2 (t0 , t1 , t2 )
two facets of S, say the facet F spanned by the points (t0 ), . . . , (tk1 ) and
F spanned by (t1 ), . . . , (tk ). These two facets intersect along a (k 2)-
dimensional face L = F F . Let = (t0 , . . . , tk ) be the angle of the
(k 1)-planes spanned by the facets F and F .
We remark that if is the angle between the outer unit normals of the facets
F and F , then the dihedral angle of the simplex at the (k 2)-dimensional
face L is defined to be the complementary angle . The angle of
the (k 1)-planes is the smaller of and the dihedral angle , that is,
= min{, } [0, /2].
(t0 , . . . , tk ) k1 (t)
lim = .
i:ti t k(tk ) (t0 )k k
sin() P P 0 EF EF
=
k(tk ) (t0 )k P P 00 ES EL
Vk (t0 , . . . , tk ) Vk2 (t1 , . . . , tk1 )
k ES EL
= .
k 1 Vk1 (t0 , . . . , tk1 ) Vk1 (t1 , . . . , tk )
EF EF
sin((t0 , . . . , tk ))
lim
i:ti t k(tk ) (t0 )k
k1 (t) k3 (t)
k1 (t) (t)
1 1 k3
1!2! k! k! 1!2! (k 2)! (k 2)!
k k1 (t)
= 2 = .
k1 k
k2 (t)
k2 (t)
1
1!2! (k 1)! (k 1)!
The above equation shows that tends to 0 as all the ti s tend to t, therefore
/ sin() tends to 1, so the equation implies the statement.
is
If we apply formula (2.17) for , the equation of the sphere can be rewritten
as
x (t0 ) 1 (t0 , t1 ) k+1 (t0 , . . . , tk+1 ) = 0.
As the limit of higher rth order divided difference of as all the parameters
tend to t exists and equals (r) (t)/r!, we have
Hence the limit of the k-sphere S(t0 , . . . , tk+1 ) exists as the parameters t0 , . . . ,
tk+1 tend to t, and the equation of the limit sphere is
Definition 2.7.2. The limit of the k-spheres S(t0 , . . . , tk+1 ) as each param-
eter ti tends to t is called the osculating k-sphere of at t.
Rk2 = 12 + + k+1
2
(2.21)
Proof. Use the same notation as in the proof of the previous proposition.
Keeping the parameters t0 , . . . , tk+1 fixed for a moment, consider the function
F : I R given by
This equation is valid for any t. Rearranging the equation we can bring it to
a simpler form
1
h (j) , ok i = (kk2 )(j) for j = 1, . . . , k + 1.
2
Assume that j k end differentiate this equation. We obtain
1
h (j+1) , ok i + h (j) , o0k i = (kk2 )(j+1) .
2
On the other hand, since j + 1 is still less than or equal to k + 1, we have
1
h (j+1) , ok i = (kk2 )(j+1) .
2
The difference of the last two equations gives that o0k is orthogonal to the
first k derivatives of , consequently, it is perpendicular to the osculating
k-plane.
Proposition 2.7.6. The coefficients i can be computed recursively by the
formulae
1 = 0,
1
2 = ,
1
1 k0
k+1 = + k1 k1 .
k v
Proof. A 0-sphere is a pair of points on a straight line, and the center is at the
midpoint. The limit of the center of the 0-sphere {(t0 ), (t1 )} as t0 , t1 t
is (t). Thus
o0 = + 1 t1 = ,
showing 1 = 0.
Differentiating (2.20) for k 1 we obtain
k+1
X
o0k = vt1 + (j0 tj + v j (j tj+1 j1 tj1 )).
j=2
116 2. Curves in En
Remark. Observe that equation (2.22) can be rearranged into a form re-
sembling the Frenet equations
1 0
= k k+1 k1 k .
v k
With the help of the proposition, we can compute the location of the centers
of the osculating k-spheres recursively. The initial cases are
1
o1 = + t2 ,
1
1 01
o2 = + t2 t3
1 v 2 21
0 !
01 01
1 2 1
o3 = + t2 t3 + t4 .
1 v 2 21 1 3 v 3 v 2 21
..
.
Proposition 2.7.7. The radius functions Rk of the osculating k-spheres obey
the following recursive rules
R02 = 0,
1
R12 = 2 ,
1
2
R2 Rk0
2
Rk+1 = Rk2 + 2 k2 .
Rk Rk1 v k+1
Proof. The initial cases are obvious. For 1 k n2, differentiate equation
(2.21). Omitting 1 = 0 and dividing by 2 we obtain
k+1
X
Rk Rk0 = j j0 .
j=2
2.8. Plane Curves 117
k+1
X
Rk Rk0 = vj (j+1 j j1 j1 )
j=2
k+1
X
= vj+1 j j vj j1 j1 = vk+2 k+1 k+1 .
j=2
2
If we square this equation and use s+1 = Rs2 Rs1
2
, then we get
t0 = vn,
n0 = vt,
where v = | 0 |.
Let us find explicit formulae for t, n and . Obviously,
1 0 0 1
t= (x , y ) = p 2 (x0 , y 0 ).
v 0
x +y 0 2
The normal vector n is the last vector of the Frenet basis so it is determined
by the condition that (t, n) is a positively oriented orthonormal basis, that is,
in our case, n is obtained from t by a 90 degree rotation in positive direction.
118 2. Curves in En
The right angled rotation in the positive direction takes the vector (a, b) to
the vector (b, a), thus
1 1
n= (y 0 , x0 ) = p 2 (y 0 , x0 ).
v x0 + y 0 2
To express , let us start from the equation
0 = vt.
Differentiating and using the first Frenet formula,
00 = v 0 t + vt0 = v 0 t + v 2 n.
Taking dot product with n and using ht, ni = 0 we get
h 00 , ni = v 2 ,
which gives
x0 y0
det
h 00 , ni x00 y 0 + y 00 x0 x00 y 00
= = = ,
v2 v3 (x0 2 + y 0 2 )3/2
in accordance with Theorem 2.5.14.
Proof. The intersection point of the normals exist if the corresponding tan-
gents are not parallel. Since t0 (t) = (t)n(t) 6= 0, the unit tangent t rotates
with nonzero angular speed at t, hence no two tangents are parallel in a small
neighborhood of t.
Assume t0 , t1 are close to t and write M (t0 , t1 ) as
using the fact that M (t0 , t1 ) is on the normal at t0 . Since M (t0 , t1 ) is also
on the normal at t1 ,
g(t1)
F(t1,t2)
g(t2)
Figure 2.5: Parallel rays reflected from a curve crossing one of the rays or-
thogonally.
evolute
Definition 2.8.6. Let be a regular plane curve with normal vector field
n. A parallel curve of is a curve of the form d = + dn, where d R is
a fixed real number.
Experiments on wave fronts show that even if the initial curve is smooth and
regular, singularities may appear on its parallel curves. As the distance d
varies, singularities on d are usually born and disappear in pairs.
2.8. Plane Curves 121
Figure 2.7: Parallel curves of an ellipse and the evolute swept out by the
singular points of them.
The length of the arc of the evolute between o1 (t1 ) and o1 (t2 ), t1 < t2 ,
is the difference of the radii of curvatures |1/(t1 ) 1/(t2 )|.
This equation shows that if 0 (t) 6= 0, then o1 is regular at t and its tangent
at t is parallel to the normal n(t) of . This proves the first part of the
122 2. Curves in En
0
= (1/) ( )d = |1/(t1 ) 1/(t2 )|.
t1
involute
evolute
Proof. We have
The first equation implies that the Frenet frame t, n of is related to that
of by t = n, n = t. Computing the curvature of ,
Exercise 2.8.16. The chain curve is the graph of the hyperbolic cosine
function
ex + ex
cosh(x) = .
2
124 2. Curves in En
Determine the involute of the chain curve through the point (0, 1). (This
curve is called tractrix.)
Let the tangent of the tractrix at P intersect the x-axis at Q. Show that
the segment P Q has unit length.
Exercise 2.8.17. Let be a regular plane curve for which the curvature
function and its derivative are positive. Show that for any t1 < t2 from
the parameter domain of the osculating circle of at (t1 ) contains the
osculating circle at (t2 ).
= 0 ,
i.e. the curvature is the derivative of the direction angle (angular speed) of
the tangent vector with respect to the arc length.
Definition 2.8.18. The total curvature of a curve is the integral of its cur-
vature function with respect to arc length
Z b
(s)ds.
a
2.8. Plane Curves 125
Convex curves can be characterized with the help of the curvature function.
Theorem 2.8.32. A simple closed curve is convex if and only if 0 or
0 everywhere along the curve.
Proof. Assume first that is a naturally parameterized convex curve. Let
(t) be a continuous direction angle for the tangent t(t). As we know, 0 = ,
thus, it suffices to show that is a weakly monotone function. This follows if
we show that if takes the same value at two different parameters t1 , t2 , then
is constant on the interval [t1 , t2 ]. The rotation number of a simple curve
is 1, hence the image of t covers the whole unit circle. As a consequence,
we can find a point at which
e
s
p x q
Figure 2.12
and
rotation number = ((b) (a))/2 = 0.
In the second case is constant on the interval [t1 , t2 ], as we wanted to show.
Now we prove the converse. Assume that is a simple closed curve with
0 everywhere and assume to the contrary that is not convex (the case
0 can be treated analogously). Then we can find a point p = (t1 ),
such that the tangent at p has curve points on both of its sides. Let us find
on each side a curve point, say q = (t2 ) and r = (t3 ) respectively, lying
at maximal distance from the tangent at p. Then the tangents at p, q and
r are different and parallel. Since the unit tangent vectors t(ti ), i = 1, 2, 3
have parallel directions, two of them, say t(ti ) and t(tj ) must be equal. The
points a = (ti ) and b = (tj ) divide the curve into two arcs. Denoting
by K1 and K2 the total curvatures of these arcs, we deduce that these total
curvatures have the form K1 = 2k1 , K2 = 2k2 , where k1 , k2 Z, since
the unit tangents at the ends of the arcs are equal. On the other hand, we
have k1 + k2 = 1 by the Umlaufsatz and k1 0, k2 0 by the assumption
0. This is possible only if one of the total curvatures K1 or K2 is equal
to zero. Since 0, this means that = 0 along one of the arcs between a
and b. But then this arc would be a straight line segment, implying that the
tangents at a and b coincide. The contradiction proves the theorem.
{P R2 | P A P B = 1/4(AB)2 },
where A 6= B are given points in the plane and show that it is a closed curve
with exactly two vertices. Determine the rotation number of the lemniscate.
Exercise 2.8.35. Find the points on the ellipse (t) = (a cos t, b sin t) at
which the curvature is minimal or maximal (a > b > 0).
130 2. Curves in En
Theorem 2.8.36 (Four Vertex Theorem). A convex closed curve has at least
4 vertices.
This result is sharp. For example, an ellipse has exactly four vertices.
Proof. Local maxima and minima of the curvature function yield vertices.
One can always find a local minimum on an arc bounded by two local max-
ima, hence if we have two local maxima or minima of the curvature then we
must have at least four vertices. Thus we have to exclude the case when the
curvature function has one absolute maximum at A and one absolute mini-
mum at B, and strictly monotone on the arcs bounded by A and B. In such
a case, choose a coordinate system with origin at A and x-axis AB.
The arcs of the curve bounded by A and B do not cut the x-axis at points
other than A and B. Indeed, if there were a further intersection point C,
then the curve would be split into three arcs by A,B and C in such a way
that on each arc we could find a point at which the tangent to the curve is
parallel to the straight line ABC. If the three tangents at these points were
different then the one in the middle would cut the curve apart contradicting
convexity, if two of the tangents coincided then we could find a straight line
segment contained in the curve yielding an infinite number of vertices.
If the two arcs bounded by A and B lied on the same side of AB then the
line AB would be a common tangent of the curve at A and B. In this case
the segment AB would be contained in the curve, yielding an infinite number
of vertices as before.
We conclude that for a suitable orientation of the y-axis, y(t)0 (t) 0 for
every t [a, b], where (t) = (x(t), y(t)), t [a, b] is a unit speed parameter-
ization of the curve. Hence we get
Z b
y(t)0 (t)dt > 0.
a
vertices:
A B
Bernoulli's lemniscate
ellipse
Integrating by parts,
Z b Z b Z b
b
y(t)0 (t)dt = [y(t)(t)]a y 0 (t)(t)dt = y 0 (t)(t)dt.
a a a
The unit tangent vector field of the curve is t = (x0 , y 0 ), the unit normal
vector field is n = (y 0 , x0 ), hence by the first Frenet formula,
x00 = y 0 .
Integrating,
Z b Z b
0 b
y (t)(t)dt = x00 (t)dt = [x0 (t)]a = 0.
a a
2.9 Curves in R3
A 3-dimensional curve is a curve of general type if its first two derivatives are
not parallel. From now on we shall suppose that the curves under consider-
ation are all of general type.
The distinguished Frenet frame vector fields t1 , t2 and t3 of a 3-dimensional
curve are denoted in classical differential geometry by t, n and b and they are
called the (unit) tangent, the principal normal and the binormal vector fields
of the curve respectively. These vector fields define a coordinate system at
each point of the curve. The coordinate planes of this coordinate system are
given the following names. We are already familiar with the plane that goes
through a given curve point and spanned by the directions of the tangent and
the principal normal. It is the osculating plane of the curve. The plane that is
spanned by the principal normal and the binormal is the plane that contains
all straight lines orthogonally intersecting the curve at the given point.
For this obvious reason, this plane is called the normal plane of the curve.
The third coordinate plane, that is the plane spanned by the tangent and
the binormal directions is the rectifying plane of the curve. The reason for
this naming will become clear later. As we know from the general theory, a
3-dimensional curve of general type has two curvature functions 1 , which is
always positive and 2 , which may have any sign. The first curvature 1 is
denoted by the classics simply by and it is referred to as the curvature of
the curve while the second curvature 2 is called the torsion of the curve and
is denoted by .
132 2. Curves in En
Using the classical notation, Frenets formulae for a space curve can be writ-
ten as follows.
t0 = vn
n0 = vt + v b
b0 = v n.
Let us find explicit formulae for the computation of these vectors and curva-
tures in an economical way. Assume that : [a, b] R3 is a curve of general
type. The unit tangent vector field t can be obtained by normalizing the
speed vector 0
0
t= 0 .
k k
To obtain the principal normal n we can use the general method based on
the GramSchmidt orthogonalization process
00 h 00 , tit k 0 k2 00 h 00 , 0 i 0
n= =
k 0 k2 00 h 00 , 0 i 0
,
k 00 h 00 , titk
and after this we can compute the binormal as a cross product of t and n
b = t n.
0 00
b= .
k 0 00 k
( 0 00 ) 0
n=bt= .
k 0 00 )k k 0 k
The formulae obtained for the curvature and the torsion in Theorem 2.5.14
can be rewritten using standard vector operations. The first curvature is
k 0 00 k k 0 00 k
= = ,
v3 v3
2.9. Curves in R3 133
det 00 k 0 k det 00
000 000
h 0 00 , 000 i
= = = .
vk 0 00 k2 k 0 00 k2 k 0 00 k2
Recall that the numerator of this fraction is the determinant of the matrix
the rows of which are 0 , 00 , 000 and geometrically it is the signed volume
of the parallelepiped spanned by 0 , 00 , 000 , where the sign is positive if and
only if ( 0 , 00 , 000 ) is a positively oriented basis.
0 = t
00 = t0 = n
000 = 0 n + n0 = 0 n + (t + b) = 2 t + 0 n + b
(t) (0) =
t3 t2 t3
= t 2 (0) t(0) + (0) + 0 (0) n(0)
6 2 6
t3
+ (0) (0) b(0) + o(t3 ).
6
134 2. Curves in En
Figure 2.14
Looking at this expansion we conclude that the projection of the curve on the
osculating plane is well approximated by the parabola tt(0) + ((0)t2 /2)n(0).
Observe that the curvature of this parabola at t = 0 is (0).
The projection onto the normal plane has approximately the same shape as
the semicubical parabola ((0)t2 /2)n(0) + ((0) (0)t3 /6)n(0), in particular,
it has a so called cusp singularity at t = 0.
Finally, the projection onto the rectifying plane has the Taylor expansion
t3
(0) + t + o(t) t(0) + (0) (0) + o(t3 ) b(0).
6
so its shape is like the graph of a cubic function. It is easy to see, that the
curvature of this projection is 0 at t = 0, thus it is almost straight around
the origin. That is the reason why the rectifying plane was given just this
name: projection of the curve onto the rectifying plane straightens the curve
and rectifying means straightening.
Exercise 2.9.1. Given a unit speed curve of general type in R3 with distin-
guished Frenet frame t1 , t2 , t3 , find a vector field along the curve such that
t0i = ti holds for i = 1, 2, 3. ( is called the Darboux vector field of the
curve.)
2.9. Curves in R3 135
Exercise 2.9.2. Suppose that the osculating planes of a curve of general type
in R3 have a point in common. Show that the curve is a plane curve. (In
particular, if a point is moving in a central force field, its motion is planar.)
Exercise 2.9.4. Show that the rectifying planes of a curve of general type
in R3 can never share a point in common.
Definition 2.9.5. The unit tangent vector field of a regular space curve
: I R3 is the vector field t = 0 /v, where v = k 0 k. The curvature of is
the function
kt0 k k 0 00 k
= = .
v v3
The total curvature of is the integral
Z
v(t)(t)dt.
I
We saw that the total curvature of a smooth closed plane curve can take only
isolated values, the integer multiples of 2. This is related to the fact that the
curvature of a planar curve is a signed function. When we try to increase the
total curvature of a plane curve making it wavier by continuously deforming it
within the class of closed regular plane curves, we shall not succeed, because
the total curvature of arcs where the curvature is negative keeps equilibrium
with the total curvature of the arcs where the curvature is negative.
In contrast to the planar case, curvature of a regular space curve is nonnega-
tive even if the curve is contained in a plane. In the latter case, the (spatial)
curvature of the curve is the absolute value of the signed planar curvature.
For this reason, the total curvature of a regular space curve can be increased
continuously and can be made arbitrarily large by making the curve wavier.
Hence for regular closed smooth space curves, the interesting question is, how
small the total curvature can be. This question is answered by a theorem of
Moritz Werner Fenchel (1905-1988).
136 2. Curves in En
Figure 2.15
Assume now that P and Q are not antipodal. Then let F be the midpoint of
the shortest great circle arc connecting them. 0F is the bisector of the angle
P 0Q and reflecting P in the line 0F gives Q.
2.9. Curves in R3 137
The surface area of S2 is 4, and the value of the integral on the right-hand
side must be less than 16, the subset of those points of S2 at which the value
of m is at most 3 must have positive measure. In particular, we can find a
vector u S2 such that the dot product hu, t(t)i vanishes at no more than 3
points.
We can think of u as a vertical vector pointing upward, and hu, (t)i as the
height of the point (t). The height of (t) increases strictly on intervals
where hu, t(t)i > 0 and decreases strictly where hu, t(t)i < 0. The roots of
hu, ti divide the image of into at most 3 simple arcs along which the height
changes in a strictly monotonous way. There must exist both an arc along
which the height increases and another one where it decreases, otherwise the
curve could not return to its initial point. If there are three arcs along which
140 2. Curves in En
the height is a strictly monotone function then two of the arcs can be joined so
that the height will strictly monotone also on the union. We can summarize
the information we got as follows. The curve traced out by can be split
into two simple arcs joining the unique highest and the unique lowest point
of the trace. As we move along any of the two arcs the height changes in a
strictly monotonous way.
Take any horizontal plane between the highest and lowest points of the trace
of . Such a plane intersects both arcs at a unique point. Connect these two
points by a segment. The union of all such horizontal segments connecting
two points of im fill a topologically embedded disc in R3 , the boundary of
which is the trace of . We conclude that is a trivial knot.
Chapter 3
Hypersurfaces in Rn
141
142 3. Hypersurfaces in Rn
TM
p
r
rn N 1 M
1
p
r
Rn
Rn1
Figure 3.1
r1 (u1 , u2 ) r2 (u1 , u2 )
N(u1 , u2 ) = .
|r1 (u1 , u2 ) r2 (u1 , u2 )|
3.1. General Theory 143
and
n1
X n1
X n1
X
v 0 t1 + v 2 1 t2 = 00 = u0i u0j rij (u) + u00i ri (u).
i=1 j=1 i=1
144 3. Hypersurfaces in Rn
Multiplying the last equation by the normal vector of the hypersurface and
using the fact that it is orthogonal to the tangent vectors t1 , r1 , . . . , rn1 , we
obtain
n1
X n1
X
v 2 1 hN(u), t2 i = hN(u), 00 i = hN(u), rij (u)iu0i u0j ,
i=1 j=1
from which
Pn1 Pn1 0 0
1 i=1 j=1 hN(u), rij (u)iui uj
1 = . (3.1)
hN(u), t2 i v2
Let us study this expression. We claim that the right hand side is determined
by the osculating plane of the curve and the hypersurface provided that the
osculating plane is not tangent to the surface.
Let us start with the expression
Pn1 Pn1 0 0
0 i=1 j=1 hN(u), rij (u)iui uj
k( ) = .
v2
Since the quantities hN(u), rij (u)i are determined by the parameterization of
the hypersurface, the functions u01 , . . . , u0n1 are the components of the speed
vector 0 of the curve with respect to the basis r1 , . . . , rn1 of the tangent
space, v is the length of the speed vector 0 , k( 0 ) depends only on the speed
vector 0 of the curve (that justifies the notation k( 0 )).
Definition 3.1.4. Let v 6= 0 be an arbitrary tangent vector of the regular
parameterized hypersurface r : Rn at r(u). The intersection curve of the
hypersurface and the plane through r(u) spanned by the direction vectors
N(u) and v is called the normal section of the hypersurface in the direction
v. Assume that the normal section is parameterized in such a way that the
speed vector of the parameterization at r(u) is v. (Such a parameterization
always exists in a neighborhood of r(u).) Orienting the cutting normal plane
by the ordered basis (v, N(u)), we may consider the signed curvature of the
normal section, which will be called the normal curvature of the hypersurface
in the direction v and will be denoted by k(v).
Applying (3.1) for normal sections one may see easily that the normal cur-
vature of a parameterized hypersurface in the direction v = v1 r1 (u) + +
vn1 rn1 (u) is just
Pn1 Pn1
i=1 j=1 hN(u), rij (u))vi vj
k(v) = , (3.2)
v2
where v = kvk. Since k(v) = k(v) for any 6= 0, the normal curvature
depends only on the direction of v.
3.1. General Theory 145
For example, the mappings r1 , . . . , rn1 and N are vector fields along the
hypersurface, and the first n 1 of them are tangential.
Given a vector field along a hypersurface, we would like to express the speed
of change of the vector field as we move along the surface, in terms of the
speed of our motion. This is achieved by the following.
n1
X
0
(X u) (0) = u0i (0)Xi (u(0)),
i=1
146 3. Hypersurfaces in Rn
where (u01 (0), . . . , u0n1 (0)) are the components of v in the basis r1 (u0 ), . . . ,
rn1 (u0 ) of the tangent space at r(u0 ), we have the following formula
n1
X
v X = vi Xi (u0 ),
i=1
where v1 , . . . , vn1 are the components of the vector v in the basis r1 (u0 ), . . . ,
rn1 (u0 ). This formula shows that the definition of v X is correct, i.e.
independent of the choice of the curve u(t).
The way in which a hypersurface curves around in Rn is closely related to
the way the normal direction changes as we move from point to point.
Lemma 3.1.8. The derivative v N of the unit normal vector field of a hy-
persurface with respect to a tangent vector v at p = r(u) is tangent to the
hypersurface at r(u).
Proof. We have to show that v N is orthogonal to N(u). Indeed, differenti-
ating the relation 1 hN, Ni, we get
0 = hv N, Ni + hN, v Ni = 2hv N, Ni.
Definition 3.1.9. Let us denote by M the parameterized hypersurface r :
Rn and by Tp M the linear space of its tangent vectors at p = r(u0 ). The
linear map
Lp : Tp M Tp M, Lp (v) = v N
is called the Weingarten map or shape operator of M at p.
We define two bilinear forms on each tangent space of the hypersurface.
Definition 3.1.10. Let M be a parameterized hypersurface r : Rn , u0
, Tp M the linear space of tangent vectors of M at p = r(u), Lp : Tp M
Tp M the Weingarten map. The first fundamental form of the hypersurface
is the bilinear function Ip on Tp M obtained by restriction of the dot product
onto Tp M
Ip (v, w) = hv, wi for v, w Tp M.
The second fundamental form of the hypersurface is the bilinear function IIp
on Tp M defined by the equality
IIp (v, w) = hLp v, wi for v, w Tp M.
Proof. We know that the normal vector field N is perpendicular to any tan-
gential vector field, thus
hN, rj i 0.
Differentiating this identity with respect to the i-th parameter we get
from which
hN, rji i hri N, rj i = hLr ri , rj i.
Proof of Theorem 3.1.11. It is enough to prove that the matrix of the second
fundamental form with respect to the basis r1 (u0 ), . . . , rn1 (u0 ) is symmetric.
However, this follows from the lemma, since by Youngs theorem rij = rji .
Comparing the identity (3.3) with (3.2) we see that the normal curvature is
the quotient of the quadratic forms of the second and first fundamental forms
Pn1 Pn1
i=1 j=1 hN(u0 ), rij (u0 )ivi vj IIp (v, v)
k(v) = = ,
|v|2 Ip (v, v)
Pn1
where v = i=1 vi ri (u0 ) is a tangent vector of the hypersurface at p = r(u0 ).
The expression
IIp (v, v)
k(v) = ,
Ip (v, v)
gives rise to a linear algebraic investigation of the normal curvature. It is
natural to ask at which directions the normal curvature attains its extrema.
Since k(v) = k(v) for any 6= 0, it is enough to consider this question for
the restriction of k onto the unit sphere S in the tangent space. The unit
sphere of a Euclidean space is a compact (i.e., closed and bounded) subset,
thus, by Weierstrass theorem, any continuous function defined on it attains
its maximum and minimum.
148 3. Hypersurfaces in Rn
d
hL((t)), (t)it=0 = hL( 0 (0)), (0)i + hL((0)), 0 (0)i
dt
= hL 0 (0), vi + hLv, 0 (0)i = 2hLv, 0 (0)i.
The determinant and trace of the Weingarten map, that is the product and
sum of the principal curvatures are of particular importance in differential
geometry.
Definition 3.1.17. For M a hypersurface, p M , the determinant K(p) of
the Weingarten map Lp is called the Gaussian or Gauss-Kronecker curvature
of M at p, H(p) = tr(Lp )/(n 1) is called the mean curvature or Minkowski
curvature.
When we want to compute the principal curvatures and directions of a hy-
persurface at a point we generally work with a matrix representation of
the Weingarten map. Dealing with a parameterized hypersurface r :
Rn , it is natural to take matrix representations with respect to the basis
r1 (u), . . . , rn1 (u) of the tangent space at r(u). Let us denote by G =
(gij )1i,jn1 , B = (bij )1i,jn1 and L = (lji )1i,jn1 the matrix rep-
resentations of the first and second fundamental forms and the Weingarten
map respectively, with respect to this basis (gij , bij and lji are functions on the
parameter domain, i is the row index, j is the column index). Components
of G and B can be calculated according to the equations
gij = hri , rj i,
bij = hN, rij i (cf. Lemma 3.1.12).
150 3. Hypersurfaces in Rn
n1
X
Lr (rj ) = lji ri i = 1, 2, . . . , n 1.
i=1
The relationship between the matrices G, B, and L follows from the following
equalities
D n1
X E n1
X n1
X
bji = bij = hLr ri , rj i = lik rk , rj = lik hrk , rj i = gjk lik
k=1 k=1 k=1
L = G 1 B
Proof. Denote by (u) the common value of the principal curvatures at r(u).
First we show that is constant. For any pair of indices 1 i < j n 1,
we have
Ni = ri and Nj = rj ,
152 3. Hypersurfaces in Rn
since ri and rj are principal directions as any tangent vector is. Differentiat-
ing the first equation with respect to the jth variable, and the second with
respect to the ith, we get
pi = ri + (1/)Ni = ri (1/)ri = 0
for any i.
Now to show that the surface lies on a hypersphere centered at p we have to
prove that the function kr pk is constant. This is clear as
1
1
kr(u) pk =
N(u)
=
is constant. The theorem is proved.
linear combination of the frame vectors. For this we have to determine the
coefficients kij , ij , jk , j in the expressions
n1
X n1
X
rij = kij rk + ij N, Nj = jk rk + j N. (3.4)
k=1 k=1
Let us begin with the simple observation that since Nj is known to be tan-
gential, and Nj = Lr (rj ), where Lr is the Weingarten map,
j = 0 for all j,
n1
and (jk )j,k=1 is the matrix L = G 1 B of the Weingarten map with respect
to the basis r1 , . . . , rn1 . Denote by gij and bij the entries of the first and
second fundamental forms as usual, and denote by g ij the components of
the inverse matrix of the matrix of the first fundamental form. (Attention!
Entries of G and G 1 are distinguished by the position of indices.) Then
n1
X
jk = ljk = g ki bij .
i=1
Taking the dot product of the first equation of (3.4) with N we gain the
equality hrij , Ni = ij and since hrij , Ni = bij ,
ij = bij for all i, j.
There is only one question left: What are the coefficients kij equal to? Let
us take the dot product of the first equation of (3.4) with rl
n1
X n1
X
hrij , rl i = kij hrk , rl i = kij gkl ,
k=1 k=1
Now let us determine the Christoffel symbols of the second type. Differen-
tiating the equality gij = hri , rj i with respect to the k-th variable and then
permuting the role of the indices i, j, k we get the equalities
Solving this linear system of equations for the secondary Christoffel symbols
standing on the right-hand side, we obtain
1
ijk = hrij , rk i = (gik,j + gjk,i gij,k )
2
and
n1 n1
X X 1
kij = ijl g lk
= (gil,j + gjl,i gij,l )g lk . (3.5)
2
l=1 l=1
Observe that the Christoffel symbols depend only on the first fundamental
form of the hypersurface.
Now we ask the following question. Suppose we are given 2(n 1)2 smooth
functions gij , bij i, j = 1, 2, . . . , n 1 on an open domain of Rn1 . When
can we find a regular parameterized hypersurface r : Rn the fundamental
forms of which are represented by the matrices G = (gij ) and B = (bij ). We
have some obvious restrictions on the functions gij and bij . First, gij = gji ,
bij = bji , and since G is the matrix of a positive definite bilinear form, the
determinants of the corner submatrices (gij )ki,j=1 must be positive for k =
1, . . . , n 1. However, several examples show that these conditions are not
enough to guarantee the existence of a hypersurface. For example, if G is
the identity matrix everywhere, while B = f G for some function on , then
the hypersurface (if exists) consists of umbilics. We know however that if
a surface consists of umbilics, then the principal curvatures are constant, so
although our choice of B and G satisfies all the conditions we have listed so
far, it does not correspond to a hypersurface unless f is constant. So there
must be some further relations between the components of B and G. Our
plan to find some of these correlations is the following. Let us express rijk
and rikj as a linear combination of the Gauss frame vectors. The coefficients
we get are functions of the entries of the first and second fundamental forms.
Since rijk = rikj , the corresponding coefficients in the expressions for these
vectors must be equal and it can be hoped that this way we arrive at further
non-trivial relations between G and B. This was the philosophy, and now let
3.1. General Theory 155
The first (n1)4 equations (we have an equation for all i, j, k, l), are the Gauss
equations for the hypersurface. The second family of (n 1)3 equations are
the CodazziMainardi equations.
Exercise 3.1.26. Express the second order derivatives Nij and Nji as a
linear combination of the Gauss frame vectors. Compare the corresponding
coefficients and prove that their equality follows from the Gauss and Codazzi
Mainardi equations.
The exercise points out that a similar try to derive new relations between G
and B does not lead to really new results. This is no wonder, since the Gauss
and CodazziMainardi equations together with the previously listed obvious
conditions on G and B form a complete system of necessary and sufficient
conditions for the existence of a hypersurface with fundamental forms G and
B.
Theorem 3.1.27 (Fundamental theorem of hypersurfaces). Let Rn be
an open connected and simply connected subset of Rn1 (e.g. an open ball or
cube), and suppose that we are given two smooth (n1)(n1) matrix valued
functions G and B on such that G = (gij ) and B = (bij ) assign to every point
a symmetric matrix, G gives the matrix of a positive definite bilinear form. In
this case, if the functions kij derived from the components of G according to
156 3. Hypersurfaces in Rn
the formulae (3.5) satisfy the Gauss and CodazziMainardi equations, then
there exists a regular parameterized hypersurface r : Rn for which the
matrix representations of the first and second fundamental forms are G and
B respectively. Furthermore, this hypersurface is unique up to rigid motions
of the whole space. In other words, if r1 and r2 are two such hypersurfaces,
then there exists an orientation preserving isometry : Rn Rn for which
r2 = r1 .
Let us denote the expressions standing on the left-hand sides of the Gauss
equations by
n1
X
l
Rkji := lij,k lik,j + (sij lsk sik lsj ).
s=1
Let us observe, that the functions Rimjk can be expressed in terms of the
first fundamental form G.
Corollary 3.1.28 (Theorema Egregium). The Gaussian curvature of a reg-
ular parameterized surface in R3 can be expressed in terms of the first funda-
mental form as follows
R1221
K= .
det G
Theorema Egregium (meaning Remarkable Theorem in Latin) is one of those
theorems of Gauss he was very proud of. The surprising fact is not the actual
form of this formula but the mere existence of a formula that expresses the
Gaussian curvature in terms of the first fundamental form. The geometrical
3.1. General Theory 157
The connection between bendings and the first fundamental form is given in
the following simple proposition.
Proof. Let u : [a, b] , u(t) = (u1 (t), . . . , un1 (t)) be a smooth curve.
Then the length of the curve = r u is equal to
v
Z b uD n1
Z bu X 0
n1
X 0
E
l = k 0 (t)kdt = t ri (u(t))ui (t), rj (u(t))uj (t) dt
a a i=1 j=1
v
u n1
Z bu X 0 0
= t ui (t)uj (t)gij (u(t)) dt,
a i,j=1
where the functions gij = hri , rj i are the entries of the matrix G. Similarly,
the length of = r u can be expressed as
v
u n1
Z bu X 0 0
l = t ui (t)uj (t)gij (u(t)) dt,
a i,j=1
158 3. Hypersurfaces in Rn
where gij = hri , rj i. Comparing these two equations we see at once that if
G = G, then l = l for all smooth curves u : [a, b] .
Conversely, assume that l = l holds for any smooth curve u : [a, b] .
Then, in particular, this holds also for the restriction of the curve u onto the
interval [a, x] for any x (a, b], consequently
v
Z xu u n1X 0 0
t ui (t)uj (t)gij (u(t)) dt
a i,j=1
v
Z x
u n1
uX 0 0
= t ui (t)uj (t)gij (u(t)) dt for all x (a, b].
a i,j=1
or equivalently,
n1 n1
X 0 0 X 0 0
ui (x0 )uj (x0 )gij (u(x0 )) = ui (x0 )uj (x0 )gij (u(x0 ))
i,j=1 i,j=1 (3.6)
for all x0 (a, b).
For a fixed u0 , both sides of this equation is a quadratic polynomial of the co-
ordinates of v. These two quadratic polynomials are equal, so the coefficients
of each monomial are equal. Comparing the coefficients of the monomial v i v j
we obtain the equality
1 1
(gij (u0 ) + gji (u0 )) = (gij (u0 ) + gji (u0 )).
1 + ij 1 + ij
X n1
n1 X n1
X n1 X
R(X, Y ; Z, W ) = Rimjk X i Y m Z j W k .
i=1 m=1 j=1 k=1
T : V V V V R
| {z } | {z }
k l
Tji11...j
...il
k
= T (ej1 , . . . , ejk ; ei1 , . . . , eil ),
which are called the components of the tensor T with respect to the basis
e1 , . . . , en .
Now consider a regular parameterized hypersurface M , r : Rn . A ten-
sor field of type (k, l) over M is a mapping T that assigns to every point
u a tensor of type (k, l) over the tangent space of M at r(u). T (u) is
uniquely determined by its components Tji11...j ...il
k
(u) with respect to the basis
i1 ...il
r1 (u), . . . , rn1 (u). The functions u 7 Tj1 ...jk (u) are called the components
of the tensor field T . T is said to be a smooth tensor field if its components
are smooth.
Examples.
Functions on M are tensor fields of type (0, 0).
Tangential vector fields are tensor fields of type (0, 1) (V is isomorphic
to V in a natural way).
160 3. Hypersurfaces in Rn
Exercise 3.1.36. Show that the curvature tensor satisfies the following sym-
metry relations:
For this reason, the surface volume of r(D) = r(h(D)) does not depend on
the way we parameterize the domain.
162 3. Hypersurfaces in Rn
where (h1 , . . . , hn1 ) are the coordinate functions of h. Taking the wedge
product of these equations and applying Proposition 1.2.31 we obtain
Computing the length of the (n 1)-vectors on the two sides, with the help
of equation (1.9), we get the required equality
p q
det G(u) = | det h0 (u)| det G(h(u)).
3.2 Surfaces in R3
A regular parameterized surface r : R3 ( is an open subset of the plane)
has two principal curvatures 1 (u, v) and 2 (u, v) at each point p = r(u, v)
of the surface. If 1 (u, v) 2 (u, v) then 1 (u, v) is the minimum of normal
curvatures in different directions at p, while 2 (u, v) is the maximum of them.
If 1 (u, v) < 2 (u, v) then the principal directions corresponding to 1 (u, v)
and 2 (u, v) are uniquely defined, however if 1 (u, v) = 2 (u, v) then the
point r(u, v) is umbilical, the normal curvature is constant in all directions
and every direction is principal.
The coordinate lines of this parameterization are the rotations of the gener-
ating curve , and the circles drawn by points of the generator curve as they
rotate about the x-axis. Rotated copies of the generating curve are called
the generators, generatrices or meridians of the surface of revolution, while
the circles are the so-called circles of latitude.
The tangent vectors ru and rv are obtained by partial differentiation
ru (u, v) = (x0 (u), y 0 (u) cos v, y 0 (u) sin v),
rv (u, v) = (0, y(u) sin v, y(u) cos v).
The matrix of the first fundamental form with respect to the basis ru , rv is
02 2
hru , ru i hru , rv i x (u) + y 0 (u) 0
G= = .
hrv , ru i hrv , rv i 0 y 2 (u)
To obtain the matrix of the second fundamental form we need the normal
vector field and the second order partial derivatives of r.
e1 e2 e3
ru rv = det x0 (u) y 0 (u) cos v y 0 (u) sin v
0 y(u) sin v y(u) cos v
= (y 0 (u)y(u), y(u)x0 (u) cos v, y(u)x0 (u) sin v),
ru rv 1
N= =q (y 0 (u), x0 (u) cos v, x0 (u) sin v).
|ru rv |
x0 2 (u) + y 0 2 (u)
The second order partial derivatives of the parameterization are
ruu (u, v) = (x00 (u), y 00 (u) cos v, y 00 (u) sin v),
ruv (u, v) = (0, y 0 (u) sin v, y 0 (u) cos v),
rvv (u, v) = (0, y(u) cos v, y(u) sin v).
The matrix of the second fundamental form is
hN, ruu i hN, ruv i
B=
hN, rvu i hN, rvv i
00
x (u)y 0 (u) y 00 (u)x0 (u)
1 0
=q .
0 x0 (u)y(u)
x0 2 (u) + y 0 2 (u)
The matrix of the Weingarten map with respect to the basis ru , rv is
00
x (u)y 0 (u) y 00 (u)x0 (u)
(x0 2 (u) + y 0 2 (u))3/2 0
L = BG 1 =
0
.
x (u)
0 0 2 0 2 1/2
y(u)(x (u) + y (u))
164 3. Hypersurfaces in Rn
We could have obtained this result in a more geometrical way. For any
point p on the surface, the plane through p and the x-axis is a symmetry
plane of the surface. Thus, reflection of a principal direction of the surface
at p is also a principal direction (with the same principal curvature). The
principal curvatures at p are either equal and then every direction is principal,
or different and then the principal directions are unique. Since a direction is
invariant under a reflection in a plane if and only if it is parallel or orthogonal
to the plane, we may conclude that ru and rv are principal directions of the
surface. Principal curvatures are the curvatures of the normal sections of the
surface in the direction ru , rv .
The normal section of the surface in the direction ru is a meridian of the
surface. Its curvature can be calculated according to the formula known for
plane curves and gives 1 up to sign. The difference in sign is due to the fact
that the unit normal of the surface and the principal normal of the meridian
are opposite to one another.
The plane passing through p perpendicular to the x-axis intersects the surface
in a circle of latitude the tangent of which at p is rv . The curvature of this
1
circle is . The normal curvature 2 = k(rv ) of the surface in the direction
y(u)
rv and the curvature of the circle intersection are related to one another by
Meusniers theorem as follows
1 1
= 2 ,
y(u) cos
where is the angle between the normal of the surface and the principal
normal of the circle. As it is easy to see, is the direction angle of the tangent
to the meridian at p, that is, by elementary calculus tan = y 0 (u)/x0 (u), from
x0 (u)
which cos = q .
x0 2 (u) + y 0 2 (u)
Therefore, we get
x0 (u)
2 (u, v) = q
y(u) x0 2 (u) + y 0 2 (u)
1 1
as before. The equation = 2 has the following consequence.
y(u) cos
3.2. Surfaces in R3 165
1
Corollary 3.2.1. The second principal radius of curvature of a surface
2
of revolution at a given point p is the length of the segment of the normal of
the surface between p and the x-axis intercept.
The tractrix is defined as the involute of the chain curve (t) = (t, cosh t)
touching the chain curve at (0, 1). The length of the chain curve arc between
(0) and (t) is
Z t Z tp Z t
0 2
k ( )kd = 1 + sinh d = cosh d = sinh t.
0 0 0
As we know from the theory of evolutes and involutes, the chain curve is the
evolute of the tractrix, the segment (t)(t) is normal to the tractrix, and its
length is the radius of curvature of the tractrix at (t). This implies from one
1
hand that the first principal curvature of the pseudosphere is 1 = sinh t.
On the other hand, we obtain that the equation of the normal line of the
tractrix at (t) is
y cosh t
= sinh t.
xt
166 3. Hypersurfaces in Rn
The x-intercept is
cosh t
t ,0 .
sinh t
According to the general results on surfaces of revolution, the second principal
radius of curvature of the pseudosphere is the distance between (t) and
(t coth t, 0), i.e.
2 !1/2
cosh2 t sinh2 t
1
1
2 = k(coth t tanh t, (cosh t)
1
)k = +
cosh t sinh t cosh2 t
1/2
1 1 1
= 2 + = .
sinh t cosh t cosh2 t
2 sinh t
gij = hri , rj i = 0 if i 6= j.
Hence, the matrix G of the first fundamental form is diagonal. The matrix L
of the Weingarten map is also diagonal by our assumption, consequently the
matrix B = LG of the second fundamental form is diagonal as well.
X1 = X11 r1 + X12 r2 ,
X2 = X21 r1 + X22 r2
system (or that of the parameterization r). There are three families of
coordinate surfaces. Each family covers the domain U . Coordinate surfaces
from different families intersect one another in coordinate lines. We say that
(or r) defines a triply orthogonal system, if coordinate surfaces from different
families intersect one another orthogonally, or equivalently, if hri , rj i = 0 for
i 6= j.
Theorem 3.2.7 (Dupins theorem). If the curvilinear coordinate system
: U R3 defines a triply orthogonal system on U , then the coordinate
lines are lines of curvature on the coordinate surfaces.
Proof. Let r : = (U ) R3 be the inverse of . We may consider without
loss of generality the coordinate surface parameterized by (u, v) 7 r(u, v, w0 ),
where w0 is constant. It is enough to show that the matrices of the first and
second fundamental forms of the surface with respect to the given parame-
terization are diagonal. The matrix of the first fundamental form is diagonal
by our assumption hr1 , r2 i = 0. The nondiagonal element of the matrix of
the second fundamental form is hr12 , Ni, where N is the unit normal of the
surface. Since r3 is parallel to N, hr12 , Ni = 0 will follow from hr12 , r3 i = 0.
Differentiating the equation hr1 , r3 i = 0 with respect to the second variable
yields
hr12 , r3 i + hr1 , r23 i = 0
and similarly, we also have
Solving this system of linear equations for the unknown quantities hr12 , r3 i,
hr23 , r1 i, hr31 , r2 i, we see that they are all zero.
The canonical equation of an ellipsoid has the form
x2 y2 z2
+ + = 1.
A B C
Suppose A > B > C. We can embed this surface into a triply orthogonal
system of second order surfaces as follows. Consider the surface
x2 y2 z2
F : + + = 1.
A+ B+ C +
F is
an ellipsoid for > C;
a one sheeted hyperboloid for C > > B;
3.2. Surfaces in R3 171
P () = (A + )(B + )(C + )
x2 (B + )(C + ) + y 2 (A + )(C + ) + z 2 (A + )(B + ) = 0.
0 = i (F r) = h(grad F ) r, ri i,
hence grad F (p) is orthogonal to the tangent vectors r1 (u0 ), r2 (u0 ) which
span the tangent plane Tp M .
or equivalently
x2 y2 z2
+ + = 0.
(A + )(A + ) (B + )(B + ) (C + )(C + 0 )
0 0
3.2. Surfaces in R3 173
We know that
x2 y2 z2 x2 y2 z2
+ + = 1 and + + = 1.
A+ B+ C + A + 0 B + 0 C + 0
Subtracting these equalities we obtain ( 0 ) times the equality to prove.
Since 6= 0 , we are ready.
and thus det B = 0 if and only if ruv is orthogonal to N i.e. if ruv is tangential.
Examples.
(1) The following surfaces are ruled but not developable:
r(u, v) = (a cos u, b sin u, 0) + v(a sin u, b cos u, c), (one sheeted hyper-
boloid);
The cone is regular only in the domain v 6= 0. The tangent plane of the
cone is spanned by the vectors
ru (u, v) = v 0 (u) and rv (u, v) = (u) p.
Since
ruv (u, v) = 0 (u) = (1/v)ru (u, v),
cones are developable.
u hrv , rv i = 2hruv , rv i = 0,
0 (u) = a0 (u) + (c0 (u)/c2 (u))b(u) b0 (u)/c(u) = (c0 (u)/c2 (u))b(u) k b(u),
H2 K
Exercise 3.2.17. Show that the Minkowski curvature of the following sur-
faces is 0 everywhere
3.2. Surfaces in R3 179
Proof. The normal curvature can be expressed with the help of the funda-
mental forms
II(t) ( 0 (t), 0 (t)) h 0 (t), L(t) ( 0 (t))i
k(t) ( 0 (t)) = = .
k 0 (t)k2 k 0 (t)k2
Thus, 0 (t) points in asymptotic direction if and only if
h 0 (t), L(t) ( 0 (t))i = 0.
3.2. Surfaces in R3 181
(c) If two asymptotic curves meet at a point p with nonparallel speed vectors
and their torsions are defined at p, then their torsions at p are opposite
to one another.
Proof. (a) Since is an asymptotic curve, both 0 (t) and 00 (t) are in the
tangent plane at (t), consequently, 0 00 = 1 k 0 00 kN and
h 0 00 , 000 i = 1 k 0 00 khN , 000 i.
Where 1 is the sign of h 0 00 , N i.
Differentiating h 00 , N i 0, we obtain
D d E
h 000 , N i = 00 , N = h 00 , L ( 0 )i.
dt
Using the linear algebraic identity
L ( 0 ) L ( 00 ) = (L ( 0 ) L ( 00 )) = (det(L ) 0 00 ) = det(L ) 0 00
and the Lagrange identity we obtain
det(L )k 0 00 k2 = h 0 00 , L ( 0 ) L ( 00 )i
0
h , L ( 0 )i h 00 , L ( 0 )i
= det
h 0 , L ( 00 )i h 00 , L ( 00 )i
= h 00 , L ( 0 )i2 .
182 3. Hypersurfaces in Rn
where = 1 2 .
If 0 (t) 00 (t) = 0 at a point, then equation (3.8) holds with any choice of
(t), as both sides are equal to 0. Assume that 0 (t) 00 (t) 6= 0. Then 00 (t)
can be decomposed into the orthogonal components
If there is a surface the fundamental forms of which have these special forms,
then it is certain, that the coordinate lines are asymptotic lines and the
surface has constant 1 Gaussian curvature. However, we know from the
fundamental theorem of curve theory, that a surface with given fundamental
184 3. Hypersurfaces in Rn
forms exist (locally) if and only if the prescribed matrices satisfy the Gauss
and CodazziMainardi equations. These equations are not automatically
satisfied for the above matrices leading to some partial differential equations
that must be satisfied by the functions A, B and . Let us compute these
equations.
First we observe that the inverse of G is
1 cos
A2 sin2
1
AB sin2
G = .
cos 1
AB sin2 B 2 sin2
Lemma 3.2.23. If a regular parameterized surface of negative curvature has
the property that the coordinate lines are asymptotic lines and buv > 0, then
the CodazziMainardi equations reduce to the equations
u (log buv ) = vuv uuu , (3.11)
v (log buv ) = vvv uvu . (3.12)
Proof. The general form of the CodazziMainardi equation is
2
X
bij,k bik,j = (lik blj lij blk ).
l=1
Since both sides are skew-symmetric in the indices j and k, these equations
are true trivially when j = k and the equation for i, j, k is equivalent to the
equation for (i0 , j 0 , k 0 ) = (i, k, j). For this reason, when we have only two
variables u and v then there are only two independent CodazziMainardi
equations for (i, j, k) = (1, 1, 2) and (i, j, k) = (2, 1, 2). The equations for
these indices are
buu,v buv,u = (uuv buu uuu buv ) + (vuv bvu vuu bvv ),
bvu,v bvv,u = (uvv buu uvu buv ) + (vvv bvu vvu bvv ).
Taking into consideration buu = bvv = 0, these equations reduce to
buv,u = uuu buv + vuv bvu ,
bvu,v = uvu buv + vvv bvu .
Dividing these equations by buv = bvu we obtain the required equalities for
the logarithmic derivative of buv
Lemma 3.2.24. The CodazziMainardi equations are satisfied by the matri-
ces (3.9) and (3.10) if and only if
v A = u B = 0.
3.2. Surfaces in R3 185
Let us compute now what the Gauss equation gives for a surface with these
fundamental form matrices. The inverse of G is
1 cos
sin2
sin2
G 1 = ,
cos 1
2 2
sin sin
the Christoffel symbols of the second kind are
1
uuu = guu,u = 0,
2
1
uuv = (2guv,u guu,v ) = sin u ,
2
1
uvu = (guv,u + guu,v guv,u ) = 0,
2
1
uvv = (guv,v + gvv,u guv,v ) = 0,
2
1
vvu = (2guv,v gvv,u ) = sin v ,
2
1
vvv = gvv,v = 0,
2
3.2. Surfaces in R3 187
cos2 u
sin2 = uv = uv sin ,
sin
hence satisfies the partial differential equation uv = sin .
Definition 3.2.27. The partial differential equation uv = sin is called
the sine-Gordon equation.
We can sum up our results as follows.
Theorem 3.2.28. A surface of constant Gaussian curvature 1 can be
parameterized in such a way that the coordinate lines are the asymptotic
lines parameterized by arc length. For such a parameterization the angle
: (0, ) between the asymptotic lines gives a solution of the sine-Gordon
equation.
188 3. Hypersurfaces in Rn
This surface has constant Gaussian curvature 1 and the coordinate lines of
the parametrization are unit speed asymptotic curves.
Roughly speaking, surfaces with constant Gaussian curvature 1 are in one
to one correspondence with solutions of the sine-Gordon equation with values
in the interval (0, ).
The integrated form of the sine-Gordon equation is the Hazzidakis formula.
Theorem 3.2.29 (Hazzidakis Formula). Let r : R3 be a surface of
constant Gaussian curvature 1 parameterized as above. Then the surface
area A of the curvilinear quadrangle bounded by asymptotic lines obtained as
the r image of the rectangle [u0 , u1 ] [v0 , v1 ] is
Proof. Suppose to the contrary that such a surface exists. Then by com-
pleteness, for any p r() both maximal asymptotic curves starting at
p have a parameterization by arc length defined on the whole real line. Let
p : R r() be a unit speed parameterization of one of the asymptotic lines
starting at (0) = p, and for each t R, denote by p (t) : R r() the unit
speed parameterization of the asymptotic line starting at p (t) (0) = p (t).
Orient the initial speed of p (t) so that 0 (t) 0 p (t) (0) be a positive mul-
tiple of the normal vector Np (t) for all t. Then the map r : R2 R3 ,
r(u, v) = p (u) (v) is a regular parameterized surface with constant Gaussian
curvature 1, for which the coordinate lines are naturally parameterized
asymptotic curves. (We do not claim that r is injective, nor that it has the
same image as r.) This map r would define a solution of the sine-Gordon
equation which would take values in the interval (0, ). However, as the next
lemma states, there are no such solutions.
Proof. Assume to the contrary that such a solution exists. We discuss three
cases separately. In the first case, suppose that (1, 0) > (0, 0). Set c =
((1, 0) (0, 0))/3 and let
Then it is clear that 0 < u1 < u2 < 1 and for any t [u1 , u2 ], we have
and
(t, v) < (1, v) + (t, 0) (1, 0) (1, v) c < c.
190 3. Hypersurfaces in Rn
Backlund transform
A Backlund transform in contemporary mathematics is a method to find
new solutions to certain partial differential equations if one particular solu-
tion is already known. Backlund transforms have their origins in differential
geometry. Namely, in the 1880s L. Bianchi and A.V. Backlund introduced
a nontrivial geometrical construction of new surfaces of constant negative
Gaussian curvature from an initial such surface using a solution of a first or-
der partial differential equation. As surfaces of constant negative curvature
can be described by solutions of the sine-Gordon equation, Backlund trans-
3.2. Surfaces in R3 191
Computing the Christoffel symbols from G we can express the second order
partial derivatives of r as linear combinations of the Gauss frame. We omit
the details of this mechanical computation, and give just the result
From the first two conditions, a Backlund transform of r must be of the form
ru rv
r = r + cos cos + sin ,
cos sin
(u0 , v0 ) with any initial condition (u0 , v0 ) = 0 if and only if the equa-
tions do not contradict the Young theorem uv = vu . As
(u )v = vv
cos cos v sin sin v sin ( sin sin v + cos cos v )
+ ,
cos
(v )u = uu
cos cos u + sin sin u sin ( sin sin u + cos cos u )
+ ,
cos
Theorem 3.2.35. For any choice of the point (u0 , v0 ) , the parameter
[/2, /2], a vector v of length cos tangent to r at (u0 , v0 ), there is
a unique Backlund transform r of r, defined in a neighborhood of (u0 , v0 ),
which shifts r(u0 , v0 ) to r(u0 , v0 ) = r(u0 , v0 ) + v.
the parameter lines of r are the lines of curvature of the Backlund trans-
form.
194 3. Hypersurfaces in Rn
This shows that singular points of the Backlund transform are the points
where is an integer multiple of /2, which means that r r is parallel to
one of the principal directions ru , rv .
Another use of the knowledge of G is that we can compute the Gaussian
curvature from it by the Theorema Egregium. It gives that the Gaussian
curvature of r is
vv uu
K = .
sin cos
K 1 is equivalent to the variant of the sine-Gordon equation uu vv =
sin cos . We saw that the sine-Gordon equation for was the integrability
condition for the partial differential equation (3.16) for the unknown . The
functions and enter into (3.16) in a symmetric way. Since is obtained
as a solution of (3.16) for a given , (3.16) does have a solution in given .
Thus satisfies the integrability condition uu vv = sin cos .
To compute the matrix of the second fundamental form of r we first compute
partial derivative of N with respect to v
cos sin
Nv = cos (v + u )ru (v + u )rv cos cos N
cos sin
+ sin Nv
From this, again some simple but tedious computation shows hru , Nv i = 0,
which means that for the parametrization r of the Backlund transform the
coordinate lines are lines of curvature.
With some more work we can compute also hrv , Nv i = sin cos . Since
the B is symmetric and its determinant is K det G = cos2 sin2 , B must
be of the form
sin cos 0
B = .
0 sin cos
This gives at once that ru rv point in asymptotic directions, from which
the rest of the theorem follows.
3.2. Surfaces in R3 195
where Gt is the Gram matrix of the partial derivatives rt,1 , . . . , rt,n1 . Dif-
ferentiating with respect to t at t = 0, we obtain
Z Z
0 dp d
Vn1 (0) = det Gt |t=0 dn1 = krt,1 rt,n1 kt=0 dn1 .
K dt K dt
(3.17)
Applying the chain rule, the Leibniz rule for the differentiation of dot and
wedge products, and Youngs theorem, we get
d
d dt rt,1 rt,n1 t=0 , r1 rn1
krt,1 rt,n1 k t=0 =
dt kr1 rn1 k
DP E
n1
i=1 r 1 r i1 t i R(., 0) r i+1 rn1 , r1 rn1
=
kr1 rn1 k
DP E
n1
i=1 r1 ri1 Xi ri+1 rn1 , r1 rn1
= ,
kr1 rn1 k
and
0
Vn1 (0)
DP E
n1
i=1 r1 ri1 Xi ri+1 rn1 , r1 rn1
Z
= dn1 .
K kr1 rn1 k
(3.18)
Decompose X into the sum of a vector field X > tangential to the hyper-
surface r and a vector field X orthogonal to it. Obviously, the orthogonal
component is X = hX, NiN. The key observation for the proof of the theo-
rem is that the tangential component X > has no contribution to the integral
in (3.18), that is,
DP E
n1 >
Z
i=1 r1 ri1 X i r i+1 rn1 , r1 rn1
dn1
K kr1 rn1 k
= 0.
Manifolds
201
202 4. Manifolds
Proof. The first part of the statement follows directly from the following set
theoretical identities.
(i) ( dom ) = ,
(ii) (X dom ) = im ,
Examples.
Rn equipped with the atlas consisting of only one chart, the identity mapping
of Rn , is an n-dimensional differentiable manifold.
Open subsets U X of an n-dimensional C r -manifold (X, A) become n-
dimensional C r -manifolds with the atlas {dom()U : A}.
G G G, (x, y) 7 x y 1
is differentiable.
Example. GL(n, R) and GL(n, C) are open subsets in the linear spaces of
all n n real or complex matrices, hence they have a differentiable manifold
structure. They also have a group structure, which is smooth since the entries
of the quotient of two matrices are rational functions of the entries of the
original matrices and rational functions are smooth. This way, general linear
groups are Lie groups.
4.1. Topological and Differentiable Manifolds 205
A rigid segment in the plane has for its configuration space the direct
product R2 S 1 , which is homeomorphic to the open solid torus.
4.1.3 Submanifolds of Rn
Theorem 4.1.12. Let F : Rn Rk be a smooth mapping, the image of
which contains 0 Rk . Consider the preimage X = {x Rn : F (x) = 0} of
the point 0.
Let us suppose that the gradient vectors
Proof. The linear space Rnk can be embedded into Rn by the mapping
Exercise 4.1.13. Check that the topology of (X, A) coincides with the sub-
space topology inherited from Rn , consequently, it is Hausdorff and second
countable.
Most important Lie groups are obtained as closed subgroups of GL(n, R),
defined by some equalities on the matrix entries. For example, GL(n, C) is
isomorphic
to the subgroup of GL(2n, R) consisting of matrices of the form
A B
, where A, B are nn matrices. The following example illustrates
B A
how we can prove that a closed subgroup of GL(n, R) is a Lie group. (Remark
that according to Cartans theorem, every closed subgroup of a Lie group is
a Lie group, therefore the computation below is an elementary verification of
a special case of Cartans theorem.)
Consider the orthogonal group
Since Fij = Fji , these equations are not independent. We get however an
independent system of equations if we restrict ourselves to the equations with
1 i j n. The number of these equations is n(n+1) 2 , so if we show the
independence of the gradients, then we obtain the required expression for the
dimension of O(n).
The gradient of Fij at A O(n) is an n n matrix with entries
n
Fij X aik ajk
= ajk + aik = ajs ir + ais jr .
ars ars ars
k=1
We show that these vectors, are orthogonal with respect to the usual scalar
product on Rn
2
= Mat(n, R). Indeed, taking the scalar product of the gradi-
ent vectors of Fij and Fkl at A we obtain
n
X
(ajs ir + ais jr )(als kr + aks lr )
r,s=1
n
X
= (ajs als ir kr + ajs aks ir lr + ais als jr kr + ais aks jr lr )
r,s=1
X n
= (jl ir kr + jk ir lr + il jr kr + ik jr lr )
r=1
= jl ik + jk il + il jk + ik jl = 2ik jl (1 + ij kl ).
4.1. Topological and Differentiable Manifolds 209
is a Lie group and determine its dimension. Prove that SU (2) is diffeomorphic
to the 3-dimensional sphere S 3 .
Exercise 4.1.22. Which surface shall we get from the classification list if
we glue to the sphere k 1 Mobius bands and l handles?
212 4. Manifolds
Figure 4.3: Self-intersecting images of the Mobius band with planar bound-
ary: the cross cap and half of the Klein bottle.
Figure 4.4: A sphere with one and two Mobius bands: the projective plane
and the Klein bottle.
M = {(z, w) C2 : z l = P (w)}.
which means that the vectors grad fi (x0 ) and 0 (0) are orthogonal.
Now we prove that if a vector v is orthogonal to the vectors grad fi (x0 ),
1 i n k, then v is a tangent vector. Let us take a smooth local
parameterization F : Rk M Rn of M around the point x0 . The curve
: t 7 F (F 1 (x0 ) + ty), where y Rk is fixed, is a curve on M passing
through x0 = (0). The speed vector of this curve at t = 0 is
The definition of tangent vectors can also be given in intrinsic terms, inde-
pendent of the embedding of M into Rn . Let us define an equivalence relation
on the set
(M, p) = { : [, ] M | (0) = p, is smooth},
consisting of curves passing through p M , by calling two curves 1 , 2
(M, p) equivalent if ( 1 )0 (0) = ( 2 )0 (0) for some chart around p.
Then this condition is true for any chart from the atlas of M . (Prove this!)
Definition B. A tangent vector to a manifold M at the point p M is an
equivalence class of curves in (M, p). The set of equivalence classes is called
the tangent space of M at p and is denoted by Tp M .
In this setting, the speed vector 0 (s) T(s) M of a smooth curve is the
equivalence class of the curve t 7 (s + t) in (M, s).
Given a chart around p, we can establish a one-to-one correspondence
between the equivalence classes and points of Rm , (m = dim M ), assigning
to the equivalence class of a curve (M, p) the vector ( )0 (0) Rm .
With the help of this identification, we can introduce a vector space structure
on the tangent space, not depending on the choice of the chart.
For embedded manifolds, Definition B agrees with Definition A. The advan-
tage of Definition B is that it is applicable also for abstract manifolds, not
embedded anywhere.
Definition 4.2.4. If = (x1 , . . . , xm ) is a chart on the manifold M around
the point p, (M, p), then the coordinates (x1 )0 (0), . . . , (xm )0 (0)
of ( )0 (0) are called the components of the speed vector 0 (0) with respect
to the chart .
The main difficulty of defining tangent vectors to a manifold is due to the
fact that an abstract manifold might not be naturally embedded into a fixed
finite dimensional linear space. Nevertheless, there is a universal embedding
of each differentiable manifold into an infinite dimensional linear space.
Let us denote by F(M ) the linear vector space of smooth functions on M ,
and by F (M ) the dual space of F(M ) that is the space of linear functions
on F(M ), and consider the embedding of M into F (M ) defined by the
formula
[(p)](f ) = f (p), where p M, f F(M ).
Having embedded the manifold M into F (M ), we can define tangent vectors
to M as elements of the linear space F (M ).
Definition 4.2.5. Let M be a differentiable manifold, p M . We say that
a linear function D F (M ) defined on F(M ) is a derivation of F(M ) at p
if the equality
D(f g) = D(f )g(p) + f (p)D(g)
4.2. The Tangent Bundle 215
Remarks.
and
gi (0) = i f (0).
Proof. Since
Z 1 Z 1X n
d f (tx)
f (x) f (0) = dt = xi i f (tx)dt
0 dt 0 i=1
n
X Z 1
= xi i f (tx)dt,
i=1 0
R1
we may take gi (x) = 0
i f (tx)dt.
Now we are ready to prove the main theorem.
Theorem 4.2.9. The tangent space to a differentiable manifold M at the
point p M coincides with the space of all derivations of F(M ) at p, which
is a linear space having the same dimension as M has.
Proof. Let us take a differentiable manifold (M, A) and a chart = (x1 , . . . ,
xn ) A defined in a neighborhood of p M . Define the derivations i (p)
as follows
[i (p)](f ) := i (f 1 )((p)).
We prove that the derivations i (p) form a basis in the space of derivations
at p. They are linearly independent since if we have
n
X
i i (p) = 0,
i=1
4.2. The Tangent Bundle 217
To finish the proof, we only have to show that every derivation at the point
p can be obtained as a speed vector of a curve passing through p. Define
the curve : [, ] M by the formula (t) := 1 ((p) + (t1 , . . . , tn )).
Pn
Then obviously the speed vector 0 (0) is just i=1 i i (p).
Proof. We derive a formula for Tp f using local coordinates which will show
both parts of the proposition.
Let = (x1 , . . . , xm ) and = (y 1 , . . . , y n ) be local coordinates in a neigh-
borhood of p M and f (p) N respectively.
If the components of D in the basis i (p) corresponding to the chart are
{i : 1 i m} then we have (xi )0 (0) = i . Observe, that i depends
only on D and the fixed chart , but not on the curve . The components
{ j : 1 j n} of Tp f (D) in the basis j (f (p)) induced by the chart
can be computed by the formula j = (y j f )0 (0). Denote by fj the j-th
coordinate function of the mapping f = f 1 , i.e.
fj = y j f 1 .
Then we have
which shows that Tp f (D) depends only on D but not on and that Tp f
is a linear mapping the matrix of which in the bases (1 (p), . . . , m
(p)) and
(1 (f (p)), . . . , n (f (p))) is
i fj ((p)) 1im .
1jn
Obviously, smooth vector fields over M form a real vector space with respect
to the operations
(f X)(p) := f (p)X(p).
f = : (a, b) Rn
(iii) (Differentiable dependence on the initial point). Let us define the set
Ut M for t R as follows
Ut = {p M : t dom p }.
and Ht1 (Ht2 (p)) = Ht1 +t2 (p). The family {Ht : t R} is called the
one-parameter family of diffeomorphisms or the flow generated by the
vector field X.
A useful consequence of the fundamental theorems is the following statement
on the straightening of vector fields.
Proposition 4.3.3. Let X be an arbitrary vector field on a manifold M
and p M be a point such that X(p) 6= 0. Then there exists a chart =
(x1 , . . . , xn ) : U Rn around p for which X|U = 1 .
This means that the derivative of the mapping turns the vector field X into
a constant vector field on Rn .
Proof. Let Ht be the flow generated by the vector field X. Choose a chart
: V Rn around p such that (p) = 0. Then the image of contains a
ball Br of radius r about the origin. Decompose X as a linear combination
X = X 1 1 + + X n n . Since X(p) 6= 0, one of the coefficients X i (p)
should also be different from 0. We may assume that X 1 (p) 6= 0 as this can
be reached by permuting the coordinates.
Let Brn1 Br be the (n 1)-dimensional ball consisting of those point of
Br the first coordinate of which is 0, and let j : Brn1 M be the inverse
restricted onto Brn1 . The set W of those points q M for which the
maximal integral curve starting at q is defined on a fixed open interval (, )
is open in M and these open sets for all positive > 0 cover M . Since j(Brn1 )
is compact, it is covered by a finite number of such sets Wi . Thus, taking
for > 0 the smallest of the i s, we obtain a positive number such that all
maximal integral curves of X starting at a point of j(Brn1 ) is defined on the
interval (, ).
Consider the smooth map F : (, ) Brn1 M which is defined by the
formula
F (x1 , . . . , xn ) = Hx1 (j(0, x2 , . . . , xn )).
The partial derivatives of F at the origin are 1 F (0) = X(p), and i F (0) =
i (p) if i 2. These vectors are linearly independent, therefore, by the
inverse function theorem, there is an open neighborhood U of p and W
(, )Brn1 of the origin in Rn such that F |W is a diffeomorphism between
W and U . In other words, F |W is a local parameterization of M and its
inverse is a local coordinate system on M . We claim that has the required
property. Indeed, if q U is an arbitrary point, and (q) = (q1 , . . . , qn ), then
the parameterized coordinate line (, ) M , t 7 F (t, q2 , . . . , qn ) is an
integral curve of X, therefore its speed vector at t = q1 is 1 (q) = X(q).
Since tangent vectors to a manifold at a point are identified with deriva-
tions at the point, vector fields can be regarded as differential operators
222 4. Manifolds
(i) [i , j ] = 0;
n
X X n
n X
[X, Y ] = i
(X(Y ) Y (X i
))i = X j j Y i Y j j X i i .
i=1 i=1 j=1
Proof. (i) The first part of the proposition is equivalent to Youngs theorem
since for any smooth function f on the domain of , we have
i (j (f )) = i (j (f 1 ) ) = i (j (f 1 ))
= j (i (f 1 )) = j (i (f )).
Suppose that we are given two vector fields X and Y on an open subset
of Rn . The corresponding flows Hs and Gt do not commute in general:
Hs Gt 6= Gt Hs .
To measure the lack of commutation of the flows Hs and Gt , we consider the
difference
(s, t; p) = Gt Hs (p) Hs Gt (p)
for a fixed point p Rn . is a differentiable function of s and t and it is 0 if
t or s is zero. This means, that in the Taylor expansion of around (0, 0; p)
(s, t; p) = (0, 0; p) + s (0, 0; p) + t (0, 0; p) +
s t
2 2
2 t2 2
s
+ (0, 0; p) + st (0, 0; p) + (0, 0; p) + o(s2 + t2 )
2 s2 st 2 t2
2
the only non-zero partial derivative is (0, 0; p).
st
Proposition 4.3.9. Through the natural identification of the tangent space
2
of Rn at p with the vectors of Rn , the vector (0, 0; p) corresponds to the
st
tangent vector [X, Y ](p).
Pn Pn
Proof. Put X = i=1 X i i , Y = i=1 Y i i . Let us compute first the vector
2
st Hs Gt (p) at s = t = 0. Since Hs is the flow of X, for any point q, the
curve q (s) = Hs (q) is an integral curve of X. Thus, we have
Hs Gt (p) (0, t) = G (p) (s) (0, t) = X(Gt (p)).
s s t
Differentiating with respect to t,
n
2 d X
Y (X i )i (p)
Hs Gt (p)s=t=0 = X(Gt (p))t=0 =
st dt i=1
2
(s, t; p) = st (0, 0, ; p) + o(s2 + t2 ) = st[X, Y ](p) + o(s2 + t2 )
st
In particular, we obtain the following expression for [X, Y ](p).
Gt Ht (p) Ht Gt (p)
[X, Y ](p) = lim
t0 t2
Definition 4.3.10. We say that two vector fields are commuting if their Lie
bracket is the zero vector field.
Proof. If both X and Y vanish at p then Hs (p) = Gt (p) = p for any s and t
and thus the assertion holds trivially. We may thus suppose that one of the
vectors X(p), Y (p), say X(p) is not zero. By the theorem on the straightening
of vector fields we may suppose that the manifold is an open subset of Rn ,
with coordinates (x1 , . . . ,P
xn ), and the vector field X coincides with the basis
n
vector field 1 . Let Y = i=1 Y i i be the decomposition of Y into a linear
combination of the basis vector fields i . By the formula for the Lie bracket
of vector fields we have
n
X
0 = [X, Y ] = (1 Y i )i 1 Y i = 0 for each i.
i=1
as we wanted to show.
Corollary 4.3.14. When f : M N is a diffeomorphism, then for any
smooth vector field X on M , there is a unique smooth vector field f (X)
4.3. The Lie Algebra of Vector Fields 227
The flow generated by the left invariant vector field X consists of the diffeo-
morphisms
Ht (A) = AeM t ,
that is, Ht is a right translation by eM t .
Now let us take two left invariant vector fields X(A) = AM and Y (A) = AN
and consider the flows Ht and Gt generated by them.
Computing Gt Ht (A) Ht Gt (A) up to o(t2 ), we get
We obtain, that the Lie algebra of GL(n, R) is isomorphic to the Lie algebra
of all matrices with Lie bracket [M, N ] = M N N M .
X(x1 , x2 ) = x1 1 + x2 2
Y (x1 , x2 ) = x2 1 x1 2 ,
compute their Lie bracket, and determine the flows generated by them.
Exercise 4.3.17. Show that the Lie algebra of SO(n) is isomorphic to the Lie
algebra of skew-symmetric nn matrices with Lie bracket [X, Y ] = XY Y X.
Exercise 4.3.20. Show that the Lie algebra of left invariant vector fields on
a Lie group is isomorphic to the Lie algebra of right invariant vector fields.
4.3. The Lie Algebra of Vector Fields 229
The theorem of Frobenius claims that the converse of this proposition is also
true.
Proof. Observe first that the properties imposed on the chart (x1 , . . . , xn ) are
equivalent to the condition that the subspace Dq is spanned for all q U by
the vectors i |q for i = 1, . . . , k. This means geometrically that any involutive
distribution can be straightened locally, that is locally diffeomorphic to a
subspace distribution on Rn assigning to each point a translate of a given
subspace.
We prove the theorem by induction on the dimension of the distribution.
If k = 0, then the statement is obvious. Assume that k 1 and that the
theorem is proved for (k 1)-dimensional distributions. In order to construct
a suitable chart around p, choose an open set V around p, over which the
distribution can be spanned by the pointwise linearly independent smooth
vector fields X1 , . . . , Xk .
By Proposition 4.3.3, there is a a local coordinate system = (y 1 , . . . , y n )
defined on an open neighborhood W V of p for which X1 |W = 1 . Denote
by W the (n 1)-dimensional submanifold defined by the equation
y 1 = 0.
Define the vector fields Y1 , . . . , Yk on W by the equations Y1 = X1 , Yi =
Xi Xi (y 1 )X1 , (i = 2, . . . , k). It is clear that Y1 , . . . , Yk span the same
subspace at each point of W as X1 , . . . , Xk . For i 2, we also have
Yi (y 1 ) = Xi (y 1 ) Xi (y 1 )X1 (y 1 ) = 0,
which means that the vector field Yi is tangent to the level sets of the first
coordinate function y 1 , in particular, it is tangent to .
Let Zi be the restriction of the vector field Yi onto . We claim that the
distribution D0 spanned by the vector fields Z2 , . . . , Zk on is involutive.
Indeed, as D is involutive, the Lie bracket [Yr , Ys ] of the vector fields Yr and
Ys is tangent to D hence can be written as
k
X
[Yr , Ys ] = cirs Yi (4.1)
i=1
4.3. The Lie Algebra of Vector Fields 231
where the coefficients cirs are smooth functions on W . Evaluating both sides of
this equation on the function y 1 for r, s 2 we obtain 0 = c1rs . Thus restrict-
Pk
ing the equation (4.1) onto gives the decomposition [Zr , Zs ] = i=2 cirs Zi
which implies easily that D0 is involutive.
Applying the induction hypothesis for the distribution D0 we obtain a local
coordinate system = (z 2 , . . . , z n ) on defined around p, such that D0 is
spanned by the vector fields i , i = 2, . . . , k. (For convenience, we shifted
the indices by 1. This means that i denotes derivation with respect to
the (i 1)st coordinate z i of .) Suppose that the transit map between the
coordinates (y 2 , . . . , y n )| and (z 2 , . . . , z n ) is given by the smooth functions
f i in such a way that
z i = f i (y 2 , . . . , y n )| , (i = 2, . . . , n). (4.2)
x1 = y 1 , (4.3)
i i 2 n
x = f (y , . . . , y ), (i = 2, . . . , n). (4.4)
k
X
1 (Yi (xr )) = Y1 (Yi (xr )) = [Y1 , Yi ](xr ) = cs1i Ys (xr ). (4.5)
s=1
Let the projection of the cube K onto the first coordinate axis be the interval
(a, a) and denote by K 0 its orthogonal projection onto the hyperplane x1 =
0. Fix an arbitrary point c = (0, c2 , . . . , cn ) K 0 and denote by qt U the
point the coordinates of which are x1 (qt ) = t, x2 (qt ) = c2 , . . . , xn (qt ) = cn .
Consider the functions hir : (a, a) R, hir (t) = (Yi (xr ))(qt ), (2 i k <
r n). According to the equations (4.5), the functions hir (i = 2, . . . , k)
232 4. Manifolds
where cs1i (t) = cs1i (qt ). We know also that at the point q0 , the vector
Yi (q0 ) Dq0 0 is a linear combination of the vectors s |q0 = s |q0 s = 2, . . . , k.
For this reason, (Yi (xr ))(q0 ) = 0, that is hir (0) = 0 for all 2 i k < r n.
The functions hir 0 also solve the differential equation (4.6) with initial
condition hir (0) = 0, thus, by the uniqueness of solutions with a given initial
condition, hir (t) = (Yi (xr ))(qt ) 0 for all 2 i k < r n. As choosing
the constants c2 , . . . , cn and t (a, a) properly qt can be any point of U ,
Yi (xr ) 0 for all 2 i k < r n, and this is what we wanted to prove.
Examples.
A trivial vector bundle over M is a product space E = M V with
the projection map = M onto the first component, where V is a
k-dimensional linear space over R.
The tangent bundle : T M M of an n-dimensional manifold M is
an n-dimensional vector bundle. It is usually not a trivial bundle.
There are several methods to construct new vector bundles from given ones.
To construct tensor bundles we can apply the following general construction
for the tangent bundle of a manifold.
Consider the category Vk of k-dimensional linear spaces in which the mor-
phisms between two k-dimensional linear spaces are the linear isomorphisms
between them and let be a covariant functor from Vk to Vl which is smooth
in the sense that for any two k-dimensional linear spaces V W , the map
: Mor(V, W ) Mor((V ), (W )) is smooth. Smoothness means that if
we fix some bases in V ,W , (V ) and (W ) to identify linear maps with their
matrices, then the coefficients of the matrix of (L) are smooth functions of
the matrix coefficients of L Mor(V, W ).
If : E M is a k-dimensional smooth vector bundle over M then applying
to each fiber of E, we can construct a new smooth vector bundle over M
with l-dimensional fibers. The total space of this bundle is the disjoint union
[
E = (Fp ),
pM
^
(, h) = ( , I (h)) : ( )1 (U ) Rn Rl
S
from and h on the set ( )1 (U ) = pM (Fp ) in the following way.
The first component of the chart (, ^ h) assigns to each element e
1
in ( ) (U ) the coordinates of the base point (e) with respect to the
chart . These coordinates are the first n coordinates of e. The symbolic
4.4. Tensor Bundles and Tensor Fields 235
notation I (h)) in the second component of the chart means the following.
Since h is a local trivialization of the original bundle, it defines for each
q U a linear isomorphism hq : Fq Rk . The functor assigns to this
isomorphism an isomorphism (hq ) : (Fq ) (Rk ). Composing (hq )
with the isomorphism I, we obtain a linear isomorphisms (Rk ) Rl . The
^
last l coordinates of (, h)(e) for e (Fq ) are the coordinates of the vector
I((hq )(e)).
The set of all charts constructed this way form an atlas on E . The transit
map between any two charts is defined on an open subset of Rn+l and it is
smooth. Indeed, if and h are another chart and local trivialization over an
1
^
open subset V , then the transit (, ^
h) (, h) map is defined on the open
set (U V ) Rl , and maps (x, v) (U V ) Rl to
1
^
(, ^
h) (, h) (x, v) = (1 (x)), I (h1 (x) )((h1 (x) ))1 I 1 (v) .
( , I ((h))) : ( )1 (U ) U Rl .
It is useful to know how one can work with tensor fields in local coordinates.
If = (x1 , . . . , xn ) : U Rn is a local coordinate system on the manifold,
then the vector fields 1 , . . . , n induced by give a basis in the tangent space
Tp M for all p M . Taking the dual basis of the basis at each point we obtain
n tensor fields of type (1, 0), which are usually denoted by dx1 , . . . , dxn . The
tensor fields dxi1 dxik j1 jl , (1 i1 , . . . , ik , j1 . . . .jl n)
yield a basis of the tensor space T (k,l) (Tp M ) at each point p U , therefore,
any smooth tensor field S of type (k, l), can be written uniquely as a linear
combination
X n
S= Sij11...i
...jl
k
dxi1 dxik j1 jl
i1 ,...,ik ,j1 ....jl =1
t S S
LX S = lim .
t0 t
The pulled back tensor field t S is defined only on Ut . Nevertheless, LX S
is properly defined on the whole manifold M , since for each p M , there
is a positive t0 , such that p Ut0 Ut0 , and then t S is defined at p for
|t| t0 , so the limit can be taken at p.
The definition shows exactly the key idea behind the Lie derivative. It is the
initial speed at which the tensor field changes when we start moving it by the
flow of the vector field X. However, the limit on the right-hand side is not
the most practical way to compute the Lie derivative, since to compute the
flow, we have to solve an ordinary differential equation. Our goal now is to
work out formulae for the computation of the Lie derivative without having
to compute the flow of the vector field explicitly. We start with the simplest
types of tensor fields.
Smooth functions on M are the tensor fields of type (0, 0), or differential 0-
forms. The Lie derivative for them is the well known differentiation of smooth
functions with respect to vector fields.
f (t (p)) f (p) d
LX f (p) = lim = f (t (p))|t=0 = Xp (f ).
t0 t dt
In the last step, we used the fact that p (t) = t (p) is an integral curve of
X, therefore its speed vector at 0 is 0 (0) = Xp .
Vector fields are tensor fields of type (0, 1). Lie derivation of vector fields is
also a familiar operation.
Proof. We use the standard trick to prove all Leibniz-type rules as follows
t ((Y )) (Y )
LX ((Y )) = lim =
t0 t
(t Y ) (t Y ) + (t Y ) (Y )
lim t
t0 t
t Y Y
= lim (lim t Y ) + lim t
t0 t t0 t0 t
= (LX )(Y ) + (LX Y ).
Tensor fields can always be multiplied by the tensor product, differential
forms can be multiplied by the wedge product pointwise. The Leibniz rule
also holds for these operations.
Proposition 4.5.5. If S1 and S2 are two arbitrary tensor fields, not nec-
essarily of the same type, , are arbitrary differential forms, then for any
smooth vector field X, we have
LX (S1 S2 ) = (LX S1 ) S2 + S1 (LX S2 ),
and
LX ( ) = (LX ) + (LX ).
Both equations can be proved by the standard trick. We leave the details to
the reader.
Proposition 4.5.6. If in addition to the assumptions of the previous propo-
sition S1 and S2 have the same type, and the differential forms and are
of the same degree, then for any two real numbers , R we have
LX (S1 + S2 ) = LX S1 + LX S2 ,
and
LX ( + ) = LX + LX .
Locally, any tensor field can be written as the sum of a finite number of
tensor products of some vector fields and differential one-forms. Similarly,
any differential form can be written locally as a sum of wedge products of
some differential 1-forms. Thus, the above rules are enough to compute the
Lie derivative of any tensor field and differential form.
The following proposition tells us how the Lie derivation operator LX depends
on the vector field X.
Proposition 4.5.7. The operator LX depends R-linearly on X, and satisfies
the commutator relation
LX LY LY LX = L[X,Y ] .
4.5. The Lie Derivative 241
Proof. The proof follows the steps we went through to work out formulae to
compute LX on more and more complicated tensor fields.
If the Lie derivations in the statement are applied to smooth functions, then
linearity in X is obvious, the commutator relation is just a reformulation of
the definition of the Lie bracket.
When Z is a vector field,
and
furthermore,
(LX LY LY LX )()(Z)
= X((LY )(Z)) (LY )([X, Z]) Y ((LX )(Z)) + (LX )([Y, Z])
= X(Y ((Z))) X(([Y, Z])) Y (([X, Z])) + ([Y, [X, Z]])
Y (X((Z))) + Y (([X, Z])) + X(([Y, Z])) ([X, [Y, Z]])
= [X, Y ]((Z)) ([[X, Y ], Z]) = (L[X,Y ] )(Z).
Next we show that if the statement is true when the Lie derivatives are
evaluated on the tensor fields S1 , S2 , or on the differential forms and ,
then it also holds on S1 S2 or on respectively. We check only for the
tensor product. The wedge product can be treated in the same way.
(LX LY LY LX )(S1 S2 )
= LX (LY (S1 ) S2 + S1 LY S2 ) LY (LX (S1 ) S2 + S1 LX S2 )
= LX (LY (S1 )) S2 + LY S1 LX S2 + LX S1 LY S2
+ S1 LX (LY (S2 )) LY (LX (S1 )) S2 LX S1 LY S2
LY S1 LX S2 S1 LY (LX (S2 ))
= (LX LY LY LX )(S1 )) S2 + S1 (LX LY LY LX )(S2 ))
= (L[X,Y ] S1 )) S2 + S1 (L[X,Y ] S2 ) = L[X,Y ] (S1 S2 ).
It is easy to see that if the identities we want to show hold on two tensor
fields or differential forms of the same type, then the identity holds also on
their sum. Since locally, e.g. in the domain of any chart, tensor fields can be
written as sums of tensor products of vector fields and 1-forms, and similarly,
differential forms can be written as sums of wedge-products of 1-forms, the
proposition is proved.
Definition 4.5.8. We say that (the flow of ) a vector field X leaves a tensor
field S invariant if t (S) = S|Ut for all t R, where t : Ut Ut (t R)
is the flow generated by X.
Proposition 4.5.9. The flow of a vector field X leaves the tensor field S
invariant if and only if LX S = 0.
Proof. If the flow leaves S invariant, then
t S S SS
LX S = lim = lim = 0.
t0 t t0 t
To prove the converse, assume that LX S = 0. Denote by St the one-
parameter family of tensor fields St = t S. S0 = S and for each p M ,
St (p) is defined on an open interval around 0, so to prove the statement, it
d
suffices to check that dt St 0. However, this is true since
St St0 h S S
lim = lim t0 = t0 (LX S) = 0.
tt0 t t0 h0 h
Corollary 4.5.10. Vector fields leaving a tensor field invariant form a Lie
subalgebra of the Lie algebra of all vector fields.
Proof. Indeed, if X and Y leave S invariant, then LX+Y S = LX S +
LY S = 0 and
= (1)kl .
When k = 0, we set X = 0.
Interior product L
with the vector field X can be thought of as a linear map
dim M k
from (M ) = k=0 (M ) into itself. This is a degree -1 map, as it
decreases the degree of any form by 1.
where the hat over li means that it is omitted. This means that
s
X
X (l1 ls ) = (1)i+1 li (X) (l1 lbi ls ).
i=1
By the introductory arguments, this special case implies the general one.
Exercise 4.6.5. Show that for any two vector fields X an Y , we have
X Y = Y X .
4.6. Differential Forms 245
Exercise 4.6.6. Show that the Lie derivation of differential forms and the in-
terior product with a vector field satisfies the commutation relation
LX Y Y LX = [X,Y ] .
0 = d(h(1 2 )) = dh (1 2 ) + h(d1 d2 ).
1i1 <<ik n
1i1 <<ik n
and set X
d = di1 ...ik dxi1 dxik .
1i1 <<ik n
Let us check that d satisfies all the requirements. It is obviously linear and
increases the degree of forms by 1, and coincides with d on smooth functions.
It is also a superderivation. Indeed, if l (M ) has the decomposition
X
= j1 ...jl dxj1 dxjl ,
1j1 <<jl n
then
d ( ) =
X X
d i1 ...ik j1 ...jl dxi1 dxik dxj1 dxjl
1i1 <<ik n 1j1 <<jl n
X X
= d(i1 ...ik j1 ...jl ) dxi1 dxik dxj1 dxjl
1i1 <<ik n 1j1 <<jl n
4.6. Differential Forms 247
= d + (1)k d.
1i1 <<ik n
X
= d (di1 ...ik ) dxi1 dxik ,
1i1 <<ik n
In the rest of this section we prove some useful formulae involving the exterior
differentiation.
It is also clear that a wedge product can be pulled back componentwise, i.e.,
f ( ) = f () f ().
If p M is an arbitrary point, then we can introduce a local coordinate
system = (x1 , . . . , xn ) : U Rn around f (p), where n = dim N (!). By
continuity of f , we can also find an open neighborhood V of p, such that
f (V ) U . Then for any k (N ), we can write
X
|U = i1 ...ik dxi1 dxik ,
1i1 <<ik n
hence
X
(f )|V = f (i1 ...ik )f (dxi1 ) f (dxik )
1i1 <<ik n
X
= (i1 ...ik f )d(xi1 f ) d(xik f ).
1i1 <<ik n
and
X
(f (d))|V = d(i1 ...ik f )d(xi1 f ) d(xik f ) = d(f ()).
1i1 <<ik n
Proof. Using the usual notation, for any differential form , we have
d d
Lx (d) = (t (d))|t=0 = d(t ())|t=0 .
dt dt
4.6. Differential Forms 249
In the last expression, we can change the order of differentiation with respect
to the parameter t and the exterior differentiation by Youngs theorem. In-
deed, if we write the one-parameter family of forms t = t using local
coordinates as
X
t (p) = i1 ...ik (p, t)dxi1 dxik ,
1i1 <<ik n
then
n
d X X
d(t )(p) = t j i1 ...ik (p, t)dxj dxi1 dxik
dt
1i1 <<ik n j=1
X Xn
= j t i1 ...ik (p, t)dxj dxi1 dxik
1i1 <<ik n j=1
d
=d t .
dt
d d
Consequently, Lx (d) = dt d(t ())|t=0 =d dt t ()|t=0 = d(Lx ).
LX ( ) = LX () + LX ()
= X (d) + d(X ) + X (d) + d(X ).
250 4. Manifolds
(X d + d X )( )
= X (d + (1)k d) + d(X + (1)k X )
= X (d) + (1)k+1 d X + (1)k X d + (1)2k X (d)+
+ d(X ) + (1)k1 X d + (1)k d X + (1)2k d(X )
= X (d) + d(X ) + X (d) + d(X ).
Since locally, in the domain of a chart, any differential form can be written
as the sum of wedge products of functions and differentials of functions, the
proposition is true.
The formula below can also be used as a coordinate free definition of the
exterior differential of a differential form.
k
X
d(X0 , . . . , Xk ) = (1)i Xi ((X0 , . . . , Xi , . . . , Xk ))+
i=0
X
+ (1)i+j ([Xi , Xj ], X0 , . . . , Xi , . . . , Xj , . . . , Xk ).
0i<jk
k
X
(LX0 )(X1 . . . , Xk ) = LX0 ((X1 . . . , Xk )) (X1 , . . . , LX0 Xi , . . . , Xk )
i=1
k
X
= X0 ((X1 . . . , Xk )) + (1)k ([X0 , Xi ], X1 , . . . , Xi , . . . , Xk ).
i=1
4.6. Differential Forms 251
The kth de Rham cohomology space is a linear space. Its elements are equiv-
alence classes of closed differential k-forms, where two forms represent the
same equivalence class, also called cohomology class if and only if their dif-
ference is exact. L
k
It is useful to consider the direct sum HdR (M ) = k=0 HdR (M ) of all de
Rham cohomology spaces, because this linear space has also a multiplicative
structure induced by the wedge product of forms. Let [] and [] be two
cohomology classes represented by the closed forms and . Then is a
closed form and the cohomology class [ ] depends only on the cohomology
classes [] and []. Indeed, if we choose other representatives +d and +d
from these cohomology classes, then their wedge product
( + d) ( + d) = + d ( + d) + (1)deg
Definition 4.6.15. The linear space HdR (M ) together with its multiplicative
structure defined by
[][] = [ ]
is called the de Rham cohomology algebra of M .
as we wanted to show.
Proof. Let f and g be the maps from the definition of homotopy equiv-
alence. Then the algebra homomorphisms f : HdR
(N ) HdR
(M ) and
g : HdR (M ) HdR (N ) have the property that f g = (g f ) = (idM ) =
(M ) and similarly, g f
idHdR = idHdR
(N ) , which means that f and g are
the inverses of one another, in particular, they are isomorphisms.
Ps
The boundary of a smooth singular k-dimensional chain j=1 s s is the
(k 1)-dimensional chain
s s
X X
s s = s s .
j=1 j=1
The boundary map is a linear map : Ck (M ) Ck1 (M ) for all k.
Proposition 4.7.5. The boundary of the boundary of a chain is 0, that is
= 0.
Proof. Since the linear space of smooth singular chains is generated by smooth
singular simplices and is linear, it is enough to check the statement for sin-
gular simplices. Let : k M be a smooth singular simplex. Then
k
X k k1
X X
() = (1)j ( ljk ) = (1)i+j ljk lik1 .
i=0 j=0 i=0
where the second integral is the integral of f over the standard simplex k
with respect to the Lebesgue measure. Ps
The integral of over a k-dimensional smooth singular chain c = i=1 i i
is defined by
Z Xs Z
= i .
c i=1 i
The pull-back of d by is
k
X
(d) = d( ) = d = dfi dx1 dx
ci dxk
i=1
k
X k
X
= j fi dxj dx1 dx
ci dxk
i=1 j=1
k
!
X
i1
= (1) i fi dx1 dxk .
i=1
To compute the ith integral on the right let us integrate with respect to
the ith variable first, and then with respect to the others. In the first step
compute the integral using the Newton-Leibniz rule. This gives us
Z
i fi (x)dx
k
Z 1(x1 ++
Z xbi ++xk )
= i fi (x1 , . . . , xk )dxi dx1 . . . dx
ci . . . dxk
(x1 ,...,xbi ,...,xk )k1 0
Z
= fi (x1 , . . . , 1 (x1 + + xbi + + xk ), . . . , xk )dx1 . . . dx
ci . . . dxk
(x1 ,...,xbi ,...,xk )k1
Z
fi (x1 , . . . , xi1 , 0, xi+1 , . . . , xk )dx1 . . . dx
ci . . . dxk .
(x1 ,...,xbi ,...,xk )k1
x1 = 1 (y1 + + yk1 ) if i 6= 1,
xj = yj1 for 2 j k and j 6= i,
The determinants of the derivative matrices are equal to 1 for both substi-
tutions, therefore
Z Z Z
i fi (x)dx = fi (l0k (y))dy fi (lik (y))dy,
k k1 k1
and
Z Z k
! k Z
X X
i1
d = (1) fi (l0k (y)) dy + (1)i fi (lik (y))dy.
k1 i=1 i=1 k1
and Z Z
= fi (lik (y))dy.
lik k1
(l0k ) (dx1 dx
ci dxk )
= d(1 y 1 y k1 ) dy 1 dy
\ i1 dy k1
= (1)i1 dy 1 dy k1 ,
consequently,
k
X
( l0k ) = (l0k ) = (1)i (fi l0k )dy 1 dy k1 ,
i=1
and !
Z Z k
X
i1 k
= (1) fi (l0 (y)) dy.
l0k k1 i=1
Adding the integrals of on the facets of with alternating signs we get
Z k
X Z
i
= (1)
i=0 lik
Z k
! k Z
X X
i1
= (1) fi (l0k (y)) dy + (1)i fi (lik (y))dy
k1 i=1 i=1 k1
Z
= d,
as we wanted to show.
Corollary 4.7.9. If c R Zk (M ) is a cycle, k (M ) is a closed differential
form, then the integral c depends only on the homology class [c] Hk (M )
k
of c and the de Rham cohomology [] HdR (M ) of .
Proof. Choose other elements c0 = c + a and 0 = + d from the same
homology and cohomology class respectively. Then
Z Z Z Z Z Z Z Z
0 = + ( +d)+ d = + (d +d(d))+ = .
c0 c c c c c c c
Z Z
[] = .
[c] c
260 4. Manifolds
(U D) = (U ) {(x1 , . . . , xn ) Rn | xn 0}.
(U D) = (U ) {(x1 , . . . , xn ) Rn | xn = 0},
and similarly, Z Z
hi hj = f(x)dx
j Rn
+ (Ui Uj )
Comparing equations (4.8) and (4.9) we see that the equation we want to
show holds up to sign. To have the equality without any sign corrections,
it remains to show that i j = sgn det H 0 . If dim M 2, then i = j = 1,
furthermore, and are positive charts, so the derivative of the transit map
between them preserves orientation, that is, it has positive determinant.
For dim M = 1, the derivative of the transit map has positive determinant if
and only if either both and are positive or none of them is positive. This
completes the proof.
Proposition 4.7.15. For any two different choices of the simplices and the
partition of unity
Xs Z Xs Z
hi = hj .
i=1 i i j=1 j j
s Z
X s Z
X Xs
hi = hi hj
i=1 i i i=1 i i j=0
s X
X s Z s X
X s Z
= hi hj = hi hj
i=1 j=1 i i i=1 j=1 j j
s Z s
! s Z
X X X
= hi hj = hj .
j=1 j j i=0 j=1 j j
264 4. Manifolds
Applying the Stokes theorem for chains, and the fact that supp(hi ) inter-
sects only the image of the nth facet of i ,
Z s Z
X s Z
X s Z
X
d = d(hi ) = d(hi ) = hi
D i=1 D i=1 i i i=1 i i
Xs Z
n
= i (1) hi .
n
i ln
i=1
The right-hand side makes sense, because the integral of compactly supported
forms can be defined in the same manner as we did also in the case when the
266 4. Manifolds
Y X = (X )0 (0) (4.10)
where i denotes the i-th coordinate vector field on Rn , X i are the compo-
nents of the vector field X with respect to i . In particular, the value of Y X
does not depend on the choice of .
It is easy to check that this differentiation of vector fields has the properties.
(1) Y1 +Y2 X = Y1 X + Y2 X,
(2) cY X = cY X,
(3) Y (X1 + X2 ) = Y X1 + Y X2 ,
(4) Y (f X) = Y (f )X + f (p)Y X,
(5) X1 X2 X2 X1 = [X1 , X2 ],
(6) Y hX1 , X2 i = hY X1 , X2 i + hX1 , Y X2 i,
where X1 , X2 X(R ), Y Tp Rn , f F(Rn ), c R.
n
(i) Y1 +Y2 X = Y1 X + Y2 X,
(ii) cY X = cY X,
(iii) Y (X1 + X2 ) = Y X1 + Y X2 ,
(iv) Y (f X) = Y (f )X + f (p)Y X,
where X1 , X2 X(M ), Y, Y1 , Y2 Tp M , f F(M ), c R.
Proof. Let us suppose that the vector fields X1 and X2 coincide on an open
neighborhood U of p. Choose a smooth function h F(M ) which is zero
outside U and constant 1 on a neighborhood of p. Then we have h(X1 X2 ) =
0, consequently
0 = (Y X1 )(p) (Y X2 )(p).
The lemma implies that an affine connection can be restricted onto any open
subset and can be recovered from its restrictions onto the elements of an open
cover.
Fix a local coordinate system = (x1 , . . . , xn ) on an open subset U of M
and let 1 = 1 ,. . . ,n = n be the basis vector fields on U induced by .
(As we work with a fixed chart , omission of from the notation should not
4.8. Differentiation of Vector Fields 269
The components kij are smooth functions on U , called the Christoffel symbols
of the covariant derivation with respect to the given chart.
Proposition 4.8.4. The restriction of a global affine connection onto an
open coordinate neighborhood U is uniquely determined by the Christoffel
symbols. Any n3 smooth functions kij on U are the Christoffel symbols of an
appropriate affine connection on U .
Pn Pn
Proof. Let X = i=1 X i i , Y = j=1 Y j j be two smooth vector fields on
U . Then by the properties of affine connections, Y X can be computed as
follows
n
! n n
!
X X X
i j i
Y X = ( n Y j j )
P X i = Y j X i
j=1
i=1 j=1 i=1
n X
X n
Y j j (X i )i + X i j i
=
i=1 j=1
n
n X n
!
X X
j i i
= Y j (X )i + X kji k
i=1 j=1 k=1
n
X n X
X n
= Y (X k ) + X i Y j kji k .
k=1 i=1 j=1
This formula shows that the knowledge of the Christoffel symbols enables us
to compute the covariant derivative of any vector field with respect to any
other one. On the other hand, if kij are an arbitrary smooth functions on U
for 1 i, j, k n, then defining the covariant derivative of a vector field by
the above formula, we obtain an affine connection on U .
Then t X( ) is given by
n n X
n
X 0 X
t X( ) := X k ( ) + X i ( )Y j ( )kji (( )) k (( )).
k=1 i=1 j=1
Remark. Several different notations are used for the covariant derivative of
a vector field X along a curve : t X, X 0
dt , X , X. The most concise
0
0
notation X does not show the dependence of this operation on . Neverthe-
less, it is convenient to use when we work with a fixed connection on M .
Notations t X and X dt can be used only when the parameter of the curve
is denoted by a fixed symbol, in our case by t. Of course, if another symbol
is used for the parameter, then t should be replaced by that. The notation
0 X is motivated by the fact that if X is a vector field defined in an open
neighborhood of ( ) which extends X in the sense that X((t))) = X(t) for
all t close enough to , then t X( ) = 0 ( ) X. In the rest of this book we
shall mainly use the notations t X and 0 X.
Since these equations are linear, the existence and uniqueness theorem for
linear differential equations guaranties that the solutions of this system of
differential equations are uniquely determined by the initial values X k (0)
and can be defined for all values of t.
If X is a parallel vector field along the curve , then the vector X((t)) is
said to be obtained from X0 by parallel transport along .
Definition 4.8.8. A connection is called symmetric or torsion free if it sat-
isfies the identity
X Y Y X = [X, Y ].
Applying this identity to the case X = i , Y = j , from [i , j ] = 0 one
obtains the relation
kij = kji
for the Christoffel symbols of a torsion free connection. Conversely, if kij =
kji then using the expression of covariant derivative with the help of Christof-
fel symbols we get
X Y Y X
Xn n X
X n
= X(Y k ) + Y i X j kji k
k=1 i=1 j=1
n
X n X
X n
Y (X k ) + X i Y j kji k
k=1 i=1 j=1
n
X
X(Y k ) Y (X k ) k = [X, Y ].
=
i=1
x (y s) = y (x s)
The partial covariant derivative of this vector field with respect to x is equal
to
n
!
X
i
x (y s) = x y s (i s)
i=1
n
X
= x (y (si )) (i s) + y si x (i s)
i=1
n
X n
X X n
n X
= x (y (si )) (i s) + y si x sj (kji s) k s.
i=1 k=1 i=1 j=1
This formula shows that interchanging the role of x and y we obtain the same
vector field for any s if and only if kij = kji .
Roughly speaking, the torsion free condition halves the degree of freedom
in the choice of Christoffel symbols, a symmetric connection is uniquely de-
2
termined by n (n+1)
2 arbitrarily chosen functions, nevertheless, the space of
symmetric affine connections on a manifold is still infinite dimensional. We
can reduce the degree of freedom further on Riemannian manifolds.
Proposition 4.8.11. Suppose that the connection is compatible with the met-
ric. Let V, W be any two vector fields along . Then
hV, W i0 = ht V, W i + hV, t W i.
4.8. Differentiation of Vector Fields 273
Therefore
n
X
ht V, W i + hV, t W i = (vi0 wi + vi wi0 ) = hV, W i0 .
i=1
Y (hX1 , X2 i) = hY X1 , X2 i + hX1 , Y X2 i.
i gjk = hi j , k i + hj , i k i.
Permuting i, j and k this gives three linear equations relating the three
quantities
hi j , k i, hj k , i i, hk i , j i.
(There are only three such quantities since i j = j i .) These equations
can be solved uniquely; yielding the first Christoffel identity
1
hi j , k i = (i gjk + j gik k gij ).
2
Pn
The left-hand side of this identity is equal to l=1 lij glk . Multiplying by
the inverse (g kl ) of the matrix (glk ) this yields the second Christoffel identity
n
X 1
kij = (i gjl + j gil l gij )g lk .
2
l=1
Conversely, defining kij by this formula, one can verify that the resulting
connection is symmetric and compatible with the metric. This completes the
proof.
1
hX Y, Zi = XhY, Zi hY, [X, Z]i + Y hZ, Xi
2 (4.12)
hZ, [Y, X]i ZhX, Y i + hX, [Z, Y ]i .
and
n1
X 1
kij = (gjl,i + gil,j gij,l )g lk .
2
l=1
4.9. Curvature 275
proved in Section 3.1.5 with the formulae derived for the Levi-Civita connec-
tion we can conclude that the Christoffel symbols of a hypersurface intro-
duced previously are the Christoffel symbols of the Levi-Civita connection
of the hypersurface. Therefore, denoting by the Levi-Civita connection of
the hypersurface, we can rewrite equation (4.13) as
ri rj = ri rj + II(ri , rj )N.
Exercise 4.8.15. Show that if and are two affine connections on a man-
ifold M , then their difference S(X, Y ) = X Y X Y is an F(M )-bilinear
mapping. (In other words, S is a tensor field of type (2, 1)). Conversely, the
sum of a connection and an F(M )-bilinear mapping S : X(M ) X(M )
X(M ) is a connection. (According to these statements, affine global connec-
tions form an affine space over the linear space of tensor fields of type (2, 1).)
T (X, Y ) = X Y Y X [X, Y ]
Exercise 4.8.18. Check that the connection defined by the Christoffel sym-
bols
n1
1X
kij = (gli,j + glj,i gij,l )g lk
2
l=1
4.9 Curvature
If is an affine connection on a manifold M , then we may consider the
operator
R(X, Y ) = [X , Y ] [X,Y ] : X(M ) X(M ),
276 4. Manifolds
Proposition 4.9.1. The curvature tensor is linear over the ring of smooth
functions in each of its arguments, and it is skew symmetric in the first two
arguments.
and
need
R(X, Y ; f Z) = X Y (f Z) Y X (f Z) [X,Y ] (f Z)
= X (Y (f )Z + f Y Z) Y (X(f )Z + f X Z)
[X, Y ](f )Z f [X,Y ] Z
= XY (f )Z + Y (f )X Z + X(f )Y Z + f X Y Z
Y X(f )Z X(f )Y Z Y (f )X Z f Y X Z
XY (f )Z + Y X(f )Z f [X,Y ] Z
= f (X Y Z Y X Z [X,Y ] Z) = f R(X, Y ; Z).
Proposition 4.9.1 is interesting, because the curvature tensor is built up from
covariant derivations, which are not linear operators over the ring of smooth
functions.
We have already introduced tensor fields over a hypersurface. We can in-
troduce tensor fields over a manifold in the same manner. A tensor field T
of type (k, l) is an assignment to every point p of a manifold M a tensor
T (p) of type (k, l) over the tangent space Tp M . If 1 , . . . , n are the basis
vector fields defined by a chart over the domain of the chart, and we denote
by dx1 (p), . . . , dxn (p) the dual basis of 1 (p), . . . , n (p), then a tensor field is
uniquely determined over the domain of the chart by the components
Tji11...j
...il
k
(p) = T (p)(dxi1 , . . . , dxil ; j1 , . . . , jk ).
We say that the tensor field is smooth, if for any chart from the atlas of M ,
the functions Tji11...j
...il
k
are smooth. We shall consider only smooth tensor fields.
Tensor fields of type (0, 1) are the vector fields, tensor fields of type (1, 0) are
the differential 1-forms. Thus, a differential 1-form assigns to every point of
the manifold a linear function on the tangent space at that point. Differential
1-forms form a module over the ring of smooth functions, which we denote
by 1 (M ).
Every tensor field of type (k, l) defines an F(M )-multilinear mapping
1 (M ) 1 (M ) X(M ) X(M ) F(M )
| {z } | {z }
l times k times
and conversely, every such F(M )-multilinear mapping comes from a tensor
field. (Check this!) Therefore, tensor fields can be identified with F(M )-
multilinear mappings 1 (M ) 1 (M ) X(M ) X(M ) F(M ).
Tensor fields of type (k, 1), that is F(M )-multilinear mappings
1 (M ) X(M ) X(M ) F(M )
can be identified in a natural way with F(M )-multilinear mappings
X(M ) X(M ) X(M ).
278 4. Manifolds
but the last expression is 0 according to the Jacobi identity on the Lie bracket
of vector fields. (At the third and fifth equality we used the torsion free
property of .)
4.9. Curvature 279
of type (k, l) with respect to a vector field X is a tensor field of the same
type, defined by
(X T )(1 , . . . , l ; X1 , . . . , Xk ) = X(T (1 , . . . , l ; X1 , . . . , Xk ))
k
X
T (1 , . . . , X i , . . . , l ; X1 , . . . , Xk )
i=1
l
X
T (1 , . . . , l ; X1 , . . . , X Xj , . . . , Xk ).
j=1
First, we have
X
X R(Y, Z) R(Y, Z) X
XY Z
X
= (X Y Z X Z Y X [Y,Z] )
XY Z
(Y Z X Z Y X [Y,Z] X )
X
= (X Y Z X Z Y X [Y,Z] )
XY Z
(X Y Z X Z Y [Y,Z] X )
X
= [Y,Z] X X [Y,Z] .
XY Z
and similarly,
Y (X(hZ, W i)) = hY X Z, W i
+ hX Z, Y W i + hY Z, X W i + hZ, Y X W i.
We also have
Subtracting from the first equality the second and the third one and applying
[X, Y ] = X Y Y X, we obtain
0 = hX Y Z Y X Z [X,Y ] Z, W i
+ hZ, X Y W Y X W [X,Y ] W i
= R(X, Y ; Z, W ) + R(X, Y ; W, Z).
282 4. Manifolds
0 = S(X, Y + Z; Y + Z, X)
= S(X, Y ; Y, X) + S(X, Y ; Z, X) + S(X, Z; Y, X) + S(X, Z; Z, X)
= S(X, Y ; Z, X) + S(X, Z; Y, X) + [S(X, Y ; Z, X) + S(Y, Z; X, X)
| {z }
=0
+ S(Z, X; Y, X)] = 2S(X, Y ; Z, X).
0 = S(X + W, Y ; Z, X + W )
= S(X, Y ; Z, X) + S(X, Y ; Z, W ) + S(W, Y ; Z, X) + S(W, Y ; Z, W )
= S(X, Y ; Z, W ) + S(W, Y ; Z, X),
4.9. Curvature 283
in other words, S is invariant under cyclic permutations of the first three vari-
ables. But the sum of the three equal quantities S(X, Y ; Z, W ), S(Y, Z; X, W )
and S(Z, X; Y, W ) is 0 because of the Bianchi symmetry, thus S(X, Y ; Z, W )
is 0.
Exercise 4.9.8. Let S be an algebraic curvature tensor, and let QS (X, Y ) :=
S(X, Y ; Y, X). Prove that QS (X, Y ) = QS (Y, X) and
6S(X, Y ; Z, W ) =QS (X + W, Y + Z) QS (Y + W, X + Z) + QS (Y + W, X)
QS (X + W, Y ) + QS (Y + W, Z) QS (X + W, Z)+
+ QS (X + Z, Y ) QS (Y + Z, X) + QS (X + Z, W )
QS (Y + Z, W ) + QS (X, Z)
QS (Y, Z) + QS (Y, W ) QS (X, W ).
Theorem 4.9.9. Assume that S is an algebraic curvature tensor. Then
S(X, Y ; Z, W ) = S(Z, W ; X, Y ).
Figure 4.5
easily from our assumptions that the sum of the labels of the vertices of
any of the shaded faces of the octahedron is 0 by a Bianchi identity, in
which the label of the face stands in the last place and the sum goes over
cyclic permutations of the other three variables. If we add these identities
284 4. Manifolds
for the upper two shaded triangles Z and W and subtract the identities
corresponding to the two lower triangles X and Y , we obtain
2S(X, Y ; Z, W ) 2S(Z, W ; X, Y ) = 0,
as we wanted to prove.
Corollary 4.9.10. For the Riemann-Christoffel tensor, the identity
R(X, Y ; Z, W ) = R(Z, W ; X, Y ) holds.
Definition 4.9.11. Let M be a Riemannian manifold, p a point on M , X
and Y two non-parallel tangent vectors at p. The number
R(X, Y ; Y, X)
K(X, Y ) =
|X|2 |Y |2 hX, Y i2
is called the sectional curvature of M at p, in the direction of the plane
spanned by the vectors X and Y in Tp M .
Proposition 4.9.12. If M is a Riemannian manifold with sectional curva-
ture K, X and Y are linearly independent tangent vectors at p M , and
are non-zero scalars, then
(i) K(X, Y ) = K(X + Y, Y );
(ii) K(X, Y ) = K(X, Y );
(iii) K(X, Y ) = K(Y, X).
Furthermore, K depends only on the plane spanned by its arguments.
Proof. (i) follows from
and
|X + Y |2 |Y |2 hX + Y, Y i2
= (|X|2 + |Y |2 + 2hX, Y i)|Y |2 (hX, Y i2 + 2hX, Y i|Y |2 + |Y |4 )
= |X|2 |Y |2 hX, Y i2 .
R(X, Y ; Y, X) = 2 2 R(X, Y ; Y, X)
and
|X|2 |Y |2 hX, Y i2 = 2 2 (|X|2 |Y |2 hX, Y i2 ).
4.9. Curvature 285
that is,
R(X, Y ; Z) = S(X, Y ; Z) for all X, Y, Z.
Differentiating with respect to a vector field U we get
(U R)(X, Y ; Z) = (U S)(X, Y ; Z)
= U (S(X, Y ; Z)) S(U X, Y ; Z)
S(X, U Y ; Z) S(X, Y ; U Z).
Since
U (S(X, Y ; Z)) = U (f )(hY, ZiX hX, ZiY ) + f U (hY, ZiX hX, ZiY )
= U (f )(hY, ZiX hX, ZiY ) + f (U hY, ZiX + hY, ZiU X
U hX, ZiY hX, ZiU Y )
= U (f )(hY, ZiX hX, ZiY ) + f (hU Y, ZiX + hY, U ZiX+
+ hY, ZiU X hU X, ZiY hX, U ZiY hX, ZiU Y ) =
= U (f )(hY, ZiX hX, ZiY ) + S(U X, Y ; Z) + S(X, U Y ; Z)+
+ S(X, Y ; U Z),
we obtain
In particular,
n
X n
X
Ric(X, Y )(p) = hR(ei , X(p); Y (p)), ei i = R(ei , X(p); Y (p), ei )
i=1 i=1
n
X n
X
= R(Y (p), ei ; ei , X(p)) = R(ei , Y (p); X(p), ei )
i=1 i=1
= Ric(Y, X)(p).
Since the Ricci tensor of a Riemannian manifold is symmetric, it is uniquely
determined by its quadratic form X Ric(X, X) (see equation (1.8)).
Definition 4.9.17. Let Xp Tp M be a non-zero tangent vector of a Rie-
mannian manifold M . The Ricci curvature of M at p in the direction Xp is
the number
Ric(Xp , Xp )
r(XXp ) = 2
.
= e1|X
, e2p,| . . . , en we can express the Ricci
p
Fixing an orthonormal basis
|Xp |
curvature as follows
n n
Ric(Xp , Xp ) X R(ei , Xp ; Xp , ei ) X
r(Xp ) = = = K(Xp , ei ).
|Xp |2 i=1
|Xp |2 i=2
288 4. Manifolds
The meaning of this formula is that the Ricci curvature in the direction Xp
is the sum of the sectional curvatures in the directions of the planes spanned
by the vectors Xp and ei , where ei runs over an orthonormal basis of the
orthogonal complement of Xp in Tp M . It is a nice geometrical corollary that
this sum is independent of the choice of the orthogonal basis.
With the help of a scalar product, one can associate to every bilinear function
a linear transformation. For the case of the Riemannian metric and the Ricci
tensor, we can find a unique F(M )-linear transformation Ric : X(M ) X(M )
defined by
that is, the scalar curvature is twice the sum of sectional curvatures taken in
the directions of all coordinate planes of an orthonormal coordinate system
in Tp M .
To finish this section with, let us study the curvature tensor of a hyper-
surface M in Rn . According to (4.14), the Levi-Civita connection of a
hypersurface can be expressed as = P , where is the derivation rule
of vector fields along the hypersurface as defined in Definition 3.1.7, P is the
orthogonal projection of a tangent vector of Rn at a hypersurface point onto
the tangent space of the hypersurface at that point. Comparing Definition
3.1.7 to formula (4.10), we see that the derivation of vector fields along a
hypersurface is induced by the Levi-Civita connection of Rn , which we also
denoted by . As the curvature of Rn is 0,
X Y Y X = [X,Y ]
We have
X Y Z = P (X Y Z) = P (X (Y Z hY Z, NiN))
= P (X Y Z) P (X(hY Z, Ni)N) P (hY Z, NiX N)
= P (X Y Z) hY Z, NiX N,
where X, Y, Z X(M ). Similarly,
Y X Z = P (Y X Z) hX Z, NiY N.
Combining these equalities with
[X,Y ] Z = P ([X,Y ] Z)
we get the following expression for the curvature tensor R of M
R(X, Y ; Z) = (X Y Z Y X Z) [X,Y ] Z =
= P ((X Y Z Y X Z) [X,Y ] Z)
hY Z, NiX N + hX Z, NiY N
= hX Z, NiY N hY Z, NiX N.
Since hZ, Ni is constant zero,
0 = X(hZ, Ni) = hX Z, Ni + hZ, X Ni
and
0 = Y (hZ, Ni) = hY Z, Ni + hZ, Y Ni.
Putting these equalities together we deduce that
R(X, Y ; Z) = hZ, Y NiX N hZ, X NiY N
= hZ, L(Y )iL(X) hZ, L(X)iL(Y ).
Comparing the formula
R(X, Y ; Z) = hZ, L(Y )iL(X) hZ, L(X)iL(Y )
relating the curvature tensor to the Weingarten map on a hypersurface with
Gauss equations we see that the curvature tensor R coincides with the curva-
ture tensor defined there. This way, the last equation can also be considered
as a coordinate free display of Gauss equations.
Exercise 4.9.19. Prove that if X1 and X2 are two nonparallel principal
directions at a given point p of a hypersurface M , 1 , 2 are the corresponding
principal curvatures, then
K(X1 , X2 ) = 1 2 .
What is the minimum and maximum of K(X, Y ), when X and Y run over
Tp M ?
290 4. Manifolds
and the cyclic permutation of (X, Y, Z) and the transposition of (Z, W ) gen-
erates the full permutation group of (X, Y, Z, W ).
4.10. Decomposition of Algebraic Curvature Tensors 291
Assume that V is a Euclidean linear space and denote by g the inner product
of V . Then the orthogonal group O(V ) acts on R by
Proposition 4.10.7.
The Ricci tensor of R does not depend on the choice of the orthonormal
basis.
Ric(R) is symmetric and the linear map Ric : R S 2 (V ) is O(V )
equivariant, that is, Ric((R)) = (Ric R)) for any R R.
Ric(h g) = (2 n)h tr(h)g, in particular, Ric(g g) = 2(1 n)g
and tr(Ric(h g)) = 2(1 n) tr(h).
Proof. Ric(R)(X, Y ) can also be obtained as the trace of the bilinear func-
tion h(Z, W ) = R(X, Z, W, Y ), thus it does not depend on the choice of
orthonormal basis.
Symmetry of Ric(R) follows from
The O(V ) equivariance follows from the first part of the proposition and the
fact that an orthogonal transformation maps an orthonormal basis to an
n
X
Ric((R))(X, Y ) = R(1 (X), 1 (ei ), 1 (ei ), 1 (Y ))
i=1
n
X
= (R(1 (X), ei , ei , 1 (Y )) = (Ric(R))(X, Y ).
i=1
4.10. Decomposition of Algebraic Curvature Tensors 293
Pn Pn
Finally, using the decompositions X = i g(X, ei )ei and Y = i g(Y, ei )ei ,
we obtain
Ric(hg)(X, Y )
Xn
= (h g)(X, ei , ei , Y )
i
n
X
= h(X, ei )g(ei , Y ) h(X, Y )g(ei , ei )
i
+ g(X, ei )h(ei , Y ) g(X, Y )h(ei , ei )
= (2 n)h(X, Y ) tr(h)g(X, Y ).
Definition 4.10.11. We call P (R) the Ricci component of the algebraic cur-
vature tensor R as it depends only on the Ricci tensor Ric(R). The algebraic
Weyl tensor W (R) W given by equation (4.17) is called the Weyl tensor
or Weyl component of R.
The action of O(V ) on S 2 (V ) is not irreducible. Since (g) = g for any
O(V ), the one dimensional subspace spanned by g is an O(V ) invariant
subspace. The trace is an O(V ) invariant linear function on S 2 (V ), so its ker-
nel S02 (V ) = {h S 2 (V ) | tr(h) = 0}, the linear space of 0-trace symmetric
bilinear functions, is an O(V ) invariant linear subspace of codimension 1.
Proposition 4.10.12. If V is an n-dimensional Euclidean linear space with
inner product g, then the linear space S 2 (V ) splits into the direct sum of the
O(V ) invariant subspaces Rg and S02 (V ). The components of h S 2 (V )
corresponding to this direct sum decomposition are (tr(h)/n)g R and h
(tr(h)/n)g S02 (V ).
We leave the proof as an exercise to the reader.
Definition 4.10.13. The scalar curvature of an algebraic curvature tensor
R is the trace s(R) = tr(Ric(R)) of the Ricci tensor of R.
Applying Proposition 4.10.12 to decompose the Ricci component of a curva-
ture tensor, we obtain the following theorem summarizing the above calcula-
tions.
Theorem 4.10.14. Let V be a Euclidean linear space with inner product g.
For an Einstein manifold with constant Ricci curvature r, the Ricci tensor
Ric is equal to rg, so the trace free Ricci component (Ric rg) is equal to 0.
The following theorem is an analog of Schurs theorem.
Proof. The trace free part of the Ricci tensor vanishes if and only if Ric = rg
for a smooth function r on the manifold M . Our goal is to show that in this
case r must be constant.
Since M is connected, it is enough to show that r has vanishing derivative
with respect to any vector field Z. To check Z(r) = 0 at a point p M , choose
a local coordinate system around p and apply the GramSchmidt orthogo-
nalization process to the vector fields 1 , . . . , n to obtain orthonormal vector
. . . , En in a neighborhood of p. Write Z Ei as a linear combination
fields E1 ,P
Z Ei = ij Ej . The coefficients ij can be computed as ij = hZ Ei , Ej i.
Differentiating the equation hEi , Ej i = ij , we obtain that ij = ji .
As a corollary of this skew symmetry, we obtain that for any two vector fields
X, Y , we have
n
X
(R(X, Z Ei , Ei , Y ) + R(X, Ei , Z Ei , Y ))
i=1
n
X
= (R(Z Ei , X, Y, Ei ) + R(Ei , X, Y, Z Ei ))
i=1
Xn
= (R(ij Ej , X, Y, Ei ) + R(Ei , X, Y, ij Ej ))
i,j=1
4.10. Decomposition of Algebraic Curvature Tensors 297
n
X
= (ij + ji )R(Ej , X, Y, Ei ) = 0. (4.18)
i,j=1
Pn
Differentiating the defining equation s = i,j=1 R(Ei , Ej , Ej , Ei ) of the
scalar curvature with respect to Z using (4.18) and the second Bianchi iden-
tity,
n
X
Z(s) = [(Z R)(Ei , Ej , Ej , Ei ) R(Z Ei , Ej , Ej , Ei )
i,j=1
R(Ei , Z Ej , Ej , Ei ) R(Ei , Ej , Z Ej , Ei )
R(Ei , Ej , Ej , Z Ei )]
n
X
= (Z R)(Ei , Ej , Ej , Ei ) (4.19)
i,j=1
Xn
= [(Ei R)(Ej , Z, Ej , Ei ) + (Ej R)(Z, Ei , Ej , Ei )]
i,j=1
X n
=2 (Ej R)(Z, Ei , Ei , Ej ).
i,j=1
Pn
Differentiate now the equation rhZ, Ej i = Ric(Z, Ej ) = i=1 R(Z, Ei , Ei , Ej )
with respect to Ej to get
Ej (r)hZ, Ej i + rhEj Z, Ej i + rhZ, Ej Ej i
X n
= (Ej R)(Z, Ei , Ei , Ej ) + R(Ej Z, Ei , Ei , Ej )+
i=1
+ R(Z, Ej Ei , Ei , Ej ) + R(Z, Ei , Ej Ei , Ej ) + R(Z, Ei , Ei , Ej Ej )
Xn
= (Ej R)(Z, Ei , Ei , Ej ) + R(Ej Z, Ei , Ei , Ej ) + R(Z, Ei , Ei , Ej Ej ) .
i=1
Since
rhEj Z, Ej i + rhZ, Ej Ej i
= Ric(Ej Z, Ej ) + Ric(Z, Ej Ej )
Xn
= R(Ej Z, Ei , Ei , Ej ) + R(Z, Ei , Ei , Ej Ej ) ,
i=1
we obtain
n
X
Ej (r)hZ, Ej i = (Ej R)(Z, Ei , Ei , Ej ).
i=1
298 4. Manifolds
On the other hand, taking the trace of the equation Ric = rg, we obtain
s = r dim M , in particular dim M Z(r) = Z(s). As dim M 3, this implies
that Z(r) = Z(s) = 0.
Thus, the unknown coefficients Gi can be found by solving the linear system
of equations
n
X
i (f ) = Gj gij , (i = 1, . . . , n).
j=1
hence
n
X
grad f = i (f )g ij j .
i,j=1
thus
X Y, Z) = e2f X(g(Y, Z)) + 2X(f )e2f g(Y, Z)+
2e2f g(
+ e2f Y (g(X, Z)) + 2Y (f )e2f g(X, Z)
e2f Z(g(X, Y )) 2Z(f )e2f g(X, Y )+
+ e2f (g([X, Y ], Z) g([X, Z], Y ) g([Y, Z], X)).
Proposition 4.11.4. The Hesse form is a tensor field indeed, that is,
Hesse(f )(X, Y ) is a bilinear function of X and Y over smooth functions.
If is torsion free, then the Hesse form is symmetric. At a critical point p
of f , i.e., at a point where the differential (df )p vanishes, Hesse(f )p does not
depend on the connection.
and
The equation
Hesse(f )p (Xp , Yp ) = Xp (Y (f ))
Now we are ready to compute the relation between the curvature tensor R
of the metric g = e2f g and the curvature tensor R of g.
From equation (4.20)
X
Y Z =
X (Y Z + Y (f )Z + Z(f )Y g(Y, Z) grad(f ))
= X (Y Z + Y (f )Z + Z(f )Y g(Y, Z) grad(f )+)
+ X(f )(Y Z + Y (f )Z + Z(f )Y g(Y, Z) grad(f ))+
+ ((Y Z)(f ) + Y (f )Z(f ) + Z(f )Y (f ) g(Y, Z)|| grad(f )||2 )X
(g(X, Y Z) + Y (f )g(X, Z) + Z(f )g(X, Y ) g(Y, Z)X(f )) grad(f )
}
X
Y Z = X Y Z + X(Y (f ))Z + Y (f )X Z + X(Z(f ))Y
+ Z(f )X Y g(X Y, Z) grad(f ) g(Y, X Z) grad(f )
}
g(Y, Z)X (grad(f )) + X(f )Y Z + X(f )Y (f )Z
+ X(f )Z(f )Y X(f )g(Y, Z) grad(f ) +
+ (Y Z)(f )X + Y (f )Z(f )X + Z(f )Y (f )X g(Y, Z)|| grad(f )||2 X
+ ( g(X, Y Z) Y (f )g(X, Z) Z(f )g(X, Y )
+ g(Y, Z)X(f ) ) grad(f )
This is the transformation rule of the curvature tensor under conformal trans-
formations of the metric. To produce from this a transformation rule for the
302 4. Manifolds
Riemannian curvature tensor, take the inner product of both sides with a
fourth vector field W with respect to g. This produces
e2f R(X, Y, Z, W )
= R(X, Y, Z, W ) + XY X(Z(f ))g(Y, W ) + g(X, Z)Y (W (f ))
Let us flip the role of X and Y in those terms, where there is no X in the
inner product g(., .). Inside the antisymmetrizer XY , this can be done with
a sign change. This results in the formula
e2f R(X, Y, Z, W )
= R(X, Y, Z, W ) + XY Y (Z(f ))g(X, W ) + g(X, Z)Y (W (f ))
e2f R(X, Y, Z, W ))
= R(X, Y, Z, W )+
+ ZW XY Y (Z(f ))g(X, W ) + (Y Z)(f )g(X, W )+
|| grad(f )||2
+ Z(f )Y (f )g(X, W ) g(Y, Z)g(X, W )
2
= R(X, Y, Z, W )
!
|| grad(f )||2
+ df df Hesse(f ) g g (X, Y, Z, W ).
2
C is called the type (3, 1) conformal Weyl tensor of the Riemannian manifold.
As a corollary of the previous theorem, we see that the type (3, 1) Weyl tensor
fields of conformally equivalent metrics are equal.
Definition 4.11.6. A Riemannian manifold is called flat if its curvature ten-
sor vanishes. A Riemannian manifold is locally conformally flat if each point
has an open neighborhood over which the metric is conformally equivalent to
a flat one.
It can be proved that
(1) a flat Riemannian manifold is locally isometric to a Euclidean space;
304 4. Manifolds
4.12 Geodesics
We define the length of a smooth curve : [a, b] M lying on a Riemannian
manifold (M, h , i) to be the integral
Z b p
l() = h 0 (t), 0 (t)idt.
a
h 0 , 0 i0 = h 0 0 , 0 i + h 0 , 0 0 i = 0.
Applying the existence and uniqueness theorem for ordinary differential equa-
tions one obtains the following.
Proposition 4.12.4. For any point p on a Riemannian manifold M and
for any tangent vector X Tp M , there exists a unique maximal geodesic
defined on an interval containing 0 such that (0) = p and 0 (0) = X.
If the maximal geodesic through a point p with initial velocity X is defined
on an interval containing [, ] then there is a neighborhood U of X in the
tangent bundle such that every maximal geodesic started from a point q with
initial velocity Y Tq M is defined on [, ].
Since a geodesic with zero initial speed can be defined on the whole real
straight line, for each point p on the manifold one can find a positive such
that for every tangent vector X Tp M with kXk < , the geodesic defined
by the conditions (0) = p, 0 (0) = X can be extended to the interval [0, 1].
The following notation will be convenient. Let X Tp M be a tangent vector
and suppose that there exists a geodesic : [0, 1] M satisfying the con-
ditions (0) = p, 0 (0) = X. Then the point (1) M will be denoted by
expp (X) and called the exponential of the tangent vector X.
Using the fact that for any c R, the curve t 7 (ct) is also a geodesic we
see that the geodesic is described by the formula
(t) = expp (tX).
4.12. Geodesics 307
As we have observed, expp (X) is defined provided that kXk is small enough.
In general however, expp (X) is not defined for large vectors X. This motivates
the following.
Definition 4.12.5. A Riemannian manifold is geodesically complete if for all
p M , expp (X) is defined for all vectors X Tp M .
This is clearly equivalent to the requirement that maximal geodesics are de-
fined on the whole real line R.
Proposition 4.12.6. For a fixed point p M , the exponential map expp
is a smooth map from an open neighborhood of 0 Tp M into the mani-
fold. Furthermore, the restriction of it onto a (possibly even smaller) open
neighborhood of 0 Tp M is a diffeomorphism.
of the radius. For small radii however, the intrinsic spheres are diffeomorphic
to the ordinary spheres in Rn , and what is more, we have the following.
Theorem 4.12.7. The normal parameterization of a manifold about a point
p maps the sphere about the origin with radius r, provided that it is contained
in the domain of the parameterization, diffeomorphically onto the intrinsic
sphere centered at p with radius r.
We prove this theorem later.
0 (b) 0 (a)
l0 (0) = 1 (0, b), 0 1 (0, a), 0 .
k (b)k k (a)k
0
D E
Observe, that the function t 7 1 (0, t), k 0 (t)k
(t)
is a primitive function
(antiderivative) of the function to be integrated. Indeed, the derivative of
this function is
0 (t) 0 (t)
d
1 (0, t), 0 = 2 1 (0, t), 0
dt k (t)k k (t)k
0
+ 1 (0, t), 0 (t) ,
k 0 k
but the second term on the right-hand side is zero since is geodesic. Con-
sequently,
b
0 (t)
Z
0
l (0) = 2 1 (0, t), 0
k (t)k
a
0 (b) 0 (a)
= 1 (0, b), 0 1 (0, a), 0 .
k (b)k k (a)k
0 (1) 0 (0)
0
0 = l (0) = 1 (0, 1), 0 1 (0, 0), 0 .
k (1)k k (0)k
0 (1)
0
0= (expp ) (0), 0 ,
| (1)|
Now we are ready to prove the theorem saying that Sr is a sphere in the
intrinsic geometry of the manifold. It is clear that d(p, q) r for any point
q on Sr , since the radial geodesic from p to r has length r, so all we need is
the following.
Theorem 4.12.11. If : [a, b] M is an arbitrary curve connecting p to a
point of Sr , then its length is r.
Proof. We may suppose without loss of generality that (b) is the only in-
tersection point of the curve with Sr and (t) 6= p for t > a. Then there is
a unique curve in the tangent space Tp M such that = expp . Let N
denote the vector field on Tp M \ {0} that is the gradient vector field of the
function f : X 7 kXk on Tp M , and therefore consists of unit vectors per-
pendicular to the spheres centered at the origin. The theorem above shows
that the derivative of the exponential map takes N into a unit vector field N
on M , perpendicular to the sets St .
We can estimate the length of a curve as follows
Z b Z b
l() = | 0 ( )|d h 0 ( ), N(( ))id.
a a
TX (Tp M ) = RN(X) TX S
at X = ( ). Therefore,
and
Z b Z b
h 0 ( ), N(( )id = (f )0 ( )d = k(b)k k(a)k = r.
a a
The proof also shows that the equality l() = r holds only for curves perpen-
dicular to the spheres S .
Exercise 4.12.12. Show that such curves are pre-geodesics.
Theorem 4.12.13. A smooth curve : [a, b] M parameterized by the
natural parameter in a Riemannian manifold is geodesic if and only if there
is a positive such that for any two values t1 , t2 [a, b] such that |t1 t2 | < ,
the restriction of onto [t1 , t2 ] is a curve of minimal length among curves
joining (t1 ) to (t2 ).
4.13. Applications to Hypersurface Theory 311
Exercise 4.12.15. Show that great circles on the sphere and helices on a
cylinder are pre-geodesics.
1
2
for k < n and n = det
..
,
.
n
where the last matrix is a matrix whose rows are the coordinates of the deriva-
tives i with respect to an arbitrary positively oriented orthonormal frame.
4.13. Applications to Hypersurface Theory 313
(The matrix depends on the choice of the frame, but its determinant does
not.)
where (x(t), y(t)) are the coordinates of (t) with respect to the standard
basis. The intrinsic counterpart of 1 is what we call geodesic curvature.
k 0 2 k
g = ,
k 0 k3
thus,
Since N(t) is orthogonal to both 0 (t) and 2 (t), the bivector 0 (t) N(t)
is orthogonal to 0 (t) 2 (t) by formula (1.9). Thus, by the Pythagorean
theorem,
where u 7 (x(u), 0, z(u)) is the generatrix in the (x, z)-plane. The group
of rotations about the z axis acts on the surface of revolution by isometries.
This isometry group is the flow of the vector field
where (t) is the angle between rv (u(t), v(t)) and 0 (t), is constant. As rv
is tangent to the circles of latitude, (t) is the angle at which the geodesic
curve intersects the circle of latitude through (t). It is also clear from
the formula for rv that krv (u, v)k = |x(u)| is the distance of the point r(u, v)
from the z-axis. Since k 0 k is constant for any geodesic curve, we obtain the
following corollary of the theorem for surfaces of revolution.
Corollary 4.13.7 (Clairaut Relation for Surfaces of Revolution). If is a
non-constant geodesic curve on a surface of revolution, r(t) is the distance of
(t) from the axis of the surface, (t) is the angle at which crosses the circle
of latitude through (t), then the function t 7 r(t) cos((t)) is constant.
4.13. Applications to Hypersurface Theory 317
n
X
dei = ij ej . (4.23)
i=1
We can arrange the 1-forms ij into a matrix using i as the row index and
j as the column index. can be thought of as a matrix valued 1-form as
j
well, which assigns to a tangent vector v Tp M the matrix i,p (v) .
The one forms i and ij are not independent of one another. The identity
d d = 0 for the exterior differentiation of differential forms yields some
318 4. Manifolds
which gives
n1
X n1
X
d j i ij = 0 for 1 j n 1, and i in = 0. (4.24)
i=1 i=1
Thus, the first fundamental form can be expressed by the differential forms
i .
To find the relation between the forms ij and the previously introduced
notions for parameterized hypersurfaces, let us start with the equation
X Y = X Y + II(X, Y )N,
where X and Y are tangential vector fields of M , and are the Levi-Civita
connections of Rn and M respectively, II is the second fundamental form. It
is clear that en = N, assume en is chosen to be equal to N. Applying this
formula for X = ei and Y = ej for 1 i, j n 1, we obtain
n1
X
ej ei = ik (ej )ek ,
k=1
and
II(ei , ej ) = in (ej ). (4.25)
It is known that the Levi-Civita connection is determined by the first funda-
mental form. Thus, there should be a formula expressing ij for 1 i, j
n 1 in terms of the forms k (1 k n 1). Let us find this formula.
The forms pk form a basis in the dual space Tp M , so we can write the 1-
Pn1
forms ij as their linear combinations ij = k=1 bjik k , where the coefficients
bjik = ij (ek ) are smooth functions on M . Substituting into (4.24) we obtain
a system of linear equations
X
d j = (bjik bjki ) i k for 1 j n 1 (4.26)
1i<kn1
and X
0= (bnik bnki ) i k .
1i<kn1
equations
1
bjik = (d i (ek , ej ) + d k (ej , ei ) d j (ei , ek )),
2
therefore,
n1
X 1
ij = (d i (ek , ej ) + d k (ej , ei ) d j (ei , ek )) k for 1 i, j n 1.
2
k=1
X e1 = 12 (X)e2 ,
X e2 = 21 (X)e1 .
d12 = 11 12 + 12 22 + 13 32 = 13 23 .
13 23 = f 1 2
According to equation (4.25), the matrix on the right-hand side is the matrix
of the second fundamental form of M with respect to the orthonormal basis
4.13. Applications to Hypersurface Theory 321
K0 (p) = 1,
K1 (p) = 1 (p) + + n1 (p),
..
.
X
Ki (p) = j1 (p) j2 (p) ji (p), (4.27)
1ji <j2 <<ji n1
4.13. Applications to Hypersurface Theory 323
..
.
Kn1 (p) = 1 (p) 2 (p) n1 (p)
Proof. As the derivative of xr at p is the linear map (ITp M rLp ), the pull-
back of the volume form of Mr to M is det(ITx(.) M rLx(.) )1 n1 .
For this reason,
Z
n1 (Dr ) = | det(ITp M rLp )|d(p),
D
where the integral is taken with respect to the volume measure of M . The
integrand can be expressed with the help of the characteristic polynomial of
Lp as
n1
Y 1
1
det(ITp M rLp )) = rn1 det ITp M Lp ) = rn1 i (p)
r i=1
r
n1 n1
X 1 X
= rn1 (1)i Ki (p) = (1)i Ki (p)ri .
i=0
rn1i i=0
as we wanted to prove.
The images of these tangent vectors under the derivative map of h are e1 (p)
sLp (e1 (p)), . . . , en1 (p)sLp (en1 (p)) and en (p). These vectors are linearly
independent if 1/s is not a principal curvature of M at p. If M has a positive
principal curvature at a point p of D, then let
n
(1)k1
X Z
= Kk1 (p)d(p) rk ,
k D
k=1
as it was to be proven.
Exercise 4.13.14. Prove that the above proposition implies Steiners for-
mula for the volume of parallel bodies of a convex body K, the boundary
M = K of which is a smooth hypersurface in Rn . Show that we get the
following explicit expression for the coefficients of the polynomial:
n Z
X 1
n (B(K, r)) = n (K) + |Kk1 (p)|d(p) rk .
k D
k=1
1
Z Z
Kn2 d = pn Kn1 d,
n1 M M
j=1 j=1
Pn p j + i , if i n 1,
j=1 j i
= P n j
j=1 pj n , if i = n.
n1
X
= (1)i+1 pi 1n n
cin n1 ,
i=1
where the hat above in means that in is omitted. Compute the differential
of . Clearly,
n1
X
d = (1)i+1 dpi 1n n
cin n1 +
i=1
n1
X X
+ (1)i+j pi 1n djn n
cin n1 +
i=1 1j<i
n1
X X
+ (1)i+j1 pi 1n
cin djn n1
n
,
i=1 i<jn1
4.13. Applications to Hypersurface Theory 327
hence
n1
XX n
d = (1)i+1 pj ij 1n n
cin n1 +
i=1 j=1
n1
X
+ (1)i+1 i 1n n
cin n1 +
i=1
n1
X n
X X
+ (1)i+j pi 1n (jk kn ) n
cin n1 +
i=1 1j<i k=1
n1
X X n
X
+ (1)i+j1 pi 1n
cin (jk kn ) n1
n
.
i=1 i<jn1 k=1
i=1 i<jn1
n1
X X
= (1)(i+j)+(j1)+(ij1) pi ji 1n n
cjn n1 +
i=1 1j<i
n1
X X
+ (1)(i+j1)+(j2)+(ji1) pi ji 1n n
cjn n1
i=1 i<jn1
X n1
n1 X
= (1)j pi ji 1n n
cjn n1 .
i=1 j=1
Pn1
The forms , pn 1n n1 n
and i=1 (1)
i+1 i
1n
cin
n
n1 are natural forms on M in the sense that they do not depend
on the choice of the orthonormal frame. Indeed, if v1 , . . . , vn2 Tp M ,
then p (v1 , . . . , vn2 ) is the signed volume of the parallelepiped spanned
Pn1
by the orthogonal projection (p) = i=1 pi (p)ei (p) of p onto Tp M and
Lp (v1 ), . . . , Lp (vn2 ), where Lp is the Weingarten map at p. The differential
form pn 1n n1 n
is the volume form of M multiplied by the product of
the GaussKronecker curvature Kn1 and the signed distance pn of the origin
from the tangent plane. Naturality of the third form follows from equation
(4.28) and naturality of the first two forms. Computing the third form using
a frame, whose vectors point in principal directions at a point, we see easily
that
n1
X
(1)i+1 i 1n n
cin n1 = Kn2 1 n1 .
i=1
1
Z Z
Kn2 d = pn Kn1 d
n1 M M
we wanted to show.
Hint: Apply Minkowskis formula with different choices of the origin of the
coordinate system.
In the special case, when M is strictly convex and the Gauss map en : M
Sn1 is a diffeomorphism between M and Sn1 , the second integral can be
rewritten as an integral on the sphere.
4.13. Applications to Hypersurface Theory 329
(1)n2
Z
nn
nWn1 (K) = n1 V1 (K) = Kn2 d = w(K).
n1 M 2
330 4. Manifolds
Proof. The easier part of the statement is that if the bending extends to an
n
orientation preserving isometry of the space, then in = 0 i . We leave the
details of this direction to the reader.
Assume now that in = in for all 1 i n 1. Choose a point p U
and an orientation preserving isometry of Rn such that (p) = h(p) and
Tp (ei (p)) = ei (h(p)) for 1 i n 1. Then h = 1 h is a bending of
M which fixes p and the frame e1 (p), . . . , en1 (p). Denote by e1 , . . . , en the
frame induced on h(U ) in the same way as the frame e1 , . . . , en was obtained
on h(U ).
Let : [0, 1] U be a smooth curve starting from (0) = p. Consider the
2
collection of the vector valued functions (e1 , . . . , en ) : [0, 1] Rn . It
4.13. Applications to Hypersurface Theory 331
The same system of differential equations is satisfied also by the vector valued
2
function (e1 h , . . . , en h ) : [0, 1] Rn . Both solutions start from
the same initial vectors at 0, therefore, by the uniqueness of the solutions of
ordinary differential equations with given initial value, ei = ei h for
i = 1, . . . , n. From this we obtain
n1
X n1
X
0 ( ) = i ( 0 ( ))ei (( )) = i ( 0 ( ))ei (h(( ))) = (h )0 ( ),
i=1 i=1
Proof. The proof follows the same line of computation as the proof of Min-
kowskis formula, the only difference is that some terms are marked with a
bar to indicate that they come from the bent surface M . Due to the similarity
of the two proofs, we skip some details. Define the differential (n 2)-form
332 4. Manifolds
on M by the equation
n1
X
= (1)i+1 pi 1n n
cin n1 ,
i=1
hence
n1
XX n
d = (1)i+1 pj ij 1n n
cin n1 +
i=1 j=1
n1
X
+ (1)i+1 i 1n n
cin n1 +
i=1
n1
X n
X X
+ (1)i+j pi 1n (jk kn ) n
cin n1 +
i=1 1j<i k=1
n1
X X n
X
+ (1)i+j1 pi 1n
cin (jk kn ) n1
n
Taking into account that M and M has the same Gaussian curvature at
corresponding points by Theorema Egregium, we get
13 23 = K2 1 2 = 13 23 ,
and Z Z Z
1
Hd Hd = p3 (13 13 ) (23 23 ).
M M 2 M
f 1 2 = (13 13 ) (23 23 )
334 4. Manifolds
is the matrix of the second fundamental form and the Weingarten map of M
at h(q) with respect to the basis (Tq h(v1 ), Tq h(v2 )), and it is also negative
definite, and has determinant K2 (q).
Consider the quadratic polynomial
must also be true, therefore we must have equality in both inequalities. Equal-
ity implies f 0, which can hold only if the matrices B and B are equal for
each q M and each choice of v1 v2 Tq M , that is, if 13 = 13 and 23 = 23 .
However this condition is equivalent to the extendability of h to an orientation
preserving isometry of the space R3 .
Bibliography
337
Index
338
Index 339
parameterized reparameterization, 83
curve, 82 by arc length, 85
hypersurface, 141 Ricci
permutation group, 2 curvature, 287
permutation rule for mixed product, tensor, 286, 292
32 RiemannChristoffel curvature tensor,
Plucker embedding, 16 280
Poincare lemma, 253 Riemannian
Poincares half-plane model of the hy- curvature tensor
perbolic plane, 160 of a hypersurface, 159
pointed topological space, 60 of a Riemannian manifold, 280
polar coordinates, 85 manifold, 264
position vector, 35 metric, 265
positive definite symmetric bilinear rotation number, 125
function, 21 theorem, 125
positively oriented basis, 11
pre-geodesic curve, 306 Sards lemma, 72
principal scalar curvature, 288, 294
curvature, 148 Schurs theorem, 285
principal axis theorem, 27 second fundamental form, 146
principal direction, 148 section of a bundle, 235
product sectional curvature, 283
measure, 69 self-adjoint linear endomorphism, 26
of manifolds, 203 separation axioms, 5658, 58
topology, 54 shape operator, 146
projection of a bundle, 233 shpere
projective plane, 210 in Euclidean space, 46
pseudosphere, 165 -algebra, 68
quadratic form of a bilinear function, sign
26 of a permutation, 8
quermassintegral, 321 simple arc, 81
sine-Gordon equation, 187
radius of curvature, 118 singular point
rank of a linear map, 6 of a curve, 121
rectifiable curve, 83 smooth singular
rectifying plane, 131 chain, 254
regular homology space, 256
domain, 260 simplex, 254
parameterized curve, 85 Sorgenfrey plane, 58
parameterized hypersurface, 142 space form, 284
reparameterization, 83 spaces of constant curvature, 284
topological space, 57 special orthogonal
Index 345
variation
of a curve, 308
of a hypersurface, 195
compactly supported , 195
vector bundle, 233
vector field
along a curve, 98, 269
binormal , 131
principal normal , 131
unit normal in plane, 117
unit tangent , 117, 131