Vector Fields
Vector Fields
Keijo Ruohonen
2013
Contents
16 II MANIFOLD
16 2.1 Graphs of Functions
17 2.2 Manifolds
18 2.3 Manifolds as Loci
22 2.4 Mapping Manifolds. Coordinate-Freeness
23 2.5 Parametrized Manifolds
30 2.6 Tangent Spaces
34 2.7 Normal Spaces
36 2.8 Manifolds and Vector Fields
38 III VOLUME
38 3.1 Volumes of Sets
41 3.2 Volumes of Parametrized Manifolds
43 3.3 Relaxed Parametrizations
46 IV FORMS
46 4.1 k-Forms
52 4.2 Form Fields
54 4.3 Forms and Orientation of Manifolds
59 4.4 Basic Form Fields of Physical Fields
83 VI POTENTIAL
83 6.1 Exact Form Fields and Potentials
87 6.2 Scalar Potential of a Vector Field in R3
92 6.3 Vector Potential of a Vector Field in R3
96 6.4 Helmholtz’s Decomposition
97 6.5 Four-Potential
98 6.6 Dipole Approximations and Dipole Potentials
i
ii
124 References
126 Index
Foreword
These lecture notes form the base text for the course ”MAT-60506 Vector Fields”. They are
translated from the older Finnish lecture notes for the course ”MAT-33351 Vektorikentät”, with
some changes and additions.
These notes deal with basic concepts of modern vector field theory, manifolds, (differential)
forms, form fields, Generalized Stokes’ Theorem, and various potentials. A special goal is a
unified coordinate-free physico-geometric representation of the subject matter. As a sufficient
background, basic univariate calculus, matrix calculus and elements of classical vector analysis
are assumed.
Classical vector analysis is one of the oldest areas of mathematical analysis.1 Modelling
structural strength, fluid flow, thermal conduction, electromagnetics, vibration etc. in the three-
space needs generalization of the familiar concepts and results of univariate calculus. There
seem to be a lot of these generalizations. Indeed, vector analysis—classical as well as modern—
has been largely shaped and created by the many needs of physics and various engineering
applications. For the latter, it is central to be able to formulate the problem as one where fast
and accurate numerical methods can be readily applied. This generally means specifying the
local behavior of the phenomenon using partial differential equations (PDEs) of a standard type,
which then can be solved globally using program libraries. Here PDEs are not extensively dealt
with, mainly via examples. On the other hand, basic concepts and results having to do with
their derivation are emphasized, and treated much more extensively.
Modern vector analysis introduces concepts which greatly unify and generalize the many
scattered things of classical vector analysis. Basically there are two machineries to do this:
1
There is a touch of history in the venerable Finnish classics TALLQVIST and VÄISÄLÄ , too.
iii
manifolds and form fields, and Clifford’s algebras. These notes deal with the former (the latter
is introduced in the course ”Geometric Analysis”).
The style, level and order of presentation of the famous textbook H UBBARD & H UBBARD
have turned out to be well-chosen, and have been followed here, too, to an extent. Many tedious
and technical derivations and proofs are meticulously worked out in this book, and are omitted
here. As another model the more advanced book L OOMIS & S TERNBERG might be mentioned,
too.
Keijo Ruohonen
”One need only know that geometric objects in spacetime
are entities that exist independently of coordinate systems
or reference frames.”
Chapter 1
1
CHAPTER 1. POINT. VECTOR. VECTOR FIELD 2
In particular −~0 = ~0. The unit of a physical vector remains the same in this operation.
In particular we define
~r + ~0 = ~0 + ~r = ~r
2
See e.g. G IBBINGS , J.C.: Dimensional Analysis. Springer–Verlag (2011) for many more details of dimen-
sional analysis.
CHAPTER 1. POINT. VECTOR. VECTOR FIELD 3
and
~r + (−~r ) = (−~r ) + ~r = ~0.
Only vectors sharing a unit can be physically added, and the unit of the sum is this unit.
These are geometrically fairly obvious. Associativity implies that long sums may be
parenthesized in any (correct) way, or written totally without parenteheses, without the
result changing.
The difference of the vectors ~r and ~s is the vector
~r − ~s = ~r + (−~s ).
For physical vectors the units again should be the same in this operation.
−→
• If ~r = hP Qi is a vector and λ a positive scalar, then λ~r is the vector obtained as follows:
If the physical scalar λ and physical vector ~r have their physical units, the unit of λ~r is
their product. With a bit of work the following laws of calculation can be (geometrically)
verified:
λ1 (λ2~r) = (λ1 λ2 )~r ,
(λ1 + λ2 )~r = λ1~r + λ2~r and
λ(~r1 + ~r2 ) = λ~r1 + λ~r2 ,
In particular d(~r, ~0 ) = |~r |. This distance also satisfies the triangle inequality
In particular
~r • ~0 = ~0 • ~r = 0 and ~r • ~r = |~r |2 .
is the (scalar) projection of ~r on ~s. (The projection of a zero vector is of course always
zero.)
• The cross product (or vector product) of the vectors ~r and ~s is the vector ~r × ~s given by
the following. First, if ~r or ~s is ~0, or ∠(~r, ~s ) is 0 or π, then ~r × ~s = ~0. Otherwise ~r × ~s is
the unique vector ~t satisfying
~r × ~s = −(~s × ~r ),
is the area of the parallelogram with lengths of sides |~r | and |~s | and spanning angle
∠(~r, ~s ). If ~r and ~s are physical vectors, then the unit of the cross product ~r × ~s is the
product of their units.
~r • (~s × ~t )
There being no danger of confusion, the scalar product is usually written without paren-
theses as ~r • ~s × ~t.
Scalar triple product is cyclically symmetric, i.e.
~r • ~s × ~t = ~s • ~t × ~r = ~t • ~r × ~s.
By this and commutativity of scalar product, the operations • and × can be interchanged,
i.e.
~r • ~s × ~t = ~r × ~s • ~t.
Geometrically it is easily noted that the scalar triple product ~r • ~s × ~t is the volume of the
parallelepiped with edges incident on the vertex P being the vectors
−→ −→ −→
~r = hP Ri , ~s = hP Si and ~t = hP T i,
with a positive sign if ~r, ~s, ~t is a right-handed system, and a negative sign otherwise. (As
special cases situations where the scalar triple product is = 0 should be included, too.)
Cyclic symmetry follows geometrically immediately from this observation.
The triple vector product expansions (also known as Lagrange’s formulas) are
These are somewhat difficult to prove geometrically, proofs using coordinates are easier.
~
Exactly as for points we can define an ~r-centered open ball B(R, ~r ) of radius R for vectors,
open and closed sets of vectors, and boundaries, closures and interiors of sets of vectors, but a
geometric intuition is then not as easily obtained.
x, . . . and r0 , r1 , . . . , etc. The coordinate point corresponding to the origin of the system is
0 = (0, 0, 0).
A coordinate system is determined by the corresponding coordinate function which maps
geometric points to the triples of R3 . We denote coordinate functions by small boldface Greek
letters, and their components by the corresponding indexed letters. If the coordinate function is
κ, then the coordinates of the point P are
κ(P ) = κ1 (P ), κ2 (P ), κ3(P ) .
then
p
d(P, Q) =
κ(P ) − κ(Q)
= (x1 − x2 )2 + (y1 − y2 )2 + (z1 − z2 )2 .
A coordinate function κ also gives a coordinate representation for vectors. The coordinate
−→
version of the vector ~r = hP Qi is
T κ 1 (Q) − κ1 (P )
κ(~r ) = κ(Q) − κ(P ) = κ2 (Q) − κ2 (P ) .
κ3 (Q) − κ3 (P )
Note that this is a column array. As is easily seen, this presentation does not depend on choices
of the representative directed line segments. In particular the zero vector has the representation
−→ −→
κ(~0 ) = (0, 0, 0)T = 0T . Also the distance of the vectors ~r = hP Qi and ~s = hP Ri (note the
choice of representatives) is obtained from the Euclidean norm of R3 :
d(~r, ~s ) =
κ(~r ) − κ(~s )
=
κ(Q) − κ(R)
= d(Q, R).
The vectors i, j, k are the basis vectors and the vector r is used as a generic variable vector. In
the background there of course is a fixed coordinate system and a coordinate function. The row
array versions of these are also used as coordinate points.
Familiar operations of coordinate vectors now correspond exactly to the geometric vector
operations in the previous section. Let us just recall that if
a1 a2
κ(~r ) = r = b1 and κ(~s ) = s = b2 ,
c1 c2
then
~r • ~s = r • s = a1 a2 + b1 b2 + c1 c2
CHAPTER 1. POINT. VECTOR. VECTOR FIELD 7
and
b1 c2 − b2 c1
κ(~r × ~s ) = r × s = c1 a2 − c2 a1 .
a1 b2 − a2 b1
The latter is often given as the more easily remembered formal determinant
i a1 a2
j b1 b2 ,
k c1 c2
κ∗ (P ) = κ(P )Q + b
and
κ(P ) = κ∗ (P )QT − bQT .
−→
The coordinate representation of a vector ~r = hP Qi is transformed similarly:
T T
κ∗ (~r ) = κ∗ (Q) − κ∗ (P ) = κ(Q)Q + b − κ(P )Q − b
T
= QT κ(Q) − κ(P ) = QT κ(~r )
and
κ(~r ) = Qκ∗ (~r ).
Note that b is the representation of the origin of the ”old” coordinate system in the ”new”
system, and that the columns of QT are the representations of the basis vectors i, j, k of the ”old”
coordinate system in the ”new” system. Similarly −bQT is the representation of the origin of
the ”new” coordinate system in the ”old” system, and the columns of Q are the representations
of the ”new” basis vectors i∗ , j∗ , k∗ in the ”old” system.
If the point of action is clear from the context, or is irrelevant, it is often omitted and only
the vector component of the pair is used, usually in a coordinate representation. This is what
we will mostly do in the sequel.
A vector field is a function mapping a point P to a tangent vector P, F~ (P ) (note the point
of action). Mostly we denote this just by F~ . A vector field may not be defined for all points of
the geometric space, i.e., its domain of definition may be smaller.
In the coordinate representation given by the coordinate function κ we denote
thus coordinate vector fields are denoted by capital boldface letters. Note that in the coordinate
transform
r∗ = rQ + b (i.e. κ∗ (P ) = κ(P )Q + b)
the vector field F = κ(F~ ) (that is, its representation) is transformed to the the field F∗ = κ∗ (F~ )
given by the formula
F∗ (r∗ ) = QT F (r∗ − b)QT .
A vector field may of course be defined in fixed coordinates in one way or another, and then
taken to other coordinate systems using the transform formula. On the other hand, definition
of a physico-geometric vector field cannot possibly depend on any coordinate system, the field
exists without any coordinates, and will automatically satisfy the transform formula.
A coordinate vector field is the vector-valued function of three arguments familiar from
basic courses of mathematics
F1 (r)
F(r) = F2 (r) ,
F3 (r)
with its components. Thus all operations and concepts defined for these apply, limits, continuity,
differentiability, integrals, etc.
A scalar field is a function f mapping a point P to a scalar (real number) f (P ), thus scalar
fields are denoted by italic letters, usually small. In the coordinate representation we denote
r = κ(P ) and just f (r) (instead of the correct f κ−1 (r) ). In the coordinate transform
r∗ = rQ + b (i.e. κ∗ (P ) = κ(P )Q + b)
A scalar field, too, can be defined in fixed coordinates, and then transformed to other coordinate
systems using the transform formula. But a physico-geometric scalar field exists without any
coordinate systems, and will automatically satisfy the transform formula.
Again a coordinate scalar field is the real-valued function of three arguments familiar from
basic courses of mathematics. Thus all operations and concepts defined for these apply, limits,
continuity, differentiability, integrals, etc.
An important observation is that all scalar and vector products in the previous section ap-
plied to vector and scalar fields will again be fields, e.g. a scalar field times a vector field is a
vector field.
CHAPTER 1. POINT. VECTOR. VECTOR FIELD 9
Using the chain rule5 we get the transforms of the derivatives in a coordinate transform
r = rQ + b:
∗
′
f ∗ ′ (r∗ ) = f (r∗ − b)QT = f ′ (r∗ − b)QT Q
and ′
F∗′ (r∗ ) = QT F (r∗ − b)QT = QT F′ (r∗ − b)QT Q.
Despite partial derivatives depending on the coordinate system used, differentiability itself
is coordinate-free: If a field has partial derivates in one coordinate system, it has them in any
other coordinate system, too. This is true for second order derivatives as well. And it is true for
continuity: Continuity in one coordinate system implies continuity in any other system. And
finally it is true for continuous differentiability: If a field has continuous partial derivatives (first
or second order) in one coordinate system, it has them in any other system, too. All this follows
from the transform formulas.
5
Assuming f and g differentiable, the familiar univariate chain rule gives the derivative of the composite func-
tion: ′
f g(x) = f ′ g(x) g ′ (x).
More generally, assuming F and G continuously differentiable (and given as column arrays), we get the derivative
of the composite function as ′
F G(r)T = F′ G(r)T G′ (r).
The arguments are here thought of as row arrays. The rule is valid in higher dimensions, too.
CHAPTER 1. POINT. VECTOR. VECTOR FIELD 10
The common differential operations for fields are the gradient (nabla) of a scalar field f , and
its Laplacian
∂2f ∂2f ∂f 2
∆f = ∇ • (∇f ) = ∇2 f = + + ,
∂x2 ∂y 2 ∂z 2
F1
and for a vector field F = F2 its divergence
F3
∇(λ1 f + λ2 g) = λ1 ∇f + λ2 ∇g and
∇ • (λ1 F + λ2 G) = λ1 ∇ • F + λ2 ∇ • G , etc.
∂f ∂f ∂f
(G • ∇)f = G • (∇f ) = G1 + G2 + G3
∂x ∂y ∂z
and (taking F1 , F2 , F3 to be scalar fields)
(G • ∇)F1
(G • ∇)F = (G • ∇)F2 = F′ G.
(G • ∇)F3
These are both coordinate-free and hence fields. Coordinate-freeness of (G • ∇)F follows from
the nabla rules below (or from the coordinate-freeness of F′ G).
Let us tabulate the familiar nabla-calculus rules:
1 1
(ii) ∇ = − 2 ∇f
f f
(iii) ∇ • (f G) = ∇f • G + f ∇ • G
(iv) ∇ × (f G) = ∇f × G + f ∇ × G
(v) ∇ • (F × G) = ∇ × F • G − F • ∇ × G
(viii) (∇ × F) × G = (F′ − F′ T )G
(ix) ∇ • (∇ × F) = 0
(x) ∇ × ∇f = 0
In formulas (ix), (x), (xi) we assume F and f are twice continuously differentiable. These
formulas are all symbolical identities, and can be verified by direct calculation, or e.g. using the
Maple symbolic computation program.
CHAPTER 1. POINT. VECTOR. VECTOR FIELD 12
r∗ = rQ + b
we denote the nabla in the new coordinates by ∇∗ . Coordinate-freeness for the basic operators
then means the following:
1. ∇∗ f (r∗ − b)QT = QT ∇f (r) (gradient)
Subtracting b and multiplying by QT we move from the new coordinates r∗ to the old
ones, get the value of f , and then the gradient in the new coordinates. The result must be
the same when the gradient is obtained in the old coordinates and then transformed to the
new ones by multiplying by QT .
2. ∇∗ • QT F (r∗ − b)QT = ∇ • F(r) (divergence)
Subtracting b and multiplying by QT we move from the new coordinates r∗ to the old
ones, get F, transform the result to the new coordinates by multiplying by QT , and get
the divergence using the new coordinates. The result must remain the same when the
divergence is obtained in the old coordinates.
3. ∇∗ × QT F (r∗ − b)QT = QT ∇ × F(r) (curl)
Subtracting b and multiplying by QT we move from the new coordinates r∗ to the old
ones, get F, transform the result to the new coordinates by multiplying by QT , and get the
curl using the new coordinates. The result must be the same when the curl is obtained in
the old coordinates and then transformed to the new ones by multiplying by QT .
Theorem 1.1. Gradient, divergence, curl and Laplacian are coordinate-free. Furthermore, if
F and G are vector fields (and thus coordinate-free), then so is (G • ∇)F.
To show formula 2. we use the trace of the Jacobian. Let us recall that the trace of a square
matrix A, denoted trace(A), is the sum of the diagonal elements. A nice property of trace6 is
that if the product matrix AB is square—whence BA is square, too—then
trace(AB) = trace(BA).
Since
∂F1 ∂F2 ∂F3
trace(F′ ) = + + = ∇ • F,
∂x ∂y ∂z
6
Denoting A = (aij ) (an n × m matrix), B = (bij ) (an m × n matrix), AB = (cij ) and BA = (dij ) we have
n
X n X
X m m X
X n m
X
trace(AB) = ckk = akl blk = blk akl = dll = trace(BA).
k=1 k=1 l=1 l=1 k=1 l=1
CHAPTER 1. POINT. VECTOR. VECTOR FIELD 13
formula 2. follows:
∇∗ • F∗ (r∗ ) = trace F∗ ′ (r∗ ) = trace QT F′ (r)Q
= QT ∇ × F(r)
1
.
Note that q2 × q3 = q1 since the new coordinate system must be right-handed, too. The other
components are dealt with similarly.
Coordinate-freeness of the Laplacian follows directly from that of gradient and divergence
for scalar fields, and for vector fields from formula (xi).
Adding formulas (vi) and (vii) on both sides we get an expression for (G • ∇)F:
1
(G • ∇)F = ∇ × (F × G) + (∇ • F)G − (∇ • G)F + ∇(F • G)
2
+ (∇ × F) × G + (∇ × G) × F .
All six terms on the right hand side are coordinate-free and thus vector fields. The left hand side
(G • ∇)F then also is coordinate-free and a vector field. (This is also easily deduced from the
matrix form (G • ∇)F = F′ G, but the formula above is of other interest, too!)
Thus for the time derivatives we get the corresponding transform formulas
∂i ∗ ∗ ∂i ∗ T
f (r , t) = f (r − b)Q , t and
∂ti ∂ti
∂i ∗ ∗ T ∂
i
∗ T
F (r , t) = Q F (r − b)Q , t ,
∂ti ∂ti
which shows that they are fields.
In addition to the familiar partial derivative rules we get for the time derivatives e.g. the
following rules, which can be verified by direct calculation:
∂ ∂F ∂G
(1) (F • G) = •G+F•
∂t ∂t ∂t
∂ ∂F ∂G
(2) (F × G) = ×G+F×
∂t ∂t ∂t
∂ ∂f ∂F
(3) (f F) = F+f
∂t ∂t ∂t
∂ ∂F ∂G ∂H
(4) (F • G × H) = •G×H+F• ×H+F•G×
∂t ∂t ∂t ∂t
∂ ∂F ∂G ∂H
(5) F × (G × H) = × (G × H) + F × ×H +F× G×
∂t ∂t ∂t ∂t
Note that now the fields are stationary, time-dependence is only in the coordinate representation
and is a consequence of the moving coordinate system.
CHAPTER 1. POINT. VECTOR. VECTOR FIELD 15
Similarly for nonstationary fields f (r, t) and F(r, t) in a moving coordinate system we get
the representations
Now part of the time-dependence comes from the time-dependence of the fields, part from the
moving coordinate system.
”The intuitive picture of a smooth surface becomes analytic
with the concept of a manifold. On the small scale a manifold
looks like Euclidean space, so that infinitesimal operations
like differentiation may be defined on it.’
Chapter 2
MANIFOLD
a subset of Rk+m . The graph is often denoted—using a slight abuse of notation—as follows:
s = f(r) (r ∈ A).
Here r contains the so-called active variables and s the so-called passive variables. Above
active variables precede the passive ones in the order of components. In a graph we also allow a
situation where the variables are scattered. A graph is smooth,1 if f is continuously differentiable
in its domain of definition A. In the sequel we only deal with smooth graphs. Note that a graph
is specifically defined using a coordinate system and definitely is not coordinate-free.
A familiar graph is the graph of a real-valued
univariate function f in an open interval (a, b), i.e., z
2
the subset of R consisting of the pairs
z = f(x,y)
x, f (x) (a < x < b),
16
CHAPTER 2. MANIFOLD 17
Similarly we allow k = 0. Then f has no arguments, and so it is constant, and the graph
consists of one point.2 Again it is agreed that such graphs are also smooth.
In what follows we will need inverse images of sets. For a function g : A → B the inverse
image of the set C is the set
g−1 (C) = r g(r) ∈ C .
Note that this has little to do with inverse functions, indeed the function g need not have an
inverse at all, and C need not be included in B.
For a continuous function defined in an open set A the inverse image of an open set is always
open.3 This implies an important property of graphs:
Theorem 2.1. If a smooth graph of a function is intersected by an open set, then the result is
either empty or a smooth graph.
Proof. This is clear if the intersection is empty, and also if k = 0 (the intersection is a point) or
m = 0 (intersection of two open sets is open). Otherwise the intersection of the graph
s = f(r) (r ∈ A)
and the open set C in Rk+m is the graph
s = f(r) (r ∈ D),
where D is the inverse image of C for the continuous4 function
g(r) = r, f(r) .
2.2 Manifolds
A subset M of Rn is a k-dimensional manifold, if it is locally a smooth graph of some function
of k variables.5
”Locally” means that for every point p of M there is an open set Bp of Rn containing the
point p such that M ∩ Bp is the graph of some function fp of some k variables. For different
points p the set Bp may be quite different, the active variables chosen in a different way, always
numbering k, however, and the function fp may be very different.
The functions fp are called charts, and the set of all charts is called an atlas. Often a small
atlas is preferable.
Example. A circle of radius R centered in the origin is a 1-dimensional manifold of R2 , since
(see the figure below) its each point is in an open arc delineated by either black dots or white
dots, and these are smooth graphs of the functions
√ p
y = ± R2 − x2 and x = ± R2 − y 2
(atlas) in certain open intervals.
2
Here we may take A to be the whole space R0 , an open set.
3
This in fact is a handy definition of continuity. Continuity of g in the point r0 means that taking C to be an
arbitrarily small g(r0 )-centered open ball, there is in A a (small) r0 -centered open ball which g maps to C.
4
It is not exactly difficult to see that if f is continuous in the point r0 , then so is g, because
g(r) − g(r0 )
2 = kr − r0 k2 +
f (r) − f (r0 )
2 .
5
There are several definitions of manifolds of different types in the literature. Ours is also used e.g. in H UB -
BARD & H UBBARD and N IKOLSKY & VOLOSOV. They are also often more specifically called ”smooth mani-
folds” or ”differentiable manifolds”. With the same underlying idea so-called abstract manifolds can be defined.
CHAPTER 2. MANIFOLD 18
Example. A circle and a sphere are loci of this kind when we set the conditions
F (x, y) = R2 − x2 − y 2 = 0
and
F (x, y, z) = R2 − x2 − y 2 − z 2 = 0
(centered in the origin and having radius R). In the circle one of the variables is always active,
in the sphere two of the three variables.
Not all loci are manifolds. For instance, the locus of points of R2 determined by the condi-
tion
y − |x| = 0
is not, and neither is the locus of points satisfying the condition
y 2 − x2 = 0.
The former is not smooth in the origin, and the latter is not a graph of any single function in the
origin (but rather of two functions: y = ±x). Actually the condition
(y − x)2 = 0
does not determine a proper manifold either. The locus is the line y = x, but counted twice!
So, in the equation F(r, s) = 0 surely F then should be continuously differentiable, and
somehow uniqueness of solution should be ensured, too, at least locally. In classical analysis
there is a result really taylor-made for this, the so-called Implicit Function Theorem, long known
and useful in many contexts. Here it is used to make the transition from local loci6 to graphs. It
should be mentioned that in the literature there are many versions of the theorem,7 we choose
just one of them.
Implicit Function Theorem. Assume that the function F : S → Rn−k , where 0 ≤ k < n,
satisfies the following conditions:
1. The domain of definition S is an open subset of Rn .
2. F is continuously differentiable.
3. F′ is of full rank in the point p0 , i.e., the rows of F′ (p0 ) are linearly independent (whence
also some n − k columns are also linearly independent).
Denote by r variables other than the ones in s. By changing the order of variables, if necessary,
we may assume that p = (r, s), and especially p0 = (r0 , s0 ). Denote further the derivative of
F with respectto the variables in r by F′r , and with respect to the variables in s by F′s . (Thus
F′ = F′r F′s in block form.)
Then there is an open subset B of Rk containing the point r0 , and a uniquely determined
function f : B → Rn−k such that
6
This is not pleonasm, although it might seem to be since ’local’ could be constructed to mean ’relating to a
locus’ etc.
7
See e.g. K RANTZ , S.G. & PARKS , H.R.: The Implicit Function Theorem. History, Theory, and Applications.
Birkhäuser (2012).
CHAPTER 2. MANIFOLD 20
(ii) p0 = r0 , f(r0 ) ,
(iii) F r, f(r) = 0 in B,
−1
f ′ (r) = −F′s r, f(r) F′r r, f(r)
Proof. The proof is long and tedious, and is omitted here, see e.g. A POSTOL or H UBBARD &
H UBBARD or N IKOLSKY & VOLOSOV. The case k = 0 is obvious, however. Then r0 is the
empty vector and f is the constant function p0 . The derivative of f in item (iv) is obtained by
implicit derivation, i.e., applying the chain rule to the left hand side of the identity
F r, f(r) = 0
Using the Implicit Function Theorem (and Theorem 2.1) we immediately get definition of
manifolds using local loci:
Corollary 2.3. If for any point p0 in the subset M of Rn there is a subset S of of Rn and a
function F : S → Rn−k such that the conditions 1.– 4. in the Implicit Function Theorem are
satisfied, and the locus condition F(p) = 0 defines the set M ∩ S, then M is a k-dimensional
manifold of Rn .
The converse holds true, too, i.e., all manifolds are local loci:
Theorem 2.4. If M is a k-dimensional manifold of Rn and k < n, then for each point p of M
there is a set S and a function F : S → Rn−k such that the conditions 1.— 4. of the Implicit
Function Theorem are satisfied and the locus condition F(p) = 0 defines the set M ∩ S.
Proof. Let us just see the case k > 0. (The case where k = 0 is similar—really a special case.)
If M is a k-dimensional manifold of Rn , then locally in some open set containing the point p0
it is the graph of some continuously differentiable function f
s = f(r) (r ∈ A)
for some choice of the k variables of r (the active variables). Reordering, if necessary, we may
assume that the active variables precede the passive ones.
Choose now the set S to be the Cartesian product A × Rn−k , i.e.
S = (r, s) r ∈ A and s ∈ Rn−k ,
G(p) = c or G(p) − c = 0,
where c is a constant, define a manifold (with the given assumptions), the so-called level mani-
fold of G.
Representation of a manifold using loci is often called an implicit representation and the
representation using local graphs of functions—as in the original definition– is called an explicit
representation.9
Example. 2-dimensional manifolds in R3 are smooth surfaces. Locally such a surface is defined
as the set determined by a condition
F (x, y, z) = 0,
is of full rank, i.e., its two rows are linearly independent. Locally we have then the curve as the
intersection of the two smooth surfaces
so that r ∈ B.
9
There is a third representation, so-called parametric representation, see Section 2.5.
CHAPTER 2. MANIFOLD 22
(locus conditions, cf. the previous example). It may be noted that the curve of intersection of
two smooth surfaces need not be a smooth manifold (curve), for this the full rank property is
needed.10
Example. The condition
F (x, y) = yey − x = 0
defines a 1-dimensional manifold (a smooth curve, actually a graph) of R2 . On the other hand,
F ′ (x, y) = − 1, (1 + y)ey ,
so, except for the point (−1/e, −1), the corresponding local graph can be taken as f (y), y
where f is one of the so-called Lambert W functions11 W0 (green upper branch) or W−1 (red
lower branch), see the figure below (Maple).
Since here
√
y = xe−y and − ln 2 > −1/e,
2
√
exists, which may seem odd because 2 > 1.
Actually it has the value
√
W0 − ln 2
√ = 2.
− ln 2
Proof. The case k = n is clear. Then M is an open subset of Rn and its inverse image g−1 (M)
is an open subset of Rm , i.e. an m-dimensional manifold of Rm .
Take then the case k < n. Consider an arbitrary point r0 of g−1(M), i.e. a point such
that g(r0 ) ∈ M. Locally near the point p0 = g(r0 ) the manifold M can be determined as
a locus, by Theorem 2.4. More specifically, there is an open subset S of Rn and a function
F : S → Rn−k , such that the conditions 1.– 4. of the Implicit Function Theorem are satisfied.
For a continuous function defined in an open set the inverse image of an open set is open,
so g−1 (S) is open. The locus condition
F g(r) = 0
determines locally some part of the set g−1 (M). In the open set g−1 (S) the composite function
F ◦ g now satisfies the conditions 1.– 4. of the Implicit Function Theorem since (via chain rule)
its derivative is
(F ◦ g)′ (r0 ) = F′ g(r0 ) g′ (r0 ) = F′ (p0 )g′ (r0 ),
and it is of full rank. Thus, by Corollary 2.3, g−1 (M) is a manifold of Rm and its dimension is
m − (n − k) (the dimension of r minus the dimension of F).
So far we have not considered coordinate-freeness of manifolds. A manifold is always
expressly defined in some coordinate system, and we can move frome one system to another
using coordinate transforms. But is a manifold in one coordinate system also a manifold in any
other, and does the dimension then remain the same?
As a consequence of Theorem 2.5 the answer is positive. To see this, take a coordinate
transform
r∗ = rQ + b,
and choose m = n and
g(r∗ ) = (r∗ − b)QT
in the theorem. Then a manifold M in coordinates r∗ is the inverse image of the manifold in
coordinates r, and thus truly a manifold. Dimension is preserved as well. Being a manifold in
one coordinate system guarantees being a manifold in any other coordinates. ”Manifoldness” is
a coordinate-free property.
γ:U →M
having a derivative matrix γ ′ of full rank. Since the derivative γ ′ is an n×k matrix and k ≤ n, it
is the columns that are linearly independent. These columns are usually interpreted as vectors.
(It is naturally assumed here that k > 0.)
12
In many textbooks manifolds are indeed defined using local parametrization, see e.g. S PIVAK or O’N EILL .
This usually requires the so-called transition rules to make sure that the chart functions are coherent. Our definition,
too, is a kind of local parametrization, but not a general one, and not requiring any transition rules.
13
This is often called a smooth parametrization.
CHAPTER 2. MANIFOLD 24
∂atan(x, y) y
=− 2 4
∂x x + y2
–2 –2
and 2
∂atan(x, y) x –1 –1
= 2 .
∂y x + y2 0
x2 + z 2 = R2 , y=0, x≥0
is removed then a manifold is obtained which can be parametrized by the familiar spherical
coordinates as
(x, y, z) = γ(θ, φ) = (R sin θ cos φ, R sin θ sin φ, R cos θ)
and the parameter domain is the open rectangle
U : 0 < θ < π , 0 < φ < 2π.
The derivative matrix
R cos θ cos φ −R sin θ sin φ
γ ′ (θ, φ) = R cos θ sin φ R sin θ cos φ
−R sin θ 0
is of full rank, and the inverse parametrization is again easy to find:
θ = arccos z
R
φ = atan(x, y).
Parametrization is at the same time more restrictive and more extensive than our earlier
definitions of manifolds: Not all manifolds can be parametrized and not all parametrizations
define manifolds. On the other hand, as noted, parametrization makes it easier to deal with
manifolds. In integration restrictions of parametrizations can be mostly neglected since they do
not affect the values of the integrals, as we will see later. Let us note, however, that if a set is
parametrized then at least it is a manifold in a certain localized fashion:
Theorem 2.6. If M ⊆ Rn , U is an open subset of Rk , u0 ∈ U, and there is a continuously
differentiable bijective function γ : U → M with a derivative γ ′ of full rank, then there is an
open subset V of U such that u0 ∈ V and γ(V) is a k-dimensional manifold of Rn .
Proof. Consider a point
p0 = γ(u0 )
of M. Then the columns of γ ′ (u0 ) are linearly independent, and thus some k rows are linearly
independent, too. Reordering, if necessary, we may assume that these rows are the first k rows
of γ ′ (u0 ).
Let us first consider the case k < n. For a general point p = (r, s) of M, r contains the k
first components. We denote further by γ 1 the function consisting of the first k components of
γ, and r0 = γ 1 (u0 ). Similarly, taking the last n − k components of γ we get the function γ 2 .
The function
F(u, r) = r − γ 1 (u)
defined in the open set S = U × Rk (cf. the proof of Theorem 2.4) then satisfies the conditions
1.– 4. of the Implicit Function Theorem. Thus there is a continuously differentiable function
g : B → Rk , defined in an open set B, whose graph u = g(r) is included in S, such that
r = γ 1 g(r) .
Finally we choose V = γ −1
1 (B), an open set. (And where, if anywhere, do we need bijectivity
of γ?)
The case k = n is similar. The function
F(u, p) = p − γ(u)
defined in the open set S = U × Rn then satisfies conditions 1.– 4. of the Implicit Function The-
orem. Hence there is an open set B, containing the point p0 , and a continuously differentiable
function g : B → Rn , whose graph u = g(p) is included in S, such that
p = γ g(p) .
γ:U →M
η : V → U,
CHAPTER 2. MANIFOLD 27
such that the derivative η ′ is nonsingular. The new parametrization is then by the composite
function γ ◦ η, that is, as
r = γ η(v) (v ∈ V).
This really is a parametrization since, by the chain rule, γ ◦ η is continuously differentiable and
its derivative
γ ′ η(v) η ′ (v)
is of full rank. n-dimensional manifolds of Rn , i.e. open subsets, often benefit from reparam-
etrization.
Example. 3-dimensional manifolds of R3 , i.e. open subsets or ’solids’, are often given using
parametrizations other than the trivial one by the identity function.
Familiar parametrizations of this type are those using cylindrical or spherical coordinates.
For instance, the slice of a ball below (an open set) can be parametrized by spherical coordi-
nates as
r = (x, y, z) = γ(ρ, θ, φ) = (ρ sin θ cos φ, ρ sin θ sin φ, ρ cos θ),
where the parameter domain is the open rectangular prism
V : 0 < ρ < R , 0 < θ < π , 0 < φ < α.
z
φ= 0 φ= α φ parameter domain
α
y U
π θ
ρ= R R
ρ
x
Different parametrizations of a manifold may come separately, without any explicit repara-
metrizations. Even then in principle, reparametrizations are there.
Theorem 2.7. Different parametrizations of a manifold can always be obtained from each other
by reparametrizations.
Proof. Consider a situation where the k-dimensional manifold M of Rn has the parametriza-
tions
r = γ 1 (u) (u ∈ U) and r = γ 2 (v) (v ∈ V).
An obvious candidate for the reparametrization is the one using η = γ −1
1 ◦ γ 2 , as
u = γ −1
1 γ 2 (v) .
This function η is bijective, we just must show that it is continuously differentiable. For this let
us first define
G(u, v) = γ 1 (u) − γ 2 (v).
CHAPTER 2. MANIFOLD 28
Then the columns of the derivative G′ corresponding to the variable u, i.e. γ ′1 , are linearly
independent.
Consider then a point
r0 = γ 1 (u0 ) = γ 2 (v0 )
of M. Since the k columns of γ ′1 (u0 ) are linearly independent, then some k rows of γ ′1 (u0 ) are
also linearly independent. Picking from G the corresponding k components we get the function
F. In the open set S = U × V the function F satisfies the conditions 1.– 4. of the Implicit
Function Theorem, and then the obtained function f clearly is η.
Since the point v0 was an arbitrary point of V, η is continuously differentiable. On the other
hand, η ′ is also nonsingular because γ 2 = γ 1 ◦ η, and by the chain rule
γ ′2 (v) = γ ′1 η(v) η ′ (v).
If now η ′ would be singular in some point of V, then γ ′2 would not have full rank there.
A parametrized manifold may be localized also in the parameter domain: Take an open
subset U ′ of U and interprete it as a new parameter domain. The thus parametrized set is again
a manifold, and it has a parameter representation (cf. Theorem 2.6).
This in fact also leads to a generalization of manifolds. Just parametrize a set N as above
using a parameter domain U and a function γ : U → N which is continuously differentiable
and whose derivative γ ′ is of full rank, but do not require that γ is bijective. If now for each
point p of N there is an open subset Up of U such that
then as in Theorem 2.6, the parametrization defines a manifold when restricted into Up . The set
N itself then need not be a manifold. Generalized manifolds of this kind are called trajectory
manifolds. A trajectory manifold can reparametreized exactly as the usual manifold.
where
φ
r(φ) = ecos φ − 2 cos 4φ + sin5 ,
12
is a complicated plane curve, but not a manifold since it passes through the origin six times, see
the left figure below (Maple). It is however a 1-dimensional trajectory manifold. The figure on
the right is the hodograph
(x, y) = γ ′ (φ)T (0 < φ < 2π).
It shows that γ ′ is of full rank (the curve does not pass through the origin). It also indicates
the parameter value φ = 0 (or φ = 2π) could have been included locally. This would not
destroy smoothness of the curve. This is common in polar parametrizations. It should also
be remembered that even though atan is discontinuous, sin atan(x, y) and cos atan(x, y)
are continuously differentiable. So the parameter interval could have been, say, 0 < φ < 4π
containg the parameter value φ = 2π. Note also that the polar parametrization allows even
negative values of the radius!
CHAPTER 2. MANIFOLD 29
3
6
2
4
1 2
–6 –4 –2 2 4 6
0
–1 0 1 2 3
–2
–1
–4
–2 –6
–8
–3
Many more self-intersections appear when the curve is drawn for the ”full” parameter
interval 0 < φ < 24π, outside of which it starts to repeat itself:
CHAPTER 2. MANIFOLD 30
f(r) ∼
= f(r0 ) + (r − r0 )f ′ (r0 )T .
The affine approximation of a function in a point gives correctly the values of the function and
its derivative in this point. Let us denote
(whence g′ (r0 ) = f ′ (r0 )). Then s = g(r) is a graph which locally touches the manifold M in
the point p0 . Geometrically this graph is part of a k-dimensional hyperplane, or a plane or a
line in lower dimensions.
The tangent space of M in the point p0 , denoted by Tp0 (M), consists of all (tangent)
vectors such that the initial point of their representative directed line segments is p0 and the
terminal point is in the graph s = g(r), i.e., it consists of exactly all vectors
T T
= r − r0 , (r − r0 )f ′ (r0 )T
r, g(r) − r0 , f(r0 )
! !
(r − r0 )T Ik
= T
= (r − r0 )T ,
f (r0 )(r − r0 )
′
f (r0 )
′
T(h) = f ′ (r0 )h
of the vector variable h. Apparently T is a linear function and f ′ (r0 ) is the corresponding
matrix.
Note that replacing the graph s = f(r) by another graph t = h(u) (as needs to be done when
moving from one chart to another) simply corresponds to a local reparametrization u = η(r)
and change of basis of the tangent space using the matrix η ′ (r0 ) (cf. Theorem 2.7 and its proof).
The space itself remains the same, of course.
Example. A smooth space curve or a 1-dimensional manifold MÊof R3 is locally a graph
(y, z) = f(x) = f1 (x), f2 (x)
(or one of the other two alternatives). The tangent space of M in the point p0 = (x0 , y0, z0 ),
where (y0 , z0 ) = f(x0 ), consists of exactly all vectors
CHAPTER 2. MANIFOLD 31
z
h
f ′ (x0 )h
1 (h ∈ R).
f2′ (x0 )h
p0 = (x0 , f1(x0) , f2(x0))
Geometrically the vectors are directed
along the line r = p0 + tv (t ∈ R), y
where
v = 1, f1′ (x0 ), f2′ (x0 ) . space curve + tangent vector
x
r = p0 + t1 v1 + t2 v2 (t1 , t2 ∈ R), z v1
tangent plane
where
∂f (x0 , y0 ) p0 = (x0 , y0 , f(x0,y0))
v1 = 1, 0,
∂x
and y
∂f (x0 , y0) v2
v2 = 0, 1, .
∂y
x
What about when a manifold M is given by local loci, say locally by the condition
F(r, s) = 0
(assuming a proper order of variables)? According to Corollary 2.3, then M is given locally
near the point p0 = (r0 , s0 ) also as a graph s = f(r) and (cf. the Implicit Function Theorem)
−1
f ′ (r0 ) = −F′s r0 , f(r0 ) F′r r0 , f(r0 )
where
F′ = F′r F′s .
The tangent space Tp0 (M) consists of the vectors
!
Ik
h.
f ′ (r0 )
i.e.
F′ (p0 )m = 0.
So we get
Theorem 2.8. If a manifold M is locally near the point p0 given as the locus defined by the
condition F(p) = 0 (with the assumptions of Corollary 2.3), then the tangent space Tp0 (M) is
the null space of the matrix F′ (p0 ).
In practice it of course suffices to find a basis for the tangent space (or null space). Coordi-
nate-freeness of tangent spaces has not been verified yet, but as a consequence of the theorem,
Corollary 2.9. Tangent spaces of manifolds are coordinate-free.
Proof. This follows because a null space is coordinate-free, and a manifold can be given as a
local locus (Theorem 2.4). More specifically, if we take a coordinate transform p∗ = pQ + b
and denote
F∗ (p∗ ) = F (p∗ − b)QT and m∗ = QT m,
F (x, y) = x2 + y 2 − R2 = 0
in the point (x0 , y0) is then the null space of the 1 × 2 matrix
F (x, y, z) = x2 + y 2 + z 2 − R2 = 0
Example. In general, the tangent space of a smooth surface (manifold) M in R3 defined im-
plicitly by the equation
F (x, y, z) = 0
in the point p0 = (x0 , y0 , z0 ) is the null space of the 1 × 3 matrix F ′ (p0 ), i.e., the set of vectors
m = (h, k, l)T satisfying
∂F (p0 ) ∂F (p0 ) ∂F (p0 )
F ′ (p0 ) • m = h+ k+ l = 0.
∂x ∂y ∂z
As a further consequence of Theorem 2.8 we see that if a manifold is parametrized, then its
tangent space can be parametrized, too.
Corollary 2.10. If the k-dimensional manifold M of Rn has the parametrization γ : U → M
and p0 = γ(u0 ), then the tangent space Tp0 (M) consists of exactly all vectors of the form
γ ′ (u0 )v (v ∈ Rk ),
that is, Tp0 (M) is the column space of γ ′ (u0 ).
Proof. Locally near the point p0 the manifold M can be given as a locus determined by some
suitable condition F(p) = 0. Thus the equation
F γ(u) = 0
is an identity valid in a neighborhood of the parameter value u0 . Applying the chain rule we get
another identity
F′ (γ u) γ ′ (u) = O,
where O is a zero matrix of appropriate dimensions. Substituting u = u0 we get the equation
F′ (p0 )γ ′ (u0 ) = O,
showing that columns of γ ′ (u0 ) are in the null space of F′ (p0 ).
On the other hand, since the dimension of the null space is k and the k columns of γ ′ (u0 )
are linearly independent, the columns of γ ′ (u0 ) span the tangent space Tp0 (M).
Example. If the parametrization of a smooth space curve C (a 1-dimensional manifold of R3 )
is
r = γ(u) (u ∈ U),
then its tangent space Tr0 (C) in the point r0 = γ(u0 ) consists of exactly all vectors
hγ ′ (u0 ) (h ∈ R).
Example. If the parametrization of a smooth surface of R3 (a 2-dimensional manifold) is
r = γ(u) (u ∈ U),
then its tangent space in the point r0 = γ(u0 ) consists of exactly all vectors
∂γ1 (u0 ) ∂γ1 (u0 )
∂u1 ∂u2
∂γ (u ) ∂γ (u )
0 h1 ∂γ(u0 ) ∂γ(u0 )
2 0 2
′
γ (u0 )h = = h (h ∈ R2 ).
∂u1 ∂u2 h2 ∂u1 ∂u2
∂γ3 (u0 ) ∂γ3 (u0 )
∂u1 ∂u2
CHAPTER 2. MANIFOLD 34
−f (r0 )T
′
k (k ∈ Rn−k ).
In−k
Some normals are more interesting than others, dealing with curvature, torsion, and the plane
most accurately containing the curve near p0 .
Example. A smooth surface in R3 is a 2-dimensional manifold M. Locally M is a graph
z = f (x, y) (or one of the other two alternatives). The normal space of M in the point
p0 = (x0 , y0, z0 ), where then z0 = f (x0 , y0 ), consists of exactly all vectors
∂f (x0 , y0 )
− ∂x
k
∂f (x , y )
0 0 (k ∈ R).
− k
∂y
k
Geometrically these vectors are in the line r = p0 + tv (t ∈ R), where
∂f (x , y ) ∂f (x , y )
0 0 0 0
v= − ,− ,1 .
∂x ∂y
Example. If the parametrization of a smooth surface (a 2-dimensional manifold of R3 ) is
r = γ(u) (u ∈ U),
then its normal space in the point r0 = γ(u0 ) consists of exactly all vectors
h
n = k
l
satisfying
∂γ1 (u0 ) ∂γ2 (u0 ) ∂γ3 (u0 )
∂u1 h
′ T ∂u1 ∂u1 0
γ (u0 ) n = k =
.
∂γ1 (u0 ) ∂γ2 (u0 ) ∂γ3 (u0 ) 0
l
∂u2 ∂u2 ∂u2
The basis vector of the null space is now obtained in the usual way using cross product, thus
the vectors are ∂γ(u ) ∂γ(u )
0 0
t × (t ∈ R).
∂u1 ∂u2
(The cross product is not the zero vector because the columns of γ ′ (u0 ) are linearly indepen-
dent.)
For instance, the normal space of a sphere parametrized by spherical coordinates as
i.e. vectors
tγ(θ0 , φ0 )T (t ∈ R).
CHAPTER 2. MANIFOLD 36
14
The least squares formulas, probably familiar from many basic courses of mathematics. Note that TT T is
nonsingular because otherwise there would be a nonzero vector c such that cT TT Tc = 0, i.e. Tc = 0.
CHAPTER 2. MANIFOLD 37
N = n , H = n • n = knk2 , Ftan
y
b = F(p0 ) • n
x
and (as is familiar from basic courses)
n n 1
Fnorm (p0 ) = F(p0 ) • =n F(p0 ) • n .
knk knk knk2
Note. A k-dimensional manifold in Rn together with its k-dimensional tangent spaces may
be thought of, at least locally, as a 2k-dimensional manifold (in Rn+k ), the so-called tan-
gent bundle, and the tangent spaces are its fibres. Similarly the manifold together with its
n − k-dimensional normal spaces can be thought of as an n-dimensional manifold (in R2n−k ),
the normal bundle.
The flux of a vector field in the manifold may then be seen locally as a cross-section of the
tangent bundle parametrized by the manifold, and similarly the flux through the manifold as
a cross-section of the normal bundle. This kind of geometric view of vector fields (and more
generally tensor fields) is popular in modern physics, but is pursued no further here.
A more general concept is the fiber bundle. See e.g. A BRAHAM & M ARSDEN & R ATIU .
”Calculating surface area is a foolhardy enterprise;
fortunately one seldom needs to know the area of a
surface. Moreover, there is a simple expression for dA
which suffices for theoretical considerations.”
Chapter 3
VOLUME
respectively. A set A is Jordan measurable if |A|out = |A|in, and the common value |A| is its
38
CHAPTER 3. VOLUME 39
Jordan’s measure1 . This measure is now defined to be the volume of A. Such a volume clearly
is coordinate-free.
The precise same construct can be used in a k-dimensional space embedded in an n-dimen-
sional space (where n > k). The k-dimensional space is then an affine subspace of Rn , i.e., a
manifold R parametrized by
k
X
γ(u) = b + ui vi (u ∈ Rk ),
i=1
where v1 , . . . , vk are linearly independent. Taking b as the origin and, if needed, orthogonaliz-
ing v1 , . . . , vk we may embed Rk in an obvious way in Rn , and thus define the k-dimensional
volume of a bounded subset of R. (The n-dimensional volume of these subsets in Rn is = 0, as
is easily seen.) Such affine subspaces are e.g. planes and lines of R3 .
Note. Not all bounded subsets of Rn have a volume! There are bounded sets not having a
Jordan measure.
The inner and outer covers used in defining Jordan’s inner and outer measures remind us
of the n-dimensional Riemann integral, familiar from basic courses of mathematics, with its
partitions and lower and upper
sums. It is indeed fairly easy to
see that whenever the volume ex-
a3
ists, it can be obtained by inte-
grating the constant 1, with im-
proper integrals, if needed:
a1
Z
|A| = 1 dr.
A
where C is the unit cube in Rn given by 0 ≤ ui ≤ 1 (i = 1, . . . , n). The volume of this cube is
= 1.
1
Also called Jordan–Peano measure or Jordan’s content.
CHAPTER 3. VOLUME 40
We already know volumes are coordinate-free. For the parallelepiped this is also clear
because the volume can be written as
p
|P| = det(G),
Being dot products, elements of a Gramian are coordinate-free. The same formula is valid
when we are dealing with the k-dimensional volume of a k-dimensional parallelepiped P as
part of Rk embedded as an affine subspace in Rn , and P is given by the n-dimensional vectors
a1 , a2 , . . . , ak . The Gramian is here, too, a k × k matrix. Note that we need no coordinate
transforms or orthogonalizations, the volume is simply given by the n-dimensional vectors.
An example would be a parallelogram embedded in R3 , with its sides given by 3-dimensional
vectors.
A bounded subset A of Rn is called a null set if its volume (or Jordan’s measure) is = 0.
For a null set A then
|A|out = inf |P | = 0
P ∈Pout
Proof. The proof is based on a tedious and somewhat complicated estimation, see e.g. H UB -
BARD & H UBBARD . Just about the only easy thing here is that if the volume of a bounded
subset A exists, then it is = 0. Otherwise |A|in > 0 and some inner cover of A would have at
least one n-dimensional cube contained in M. But near the center of the cube M is locally a
graph of a function, which is not possible.
Example. Bounded subsets of smooth surfaces and curves of R3 (1- and 2-dimensional mani-
folds) are null sets.
Null sets—as well as the k-null-sets to be defined below—are very important for integration
since they can be included and excluded freely in the region of integration without changing
values of integrals.
Rather than the ”full” n-dimensional one, it is possible to define a lower-dimensional vol-
ume for subsets of Rn , but this is fairly complicated in the general case.2 On the other hand, it
is easy to define the k-dimensional volume of an n-cube: It is hk if the edge length of the cube
is h. This gives us the k-dimensional volume of an outer cover P of a subset A of Rn , and then,
via the infimum, the k-dimensional Jordan outer measure. The corresponding inner measure is
of course always zero if A does not contain an open set, i.e. its interior A◦ is empty.
Thus, at least it is easy to define the k-dimensional zero volume for a subset A of Rn , that
is, the k-null-set of Rn : It is a set having a zero k-dimensional Jordan outer measure. Again
this definition is not that easy to use. Theorem 3.2 can be generalized, however, the proof then
becoming even more complicated (see H UBBARD & H UBBARD):
2
And has to do e.g. with fractals.
CHAPTER 3. VOLUME 41
For instance, bounded subsets of smooth curves of R3 (1-dimensional manifolds) are 2-null-sets
of R3 .
r = γ(u) (u ∈ U),
then Z q
|M|k = det γ ′ (u)T γ ′ (u) du,
U
or briefly denoted Z
|M|k = dr.
M
Such a volume may be infinite. Note that inside the square root there is a Gramian determinant
which is coordinate-free. The whole integral then is coordinate-free.
Comparing with the k-dimensional volume of a parallelepiped
p
|P|k = det(AT A)
in the previous section we notice that in a sense we obtain |M|k by ”summing” over the points
r = γ(u) of the manifold k-dimensional volumes of parallelepipeds whose edges are given by
the vectors
∂γ(u)
dui (i = 1, . . . , k).
∂ui
Moreover, (as a directed line segment) the vector
∂γ(u)
dui
∂ui
approximatively defines the movement of a point in the manifold when the parameter ui changes
a bit the other parameters remaining unchanged.
Note. In a certain sense definition of the volume of a parametrized manifold then is natural,
but we must remember that it is just that, a definition. Though various problems concern-
ing volumes are largely solved for parametrized manifolds—definition, computing, etc.—others
remain. There are parametrizations giving a volume for a manifold even when it does not oth-
erwise exist (e.g. as Jordan measure). On the other hand, a manifold having a volume may
sometimes be given a parametrization such that the above integral does not exist.
Things thus depend on the parametrization. In the sequel we tacitly assume such ”patho-
logical” cases are avoided.
CHAPTER 3. VOLUME 42
It is important to check that the definition above gives the open parallelepiped P the same
volume as before. The parallelepiped is a manifold and can be parametrized naturally as
k
X
r = γ(u) = b + ui aTi (U : 0 < u1 , . . . , uk < 1),
i=1
whence Z p
′
p
γ (u) = A and |P|k = det(AT A) du = det(AT A).
U
Another important fact to verify is freeness of parametrization.
Theorem 3.4. The volume of a parametrized manifold does not depend on the parametrization.3
Proof. By Theorem 2.7 different parametrizations can be obtained from each other by repara-
metrization. When reparametrizing a k-dimensional manifold M originally parametrized by
some γ : U → M a new parameter domain V ⊆ Rk is taken and a continuously differentiable
bijective function η : V → U having a nonsingular derivative matrix η ′ . The new parametriza-
tion then is given by the composite function δ = γ ◦ η as
r = γ η(v) = δ(v) (v ∈ V).
Via the chain rule
δ ′ (v) = γ ′ η(v) η ′ (v).
In the integral giving |M|k this corresponds to the change of variables u = η(v) and the
corresponding transform of the region of integration from U to V. Thus we only need to check4
the form of the new integrand:
Z q
|M|k = det γ ′ (u)T γ ′ (u) du
U
Z q T
det γ ′ η(v) γ ′ η(v) det η ′ (v) dv
=
V
Z q T 2
= det γ ′ η(v) γ ′ η(v) det η ′ (v) dv
V
Z q T
= det η ′ (v)T det γ ′ η(v) γ ′ η(v) det η ′ (v) dv
V
Z q
det δ ′ (v)T δ ′ (v) dv.
=
V
U U U
3
Remember that we tacitly assume all parametrizations do give a volume.
4
Recall that for square matrices det(AB) = det(A) det(B) and det(AT ) = det(A).
CHAPTER 3. VOLUME 43
S : r = γ(u) (u ∈ U)
is Z q
|S|2 = det γ ′ (u)T γ ′ (u) du.
U
∂γ(u) ∂γ(u)
Z
×
du
∂u1 ∂u2
U
since the area (2-dimensional volume) of the parallelogram in the integrand (a 2-dimensional
parallelepiped) can be given as the length of the cross product.
3. X is a null set of Rk .
4. γ(X ) is a k-null-set of Rn .
r = γ(u) (u ∈ U − X )
where the parameter domain is U : −4 ≤ u ≤ 4. The exception set is X = {−4, −2, 0, 2, 4}.
Similarly, relaxed parametrizations could be obtained for surfaces of cubes, or more gener-
ally, surfaces of polyhedra.
In volume computations the exception set X has no contribution since it is mapped to γ(X ),
a k-null-set of Rn : Z q
′
|M|k = |M |k = det γ ′ (u)T γ ′ (u) du.
U −X
q
If the integrand det γ ′ (u)T γ ′ (u) has a continuous extension onto the whole U, as is some-
times the case, we can write further
Z q
|M|k = det γ ′ (u)T γ ′ (u) du,
U
since X is a null set of Rk . Improper integrals are here allowed, too, giving even more ”re-
laxation”. Thus relaxed parametrizations can be used in volume computations and integrations
more or less as the ”usual” parametrizations of Section 2.5.
CHAPTER 3. VOLUME 45
Even relaxed reparametrization7 is possible and preserves volumes. A new extended pa-
rameter domain V ⊆ Rk , an exception set Y ⊂ V, and a continuous function η : V → U are
then sought, such that
In such a relaxed reparametrization the exception set is mapped to an exception set, which
guarantees preservation of the volume in reparametrization (as a consequence of Theorem 3.4).
To make reparametrization in this sense possible, exception sets often need to be modified.
(the usual polar coordinate parametrization with the exception set {0}), and
p
− u − 2, 1 − (u + 2)2 for −3 ≤ u ≤ −1
r = γ 2 (u) = (U : −3 ≤ u < 1)
u, −√1 − u2 for −1 ≤ u < 1
(the exception set is {−3, −1}). These two relaxed parametrizations cannot be directly repara-
metrized to each other. If, however, in the exception set of the first parametrization the ”unnec-
essary” number π is added, then a reparametrization is possible by the function
(
−2 − cos φ for 0 ≤ φ ≤ π
u = η(φ) =
cos φ for π ≤ φ < 2π.
Note. Relaxed parametrizations are possible for trajectory manifolds, too. And volumes will
then also be preserved in relaxed reparametrizations.
7
Again in literature this is defined in many different ways.
”Gradient a 1-form? How so? Hasn’t one always known
the gradient as a vector? Yes, indeed, but only because
one was not familiar with the more appropriate 1-form
concept. The more familiar gradient is the vector
corresponding to the 1-form gradient.”
Chapter 4
FORMS
4.1 k-Forms
Whenever vector fields are integrated—the result being a scalar—the fields first need to be
”prepared” in one way or another, so that the integrand is scalar-valued. A pertinent property
of a vector field is its direction, which then must be included somehow in the preparation.
Integration regions are here parametrized manifolds, possibly in a relaxed sense. Directions
related to the manifold can be included via tangent spaces or normal spaces. And in many
cases orientation of the manifold should be fixed, too. Examples of integrals of this type are the
familiar line and surface integrals
Z Z
F(r) • ds and F(r) • dS.
C S
A general apparatus for the ”preparation” is given by so-called forms. In integration they
appear formally as a kind of differentials, and are therefore often called differential forms. In
various notations, too, differentials appear for this reason. Other than this, they do not have any
real connection with differentials.
An n-dimensional k-form is a function φ mapping n-dimensional vectors1 r1 , . . . , rk (thus
k vectors, and the order is relevant!) to a real number φ(r1 , . . . , rk ), satisfying the conditions
1. φ is antisymmetric, i.e., interchanging any two vectors ri and rj changes the sign of the
value φ(r1 , . . . , rk ). Thus if for instance k = 4, then
φ(r3 , r2 , r1 , r4 ) = −φ(r1 , r2 , r3 , r4 ).
As a limiting case, 0-forms can be included, too: they are constant functions which do not have
any variable vectors at all.
1
Often tangent vectors where only the vector part is used. Note that in 1-, 2- and 3-dimensional spaces geomet-
ric vectors could be used.
46
CHAPTER 4. FORMS 47
Thinking about the intended use of k-forms—as devices for coupling vector fields and
k-dimensional volumes for integration—this definition looks fairly minimal. Antisymmetry
makes it possible the change direction by ”reflecting”, that is, interchanging two vector vari-
ables. Multilinearity on the one hand guarantees the volume of a combination of two disjoint
sets is duly obtained by adding the volumes of the two parts, and on the other hand implies that
scaling works correctly, i.e., if the vector variables are multiplied (scaled) by the scalar λ, then
the value is scaled by λk .
Example. A familiar example of an n-dimensional n-form is the determinant det. Take n vec-
tors r1 , . . . , rn , consider them as columns of a matrix: (r1 , . . . , rn ), and compute its determinant
det(r1 , . . . , rn ). It follows directly from basic properties of determinants that this is an n-form.
Example. An equally familiar example of an n-dimensional 1-form is dot product with a fixed
vector. If a is a fixed constant vector, then the function φ(r) = a • r is a 1-form. Note that
antisymmetry is not relevant here because there is only on vector variable.
The dot product r1 • r2 is however not a 2-form (and why not?).
Example. A further familiar example of a 3-dimensional 2-form is the first component (or any
other component) of the cross product r1 × r2 . Properties of the cross product immediately
imply that
(r2 × r1 )1 = (−r1 × r2 )1 = −(r1 × r2 )1
(antisymmetry) and that
r1 × (c1 r21 + c2 r22 ) 1
= (c1 r1 × r21 + c2 r1 × r22 )1 = c1 (r1 × r21 )1 + c2 (r1 × r22 )1
(multilinearity).
(ii) If φ1 (r1 , . . . , rk ) and φ2 (r1 , . . . , rk ) are n-dimensional k-forms, then so is their sum
φ1 (r1 , . . . , rk ) + φ2 (r1 , . . . , rk ).
(iv) If the vectors a1 , . . . , ak are linearly dependent, then the value of a k-form φ(a1 , . . . , ak )
is = 0. In particular, if some ai is the zero vector, then φ(a1 , . . . , ak ) = 0.
Proof. Items (i), (ii) and (iii) are immediate, so let us proceed to item (iv). We notice first that
if two of the vectors a1 , . . . , ak are the same, then the value is = 0. Indeed, interchanging these
vectors the value changes its sign, and also remains the same. If then a1 , . . . , ak are linearly
dependent, then one of them can be expressed as a linear combination of the others, say,
k−1
X
ak = ci ai .
i=1
CHAPTER 4. FORMS 48
Note that all terms having i = j can be omitted in the sum. In this way we finally get a
representation of φ(r1 , . . . , rk ) as a sum of terms of the form
where a = φ(e1 , e2 , . . . , ek ) and the sign ± is determined by the parity of the number of uses
of antisymmetry (’even’ = +, ’odd’ = −). On the other hand, this is exactly the way the
determinant
x1,1 x1,2 · · · x1,k
x2,1 x2,2 · · · x2,k
.. .. . . ..
. . . .
xk,1 xk,2 · · · xk,k
CHAPTER 4. FORMS 49
Thus we get the value of the form by taking elements having the indices j1 , j2 , . . . , jk from
r1 , . . . , rk , forming the corresponding k × k determinant, and computing the value of the deter-
minant. As a special case an elementary 0-form is included, it always has the value = 1.
Obviously there are
n n!
=
k k!(n − k)!
elementary n-dimensional k-forms. In particular there is one elementary 0-form and one ele-
mentary n-form. This makes it possible to ”embed” single scalars in forms in two ways. On the
other hand there are n elementary 1-forms, and also n elementary n − 1-forms, which makes it
possible to ”embed” vectors in forms via the coefficients, again in two ways (even when n = 2,
in a sense!).
using a determinant actually works even when the indices j1 , j2 , . . . , jk are not in order of
magnitude. This possibility is a convenient one to allow. For instance, interchanging dxjl
and dxjh interchanges the ith and the hth rows in the determinant and thus changes its sign.
Furthermore, if two (or more) of the indices j1 , j2 , . . . , jk are the same, then we naturally agree
that the k-form dxj1 ∧ dxj2 ∧ · · · ∧ dxjk has the value 0.
2
This particular notation is traditional. It does not have much to do with differentials. The symbol ’∧’ is read
”wedge”. The corresponding binary operation, the so-called wedge product (or sometimes the exterior product), is
generally available for forms, as will be seen.
CHAPTER 4. FORMS 50
Here in the 2-form dz ∧ dx the variables are not in the correct order, cf. the above note.
If the variables are not in the correct order and/or are repeated, then
since it does not matter whether it is the rows or the columns that are interchanged.
where
aj1 ,j2,...,jk = φ(ej1 , ej2 , . . . , ejk ).
The representation is unique, i.e., the coefficients aj1 ,j2,...,jk can only be chosen in one way (and
why is that?).
i.e., exactly all forms obtained by taking a dot product with a constant vector a = (a1 , . . . , an )T .
It might be a bit confusing to consider this using elementary 1-forms:
(a k-form) and
X
ψ(s1 , . . . , sl ) = bh1 ,h2 ,...,hl (dxh1 ∧ dxh2 ∧ · · · ∧ dxhl )(s1 , . . . , sl )
1≤h1 <h2 <···<hl ≤n
with the proviso of the above note concerning order and/or repetetition of indices. If k = 0, that
is φ is a 0-form (constant), this simply is multiplication by a constant. Note especially that the
wedge product is associative and distributive, i.e.,
φ ∧ (ψ ∧ ξ) = (φ ∧ ψ) ∧ ξ ,
φ ∧ (ψ + ξ) = φ ∧ ψ + φ ∧ ξ and
(φ + ψ) ∧ ξ = φ ∧ ξ + φ ∧ ξ.
Note. The concept of a form was defined without using any coordinate systems. On the other
hand, the wedge product above was defined via elementary forms, and thus using a coordinate
system. It can, however, also be defined without elementary forms, starting from the original
definition of forms, and so is coordinate-free. See e.g. H UBBARD & H UBBARD .
CHAPTER 4. FORMS 52
Φ(p; r1 , . . . , rk ).
Here r1 , . . . , rk are the vector variables for the k-form given by the form field for the point p.
Often they are considered as tangent vectors with the point of action p. By Theorem 4.2, to
define a form field Φ it suffices to give the coefficients in the presentation by elementary forms,
and they will depend on p:
X
Φ(p; r1 , . . . , rk ) = aj1 ,j2 ,...,jk (p)(dxj1 ∧ dxj2 ∧ · · · ∧ dxjk )(r1 , . . . , rk ),
1≤j1 <j2 <···<jk ≤n
n
where the needed coefficients are
k
aj1 ,j2 ,...,jk (p) = Φ(p; ej1 , ej2 , . . . , ejk ).
Often a form field is not defined in the whole space Rn but only in a subset, e.g. in a manifold.
Example. A constant k-form field is one depending only on the vector variables and not on the
point p, i.e. a form φ(r1 , . . . , rk ).
Φ(p; r) = F(p) • r,
where a is a real-valued valued function (scalar field) defined in a suitable subset of Rn (say, a
manifold).
3
For some reason form fields are also called differential forms in some literature, though they have little to do
with differentials.
CHAPTER 4. FORMS 53
These examples already indicate how manifolds, vector and scalar fields and forms are
connected: Using forms vector and scalar fields are transformed into form fields which then
are integrated over a region, often a manifold or a parameter domain, possibly in a relaxed
parametrization. In the general case such integrals are hard to deal with. For parametrizations
it is easier, in principle anyway.
If a k-dimensional manifold M of Rn has a (relaxed) parametrization
r = γ(u) (u ∈ U)
and Φ(p; r1 , . . . , rk ) is an n-dimensional k-form field defined in M, then its integral over M is
Z Z
Φ = Φ γ(u); γ ′ (u) du.
M U
′
Here the k columns of γ (u) are interpreted as the vector variables of the form field. Many
integrals familiar from basic courses on mathematics are of this type, and they now have a
uniform formulation.
Example. The line integral of a vector field F over the smooth curve
C : r = γ(u) (u ∈ U),
i.e., Z Z
F γ(u) • γ ′ (u) du
F(r) • ds =
C U
is the integral of the 1-form field
Φ(p; r) = F(p) • r
over C. The curve C could here also be only piecewise smooth via a relaxed parametrization. It
certainly should be obvious why this integral often is given (in R3 ) as
Z
F1 (r) dx + F2 (r) dy + F3 (r) dz.
C
Example. Similarly the surface integral of the vector field F over the smooth surface
S : r = γ(u) (u ∈ U),
i.e.,
∂γ(u) ∂γ(u)
Z Z
F(r) • dS = F γ(u) • × du
∂u1 ∂u2
S U
is the integral of the 2-form field
Φ(p; r1 , r2 ) = F(p) • r1 × r2
over S. Here, too, the surface S could be only piecewise smooth via a relaxed parametrization.
Thinking about the connection of the cross product and elementary 2-forms, it should not be
surprising that this integral is sometimes given as
Z
F1 (r) dy ∧ dz + F2 (r) dz ∧ dx + F3 (r) dx ∧ dy,
S
or even Z
F1 (r) dy dz + F2 (r) dz dx + F3 (r) dx dy.
S
CHAPTER 4. FORMS 54
M U
(recall change of variables in an integral), is not the integral of a form field over M, because
of the absolute value of the determinant. But the integral
Z
f γ(u) det γ ′ (u) du
over M.
And what about the case of a 0-form field Φ of Rn ? It simply is a real-valued function Φ(p)
of n variables. On the other hand, a 0-dimensional manifold M of Rn consists of separate
points (possibly infinitely many), so
Z X
Φ= Φ(p).
M p∈M
Φ(p0 ; r1 , . . . , rk )
is defined in the tangent space Tp0 (M). I.e., Φ(p0 ; r1 , . . . , rk ) is defined for all vectors
r1 , . . . , rk in the tangent space Tp0 (M), cf. Section 2.6. Of course, the k-form may well
be defined for other vectors as well but they are not needed here.
CHAPTER 4. FORMS 55
2. Φ is continuous with respect to the variable p in M and does not change sign.
More specifically, for every point p of M there exist vectors
t1 (p) , . . . , tk (p)
of the tangent space Tp (M) such that
Φ p; t1 (p), . . . , tk (p)
is continuous with respect to p in M, and is either positive in M or negative in M.
The sign can then used for orientation. Note that the vectors t1 (p), . . . , tk (p) then must
constitute a basis for the tangent space Tp (M), otherwise the form would have the value
= 0. The question thus is whether or not the tangent space can always be chosen a basis
with the same orientation (same ”handedness”).
If a manifold M is parametrized by
p = γ(u) (u ∈ U),
then it is possible to use this to orient M. Item 1. remains the same but item 2. is replaced by
2.’ Φ γ(u); γ ′ (u) is continuous with respect to the parameter u, and does not change sign.
This is not a different type of orientation than the general one above, parametrization is just used
to choose the bases for the tangent spaces. As was seen in the proof of Theorem 2.6, locally
the parameter u can be given as a continuous—and even continuously differentiable—function
u = g(r) of some components r of p. Thus γ −1 is continuous
in M, and according to item 2.,
the orientation is given by the value Φ p; γ ′ γ −1 (p) .
Not all manifolds can be oriented. Classical examples are various Möbius’ bands. An
example is given by the relaxed parametrization
u2
γ 1 (u) = 1 + u 1 cos cos u2
2
u2
γ2 (u) = 1 + u1 cos sin u2 (U : −1 < u1 < 1, 0 ≤ u2 < 2π).
2
γ (u) = u sin u2
3 1
2
If the parameter value u2 = 0 is omitted, the band
is ”severed” and is a parametrized 2-dimensional
manifold of R3 . (Proving this takes a bit of tedious
calculation.) The severed band can be oriented us-
ing the parametrization. On the other hand, if in
the relaxed parametrization we take another inter-
val for the parameter u2 , say π ≤ u2 < 3π, the
break would move elsewhere. Since ”manifold-
ness” is a local property, this basically means that
the ”unsevered” Möbius band is a manifold, too—
but not a parametrized one. Looking at the fig-
ure on the right (Maple) it is at least geometrically
fairly obvious that the band cannot be oriented (the
exact proof again being tedious).
Manifolds that can be oriented are called orientable. If the orientation of an orientable
−→
manifold M is fixed (using some form field), this is indicated by denoting M. The opposite
−→ ←−
orientation is then often denoted by −M (sometimes by M).
CHAPTER 4. FORMS 56
Orientations of surfaces and curves are actually examples of this method of orientation.
∂γ(u) ∂γ(u)
×
∂u1 ∂u2
always points to the same side of the surface.
But what is the 2-form field needed for the orientation? The field is known to be of the form
and it is always positive. Other choices are of course possible, too, but maybe not quite as
natural.
Let us try the orientation on the torus. A relaxed parametrization of a torus is
γ1 (u) = (2 + cos u2 ) cos u1
4
γ2 (u) = (2 + cos u2 ) sin u1 3
γ3 (u) = sin u2 , –4 2 –4
–3 –3
–2 1 –2
where the parameter domain is U : 0 ≤ u1 , u2 < 2π. As –1 0 –1
for the Möbius band, it may be noticed that this torus is a
manifold, but not a parametrized one. For the orientation we 1 –1 1
2 –2 2
omit the value u1 = 0 (although the whole torus is of course 3 3
–3
orientable, too). See the figure on the right (Maple). Basis 4 4
–4
vectors of the tangent space are (in this order)
−(2 + cos u2 ) sin u1 − sin u2 cos u1
∂γ(u) ∂γ(u)
= (2 + cos u2 ) cos u1 and = − sin u2 sin u1 .
∂u1 ∂u2
0 cos u2
CHAPTER 4. FORMS 57
Which side of the torus is the one where the basis vectors are positively ordered (as in the
xy-coordinate system)? Take the basis vectors for the parameter values u1 = u2 = π, i.e. in the
point (−1, 0, 0) of the x-axis:
0 0
∂γ(π, π) ∂γ(π, π)
= −1 and = 0 .
∂u1 ∂u2
0 −1
These basis vectors are ordered as the negative y- and z-axes, thus this side is the ”outside”.
(In the ”inside” view the axes are interchanged).
C : p = γ(u) (u ∈ U)
can be oriented similarly. In the point γ(u) the basis vector γ ′ (u) of the tangent space is chosen,
and the curve is oriented. The 1-form field needed is of the form
Φ(p; r) = F(p) • r
and we choose
F γ(u) = γ ′ (u),
2
the value of the 1-form is then
γ ′ (u)
> 0. We could have chosen, say,
F γ(u) = 2γ ′ (u)
F γ(u) = −γ ′ (u)
If a manifold consists of several disjoint parts, orientation in these parts can be chosen
arbitrarily, the continuity condition does not connect the parts in any way. If for instance in a
3-dimensional manifold of R3 consisting of two disjoint parts, coordinate system in one of the
parts is right-handed, nothing prevents us from taking a left-handed system for the other part
(unless it is the general preference for right-handed systems).
For a connected manifold orientation is coherent all over the manifold. A manifold M of
Rn is said to be connected, if for each two points p0 and p1 of the manifold there is a continuous
function f : [0, 1] → M such that
In other words, any two points of the manifold can always be connected by a continuous curve
in the manifold.
Theorem 4.3. If the oriented manifold M is connected then an orientation in one point gives
the orientation everywhere in the manifold.
Proof. Suppose M is oriented as above. The function of the point p of the manifold in item 2.
Φ p; t1 (p), . . . , tk (p)
CHAPTER 4. FORMS 58
is then in the chosen point p0 either positive or negative. Any other point p1 of the manifold
can be connected to p0 using some continuous function f : [0, 1] → M, where p0 = f(0) and
p1 = f(1). But then the function
Φ f(s); t1 f(s) , . . . , tk f(s)
is a continuous function of s in the interval [0, 1] and does not attain the value 0. Thus
Φ p0 ; t1 (p0 ), . . . , tk (p0 ) and Φ p1 ; t1 (p1 ), . . . , tk (p1 )
have the same sign.
In reparametrization the orientation of a parametrized manifold may be changed for the
opposite. This can be prevented by setting a further condition:
Theorem 4.4. If the oriented parametrized k-dimensional manifold
M : r = γ(u) (u ∈ U)
of Rn is reparametrized by
u = η(v) (v ∈ V),
where
det η ′ (v) > 0,
then the orientation remains the same. If, on the other hand, det η ′ (v) < 0, then the orien-
tation will change. (It is naturally assumed here that the orientation is defined using the same
k-form field.)
Proof. As above, the manifold M is oriented using some k-form field Φ(p; r1 , . . . , rk ). Con-
sider then orientation in some point p0 = γ(u0 ) of M, given by the value Φ p0 ; γ ′ (u0 ) . By
Theorem 4.2 the k-form Φ(p0 ; r1 , . . . , rk ) can be represented as a linear combination of ele-
mentary k-forms. Let us show that these forms will be multiplied by the same positive number
in reparametrization.
The elementary k-forms above are of the form
xj ,1 xj ,2 · · · xj ,k
1 1 1
xj ,1 xj ,2 · · · xj ,k
2 2 2
(dxj1 ∧ dxj2 ∧ · · · ∧ dxjk )(r1 , . . . , rk ) = .. .. .
.. ..
. . . .
xjk ,1 xjk ,2 · · · xjk ,k
On the other hand, in the new parametrization a basis for the tangent space is given by the
columns of the matrix
γ ′ (u0 )η ′ (v0 )
(via the chain rule), where u0 = η(v0 ), and the value of the corresponding new elementary
k-form is
(dxj1 ∧ dxj2 ∧ · · · ∧ dxjk ) γ ′ (u0 )η ′ (v0 ) = (dxj1 ∧ dxj2 ∧ · · · ∧ dxjk ) γ ′ (u0 ) det η ′ (v0 ) .
(Remember matrix multiplication and that for square matrices det(AB) = det(A) det(B).)
All elementary
k-forms appearing here are thus multiplied by the same positive number
′
det η (v0 ) . So
Φ p0 ; γ ′ η(v0 ) η ′ (v0 ) = Φ p0 ; γ ′ (u0 ) det η ′ (v0 ) ,
Except for sketching the geometric picture, orientation is important for integration in repara-
metrization.
Theorem 4.5. Orientation preserving reparametrization of a parametrized oriented manifold
does not change integrals of form fields over the manifold. If, on the other hand, reparametriza-
tion changes the orientation, integrals change sign.
Proof. If the parametrized oriented manifold
−
→
M : r = γ(u) (u ∈ U)
u = η(v) (v ∈ V)
−
→ U V
M
is obtained. Since the reparametrization was orientation preserving, det η ′ (v) must be posi-
tive by the previous theorem. Thus the integral can be written as
Z Z Z
Φ = Φ γ η(v) ; γ η(v) det η (v) dv = Φ γ η(v) ; γ ′ η(v) η ′ (v) dv
′ ′
−
→ V V
M
(cf. the previous proof), and this is the integral of the form field Φ over M given in the new
parametrization.
On the other hand, if the reparametrization does change the orientation, then det η ′ (v) is
negative, and the sign is changed.
Change of sign of an integral by change of orientation is denoted as
Z Z Z
Φ = Φ = − Φ.
−
→ ←
− −
→
−M M M
1. For scalar fields f and g and vector field F we of course have f ∧ g = f g, and also
f ∧ ΦF–work = Φf F–work ,
f ∧ Φg–density = Φf g–density .
Further convenience is given by the Hodge star operator ∗ which transforms a k-form field
Φ in an n-dimensional space to an n − k-form field ∗Φ (the Hodge dual) as follows. If
X
Φ= aj1 ,...,jk (p) dxj1 ∧ · · · ∧ dxjk ,
j1 <···<jk
then X
∗Φ = ±aj1 ,...,jk (p) dxjk+1 ∧ · · · ∧ dxjn ,
j1 <···<jk
where in each summand the indices j1 , . . . , jn are exactly all numbers 1, . . . , n and the sign ±
is determined by the condition
For a 0-form field (scalar field) f and an n-form field Φf –density we define ∗f = Φf –density and
∗Φf –density = f . In R3 we then see (check!) that ∗ΦF–work = ΦF–flux and ∗ΦF–flux = ΦF–work . All
this makes it easy to change types of fields. Note also that for a k-form field Φ of Rn the dual
of the dual is ∗∗Φ = (−1)k(n−k) Φ, and in particular in R3 then ∗∗Φ = Φ.
CHAPTER 4. FORMS 61
There are physical fields which are neither scalar nor vector fields. For instance, the electric
field E is not a vector field (and of course not a scalar field either). Lorentz’s law says that the
force acting on a charged particle with charge q in the point r is
q E(r) + v × B(r) ,
where v is the speed of the particle and B is the magnetic flux density. If there is no ambient
electric field, the particle does not move, and the coordinate frame is stationary, there is no force
(i.e. the force is a zero vector). In a coordinate system moving with a constant speed (i.e. an
inertial frame) the force remains a zero vector (the particle does not no accelerate), but now v
is not a zero vector, the moving charged particle creates a magnetic field and a compensating
electric field. Seen from a moving frame then E is not a zero field. No physical vector field
can behave in this way, indeed, electric and magnetic fields should be dealt with together in
space-time (in R4 , the xyzt-space).
Even though electro-magnetic fields then cannot be modelled as vector fields, they can be
modelled as 2-form fields of R4 , the so-called Faraday form field and Maxwell form field 4
E1 dx ∧ dt + E2 dy ∧ dt + E3 dz ∧ dt + B1 dy ∧ dz + B2 dz ∧ dx + B3 dx ∧ dy
and
−c2 B1 dx ∧ dt − c2 B2 dy ∧ dt − c2 B3 dz ∧ dt + E1 dy ∧ dz + E2 dz ∧ dx + E3 dx ∧ dy,
where c is the speed of light in vacuum, often written simply—with only a slight abuse of
notation—as
ΦFaraday = ΦE–work ∧ dt + ΦB–flux
and
ΦMaxwell = ΦE–flux − c2 ΦB–work ∧ dt.
4
Michael Faraday and James Clerk Maxwell certainly did not consider form fields like these. The names
probably originated in the celebrated book M ISNER & T HORNE & W HEELER . The exact definitions of these
forms seem to vary a bit in the literature, the gist remaining the same, of course.
”Stokes theorem was stated by Sir George Stokes as
a Cambridge examination question, having been raised
by Lord Kelvin in a letter to Stokes in 1850.”
Chapter 5
Many integration theorems in basic courses on mathematics share a common overall form:
Zb
df (x)
dx = f (b) − f (a)
dx
a
(Fundamental Theorem of Integral Calculus)
Zr2
∇f • ds = f (r2 ) − f (r1) (Gradient Theorem)
r1
∂F2 (r) ∂F1 (r)
Z I
− dr = F(r) • ds (Green’s Theorem)
∂x ∂y
A −→
∂A
Z I
∇ • F(r) dr = F(r) • dS (Gauß’ Theorem)
A −→
∂A
Z I
∇ × F(r) • dS = F(r) • ds (Stokes’ Theorem)
−
→ −
→
S ∂S
• In the left hand side the integration region is a manifold, possibly in a relaxed parametriza-
tion, and oriented.
• In the left hand side the integrand is a derivative of some sort, coupled with the manifold
via tangent space and a suitable form (as a form field).
• In the right hand side the integration region (or summation region) is the oriented ”bound-
ary” of the left hand side manifold, possibly in a relaxed parametrization. Note that in
the two first theorems an oriented boundary also appears, and one of the points has a plus
sign and the other one a minus sign (orientation).
• In the right hand side the integrand (or summand) is a field appearing in the left hand side
differentiated in some way.
To get a coherent form for all these theorems, and others, we must define a general deriva-
tive for form fields, and the concept of an oriented boundary over which form fields can be
integrated. The final result then is the Generalized Stokes Theorem.
62
CHAPTER 5. GENERALIZED STOKES’ THEOREM 63
(i) p is not in M.
(ii) Any open subset containing p also contains a point of M. In particular this is true for
p-centered open balls.
The boundary of a manifold may be empty. E.g. the boundary of a sphere in R3 is empty. On
the other hand, the boundary of a manifold need not be a (lower-dimensional) manifold. E.g.
the boundary of the open square 0 < x, y < 1 of R2 is the perimeter of the square, and thus not
a manifold. In fact, the boundary of a bounded manifold can have a very complicated structure,
it may e.g. not have a Jordan measure. A more serious problem often is existence of tangent
spaces, needed for the orientation.
A better way to start is to define the boundary of a subset within a manifold. This is so,
because inside the manifold its smoothness and tangent spaces can be used for the orientation.
The boundary of a subset A ⊂ M
within the manifold, denoted by ∂M A,
consists of exactly all points of M such
that in every open set containing the point a nonsmooth point
of the boundary
there is a point of A and a point of M−A. M
(Cf. the condition (ii) above.) Note that a a smooth point
point of the boundary ∂M A may itself be A of the boundary
in A.
Consider then a point p of the bound-
ary ∂M A. Since the point p0 is in the MA
manifold M, for some open subset B con-
taining the point the intersection M ∩ B
is a local locus, cf. Theorem 2.4. In other
words, there is a continuously differentiable function Fp0 : B → Rn−k such that F′p0 has full
rank and p is in M ∩ B if and only if Fp0 (p) = 0. There may be several such functions Fp0
to choose from. The point p0 is called a smooth point of the boundary ∂M A if, in addition to
the function Fp0 , another continuously differentiable function gp0 : B → R can be chosen such
that
The subset of the boundary ∂M A consisting of exactly all smooth points is called the smooth
s
boundary of A in the manifold M, and denoted by ∂M A.
The definition is easily interpreted even when k = n. Item 1. is then not needed and the
only condition is gp0 (p) ≥ 0.
1
Note that this (usually) is not the boundary as a boundary of a set in Rn .
CHAPTER 5. GENERALIZED STOKES’ THEOREM 64
F (x, y, z) = z − xy = 0 0.2
0
determines in R3 a smooth surface (a 2-di- –0.2
mensional manifold) S, one of the so-called –0.4
saddle surfaces. The condition kpk ≤ 1 –1
specifies a subset of A. The boundary ∂S A x 0
is then the locus given by the conditions 1
1 0.5 0 –0.5 –1
( y
z − xy = 0
1 − x2 − y 2 − z 2 = 0,
in fact a 1-dimensional manifold. See the figure above (Maple). All points of the boundary ∂S A
are smooth points, since g can be chosen as
g(x, y, z) = 1 − x2 − y 2 − z 2 ,
and
F ′ (x, y, z) = (−y, −x, 1) and g ′(x, y, z) = (−2x, −2y, −2z)
are locally linearly independent (check!). Thus ∂Ss A = ∂S A.
N : 0 ≤ x ≤ 1, 0 ≤ y ≤ 1, z = 1
is a subset of the plane z = 1 (a 2-dimensional manifold). Its boundary is the perimeter, i.e.,
the union of the four line segments
y , 1−y , x and 1 − x,
respectively, and F as
F (x, y, z) = z − 1.
Proof. We use the (local) notation above, and consider an arbitrary point p0 of the smooth
boundary. By Corollary 2.3, the condition
(
Fp0 (p) = 0
gp0 (p) = 0
gp0 (p1 ) 6= 0.
(Since p1 is in M, we have Fp0 (p1 ) = 0.) By continuity then gp0 (p) 6= 0 also in some (small)
open subset B2 containing p1 . Thus in the set B2 , gp0 has only positive values or only negative
values. But then B2 cannot contain both points of M and points of M − A, and p1 cannot be a
point of the boundary of A in M. This contradiction shows that p1 ∈ K ∩ B1 .
Proving inclusion in the other direction we take a point p2 in the intersection K ∩ B1 , and
s s
show that it is also in the intersection ∂M A ∩ B1 . Assume again the contrary: p2 ∈
/ ∂M A ∩ B1 .
Now gp0 (p2 ) = 0, and in some (small) open subset B3 containing p2 , the function gp0 has either
only positive values or only negative values. Otherwise p2 would be a point of the smooth
boundary (take Fp2 = Fp0 and gp2 = gp0 ). This means that p2 is a local solution of the
constrained optimization problem
(
gp0 (p) = extremum!
Fp0 (p) = 0
(either a point of local maximum or local minimum, the local extremum value being = 0).
Using the method of Lagrange’s multipliers, we introduce the Lagrange function
where λ is the (row) vector of Lagrange multipliers. In the point (p2 , λ2 ) of local extremum its
derivative is zero:
∂L(p2 , λ2 )
= gp′ 0 (p2 ) + λ2 F′p0 (p2 ) = 0
∂p
then the direction of the vector t is opposite to the direction of increasing gp0 , that is, away from
the region A. Similarly, if
gp′ 0 (p0 ) • t > 0,
CHAPTER 5. GENERALIZED STOKES’ THEOREM 67
then the direction of the vector t is the direction of increasing gp0 , into the regionA.
s
Using exterior vectors we can couple an orientation for the smooth boundary ∂M A and an
orientation of M. Cf. Section 4.3. Consider then the case of a manifold M oriented by a k-form
field Φ. Thus in its points p vectors
t1 (p) , . . . , tk (p),
where texterior (p) is a suitable exterior vector depending on p. This choice couples orientations
s
of the boundary ∂M A and the manifold M. This oriented boundary of a region with boundary
−→ →
−
of an oriented manifold M is denoted by ∂ s−→ A.
M
Note that the vectors
s1 (p), . . . , sk−1(p)
are in the tangent space of the manifold M in the point p, so that the form field Φ is available.
s
Indeed, the tangent space of of the smooth boundary ∂M A in the point p is the null space of the
matrix
F′p0 (p)
!
,
gp′ 0 (p)
cf. Theorem 2.8 and the proof of Theorem 5.1. Thus it is included in the null space of the matrix
F′p0 (p), which again is the tangent space of the manifold M. What then is needed is a basis
for the tangent space of M in the points p of its smooth boundary having the same ”handidness”
as the basis t1 (p), . . . , tk (p).
Example. A parametrized smooth space curve
C : p = γ(u) (u ∈ U)
Φ(p; r1 , r2 ) = n(p) • r1 × r2 ,
y
where texterior(p)
∂γ(u) ∂γ(u)
x
n(p) = n γ(u) = ×
∂u1 ∂u2
is the normal vector in the point p. The value of the 2-form is then
∂γ(u) ∂γ(u)
2
×
∂u1 ∂u2
and it is always positive. The smooth boundary of a region with orientation A of S consists of
smooth space curves, and it is oriented in a smooth point p = γ(u) by the form field condition
Φboundary p; s(p) = n(p) • texterior (p) × s(p) > 0,
where s(p) is the tangent vector of the boundary (curve) in the point p. The vectors
then form a right-handed system, and the orientation is given by the familiar ”right hand rule”,
see the figure below.
Example. The tangent space of a 3-dimensional manifold (or open subset) M of R3 always
consists of the whole vector space R3 . A region with boundary (or a solid body) A of M can
be oriented, say, by the 3-form field condition
then it is all about orienting the boundary (surface) in such a way that the normal vector n(p)
is the exterior normal, and that
where n(p) is a vector (exterior normal) of the normal space and s1 (p), . . . , sn−1 (p) form a
s
basis of the tangent space of the smooth boundary ∂M A in the point p.
f (x + h) − f (x)
f ′ (x) = lim .
h→0 h
In basic courses of calculus also the corresponding 1-form field, the so-called differential
f (x + hr) − f (x)
f ′ (x)r = lim .
h→0 h
The exterior derivative is an extension of this for k-form fields. Note in particular that the
function f is a 0-form field and that the differential df is a 1-form field. Taking an exterior
derivative increases the degree of a form by one.
The interval
[x, x + h] (or [x + h, x])
appearing in the difference quotient is replaced by a k + 1-dimensional (closed) parallelepiped
k+1
X
P(p; r1 , . . . , rk+1) = p+ ui rTi 0 ≤ u1 , . . . , uk+1 ≤ 1 .
i=1
For the real line this is a closed interval, in the plane R2 a parallelogram and in the space R3
a ”geometric” parallelepiped. Here k + 1 ≤ n, so we could be dealing with a parallelepiped
embedded in a higher-dimensional space.
Embedded in the affine subspace
k+1
X
T
A(p; r1, . . . , rk+1 ) = p + ui ri u1 , . . . , uk+1 ∈ R
i=1
CHAPTER 5. GENERALIZED STOKES’ THEOREM 70
In the former face ri is an interior vector and in the latter it is an exterior vector. These faces
thus have opposite orientations, and it is agreed that A(p; r1, . . . , rk+1 ) is oriented according to
the orientations of the faces Ti+ by the exterior vectors ri .
The smooth boundary consists of faces minus their edges. By Theorem 3.3 the edges are
k-null-sets, and by Theorem 5.1 the smooth boundary is a k-dimensional manifold. The oriented
sum in the difference quotient, where f (x + hr) has a plus-sign and f (x) a minus-sign, is now
replaced by an integral over the oriented (smooth) boundary of the parallelepiped. Alternation
of the signs in included, too, since the faces Ti− and Ti+ always have opposite orientations.
We thus get the exterior derivative of a k-form field Φ as
1
Z
(dΦ)(p; r1 , . . . , rk+1) = lim k+1 Φ,
h→0 h
−
→
∂→− Ph
A
Note that the integration region can be given by a relaxed parametrization, so that the integral
is of the form in Section 4.2. (As an oriented sum if k = 0.) In addition it is agreed that if
k + 1 > n, then the exterior derivative is = 0.
Exterior derivative calculation is tedious starting from the very definition, as often is calcu-
lation of the usual univariate derivative, too. To make things simpler, differentiation rules can
be given. Together they form a sufficiently extensive toolbox to make derivative calculations
relatively easy. Proving these rules then can be quite tedious.
Let us give a basic collection of differentiation rules. The vector variables r1 , . . . , rk are
mostly omitted for brevity.
(I) If Φ and Ψ are k-form fields, and c1 and c2 are constants, then
Taking an exterior derivative thus is a linear operation. This follows directly from the
definition by the integral.
(II) If a k-form field Φ is constant (i.e. a constant form), then dΦ is a k + 1-form field having
only the value 0 (i.e. a zero form field).
As is usual, derivative of a constant is zero. This, too, follows directly from the definition
by the integral, since as a consequence of the opposite orientations of the pairs of faces
the integral then is always = 0. (The integrand having the same value in the matching
points of the faces.)
CHAPTER 5. GENERALIZED STOKES’ THEOREM 71
f (p + hrT ) − f (p) d
lim = f (p + tr ) = f ′ (p)r
T
h→0 h dt t=0
(by the chain rule). In particular, if krk = 1, this is the usual directional derivative in the
direction of r.
This is the key rule since, combined with the previous rules, it makes it possible to cal-
culate the exterior derivative starting from the expansion of the form field as a linear
combination of elementary form fields, and in fact also shows that the exterior derivative
is a k + 1-form field. The proof of this rule (even via Theorem 6.1) is a bit tedious, see
e.g. H UBBARD & H UBBARD .
(V) The k + 2-form d2 Φ = d(dΦ) field obtained by taking the exterior derivative of a k-form
field Φ twice is a constant form field having the value 0 (a zero form field).
Form fields are thus always ”of the first degree” in a sense. Differentiating twice always
gives zero. This is a consequence of the previous rules, since if f is a function, then
n n X n
X ∂f (p) X ∂ 2 f (p)
(d2 f )(p; ·, ·) = d dxi = dxj ∧ dxi = 0.
i=1
∂x i j=1 i=1
∂x j ∂xi
(Recall that dxi ∧ dxj = −dxj ∧ dxi and dxi ∧ dxi = 0.)
Note. As a matter of fact, the whole exterior derivative could be defined via these rules. It is,
however, nice to see that it is a generalization of the ordinary univariate derivative. And the
definition by the integral does give a ”geometric” way of considering the exterior derivative,
not just a symbolical machinery.
of R4 . (For brevity p = (x1 , x2 , x3 , x4 ) as well as r1 and r2 are omitted.) Applying the rules we
get
For exterior derivatives of wedge products of form fields we have the famous
Cartan’s Magic Formula.2 The exterior derivative of the wedge product of a k-form field Φ
and an l-form field Ψ is given by
Proof. By the definition of wedge product it suffices to show the rule for elementary form fields
By rule (IV)
n
X ∂ a(p)b(p)
d(Φ ∧ Ψ) = dxi ∧ dxj1 ∧ · · · ∧ dxjk ∧ dxh1 ∧ · · · ∧ dxhl
i=1
∂xi
n
X ∂a(p) ∂b(p)
= b(p) + a(p) dxi ∧ dxj1 ∧ · · · ∧ dxjk ∧ dxh1 ∧ · · · ∧ dxhl
i=1
∂xi ∂xi
n
X ∂a(p)
= dxi ∧ dxj1 ∧ · · · ∧ dxjk ∧ b(p) dxh1 ∧ · · · ∧ dxhl
i=1
∂xi
n
X ∂b(p)
+ a(p) dxj1 ∧ · · · ∧ dxjk ∧ (−1)k
dxi ∧ dxh1 ∧ · · · ∧ dxhl
i=1
∂xi
2
La ”formule magique de Cartan”. Élie Cartan was the father of form theory.
CHAPTER 5. GENERALIZED STOKES’ THEOREM 73
(a 2-form field), and for a scalar field f (p) the basic form field (in addition to itself as a 0-form
field) is the density form field
(a 3-form field). The connection between these basic form fields and exterior derivatives is the
following:
(A) df = Φ∇f –work (df is the work form field of the gradient ∇f .)
(B) dΦF–work = Φ∇×F–flux (dΦF–work is the flux form field of the curl ∇ × F.)
(C) dΦF–flux = Φ∇•F–density (dΦF–flux is the density form of the divergence ∇ • F.)
All these can be proved by straightforward calculation. (A) follows immediately from rule (III).
(B) is shown as follows:
According to the scheme, curl of a gradient corresponds to exterior derivation twice, as does
divergence of a curl. By the rule (V) these are zero form fields, and indeed
∇ × ∇f = 0 and ∇ • (∇ × F) = 0.
On the other hand, the other second order derivatives ∇ • (∇f ) = ∇2 f = ∆f (Laplacian),
∇(∇ • F) and ∇ × (∇ × F) (double curl) do not correspond to second exterior derivatives, and
in general do not vanish.
Cartan’s Magic Formula gives fairly directly nabla formulas (i), (iii), (iv) and (v) in Section
1.5 since, as noted in Section 4.4, for scalar fields f and g, and vector fields F and G we have
f ∧ g = fg , ΦF–work ∧ ΦG–work = ΦF×G–flux ,
f ∧ ΦF–work = Φf F–work , ΦF–work ∧ ΦG–flux = ΦF•G–density ,
f ∧ ΦF–flux = Φf F–flux .
For instance, the formula
(v) ∇ • F × G = ∇ × F • G − F • ∇ × G
is obtained as follows:
Φ∇•F×G–density = dΦF×G–flux = dΦF–work ∧ ΦG–work + (−1)1 ΦF–work ∧ dΦG–work
= Φ∇×F–flux ∧ ΦG–work − ΦF–work ∧ Φ∇×G–flux
= Φ∇×F•G–density − ΦF•∇×G–density
= Φ(∇×F•G−F•∇×G)–density .
Note. Formulas (vi) and (vii) in Section 1.5, on the other hand, are of a quite different form,
and do not follow directly from the Magic Formula. They are more closely related to another
rule of derivation given in Theorem 6.1.
Let us then give some physical intuition for the differential operators. Thinking about the
definition of the exterior derivative as the limit of the integral in the previous section, it may
be noticed that in a very small parallelogram Ph (p; r1 , r2 ) (where h is small) embedded in R3
hr2 hr1
p
p Ph Ph
hr1 hr3
the scaled projection of the curl on the normal of the parallelogram (see the figure on the left
above) is
Z Z I
∇ × F(p) • (hr1 ) × (hr1 ) ∼ ′
= ΦF–work = F γ(u) • γ (u) du = F(r) • ds,
−
→ U −
→
∂→− Ph ∂→− Ph
A A
CHAPTER 5. GENERALIZED STOKES’ THEOREM 75
→
−
where the (relaxed) parametrization of ∂ − → Ph is r = γ(u) (u ∈ U). The projection thus is
A
approximatively the line integral of the vector field around the perimeter of the parallelogram
in the direction given by the right hand rule. For this reason the curl is sometimes also called
vorticity or vortex density.
Similarly in a very small (meaning h is small) parallelepiped Ph (p; r1 , r2 , r3 ) of R3 the
scaled divergence is
Z I
∼
∇ • F(p) det(hr1 , hr2 , hr3 ) = ΦF–flux = F(r) • dS.
−
→ −
→
∂→− Ph ∂→− Ph
A A
Locally the divergence thus is the flux through the boundary of a very small parallelepiped from
the inside to the outside, see the figure on the right above. This could be a change of mass of
a fluid in the parallelepiped. For this reason the divergence is also called source density. It is a
typical density since det(hr1 , hr2 , hr3 ) is the volume of the parallelepiped.
Combining the Hodge star (see Section 4.4) and exterior derivation gives a handy tool, and
a way to get higher derivatives. We see (check!) that in R3
d∗ΦF–work = Φ∇•F–density and ∗d∗ΦF–work = ∇ • F ,
d∗d∗ΦF–work = Φ∇(∇•F)–work and ∗d∗d∗ΦF–work = Φ∇(∇•F)–flux ,
d∗ΦF–flux = Φ∇×F–flux and ∗d∗ΦF–flux = Φ∇×F–work ,
d∗d∗ΦF–flux = Φ∇×(∇×F)–flux and ∗d∗d∗ΦF–flux = Φ∇×(∇×F)–work ,
d∗Φf –density = Φ∇f –work and ∗d∗Φf –density = Φ∇f –flux ,
d∗d∗Φf –density = Φ∆f –density and ∗d∗d∗Φf –density = ∆f.
Electro-magnetic fields are governed by the famous Maxwell equations:
(M1) ∇ • D = ρ (Gauß’ law)
3
We are dealing with the isotropic case only. Here E is the electric field [V/m], D is the electric flux density
[As/m2 ], H is the magnetic field [A/m], B is the magnetic flux density [Vs/m2 ] (tesla), ε is the permittivity
[As/V/m], µ is the permeability [Vs/A/m], ρ is the charge density [As/m3 ], J is the current density [A/m2 ], and σ
is the conductivity [A/V/m].
CHAPTER 5. GENERALIZED STOKES’ THEOREM 76
and applying the matter laws and (M1) and (M4), the exterior derivative is simplified (check!)
to
1
dΦMaxwell = (Φρ–density − ΦJ–flux ∧ dt).
ε
This exterior derivative, the so-called current density form field, is often denoted as a Hodge
dual ∗J. (What is then J?). Second exterior derivatives being zeros, we have then also
All in all, we then note that (in the homogeneous case) Maxwell’s equations can be written
using the Faraday form field and the Maxwell form field in an utterly simple form:
(
dΦFaraday = 0
dΦMaxwell = ∗J.
(We did need all Maxwell’s equations in deriving these two equations!)
CHAPTER 5. GENERALIZED STOKES’ THEOREM 77
→
−
In case the smooth boundary ∂ s−→ A is empty, the right-hand side integral is = 0.
M
Proof. A rigorous proof consists of a long and complicated estimation5, see e.g. H UBBARD &
H UBBARD . The rough idea however is the following. We approximate A by a set of very small
k + 1-dimensional parallelepipeds6 :
whence
γ ′j (u) = rj,1 , · · · , rj,k+1 .
The larger the number of parallelepipeds is, the better the approximation:
4
This is often written in the short form
Z Z Z Z
dΦ = Φ or even dΦ = Φ
→
− →
− A ∂A
A ∂A
(figure by Maple). Thus, by the definition of the parametrized integral as a Riemann integral,
Z N
dΦ ∼
X
= (dΦ)(pj ; rj,1, . . . , rj,k+1).
A j=1
On the other hand, by the definition of the exterior derivative (taking h = 1, since the paral-
lelepipeds are very small),
Z
∼
(dΦ)(pj ; rj,1, . . . , rj,k+1) = Φ.
−
→s
∂−→ Pj
M
Exterior normals of faces shared by adjacent parallelepipeds are opposite, i.e. the faces have
opposite orientations. Thus integrals over such faces are cancelled, and only integrals over
faces approximating the smooth boundary of A remain. So
N Z Z
Φ∼
X
= Φ.
j=1 −
→s −
→s
∂−→ Pj ∂−→A
M M
dΦF–flux = Φ∇•F–density ,
(note the brief notation). It is important here that the exterior normal corresponds to the right-
handedness of the coordinate system.
Example. (General Divergence Theorem) The Divergence Theorem is valid in any dimension
in the form Z I
Φ∇•F–density = ΦF–flux
−
→ −
→s
K ∂−→K
M
(cf. also the example in Section A2.2). The 1-dimensional version is the usual Fundamental
Theorem of Integral Calculus.
In particular the Divergence Theorem is valid in two dimensions. Then
and so
d − F2 (p)dx + F1 (p)dy = ∇ • F(p) dx ∧ dy.
CHAPTER 5. GENERALIZED STOKES’ THEOREM 79
Here the manifold is an open subset of R2 , and the region A is its subset whose smooth boundary
is oriented by exterior normals. Note (check!) that then integration around the boundary (curve)
of A is in the counterclockwise direction. The right hand side is sometimes written in one of the
forms I I I
F1 (r) dx
F(r) • dn = −F2 (r) dx + F1 (r) dy = .
F2 (r) dy
−
→ −
→ −
→
∂A ∂A ∂A
Example. (Green’s Theorem) Applying the 2-dimensional Divergence Theorem above to the
vector field
F2 (r)
G(r) =
−F1 (r)
(i.e., F(r) rotated by −90◦ ) we get Green’s Theorem:
I
∂F2 ∂F1
Z Z I
−G2 (r)
− dr = ∇ • G(r) dr = • ds = F(r) • ds.
∂x ∂y G1 (r)
−
→ −
→ −
→ −
→
A A ∂A ∂A
Again integration around the boundary (curve) of A is in the counterclockwise direction (cor-
responding to the exterior normals).
→
−
Example. (Stokes’ Theorem) As the manifold take an oriented surface S of R3 , and as the
→
− →
−
region its correspondingly oriented part A whose smooth boundary ∂ s− → A is oriented by the
S
right-hand rule. Since
dΦF–work = Φ∇×F–flux ,
then according to the Generalized Stokes Theorem
Z Z I I
∇ × F(r) • dS = Φ∇×F–flux = ΦF–work = F(r) • ds.
−
→ −
→ −
→s −
→
A A ∂→−A ∂A
S
In case A = S is a closed surface having no smooth boundary, the right hand side is = 0.
These formulations of the classical integral theorems are fairly general, relaxed parametriza-
tions and sufficiently continuous integrands is more or less what is needed. Manifolds and their
regions with boundaries may consist of several separate parts, as may their smooth boundaries.
Note however, that in Stokes’ Theorem the surface S must be orientable, so e.g. the Möbius
band does not qualify.
There are also nonorientable closed surfaces, e.g. the so-called Klein bottles, a version of
which is depicted below (Maple). This surface intersects itself, in R4 there however are versions
of Klein’s bottles which do not. Gauß’ Theorem is not applicable to Klein’s bottles.
CHAPTER 5. GENERALIZED STOKES’ THEOREM 80
f (p)r1 • r2 × r3
Since
∇ • f (r)a = ∇f (r) • a
and a was an arbitrary constant vector (choosing a = i, j, k we get the coordinates), this result
is often written in the vectoral form
Z I
∇f (r) dr = f (r) dS.
−
→ −
→
K ∂K
The classical integral theorems, and others, find their use in the manipulation of physical
formulas.
∇ × H = J.
→
− −
→
If the closed curve C (loop) is sufficiently smooth and oriented, there is an oriented surface S ,
→
−
having C as its oriented boundary (via the right hand rule) and satisfying the assumptions of
Stokes’s Theorem, and
I Z Z
H(r) • ds = ∇ × H(r) • dS = J(r) • dS.
−
→ −
→ −
→
C S S
The left hand side is the integral of the magnetic field around the loop, and the right hand side
tells it is equal to the net current through the loop. This is the so-called Ampère Piercing law.
Thus the electromotive force around a loop equals the time derivative of the net magnetic flux
through the loop.
Example. As a first approximation—and often as a last one, too—in many fluid dynamics prob-
lems the fluid is assumed to be ideal, inviscid and incompressible. Let us consider a stationary
flow. The so-called Thomson (Kelvin) circulation law tells us that, with certain assumptions, if
the curl of the velocity is a zero vector initially it will remain so afterwards. Thus then all the
time
∇ × v = 0.
If C is a closed curve (loop) and a boundary of a surface satisfying the assumptions of the Stokes
Theorem, and v is defined and continuously differentiable, then the integral around the loop is
zero: I
v(r) • ds = 0.
−
→
C
always has the same value, and this can then be calculated using the best loop for the purpose.
This is important e.g. when the Zhukovsky7 lift law is applied. The law tells the lift per unit of
length of an airfoil is
F = ρΓv∞
for air of density ρ flowing from infinity with speed v∞ . The force is perpendiculat to the velocity,
as is familiar from airplane wings.
Independence of the value of the line integral of the loop is seen roughly as follows. If
→
− →
−
C1 and C2 are two loops encircling the same (infinitely long) obstacle, they can be cut and
7
Often Joukowsky or Joukowski or Schukowski etc.!
CHAPTER 5. GENERALIZED STOKES’ THEOREM 82
→
−
recombined using a double connecting line L to a single closed oriented curve C as in the
figure below. (The connecting line is drawn here as two lines for clarity.) The closed loop
C then is a boundary curve of a surface where the velocity v is defined, and since the curl
vanishes,
I
y y L
v(r) • ds = 0.
C
−
→
C C1 C2
On the other hand, in the line integral
→
−
over C the connecting line L is traversed
twice in opposite directions. These two
integrals over L cancel each other, and x
z x z
so (think about the orientations)
I I I
v(r) • ds = v(r) • ds − v(r) • ds = 0.
−
→ −
→ −
→
C C2 C1
Ψ is called the (standard) symplectic potential8 (see the next chapter). Consider then a
−→
2-dimensional oriented manifold M of R2m and its bounded region A, where the boundary
of A has the (relaxed) parametrization
C : r = γ 1 (u), γ 2 (u) (u ∈ U)
and γ 1 contains the first m coordinates. By the Generalized Stokes Theorem then
Z Z I Z
Φ = dΨ = Ψ = γ 1 (u)γ ′2 (u) du.
−
→ −
→ −
→ U
A A C
In R2 this also follows from Green’s Theorem (and then the integral equals the area of A).
Symplectic forms are important in mechanics where the phase spaces may have a quite high
(even) dimension.
Note. The foremost problem in results like these is that in our formalism the regions of inte-
gration must in principle be embedded in larger manifolds, and that the fields must exist (as
continuous or continuously differentiable) in regions larger than the region of integration. In
many practical situations there are no such extensions, not even theoretical or artificial ones.
8
And also the Liouville 1-form, the Poincaré 1-form, the canonical 1-form or the tautological 1-form, etc. As
they say: ”A dear child has many names.”
”Truth in the world resides only in
mathematical proofs and physics labs.”
Chapter 6
POTENTIAL
1
A form field whose exterior derivative is = 0, is often called closed.
83
CHAPTER 6. POTENTIAL 84
i.e. the work form field ∇f (p) • r. On the other hand, we have dΦ = 0, since the exterior
derivative is the flux form field of the curl
−C(p − p1 )
∇× =0 (check!).
kp − p1 k3
Here of course
(p − p1 )
−C = ∇f (p)
kp − p1 k3
is the familiar Newton vector field. It will be remembered that according to Newton’s law of
gravitation Earth attracts a point mass m in p by the force
p − p1
−GmM ,
kp − p1 k3
where p1 is the center of Earth, M is Earth’s mass, and G is the gravitational constant.
Example. Newton’s vector field can also be connected to the 2-form field
(p − p1 ) • r1 × r2
Φ(p; r1 , r2 ) = −C .
kp − p1 k3
This form field is exact, too, since it is a flux form field and its exterior derivative is the density
form field of the divergence
−C(p − p1 )
∇• = 0 (check!).
kp − p1 k3
Here, however, existence of a potential may actually depend on the region, as will be seen.
To deal with general potentials one more exterior differentiation rule is needed, in addition
to those in Section 5.2. This could also be used as a (coordinate-free) definition of the exterior
derivative.
Theorem 6.1. The exterior derivative of a k-form field Φ can be written as
k+1
X
(dΦ)(p; r1 , . . . , rk+1) = (−1)i−1 Φ′ (p; r1 , . . . , ri−1 , rbi , ri+1, . . . , rk+1)ri ,
i=1
where Φ′ is the derivative of Φ with respect to the point variable p and the circumflex above a
vector variable means that it is omitted from the list.
Proof. By Theorem 4.2, Φ can be expanded as a combination of elementary form fields as
X
Φ(p; r1 , . . . , rk ) = aj1 ,j2 ,...,jk (p)(dxj1 ∧ dxj2 ∧ · · · ∧ dxjk )(r1 , . . . , rk ),
1≤j1 <j2 <···<jk ≤n
When this determinant is expanded along the first row, we get the sum
k+1
X
(−1)i−1 xl,i (dxj1 ∧ · · · ∧ dxjk )(r1 , . . . , ri−1 , rbi , ri+1, . . . , rk+1 ).
i=1
Pn Pk+1
Exchanging the summings l=1 and i=1 we then get the differentiation formula.
The general result will be stronger than just a local one: It will deal with the so-called star-
shaped sets. A subset A of Rn is said to be star-shaped with respect to the point p0 , if whenever
a point p is in A, then the whole line segment connecting it to p0 , i.e. the set
p0 + t(p − p0 ) 0 ≤ t ≤ 1 ,
whence by multilinearity
n
X
Φ(tp; p, r1 , . . . , rk ) = pl Φ(tp; el , r1 , . . . , rk )
l=1
CHAPTER 6. POTENTIAL 86
and by differentiating
n
X
′
Φ(tp; p, r1 , . . . , rk ) = q + tpl Φ′ (tp; el , r1 , . . . , rk ) = q + tΦ′ (tp; p, r1 , . . . , rk ),
l=1
where
q = Φ(tp; e1 , r1, . . . , rk ), . . . , Φ(tp; en , r1, . . . , rk ) .
Let us first calculate d(IΦ) as in Theorem 6.1 using the above derivative and exchanging
exterior derivation and integration2:
(d(IΦ))(p; r1 , . . . , rk+1)
k+1
X
= (−1)i−1 (IΦ)′ (p; r1 , . . . , ri−1 , rbi , ri+1 , . . . , rk+1)ri
i=1
k+1
X Z1
i−1
= (−1) tk Φ(tp; ri , r1 , . . . , ri−1 , rbi , ri+1, . . . , rk+1 ) dt
i=1 0
k+1
X Z1
+ (−1)i−1 tk+1 Φ′ (tp; p, r1 , . . . , ri−1 , rbi , ri+1 , . . . , rk+1)ri dt.
i=1 0
Note that
qri = Φ(tp; ri , r1 , . . . , ri−1 , rbi , ri+1 , . . . , rk+1).
The first sum is simplified a lot by antisymmetry when the ri ’s are moved to their ”correct”
positions:
k+1
X Z1 Z1
(−1)i−1 tk Φ(tp; ri , r1 , . . . , ri−1, rbi , ri+1 , . . . , rk+1) dt = (k+1) tk Φ(tp; r1 , . . . , rk+1) dt.
i=1 0 0
(Note the powers t and (−1) . Why these?) Combining these two results we see that
k+1 i
d(IΦ) (p; r1 , . . . , rk+1) + I(dΦ) (p; r1 , . . . , rk+1)
Z1 Z1
= (k + 1)t Φ(tp; r1 , . . . , rk+1) dt + tk+1 Φ′ (tp; r1 , . . . , rk+1 )p dt
k
0 0
Z1
d k+1
= t Φ(tp; r1 , . . . , rk+1) dt = Φ(p; r1 , . . . , rk+1 ).
dt
0
Example. If the n-form field Φ is the density form field of the sclar field f , i.e. of the form
then dΦ = 0 and Φ is exact. By Poincaré’s Lemma in a star-shaped region with respect to some
point p0 then Φ has the potential
Z1
(IΦ)(p; r1 , . . . , rn−1 ) = tn−1 f p0 + t(p − p0 ) dt det(p − p0 , r1 , . . . , rn−1 ),
0
which is the flux-form field ΦF–flux (p; r1 , . . . , rn−1 ) of the vector field
Z1
F(p) = tn−1 f p0 + t(p − p0 ) dt (p − p0 ).
0
In a sense this is the one and only analogue of the integral function of a univariate function.
For instance, in R3 Gauß’ Theorem implies
Z Z I
f (r) dr = ∇ • F(r) dr = F(r) • dS
−
→ −
→ −
→
K K ∂K
f (x, y, z) = x + y + z and p0 = 0.
The result is
Z1 x x
2 x+y+z
F(x, y, z) = t (tx + ty + tz) dt y = y .
4
0 z z
Φ(p; r) = F(p) • r
(the work form field of the vector field F) has a potential in star-shaped regions with respect to
some point p0 whenever
dΦ = Φ∇×F–flux = 0,
i.e. whenever ∇ × F = 0, or the field is irrotational, and one such potential is
Z1
f (p) = F p0 + t(p − p0 ) • (p − p0 ) dt.
0
The potential (a 0-form field, or a scalar field) f is then called a scalar potential of the vector
field F, and F = ∇f .
CHAPTER 6. POTENTIAL 88
In engineering applications the vector field F is in general unknown, and the task is to find
it—analytically or numerically—based on given initial data. This would mean finding three
scalar-valued functions (the components of the field). In the irrotational case the unknown
vector field can be given as the gradient of an unknown scalar field: F = ∇f . To find the vector
field it thus suffices to find only one scalar-valued function, the potential f . Irrotationality
usually follows from the statement of the problem (by Thomson’s circulation law, Maxwell’s
equations, etc.).
As noticed, if f is a scalar potential and c is a constant, then f + c is another scalar potential
for the same vector field, because
∇(f + c) = ∇f.
On the other hand, if f and g both are scalar potentials of a vector field, then ∇(f − g) = 0 and
f − g is thus constant. A scalar potential is thus unique modulo a scalar additive constant.
There is a connection between existence of a scalar potential and line integrals:
Theorem 6.2. The following three conditions are equivalent for a continuous vector field F
defined in an open subset (manifold) K of R3 .
(i) F has in K a scalar potential f .
(iii) If the points r0 and r1 are in K and the curve (1-dimensional oriented manifold in relaxed
→
−
parametrization) C in K connects them, i.e., in the orientation r0 is a fixed initial point
and r1 a variable terminal point, then the value of integral
Z
F(r) • ds = h(r1 )
−
→
C
→
−
depends on the choice of the curve C only through the terminal point r1 , and thus defines
a function h(r1 ) of r1 .
The function h in item (iii) is a potential of F in K.
Proof. Let us first establish the implication (i) ⇒ (iii). So, let us assume item (i), and take a
→
−
relaxed parametrization r = γ(u) (a ≤ u ≤ b) of C . Using the assumed potential f we get
Z Z Zb
F(r) • ds = ∇f (r) • ds = f ′ γ(u) γ ′ (u) du
−
→ −
→ a
C C
Zb . b
d
= f γ(u) du = f γ(u) = f (r1 ) − f (r0 ),
du
a a
∂h(r1 )
= F1 (r1 ), C1
∂x1
r0
the other partial derivatives being treated similarly. x
→
−
For this, assume r1 is in the manifold K and take C to
be an oriented curve where the last ”piece” is a short
line segment z
→
−
C 2 : r = r2 ± u(1, 0, 0) (0 < u ≤ ±(x1 − x2 ))
parallel to the x-axis, where r2 = (x2 , y2 , z2 ) and the sign depends on the direction (i.e., plus if
→
−
x1 > x2 , minus otherwise), see the figure below. Clearly ±i is a tangent vector of C2 . The value
of the integral does not depend on the path, so this choice is possible. Since K is an open subset,
→
− →
− → −
− → − →
there is the space available, too. Let us denote the initial part of C by C1 , so C = C 1 + C2 .
If the direction is from left to right (meaning that the sign is plus, the case of a minus sign is
similar), then
Z Z Z
h(r1 ) = F(r) • ds = F(r) • ds + F(r) • ds
−
→ −
→ −
→
C C1 C2
Z xZ
1 −x2
= F(r) • ds + F r2 + u(1, 0, 0) • i du
−
→ 0
C1
Z xZ
1 −x2
by item (ii).
−
→
Implication (iii) ⇒ (ii) is also fairly obvious. Assume item (iii) and take a closed curve C ,
→
−
and two points r0 and r1 of C. Thus C is divided into two parts, the oriented curve C ′ connecting
→
−
r0 and r1 , and the oriented curve C ′′ connecting r1 and r0 . Reversing the orientation of the latter
→
− →
−
curve we get two curves C ′ and − C ′′ , as in item (iii), see the figure below. It then follows from
CHAPTER 6. POTENTIAL 90
It may also be noticed that the surface S can be globally extremely complicated, just think
of a case where C is a knot (figure: Maple):
Example. As an example of a manifold which is not simply connected take the inside of a torus,
say the one parametrized as
4
x = (R1 + R2 cos u2 ) cos u1 3
–4 –4
y = (R1 + R2 cos u2 ) sin u1 (0 ≤ u1 , u2 ≤ 2π). –3
2
–3
–2 1 –2
z = R2 sin u2 –1 –1 0
1 –1
See the figure on the right (Maple). 1
2 –2 2
Then e.g. the vector field 3 3
–3
−y 4 4
1 –4
F(r) = 2 x
x + y2
0
does not have a potential even though it is easily verified that it is irrotational (check!). This
follows from Theorem 6.2 since for e.g. the centre circle
→
−
C : γ(u1 ) = (R1 cos u1 , R1 sin u1 , 0) (0 ≤ u1 ≤ 2π)
we have
I Z2π −R1 sin u1 −R1 sin u1 Z2π
1
F(r) • ds = R1 cos u1 • R1 cos u1 du1 = du1 = 2π 6= 0.
R12
−
→ 0 0 0 0
C
CHAPTER 6. POTENTIAL 92
On the other hand, after removing the part of the torus in the right xy-plane the potential
appears, it is atan (see the example in Section 2.5).
If the manifold is not simply connected, an irrotational vector field may still have a local
scalar potential, but it need not be globally unique modulo an additive constant. Since nu-
merical methods can only find unique solutions, the manifold must be made artificially simply
connected by ”cutting it open” using a suitable surface, as was done in the above example.
Note. The condition in Theorem 6.3 is sufficient but not necessary. In some cases an irrotational
vector field has a unique scalar potential (modulo the additive constant) even when the manifold
is not simply connected. This is the case e.g. for some problems dealing with the electric field
E when the magnetic field is stationary and there are no electromotive forces. For reasons of
energy conservation, the integral around a closed curve C is then
I
E(r) • ds = 0.
C
(the work form field of the vector field U). A vector field U such that ΦU–work is a potential of
Φ is called a vector potential of the vector field F, and then F = ∇ × U.
Example. Let us find a vector potential for the vector field
x
F(x, y, z) = y
−2z
in R3 . It is easily verified that this is solenoidal. Choosing p0 = 0 and integrating we get
(omitting the vector variable r)
Z1 tx x Z1 3yz yz
U(x, y, z) = t ty × y dt = t2 −3xz dt = −xz .
0 −2tz z 0 0 0
CHAPTER 6. POTENTIAL 93
Example. As another example let us take a more complicated but sometimes very useful case
(cf. Appendix 3). Let us find a vector potential for the Newton vector field
p − p1
F(p) = −C ,
kp − p1 k3
in the manifold which we get by removing from R3 the ray (so-called Dirac’s string)
p1 + u(p1 − p0 ) (u ≥ 0).
The resulting region is star-shaped with respect to the point p0 assuming—as is done—that
p0 6= p1 . Note that the point p1 is then removed but the point p0 is not. For brevity we denote
4a + 2b = 4(p0 − p1 ) • (p − p1 ) ,
a + b + c = kp − p1 k2 and
2
d = 4kp0 − p1 k2 kp − p1 k2 − 4 (p0 − p1 ) • (p − p1 ) .
Calculate then the vector potential (omitting the vector variable r and the constant −C):
Z1
p0 + t(p − p0 ) − p1
U(p) = t
× (p − p0 ) dt
p0 + t(p − p0 ) − p1
3
0
Z1
t dt
= (p0 − p1 ) × (p − p1 )
(a + bt + ct2 )3/2
0
.1 −2(2a + bt)
= (p0 − p1 ) × (p − p1 ) √
0
d a + bt + ct2
√ √
4 a a + b + c − (4a + 2b)
= (p0 − p1 ) × (p − p1 ) √
d a+b+c
kp0 − p1 kkp − p1 k − (p0 − p1 ) • (p − p1 )
= (p0 − p1 ) × (p − p1 ) 2
kp − p1 k kp0 − p1 k2 kp − p1 k2 − (p0 − p1 ) • (p − p1 )
(p0 − p1 ) × (p − p1 )
= .
kp0 − p1 kkp − p1 k2 + (p0 − p1 ) • (p − p1 ) kp − p1 k
This result is in many ways the best possible, it is scarcely possible to remove anything less than
the ray from R3 to get the vector potential. (See the example in Section 3 of Appendix 3.)
CHAPTER 6. POTENTIAL 94
An unknown solenoidal vector field can then be given by as the curl of another (also so-far
unknown) vector field, the vector potential: F = ∇ × U. Solenoidality usually follows from
the statement of the problem. For form fields, often a work form field is somehow ”simpler”
than a flux form field.
If U is a vector potential of the vector field F and ∇f is a gradient field, then U + ∇f is
also a vector potential of F, because
∇ × (U + ∇f ) = ∇ × U + ∇ × ∇f = ∇ × U.
On the other hand, if both U and V are vector potentials of the same vector field, then
∇ × (U − V) = 0 and U − V is irrotational. A vector potential is thus unique modulo an
additive irrotational vector field. In case the manifold allows existence of a scalar potential—
e.g. being simply connected—vector potential is unique modulo an additive gradient field. But
if scalar potentials do not exist, then this might not be true. For instance, in the torus of the
example in the previous section the zero vector field 0 surely has a vector potential, one such is
of course 0, but
−y
1 x
F(r) = 2
x + y2
0
is also a vector potential of 0, and it is not a gradient field.
There is a connection between existence of vector potentials and surface integrals given by
the Generalized Stokes Theorem, once we remember that a surface integral is the integral of
the flux form field of the curl of the vector potential, and that the flux form field is the exterior
derivative of the work form field of the vector potential:
Theorem 6.4. If a continuously differentiable vector field F has a vector potential U in the open
→
−
subset K of R3 , and S is a closed oriented surface in K (a 2-dimensional oriented manifold in
extended parametrization having an empty smooth boundary), then
I I
F(r) • dS = ∇ × U(r) • dS = 0.
−
→ −
→
S S
As was the case for the scalar potential, existence of a vector potential does not depend only
on the field but also on the manifold K. In star-shaped manifolds solenoidal vector fields have a
vector potential by Poincaré’s Lemma. In other cases the situation is more complicated. In any
case, the necessary condition of the theorem above must hold true.
Example. A typical example of a case where a solenoidal vector field does not have a vector
potential is the Newton vector field
r − r0
F(r) =
kr − r0 k3
in a ”punctured” manifold K, where the point r0 is removed but which otherwise completely
”surrounds” the point r0 . As was noted before, this vector field is solenoidal. Integrating over
a small r0 -centered sphere
The conclusion of Theorem 6.4 is thus not valid, so no vector potential exists.
Gauß’ Theorem, too, has its consequences for vector potentials. If the closed oriented sur-
→
−
face S in Theorem 6.4 encloses a solid L which is included in the manifold K, then, in line
with the theorem, according to Gauß’ Theorem
I Z
F(r) • dS = ∇ • F(r) dr = 0,
−
→ L
S
because in order to have a vector potential the vector field F must be solenoidal.
But what if not all of L is contained in the manifold K, but K has ”cavities” which are
included in L? Taking a sufficiently small surface S we may assume that L contains only one
such cavity. Let us denote by N the part of the manifold K included in L, and the boundary
→
− →
−
of the cavity by T . Then T is a closed surface, oriented by the exterior normal directed to the
→ −
− →
inside of the cavity. When we now apply Gauß’ Theorem to N and its boundary S + T (where
→
−
S is oriented using an exterior normal) and also Theorem 6.4 we get
Z I I I I
0 = ∇ • F(r) dr = F(r) • dS = F(r) • dS + F(r) • dS = F(r) • dS.
N − −
→ → −
→ →
− −
→
S +T S T T
Thus even through the boundary of each cavity inside K the flux of the vector field must be = 0,
in order for the field to have a vector potential in K.
Ideally this necessary condition for existence of a vector potential should also be sufficient
for solenoidal vector fields. The situation is however complicated. In the literature there are
some sufficient conditions of this classical type—in addition to the case of star-shaped mani-
folds—e.g. the following:
• The vector field F is continuously differentiable and solenoidal. Each manifold L en-
closed by a closed surface inside K ∪ ∂K (manifold + boundary) is included in K (no
cavities) (A POSTOL).
• A magnetic field B in vacuum where there are moving charges (F EYNMAN , R.P. et al.:
The Feynman Lectures on Physics. Volume II. Addison–Wesley (1998)).
• The vector field F is differentiable and its flux through any closed piecewise smooth
surface inside the manifold K vanishes. K can be deformed by continuous transforms
to a star-shaped manifold possibly with finitely many ball-shaped cavities (TON , T.-C.:
On the Potential of a Solenoidal Vector Field. Journal of Mathematical Analysis and
Applications 151 (1990), 557–580).
As a more physical treatment, the classic reference M ISNER , C.W. & W HEELER , J.A.: Clas-
sical Physics as Geometry: Gravitation, Electromagnetism, Unquantified Charge, and Mass as
CHAPTER 6. POTENTIAL 96
Properties of Curved Empty Space. Annals of Physics 2 (1957), 525–603, gives a relativity-
theoretic abstract characterization for existence of a vector potential in some fairly general
cases.
Advances have been made using manifold theory, too. A classic reference is W EYL , H.: The
Method of Orthogonal Projection in Potential Theory. Duke Mathematical Journal 7 (1940),
411–444. More modern ideas are based on the fundamental work of Georges de Rham. See
DE R HAM , G.H.: Differentiable Manifolds. Springer–Verlag (1984). It could be said that in
this way we are already very close to the above mentioned ideal characterization for existence
of vector potential.
(The case g = 0 is of course clear anyway, and can be omitted.) Thus for the needed scalar field
φ we get the equation
∆φ(r) = −g(r).
This is a partial differential equation, a so-called Poisson equation. The solution of the equation
is not unique, in fact any function φ+h, where ∆h = 0, is also a solution. The partial differential
equation
∆h(r) = 0
in turn is called Laplace’s equation, and its solutions are called harmonic functions 3 , see also
Chapter 7 and Appendix 1. So any φ (if it exists) is unique modulo an additive harmonic
function. To fix a φ, some extra conditions—so-called boundary conditions—for the Poisson
equation need to be set, an important topic not dealt with here any further, however.
All in all, since Poisson’s equations have solutions in very general situations, if there is a
vector potential, there also is a solenoidal vector potential.4
A similar idea can be used to get an irrotational vector field U2 with a given divergence f .
Then U2 = ∇ψ and
∇ • U2 (r) = ∆ψ(r) = f (r).
So again we have a Poisson equation whose solution ψ is unique up to an additive harmonic
function. And finally, if we want a vector field U with a given curl ∇ × U = F and a given
divergence ∇ • U = f , then it is simply U = U1 + U2 . To fix U uniquely, extra (boundary)
conditions are needed.
From these we get the celebrated Helmholtz’s decomposition. In a fairly general situation a
vector field U can be expressed as a sum of a solenoidal vector field and an irrotational vector
field. On way to get decomposition would be the following:
3
Harmonic functions have a central rôle in the investigation and modelling of scalar fields.
4
And as a consequence, if F has a vector potential, then it also has a double vector potential V such that
F = ∇ × (∇ × V), and a triple vector potential W such that F = ∇ × ∇ × (∇ × W) , and so on.
CHAPTER 6. POTENTIAL 97
6.5 Four-Potential
A four-potential is the potential of a 4-dimensional 2-form field Φ. By Poincaré’s Lemma it
exists in a star-shaped region if dΦ = 0.
It was noted in Section 5.3 that dΦFaraday = 0. Thus in a star-shaped region the Faraday
form field has a potential A, the so-called electromagnetic four-potential. This potential is a
1-form field, (traditionally) written as
A = A1 dx + A2 dy + A3 dz − φ dt.
Let us compute the exterior derivative and find the coefficients A1 , A2 , A3 and φ. By rule
(IV) in Section 5.2,
∂A ∂A1 ∂A1 ∂A1
1
dA = dx + dy + dz + dt ∧ dx
∂x ∂y ∂z ∂t
∂A ∂A2 ∂A2 ∂A2
2
+ dx + dy + dz + dt ∧ dy
∂x ∂y ∂z ∂t
∂A ∂A3 ∂A3 ∂A3
3
+ dx + dy + dz + dt ∧ dz
∂x ∂y ∂z ∂t
∂φ ∂φ ∂φ ∂φ
− dx + dy + dz + dt ∧ dt,
∂x ∂y ∂z ∂t
and this must be
E1 dx ∧ dt + E2 dy ∧ dt + E3 dz ∧ dt + B1 dy ∧ dz + B2 dz ∧ dx + B3 dx ∧ dy.
CHAPTER 6. POTENTIAL 98
Thus we get A by first finding a vector potential A for the magnetic flux density B (by Maxwell’s
equation (M2) B is solenoidal), and then a scalar potential φ for the vector field
∂A
− − E.
∂t
This is possible since by Maxwell’s equations (and some weak extra conditions)
∂A ∂ ∂B ∂B
∇× − − E = − (∇ × A) − ∇ × E = − + = 0.
∂t ∂t ∂t ∂t
The four-potential A is not unique, we can add a gradient field ∇ψ in A, and replace φ by
φ − ∂ψ/∂t, without the exterior derivative changing. Often A and φ are chosen to satisfy the
so-called Lorenz gauge condition
1 ∂φ
∇•A+ = 0.
c2 ∂t
In a sense this separates A and φ. If originally
1 ∂φ
∇•A+ = g 6= 0,
c2 ∂t
then Lorenz’s gauge condition will be satisfied if ψ is a solution of the partial differential equa-
tion
1 ∂ 2 ψ(r, t)
∆ψ(r, t) − 2 = −g(r, t)
c ∂t2
(check!). This equation is a so-called wave equation and it has a solution in some very general
situations. Lorenz’s gauge condition can then be assumed quite generally.
On the other hand, it was noticed in Section 5.3 that the exterior derivative of the function (or
0-form field)
q 1
f (r′) =
4πε kr − r′ k
is the work form field of the gradient of f , i.e.
where the derivation is with respect to r′ (the primed nabla). The approximation follows imme-
diately from the definition of the exterior derivative. Thus in the point of observation r
q ′ 1 q r − r′
φ(r) ∼
= ∇ • h = • h.
4πε kr − r′ k 4πε kr − r′ k3
Here
qh = p′
is called the electric dipole moment. With this notation we get the usual dipole approximation
of the potential as
1 p′ • (r − r′ )
φ(r) ∼
= ,
4πε kr − r′ k3
where the connection between r′ and p′ is indicated by the primes.
But what about the electric field in r? In electrophysics negative potentials are used, so
E(r) = −∇φ(r), and
1 p′ • (r − r′ )
E(r) ∼=− ∇ .
4πε kr − r′ k3
The gradient is is taken with respect to the variable r. The vector p′ does not depend on r.
Applying the derivation rule (i) in Section 1.5 we get
p′ • (r − r′ ) ′ ′
1 1 ′ ′
∇ = p • (r − r ) ∇ + ∇ p • (r − r )
kr − r′ k3 kr − r′ k3 kr − r′ k3
p′ • (r − r′ ) ′ p′
= −3 (r − r ) + .
kr − r′ k5 kr − r′ k3
Except for the dipoles themselves, such an approximation can be useful for other fields closely
resembling dipoles. The dipole moment is then obtained by physical considerations.
A magnetic dipole is very similar to the electric dipole above. We approach it using a current
loop remembering the definition of curl via exterior derivation in Section 5.3.
→
−
So, let us consider a small parallelogram-shaped current loop ∂P with current I. A normal
vector for the plane of the parallelogram is given by r1 × r2 = An, where A is the area of the
parallelogram and n is the corresponding unit normal vector, correctly oriented with respect to
the direction of the current. The point of observation r is far away compared with the size of
the loop. We use the vertex r0 to characterize the position of the loop, see the figure below.
Points of the loop are denoted by the variable r′ . The point of observation r being ”far away”
then means that kr − r′ k ≫ kr′ − r0 k.
CHAPTER 6. POTENTIAL 100
I a × (r′ − r)
4π kr − r′ k3
around the loop and—as noted in Section 5.3—it is approximately the flux form field of the curl
in the point r0 , i.e.
I a × (r′ − r) AI a × (r′ − r)
∼ ′
a • H(r) = ∇ × • r1 × r2 = ′
n• ∇ × ,
4π kr − r′ k3 r′ =r0 4π kr − r′ k3 r′ =r0
where ∇′ is with respect to r′ . Using the derivation rule (vi) in Section 1.55 and remembering
that the Newton vector field is solenoidal, the curl will be seen to be
a a • (r − r′ )
− ′ 3
+ 3 ′ 5
(r − r′ )
kr − r k kr − r k
(check!). Thus
∼ IA a•n (r − r0 ) • n
a • H(r) = − +3 (a • (r − r0 )) .
4π kr − r0 k3 kr − r0 k5
IA n (r − r0 ) • n
H(r) ∼
= − + 3 (r − r 0 ) .
4π kr − r0 k3 kr − r0 k5
m′ = IAn
of the loop, we get the dipole approximation of the magnetic field in a far away point r as
1
H(r) ∼
= 5
− kr − r0 k2 m′ + 3 (r − r0 ) • m′ (r − r0 ) .
4πkr − r0 k
This expression is of exactly the same form as it was for the electric dipole field! So, reverse
argumentation as above shows that the magnetic field has here an approximative scalar potential
φ, i.e.
H(r) ∼= −∇φ(r),
5
Or using Theorem 6.1.
CHAPTER 6. POTENTIAL 101
where
1 m′ • (r − r0 )
φ(r) = .
4π kr − r0 k3
Using the derivation rule (iii) in Section 1.5 we may check that the dipole approximation
field is solenoidal. Thus it also has a vector potential (in appropriate regions). We could find
it more or less as we found a vector potential for the Newton vector field in the example in
Section 6.3. Looking at the curl expression above it should become clear, however, that the
vector potential is
1 m′ × (r − r0 )
A(r) = .
4π kr − r0 k3
Note. Dipole approximation is just a part of the more general multipole expansion of a vector
field. Its first term is the so-called monopole approximation, a Newton vector field. The second
term is the dipole approximation of a combination of two Newton vector fields, as above. The
third term is the quadrupole approximation of the combination of four Newton vector fields,
with pairwise opposite signs. The fourth term is the octupole approximation, and so on.
”The miracle of the appropriateness of the
language of mathematics for the formulation
of the laws of physics is a wonderful gift which
we neither understand nor deserve.”
Chapter 7
• f > 0 and g > 0: telegraph equation, lossy wave equation, hyperbolic heat equation.
The order of a PDE is the order of the highest partial derivative appearing in the equation.
The above PDEs are all of second order. There are, of course, many important kinds of PDEs
of first order, e.g. advection equations, and also of order higher than two, e.g. the biharmonic
equation (elasticity), Korteweg–de Vries equation (shallow water waves), and Dym’s equation
(solitons). Theories and numerical solution methods for these other orders are very different
from the ones for the above PDEs.
7.2 Examples
Let us take examples of field modelling problems leading to second order PDEs of the above
types. For electric and magnetic fields (M1)–(M4) refer to the Maxwell equations in Section
5.3.
102
CHAPTER 7. PARTIAL DIFFERENTIAL EQUATIONS 103
Example. (Electrostatic field) Since in the stationary case the electric field E is irrotational
(M3), we have ∇ × E = 0 and the field has a scalar potential: E = −∇V . (In electrophysics
the negative potential Φ = −V is used.)
On the other hand (M1),
Here ε is the permittivity. If the charge density—and of course the permittivity—is known,
∇ • ε(r)∇V (r) = −ρ(r).
Example. (Stationary electric current) By Ohm’s law, the current density is J = σE, where
σ is the conductance. According to the Kirchhoff law electricity is not accumulated anywhere,
→
−
i.e., the net current through a closed surface is zero. For the boundary (a closed surface) S of
the solid K, I Z
J(r) • dS = 0 = ∇ • J(r) dr
−
→ K
S
(by Gauß’ Theorem). In these problems the field usually is assumed continuously differentiable,
thus since K is arbitrary we have further ∇ • J = 0.
The field E being irrotational (M3) it has a scalar potential. i.e. E = −∇V , and so
∇ • J(r) = ∇ • σ(r)∇V (r) = 0,
Example. (Magnetic field with a scalar potential) In a region with no conductive material
charge density is zero, and (M4) ∇ × H = J = 0. There is thus a scalar potential, i.e.
H = ∇Φ.
On the other hand (M2)
∇ • B = ∇ • (µH) = 0,
where µ is the permeability. We have again a PDE of the same type since now
∇ • µ(r)∇Φ(r) = 0.
Example. (Stationary incompressible irrotational fluid flow) Here the flow is known to be
irrotational by nature, that is ∇ × v = 0. The velocity thus has a scalar potential: v = ∇φ. In
an incompressible flow fluid is not accumulated in any body K bounded by the closed surface
→
−
S . What comes in also goes out, and so fluid leaves the body at zero net rate;
I Z
v(r) • dS = ∇ • v(r) dr = 0.
−
→ K
S
Since the body K is arbitrary, then as above ∇ • v = 0, and we have the Laplace equation
∇ • ∇φ(r) = ∆φ(r) = 0.
CHAPTER 7. PARTIAL DIFFERENTIAL EQUATIONS 104
Example. (Stationary heat conduction) The heat flow is v = −k∇T , where T is the temper-
ature and k is the thermal conductivity. Here the model is an empirical one, valid for isotropic
material (one where local heat flow is the same in all directions).
If the temperature is stationary, heat is not accumulated anywhere, and as above with the
fluid flow or with the Kirchhoff law, ∇ • v = 0. Thus we have
∇ • k(r)∇T (r) = 0.
Example. (Nonstationary heat conduction) As above, the heat flow is v = −k∇T . The net
→
−
heat flow out of a body K with boundary surface S is
I
v(r, t) • dS,
−
→
S
(again by Gauß’ Theorem). Energy conservation forces this to be the same as the power needed
to raise the temperature of the body.
∂T
The power per volume unit needed to raise the temperature at the rate is
∂t
∂T (r, t)
C(r)ρ(r) ,
∂t
where C(r) is the thermal capacity and ρ(r) the mass density in the point r. For the whole body
the power is thus Z
dE ∂T (r, t)
= C(r)ρ(r) dr.
dt ∂t
K
Comparing the powers (and again remembering that K was arbitrary) we deduce that the inte-
grands must be the same, and we get the heat equation
∂T (r, t)
∇ • k(r)∇T (r, t) = C(r)ρ(r) .
∂t
It is interesting how mathematical modelling in different areas leads to the precise same
types of PDEs. Some PDEs contain second time derivatives originating from Newton’s second
law. This would be the case e.g. for nonstationary fluid flow. Our example, however, is an
acoustic wave equation.
Example. (Small amplitude acoustic plane wave) A plane wave is a planar wave front. We
set the coordinates in such a way that the wave front ”proceeds” to the direction of the positive
x-axis, and is thus parallel to the yz-plane. We then need only the x-coordinate and time t,
within the front there is no change. This would correspond to a sound source far away in the
direction of the negative x-axis.
Let us denote the position of an air molecule at time t, initially in the point x, by x + u(x, t).
Thus u(x, t) is the deviation from the initial position, and u(x, 0) = 0. Initially the left edge of
a layer of thickness dx is in the point x, and at time t it is in the point x + u(x, t). The thickness
of the layer is then
∂u(x, t)
dx + dx,
∂x
see the figure below.
CHAPTER 7. PARTIAL DIFFERENTIAL EQUATIONS 105
The mass of the layer per area unit remains the same, y dx dx + ux dx
i.e. ∂u(x, t)
ρ0 dx = ρ(x, t) 1 + dx,
∂x
where ρ0 is air density at time t = 0 and ρ(x, t) at
time t. We see thus that
ρ0 ρ
ρ(x, t) ∂u(x, t) −1
= 1+ .
ρ0 ∂x
In an adiabatic process, where there is no heat loss
nor heat gain, is is known that
u(x,t)
p(x, t) ρ(x, t) γ x
= , z x
p0 ρ0
where p0 is the air pressure at time t = 0 and p(x, t) at time t, and γ is the so-called adiabatic
constant, for air γ = 1.40. Differentiating we get
∂p(x, t) ∂ ∂u(x, t) −γ ∂u(x, t) −γ−1 ∂ 2 u(x, t)
= p0 1+ = −γp0 1 +
∂x ∂x ∂x ∂x ∂x2
γp0 ∂u(x, t) −γ ∂ 2 u(x, t)
=− ρ(x, t) 1 + .
ρ0 ∂x ∂x2
For small amplitudes
γp0
= c2
ρ0
is constant, speed of sound squared. This follows from the ideal gas law, since for small ampli-
tudes temperature is constant and density is inversely proportional to volume. For air, at 1 at
and at sea level, this gives c = 340.3 ms−1.
A layer is moved by force dictated by the pressure difference (per area unit) between the
edges, and this must be the same as the force given by Newton’s second law:
∂p(x, t) ∂u(x, t) −γ ∂ 2 u(x, t)
dx = −c2 ρ(x, t) 1 + dx
∂x ∂x ∂x2
∂u(x, t) ∂ 2 u(x, t)
= −ρ(x, t) 1 + dx.
∂x ∂t2
Note the sign, molecules are moving against the pressure difference. Thus finally we get the
PDE γ+1 2
∂ 2 u(x, t) 1 ∂u(x, t) ∂ u(x, t)
= 1 + ,
∂x2 c2 ∂x ∂t2
and it is of the type indicated. The coefficient function g depends here also on ∂u/∂x.
For small amplitudes and acoustic frequencies ∂u(x, t)/∂x is also small, so at least approx-
imatively
∂ 2 u(x, t) 1 ∂ 2 u(x, t)
= 2 .
∂x2 c ∂t2
This last PDE is a so-called wave equation. Its general solution is of the form
∂u(r, t) ∂u(r, t)
= = 0.
∂y ∂z
Thus
∂ 2 u(r, t) ∂ 2 u(r, t) ∂ 2 u(r, t)
= = 0 and = ∆u(r, t),
∂y 2 ∂z 2 ∂x2
and the wave equation can be written as
1 ∂ 2 u(r, t)
∆u(r, t) = .
c2 ∂t2
The Laplacian is coordinate-free, so this equation holds true for any direction of the planar
wave front.
A general wave equation in R3 has solutions other than the plane waves, e.g. the so-called
(r0 -centered) spherical waves
1 1
u(r, t) = f kr − r0 k − ct + g kr − r0 k + ct ,
kr − r0 k kr − r0 k
where f and g are arbitrary twice continuously differentiable functions (check, and see also
Section A2.4). These may be interpreted as expanding/contracting spherical wave fronts with a
point source in r0 . Strangely such ”spreading sharp signals” are only possible in odd dimen-
sions1—though in another form in higher dimensions—and thus in particular not in R2 .
Often the units are chosen so that c = 1, and the notation
∂2u
u = ∆u − 2 (so-called d’Alembertian)
∂t
is adopted.2 The wave equation is then simply
u = 0 or u = F (r, t),
It is again interesting that the same wave equation is obtained for small amplitude vibra-
tions in general, for vibrating strings and membranes, for longitudinal, transverse and torsional
vibrations of rods, for electromagnetic waves (as a consequence of Maxwell’s equations), for
oscillating electric circuits and mechanical systems, for vibrating columns of air and all kinds
of acoustic pressure waves, etc.
1
As a consequence of the Hyugens Principle.
2
Sometimes with the signs reversed or denoted by 2 .
hi i need some help here
it’s due next monday
how to use green’s first identity to show if a function
is harmonic on D(a surface), then the line intergral of
”normal derivative of” is equal to zero.
thanks!
can be generalized4 using the Generalized Stokes Theorem and Cartan’s magic formula as
Z Z Z
k+1
dΦ ∧ Ψ = Φ ∧ Ψ + (−1) Φ ∧ dΨ,
A −
→s A
∂−→A
M
4
There is a generalization for indefinite integrals (potentials) as well but it does not seem to be that useful.
107
APPENDIX 1. PARTIAL INTEGRATION AND GREEN’S IDENTITIES 108
Z Z Z
1. Φg∇f –work = fg − Φf ∇g–work
A −s
→ A
∂−→A
M
→
−
I.e., for an oriented curve C with end vertices r1 (initial) and r2 (terminal),
Z Z
g(r)∇f (r) • ds = f (r2 )g(r2) − f (r1 )g(r1) − f (r)∇g(r) • ds.
−
→ −
→
C C
Z Z Z
2. Φ∇f ×G–flux = Φf G–work − Φf ∇×G–flux
A −s
→ A
∂−→A
M
→
−
I.e., for an oriented surface S ,
Z I Z
∇f (r) × G(r) • dS = f (r)G(r) • ds − f (r)∇ × G(r) • dS.
−
→ −
→ −
→
S ∂S S
Z Z Z
3. ΦG•∇f –density = Φf G–flux − Φf ∇•G–density
A −s
→ A
∂−→A
M
→
−
I.e., for an oriented solid body K ,
Z I Z
G(r) • ∇f (r) dr = f (r)G(r) • dS − f (r)∇ • G(r) dr.
−
→ −
→ −
→
K ∂K K
Z Z Z
4. ΦG•∇×F–density = ΦF×G–flux + ΦF•∇×G–density
A −s
→ A
∂−→A
M
→
−
I.e., for an oriented solid body K ,
Z I Z
G(r) • ∇ × F(r) dr = F(r) × G(r) • dS + F(r) • ∇ × G(r) dr.
−
→ −
→ →
−
K ∂K K
where
∂u(r)
= ∇u(r) • n
∂n
is the normal derivative of the function u, i.e., its directional derivative in the direction of the
exterior unit normal n.
Partial integration is significant in two ways when PDEs are solved using FEM. First, known
boundary conditions render the right hand side surface integral known. Hence boundary condi-
tions are explicitly included in the solution. Second, the integral
Z
k(r)∇u(r) • ∇v(r) dr
K
only requires continuity of the function k and the partial derivatives of the functions u and v, if
not even that!5 Using partial integration makes it possible to have much less stringent continuity
requirements.
and subtracting these equalities on both sides we get Green’s second identity
Z Z
∂u(r) ∂v(r)
(v(r)∆u(r) − u(r)∆v(r)) dr = v(r) − u(r) dS.
∂n ∂n
K S
To get the third identity we need to first derive a property of harmonic functions. Consider
a scalar field u which is twice continuously differentiable in K and harmonic, i.e. ∆u = 0 (cf.
Section 6.4). Take a point r0 in the interior of the solid body K, and take v to be the Newton
potential
1
v(r) = .
kr − r0 k
Since v has a singularity (the point r0 ) in K, we specify another solid body K1 by removing
from K a small r0 -centered ball K2 of radius δ. The corresponding sphere S2 is oriented using
its exterior normal, that is the normal pointing into the body K1 .
5
In FEM the functions u and v would be so-called element functions, and the approximate solution is formulated
as a linear combination of the element functions ui .
APPENDIX 1. PARTIAL INTEGRATION AND GREEN’S IDENTITIES 110
The body K1 has two boundary surfaces, the inner one and the outer one. In the body
K1 both functions u and v are harmonic (it will be remembered that Newton’s vector field is
solenoidal). We now apply Green’s second identity to the body K1 :
Z Z
∂u(r) ∂v(r) ∂u(r) ∂v(r)
0= v(r) − u(r) dS − v(r) − u(r) dS.
∂n ∂n ∂n ∂n
S S2
Appendix 2
constitutes a so-called curvilinear coordinate system. To get the axes (curves) in the point γ(u1 )
just fix values of the parameters to u = u1 except for one.
The tangent space Tr (M) of M in the point r = γ(u) is the whole Rn . On the other hand it
is the column space of the derivative matrix γ ′ (u), see Section 2.6. Thus the columns of γ ′ (u)
give a basis for Rn which depends on the point r, and will change (in general) with the point.
This basis gives a so-called local coordinate system.
In what follows we only deal with the case where the local coordinate system is orthogonal,
i.e., the columns of γ ′ (u) are mutually orthogonal and positively oriented. Let us denote by
Q(u) the matrix which we get by normalizing the columns of γ ′ (u), i.e. dividing each column
by its length. Then Q(u) is an orthogonal matrix, i.e. Q(u) = Q(u) , and (for the positive
−1 T
Q(u) = γ ′ (u)Λ(u)−1 .
111
APPENDIX 2. PULLBACKS AND CURVILINEAR COORDINATES 112
and
cos φ − sin φ 0
Q(r, φ, z) = sin φ cos φ 0 .
0 0 1
Apparently Q(r, φ, z) is orthogonal and det Q(r, φ, z) = 1. The local coordinate vectors are
then
cos φ − sin φ 0
er = sin φ , eφ = cos φ and ez = 0 .
0 0 1
and
sin θ cos φ cos θ cos φ − sin φ
Q(ρ, θ, φ) = sin θ sin φ cos θ sin φ cos φ .
cos θ − sin θ 0
in the tangent space in the point). Columns of the matrix Q(u) are the new coordinate vectors
in the old coordinates. The required representation is thus (cf. Section 1.3)
G(u) = Q(u)T F γ(u) .
The gradient and the Laplacian of a scalar field f in curvilinear local coordinates are thus
Q(u)T ∇f γ(u) and ∆f γ(u) ,
A2.2 Pullbacks
The correct tool for dealing with curvilinear coordinates is the so-called pullback. If Φ is a
k-form field of Rn and δ : Rm → Rn is a continuously differentiable function, then the k-form
field
(δ ∗ Φ)(u; r1 , . . . , rk ) = Φ δ(u); δ ′ (u)r1 , . . . , δ ′ (u)rk
of Rm is the so-called pullback form field of Φ with respect to the function δ. The notation δ ∗ Φ
is conventional. Note in particular that the δ ′ (u)ri ’s are n-dimensional vectors, as they should
be. Pullbacks make it easy to define new form fields, more or less as Theorem 2.5 was used
to define manifolds. However, to make the pullback of a continuously differentiable form field
continuously differentiable, the function δ should be twice continuously differentiable.
Pullbacks are clearly left-distributive over addition, i.e.
δ ∗ (Φ + Ψ) = δ ∗ Φ + δ ∗ Ψ,
Proof. By Theorem 4.2, a form field Φ can be expanded as a combination of elementary form
fields as
X
Φ(p; r1 , . . . , rk ) = aj1 ,j2 ,...,jk (p)(dxj1 ∧ dxj2 ∧ · · · ∧ dxjk )(r1 , . . . , rk ).
1≤j1 <j2 <···<jk ≤n
It is seen then that it suffices to prove the properties for form fields of the form
Left distributivity over the wedge product of such form fields follows fairly directly from the
determinant definition of elementary forms (we leave the details to the reader but cf. also the
proof of Theorem 4.4).
APPENDIX 2. PULLBACKS AND CURVILINEAR COORDINATES 114
By the induction hypothesis, d(δ ∗ Φ) = δ ∗ (dΦ). Again applying the Magic Formula we get
d δ ∗ (Φ ∧ dxm+1 ) = d(δ ∗ Φ ∧ δ ∗ dxm+1 ) = d(δ ∗ Φ) ∧ δ ∗ dxm+1 + (−1)m δ ∗ Φ ∧ d(δ ∗ dxm+1 )
= δ ∗ (dΦ) ∧ δ ∗ dxm+1 + (−1)m δ ∗ Φ ∧ dd(δ ∗ xm+1 )
= δ ∗ (dΦ) ∧ δ ∗ dxm+1 + 0 = δ ∗ (dΦ ∧ dxm+1 ) = δ ∗ d(Φ ∧ dxm+1 ) .
As an example of the many uses of pullbacks we take a reduction of the Generalized Stokes
Theorem.
Example. For the notation we refer to the Generalized Stokes Theorem in Section 5.4. Assume
that the manifold M has the (relaxed) parametrization p = γ(u) (u ∈ U), and further that the
bounded region with boundary A has the (relaxed) parametrization p = γ(u) (u ∈ V).
Let us first consider orientations. The manifold M is oriented by a k-form field Ψ via
the sign of Ψ(p; t1 , . . . , tk ) where t1 , . . . , tk are in the tangent space Tp (M), as explained in
Section 4.3. Writing ti = γ ′ (u)vi (i = 1, . . . , k) we get the corresponding orientation for U
(and V) by the pullback k-form field
Ψ γ(u); γ ′ (u)v1 , . . . , γ ′ (u)vk = (γ ∗ Ψ)(u; v1 , . . . , vk ).
s
The smooth boundary ∂M A is defined locally, near the point p0 = γ(u0 ), by a continuously
differentiable function gp0 via the sign of gp0 (p): + (p inside A), − (p outside A), 0 (p in
the boundary), see Section 5.1. For the orientation of the smooth boundary, an exterior vector
texterior (p), satisfying gp′ 0 (p)texterior (p) < 0, is chosen as t1 , and the k − 1-form field
Ψ p; texterior (p), t2 , . . . , tk .
is used. Returning to the parameter domain, for orienting ∂Us V the function gp0 (p) is replaced
′
by hu0 (u) = gp0 γ(u) . Writing texterior (p) = γ (u)vexterior (u) we see first that
h′u0 (u)vexterior (u) = gp′ 0 γ(u) γ ′ (u)vexterior (u) = gp′ (p0 )texterior (p) < 0,
i.e. vexterior (u) will indeed be an exterior vector, and second that ∂Us V is oriented by the pullback
k − 1-form field
Ψ γ(u); γ ′ (u)vexterior (u), γ ′ (u)v2 , . . . , γ ′ (u)vk = (γ ∗ Ψ) u; vexterior (u), v2 , . . . , vk .
s
Thus all four orientations, that of M, of ∂M A, of U, and of ∂Us V, agree.
APPENDIX 2. PULLBACKS AND CURVILINEAR COORDINATES 115
Second we consider the integrals in the theorem. For any k-form field Ξ defined on M we
have Z Z Z Z
′ ∗
Ξ = Ξ γ(u); γ (u) du = (γ Ξ)(u; Ik ) du = γ ∗ Ξ
−
→ V V −
→
A V
(we used here the identity parametrization for U for which the derivative is the k × k identity
matrix Ik ). Assuming γ is twice continuously differentiable, applying this to dΦ we thus get, by
Theorem A2.1, Z Z Z
dΦ = γ (dΦ) = d(γ ∗ Φ).
∗
−
→ −
→ −
→
A V V
Let then the smooth boundary ∂Us V have the (relaxed) parametrization
u = ε(s) (s ∈ S),
whence ∂M A has the (relaxed) parametrization p = γ ε(s) (s ∈ S). So
s
Z Z Z Z
′ ′ ∗ ′
Φ = Φ γ ε(s) ; γ ε(s) ε (s) ds = (γ Φ) ε(s); ε (s) ds = γ ∗ Φ.
−
→ S S −
→
∂A ∂V
This means we can prove the Generalized Stokes Theorem by first taking the pullback to the
parameter domain and then proving the theorem there which is very much easier (the theorem
becomes essentially the Generalized Divergence Theorem). This however works only with the
annoying assumption of γ being twice continuously differentiable.
and
Normalizing γ ′ (u) we get the desired curl given in the local coordinates:
Λ(u)
Q(u)T ∇ × F γ(u) = ′
∇u × Λ(u)G(u) .
det γ (u)
and
and thus
1
∇ • F γ(u) = ∇u • det γ ′ (u) γ ′ (u)−1 F γ(u) .
det γ ′ (u)
1
∇ • F γ(u) = ∇u • det γ ′ (u) Λ(u)−1 G(u) .
det γ ′ (u)
APPENDIX 2. PULLBACKS AND CURVILINEAR COORDINATES 117
Combining the representations of the gradient and the divergence we also get the Laplacian
in the local coordinates:
1
∆f γ(u) = ∇u • det γ ′ (u) Λ(u)−2 ∇u g(u) .
′
det γ (u)
Note that this expression contains both second order and first order partial derivatives.
∂g 1 ∂g ∂g
∇f = er + eφ + ez
∂r r ∂φ ∂z
1 ∂Gr 1 ∂Gφ ∂Gz
∇•F= Gr + + +
r ∂r r ∂φ ∂z
1 ∂G ∂Gφ ∂Gr ∂Gz 1 ∂Gφ ∂Gr
z
∇×F= − er + − eφ + Gφ + r − ez
r ∂φ ∂z ∂z ∂r r ∂r ∂φ
1 ∂g ∂ 2 g 1 ∂2g ∂2g
∆f = + 2+ 2 2+ 2
r ∂r ∂r r ∂φ ∂z
These representations of the derivative operations are very handy for axially symmetric fields.
The spherical coordinates in turn are very handy for radially symmetric fields. The local
basis vectors are eρ , eθ and eφ . Writing
G = Gρ eρ + Gθ eθ + Gφ eφ
we have
∂g 1 ∂g 1 ∂g
∇f = eρ + eθ + eφ
∂ρ ρ ∂θ ρ sin θ ∂φ
2 ∂Gρ 1 1 ∂Gθ 1 ∂Gφ
∇•F= Gρ + + Gθ + +
ρ ∂ρ ρ tan θ ρ ∂θ ρ sin θ ∂φ
1 ∂Gφ ∂Gθ 1 1 ∂Gρ ∂Gφ
∇×F= cos θ Gφ + sin θ − eρ + − Gφ − ρ eθ
ρ sin θ ∂θ ∂φ ρ sin θ ∂φ ∂ρ
1 ∂Gθ ∂Gρ
+ Gθ + ρ − eφ
ρ ∂ρ ∂θ
2 ∂g ∂ 2 g 1 ∂g 1 ∂2g 1 ∂2g
∆f = + 2+ 2 + 2 2+ 2 2
ρ ∂ρ ∂ρ ρ tan θ ∂θ ρ ∂θ ρ sin θ ∂φ2
”For Angling may be said to be so like the Mathematics that it can
never be fully learned...”
Appendix 3
ANGLE
The angle is thus determined by the boundary of the manifold, and can be computed by integra-
tion over it.
The case n = 1 is included here, although it is very simple and subject to interpretation.
Then
p
Φ1−angle (p) = = signum(p),
|p|
−→
and an oriented 0-dimensional oriented manifold is a finite set M = {p1 , . . . , pm } of oriented
points. The orientation of a point pi is denoted by ω(pi ). When chosen as the sign of pi it
−→
contributes positively to the angle, negatively otherwise. Then the (net) angle spanned by M is
Z Xm
Φ1−angle = ω(pi )signum(pi ).
−
→ i=1
M
Note that if the orientation of each point pi is its sign, then the angle is = m.
118
APPENDIX 3. ANGLE 119
(a relaxed parametrization). We assume in addition that the curve does not contain the origin.
Then Z Z Z
′
′ d
Φ2−angle = atan γ(u) γ (u) du = atan γ(u) du.
du
−
→ U U
C
y
This is the net angle that a point moving along the curve C
scans seen from the origin, counted as positive in the posi-
tive rotation and negative in the negative rotation. Note that
discontinuity of atan does not matter because when mov-
ing over the positive x-axis the value of atan changes by
2π, but its derivative is continuous. The star-shaped region
indicated above could be replaced by any region we get by
removing from R2 a ray starting from the origin, and mak- x
ing the corresponding changes in the definition of atan.
The angle potential in R2 is a scalar potential, essentially atan or any function obtained by
adding a constant. The boundary is formed of separate
points, and the net angle spanned is always in principle ob- 1
tained by addition/subtraction.
x
1 2 3 4
curve
–1
→
− y
C : r = γ(u) = u cos(cos u), u sin(cos u) (1 ≤ u ≤ 5),
–2
as it should.
Note. For those familiar with complex analysis this brings to mind the complex logarithm and
its derivative.
pT
F(p) = ,
kpk3
and Φ3−angle is the flux form field of F. The angle here is the so-called solid angle.
Geometrically the solid angle spanned
→
− z
by the surface S seen from the origin is
the area of the part of the origin-centered
→
−
unit sphere P that exactly covers S (see the
figure on the right). The general situation S
is however more complex. We must allow
the possibility that some part of the sphere
is seen ”twice” or even more times. We P
also count as positive those points p of the
surface where the normal n points away y
from the origin, i.e. p • n > 0, and as
negative points where the normal points
x
towards the origin, i.e. p • n < 0. The whole
space obviously spans the solid angle 4π (the
area of a unit sphere). The unit of solid angle is the steradian (sr).
Mathematically the solid angle seen from the origin spanned by the oriented 2-dimensional
manifold
→
−
S : r = γ(u) (u ∈ U)
(an oriented surface in a relaxed parametrization) of R3 is
Z
→
− ∂δ(u) ∂δ(u)
Ω( S ) = δ(u) • × du,
∂u1 ∂u2
U
where
γ(u)
δ(u) =
γ(u)
.
Here we again assume that S does not contain the origin. Note that the values of δ(u) are in the
unit sphere x2 + y 2 + z 2 = 1.
On the other hand,
→
−
P : r = δ(u) (u ∈ U)
APPENDIX 3. ANGLE 121
is not (necessarily) a parametrized manifold, even in a relaxed sense. A part of the unit sphere
may appear several times, possibly with opposite orientations, depending on how many times a
→
−
ray starting from the origin intersects the surface S and in which direction. Generally however
→
−
P can be considered as an oriented trajectory manifold in a relaxed parametrization.
The unit normal of the origin-centered unit sphere in the point δ(u) is either δ(u)T (the
exterior normal) or −δ(u)T (the interior normal). Thus
∂δ(u) ∂δ(u)
∂δ(u) ∂δ(u)
× = ±
∂u1 × ∂u2
δ(u),
∂u1 ∂u2
follows as special case of the general result in the next section (but can be verified quite easily
separately, too).
Example. Using the result in the example in Section 6.3 we get a vector potential U for the
vector field
p
F(p) =
kpk3
in the star-shaped region that we get by removing from the space one of the rays −up0 (u ≥ 0)
where p0 6= 0, as
p0 × p
U(p) = 2
kp0 kkpk + (p0 • p)kpk
(we chose p1 = 0).6 The corresponding angle potential is the 1-form field
p0 × p • r
Ψ3−angle (p; r) = U(p) • r = .
kp0 kkpk2 + (p0 • p)kpk
→
−
The solid angle spanned by the oriented surface S is thus obtained by Stokes’ Theorem by
→ −
− →
integrating over the correspondingly oriented boundary curve C of S :
I I
→
− p0 × p
Ω( S ) = U(p) • ds = 2
• ds.
kp0 kkpk + (p0 • p)kpk
−
→ −
→
C C
We can choose the ray (the Dirac string) removed from the space to be one pointing to the
opposite of the direction of sight, i.e. p0 is in the direction of sight.
As a concrete example, let us compute the solid angle spanned by the ellipse
−
→
C : p = γ(u) = (cos u, 2 sin u, 2) (0 ≤ u ≤ 2π)
6
For simplicity we again use the same notation for points and vectors.
APPENDIX 3. ANGLE 122
–2
and the solid angle is –1.5
–1
–0.5 –1
–0.5
Z2π 1
0.5 0.5
2 du x y 1
Ω= √ 1.5
cos2 u + 4 sin u + 4 + 2 cos2 u + 4 sin2 u + 4
2 2
0
Z2π
2 du ∼
= √ = 1.1 sr.
3 sin u + 5 + 2 3 sin2 u + 5
2
0
(The integral here is a so-called elliptic integral and it cannot be given by elementary functions.
Numerical calculation is of course possible and easy.)
The part of the space not spanned by the ellipse then determines the solid angle
4π − Ω ∼
= 11.5 sr.
Note how the ray removed from the space—here the negative z-axis—deftly chooses which one
of these two conjugate solid angles, Ω or 4π − Ω, is computed!
A3.4 Angles in Rn
Mathematically the angle ”seen” from the origin spanned by the oriented n − 1-dimensional
manifold
−
→
M : r = γ(u) (u ∈ U)
(an oriented hypersurface in a relaxed parametrization) of Rn is
Z q
−
→
Ω(M) = ± det δ ′ (u)T δ ′ (u) du,
U
where
γ(u)
δ(u) =
kγ(u)k
and the sign is chosen by the particular local orientation. Here we again assume that S does not
−→
contain the origin. Thus Ω(M) is the volume of the part P of a unit hypersphere kpk = 1 of
Rn exactly covering M as ”seen” from the origin. Here, too,
−
→
P : r = δ(u) (u ∈ U)
is not (necessarily) a parametrized manifold, even in a relaxed sense, but it may be considered
as an oriented trajectory manifold in a relaxed parametrization.
Writing first
γ(u)
δ(u) = 1/2
γ(u) • γ(u)
APPENDIX 3. ANGLE 123
1. A BRAHAM , R. & M ARSDEN , J.E. & R ATIU , T.: Manifolds, Tensor Analysis, and Appli-
cations. Springer–Verlag (1993)
2. A KCOGLU , M.A. & BARTHA , P.F.A. & H A , D.M.: Analysis in Vector Spaces. A Course
in Advanced Calculus. Wiley (2009)
3. A POSTOL , T.M.: Calculus. Vol. II: Multi-Variable Calculus and Linear Algebra with
Applications to Differential Equations. Wiley (1975)
5. C HERN , S.-S. & C HEN , W.H. & L AM , K.S.: Lectures on Differential Geometry. World
Scientific (1999)
6. H UBBARD , J.H. & H UBBARD , B.B.: Vector Calculus, Linear Algebra, and Differential
Forms. Matrix Editions (2009)
8. L OOMIS , L.H. & S TERNBERG , S.: Advanced Calculus. Jones and Bartlett (1990)
10. M ARSDEN , J.E. & T ROMBA , A.J.: Vector Calculus. W.H. Freeman (2003)
11. M ISNER , C.W. & T HORNE , K.S. & W HEELER , J.A.: Gravitation. W.H. Freeman (1973)
12. N IKOLSKY, S.M. & VOLOSOV, V.M.: A Course of Mathematical Analysis. MIR Publish-
ers (1987)
13. O’N EILL B.: Elementary Differential Geometry. Academic Press (2006)
14. O STROWSKI , A.: Vorlesungen über Differential- und Integralrechnung. Birkhäuser (1972)
16. S CHEY, H.M.: Div, Grad, Curl, and All That. An Informal Text on Vector Calculus. W.W.
Norton (1997)
17. S PIVAK , M.: Calculus on Manifolds. A Modern Approach to Classical Theorems of Ad-
vanced Calculus. W.A. Benjamin (1971)
124
References 125
Index
active variable . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16 direction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
addition of vectors . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2 directional derivative . . . . . . . . . . . . . . . . . . . . . . . . 10
advection equation . . . . . . . . . . . . . . . . . . . . . . . . . 102 distance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1,4
affine approximation . . . . . . . . . . . . . . . . . . . . . . . . 30 Divergence Theorem . . . . . . . . . . . . . . . . . . . . . . . . 78
affine subspace . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39 divergence . . . . . . . . . . . . . . . . . . . . . . . . . . . 10,74,116
Ampère’s law . . . . . . . . . . . . . . . . . . . . . . . . . . . . 75,81 division by scalar . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
angle form field . . . . . . . . . . . . . . . . . . . . . . . . . . . . 118 dot product . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4,6
angle potential . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 118 double-curl expansion . . . . . . . . . . . . . . . . . . . . 11,97
angle . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3,118 Dym’s equation . . . . . . . . . . . . . . . . . . . . . . . . . . . . 102
atan . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25,92,119 dynamical . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
anticommutativity . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4 electric dipole moment . . . . . . . . . . . . . . . . . . . . . . 99
antisymmetry . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 46 elementary form . . . . . . . . . . . . . . . . . . . . . . . . . . . . 49
atlas . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18 elliptic PDE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 102
ball . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1 exact form field . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 83
biharmonic equation . . . . . . . . . . . . . . . . . . . . . . . . 102 exception set . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
bilinearity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4 explicit representation . . . . . . . . . . . . . . . . . . . . . . . 21
Biot–Savart law . . . . . . . . . . . . . . . . . . . . . . . . . . . . 101 extended parameter domain . . . . . . . . . . . . . . . . . . 43
Black–Scholes equation . . . . . . . . . . . . . . . . . . . . 102 exterior derivative . . . . . . . . . . . . . . . . . . . . . 69,70,83
boundary . . . . . . . . . . . . . . . . . . . . . . . . . . 1,43,63,115 exterior vector . . . . . . . . . . . . . . . . . . . . . . . . . . 66,114
Cartan’s Magic Formula . . . . . . . . . . . . . . . . . 72,107 face . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 70
chain rule . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9 Faraday’s form field . . . . . . . . . . . . . . . . . . . . . . 61,76
charge conservation law . . . . . . . . . . . . . . . . . . . . . 76 Faraday’s law . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 75
chart . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18 FEM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 109
closed form field . . . . . . . . . . . . . . . . . . . . . . . . . . . . 83 fiber . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37
closed set . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1 fiber bundle . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37
closure of a set . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1 finite element method . . . . . . . . . . . . . . . . . . . . . . 110
connected manifold . . . . . . . . . . . . . . . . . . . . . . . . . 58 flux form field . . . . . . . . . . . . . . . . . . . . . . . . . . . 59,73
conservative vector field . . . . . . . . . . . . . . . . . . . . . 90 flux . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36
continuity law . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 75 forcing function . . . . . . . . . . . . . . . . . . . . . . . . . . . . 102
coordinate function . . . . . . . . . . . . . . . . . . . . . . . . . . . 6 form field . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52
coordinate point . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5 form . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 46
coordinate transform . . . . . . . . . . . . . . . . . . . . . . . . 7,9 four-potential . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 97
coordinate vector . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6 Fundamental Theorem of Integral Calculus 62,83
coordinate-freeness . . . . . . . . . . . . . . . . . . . . . . . 12,22 Gauß’ law . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 75
cover . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38 Gauß’ Theorem . . . . . . . . . . . . . . . . . . . . . . . . . . 62,78
cross product . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4,7 General Divergence Theorem . . . . . . . . . . . . . 78,115
curl . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10,74,115 Generalized Stokes’ Theorem . . . . . . . . . 62,77,114
current density form field . . . . . . . . . . . . . . . . . . . . 76 geometric vector . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
curvilinear coordinate system . . . . . . . . . . . . . . . 111 gradient . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9,73,115
cyclical symmetry . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5 Gradient Theorem . . . . . . . . . . . . . . . . . . . . . . . . . . . 62
cylindrical coordinates . . . . . . . . . . . . . . . . . . 111,117 Gramian . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37,40,41
d’Alembertian . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 106 graph . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
density form field . . . . . . . . . . . . . . . . . . . . . . . . 59,73 Green’s first identity . . . . . . . . . . . . . . . . . . . . . . . . 109
derivative . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9,70 Green’s second identity . . . . . . . . . . . . . . . . . . . . . 109
difference of vectors . . . . . . . . . . . . . . . . . . . . . . . . . . 3 Green’s Theorem . . . . . . . . . . . . . . . . . . . . . . . . . 62,79
differential . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 70 Green’s third identity . . . . . . . . . . . . . . . . . . . . . . . 110
differential form . . . . . . . . . . . . . . . . . . . . . . . . . . . . 46 harmonic function . . . . . . . . . . . . . . . . . . . . . . . 97,109
diffusion equation . . . . . . . . . . . . . . . . . . . . . . . . . . 103 heat equation . . . . . . . . . . . . . . . . . . . . . . . . . . 103,104
dipole approximation . . . . . . . . . . . . . . . . . . . . . . . . 98 Helmholtz’s decomposition . . . . . . . . . . . . . . . . . . 96
Dirac’s string . . . . . . . . . . . . . . . . . . . . . . . . . . . 93,121 Hodge’s decomposition . . . . . . . . . . . . . . . . . . . . . . 97
Index 127