Lecture Notes Vector Analysis MATH 332

Download as pdf or txt
Download as pdf or txt
You are on page 1of 118

Lecture Notes

Vector Analysis
MATH 332
Ivan Avramidi
New Mexico Institute of Mining and Technology

Socorro, NM 87801

May 19, 2004

Author: Ivan Avramidi; File: vecanal4.tex; Date: June 30, 2005; Time: 19:16
Contents

1 Linear Algebra 1
1.1 Vectors in Rn and Matrix Algebra . . . . . . . . . . . . . . . . . . . 1
1.1.1 Vectors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.1.2 Matrices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
1.1.3 Determinant . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
1.1.4 Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
1.2 Vector Spaces . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11
1.2.1 Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
1.3 Inner Product and Norm . . . . . . . . . . . . . . . . . . . . . . . . 14
1.3.1 Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
1.4 Linear Operators . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
1.4.1 Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24

2 Vector and Tensor Algebra 27


2.1 Metric Tensor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27
2.2 Dual Space and Covectors . . . . . . . . . . . . . . . . . . . . . . . 29
2.2.1 Einstein Summation Convention . . . . . . . . . . . . . . . . 31
2.3 General Definition of a Tensor . . . . . . . . . . . . . . . . . . . . . 34
2.3.1 Orientation, Pseudotensors and Volume . . . . . . . . . . . . 37
2.4 Operators and Tensors . . . . . . . . . . . . . . . . . . . . . . . . . 41
2.5 Vector Algebra in R3 . . . . . . . . . . . . . . . . . . . . . . . . . . 44

3 Geometry 49
3.1 Geometry of Euclidean Space . . . . . . . . . . . . . . . . . . . . . 49
3.2 Basic Topology of Rn . . . . . . . . . . . . . . . . . . . . . . . . . . 53
3.3 Curvilinear Coordinate Systems . . . . . . . . . . . . . . . . . . . . 54
3.3.1 Change of Coordinates . . . . . . . . . . . . . . . . . . . . . 56
3.3.2 Examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . 57
3.4 Vector Functions of a Single Variable . . . . . . . . . . . . . . . . . 59
3.5 Geometry of Curves . . . . . . . . . . . . . . . . . . . . . . . . . . . 61
3.6 Geometry of Surfaces . . . . . . . . . . . . . . . . . . . . . . . . . 65

I
II CONTENTS

4 Vector Analysis 69
4.1 Vector Functions of Several Variables . . . . . . . . . . . . . . . . . 69
4.2 Directional Derivative and the Gradient . . . . . . . . . . . . . . . . 71
4.3 Exterior Derivative . . . . . . . . . . . . . . . . . . . . . . . . . . . 73
4.4 Divergence . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 76
4.5 Curl . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 77
4.6 Laplacian . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 78
4.7 Differential Vector Identities . . . . . . . . . . . . . . . . . . . . . . 79
4.8 Orthogonal Curvilinear Coordinate Systems in R3 . . . . . . . . . . . 80

5 Integration 83
5.1 Line Integrals . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 83
5.2 Surface Integrals . . . . . . . . . . . . . . . . . . . . . . . . . . . . 84
5.3 Volume Integrals . . . . . . . . . . . . . . . . . . . . . . . . . . . . 86
5.4 Fundamental Integral Theorems . . . . . . . . . . . . . . . . . . . . 87
5.4.1 Fundamental Theorem of Line Integrals . . . . . . . . . . . . 87
5.4.2 Green’s Theorem . . . . . . . . . . . . . . . . . . . . . . . . 87
5.4.3 Stokes’s Theorem . . . . . . . . . . . . . . . . . . . . . . . . 87
5.4.4 Gauss’s Theorem . . . . . . . . . . . . . . . . . . . . . . . . 87
5.4.5 General Stokes’s Theorem . . . . . . . . . . . . . . . . . . . 88

6 Potential Theory 89
6.1 Simply Connected Domains . . . . . . . . . . . . . . . . . . . . . . 90
6.2 Conservative Vector Fields . . . . . . . . . . . . . . . . . . . . . . . 91
6.2.1 Scalar Potential . . . . . . . . . . . . . . . . . . . . . . . . . 91
6.3 Irrotational Vector Fields . . . . . . . . . . . . . . . . . . . . . . . . 92
6.4 Solenoidal Vector Fields . . . . . . . . . . . . . . . . . . . . . . . . 93
6.4.1 Vector Potential . . . . . . . . . . . . . . . . . . . . . . . . . 93
6.5 Laplace Equation . . . . . . . . . . . . . . . . . . . . . . . . . . . . 94
6.5.1 Harmonic Functions . . . . . . . . . . . . . . . . . . . . . . 94
6.6 Poisson Equation . . . . . . . . . . . . . . . . . . . . . . . . . . . . 95
6.6.1 Dirac Delta Function . . . . . . . . . . . . . . . . . . . . . . 95
6.6.2 Point Sources . . . . . . . . . . . . . . . . . . . . . . . . . . 95
6.6.3 Dirichlet Problem . . . . . . . . . . . . . . . . . . . . . . . . 95
6.6.4 Neumann Problem . . . . . . . . . . . . . . . . . . . . . . . 95
6.6.5 Green’s Functions . . . . . . . . . . . . . . . . . . . . . . . 95
6.7 Fundamental Theorem of Vector Analysis . . . . . . . . . . . . . . . 96

7 Basic Concepts of Differential Geometry 97


7.1 Manifolds . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 98
7.2 Differential Forms . . . . . . . . . . . . . . . . . . . . . . . . . . . . 99
7.2.1 Exterior Product . . . . . . . . . . . . . . . . . . . . . . . . 99
7.2.2 Exterior Derivative . . . . . . . . . . . . . . . . . . . . . . . 99
7.3 Integration of Differential Forms . . . . . . . . . . . . . . . . . . . . 100
7.4 General Stokes’s Theorem . . . . . . . . . . . . . . . . . . . . . . . 101
7.5 Tensors in General Curvilinear Coordinate Systems . . . . . . . . . . 102

vecanal4.tex; June 30, 2005; 19:16; p. 1


CONTENTS III

7.5.1 Covariant Derivative . . . . . . . . . . . . . . . . . . . . . . 102

8 Applications 103
8.1 Mechanics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 104
8.1.1 Inertia Tensor . . . . . . . . . . . . . . . . . . . . . . . . . . 104
8.1.2 Angular Momentum Tensor . . . . . . . . . . . . . . . . . . 104
8.2 Elasticity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 105
8.2.1 Strain Tensor . . . . . . . . . . . . . . . . . . . . . . . . . . 105
8.2.2 Stress Tensor . . . . . . . . . . . . . . . . . . . . . . . . . . 105
8.3 Fluid Dynamics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 106
8.3.1 Continuity Equation . . . . . . . . . . . . . . . . . . . . . . 106
8.3.2 Tensor of Momentum Flux Density . . . . . . . . . . . . . . 106
8.3.3 Euler’s Equations . . . . . . . . . . . . . . . . . . . . . . . . 106
8.3.4 Rate of Deformation Tensor . . . . . . . . . . . . . . . . . . 106
8.3.5 Navier-Stokes Equations . . . . . . . . . . . . . . . . . . . . 106
8.4 Heat and Diffusion Equations . . . . . . . . . . . . . . . . . . . . . . 107
8.5 Electrodynamics . . . . . . . . . . . . . . . . . . . . . . . . . . . . 108
8.5.1 Tensor of Electromagnetic Field . . . . . . . . . . . . . . . . 108
8.5.2 Maxwell Equations . . . . . . . . . . . . . . . . . . . . . . . 108
8.5.3 Scalar and Vector Potentials . . . . . . . . . . . . . . . . . . 108
8.5.4 Wave Equations . . . . . . . . . . . . . . . . . . . . . . . . . 108
8.5.5 D’Alambert Operator . . . . . . . . . . . . . . . . . . . . . . 108
8.5.6 Energy-Momentum Tensor . . . . . . . . . . . . . . . . . . . 108
8.6 Basic Concepts of Special and General Relativity . . . . . . . . . . . 109

Bibliography 111

Notation 113

Index 113

vecanal4.tex; June 30, 2005; 19:16; p. 2


IV CONTENTS

vecanal4.tex; June 30, 2005; 19:16; p. 3


Chapter 1

Linear Algebra

1.1 Vectors in Rn and Matrix Algebra


1.1.1 Vectors
• Rn is the set of all ordered n-tuples of real numbers, which can be assembled as
columns or as rows.

• Let x1 , . . . , xn be n real numbers. Then the column-vector (or just vector) is an


ordered n-tuple of the form
 
 v1 
 v2 
v =  .  ,
 .. 
 
vn
and the row-vector (also called a covector) is an ordered n-tuple of the form

vT = (v1 , v2 , . . . , vn ) .

The real numbers x1 , . . . xn are called the components of the vectors.

• The operation that converts column-vectors into row-vectors and vice versa pre-
serving the order of the components is called the transposition and denoted by
T . That is
 T  
 v1   v1 
 v2   v2 
  = (v , v , . . . , v )  
 ..  1 2 n and (v1 , v2 , . . . , vn )T =  ..  .
 .   . 
  
vn vn

Of course, for any vector v


(vT )T = v .

1
2 CHAPTER 1. LINEAR ALGEBRA

• The addition of vectors is defined by


 
 u1 + v1 
 u + v 
 2 
u + v = 
2
..  ,
 . 
 
un + vn
and
u + v = (u1 + v1 , . . . , un + vn ) .
• Notice that one cannot add a column-vector and a row-vector!
• The multiplication of vectors by a real constant, called a scalar, is defined by
 
 av1 
 av2 
av =  .  , av = (av1 , . . . , avn ) .
 .. 
 
avn

• The vectors that have only zero elements are called zero vectors, that is
 
 0 
 0 
 
0 =  .  , 0T = (0, . . . , 0) .
 .. 
 
0

• The set of column-vectors


     
 1   0   0 
 0   1   .. 
     . 
     
e1 =  0  , e2 =  0  , ··· , en =  0 
 .   ..   
 .. 
 

 .  
 0 
0 0 1
and the set of row-vectors
eT1 = (1, 0, . . . , 0), eT2 = (0, 1, . . . , 0), eTn = (0, 0, . . . , 1)
are called the standard (or canonical) bases in Rn .
• There is a natural product of column-vectors and row-vectors that assigns to a
row-vector and a column-vector a real number
 
 v1 
 v  X n
 2 
huT , vi = (u1 , u2 , . . . , un )  .  = ui vi = u1 v1 + u2 v2 + · · · + un vn .
 ..  i=1
 
vn
This is the simplest instance of a more general multiplication rule for matrices
which can be summarized by saying that one multiplies row by column.

vecanal4.tex; June 30, 2005; 19:16; p. 4


1.1. VECTORS IN RN AND MATRIX ALGEBRA 3

• The product of two column-vectors and the product of two row-vectors, called
the inner product (or the scalar product), is defined then by
X
n
(u, v) = (uT , vT ) = huT , vi = ui vi = u1 v1 + · · · + un vn .
i=1

• Finally, we define the norm (or the length) of both column-vectors and row-
vectors is defined by
 n 1/2 q
p X 
||v|| = ||v || = hv , vi =  v2i  = v2 + · · · + v2n .
T T
1
i=1

1.1.2 Matrices
• A set of n2 real numbers Ai j , i, j = 1, . . . , n, arranged in an array that has n
columns and n rows
 
 A11 A12 · · · A1n 
 A 
 21 A22 · · · A2n 
A =  . .. .. .. 
 .. . . . 

An1 An2 · · · Ann
is called a square n × n real matrix.
• The set of all real square n × n matrices is denoted by Mat(n, R).
• The number Ai j (also called an entry of the matrix) appears in the i-th row and
the j-th column of the matrix A
 
 A11 A12 · · · A1 j · · · A1n 
 A21 A22 · · · A2 j · · · A2n 
 
 .. .. .. .. .. .. 
 . . . . . . 
A =  
 Ai1 Ai2 · · · Ai j · · · Ain 
 
 .. .. .. .. .. .. 
 . . . . . . 
An1 An2 · · · An j · · · Ann

• Remark. Notice that the first index indicates the row and the second index
indicates the column of the matrix.
• The matrix whose all entries are equal to zero is called the zero matrix.
• The addition of matrices is defined by
 
 A11 + B11 A12 + B12 ··· A1n + B1n 
 A + B A22 + B22 ··· A2n + B2n 
 21 
A + B = 
21
.. .. .. .. 
 . . . . 
 
An1 + Bn1 An2 + Bn2 ··· Ann + Bnn

vecanal4.tex; June 30, 2005; 19:16; p. 5


4 CHAPTER 1. LINEAR ALGEBRA

and the multiplication by scalars by


 
 cA11 cA12 ··· cA1n 
 cA cA22 ··· cA2n 
 
cA =  .
21
.. .. .. 
 .. . . . 
 
cAn1 cAn2 ··· cAnn

• The numbers Aii are called the diagonal entries. Of course, there are n diagonal
entries. The set of diagonal entries is called the diagonal of the matrix A.
• The numbers Ai j with i , j are called off-diagonal entries; there are n(n − 1)
off-diagonal entries.
• The numbers Ai j with i < j are called the upper triangular entries. The set of
upper triangular entries is called the upper triangular part of the matrix A.
• The numbers Ai j with i > j are called the lower triangular entries. The set of
lower triangular entries is called the lower triangular part of the matrix A.
• The number of upper-triangular entries and the lower-triangular entries is the
same and is equal to n(n − 1)/2.
• A matrix whose only non-zero entries are on the diagonal is called a diagonal
matrix. For a diagonal matrix

Ai j = 0 if i , j.

• The diagonal matrix


 
 λ1 0 ··· 0 
 0 λ2 ··· 0 
 
A =  .. .. .. .. 
 . . . . 
 
0 0 ··· λn
is also denoted by
A = diag (λ1 , λ2 , . . . , λn )

• A diagonal matrix whose all diagonal entries are equal to 1


 
 1 0 ··· 0 
 0 1 ··· 0 
 
I =  .. .. .. .. 
 . . . . 
 
0 0 ··· 1

is called the identity matrix. The elements of the identity matrix are


 1, if i = j

Ii j = 

 0, if i , j .

vecanal4.tex; June 30, 2005; 19:16; p. 6


1.1. VECTORS IN RN AND MATRIX ALGEBRA 5

• A matrix A of the form  


 ∗ ∗ ··· ∗ 
 0 ∗ ··· ∗ 
 
A =  .. .. .. .. 
 . . . . 
 
0 0 ··· ∗
where ∗ represents nonzero entries is called an upper triangular matrix. Its
lower triangular part is zero, that is,

Ai j = 0 if i < j.

• A matrix A of the form  


 ∗ 0 ··· 0 
 ∗ ∗ ··· 0 
 
A =  .. .. . . .. 
 . . . . 
 
∗ ∗ ··· ∗
whose upper triangular part is zero, that is,

Ai j = 0 if i > j,

is called a lower triangular matrix.

• The transpose of a matrix A whose i j-th entry is Ai j is the matrix AT whose


i j-th entry is A ji . That is, AT obtained from A by switching the roles of rows and
columns of A:
 
 A11 A21 · · · A j1 · · · An1 
 A 
 12 A22 · · · A j2 · · · An2 
 .. .. .. .. .. .. 
 . . . . . . 
A = 
T 
 A1i A2i · · · A ji · · · Ani 
 
 .. .. .. .. .. .. 
 . . . . . . 
A11 A2n · · · A jn · · · Ann
or
(AT )i j = A ji .

• A matrix A is called symmetric if

AT = A

and anti-symmetric if
AT = −A .

• The number of independent entries of an anti-symmetric matrix is n(n − 1)/2.

• The number of independent entries of a symmetric matrix is n(n + 1)/2.

vecanal4.tex; June 30, 2005; 19:16; p. 7


6 CHAPTER 1. LINEAR ALGEBRA

• Every matrix A can be uniquely decomposed as the sum of its diagonal part AD ,
the lower triangular part AL and the upper triangular part AU

A = AD + AL + AU .

• For an anti-symmetric matrix

ATU = −AL and AD = 0 .

• For a symmetric matrix


ATU = AL .

• Every matrix A can be uniquely decomposed as the sum of its symmetric part AS
and its anti-symmetric part AA

A = AS + A A ,

where
1 1
AS = (A + AT ) , AA = (A − AT ) .
2 2
• The product of matrices is defined as follows. The i j-th entry of the product
C = AB of two matrices A and B is
X
n
Ci j = Aik Bk j = Ai1 B1 j + Ai2 B2 j + · · · + Ain Bn j .
k=1

This is again a multiplication of the “i-th row of the matrix A by the j-th column
of the matrix B”.
• Theorem 1.1.1 The product of matrices is associative, that is, for any matrices
A, B, C
(AB)C = A(BC) .

• Theorem 1.1.2 For any two matrices A and B

(AB)T = BT AT .

• A matrix A is called invertible if there is another matrix A−1 such that

AA−1 = A−1 A = I .

The matrix A−1 is called the inverse of A.


• Theorem 1.1.3 For any two invertible matrices A and B

(AB)−1 = B−1 A−1 ,

and
(A−1 )T = (AT )−1 .

vecanal4.tex; June 30, 2005; 19:16; p. 8


1.1. VECTORS IN RN AND MATRIX ALGEBRA 7

• A matrix A is called orthogonal if

AT A = AAT = I ,

which means AT = A−1 .

• The trace is a map tr : Mat(n, R) that assigns to each matrix A = (Ai j ) a real
number tr A equal to the sum of the diagonal elements of a matrix
X
n
tr A = Akk .
k=1

• Theorem 1.1.4 The trace has the properties

tr (AB) = tr (BA) ,

and
tr AT = tr A .

• Obviously, the trace of an anti-symmetric matrix is equal to zero.

• Finally, we define the multiplication of column-vectors by matrices from the left


and the multiplication of row-vectors by matrices from the right as follows.
• Each matrix defines a natural left action on a column-vector and a right action
on a row-vector.
• For each column-vector v and a matrix A = (Ai j ) the column-vector u = Av is
given by
      
 u1   A11 A12 ··· A1n   v1   A11 v1 + A12 v2 + · · · + A1n vn 
 u2   A A22 ··· A2n   v2   A v + A v + · · · + A v 
   21     21 1 22 2 2n n 
 ..   .. .. .. ..   ..   .. 
 .   . . . .   .   . 
  =     =  
 ui   Ai1 Ai2 ··· Ain   vi   Ai1 v1 + Ai2 v2 + · · · + Ain vn 
 ..   . .. .. ..   ..   .. 
 .   .. . . .   .   . 
      
un An1 An2 ··· Ann vn An1 v1 + An2 v2 + · · · + Ann vn

• The components of the vector u are


X
n
ui = Ai j v j = Ai1 v1 + Ai2 v2 + · · · + Ain vn .
j=1

• Similarly, for a row vector vT the components of the row-vector uT = vT A are


defined by
X
n
ui = v j A ji = v1 A1i + v2 A2i + · · · + vn Ani .
j=1

vecanal4.tex; June 30, 2005; 19:16; p. 9


8 CHAPTER 1. LINEAR ALGEBRA

1.1.3 Determinant
• Consider the set Zn = {1, 2, . . . , n} of the first n integers. A permutation ϕ of the
set {1, 2, . . . , n} is an ordered n-tuple (ϕ(1), . . . , ϕ(n)) of these numbers.
• That is, a permutation is a bijective (one-to-one and onto) function

ϕ : Zn → Zn

that assigns to each number i from the set Zn = {1, . . . , n} another number ϕ(i)
from this set.
• An elementary permutation is a permutation that exchanges the order of only
two numbers.
• Every permutation can be realized as a product (or a composition) of elemen-
tary permutations. A permutation that can be realized by an even number of
elementary permutations is called an even permutation. A permutation that
can be realized by an odd number of elementary permutations is called an odd
permutation.
• Proposition 1.1.1 The parity of a permutation does not depend on the repre-
sentation of a permutation by a product of the elementary ones.
• That is, each representation of an even permutation has even number of elemen-
tary permutations, and similarly for odd permutations.
• The sign of a permutation ϕ, denoted by sign(ϕ) (or simply (−1)ϕ ), is defined
by (
ϕ +1, if ϕ is even,
sign(ϕ) = (−1) =
−1, if ϕ is odd
• The set of all permutations of n numbers is denoted by S n .
• Theorem 1.1.5 The cardinality of this set, that is, the number of different per-
mutations, is
|S n | = n! .

• The determinant is a map det : Mat(n, R) → R that assigns to each matrix


A = (Ai j ) a real number det A defined by
X
det A = sign (ϕ)A1ϕ(1) · · · Anϕ(n) ,
ϕ∈S n

where the summation goes over all n! permutations.


• The most important properties of the determinant are listed below:

Theorem 1.1.6 1. The determinant of the product of matrices is equal to the


product of the determinants:
det(AB) = det A det B .

vecanal4.tex; June 30, 2005; 19:16; p. 10


1.1. VECTORS IN RN AND MATRIX ALGEBRA 9

2. The determinants of a matrix A and of its transpose AT are equal:

det A = det AT .

3. The determinant of the inverse A−1 of an invertible matrix A is equal to the


inverse of the determinant of A:

det A−1 = (det A)−1

4. A matrix is invertible if and only if its determinant is non-zero.

• The set of real invertible matrices (with non-zero determinant) is denoted by


GL(n, R). The set of matrices with positive determinant is denoted by GL+ (n, R).
• A matrix with unit determinant is called unimodular.
• The set of real matrices with unit determinant is denoted by S L(n, R).
• The set of real orthogonal matrices is denoted by O(n).
• Theorem 1.1.7 The determinant of an orthogonal matrix is equal to either 1 or
−1.
• An orthogonal matrix with unit determinant (a unimodular orthogonal matrix) is
called a proper orthogonal matrix or just a rotation.
• The set of real orthogonal matrices with unit determinant is denoted by S O(n).
• A set G of invertible matrices forms a group if it is closed under taking inverse
and matrix multiplication, that is, if the inverse A−1 of any matrix A in G belongs
to the set G and the product AB of any two matrices A and B in G belongs to G.

1.1.4 Exercises
1. Show that the product of invertible matrices is an invertible matrix.
2. Show that the product of matrices with positive determinant is a matrix with positive
determinant.
3. Show that the inverse of a matrix with positive determinant is a matrix with positive
determinant.
4. Show that GL(n, R) forms a group (called the general linear group).
5. Show that GL+ (n, R) is a group (called the proper general linear group).
6. Show that the inverse of a matrix with negative determinant is a matrix with negative
determinant.
7. Show that: a) the product of an even number of matrices with negative determinant is a
matrix with positive determinant, b) the product of odd matrices with negative determinant
is a matrix with negative determinant.
8. Show that the product of matrices with unit determinant is a matrix with unit determinant.
9. Show that the inverse of a matrix with unit determinant is a matrix with unit determinant.

vecanal4.tex; June 30, 2005; 19:16; p. 11


10 CHAPTER 1. LINEAR ALGEBRA

10. Show that S L(n, R) forms a group (called the special linear group or the unimodular
group).
11. Show that the product of orthogonal matrices is an orthogonal matrix.
12. Show that the inverse of an orthogonal matrix is an orthogonal matrix.
13. Show that O(n) forms a group (called the orthogonal group).
14. Show that orthogonal matrices have determinant equal to either +1 or −1.
15. Show that the product of orthogonal matrices with unit determinant is an orthogonal ma-
trix with unit determinant.
16. Show that the inverse of an orthogonal matrix with unit determinant is an orthogonal
matrix with unit determinant.
17. Show that S O(n) forms a group (called the proper orthogonal group or the rotation
group).

vecanal4.tex; June 30, 2005; 19:16; p. 12


1.2. VECTOR SPACES 11

1.2 Vector Spaces


• A real vector space consists of a set E, whose elements are called vectors, and
the set of real numbers R, whose elements are called scalars. There are two
operations on a vector space:
1. Vector addition, + : E × E → E, that assigns to two vectors u, v ∈ E
another vector u + v, and
2. Multiplication by scalars, · : R × E → E, that assigns to a vector v ∈ E
and a scalar a ∈ R a new vector av ∈ E.
The vector addition is an associative commutative operation with an additive
identity. It satisfies the following conditions:
1. u + v = v + u, ∀u, v, ∈ E
2. (u + v) + w = u + (v + w), ∀u, v, w ∈ E
3. There is a vector 0 ∈ E, called the zero vector, such that for any v ∈ E
there holds v + 0 = v.
4. For any vector v ∈ E, there is a vector (−v) ∈ E, called the opposite of v,
such that v + (−v) = 0.
The multiplication by scalars satisfies the following conditions:
1. a(bv) = (ab)v, ∀v ∈ E, ∀a, bR,
2. (a + b)v = av + bv, ∀v ∈ E, ∀a, bR,
3. a(u + v) = au + av, ∀u, v ∈ E, ∀aR,
4. 1 v = v ∀v ∈ E.
• The zero vector is unique.
• For any u, v ∈ E there is a unique vector denoted by w = v − u, called the
difference of v and u, such that u + w = v.
• For any v ∈ E,
0v = 0 , and (−1)v = −v .

• Let E be a real vector space and A = {e1 , . . . , ek } be a finite collection of vectors


from E. A linear combination of these vectors is a vector

a1 e1 + · · · + ak ek ,

where {a1 , . . . , an } are scalars.


• A finite collection of vectors A = {e1 , . . . , ek } is linearly independent if

a1 e1 + · · · + ak ek = 0

implies a1 = · · · = ak = 0.

vecanal4.tex; June 30, 2005; 19:16; p. 13


12 CHAPTER 1. LINEAR ALGEBRA

• A collection A of vectors is linearly dependent if it is not linearly independent.

• Two non-zero vectors u and v which are linearly dependent are also called par-
allel, denoted by u||v.

• A collection A of vectors is linearly independent if no vector of A is a linear


combination of a finite number of vectors from A.

• Let A be a subset of a vector space E. The span of A, denoted by span A, is the


subset of E consisting of all finite linear combinations of vectors from A, i.e.

span A = {v ∈ E | v = a1 e1 + · · · + ak ek , ei ∈ A, ai ∈ R} .

We say that the subset span A is spanned by A.

• Theorem 1.2.1 The span of any subset of a vector space is a vector space.

• A vector subspace of a vector space E is a subset S ⊆ E of E which is itself a


vector space.

• Theorem 1.2.2 A subset S of E is a vector subspace of E if and only if span S =


S.

• Span of A is the smallest subspace of E containing A.

• A collection B of vectors of a vector space E is a basis of E if B is linearly


independent and span B = E.

• A vector space E is finite-dimensional if it has a finite basis.

• Theorem 1.2.3 If the vector space E is finite-dimensional, then the number of


vectors in any basis is the same.

• The dimension of a finite-dimensional real vector space E, denoted by dim E, is


the number of vectors in a basis.

• Theorem 1.2.4 If {e1 , . . . , en } is a basis in E, then for every vector v ∈ E there


is a unique set of real numbers (vi ) = (v1 , . . . , vn ) such that
X
n
v= vi ei = v1 e1 + · · · + vn en .
i=1

• The real numbers vi , i = 1, . . . , n, are called the components of the vector v


with respect to the basis {ei }.

• It is customary to denote the components of vectors by superscripts, which


should not be confused with powers of real numbers

v2 , (v)2 = vv, ..., vn , (v)n .

vecanal4.tex; June 30, 2005; 19:16; p. 14


1.2. VECTOR SPACES 13

Examples of Vector Subspaces


• Zero subspace {0}.
• Line with a tangent vector u:

S 1 = span {u} = {v ∈ E | v = tu, t ∈ R} .

• Plane spanned by two nonparallel vectors u1 and u2

S 2 = span {u1 , u2 } = {v ∈ E | v = tu1 + su2 , t, s ∈ R} .

• More generally, a k-plane spanned by a linearly independent collection of k vec-


tors {u1 , . . . , uk }

S k = span {u1 , . . . , uk } = {v ∈ E | v = t1 u1 + · · · + tk uk , t1 , . . . , tk ∈ R} .

• An (n − 1)-plane in an n-dimensional vector space is called a hyperplane.

1.2.1 Exercises
1. Show that if λv = 0, then either v = 0 or λ = 0.
2. Prove that the span of a collection of vectors is a vector subspace.

vecanal4.tex; June 30, 2005; 19:16; p. 15


14 CHAPTER 1. LINEAR ALGEBRA

1.3 Inner Product and Norm


• A real vector space E is called an inner product space if there is a function
(·, ·) : E × E → R, called the inner product, that assigns to every two vectors u
and v a real number (u, v) and satisfies the conditions: ∀u, v, w ∈ E, ∀a ∈ R:

1. (v, v) ≥ 0
2. (v, v) = 0 if and only if v = 0
3. (u, v) = (v, u)
4. (u + v, w) = (u, w) + (v, w)
5. (au, v) = (u, av) = a(u, v)

A finite-dimensional inner product space is called a Euclidean space.

• The inner product is often called the dot product, or the scalar product, and is
denoted by
(u, v) = u · v .

• All spaces considered below are Euclidean spaces. Henceforth, E will denote an
n-dimensional Euclidean space if not specified otherwise.

• The Euclidean norm is a function || · || : E → R that assigns to every vector


v ∈ E a real number ||v|| defined by
p
||v|| = (v, v).

• The norm of a vector is also called the length.

• A vector with unit norm is called a unit vector.

• Theorem 1.3.1 For any u, v ∈ E there holds

||u + v||2 = ||u||2 + 2(u, v) + ||v||2 .

• Theorem 1.3.2 Cauchy-Schwarz’s Inequality. For any u, v ∈ E there holds

|(u, v)| ≤ ||u|| ||v|| .

The equality

|(u, v)| = ||u|| ||v||

holds if and only if u and v are parallel.

• Corollary 1.3.1 Triangle Inequality. For any u, v ∈ E there holds

||u + v|| ≤ ||u|| + ||v|| .

vecanal4.tex; June 30, 2005; 19:16; p. 16


1.3. INNER PRODUCT AND NORM 15

• The angle between two non-zero vectors u and v is defined by


(u, v)
cos θ = , 0 ≤ θ ≤ π.
||u|| ||v||
Then the inner product can be written in the form

(u, v) = ||u|| ||v|| cos θ .

• Two non-zero vectors u, v ∈ E are orthogonal, denoted by u ⊥ v, if

(u, v) = 0.

• A basis {e1 , . . . , en } is called orthonormal if each vector of the basis is a unit


vector and any two distinct vectors are orthogonal to each other, that is,
(
1, if i = j
(ei , e j ) = .
0, if i , j

• Theorem 1.3.3 Every Euclidean space has an orthonormal basis.

• Let S ⊂ E be a nonempty subset of E. We say that x ∈ E is orthogonal to S ,


denoted by x ⊥ S , if x is orthogonal to every vector of S .
• The set
S ⊥ = {x ∈ E | x ⊥ S }
of all vectors orthogonal to S is called the orthogonal complement of S .
• Theorem 1.3.4 The orthogonal complement of any subset of a Euclidean space
is a vector subspace.
• Two subsets A and B of E are orthogonal, denoted by A ⊥ B, if every vector of
A is orthogonal to every vector of B.

• Let S be a subspace of E and S ⊥ be its orthogonal complement. If every element


of E can be uniquely represented as the sum of an element of S and an element
of S ⊥ , then E is the direct sum of S and S ⊥ , which is denoted by

E = S ⊕ S⊥ .

• The union of a basis of S and a basis of S ⊥ gives a basis of E.

1.3.1 Exercises
1. Show that the Euclidean norm has the following properties
(a) ||v|| ≥ 0, ∀v ∈ E;
(b) ||v|| = 0 if and only if v = 0;

vecanal4.tex; June 30, 2005; 19:16; p. 17


16 CHAPTER 1. LINEAR ALGEBRA

(c) ||av|| = |a| ||v||, ∀v ∈ E, ∀a ∈ R.


2. Parallelogram Law. Show that for any u, v ∈ E
 
||u + v||2 + ||u − v||2 = 2 ||u||2 + ||v||2

3. Show that any orthogonal system in E is linearly independent.


4. Gram-Schmidt orthonormalization process. Let G = {u1 , · · · , uk } be a linearly inde-
pendent collection of vectors. Let O = {v1 , · · · , vk } be a new collection of vectors defined
recursively by

v1 = u1 ,
X
j−1
vi (vi , u j )
vj = uj − , 2 ≤ j ≤ k,
i=1
||vi ||2

and the collection B = {e1 , . . . , ek } be defined by


vi
ei = .
||vi ||
Show that: a) O is an orthogonal system and b) B is an orthonormal system.
5. Pythagorean Theorem. Show that if u ⊥ v, then

||u + v||2 = ||u||2 + ||v||2 .

6. Let B = {e1 , · · · en } be an orthonormal basis in E. Show that for any vector v ∈ E


X
n
v= ei (ei , v)
i=1

and
X
n
||v||2 = (ei , v)2 .
i=1

7. Prove that the orthogonal complement of a subset S of E is a vector subspace of E.


8. Let S be a subspace in E. Prove that
a) E ⊥ = {0}, b) {0}⊥ = E, c) (S ⊥ )⊥ = S .
9. Show that the intersection of orthogonal subsets of a Euclidean space is either empty or
consists of only the zero vector. That is, for two subsets A and B, if A ⊥ B, then A∩B = {0}
or ∅.

vecanal4.tex; June 30, 2005; 19:16; p. 18


1.4. LINEAR OPERATORS 17

1.4 Linear Operators


• A linear operator on a vector space E is a mapping L : E → E satisfying the
condition ∀u, v ∈ E, ∀a ∈ R,

L(u + v) = L(u) + L(v) and L(av) = aL(v).

• Identity operator I on E is defined by

I v = v, ∀v ∈ E

• Null operator 0 : E → E is defined by

0v = 0, ∀v ∈ E

• The vector u = L(v) is the image of the vector v.


• If S is a subset of E, then the set

L(S ) = {u ∈ E | u = L(v) for some v ∈ S }

is the image of the set S and the set

L−1 (S ) = {v ∈ E | L(v) ∈ S }

is the inverse image of the set A.


• The image of the whole space E of a linear operator L is the range (or the image)
of L, denoted by

Im(L) = L(E) = {u ∈ E | u = L(v) for some v ∈ E} .

• The kernel Ker(L) (or the null space) of an operator L is the set of all vectors in
E which are mapped to zero, that is

Ker (L) = L−1 ({0}) = {v ∈ E | L(v) = 0} .

• Theorem 1.4.1 For any operator L the sets Im(L) and Ker (L) are vector sub-
spaces.
• The dimension of the kernel Ker (L) of an operator L

null (L) = dim Ker (L)

is called the nullity of the operator L.


• The dimension of the range Im(L) of an operator L

rank (L) = dim Ker (L)

is called the rank of the operator L.

vecanal4.tex; June 30, 2005; 19:16; p. 19


18 CHAPTER 1. LINEAR ALGEBRA

• Theorem 1.4.2 For any operator L on an n-dimensional Euclidean space E

rank (L) + null (L) = n

• The set L(E) of all linear operators on a vector space E is a vector space with
the addition of operators and multiplication by scalars defined by

(L1 + L2 )(x) = L1 (x) + L2 (x), and (aL)(x) = aL(x) .

• The product of the operators A and B is the composition of A and B.


• Since the product of operators is defined as a composition of linear mappings,
it is automatically associative, which means that for any operators A, B and C,
there holds
(AB)C = A(BC) .

• The integer powers of an operator are defined as the multiple composition of the
operator with itself, i.e.

A0 = I A1 = A, A2 = AA, . . .

• The operator A on E is invertible if there exists an operator A−1 on E, called the


inverse of A, such that
A−1 A = AA−1 = I .

• Theorem 1.4.3 Let A and B be invertible operators. Then:

(A−1 )−1 = A , (AB)−1 = B−1 A−1 .

• The operators A and B are commuting if

AB = BA

and anti-commuting if
AB = −BA .

• The operators A and B are said to be orthogonal to each other if

AB = BA = 0 .

• An operator A is involutive if
A2 = I
idempotent if
A2 = A ,
and nilpotent if for some integer k

Ak = 0 .

vecanal4.tex; June 30, 2005; 19:16; p. 20


1.4. LINEAR OPERATORS 19

Selfadjoint Operators
• The adjoint A∗ of an operator A is defined by

(Au, v) = (u, A∗ v), ∀u, v ∈ E.

• Theorem 1.4.4 For any two operators A and B

(A∗ )∗ = A , (AB)∗ = B∗ A∗ .

• An operator A is self-adjoint if

A∗ = A

and anti-selfadjoint if

A∗ = −A

• Every operator A can be decomposed as the sum

A = AS + AA

of its selfadjoint part AS and its anti-selfadjoint part AA


1 1
AS = (A + A∗ ) , AA = (A − A∗ ) .
2 2

• An operator A is called unitary if

AA∗ = A∗ A = I .

• An operator A on E is called positive, denoted by A ≥ 0, if it is selfdadjoint and


∀v ∈ E
(Av, v) ≥ 0.

Projection Operators
• Let S be a subspace of E and E = S ⊕ S ⊥ . Then for any u ∈ E there exist unique
v ∈ S and w ∈ S ⊥ such that
u = v + w.
The vector v is called the projection of u onto S .

• The operator P on E defined by

Pu = v

is called the projection operator onto S .

vecanal4.tex; June 30, 2005; 19:16; p. 21


20 CHAPTER 1. LINEAR ALGEBRA

• The operator P⊥ defined by


P⊥ u = w

is the projection operator onto S ⊥ .

• The operators P and P⊥ are called complementary projections. They have the
properties:
P∗ = P, ( P⊥ ) ∗ = P⊥ ,

P + P⊥ = I ,

P2 = P , ( P⊥ ) 2 = P⊥ ,

PP⊥ = P⊥ P = 0 .

• Theorem 1.4.5 An operator P is a projection if and only if P is idempotent and


self-adjoint.

• More generally, a collection of projections {P1 , . . . , Pk } is a complete orthogo-


nal system of complimentary projections if

P i Pk = 0 if i,k

and
X
k
Pi = P1 + · · · + P k = I .
i=1

• A complete orthogonal system of projections defines the orthogonal decomposi-


tion of the vector space
E = E1 ⊕ · · · ⊕ Ek ,

where Ei is the subspace the projection Pi projects onto.

• Theorem 1.4.6 1. The dimension of the subspaces Ei are equal to the ranks
of the projections Pi
dim Ei = rank Pi .

2. The sum of dimensions of the vector subspaces Ei equals the dimension of


the vector space E

X
n
dim Ei = dim E1 + · · · + dim Ek = dim E .
i=1

vecanal4.tex; June 30, 2005; 19:16; p. 22


1.4. LINEAR OPERATORS 21

Spectral Decomposition Theorem


• A real number λ is called an eigenvalue of an operator A if there is a unit vector
u ∈ E such that
Au = λu .
The vector u is called the eigenvector corresponding to the eigenvalue λ.
• The span of all eigenvectors corresponding to the eigenvalue λ of an operator A
is called the eigenspace of λ.
• The dimension of the eigenspace of the eigenvalue λ is called the multiplicity
(also called the geometric multiplicity) of λ.
• An eigenvalue of multiplicity 1 is called simple (or non-degenerate).
• An eigenvalue of multiplicity greater than 1 is called multiple (or degenerate).
• The set of all eigenvalues of an operator is called the spectrum of the operator.
• Theorem 1.4.7 Let A be a selfadjoint operator. Then:
1. The number of eigenvalues counted with multiplicity is equal to the dimen-
sion n = dim E of the vector space E.
2. The eigenvectors corresponding to distinct eigenvalues are orthogonal to
each other.
• Theorem 1.4.8 Spectral Decomposition of Self-Adjoint Operators. Let A
be a selfadjoint operator on E. Then there exists an orthonormal basis B =
{e1 , . . . , en } in E consisting of eigenvectors of A corresponding to the eigenvalues
{λ1 , . . . λn }, and the corresponding system of orthogonal complimentary projec-
tions {P1 , . . . , Pn } onto the one-dimensional eigenspaces Ei ,
X
n
A= λ i Pi .
i=1

The projections {Pi } are defined by


Pi v = ei (ei , v) .
and satisfy the equations
X
n
Pi = I , and Pi P j = 0 if i , j .
i=1

• In other words, for any


X
n
v= ei (ei , v) ,
i=1
we have
X
n
Av = λi ei (ei , v) .
i=1

vecanal4.tex; June 30, 2005; 19:16; p. 23


22 CHAPTER 1. LINEAR ALGEBRA

• Let f : R → R be a real-valued function on R. Let A be a selfadjoint operator on


a Euclidean space E given by its spectral decomposition
X
n
A= λ i Pi ,
i=1

where Pi are the one-dimensional projections. Then one can define a function
of the self-adjoint operator f (A) on E by
X
n
f (A) = f (λi )Pi .
i=1

• The exponential of an operator A is defined by


X∞
1 k X λi
n
exp A = A = e Pi
k=1
k! i=1

• Theorem 1.4.9 Let U be a unitary operator on a real vector space E. Then there
exists an ani-selfadjoint operator A such that

U = exp A .

• Recall that the operators U and A satisfy the equations

U∗ = U−1 and A∗ = −A.

• Let A be a self-adjoint operator with the eigenvalues {λ1 , . . . , λn } . Then the trace
of the operator and the determinant of the operator A are defined by
X
n
tr A = λi , det A = λ1 · · · λn .
i=1

• Note that
tr I = n , det I = 1 .

• The trace of a projection P onto a vector subspace S is equal to its rank, or the
dimension of the vector subspace S ,

tr P = rank P = dim S .

• The trace of a function of a self-adjoint operator A is then


X
n
tr f (A) = f (λi ) .
i=1

If there are multiple eigenvalues, then each eigenvalue should be counted with
its multiplicity.

vecanal4.tex; June 30, 2005; 19:16; p. 24


1.4. LINEAR OPERATORS 23

• Theorem 1.4.10 Let A be a self-adjoint operator. Then

det exp A = etr A .

• Let A be a positive definite operator, A > 0. The zeta-function of the operator


A is defined by
Xn
1
ζ(s) = tr A−s = .
λs
i=1 i

• Theorem 1.4.11 The zeta-functions has the properties

ζ(0) = n ,

and
ζ 0 (0) = − log det A .

Examples
• Let u be a unit vector and Pu be the projection onto the one-dimensional subspace
(line) S u spanned by u defined by

Pu v = u(u, v) .

The orthogonal complement S u⊥ is the hyperplane with the normal u. The oper-
ator Ju defined by
Ju = I − 2Pu
is called the reflection operator with respect to the hyperplane S u⊥ . The reflec-
tion operator is a self-adjoint involution, that is, it has the following properties

J∗u = Ju , J2u = I .

The reflection operator has the eigenvalue −1 with multiplicity 1 and the eigenspace
S u , and the eigenvalue +1 with multiplicity (n − 1) and with eigenspace S u⊥ .

• Let u1 and u2 be an orthonormal system of two vectors and Pu1 ,u2 be the projec-
tion operator onto the two-dimensional space (plane) S u1 ,u2 spanned by u1 and
u2
Pu1 ,u2 v = u1 (u1 , v) + u2 (u2 , v) .
Let Nu1 ,u2 be an operator defined by

Nu1 ,u2 v = u1 (u2 , v) − u2 (u1 , v) .

Then
Nu1 ,u2 Pu1 ,u2 = Pu1 ,u2 Nu1 ,u2 = Nu1 ,u2
and
N2u1 ,u2 = −Pu1 ,u2 .

vecanal4.tex; June 30, 2005; 19:16; p. 25


24 CHAPTER 1. LINEAR ALGEBRA

A rotation operator Ru1 ,u2 (θ) with the angle θ in the plane S u1 ,u2 is defined by

Ru1 ,u2 (θ) = I − Pu1 ,u2 + cos θ Pu1 ,u2 + sin θ Nu1 ,u2 .

The rotation operator is unitary, that is, it satisfies the equation

R∗u1 ,u2 Ru1 ,u2 = I .

• Theorem 1.4.12 Spectral Decomposition of Unitary Operators on Real Vec-


tor Spaces. Let U be a unitary operator on a real vector space E. Then the only
eigenvalues of U are +1 and −1 (possibly multiple) and there exists an orthogo-
nal decomposition
E = E + ⊕ E − ⊕ V1 ⊕ · · · ⊕ Vk ,
where E+ and E− are the eigenspaces corresponding to the eigenvalues 1 and
−1, and {V1 , . . . , vk } are two-dimensional subspaces such that

dim E = dim E+ + dim E− + 2k .

Let P+ , P− , P1 , . . . , Pk be the corresponding orthogonal complimentary system


of projections, that is,
X
k
P+ + P − + Pi = I .
i=1

Then there exists a corresponding system of operators N1 , . . . , Nk satisfying the


equations
N2i = −Pi , Ni Pi = Pi Ni = Ni ,
Ni P j = P j Ni = 0 , if i, j
and the angles θ1 , . . . θk such that

X
k
U = P+ − P− + (cos θi Pi + sin θi Ni ) .
i=1

1.4.1 Exercises
1. Prove that the range and the kernel of any operator are vector spaces.
2. Show that

(aA + bB)∗ = aA∗ + bB∗ ∀a, b ∈ R ,


∗ ∗
(A ) = A
(AB)∗ = B∗ A∗

3. Show that for any operator A the operators AA∗ and A + A∗ are selfadjoint.
4. Show that the product of two selfadjoint operators is selfadjoint if and only if they com-
mute.
5. Show that a polynomial p(A) of a selfadjoint operator A is a selfadjoint operator.

vecanal4.tex; June 30, 2005; 19:16; p. 26


1.4. LINEAR OPERATORS 25

6. Prove that the inverse of an invertible operator is unique.


7. Prove that an operator A is invertible if and only if Ker A = {0}, that is, Av = 0 implies
v = 0.
8. Prove that for an invertible operator A, Im(A) = E, that is, for any vector v ∈ E there is a
vector u ∈ E such that v = Au.
9. Show that if an operator A is invertible, then

(A−1 )−1 = A .

10. Show that the product AB of two invertible operators A and B is invertible and

(AB)−1 = B−1 A−1

11. Prove that the adjoint A∗ of any invertible operator A is invertible and

(A∗ )−1 = (A−1 )∗ .

12. Prove that the inverse A−1 of a selfadjoint invertible operator is selfadjoint.
13. An operator A on E is called isometric if ∀v ∈ E,

||Av|| = ||v|| .

Prove that an operator is unitary if and only if it is isometric.


14. Prove that unitary operators preserves inner product. That is, show that if A is a unitary
operator, then ∀u, v ∈ E
(Au, Av) = (u, v) .
15. Show that for every unitary operator A both A−1 and A∗ are unitary.
16. Show that for any operator A the operators AA∗ and A∗ A are positive.
17. What subspaces do the null operator 0 and the identity operator I project onto?
18. Show that for any two projection operators P and Q, PQ = 0 if and only if QP = 0.
19. Prove the following properties of orthogonal projections

P∗ = P, (P⊥ )∗ = P⊥ , P⊥ + P = I, PP⊥ = P⊥ P = 0 .

20. Prove that an operator is projection if and only if it is idempotent and selfadjoint.
21. Give an example of an idempotent operator in R2 which is not a projection.
22. Show that any projection operator P is positive. Moreover, show that ∀v ∈ E

(Pv, v) = ||Pv||2 .

23. Prove that the sum P = P1 + P2 of two projections P1 and P2 is a projection operator if
and only if P1 and P2 are orthogonal.
24. Prove that the product P = P1 P2 of two projections P1 and P2 is a projection operator if
and only if P1 and P2 commute.
25. Find the eigenvalues of a projection operator.
26. Prove that the span of all eigenvectors corresponding to the eigenvalue λ of an operator A
is a vector space.

vecanal4.tex; June 30, 2005; 19:16; p. 27


26 CHAPTER 1. LINEAR ALGEBRA

27. Let
E(λ) = Ker (A − λI) .
Show that: a) if λ is not an eigenvalue of A, then E(λ) = ∅, and b) if λ is an eigenvalue of
A, then E(λ) is the eigenspace corresponding to the eigenvalue λ.
28. Show that the operator A − λI is invertible if and only if λ is not an eigenvalue of the
operator A.
29. Let T be a unitary operator. Then the operators A and

à = TAT−1

are called similar. Show that the eigenvalues of similar operators are the same.
30. Show that an operator similar to a selfadjoint operator is selfadjoint and an operator sim-
ilar to an anti-selfadjoint operator is anti-selfadjoint.
31. Show that all eigenvalues of a positive operator A are non-negative.
32. Show that the eigenvectors corresponding to distinct eigenvalues of a unitary operator are
orthogonal to each other.
33. Show that the eigenvectors corresponding to distinct eigenvalues of a selfadjoint operator
are orthogonal to each other.
34. Show that all eigenvalues of a unitary operator A have absolute value equal to 1.
35. Show that if A is a projection, then it can only have two eigenvalues: 1 and 0.

vecanal4.tex; June 30, 2005; 19:16; p. 28


Chapter 2

Vector and Tensor Algebra

2.1 Metric Tensor


• Let E be a Euclidean space and {ei } = {e1 , . . . en } be a basis (not necessarily
orthonormal). Then each vector v ∈ E can be represented as a linear combination
X
n
v= vi ei ,
i=1

where vi , i = 1, . . . , n, are the components of the vector v with respect to the


basis {ei } (or contravariant components of the vector v). We stress once again
that contravariant components of vectors are denoted by upper indices (super-
scripts).
• Let G = (gi j ) be a matrix whose entries are defined by

gi j = (ei , e j ) .

These numbers are called the components of the metric tensor with respect to
the basis {ei } (also called covariant components of the metric).
• Notice that the matrix G is symmetric, that is,

gi j = g ji , GT = G.

• Theorem 2.1.1 The matrix G is invertible and

det G > 0 .

• The elements of the inverse matrix G−1 = (gi j ) are called the contravariant
components of the metric. They satisfy the equations
X
n
gi j g jk = δij ,
j=1

27
28 CHAPTER 2. VECTOR AND TENSOR ALGEBRA

where δij is the Kronecker symbol defined by




 1, if i = j

δij =

 0, if i , j

• Since the inverse of a symmetric matrix is symmetric, we have

gi j = g ji .

• In orthonormal basis

gi j = gi j = δi j , G = G−1 = I .

• Let v ∈ E be a vector. The real numbers

vi = (ei , v)

are called the covariant components of the vector v. Notice that covariant com-
ponents of vectors are denoted by lower indices (subscripts).
• Theorem 2.1.2 Let v ∈ E be a vector. The covariant and the contravariant
components of v are related by
X
n X
n
vi = gi j v j , vi = gi j v j .
j=1 j=1

• Theorem 2.1.3 The metric determines the inner product and the norm by
X
n X
n X
n X
n
(u, v) = gi j ui v j = gi j ui v j .
j=1 i=1 j=1 i=1

X
n X
n X
n X
n
||v||2 = gi j vi v j = gi j vi v j .
i=1 j=1 j=1 i=1

vecanal4.tex; June 30, 2005; 19:16; p. 29


2.2. DUAL SPACE AND COVECTORS 29

2.2 Dual Space and Covectors


• A linear mapping ω : E → R that assigns to a vector v ∈ E a real number hω, vi
and satisfies the condition: ∀u, v ∈ E, ∀a ∈ R

hω, u + vi = hω, ui + hω, vi , and hω, avi = ahω, vi ,

is called a linear functional.


• The space of linear functionals is a vector space, called the dual space of E
and denoted by E ∗ , with the addition and multiplication by scalars defined by:
∀ω, σ ∈ E ∗ , ∀v ∈ E, ∀a ∈ R,

hω + σ, vi = hω, vi + hσ, vi , and haω, vi = ahω, vi .

The elements of the dual space E ∗ are also called covectors or 1-forms. In
keeping with tradition we will denote covectors by Greek letters.
• Theorem 2.2.1 The dual space E ∗ of a real vector space E is a real vector space
of the same dimension.
• Let {ei } = {e1 , . . . , en } be a basis in E. A basis {ωi } = {ω1 , . . . ωn } in E ∗ such that

hωi , e j i = δij

is called the dual basis.


• The dual {ωi } of an orthonormal basis is also orthonormal.
• Given a dual basis {ωi } every covector σ in E ∗ can be represented in a unique
way as
Xn
σ= σi ωi ,
i=1
where the real numbers (σi ) = (σ1 , . . . , σn ) are called the components of the
covector σ with respect to the basis {ωi }.
• The advantage of using the dual basis is that it allows one to compute the com-
ponents of a vector v and a covector σ by

vi = hωi , vi

and
σi = hσ, ei i .
That is
X
n
v= ei hωi , vi
i=1
and
X
n
σ= hσ, ei iωi .
i=1

vecanal4.tex; June 30, 2005; 19:16; p. 30


30 CHAPTER 2. VECTOR AND TENSOR ALGEBRA

• More generally, the action of a covector σ on a vector v has the form


X
n X
n
hσ, vi = hσ, ei ihωi , vi = σi vi .
i=1 i=1

• The existense of the metric allows us to define the following map

g : E → E∗

that assigns to each vector v a covector g(v) such that

hg(v), ui = (v, u) .

Then
X
n
g(v) = (v, ei )ωi .
i=1

In particular,
X
n
g(ek ) = gki ωi .
i=1

• Let v be a vector and σ be the corresponding covector, so σ = g(v) and v =


g−1 (σ). Then their components are related by
X
n X
n
σi = gi j v j , vi = gi j σ j .
j=1 j=1

• The inverse map g−1 : E ∗ → E that assigns to each covector σ a vector g−1 (σ)
such that
hσ, ui = (g−1 (σ), u) ,
can be defined as follows. First, we define
X
n
−1 k
g (ω ) = gki ei .
i=1

Then
X
n X
n
−1
g (σ) = hσ, ek igki ei .
k=1 i=1

• The inner product on the dual space E ∗ is defined so that for any two covectors
α and σ
(α, σ) = hα, g−1 (σ)i = (g−1 (α), g−1 (α)) .

• This definition leads to


gi j = (ωi , ω j ) .

vecanal4.tex; June 30, 2005; 19:16; p. 31


2.2. DUAL SPACE AND COVECTORS 31

• Theorem 2.2.2 The inner product on the dual space E ∗ is determined by


X
n X
n
(α, σ) = gi j αi σ j .
i=1 j=1

In particular,
X
n
(ωi , σ) = gi j σ j
j=1

• The inverse map g−1 can be defined in terms of the inner product of covectors as
X
n
g−1 (σ) = ei (ωi , σ) .
i=1

• Since there is a one-to-one correspondence between vectors and covectors, we


can treat a vector v and the corresponding covector g(v) as a single object and
denote the components vi of the vector v and the components of the covector
g(v) by the same letter, that is,
X
n X
n
vi = gi j v j , vi = gi j v j .
j=1 j=1

• We call vi the contravaraint components and vi the covariant components.


This operation is called raising and lowering an index; we use gi j to raise an
index and gi j to lower an index.

2.2.1 Einstein Summation Convention


• In many equations of vector and tensor calculus summation over components of
vectors, covectors and, more generally, tensors, with respect to a given basis fre-
quently appear. Such a summation usually occurs on a pair of equal indices, one
lower index and one upper index, and one sums over all values of indices from 1
P
to n. The number of summation symbols ni=1 is equal to the number of pairs of
repeated indices. That is why even simple equations become cumbersome and
uncomfortable to work with. This lead Einstein to drop all summation signs and
to adopt the following summation convention:

1. In any expression there are two types of indices: free indices and repeated
indices.
2. Free indices appear only once in an expression; they are assumed to take
all possible values from 1 to n. For example, in the expression

gi j v j

the index i is a free index.

vecanal4.tex; June 30, 2005; 19:16; p. 32


32 CHAPTER 2. VECTOR AND TENSOR ALGEBRA

3. The position of all free indices in all terms in an equation must be the same.
For example,
gi j v j + α i = σ i
is a correct equation, while the equation

gi j v j + α i = σ i

is a wrong equation.
4. Repeated indices appear twice in an expression. It is assumed that there is a
summation over each repeated pair of indices from 1 to n. The summation
over a pair of repeated indices in an expression is called the contraction.
For example, in the expression

gi j v j

the index j is a repeated index. It actually means


X
n
gi j v j .
j=1

This is the result of the contraction of the indices k and l in the expression

gik vl .

5. Repeated indices are dummy indices: they can be replaced by any other
letter (not already used in the expression) without changing the meaning of
the expression. For example

gi j v j = gik vk

just means
X
n
gi j v j = gi1 v1 + · · · + gin vn ,
j=1

no matter how the repeated index is called.


6. Indices cannot be repeated on the same level. That is, in a pair of repeated
indices one index is in upper position and another is in the lower position.
For example,
vi vi
is a wrong expression.
7. There cannot be indices occuring three or more times in any expression.
For example, the expression
gii vi
does not make sense.

• From now on we will use the Einstein summation convention. We will say that
an equation is written in tensor notation.

vecanal4.tex; June 30, 2005; 19:16; p. 33


2.2. DUAL SPACE AND COVECTORS 33

Examples
• First, we list the equations we already obtained above

vi = gi j v j , v j = g ji vi ,

(u, v) = gi j ui v j = ui vi = ui vi = gi j ui v j ,
(α, β) = gi j αi β j = αi βi = αi βi = gi j αi β j .
gi j g jk = δki .

• A contraction of indices one of which belongs to the Kronecker symbol just


renames the index. For example:

δij v j = vi , δij δkj = δik ,

etc.

• The contraction of the Kronecker symbol gives


X
n
δii = 1 = n.
i=1

vecanal4.tex; June 30, 2005; 19:16; p. 34


34 CHAPTER 2. VECTOR AND TENSOR ALGEBRA

2.3 General Definition of a Tensor


• It should be realized that a vector is an invariant geometric object that does not
depend on the basis; it exists by itself independently of the basis. The basis is just
a convenient tool to represent vectors by its components. The componenets of a
vector do depend on the basis. It is this transformation law of the components of
a vector (and, more generally, a tensor as we will see later) that makes an n-tuple
of real numbers (v1 , . . . , vn ) a vector. Not every collection of n real numbers is
a vector. To represent a vector, a geometric object that does not depend on the
basis, these numbers should transform according to a very special rule under a
change of basis.
• Let {ei } = {e1 , . . . , en } and {e0j } = {e01 , . . . , e0n } be two different bases in E. Obvi-
ously, the vectors from one basis can be decomposed as a linear combination of
vectors from another basis, that is

ei = Λ j i e0j ,

where Λ j i , i, j = 1, . . . , n, is a set of n2 real numbers, forming the transforma-


tion matrix Λ = (Λi j ). Of course, we also have the inverse transformation

e0j = Λ̃k j ek

where Λ̃k j , k, j = 1, . . . , n, is another set of n2 real numbers.


• The dual bases {ωi } and {ω0 j } are related by

ω0i = Λi j ω j , ωi = Λ̃i j ω0 j .

• By using the second equation in the first and vice versa we obtain

ei = Λ̃k j Λ j i ek , e0j = Λi k Λ̃k j e0i ,

which means that


Λ̃k j Λ j i = δki , Λi k Λ̃k j = δij .
Thus, the matrix Λ̃ = (Λ̃i j ) is the inverse transformation matrix.
• In matrix notation this becomes

Λ̃Λ = I, ΛΛ̃ = I ,

which means that the matrix Λ is invertible and

Λ̃ = Λ−1 .

• The components of a vector v with respect to the basis {e0i } are

v0i = hω0i , vi = Λi j hω j , vi .

vecanal4.tex; June 30, 2005; 19:16; p. 35


2.3. GENERAL DEFINITION OF A TENSOR 35

This immediately gives


v0i = Λi j v j .
This is the transformation law of contravariant components. It is easy to
recognize this as the action of the transformation matrix on the column-vector of
the vector components from the left.

• We can compute the transformation law of the components of a covector σ as


follows
σ0i = hσ, e0i i = Λ̃ j i hσ, e j i ,
which gives
σ0i = Λ̃ j i σ j .
This is the transformation law of covariant components. It is the action of
the inverse transformation matrix on the row-vector from the right. That is, the
components of covectors are transformed with the transpose of the inverse trans-
formation matrix!

• Now let us compute the transformation law of the covariant components of


the metric tensor gi j . By the definition we have

g0i j = (e0i , e0j ) = Λ̃k i Λ̃l j (ek , el ) .

This leads to
g0i j = Λ̃k i Λ̃l j gkl .

• Similarly, the contravariant components of the metric tensor gi j transform


according to
g0i j = Λi k Λ j l gkl .

• The transformation law of the metric components in matrix notation reads

G0 = (Λ−1 )T GΛ−1

and
G0−1 = ΛG−1 ΛT .

• We denote the determinant of the covariant metric components G = (gi j ) by

|g| = det G = det(gi j ) .

• Taking the determinant of this equation we obtain the transformation law of


the determinant of the metric

|g0 | = (det Λ)−2 |g| .

vecanal4.tex; June 30, 2005; 19:16; p. 36


36 CHAPTER 2. VECTOR AND TENSOR ALGEBRA

• More generally, a set of real numbers T i1 ...i p j1 ... jq is said to represent components
of a tensor of type (p, q) (p times contravariant and q times covariant) if they
transform under a change of the basis according to
0i ...i l ...l
T j11... jqp = Λi1 l1 · · · Λi p l p Λ̃m1 j1 · · · Λ̃mq jq T m11 ...m
p
q
.

This is the general transformation law of the components of a tensor of type


(p, q).

• The rank of the tensor of type (p, q) is the number (p + q).

• A tensor product of a tensor A of type (p, q) and a tensor B of type (r, s) is a


tensor A ⊗ B of type (p + r, q + s) with components
i ...i l ...l i ...i
(A ⊗ B) j11 ... qjq 1k1 ...kr s = A j11 ... jpq Blk11...l
...k s .
r

• The symmetrization of a tensor of type (0, k) with components Ai1 ...ik is another
tensor of the same type with components

1 X
A(i1 ...ik ) = Ai ...i ,
k! ϕ∈S ϕ(1) ϕ(k)
k

where summation goes over all permutations of k indices. The symmetrization


is denoted by parenthesis.

• The antisymmetrization of a tensor of type (0, k) with components Ai1 ...ik is


another tensor of the same type with components

1 X
A[i1 ...ik ] = sign (ϕ)Aiϕ(1) ...iϕ(k) ,
k! ϕ∈S
k

where summation goes over all permutations of k indices. The antisymmetriza-


tion is denoted by square brackets.

• A tensor Ai1 ...ik is symmetric if

A(i1 ...ik ) = Ai1 ...ik

and anti-symmetric if
A[i1 ...ik ] = Ai1 ...ik .

• Anti-symmetric tensors of type (0, p) are called p-forms.

• Anti-symmetric tensors of type (p, 0) are called p-vectors.

• A tensor is isotropic if it is a tensor product of gi j , gi j and δij .

• Every isotropic tensor has an even rank.

vecanal4.tex; June 30, 2005; 19:16; p. 37


2.3. GENERAL DEFINITION OF A TENSOR 37

• For example, the most general isotropic tensor of rank two is

Ai j = aδij ,

where a is a scalar, and the most general isotropic tensor of rank four is

Ai j kl = agi j gkl + bδik δlj + cδil δkj ,

where a, b, c are scalars.

2.3.1 Orientation, Pseudotensors and Volume


• Since the transformation matrix Λ is invertible, then the determinant det Λ is
either positive or negative. If det Λ > 0 then we say that the bases {ei } and {e0i }
have the same orientation, and if det Λ < 0 then we say that the bases {ei } and
{e0i } have the opposite orientation.
• This defines an equivalence relation on the set of all bases on E called the ori-
entation of the vector space E. This equivalence relation divides the set of all
bases in two equivalence classes, called the positively oriented and negatively
oriented bases.
• A vector space together with a choice of what equivalence class is positively
oriented is called an oriented vector space.
i ...i
• A set of real numbers A j11 ... jpq is said to represent components of a pseudo-tensor
of type (p, q) if they transform under a change of the basis according to
0i ...i l ...l
A j11... jqp = sign(det Λ)Λi1 l1 · · · Λi p l p Λ̃m1 j1 · · · Λ̃mq jq Am1 1 ...m
p
q
,

where sign(x) = +1 if x > 0 and sign(x) = −1 if x < 0.


• The Levi-Civita symbol (also called alternating symbol) is defined by



 +1, if (i1 , . . . , in ) is an even permutation of (1, . . . , n),



i1 ...in
εi1 ...in = ε =
 −1, if (i1 , . . . , in ) is an odd permutation of (1, . . . , n),



 0, if two or more indices are the same .

• The Levi-Civita symbols εi1 ...in and εi1 ...in do not represent tensors! They have the
same values in all bases.
• Theorem 2.3.1 The determinant of a matrix A = (Ai j ) can be written as

det A = εi1 ...in A1 i1 . . . An in


= ε j1 ... jn A j1 1 . . . A jn n
1 i1 ...in
= ε ε j1 ... jn A j1 i1 . . . A jn in .
n!
Here, as usual, a summation over all repeated indices is assumed from 1 to n.

vecanal4.tex; June 30, 2005; 19:16; p. 38


38 CHAPTER 2. VECTOR AND TENSOR ALGEBRA

• Theorem 2.3.2 There holds the identity


X
εi1 ...in ε j1 ... jn = sign(ϕ) δij1ϕ(1) · · · δijnϕ(n)
ϕ∈S n

= n!δi[1j1 · · · δijnn ] .

The contraction of this identity over k indices gives

εi1 ...in−k m1 ...mk ε j1 ... jn−k m1 ...mk = k!(n − k)!δi[1j1 · · · δijn−k


n−k ]
.

In particular,
εm1 ...mn εm1 ...mn = n! .

• Theorem 2.3.3 The sets of real numbers Ei1 ...in and E i1 ...in defined by
p
Ei1 ...in = |g| εi1 ...in

1
E i1 ...in = p εi1 ...in ,
|g|
where |g| = det(gi j ), define (pseudo)-tensors of type (0, n) and (n, 0) respectively.

• Let {v1 , . . . , vn } be an ordered n-tuple of vectors. The volume of the paral-


lelepiped spanned by the vectors {v1 , . . . , vn } is a real number defined by
q
|vol (v1 , . . . , vn )| = det((vi , v j )) .

• Theorem 2.3.4 Let {ei } be a basis in E, {ωi } be the dual basis, and {v1 , . . . , vn }
be a set of n vectors. Let V = (vi j ) be the matrix of contravariant components of
the vectors {v j }
vi j = hωi , v j i,
and W = (vi j ) be the matrix of covariant components of the vectors {v j }

vi j = (ei , v j ) = gik vk j .

Then the volume of the parallelepiped spanned by the vectors {v1 , . . . , vn } is


p | det W|
|vol (v1 , . . . , vn )| = |g| | det V| = p .
|g|

• If the vectors {v1 , . . . , vn } are linearly dependent, then

vol (v1 , . . . , vn ) = 0 .

• If the vectors {v1 , . . . , vn } are linearly independent, then the volume is a positive
real number that does not depend on the orientation of the vectors.

vecanal4.tex; June 30, 2005; 19:16; p. 39


2.3. GENERAL DEFINITION OF A TENSOR 39

• The signed volume of the parallelepiped spanned by an ordered n-tuple of


vectors {v1 , . . . , vn } is
p
vol (v1 , . . . , vn ) = |g| det V
= sign(v1 , . . . , vn ) |vol (v1 , . . . , vn )| , .

The sign of the signed volume depends on the orientation of the vectors {v1 , . . . , vn }:


 +1, if {v1 , . . . , vn } is positively oriented

sign(v1 , . . . , vn ) = sign(det V) = 

 −1, if {v1 , . . . , vn } is negatively oriented

• Theorem 2.3.5 The signed volume is equal to

vol (v1 , . . . , vn ) = Ei1 ...in vi1 1 · · · vin n = E i1 ...in vi1 1 · · · vin n ,

where vi j = hωi , v j i and vi j = (ei , v j ).


That is why the pseudo-tensor Ei1 ...in is also called the volume form.

Exterior Product and Duality


• The volume form allows one to define the duality of k-forms and (n − k)-vectors
as follows. For each k-form Ai1 ...ik one assigns the dual (n − k)-vector by

1 j1 ... jn−k i1 ...ik


∗A j1 ... jn−k = E Ai1 ...ik .
k!
Similarly, for each k-vector Ai1 ...ik one assigns the dual (n − k)-form

1
∗A j1 ... jn−k = E j ... j i ...i Ai1 ...ik .
k! 1 n−k 1 k

• Theorem 2.3.6 For each k-form α there holds

∗ ∗ α = (−1)k(n−k) α .

That is,
∗∗ = (−1)k(n−k) .

• The exterior product of a k-form A and a m-form B is a (k + m)-form A ∧ B


defined by
(k + m)!
(A ∧ B)i1 ...ik j1 ... jm = A[i1 ...ik B j1 ... jm ]
k!m!
Similarly, one can define the exterior product of p-vectors.

• Theorem 2.3.7 The exterior product is associative, that is,

(A ∧ B) ∧ C = A ∧ (B ∧ C) .

vecanal4.tex; June 30, 2005; 19:16; p. 40


40 CHAPTER 2. VECTOR AND TENSOR ALGEBRA

• A collection {v1 , . . . , vn−1 } of (n − 1) vectors defines a covector α by

α = ∗(v1 ∧ · · · ∧ vn−1 )

or, in components,
α j = E ji1 ...in−1 vi1 1 · · · vin−1 n−1 .

• Theorem 2.3.8 Let {v1 , . . . , vn−1 } be a collection of (n − 1) vectors and S =


span {v1 , . . . vn−1 } be the hyperplane spanned by these vectors. Let e be a unit
vector orthogonal to S oriented in such a way that {v1 , . . . , vn−1 , e} is oriented
positively. Then the vector u = g−1 (α) corresponding to the 1-form α = ∗(v1 ∧
· · · ∧ vn−1 ) is parallel to e (with the same orientation)

u = e ||u||

and has the norm


||u|| = vol (v1 , . . . , vn−1 , e) .

• In three dimensions, i.e. when n = 3, this defines a binary operation ×, called


the vector product, that is

u = v × w = ∗(v ∧ w) ,

or p
u j = E jik vi wk = |g| ε jik vi wk .

vecanal4.tex; June 30, 2005; 19:16; p. 41


2.4. OPERATORS AND TENSORS 41

2.4 Operators and Tensors


• Let A be an operator on E. Let {ei } = {e1 , . . . , en } be a basis in a Euclidean
space and {ωi } = {ω1 , . . . , ωn } be the dual basis in E ∗ . The real square matrix
A = (Ai j ), i, j = 1, . . . , n, defined by

Ae j = Ai j ei ,

is called the matrix of the operator A.


• Therefore, there is a one-to-one correspondence between the operators on E and
the real square n × n matrices A = (Ai j ).
• It can be computed by

Ai j = hωi , Ae j i = gik (ek , Ae j ) .

• Remark. Notice that the upper index, which is the first one, indicates the row
and the lower index is the second one indicating the column of the matrix. The
convenience of this notation comes from the fact that all upper indices (also
called contravariant indices) indicate the components of vectors and “belong”
to the vector space E while all lower indices (called covariant indices) indicate
components of covectors and “belong” to the dual space E ∗ .
• The matrix of the identity operator I is

I i j = δij .

• For any v ∈ E
v = v je j, v j = hω j , vi
we have
Av = Ai j v j ei .
That is, the components ui of the vector u = Av are given by

ui = Ai j v j .

Transformation Law of Matrix of an Operator


• Under a change of the basis ei = Λ j i e0j , the matrix Ai j of an operator A transforms
according to
A0i j = Λi k Ak m Λ̃m j ,
which in matrix notation reads

A0 = ΛAΛ−1 .

• Therefore, the matrix A = (Ai j ) of an operator A represents the components of


a tensor of type (1, 1). Conversely, such tensors naturally define linear operators
on E. Thus, linear operators on E and tensors of type (1, 1) can be identified.

vecanal4.tex; June 30, 2005; 19:16; p. 42


42 CHAPTER 2. VECTOR AND TENSOR ALGEBRA

• The determinant and the trace of the matrix of an operator are invariant under the
change of the basis, that is

det A0 = det A , tr A0 = tr A .

• Therefore, one can define the determinant of the operator A and the trace of
the operator A by the determinant and the trace of its matrix, that is,

det A = det A , tr A = tr A .

• For self-adjoint operators these definitions are consistent with the definition in
terms of the eigenvalues given before.

• The matrix of the sum A + B of two operators A and B is the sum of matrices A
and B of the operatos A and B.

• The matrix of a scalar multiple cA is equal to cA, where A is the matrix of the
operator A and c ∈ R.

Matrix of the Product of Operators


• The matrix of the product C = AB of two operators reads

C i j = hωi , ABe j i = hωi , Aek ihωk , Be j i = Ai k Bk j ,

which is exactly the product of matrices A and B.

• Thus, the matrix of the product AB of the operators A and B is equal to the
product AB of matrices of these operators in the same order.

• The matrix of the inverse A−1 of an invertible operator A is equal to the inverse
A−1 of the matrix A of the operator A.

• Theorem 2.4.1 The algebra L(E) of linear operators on E is isomorphic to the


algebra Mat(n, R) of real square n × n matrices.

Matrix of the Adjoint Operator


• For the adjoint operator A∗ we have

(ei , A∗ e j ) = (Aei , e j ) = (e j , Aei ) .

Therefore, the matrix of the adjoint operator is

A∗k j = gki (ei , A∗ e j ) = gki Al i gl j .

In matrix notation this reads

A∗ = G−1 AT G .

vecanal4.tex; June 30, 2005; 19:16; p. 43


2.4. OPERATORS AND TENSORS 43

• Thus, the matrix of a self-adjoint operator A satisfies the equation

Ak j = gki Al i gl j or gik Ak j = Al i gl j ,

which in matrix notation reads

A = G−1 AT G , or GA = AT G .

• The matrix of a unitary operator A satisfies the equation

gki Al i gl j A j m = δkm or Al i gl j A j m = gim ,

which in matrix notation has the form

G−1 AT GA = I , or AT GA = G .

vecanal4.tex; June 30, 2005; 19:16; p. 44


44 CHAPTER 2. VECTOR AND TENSOR ALGEBRA

2.5 Vector Algebra in R3


• We denote the standard orthonormal basis in R3 by
e1 = i , e2 = j , e3 = k ,
so that
ei · e j = δi j .
• Each vector v is decomposed as
v = v1 i + v2 j + v3 k .
The components are computed by
v1 = v · i , v2 = v · j , v3 = v · k .

• The norm of the vector q


||v|| = v21 + v22 + v23 .

• Scalar product is defined by


v · u = v1 u1 + v2 u2 + v3 u3 .

• The angle between vectors


u·v
cos θ = .
||u|| ||v||
• The orthogonal decomposition of a vector v with respect to a given unit vector u
is
v = vk + v⊥ ,
where
vk = u(u · v) , v⊥ = v − u(u · v) .
• We denote the Cartesian coordinates in R3 by
x1 = x , x2 = y , x3 = z .
The radius vector (the position vector) is
r = xi + yj + zk .

• The parametric equation of a line parallel to a vector u = a i + b j + c k is


r = r 0 + tu ,
where r 0 = x0 i + y0 j + z0 k is a fixed vector and t is a real parameter. In
components,
x = x0 + at , y = y0 + bt , z = z0 + ct .
The non-parametric equation of a line (if a, b, c are non-zero) is
x − x0 y − y0 z − z0
= = .
a b c

vecanal4.tex; June 30, 2005; 19:16; p. 45


2.5. VECTOR ALGEBRA IN R3 45

• The parametric equation of a plane spanned by two non-parallel vectors u and


v is
r = r 0 + tu + sv ,
where t and s are real parameters.
• A vector n that is perpendicular to both vectors u and v is normal to the plane.
• The non-parametric equation of a plane with the normal n = a i + b j + c k is

( r − r 0) · n = 0

or
a(x − x0 ) + b(y − y0 ) + (z − z0 ) = 0 ,
which can also be written as

ax + by + cz = d ,

where
d = ax0 + by0 + cz0 .

• The positive (right-handed) orientation of a plane is defined by the right hand


(or counterclockwise) rule. That is, if u1 and u2 span a plane then we orient the
plane by saying which vector is the first and which is the second. The orientation
is positive if the rotation from u1 to u2 is counterclockwise and negative if it is
clockwise. A plane has two sides. The positive side of the plane is the side with
the positive orientation, the other side has the negative (left-handed) orientation.
• The vector product of two vectors is defined by

i j k
w = u × v = det u1 u2 u3 ,
v1 v2 v3

or, in components,
1 i jk
wi = εi jk u j vk = ε (u j vk − uk v j ) .
2

• The vector products of the basis vectors are

ei × e j = εi jk ek .

• If u and v are two nonzero nonparallel vectors, then the vector w = u × v is


orthogonal to both vectors, u and v, and, hence, to the plane spanned by these
vectors. It defines a normal to this plane.
• The area of the parallelogram spanned by two vectors u and v is

area(u, v) = |u × v| = ||u|| ||v|| sin θ .

vecanal4.tex; June 30, 2005; 19:16; p. 46


46 CHAPTER 2. VECTOR AND TENSOR ALGEBRA

• The signed volume of the parallelepiped spanned by three vectors u, v and w


is
u1 u2 u3
vol(u, v, w) = u · (v, w) = det v1 v2 v3 = εi jk ui v j wk .
w1 w2 w3
The signed volume is also called the scalar triple product and denoted by

[u, v, w] = u · (v × w) .

• The signed volume is zero if and only if the vectors are linearly dependent, that
is, coplanar.

• For linearly independent vectors its sign depends on the orientation of the triple
of vectors {u, vw}

vol(u, v, w) = sign (u, v, w)|vol(u, v, w)| ,

where (
1 if {u, v, w} is positively oriented
sign (u, v, w) =
−1 if {u, v, w} is negatively oriented

• The scalar triple product is linear in each argument, anti-symmetric

[u, v, w] = −[v, u, w] = −[u, w, v] = −[w, v, u]

cyclic
[u, v, w] = [v, w, u] = [w, u, v] .
It is normalized so that
[i, j, k] = 1.

• The orthogonal decomposition of a vector v with respect to a unit vector u can


be written in the form

v = u(u · v) − u × (u × v) .

• The Levi-Civita symbol in three dimensions





 +1 if (i, j, k) = (1, 2, 3), (2, 3, 1), (3, 1, 2)



εi jk = εi jk =
 −1 if (i, j, k) = (2, 1, 3), (3, 2, 1), (1, 3, 2)



 0 otherwise

has the following properties:

εi jk = −ε jik = −εik j = −εk ji

εi jk = ε jki = εki j

vecanal4.tex; June 30, 2005; 19:16; p. 47


2.5. VECTOR ALGEBRA IN R3 47

εi jk εmnl = 6δm n l
[i δ j δk]

= δm n l m n l m n l m n l m n l m n l
i δ j δk + δ j δk δi + δk δi δ j − δi δk δ j − δ j δi δk − δk δ j δi

εi jk εmnk = 2δm n m n m n
[i δ j] = δi δ j − δ j δi

εi jk εm jk = 2δm
i
εi jk εi jk = 6

• This leads to many vector identities that express double vector product in terms
of scalar product. For example,

u × (v × w) = (u · w)v − (u · v)w
u × (v × w) + v × (w × u) + w × (u × v) = 0
(u × v) × (w × n) = v[u, w, n] − u[v, w, n]
(u × v) · (w × n) = (u · w)(v · n) − (u · n)(v · w)

vecanal4.tex; June 30, 2005; 19:16; p. 48


48 CHAPTER 2. VECTOR AND TENSOR ALGEBRA

vecanal4.tex; June 30, 2005; 19:16; p. 49


Chapter 3

Geometry

3.1 Geometry of Euclidean Space


• The set Rn can be viewed geometrically as a set of points, that we will denote
by P, Q, etc. With each point P we associate an ordered n-tuple of real numbers
(xiP ) = (x1P , . . . , xnP ), called the coordinates of the point P. The assignment of
n-tuples of real numbers to the points in space should be bijective. That is,
different points are assigned different n-tuples, and for every n-tuple there is a
point in space with such coordinates. Such a map is called a coordinate system.

• A space Rn with a coordinate system is a Euclidean space if the distance be-


tween any two points P and Q is determined by
v
t n
X
d(P, Q) = (xiP − xiQ )2 .
i=1

Such coordinate system is called Cartesian.

• The point O with the zero coordinates (0, . . . , 0) is called the origin of the Carte-
sian coordinate system.

• In Rn it is convenient to associate vectors with points in space. With each point


P with Cartesian coordinates (x1 , . . . , xn ) in Rn we associate the column-vector
r = (xi ) with the components equal to the Cartesian coordinates of the point P.
We say that this vector points from the origin O to the point P; it has its tail
at the point O and its tip at the point P. This vector is often called the radius
vector, or the position vector, of the point P and denoted by r .

• Similarly, with every two points P and Q with the coordinates (xiP ) and (xiQ ) we
associate the vector
uPQ = r Q − r P = (xiQ − xiP )
that points from the point P to the point Q.

49
50 CHAPTER 3. GEOMETRY

• Obviously, the Euclidean distance is given by

d(P, Q) = || r P − r Q || .

• The standard (orthonormal) basis {e1 , . . . , en } of Rn are the unit vectors that con-
nect the origin O with the points {(1, 0, . . . , 0), . . . , (0, . . . , 0, 1)} that have only
one nonzero coordinate which is equal to 1.

• The one-dimensional subspaces Li spanned by a single basis vector ei ,

Li = span {ei } = {P | r P = tei , t ∈ R},

are the lines called the coordinate axes. There are n coordinate axes; they are
mutually orthogonal and intersect at only one point, the origin O.

• The two-dimensional subspaces Pi j spanned by a couple of basis vectors ei and


e j,
Pi j = span {ei , e j } = {P | r P = tei + se j , t, s ∈ R},
are the planes called the coordinate planes. There are n(n − 1)/2 coordinate
planes; the coordinate planes are mutually orthogonal and intersect along the
coordinate axes.

• Let a and b be real numbers such that a < b. The set [a, b] is a closed interval
in R. A parametrized curve C in Rn is a map C : [a, b] → Rn which assigns a
point in Rn
C : r (t) = (xi (t))
to each real number t ∈ [a, b].

• The positive orientation of the curve C is determined by the standard orientation


of R, that is, by the direction of increasing values of the parameter t.

• The point r (a) is the initial point and the point r (b) is the endpoint of the
curve.

• The curve (−C) is the parametrized curve with the opposite orientation. If the
curve C is parametrized by r (t), a ≤ t ≤ b, then the curve (−C) is parametrized
by
(−C) : r (−t + a + b) .

• The boundary ∂C of the curve C consists of two points C0 and C1 correspond-


ing to r (a) and r (b), that is,

∂C = C1 − C0 .

• A curve C is continuous if all the functions xi (t) are continuous for any t on
[a, b].

vecanal4.tex; June 30, 2005; 19:16; p. 50


3.1. GEOMETRY OF EUCLIDEAN SPACE 51

• Let a1 , a2 and b1 , b2 be real numbers such that


a1 < b1 , and a2 < b2 .
The set D = [a1 , b1 ] × [a2 , b2 ] is a closed rectangle in the plane R2 .
• A parametrized surface S in Rn is a map S : D → Rn which assigns a point
S : r (u) = (xi (u))
in Rn to each point u = (u1 , u2 ) in the rectangle D.
• The positive orientation of the surface S is determined by the positive orienta-
tion of the standard basis in R2 . The surface (−S ) is the surface with the opposite
orientation.
• The boundary ∂S of the surface S consists of four curves S (1),0 , S (1),1 , S (2),0 ,
and S (2),1 parametrized by r (a1 , v), r (b1 , v), r (u, a2 ) and r (u, b2 ) respectively.
Taking into account the orientation, the boundary of the surface S is
∂S = S (2),0 + S (1),1 − S (2),1 − S (1),0 .

• Let a1 , . . . , ak and b1 , . . . , bk be real numbers such that


ai < bi , i = 1, . . . , k.
The set D = [a1 , b1 ] × · · · × [ak , bk ] is called a closed k-rectangle in Rk . In
particular, the set [0, 1]k = [0, 1] × · · · × [0, 1] is the standard k-cube.
• Let D = [a1 , b1 ] × · · · × [ak , bk ] be a closed rectangle in Rk . A parametrized
k-dimensional surface S in Rn is a continuous map S : D → Rn which assigns
a point
S : r (u) = (xi (u))
in Rn to each point u = (u1 , . . . , uk ) in the rectangle D.
• A (n − 1)-dimensional surface is called the hypersurface. A non-parametrized
hypersurface can be described by a single equation
F(x) = F(x1 , . . . , xn ) = 0 ,
where F : Rn → R is a real-valued function of n coordinates.
• The boundary ∂S of S consists of (k − 1)-surfaces, S (i),0 and S (i),1 , i = 1, . . . , k,
called the faces of the k-surface S . Of course, a k-surface S has 2k faces. The
face S (i),0 is parametrized by
S (i),0 : r (u1 , . . . , ui−1 , ai , ui+1 , . . . , uk ) ,
where the i-th parameter ui is fixed at the initial point, i.e. ui = ai , and the face
S (i),0 is parametrized by
S (i),1 : r (u1 , . . . , ui−1 , bi , ui+1 , . . . , uk ) ,
where the i-th parameter ui is fixed at the endpoint, i.e. ui = bi .

vecanal4.tex; June 30, 2005; 19:16; p. 51


52 CHAPTER 3. GEOMETRY

• The boundary of the surface S is defined by

X
k
∂S = (−1)i (S (i),0 − S (i),1 ) .
i=1

• Let S 1 , . . . , S m be parametrized k-surfaces. A formal sum


X
m
S = ai S i
i=1

with integer coefficients a1 , . . . , am , is called a k-chain. Usually (but not always)


the integers ai are equal to 1, (−1) or 0.
• The product of any k-chain S with zero is called the zero chain

0S = 0 .

• The addition of k-chains and multiplication by integers is defined by


X
m X
m X
m
ai S i + bi S i = (ai + bi )S i ,
i=1 i=1 i=1
 m 
X  Xm
b  ai S i  = (bai )S i .
i=1 i=1

• The boundary of a k-chain S is an (k − 1)-chain ∂S defined by


 m 
X  Xm
∂  ai S i  = ai ∂S i .
i=1 i=1

• Theorem 3.1.1 For any k-chain S there holds

∂(∂S ) = 0 .

vecanal4.tex; June 30, 2005; 19:16; p. 52


3.2. BASIC TOPOLOGY OF RN 53

3.2 Basic Topology of Rn


• Let P0 be a point in a Euclidean space Rn and ε > 0 be a positive real number.
The open ball Bε (P0 ) or radius  with the center at P0 is the set of all points
whose distance from the point P0 is less than ε, that is,

Bε (P0 ) = {P | d(P, P0 ) < ε} .

• A neighborhood of a point P0 is any set that contains an open ball centered at


P0 .
• Let S be a subset of a Euclidean space Rn . A point P is an interior point of S if
there is a neighborhood of P that lies completely in S .

• A point P is an exterior point of S if there is a neighborhood of P that lies


completely outside of S .
• A point P is a boundary point of S if it is neither an interior nor an exterior
point. If P is a boundary point of S , then every neighborhood of P contains
points in S and points not in S .
• The set of boundary points of S is called the boundary of S , denoted by ∂S .

• The set of all interior points of S is called the interior of S , denoted by S o .


• A set S is called open if every point of S is an interior point of S , that is, S = S o .
• A set S is closed if it contains all its boundary points, that is, S = S o ∪ ∂S .

• Henceforth, we will consider only open sets and call them regions of space.
• A region S is called connected (or arc-wise connected) if for any two points P
and Q in S there is an arc joining P and Q that lies within S .
• A connected region, that is a connected open set, is called a domain.

• A domain S is said to be simply-connected if every closed curve lying within S


can be continuously deformed to a point in the domain without any part of the
curve passing through regions outside the domain.
• A domain is simply connected if for any closed curve lying in the domain there
can be found a surface within the domain that has that curve as its boundary.
• A domain is said to be star-shaped if there is a point P in the domain such that
for any other point in the domain the entire line segment joining these two points
lies in the domain.

vecanal4.tex; June 30, 2005; 19:16; p. 53


54 CHAPTER 3. GEOMETRY

3.3 Curvilinear Coordinate Systems


• We say that a function f (x) = f (x1 , . . . , xn ) is smooth if it has continuous partial
derivatives of all orders.

• Let P be a point with Cartesian coordinates (xi ). Suppose that we assign another
n-tuple of real numbers (qi ) = (q1 , . . . , qn ) to the point P, so that

xi = f i (q) ,

where f i (q) = f i (q1 , . . . , qn ) are smooth functions of the variables xi . We will


call this a change of coordinates.

• The matrix !
∂xi
J=
∂q j
is called the Jacobian matrix. The determinant of this matrix is called the Ja-
cobian.

• A point P0 at which the Jacobian matrix is invertible, that is, the Jacobian is not
zero, det J , 0, is called a nonsingular point of the new coordinate system (qi ).

• Theorem 3.3.1 (Inverse Function Theorem) In a neighborhood of any nonsin-


gular point P0 the change of coordinates is invertible.

That is, if xi = f i (q) and


!
∂xi
det , 0,
∂q j P0

then for all points sufficiently close to P0 there exist n smooth functions

qi = hi (x) = hi (x1 , . . . , xn ) ,

of the variables (xi ) such that

f i (h1 (x), . . . , hn (x)) = xi , hi ( f 1 (q), . . . , f n (q)) = qi .

The Jacobian matrix of the inverse transformation is the inverse matrix of the
Jacobian matrix of direct transformation, i.e.

∂xi ∂q j ∂qi ∂x j
= δik , and = δik .
∂q j ∂xk ∂x j ∂qk

• The curves Ci along which only one coordinate is varied, while all other are
fixed, are called the coordinate curves, that is,

xi = xi (q10 , . . . , qi−1 i i+1 n


0 , q , q0 , . . . , q0 )} .

vecanal4.tex; June 30, 2005; 19:16; p. 54


3.3. CURVILINEAR COORDINATE SYSTEMS 55

• The vectors
∂r
ei =
∂qi
are tangent to the coordinate curves.
• The surfaces S i j along which only two coordinates are varied, while all other are
fixed, are called the coordinate surfaces, that is,
j−1 j+1
xi = xi (q10 , . . . , qi−1 i i+1 j n
0 , q , q0 , . . . , q0 , q , q0 , . . . , q0 )} .

• Theorem 3.3.2 For each point P there are n coordinate curves that pass through
P. The set of tangent vectors {ei } to these coordinate curves is linearly indepen-
dent and forms a basis.
• The basis {ei } is not necessarily orthonormal.
• The metric tensor is defined as usual by
Xn
∂xk ∂xk
gi j = ei · e j = .
k=1
∂qi ∂q j

• The dual basis of 1-forms is defined by


∂qi j
ωi = dqi = σ ,
∂x j
where σ j is the standard dual basis.
• The vector
∂r
d r = ei dqi = dqi
∂qi
is called the infinitesimal displacement.
• The arc length, called the interval, is determined by

ds2 = ||d r ||2 = d r · d r = gi j dqi dq j .

• The volume of the parallelepiped spanned by the vectors {e1 dq1 , . . . , en dqn },
called the volume element, is
p
dV = |g| dq1 · · · dqn ,

where, as usual, |g| = det(gi j ).


• A coordinate system is called orthogonal if the vectors ∂ r /∂qi are mutually
orthogonal. The norms of these vectors
∂r
hi =
∂qi
are called the scale factors.

vecanal4.tex; June 30, 2005; 19:16; p. 55


56 CHAPTER 3. GEOMETRY

• Then one can introduce the orthonormal basis {ei } by


−1
∂r ∂r 1 ∂r
ei = = .
∂qi ∂q i hi ∂qi

• For an orthonormal system the vector components are (there is no difference


between contravariant and covariant components)

vi = vi = v · ei .

• Then the interval has the form


X
n
ds2 = h2i (dqi )2 .
i=1

• The volume element in orthogonal coordinate system is

dV = h1 · · · hn dq1 · · · dqn .

3.3.1 Change of Coordinates


• Let (qi ) and (q0 j) be two curvilinear coordinate systems. Then they should be
related by a smooth invertible transformation

q0i = f i (q) = f i (q1 , . . . , qn ) , qi = hi (q0 ) = f i (q01 , . . . , q0n ) ,

such that
f i (h(q0 )) = q0i , hi ( f (q)) = qi .

• The Jacobian matrices are related by

∂q0i ∂q j i ∂qi ∂q0 j


= δ , = δik ,
∂q j ∂q0k k
∂q0 j ∂qk
 0i   i
so that the matrix ∂q
∂q j is inverse to ∂q
∂q0 j
.

• The basis vectors in these coordinate systems are


∂r ∂r
e0i = , ei = .
∂q0i ∂qi
Therefore, they are related by a linear transformation

∂q j ∂q0i 0
e0i = ej , ej = e .
∂q0i ∂q j j
They have the same orientation if the Jacobian of the change of coordinates is
positive and oppositive orientation if the Jacobian is negative.

vecanal4.tex; June 30, 2005; 19:16; p. 56


3.3. CURVILINEAR COORDINATE SYSTEMS 57

• Thus, a set of real numbers T i1 ...i p j1 ... jq is said to represent components of a ten-
sor of type (p, q) (p times contravariant and q times covariant) if they transform
under a change of coordinates according to

0i ...i ∂q0i1 ∂q0i p ∂qm1 ∂qmq l1 ...l p


T j11... jqp = · · · · · · T m ...m .
∂ql1 ∂ql p ∂q0 j1 ∂q0 jq 1 q
This is the general transformation law of the components of a tensor of type
(p, q) with respect to a change of curvilinear coordinates.
• A pseudo-tensor has an additional factor equal to the sign of the Jacobian, that
is, the components of a pseudo-tensor of type (p, q) transform as
" !#
0i ...i ∂q0i1 ∂q0i1 ∂q0i p ∂qm1 ∂qmq l ...l p
T j11... jqp = sign det ··· l 0
· · · 0 j T m11 ...m .
∂q l 1 ∂ql 1 ∂q ∂q
p j 1 ∂q q q

This is the general transformation law of the components of a pseudo-tensor


of type (p, q) with respect to a change of curvilinear coordinates.

3.3.2 Examples
• The polar coordinates in R2 are introduced by

x1 = ρ cos ϕ , x2 = ρ sin ϕ ,

where ρ ≥ 0 and 0 ≤ ϕ < 2π. The Jacobian matrix is


!
cos ϕ −ρ sin ϕ
J= .
sin ϕ ρ cos ϕ
The Jacobian is
det J = ρ .
Thus, the only singular point of the polar coordinate system is the origin ρ = 0.
At all nonsingular points the change of variables is invertible and we have
p 1
! 2
!
−1 x −1 x
ρ = (x ) + (x ) ,
1 2 2 2 ϕ = cos = sin .
ρ ρ
The coordinate curves of ρ are half-lines (rays) going through origin with the
slope tan ϕ. The coordinate curves of ϕ are circles with the radius ρ centered at
the origin.
• The cylindrical coordinates in R3 are introduced by

x1 = ρ cos ϕ , x2 = ρ sin ϕ , x3 = z

where ρ ≥ 0, 0 ≤ ϕ < 2π and z ∈ R. The Jacobian matrix is


 
 cos ϕ −ρ sin ϕ 0 
 
J =  sin ϕ ρ cos ϕ 0  .
 
0 0 1

vecanal4.tex; June 30, 2005; 19:16; p. 57


58 CHAPTER 3. GEOMETRY

The Jacobian is
det J = ρ .
Thus, the only singular point of the cylindrical coordinate system is the origin
ρ = 0. At all nonsingular points the change of variables is invertible and we have
p ! !
x1 x2
ρ= (x1 )2 + (x2 )2 , ϕ = cos−1 = sin−1 , z = x3 .
ρ ρ

The coordinate curves of ρ are horizontal half-lines in the plane z = const going
through the z-axis. The coordinate curves of ϕ are circles in the plane z = const
of radius ρ centered at the z axis. The coordinate curves of z are vertical lines.
The coordinate surfaces of ρ, ϕ are horizontal planes. The coordinate surfaces of
ρ, z are vertical half-planes going through the z-axis. The coordinate surfaces of
ϕ, z are vertical cylinders centered at the origin.

• The spherical coordinates in R3 are introduced by

x1 = r sin θ cos ϕ , x2 = r sin θ sin ϕ , x3 = r cos θ

where r ≥ 0, 0 ≤ ϕ < 2π and 0 ≤ θ ≤ π. The Jacobian matrix is


 
 sin θ cos ϕ r cos θ cos ϕ −r sin θ sin ϕ 
 
J =  sin θ sin ϕ r cos θ sin ϕ r sin θ cos ϕ  .

cos θ −r sin θ 0

The Jacobian is
det J = r2 sin θ .
Thus, the singular points of the spherical coordinate system are the points where
either r = 0, which is the origin, or θ = 0 or θ = π, which is the whole z-axis. At
all nonsingular points the change of variables is invertible and we have
p
r = (x1 )2 + (x2 )2 + (x3 )2 ,
1
! 2
!
−1 x −1 x
ϕ = cos = sin ,
ρ ρ
!
x3
θ = cos−1 ,
r
p
where ρ = (x1 )2 + (x2 )2 .
The coordinate curves of r are half-lines going through the origin. The coordi-
nate curves of ϕ are circles of radius r sin θ centered at the z axis. The coordinate
curves of θ are vertical half-circles of radius r centered at the origin. The coor-
dinate surfaces of r, ϕ are half-cones around the z-axis going through the origin.
The coordinate surfaces of r, θ are vertical half-planes going through the z-axis.
The coordinates surfaces of ϕ, θ are spheres of radius r centered at the origin.

vecanal4.tex; June 30, 2005; 19:16; p. 58


3.4. VECTOR FUNCTIONS OF A SINGLE VARIABLE 59

3.4 Vector Functions of a Single Variable


• A vector-valued function is a map v : [a, b] → E from an interval [a, b] of
real numbers to a vector space E that assigns a vector v(t) to each real number
t ∈ [a, b].

• We say that a vector valued function v(t) has a limit v0 as t → t0 , denoted by

lim v(t) = v0
t→t0

if
lim ||v(t) − v0 || = 0 .
t→t0

• A vector valued function v(t) is continuous at t0 if

lim v(t) = v(t0 ) .


t→t0

A vector valued function v(t) is continuous on the interval [a, b] if it is continuous


at every point t of this interval.

• A vector valued function v(t) is differentiable at t0 if there exists the limit


v(t0 + h) − v(t0 )
lim .
h→0 h
If this limit exists it is called the derivative of the function v(t) at t0 and denoted
by
dv v(t0 + h) − v(t0 )
v0 (t0 ) = = lim .
dt h→0 h
If the function v(t) is differentiable at every t in an interval [a, b], then it is called
differentiable on that interval.

• Let {ei } be a constant basis in E that does not depend on t. Then a vector valued
function v(t) is represented by its components

v(t) = vi (t)ei

and the derivative of v can be computed componentwise

dv dvi
= ei .
dt dt

• The derivative is a linear operation, that is,

d du dv d dv
(u + v) = + , (cv) = c ,
dt dt dt dt dt
where c is a scalar constant.

vecanal4.tex; June 30, 2005; 19:16; p. 59


60 CHAPTER 3. GEOMETRY

• More generally, the derivative satisfies the product rules


d dv d f
( f v) = f + v,
dt dt dt
! !
d du dv
(u, v) = , v + u,
dt dt dt
Similarly for the exterior product
d dω dσ
ω∧σ= ∧σ+ω∧
dt dt dt
By taking the dual of this equation we obtain in R3 the product rule for the vector
product
d du dv
u×v= ×v+u×
dt dt dt
• Theorem 3.4.1 The derivative v0 (t) of a vector valued function v(t) with the
constant norm is orthogonal to v(t). That is, if ||v(t)|| = const , then for any t

(v0 (t), v(t)) = 0 .

vecanal4.tex; June 30, 2005; 19:16; p. 60


3.5. GEOMETRY OF CURVES 61

3.5 Geometry of Curves


• Let r = r (t) be a parametrized curve.

• A curve
r = r0 + t u
is a straight line parallel to the vector u passing through the point r 0 .

• Let u and v be two orthonormal vectors. Then the curve

r = r 0 + a (cos t u + sin t v)

is a circle of radius a with the center at r 0 in the plane spanned by the vectors u
and v.

• Let {u, v, w} be an orthonormal triple of vectors. Then the curve

r = r 0 + b (cos t u + sin t v) + at w

is the helix of radius b with the axis passing through the point r 0 and parallel to
w.

• The vertical distance between the coils of the helix, equal to 2π|a|, is called the
pitch.

• Let (qi ) be a curvilinear coordinate system. Then a curve can be described by


qi = qi (t), which in Cartesian coordinates becomes r = r (q(t)).

• The derivative
dr ∂ r dqi
= i
dt ∂q dt
of the vector valued function r (t) is called the tangent vector. If r (t) represents
the position of a particle at the time t, then r 0 is the velocity of the particle.

• The norm r
dr dqi dq j
= gi j (q(t))
dt dt dt
of the velocity is called the speed. Here, as usual

∂r ∂r X ∂xk ∂xk
n
gi j = · =
∂qi ∂q j k=1 ∂qi ∂q j

is the metric tensor in the coordinate system (qi ).

• We will say that a curve r = r (t) is smooth if:


a) it has continuous derivatives of all orders,
b) there are no self-intersections, and
c) the speed is non-zero, i.e. || r 0 (t)|| , 0, at every point on the curve.

vecanal4.tex; June 30, 2005; 19:16; p. 61


62 CHAPTER 3. GEOMETRY

• For a curve r : [a, b] → Rn , the possibility that r (a) = r (b) is allowed. Then it
is called a closed curve. A closed curve does not have a boundary.

• A curve consisting of a finite number of smooth arcs joined together without


self-intersections is called piece-wise smooth, or just regular.

• For each regular curve there is a natural parametrization, or the unit-speed


parametrization with a natural parameter s such that

dr
= 1.
ds

• The orientation of a parametrized curve is determined by the direction of in-


creasing parameter. The point r (a) is called the initial point and the point r (b)
is called the endpoint.

• Nonparametric curves are not oriented.

• The unit tangent is determined by


−1
dr dr
T = .
dt dt
For the natural parametrization the tangent is the unit tangent, i.e.

dr ∂ r dqi
T = = i .
ds ∂q ds

• The norm of the displacement vector d r = r 0 dt


r
dr dqi dq j
ds = dt = ||d r || = gi j (q(t)) dt .
dt dt dt
is called the length element.

• The length of a smooth curve r : [a, b] → Rn is defined by


Z Z b Z br
dr dqi dq j
L= ds = dt = gi j (q(t)) dt
C a dt a dt dt

For the natural parametrization the length of the curve is simply

L = b − a.

That is why, the parameter s is nothing but the length of the arc of the curve
from the initial point r (a) to the current point r (t)
Z t
dr
s(t) = dτ .
a dτ

vecanal4.tex; June 30, 2005; 19:16; p. 62


3.5. GEOMETRY OF CURVES 63

This means that


ds dr
= ,
dt dt
and
dr ds d r
= .
dt dt ds
• The second derivative
!
d2 r
00 d ∂ r dqi ∂ r d2 qi
r = 2 = + i 2 .
dt dt ∂qi dt ∂q dt
is called the acceleration.
• In the natural parametrization this gives the natural rate of change of the unit
tangent !
dT d2 r d ∂ r dqi ∂ r d2 qi
= = + .
ds ds2 ds ∂qi ds ∂qi ds2
• The norm of this vector is called the curvature of the curve
−1
dT dr dT
κ= = .
ds dt dt
The radius of curvature is defined by
1
ρ= .
κ

• The normalized rate of change of the unit tangent defines the principal normal
−1
dT dT dT
N =ρ = .
ds dt dt

• The unit tangent and the principal normal are orthogonal to each other. They
form an orthonormal system.
• Theorem 3.5.1 For any smooth curve r = r (t), the acceleration r 00 lies in the
plane spanned by the vectors T and N . The orthogonal decomposition of r 00
with respect to T and N has the form
d || r 0 ||
r 00 = T + κ || r 0 ||2 N .
dt

• The vector
dN
+κT
ds
is orthogonal to both vectors T and N , and hence, to the plane spanned by these
vectors. In a general space Rn this vector could be decomposed with respect to
a basis in the (n − 2)-dimensional subspace orthogonal to this plane. We will
restrict below to the case n = 3.

vecanal4.tex; June 30, 2005; 19:16; p. 63


64 CHAPTER 3. GEOMETRY

• In R3 one defines the binormal

B = T×N.

Then the triple { T , N , B } is a right-handed orthonormal system called a moving


frame.
• By using the orthogonal decomposition of the acceleration one can obtain an
alternative formula for the curvature of a curve in R3 as follows. We compute

r 0 × r 00 = κ || r 0 ||3 B .

Therefore,
|| r 0 × r 00 ||
κ= .
|| r 0 ||3
• The scalar quantity
dN dB
τ= B · = −N ·
ds ds
is called the torsion of the curve.
• Theorem 3.5.2 (Frenet-Serret Equations) For any smooth curve in R3 there
hold
dT
= κN
ds
dN
= −κ T + τ B
ds
dB
= −τ N .
ds

• Theorem 3.5.3 Any two curves in R3 with identical curvature and torsion are
congruent.

vecanal4.tex; June 30, 2005; 19:16; p. 64


3.6. GEOMETRY OF SURFACES 65

3.6 Geometry of Surfaces


• Let S be a parametrized surface. It can be described in general curvilinear coor-
dinates by qi = qi (u, v), where u ∈ [a, b] and v ∈ [c, d]. Then r = r (q(u, v)).

• The parameters u and v are called the local coordinates on the surface.

• The curves r (u, v0 ) and r (u0 , v), with one coordinate being fixed, are called the
coordinate curves.

• The tangent vectors to the coordinate curves

∂r ∂ r ∂qi ∂r ∂ r ∂qi
ru = = i and rv = = i
∂u ∂q ∂u ∂v ∂q ∂v

are tangent to the surface.

• A surface is smooth if:


a) r (u, v) has continuous partial derivatives of all orders,
b) the tangent vectors r u and r v are non-zero and linearly independent,
c) there are no self-intersections.

• It is allowed that r (a, v) = r (b, v) and r (u, c) = r (u, d).

• A plane T P spanned by the tangent vectors r u and r v at a point P on a smooth


surface S is called the tangent plane.

• A surface is smooth if the tangent plane is well defined, that is, the tangent vec-
tors are linearly independent (nonparallel), which means that it does not degen-
erate to a line or a point at every point of the surface.

• A surface is piece-wise smooth if it consists of a finite number of smooth pieces


joined together.

• The orientation of the surface is achived by cutting it in small pieces and orient-
ing the small pieces separately. If this can be made consistenly for the whole
surface, then it is called orientable.

• The boundary ∂S of the surface r = r (u, v), where u ∈ [a, b], v ∈ [c, d] consists
of the curves r (a, v), r (b, v), r (u, c) and r (u, d). A surface without boundary
is called closed.

• Remark. There are non-orientable smooth surfaces.

• In R3 one can define the unit normal vector to the surface by

n = || r u × r v ||−1 r u × r v .

Notice that p
|| r u × r v || = || r u ||2 || r v ||2 − ( r u · r v )2 .

vecanal4.tex; June 30, 2005; 19:16; p. 65


66 CHAPTER 3. GEOMETRY

In components,
√ ∂ql ∂qm
( r u × r v )i = gεilm ,
∂u ∂v
∂qi ∂q j ∂qi ∂q j ∂qi ∂q j
r u · r u = gi j , r v · r v = gi j , r u · r v = gi j .
∂u ∂u ∂v ∂v ∂u ∂v
• The sign of the normal is determined by the orientation of the surface.
• For a smooth surface the unit normal vector field n varies smoothly over the
surface.
• The normal to a closed surface in R3 is usually oriented in the outward direction.
• In R3 a surface can also be described by a single equation

F(x, y, z) = 0 .

This equation does not prescribe the orientation though. Then


∂F ∂xi ∂F ∂xi
= 0, = 0.
∂xi ∂u ∂xi ∂v
The unit normal vector is then
grad F
n=± .
|| grad F||
The sign here is fixed by the choice of the orientation. In components,
∂F
ni = .
∂xi
• Let r (u, v) be a surface, u = u(t), v = v(t) be a curve in the rectangle D =
[a, b] × [c, d], and r ((u(t), v(t)) be the image of that curve on the surface S . Then
the arc length of this curve is
dr
dl = dt ,
dt
or
dl2 = ||d r ||2 = hab dua dub ,
where u1 = u and u2 = v, and
∂qi ∂q j
hab = gi j ,
∂ua ∂ub
is the induced metric on the surface and the indices a, b take only the values
1, 2. In more detail,

dl2 = h11 du2 + 2h12 du dv + h22 dv2 ,

and
h11 = r u · r u , h12 = h21 = r u · r v , h22 = r v · r v .

vecanal4.tex; June 30, 2005; 19:16; p. 66


3.6. GEOMETRY OF SURFACES 67

• The area of a plane spanned by the vectors r u du and r v dv is called the area
element √
dS = h du dv ,
where
h = det hab = || r u ||2 || r v ||2 − ( r u · r v )2 .

• In R3 the area element can also be written as

dS = || r u × r v || du dv

• The area element of a surface in R3 parametrized by

x = u, y = v, z = f (u, v) ,

is s
!2 !2
∂f ∂f
dS = 1+ + dx dy .
∂x ∂y

• The area element of a surface in R3 described by one equation F(x, y, z) = 0 is


−1
∂F
dS = || grad F|| dx dy
∂z
s
!2 !2 !2
∂F −1 ∂F ∂F ∂F
= + + dx dy
∂z ∂x ∂y ∂z

if ∂F/∂z , 0.

• The area of a surface S described by r : D → Rn , where D = [a, b] × [c, d], is


Z Z b Z d √
S = dS = h du dv .
S a c

• Let S be a parametrized hypersurface defined by r = r (u) = r (u1 , . . . , r n−1 ).


The tangent vectors to the hypersurface are

∂r
,
∂ua
where a = 1, 2, . . . , (n − 1).
The tangent space at a point P on the hypersurface is the hyperplane equal to the
span of these vectors ( )
∂r ∂r
T = span ,..., a .
∂ua ∂u

The unit normal to the hypersurface S at the point P is the unit vector n orthog-
onal to T .

vecanal4.tex; June 30, 2005; 19:16; p. 67


68 CHAPTER 3. GEOMETRY

If the hypersurface is described by a single equation

F(x) = F(x1 , . . . , xn ) = 0 ,

then the normal is


grad F
n=± .
|| grad F||
The sign here is fixed by the choice of the orientation. In components,

∂F
(dF)i = .
∂xi

The arc length of a curve on the hypersurface is

dl2 = hab dua dub ,

where
∂qi ∂q j
hab = gi j .
∂ua ∂ub
The area element of the hypersurface is

dS = h du1 . . . dun−1 .

vecanal4.tex; June 30, 2005; 19:16; p. 68


Chapter 4

Vector Analysis

4.1 Vector Functions of Several Variables


• The set Rn is the set of ordered n-tuples of real numbers x = (x1 , . . . , xn ). We
call such n-tuples points in the space Rn . Note that the points in space (although
related to vectors) are not vectors themselves!
• Let D = [a1 , b1 ] × · · · × [an , bn ] be a closed rectangle in Rn .
• A scalar field is a scalar-valued function of n variables. In other words, it is a
map f : D → R, which assigns a real number f (x) = f (x1 , . . . , xn ) to each point
x = (x1 , . . . , xn ) of D.
• The hypersurfaces defined by
f (x) = c ,
where c is a constant, are called level surfaces of the scalar field f .
• The level surfaces do not intersect.
• A vector field is a vector-valued function of n variables; it is a map v : D → Rn
that assigns a vector v(x) to each point x = (x1 , . . . , xn ) in D.
• A tensor field is a tensor-valued function on D.
• Let v be a vector field. A point x0 in Rn such that v(x0 ) = 0 is called a singular
point (or a critical point) of the vector field v. A point x is called regular if it
is not singular, that is, if v(x) , 0.
• In a neighborhood of a regular point of a vector field v there is a family of
parametrized curves r (t) such that at each point the vector v is tangent to the
curves, that is,
dr
= fv,
dt
where f is a scalar field. Such curves are called the flow lines, or stream lines,
or characteristic curves, of the vector field v.

69
70 CHAPTER 4. VECTOR ANALYSIS

• Flow lines do not intersect.


• No flow lines pass through a singular point.

• The flow lines of a vector field v = vi (x)ei can be found from the differential
equations
dx1 dxn
= · · · = .
v1 vn

vecanal4.tex; June 30, 2005; 19:16; p. 69


4.2. DIRECTIONAL DERIVATIVE AND THE GRADIENT 71

4.2 Directional Derivative and the Gradient


• Let P0 be a point and u be a unit vector. Then r (s) = r 0 + su is the equation of
the oriented line passing through P0 with the unit tangent u.
• Let f (x) be a scalar field. Then the derivative
d
f (x(s))
ds s=0

at s = 0 is called the directional derivative of f at the point P0 in the direction


of u and denoted by
d
∇u f = f (x(s)) .
ds s=0

• The directional derivatives in the direction of the basis vectors ei are the partial
derivatives
∂f
∇ei f = i ,
∂x
which are also denoted by
∂f
∂i f = i .
∂x
• More generally, let r (s) be a parametrized curve in the natural parametrization
and u = d r /ds be the unit tangent. Then

∂ f dxi
∇u f =
∂xi ds
In curvilinear coordinates
∂ f dqi
∇u f = .
∂qi ds

• The covector (1-form) field with the components ∂ f /∂qi is denoted by


∂f i ∂f i
df = dx = i dq .
∂xi ∂q

• Therefore, the 1-forms dqi form a basis in the dual space of covectors.
• The vector field corresponding to the 1-form d f is called the gradient of the
scalar field f and denoted by
∂f
grad f = ∇ f = gi j ej .
∂qi

• The directional derivative is simply the action of the covector d f on the vector u
(or the inner product of the vectors grad f and u)

∇u f = hd f, ui = ( grad f, u) .

vecanal4.tex; June 30, 2005; 19:16; p. 70


72 CHAPTER 4. VECTOR ANALYSIS

• Therefore,
∇u f = || grad f || cos θ ,
where θ is the angle between the gradient and the unit tangent u.
• Gradient of a scalar field points in the direction of the maximum rate of increase
of the scalar field.

• The maximum value of the directional derivative at a fixed point is equal to the
norm of the gradient
max ∇u f = || grad f || .
u

• The minimum value of the directional derivative at a fixed point is equal to the
negative norm of the gradient

min ∇u f = −|| grad f || .


u

• Let f be a scalar field and P0 be a point where grad f , 0. Let r = r (s) be


a curve passing through P with the unit tangent u = d r /ds. Suppose that the
directional derivative vanishes, ∇u f = 0. Then the unit tangent u is orthogonal
to the gradient grad f at P. The set of all such curves forms a level surface
f (x) = c, where c = f (P0 ). The gradient grad f is orthogonal to the tangent
plane to the this surface at P0 .
• Theorem 4.2.1 For any smooth scalar field f there is a level surface f (x) = c
passing through every point where the gradient of f is non-zero, grad f , 0.
The gradient grad f is orthogonal to this surface at this point.

• A vector field v is called conservative if there is a scalar field f such that

v = grad f .

The scalar field f is called the scalar potential of v.

vecanal4.tex; June 30, 2005; 19:16; p. 71


4.3. EXTERIOR DERIVATIVE 73

4.3 Exterior Derivative


• Recall that antisymmetric tensors of type (0, k) are called the k-forms, and the
antisymmetric tensors of type (k, 0) are called k-vectors. We denote the space of
all k-forms by Λk and the space of all k-vectors by Λk .
• The exterior derivative d is an operator
d : Λk → Λk+1 ,
that assigns a (k + 1)-form to each k-form. It is defined as follows.
• A scalar field can be also called a 0-form. The exterior derivative of a zero form
f is a 1-form
∂f
d f = i dxi
∂q
with components
∂f
(d f )i = i .
∂q
• The exterior derivative of a 1-form is a 2-form dσ defined by
∂σ j ∂σi
(dσ)i j = − j.
∂qi ∂q

• The exterior derivative of a k-form σ is a (k + 1)-form dσ with components


∂ X ∂
(dσ)i1 i2 ...ik+1 = (k + 1) [i σi2 ...ik+1 ] = sign (ϕ) i σiϕ(2) ...iϕ(k+1) .
∂q 1 ϕ∈S
∂q ϕ(1)
k+1

• Theorem 4.3.1 The exterior derivative of a k-form is a (k + 1)-form.


• The exterior derivative plays the role of the gradient for k-forms.
• Theorem 4.3.2 The exterior derivative has the property
d2 = 0

• Recall that the duality operator ∗ assigns a (n − k)-vector to each k-form and an
(n − k)-form to each k-vector:
∗ : Λk → Λn−k , ∗ : Λk → Λn−k .

• Therefore, one can define the operator


∗d : Λk → Λn−k−1 ,
which assigns a (n − k − 1)-vector to each k-form by
1 −1/2 i1 ...in−k−1 j1 j2 ... jk+1 ∂
(∗dσ)i1 ...in−k−1 = g ε σ j ... j
k! ∂q j1 2 k+1

vecanal4.tex; June 30, 2005; 19:16; p. 72


74 CHAPTER 4. VECTOR ANALYSIS

• We can also define the operator

∗d∗ : Λk → Λk−1

acting on k-vectors, which assigns a (k − 1) vector to a k-vector.


• Theorem 4.3.3 For any k-vector A with components Ai1 ...ik there holds
∂  1/2 ji1 ...ik−1 
(∗d ∗ A)i1 ...ik−1 = (−1)nk+1 g−1/2 g A .
∂q j

• The operator ∗d∗ plays the role of the divergence of k-vectors.


• Theorem 4.3.4 The operator ∗d∗ has the property

(∗d∗)2 = 0

• Let G denote the operator that converts k-vectors to k-forms,

G : Λk → Λk .

That is, if A j1 ... jk are the components of a k-vector, then the corresponding k-form
σ = GA has components

σi1 ...ik = gi1 j1 . . . gik jk A j1 ... jk .

• Then the operator


G ∗ d : Λk → Λn−k−1
assigns a (n − k − 1)-form to each k-form by
1 −1/2 ∂
(G ∗ dσ)i1 ...in−k−1 = g gi1 m1 · · · gin−k−1 mn−k−1 εm1 ...mn−k−1 j1 j2 ... jk+1 j σ j2 ... jk+1
k! ∂q 1

• The operator ∗d plays the role of the curl of k-forms.


• Further, we can define the operator

δ = G ∗ dG∗ : Λk → Λk−1 ,

which assigns a (k − 1)-form to each k-form.


• Theorem 4.3.5 For any k-form A with components σi1 ...ik there holds
∂  1/2 jp j1 m1 
(δσ)i1 ...ik−1 = (−1)nk+1 g−1/2 gi1 j1 · · · gik−1 jk−1 g g g · · · g jk−1 mk−1 σ pm1 ...mk−1 .
∂q j

• The operator δ plays the role of the divergence of k-forms.


• Theorem 4.3.6 The operator δ has the property

δ2 = 0 .

vecanal4.tex; June 30, 2005; 19:16; p. 73


4.3. EXTERIOR DERIVATIVE 75

• Therefore the operator

L = dδ + δd

assigns a k-form to each k-form, that is,

L : Λk → Λk .

• This operator plays the role of the Laplacian of k-forms.

• A k-form σ is called closed if dσ = 0.


• A k-form σ is called exact if there is a (k − 1)-form α such that σ = dα.

• A 1-form σ corresponding to conservative vector field v is exact, that is, σ = d f .


• Every exact k-form is closed.

vecanal4.tex; June 30, 2005; 19:16; p. 74


76 CHAPTER 4. VECTOR ANALYSIS

4.4 Divergence
• The divergence of a vector field v is a scalar field defined by

div v = (−1)n+1 ∗ d ∗ v ,

which in local coordinates becomes


∂ 1/2 i
div v = g−1/2 (g v ) .
∂qi
where g = det gi j .
• Theorem 4.4.1 For any vector field v the divergence div v is a scalar field.
• The divergence of a covector field σ is

∂ 1/2 i j
div σ = g−1/2 (g g σ j ) .
∂qi

• In Cartesian coordinates this gives simply

div v = ∂i vi .

• A vector field v is called solenoidal if

div v = 0 .

• The 2-form ∗v dual to a solenoidal vector field v is closed, that is, d ∗ v = 0.

Physical Interpretation of Divergence


• The divergence of a vector field is the net outflux of the vector field per unit
volume.

vecanal4.tex; June 30, 2005; 19:16; p. 75


4.5. CURL 77

4.5 Curl
• Recall that the operator ∗d assigns a (n − k − 1)-vector to a k-form. In case n = 3
and k = 1 this operator assigns a vector to a 1-form. This enables one to define
the curl operator in R3 , which assigns a vector to a covector by
curl σ = ∗dσ ,
or, in components,
e1 e2 e3
∂ ∂ ∂ ∂
( curl σ)i = g−1/2 εi jk j σk = g−1/2 det .
∂q ∂q1 ∂q2 ∂q3
σ1 σ2 σ3

• We can also define the curl of a vector field v by



( curl v)i = g−1/2 εi jk (gkm vm ) .
∂q j
• In Cartesian coordinates we have simply
( curl σ)i = εi jk ∂ j σk .
This can also be written in the form
i j k
curl σ = det ∂ x ∂y ∂z
σ1 σ2 σ3

• A vector field v in R3 is called irrotational if


curl v = 0 .

• The one-form σ corresponding to an irrotational vector field v is closed, that is


dσ = 0.
• Each conservative vector field is irrotational.
• Let v be a vector field. If there is a vector field A such that
v = curl A ,
when A is called the vector potential of v.
• If v has a vector potential, then it is solenoidal.
• If A is a vector potential for v, then the 2-form ∗v dual to v is exact, that is,
∗v = dα, where α is the 1-form corresponding to A .

Physical Interpretation of the Curl


• The curl of a vector field measures its tendency to swirl; it is the swirl per unit
area.

vecanal4.tex; June 30, 2005; 19:16; p. 76


78 CHAPTER 4. VECTOR ANALYSIS

4.6 Laplacian
• The scalar Laplace operator (or the Laplacian) is the map ∆ : C ∞ (Rn ) →
C ∞ (Rn ) that assigns a scalar field to a scalar field. It is defined by

∆ f = div grad f = g−1/2 ∂i g1/2 gi j ∂ j f .

• In Cartesian coordinates it is simply

∆ f = ∂i ∂i f .

• The Laplacian of a 1-form (covector field) σ is defined as follows. First, one


obtains a 2-form dσ by the exterior derivative. Then one take the dual of this
2-form to get a (n − 2)-form ∗dσ. Then one acts by exterior derivative to get a
(n − 1)-form d ∗ dσ, and, finally, by taking the dual again one gets the 1-form
∗d ∗ dσ. Similarly, reversing the order of operations one gets the 1-form d ∗ d ∗ σ.
The Laplacian is the sum of these 1-forms, i.e.

∆σ = −(G ∗ dG ∗ d + dG ∗ dG∗)σ .

The expression of this Laplacian in components is too complicated, in general.

• The components expression for this is


 ∂ −1/2 ∂ 1/2 ∂ ∂
(∆v)i = gi jg g − g−1/2 j g1/2 g pi gq j p gqk
∂q j ∂q k ∂q ∂q
∂ ∂ 
+g−1/2 j g1/2 g p j gqi p gqk vk .
∂q ∂q

• Of course, in Cartesian coordinates this simpifies significantly

(∆v)i = ∂ j ∂ j vi .

• In R3 it can be written as

∆v = grad div v − curl curl v .

Interpretation of the Laplacian


• The Laplacian ∆ measures the difference between the value of a scalar field f (P)
at a point P and the average of f around this point.

vecanal4.tex; June 30, 2005; 19:16; p. 77


4.7. DIFFERENTIAL VECTOR IDENTITIES 79

4.7 Differential Vector Identities


• The identities below that involve the vector product and the curl apply only for
R3 . Other formulas are valid for arbitrary Rn in arbitrary coordinate systems:

grad ( f h) = ( grad f )h + f grad h


div ( f v) = ( grad f ) · v + f div v
df
grad f (h(x)) = grad h
dh
curl ( f v) = ( grad f ) × v + f curl v
div (u × v) = ( curl u) · v − u · ( curl v)
curl grad f = 0
div curl v = 0
div ( grad f × grad h) = 0

• Let ei be the standard basis in Rn , xi be the Cartesian


p coordinates, r = xi ei be
the position (radius) vector field and r = || r || = xi xi . Scalar fields that depend
only on r and vector fields that depend on x and r are called radial fields. Below
a is a constant vector field.

div r = n
curl r = 0
grad (a · r ) = a
curl (a × r ) = 2a
r
grad r =
r
df r
grad f (r) =
dr r
1 r
grad = − 3
r r
grad r = krk−2 r
k

(n − 1) 0
∆ f (r) = f 00 + f
r
∆rk = k(k + n − 2)rk−2
1
∆ n−2 = 0
r

• Some useful formulas when working with radial fields are

∂i xk = δki , δii = n .

vecanal4.tex; June 30, 2005; 19:16; p. 78


80 CHAPTER 4. VECTOR ANALYSIS

4.8 Orthogonal Curvilinear Coordinate Systems in R3


• Let (q1 , q2 , q3 ) be an orthogonal coordinate system in R3 and {ê1 , ê2 , ê3 } be the
corresponding orthonormal basis

1 ∂r
êi = .
hi ∂qi
where
∂r
hi =
∂qi
are the scale factors.
Then for any vector v = vi êi the contravariant and the covariant components
coincide
vi = vi = êi · v .

• The displacement vector, the interval and the volume element in the orthogonal
coordinate system are
d r = h1 ê1 + h2 ê2 + h3 ê3 ,
ds2 = h21 (dq1 )2 + h22 (dq2 )2 + h23 (dq3 )2 ,
dV = h1 h2 h3 dq1 dq2 dq3 .

• The differential operators introduced above take the following form

1 ∂ 1 ∂ 1 ∂
grad f = ê1 f + ê2 f + ê3 f
h1 ∂q1 h2 ∂q2 h3 ∂q3
( )
1 ∂ ∂ ∂
div v = (h h v
2 3 1 ) + (h h v
3 1 2 ) + (h h v
1 2 3 )
h1 h2 h3 ∂q1 ∂q2 ∂q3

" #
1 ∂ ∂
curl v = ê1 (h3 v3 ) − 3 (h2 v2 )
h2 h3 ∂q2 ∂q
" #
1 ∂ ∂
+ ê2 (h1 v1 ) − 1 (h3 v3 )
h3 h1 ∂q3 ∂q
" #
1 ∂ ∂
+ ê3 (h2 v2 ) − 2 (h1 v1 )
h1 h2 ∂q1 ∂q
( ! ! !)
1 ∂ h2 h3 ∂ ∂ h2 h3 ∂ ∂ h2 h3 ∂
∆f = + 1 + 1 f
h1 h2 h3 ∂q1 h1 ∂q1 ∂q h1 ∂q1 ∂q h1 ∂q1

• Cylindrical coordinates:

d r = dρ êρ + ρdϕ êϕ + dz êz

ds2 = dρ2 + ρ2 dϕ2 + dz2

vecanal4.tex; June 30, 2005; 19:16; p. 79


4.8. ORTHOGONAL CURVILINEAR COORDINATE SYSTEMS IN R3 81

dV = ρ dρ dϕ dz
1
grad f = êρ ∂ρ f + êϕ ∂ϕ f + êz ∂z f
ρ
1 1
div v = ∂ρ (ρvρ ) + ∂ϕ vϕ + ∂z vz
ρ ρ
êρ ρêϕ êz
1
curl v = ∂ρ ∂ϕ ∂z
ρ vρ ρvϕ vz
1 1
∆f = ∂ρ (ρ ∂ρ f ) + 2 ∂2ϕ f + ∂2z f
ρ ρ
• Spherical coordinates:

d r = dr êr + rdθ êθ + r sin θ dϕ êϕ

ds2 = dr2 + r2 dθ2 + r2 sin2 θ dϕ2


dV = r2 sin θ dr dθ dϕ
1 1
grad f = êr ∂r f + êθ ∂θ f + êϕ ∂ϕ f
r r sin θ
1 1 1
div v = 2 ∂r (r2 vr ) + ∂θ (sin θ vθ ) + ∂ϕ vϕ
r r sin θ r sin θ
êr rêθ r sin θ êϕ
1
curl v = ∂r ∂θ ∂ϕ
r sin θ
2
vr rvθ r sin θ vϕ
1 1 1
∆f = ∂r (r2 ∂r f ) + 2 ∂θ (sin θ∂θ f ) + ∂2ϕ f
r 2 r sin θ r sin2 θ
2

vecanal4.tex; June 30, 2005; 19:16; p. 80


82 CHAPTER 4. VECTOR ANALYSIS

vecanal4.tex; June 30, 2005; 19:16; p. 81


Chapter 5

Integration

5.1 Line Integrals


• Let C be a smooth curve described by r (t), where t ∈ [a, b]. The length of the
curve is defined by
Z Z b
dr
L= ds = dt .
C a dt
• Let f be a scalar field. Then the line integral of the scalar field f is
Z Z b
dr
f ds = f (x(t)) dt .
C a dt
• If v is a vector field, then the line integral of the vector field v along the curve
C is defined by
Z Z b
dr
v·dr = v(x(t)) · dt .
C a dt
• In components, the line integral of a vector field takes the form
Z Z Z  
i
v·dr = vi dq = v1 dq1 + · · · + vn dqn ,
C C C

where vi = gi j v are the covariant components of the vector field.


j

• The expression
σ = vi dqi = v1 dq1 + · · · + vn dqn
is called a differential 1-form. Each covector naturally defines a differential
form. That is why it is also called a 1-form.
• If C is a closed curve, then the line integral of a vector field is denoted by
I
v·dr
C

and is called the circulation of the vector field v about the closed curve C.

83
84 CHAPTER 5. INTEGRATION

5.2 Surface Integrals


• Let S be a smooth parametrized surface described by r : D → Rn , where
D = [a, b] × [c, d]. The surface integral of a scalar field f is
Z Z bZ d √
f dS = f (x(u, v)) h du dv ,
S a c

where u1 = u, u2 = v, h = det hab and hab is the induced metric on the surface

∂qi ∂qi
hab = gi j .
∂ua ∂ub

• Let A be an antisymmetric tensor field of type (0, 2) with components Ai j . It


naturally defines the differential 2-form
X
α = Ai j dqi ∧ dq j
i< j

= A12 dq1 ∧ dq2 + · · · + A1n dq1 ∧ dqn


+A23 dq2 ∧ dq3 + · · · + A2n dq2 ∧ dqn
+ · · · + An−1,n dqn−1 ∧ dqn .

• Then the surface integral of a 2-form α is defined by


Z Z X Z bZ dX
α= Ai j dqi ∧ dq j = Ai j J i j du dv ,
S S i< j a c i< j

where
∂qi ∂q j ∂q j ∂qi
Ji j = − .
∂u ∂v ∂u ∂v
• In R3 every 2-form defines a dual vector. Therefore, one can integrate vectors
over a surface. Let v be a vector field in R3 . Then the dual two form is

Ai j = g εi jk vk ,

or
√ 3 √ √
A12 = gv , A13 = − g v2 , A23 = g v1 .
Ttherefore,
√  3 1 
α= g v dq ∧ dq2 − v2 dq1 ∧ dq3 + v1 dq2 ∧ dq3 .

Then the surface integral of the vector field v, called the total flux of the
vector field through the surface, is
Z Z Z bZ d
α= v · n dS = [v, r u , r v ] du dv ,
S S a c

vecanal4.tex; June 30, 2005; 19:16; p. 82


5.2. SURFACE INTEGRALS 85

where
n = || r u × r v ||−1 r u × r v
is the unit normal to the surface and
√ ∂q j ∂qk
[v, r u , r v ] = vol (v, r u , r v ) = gεi jk vi .
∂u ∂v

• Similarly, the integrals of a differential k-form


X
α= Ai1 ...ik dqi1 ∧ · · · ∧ dqik
i1 <···<ik

with components Ai1 ...ik over a k-dimensional surface qi = qi (u1 , . . . , uk ), ui ∈


[ai , bi ] are defined by
Z Z bk Z b1 X
α= ··· Ai1 ...ik J i1 ...ik du1 · · · duk ,
S ak a1 i1 <···<ik

where
∂q[i1 ∂qik ]
J i1 ...ik = k! ··· k .
∂u 1 ∂u
• The surface integral over a closed surface S without boundary is denoted by
I
α.
S

• In the case k = n−1 we obtain the integral of a (n−1)-form α over a hypersurface


Z Z bn−1 Z b1 X
α= ··· Ai1 ...in−1 J i1 ...in−1 du1 · · · dun−1 .
S an−1 a1 i1 <···<in−1

Let n be the unit vector orthogonal to the hypersurface and v = ∗α be the vector
field dual to the (n − 1)-form α. Then
Z Z bn−1 Z b1 √
α= ··· v · n h du1 · · · dun−1 .
S an−1 a1

This defines the total flux of the vector field v through the hypersurface S .
The normal can be determined by
√ 1 √ ∂qi1 ∂qin−1
h nj = gε ji1 ...in−1 1 · · · n−1 .
(n − 1)! ∂u ∂u

vecanal4.tex; June 30, 2005; 19:16; p. 83


86 CHAPTER 5. INTEGRATION

5.3 Volume Integrals


• Let D = [a1 , b1 ] × · · · × [an , bn ] be a domain in Rn described in local coordinates
(qi ) by
qi ∈ [ai , bi ] .

• The volume element in general curvilinear coordinates is



dV = g dq1 · · · dqn ,

where g = det(gi j ).

• The volume of the region D is


Z Z b1 Z bn

V= dV = ··· g dq1 · · · dqn .
D a1 an

• The volume integral of a scalar field f (x) is


Z Z b1 Z bn

f dV = ··· f (x(q)) g dq1 · · · dqn .
D a1 an

vecanal4.tex; June 30, 2005; 19:16; p. 84


5.4. FUNDAMENTAL INTEGRAL THEOREMS 87

5.4 Fundamental Integral Theorems


5.4.1 Fundamental Theorem of Line Integrals
• Theorem 5.4.1 Let C be a smooth curve parametrized by r (t), t ∈ [a, b]. Then
for any scalar field f (a 0-form)
Z Z Z
∂f i
df = dq = grad f · d r = f (x(b)) − f (x(a)) .
C C ∂qi C

• The line integral of a conservative vector field does not depend on the interior of
the curve but only on the endpoints of the curve.

• Corollary 5.4.1 The circulation of a smooth conservative vector field over a


closed smooth curve is zero,
I
grad f · d r = 0 .
C

5.4.2 Green’s Theorem


• Theorem 5.4.2 Let x and y be the Cartesian coordinates in R2 . Let U be a
bounded region in R2 with the boundary ∂U, which is a closed curve oriented
couterclockwise. Then for any 1-form α = Ai dxi = A1 dx + A2 dy
Z ! I
∂A2 ∂A1
− dx dy = (A1 dx + A2 dy) .
U ∂x ∂y ∂U

5.4.3 Stokes’s Theorem


• Theorem 5.4.3 Let S be a bounded surface in R3 with the boundary ∂S oriented
consistently with the the surface S . Then for any vector field v
Z I
curl v · n dS = v·dr .
S ∂S

5.4.4 Gauss’s Theorem


• Theorem 5.4.4 Let D be a bounded domain in R3 with the boundary ∂D ori-
ented by an outward normal. Then for any vector field v
Z I
div v dV = v · n dS
D ∂D

vecanal4.tex; June 30, 2005; 19:16; p. 85


88 CHAPTER 5. INTEGRATION

5.4.5 General Stokes’s Theorem


• Theorem 5.4.5 Let S be a bounded smooth k-dimensional surface in Rn with the
boundary ∂S , which is a closed (k − 1)-dimensional surface oriented consistently
with S . Let X
α= Ai1 ...ik−1 dqi1 ∧ · · · ∧ dqik−1
i1 <···<ik−1

be a smooth (k − 1)-form. Then


Z I
dα = α.
S ∂S

• In components this formula takes the form


Z Z
∂Ai1 ...ik−1 ∂q[ j ∂qi1 ∂qik−1 ] 1 k ∂q[i1 ∂qik−1 ] 1
· · · du · · · du = A i ...i · · · du · · · duk−1 .
S ∂q j ∂u1 ∂u2 ∂uk ∂S
1 k−1
∂u1 ∂uk−1

vecanal4.tex; June 30, 2005; 19:16; p. 86


Chapter 6

Potential Theory

89
90 CHAPTER 6. POTENTIAL THEORY

6.1 Simply Connected Domains

vecanal4.tex; June 30, 2005; 19:16; p. 87


6.2. CONSERVATIVE VECTOR FIELDS 91

6.2 Conservative Vector Fields


6.2.1 Scalar Potential

vecanal4.tex; June 30, 2005; 19:16; p. 88


92 CHAPTER 6. POTENTIAL THEORY

6.3 Irrotational Vector Fields

vecanal4.tex; June 30, 2005; 19:16; p. 89


6.4. SOLENOIDAL VECTOR FIELDS 93

6.4 Solenoidal Vector Fields


6.4.1 Vector Potential

vecanal4.tex; June 30, 2005; 19:16; p. 90


94 CHAPTER 6. POTENTIAL THEORY

6.5 Laplace Equation


6.5.1 Harmonic Functions

vecanal4.tex; June 30, 2005; 19:16; p. 91


6.6. POISSON EQUATION 95

6.6 Poisson Equation


6.6.1 Dirac Delta Function
6.6.2 Point Sources
6.6.3 Dirichlet Problem
6.6.4 Neumann Problem
6.6.5 Green’s Functions

vecanal4.tex; June 30, 2005; 19:16; p. 92


96 CHAPTER 6. POTENTIAL THEORY

6.7 Fundamental Theorem of Vector Analysis

vecanal4.tex; June 30, 2005; 19:16; p. 93


Chapter 7

Basic Concepts of Differential


Geometry

97
98 CHAPTER 7. BASIC CONCEPTS OF DIFFERENTIAL GEOMETRY

7.1 Manifolds

vecanal4.tex; June 30, 2005; 19:16; p. 94


7.2. DIFFERENTIAL FORMS 99

7.2 Differential Forms


7.2.1 Exterior Product
7.2.2 Exterior Derivative

vecanal4.tex; June 30, 2005; 19:16; p. 95


100 CHAPTER 7. BASIC CONCEPTS OF DIFFERENTIAL GEOMETRY

7.3 Integration of Differential Forms

vecanal4.tex; June 30, 2005; 19:16; p. 96


7.4. GENERAL STOKES’S THEOREM 101

7.4 General Stokes’s Theorem

vecanal4.tex; June 30, 2005; 19:16; p. 97


102 CHAPTER 7. BASIC CONCEPTS OF DIFFERENTIAL GEOMETRY

7.5 Tensors in General Curvilinear Coordinate Systems


7.5.1 Covariant Derivative

vecanal4.tex; June 30, 2005; 19:16; p. 98


Chapter 8

Applications

103
104 CHAPTER 8. APPLICATIONS

8.1 Mechanics
8.1.1 Inertia Tensor
8.1.2 Angular Momentum Tensor

vecanal4.tex; June 30, 2005; 19:16; p. 99


8.2. ELASTICITY 105

8.2 Elasticity
8.2.1 Strain Tensor
8.2.2 Stress Tensor

vecanal4.tex; June 30, 2005; 19:16; p. 100


106 CHAPTER 8. APPLICATIONS

8.3 Fluid Dynamics


8.3.1 Continuity Equation
8.3.2 Tensor of Momentum Flux Density
8.3.3 Euler’s Equations
8.3.4 Rate of Deformation Tensor
8.3.5 Navier-Stokes Equations

vecanal4.tex; June 30, 2005; 19:16; p. 101


8.4. HEAT AND DIFFUSION EQUATIONS 107

8.4 Heat and Diffusion Equations

vecanal4.tex; June 30, 2005; 19:16; p. 102


108 CHAPTER 8. APPLICATIONS

8.5 Electrodynamics
8.5.1 Tensor of Electromagnetic Field
8.5.2 Maxwell Equations
8.5.3 Scalar and Vector Potentials
8.5.4 Wave Equations
8.5.5 D’Alambert Operator
8.5.6 Energy-Momentum Tensor

vecanal4.tex; June 30, 2005; 19:16; p. 103


8.6. BASIC CONCEPTS OF SPECIAL AND GENERAL RELATIVITY 109

8.6 Basic Concepts of Special and General Relativity

vecanal4.tex; June 30, 2005; 19:16; p. 104


110 CHAPTER 8. APPLICATIONS

vecanal4.tex; June 30, 2005; 19:16; p. 105


Bibliography

[1] A. I. Borisenko and I. E. Tarapov, Vector and Tensor Analysis, Dover, 1979
[2] D. E. Bourne and P. C. Kendall, Vector Analysis and Cartesian Tensors, Nelson,
1977
[3] H. F. Davis and A. D. Snider, Introduction to Vector Analysis, 7th Edition,
Brown Publishers, 1995

[4] J. H. Hubbard and B. B. Hubbard, Vector Calculus, Linear Algebra, and Differ-
ential Forms, (2nd Edition), Prentice Hall, 2001

[5] H. M. Schey, Div, grad, curl, and all that: an informal text on vector calculus,
W.W. Norton, 1997
[6] Th. Shifrin, Multivariable Mathematics, Wiley, 2005
[7] M. Spivak, Calculus on Manifolds: A Modern Approach to Classical Theorems
of Advanced Calculus, HarperCollins Publishers, 1965

[8] J. L. Synge and A. Schild, Tensor Calculus, Dover, 1978


[9] R. C. Wrede, Introduction to Vector and Tensor Analysis, Dover, 1972

111
112 BIBLIOGRAPHY

vecanal4.tex; June 30, 2005; 19:16; p. 106


Notation

113

You might also like