Electrodynamics
Electrodynamics
Oscar Reula1
I First Course 13
1 Maxwell’s Equations: An overview 15
1.1 Maxwell’s equations in vacuum . . . . . . . . . . . . . . . . . . . . . . . . . . 15
1.1.1 Maxwell’s evolution equations . . . . . . . . . . . . . . . . . . . . . . . 16
1.1.2 The Constraint Equations . . . . . . . . . . . . . . . . . . . . . . . . . 30
1.2 Initial Value Formulation II: Sources . . . . . . . . . . . . . . . . . . . . . . . 32
1.2.1 Charge Conservation . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
1.2.2 Existence and Uniqueness . . . . . . . . . . . . . . . . . . . . . . . . . 34
3
4 CONTENTS
4.4 Capacities . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 73
8 Dielectrics 119
8.1 The Nature of the Problem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 119
8.2 A Microscopic Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 122
8.2.1 Matching conditions for dielectrics . . . . . . . . . . . . . . . . . . . . 127
8.3 The Electrostatic Energy of Dielectrics . . . . . . . . . . . . . . . . . . . . . . 131
5.1 A conducting box with only its upper face at non-zero potential. . . . . . . . . 78
5.2 The value of the potential at z=c with 10 Fourier modes . . . . . . . . . . . . 80
5.3 The value of the potential at y=b/2 (0.5) . . . . . . . . . . . . . . . . . . . . . 81
5.4 The value of the potential at y=b/2 (5) for a wide box . . . . . . . . . . . . . 81
5.5 The value of the potential at y=5 with only 3 Fourier modes . . . . . . . . . . 81
5.6 Potential at cusps with 9 coefficients. . . . . . . . . . . . . . . . . . . . . . . . 87
5.7 Potential with 20 coefficients. . . . . . . . . . . . . . . . . . . . . . . . . . . . 87
5.8 Absolute value of the real part of Y20 (θ, φ) . . . . . . . . . . . . . . . . . . . . 91
5.9 Absolute value of the real part of Y21 (θ, φ) . . . . . . . . . . . . . . . . . . . . 91
5.10 Absolute value of the real part of Y30 (θ, φ) . . . . . . . . . . . . . . . . . . . . 92
5.11 Absolute value of the real part of Y31 (θ, φ) . . . . . . . . . . . . . . . . . . . . 92
5.12 The relation among angles . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 98
9
10 LIST OF FIGURES
21.3 Measured extinction curves of F and M color centers in a KBr crystal at dif-
ferent temperatures (a); the corresponding changes of the refractive index are
computed via Kramers-Kronig relation (b). The curves correspond to 1) 250
K, 2) 170 K, 3) 100 K, and 4) 35 K, with peaks becoming sharper at low
temperature. Adapted from Peiponen et al. (1984). . . . . . . . . . . . . . . . 304
21.4 Path deformation for Einstein Causality argument. . . . . . . . . . . . . . . . 306
First Course
13
Chapter 1
In this chapter we shall study general aspects of Maxwell’s equations and their solutions.
In particular we shall concentrate on their dynamical properties. For simplicity, but also
to stress that the existence of the electromagnetic fields is independent of the existence of
sources, that is, that they are an entity in their own right, we shall first consider these
equations in vacuum. For them we shall consider their initial value formulation, showing
that the prescription of the electric and magnetic field vectors at a given time, say t = 0,
~ ~x) = F~ (~x) and B(0,
E(0, ~ ~x) = G(~ ~ x) determines uniquely a solution for all future and past
times. To accomplish this we shall first make a detour into the wave equation and using the
time and space translation invariance of that equation, write down the general solution for it,
as a function of its values at a given initial time. We shall them use that formula to obtain the
general solution to Maxwell’s equations. Second we shall define the energy of electromagnetic
fields, and use its conservation in time to show that the solution obtained is the only one with
the prescribed values at t = 0.
In the last part of the chapter we shall consider sources for Maxwell equations, - from a
microscopical point of view - and depending on the nature of them, discuss to what extent
the results already found for the vacuum are still valid.
~
∂E ~ ∧B
= c∇ ~ (1.1)
∂t
~
∂B ~ ∧E
= −c∇ ~ (1.2)
∂t
~ ·E
∇ ~ =0 (1.3)
~ ·B
∇ ~ =0 (1.4)
There are several remarks to be made:
15
16 CHAPTER 1. MAXWELL’S EQUATIONS: AN OVERVIEW
1. The first two equations include time derivatives, in fact tell us what is the instantaneous
~ and B
time evolution of E ~ as a function of the values of their space derivatives at the
present time. They shall be called the evolution equations.
2. The other two are equations relating different space derivatives of the fields at the same
time, that is they constrain the possible value the electric and magnetic fields can have
at any given time, and so shall be called the constraint equations.
3. The equations are linear, that is if (E ~ 1, B
~ 1 ) and (E
~ 2, B
~ 2 ) are two solutions, then (E
~1 +
αE~ 2, B
~ 1 + αB
~ 2 ), where α is any constant, is also a solution. This property is very useful
since it allows finding complicated solutions as sum of simpler ones.
4. There is an asymmetric under interchange of E ~ and B ~ on the evolution equations, [a
sign difference between 1.1 and 1.2]. This asymmetric on the evolution equations is
crucial: Without it they would be inconsistent, not just because they would give a
different evolution than the one observed, but because they would – for generic values
of E~ and B~ at any given initial time – give no evolution at all! We shall come back to
this point later in this chapter. This asymmetry between E ~ and B~ is different from the
one that appears in the constraint equations when sources are present, reflecting the
fact that no magnetic charges have been observed in nature.
5. In the equations there is a constant, c, which has the dimensions of a speed. The
presence of this constant has deep consequences on our present conception of space and
time, but we shall discuss that problem later in the book.
6. If we count the unknowns we see we have six, the three components of E ~ and three of
~
B. But we have 8 equations, 6 evolutions + 2 constraints. So, in order for the system
not to be over-determined two of them should be consequences of the other six. This is
indeed the case, but in a very subtle way. We shall see this later in the chapter.
~
~
E(∆t, ~ ~x) + ∂ E (t, ~x) |t=0 ·∆t + O((∆t)2 )
~x) = E(0,
∂t
~
= E(0, ~x) + c∆t∇~ ∧ B(0,
~ ~x) + O((∆t)2 ),
and correspondingly
1
The origin of time is irrelevant for this discussion as will become apparent in chapter II.
1.1. MAXWELL’S EQUATIONS IN VACUUM 17
~
B(∆t, ~ ~x) − c∆t∇
~x) = B(0, ~ ∧ E(0,
~ ~x) + O((∆t)2 ).
The symbol O((∆t)2 ) means the presence of extra terms, which are bounded and go to
zero as (∆t)2 when ∆t → 0.
Once we have the values of E ~ and B ~ at t = ∆t (to that given order) we can take their
space derivatives at t + ∆t and repeat the above argument to obtain the values of E ~ and B ~
at t = 2∆t and so on until obtaining a “solution" for a given interval of time T .
Of course the solution is only an approximated one, but we could think –in analogy with
ordinary differential equations– of improving on it by taking smaller and smaller ∆’s and
more and more of them as to keep reaching T in the last step.
One would hope, in the limit ∆t → 0, to obtain a unique solution to the evolution equations
~ ~x) and B(0,
corresponding to the given initial data, E(0, ~ ~x) and which depends continuously
under arbitrary variations on that initial data. This argument is in general misleading, for in
each step we must take spatial derivatives. If we trace the formulas to the bare dependence on
~ and B
initial data we see that if we are taking n steps to reach T , then the values of E ~ there
depend on n space derivatives of the initial data. So in the limit ∆t → 0 it would depend on
an infinity of derivatives of the initial data values. One could have then the situation were
very bumpy, but tiny variations, in, say, the 1079th. derivative of E~ would significantly affect
the value of E~ at a later time. One would consider such a situation very unphysical and
should be prepare to through away any theory with such a pathology. It can be shown that
for electromagnetism this awkward situation does not appear, and that this happens because
very precise cancellations occur due to the asymmetry already mentioned of the evolution
equations 2 .
For the problem at hand one does not need to use an argument along the above lines to
show the existence of solutions, for one can find the general solution to the problem, which
we shall display in the proof of the next theorem.
Theorem 1.1 (The Cauchy Problem for electromagnetism) Given (F~ (~x), G(~ ~ x)),
3 3 ~ ~x), B(t,
smooth vectorial functions in space (lR ). There exists a unique smooth solution (E(t, ~ ~x))
to Maxwell’s vacuum evolution equations, 1.1 and 1.2, satisfying
~ ~x), B(0,
(E(0, ~ ~x)) = (F~ (~x), G(~
~ x))
This class of theorem 4 is without doubt one of the more important and basic results of all
physics. It gives ground to our causality conceptions about physical phenomena, and so to the
experimental method in sciences. It says that if we know “the present", that is the initial data
at a given time we can predict the future (the corresponding unique solution to that data) and
that it is a continuous function of the present, that is, if we change things at present by an
small amount the change this will in turn produce in the future shall be correspondingly small.
In particular measurement errors at present, if sufficiently small, would result in small errors
on our predictions about the future. About the experimental method as basic to sciences, it
says that if we carefully prepare an experiment, we can then predict its outcome.
Proof: As already mentioned the proof of this theorem will not follow the argument given
earlier in the chapter, for this would require a fair amount of previous mathematical steps
which are not the subject of this course and which although very interesting would take us
into a long detour. We shall follow instead a much more direct approach, which, although
would take us into some detours -but always within the main subject of the course-, it would
give us also, as a by product, the general solution to Maxwell’s equations! This method will
have the disadvantage that it would be only applicable to the vacuum equations and to some
very particular type of sources, but not to the most general type of problem encountered in
electromagnetism, let alone the rest of physics!
We shall split the proof into two Lemmas. We shall first find a solution to the problem
and second show that this is the solution, e.i. show uniqueness.
~ ~x) = ∂
E(t, (tMct (F~ (~x))) + ctMct (∇ ~ ∧ G(~
~ x))
∂t Z
t
~
− ∇( ~ · F~ (~x))]cdt̃)
[ct̃Mct̃ (∇
0
~ ~x) = F~ (~x)
E(0,
~ ~x) = G(~
B(0, ~ x)
1 Z
Mt (f (~x)) =
f (~x + tn̂) dΩ,
4π Ω
where Ω is the sphere of unit radius, that is,
1 π 2π
Z Z
Mt (f (~x)) = f (x + t cos θ sin ϕ, y + t cos θ cos ϕ, z + t sin θ) sin θdθdϕ.
4π 0 0
Proof: This Lemma can be proven by brute force, namely applying directly the equations to
the proposed solution, but here me propose another method which will teach us some things
about the wave equation. Taking a time derivative of (1.1) and using (1.2) we obtain,
1 ∂2E~ 1 ~ ∂B ~
= ∇ ∧
c2 ∂t2 c ∂t
~
= −∇ ∧ (∇~ ∧ E)
~
= ∆E~ − ∇(
~ ∇ ~ · E),
~
where in the third step we have used the vectorial calculus identity:
~ ∧ (∇
∇ ~ ∧V
~ ) = ∇(
~ ∇~ ·V
~ ) − (∇
~ · ∇)
~ V~
.
At this point we could use equation 1.3 to drop the second term in the r.h.s. of the above
equation, but since at the moment we are only solving for the evolution equations, we prefer
to eliminate it by taking another time derivative.
~
Taking it and calling Y~ = ∂ E we obtain,∂t
1 ∂ 2 Y~
= ∆Y~ − ∇(
~ ∇~ · (c∇
~ ∧ B))
~
c2 ∂t2
= ∆Y~
~ to substitute in the second
where in the first step we have use the evolution equation for E
term on the r.h.s. Y~ by c∇ ~ ∧ B,
~ and in the second step the vectorial calculus identity
~ ~ ~
∇ · (∇ ∧ A) = 0.
1 ∂2
( 2 2 − ∆)Y~ = 0. (1.5)
c ∂t
20 CHAPTER 1. MAXWELL’S EQUATIONS: AN OVERVIEW
If we express Y~ and the above equation in Cartesian coordinates it is easy to see that each
individual component of Y~ , Y k , satisfies the scalar wave equation:
1 ∂2 ij ∂2 k 1 ∂2 ∂2 ∂2 ∂2
( − δ )Y = ( − − − )Y k = 0,
c2 ∂t2 ∂xi ∂xj c2 ∂t2 ∂x2 ∂y 2 ∂z 2
where by Y k we are denoting the Cartesian component of Y~ along the axis, ~xk , and
(
ij 1 i=j
δ =
0 i=
6 j
denotes the Euclidean metric of lR3 and we are using the Einstein’s notation, by which repeated
indices, one up one down, imply summation.
The scalar wave equation is the small brother of Maxwell’s equation and we shall have the
opportunity to encounter it many times during this course, so it is convenient to make the
following:
Exercise: Show that this invariance plus linearity implies that if φ is a smooth solution, then
∂φ ~ also satisfy the wave equation.
and ∇φ
∂t
To see that the wave equation has these invariances consider the wave equation applied
to φT (t, ~x). To apply it we need to compute ∂φ
∂t
T
in terms of φ,
1 ∂2 1 ∂2
−(∆)φ T (t, ~
x ) = ( − ∆)φ(u, ~x) = 0,
c2 ∂t2 c2 ∂u2
proving that φT is also a solution. The case of space translations is similar using the trick of
aligning one of the coordinate axis with the space translation under consideration.
We shall make use of this invariance to find the general solution to the wave equation.
5
We shall study this type of invariance for Maxwell’s equations in the next chapter.
1.1. MAXWELL’S EQUATIONS IN VACUUM 21
1 ∂2φ 1 ∂2 1 ∂ ∂φ 1 ∂2φ
− (rφ) + (sin θ ) + =0
c2 ∂t2 r ∂r 2 r 2 sin θ ∂θ ∂θ r sin2 θ ∂φ2
So if we look for spherically symmetric solutions, that is solutions which only depend on
t and r the equation reduces to:
1 ∂ 2 φ 1 ∂ 2 (rφ)
− =0
c2 ∂t2 r ∂r 2
which can be further simplified, by defining. ψ(t, r) = rφ(t, r), to
1 ∂2ψ ∂2ψ
− 2 = 0,
c2 ∂t2 ∂r
that is, a one dimensional wave equation. A particular solution to this equation is ψ(t, r) :=
F (ct − r), where F is any sufficiently smooth function.
Exercise: Show that ψ(t, r) := G(ct + r), with G an arbitrary smooth function is also a
solution.
Thus we obtain a solution to the three dimensional wave equation, φ(t, r) = F (ct−r)r
, at all
points besides the origin, which represents a spherical wave going away from the coordinate
axis and meanwhile decaying in intensity as 1r if F is taken to be of compact support. Note
that this solution will, in general, be unbounded at the origin. This can be arranged by
instead choosing alternatively the solution φ(t, r) = F (ct−r)−F
r
(ct+r)
. The problem at the origin
will not be important for our present application.
Because of the translational invariance we can take any coordinate origin we please and
get a solution of this type. Of course their sum would also be a solution and so we conclude
that
is also a solution, where φ̃xj is a weight constant for each “symmetry center". One can picture
this type of solutions by throwing a hand full of marbles into a lake, roughly, the weight factors
there are related to the actual weight of each marble. The picture is not totally correct, for
those waves decay at a different rate with the distance.
There is no reason not to consider a smooth distribution of these symmetry centers and
so passing from the sum to an integral we conclude that
22 CHAPTER 1. MAXWELL’S EQUATIONS: AN OVERVIEW
Z
φ̃(~x′ )F (ct − |~x − ~x′ |) 3 ′
φ(t, ~x) = d ~x ,
lR3 |~x − ~x′ |
is also a solution to the wave equation.
Using spherical coordinates centered at ~x, ~x′ = ~x + rn̂, so that d3~x′ = dΩ2 r 2 dr the integral
becomes,
noticing that when ε → 0 the function becomes more and more concentrated at s = 0 while
its integral remains finite,
1
Z ∞
Fε (s)ds = .
0 4π
Thus, in the above integral the values which contribute are more and more concentrated at
the values of r for which ct − r is near zero, namely r = ct. Taking such a limit, we reach the
conclusion that the following is also a solution,
ct
Z
φ(t, ~x) = φ̃(~x + ctn̂)dΩ
4π Ω
:= ctMct (φ̃).
But these are still very particular ones, for at t = 0 they all vanish. To enlarge that set we
consider the bigger set obtained by adding to the ones found the sets of all its time derivatives.
We now claim:
Any solution of the wave equation can be written as
∂
φ(t, ~x) = tMct (φ1 (~x)) + (tMct (φ0 (~x))),
∂t
∂φ
noticing that ∂t
(t, ~x) |t=0 = φ1 (~x) and φ(0, ~x) = φ0 (~x).
1.1. MAXWELL’S EQUATIONS IN VACUUM 23
We have already shown that the first term is a solution. The second term is also a solution
for time translational invariance plus linearity implies that any time derivative of a solution
is also a solution.
Exercise: Show the statement about the values of φ0 (~x) and φ1 (~x).
But any solution of the wave equation can be expressed in the above form, since any two
solutions which coincide and have the same time derivatives at t = 0, necessarily also coincide
in the whole space. We shall not proof that this statement, since, although the proof is not
difficult, it is very similar to the one we shall give for Maxwell’s equations, and we shall not
need it in what follows. Notice that since all solutions have this form we can “classify" them
by listing the values and the values of their time derivatives at a t = const. surface. These
values constitute the initial data set of the equation, and we have just asserted that to each
element on this set, that is, a pair (φ0 (~x), φ1 (~x)), there corresponds a unique solution to the
equation.
The form of the general solution we have found underlines an important property of the
wave equation which it is also shared by Maxwell’s equations, namely that the value of a
solution at a point ~x at time t depends only on the value of the solution at the points ~x′ at
t = 0 such that |~x − ~x′ | = ct, that is, the waves travel at the speed c.
(t, ~
x)
|~ x′ =ct
x−~
t=0
Figure 1.1: Set of points which influence the event (t, ~x)
The diagram above represents points of space and time, usually called events, horizontal
displacements mean displacements along space, vertical ones along time. The points drawn
represent those points in space time which can influence the value of any solution of the wave
equation at the point (t, ~x). It is clearly a cone,
C − (t, ~x) = {(t′ , ~x′ ) | |~x − ~x′ | = c(t − t′ )}
Exercise: Let (φ0 (~x), φ1 (~x)) of compact support, show that given any point ~x0 there exists
a time t0 such that for all t > t0 , φ(t, ~x0 ) = 0.
24 CHAPTER 1. MAXWELL’S EQUATIONS: AN OVERVIEW
Exercise: Find all the points in space-time which can be influenced from data given at
(t = 0, ~x0 ).
We finish this detour into the wave equation giving a direct proof of Lemma 1.2:
Given any smooth function φ̃(~x) : R3 → R, the function R3 × R → R,
φ(t, ~x) := t Mt (φ̃(~x))
satisfies the wave equation with initial data (φ(0, ~x) = 0, ∂t φ(t, ~x)|t=0 = φ̃(~x)).
Proof:
First notice that
1
Z
∂t Mt (φ̃(~x)) = ∂t φ̃(~x + tn̂) dΩ
4π ZS 2
1 ~ φ̃(~x + tn̂) · n̂ dΩ
= ∇
4π SZ2
1 ~ φ̃(~x + tn̂) · n̂ t2 dΩ
= ∇ (1.6)
4πt S
2 2
The last term above is just an integral on a sphere of radius t of the gradient of a function
normal to the surface, so we can use Gauss theorem to conclude that:
1
Z
∂t Mt (φ̃(~x)) = ∆φ̃(~x + tn̂) dV
4πt2 Bt (~x)
1
Z
= ∆x φ̃(~x + tn̂) dV
4πt2 Bt (~x)
1
Z tZ
= ∆x φ̃(~x + τ n̂) dΩ τ 2 dτ
4πt2 0 S 2
1 t
Z
= 2 ∆x Mτ (φ̃(~x) τ 2 dτ (1.7)
t 0
where Bt (~x) is a ball of radius t centered at ~x, in the second line we change to the Laplacian
with respect to ~x (just a coordinate change), and in the third line we had written the integral
in spherical coordinates. Finally we re-write the expression in terms of spherical means. Thus,
multiplying by t2 and taking a derivative we get,
1 ∂ 2 φ(t, ~x)
− ∆φ(t, ~x) = 4πf (t, ~x)
c2 ∂t2
has the general solution,
∂
φ(t, ~x) = (tMct (φ0 (~x))) + tMct (φ1 (~x))
∂t Z
t
+ 4πc2 t̃Mct̃ (f (t − t̃, ~x))dt̃, (1.8)
0
∂φ
with φ(0, ~x) = φ0 (~x), ∂t
(t, ~x) |t=0 = φ1 (~x).
Exercise: Show that the inhomogeneous term can be re-written, via a change of variables,
as
Z tc Z
4π rf (t − r/c, ~x + rn̂)drdΩ (1.9)
0 Ω
Draw the integration region.
Exercise: Check that in one dimension, the situation is different. Namely that the inhomo-
geneous solution to the corresponding wave equation,
1
∂tt φ − ∂rr φ = 4πf (t, r) (1.10)
c2
is,
Z tZ r+ct̃
φI = 2πc f (t̃, r̃) dr̃dt̃ (1.11)
0 r−ct̃
Exercise: NOT AT ALL EASY! We have seen that the first two terms are solutions to the
homogeneous (e.i. without sources) equations, with the correct initial data. Show that the
third term is a solution to the inhomogeneous (e.i. with sources) wave equation with zero
initial data.
du
= Au,
dt
26 CHAPTER 1. MAXWELL’S EQUATIONS: AN OVERVIEW
where u is a vector in some space and A a linear operator independent of t. The solutions
are then of the form,
u(t) = Pst (φ(s))
That is, this is the unique solution with initial data φ(s) at time t = s, where the operator
Pst satisfies the equation,
d t
P (φ(s)) = APst (φ(s)), with initial condition Ptt = id
dt s
We can use this solution representation to find the solutions to the inhomogeneous equation,
du
= Au + f,
dt
with homogeneous initial data. They are given by,
Z t
v(t) = Pst (f (s)) ds. (1.12)
0
Indeed,
d t d t
Z
v = Ptt (f (t)) + P (f (s)) ds
dt 0 dt s
Z t
= f (t) + APst (f (s)) ds
0
Z t
= f (t) + A Pst (f (s)) ds
0
= f (t) + Av(t).
For the case at hand, we can write the wave equation as a first order system as,
! ! ! !
∂ φ 0 1 φ 0
= 2 +
∂t π c ∆ 0 π 4πc2 f
For the homogeneous equation we have,
! !
φ(t) ∂t (tMct (φ(0))) + tMct (π(0))
u(t) = P0t (u(0)) = = .
π(t) ∂t (∂t (tMct (φ(0))) + tMct (π(0)))
Thus, applying the above result we get, (the source f has to be replaced for π),
Z t Z t
2 2
φ(t) = 4πc (t − s)Mc(t−s) (f (s)) ds = 4πc τ Mcτ (f (t − τ )) dτ,
0 0
End of detour
1.1. MAXWELL’S EQUATIONS IN VACUUM 27
We now continue with the proof of our lemma. Since each Cartesian component of Y~ (t, ~x)
satisfies the wave equation we can readily write down some solutions:
∂
Y~ (t, ~x) = (tMct (Y~ 0 (~x))) + tMct (Y~ 1 (~x))
∂t
Y~ (t, ~x)|
with Y~ 0 (~x) ≡ Y~ (0, ~x) and Y~ 1 (~x) ≡ ∂∂t t=0
Using now the evolution equations, 1.1, 1.2, we obtain,
~
∂E
Y~ 0 (~x) = Y~ (0, ~x) = (t, ~x) |t=0 = c∇ ~ ∧ B(t,
~ ~x) |t=0
∂t
= c∇~ ∧ G(~~ x),
∂ Y~ ∂2E ~ ~
~ ∧ ∂ B (t, ~x) |t=0
Y~ 1 (~x) = (t, ~x) |t=0 = (t, ~
x ) | t=0 = c ∇
∂t ∂t2 ∂t
2~ ~ ∧ E(t,
~ ~x) |t=0 = −c ∇ 2 ~ ∧∇ ~ ∧ F~ (~x),
= −c ∇ ∧ ∇
And so,
∂ ~ ∧ G(~
Y~ (t, ~x) = (ctMct (∇ ~ x))) − c2 tMct (∇
~ ∧∇
~ ∧ F~ (~x)).
∂t
But by definition, Z t
~ ~x) = E(0,
E(t, ~ ~x) + Y~ (t̃, ~x)dt̃,
0
and so,
~ ~x) = F~ (~x) + ctMct (∇
E(t, ~ ∧ G)(~
~ x))
Z t
− c ~ ∧∇
ct̃Mct̃ (∇ ~ ∧ F~ (~x))dt̃.
0
Z t
~ ~x) = G(~
B(t, ~ x) − c ~ ∧ E(
∇ ~ t̃, ~x)dt̃
0
Z t
~ x) − c
= G(~ ~ ∧ F~ (~x) − ct̃Mct̃ (∇
[∇ ~ ∧∇
~ ∧ G(~
~ x))
0
Z t̃
− c ˜ ˜(∇
ct̃M ~ ∧∇
~ ∧∇ ˜ t̃
~ F~ (~x))dt̃]d
ct̃
0
~ ~x) = ∂
E(t, (tMct (F~ (~x))) + ctMct (∇ ~ ∧ G(~
~ x))
∂t Z
t
~
− ∇( ~ · F~ (~x))]cdt̃)
[ct̃Mct̃ (∇ (1.13)
0
28 CHAPTER 1. MAXWELL’S EQUATIONS: AN OVERVIEW
Hint: Use that t̃Mct̃ (F~ (~x)) satisfies the wave equation.
To complete the proof of the theorem we must prove now that the general solution we
have obtained are the only possible.
~ 1 (t, ~x) − E
E(t, ~x) = E ~ 2 (t, ~x),
~ 1 (t, ~x) − B
B(t, ~x) = B ~ 2 (t, ~x).
Since Maxwell’s equations are linear these differences are also a solution, and
~ 0 , ~x) = B(t
E(t ~ 0 , ~x) = 0.
So, because the linearity of the equations, the proof reduces to show that the only solution
with vanishing initial data is the zero solution. If these fields decay sufficiently fast at infinity,
then the following integral is finite:
1
Z
E(t) := ~ ~x) · E(t,
{E(t, ~ ~x) + B(t,
~ ~x) · B(t,
~ ~x)}d3~x.
2 lR3
6
This last requirement can be lifted away using an argument involving the finite velocity of propagation of
the electromagnetic waves, but the proof would become too complicated.
1.1. MAXWELL’S EQUATIONS IN VACUUM 29
d
Z
~ ~x) · ∇
E(t) = c 3 {E(t, ~ ∧ B(t,
~ ~x) − B(t,
~ ~x) · ∇
~ ∧ E(t,
~ ~x)}d3~x.
dt lR
~ · (V
Using now the vectorial calculus identity: ∇ ~ ∧W
~)=W
~ · (∇
~ ∧V
~)−V
~ · (∇
~ ∧W
~ ),
and Gauss theorem, we see that
d
Z
E(t) = lim ~ ~x) ∧ B(t,
−c(E(t, ~ ~x)) · n̂ dS
dt r→∞ S 2 (r)
where the integral is over a sphere of radius r, and n̂ is the outward unit normal to it. The
limit is taken after making the integral.
~ and B
If the field differences E ~ decay sufficiently fast at infinity, then in the limit the
surface integral becomes zero and so E(t) is conserved, E(t) = E(t0 ). But at t = t0 E ~ and B ~
were zero, so that E(t0 ) = 0 and so we conclude that E(t) = 0 ∀t. Since E(t) is the integral of
a sum of positive definite terms we conclude that each one of them must vanish point wise 7
and so we have
~ ~x) = B(t,
E(t, ~ ~x) = 0
Exercise: Prove a similar Lemma for the wave equation using as the energy functional,
1 ∂φ
Z
E(t) = ~ · ∇φ}d
{( )2 + ∇φ ~ 3
~x.
2 lR 3 ∂t
One asks oneself what is behind such a neat proof of uniqueness? Usually simplicity
and beauty are correlated with some deep aspect of nature. In this case it happens that the
integral which so conveniently helped us to prove uniqueness is in fact the total energy carried
~ ~x), B(t,
by the electromagnetic configuration (E(t, ~ ~x)), and the fact that it is constant in time
is nothing else than energy conservation. The surface integral is nothing else but the energy
by unit of time radiated away from the region, and its integrand, S ~ = cE ~ ∧ B,
~ called the
Poynting vector, is the flux of energy.
Exercise: Give an argument, using the fact that the solutions propagate at finite speed to
show, first the if the initial data vanishes in, say, inside ball of radius R, then after some
interval δt it vanishes in a ball of radius R − cδt, and second that the requirement of the fields
decaying at infinity of the previous result is superfluous.
~ ~x), B(t,
What have we found so far? Consider the set, S, of all pairs of vectors (E(t, ~ ~x)),
defined at all points of space and for all times. This set has a subset, SE, which consists of
all pairs which satisfies the evolution Maxwell’s equations. If we evaluate those pairs in SE
7
Since we are assuming all fields are smooth.
30 CHAPTER 1. MAXWELL’S EQUATIONS: AN OVERVIEW
at t = 0, we see that we have a map, Φ, from SE to the set of “free initial data", F ID, that
is the set of all pairs of vector (F~ (~x), G(~
~ x)) defined at point of lR3 , the map given by
~ ~x), B(t,
Φ((E(t, ~ ~x))) = (E(0,
~ ~x), B(0,
~ ~x)) = (F~ (~x), G(~
~ x)).
What we have shown so far is that this map is one to one, and so invertible. That is, each
free initial data pair (F~ (~x), G(~
~ x)) gives rice to a unique solution to the evolution equation.
Solutions are uniquely characterized by their initial data. To speak of a solution or of its
initial data is completely equivalent.
S
FID
SE
Figure 1.2: The map between solutions and initial data sets
Theorem 1.2 P is uniquely characterized by ID. That is, if the initial data satisfies the con-
straint equations, then the solution to the evolution equations they give rise to, automatically
satisfy the constraint equations for all times.
1.1. MAXWELL’S EQUATIONS IN VACUUM 31
F ID
ID
P SE SC
Figure 1.3: Map between true solutions and constrained initial data sets
Proof: Assume we are given any pair in ID, that is a pair (F~ (~x), G(~
~ x)) such that ∇
~ · F~ (~x) =
~ · G(~
∇ ~ x) = 0. We want to prove that the solution to the evolution equations this pair gives
~ ~x), B(t,
rise to,9 (E(t, ~ ~x)) satisfies the constraints equations for all times.
Taking a time derivative to each one of the constraint equations and using the evolution
equations we have,
∂ ~ ~ ~ · ∂ E(t,
~ ~x) = c∇
~ · (∇
~ ∧ B(t,
~ ~x)) = 0,
(∇ · E(t, ~x)) = ∇
∂t ∂t
∂ ~ ~ ~ · ∂ B(t,
~ ~x) = −c∇~ · (∇
~ ∧ E(t,
~ ~x)) = 0,
(∇ · B(t, ~x)) = ∇
∂t ∂t
~ · (∇
where we have used the vector calculus identity, ∇ ~ ∧ A)
~ = 0. Therefore,
~ · E(t,
∇ ~ ~x) = ∇
~ · E(0,
~ ~x) = ∇
~ · F~ (~x) = 0,
~ · B(t,
∇ ~ ~x) = ∇
~ · B(0,
~ ~x) = ∇
~ · G(~
~ x) = 0
Remark: From this result we see that the name “constraint equations" is justified. These
equations relate different components of the initial data, that is, it is not possible to prescribe
arbitrarily the three components of the electric and the three components magnetic field as
initial data for Maxwell’s equations. At most one could prescribe two for each one of them,
the rest been determined from the ones given. But this counting of freely given components
is only approximate, as the following exercise shows.
Exercise: Let V~ (~x) and W ~ (~x) two arbitrary vector functions in lR3 . Show that the pair
~ ∧V
(∇ ~ (~x), ∇
~ ∧W
~ (~x)) is in ID.
8
P is not empty because the zero pair bellow to it.
9
Whose existence and uniqueness has been already shown.
32 CHAPTER 1. MAXWELL’S EQUATIONS: AN OVERVIEW
The above example also shows that there are plenty of elements of P . These solutions
represents “pure radiation fields", since they do not arise from any source whatsoever.
Exercise: Find expressions (1.13,1.14) again, this time using that the divergences cancell.
Hint: Check that now both E,~ and B~ satisfy the wave equation and write the general solutions
for them from the initial data. You will need to use equations (1.1,1.2).
~
∂E ~ ∧ B(t,
(t, ~x) = c∇ ~ ~x) − 4π J(t,
~ ~x) (1.15)
∂t
~
∂B ~ ∧ E(t,
(t, ~x) = −c∇ ~ ~x) (1.16)
∂t
~ · E(t,
∇ ~ ~x) = 4πρ(t, ~x) (1.17)
~ · B(t,
∇ ~ ~x) = 0, (1.18)
To see this take the divergence of 1.1 and substract from it the time derivative of 1.17,
∂ ~ ~ ~ · ( ∂ E)
~
0 = (∇ · E) − ∇
∂t ∂t
∂ρ ~ ~ ∧B ~ − 4π J)
~
= 4π − ∇ · (c∇
∂t
∂ρ ~ ~
= 4π( + ∇ · J),
∂t
where in the second step we have used the vector identity, ∇~ · (∇
~ ∧ A) ≡ 0.
Thus if this relation between the charge density and the charge current is not satisfied,
then there can not be solutions to Maxwell’s equations.
The continuity equation has a simple interpretation, it just says that charge can not
just disappear from a given volume, if it decreases there, it has to be because it leaves the
boundaries of that volume as a current thought it boundaries. Indeed taking a fix volume, V ,
the total charge enclosed on it is,
Z
QV (t) = ρ(t, ~x)d~x3 ,
V
where after using the continuity equation we have used Stoke’s theorem. ∂V denotes the
boundary of V , and n̂ its outer normal. Thus, the sources of Maxwell’s equations can not be
anything we like, they must have a “material" entity in the sense that if their amount change
in a given volume then a correlated flux of the quantity must be present at the boundary of
it. In particular we have global charge conservation:
Lemma 1.4 If the sources have compact support along space directions, then the total charge
is constant.
~ ~x), and ρ(0, ~x) = ρ0 (~x) are given then we also have ρ(t, ~x), indeed the
Note that if J(t,
continuity equation asserts,
Z t
ρ(t, ~x) = ρ0 (~x) − ~ · J(
∇ ~ t̃, ~x)dt̃.
0
This fact is sometimes useful in constructing consistent sources for Maxwell’s equations.
Theorem 1.3 Let ρ(t, ~x), J(t, ~ ~x)) satisfy the continuity equation, and let the initial data
(F~ (~x), G(~
~ x)) satisfies the constraint equations, ∇
~ · F~ (~x) = 4πρ(0, ~x), ∇
~ · G(~
~ x) = 0. Then
the solution to the evolution equations it generates (if it exist) satisfies the constraint equa-
tions for all times.
Proof: Taking a time derivative of the constraint equations, using the evolution equations,
~ · (∇
and the identity, ∇ ~ ∧ A)
~ = 0, we obtain,
10
The sources in general depend on the electromagnetic fields at past times, so in general the new equations
become integro–differential equations. They become local only when Fourier transformed.
1.2. INITIAL VALUE FORMULATION II: SOURCES 35
∂ ~ ~ ~
(∇ · E − 4πρ) = ∇~ · ∂ E − 4π ∂ρ
∂t ∂t ∂t
~ · (c∇
= ∇ ~ ∧B ~ − 4π ∂ρ
~ − 4π J)
∂t
~ · J~ + )∂ρ
= −4π(∇
∂t
= 0
∂ ~ ~ ~
~ · ∂B
(∇ · B) = ∇
∂t ∂t
~
= ∇ · (−c∇~ ∧ E)~
= 0.
Thus
~ · E(t,
∇ ~ ~x) − 4πρ(t, ~x) = ∇
~ · E(0,
~ ~x) − 4πρ(0, ~x)
~ · F~ (0, ~x) − 4πρ(0, ~x)
= ∇
= 0,
and,
~ · B(t,
∇ ~ ~x) = ∇
~ · G(~
~ x) = 0.
Remarks
1. The equation of continuity is a key ingredient here. If it were not satisfied by the sources
then Maxwell’s equations would be inconsistent, namely P = Φ
2. The above proof does not use anything about the character of the equations for the
sources or their nature, so the conclusion of the theorem are general, as long as the
solution exists, as is the necessity of the continuity equation to hold for consistency.
Then
~ ~x) = ∂
E(t, (tMct (F~ (~x))) + ctMct (∇ ~ ∧ G(~
~ x))
∂t
∂ J~(t − t̃, ~x)
Z t
− 4π ~
t̃Mct̃ (c2 ∇ρ(t − t̃, ~x) + )dt̃
0 ∂t
~ ~x) = ∂ (tMct (G(~
B(t, ~ x))) − ctMct (∇ ~ ∧ F~ (~x))
∂t Z
t
+ 4πc t̃Mct̃ (∇ ~ ∧ J(t
~ − t̃, ~x))dt̃
0
~ ~x))
(ρ(t, ~x), J(t,
and initial data
~ ~x) = F~ (~x),
E(0, ~ ~x) = G(~
B(0, ~ x).
~
∂2B ~ ∧ (∇
= −c2 ∇ ~ ∧ B)
~ + 4πc∇~ ∧ J~
∂t2
= +c2 (∆B~ − ∇(
~ ∇~ · B))
~ + 4πc∇ ~ ∧ J~
= +c2 ∆B~ + 4πc∇
~ ∧ J,~
~ ·B
where in the last step we have used the already shown fact that ∇ ~ = 0. Thus we have a
~
wave equation for B, but this time with a source:
1 ∂2 ~ ~ = 4π ∇
~ ∧ J~
2 2
B − ∆B
c ∂t c
Thus, using the formula for the solutions to the wave equation with sources, (1.8), and
~ ~x) to express its time derivative in terms of E(t,
again the evolution equation for B(t, ~ ~x) at
t = 0, and so in terms of F~ (~x) we get,
~ ~x)
Using this expression and integrating forward in time the evolution equation for E(t,
we obtain the values of the electric field for all times.
Exercise: Make this last calculation. Redo the calculation getting first a wave equation for
~ ~x), using its constraint equation, and see that one arrives to the same results.
E(t,
Exercise: Use Duhamel’s formula, 1.12, and the homogeneous solution to Maxwell’s equa-
~ and B)
tions (in terms of E, ~ to find the inhomogeneous solution directly.
Remark: We see now that the solution has two different type of terms: One type only
depends on the initial data, and are vacuum (homogeneous) solutions with the sought initial
data. The other is a solution to the equations with sources (inhomogeneous equation) but
with vanishing initial data. The validity of this splitting is a generic and very useful property
of linear systems.
1.2. INITIAL VALUE FORMULATION II: SOURCES 37
To summarize we have shown that given a current density everywhere in space and time,
~
J(t, ~x) and an initial charge density, ρ(0, ~x) = ρ0 (~x), (and obtaining ρ(t, ~x) by integration
of the continuity equation), we have constructed the set P(ρ0 ,J) of all solutions to Maxwell’s
equations, and display its explicit dependence on initial data sets, ID(ρ0 ). Note that the map
~ ~x), that is, Φ = Φ .
relating these sets now depends on J(t,
J~
Even taking into account the discussion at the beginning of this section warning the reader
that this type of sources are very special, it would seem we have made a big break through
into Maxwell’s theory and it would only remain just to explicitly display the set ID(ρ0 ), which
in fact can be done quite easily. This is not so for various reasons:
• As we have discussed in general the motion of the sources do depend on the electro-
magnetic fields they generate or are acted upon from the outside. This includes all
macroscopic media discussions.
• The presence of sources brings into the problem different time and space scales. In
general, it is necessary to use some extra physical arguments to say in which scale the
equations are to hold. We shall see this when we study continuous media.
• Even if we could find all solutions for the case where the interaction of the sources,
between each other and with the electromagnetic fields were taken into account, (some-
thing which probably could be done in a few years thanks to the ever growing computer
power) it is very difficult to extract physically relevant information from general so-
lutions, or “tables of solutions" as the computers would generate. As we shall see in
the following chapters, Maxwell’s equations have an incredible richness which hardly
could be extracted by these methods. Such richness is best discovered by studying the
equation (rather than the specific solutions), and by studying many simple solutions,
which in turn act as building blocks of general, very complicated ones, and at the same
time are suitable for displaying all the physical phenomena behind electromagnetism.
38 CHAPTER 1. MAXWELL’S EQUATIONS: AN OVERVIEW
Chapter 2
~v ~
Z Z
PTo charges = ~+
∧ B) · ~v d3~x =
ρ(E ~ · ~v d3~x,
ρE
V c V
which includes the well known fact that the magnetic field does not exert work upon charges.
1
Some times they are not conserved, but even in many those cases they are useful for their variation can
be computed without difficult calculations.
39
40CHAPTER 2. ENERGY AND MOMENTUM OF THE ELECTROMAGNETIC FIELD.
This should be equal to the time derivative of the total energy of the electromagnetic field
configuration. Using now Maxwell’s evolution equation for E, ~
~
∂E ~ ∧B
= c∇ ~ − 4π J,
~
∂t
to replace the J~ in the expression above we get,
1 ~
∂E ~ d3~x = 1 ∂ ~ ~
Z Z
PFrom fields = ( ~ ∧ B)
− c∇ ~ ·E { ~ ∧ B)
(E · E) − 2c(∇ ~ · E}
~ d3~x,
4π V ∂t 8π V ∂t
~ ∧W
to handle the second term we use the identity, (∇ ~ )·V
~ − (∇
~ ∧V
~)·W
~ = −∇
~ · (V
~ ∧W
~ ),
and obtain,
1 Z ∂ ~ ~ ~ ∧ E)
~ ·B
~ −∇
~ · (E
~ ∧ B))}
~
PFrom fields = { (E · E) − 2c((∇ d3~x
8π v ∂t
~
We now use the evolution equation for B,
∂ ~ ~ ∧ E,
~
B = −c∇
∂t
and get,
1 ∂ ~ ~
Z
PFrom fields = { (E ·E+B ~ · B)
~ + 2c∇
~ · (E
~ ∧ B)}
~ d3~x.
8π V ∂t
Finally, we integrate by parts the last term using Gauss theorem and take out of the
integral sign the time derivative to obtain:
d d 1 Z ~ ~ ~ ~ 3
Z
~ · ~nd2~s.
E(t) = {E · E + B · B} d ~x = PFrom fields − S
dt dtZ8π V Z ∂V
~ ~ 3
= − J · Ed ~x − ~ · d2~s,
S
V ∂V
Remark: • Lorentz formula appeared only to assert that the work (density) done by unit
of time was,
W
Z
= − J~ · Ed~ 3~x.
δt V
In essence, it is used to “glue" the electromagnetic energy to other energies, like the mecha-
1
nical one, through the interaction of the systems. In particular this gives us the constant 8π
in front of the definition. Once we have this expression for the energy we can even apply to
the case where no sources are present, as we did in the previous chapter.
• The existence of surface term, the integral of the scalar product of the Poynting vector
~
S, with the unit normal to the boundary of the region, shows that in moving charges -or
fields- around we also produce electromagnetic radiation which leaves the region. That is,
after the redistribution of the fields in general there will be less or more energy stored in the
region than the one remaining after taking into account the work strictly needed for doing
the job.
• Care has to be exercised in interpreting Poynting vector as a radiation flux. This is not
a very precise interpretation, for instance the Maxwell solution,
E~ = (0, E, 0), E = cte, B ~ = (0, 0, B), B = cte., in Cartesian coordinates has a non-zero
Poynting vector, S ~ = ( c EB, 0, 0), but, since the solution is static there is no energy exchange
4π
and therefor no energy flux.
It is only theR integral of the Poynting vector the one which makes physical sense. In the
example above, ∂V S ~ · d2~s = 0.
• In the presence of a material obeying Ohm’s law, J~ = σ E, ~ we see that the presence of
any electric field in the region of that material would cause Joule dissipation,
dE
Z Z
= 3
−J~ · E
~ d3~x = −
3
~ ·E
σE ~ d3~x < 0,
dt lR lR
where we have integrated over the whole of lR3 , (although the only regions which contribute
to the integral are the ones where σ 6= 0), and assumed that the Poynting vector vanishes
near infinity.
The above argument tell us that in the presence of finite resistivity materials the stationary
solutions have zero electric field in the regions occupied by the material.
• We have only used the evolution equations. We shall see latter what is the role of the
constraint equations.
• If we integrate the above expression along a time interval [0, t] we get,
Z tZ Z tZ
EV (t) − EV (0) = −J~ · E
~ d3~xdt − ~ · d2~s.
S
0 V 0 ∂V
See figure XXX, thus the change in field energy in a fixed in time region V has two contri-
butions, a bulk one due to Joule work, and a boundary one, interpreted as the energy flux
leaving or entering the volume boundary as the time goes.
42CHAPTER 2. ENERGY AND MOMENTUM OF THE ELECTROMAGNETIC FIELD.
There is a minus sign because it is the force caused by the charge density upon the field.
In analogy with the case for the energy, we want to find an expression P~ V such that, when
contracted with any constant vector ~k, we get,
d(P~ V · ~k)
Z Z
~ ~ ~ ~ 3
= − (ρE + J/c ∧ B) · k d ~x − T~ ~ · n̂ d2 S,
dt V ∂V k
where T~ ~ is a vector which also depends linearly on ~k.
k
The last term in the above equation, the surface term, will be needed, for radiation will
not only take energy away, but also momentum. Let us see that P~ = c12 S ~ satisfies our desired
equation:
∂ P~ · ~k 1 ∂E~
~ + (E
~
~ ∧ ∂ B )} · ~k
= {( ∧ B)
∂t 4πc ∂t ∂t
1
= {[c∇~ ∧B~ − 4π J]
~ ∧B ~ + (Ê ∧ [−c∇ ~ ∧ E])}
~ · ~k
4πc
=
−1 ~
{ B ∧ (∇ ~ + 4π J~ ∧ B
~ ∧ B) ~ +E ~ ∧ (∇ ~ · ~k
~ ∧ E)}
4π c
−1 ~ ~ ~ ~ ~ · ∇)(
~ B)~ · ~k + (~k · ∇)(
~ E) ~ ·E ~ − (E~ · ∇)(
~ E) ~ · ~k} + (J/c
~ ∧ B)~ · ~k,
= {(k · ∇)(B) · B − (B
4π
−1 1 ~ ~ ~ ~ ~ · ∇)(
~ B ~ · ~k) + 1 (~k · ∇)(
~ E ~ · E)
~ − (E ~ · ∇)(
~ E ~ · ~k)} + (J/c
~ ∧ B)~ · ~k,
= { (k · ∇)(B · B) − (B
4π 2 2
∂ P~ · ~k −1 1 ~ ~ ~ ~ ~ · ∇)(
~ B ~ · ~k) − (B~ · ~k)(∇
~ · B)
~ +
= { (k · ∇)(B · B) − (B
∂t 4π 2
1 ~ ~ ~ ~ ~ · ∇)(
~ E ~ · ~k) − (E~ · ~k)(∇
~ · E)
~ + 4πρ(E~ · ~k)} + (J/c ~ · ~k).
~ ∧ B)
(k · ∇)(E · E) − (E
2
regrouping terms we get,
2.2. THE ELECTROMAGNETIC FIELD MOMENTUM 43
∂ P~ · ~k −1 ~ ~ ~ ~ ~ · B)]
~ − 2[E(
~ E~ · ~k) + B(
~ B~ · ~k)]} − (ρ + J~/c ∧ B)
~ E~ · ~k).
= ∇ · {k[(E · E) + (B
∂t 8π
Integrating over a volume V and using Gauss theorem, we get
d(P~ V · k) Z
~ · ( 1 {2[E(
~ E~ · ~k) + B( ~ · ~k)] − ~k(E
~ B ~ ·E
~ +B
~ · B)}
~
= {∇
dt V 8π
− (ρE~ + J/c
~ ∧ B)~ · ~k} d3~x
Z
= − (ρE ~ ∧ B) · ~k d3~x
~ + J/c
V
Z
1 ~ ~ ~ ~ B~ · ~k) − ~k(E
~ ·E
~ +B
~ · B)]
~ · ~n d2 S.
+ [2E(E · k) + 2B(
∂V 8π
A very important tool for disentangling physics phenomena has been the study of their symme-
tries. Through Noether’s theorem symmetries and conserved quantities are intimately related.
Observable conserved charges on physical systems imply symmetries which put constraints in
the possible mathematical models describing them.
What is a symmetry? It is a property of the equations which allows to find more solutions
once one is given. Sometimes, as we saw in the case of the wave equation, all solutions. More
precisely, it is a map which sends solutions into solutions and preserves certain structure, the
precise description of which is beyond the scope of this book.
~T
∂E ∂E~ ~
∂u ∂ E
(t, ~x) = (u, ~x) = (u, ~x)
∂t ∂t ∂t ∂u
∂E~
= (u, ~x) = c∇ ~ ∧ B(u,
~ ~ ~x)
~x) − 4π J(u,
∂u
~ ∧B
= c∇ ~ T (t, ~x) − 4π J~T (t, ~x).
For the other equations the procedure is identical and is left as an exercise. Time translation
symmetry is associated to energy conservation.
45
46 CHAPTER 3. THE SYMMETRIES OF MAXWELL’S EQUATIONS
Exercise: To see that there are equations which do not have such a symmetry consider the
E~ ~ ∧ E.
~ Show that this one does not have the above time symmetry. Does it
equation ∂∂t = t∇
have another symmetry?
~
∂E ~
~
L ∂E ~
(t, ~x) = (t, ~x − L)
∂t ∂t
∂E~
= (t, ~x′ )
∂t
~ ′ ∧ B(t,
= c∇ ~ ~x′ ) − 4π J(t,
~ ~x′ ),
~ Since
where we have defined ~x′ := ~x − L.
~
∂B ~
∂x′j ∂ B ~ ~
j ∂B ∂B
= = δi = ,
∂xi ∂xi ∂x′j ∂x′ j ∂x′i
~ ′ ∧ B(t,
∇ ~ ~x′ ) = ∇
~ ∧ B(t,
~ ~x′ (~x)) = ∇
~ ∧ B(t,
~ ~x − L)
~ =∇
~ ∧B
~ (t, ~x),
~
L
and the result follow immediately. We attribute this symmetry to the homogeneity of space.
It intimately related to momentum conservation.
3.3 Rotation
Space not only does not have any preferred origin, but neither have any preferred direction
or orientation, all of them are alike. Thus, if we rotate any given solution we should expect
to get another solution, and this should be so irrespectively of origin we choose to perform it.
3.3. ROTATION 47
Since in Maxwell’s equations we are dealing with vectorial quantities, we must be a bit more
careful and note that the vectors also must rotate! See figure.
The rotated solution would be
~ R (t, ~x), B
(E ~ R (t, ~x)) = (R(E(t,
~ R−1 (~x))), R(B(t,
~ R−1 (~x))))
(ρR (t, ~x), J~R (t, ~x)) = (ρ(t, R−1 (~x)), R(J(t,
~ R−1 (~x))))
~ ·E
∇ ~ R (t, ~x) = ∇
~ · R(E(t,~ R−1 (~x)))
∂ l ~
= R k [E(t, ~x′ )]k
∂xl
∂x′i l ∂ ~
= R k ′ i [E(t, ~x′ )]k
∂xl ∂x
∂ ~
= [R−1 ]i l Rl k ′ i [E(t, ~x′ )]k
∂x
∂ ~
= [E(t, ~x′ )]k
∂x′k
~ ′ · E(t,
= ∇ ~ ~x′ )
= 4πρ(t, ~x′ )
= 4πρ(t, R−1 (~x))
= 4πρR (t, ~x).
~ we have,
For the evolution equation for E
~R
∂E ∂ ~
(t, ~x) = ~ R−1 (~x))) = R( ∂ E (t, R−1 (~x)))
R(E(t,
∂t ∂t ∂t
′
~ ~ ~
= cR(∇ ∧ B(t, ~x ) − 4πRJ(t, R−1~x′ )
′
= cR(∇~ ′ ∧ B(t,
~ R−1 (~x)) − 4πRJ(t,
~ R−1 (~x)),
where the curl is with respect to the variable ~x′ = R−1 (~x). We write now the curl in compo-
nents,
e
~ ~x′ )]i = E ijk ∂ Bk (t, ~x′ ) = E ijk ∂x′ ∂Bk (t, ~x),
~ ′ ∧ B(t,
[∇
∂x′j ∂x j ∂xe
′ ∂xe
but xe = Re n x n and so ∂x′j
= Re j .
3.4. DISCRETE SYMMETRIES 49
On the other hand, since B ~ appears in the curl with its “index down", and B ~ ·B
~ = B k Bk
must be a scalar, we must have that B ~ ·B~ =B ~R · B
~ R , so the index down version of B
~ must
change with the inverse transformation, [BR ]k = [R ] k [B]l . Therefore, [B]k = Rkl [BR ]l ,
−1 l
[Notice that this is the same change as the one for the differential ∇i .] so,
~ ′ ∧ B]
[R∇ ~ i = Rli E ljk Re j Rn k [∇]
~ e [B
~ R ]n .
But
E ljk Ri l Re j Rn k = det(R)E ien ,
and for a rotation, detR = 1. Another way to see this is to notice that the vector product is
an invariant operation, and so the vector product of two rotated vectors should be the rotated
of the vector product of the vectors prior to the rotation, that is R(v) ∧ R(w) = R(v ∧ w),
but distinguishing between vectors with the index up to the ones with the index down in the
formulae, this implies,
E ijk Re j Rn k = [R−1 ]i m E men .
With this formula, we see that,
~ ′ ∧ B(t,
cR(∇ ~ ~x′ ) − 4π/cJ(t, ~ ~x′ ))
~ ∧B
= c∇ ~ R (t, ~x) − 4π J~R (t, ~x)
and the result follows. We attribute this symmetry to the isotropy of space.
Exercise: Check that: E ijk Rl i Rm j Rn k = det(R)E lmn , for any 3 × 3 matrix Rl i ; and that
det(R) = 1 for rotations. Hint: for the last part use that any rotation can be expressed terms
of five known consecutive transformations, each one of determinant one. Or realize that the
final expression is independent of the coordinate system and choose the coordinates after
giving the rotation so as to leave invariant one of the axis. What is the general expression for
the determinant of a n × n matrix?
as if we were watching a movie starting from the end. As in the case of rotations, in this
case the symmetry does not only involve the change t → −t, but it also must involve same
~ and B.
change on the fields, E ~ For simplicity in the presentation we shall assume each of one
these vectors changes just by a constant factor. Since the application of two consecutive time
inversion is the identity (t → −t, (−t) → −(−t) = t), those constant factors must have unit
square, and so they must be +1 or −1. Thus, we assume the symmetry has the form:
~ I (t, ~x), B
(E ~ I (t, ~x)) = (aE E(−t,
~ ~
~x), aB B(−t, ~x))
with a2E = a2B = 1, both real. Using now the equations we shall fix these values.
Since this symmetry concerns only with the time variable the constraint equations must
hold identically, so in particular we must have the obvious result, ρI (t, ~x) = aE ρ(−t, ~x).
Since
∂E~I ~
∂E
(t, ~x) = aE (−t, ~x)
∂t ∂t
∂(−t) ∂ E ~
= aE (−t, ~x)
∂t ∂(−t)
~
∂E
= −aE (u, ~x),
∂u
we have,
∂E~I
~ ∧ B(u,
(t, ~x) = −aE [c∇ ~ ~ ~x)]
~x) − 4π J(u,
∂t
~ ∧B
= −aE [aB c∇ ~ I (t, ~x) − 4π J(−t,
~ ~x)]
similarly,
∂B~I ∂B~
(t, ~x) = −aB (u, ~x)
∂t ∂u
= −aB [−c∇~ ∧ E(u,
~ ~x)]
~ ∧E
= +aB aE [c∇ ~ I (t, ~x)]
linear transformation which changes -inverts- the ẑ axis and leave invariant the other two, so
it is like a rotation, but with determinant equal to −1.
Exercise: Check this last assertion writing down explicitly the inversion matrix. Namely,
~ Invz (t, ~x) = RInvz E(t,
(E ~ RInvz ~x), H
~ Invz (t, ~x) = −RInvz H(t,
~ RInvz ~x)).
Exercise: Check that inverting two axis is a transformation with determinant one. Check
that indeed this can be accomplished by a rotation.
Since the transformation is linear we do not have to redo the calculation we made for
rotations, it is similar to it except that since the determinant of the transformation is minus
one, in analogy with the time inversion symmetry, B ~ must have an extra −1 change!
As we shall see latter this comes about because B ~ itself is in fact made out of the curl
of a proper vector, i.e. one which changes like E ~ under a space inversion, and so, since the
Levi-Civita symbol changes sign under space inversion, so does B. ~
d2~x1 (t)
m = F~ 1 (~x1 (t) − ~x2 (t))
dt2
d2~x2 (t)
m = F~ 2 (~x1 (t) − ~x2 (t)),
dt2
as it would be the case for motion under the influence of their gravitational attraction. The
Galilean principle of relativity, for this case would then imply that (~x′1 (t), ~x′2 (t)) = (~x1 (t) −
~v t, ~x2 (t) − ~vt) is also a solution, if (~x′1 (t), ~x′2 (t)) was one. Indeed,
d2~x′1 d2
m = m (~x1 (t) − ~v t)
dt2 dt2
d2
= m 2 ~x1 (t)
dt
~
= F (~x1 (t) − ~x2 (t))
= F~ (~x1 (t) − ~v t − (~x2 (t) − ~v t))
= F~ (~x′1 (t) − ~x′2 (t)).
Exercise: Check this for the case of two particles interacting through a spring of Hooke
52 CHAPTER 3. THE SYMMETRIES OF MAXWELL’S EQUATIONS
constant k. That is, find the solution corresponding to the particles (and the spring) traveling
with constant velocity ~v .
In the case of electromagnetism this principle would imply, for instance, that if one has
two equal capacitors with the same potential difference between its plates, and one is in the
surface of the earth and the other over a ship sailing nearby, then the forces between the
plates should be equal for both capacitors. In Galilean terms we are then saying that, if
~ ~x), B(t,
(E(t, ~ ~x)) is a solution, then,
~ ~x − ~v t) + G B(t,
(Eg (t, ~x) := F~v E(t, ~ ~x − ~v t), B
~ j (t, ~x) := F̃ E(t,
~ ~x − ~v t) + G̃ B(t,
~ ~x − ~v t)
~v ~v ~v
should be also a solution, for some set of linear transformations, (F~v , G~v , F̃~v , G̃~v ) where
~ i = F i [E]
[F~v E] ~ j , etc.
~v j
Since ~v is a constant velocity, not only in time but also in space, and because of the
linearity of Maxwell’s equations the linear transformations can neither depend on the fields
E~ nor B,~ they must be constant in both time and space and only depend on ~v. Furthermore,
since the transformation should be the identity for vanishing velocities,
Notice that we have to allow for this type of transformation, which mixes both E~ and B,~
for in the above example the capacitor’s plates would be moving, and so generating currents,
this in turn, would generate magnetic fields, which in turn, would contribute to the total
force.
Let us see now whether these new fields we have defined also satisfy Maxwell’s equations.
To simplify the calculation we consider only vacuum solutions (or just look at the solution in
an empty region), and decouple them taking a time derivative to the equations to get,
∂2 ~ ~ =0
E − c2 ∆E
∂t2
∂2 ~ ~ = 0.
B − c2 ∆B
∂t2
2 2
(∂t2 E ~ g ) = F ( ∂ E(t,
~ g − c2 ∆E ~ ~x′ )) + G ( ∂ B(t,
~ ~x′ ) − c2 ∆′ E(t, ~ ~x′ ) − c2 ∆′ B(t,
~ ~x′ ))
~v ∂t2 ~v ∂t2
~ ′ )(F (∇
−2c(~v · ∇ ~ ′ ∧ B(t,
~ ~x′ )) − G (∇~ ∧ E(t,
~ ~x′ )))
~v ~v
~ ′ )(~v · ∇
+(~v · ∇ ~ ′ )(F (E(t,
~ ~x′ ) + G (B(t, ~x′ ))) = 0
~v ~v
Consider now this relation at t = 0. We see that, since there are no time derivatives, this
is a relation among the initial data, and that this relation is different from the constraint
equations (for it involves curls). So this relation is enforcing further constraints to the ones
already imposed by Maxwell’s equations. So there are plenty of solutions to Maxwell’s equa-
tions which do not conform with Galilean relativity, namely all solutions coming form initial
data sets not satisfying the above relatio 1 .
Solution at
rest in syst.
(t,x) Solution at
rest in syst.
(t,x´)
We are in trouble, either Galilean Relativity is wrong or Maxwell’s equations are wrong.
The two together are inconsistent. One has two logically consistent formalism, only exper-
iments can decide. The incredible amount of phenomena described by Maxwell’s equations
tell us that it is the Galilean Principle of Relativity the one which must be abandoned. We
shall come back to this point latter in the book.
One could have guessed the result of our calculation, for we already remarked that in
Maxwell’s equation there appears a constant with dimensions of velocity. At that moment
we should have asked: What is moving with that velocity? And: With respect to what are
1
To see this in more detail, assuming our transformations are smooth note that F~v = Id + O(| ~v |) and
G~v = O(| ~v |), since the relation has to vanish order by order we see that to first order we must have,
~ ′ )(∇
(~v · ∇ ~ ′ ∧ B(t,
~ ~x′ )) = 0, which is clearly an extra constraint.
54 CHAPTER 3. THE SYMMETRIES OF MAXWELL’S EQUATIONS
we computing that velocity? The answer to the first was already given, it is the velocity
of propagation of electromagnetic waves. The second will be answered latter in the book.
For the moment we shall content ourselves to seeing how the constant c appears on Maxwell
equation.
[F ] [F ]T 2
[k2 ] = 2 = .
[I] [q]2
Therefore,
[k1] L2
= 2 = [c]2
[k2] T
Of course these laws not only allow to know the dimensions of c, put also its value! In
spite of the fact that they are basically stationary experiments, where nothing travels to the
speed c. In fact, when Maxwell unified electricity and magnetism, and found this quantity
noticed that its value was close to the speed of light! It took some time to relate both, for
the optical phenomena were really not thought, at that time, as having anything to do with
the electric or magnetic phenomena.
2
Square brackets here means “the dimension of" the quantity inside them.
3.6. OTHER SYMMETRIES 55
Exercise: Check that in vacuum Maxwell’s equations also have the symmetry:
~˜ = cos(α)E
E ~ + sin(α)B
~
~˜ = cos(α)B
B ~ − sin(α)E.
~ (3.1)
Generalize it further.
Exercise: Check that the energy does not change under the above symmetry.
Exercise: Use the vacuum Maxwell’s equations to find the time derivative and divergence of
~ := E
W ~ + iB
~
56 CHAPTER 3. THE SYMMETRIES OF MAXWELL’S EQUATIONS
Chapter 4
that is, they are independent of time. Naturally for this to happens the sources must also be
time independent. For these solutions all time derivative vanish and so the equations becomes,
~ ∧ B(~
c∇ ~ x) = 4π J(~
~ x)
~ ∧ E(~
c∇ ~ x) = 0
~ · E(~
∇ ~ x) = 4πρ(~x)
~ · B(~
∇ ~ x) = 0.
Note that the equations are now decoupled, we have two equations for E(~ ~ x) and two
~ ∂ρ
equations for B(~x). Note also that since ∂t = 0, the continuity equation implies ∇~ · J(~
~ x) = 0,
a necessary condition for the first equation above to have a solution. If furthermore we require
the solutions to be invariant under time inversion, that is,
~ x) = J(~
we see that the independence of the solution on the time variable imply B(~ ~ x) = 0, and
~ remain. We shall call these solutions static solutions, and
so only the two equations for E
their study electrostatics. The study of the stationary solution is completed by studying
~ x)) and is called magnetostatics.
the remaining two equations (for B(~
57
58 CHAPTER 4. STATIONARY SOLUTIONS: ELECTROSTATICS
4.2 Electrostatics
The equations of electrostatics are:
∇~ ∧ E(~
~ x) = 0 , ∇
~ · E(~
~ x) = 4πρ(~x).
To solve them we make the ansatz, E ~ = −∇φ.
~ Then it is ease to see that the first equation
~ ~ ~ ~
is trivially satisfied, ∇ ∧ E = −∇ ∧ ∇φ ≡ 0, while the second becomes Poison’s equation:
∆φ(~x) = −4πρ(~x).
We shall discuss this equation during the rest of the chapter.
How general is the ansatz we have used? Or in other words, are all solutions of the
electrostatic equations gradients of some functions?
The answer to this problem is completely known and there are precise conditions which
guarantee uniqueness. When these conditions are not meet, there are other solutions. In
most cases in electrostatics the ansatz suffices, but we shall see in magnetostatics an example
where it fails.
and this, we claim, is the general solution to Poison’s equation in lR3 . It is instructive to
check this directly, we do that next:
ρ(~x′ ) 1
Z Z
3 ′
∆φ(~x) = 3
∆( ′ ) d ~
x = 3
∆( x′ ) d3~x′
′ )ρ(~
lR | ~x − ~x | lR | ~
x − ~
x |
But
1 ~ ~x − ~x′ 3 (~x − ~x′ ) · (~x − ~x′ )
∆( ) = −∇ · ( )=− +3 =0
| ~x − ~x′ | | ~x − ~x′ |3 | ~x − ~x′ |3 | ~x − ~x′ |5
if ~x 6= ~x′ .
Therefore, if ∆ 1
′ were a function at ~x = ~x′ , as is for ~x 6= ~x′ , then the integral above
|~
x−~x |
3 1
would give zero, and we would have, ∆φ(~x) = 0! Thus, we conclude that ∆ ′ is not a
|~
x−~x |
1
function. The problem arises because we are trying to take derivatives of a function, ( ′ ),
|~
x−~x |
at a point where it is not differentiable, and then we further pretend to integrate is as if it
were.
Thus, we must proceed with certain care and treat the equation in the sense of distribu-
tions. That is, we think about it as a map, as applied to a smooth, compactly supported, test
function ψ(~x), into the reals. We define,
Z Z
ρ(~x′ )ψ(~x) 3 ′ 3
TKρ (ψ) := d ~x d ~x (4.2)
lR3 lR3 |~x − ~x′ |
We want to show now that,
Z
∆TKρ (ψ) = −4πTρ (ψ) := −4π ρ(~x)ψ(~x) d3~x. (4.3)
lR3
But by definition,
3
The integral of a function which zero everywhere except at a point vanishes.
60 CHAPTER 4. STATIONARY SOLUTIONS: ELECTROSTATICS
Z Z ~ y + ~x′ )
~y · ∇ψ(~
∆TKρ (ψ) = ρ(~x′ ) d3~x′ d3~y
lR3 lR 3 |~y |3
Z Z ∞ Z
= 3
ρ(~x′ )∂r ψ(r~n + ~x′ ) drdΩ d3~x′
ZlR Z0 S2
1
Z
3
x′ ) d3~x′ = −4πψ(~x) ∀ smooth and compactly supported φ(~x)
′ ∆ψ(~ (4.7)
lR |~x − ~x |
Notice that the calculation done would imply (if the Dirac’s Delta were a function),
Z
3
δ(~x − ~x′ )ρ(~x′ ) d3~x′ = ρ(~x),
lR
for ρ(~x) sufficiently smooth. This is a convenient abuse of notation for doing some calculations,
but many times leads to nonsense.
4.2. ELECTROSTATICS 61
4.2.3 Uniqueness
So far we have found a solution corresponding to an isolated system of charges, how unique
is that solution? Are there more? The answer is that clearly there are lots of solutions, any
solution to Laplace’s equation (the homogeneous, i.e. vacuum, Poison’s equation) can be also
added to the one already found and the result will also be a solution. For instance, we could
add:
φ0 (~x) = φ0 = cnst,
φ1 (~x) = ax + by + cz,
φ2 (~x) = a1 yz + a2 xz + a3 xy + ã1 (y 2 − z 2 ) + ã2 (x2 − z 2 ) + ã3 (x2 − y 2 ),
Exercise: Show that in the class φ2 (~x) there are only five linearly independent terms.
Exercise: Find all (linearly independent) solutions which are cubic in the combination
x, y, z, that is terms with xα , y β , z γ and α + β + γ = 3, with α, β, γ positive or null integers.
Thus, to single out a unique solution one must impose boundary conditions, in this
case, asymptotic conditions, which are given by the physics of the problem. As we already
saw, the condition in this case should be: φ(~x) → 0 as | ~x |→ ∞. If we do impose them, we
do get uniqueness.
62 CHAPTER 4. STATIONARY SOLUTIONS: ELECTROSTATICS
Theorem 4.1 If φ(~x) decays sufficiently fast at infinity, and satisfies Poisons equation 4 ,
∆φ(~x) = 4πρ(~x).
ρ(~x′ )
Z
φ(~x) = d3~x′ .
| ~x − ~x′ |
Proof: Assume that there are two solutions with the required decay properties, φ1 (~x) and
φ2 (~x). Then their difference, ϕ(~x) = φ1 (~x) − φ2 (~x) also has the required decay properties and
satisfies Laplace’s equation,
∆ϕ(~x) = 0
Multiply this equation by −ϕ and integrate over all space,
Z
0 = 3
−ϕ∆ϕ d3~x
ZlR
= ~ · ∇ϕ
−ϕ∇ ~ d3~x
3
ZlR
= ~ · (−ϕ∇ϕ)
[∇ ~ + ∇ϕ
~ · ∇ϕ]
~ d3~x
3
ZlR Z
= ~ · ∇ϕ
∇ϕ ~ d3~x − lim ~ d2 S,
−ϕ ~n · ∇ϕ
3
lR r→∞ s(r)
where in the second line we have defined the inverse of a curve, namely the curve which runs
in the opposite direction, from ~x to the initial point. The union of the two curves results in
a closed loop that we have called γ2−1 ∗ γ1 . It is the boundary of a surface, which we call S.
Stokes theorem is used in the fourth line to express that circulation integral in terms of a
~ Finally, the fact that E
surface integral of the curl of E. ~ is curl-free has been used to conclude
that the difference vanishes. Thus, we are constructing a well-defined function. If we take
different curves, so as to cover the whole of lR3 with them, all of them starting from the same
point, we would get a function on the whole of lR3 .
~ = −∇φ.
It remains to see that this function satisfies E ~ To see this connect any two nearby
points by a curve, γ(s), we then have,
~ x) = −E(~
∇φ(~ ~ x),
and so that in lR3 we can always write a curl-free vector field as the gradient of a function.
64 CHAPTER 4. STATIONARY SOLUTIONS: ELECTROSTATICS
4.3 Conductors
So far we have shown that given a system or distribution of charges in an isolated region of
space we have a unique solution, and we have an explicit formula for it. One could think
that now it is a question of just doing the integral for the given source. Unfortunately this is
not so for in most cases we do not know where the charges are, for charges, if not completely
fixed, are going to move and accommodate until an equilibrium configuration is found, and
that equilibrium configuration depends on the electric field.
To see an example of the type of problem we envision consider any object, of copper say,
of which you know its shape and which you know it has a given amount of charge, q. One
does not know before finding the solution where the charges will be, for since the object is a
conductor, the charges would move in its interior until all electric field inside vanishes, that
is, until all forces upon them vanish, which is in this case the equilibrium configuration.
So this problem can not be solved using the formulas we gave. Although it will satisfy the
equations!
How do we solve it? Let us see what we know about the solution. First we know that
whatever the solution is, it should go to zero at infinity. Second that inside the conductor
~ = 0 or φ(~x) = φ0 inside V , where V is the volume
there should no be electric fields, that is ∇φ
occupied by it. But since the charges are in the conductor and since ∆φ = −4πρ, we see that
they must be at the surface of it. This implies, in the mathematical approximation we are
making of a conductor, that ∇φ ~ must be discontinuous across the surface of the conductor,
for ρ is discontinuous there, but it is only discontinuous along the normal to the conductor,
so one expects that at the conductor’s surface only the normal component of ∇φ ~ will be
discontinuous. To see this we consider the following line integral along γ, see figure.
Z
L= ~l · ∇φds,
~
γ
where γ = γ(s), l = dds~x , that is “the velocity at which we circulate along γ if the parameter s
were the time".
4.3. CONDUCTORS 65
CONDUCTOR VACUUM
γ2
γ1 γ3
γ4
The value of that integral is the difference of the value of φ at the end of the curve and
the value of φ at the beginning of the curve, so, since in this case the curve is closed we have,
L = 0. That is
Z Z Z Z Z
0= ~l · ∇φ
~ dS = ~l · ∇φ
~ dS + ~l · ∇φ
~ dS + ~l · ∇φ
~ dS + ~l · ∇φ
~ dS,
γ γ1 γ2 γ3 γ4
where we have split the integral into four ones, we now consider the limit when the curve
approaches more and more the boundary of the conductor, in that case, γ1 → γ3 and γ2 →
~ discontinuous but finite, we see that the integrals along
0, γ4 → 0. Since we are assuming ∇φ
γ2 and γ4 go to zero, for γ2 and γ4 go to zero and so,
Z
0= ~l · (∇φ
~ i − ∇φ
~ e ) dS,
γ3
CONDUCTOR
VACUUM
normal
where ~n is the unit normal to the surface of the box, denoted by S, pointing towards the
outside of the box.
Using Gauss theorem we see that its value is
Z Z
~ · (∇φ)
∇ ~ d3~x = −4π ρ(~x) d3~x
box box
Taking now the limit when the faces of the box go flat on the boundary we see that the rim
of the box in that limit does not contribute. Thus,
Z Z Z Z
~ i+
n̂i · ((∇φ) ~ e ) d2 s =
n̂e (∇φ) ~ i − (∇φ)
n̂e · ((∇φ) ~ e ) d2 S = −4π σ(~xs ) d2 S,
S S S S
5
where σ(~xs ) is now the surface charge density, that is ρ(~x) = δ(~x − ~xs ) · σ(~xs ).
~ i = 0 and so we have,
In the present case, (∇φ)
Z Z Z
~ e d2 S = −
n̂e · (∇φ) ~ d2 S = −4π
n̂ · E σ(~xs ) d2 S.
S S S
2.
R
~ d2 S = −4πq.
n̂ · ∇φ
S
3. φ(~x) → 0 when | ~x |→ ∞.
Are these three conditions enough to determine φ(~x) everywhere? The answer is in the
following theorem:
2.
R
~ d2 S = −4πq
n̂ · ∇φ
S
3. φ(~x) → 0 | ~x |→ ∞
Proof: We shall only prove uniqueness, to show existence amounts to know the core of
the theory of elliptic equations, Let φ1 and φ2 two solutions to ∆φi = 0 in D = lR3 − V , where
V is the volume occupied by conductor.
D = lR3 − V
Conductor n̂
S = ∂V
Thus, ϕ = φ1 − φ2 satisfies ϕ |s = c = c1 − c2 ,
R
~ = 0 and ϕ(~x) → 0 as | ~x |→ ∞, and
n̂ · ∇ϕ
s
furthermore 6 ∆ϕ = 0 in D.
6
Note that the same proof of uniqueness holds if there are (fixed) charges in the region D, for the difference,
ϕ, also satisfies Laplace’s equation.
68 CHAPTER 4. STATIONARY SOLUTIONS: ELECTROSTATICS
Z
0 = [−ϕ∆ϕ] d3~x
ZD
= ~ · (ϕ∇ϕ)
[−∇ ~ + ∇ϕ
~ · ∇ϕ]
~ d3~x
ZD Z Z
= ~ · ∇ϕ]
[∇ϕ ~ d3~x − ~ d2 S +
ϕ n̂ · ∇ϕ ~ d2 S.
ϕ n̂ · ∇ϕ
D S(∞) S
Where in the second step we have used Gauss theorem, and in the integral over S we use
as ~n the outer normal. From the decay assumptions we have that the first surface integral
(at infinity) vanishes, while the second gives,
Z Z
~ d2 S = c
ϕn̂ · ∇ϕ ~ d2 S = 0
n̂ · ∇ϕ
S s
Thus,
Z
~ · ∇ϕ]
[∇ϕ ~ d3~x = 0
D
~ = 0 in D, and ϕ(~x) = cnst in D. But ϕ(~x) → 0 as | ~x |→ ∞ and so ϕ(~x) = 0 in
and so ∇ϕ
D.
We are given a ball of radius a, say, and charge Q and ask to find the potential outside it.
We know that at the surface’s ball the electric field must be normal to it. That is, taking
the coordinates origin at the center of the ball, it must be radial so, φ(~x)||~x|=a = φ0 , a
constant.
Thus the exterior problem is:
3.
R
~ d2 S = −4πQ
n̂ · ∇φ
S
4. φ(~x) → 0 | ~x |→ ∞,
solution, and φ(r) = 1r Since the solution must go to zero at infinity the constant contribution
is ruled out and so φ(r) = rc . To find the value for c we must impose the condition on the
total charge, namely,
Z
~ d2 S = −
Z
c
4πQ = − n̂ · ∇φ ∂r ( )|r=a r 2 d2 Ω = 4πc
S S r
Q
Thus the solution is φ(~x) = r , and the value of the potential at the boundary is φ0 = Qa .
Notice that the relation among the potential and the charge is Q = aφ0 so the sphere
capacity is C = a.
Exercise: Find an example that shows that, even if φ is known everywhere outside a surface
S, we do not know its value inside, nor can determine ρ inside. This is particularly strange
because if ρ were an analytic function, then φ would be analytic also, and so the knowledge of
φ in an arbitrary open neighborhood of any point would determine completely φ everywhere.
Exercise: Show that nevertheless some knowledge about ρ inside S we can obtain from the
knowledge of φ at S, although indirectly. Hint: Use Coulomb’s Law.
This seems to be complicated, but we can reduce it to simpler problems: We look for
N + 1 solutions. The first one satisfying,
1. φ|Si = 0, i = 1..N
2. ∆φ = −4πρ
70 CHAPTER 4. STATIONARY SOLUTIONS: ELECTROSTATICS
1. φj |Sj = 1
2. φj |Si = 0, j 6= i
3. ∆φj = 0
They also have unique solutions, φ(~x)i . We now scale them by the constants φi0 so that they
satisfy the required boundary conditions. Then,
φi0 φi (~x)
X
φ(~x) = φI (~x) +
i
satisfy the required boundary conditions for the first problem. Shortly we shall see how to do
it when the total charges at each conductor is given.
We could have also split the sources in simpler ones and solve a problem for each one of
them and then sum all of them together. This is also very handy in solving some problems.
Exercise: Show using these techniques, and Gauss theorem the solution to the following
situation: There are two infinite flat conducting slabs held parallel to each other at a distance
L. Each one of thickness l1 and l2 respectively and charges q1 and q2 . Show:
• The charge densities on each of the internal faces of each slab are equal and of opposite
sign.
• The charge densities on each of the two external faces are equal.
• The value of the fields and charge densities are independent of the length.
• What happens when the two charges are equal but opposite?
Hint: Think on the charge densities and forget about the conductors, except to impose
the conditions that the field is zero inside them.
resulting function would also be a solution to Poison’s equations with a charge at ~x = (r, 0, 0)
and would also vanish at the surface of radius a. But there could be at most one solution
with those characteristics, so the rotated solution must be the same as the original, thus it
must have the required symmetry.
~x
a n̂
q′ q
r′ r
If we pretend to continue the solution inside the sphere, then at the symmetry axis the
equipotential surfaces can only become points or cross orthonormal to it. This suggests that
we look there for a solution with point-like sources along the symmetry axis. We try with
just one charge there, say at the point ~xq′ = (r ′ , 0, 0). Then, the potential due to these two
charges is:
q q′
φ(~x) = + ,
| ~x − rx̂ | | ~x − r ′ x̂ |
where x̂ = (1, 0, 0).
At the surface of the sphere we have ~x = an̂ and so,
q q′
φ(~x) = +
| an̂ − rx̂ | | an̂ − r ′ x̂ |
q q′
= √ + √
a2 + r 2 − 2a rn̂ · x̂ a2 + r ′2 − 2a r ′ n̂ · x̂
q q′
= q + q . (4.9)
a 1 + r 2 /a2 − 2r n̂·x̂
a
r ′ 1 + a2 /r ′2 − 2a n̂·x̂
r′
q′ q
′
=− ,
r a
and
a r
′
= ,
r a
72 CHAPTER 4. STATIONARY SOLUTIONS: ELECTROSTATICS
Figure 4.5: Contour plot on the z = 0 plane of the potential of two charges as in the problem.
′ 2
that is q ′ = − qra = − qa
r
, and r ′ = ar < a. See figure (4.5) with the contour plots in the z = 0
plane.
What have we done? We have constructed a solution to Poison’s equation which vanishes
at | ~x |= a and which outside the sphere of radius a has just a single point-like source at
~xq = (r, 0, 0). But this is what we were looking for! For the solution to the problem outside
the sphere is unique.
What we have done is to mimic the effect of the surface charge density at the surface of
the conducting sphere with a point-like source in its interior, by giving to it a special location
and strength. Indeed, we can now use the formula +4πσ = −n̂ · ∇φ ~ |S to compute such
distribution.
qa
4πσ d2 S = q ′ =
R
Exercise: Compute σ using the formula above and check that r=a r
.
Exercise: How would you find a solution for the same problem but assuming now that the
sphere is at potential φ 6= V ? Compute the total charge that would have to be at the sphere.
Note that it is easy to compute the total force the sphere exerts on the charge q, it is just
the force q ′ would exert on q would it exist, namely
F~ = (F, 0, 0),
−q 2 a −q 2 ra
F = =
r | r − a2 /r |2 (r 2 − a2 )2
Exercise: Use the principle of action-reaction to compute the total force exerted by the
4.4. CAPACITIES 73
charge q on the conducting sphere. Compare with the result obtained using Maxwell stress
tensor.
4.4 Capacities
Consider a configuration of conductors as shown in the figure (4.6) below, on an otherwise
empty space.
Imagine that to get to this distribution we fixed the potentials, V i , i = 1...n on each
conductor and solved,
D = lR3 − Vi
S
i
V1
V2
S1 = ∂V1
V3
Correspondingly there would be fixed ammount of charges on the surface of each conductor,
given by,
1
Z
Qi = ~ dS 2
n̂ · ∇φ (4.12)
4π ∂Si
If we had fixed another set of potentials, Ṽ i , then we would have obtained another poten-
tial, φ̃ and correspondingly another set of charges Q̃i . Now, if we consider the configuration
where the potentials at the conductors are given by V̂ i = V i + cṼ i where c is any constant,
then the the potential field would be φ̂ = φ + cφ̃. Indeed, it follows from the linearity of the
equations and of the boundary conditions that φ̂(~x) satisfies the equation and the boundary
conditions, but uniqueness of the solutions implies this is the right one. But then, from the
linear dependence of the charges on the potential field, (4.12), the total charge would be,
Q̂i = Qi + cQ̃i . That is, there exists a linear relation between potentials at conductors and
their charges! Thus, there must exist a matrix, Cij such that,
Qi = Cij V i (4.13)
This matrix, called the capacities matrix, can only depend on the geometrical configuration
of the conductors distribution.
Exercise: Show that this matrix in invertible by setting the problem of fixing the charges
and finding the potentials.
Exercise: Show that the matrix is symmetric, that is, Cij = Cji. Hint: choose any pair
of conductors, say the i − th and the j − th, solve the potential problem for the boundary
conditions {V i = 1, v k = 0, k 6= i} call that solution φi . Do the same for the j − th conductor
and call the solution so obtained, φj . We then have, for the j − th solution,
1 Z ~ j dS 2 = 1
Z
~ j dS 2 ,
Cik V k = Cij V j = Cij = n̂ · ∇φ φi n̂ · ∇φ (4.14)
4π Si 4π Si
4.4. CAPACITIES 75
where in the last equality we have used that φi |Si = 1. Find a similar formula for Cji, subtract
the first from the second and use the divergence theorem to convert the two surface integrals
in a volume one. Finally show that the volume integral vanishes.
76 CHAPTER 4. STATIONARY SOLUTIONS: ELECTROSTATICS
Chapter 5
∂2 ∂2 ∂2
∆φ = φ + φ + φ.
∂x2 ∂y 2 ∂z 2
Thus, if we assume a solution of the form,
∂2 ∂2 ∂2
∆φ = Y (y)Z(z) X(x) + X(x)Z(z) Y (y) + X(x)Y (y) Z(z)
∂x2 ∂y 2 ∂z 2
or
1 ∂2 1 ∂2 1 ∂2
∆φ = [ X(x) + Y (y) + Z(z)]φ(x, y, z)
X(x) ∂x2 Y (y) ∂y 2 Z(z) ∂z 2
77
78CHAPTER 5. STATIC PROBLEMS – CONSTRUCTING INTERESTING SOLUTIONS
b y
Figure 5.1: A conducting box with only its upper face at non-zero potential.
Since each term inside the brackets is a function of a different variable, each term must be
constant and the sum of them must be zero. That is, we need,
∂2 ∂2 1 ∂2
X(x) = α2 X(x) Y (y) = β 2 Y (y) Z(z) = γ 2 Z(z)
∂x2 ∂y 2 Z(z) ∂z 2
with the constants satisfying,
α2 + β 2 + γ 2 = 0.
We have used squared quantities for later convenience, there is no loss of generality on that for
we shall consider them as complex numbers. We shall see later how to construct the general
solution as linear combination of these basic ones.
Consider any of the above equations, say,
∂2
2
X(x) = α2 X(x)
∂x
The general solution to it is given by, 1
and
A+ + A− = 0 A+ eαa + A− e−αa = 0.
That is,
A+ (eαa − e−αa ) = 0.
Thus, to have a nontrivial solution we need,
3
Thus, the coefficients Cnm just proportional to the Fourier coefficients of V (x, y),
4 Z aZ b
πnx πmy
Cnm = V (x, y) sin( ) sin( ) dx dy.
ab sinh(γnm c) 0 0 a b
Example: Box with all faces at zero potential except one at V0 . We take the
coordinate system with the axis perpendicular to the faces and centered at one of boxes
corner, so that the non-zero potential face corresponds to z = c. The Fourier coefficients are
obtained from the above formula or from a table if one considers a potential function defined
in the rectangle [−a, a] × [−b, b] in such a way that is odd under x → −x or y → −y. In any
case, the coefficients are:
16V0
Cnm = n, m odd Cnm = 0, all other cases
π 2 nm
Figure 5.2: The value of the potential at z=c with 10 Fourier modes
3
To see this consider the Fourier transform on the rectangle [−a, a]×[−b, b] and describe only odd functions
in both variables, thus you need only the sin function and can dispense all the cos ones. The factor 4 that
appears in the formula is due to the fact that we perform the integrals just in one quarter of the whole
rectangle.
5.1. METHOD OF SEPARATION OF VARIABLES IN CARTESIAN COORDINATES 81
Figure 5.4: The value of the potential at y=b/2 (5) for a wide box
Figure 5.5: The value of the potential at y=5 with only 3 Fourier modes
82CHAPTER 5. STATIC PROBLEMS – CONSTRUCTING INTERESTING SOLUTIONS
1 ∂2 1 ∂ ∂φ 1 ∂2φ
∆φ = (rφ) + (sin θ ) +
r ∂r 2 r 2 sin θ ∂θ ∂θ r 2 sin2 θ ∂ϕ2
We shall look for solutions to Laplace’s equation using the method of separation of vari-
ables, with the hope to find enough of them to account, using linear combination, for all
possible solution.
To this end we propose solutions of the form φ(r, θ, ϕ) = U r(r) P (θ)Q(ϕ). Substituting in
2 sin2 θ
the above formula for the Laplacian and multiplying the result by rφ(r,θ,ϕ) we obtain:
1 d2 U 1 d dP 1 d2 Q
r 2 sin2 θ[ + (sin θ )] + = 0.
U dr 2 P r 2 sin θ dθ dθ Q dϕ2
Since the first term can only be a function of r and θ, while the second only a function of ϕ
and r, we conclude that in order to cancel each one of them must be constant, thus:
d2 Q
2
= α2 Q, for some constant α,
dϕ
that is Q(ϕ) = Qα eαϕ . If we look for solutions defined for all angles ϕε[0, 2π], then we must
have Q(ϕ) = Q(ϕ + 2π) = Q(ϕ)e2πα and therefore α = ±im, that is α2 = −m2 , and the
general solution is:
Q(ϕ) = Q+ me
imϕ
+ Q−me
−imϕ
,
where Q+ −
m and Qm are two arbitrary constants to be determined latter when imposing bound-
ary conditions.
The rest of the equation becomes then,
r 2 sin2 θ d2 U sin θ d dP
2
+ (sin θ )] + α2 = 0.
U dr P dθ dθ
Thus, since the first term is the only one that depends on r, we conclude that
d2 U βU
2
= 2, for some constant β.
dr r
The solutions to this equation are:
d dP
(1 − x2 ) ((1 − x2 ) ) + (α2 + l(l + 1)(1 − x2 ))P = 0,
dx dx
or
d dP α2
((1 − x2 ) ) + ( + l(l + 1)P = 0,
dx dx 1 − x2
which is known as Legendre’s equation.
We shall first treat the case α = 0, that is we shall consider only solutions with azimuthal
symmetry. Latter we shall generalize to α2 6= 0. In this case the equation becomes,
d dP
((1 − x2 ) ) + l(l + 1)P = 0.
dx dx
We shall look for solutions valid for the whole interval of θ, namely [0, π], that is for
solutions valid in the x-interval [−1, 1]. Assuming that in that interval the solutions admit a
convergent power series representation,
∞
aj xj ,
X
P (x) =
j=0
j(j + 1) − l(l + 1)
aj+2 = aj
(j + 1)(j + 2)
84CHAPTER 5. STATIC PROBLEMS – CONSTRUCTING INTERESTING SOLUTIONS
Thus, we can split the solution into two series, one with a0 6= 0, a1 = 0 and the other with
a0 = 0, a1 6= 0. One is even in x, the other odd, so they are linearly independent solutions.
Since every second order ordinary differential equation has two linearly independent solutions
at every point, we know we have found all possible solutions.
j
For j > l, |aj+2 | < j+2 |aj | < |aj |, so |aj | < c for some c ≥ 0. Thus, in the interval of
interest, x ∈ [−1, 1], we have,
∞ ∞ ∞
|aj xj | ≤ C + |aj xj | ≤ C + c |x|j
X X X
where the series in the last term converges for all |x| < 1. Thus, the series converges for all
|x| < 1. But in general it diverges at the boundaries, (see exercise below) unless it terminates
for finite j. So if we want to consider solutions which are everywhere smooth, including the
poles of the sphere, then the series must terminate. This happens when l is an integer, and
the series becomes a polynomial of order l in x. They are called the Legendre Polynomials,
and the first ones are:
P0 (x) = 1,
P1 (x) = x,
P2 (x) = 1/2(3x2 − 1),
Notice that the first corresponds to the choice a0 = 1, a1 = 0,, l = 0, the second to
a0 = 0, a1 = 1, l = 1, and the third to a0 = −1/2, a1 = 0, l = 3. They are normalized in
such a way that,
Pl (1) = 1
Exercise: Compute P3 and P4 using the recursion relation found and the normalization
condition
Exercise: Convince yourself that indeed the above series diverge at |x| = 1 unless they
terminate. Hint: First show that for any ε > 0 there exists integer J such that |aj+2 | >
(1 − ε)|aj | ∀ j > J. Second use the previous point to establish that |aj+J | > (1−εj ) J |aJ |.
j
j
j+2
2j 2j+1
Compare this result with the series ∞ s
j=0 2j , and
∞ s
j=0 2j+1 . Show, using that 1−s =
1 ∞
j=0 s
j
P P P
that the above series represent the functions 12 ln(1 + s)(1 − s) and 12 ln( 1−s
1+s
) respectively and
so that they diverge for |s| = 1. Conclude, using s = x(1 − ε), and bounding one series with
the other, that the series diverges for all |x| ≥ 1.
Since for each l we obtain a polynomial of degree l, with linear combinations of then
we can obtain all powers of x. But since every smooth function on a compact set can be
approximated to arbitrary precision with a polynomial with appropriated coefficients. We
can approximate any function by linear combination of the Legendre polynomials, thus they
5.2. METHOD OF SEPARATION OF VARIABLES IN SPHERICAL COORDINATES 85
form a complete basis to expand smooth functions in the interval [−1, 1]. If we consider the
scalar product in the space of smooth functions in [−1, 1] given by,
Z 1
< f, g >= f¯(x)g(x)dx,
−1
then one can see that the Legendre polynomials form a orthogonal basis with respect to this
scalar product, indeed,
Z 1
− [l(l + 1) − l′ (l′ + 1)] Pl (x)Pl′ (x)dx
−1
d1 dPl d dPl′
Z
= [(1 − x2 )
[ ] Pl′ − [(1 − x2 ) ] Pl ] dx
−1 dx dx dx dx
dPl dPl′ dPl′ dPl
Z 1
= [−(1 − x2 ) + (1 − x2 ) ] dx
−1 dx dx dx dx
dPl dPl′
+ (1 − x2 ) Pl′ |1−1 − (1 − x2 ) Pl |1−1 = 0
dx dx
thus, Z 1
Pl (x)Pl′ (x)dx = 0 if l 6= l′ .
−1
On the other hand it can be seen using special properties of these polynomials that
1 2
Z
Pl2 (x)dx = .
−1 2l + 1
Exercise: Use the explicit formula of the first four Legendre polynomials and the orthogo-
nality relations to compute Z 1
x4 Pl (x)dx.
−1
∆φ = 0 at V = lR3 − Ba
φ|∂V + = φ0
φ|∂V − = −φ0
φ(~x) → 0 as |~x| → ∞ (5.2)
Where we have divided the sphere into two parts, V + , and V − according to the the sign of
the coordinate z. The asymptotic condition implies Ul+ = 0 ∀l so we have only to find the
coefficients Ul± from the other two boundary conditions. In terms of the coordinates used
these two boundary conditions result in the following one:
∞
Ul− a−(l+1) Pl (cos(θ)) = φ0 (cos(θ))
X
φ(a, θ) = (5.3)
l=0
Exercise: Find the solution inside the cups (assuming them hollow). Then apply Gauss law
and compute the surface charge distribution.
Exercise: Integrate the surface charge distribution and find the configuration’s capacity.
with the z axis along the symmetry axis. The parameters describing the problem are then,
the charge density, σ, the radius of the crust, a, and the angular aperture of the hole, α.
The procedure to solve this problem is as follows, we know that inside the crust the
potential satisfies Laplace’s equation, and also outside. Thus, we can propose an internal and
an external solution and match them at the interface r = a. They are given by,
∞
Ulint r l Pl (cos(θ))
X
φint (r, θ) := r < a
l=0
∞
Ulext r −(l+1) Pl (cos(θ))
X
φext (r, θ) := r > a
l=0
To perform the integral it is useful to use the following relation among Legendre polynomials,
∂ ∂
Pl+1 (x) − Pl−1 (x) − (2l + 1)Pl (x) = 0.
∂x ∂x
Using it we get,
σ cos(α) ∂ ∂
Z
σl = [ Pl+1 (x) − Pl−1 (x)] dx
2 −1 ∂x ∂x
σ
= [Pl+1 (cos(α)) − Pl+1 (−1) − Pl−1 (cos(α)) + Pl−1 (−1)]
2
σ
= [Pl+1 (cos(α)) − (−1)l+2 − Pl−1 (cos(α)) + (−1)l ]
2
σ
= [Pl+1 (cos(α)) − Pl−1 (cos(α))]
2
Thus, the second interface conditions shields,
4π 2π
Ulint = σl = [Pl+1 (cos(α)) − Pl−1 (cos(α))],
2l + 1 2l + 1
and we have solved the problem.
Exercise: Using the above result find the first two multipole moments in terms of the total
charge of the crust, Q, the radius, a, and α.
Exercise: Find the solutions when there are two holes, one in the north pole and one on the
south, of aperture α+ and α− respectively. Look at the limit, when the angles go to π/2 and
σ → ∞ but keeping the total charge constant.
d dP m2
[(1 − x2 ) ] − P = −l(l + 1)P,
dx dx 1 − x2
where we have already assumed l integer for this is needed to obtain smooth solutions in the
whole sphere.
It can be seen that the following are solutions to this equation:
90CHAPTER 5. STATIC PROBLEMS – CONSTRUCTING INTERESTING SOLUTIONS
dm m 2 m/2
Plm (x) := (−1) (1 − x ) Pl (x) m ≥ 0.
dxm
The coefficient, (−1)m , appears only for convention and does not play any important role.
dm+2 dm+1 dm dm
(1 − x2 ) P l (x) − 2(m + 1) P l (x) − m(m + 1) P l (x) = l(l + 1) Pl (x)
dxm+2 dxm+1 dxm dxm
Remark: Note that the solutions are no longer polynomials for m odd.
Remark: Note also that Plm = 0 for m > l.
For fixed m these functions, called generalized Legendre functions, satisfy the following
orthogonality condition (same m):
1 2 (l + m)!
Z
Pl′ m Plm dx = δll′ .
−1 2l + 1 l − m)!
In practice, it is convenient to work with the whole angular part of the solutions to the
Laplace’s equation, that is to multiply Plm by Qm (ϕ) = eimϕ .
Renormalizing them for convenience we define the spherical harmonics of type (l, m):
v
u 2l + 1 (l − m)!
u
Ylm (θ, ϕ) := t Plm (cos θ)eimϕ , m ≥ 0.
4π (l + m)!
But we also need solutions with m negative, those we define as:
v
u 2l + 1 (l−| m |)!
u
Ylm (θϕ) := t Pl |m| (cos θ)(−1)m eimϕ , m < 0,
4π (l+ | m |)!
or in other words:
Yl−m (θ, ϕ) = (−1)m Ylm
⋆
(θ, ϕ),
where ⋆ means complex conjugation.
Exercise: Compute all the spherical harmonic functions for l = 0, 1, 2. Express them also in
Cartesian coordinates.
∞ X
X m
f (θ, ϕ) = Clm Ylm (θ, ϕ),
l=0 l=−m
with Z
Clm = f (θ′ , ϕ′ )Ylm
⋆
(θ′ , ϕ′ )dΩ.
S2
But then,
∞ X
m Z
f (θ′ , ϕ′ )Ylm
⋆
(θ′ , ϕ′ )dΩYlm (θ, ϕ)
X
f (θ, ϕ) =
2
l=0 l=−m S
Z ∞ X
m
f (θ′ , ϕ′ ) ⋆
(θ′ , ϕ′ )Ylm (θ, ϕ)dΩ,
X
= Ylm
S2 l=0 l=−m
and so,
∞ X
l
⋆
(θ′ , ϕ′ )Ylm (θ, ϕ) = δ(ϕ − ϕ′ )δ(cos θ − cos θ′ ).
X
Ylm
l=0 m=−l
This is a convenient way of writing Dirac’s delta on the surface of a sphere. It shows also
that the commutation of the integral with the series is not an allowed step, and that the last
equality is just formal, neither the double series converges nor the integral makes sense, since
the integrand is not an integrable function.
What are these spherical harmonic functions on the sphere? Can we characterize them in
some invariant way? If we take the Laplacian in lR3 in spherical coordinates, set r = 1 and
dismiss all “derivatives" along the radial direction, we obtain the Laplacian of the unit sphere,
1 ∂ ∂φ 1 ∂2φ
∆S 2 φ := (sin θ ) + .
sin θ ∂θ ∂θ sin2 θ ∂ϕ2
The spherical harmonics we have defined are a complete, linearly independent set of solu-
tions to the following eigenfunction - eigenvalue problem:
∆S 2 Ui = λi Ui ,
that is,
∆S 2 Ylm (θ, ϕ) = −l(l + 1)Ylm (θ, ϕ).
Note that there are several, 2l + 1, solutions for each eigenvalue, as m ranges from −l to l.
So the Laplacian in the sphere splits naturally the space of smooth functions into invariant
subspaces of dimension 2l + 1.
94CHAPTER 5. STATIC PROBLEMS – CONSTRUCTING INTERESTING SOLUTIONS
Exercise: Show the orthogonality property of the spherical harmonics with different l using
the previous exercise. Exercise: Use that for each l the spherical harmonic functions satisfy
∂2
Y
∂ϕ2 lm
= −m2 Ylm to show their orthogonality.
ρ(~x′ ) 3 ′
Z
φ(~x) = d ~x . (5.8)
lR3 |~x − ~x′ |
This solution is unique and smooth outside D.
Outside D the potential φ satisfies the homogeneous solution, ∆φ(~x) = 0, so it must be
of the form:
∞ X
l
4π
Clm r −(l+1) Ylm (θ, ϕ),
X
φ(~x) = ~x ∈
/D (5.9)
l=0 m=−l 2l + 1
4π
for some constants Clm (the factor 2l+1 is for compatibility with standard definitions). These
constants are called the multipole moments of the electrostatic field. They characterize
uniquely the field outside the support of the static sources region. The bigger the l the faster
they decay at large distances and each one of them has a characteristic angular dependence.
We shall see now how to compute these constants as functions of the source distribution ρ.
The first one, C00 is immediate by applying Gauss theorem. Indeed, if we perform a surface
integral of the normal component of the electric field at a surface r = R surrounding the
sources, that is, outside D, we have:
5.3. APPLICATION: MULTIPOLE MOMENTS OF A STATIC CONFIGURATION 95
Z Z
3
4πQ = 4πρ(~x) d ~x = − ∆φ(~x) d3~x
lRI3 I lR
3
= − ~ · d~s = −
∇φ ∂r φ|r=R R2 dΩ
S(R) S2
∞ X
l
4π
I
Clm (l + 1)R−(l+2) Ylm (θ, ϕ)R2 dΩ
X
=
S 2 l=0 m=−l 2l + 1
3
∞ X
l
(4π) 2
I
Clm (l + 1)R−l Ylm (θ, ϕ)Y ∗00 (θ, ϕ) dΩ
X
=
S 2 l=0 m=−l 2l + 1
3
= (4π) 2 C00 (5.10)
Where in the last step we have included Y00∗ (θ, ϕ) = √14π and use the orthogonality relations
among spherical harmonics. So we see that the first one, the one that decays as 1/r at long
distances, carries the information about the total mass of the charge system. We use now the
same strategy to find the others, recalling that ψ = r l Ylm
∗
is a solution to Laplace’s equation
we get,
Z Z
l ∗
4πqlm := 3
4πρ(r, θ, ϕ)r Ylm (θ, ϕ) dV = − 3
∆φ(r, θ, ϕ)r l Ylm
∗
(θ, ϕ) dV
ZlR IlR
= ~
∇φ(r, ~ l Y ∗ (θ, ϕ)) dV −
θ, ϕ)∇(r ∂r φ|r=R Rl Ylm
∗
(θ, ϕ) R2 dΩ
3 lm
lR S(R)
Z I
= − 3
∆(r l Y ∗lm (θ, ϕ))φ(r, θ, ϕ) dV + φ|r=R ∂r (r l Ylm
∗
(θ, ϕ))|r=R R2 dΩ
lR S(R)
I
− ∂r φ|r=R Rl Ylm
∗
(θ, ϕ) R2 dΩ
S(R)
∞ l′
I
4π ′ ∗
Cl′ m′ l R(l +1)−(l−1) Yl′ m′ (θ, ϕ)Ylm
X X
= (θ, ϕ)) dΩ
S 2 l′ =0 m′ =−l′ 2l′+1
∞ l′
4π
I
′ ∗
Cl′ m′ (l′ + 1) R(l +2)−(l+2) Yl′ m′ (θ, ϕ)Ylm
X X
+ (θ, ϕ)) dΩ
S2 l′ =0 m′ =−l′
2l′+1
∞ l′
4π
I
′ ∗
Cl′ m′ (l + l′ + 1) Rl −l Yl′ m′ (θ, ϕ)Ylm
X X
= (θ, ϕ)) dΩ
S 2 l′ =0 m′ =−l′ 2l′+1
= 4π Clm (5.11)
So we see that we have an explicit and simple relation between the multipole moments
of the fields and certain integrals over the sources, which are called the multipole moments
of the sources. Do this multipole moments of the sources characterize completely them?,
namely, can we recuperate ρ from our knowledge of the {qlm }’s? Notice that this would imply
that by knowing the field outside the sources we could know the sources themselves! This is
not so, because it is easy to find different source distributions for which the corresponding
fields outside them are identical. The simplest example is to take two different spherically
symmetric charge distributions with the same total charge, but it is easy to build many
examples. Nevertheless, the multipole moments are a very important tool to describe the
qualitative behavior of the fields knowing the source distributions and also the other way
around, knowing the field distribution obtain information about the source distribution. This
is mostly used to perform geological studies, where instead of the electric field potential people
look at the gravitational potential, which satisfies the same equation but the matter density
as the source.
Exercise: Find two different charge distributions with no symmetry at all but giving the
same external field (which can have symmetries). Hint: do not construct them explicitly, but
just start from their potentials and work backwards.
Exercise: Show that for real charge distributions we have qlm = (−1)m ql,−m
∗
.
Exercise: Compute the expressions in Cartesian coordinates of the first three source mo-
ments.
l
∞ X
4π
qlm r −(l+1) Ylm (θ, ϕ)
X
φ(~x) =
l=0 m=−l 2l + 1
∞ X
l
4π Z
[ ρ(~x′ ) (r ′ )l Ylm
∗
(θ′ , ϕ′ ) d3~x′ ] r −(l+1) Ylm (θ, ϕ)
X
=
l=0 m=−l 2l + 1 V
∞ X
l
4π (r ′ )l ∗ ′ ′
Z
ρ(~x′ ) Y (θ , ϕ ) Ylm (θ, ϕ) d3~x′
X
= (l+1) lm
V l=0 m=−l 2l + 1 r
(5.14)
5.3. APPLICATION: MULTIPOLE MOMENTS OF A STATIC CONFIGURATION 97
Since we are in the region far from the sources, r ′ < r and one can show the expression
′
converges (the terms, taking a factor 4π
r
out, are bounded by ( rr )l ).
But since (5.8) holds, and these equations are for arbitrary ρ(~x′ ), it follows that,
∞ l
1 1 (r ′ )l X ⋆
(θ′ , ϕ′ )Ylm (θ, ϕ),
X
′ = 4π Ylm
l=0 2l + 1 (r)
| ~x − ~x | l+1
m=−l
whenever r ′ < r. But this expression clearly diverges for r ′ > r, so can not be true in this case
and our claimed identity fails, the reason being that in the previous calculations we commute
some integrals with series and in the present case this is not allowed. From the symmetry of
the expression on the left, it is clear that the same expression holds when we interchange ~x
and ~x′ , that is r and r ′ , and (θ, ϕ) and (θ′ , ϕ′ ). Since only matters the angle difference we
must have,
∞ l
1 1 (r < )l X ⋆
(θ′ , ϕ′)Ylm (θ, ϕ),
X
= 4π Ylm
| ~x − ~x′ | l=0 2l + 1 (r > )l+1
m=−l
where r < = min{r, r ′ }, r > = max{r, r ′ }. This expression being always convergent whenever
~x 6= ~x′ . This useful relation is known (or rather equivalent to) as the addition theorem for
spherical harmonics. A direct proof follows:
with Z
⋆
Alm = f (θ, ϕ)Ylm (θ, ϕ)dΩ
S2
Note that at θ = 0, ( q
2l+1
m=0
Ylm (θ = 0, ϕ) = 4π
0 m 6= 0,
and so, s
∞
X 2l + 1
f (θ, ϕ) |θ=0 = Al0 ,
l=0 4π
with s
2l + 1
Z
Al0 = f (θ, ϕ)Pl (cos θ)dΩ.
4π S 2
We shall use this property in what follows. We now want to see that if we define the angle
γ by cos γ = n̂′ · n̂, then
l
4π X
Pl (cos γ) = Y ⋆ (θ′ , ϕ′ )Ylm (θ, ϕ),
2l + 1 m=−l lm
98CHAPTER 5. STATIC PROBLEMS – CONSTRUCTING INTERESTING SOLUTIONS
n′
~
θ′
~
n
γ
θ
φ′
where (θ, ϕ) and (θ′ , ϕ′ ) are the angular coordinates defining n̂ and n̂′ respectively, see figure.
Since
we have,
cos γ = sin θ′ cos ϕ cos ϕ′ + sin θ sin θ′ sin ϕ sin ϕ′ + cos θ cos θ′
= sin θ sin θ′ [cos ϕ cos ϕ′ + sin ϕ sin ϕ′ ] + cos θ cos θ′
= sin θ sin θ′ [n̂xy · n̂′xy ] + cos θ cos θ′
= sin θ sin θ′ cos(ϕ − ϕ′ ) + cos θ cos θ′ ,
To get the result we are seeking, we first write the general expression,
∞ X
l ′
where we think of Pl (cos γ) as a function of (θ, ϕ), where (θ′ , ϕ′ ) are some fixed parameters,
the right-hand side is just the (true) assumption that any smooth function an the unit sphere
can be expressed as a linear combination of spherical harmonics.
But Pl (cos γ), being just a rotated Legendre polynomial, is a solution to
∆S 2 U = −l(l + 1)U
It must be a linear combination of its eigenfunction, thus,
l
X
Pl (cos γ) = Am Ylm (θ, ϕ),
m=−l
we see that only the l′ = l terms can be different from zero on the above expression, that is,
Pl (cos γ) can be written by just a linear combination of spherical harmonics with the same l,
l
Alm (θ′ , ϕ′ )Ylm (θ, ϕ),
X
Pl (cos γ) =
m=−l
with Z
′ ′ ⋆
Alm (θ , ϕ ) = Pl (cos γ)Ylm (θ, ϕ) dΩ.
S2
We now, for fixed (θ′ , ϕ′ ), make a coordinate change in the integration variables and integrate
with respect to (γ, β) where β is the new angular variable around the axis defined by n̂′ .
⋆
Thus, we interpret Ylm (θ, ϕ) as a function of (γ, β), i.e. θ = θ(γ, β), ϕ = ϕ(γ, β).
It can be shown that this change of variables leaves the surface element unchanged.
Thus, Z
Alm (θ′ , φ′ ) = ⋆
Pl (cos γ)Ylm (θ(γ, β), ϕ(γ, β))dΩ(γ, β).
S2
⋆
But then we can think of this as the computation of the l coefficient of Ylm (θ(γ, β), ϕ(γ, β)),
for
s
2l + 1
Z
⋆
Al0 = Pl (cos γ)Ylm (θ(γ, β), ϕ(γ, β))dΩ(γ, β)
4π S2
s
2l + 1
= Alm
4π
⋆
Since there is only one l in the expansion of Ylm (θ(γ, β), ϕ(γ, β)) [for the same reason that
there was only one l for the expansion of Pl (cos γ).] We have
s s
∞
⋆
X 2l′ + 1 2l + 1 2l + 1
Ylm (θ(γ, β), ϕ(γ, β)) |γ=0 = Al′ 0 = Al0 = Alm
l′ =0 4π 4π 4π
100CHAPTER 5. STATIC PROBLEMS – CONSTRUCTING INTERESTING SOLUTIONS
But when γ = 0 θ = θ′ , ϕ = ϕ′ , so
4π
Alm = Y ⋆ (θ′ , ϕ′ ),
2l + 1 lm
and the result follows.
1
The above identity allows to expand the function ′ with respect to arbitrary axes:
|~
x−~x |
∞ ∞
1 1 (r < )l X ⋆
(θ′ , ϕ′ )Ylm (θ, ϕ),
X
′ = 4π Ylm
l=0 2l + 1 (r )
| ~x − ~x | > l+1
m=−l
where (θ, ϕ) and (θ′ , ϕ′ ) are respectively the angular coordinates of ~x and ~x′ with respect to
an arbitrary coordinate system.
To show this result, first choose axes so that ~x′ points in the ẑ direction. Thus, the above
expression satisfies Laplace’s equation away from ~x′ , and has azimuthal symmetry (think of
it as the potential of a point charge which is on the z axis (at ~x′ = (0, 0, r ′)). And so it can
be written as,
∞
1
Cl r −(l+1) Pl (cos(γ))
X
′ =
| ~x − ~x | l=0
where γ is the angle between both directions. We shall assume that r := |~x| > r ′ |. To find
the coefficients we take the case where ~x is also on the z axis, namely, ~x = (0, 0, r). In that
case we should have,
∞ ∞
1 −(l+1)
Cl r −(l+1)
X X
′
= Cl r Pl (1) =
| r − r | l=0 l=0
Comparison with the geometric series gives Cl = (r ′ )l and the result is established The
symmetry of the expression under interchange of ~x with ~x′ implies that if r ′ > r the result is
obtained by interchange of r with r ′ in the final expression, that is,
∞
1 X (r < )l
= Pl (cos γ),
| ~x − ~x′ | l=0 (r > )l+1
where r < = min{r, r ′ }, r > = max{r, r ′ } and γ was the angle between ~x and ~x′ .
Chapter 6
Eigenfunction Expansions
AUi = λi Ui .
d 2 2
2 d d
In our examples A was dx 2 , (1 − x ) dx2 − 2x dx , or the Laplacian on the sphere.
• We found a large enough number of solution such that we could expand any (smooth)
function in terms of them. In the first of our examples we found the trigonometric functions,
Un (x) = sin( πnx
a
), and the theory of Fourier Series granted us that using them we could
expand an arbitrary continuous function in the interval [0, a], provided it vanished on the
extremes. In the second example the fact that we got polynomials of arbitrary order and
Weierstrass approximation theorem granted us similar expansions properties.
We shall see now that these three phenomena are deeply related, and are consequences of
certain property of the operators A which appear in physics.
101
102 CHAPTER 6. EIGENFUNCTION EXPANSIONS
We shall treat in detail the finite dimensional case, that in when A is just a n × n matrix
and so Ui are n-vectors. Then we discuss the infinite dimensional case remarking the analogies
and differences.
AU = λU,
or
Dλ U = 0,
with Dλ := A − λI where I is the identity matrix. Clearly eigenvectors are only determined
up to a factor, if U is one, so is cU, with the same eigenvalue.
Using the fundamental theorem of linear algebra we have:
The above equation has a solution if and only if detDλ = det(A − λI) = 0.
Now this determinant is a polynomial of degree n in λ, and so we know it has at least one
solution, λ0 . Thus we conclude:
There is at least one pair (λ0 , U0 ) solution to the above eigenvalue – eigenvector
problem.
and so for them to be orthogonal it is necessary that the right-hand side vanishes. This is
automatically satisfied if < AU, V >=< U, AV >, in that case we say that A is self-adjoint.
Notice that this property depends on what scalar product we are using, with the conventional
one the self-adjoint matrices are the symmetric ones. Most of the operators which appear in
physics are self-adjoint, so we restrict the discussion to them form now on.
What happens if there are several eigenvectors with identical eigenvalues? Let the set
{Ui } i = 1, . . . m, be eigenvectors with the same λ as eigenvalue. Then any linear combina-
tion of then is also an eigenvector with the same eigenvalue, indeed,
m m m m
ci Ui ) = ci AUi = ci λUi = λ ci Ui .
X X X X
A(
i=1 i=1 i=1 i=1
Thus, we can use the Gram - Schmidt procedure to get a orthonormal base for the subspace
expanded by the original ones. Thus, we conclude: The eigenvectors of an arbitrary self-
adjoint matrix A, can be chosen to be orthonormal.
We turn now to the question of completeness, namely whether the above set of eigenvectors
suffices to expand any vector in lRn . In finite dimensions this is simple, we only must check
that we have n linearly independent eigenvectors. Again this is true if A is self adjoint with
respect to some scalar product, <, >.
6.2. THE FINITE DIMENSIONAL CASE 103
Exercise: Find a 2 × 2 matrix where this is not true, namely a matrix which has only a
single l.i. eigenvector.
Alternative to the proof given above we have the following, which admits an easy gener-
alization to infinite dimensional vector spaces:
Let (H, <, >) a vector space with scalar product, and A : H → H a linear map which
is self-adjoint, < V, AU >=< AV, U >. Then their eigenvectors form an orthonormal base
which expands the space.
104 CHAPTER 6. EIGENFUNCTION EXPANSIONS
AU0 = λ0 U0
Consider the space H0 = {V ∈ H, < V, U0 >= 0} that is the space of all vectors perpendicular
to U0 . This is a subspace of H and A : H0 → H0 , indeed, < AV, U0 >=< V, AU0 >= λ0 <
V, U0 >= 0 ∀V ∈ H0 . Thus, AV ∈ H0 . Therefore, by the same argument as above, there
exist an eigenvector-eigenvalue pair in this subspace,
AU1 = λ1 U1 .
Continuing that way we get a complete set of eigenvectors and eigenvalues pairs.
where we take the complex conjugate of the first member to allow for the case of complex
valued solutions, and we are considering functions defined over lR, the case for lRn or other
spaces follows by trivial generalizations.
The first difference with the finite dimensional case is that here there exist vectors whose
norm is infinite! So the first step is to exclude them from consideration: From now on
we consider only classes of functions whose norms are finite w.r.t. the scalar product under
consideration.
There are other very important differences with the finite dimensional case, but they are
subtle, and we shall not discuss them here. In the infinite dimensional case, linear differential
operators can be taken as our A operator in the eigenvalue - eigenfunction problem. A straight
forward generalization to self adjoint operators follows. Repeating the proof of orthonormality
we made for eigenvectors, we also obtain in this case the orthogonality of eigenfunctions as a
consequence of the selfadjointness of the operator under consideration.
We now turn to the problem of completeness of eigenfunctions. The first difference here
is that, since the expression
∞ N
al Ul (x) := lim al Ul (x),
X X
f (x) =
N →∞
l=0 l=0
involves a limiting procedure, it is not well-defined until we give the norm with respect to
which the limiting procedure holds.
There are at least two important norms which one should consider:
6.3. THE INFINITE DIMENSIONAL CASE 105
N
d
cN 2
X
0 = kf− i (λ)Ui k |λ=0
dλ i=1
N N
d
cN cN
X X
= <f− i (λ)Ui , f − j (λ)Uj >|λ=0
dλ i=1 j=1
N N N N
δi⋆ < Ui , f − cN cN
X X X X
= − j Uj > − δj < f − i Ui , Uj >
i=1 j=1 j=1 i=1
N
δi⋆ (< Ui , f > − cN
X
= − i ) + complex conjugate.
i=1
N
cN
X
kf− i Ui k0 < ǫ.
i=1
In this norm is not so clear what the best choice of coefficients is, but if we take the ones
of the case before, which computationally is very handy, in many cases of interest one can see
that continuous functions are well approximated by eigenfunctions of self adjoint operators.
Chapter 7
One of our first results in electrostatics was the general solution to Poisson’s equation for
isolated systems,
ρ(~x′ ) 3 ′
Z
φ(~x) = d ~x .
lR3 |~x − ~x′ |
We shall call ψ(~x, ~x′ ) = 1
′ the Green function of the problem and notice that,
|~
x−~x |
∆ψ(~x, ~x′ ) = −4πδ(~x − ~x′ ).
So, the function which one integrates to obtain de general solution is just the solution
corresponding to a point source at point ~x′ . Thus, if we know those solutions for all points
where the sources can be, we can integrate and obtain the generic solution. Can one find
general solutions for other boundaries conditions, for instance and to fix ideas, for arbitrary
distributions of charges in the presence of a conducting body? The answer is yes, and the
machinery to find them is based in the following identity, called the second Green Identity:
Z I
3
(ψ∆φ − φ∆ψ) d ~x = [ψn̂ · ∇φ − φn̂∇ψ] dS,
V ∂V
That is,
1
Z I
φ(~x′ ) = ψ(~x, ~x′ )ρ(~x) d3~x + ~ x) − φ(~x)n̂ · ∇ψ(~x, ~x′ )] dS.
[ψ(~x, ~x′ )n̂ · ∇φ(~
V 4π ∂V
107
108 CHAPTER 7. THE THEORY OF GREEN FUNCTIONS
Notice that this is not yet a formula for a solution, for it is still an equation, since φ(~x′ )
appears in both sides of the expression. But we can transform it into a formula for a solution
if we choose conveniently ψ(~x, ~x′ ), that is F (~x, ~x′ ) as to cancel, with the help of the boundary
conditions we want to impose on φ(~x), the boundary integrals.
For instance, if we would want to give as a boundary condition the value of φ(~x) at
∂V, φ(~x)|∂V = f (~x), we would choose F (~x, ~x′ ) such that ψ(~x, ~x′ )|∂V = 0 and so we would then
have,
1
Z I
φ(~x′ ) = ψ(~x, ~x′ )ρ(~x) d3~x − ~ x, ~x′ ) dS.
f (~x)n̂ · ∇ψ(~
V 4π ∂V
We call this function, ψ(~x, ~x′ ) the Green function of the Dirichlet problem for V , and no-
tice that it depends on V and its boundary.
Notice that this function satisfies:
that is satisfies the Dirichlet problem for a point-like source of strength one at ~x′ with homo-
geneous boundary conditions.
Similarly, we can find other Green functions adapted to other boundary problems.
~
x
a
x′
~
q
Indeed,
For the Dirichlet Problem ψ(~y , ~x) |~yǫ∂V = 0, therefore the surface term vanishes and so the
result follows. It is clear that this is also true for the homogeneous Neumann problem. Had we
had considered complex green functions, then the result would have been, ψ(~x′ , ~x) = ψ̄(~x, ~x′ ).
We now return to the problem at hand. It is clear that it is convenient to write all
expressions in spherical coordinates. We already know the expression for the right-hand side,
1
δ(~x − ~x′ ) = δ(r − r ′ ) δ(ϕ − ϕ′ )δ(cos θ − cos θ′ )
r2
∞ X l
1 X
= δ(r − r ′ ) 2 ⋆
Ylm (θ, ϕ)Ylm (θ′ , ϕ′ )
r l=0 m=−l
:= δ(r − r ′ )σ(r, θ, θ′ , ϕ, ϕ′ ), (7.2)
110 CHAPTER 7. THE THEORY OF GREEN FUNCTIONS
where we can think of σ(r, θ, θ′ , ϕ, ϕ′) as a surface distribution. We also know the general
expression for a solution to Laplace’s equation, which will be valid at all points where ~x 6= ~x′ ,
∞ X
l
ψ(~x, ~x′ ) = [Alm (r ′ , θ′ , ϕ′ )r l + Blm (r ′ , θ′ ϕ′ )r −(l+1) ]Ylm (θ, ϕ).
X
l=0 m=−l
It will be discontinuous at ~x = ~x′ , so we expect the solution to have a given form for r < r ′
and a different one for r > r ′ Since ψ(~x, ~x′ )|r=a = ψ(~x, ~x′ )|r=b = 0 these expressions must be,
∞ X
l
a2l+1
ψ(~x, ~x′ ) = [A< ′ ′ ′ l
r < r′
X
lm (r , θ , ϕ )(r − )]Ylm (θ, ϕ)
l=0 m=−l r l+1
∞ X
l
b2l+1
ψ(~x, ~x′ ) = [A> ′ ′ ′ l
r > r′
X
lm (r , θ , ϕ )(r − )]Ylm (θ, ϕ)
l=0 m=−l r l+1
Continuity of the potential at r = r ′ (at points other than where the point charge is)
implies,
∞ X
l ∞ X l
a2l+1 b2l+1
[A< ′ ′ l > ′ ′ l
X X
lm (r, θ , ϕ )(r − )]Y lm (θ, ϕ) = [Alm (r, θ , ϕ )(r − )]Ylm (θ, ϕ)
l=0 m=−l
r l+1 l=0 m=−l
r l+1
Using the orthogonality of the spherical harmonics we see that for each pair l, m we must
have,
a2l+1 b2l+1
A< ′ ′ l
lm (r, θ , ϕ )(r − ) = A>
lm (r, θ ′
, ϕ ′
)(r l
− ),
r l+1 r l+1
and so,
l b2l+1
A< ′ ′ ′ ′ ′ ′ ′
lm (r , θ , ϕ ) = Elm (r , θ , ϕ )(r − )
r ′ (l+1)
′l a2l+1
A>
lm (r ′ ′
, θ , ϕ ′
) = Elm (r ′ ′
, θ , ϕ ′
)(r − )
r ′ (l+1)
For some functions Elm (r ′ , θ′ , ϕ′ ). To find them we look now at the matching conditions at
r = r ′ , we can consider the delta function there as a surface distribution (which in itself, at
(θ, ϕ) = (θ′ , ϕ′ ), is a distribution). So we must have,
Exercise: Use the symmetry under interchange of variables of the Green function to get the
solution up to constants depending only on l and m.
∞ X
l
′
X 1 l b2l+1 l ⋆ ′ ′
ψ(~x, ~x ) = −4π (r > − l+1 )r< Ylm (θ, ϕ)Ylm (θ , ϕ )
l=0 m=−l
(2l + 1)b2l+1
r>
∞ X
l
1 (rr ′ )l r<l
⋆
(θ′ , ϕ′ )
X
= −4π (b 2(l+1) − l+1 )Ylm (θ, ϕ)Ylm
l=0 m=−l 2l + 1 b r >
−b 1
= + ′ ,
b2 ~
x
|~x||~x − 2 | |~x − ~x |
′
|~
x|
112 CHAPTER 7. THE THEORY OF GREEN FUNCTIONS
1
Exercise: Use the expansion of ′ and the substitution r, r ′ → r, b2 → r ′ and r < r ′ to
|~
x−~x |
check the last line in the above deduction.
∞ X
l
1 a2l+1 1
ψ(~x, ~x′ ) = 4π l ⋆
(θ′ , ϕ′ )
X
(r< − l+1 ) l+1 Ylm (θ, ϕ)Ylm
l=0 m=l 2l + 1 r< r>
∞ X
l l
1 r< a a2 l ⋆
(θ′ , ϕ′ )
X
= 4π ( l+1 − ′ ( ′ ) )Ylm (θ, ϕ)Ylm
l=0 m=−l 2l + 1 r> rr rr
1 a
= ′ − .
|~x − ~x | |~x||~x − a2~x |
′
|~
x|2
It is interesting to recall what we have done in solving the equation for ψ(~x, ~x′ ). That
equation is second order so in general one should specify two boundary conditions, for in
general there would be two linearly independent solutions. If we depart from r = b inward
and towards r = r ′ , we must spend one boundary condition in giving ψ(~x, ~x′ )|r=b = 0. If we
depart from r = a outwards and towards r = r ′ we also waste one in setting ψ(~x, ~x′ )|r=a = 0.
For fix r ′ we must use another to set equal both solutions at r = r ′ . It remains one boundary
condition to be settled. One could imagine using it to set both first derivatives equal at r = r ′,
but that can not be done, unless all boundary conditions are taken to be zero, for the only
solution to the homogeneous equation with zero boundary conditions at r = a and r = b is
the zero solution. The last parameter has to be chosen then as to regulate the strength of the
source of the equation, it sets the jump on first derivatives. See figure.
1
Z I
φ(~x) = ′ ′
ρ(~x ) ψ(~x, ~x ) d ~x − 3 ′ ~ ′ ψ(x, ~x′ ) d2 S,
V (~x′ )n̂ · ∇
V 4π ∂V ~x
where is the volume inside the sphere of radius b, satisfies
∆φ(~x) = −4πρ(~x)
φ(~x)|∂V = V (~x).
~ ′ ψ(~x, ~x′ ).
Thus, we must compute n̂ · ∇~x
7.2. SOME EXAMPLES USING THE SPHERICAL GREEN FUNCTIONS 113
~
x
q
x′
~
r′
For the Green function corresponding to Case 2, which is well suited to study interior (to
the sphere) problems we have,
Thus, for instance, the solution to Laplace’s equation (ρ ≡ 0) inside a sphere at potential
V (θ, ϕ) is
∞ X
l
r
Z
( )l Ylm (θ, ϕ)[ V (θ′ , ϕ′ )Ylm
⋆
(θ′ , ϕ′ )dΩ′ ]
X
φ(r, θ, ϕ) =
l=0 m=−l b S 2 (r=b)
where a is the ring radius and b the one of the sphere. See figure.
∞ l
Q 1 ra r<
Z
ρ(~x′ )ψ(~x, ~x′ )d3~x′ = (−4π) ( ( 2 )l − l+1
X
φ(~x) = )Pl (cos θ)Pl (0),
V 2π l=0 b b r>
with r> = max{a, r}, r< = min{a, r}.
But
0 l odd
Pl (0) = (−1)l/2 (l−1)!!
2l/2 (l/2)!
l even
so,
∞
X (−1)l (2l − 1)!! 1 ra 2l 2l
r<
φ(r, θ, ϕ) = −2Q ( ( ) − 2l+1 )P2l (cos θ).
l=0 2l l! b b2 r>
Q
a
In this case,
Q 1
ρ(~x′ ) = [δ(cos θ′ − 1) + δ(cos θ′ + 1)]
2b 2πr ′ 2
Using the Green function of the problem we obtain,
∞
−Q X Z b
1 rr ′ l
r<
φ(~x) = [Pl (1) + Pl (−1)]Pl (cos θ) [ ( 2 )l − l+1 ]dr ′ .
b l=0 0 b b r>
Let
b 1 rr ′ l
r< b 1 rr ′ r (r ′ )l b rl
Z Z Z Z
Il = [ ( 2 )l − l+1 ] dr ′ = [ ( 2 )l dr ′ − ] dr ′ − ] dr ′ ,
0 b b r> 0 b b 0 r l+1 r (r ′ )l+1
then for l 6= 0,
rl bl+1 1 r l+1 1 r l b
Il = − + ′l |r
b2l+1 l + 1 l + 1 r l+1 lr
r l 1 1 r l1 1 r 2l + 1
= ( ) − + ( ) − = [( )l − 1] ,
b l+1 l+1 b l l b l(l + 1)
while,
b r
I0 = − ln = ln
r b
Since Pl (1) = 1 and Pl (−1) = (−1)l we have,
∞
Q r r 4l + 1
φ(~x) = [− ln + 2 [1 − ( )2l ]
X
P2l (cos θ)]
b b l=1 b 2l(2l + 1)
Bx ψ(~x, ~x′ ) = 0, in ∂V
116 CHAPTER 7. THE THEORY OF GREEN FUNCTIONS
where Bx is some linear, possible differential operator in ∂V , which gives the boundary con-
ditions.
For instance Bx = 1 gives Dirichlet’s boundary condition, while Bx = n̂ · ∇ ˆ x gives Neu-
mann’s.
To find ψ(~x, ~x′ ) one can look at the associated problem:
That is, the eigenvalue - eigenfunction problem for Ax with the same boundary conditions
as we want for the Green function.
If Ax is self-adjoint for these boundary conditions, then the set of eigenfunctions, {Ui }
i = 0, . . . form a complete set and therefore the Green function, as a function of ~x, can be
expanded as:
∞
′
ai (~x′ )Ui (~x),
X
ψ(~x, ~x ) =
i=0
i=0
~ j (~x), integrating over V and using the orthonormality relations for the Uj′ s
Multiplying by U
we obtain:
aj (~x′ )λj = −4πUj∗ (~x′ )
and so ∞
4π ∗ ′
ψ(~x, ~x′ ) =
X
(Uj (~x )Uj (~x)).
i=0 λj
Remark: For this formula to be valid the operator Ax does not have to have an eigenfunction
with zero eigenvalue for the required boundary conditions. If there is an eigenfunction U0 with
zero eigenvalue, then we can consider
but now only be able to find solutions for sources which do not have components along U0 ,
that is, sources such that: Z
U0 (~x)ρ(~x) d3~x = 0
V
If this condition, called an obstruction, is not satisfied, then there is no solution to the
problem. Note also that solutions are now not unique, if φ is a solution to our problem, then
φ + U0 is also a solution.
Remark: If the operator Ax is self-adjoint for the boundary conditions Bx , then if Ui is an
eigenfunction, so is Ūi , and both for the same real eigenvalue. This implies that ψ(~x, ~x′ ) is
real and so that ψ(~x, ~x′ ) = ψ(~x′ , ~x).
7.3. CONSTRUCTION OF GREEN FUNCTIONS USING EIGENFUNCTIONS. 117
Remark: For most self-adjoint operators in compact sets the set of eigenvalues do not
accumulate around any finite value and so the number of obstructions are finite.
~ =g
Exercise: Analyze the obstruction for the Neumann problem ∆φ = −4πρ in V , n̂ · ∇φ
in ∂V for finite V . Give a physical interpretation to it.
∆Ui = λi Ui ∈ V, ~ i |∂V = 0,
n̂ · ∇U
q
has a eigenfunction with zero eigenvalue, U0 = 1/ vol(V ). This is the unique one, indeed
assume U is an eigenfunction with zero eigenvector, then ∆U = 0, so, integrating over the
~ =0
volume UδU, integrating by parts, and using the boundary condition we find that ∇U
from which we conclude that U is constant.
Thus, the obstruction is that
Q
Z
ρ U0 = q =0
V vol(V )
So we can apply our Green’s function construction. Indeed, if (7.3) holds, then
Z
ρ̃ d3~x = 0.
V
Z
δφ(~x) = ψ(~x, ~x′ )ρ̃(~x′ ) d3~x′
V
1
Z
= ψ(~x, ~x′ )(ρ(~x′ ) +
∆φ̃) d3~x′
V 4π
1 1
Z I
= ′ ′
ψ(~x, ~x )ρ(~x ) + ′ 3 ′
∆ψ(~x, ~x )φ̃ d ~x + ~ φ̃ − φ̃ n̂ · ∇ψ]
[ψ n̂ · ∇ ~ d2 S
V 4π 4π ∂V
Z
′ ′ 1 I ~ φ̃ d2 S
= ψ(~x, ~x )ρ(~x ) − φ̃(~x) + ψn̂ · ∇ (7.4)
V 4π ∂V
where in the second and the last line we have used the properties of Green’s functions, and on
the third Green’s identity. Thus, we reach to the known expression for this class of problems,
but knowing that it is only valid when the integrability condition (7.3) is satisfied.
If the condition is not satisfied, the above expression will give some function, but it will
not satisfy the equations for the ρ and g given, but rather for a pair projected into a subspace
where the condition holds.
Chapter 8
Dielectrics
119
120 CHAPTER 8. DIELECTRICS
a ~
E
we shall modify the problem and assume the we have now two spheres superimposed with
charges Q and −Q respectively, so that initially the field they produce cancel each other and
so the total field due to the charges vanishes.
Of course the answer is simple, the new configuration is the sum of the external field
and of the fields corresponding to the two sphere of uniform charge in their new positions,
~
d~ = ± QE 0 . To first order the new field satisfies:
k
~ · δ E(~
∇ ~ − ρ0 (~x + d)]
~ x) = 4π[ρ0 (~x − d) ~ (8.1)
≃ −8π d~ · ∇ρ
~ 0 (~x) (8.2)
≃ −8π ∇ ~ 0 (~x)),
~ · (dρ (8.3)
or
∆δφ(~x) = 8π d~ · ∇ρ
~ 0 (~x)
−3Q
= δ(r − a)d~ · n̂
a3
= −4πσ(~x)δ(r − a),
+3Q ~
with σ(~x) := d · n̂.
4πa3
Thus, our problem is to solve ∆δφ = 0 inside and outside the sphere, with the boundary
conditions:
~
n̂ · ∇(δφ out − δφin )|r=a = −4πσ
(δφout − δφin )|r=a = 0.
8.1. THE NATURE OF THE PROBLEM 121
a ~
E
d~ · ~x
δφout (~x) = B ,
r3
δφin (~x) = C d~ · ~x.
∂ B~
~
n̂ · ∇(δφ out − δφin )|r=a = ( d · n̂ − Cr d~ · n̂)|r=a
∂r r 2
2B 3Q
= d~ · n̂(C + 3 ) = 3 d~ · n̂
a a
B ~
and (δφout − δφin )|r=a = a(−C + 3 )d · n̂ + 0
a
Thus, B = Q, C= Q
, ~ 0,
or in terms of E
a3
~ 0 ·~x
Q2 E
r>a
δφ = k
Q2 ~
r3
E
ka3 0
· ~x r<a
We see that outside the sphere the electric field difference, besides the constant field, is of
2~
a dipolar field configuration corresponding to a dipole ~p = −QkE 0 , while inside the constant
2
field has diminished its strength due to the presence of an extra constant field −Q ~ 0.
E
ka3
The total field is (again dismissing the field of the initial charge configuration):
~ 0 ·~n ~n−E
~0
~ =
~ 0 + 3 Q2 E
E r>a
E k r3
Q2 ~
(1 − ka3 )E 0 r<a
Remark:
1.) If we define
~ 0 (~x)
P~ (~x) = 2dρ
Q ~ 0,
= ρ0 (~x)E
k
3Q2 ~
= E 0 θ(a − r)
4πka3
~ = E(~
then it follows from 8.1 that, D ~ x) + 4π P
~ (~x) satisfies,
~ ·D
∇ ~ = 0.
This field, called the electric displacement vector has a continuous radial component, but
the angular components have jumps across r = a. To see this, contract the electric field with
the normal to the sphere (r = a) both inside and outside it and compare.
2.) Note that E ~ and P~ are linear in E
~ 0 , to the approximation we are working with, and
~ ~
so P is linear in E. We define the electric susceptibility χe by the relation:
P~ = χe E
~
3Q2
For this case: χe = 4π(ka3 −Q2 )
.
~ = εE,
Defining, ε, the electric permittivity by the relation, D ~ in this case we have,
ka3 + 2Q2
ε = 1 + 4πχe =
ka3 − Q2
some averaged fields and, after finding equations for them, impose certain phenomenological
relations between them.
To fix ideas we consider a solid with a charge density given by
XX
ρ(~x) = ρf ree (~x) + qj(m) δ(~x − ~xm − ~xj(m) ),
m j(m)
where, ρf ree (~x), is the charge density of freely moving charges, ~xm is the center of mass
coordinate of the present, fixed molecules, and ~xj(m) is the, relative to the center of mass,
displacement of the “j - sphere" in the molecule m. These last displacements are supposed to
change when external fields are applied.
~
xm,1 ~
xm,3
~
xm,2
~
xm,4
~
xm
The free charge contribution to the solution is dealt with as in conductors and so will not
be discussed further. Notice that if that free charge is present in any reasonable amount, then
the material would behave as a conductor and so no polarization effect would be present in
the bulk of the material. It basically amounts to have an infinite value for ε.
We shall consider now space averages over distances which are big compared with the size
of individual molecules, but still very small in comparison with the smallest regions on which
we measure macroscopic electric fields. A typical distance in solids is L = 10−6 cm, in the
corresponding volume are housed typically about 106 molecules.
Given smooth, compactly supported function in lR+ , ϕ(s), such that ϕ(0) = 1,
R∞
4π 0 ϕ(s)s2 ds = 1, we define for each function f : lR3 → lR
Z
fϕ (~x) = 3
ϕ(|~x − ~y |)f (~y)d3~y
lR
Averaging region
Z
~ ϕ (~x) =
∇f ~ x ϕ(|~x − ~y |)f (~y )d3~y
∇
lRZ3
= − ∇~ y ϕ(|~x − ~y |)f (~y)d3~y
3
Z lR
= ϕ(|~x − ~y |)∇ ~ y f (~y )d3~y
3
lR
= (∇~ x f (~x))ϕ .
Since the surface integral at infinity that results from the application of Gauss theorem
in the third step vanishes, for ϕ is compactly supported. If we apply these averages to the
static Maxwell’s equations we obtain,
~ ∧E
∇ ~ϕ = 0
~ ·E
∇ ~ ϕ = 4πρϕ .
These are now equations for averaged quantities over much larger scales than those charac-
teristics of the matter scales. 1 This field is not the field which is felt by individual molecules,
but rather the averaged field produced by other molecules and external fields, the own molec-
ular field can be much bigger, but mostly participate in creating the initial molecular equi-
librium configuration.
The task now is to accurately enough describe ρϕ in terms of known quantities. Taking
the molecular charge distribution we had defined above we find:
Z
(ρm )ϕ (~x) = ϕ(|~x − ~y |)ρm (~y )d3~y
lR3
X
= qj(m) ϕ(|~x − ~xm − ~xj(m) |).
j(m)
1
The fact that we get the same averaged equations for any average is due to the linear character of the
equations, which we shall also assume in the way the electrostatic field affects the matter.
8.2. A MICROSCOPIC MODEL 125
~
X
(ρm )ϕ (~x) = qj(m) [ϕ(|~x − ~xm |) − ~xj(m) · ∇ϕ(|~
x − ~xm |) +
j(m)
1 i ∂2
xj(m) xkj(m) i k ϕ(|~x − ~xm |) + . . .
2 ∂x ∂x
~
= qm ϕ(|~x − ~xm |) − p~m · ∇ϕ(|~ x − ~xm |)
2
1 ik ∂
+ Sm ϕ(|~x − ~xm |) + · · · ,
2 ∂xi ∂k k
with
X
qm := qj(m)
j(m)
X
p~m := qj(m)~xj(m)
j(m)
ik
qj(m) xij(m) xkj(m) .
X
Sm :=
j(m)
Z
~ x − ~xm |) =
−~pm · ∇ϕ(|~ ~
−~p · ∇ϕ(|~ x − ~y |)δ(~xm − ~y ) d3~y
lR3 Z m
= −∇~ · p~ ϕ(|~x − ~y |)δ(~xm − ~y ) d3~y
3 m
lR
~
= −∇ · ~pmϕ .
~ · P~ ϕ + 1 ∇
ρϕ = ρ̄ϕ − ∇ ~ · (∇
~ ·S
~ ϕ) + · · · ,
2
where, P~ ϕ := m ~pmφ , etc. Since all except the first term are divergences we could write:
P
~ ·D
∇ ~ = 4π(ρ̄ + ρf ree )
126 CHAPTER 8. DIELECTRICS
with D~ϕ = E~ ϕ + 4π P~ ϕ − 2π ∇
~ ·S
~ϕ + · · ·
The above, plus the equation ∇ ~ ∧E~ ϕ = 0 are the two equations we shall now be interested
in. But they can not be solved until we relate D ~ ϕ and E~ ϕ.
The fundamental assumption we shall be doing is that P~ , S,
~ and all other similar averaged
quantities are caused by the applied external electric field, and that these relations are all
local. That is, we shall assume there exists a smooth relation fo the form:
~ x) = D(
D(~ ~ E(~
~ x))
Expanding this relation in Taylor series,
~ x)]i = [D
[D(~ ~ 0 (~x)]i + εi (~x)[E(~
~ x)]j + εi (~x)[E(~
~ x)]j [E(~
~ x)]k + · · · ,
1j 2jk
we note that usually the averaged interior fields vanish, and so D ~ 0 (~x) = 0. This is for sure
the case for isotropic media, for there is no privileged directions and so necessarily D~ 0 (~x) = 0.
In general the first non-trivial term in the series suffices to treat most materials with weak
external fields, and so
~ x)]i = εi (~x)[E(~
[D(~ ~ x)]j .
1j
On isotropic media εi1j , as a linear map, can not have any preferred subspaces and so
εi1j = εδ i j, thus,
~ x) = ε(~x)E(~
D(~ ~ x).
For the isotropic case then the equations to solve now are:
~ ∧E
∇ ~ = 0
~ · (εE)
∇ ~ = 4π(ρf ree + ρ̄),
~ = −∇φ,
or with the ansatz E ~
~ · (ε∇φ)
∇ ~ = ε∆φ + ∇ε
~ · ∇φ
~ = −4π(ρf ree + ρ̄).
that is,
q
D(r) = ,
r2
2
Actually this is not a free charge, in the sense that we have fixed it to stay in a point, but we call it free
to distinguish it from the averaged charges, for here it is used just to create an external field
8.2. A MICROSCOPIC MODEL 127
and
E~ = 1D ~ = q n̂ .
ε ε r2
Since ε > 1 we see that the electric field the charge generates is in this case smaller than
if the charge had been in vacuum. These phenomena are called screening and is due to the
microscopic dipole alignment along the electric field direction and so any sphere centered at
the point-like source has a real charge given by qε .
ε ε=1
Inside the sphere the result must be identical to the one of the prior example, that is:
D ~ = qn̂2 . Outside D
~ = qn̂2 , E ~ = E(ε
~ = 1) and E ~ = qn̂2 . We see that D
~ is continuous across
r εr r
the surface delimiting the dielectric, that is, where ε jumps, while E ~ is not. The jump in E
~
corresponds to a surface charge density given by:
~ in )|r=a = q (1 − 1 ).
~ out − E
4πσ = n̂ · (E
a2 ε
or
~ 1 = ε2 n̂ · ∇φ
ε1 n̂ · ∇φ ~ 2,
~1 − E
~ 2 ) · n̂ = ( 1 1 ~
4πσ = (E − )D · n̂.
ε1 ε2
This can be seen applying Gauss theorem to an infinitely thin pillbox surface enclosing a
section of the discontinuity surface.
DIELECTRIC
VACUUM
normal
~ ∧E
Since ∇ ~ = 0, the same argument as the one used in the vacuum case can be applied
and integration along a very thin loop at the surface implies:
~1 − E
(E ~ 2 ) ∧ n̂|S = 0
or
~ 1 − ∇φ
(∇φ ~ 2 ) ∧ n̂|S = 0.
Thus, all derivatives tangential to the interface S of φ1 − φ2 vanish and so (φ1 − φ2 )|S = c for
some constant c. Therefore, if we can arrange for φ1 − φ2 = 0 at some point of S,
(φ1 − φ2 )|S = 0
This in general can be done, for it is just the requirement that there will be no discontinuous
increase of energy when bringing a test charge from infinity through the interface. Thus, we
find that the potential must be continuous along the boundary.
8.2. A MICROSCOPIC MODEL 129
DIELECTRIC VACUUM
γ2
γ1 γ3
γ4
~0
E
a
~ in = 3εout ~
E E0
2εout + εin
3 3
~ out = E
E ~ 0 (1 + a εin − εout ) − 3a εin − εout (E ~ 0 · n̂)n̂ (8.4)
r 3 εin + 2εout r 3 εin + 2εout
When εout = εin we just get the external field everywhere. When εout = 1 and εin → ∞
(uncharged conductor), E ~ in → 0 and E ~ out → E
~ 0 + a33 (E~ 0 − 3(E
~ 0 · n̂)n̂).
r
Notice also that if εout < εin , as is the case if outside the sphere we are in vacuum, them
3εout
2εout +εin
< 1 and again we have screening.
3
Recall that Laplace’s equation solutions which are regular at the origin are of the form S ij...n xi xj . . . xn ,
~ 0 , and linearity
with S ij...k symmetric and without trace. Thus, in our case S ij...k can only be build out of E
implies that only the dipolar field survives. Without imposing linearity we would have also, for instance
S ij = E0i E0j − 13 δ ij E
~0 · E
~ 0.
8.3. THE ELECTROSTATIC ENERGY OF DIELECTRICS 131
If we bring from infinity to the surface of, say, a conductor C1 a charge δq, then we would
be giving to the system an energy given by
−V 1
I
1
δE = V δq = ~ · n̂ dS,
δD
4π ∂C1
where we have used that, since C1 is a conductor, δq would distribute in its surface creating
an increment in the surface charge distribution, δσ, which in turn can be expressed as an
~ δσ = − 1 (δ D
increment in D, ~ · n̂)|∂C1 , where we have taken the normal towards the inside of
4π
the conductor.
Since the potential φ|∂C 1 = V 1 we have,
1 I ~ · n̂ dS
δE = − φδ D
4π ∂C1
1 Z ~ ~ d3~x
= − ∇ · (φδ D)
4πZ V
−1 ~ · ∇φ
~ + φ∇ ~ · δ D]
~ d3~x,
= [δ D
4π V
where V is the space outside the conductors.
Using now that E ~ = −∇φ~ and ∇ ~ ·D
~ = 0 – we are assuming there are no free charges
outside the conductors – we have,
1
Z
δE = ~ · δD
E ~ d3~x.
4π V
Thus, we have an expression for the infinitesimal change in energy in the above configu-
~ and D.
ration of dielectric and conductors in terms of E ~
~ ~ ~ ~
If we assume a linear between E and D, D = εE, then,
132 CHAPTER 8. DIELECTRICS
1 Z ~ ~ 3
δE = εE · δ E d ~x.
4π V
This variation corresponds to an energy given by,
1 Z ~ ~ 3 1 Z ~ ~ 3
E= εE · E d ~x = E · D d ~x.
8π V 8π V
To see this in some detail, consider reaching the final configuration of conductors and
dielectrics by slowly increasing the potential of all conductors simultaneously and at the same
rate, thus the change in the electric potential at the conductors surface will be of the form,
~
V i (λ) = λV i , thus, E(λ) =λE ~ and correspondingly we will have δ E ~ = Edλ.
~ That is, the
field will increase from zero to a final value without changing its direction nor their respective
magnitudes from point-to-point. We then have,
1 1
Z Z
E = [ λεE ~ ·E~ dλ] d3~x
4π V 0
1
Z 1 Z
= [ λ dλ] εE ~ ·E
~ d3~x
4π Z 0 V
1 ~ ·E~ d3~x
= εE
8π ZV
1 ~ ·D~ d3~x.
= E (8.5)
8π V
Chapter 9
~ ∧E
∇ ~ = 0
~ ·E
∇ ~ = 4πρ
~ ∧B
∇ ~ = 4π J~
c
~ ~
∇ · B = 0.
~ ~x) = J~(~x) , with ∇
We shall assume that ρ(t, ~x) = ρ(~x), J(t, ~ · J(~
~ x) = 0, are given.
We already know how to solve the first pair of equations, so we now concentrate in the
second pair, called the magnetostatics equations:
~ ∧B
∇ ~ = 4π J~ (9.1)
c
~ ~
∇ · B = 0, (9.2)
~ ~x) = J(~
where J(t, ~ x), with ∇~ · J(~
~ x) = 0, is given.
~ x) is needed, for the identity ∇
Notice that this last condition in J(~ ~ · (∇
~ ∧ V~ ) = 0 applied
to the first equation above implies it.
Notice that in contrast with the static equations the sources in this case have vectorial
character, and so they appear in the vectorial equation and not in the scalar one as in the
static case.
As in the electrostatic case our strategy shall be to first find a way to solve trivially the
source-less equation, in this case the scalar one, and then concentrate in the other one. To
do that we introduce the vector potential, A, ~ that is, assume the magnetic field is of the form
~ ~
B = ∇ ∧ A. ~
With this ansatz, then the second equation, the source-less one, is identically satisfied, as
follows from the vector calculus identity mentioned above. The curl equation becomes now,
~ ∇
∇( ~ · A)
~ − ∆A ~ = 4π J,
~
c
133
134 CHAPTER 9. STATIONARY SOLUTIONS: MAGNETOSTATICS
The above equation can not be considered a system of three equations, one for each
component of J, ~ for three unknown, each component of A, ~ for they are not independent
equations. Indeed, if we take the divergence of the left-hand side we get identically zero. In
fact, it is easy to see that the above equation does not determine uniquely A,~ for if A~ is a
′
~ then A
solution for a given J, ~ =A ~ + ∇λ
~ is also a solution, for any given smooth function λ,
~ ∇
for ∇( ~ · ∇λ)
~ − ∆∇λ ~ = ∇(∆λ)
~ ~ = 0.
− ∆∇λ
~
The lack of uniqueness of A does not affect the uniqueness of B ~ as solutions to the
magnetostatics equations, for
~′ = ∇
B ~′ = ∇
~ ∧A ~ ∧A
~+∇
~ ∧ (∇λ)
~ =∇~ ∧A
~ = B.
~
We can use this freedom in the choice of vector potential to get a simpler equation for it.
We do that imposing to the potential an extra condition which makes it unique. That extra
condition is only for mathematical convenience and has no physical meaning. The most
~ ·A
convenient one is to require that ∇ ~ = 0, which is called the Coulomb Gauge. In this case
~ becomes,
the equation for A
~ x) = − 4π ~
∆A(~ J(~x), (9.3)
c
which, if we express the vectors in Cartesian components, is a system of three decoupled
Poisson equations. In cases where this equation has a unique solution we obtain a unique
~ x).
vector potential A(~
Can one always find a gauge where ∇ ~ ·A~ = 0? The answer in most cases is affirmative,
suppose you have a solution in some other gauge, that is a A ~ ′ such that its divergence is not
zero, then one can solve ∆λ = −∇ ~ ′ for some field λ (provided ∇
~ ·A ~ ′ decays sufficiently fast
~ ·A
asymptotically), and so the new potential, A ~=A ~ ′ + ∇λ
~ will have, ∇~ ·A ~=∇ ~ ·A~ ′ + ∆λ = 0
and so will be divergence-less
1
Z ~ y )d3~y
J(~
~
A(~x) = ,
c lR3 |~x − ~y |
~ and J~ are expressed in Cartesian coordinates.
provided that A
But, do we get by this procedure all solution to the original magnetostatics equations?
For isolated systems this is so, as the following theorem asserts:
9.1. THE GENERAL PROBLEM 135
Theorem 9.1 Systems (9.1 - 9.2) and (??) are equivalent, in the sense that given J(~ ~ x) in
~ · J(~
lR3 , of compact support and with ∇ ~ x) = 0, for each solution B(~
~ x) of (9.1- 9.2) there exists
~ x) of (??) and vice-versa, provided both decay asymptotically sufficiently
a unique solution A(~
fast.
~ ∧ (∇
0=∇ ~ ∧ δ B)
~ = ∇(
~ ∇~ · δ B)
~ − ∆δ B,
~
~ = 0.
∆δ B
Expressed in Cartesian coordinates this is a decoupled system of three Poisson equations and
from their uniqueness, for the case that the magnetic field decays asymptotically, it follows
~ = 0. Thus, if we show that B
that δ B ~ ′ := ∇
~ ∧A ~ also satisfies the (9.1 - 9.2 ) equations, we
′
would conclude that B~ =B ~ =∇ ~ ∧ A.
~ The curl of B ~ ′ is
∇ ~′ = ∇
~ ∧B ~ ∧ (∇~ ∧ A)
~ = ∇(~ ∇
~ · A)
~ − ∆A
~
~ ∇
= ∇( ~ + 4π J.
~ · A) ~
c
~ · J~ = 0 we get,
But, taking the divergence of equation (??), and using that ∇
~ · A)
∆(∇ ~ = 0,
~ ·A
and the uniqueness of Poisson’s equations then implies ∇ ~ = 0, so we conclude B
~ satisfies
system (9.1 9.2).
J~ = Jϕ êϕ
= Jϕ (− sin ϕ êx + cos ϕ êy )
′
with Jϕ = I sin θ′ δ(cos θ′ ) δ(r a−a) .
z
θ
ϕ
x
1
Expanding ′ is spherical harmonics we get,
|~
x−~x |
l l
~ θ, ϕ) = I sin θ′ δ(cos θ′ )δ(r ′ − a) X ⋆
Ylm (θ, ϕ)Ylm (θ′ , ϕ′)
Z X
A(r, 4π
c lR3 a l=0 m=−l
2l + 1
l
r< ′ ′ ′2 ′ ′ ′ ′
l+1 (− sin ϕ êx + cos ϕ êy )r dr sin θ dθ dϕ
r>
∞ X l l Z 2π
4πIa X Ylm (θ, ϕ) r< ⋆ π
= l+1 Ylm ( , ϕ′ )(sin ϕ′ êx + cos ϕ′ êy ) dϕ′
c l=0 m=−l 2l + 1 r> 0 2
But,
⋆ π
Z 2π
Ylm ( , ϕ′ )(sin ϕ′ êx + cos ϕ′ êy ) dϕ′
0 2
iϕ′ ′ ′ ′
Z 2π
⋆ π ′ e − e−iϕ eiϕ + e−iϕ
= Ylm ( , ϕ )( êx + êy ) dϕ′
0 2 2i 2
and
⋆ π
Z 2π ′
Ylm ( , ϕ′ )eiϕ dϕ′
0 2 v
s
2l + 1 u (l − 1)!
u
= πδm,1 t Pl1 (0)
4π (l + 1)!
and so,
∞ l
~ θ, ϕ) = πIa
X r< 1 1 (l − 1)! iϕ
A(r, l+1 [Pl (0)Pl (cos θ) e (iêx + êy )
c l=0 r> (l + 1)!
(l + 1)! −iϕ
+Pl−1(0)Pl−1 (cos θ) e (−iêx + êy )].
(l − 1)!
9.2. BOUNDARY CONDITIONS - SUPER CONDUCTORS 137
Using now, Pl−1(x) = (−1) (l−1)! P 1(x), the above expression becomes,
(l+1)! l
∞ l
~ θ, ϕ) = πIa r< 1
P 1 (0)Pl1(cos θ)[−2 sin ϕêx + 2 cos ϕêy ]
X
A(r,
c l=0 r> l(l + 1) l
l+1
∞ l
2πIa X r< 1
= êϕ [ l+1 Pl1 (0)Pl1 (cos θ)],
c l=0 r> l(l + 1)
where,
0 l even
Pl1 (0) = (−1)n+1 Γ(n+ 23 )
Γ(n+1)Γ( 32 )
l = 2n + 1
Recall that the Gamma function takes the following values,
√
1 π
Γ(n) = n!, Γ(n + ) = n (2n − 1)!!
2 2
Far away from the source, (r> = r, r< = a), the leading contribution will come from the
first non-null term, that is,
2 2
~ θ, ϕ) ≡ − πIa P 1 (cos θ)êϕ = πIa sin θêϕ
A(r,
cr 2 1 cr 2
2
πIa
≡ (−yêx + xêy )
cr 3
m
~ ∧ ~x
≡ ,
r3
πa2 I
~ :=
with m c
k̂ = (Area of circular loop) Ic k̂. To this approximation the magnetic field is,
~ =∇
B ~ ∧A ~ ∧ (m
~ ≡ ∇ ~ ∧ ~x
) = ~ 1 ) ∧ (m
∇( ~ ∧ ~
x ) +
1
(m(
~ ∇~ · ~x) − (m ~ x)
~ · ∇)~
r3 r3 r3
−3~x 3m~ m
~
≡ ( 5 ) ∧ (m ~ ∧ ~x) + 3 − 3
r r r
−3 2m~
≡ 5
(m(~
~ x · ~x) − ~x(~x · m))
~ + 3
r r
3(n̂ · m~ )n̂ − m~
≡ ,
r3
which in analogy with the electrostatic expression is called a magnetic dipole.
The same integral argument used for conductor boundaries in the electrostatic case (8.2.1)
says here that:
1. From ∇~ ·B~ = 0, the normal component of the magnetic field must be continuous and
~ · n̂|∂V = 0.
therefore B
~ B
2. From ∇∧ ~ = 4πJ~ , the jump on the tangential component is due to a superficial current
c
density, 4πc k̂ = −B~ ∧ n̂|∂V .
Since the second condition now includes a vector source we deduce it again. Following the
notation in figure 8.7 we integrate the magnetic field along the curve γ, as in that case we
call n̂ the unit normal, ˆl the unit vector along the outer part of the curve γ. It is understood
that we are in the limit of a small curve so that their pieces along both sides of the interface
are almost straight lines and the pieces entering the interfaces are very small. Since these two
vectors are perpendicular among each other, the vector ŝ = n̂ ∧ ˆl is unitary, perpendicular to
both, and normal to the surface enclosed by γ. In our figure it would be a vector entering the
page.
Using Stokes theorem we get:
Z l1 Z
~ · ˆl dl =
B ~ · d~l
B
l0 γ
Z
= ~ ∧ B)
(∇ ~ · dS
~
S(γ)
4π
Z
= J~ · dS
~
c S(γ)
4π l1 re ~
Z Z
= J · ŝ dl dr
c l0 ri
~ · ˆl dl = 4π
Z l1 l1 Z
B ~k · ŝ dl
l0 c l0
Since the integrals have arbitrary length we conclude that,
4π ~ ~ · ˆl
k · ŝ = B
c
=~ · (ŝ ∧ n̂)
B
= −B~ ∧ n̂ · ŝ,
from this, and the arbitrariness of ˆl, hence, ŝ, a the above interface matching condition follows.
9.2. BOUNDARY CONDITIONS - SUPER CONDUCTORS 139
This type of problems can be solved using the vector potential equation with the following
boundary conditions:
A~ ∧ n̂|∂V = 0
~ · [n̂(n̂ · A)]|
∇ ~ ∂V = 0.
The first condition guarantees B ~ · n̂|∂V = 0, for B~ · n̂ contains only tangential derivatives
of the tangential components of A. ~ Depending on the physical situation this condition is too
restrictive, we shall see this latter in an example. The second implies, together with the first,
that ∇~ · A|
~ ∂V = 0, a boundary condition sufficient to ensure that ∇ ~ ·A
~ = 0 everywhere outside
the superconductor, and so that B ~ =∇ ~ ∧A ~ is the solution sought.
~k
~0
B
a
Choosing the z axis along the constant external magnetic field we have,
~ 0 = B0 k̂.
B
This field has a vector potential given by
~ 0 (~x) = B0 ∇
~ ∧A
∇ ~ ∧ (k̂ ∧ ~x) = B0 (k̂ ∇
~ · ~x − (~k · ∇)~
~ x)
2 2
B0
= (3k̂ − k̂) = B0 k̂.
2
140 CHAPTER 9. STATIONARY SOLUTIONS: MAGNETOSTATICS
~ 0 does not single out any point in space, and so any coordinate origin,
Notice that while B
~ 0 (~x) does. That means that the difference between two vector potentials for B
A ~ 0 with
different origins must be just the gradient of a function, indeed,
~ ~ ~
~ −A
A ~ 0 = B 0 (k̂ ∧ ((~x + ~r)) − B 0 (k̂ ∧ ~x) = B 0 k̂ ∧ ~r
~r 2 2 2
~
= ∇(~ B 0 (k̂ ∧ ~r) · ~x).
2
For this problem it is convenient to choose as coordinate origin the center of the supercon-
ducting sphere and a external vector potential centered on it.
The external field would induce currents on the surface of the superconducting sphere
which would rotate along circular loops, all of them perpendicular to the k̂ direction, these
in turn would generate a vector potential, which as in the case of the single circular current
loop, would have only component along the eˆϕ direction. Thus, the induced vector potential
would have the form:
A~ I (~x) = AI (r, θ)(k̂ ∧ ~x).
This vector is tangent to all spheres centered at the origin, and so the boundary condition,
~ · [n̂(A
∇ ~ · n̂)]|r=a = 0
is satisfied trivially.
The other boundary condition then implies,
~ θ, ϕ)|r=a = ( B0 + AI (r, θ))(k̂ ∧ ~x)|r=a = 0,
A(r,
2
but then AI (a, θ) = − B20 , independent of θ. But we have already found a solution with these
characteristics, namely,
~ I (~x) = m
A
~ ∧ ~x
,
r3
3
~ = − B02a k̂.
with m
The total solution is then,
3
~ x) = (k̂ ∧ ~x)( B0 − B0 a ),
A(~
2 2r 3
and
~ B0 a3
B(~x) = B0 k̂ − (3(n̂ · k̂)n̂ − k̂).
2r 3
Notice that A~ has only component along φ̂, vanishes at r = a, and growth as r → ∞.
This last property is not relevant, for it is only a property of the gauge.
The induced surface current is:
~k = c (B~ ∧ n̂)|r=a = c B0 (k̂ ∧ n̂ − 3 (n̂ · k̂)(n̂ ∧ n̂) + 1 k̂ ∧ n̂) = 3cB0 k̂ ∧ n̂.
4π 4π 2 2 8π
9.3. THE MAGNETIC SCALAR POTENTIAL 141
That is, a Neumann boundary value problem. In this case, ϕ0 (~x) = −B0 k̂·~x and ϕI (~x) = α k̂·r~3x ,
the only other solution to Poisson’s equation with this angular dependence. The boundary
condition implies,
3α ~
~ 0 + ϕI )|r=0 = −B0 k̂ · ~n − α~
n̂ · ∇(ϕ k · ~
n + k · ~n
a3 a3
2α
= (−B0 − 3 )~k · ~n = 0
a
3
Thus, α = − Bo2a , and the problem is solved. We see that for boundary value problems in
vacuum this method is very useful, for it reduces the problem to a single Poisson equation
which we can handle very easily with the techniques already learn for electrostatics. But one
has to be very careful, for there are situations where the magnetic potential can not be defined
everywhere outside the superconducting bodies.
To see how this problem arises we shall first give an argument showing why the definition
of a potential does always works in the electrostatic case. Given a curve γ in lR3 with starting
142 CHAPTER 9. STATIONARY SOLUTIONS: MAGNETOSTATICS
~x
~
E
~
E
γ
~
E
~x0
If we take any another curve γ̃, also starting at ~x0 and ending at ~x, we can also define
φγ̃ (~x), we claim φγ (~x) = φγ̃ (~x) and so this procedure really define a function in lR3 , φ(~x),
independent of any particular curve chosen to compute it.
~x
γ̃
~
E
~
E
γ
~
E
~x0
Indeed,
Z Z
φγ (~x) − φγ̃ (~x) = ~ · d~l −
E ~ · d~l
E
γ γ̃
I
= ~ · d~l,
E
γ̃ −1 γ
9.3. THE MAGNETIC SCALAR POTENTIAL 143
where γ̃ −1 is the curve “going backwards" of γ̃, i.e. if γ̃ is given by ~x˜(s), then γ̃ −1 is given by,
(~x˜−1 (s) = ~x˜(1 − s). But then the curve γ̃ −1 γ is a closed curve, starting at ~x0 , going (with γ)
up to ~x and returning (with γ̃ −1 ) to ~x0 . Using Stokes theorem we then have,
Z
φγ (~x) − φγ̃ (~x) = ~ ∧ E)
(∇ ~ · n̂ dS = 0,
S
where S is any surface having as boundary the curve γ̃ −1 γ, [see figure], and we have used the
~ ∧E
electrostatic equation, ∇ ~ = 0.
~x
γ̃ −1
~
E
~
E
γ
~
E
~x0
Exercise: Choose families of curves which go along the three coordinate axis to show that
~ = −E.
∇φ ~
Thus, we conclude that the electrostatic potential is always well-defined on lR3 . On the
contrary this is not the case for the magnetic potential, for, if we similarly define
Z
ϕmγ (~x) = − ~ · d~l,
B
γ
then it is clear from the above argument that if currents are somewhere present, then the
potential does depends on the loop. To see this consider the following current distribution,
(see figure) where we have a closed current loop.
Thus, if we go to the point ~x along γ we get some value for φm (~x), φmγ (~x), if we go along
γ̃ we get some other value, its difference is
I Z
ϕmγ (~x) − ϕmγ̃ (~x) = ~ · d~l =
B ~ ∧ B)
(∇ ~ · n̂ dS
γ̃ −1 γ S
4π 4π
Z
= J~ · n̂ dS = I,
c S c
that is, proportional to the total current along the loop.
144 CHAPTER 9. STATIONARY SOLUTIONS: MAGNETOSTATICS
~
B ~x
γ̃
~
B
γ
~
B
~x0
9.3.1 Wires
We consider now the idealization of an infinitely thin current loop, that is a current line o
wire. If the current circulates along a closed loop γ : [0, 1], given, say, by ~x(s) : [0, 1] → lR
with ~x(1) = ~x(0). Then a current density is a distribution given by:
1 d~x(s)
Z
~ x) =
J(~ I δ(~x − ~x(s)) ds,
0 ds
where I is a constant denoting the total current flowing along the loop, and d~xds(s) is the
circulation velocity of the chosen parametrization of the curve. On the above expression only
the tangency of that velocity is relevant, for one can see that the integral does not depend on
the parametrization chosen to describe the curve.
wire
d~
x(s)
ds
~x(s)
To see how to deal with such distributions, we shall check now that it satisfies charge
conservation, namely,
9.3. THE MAGNETIC SCALAR POTENTIAL 145
d~x 1
Z Z Z
~ x) · ∇ϕ(~
J(~ ~ x) d3~x = ~ x) ·
I ∇ϕ(~ dsδ(~x − ~x(s)) d3~x
3 3
lR lR 0 ds
Z 1
d~x ~ Z 1
dϕ(~x(s))
= I · ∇ϕ(~x(s))ds = I ds
0 ds 0 ds
= I[ϕ(~x(0)) − ϕ(~x(1))] = 0,
~ x) = 1
~ y )d3~y
J(~ I d~l
Z I
A(~ = ,
c |~x − ~y | c γ |~x − ~x(s)|
~ x)
dA(~
~x − ~x(s) d~
x(s)
ds
~x
~x(s)
Similarly,
~ x) = −I (~x − ~y ) ∧ d~l
I
~ x) = ∇
B(~ ~ ∧ A(~ ,
c γ |~x − ~y |3
which is Biot - Savart’s law.
Contracting the vector potential with a constant vector field, ~k, using Stokes theorem we
146 CHAPTER 9. STATIONARY SOLUTIONS: MAGNETOSTATICS
have,
I ~
~ x) = I
~k · A(~ k · d~l IZ ~ ~k
= ∇~y ∧ ( ) · n̂ dS
c γ |~x − ~y | c S |~x − ~y |
IZ ~ 1
= (∇~y ( ) ∧ ~k) · n̂ dS
c S |~x − ~y |
I ~ ~ ( 1 )) dS,
Z
= k · (n̂ ∧ ∇ ~y |~x − ~y |
c S
where S is any surface whose boundary is γ, and n̂ its unit normal. Thus,
~ x) = − I ~ ( 1 ) ∧ n̂ dS,
Z
A(~ ∇~y |~x − ~y |
c S
and so,
~ x) = −I ∇ ~ ( 1 ) ∧ n̂) dS
Z
B(~ ~ ∧ (∇
c S ~
x ~y |~x − ~y |
−I ~ ( 1 )) + (n̂ · ∇ 1
Z
= ~ ·∇
[−n̂(∇ ~ )(∇ ~
c S ~x ~y |~x − ~y | ~x ~y |~x − ~y | )] dS
I ~ ( 1 )) + (n̂ · ∇ 1
Z
= [−n̂(∇ ~ ·∇ ~ )(∇ ~ )] dS
c S ~
y ~
y |~x − ~y | ~
y ~
y |~x − ~y |
I ~ ( 1 )] dS
Z
= [n̂ 4πδ(~x − ~y ) − (n̂ · ∇ ~ )∇
c S ~
y ~x |~x − ~y |
−I ~ ~ ( 1 ) dS]
Z
= ∇~x [ n̂ · ∇ ~y |~x − ~y |
c S
1
where in the second step we have used that, ∆( ) = −4πδ(~x − ~y ), and in the third that
|~
x−~y |
~x is not a point along the line current. So we see that
I 1 I n̂ · (~x − ~y )
Z Z
ϕm (~x) = ~ (
n̂ · ∇ ) dS = dS.
c S
~y |~x − ~y | c S |~x − ~y |3
From the figure, and changing the coordinate origin to the point ~x, that is, using a new
integration variable ~y˜ = ~y − ~x, the above integral becomes,
I n̂ · m̂
Z
ϕm (~x) = dS,
c S |~y˜|2
˜
with m̂ = ~y˜ = (~x−~y ) . Note that n̂ · m̂ dS is just the area as seen from ~x, that is, the solid
|~
y| |~
x−~y |
angle that differential spans times the square of the distance to ~x. So, n̂·˜m̂2 dS is just the
|~
y|
differential of solid angle spun by dS as seeing from ~x. Correspondingly the integral is just
the solid angle spun by the whole surface S. If we approach that surface from bellow n̂· m̂ > 0
the value of that solid angle in the integration tends to 2π, while if we approach the surface
9.3. THE MAGNETIC SCALAR POTENTIAL 147
n̂ S
~x − ~y
γ
~x
~y
n̂ S
m̂
from above, n̂ · m̂ < 0, the solid angle covered tends to −2π, and so we have a jump of 4πI
c
on
ϕm . The surface S is arbitrary, as long as its boundary is γ, and so we can choose the jump
whatever we please, but it has to be somewhere.
1 d~x(s)
Z
~ x) = I
J(~ δ(x − x(s)) ds,
0 ds
then,
Z
J~ · n̂ dS = I.
S
1 d~x(s)
Z
~
J(φ) =I φ(x(s)) ds,
0 ds
and this distribution is not so wild that one can even apply to other distributions, in particular
to surface distributions. Among surface distributions is n̂S which has support just on a smooth
surface S and there is its unit normal. When applied to a smooth compactly supported test
vector li it gives the flux of such a vector across S. We claim that J i (n̂i ) = I and so represents
the integral flux of J~ across S. To see this consider a thickening of the surface S and the
following distributional normal,
1 ~
(
∇r 0≤r<ε
n̂εi = 2ε (9.4)
0 r≥ε
where r is the distance from S into a neighborhood of it, and we are taking ε small enough
so that the distance is well-defined and smooth. In the limit ε → 0 this gives a good integral
representation of n̂S .
But,
i ε I Z s(ε)+ d~x(s) ~
J (n̂i ) = · ∇rds,
2ε s(ε)− ds
where s(ε)± is the value of s for which r(~x(s)) = ε before and after the loop goes into the
region where r < ε. If ε is small enough the relation r = r(~x(s)) can be inverted, and we can
define s(r), so after using r as variable in the above integral we get,
I
Z +ε
d~x(r) ~
J (n̂εi ) =
i
· ∇rdr = I.
2ε −ε dr
Thus we can take the limit ε → 0 and get the correct result.
9.4. NON-SIMPLY CONNECTED SUPER-CONDUCTORS 149
Multiplying by φm the first equation, integrating on the whole space outside the sphere, and
assuming the fields decay at infinity we find that φm = cons, and so that B ~ ≡ 0. The same
argument follows for any superconductor body whose topology implies that φm is well-defined
everywhere outside it. But this is true provided any closed loop outside the body can be
continuously deformed to zero. Bodies with this property are called simply connected.
What happens in the case of bodies which are not simply connected? As is the case of a
superconducting ring?
Physically it is reasonable that we can have configurations where a current of arbitrary
total intensity flows along the ring. In fact, we shall show latter how to build such a configu-
ration. Thus, we expect in this case to have many non-trivial solutions.
Mathematically we can look for these solutions in the following way: Since the magnetic
potential will necessarily have discontinuities, –but we can choose where they are going to
be– we set the following boundary value problem (see figure):
SR
D+
D−
Proof: We only prove uniqueness and not existence. First notice that at the disk inside the
~ + = n̂ · ∇φ
ring, D, we must have n̂ · ∇φ ~ − . Indeed, applying Gauss theorem to ∇ ~ ·B~ = 0 in
m m
a pill-box containing a piece of the disk and taking the limit on which the pill-box flattens
over the disk the assertion follows. Assume now we have two solutions, φ1 and φ2 satisfying
the above problem, then δφ ≡ φ1 − φ2 satisfies,
But then,
Z
0 = − δφ∆δφ d3~x (9.16)
V
Z Z
= ~
∇δφ ~
· ∇δφ d3~x − ~
δφn̂ · ∇δφ ~
d2 S (9.17)
V SR
Z Z
− ~ + d2 S
δφ+ n̂ · ∇δφ ~− ~ − d2 S
δφ− n̂ · ∇δφ ~ (9.18)
D+ D−
Z
= ~
|∇δφ| 2 3
d ~x (9.19)
ZV
− ~ + − n̂ · ∇δφ
δφ+ (n̂ · ∇δφ ~
~ − ) d2 S, (9.20)
ZD
= ~
|∇δφ| 2 3
d ~x, (9.21)
V
where in the third equality it was used that at the disk, δφ+ = δφ− , and in the fourth that
at the disk both gradients were also the same, and we have taken the normal to be the one
incoming into the upper disk. This shows that ∇δφ~ = 0, and since δφ → 0 as |~x| → ∞, we
conclude δφ = 0 everywhere. Thus, the total current flow suffices to determine uniquely the
solution. 1
In electrostatics, we saw that we could give either the potential V or the total charge Q
on a conductor (or the corresponding arrays of potential or charges in the case of an array of
conductors) and that –together with the geometry of the bodies– would determine a unique
solution. Does there exist in the case of superconductors another quantity that one could
specify and so determine a unique solution? The answer is affirmative and the other quantity
is the total magnetic flux, Z
Φ≡ B ~ · n̂ d2 S, (9.22)
S
1
In fact one could have avoided this calculation since δφ is smooth at the disk surface, and so we can
extend the Gauss surface pass that surface and just wrap up the superconductor, but at that surface we just
~
use n̂ · ∇δφ|S = 0 to finish the argument.
9.4. NON-SIMPLY CONNECTED SUPER-CONDUCTORS 151
where S is any surface whose boundary, γ, meets the body. In the case of the ring one could
take, for instance, the disk D.
Using that B ~ =∇ ~ ∧ A,
~ and Stokes theorem we get,
I
Φ= ~ · d~l.
A (9.23)
γ
~
B ~
B
~
S B
n̂
~
B
~
A
γ
Notice that this can be taken as an integral condition on the magnetic field, and so it is, in
some sense, like the condition that the integral of the normal to the electric field on the surface
of a conductor is the total charge it contains. In fact, it is the total flux which is the analog to
the total charge and not the total current, for if we have an array of conductors with potentials
and charges, (V i , Qi ), and move them around changing their geometrical configurations, then
their potentials V i will change values, but not their total charges Qi . On the other hand, if
we have an array of superconducting bodies and move them around, then their currents I i
will change, but not their fluxes Φi , as follows from the following calculation,
dΦ ∂ ~
Z
= B · n̂ d2 S (9.24)
dt S ∂t
Z
= −c (∇ ~ ∧ E)~ · n̂ d2 S (9.25)
IS
= −c ~ · d~l
E (9.26)
γ
= 0. (9.27)
Since the last integral is along the border of the superconductor and there E~ can only have
normal component. Since this calculation is valid for any one of the bodies it shows that the
flux on each one of them is constant in time, and so they will not change if we change the
superconductor configuration.
152 CHAPTER 9. STATIONARY SOLUTIONS: MAGNETOSTATICS
Φi = L̃ij I j , (9.28)
for some matrix L̃ij , that is, the flux is a function of the currents, and it is a linear function. We
claim, without proving it, that this relation is also invertible. The coefficients Lij ≡ cL̃ij are
called inductances and, as the capacities, they only depend on the geometrical configuration
of the system.
Φ1 I1 Φ2
I2
Φ3
I3
1
Z ~ y)
J(~
~
A(~x) = d3~y ,
c lR3 |~x − ~y |
valid in Cartesian coordinates, or contracting with an arbitrary constant vector ~k,
~k · J(~
~ y)
~ x) = 1
Z
~k · A(~ d3~y ,
c lR3 |~x − ~y |
in an arbitrary coordinate system.
9.5. MULTIPOLAR EXPANSION OF THE MAGNETOSTATICS FIELD 153
But this expression is identical to the expression for the scalar potential en electrostatics,
so we can proceed as in electrostatics and make a Taylor series expansion of the function
1
,
|~
x−~y |
1 1 ~ 1 ) + ...
= − ~y · ∇(
|~x − ~y | |~x| |~x|
and so obtain the leading behavior of the magnetostatic field at large distances away from
the current sources,
~ x) = 1 ~ 1 ) + . . .] d3~y ,
~k · J~(~y )[ 1 − ~y · ∇(
Z
~k · A(~
c lR3 |~x| |~x|
1 1 ~ 1 )·
Z Z
= [ ~k · J(~
~ y ) d3~y − ∇( ~y ~k · J(~
~ y ) d3~y + . . .]. (9.29)
c |~x| lR3 |~x| lR3
~ ~k · ~x)) = (~k · ~x)∇
~ · (J(
But ∇ ~ · J~ + J~ · ~k = J~ · ~k, for ∇
~ · J~ = 0, so
Z Z Z
~k · J(~
~ y ) d3~y = ~ ~k · ~y )) d3~y =
~ · (J(
∇ J~(~k · ~y ) · n̂ dS 2 = 0, (9.30)
3 3
lR lR S 2 (∞)
since the sources are assumed to have compact support. Thus, since −∇( ~ 1 ) = n̂ , we have
|~
x| |~
x|2
the very important fact that the magnetostatic field does not have any monopole contribution,
that is, it decays one order (in 1r ) faster than the electric field, as we go away from the sources.
The divergence free property of J~ can also be used to write in a more transparent way the
subsequent terms in the series. To do that, notice that in general,
~ · (Jf
∇ ~ ) = f∇
~ · J~ + J~ · ∇f
~ = J~ · ∇f.
~
And since J~ has compact support,
Z Z
3
J~ · ∇f
~ d3~y =
3
~ · (Jf
∇ ~ ) d3~y = 0.
lR lR
To handle the first term we used f = ~x · ~k. For the second it is convenient to use a vector,
namely f = ~x(~x · ~k), then,
~ x) ≡ m
A(~
~ ∧ ~x
,
|~x|3
and correspondingly,
~ x) = 3n̂(n̂ · m)
B(~
~ −m~
.
|~x| 3
−1 I
Z I
~ =
m 3
J~ ∧ ~x d3~x = ~x ∧ d~l.
2c lR 2c
~l|
If the current loop is contained in a plane, then |~x∧d2
is the area of the infinitesimal triangle
2
of the figure below , and so
I
× Circuit area
~ =
|m|
c
while the direction of m
~ is perpendicular to the plane where the circuit lies.
If the current is due to point like charges in motion, then
J~ =
X
qi~vi δ(~x − ~xi ),
i
and so,
1 X
~ =
m qi (~xi ∧ ~v i ).
2c i
If the charges have mass mi and therefore angular momentum,
2
To see this, first notice that Area = h(b12+b2 ) . Choosing the ê1 vector along the vector ~x + d~l, we have,
h = −d~l · ê2 = ~x · ê2 , b1 = d~l · ê1 , b2 = ~x · ê1 . Thus, h b1 = ~x · ê2 d~l · ê1 , h b2 = −d~l · ê2 ~x · ê1 . And so,
h(b1 + b2 ) = |d~l ∧ ~x|.
9.5. MULTIPOLAR EXPANSION OF THE MAGNETOSTATICS FIELD 155
d~l
~x + d~l
~x
ê1
ê2 b1 d~l
~x + d~l h
b2
~x
~ i = mi (~xi ∧ ~v i ),
L
then,
1 X qi ~
~ =
m Li .
2c i mi
In the case of identical particles, or in the more general case of particles with the same charge
to mass ratio, we have,
q ~
~ =
m L,
2cm
where L~ is the total angular momentum of the system. This relation is called the giro-magnetic
momenta. It works well for even orbital electrons, if we do not take into account their intrinsic
angular momentum or spin. For their intrinsic angular momentum this relation fails in the
sense that the numerical factor (in the case above 21 ) takes another value.
Chapter 10
As we have seen, the energy stored in a magnetic field in a volume V is given by,
1 Z ~ ~ 3
E= B · B d ~x. (10.1)
8π V
In the case of a stationary field this quantity can be expressed in terms of the currents which
~ B
generate the fields. To do that we first use that since ∇· ~ = 0, there exists a vector potential,
~ ~ ~ ~
A such that B = ∇ ∧ A. Second we substitute it for one of the B’s ~ above and use the identity,
~ · (∇
V ~ ∧W ~ )−W ~ · (∇
~ ∧V~)=∇ ~ · (W
~ ∧V ~ ). Integrating by parts we obtain,
1 Z ~ ~ ~ d3~x + 1
I
~ ∧ B)
~ d2 S.
E= (A · ∇ ∧ B) n̂ · (A (10.2)
8π V 8π ∂V
Finally we use now one of Maxwell’s stationary equations, ∇ ~ ∧B~ = 4π J~, to obtain,
c
1
Z I
E= ~ · J~ d3~x +
A ~ ∧ B)
n̂ · (A ~ d2 S. (10.3)
2c V ∂V
Exercise: Check that the above expression, as the original one, does not depend upon any
~ That is, check that the expression remains the same if we use
particular gauge chosen for A.
′
~ =A
A ~ + ∇λ,
~ for any arbitrary smooth function λ. a.-) Consider only the case V = lR3 and
assume the fields decay sufficiently fast at infinity so that no surface integral contributes. b.-)
Consider the case of arbitrary region V .
157
158 CHAPTER 10. THE ENERGY OF THE MAGNETOSTATIC FIELD
1 Z ~ ~ I )2 d3~x − 1
Z
~ 2 d3~x = 1
Z
~ I + 2B~ 0) · B
~ I d3~x
∆E2 = (B 0 + B B (B
8π V 8π V 0 8π V
where B~ 0 is the constant field and B ~ I the induced one due to the presence of the sphere. We
now proceed as in the general case and use that B ~ I + 2B
~0 = ∇~ ∧ (A~ I + 2A ~ 0 ) to get, after
integration by parts,
1 Z ~ ~ I + 2A ~ 0) · B~ I d3~x
∆E2 = ∇ ∧ (A
8π ZV
1 ~ · ((A~ I + 2A ~ 0) ∧ B ~ I ) − (∇~ ∧B
~ I ) · (A
~ I + 2A
~ 0 )) d3~x
= ∇
8π ZV
−1 ~ I + 2A ~ 0) ∧ B ~ I ) · n̂) d2 S
= ((A
8π ZS
−1 ~0 ∧ B~ I ) · n̂) d2 S
= (A (10.5)
8π S
where we are using the normal outgoing from the sphere surface, that is incoming into V . In
the third equality we used that the current is zero in V and in the fourth that at the surface
~I + A
of the sphere the total potential, A ~ 0 vanishes. From the previous chapter computation
we know that
k̂
$region 2$
n̂
$region 1$
~k
a3 B02
Thus, ∆E = ∆E1 + ∆E2 = − 12
.
1 ~ x) · J(~
J(~
Z Z ~ y)
E = d3~x d3~y , (10.7)
2c2
V V |~x − ~y |
1 X i j d~li · d~lj
I I
= I I , (10.8)
2c2 i,j γi γj |~ xi − ~xj |
where γi and I i are respectively the path and current of the i-th current line.
This quantity is ill-defined, for each of the integrals with i = j diverges. As in the case
of point charges in electrostatics we redefine the energy by dropping all self-energies, and get
the interaction energy ,
X
EI = Eij , (10.9)
i<j
where,
I iI j d~li · d~lj
I I
Eij = 2 . (10.10)
c γi γj |~xi − ~xj |
160 CHAPTER 10. THE ENERGY OF THE MAGNETOSTATIC FIELD
J~1
J~2
J~3
1
Z
Eij = J~i · A
~ j d3~x, (10.11)
c V
Ii I ~
Eij = Aj · d~li (10.12)
c γi
Ii
Z
= ~ j · n̂i d2 Si
B (10.13)
c Si
Ii
= Φij (10.14)
c
where in the first step we have used the explicit formula for the i-th current distribution, in
the second Stokes theorem to transform a line integral into a surface integral, where Si is
any surface such that its boundary, ∂Si = γi , and in the last step we have defined Φij as the
magnetic flux due to the j-th current across the i-th current loop.
d~li · d~lj
I I
Lij = , (10.15)
γi γj |~xi − ~xj |
called the inductance of the i-th circuit with respect to the j-th circuit. It is clear that
Lij = Lji and that,
1 X
EI = 2
Lij I i I j . (10.16)
c i<j
Comparing with the expression for Eij in terms of the magnetic flux we see that the total flux
across the i-th loop due to the magnetic fields generated by the other loops is,
1X
Φi ≡ Lij I j . (10.17)
c j6=i
Notice that there is a close analogy between the triplets (I i , Φi , Lij ) and (Qi , V i , Cij ) of
electrostatics of perfect conductors. Can this analogy be extended to more realistic circuits?
The answer is yes, but one has to include the more general case the self inductances of circuits,
for now they are no longer divergent and so play a role in the energetics and fluxes. We treat
now superconducting circuits, for there the idealization allows for a strong argument. We
start with an example:
−1
Z
E = B~ · ∇φ
~ m d3~x (10.19)
8π V
1 1
Z I
= ~ ~ 3
φm ∇ · B d ~x − ~ · n̂ d2 S,
φm B (10.20)
8π V 8π ∂V
where we have used Gauss theorem and taken as the boundary for V the surface shown in
the figure, for φm is taken to be discontinuous on the inner ring plane. The contribution from
the part of ∂V at the surface of the ring vanishes, for there B ~ · n̂ = 0, but the part in the
plane gives,
1
Z
E = (φ+ − φ− ~ 2
m )B · n̂ d S (10.21)
8π ∂V m
φ+m − φm
− Z
~ · n̂ d2 S
= B (10.22)
8π ∂V
I
= Φ, (10.23)
2c
162 CHAPTER 10. THE ENERGY OF THE MAGNETOSTATIC FIELD
SR
D+
D−
4π
where we have used that φ+ −
m − φm = c
I, and defined,
Z
Φ= ~ · n̂ d2 S,
B (10.24)
∂V
that is, the self-flux across the ring. But on the other hand, by definition,
LI
Φ= . (10.25)
c
so we see that
1
E= LI 2 , (10.26)
2c2
For a superconducting ring of radius a and section radius b one can compute L and find,
8a
L = 4πa[ln( ) − 2]. (10.27)
b
In the line current limit, (a → 0), L → ∞, and correspondingly Φ, but in such a way that
their ratio, which is Ic , stays constant.
Notice that,
Z
Φ = ~ · n̂ d2 S
B (10.28)
ZS
= ~ ∧ A)
(∇ ~ · n̂ d2 S (10.29)
IS
= ~ · d~l,
A (10.30)
γ
where S is any surface with boundary a loop γ at the surface of the ring, and we have used
in the last equality Stokes theorem. Thus, since L is a geometrical factor, a non-zero current
~ must have a non-zero component along
results then in a non-zero flux and so, in this case, A
the boundary of the ring.
Chapter 11
Magnetic Materials
In this chapter we treat macroscopic fields resulting from averages upon materials which
interact with magnetic fields producing changes which can be accounted by defining new
averaged fields and a constitutive relation among them.
As in the treatment of macroscopic fields on electrically susceptible materials we start
with the microscopic equations,
~ ∧B
∇ ~ = 4π J~ (11.1)
c
~ ~
∇ · B = 0, (11.2)
and take averages, since them commute with derivatives the only effect on the equations is to
change the sources, in this case the currents. We won’t go into the details of that calculation
for two reasons, first because the one that can be done it is very similar to the one performed
for dielectrics, second because in magnetic materials most of the effect is quantum in nature,
basically due to the presence of spins on electrons and orbitals which get only discrete values,
and all of them non-vanishing, thus, having an influence even in the absence of external fields.
On making the calculations one obtains to first order on the field strength,
~ ∧B
∇ ~ ϕ = 4π J~ϕ = 4π J~f ree + 4π ∇
~ ∧M
~ (11.3)
c c
~ ·B
∇ ~ϕ = 0 (11.4)
where
~ x) =
X
M(~ Ni (x) < m
~i> (11.5)
i
I I
I
~ · ~k =
m (~x ∧ d~l) · ~k = Areak , (11.6)
2c c
163
164 CHAPTER 11. MAGNETIC MATERIALS
~ ≡B
H ~ − 4π M,
~ (11.7)
and write the equations as:
∇ ~ = 4π J~f ree
~ ∧H (11.8)
c
~ ~
∇·B = 0 (11.9)
~ = B(
2. Hysteresis curve: B ~ H)
~ (see figure)
The first relation applies to normal materials and is only valid for small values of the
fields, the second has a saturation range, which is expected of many materials but also has
two more special features, one is that gives a nonzero value for B ~ even when there is no H,~
that is, it describes magnets. The other is that it is not a single valued function, the value of
B~ depends on the history of the material.
~2 − B
(B ~ 1 ) · n̂|S = 0 (11.10)
~2−H
(H ~ 1 ) ∧ n̂|S = 4π ~k l (11.11)
c
where ~k l is a possible free current on the interface. We see that the normal component of B ~
~
is continuous while the tangential components of H have a jump proportional to the possible
free current distributions there.
~ = µH,
If the relation is linear, B ~ and there is no superficial free currents, then we have,
~2 − B
(B ~ 1 ) · n̂|S = 0 (11.12)
~ 2 − µ2 B
(B ~ 1 ) ∧ n̂|S = 0 (11.13)
µ1
or
~2− µ1 ~
(H H 1 ) · n̂|S = 0 (11.14)
µ2
~2−H
(H ~ 1 ) ∧ n̂|S = 0 (11.15)
~ field.
if one prefers to work with the H
~ ∧H
∇ ~ = 0, (11.16)
~ ·B
∇ ~ = ∇
~ · (H
~ + 4π M)
~ = 0, (11.17)
~ = −∇φ
and –for simply connected bodies– we can use a magnetic potential φm such that H ~ m.
Thus we obtain,
~ ·M
∆φm = 4π ∇ ~ ≡ −4πρM , (11.18)
~ ·M
with ρM ≡ −∇ ~ . Thus, if M
~ is smooth,
~ ·M
∇ ~ (~y )
~y
Z
φm (~x) = − 3
d3~y , (11.19)
lR |~x − ~y |
which, integrating by part can be transformed into,
166 CHAPTER 11. MAGNETIC MATERIALS
1
Z
φm (~x) = ~ (~y ) · ∇
M ~ ( )d3~y (11.20)
lR 3 ~y
|~x − ~y |
~ ( 1 )d3~y
Z
= − 3M ~ (~y ) · ∇ (11.21)
lR ~x |~x − ~y |
Z
M~ (~y )
= −∇~ ·( d3~y ). (11.22)
3 |~x − ~
y |
lR
In particular, far away from the source,
1
Z
~
φm (~x) ≈ −∇( ) ~ y )d3~y ,
M(~ (11.23)
|~x| lR3
m
~ · ~x
≈ , (11.24)
r3
where m
R
~ ≡ lR3 M ~ (~y )d3~y is the total magnetic moment of the magnetic source.
If M~ is discontinuous, as is the case if the material ends abruptly at a surface S, some of
the formulas above are incorrect, and we have to proceed with care, for in this case ∇ ~ ·M~ is
only a distribution. In this case we interpret equation (11.18) as a distributional equation in
the following sense: If φm and M ~ were smooth, then equation (11.18) would be equivalent to
the following infinite set of relations:
Z
~ · M(~
u(~x)[∆φm (~x) − 4π ∇ ~ x)]d3~x = 0, for all smooth u(~x) of compact support. (11.25)
lR3
But this set of equations in turn are equivalent to:
Z
~ x)·M(~
[∆u(~x)φm (~x)+4π ∇u(~ ~ x)]d3~x = 0, for all smooth u(~x) of compact support, (11.26)
3
lR
where we have just integrated by parts. But these expressions make sense even when M ~ , and
~
φm are discontinuous, so we interpret the above equation for discontinuous M ’s as these set
of relations. That is, as distributions.
In the case that we have a continuous medium which ends abruptly in a 2-surface S, then
~
M is only piece-wise continuous and it can be seen that in that case, φm is continuous but
not differentiable at S. So we should use the equation in the distributional sense.
Recalling that the equation
1
∆~x ( ) = −4πδ(~x − ~y ), (11.27)
|~x − ~y |
should in fact be interpreted as,
1
Z
3
(∆u(~x)) d3~x = −4πu(~y ) for all smooth u(~x) of compact support. (11.28)
lR |~x − ~y |
11.1. CONSTANT MAGNETIZATION 167
From (11.26), and using (11.28), we get, for all smooth u(~x) of compact support,
Z Z
3
∆u(~x)φm (~x) d ~x = −4π ~ u(~y )) · M(~
(∇ ~ y ) d3~y
lR3 lR3 ~y
1
Z Z
= (∆∇ ~ (~y )(
~ u(~x)) · M ) d3~xd3~y
3
lR lR 3 ~x |~x − ~y |
~ (~y )( 1 ) d3~xd3~y
Z Z
= ~ (∆u(~x)) · M
∇
3
lR lR 3 ~
x |~x − ~y |
~ ( 1 ) d3~xd3~y
Z Z
= − 3 ~ y) · ∇
∆u(~x)M(~
lR lR 3 ~x |~x − ~y |
~ ( 1 ) d3~xd3~y
Z Z
= ~ y) · ∇
∆u(~x)M(~
3
lR lR 3 ~y |~x − ~y |
~ ( 1 ) d3 ~y ] d3~x
Z Z
= ∆u(~x)[ 3 M ~ (~y ) · ∇ (11.29)
lR 3
lR ~y |~x − ~y |
which is equation equivalent to (11.20). Thus, we see that the correct, or more general,
expression for the potential is,
1
Z
φm (~x) = − ~ (~y ) · ∇
M ~ d3~y (11.30)
lR 3 ~x |~x − ~y |
which is a valid expression, even when M ~ is only piece-wise continuous. Using Gauss theorem
in both sides of a discontinuity surface S we get,
~
~ (~y ) 1 d3~y + n̂ · [M]
Z Z
φm (~x) = ~ ·M
∇ d2 S, (11.31)
3
lR −S ~y |~x − ~y | S |~x − ~y |
where [M ~]≡M ~ in − M
~ out measures the jump of the magnetization across the discontinuity
and n̂ is the outward normal to S (that is from in to out).
Had we chosen to solve the above problem using the vector potential, A,~ such that B ~ =
~ ∧ A,
∇ ~ then,
~ ∧H
∇ ~ =∇
~ ∧ (B
~ − 4π M)
~ = 0, (11.32)
and so,
~=− 4π ~
∆A JM, (11.33)
c
with J~M = c∇
~ ∧M
~ . Thus, if M
~ is smooth,
~ ∧M
∇ ~ (~y )
~y
Z
~ x) =
A(~ d3~y . (11.34)
3 |~x − ~y |
lR
If the magnetization is not smooth, then an argument parallel to the one given above
shows that the correct expression for the vector potential is,
168 CHAPTER 11. MAGNETIC MATERIALS
~ x) =
Z
~ (~y ) ∧ ∇
~ ( 1
A(~ 3
M ~x )d3~y , (11.35)
lR |~x − ~y |
and, in the presence of a surface discontinuity,
Z ~ ∧M
∇ ~ (~y ) Z ~ ] ∧ n̂
[M
~ x) =
A(~ 3
d ~y + d2 S. (11.36)
lR3 −S |~x − ~y | S |~
x − ~y |
~ out − B
(B ~ in )|r=a · n̂ = (H
~ out − H
~ in )|r=a · n̂ + 4π(M
~ out − M
~ in ) · n̂ = 0, (11.37)
that is,
(φout in
m − φm )|r=a = 0. (11.39)
~ = k̂M), and only
Since the solution can only depend on k̂, the magnetization direction, (M
linearly, we must have,
φin in
m = A r k̂ · n̂, (11.40)
k̂ · n̂
φout
m = Aout . (11.41)
r2
Therefore, continuity implies,
inAout
A = 3 , (11.42)
a
and the jump on the normal derivatives,
Aout in
A + 2 3 = 4πM. (11.43)
a
Therefore Ain = 4π
3
M, and Aout = 4π
3
a3 M.
~ = H,
Outside the sphere, B ~ is a magnetic dipole with magnetic moment given by, m
~ =
4π 3 ~
3
a M.
4π 3 ~
~ =
m a M. (11.44)
3
Inside the sphere,
11.2. PERMANENT MAGNETS 169
~ = − 4π M
H ~, (11.45)
3
~ = H
B ~ + 4π M~ = 8π M.
~ (11.46)
3
8π
Bin = B0 + M, (11.47)
3
4π
Hin = B0 − M, (11.48)
3
where we have just used the norms of the vectors for all of them point along the same direction.
To solve this system we need to specify some relation between the three unknowns. In practice
the following to cases are of interest:
Case 1: If B ~ = µH,~ then, since Bin + 2Hin = 3B0 , we have(µ + 2)Hin = 3B0 , and so,
3B0
Hin = , (11.49)
µ+2
3µB0
Bin = , (11.50)
µ+2
3(µ − 1)B0
M = . (11.51)
4π(µ + 2)
We see that for linear relations between the fields both fields, H~ in , and B
~ in vanish when the
external field is not present.
Case 2: If we have a nonlinear relation like the one in the figure bellow –called hysteresis
curve–, Bin = Bin (Hin ), then the solution would be the intersection of that graph and the
above linear relation, Bin + 2Hin = 3B0 .
1
This is a property of the sphere and does not hold for bodies of other shapes.
170 CHAPTER 11. MAGNETIC MATERIALS
B B
B0 = 00
B = B(H) B = µH
H
B = B0 − 2H
We see that there could be one solution, or two solutions, depending on the value of B0 .
In particular for B0 = 0 there are two nontrivial solutions, one the reverse of the other,
corresponding to permanent magnets. In particular, we see that if we start at B0 = 0 in a
unmagnetized material and slowly increase its value we shall be moving along the point which
crosses the hysteresis curve Bin (Hin ) and the linear relation Bin + 2Hin = 3B0 . We see that
when we return to B0 = 0 the material would remain magnetized.
∂ET
F~ = − , (11.53)
∂~x
where ET is the total energy of the system, and not just the electromagnetic energy. Let us
see an example.
Let us assume we have a pair of infinite parallel plates separated a distance L apart, with
potential difference V . We want to compute the force by unit area between them, that is, the
pressure on one of the plates
To do this we assume the displacement is done keeping the plates isolated, that is, their
charges by unit area σ, is constant, and compute the change in its total energy, which in this
case is just the change in its electromagnetic energy
ET E 1 ZL~ ~
= = E · E dx. (11.54)
Area Area 8π 0
~ = k̂E, E =
Now, E V
= 4πσ, therefore,
L
E
= 2πσ 2 L, (11.55)
Area
and
E 1 V2
( )∆L = 2πσ 2 ∆L = ∆L. (11.56)
Area 8π L2
Thus, the pressure, namely the force per unit area is given by
−1 ~ ~ V2
P = E ·E = − 2. (11.57)
8π L
Notice that to compute the change in the total energy we first have to decide how to do
the displacement. We choose to make it keeping the charges constant. Since in that case the
only energy change is in the electrostatic energy, we just did,
∂ET ∂E
= ( )Q . (11.58)
∂~x ∂~x
We could have done the displacement in another way. We could have connected a battery
with a potential difference V to the plates and then make the displacement. In that case
the change in the total energy would have to include the change in the energy stored in the
battery, because in this case there would certainly have been a current across the circuit. So
even if we could compute
∂E
( )V , (11.59)
∂~x
this would not had help us much, for it is different from ET . 2
∂~
x
Nevertheless, notice that, since
2
Notice that since we are computing the force at the given configuration, the result is independent on the
way we choose to make the displacement. It is just a virtual displacement.
172 CHAPTER 11. MAGNETIC MATERIALS
1
E = Qi V i (11.60)
2
1
= Cij V i V j (11.61)
2
1 −1 ij
= (C ) Qi Qj , (11.62)
2
where (C −1 )ij Cjk = δ i j , and summation over repeated indices is assumed. Thus,
∂E 1 ∂Cij i j
( )V = ( )V V (11.63)
∂~x 2 ∂~x
−1 ∂(C −1 )kl
= Cik Clj V i V j (11.64)
2 ∂~x
−1 ∂(C −1 )kl
= Qk Ql (11.65)
2 ∂~x
∂E
= −( )Q , (11.66)
∂~x
where we have used that 0 = ∆δ i k = ∆(C −1 )ij Cjk + (C −1 )ij ∆Cjk , and Qk = Ckj V j .
Thus, we see that if in some circumstances it is simpler to compute ( E )V than to compute
∂~
x
( E )Q , we can go ahead and compute it, and then use the relation above to compute ∂ET using
∂~
x ∂~
x
( ∂E )Q . This is usually formalized by introducing the free energy, F = E − Qi V i = −1 C i j
ij V V .
∂~
x 2
For then ∂F | = E∂xT .
∂x V
We consider now the case of circuits. Here we have a similar situation with currents and
vector potentials (or fluxes) instead of charges and potentials, and inductances instead of
capacities. Thus, the following relation must also hold,
∂E ∂E
(
) ~ = −( ) ~ . (11.67)
∂~x A ∂~x J
But in contrast with electrostatics here the adiabatic displacement, that is the displacement
where no external energy sources are needed is at constant vector potential.
There are basically two ways to reach this conclusion: One is to do the calculation at
constant current, keeping truck of all external electromotive forces needed for this to happens.
If this is done one finds,
∂ET ∂E
) ~ = −( ) ~,
( (11.68)
∂~x J ∂~x J
that is the external energy sources must do twice –and with the opposite sign– the work of
the circuit. Thus,
∂ET ∂ET ∂E ∂E
F =− = −( ) ~ = ( ) ~ = −( ) ~ . (11.69)
∂~x ∂~x J ∂~x J ∂~x A
The other way to reach this conclusion is to replace the original circuit by a superconduct-
ing one. In order for this not to change significantly the configuration we imagine replacing
11.4. THE ENERGY OF MAGNETIC MATERIALS 173
the original circuit for a bunch of very thin superconducting wires, thus modeling more and
more precisely, in the limit when the section of the wires goes to zero, the original current
distribution. But for superconducting circuits the magnetic flux is constant under displace-
ments, and so is the vector potential at the surface. Therefore, this is the natural quantity to
keep constant in the adiabatic variation.
1Z
F ≡ E− J~ · A
~ d3~x (11.73)
c lR3
= −E, (11.74)
and then,
∂F δE ∂ A ~ ∂E 1 ~
~ · ∂ A d3~x
Z
( ) = ( ) ( ) + ( ) − J (11.75)
∂s J~ ~
s
~
δA ∂s ∂s A c lR3 ∂s
∂E
= ( ) ~. (11.76)
∂s A
Thus, for calculations it is equivalent to know F as a function of J~, than to know E as a
~ But,
function of A.
~ ~
δF ~ δE δ A ~ −1 ~ · δ A · δ J~ d3~x − 1
Z Z
(δ J) = ( · δ J) J ~ · δ J~ d3~x
A (11.77)
~
δJ ~ ~
δ A Zδ J c lR 3
δJ~ c lR 3
1 ~ · δ J~ d ~x.
3
= − A (11.78)
c lR3
174 CHAPTER 11. MAGNETIC MATERIALS
δF ~ 1
Z
(δ J) = − ~ ·∇
A ~ ∧ δH
~ d3~x. (11.79)
δ J~ 4π lR3
1
Z
= − ~ · (∇
{δ H ~ ∧ A)
~ +∇
~ · (δ H
~ ∧ A)}
~ d3~x (11.80)
4π 3
lR
1
Z
= − ~ · δH
B ~ d3~x. (11.81)
4π 3
lR
~ = µH,
When B ~ then
1
Z
F =− ~ ·H
B ~ d3~x. (11.82)
8π lR3
~ = B(
In the case that B ~ H),
~
1 Z I ~ ~ ~ d3~x,
∆F = − H(B) · δ B (11.83)
4π lR3
where the integral is along a hysteresis cycle, that is, it is just the area inside the hysteresis
curve.
Example: The difference between the free energies of a set of circuits in vacuum
and the same set in the presence of a magnetizable medium.
−1
Z
∆δF = [B ~ − ~h · δ~h] d3~x,
~ · δH (11.84)
4π lR3
where ~h is the field generated by the circuit, given by certain J(~ ~ x), as if it were in vacuum,
i.e. µ = 1.
The above expression can be rewritten as,
−1 Z ~ − ~h) · δ~h + B ~ − δ~h) + (B
~ · (δ H ~ · δ~h] d3~x.
~ − H)
∆δF = [(H (11.85)
4π lR3
The first term is
Z Z
~ − ~h) · δ~h d3~x =
(H ~ − ~h) · (∇
(H ~ ∧ δ A)
~ d3~x (11.86)
3 3
lR ZlR
= ~ · (δ A
[∇ ~ − ~h)) + δ A
~ ∧ (H ~·∇ ~ − ~h)] d3~x (11.87)
~ ∧ (H
lR3
= 0, (11.88)
11.4. THE ENERGY OF MAGNETIC MATERIALS 175
~ H
for the first term vanishes upon application of Gauss theorem and the second because ∇∧ ~ =
~ ∧ ~h = 4π J.
∇ ~
c
The second term also vanishes, as can be seen using similar arguments upon substitution
~
of B by ∇ ~ ∧ A.~ Thus,
1
Z
∆δF = ~ − H)
(B ~ · δ~h d3~x (11.89)
4πZ lR 3
−1
Z
∆F = M~ · ~h d3~x (11.91)
2 lR 3
−1
≈ ~ · ~h,
m (11.92)
2
where
Z
m
~ ≡ ~ d3~x,
M (11.93)
lR3
and we have assumed that near the medium ~h ≈ ~h0 , a constant field.
If the magnetization does not depend on the external field, then
~ · ~h.
∆F ≈ −m (11.94)
176 CHAPTER 11. MAGNETIC MATERIALS
Chapter 12
Examination Questions 1
Problem 1 State the Cauchy problem (1.1) for electromagnetism and explain it.
Problem 2 Prove uniqueness of solutions in the Cauchy problem for electromagnetism (1.1).
a)
∂
φ(t, ~x) = tMt (φ1 (~x)) + (tMt (φ0 (~x))),
∂t
satisfies also the wave equation.
∂φ
b) Furthermore, it has as initial conditions: φ1 (~x) = ∂t
(t, ~x) |t=0 and φ0 (~x) = φ(0, ~x).
Problem 6 Show that if the constraint equations are satisfied at t = 0 then they are satisfied
for all times provided the fields satisfy the evolution equations.
177
178 CHAPTER 12. EXAMINATION QUESTIONS 1
Problem 7 Given two smooth vector fields in lR3 , ~l(~x) y m(~x) produce everywhere smooth
~ and B.
solutions to the vacuum constraint equations for E ~
Problem 8 Find the expression for the energy of the electromagnetic field starting from the
expression of the power in terms of the work done by the Lorentz force.
Problem 10 You are left alone in a region without any house or other reference, except a
power line. To reach a city you decide to follow the line in the assumption that it is feeding
power to it. How you determine the direction you should go without cutting the power line?
Problem 11 Show that Maxwell’s equations are invariant under time and space translations.
Problem 12 Show that Maxwell’s equations are not invariant under Galilean transforma-
tions.
Problem 18 Deduce the existence of the Capacities matrix, eqn. 4.13. Given three conduc-
tors, how would you measure the component C23 of the capacity matrix? In the experiment
you can set up any given potential distribution among the conductors (only once) and can
measure the induced charge at only one of the conductors of your choice.
Problem 20 Compute P3 and P4 using the recursion relation found and the normalization
condition
Problem 21 Deduce the expression for the multipole constants, eqn. 5.11
Problem 22 Find two different charge distributions with no symmetry at all but giving the
same external field (which can have symmetries). Hint: do not construct them explicitly but
just start from their potentials and work backwards.
179
Problem 23 Deduce the expression for the Dirichlet Green’s function corresponding to two
concentric conducting spheres.
Problem 24 Deduce the symmetry of the Dirichlet Green function, equation 7.1
Problem 25 It is required to solve the Poisson equation inside a volume V where there is
defined a smooth source function ρ(~x), and a Neumann boundary condition ∂n φ(~x)|∂V = g.
What condition must g satisfy in order for a solution to exist?
Problem 29 Deduce the boundary conditions at the surface of a superconductor for the mag-
netic field.
Problem 30 Find the vector potential corresponding to a constant magnetic field, show that
the dependence on a given origin is pure gauge.
Problem 31 Deduce the boundary conditions at the surface of a superconductor for the scalar
magnetic potential. Explain the multi-valuate nature of it, and give the expression for the value
of the needed discontinuity when a non simply connected superconductor is present.
Problem 32 Deduce that ∇ · J~ = 0 (in the sense of distributions) for a wire current distri-
bution.
Problem 34 Prove that the definition of magnetic flux using the vector potential eqn. (9.23)
is gauge independent.
Problem 35 Show that the flux across a surface bounded by a superconductor is constant in
time.
Problem 37 Deduce that the first term in the multipolar expansion of the magnetic potential
vanishes, and find the following two terms.
Problem 38 Find the expression for the energy of the magnetostatic field in terms of the
inductance matrix.
Problem 40 Deduce, using distributions, the true formula for the magnetic potential when
the magnetization is not differentiable.
180 CHAPTER 12. EXAMINATION QUESTIONS 1
Part II
Second Course
181
Chapter 13
13.1 Introduction
In chapter 3 we saw Maxwell’s equations have a number of symmetries, that is, transformations
that take one solution into another. It was clear that most of them arise because of the
underlying symmetries of space and time. Indeed, time translation, or the homogeneity of
time implied that if (E(~~ x, t), B(~
~ x, t)) was a solution, then (E
~ T (~x, t), B
~ T (~x, t)) = (E(~
~ x, t +
~
T ), B(~x, t + T )) was also a solution (with the corresponding time translated sources, if any
present). The same for space translations or rotations.
But we also saw that the assumed Galilean symmetry of space-time was not a symmetry
of these equations. Galilean symmetry can be stated in different ways, the ampler is the
statement that by means of local experiments one can not determine the state motion of our
system with respect to others objects outside it, namely there is no notion of absolute speed.
This conception, very obvious today, was sharpened by Galileo as a justification why we do not
feel earth motion in its orbit around the sun. The previous conception of space and time was
due to Aristotle, which though there was a natural motion state for earthly bodies, namely
to be at rest, meaning with respect to earth. 1 Since then this conception has been central
for our understanding of physics. But here we are presented with equations which describe
with incredible precision the electromagnetic phenomena and yet have on it a parameter with
dimension of velocity and its solutions propagate which such a speed. So the natural way
of interpreting this was to throw away Galileo’s conception and say that after all there was
an ether, namely, something with respect to which these solutions were moving. From a
mechanistic conception, still strong at that time, electromagnetic waves where mechanical
waves of the ether, that is local modifications of the state of such a material which would
propagate along it with the "sound speed" characteristic of it. A very stiff material indeed! 2
1
Galileo’s reasoning against this beliefs, thinking on situations where friction forces where smaller and
smaller imagining a limit where they were absent and so that in this idealized situation the bodies would
continue to move forever, marks the beginning of one of the pillars
q of the scientific thinking.
2 Y
Recall that the sound speed of a material is given by Cs = ρ where Y is Young’s module and measures
the stiffness of a material. For steel Cs ≈ 6 106 m
s .
183
184 CHAPTER 13. THE SYMMETRIES OF MAXWELL’S EQUATIONS CONTINUED
Here, instead, we shall investigate whether there is a symmetry after all reflecting Galileo’s
principle of relativity of motion more general than the one already tried in chapter 3.
To that end, and to facilitate the calculations, we shall consider the symmetries of the
wave equation,
∂2 ∂2 ∂2
✷φ := ( ∂t2 − c2∆)φ := ( ∂t2 − c2δij ∂xi ∂xj )φ = 0
where the last expression is valid in Cartesian coordinates. We know that in these coordinates
each component E ~ or B~ satisfies the wave equation, so we know that a symmetry of Maxwell’s
equation necessarily must be a symmetry of the wave equation. On the other hand in this
simpler equation we already have the fact that solutions propagate at a given speed.
So we want to find a transformation among solutions which represent a given solution
but as seen in constant motion from the original coordinate system. So we want to look for
a coordinate system transformation which depend only on a constant vector ~v, namely the
relative speed between the solutions.
The most general linear transformation we can write depending on just a vector is:
˜v ∧ ~x
~x′ = a~x + b~v t + d(~v · ~x)~v + d~ (13.1)
t′ = et + f (~v · ~x) (13.2)
Redefining a the last term in the first transformation can be seen to be a rotation, and we
have already seen that rotations are symmetries by themselves, so we can safely remove that
term. We also know that changing the value of a amounts to a redefinition of the space scale,
and that would give rise to a trivial symmetry (assuming c or t are also scaled, and so also
the equation sources, if there were). So we can set a = 1 without lost of generality. Choosing
to keep the value of c as fixed, we see we can not re-scale t. Furthermore, if we want this
transformation to really represent a motion with speed ~v , then it should be the case that a
trajectory of the form ~x(t) = ~xo −~v t should transform into a stationary trajectory, namely we
should have dtd ~x′ (t) = 0. Plugging these two conditions on the above transformation (where
we have already set a = 1, d˜ = 0) we find,
d ′ d d
~x = ~x + b~v + d(~v · ~x)~v (13.3)
dt dt dt
2
= −(1 + dv )~v + b~v (13.4)
b = 1 + dv 2 . (13.5)
So now we assume we have a solution to the wave equation, φ(~x, t) and want to see
whether φ~v (~x, t) := φ(~x′ (~x, t), t′ (~x, t)) is also a solution. This will impose conditions on the
remaining free coefficients and determine them completely. Notice that these coefficients √ can
only depend on the two scalar parameters of the problem, namely c, and v := ~v · ~v .
13.1. INTRODUCTION 185
∂ ∂t′ ∂ ∂x′i ∂
φ~v = φ + φ
∂t ∂t ∂t′ ∂t ∂x′i
∂ ∂t′ ∂ ∂x′l ∂
φ = φ + φ
∂xi ~v ∂xi ∂t′ ∂xi ∂x′l
and for the second derivatives, using the fact that the Jacobian of the transformation is
constant,
∂x′l
= δ l i + dv l vi
∂xi
∂x′i
= bv i
∂t
∂t′
= e
∂t
∂t′
= f vj
∂xj
where vi := δij v j , being δij the inverse of δ ij , δij δ jk = δi k , ~x · ~v = δij xi v j , etc.
Therefore, we have,
∂2 2 2 2 2 ∂2
φ
✷ ~v = ′2
φ[e − f c v ] + ′ ′i
φ[2ebv i − 2f c2 (1 + dv 2 )v i]
∂t ∂t ∂x
∂2
+ φ[−c2 δ lm − v l v m (2adc2 + d2 v 2 c2 − b2 )] (13.6)
∂x′l ∂x′m
2 2
Using that φ satisfies the wave equation and so, −c2 δ lm ∂x′l∂∂x′m φ = − ∂t∂ ′2 φ we get,
∂2 2 2 2 2 ∂2
φ
✷ ~v = φ[e − f c v − 1] + φ[2eb − 2f c2 (1 + dv 2)]v i
∂t′2 ∂t′ ∂x′i
∂2
+ φ[v l v m (−(2 + dv 2 )dc2 + b2 )] (13.7)
∂x′l ∂x′m
So, if we want that φ~v satisfies the wave equation we need that every term on the right-hand
side must vanish. At any given point in space and any time we can find solutions to the
wave equation for which any of the second partial derivatives are zero except for one of them.
Thus, every term in brackets must vanish.
186 CHAPTER 13. THE SYMMETRIES OF MAXWELL’S EQUATIONS CONTINUED
Exercise: Check this! Hint: choose Cartesian coordinates (x, y, z) such that ~v = (1, 0, 0)
and try with the following solutions to the wave equation: tx, c2 t2 − y 2 , c2 t2 − x2 .
Using that b = (1 + dv 2) in the third term we get a second order polynomial in d. From
which we get that,
s
±γ − 1 v2
d= γ := 1−
v2 c2
and correspondingly,
b = ±γ.
We take the plus sign, for this is the value that gives the Galilean transformation in the limit
of small ~v . Since b ≥ 1 we can factorize it in the second term and get,
e
f= .
c2
Substitution of this relation on the first term gives,
1
e2 = 2 or e = ±γ.
1 − vc2
We shall keep the plus sign, the other corresponds to a time inversion.
So, we have found all coefficients and the final transformation is:
~x′ = ~x + γ~vt + d(~v · ~x)~v = γ(n̂ · ~x)n̂ + γ~v t + (~x − (n̂ · ~x)n̂) (13.8)
(~v · ~x)
t′ = γ(t + ) (13.9)
c2
where n̂ := ~v .
|~
v|
Finally, then we have found a symmetry transformation for the wave equation that re-
sembles a Galilean transformation as close as possible. Together with the transformations
already found en chapter 3 they are all symmetry transformations of the wave equation (of
course one can compose many of them one after the other and obtain more transformations,
but they are all generated, in this way, from the ones already found).
Let us analyze it in detail. Choose coordinate axis so that the velocity is in the ~e1 direction,
~v = (−v, 0, 0). Then the symmetry transformation becomes,
x′ = γ(x − vt)
y′ = y
z′ = z
vx
t′ = γ(t − 2 )
c
13.1. INTRODUCTION 187
ct
ct′
x′
ct ct′
x′
v
In the following figure we see what such a transformation does for a value of c
= 54 , that
is, γ = √ 1 16 = 35
1− 25
Exercise: The above plot uses the value c = 1, or alternatively it is plotted using as variable
ct instead of t. Make a similar plot but now with take c = 0.1.
We can see from the example that points along the line x = ct are transformed among
each other. Let us see this in more detail.
vx 2
−c2 (t′ )2 + (x′ )2 + (y ′ )2 + (dz ′ )2 = −c2 (γ(t + 2
)) + (γ(x + vt))2 + y 2 + z 2
c
= t2 (γ 2 (−c2 + v 2 )) + x2 (γ 2 (−v 2 c−2 + 1) + y 2 + z 2
= −c2 t2 + x2 + y 2 + z 2 (13.10)
Thus we see that not only the lines x = ct are transformed among each other, but also the
lines x = −ct and all the hyperbolas of the form −c2 t2 + x2 + y 2 + z 2 = const..
Exercise: Draw in a diagram these hyperbolas for positive and negative constant values.
Chapter 14
Special Relativity
14.1 Introduction
In the previous chapter we found a new symmetry for the wave equation, which can be
extended also to Maxwell’s equations. This symmetry is a transformation on space and time
which leaves invariant the lines which move at the speed of light. To understand what all this
means we have to change our concepts of space and time.
The building block of our description are happenings in phenomena, they will be called
"events", that is an idealization of something happening in a limit where the duration of it
goes to zero, as well as the size of the space where this is taken place. We imagine we have
a big bag with all events which have occurred in history, not only the important events, but
also all events where nothing important happened, or just where nothing happened at all! We
can also imagine having all events that will happens in the future. That will be the sand-box
where we want to describe phenomena, we shall refer to it as space-time. If we think on
our usual description of events, we see that we can describe them using four numbers, for
instance if I have an appointment at the doctor, I need the time, the street name (which I
can code with some number), the building number and the floor. In general, besides the time,
one needs to define a position relative to some other objects, and for that it suffices with
three other numbers. This way we can mark events with four numbers, and so we have some
notion that they form a continuum in the sense that given two events with their respective
numbers we can find some other whose numbers are all equidistant to those of both events.
That way we can describe, say the events of a given molecule, by a continuous line, centered
at the events occupied by its center of mass. That assumption might ultimately be wrong as
we probe further and further smaller distances, but it suffices for the present description of
the world. This is all what we meant when we say that we live in four dimensions.
We can put some order on these events and draw them on a four dimensional space.
Representing the label time going in an upright direction, so that events at the same time
are represented by horizontal hyper-surfaces, which in our drawings have one dimension sup-
pressed.
We will now describe some phenomena with these tools, so as to get acquainted with the
concepts.
In figure 14.1 we describe the encounter of two persons in space-time by marking the events
189
190 CHAPTER 14. SPECIAL RELATIVITY
hart-beat of each one of them. They are not proper events for hart-beats are not localizable
in time or space in arbitrarily small regions, but they suffice to describe the encounter pretty
well. Of course, we can imagine that we look at more and more localized events, like the
firing of a nerve cell in the hart, some given molecule passing a cell membrane, and so on
(disregarding the quantum world for this discussion) and so as to describe a continuum of
events. In that limit each of our persons is geometrically characterized in our space-time as
a line. Each line is called the world–line of the person, the encounter is the point where
these lines intersect. The important thing to realize here is that each person, or molecule
inside the person is a one-dimensional object! The event “encounter” is a point, that is zero-
dimensional, but each person, in our rough description, is a line. In figure 14.2 we show this
idealized situation.
Consider now a hair belonging to one of the persons in the previous encounter. If we
imagine events taking place along the hair, say in its cells, then they would describe in our
space-time a two-surface! We see then that in this way of describing phenomena a hair is no
longer though as a “line”, that is a one dimensional object, but rather as a two-dimensional
one. If one develops the intuition to picture this situation, one is half-way in understanding
relativity. Here is a figure 14.3 showing our hair in space-time.
t
a hair
Having introduced the concepts of event and of space-time we can now interpret on it our
usual concepts of space and time.
an absolute time, namely a label for these hyper-surfaces, namely a function from space-time
into the reals. The real value this function has on each of these hyper-surfaces is not very
important for now, the important thing is that it distinguishes each one of these hyper-surfaces
among them. Thus, we have the following picture of this view, see figure 14.5. Notice that
since the time changes for the objects at rest, their world-lines transverse the simultaneity
hyper-surfaces and through these world-lines we can identify points at different hyper-surfaces
as occupying the “same space along time”. The identification of all these hyper-surfaces with
each other along the preferred world-lines is what we call “space”.
t
t=3
t=2
t=1
t=0
We now describe in mathematical terms the structure this view assumes. We have al-
ready introduced a time function and its level surfaces, the simultaneity hyper-surfaces. The
differential of this function, dt, allows to say which vectors are tangent to the simultaneity
hyper-surfaces. If we take coordinates describing points of space-time, xµ , µ = 0, .., 3. Then
∂t
a vector a of components aµ is tangent to the simultaneity surfaces if a(dt) := aµ ∂x µ = 0.
This follows from the fact that the previous expression is the derivative of the function t on
the direction of a. And this derivative vanishes if and only if this vector is tangent to the
level surfaces of t. Then we have the preferred world-lines of objects at rest, their tangent
vectors, (four-dimensional vectors) provide a vector field over the whole space-time. To obtain
them we use as parametrization of these lines the values of the time function as they cross the
simultaneity surfaces. Then the preferred world lines are given by xµ (t) where t is the value of
µ ∂xµ (t) ∂t
the function t. Thus, we have a vector tµ = ∂xdt(t) , and tµ (dt)µ := tµ ∂x
∂t
µ = dt ∂xµ
dt
= dt = 1.
There is an extra mathematical structure present in this view, namely a notion of distance
between events. This notion is very specific, it says that the distance between two events (in
the same simultaneity hyper-surface) is given by the square root of the sum of the square
of the components of the vector connecting the events when expressed in some preferred co-
ordinate systems. These are the so-called Cartesian coordinate systems of space. One way
of encoding this information without referring to particular coordinate systems is to use a
metric tensor, namely an object which takes two vectors, and provides a number, called
14.1. INTRODUCTION 193
t
t=3
t=2
t=1
t=0
space
the square of its norm. This is a bi-function on a vector space V , (an entry for each vector),
h(·, ·) which is linear on each entry, namely,
and symmetric, h(v, w) = h(w, v) ∀u, v ∈ V . In Cartesian coordinates this tensor has
components, hij = δij , the Kronecker delta. 1 The object just described lives in our “space”,
now how do we extend it to the space-time? One procedure is the following, we define the
Cartesian coordinates in one of the simultaneity hyper-surfaces, that is, at a given time. And
then transport it in time requiring that the points along the preferred world-lines have the
same coordinate values for all times. We then extend the metric to all space-time by requiring
that its component in the Cartesian coordinates remain constant for all times. This is not the
whole prescription for symmetric two-tensors in space-time have more components, 10 instead
of 6. We define the other components completing the four dimensional coordinate system with
the time function as the zero-index coordinate, x0 = t, and defining h00 = 0, h0i = 0, i =
1 . . . 3. Notice that in this way, the distance between two events which are occurring at the
same space point but at different time can be measured, and it has the value zero. While the
distance between two events at different space points have a given, positive distance among
them. The “time” between two events is just the difference between the times of each of their
simultaneity hyper-surfaces, so for Aristotle it makes perfect sense to measure time intervals
and distances between arbitrary intervals. Alternatively given a vector in space-time we can
decompose it in a unique way into a vector along t and another perpendicular to it,
1
The existence of Cartesian coordinates, in particular global ones, is a very nontrivial topic, and implies a
certain structure for the space-time. Since this topic is too complex to be dealt with here, we just postulate
its existence.
194 CHAPTER 14. SPECIAL RELATIVITY
X = [X(dt)] t + X̃
since
t t̃
t=3
t=2
t=1
t=0
This completes the description of the mathematical structure of the Aristotelian view of
space-time. We have the set of all events which we call space-time. In space-time, we have
a set of preferred world-lines, representing the events taking part at objects at rest. We also
have a time function t from this set to the reals, and we call simultaneity hyper-surfaces at the
level set of it. With this function we can define the tangent vector to each preferred world-line
and so have a vector field t. We can identify these hyper-surfaces with each other through
the preferred world-lines thus having a notion of “space”. In “space” we have a notion of
Cartesian coordinates and distance among objects. Which we uplift to space-time using the
present structure, giving a (degenerate) metric tensor h in space-time. Time is parametrized
by a clock carried by any of the observer at rest. Note that time and space translations, and
space rotations leave invariant all this structure.
the ball will roll indefinitely, showing that the concept of “natural rest” is not fundamental, it
exists in our mind due to our inability to experience the limit of frictionless motion. Once he
realized this, he proposed the relativity principle which with our present concepts we can re-
state as saying that there are no preferred world-lines and so no natural identification among
simultaneity hyper-surfaces, and so no “space” in the way we have introduced it before.
If no preferred world-lines exist what does remains from the Aristotelian view? We still
have simultaneity hyper-surfaces and a time function, but no preferred vector t. What about
the notion of distance? Assume we have a given Aristotelian frame, that is a vector t, then
we are in the Aristotelian framework and can assume there is a space with its 3-metric h
defined. We also have a coordinate system xµ obtained by pulling up from one simultaneity
hyper-surface a Cartesian system and declaring the other coordinate to be the label of the
hyper-surfaces. Then in this coordinate system we can also lift the metric tensor hij = δij ,
other components vanishing. Now imagine another family of world-lines moving at a velocity
v i = (v, 0, 0) with respect to the previous one, that is, we now have coordinates which relate
to the previous one as,
t′ = t
x′ = x + vt
y′ = y
z′ = z
Then the components of the metric tensor above introduced in this coordinates would
∂xσ ∂xρ 2
be, (using that components of a tensor transform like hµ′ ν ′ = ∂x µ′ ∂xν ′ hσρ ), h0 0 = v hxx ,
′ ′
h0′ i′ = −2vhxi , hi′ j ′ = hij . Thus, we see that the possible lifts will generate different 4-
tensors, so there is no preferred one. The only intrinsic tensor is the one acting on tangent
vectors to the simultaneity hyper-surfaces. The reason is simple: given any two points at
different simultaneity hyper-surfaces and an Aristotelian frame we can project one of them
into the simultaneity hyper-surface of the other one by following the rest lines of that frame,
and then measure the distance among them at that hyper-surface. But this prescription for
measuring distances depends on the Aristotelian frame chosen, in fact, taking v = ∆x AB
∆tAB
this
number can be made to vanish! Thus, we see that in the Galilean framework we can only
measure distances among events when they are at the same simultaneity hyper-surface, still
we can measure the time between events as before. This does not mean that we are restricted
on what we can describe in nature, on the contrary this framework tells us what we should
expect from nature, namely that natural phenomena can not depend on “distances” among
events at different times, but only on distances between events at the same simultaneity
hyper-surface. Thus, for instance, the force between two bodies in the Newtonian theory of
gravitation can not depend on their distances other that when measured at the same time.
This framework order our ideas about how nature behaves.
Exercise: Check that given any two non-simultaneous events there is a world-line for which
the distance between events can be made to vanish. Check that nevertheless it can not be
made to be negative.
196 CHAPTER 14. SPECIAL RELATIVITY
The Galilean view is democratic, in the sense that any two sets of world-lines, and so
any two sets of Aristotelian frameworks are equivalent. Physics can not depend on them,
and one way to make sure that one is considering physically relevant concepts is to check
such equivalence. That is why in classical mechanics one requires the theories to not only to
be invariant under space and time translations and space rotations, but also under arbitrary
Galilean transformations.
We are ready to define now our subject, special relativity, which is just the following
assumptions.
198 CHAPTER 14. SPECIAL RELATIVITY
Figure 14.7: Different causal relations among events: events A and B are space-related,
events A and B are light-related, events A and D future-time-related, and event A and E
past-time-related.
1. The set of events is four dimensional, that is it suffices four numbers to label all
events and describe their interactions. More specifically at each neighborhood Up of
an event p there exist four functions (x0 , x1 , x2 , x3 ) form Up into the reals (a coor-
dinate system), such that any function, f from Up into the reals can be written as
˜ 0 (q), x1 (q), x2 (q), x3 (q)) ∀q ∈ Up .
f (q) = f((x
2. At each point of this space of events there is a metric tensor, namely an invertible,
symmetric, by-tensor of signature (+, −, −, −).
3. There are global coordinate systems in which the metric tensor is constant and is diag-
onal with components, (1, −1, −1, −1). We call them Cartesian systems.
This is all what we need to have the arena where physics is played. This space-time is called
the Minkowski space-time . The first two assumptions are though to be very basic. But
they are just assumptions and so are constantly contested in modern physics, in particular the
notion of a four dimensional space-time is abandoned in string theory, although these theories
have not yet succeeded in describing more physics than before. The third assumption is clearly
false because implies the space-time geometry does not have any curvature, and we know that
matter curves space-time, for instance it is known form Eddington expedition to the island of
PrÃncipe near Africa to watch the solar eclipse of 29 May 1919 that sun’s gravity pull bends
light rays. Thus, they do not travel along the straight lines a global coordinate system as the
one above would imply. Nevertheless, this assumption is a good first step in the sense that
with it we can model many physical situations (where gravity is not important) with very
high accuracy and with simple mathematical tools. In particular for this book this is enough.
14.2. THE GEOMETRY OF MINKOWSKI SPACE-TIME. 199
In the next sections we shall develop the geometry corresponding to this new concepts, in
particular the kinematics and dynamics of particles and other fields.
Clearly if we multiply any vector by any real number the vector remains in the same set,
its nature does not change. The zero vector is the only vector in the intersection of these sets.
Exercise: Show that the time-like vectors form a double cone. That is, if v 1 and v 2 are
time-like then either v 1 + αv 2 is time-like ∀α ∈ [0, ∞) or v 1 − αv 2 is time-like ∀α ∈ [0, ∞).
Conclude that the non-zero time-like vectors are in either of two cones. One of them is called
the future cone, the other the past one. Reach the same conclusion for the non-space-like
vectors.
Exercise: Show that the sum of two null vectors is never a null vector unless they are
proportional to each other.
Exercise: Show that the sum of two future directed null vectors is a future-directed time-like
vector.
e = −at + k
(14.1)
e = at − l
(14.2)
for a > 0, and l, and k two future directed vectors. We immediately see that,
1
e = (k − l)
2
(14.3)
1
t = (k + l).
2a
(14.4)
14.2. THE GEOMETRY OF MINKOWSKI SPACE-TIME. 201
Therefore,
−1
η(e, e) = η(k, l) ≤ 0
2
(14.5)
1
η(t, e) = η(k + l, k − l) = 0.
4a
(14.6)
Thus, e is space-like (the scalar product of two future directed null vectors is positive or zero,
being zero only when the vectors are zero or proportional to each other, this last condition
can not hold since t was taken to be time-like), and it is perpendicular to t. Thus, the
simultaneity hyper-surfaces are space-like and perpendicular to the observer world line.
Therefore the space of vectors perpendicular to a time-like vector t is a three-dimensional
space with a norm, indeed −η(x, x) > 0 if x 6= 0 for all the above vectors. When appropriate,
for vectors at a simultaneity space we shall write,
−η(x, y) = ~x · ~y ,
dxi i ũi
:= β :=
dx0 η(t, u)
Note that we use β i to express this velocity, the reason is that we are using as time function
x0 = ct so the velocity is dimensionless, in terms of t the velocity is v i = cβ i . Note also that
this 3-velocity does not depend on the parametrization of the curve, and therefore not in the
magnitude of the 4-velocity vector.
Thus, it is better to express the above splitting in the following way,
u = η(t, u)(t + β)
Contracting u with itself we get,
Therefore,
q
u= η(u, u)γ(t + β)
where,
η(t, u) 1
γ=q =q ≥ 1.
η(u, u) ~
1−β·β~
t̂ = γ(t + β),
14.2. THE GEOMETRY OF MINKOWSKI SPACE-TIME. 203
t̃ l̃ := t̃ − ẽ1
l := t − e1
t
ẽ1
e1
k̃ := t̃ + ẽ1
k := t + e1
where β is the relative velocity of t̂ with respect to t, and so is a vector in the space of the
simultaneity surface perpendicular to t. Likewise, we can write, t in terms of t̂,
t = γ̂(t̂ + β̂).
We want to find now the relation among these quantities. Since both vectors are unit, we
have,
β̂ = t/γ − t̂
= t/γ − γt − γβ
1
= −γ([1 − 2 ]t + β)
γ
2
= −γ(β t + β),
so β̂ is a vector with a part sticking out of the hyper-surface perpendicular to t, and indeed
it is perpendicular to t̂. To see the components of this vector in that hyper-surface we take a
frame (and its associated coordinate system) {êi }, i = 0..3, with ê0 = t̂, and ê1 in the plane
spanned by t̂ and t. The orthonormality condition determines it to be,
ê1 = γ(βt + e1 ),
where e1 is the frame component perpendicular to t and in the plane spanned by t̂ and t,
(e1 = β/β, or β = βe1 ). Thus, we see that
β̂ = −β ê1 ,
so indeed, in components the relative velocities are identical except that the directions are
opposite. This could have been guessed from the fact that,
Relative Velocities
Frequently one is given the world lines of two bodies in some arbitrary reference frame, t, and
would like to find their relative velocities, namely the velocity of one as seen from the other.
That is, we are given their velocities,
14.2. THE GEOMETRY OF MINKOWSKI SPACE-TIME. 205
u′ = γuu′ (u + β̂),
with η(β̂, u) = 0. The normalization factor is easy,
We also have,
β̂ = u′ /γuu′ − u,
form which it follows,
β̂ 2 := −η(β̂, β̂)
1 2η(u′ , u)
= −[ 2 + 1 − ]
γuu′ γuu′
1
= 1− 2
γuu′
(1 − β 2 )(1 − β ′2 )
= 1−
1 + η(β, β ′ )
−η(β − β ′ , β − β ′ ) + β 2 β 2 sin(θ)2
=
(1 − ββ ′ cos(θ))2
β̂ 2 ≈ −η(β − β ′ , β − β ′ ),
which corresponds to the usual addition of velocities en Galilean kinematics.
We would like now to find the components of β̂ in the frame associated to u. If {ei },
i = 0..3 is the frame associated to t, and {êi } is the frame associated to u, (e0 = t, and
ê0 = u), such that both e1 and ê1 are in the plane spanned by these two vectors, and
ei = êi , i = 2, 3, we have,
u′ = γu′ (e0 + β ′i ei ) = γu′ [γu (ê0 − β ê1 ) + γu β ′1 (ê1 − β ê0 ) + β ′2 ê2 + β ′2 ê2 ],
206 CHAPTER 14. SPECIAL RELATIVITY
and,
−β + β ′1
β̂ 1 := −η(β̂, ê1 ) = −η(u′ /γuu′ − u, ê1 ) = ,
1 + η(β, β ′ )
(14.8)
′i
β
β̂ i := −η(β̂, êi ) = i = 2, 3.
γu (1 + η(β, β ′ ))
(14.9)
xµ
1
γ0
γ1
Lemma 14.1 The maximal interval corresponds to the straight line connecting both events.
The value of this interval corresponds to the proper time of an observer moving along such
curve.
Proof: Assume xµ0 (s) is the curve which maximizes the above quantity, and consider nearby
curves of the type xµλ (s) = xµ0 (s) + λxµ1 (s). Since they all must leave from A and reach B we
assume xµ1 (0) = xµ1 (1) = 0. Using them we construct a function,
s
dxµλ dxνλ
1
Z
TAB (λ) := dτ.
ηµν
0 dτ dτ
taking the derivative with respect to λ and setting it to zero when λ = 0, (the condition for
dAB (λ = 0) to be an extreme. We get,
dxµ
λ d dxνλ
dTAB (λ)
Z 1 ηµν ( )
|λ=0 = r ds dλ ds ds|λ=0
dλ 0 dxµ dxν
ηµν dsλ dsλ
dxµ ν
0 dx1
Z 1 ηµνds ds
= ds.
dxµ dxν
q
0 ηµν ds0 ds0
choosing, for simplicity, the parametrization so that the denominator is constant, and inte-
grating by parts we get,
where in the last step we have used the initial and final conditions for both curves to coincide
at A and B. Since xν1 (s) is arbitrary, and ηµν invertible we reach the conclusion that along
the maximum the second derivatives of the curve, that is the acceleration, must be zero. So
the curve has constant velocity vector, and therefore it is a straight line.
Exercise: Show with an example that if ηµν were not invertible we would not be able to
reach the above conclusion.
We shall say that in absence of forces particles travel along straight lines of this coordinate
systems. Not only that but with the parametrization used those lines can be described as
linear functions of the parameters, in particular their internal clocks mark the time interval
between the events they occur. Thus, imagine two persons which meet at event A at which
they synchronized their watches, then each one of them went on different paths until meeting
again at event B. In general their watches will mark different times, the shorter time would
correspond to the person which accelerated the most during its path. So if you move around,
you stay younger!
Exercise: Show that the time difference between person going through the longest past and
a person going along a nearby path is given by
Z s
∆TAB = η(δu, δu) ds,
0
where δu := λ dx
ds
1
is the variation of the velocity in the direction perpendicular to it. Notice
that if we parametrize the curve with the proper time, then, since η(u, u) = 1, η(u, δu) = 0
and so, η(δu, δu) is negative definite unless δu = 0. Hint, perform a second variation of the
above formula, and use Taylor’s formula to second order to approximate TAB in terms of λ.
each one of them. In the next figure 14.12 we show the situation, we have our chalk moving
with velocity u from another frame, and show the two different representations by the vectors
x, and x̃.
t
u
chalk
x̃
From the figure we see that there must exist α such that
x̃ = x + αu.
Since u = γ(t + v), η(t, x) = 0, and η(u, x̃) = 0, contracting the above expression with u
we obtain α, and so,
x = η(u, x) u + x̃ = γη(β, x) u + x̃
therefore,
η(x, x) = η(x̃, x̃) + γ 2 η(β, x)2 η(u, u) = η(x̃, x̃) + γ 2 η(β, x)2
and so,
x̃ · x̃ = −η(x̃, x̃) = −η(x, x) + γ 2 η(β, x)2 = x · x + γ 2 (β · x)2 .
we see that the largest distance the object has is precisely when viewed from the frame at
which it is at rest.
If we choose coordinates so that the relative velocity is along the x axis, v = (v, 0, 0), then
in the other frame we have,
x2 v 2
x̃2 + ỹ 2 + z̃ 2 = x2 + y 2 + z 2 + = γ 2 x2 + y 2 + z 2
c2 (1 − v 2 /c2 )
210 CHAPTER 14. SPECIAL RELATIVITY
Consider now a set of N particles in a given volume, then the particle number density,
ρ := N
V
will depend on the observer that is viewing it, indeed, in some other reference frame
we will see a density given by ρ = VN0 = NV0γ = γρ0 . Nevertheless, there is an object which is
γ
invariant, the current density, defined by,
j := ρ0 u
where u is the four-velocity of the center of mass of the particles and ρ0 the density as
measured on that frame. If t is another observer, then, u = γ(t + v) with η(v, t) = 0 and
therefore,
∂j µ 1 ∂ρ ∂(ρv i )
= [ + ]
∂xµ c ∂t ∂xi
14.3. RELATIVISTIC OPTICS 211
So we see we get the expression for particle number conservation when we set this four-
divergence to zero. In general, we can interpret it as the rate of particle creation. It also
applies to other currents, as the charge density and electric currents, which we see they get
together into a single four-vector.
φ(xµ ) = f (xµ kµ )
for some smooth function f : lR → lR. Clearly the surface of constant f are hyperplanes in
space-time. Given a reference frame with Cartesian coordinates xµ = (ct, xi ) we have,
where kµ = ( ωc , ~k) . We see that expression represents a function which is constant along the
lines perpendicular to k, namely the position vectors of the form,
−~k · ~x
t=
ω
These hyperplanes are the level sets of f .
The intersection of a simultaneity hyper-surface of this coordinate system with a level sets
of f consists the sum of a fixed (space) position vector plus an arbitrary vector perpendicular
to ~k. That is on each simultaneity hyper-surface f is constant along planes perpendicular to
~k. That is, the intersection of two hyper-planes is a 2-plane. We can think of those space 2-
planes as moving along the lines (t, ~x = −ω
k2
t). In the case that f (s) := f0 eis , ω is interpreted
as the wave frequency that observer at rest in such a frame will see the wave has.
If we apply the wave equation to this function we get,
∂2φ ω 2 ~ ~ ′′
0 = ✷φ(t, ~x) = − ∆φ = ( − k · k)f
c2 ∂t2 c2
thus we see that in order that this be a solution we need that the four-co-vector k be a null
vector, indeed the above expression is just
η µν kµ kν = 0
1 0 0 0
0 −1 0 0
η µν = (14.11)
0 0 −1 0
0 0 0 −1
Therefore, we must consider ray co-vectors of the form, kµ = (± ωc , ωc k̂) with k̂ any unit
vector. We shall take the plus sign for the time like part, which corresponds to a wave
moving in the direction opposite to k̂. This is so because the vector form of kµ has the form,
k µ := η µν kν = ωc (1, −k̂). This will be used when discussing aberration §14.3.2. Note that
given any time like vector, t, any null vector can be decomposed as,
k = a(t + k̂).
Contracting with t we obtain, as for the case of time-like vectors, a = η(t, k). Contrary to
the case for time-like vectors, the value of this quantity can not be expressed in terms of the
norm of the space part of k. It is just the frequency the observer t sees the wave has in its
reference frame. A property of the whole vector k. Indeed, the null character of the vector
only implies that η(k̂, k̂) = −1, or k̂ · k̂ = 1.
The geometrical description of this is simple, if one marks, null planes at which the function
f reaches, say, its maximum, and several unit vectors, then the number of those planes pierced
through by that those vectors will be proportional to the frequency an observer moving along
the line defined by each one of those vectors will measure that frequency. See figure 14.14.
Figure 14.14: Null planes and unit vectors. The number of planes the unit vector pierced
through by the unit vector is proportional to the frequency the observer moving along it will
measure.
ω = ct(k) = ctµ kµ ,
the speed of light here appears because x0 = ct. Like wise, for any other observer, t̃ = γ(t+β)
the frequency will be,
ω̃ 1 + ~vc·k̂
=q ,
ω 1 − vc2
2
as expected from the usual derivation for sound waves, the frequency is bigger when the ob-
server moves at a faster speed towards the light ray than the previous observer. In particular,
we can consider a light source which emits at a given frequency determined by its internal
state. So on the rest frame of the source we will observe the real frequency as determined by
its internal state, ω0 . Any other observer will see this ray with a different frequency given by,
1 − ~vc·k̂
ω = ω0 q 2
1 − vc2
where ~v is the velocity of the observer respect to the light source, and k̂ is the wave direction
as pointing in the direction to the other observer (the negative of the previously used direction
vector). Notice that the Doppler effect has two different contributions, one is the usual we
know from sound waves, and is related to the relative velocities of sources and receivers, and
depends on the sign of that velocity. The other is a purely relativistic effect, a γ factor, which
is present even in the case when the source moves perpendicular to the observer direction.
For relativistic situations this factor can be dominant.
t t̃
light ray
k̂
tr
ts
k̂
receiver path
Exercise: At rest a interstellar hydrogen cloud absorbs light at some given spectral line
frequencies, can we infer its relative motion with respect to a further away background star?
Exercise: Millisecond pulsars binaries. Using Newtonian gravity estimate the Doppler shift
of a millisecond pulsar rotating around a normal star with a period of 1 hour.
Exercise: What happens when the observer velocity is perpendicular to the line of sight of
the light source object?
14.3.2 Aberration
Light aberration is also a classical (pre-relativistic) phenomena which has also its relativistic
counterpart. In fact was used for the second measurement of the light speed by Bradley in
1729.
The observation of this effect consists in realizing that the angle at which a star is seen in
the sky depends on the velocity the observer has with respect to the star. Indeed, when the
star is sufficiently far away all light coming from it can be assumed to reach the telescope at
parallel rays. If the telescope is at rest with respect to the star one would measure the correct
14.3. RELATIVISTIC OPTICS 215
angles, in particular the declination angle, θ. If the earth, and so the telescope is moving, for
simplicity in the same direction as the star is, then the light entering the telescope aperture
will arrive at the ocular when this has already traveled a distance from the original position,
thus resulting in a measurement of the declination angle smaller that when at rest with respect
to the faraway star.
L ~v
θ′
θ
B
R cos θ
L cos θ ′
From the figure 14.17 we see that A = L sin θ′ = R sin θ, and R = c∆t, on the other hand,
B = L cos θ′ − R cos θ = v∆t. Therefore,
v
cos θ′ sin θ − cos θ sin θ′ =sin θ′
c
For small effects, defining dθ = θ − θ and to first order in vc we get,
′
v
dθ = − sin θ.
c
Using measurements of this effect when the earth was in opposite sides of its orbit around
the sun, Bradley, who discovered it, was able to measure the speed of light. Notice that,
as Doppler’s, it is an effect of order vc , so not so difficult to measure. It is not due to the
2
Minkowski metric, which only can give effects of order vc2 , and so only relativistic in the sense
that the speed that appears is a relative speed.
We now consider the full relativistic version, for this it is better to consider two stars at
slightly different positions and relate all angles to the angles between them. This is actually
216 CHAPTER 14. SPECIAL RELATIVITY
much closer to the real measurement, a plate is taken of a sector of the sky, and then a second
one 6 month latter. One then looks for the difference on the plate distance among the stars,
which is just proportional to the angles they have on the sky on both situations. Following the
situation in figure 14.18, we have two stars, each one sending light according to null vectors
k1 and k2 . At event A they are registered by two observers moving with four-velocities t and
t′ respectively. In term of these two vectors we have,
t
t′ stars
light rays
k̂2
k̂1
ω1 ω1′ ′ ′
k1 = (t + k̂1 ) = (t + k̂1 )
c c
ω2 ω2′ ′ ′
k2 = (t + k̂2 ) = (t + k̂2 )
c c
′
where k̂i ( k̂i ) are unit vectors perpendicular to t (respectively t′ ). Therefore,
ω1 ω2 ω1′ ω2′
η(k1 , k2 ) = (1 − k̂ 1 · k̂ 2 ) = (1 − k̂1′ · k̂2′ )
c2 c2
from where we get,
(sin 2θ )2 (1 + v cos θ1
c
)(1 + v cosc θ2 )
′ = 2 ,
(sin θ2 )2 1 − vc2
This is the relativistic version of the aberration effect.
dη(u, u)
= 0,
ds
u is perpendicular to the 4-velocity,
so the 4-acceleration a := dds
dη(u, u) du
0= = 2η( , u) = 2η(a, u).
ds ds
Thus, we expect to have only three equations out of a vectorial equation determining a. The
4-acceleration is a spatial vector with respect to the frame instantaneously co-moving with u,
that is, a = (0, c~a2 ). On it we impose Newtons equations,
m0~a = F~ , (14.12)
Here m0 is the mass of the particle as measured in the instantaneous rest frame of the particle,
an additive property of the matter. If no forces are present, then we see that the velocity
vector remains constant and so, in this coordinate system we have a linear relation among
the coordinates along the trajectory. That is why we choose the proper time parametrization,
any other parametrization not proportional to proper time would make the above relation
undetermined, unless the force would also have information on the parametrization chosen.
So we can remove now our frames and write an invariant four dimensional expression for the
above law.
218 CHAPTER 14. SPECIAL RELATIVITY
du
m0 = f ⊥ := f − η(u, f )u (14.13)
ds
where f is any given 4-vector in space-time representing forces, and we have taken in the
right-hand side of the equation the part of it perpendicular to u, since we have, η(a, u) = 0.
We have written explicitly the acceleration to remark that the derivative we are taking is with
respect to proper time and not coordinate time, which would be wrong here, for the above
expression is valid for general coordinate systems (so no reference to a particular coordinate
system can occur).
How this expression results in an arbitrary system? Consider then a system whose constant
4-velocity is t, then with respect to these systems u = γ(t + β), with γ = η(t, u) = q 1 2 .
1− v2
c
For the force we write,
F
f ⊥ := γ(λt + ),
c2
where, from the perpendicularity condition,
η(F , u) c2 η(F , u) 1
λ=− 2
= = 3 F~ · ~v,
c η(t, u) γ c
is the work generated by F~ by unit coordinate time. There result two equations, one when
contracting 14.13 with t, and the other when taking the perpendicular part, (recalling that t
is a constant vector),
dγ
m0 c2 = F ·v (14.14)
dt
dγv
m0 = F (14.15)
dt
Where we have replaced the proper time interval ds by cdt γ
. As already mentioned, there
are only three independent equations, and in fact it is easy to see that the first one is a
consequence of the others.
To see this consider the limit of v << c and perform a Taylor expansion of both equations
v2
above, since γ ≈ 1 + 2c 2,
2 2
2 dγ d( v ) d( m02v )
m0 c ≈ m0 2 ≈ ≈F ·v (14.16)
dt dt dt
dγv dv dv
m0 ≈ m0 ≈ m0 ≈F (14.17)
dt dt dt
therefore the first equation follows, up to an integration constant, from integrating the second
when contracted with v. It is the well known equation giving the variation of the kinetic
energy due to work done on the system.
Exercise: Show that the first equation follows from the second one. To see this, get an
2
equation relating γ dγ
ds
, v 2 and dv
ds
. Integrate it to get back the expression of γ as a function of
14.4. ENERGY-MOMENTUM 4-VECTOR 219
v 2 . Compare with the Newtonian version, obtained by Taylor expanding the above expression
and retaining the first non-zero terms.
We can take a further step and define p := m0 u, the 4-momentum of the particle. As in
Newtonian mechanics, in cases where the masses can change we generalize the above system
to,
dp
= f, (14.18)
ds
in terms of this 4-vector,
q
m0 := η(p, p) (14.19)
can be taken to be the definition of rest mass. We shall come back to this formula once we
formulate Maxwell’s fields as objects in four dimensions and introduce Lorentz force.
If we have several particles, then the total 4-momentum, is given by,
N
X
pT := pi (14.20)
i=1
Now, this vector is the sum of vectors of the form mi ui , with mi positive and ui , that is
vectors pointing into the future of the light cone, namely the cone whose boundary consists of
all vectors of norm zero, see figure 14.8. But the sum of any two vectors in this cone produces
another vector inside the cone, thus any sum of vectors inside the cone gives a vector also
inside it and so the total momentum vector is inside the light cone and future directed. So
we can use its direction to define a rest frame, the center of mass rest frame of the particle
system.
We shall assume it is conserved in absence of external forces. This statement does not
follow from the similar one in Newtonian mechanics, for it assumes contact forces (so as to
be taking place at the same event) which in general is not the case. The general proof of this
statement needs a different formalism, the introduction of interaction fields in a variational
setting. This conservation law unifies both energy and momentum conservation in just one
law. As it stands, without the existence of extra fields carrying the interactions, it can not
possibly be true. For instance consider the two situations in the figure 14.19, clearly on each
one of these two simultaneity surfaces the total momentum is different, for they are considered
at different times. The only unambiguous case is when all interactions happen at the same
event, and then all momenta are at the same point, at which, of course the calculation does
not make any sense. The situation in which does makes sense is when the interaction among
particles is limited to a region of space-time, and we look at the total 4-momentum much
before and much later the interaction region, in that case the trajectories are all constants
and it makes sense to add vectors at different points in space-time, this is what is called a
scattering situation, see figure 14.20.
220 CHAPTER 14. SPECIAL RELATIVITY
interaction region
Figure 14.20: Scattering situation, the sums are done before and after all interactions have
finished.
Chapter 15
Relativistic Electrodynamics
j µ := ρ0 uµ
the charge density measured by an observer for which the charges are momentarily at rest,
times its 4-velocity. An observer with 4-velocity tµ will see a charge density given by ρ =
tµ jµ = ρ0 γ, and a charge current given by J~ = c(j − tρ) = ρ~v .
The objects we are seeking for describing the electromagnetic fields can not be a couple
of 4-vectors, this is because a Lorentz transformation mixes both components. This can be
seen from the following simple observation:
Consider an infinite conducting flat plate with constant surface charge density σ, say at x =
0. Then we know it will generate a constant electric field perpendicular to it, E ~ = (4πσ, 0, 0).
But an observer moving in, say the y direction, ~v = (0, v, 0), not only will see a bigger field,
E~ ′ = (4πσ ′ , 0, 0) = (4πσγ, 0, 0), but also a magnetic field, for he will see the charge density
′
moving and so generating a current in the opposite direction, namely, J~ = (0, −σγv, 0), which
in turn will generate a constant magnetic field given by B ~ ′ = (0, 0, −4πσγ v ) = γ ~v ∧ E.
~ Thus,
c c
both vectors must be part of a single four dimensional geometrical object.
Since at every point of space-time we have six independent components, we must ask
ourselves which kind of tensor have that many independent
components.
Since a single 4-
1 0 2
vector would not do we try next with tensors of type 1 , or 2 or 0 . Since we can transform
any of this kind of tensors in any other of them by contracting with the metric or its inverse,
to consider just one kind of them is enough. But they have 4 × 4 = 16 components! So
221
222 CHAPTER 15. RELATIVISTIC ELECTRODYNAMICS
for this to work we must impose some conditions among their components so as to get only
6 independent ones. These conditions must be valid for the components in any coordinate
system, (in particular Lorentz transformations), that is they must be geometrically invariant.
There are only three type of conditions which satisfy this, namely to ask for the tensor to be
symmetric: tµν = tνµ , antisymmetric: tµν = −tνµ , and/or trace free: η µν tµν = 0. Symmetric
four-dimensional bi-tensors, like the metric, have 10 independent components, while the trace
free condition only imposes 1 condition, so even if we require both conditions we are left
with 9 components. They are no good. But anti-symmetric four-dimensional bi-tensors
have precisely 6 components. So it seems that Maxwell’s fields must have a four-dimensional
realization as an antisymmetric bi-tensor, Fµν = −Fνµ . How do we relate it to our old friends,
~ and B?
E, ~ Given an observer t we can make,
Eµ := Fµν tν
notice that this is a vector in the space perpendicular to t, indeed from the anti-symmetry
of F we see that Eµ tµ := Fµν tν tµ = 0. So we assert that this will be our electric field for
this observer. How do we obtain B ~ now? Out of t, and F we need to obtain another space
vector, this can be done with the help of the completely antisymmetric tensor εµνσρ with
ε0123 = −1 in one of our Minkowskian Cartesian systems. 1 This is the Levi-Civita tensor in
four dimensions. Once we have fixed the value of this component all other components are
fixed because the antisymmetry of it. With the help of this tensor we define,
1⋆
F µν := εµνσρ Fσρ ,
2
This new tensor is also antisymmetric and so,
Bµ := −⋆ Fµν tν
~
is also a spatial vector perpendicular to t, this one is B!
Exercise: Find the components of ⋆ Fi0 . To which vector corresponds 12 εijk⋆ Fjk ?
1
One could choose the opposite sign, this corresponds to a change in the direction of one of the coordinate
axis, that is to a parity transformation. With this sign tρ ερµνσ is the conventional volume 3-form.
15.2. MAXWELL’S EQUATIONS 223
⋆
F 00 ⋆
F 01 ⋆
F 02 ⋆
F 03 0 −F23 F13 −F12
⋆ 10 ⋆ 11 ⋆ 12 ⋆ 13
⋆
F F F F F23 0 F30 −F02
F µν = =
⋆ 20 ⋆ 21 ⋆ 22 ⋆ 23
F F F F −F13 −F30 0 F10
⋆ 30 ⋆ 31 ⋆ 32 ⋆ 33
F F F F F12 F02 −F10 0
0 B1 B2 B3
−B 1 0 E3 −E2
= (15.2)
2
−B −E3 0 E1
−B 3 E2 −E1 0
η µν ∂µ Fνσ , (15.3)
and
−1 µνσρ
ε ∂ν Fσρ = ∂ν ⋆ F νµ . (15.4)
2
The first expression is a vector, while the second is a pseudo-vector, since it is built out of
Fµν , and εµνσρ , the Levi-Civita tensor, a pseudo-tensor since it changes sign when we perform
time or space inversion. Thus, since matter is vectorial in nature, we must have,
∂ν ⋆ F νµ = 0. (15.6)
To see that they are indeed Maxwell’s equations we write them in Cartesian coordinates,
x = (ct, x, y, z) (so that the metric is now ηµν = diag(1, −1, −1, −1)). Then tµ = (1, 0, 0, 0), is
µ
unitary, and we can write a 4-velocity vector as uµ = γ(tµ + vcµ ) (recalling that tµ = (1, 0, 0, 0))
The zero component, ν = 0, of 15.5 is,
∂t B i + cεijk ∂j Ek = 0 ∂i B i = 0
and again we see that the separations between constraint and evolution equations is observer
dependent.
Otherwise, for simplicity, we can take the alternative version of 15.6, 15.8 and compute
some of its components to see that they correspond to the above expression. Indeed, com-
puting
and choosing i = 1, and j = 2 we see that we get the third component of the first of the
above equations.
Exercise: Compute ∂[i Fjk] , how many equations are these? To what do they correspond?
So we see that the tensor Fµν has all the information of Maxwell fields and also can encode
in a very simple manner Maxwell’s equations. It is called the Maxwell tensor, and it is the
real four dimensional object behind the electromagnetic phenomena. It unifies in one entity
all magnetic and electric phenomena in a natural way.
Exercise: Use the above expression of F to show that B µ = −⋆ F µν tν . For that prove first
that
εµνσρ εκχσρ = −2(δµ κ δν χ − δν κ δµ χ ),
(Hint: the above expression is coordinate independent, so choose any easy coordinate basis
and compute it.)
Fµν ⋆ F µν = 2E µ Bµ .
How do we know there are no more invariants? What about, for instance, Fµν F µσ Fσκ F κν ?
Notice that since, given any t, unit time-like, F , is equivalent to two vectors perpendicular to
it, then it follows, since vectors are not invariant under rotations on the space perpendicular
to t that all invariants must be also invariant under rotations and so can only depend on
E µ Eµ , B µ Bµ , or E µ Bµ . We already saw that the combinations Eµ E µ − Bµ B µ , and E µ Bµ are
invariants. We shall see later that the remaining combination, namely Eµ E µ + Bµ B µ , is not
an invariant (it changes values for different observers), but rather the component of a tensor.
So there are just two invariants.
Exercise: Show that if at some point of space-time Fµν F µν > 0 and Fµν ⋆ F µν = 0, that is,
if for any observer at that point the electric and magnetic fields are perpendicular and the
magnetic field is bigger in norm than the electric field, then there is an observer for which the
electric field vanishes.
Eµ = Fµν tν Bµ = −⋆ Fµν tν
while for another, t̃, t̃µ = γ(tµ + β µ ) with tµ βµ = 0, tµ tµ = −1,
Since E µ Bµ = 0, we can choose the space coordinate axis so that E µ = (0, E, 0, 0), B µ =
(0, 0, B, 0), we further choose a velocity in the perpendicular direction, β µ = (0, 0, 0, β), we
E
get, Ẽ0 = Ẽ2 = Ẽ3 = 0 and Ẽ1 = γ(E − βB). So, since |B| > |E|, taking β = B < 1 we
obtain Ẽ1 = 0.
In this frame there is no electric force, this is the case when studying electromagnetic fields
in a plasma. There the high mobility of electrons implies that there are no electrical forces
present, that is, the plasma acts like a perfect conductor for time scales larger that those
for which the electrons re-accommodate. Nevertheless, in this situation, magnetic fields are
important and the resulting theory is magneto-hydrodynamics.
15.4. THE ENERGY-MOMENTUM TENSOR 227
Exercise: Show that if for some time-like vector uµ , Fµν uν = 0 then ⋆ Fµν = 2b[µ uν] for some
space-like vector bµ . Show that it also follows that F µν ⋆ Fµν = 0 and F µν Fµν ≤ 0
−1 µσ νρ 1
T µν := (F F ησρ − η µν F σρ Fσρ ) (15.13)
4π 4
Note that it is symmetric and its trace vanishes, T := T µν ηµν = 0.
If we contract it twice with a unit, time-like vector tµ we get,
1 1 1
e := T µν tµ tν = (E 2 + (2(−E 2 + B 2 ))) = (E 2 + B 2 ),
4π 4 8π
which we recognize as the energy density of the electromagnetic field that the observer with
4-velocity t sees. Thus, the vector
pµ := T µν tν (15.14)
must be the 4-momentum of the electromagnetic field as seen by that observer. Thus, we see
that for the electromagnetic field the 4-momentum is not a basic object, but it depends on
the observer. Aside for particles all other physical objects have energy momentum tensors
from which observer dependent 4-momenta can be built.
~ and B
Exercise: Check that pµ in terms of E ~ is given by,
P~ 1
pµ := etµ + = ((E 2 + B 2 )tµ − 2tρ ερµσν Eσ Bν ),
c 8π
Proof: We have seen that its time component is positive, so it remains to see whether
~ ~
the vector is time-like or null. That is pµ pµ ≥ 0. So we compute pµ pµ = e2 − Pc·2P =
1 ~ ∧ B|
((E 2 + B 2 )2 − 4|E ~ 2 ). But |E
~ ∧ B|
~ ≤ |E||
~ B|,
~ so (E 2 + B 2 ) ≥ 2|E||
~ B|~ ≥ 2|E
~ ∧ B|
~ and
(8π)2
µ
p pµ ≥ 0.
Exercise: Check that if pµ tν = 0 for any future directed vector t, then at that point, Fµν = 0.
Hint, split the Maxwell tensor in terms of its electric and magnetic parts with respect to tµ .
228 CHAPTER 15. RELATIVISTIC ELECTRODYNAMICS
Exercise: Show that if a given time-like or null future directed vector when contracted with
another time-like future directed vector vanishes, then the given vector itself vanishes.
~ and B.
Exercise: Write down all components of T µν in terms of E ~
⋆ 1
F µσ ⋆ Fσν = δ µ ν F σρ Fσρ + F µσ Fσν . (15.15)
2
b) Check that,
⋆
F µν ⋆ Fµν = −F µν Fµν .
c) Write the energy momentum in terms of ⋆ F .
∂µ T µ ν = j µ Fµν (15.16)
by virtue of Maxwell’s equations. Indeed,
−1 1
∂µ T µ ν = ((∂µ F µσ )Fνσ + F µσ ∂µ Fνσ − ∂ν (F σρ Fσρ ))
4π 4
−1 σ µσ 1
= (−4πj Fνσ + F (∂µ Fνσ − ∂ν Fµσ ))
4π 2
σ 1 µσ
= −j Fσν − F (∂µ Fνσ − ∂µ Fσν − ∂ν Fµσ )
8π
1 µσ
= −j σ Fσν − F (∂µ Fνσ + ∂σ Fµν + ∂ν Fσν )
8π
3 µσ
= −j σ Fσν − F (∂[µ Fνσ] )
8π
= −j σ Fσν (15.17)
where in the second line we have used one of Maxwell’s equations, in the third and fourth the
antisymmetry of F and in the fifth the other Maxwell equation.
To see the relevance of the above formula take a constant unit time-like vector t, (so in
some Cartesian coordinate system it has components tµ = (1, 0, 0, 0), when writing coordinate
15.4. THE ENERGY-MOMENTUM TENSOR 229
components we shall refer to them), and define pµ := T µν tν , the 4-momentum that this
observer assigns to the electromagnetic field. We have,
∂µ pµ = ∂µ (T µν tν )
= (∂µ T µν )tν T µν ∂µ tν
= −j σ Fσ ν tν
J~ ~
= · E, (15.18)
c
where we have used the constancy of t, namely ∂µ tν = 0. So we see that locally the energy
~ in the times the velocity of the
momentum loss is just the work done by the electric force ρE
particles, ~v .
We shall consider what happens with this result when considering a finite region, as the
one in the figure 15.1
Σ1 −t
t=T
n
T
t
Σ0
t=0
where the integral on the left is over a volume in space-time, and the others are over three-
dimensional hypersurfaces as shown in the figure. The normal vectors point outwards to
the hypersurfaces at t = 0 and t = T are respectively, n0µ = (−1, 0, 0, 0) = −tµ , and nTµ =
(1, 0, 0, 0) = tµ . 2 Using the above result we then have,
Z Z Z Z
pµ tµ d3 x = pµ tµ d3 x + pµ Nµ d3 x − J~ · E
~ d4 x. (15.20)
ΣT Σ0 T V
That is,
2
To see that these are the correct signs for the integrals over the t = 0 and t = T hypersurfaces assume pµ
has only time component and perform the integral of ∂0 p0 along x0 .
230 CHAPTER 15. RELATIVISTIC ELECTRODYNAMICS
Z Z Z Z
p0 d 3 x = p0 d 3 x + pµ Nµ d3 x − J~ · E
~ d4 x, (15.21)
ΣT Σ0 T V
the energy at ΣT is the energy at Σ0 minus the integral over the side-boundary and the work
done by the field to the charges, if any is present.
So we now examine the side-boundary term, taking any vector N perpendicular to t, that
is spatial, one sees that,
1 µσ 1
pµ Nµ = [F Fνσ tν − tµ F σρ Fσρ ]Nµ
4π 4
−1
= Eσ F µσ Nµ
4π
−1
= Eν Nµ (−2E [µ tν] − εµν σρ B σ tρ )
4π
1
= Eν Nµ εµν σρ B σ tρ
4π
−1 ρ
= t ερµνσ N µ E ν B σ
4π
1 ~ ~ ~
= N · (E ∧ B),
4π
1~ ~
= N ·S (15.22)
c
is Poynting’s vector, S ~ = cE ~ ∧B~ 3 Thus, we conclude that the change in energy form
4π
hypersurface Σ0 to hypersurface ΣT due to volumetric work done upon the charges integrated
over time and energy flux losses, the (time×2-surface) integral along the boundary of the
region. In the case the boundary is tangent to the time direction (so its normal is perpendicular
to t) that is the space integration region is the same at all times for this observer, the energy
flux is just Poynting’s vector.
Exercise: Assume the region under consideration is moving with four velocity uµ = γ(tµ +βµ )
with respect to the observer t with respect to which we measure energy. See figure 15.2. What
is the energy flux in this case?
Exercise: Show that pµ uµ = T µν tν uµ > 0 if both t, and u are time-like and future directed.
Proof: This follows from the fact that p is time-like or null and future directed, but it is
instructive to do the following calculation, taking, for simplicity, both vectors to have norm
one we and so uµ = γ(tµ + β µ ), with η(t, β) = 0, we have,
3 1
The factor c in the expression above is due to the fact that on the right-hand side we integrate power in
units of ct.
15.4. THE ENERGY-MOMENTUM TENSOR 231
Σ1
−t t=T
u
n T
t
Σ0
t=0
pµ uµ = γ(pµ tµ + pµ βµ )
γ
= (E 2 + B 2 − 2β~ · (E
~ ∧ B))
~
8π
γ
≥ (E 2 + B 2 − 2βEB)
8π
γ
≥ (E 2 + B 2 )(1 − β)
8π
≥ 0,
~ ∧ B)|
since β < 1, |(E ~ ≤ |E||
~ B|,
~ and E 2 + B 2 ≥ EB. Notice that this quantity is zero only
when both electric and magnetic fields vanishes.
Exercise: Use the above result to show, that if two solutions to Maxwell equations coincide
at Σ0 then they must coincide inside a region like the one shown in figure 15.3 as long as the
normal to ΣT is time-like.
t
ΣT u
Σ0
t=0
Tk := T µν kν satisfies,
∂µ Tkµ = ∂µ (T µν kν ) = j σ Fσ ν kν
Performing the same integral as before over a surface perpendicular to t, we get,
Z Z Z Z
− ∂µ Tkµ d4 x = T µν kν n0µ d3 x + T µν kν nTµ d3 x + T µν kν Nµ d3 x
V Σ0 ΣT T
Z Z Z
µν 3 µν 3
= T kν tµ d x − T kν tµ d x + T µν kν Nµ d3 x
Σ0 ΣT T
Z Z Z
ν 3 ν 3
= −p kν d x − −p kν d x + T µν kν Nµ d3 x
Σ0 ΣT T
(15.23)
So we see that from this equation we can get the change in the space components of the
energy-momentum 4-vector. Re-arranging this expression we get,
Z Z Z Z
pν k ν d 3 x = pν k ν d 3 x + T µν kν Nµ d3 x − j σ Fσ ν kν d4 x (15.24)
ΣT Σ0 T V
Again we see that the change in momentum has a volumetric term and a flux term. Notice
~ · ~k = 1 (E
that pµ kµ = −T µν tν kµ = 1c S ~ · ~k, so that the 3-momentum density that the
~ ∧ B)
4π
observer with 4-velocity t sees is the Poynting vector.
We now compute the momentum flux along k, −T µν Nµ kν , and the volumetric source,
−jµ F µν kν .
We first compute,
~ · ~k − tρ ερµσν Bσ kν
F µν kν = −(E µ tν − E ν tµ )kν − εµνσρ Bσ tρ kν = tµ E
So the momentum source becomes,
~ ~
~ · ~k − J · (B
−jµ F µν kν = jµ tµ E ~ · ~k − ( J ∧ B)
~ ∧ ~k) = −ρE ~ + 1 (J~ ∧ B))
~ · ~k = −(ρE ~ · ~k
c c c
So we see that the source of 3-momentum change is the Lorentz force. We turn now to
momentum flux,
1 µσ ν 1
T µν Nµ kν = [F F σ Nµ kν − ~k · N ~ F µν Fµν ]
4π 4
1 ~ ·N
~ − tρ ερσµν Bµ Nν )(−tσ E ~ · ~k − tρ ερσ µν Bµ kν ) + 1 (~k · N
~ )F µν Fµν ]
= [(−tσ E
4π 4
1 ~ · N)(
~ E ~ · ~k) + (B
~ ∧N ~ ∧ ~k) + (E
~ ) · (B 1 ~ ·E~ −B ~ ~k · N)]
~ · B)( ~
= [−(E
4π 2
1 ~ · N)(
~ E ~ · ~k) + B ~ ~k · N
~ · B( ~ ) − (B ~ · N)(
~ B ~ · ~k) + 1 (E~ ·E ~ −B ~ ~k · N)]
~ · B)( ~
= [−(E
4π 2
1 ~ ~ ~ ~k · N)
~ · B)( ~ − 2(E ~ ·N ~ )(E~ · ~k) − 2(B~ ·N ~ · ~k)]
~ )(B
= [(E · E + B (15.25)
8π
15.4. THE ENERGY-MOMENTUM TENSOR 233
1. Momentum density is proportional to the energy flux, i.e. the Poynting vector.
3. The momentum flux is Maxwell’s stress tensor, so in a static situation, when no mo-
mentum change occur, the total electromagnetic force over an object is proportional to
the surface integral of Maxwell stress tensor. T ij = −1
4π
[E i E j + B i B j − 21 eij (E 2 + B 2 )]
∂µ (T µν kν ) = (∂µ T µν )kν ) + T µν ∂µ kν .
Thus, to have a conservation law we only need a vector kµ such that T µν ∂µ kν = 0, in particular,
since T µν is symmetric, T µν = T νµ , it follows that there are conservation laws for all vectors
such that ∂(µ kν) := 12 (∂µ kν + ∂ν kµ ) = 0. They are called Killing vectors. It can be shown
that in Minkowski space-time there are 10 linearly independent vectors with this property.
We have already seen four of them, one time translation and three space translations. The
rest can be constructed taking arbitrary constant antisymmetric tensors Aµν and defining,
k µ = εµνρσ xν Aρσ ,
where xµ = ηµν xν , are Cartesian coordinate functions.
Exercise: Using a constant vectors base construct the remaining five quantities. Interpret
them as motions in space-time.
234 CHAPTER 15. RELATIVISTIC ELECTRODYNAMICS
Exercise: The above definitions depend on the coordinate origin, show that if one change
origin then the above vectors change by a translation.
The formulas used to express the conserved quantities, fluxes and sources for space and
time translations can be used also for angular momentum, making the appropriate substitu-
tions, of course. They shall be used latter in an example.
~ and B
Exercise: Express in terms of E ~ the angular momentum along the ẑ axis.
Exercise: Find the angular momentum flux. Find the total torque an electromagnetic field
exerts over a static source in terms of a surface integral.
∂(µ kν) = 0.
In the presiding section we have seen that there is at least ten (10) of them which are
linearly independent, four (4), are just constant vectors, taking a constant time-like vector
t, we can split the remaining six into three (3) infinitesimal rotations, given by Aµν = tµ eνi
where the vectors ei represent the rotation axis, which for simplicity we take to be three
orthonormal space-like vectors perpendicular to t, and three (3) boosts , which we take to
be, those vectors obtained from Aµν = eµi eνj , i 6= j. They are just proper infinitesimal Lorentz
transformations along a 2-plane perpendicular to both frame vectors.
These vectors represent symmetries of the space-time and so are related to conserved
quantities. So it is important to know how many of them there are, for we will have as many
conserved quantities as Killing vectors.
To see this, we shall the following argument [2]: note that if we define
∂[σ Lµν] = 0.
Thus,
anti-symmetric matrices Lµν (there are six linearly-independent) we obtain the other six
Killing vector fields. So we see that these are all possible (up to linear constant combinations)
infinitesimal symmetries of space-time.
Exercise: Prove that if a Killing vector field vanishes at a point together with its first
derivatives then it is the trivial vector field.
∂µ (T µν kν ) = (∂µ T µν )kν + T µν ∂µ kν ,
we see that not only the last term vanishes when ∂(µ kν) vanishes, but also if it is proportional
to ηµν , that is
∂ρ Lµ = 0.
Thus, we have another five possibilities, namely we can give the value of φ at some point and
the constants Lµ and obtain 5 extra conformal killing vector fields.
236 CHAPTER 15. RELATIVISTIC ELECTRODYNAMICS
kµ = ηµν xν or k µ = xµ ,
this corresponds to an infinitesimal dilation or scale symmetry, namely a transformation
which changes the units of time and space in the same way.
If we choose at a point φ = 0, (and set it as the coordinate origin), and some value for
Lµ , then we have,
φ = Lµ xµ ,
and,
16.1 Introduction
In the 3 + 1 dimensional treatment of Maxwell’s equations it is very useful to introduce
potentials so that certain set of equations are solved automatically and so, one can concentrate
in the rest. This idea can be further carried on in the four dimensional formalism but here it
acquires a different and more profound meaning. The potentials are not merely a simple way
of obtaining solutions or solving some equations, but have a geometrical meaning as curvature
carriers for a whole set of theories, called en general Gauge theories or Yang Mills theories.
In their more extended version they represent all force fields (bosonic fields) in the Standard
Model of particle physics, which, of course, includes among them the electromagnetic forces.
In the four dimensional setting we have just two set of Maxwell’s equations,
∂ν ⋆ F νµ = 0. (16.2)
While there has been numerous experiments trying to detect sources for the second equations,
notably magnetic monopoles, none of them have give any positive result. The tendency of
theoretical physicists is to think about the second equation as intrinsically source-less and in
that case as acting as a integrability condition or compatibility condition to the existence of
another field of a more fundamental character than the tensor F , namely the existence of a
4-vector, A, such that
Fµν = 2∂[µ Aν] . (16.3)
Indeed, it easily follows, from the fact that partial derivatives commute, that
237
238 CHAPTER 16. VECTOR POTENTIAL - CONNECTION
∂[µ Fνσ] = 0,
namely 16.2, then locally there always exists a vector potential, Aµ , in fact infinitely many,
such that 16.3 holds. Thus, we see that this equation is equivalent to the existence of vector
potentials.
Suppose now we have an Fµν and we have a potential for it, Aν , that is Fµν = 2∂[µ Aν] ,
then given any smooth function λ, õ := Aµ + ∂µ λ will also give the same Maxwell field,
indeed,
Exercise: Given a constant unit time-like vector t define Aµ = −φtµ + õ that is, φ = A0 =
A · t = A0 and à · t = 0. Check that,
B µ = ε̃µνσ ∂ν Aσ := tρ ερµνσ ∂ν Aσ ,
and
Eµ = ∂µ φ − tµ ∂0 φ − ∂0 õ = ∂˜µ φ − ∂0 õ ,
where ˜ means projection to the simultaneity surface. Check explicitly that both vectors
~ = −∇φ,
are gauge invariant. Notice that with this definition E ~ consistent with the usual
definition.
∂µ Aµ = 0. (16.5)
To see that this is possible consider an arbitrary potential, Aµ , then make a gauge transfor-
mation, õ = Aµ + ∂µ λ, then we have,
∂µ õ = ∂µ Aµ + ∂ µ ∂µ λ
So we can solve for λ the equation,
✷λ = −∂µ Aµ
and the new potential would satisfy Lorentz gauge.
Notice that this does not fix completely the gauge freedom, the arbitrariness is the same
as the set of solutions of the wave equation, namely we can prescribe the value of λ and of
∂0 λ at some t = t0 hypersurface.
For this choice of gauge Maxwell’s equations reduces to just the wave equation,
∂0 ∂µ Aµ = 0 ⇒ ∂02 A0 = −∂i ∂0 Ai ,
so we have,
Thus, we see that given Ai and ∂0 Ai at t = t0 , then values for A0 and ∂0 A0 are determined
by the above equations. So in this gauge the free initial data for Maxwell’s are the pairs
(Ai , ∂0 Ai ).
Exercise: Check that the equation for A0 (last equation) is gauge invariant. Confirm that
~ ·E
equation is just ∇ ~ = 4πρ
∂0 A0 |t=t0 = 0
∆(A0 |t=t0 ) = −4πρ|t=t0 . (16.10)
∂i Ai = 0 (16.11)
This gauge is not space-time invariant, but it is very convenient in some calculations for it
resembles what happens in electrostatics and magnetostatics.
In this case the equations become,
∂ 0 ∂0 A0 − ∂ i ∂i A0 − ∂ 0 ∂0 A0 = −∂ i ∂i A0 = 4πρ
✷Ai − ∂ i (∂0 A0 ) = 4πj i (16.13)
In this gauge one solves for A0 at each time slice, t = cons., using the first equation above
and then plugs the result into the second equations as a source to solve for Ai .
The consistency of the gauge fixing condition is granted by requiring for the initial condi-
tions for the second equation (Ai , ∂0 Ai )|t=t0 that they satisfy,
✷∂iAi = ∂i ✷Ai
= ∂i ∂ i (∂0 A0 ) + 4π∂i j i
= ∂0 ∂i ∂ i A0 + 4π∂i j i
= 4π∂0 ρ + 4π∂i j i
= 4π∂µ j µ
= 0. (16.15)
So the conditions on the initial data ensure that the unique solution to this equation is
∂i Ai = 0 and the gauge fixing condition is granted.
It can be shown that given Fµν |t=t0 there is a unique initial data pair, (Ai , ∂0 Ai )|t=t0 such
that 16.14 holds. Notice that in this gauge,
~ = −∇φ
E ~ + ∂0 A,
~
with,
∆φ = −4πρ, ~ · (∂0 A)
and ∇ ~ = 0,
at all times.
Exercise: Check that given Fµν |t=t0 satisfying initially Maxwell’s equations, with ∂ i F0i and
∂ i Fij of compact support. There is a unique pair (Aµ |t=t0 , (∂0 Aµ )|t=t0 ) satisfying, Fµν =
2∂[µ Aν] , ∂µ Aµ = 0, (∆A0 + 4πρ)|t=t0 = 0, ∂0 A0 |t=t0 = 0, and decaying sufficiently fast at
infinity.
Proof: The Lorentz gauge plus the condition, ∂0 A0 |t=t0 = 0 implies, that ∂ i Ai |t=t0 = 0,
so we assume this. Consider now the relation, Fij = 2∂[i Aj] . Taking the divergence of this
expression we get,
∂ i Fij = ∆Aj − ∂j ∂ i Ai = ∆Aj
This is Poisson’s equation for each Aj , so if ∂ i Fij is of compact support, as assumed, there is
a unique Ai satisfying it and decaying at infinity.
242 CHAPTER 16. VECTOR POTENTIAL - CONNECTION
1 ~ ~ 1 i
The equation (∆A0 +4πρ)|t=t0 = 0 fixes uniquely A0 if ρ = 4π ∇· E = 4π ∂ Fi0 is of compact
0
support and A is assumed to decay sufficiently fast. Thus, Fi0 = ∂i A0 − ∂0 Ai determines
∂0 Ai uniquely. Notice, that
0 = ∂ i Fi0 − ∆A0 = ∂ i ∂0 Ai .
So this vector potential initially satisfies Coulomb gauge.
∂0 (∂i Ai ) = 4πρ
✷Ai − ∂ i (∂j Aj ) = 4πj i, (16.16)
so the first equation can be integrated in time to find ∂i Ai , this value can then be inserted into
the second equation and then this one is solved for Ai . Again, charge conservation ensures
consistency.
Variational Formalisms
17.1 Introduction
Variational formalisms, that is the treatment of systems of equations as coming from a varia-
tional principle, are central in physics for they allow a different view point to understand these
systems, and, in many cases, to apply powerful mathematical techniques to solve them or to
find out how they behave. We start with a mechanical, finite dimensional system, namely
the variational principle ruling the motion of a charged particle in the presence of a given,
fixed, electromagnetic field. We then present the variational principle leading to Maxwell’s
equations interacting with charged matter. What we show here is just a glimpse of the beauty
of these theories and, in order not to lose the scope and purpose of this book we shall not dis-
cuss the many interesting properties and ramifications these approach renders. But we hope
the reader will get interested and will return to these topics at future instances of his/her
formation.
243
244 CHAPTER 17. VARIATIONAL FORMALISMS
The simplest expression which we can build out of the four velocity and the electromagnetic
field which is re-parametrization invariant is,
s
1 dxµ dxν e dxµ
Z
IAB := [m0 c ηµν + Aµ (xν (τ )) ] dτ. (17.2)
0 dτ dτ c dτ
This expression depends on the potential and so the first question we must ask is whether it is
gauge invariant, otherwise any conclusion we could obtain from it would not be physical. We
have included the mass coefficient in the first term, m0 c so that the function has the dimensions
of an action, and have included another coefficient, e, which would be the definition of charge.
Since the equations will appear as a result of an extremal principle they can not depend on
the over all magnitude of this integral. Thus, the interaction will only depend on the ratio
e
m0 c2
.
Changing the 4-potential by the gradient of a function, Aµ → Aµ + ∂µ λ, we obtain,
Thus we see that this quantity is not gauge invariant, so it is not a physical quantity. Nev-
ertheless, the gauge dependence is only through the values of the field at the extremes of
the integration curve, and so variations which vanish at the extreme will still give gauge
independent equations. Indeed, looking at a one parameter families of curves of the form,
xµε := xµ0 + εxµ1 , with fixed extremes, namely, xµ1 (0) = xµ1 (1) = 0, we obtain,
dxµ dxν
dIAB 1 m0 c 0
ηµν dτ1 e dxµ dxµ
Z
|ε=0 = [ dτ
+ (xν1 ∂ν Aµ 0 + Aµ 1 )] dτ, (17.4)
dxµ dxν
q
dε 0 ηµν dτ0 dτ0 c dτ dτ
dxµ
As for the case of a free particle we now take a parametrization on which the norm of dτ
0
d2 xµ
dIAB 1 m0 c
2 ηµν x
0 ν
e ν dxµ0 dAµ µ
Z
|ε=0 = [− q dτ µ dxν1 + (x1 ∂ν Aµ − x1 )] dτ − eAµ xµ1 |10
dε 0 ηµν d0dτx dτ0 c dτ dτ
d2 xµ
m0 c dτ 20 ηµν xν1 e dxµ ∂Aµ dxν0 µ
Z 1
= [− q + (xν1 ∂ν Aµ 0 − x )] dτ
0 dxµ dxν
ηµν dτ0 dτ0 c dτ ∂xν dτ 1
d2 xµ
m0 c dτ 20 ηµν xν1 e dxµ0 ν
Z 1
= [− q + (∂ν Aµ − ∂µ Aν ) x ] dτ
0 dxµ dxν
ηµν dτ0 dτ0 c dτ 1
d2 xµ
m0 c dτ 20 ηµν e dxµ0 ν
Z 1
= [− q + (∂ A
ν µ − ∂ A
µ ν ) ]x1 dτ
0 dxµ ν
0 dx0 c dτ
ηµν dτ dτ
d2 xµ
1
2 ηµν
m0 c
0
e dxµ
Z
= [− q dτ µ ν + Fνµ 0 ]xν1 dτ
0 dx dx
ηµν dτ0 dτ0 c dτ
So the condition for the curve xµ0 (τ ) to be an extremal for the above quantity under arbitrary
perturbation curves with fixed endpoints implies it must satisfy,
µ
d2 xν0 ν dx0
m0 2 = eF µ = F ν µj µ,
dτ dτ
where we have already chosen the proper time parametrization for the extremal curve. We
see that indeed we get the Lorentz force expression on the right-hand side.
Notice that the expression for the action, 17.2, depends on the potential A instead of
the Maxwell tensor F . So we see that variational principles need the potential as a basic
component of the theory and here it is no longer a helpful way of getting rid of an equation,
acquiring a more foundational character. It is interesting that defining the Lagrangian for
this theory as,
s
dxµ dxν e dxµ
L := m0 c ηµν + Aµ (xν (τ )) ,
dτ dτ c dτ
we get as the associated momentum,
∂L dxµ e
pµ := dxµ = m0 c + Aµ
∂ dτ dτ c
and we see that now the momentum is no longer a property of the particle itself, but also
depends on the electromagnetic field present at the particle position. This property is related
to the gauge character of the theory and has a natural extension to other theories where
particles are seen as fields.
in Maxwell’s equations?
Notice that we want to vary an object with respect to some tensorial fields and obtain
equations, another tensorial objects. These equations are partial differential equations, so the
object we vary can not be a line integral, as was conventional in classical mechanics, that
would only give us ordinary differential equations, namely equations where the derivatives are
with respect to the parameters involved in the integration. So for this case the integral should
be over space-time, or at least part of it. The integrand should then be a scalar function over
the basic fields of the theory, F , or the vector potential associated to it, A. But we have
already seen that there are at most two scalars one can construct out of the tensor F , Fµν F µν
and Fµν ⋆ F µν . The same result holds if we use A and furthermore require gauge invariance.
Furthermore, we want the action to include only terms up to second order in powers of F ,
since otherwise the variation would give non-linear equations.
Thus, not much room is left and the action integral has to be of the form,
1
Z
I := [ (Fµν F µν + RFµν ⋆ F µν ) + j µ Aµ ] d4 x. (17.5)
space−time 16π
We have also added the same factor we had for the charged particle, it is the same factor, but
now to represent the interaction between charge matter and the electromagnetic field from
the side of the electromagnetic source terms. Again this is the only scalar term we can make
which is both linear in the fields and in the sources, notice that we need to use the 4-vector
1
potential, we can not get a scalar out of j and F alone. The factor 16π is to get the correct
equations when coupled to the sources, the other factor, R, is arbitrary, and we shall see that
its value is irrelevant for that part of the action does not contribute to the variation.
Again we have to check whether this term is gauge invariant, under a gauge transformation
Aµ → Aµ + ∂µ λ, we get,
Z
δI = j µ ∂µ λ d4 x
space−time
Z
= [∂µ (j µ λ) + λ∂µ j µ ] d4 x
space−time
Z
= Nµ j µ λ d3 x,
∂(space−time)
where we have used charge conservation in the second line. The integral √ of the last line has to
be understood as the limit r → ∞ of an integral over the surface r = t2 + x2 + y 2 + z 2 . So,
again, for gauge transformations which are of compact support, or which vanish sufficiently
fast at both, space and time directions the action is gauge invariant. 1
We shall perform variations over a region D of space-time bounded by two space-like
hyperplanes, Σ1 and Σ2 , one to the future of the other, both going like parallel planes towards
infinity. Imagine now we have some field A0 which we assume are an extrema of the action
1
It is difficult to make these formalisms to be mathematically correct in the sense that some of these space-
time integrals might be infinite at the extreme points, for there are many solutions which are of compact
support on space directions, but can not be so along time directions (basically by uniqueness or energy
conservation). So they are usually formal in more than one sense.
17.3. VARIATIONAL DYNAMICS OF MAXWELL’S FIELDS 247
and look at variations along arbitrary one parameter families of fields of the form Aε :=
A0 + εA1 , and consequently F ε := F 0 + εF 1 . We shall request the variations to vanish at
both hyperplanes, A1 |Σ1 = A1 |Σ2 = 0, and to decay sufficiently fast at infinity. 2
We then get,
dI Z
1
|ε=0 = [ (F0µν 2∂µ A1ν + 2R ⋆ F0µν ∂µ A1ν ) + j ν A1ν ] d4 x
dε D 8π
Z
1 1
= [ ∂µ (F0µν A1ν + R ⋆ F0µν A1ν ) + ∂µ (F0µν + R ⋆ F0µν )A1ν ) + j ν A1ν ] d4 x
D 4π 4π
1 1
Z Z
= [− ∂µ F0µν + j ν ]A1ν d4 x + Nµ (F0µν A1ν + R ⋆ F0µν A1ν ) d3 x
D 4π 4π ∂D
where in the last line we have used that ∂µ ⋆ F0µν = 21 ∂µ εµνσρ Fσρ = ∂µ εµνσρ ∂σ Aρ = 0 . Thus,
for the field A0µ to be an action extrema under arbitrary variations of compact support we
need,
∂µ F0µν = 4πj ν
and so Maxwell’s equations are satisfied.
2
Again, mathematically, this kind of variations are difficult to define, the solutions do not decay as fast
as we would like to at infinity and so the finiteness of the integral is delicate, furthermore the set of field
solutions which vanish at the extreme surfaces does not have a nice topology.
248 CHAPTER 17. VARIATIONAL FORMALISMS
Chapter 18
Plane Waves
d d ′
∂σ Fµν = ∂σ s Fµν = kσ Fµν := kσ Fµν
ds ds
where we have denoted, for brevity, by a ′ the derivative with respect to s.
The vacuum Maxwell’s equations become,
∂ µ Fµν = 0 ⇒ k µ Fµν′
=0
′
∂[σ Fµν] = 0 ⇒ k[σ Fµν] = 0
′
Fµν = 2k[µ uν] (18.1)
249
250 CHAPTER 18. PLANE WAVES
Notice that the vector uν is not completely determined, for we can add to it any other
vector proportional to kν and still obtain the same Fµν . But obviously can not be just
proportional to kν , otherwise we will get the trivial solution.
The first equation now says,
k µ kµ = 0 and k µ uµ = 0
the first says that kµ is a null vector, the second that uµ is perpendicular to kµ . Notice that
we still have the freedom of adding to u a vector in the k direction for k is null.
Thus, we see that the general solution depends on two vectors, the vector k determining
the plane, and we see that the plane has to be null, and another vector, u(s) perpendicular
to it and which is just defined up to a multiple of k. So there are just two free functions left.
We can represent the solutions in the following figure, 18.1 the tensor F is constant along
the null planes and changes along the s function s from plane to plane, representing waves
which move to the right in the k direction.
s=2
s=1
s=0
To obtain the tensor F we must now integrate u along s to obtain another vector, v,
which will also be perpendicular to k (appropriately choosing the integrations constants) and
defined up to a multiple of k. So we will have,
k
t
s=2
s=1
s=0
Fµν ⋆ F µν = 2εµνσρ kµ vν kσ vρ = 0,
due to the total antisymmetry of the Levi-Civita tensor. So we see that these solutions are
very special indeed.
It is interesting to compute the energy-momentum tensor of this solution,
−1 µ σ −1 µ ν σ
T µν = (k v − k σ v µ )(k ν vσ − kσ v ν ) = k k v vσ . (18.3)
2π 2π
Thus, the momentum any observer with 4-velocity u would assign to this solution is,
−1 σ
pµ = v vσ k ν uν k µ .
2π
So it is a null vector, pµ pµ = 0. Furthermore, notice that v σ vσ ≤ 0, vanishing only if v is
proportional to k in which case F vanishes. So the waves always carry energy-momentum,
and it is in the form of a null four-vector. If k is future directed, then we say the plane wave
is moving into the future.
Energy flux, which is proportional to Poynting vector, is given by the space component of
µ
p , namely, writing kµ = ω(uµ + k̂µ )
ω σ
pµ⊥ = v vσ k ν uν k̂ µ
2π
Exercise: Check that v σ vσ ≤ 0 by choosing a Cartesian base so that k µ = ω(t̂µ + x̂µ ) and so
we have, t̂µ vµ = −x̂µ vµ . Write now the metric in this base and compute v σ vσ .
A′ν = vν .
18.4. HELICITY 253
That is the vector potential is a first integral of the vector v. Since v is perpendicular to k
we can choose the integration constants so that A is also perpendicular to it,
∂µ Aµ = kµ v µ = 0.
So this 4-potential is in Lorentz gauge! The freedom we have in adding to v a vector propor-
tional to k is just the freedom to add to A the differential of a function which depends only
on s, and those are just (plane) solutions to the wave equation, so it is just the freedom one
has on the choice of Lorentz gauge 1
18.4 Helicity
We have seen that plane waves are defined by two vectors, k and v, the individual size of
them is irrelevant. If we scale one and scale the other with the inverse factor the Maxwell
tensor they define will not change. Both lie at the symmetry plane. The vector k determines
the symmetry plane of the wave and the propagation direction, so its direction is well-defined.
On the contrary the direction of v is not, for we can change it by a multiple of k and still
would produce the same F . Thus, the object it defines is not a direction, but rather a plane,
the 2-plane perpendicular to k of vectors of the form αv + βk, for some vectors v and k. This
plane is called the helicity plane of the wave.
t
k ũ
u
αk
1
One can of course choose other gauge functions solutions to the wave equation which do not satisfy the
plane symmetry, but we are restricting attention here only to the ones that do.
254 CHAPTER 18. PLANE WAVES
Fµν ⋆ F µν = 2E µ Bµ = 0,
so the vectors are perpendicular to each other. Furthermore, since F and ⋆ F are per-
pendicular to k, both electric and magnetic vectors are also perpendicular to it. That is
Eµ k µ = (Fµν tν )k µ = (Fµν k µ )tν = 0, the same for Bµ using ⋆ F .
Choosing an observer, t we have,
kµ = ω(tµ + k̂µ )
with k̂ perpendicular to t. This is the propagation direction this observer assigns to the wave
in his simultaneity surface. Since E~ and B~ will be perpendicular to both, t and k they will
be also perpendicular to k̂. Given any vector ṽ we can find another,
(ṽ · t)
v = ṽ − k ,
k·t
which belongs to same equivalent class and furthermore is perpendicular to t. This vector is
~ indeed,
proportional to E,
~
B
~
E
the normal magnetic field must both vanish. Since we can take any time direction to split
Maxwell’s tensor into its electric and magnetic parts there must be a covariant equation to
express this boundary condition. Indeed, choosing any time direction perpendicular to n̂, one
can see that the condition is,
Exercise: Using 15.15, show that the above condition means that the energy flux across the
superconductor vanishes for all energy densities defined with respect to observers tangent to
it. That is,
T µν tµ nν = 0 with tµ nµ = 0.
Exercise: Show that there is no momentum flux for the components normal to the boundary.
Compute the momentum flux for the normal component. Does it has a definite sign?
Exercise: Show that if a superconductor body has axial symmetry then there is no angular
momentum flux for the axial component of it.
0 = η(kin , kin )
= η(kin in in in
|| , k|| ) + η(k⊥ , k⊥ )
= η(kout out in in
|| , k|| ) + η(k⊥ , k⊥ )
= −η(kout out in in
⊥ , k⊥ ) + η(k⊥ , k⊥ ).
If they coincide then both wave vectors are the same, and we know there is no solution unless
the wave is tangent to the superconductor. So the general solution must have two wave vector
whose tangent components coincide and whose normal components are opposite.
kin out
⊥ = −k⊥ , (18.8)
18.6. COVARIANT BOUNDARY CONDITIONS FOR SUPERCONDUCTORS AND TOTAL REFLE
where from now on the subscript ⊥ means the part of the vector perpendicular to the hyper-
surface, that is, proportional to n. Therefore,
kout = kin in
|| − k⊥ (18.9)
Exercise: Choose a time direction and coordinates adapted to it and to the hyper-surface
and recuperate Snell’s reflection law.
Exercise: Show that the planes generated by n, and kin , coincides with the one generated
by n, and kout , that is, span{n, kin } = span{n, kout }. Once an observer is chosen, the
intersection of its simultaneity hyper-surface with this plane is a 2-plane usually called the
incidence plane.
kin kout
uout
n
uin
uin out
|| = −u||
258 CHAPTER 18. PLANE WAVES
0 = η(kin , uin )
= η(kin in in in
|| , u|| ) + η(k⊥ , u⊥ )
= −η(kin out in in
|| , u|| ) + η(k⊥ , u⊥ )
where in the third line we have used the above condition, in the fourth the relation between kin ||
and kout
|| , in the fifth other perpendicularity condition η(k out
, u out
) = 0, and in the sixth the
relation between kin out
⊥ and k⊥ . Since we are assuming k not to be in the plane perpendicular
to n, this implies that uout in
⊥ = u⊥ . We see then that the relation between the incoming and
outgoing field strength is:
Exercise: Check that the above condition is invariant under the addition to uin of a vector
proportional to kin and a corresponding vector kout to uout .
Exercise: Use the boundary condition, 18.4, and the expression of the energy momentum
tensor in terms of ⋆ F to compute the momentum flux across the conductor boundary. Is
there any net momentum transfer for the normal component of it? Does it has a definite
sign? Then imagine the conductor is a finite body with finite mass, if the wave transfers to
it a finite momentum, how is that this is compatible with the fact that there is initially no
energy flux across it?
Exercise: Choose an observer and adapted coordinates and express the above result in terms
~ and B
of the incoming E ~ fields.
v(s) = v 0 eis
18.7. MONOCHROMATIC WAVES AND FOURIER DECOMPOSITION 259
Since s = kµ xµ +s0 we see that an observer for which kµ = ω(tµ + k̂µ ) will see a wave oscillating
with frequency ω = η(k, t). When considering plane waves moving along different directions
it is customary then to label the coefficients of them with the wave vector k,
v(s) = v k eis
and in general consider them as complex, so after a sum over different modes one takes the
real part of it.
We shall show now that any vacuum solution to Maxwell’s equations can be written as a
liner combination of monochromatic plane waves. To see this we shall pick a time direction t
and a Cartesian coordinate system compatible with it.
We recall that any smooth function f in R3 which suitable decay at infinity can be de-
composed in Fourier modes, namely,
1 ~
Z
f (~x) = 3 fˆ(~k)eik ·~x d3 k
(2π) 2 R3
1 ~
Z
ˆ ~k) =
f( 3 f (~x)e−ik ·~x d3 x.
(2π) 2 R3
¯
Notice that if f (~x) is a real function then we have, fˆ(~k) = fˆ(−~k), thus a real function is
represented by a complex one, but within the class having the above property.
The same hold for vectors, as long as we write them in Cartesian coordinates, so that inte-
grals can be taken, that is so that we can add vectors at different points in space. Thus, given
a vector vµ (~x) we have its Fourier components, v̂µ (~k). Now, and arbitrary linear combination
of solutions to the wave equation will have an expression,
1 + + ~ i(|~ ~ − − ~ i(−|~ ~
Z
Fµν (xσ ) = 3 [2k[µ v̂ν] (k)e k |t+k ·~x) + 2k[µ v̂ν] (k)e k |t+k ·~x) ] d3 k
(2π) 2 R3
where we now label the waves by its 3-dimensional wave number vector ~k, and since there
are two null planes with the same 3-vector, namely kµ+ = (|~k|, ~k), and kµ− = (−|~k|, ~k), we
have taken care of that by labeling those contributions v̂µ+ (~k) and v̂µ− (~k). Notice that these
contributions must satisfy, v̂ ±µ (~k)kµ± = 0, and since they can be chosen such that they are
perpendicular to the time direction, they result to be perpendicular to ~k.
Consider now arbitrary solutions to vacuum Maxwell’s equations, if one assumes some
asymptotic and regularity conditions, then we claim all such solutions can be written in this
way, namely as linear combination of plane waves. To see this recall that given F at t = 0,
with the conditions assumed above there is a unique solution to vacuum Maxwell’s equations,
for they are determined by their initial data. So we assume Ej (~x, 0) and Flm (~x, 0) are given
and have their corresponding Fourier transform, Êj (~k), and F̂lm (~k). Using the expression
260 CHAPTER 18. PLANE WAVES
In particular,
1 kl
v̂j± = ( F̂lj ± Êj )
2|~k| |~k|
Therefore we get the whole solution in terms of an integral over plane waves.
For computing energy fluxes and other physical quantities some time it is useful to take
time averages of harmonic functions, in that case, we are given two functions A(t) = 21 (A0 eiωt +
Ā0 e−iωt ) and B(t) = 12 (B0 eiωt + B̄0 e−iωt ), then
1 T
Z
< A(t)B(t) > := limT →∞ A(t)B(t)dt
T 0
1 T 1
Z
= limT →∞ (A0 eiωt + Ā0 e−iωt )(B0 eiωt + B̄0 e−iωt )dt
T 0 4
1ZT 1
= limT →∞ (A0 B̄0 + Ā0 B0 + A0 B0 ei2ωt + Ā0 B̄0 e−i2ωt )dt
T 0 4
1
= ℜ(A0 B̄0 ). (18.12)
2
Exercise: Compute the average energy flux of a plane wave along an arbitrary time-like
plane.
Chapter 19
~
∂E ~ ∧B
= c∇ ~ (19.1)
∂t
~
∂B ~ ∧E
= −c∇ ~ (19.2)
∂t
~ ·E
∇ ~ =0 (19.3)
∇ ~ =0
~ ·B (19.4)
In the approximation where the conductor is perfect (something which depends on the
frequency of the waves, but that here we assume valid for all frequencies) we have that at the
boundary the electric field must be perpendicular to it,
~ ∂V = 0
n̂ ∧ E| (19.5)
this is the boundary condition we have. It is therefore simple to express the equations and
their solutions in terms of the electric field. To do that we take another time derivative to
the equation for the electric field, 19.1, and use 19.2 to get,
−1 2 i
∂ E + ∆E i = 0 (19.6)
c2 t
~ ~x) = F~ (~x), and ∂t E(t,
We add to it equation 19.3. Given as initial data, E(0, ~ ~x)|t=0 =
~ ∧ B(0,
c∇ ~ ~x) = c∇ ~ ∧ G(~
~ x), satisfying ∇
~ · F~ = ∇
~ ·G
~ = 0 and the boundary condition, 19.3,
there is a unique solution to the wave equation in [0, ∞) × V satisfying the constraints, and
the boundary condition for all times. Once the electric field is computed, the magnetic field
261
262 CHAPTER 19. RESONANT CAVITIES AND WAVE GUIDES
n̂
~ ~x) = E
E(t, ~ ω (~x)eiωt ~ ~x) = B
B(t, ~ ω (~x)eiωt (19.7)
We need to solve then the following eigenvalue-eigenfunctions system,
~ ∧E
∇ ~ ω = −iω B~ω (19.8)
c
~ ∧B
∇ ~ ω = iω E~ω (19.9)
c
~ ·E
∇ ~ω = 0 (19.10)
~ ·B
∇ ~ω = 0 (19.11)
~ ω |∂V
n̂ ∧ E = 0 (19.12)
19.1. RESONANT CAVITIES 263
Note that the divergence equations are consistent with the first two equations.
We now assert: There is a countable, and infinite set of real frequencies, {ωi }, i =
0..∞, for which this system of equations has solutions (E ~ li , B
~ li ), the li supra-index meaning
ωi ωi
that in general there will be two solutions for each ωi . Given an arbitrary initial data set
~ 0 (~x), B
(E ~ 0 (~x), satisfying ∇·E
~ 0 = 0, ∇·B
~ 0 = 0, and the boundary condition, 19.5, the resulting
solution can be expressed as a linear combination of the above monochromatic solutions.
This assertion is based in the following observations:
• The above system can be considered as a system in the Hilbert space of square integrable
fields with zero divergence (in the distributional sense),
~ B)
H = {(E, ~ ∈ L2 |∇
~ ·E
~ω = ∇
~ ·B
~ ω }.
~ ~
! !
~ B)
~ := 0 −i∇∧ E
A(E, ~ ~ .
i∇∧ 0 B
~
• In H the operator is elliptic. For plane solutions with dependence eik ·~x and ~k · E
~ =
~k · B
~ = 0 we have,
|A(E, ~ = |~k|2 [|E|
~ B)| ~ 2 + |B|
~ 2 ].
• With the above boundary condition the operator A is formally self adjoint in H.
~¯2 · ∇ ~¯2 · ∇
Z
~ 2, B
h(E ~ 2 ), A(E
~ 1, B
~ 1 )i = (−iE ~ ∧B
~ 1 + iB ~ ∧E
~ 1 ) dV
ZV
= ~1 · ∇
[−iB ~¯2 + iE
~ ∧E ~1 · ∇ ~¯2
~ ∧B
Σ
+ i∇ ~¯2 ∧ B
~ · (E ~1 + E ~¯2 )] dV
~1 ∧ B
Z
= ~ ∧B
[−i(∇ ~ 2) · E
~ 1 + i(∇
~ ∧E
~ 2) · B
~ 1 ] dV
Σ
~¯2 ∧ B ~¯2 ] dΣ
I
+ i n̂ · [E ~1 + E
~1 ∧ B
∂V
~ 2, B
= hA(E ~ 2 ), (E
~ 1, B
~ 1 )i. (19.13)
Alternatively we can take the curl of 19.8, and use 19.9, to get,
~ ∧ (∇
∇ ~ ∧E ~ ω ) = ω2E
~ (19.14)
~ ω |∂V = 0
n̂ ∧ E (19.15)
264 CHAPTER 19. RESONANT CAVITIES AND WAVE GUIDES
~ ω = c∇
Exercise: Check that using that iω B ~ ∧E
~ ω that the other boundary condition,
~ ω · n̂|∂V = 0 is also satisfied.
B
We shall not prove the above assertion, but show its implication in an example.
∂2X
2
= −kx2 X
∂x
∂2Y
= −ky2 Y
∂y 2
∂2Z
= −kz2 Z
∂z 2
and,
ω2
= kx2 + ky2 + kz2 (19.17)
c2
The solutions to the above system are, X(x) = A+ eikx x + A− e−ikx x , and similarly for the
other coordinates. In particular if we want, for instance, that X(0) = X(a) = 0, then the
solution is X(x) = A sin(kx x) with kx = πl a
.
We now impose the boundary conditions, Eω1 (x, 0, z) = Eω1 (x, b, z) = Eω1 (x, y, 0) = Eω1 (x, y, c) =
0, Eω2 (0, y, z) = Eω1 (a, y, z) = Eω1 (x, y, 0) = Eω1 (x, y, c) = 0, etc. We have then,
where for each solution we have, in principle, different wave vectors ~k = (kx , ky , kz ) each one
of them satisfying the above dispersion relation 19.16, and furthermore for Eω1 , ky = πn b
, and
kz = πmc
, etc.
It remains now to impose the divergence free condition,
~ 2
E ω = E cos(ky y) sin(kx x) sin(kz z)
(19.20)
E 3 cos(kz z) sin(kx x) sin(ky y)
with,
kx E 1 + ky E 2 + kz E 3 = 0, (19.21)
the last requirement from the divergence free condition, and
~k = ( πl , πn , πm ). (19.22)
a b c
Since s
l2 m2 n2
ωlnm = ±π + + 2,
a2 b2 c
we see that it is real and there are a countable infinite number of them. For each vector ~k lmn
there are at most two linearly independent solutions for we have also to satisfy 19.21. To do
that we can choose, for instance, any two vectors perpendicular to ~k lmn , except for the case
where one of the integers vanish, in that case the vector direction is completely determined.
For instance if l = 0 then the vector is:
+ −
~ ~x) = ++ iωlmn t −+ −iωlmn t ~ +− iωlmn t −− −iωlmn t ~
X
E(t, [(Clmn e + Clmn e )E lmn (~x) + (Clmn e + Clmn e )E lmn (~x)]
l,n,m
(19.24)
1 ~ + + iE
Actually is better to choose a complex vector, E ~ − for performing calculations.
266 CHAPTER 19. RESONANT CAVITIES AND WAVE GUIDES
Notice that these are all standing waves, namely waves which do not propagate, their nodes
remain at fixed points in space. Nevertheless, they can describe general solutions which can
travel along the cavity bouncing off their walls.
∆2 V = −k 2 V (19.26)
with,
ω2
k 2 :=− kz2 , (19.27)
c2
where ∆Σ := ∂x2 + ∂y2 is the two-dimensional Laplacian in the guide section, Σ. We recall
that not all components are independent since the divergence of the both fields must vanish,
and we have the evolution equations which give one of the field by time integration when the
other is known.
We shall consider forward going solutions, that is with the dependence e−i(ωt−kz z) , where
we are assuming sign(ω) = sign(kz ). The treatment for the others is completely analogous.
19.2. WAVE GUIDES 267
k̂
n̂
iω
∂y Ez − ikz Ey = Bx (19.28)
c
iω
−∂x Ez + ikz Ex = By (19.29)
c
iω
∂y Ex − ∂x Ey = Bz (19.30)
c
and
−iω
∂y Bz − ikz By = Ex (19.31)
c
−iω
−∂x Bz + ikz Bx = Ey (19.32)
c
−iω
∂y Bx − ∂x By = Ez . (19.33)
c
We further have the two constraint equations,
∂x Ex + ∂y Ey + ikz Ez = 0 (19.34)
∂x Bx + ∂y By + ikz Bz = 0 (19.35)
268 CHAPTER 19. RESONANT CAVITIES AND WAVE GUIDES
To study the different solutions it is helpful to classify them according as to whether they
have components along the symmetry direction, k̂.
ω2
kz2 = or k 2 = 0 (19.38)
c2
Thus these waves travel to the speed of light. Since we are assuming sign(kz ) = sign(ω),
we have, Bx = −Ey , and By = Ex , that is,
~ = k̂ ∧ E,
B ~ (19.39)
~ is completely determined once we know E.
and so B ~
The remaining two equations are:
∂y Ex − ∂x Ey = 0
∂x Ex + ∂y Ey = 0
and identical equations for B ~ which we shall not need. The first equation tell us that there
~ = −∇φ.
exists a scalar field φ(x, y) such that E ~ The second that it must satisfy,
∆Σ φ = 0, (19.40)
in the guide section. Since we are assuming the conductor is perfect at its boundary we must
have,
~ ∂Σ = (n̂ ∧ ∇φ)|
(n̂ ∧ E)| ~ ∂Σ = 0,
Thus, the boundary condition implies that on each connected component of the boundary,
configuration allows for these waves they propagate at the speed of light along the k̂ direction
with the electric and magnetic fields perpendicular to it and to each other.
Notice that the following condition on the magnetic field follows from the first, namely
~ · n̂)|Σ = 0.
(B
Indeed,
~ · n̂ = (k̂ ∧ E)
B ~ · n̂ = k̂ · (E
~ ∧ n̂). (19.42)
Exercise: Make a qualitative drawing of the fields for the wave guide in the right.
Exercise: Show that the magnetic field satisfy the remaining equations.
ω −ω ω2 k2
∂y Ez = [ikz + i ( )]Ey = ikz [1 − 2 2 ]Ey = Ey . (19.45)
c kz c c kz ikz
If k 2 = 0 then ∂y Ez = 0 and the boundary condition Ez |Σ = 0 implies Ez = 0 in the whole
section, and we are back to Case 1. So we consider k 2 6= 0. In this case then Ey is completely
determined by Ez ,
ikz
Ey = ∂y Ez (19.46)
k2
Similarly, we obtain,
ikz
Ex = ∂x Ez (19.47)
k2
So,
E~ Σ = ikz ∇
~ Σ Ez . (19.48)
k2
where the label Σ means projection into the two-dimensional section Σ.
270 CHAPTER 19. RESONANT CAVITIES AND WAVE GUIDES
∆Σ Ez = −k 2 Ez . (19.50)
One can check that this is equivalent to the divergence free equation, 19.34.
To see which is the appropriate boundary condition for this case, we compute,
~ ∂Σ = (n̂ ∧ (E
(n̂ ∧ E)| ~ Σ + k̂Ez ))|∂Σ
ikz ~
= (n̂ ∧ ( 2 ∇ Σ Ez + k̂Ez ))|∂Σ
k
ikz ~ Σ Ez )|∂Σ + (n̂ ∧ k̂)Ez |∂Σ
= (n̂ ∧ ∇ (19.51)
k2
each term points in a different direction, so they both must cancel. This is the case if we
require,
Ez |∂Σ = 0 (19.52)
then both conditions are satisfied. As before we also have,
B ~ Σ ) · n̂|∂Σ = ω k̂ · (E
~ · n̂|∂Σ = ω (k̂ ∧ E ~ ∧ n̂)|∂Σ . (19.53)
ckz ckz
~ ·E
Exercise: Check that if this equation is satisfied, then we also have, ∇ ~ =∇
~ ·B
~ = 0.
So we must solve now equation 19.50 subject to the boundary condition 19.52. We know
from the general theory of self-adjoint operators that this system has an infinite (countable)
number of eigenvalues-eigenfunctions {kn2 , Ezn } which form a complete orthonormal eigenfunc-
tion base.
Multiplying both sides of 19.50 by −Ez and integrating over σ we obtain,
Z Z
k 2 Ez2 d3~x = −Ez ∆Σ Ez d3~x
Σ Σ
Z Z
= ~ Σ Ez · ∇
∇ ~ Σ Ez d3~x − Ez ∂n Ez dΣ
ZΣ ∂Σ
= ~ Σ Ez · ∇
∇ ~ Σ Ez d3~x
Σ
≥ 0.
~ z=0
Thus, k 2 ≥ 0. The case k 2 = 0 reduces to the Case 1, for the above equation implies ∇E
2
and then the boundary condition implies Ez = 0. Thus, we take k > 0, thus k is real, and
we have traveling waves. We also have,
19.2. WAVE GUIDES 271
ω2
= k 2 + kz2 > kz2 , (19.54)
c2
so the phase speed of these waves, vp := kωz , is greater than c. Nevertheless, the group velocity,
dω 2
dkz
= c ωkz ≤ c. Since the eigenvalues are numerable, and they can at most accumulate at
infinity, there is a minimal one, kmin ≈ L1 where L is the maximal transversal length of Σ,
and therefore a minimal frequency, ωmin = kmin c bellow which there can not be transmission
of TM waves.
∆Σ Bz + k 2 Bz = 0 (19.55)
To complete the system we need to impose some boundary condition to the above equation.
Since for this case,
~ Σ = ikz ∇
B ~ Σ Bz .
k2
We have,
E ~ Σ ) = −ω (k̂ ∧ B),
~ = −ω (k̂ ∧ B ~ (19.56)
ckz ckz
we have,
~ ∂Σ = −ω n̂ ∧ (k̂ ∧ B)
n̂ ∧ E| ~
ckz
−ω ~ k̂ − (n̂ · k̂)B]|
~ ∂Σ
= [(n̂ · B)
ckz
∂n Bz |∂Σ = 0.
In this case we also have an infinite number of wave modes. They are obviously different from
those found on the other cases.
Exercise: By an argument similar to the one employed for the TM waves, see that k 2 ≥ 0.
272 CHAPTER 19. RESONANT CAVITIES AND WAVE GUIDES
~ T · k̂ =c ~ ~⋆ c ωkz ~ ωkz ~
hSi (E ∧ B ) · k̂ = |∇Σ Ez |2 = |∇Σ Ez |2 ,
8π 8π ck 4 8πk 4
and so the energy across a section Σ per unit time (averaged) is,
Z
< P >T = ~ TΣ
k̂ · hSi
Σ
ωkz Z ~
= |∇Σ Ez |2 dΣ
8πk 4 Σ
ωkz
Z Z
= [ Ēz n̂ · ~
∇ E
Σ z dl − Ēz ∆2 Ez dΣ]
8πk 4 Σ Σ
ωkz
Z
= Ēz Ez dΣ
8πk 2 Σ
~ = k̂ ∧ E
Answer: For a T wave, B ~ and E
~ =∇
~ Σ φ, therefore
~ T = c ~ 2
hSi |E| k̂,
8π
and
c
Z
< P >T = ~ 2 dΣ
|E|
8π ZΣ
c ~ Σ φ|2 dΣ
= |∇
8π ZΣ
c ~ dΣ
= φ̄ n̂ · E
8πZ Σ
c
= φ̄ σ dΣ
2 Σ
cX Z
= φ̄i σ dΣ
2 i Σi
cX
= φ̄i Qi
2 i
cX
= hφi(t) Qi (t)i
2 i
19.3. CAVITIES WITH CONSTANT SECTION 273
where Qi , and φi are respectively the total charges by unit length, and the potentials (con-
stants) on each of the conducting boundaries.
~ + (x, y)eiωt+kz z + E
~ ~x) = E
E(t, ~ − eiωt−kz z ,
~ Σ = 0, Bz = 0), would imply that,
and the boundary conditions at z = 0 and z = z0 , (E
~ + (x, y) + E
E ~ − (x, y) (z = 0) ~ + (x, y)eikz z0 + E
E ~ − (x, y)e−ikz z0 = 0 (z = z0 ).
kz = πn/z0 .
So, as expected, we get only discrete modes.
We can now analyze each case:
19.3.1 T modes
In that case Bz = 0 and so no extra condition is needed. The modes are given by the solution
to the 2-dimensional Laplace equation times the solutions just find above.
~ ~x)n = −∇φ(x,
~ i πnz
πcnt
E(t, y)e z0
)sin(
z0
Note that again in this case, the boundary to the cavity section must have at least two
disconnected components.
19.3.2 TE
In that case we still have Bz = 0, so we just solve for 19.50 with Dirichlet homogeneous
boundary conditions. The frequency of the resulting modes will be:
s
πn 2
ωni = c ( ) + ki2 ,
z0
and the mode, is built out of,
274 CHAPTER 19. RESONANT CAVITIES AND WAVE GUIDES
πnz
Ezni (t, ~x) = Ezi (x, y)eiωni t sin( ).
z0
19.3.3 TM
In this case we must make sure Ez vanishes at the boundary. We solve 19.2.3 with Neumann
homogeneous boundary conditions, since Bz = 0 2 at the extremes would mean EΣ = 0 there.
We again use the same kz as in the other cases and obtain the same form for the standing
waves.
Exercise: Check that all these modes are present are the same as those we found for the
case of the cubical resonant cavity.
2
One could consider Bz = cons at the extremes, but this just generates the trivial solution.
Chapter 20
~ x, t) = E
E(~ ~ 0 (~x)e−iωt ,
and consider long enough wave lengths so that the spatial dependence of the field can be
~ ~
ignored, ( |∇E | << |~xl |). Under these assumptions the solution is approximately given by:
~
E
−e E~ 0 (~x)e−iωt
~x(t)l = ℜ[ 2 ].
m ω0 − ω 2 − iωγ
275
276 CHAPTER 20. WAVE PROPAGATION IN CONTINUUM MEDIA
besides this solution there are homogeneous ones, but they decay exponentially to the equi-
librium position (~xl = 0) on times of the order of T = γ1 , and so, after that transient they do
not play any role.
Exercise: Find the homogeneous solutions and check the above assertion.
e2 ~ 0 e−iωt
E
p~(~x, t)l = −e~xl ≈ ,
m ω02 − ω 2 − iωγ
which oscillates to the same frequency as the electric field, but with a different phase and an
amplitude that depends on the frequency.
In fact, in Fourier–Laplace space, summing over all contributions,
e2 N 2
~ ω (~x) = ~ ω (~x),
~pω (~x) = χe (ω)E [ω − ω 2 − iωγ]−1 E
m 0
where N is the particle number density. Thus,
~ ω (~x) = ε(ω)E
D ~ ω (~x),
with,
4πe2 N 2
ε(ω) = 1 + 4πχe (ω) = 1 + [ω0 − ω 2 − iωγ]−1 .
m
We have then:
X e2j Nj 1 X 4πe2j Nj 1
χe (ω) = 2 2
, ε(ω) = 1 + 2 2
,
j mj ωj − ω − iγj ω j mj ωj − ω − iγj ω
where Nj is the number density of dipole sites times the number of charged particles
per site.
~
∂D ~
∂E
6= ε .
∂t ∂t
Indeed,
~ ~x) := √1
∞ Z
D(t, ε(ω)E ~ ω (~x)e−iωt dω
2π −∞
1 Z∞Z∞Z∞ ~ ω (~x)e−iωt dωdω̃dt̃
= ε(ω̃)e−i(ω̃−ω)t̃ E
2π −∞ −∞ −∞
1 ∞ ∞ ∞
Z Z Z
= ~ ω (~x)e−iω(t−t̃) dωdω̃dt̃
ε(ω̃)e−iω̃t̃ E
2π −∞ −∞ −∞
Z ∞
= ε(t̃)E~ ω (~x, t − t̃)dt̃
−∞
1 ∞
Z
√ e−i(ω̃−ω)t̃ dt̃ = δ(ω̃ − ω).
2π −∞
Thus,
~
∂D
Z ∞ ~ ω (~x, t − t̃)
∂E ~
∂E
= ε(t̃) dt̃ 6= ε ,
∂t −∞ ∂t ∂t
unless ε(t) = ε0 δ(t). Which in general, as in our model, it is not the case. This implies
that in general the space-time equations describing these phenomena are no longer
partial differential equations, and are, at best, integro-differential ones.
5. Dielectrics: If ωj 6= 0 ∀j, so that there are no free electrons, then at low frequencies,
X e2j Nj 1
ε(ω) = 1 + 4π := ε̃,
j mj ωj2
is real, and we have the dielectric model already studied in electrostatics.
i4πe2 N 1
ε(ω) = ε̃ + ,
m ω(γ0 − iω)
where in ε̃ we have included the contributions due to the other dipoles with ωj 6= 0.
~ ω = εω E
This term can be related to the media conductivity. Indeed, substituting D ~ω
in,
278 CHAPTER 20. WAVE PROPAGATION IN CONTINUUM MEDIA
∇ ~ ω = − iω D
~ ∧B ~ ω.
c
(where we have assumed no free current), we get,
2 2
~ ∧B
∇ ~ ω = −[ iω ε̃ − 4πe N ]E
~ ω = − iω [ε̃ + i4πe N ]E ~ ω + 4π J~l ,
~ ω = − iω ε̃E
c mωγ0 c mcγ0 c c
Where in the last step we have assumed Ohm’s law to hold at these frequencies,
J~l = σ0 E,
~
with
e2 N
σ0 = .
mγ0
So we see that we recuperate this way Maxwell’s equations at the given frequency. In
the low frequency limit, ε̃ does not depend on the frequency, and so we can recuperate
a partial differential equation, namely,
~ ~
~ ~x) = 1 ε̃ ∂ E (t, ~x) + 4π J~l (t, ~x) = 1 ∂ D (t, ~x) + 4π J~l (t, ~x).
~ ∧ B(t,
∇
c ∂t c c ∂t c
With the understanding that is only valid for low frequencies, ω << ωj .
ωp2
ε(ω) ≈ 1 −
ω2
with
4πe2 N
ωp2 = ,
m
where N is the total electron number density, usually N = NA Z (= # atoms ×
atomic # per unit volume).
For dielectrics, in general ωp ≤ ωj and so the above formula only says that
ε(ω) ≤ 1,
for it only holds when ω >> ωj ≥ ωp .
But for an electron gas (plasma) or in a metal, which in most aspects behaves like an
electron gas, the above expression is valid for ω ≈ ωp , and so ε(ω) ≈ 0. We shall see
later some consequences of this.
20.2. PLANE WAVES IN HOMOGENEOUS AND ISOTROPIC MEDIA. 279
~ ω e−i(ωt−~k ·~x) )
~ ~x) = ℜ(E
E(t,
~ ω e−i(ωt−~k ·~x) ),
~ ~x) = ℜ(B
B(t,
that is plane waves traveling along the ~k direction. The phase speed of these waves would be,
(keeping the phase, ωt − ~k · ~x constant),
ω ω
|~v| = = . (20.1)
|~k| k
Inserting this ansatz in Maxwell’s equations we get,
−iωµε ~
Eω = i~k ∧ B
~ω (20.2)
c
−iω ~
Bω = −i~k ∧ E
~ω (20.3)
c
~k · E
~ω = 0 (20.4)
~k · B
~ω = 0 (20.5)
Notice that when ω 6= 0 the last two equations are a consequence of the first two.
This is an eigenvalue-eigenvector problem identical to the one studied for the case of
monochromatic modes in resonant cavities. To obtain the eigenvectors we proceed in a similar
way, taking the vector product of the first equation with ~k we get,
−iωµε ~ ~
k ∧ E ω = i~k ∧ (~k ∧ B
~ ω ) = i(~k · B
~ ω )~k − (~k · ~k)B
~ ω, (20.6)
c
using now the second equation on the left and the fourth equation on right we get,
−ω 2 µε ~
B ω = −(~k · ~k)B
~ ω, (20.7)
c2
from which it follows that
ω 2µε ~ ~
=k·k (20.8)
c2
280 CHAPTER 20. WAVE PROPAGATION IN CONTINUUM MEDIA
ω c
|~v | =
= . (20.10)
k n
Although this speed can be larger than the speed of light, it is just the phase speed and
not the real propagation speeds of perturbations of compact support, which is called
the group velocity and it is given by,
dω
vg = , (20.11)
dk
and this is in general smaller or equal to c. We shall prove in the next chapter that
under very general conditions in ε(ω) all solutions propagate at speeds limited by the
vacuum speed of light.
2. Since ε can be in general complex (or also real but negative), n, and therefore ~k, are
complex. So care must be exercised when asymptotic conditions are imposed, since
there could be asymptotically growing modes.
Indeed, considering for simplicity, the case µ = 1, we see that there are three possibilities:
a) ε real, ε > 0. In this case n > 0 and so ~k can be taken to be real. In our simple
matter model this corresponds to systems with no dissipation, γ = 0, and indeed
the resulting wave propagates with no dissipation along any arbitrary k̂ direction.
We call this a transparent medium.
b) ε real, ε < 0. In this case n is purely imaginary, n = iκ, κ real. In general in this
case ~k is a complex vector, but there are solutions where ~k = iκk̂, with k̂ a real
unit vector. For instance the vector ~k = (1, 2i, 0), has ~k · ~k = −3.
c) ε complex, n = ñ + iκ, and so ~k is complex, as in the former case there are
solutions with ~k = (ñ + iκ)k̂, with k̂ a real unit vector. But there are many others.
These other solutions are not plane waves, for there is no real plane perpendicular
to a generic complex vector, so the generic solution depends on three variables,
the time and two other spatial directions. There are cases where these solutions
must be considered. In the case a real direction can be defined, namely when
~k = (ñ + iκ)k̂, the quantity ñ is also called the refractive index of the medium,
and κ the absorption coefficient .
3. Notice that already in cases b, and c, above the solutions grow exponentially in one
direction and decay exponentially in the opposite. In the case ~k = (ñ + iκ)k̂, with k̂ a
20.2. PLANE WAVES IN HOMOGENEOUS AND ISOTROPIC MEDIA. 281
real unit vector, if sign(ñ) = sign(κ), say positive, then we have a wave traveling along
k̂ and decaying along the same direction. This is acceptable because it means there is
absorption from the media, but if sign(κ) is the opposite to sign(ñ), then we should be
having a medium producing the wave and this one growing exponentially as it travels.
This is not physically acceptable, so we should always have, ℑε ≥ 0.
Indeed, if
ε = |ε|eiθ , (20.12)
then
q
n = ± |ε|eiθ/2 , (20.13)
and so
π
2
n
ε
0
π
ℜ
n
3π
2
~ ω = − c ~k ∧ E
4. Since B ~ ω , and ~k · E
~ ω = 0, we have,
ω
~⋆ = − c E
~ω ∧ B
E ~ ⋆ ) = c [−(E
~ ω ∧ (~k ⋆ ∧ E ~ ω · ~k ⋆ )E
~ ⋆ + ~k ⋆ E ~ ⋆ ].
~ω · E
ω ω ω ω
ω ω
Thus, the time averaged Poynting vector is given by,
~ c ~ ~ ⋆ c2 1 ~ ⋆ ~ ~ ⋆ ~ ~ ω · (~k − ~k ⋆ ))E
~ ⋆ } + ℜ{~k}|E
~ ω |2 ]
hSi = ℜ{E ω ∧ B ω } = [ {(E ω · (k − k ))E ω − (E ω
8π 8πω 2
282 CHAPTER 20. WAVE PROPAGATION IN CONTINUUM MEDIA
⋆
In particular, along the real direction ~l := −i(~k − ~k ) we have,
2
~ · ~l = c (ℜ{~k} · ~l)|E
hSi ~ ω |2
8πω
Thus, we see that the standard expression for Poynting’s vector is only valid for trans-
parent media, κ = 0.
Exercise: Check that our model satisfies the above condition, 20.15.
Exercise: Check that the group velocity in the high frequency limit is smaller than c.
Where[·] indicates the field difference across the interface, and n2 := εµ. So we see that in
this case the only lack of continuity is in the normal component of the electric field.
20.3. REFLECTION AND REFRACTION ACROSS INTERFACES, THE PERPENDICULAR CASE.
We shall choose coordinates so that the plane z = 0 is the interface plane, to the left we
shall have a medium with refraction index n, and an incident wave on it. To the right we
shall denote the refraction index by n′ .
Since the jumps are fixed, we must have that the waves have all the same time dependence,
so the ω’s are all the same. Thus, we expect to have three waves, two on the side (left) where
the wave is coming, (the incoming one and the reflected one), and one on the side the wave
is leaving the interface.
~ I e−i(ωt−kI z) + E
~ R e−i(ωt−kR z) z < 0
(
~ ~x) = E
E(t, ~ T e−i(ωt−kT z) (20.20)
E z>0
From the conditions,
ω 2 n2 ω 2 n′2 n′2 2
kI2 = kR
2
= and k 2
T = = k
c2 c2 n2 I
where n, (n′ ) are the left, respectively right, refraction indexes. From this we see we can take
′
kR = −kI (the other sign would just produce two waves in the same direction), and kT = nn kI .
We must see now what are the relative field strengths. From the above boundary conditions
and equations 20.2-20.5 we get the following relations:
~T = 2n ~
E EI (20.23)
n + n′
~R n − n′ ~
E = EI (20.24)
n + n′
For transparent media (ε real and positive), we have three plane waves without any dis-
persion nor dissipation and without any phase difference. In that case the energy fluxes of the
incoming wave minus the one of the reflected wave should equal the one of the transmitted
~ = −n~kˆ ∧ E
wave, indeed, recalling that B ~ we have,
hSlef t iT · n̂ = (hS~ I iT − hS ~ R iT ) · n̂
cℜn ~ 2 ~ R |2 }
= {||E I | − |E
8π
c |n − n′ |2 ~ 2 c 4ℜn|n||n′ | ~ 2
= ℜn(1 − )| E I | = |E I |
8π |n + n′ |2 8π |n + n′ |2
while,
′ 2 ′
~ T iT · n̂ = cℜn |E
hSright iT · n̂ = hS ~ T |2 = c 4|n| ℜn |E ~ I |2
8π 8π |n + n′ |2
284 CHAPTER 20. WAVE PROPAGATION IN CONTINUUM MEDIA
Exercise: Assume the left side medium is transparent while the one at the right is dissipative.
Should the energy fluxes also coincide at the interface? Compute the energy flux as a function
of the depth in the right media.
Exercise: If the incident media is not transparent both expressions do not coincide. Nev-
ertheless energy conservation must hold, so there is an assumption which is invalid. Which
one?
Exercise: Compute Poynting’s vector for the total field on the left (incident plus reflected)
and see under which conditions holds that it is the sum of individual Poynting’s vector for
each of the two fields.
~ ω e−i(ωt−~k ·~x) )
~ ~x) = ℜ(E
E(t,
~ ω e−i(ωt−~k ·~x) ),
~ ~x) = ℜ(B
B(t,
~~ 2
with ω 2 = k ·nk2c , and ~k · E
~ ω = 0. For some cases we shall allow ~k to be complex, and not only
′ ′
of the form ~k = a~k with a complex and ~k real. When ~k does not define a real direction the
solutions are strictly speaking no longer plane waves, for they do not have a planar symmetry
in real space.
We shall choose coordinates so that the plane z = 0 is the interface plane, to the left we
shall have a medium with refraction index n, and an incident wave on it. To the right we
shall denote the refraction index by n′ . Let us assume that the incident wave is a plane wave
with wave number, ~k I , real. Let us denote by n̂ the unit normal to the discontinuity surface.
If n̂ is parallel to ~k I we are in the case already considered, so we assume they are not parallel.
Thus, both define a plane, called the incidence plane, in our figure (20.1), the (x, z) plane.
As before we shall propose an ansatz consisting in a transmitted wave, with wave number
vector ~k T , and a reflected wave, with wave number vector ~k R . The direction perpendicular to
the incident plane is common in both sides, and so will remain a symmetry direction, in our
scheme that coincides with the ŷ direction, so we expect the wave vectors will not have any
component along it, kIy = kT y = kRy = 0.
20.4. REFLECTION AND REFRACTION, THE GENERAL CASE. 285
kT kI
n
kI
z
~kT
θT
~kR
θI θR
~k I
In order to be able to satisfy the jump conditions at the interface we need the waves at both
sides to have the same functional dependence both in time and in the directions tangent to
the interface, that is at the plane z = 0. This means that ωI = ωR = ωT and kIx = kT x = kRx .
Furthermore, we have that
2 2 ′2 2
~k I · ~k I = ~k R · ~k R = n ω ~k T · ~k T = n ω . (20.25)
c2 c2
We shall take kIz = −kRz for we know we need a reflected wave (traveling to the left) to have
a consistent solution.
′2 2 ′2 2 2 ′2
Since kT2 = n c2ω = nn2 n c2ω = nn2 kI2 , in the case the right medium is transparent, we have
kT x kIx n kIx n
sin θT :== = ′ := ′ sin θI , (20.29)
kT kT n kI n
which in optics is known as Snell’s law. Notice also that,
n′2 2 2 n
′2
kT2 z
= kT2
− = kT2 x
− kT2 2
= 2 kI − kIx = kI ( 2 − sin2 θI ).
kIx 2
(20.30)
n n
Thus, if n′ < n, both real, then there are incidence angles for which kT z vanishes or becomes
purely imaginary. We call the limiting angle, for which kT z = 0, the total reflection angle, θr ,
it is defined by the relation,
n′
sin θr =
.
n
For those angles with θI > θr the field decays exponentially inside the medium on the right,
and so we have total reflection. On the above equation we must take the positive root, that
is ℜkT z ≥ 0, in order to have a wave traveling to the right. If ℜkT z = 0, that is when n′ < n
′2
both real, and sin2 θI > nn2 , then one must choose, for consistency, the root with ℑkT z ≥ 0,
so that the wave decays to the right. If both ℜkT z 6= 0, and ℑkT z 6= 0 then we can only make
one choice and check for the other, that is, if we choose the root with ℜkT z ≥ 0, then we must
check that ℑkT z ≥ 0. If this does not happen, that is if ℑ( nn2 − sin2 θI ) = ℑ εε < 0, then we
′2 ′
We now look for the relative field strengths. We consider two separate cases, in the first
we shall assume the electric field is perpendicular to the incidence plane, that is in the ŷ
direction, in the second that it is in the incidence plane, the general case can be obtained by
decomposing the general wave into these two directions.
Where we have only considered the x component of equation 20.3, since the z component is
automatically satisfied from the continuity of the electric field and the tangential component
of the wave vector. As before we obtain,
that is, Fresnel’s law. The second formulas are valid when the incident medium is transparent.
The third formulas, in terms of incidence and transmission angles, are valid only when both
media are transparent.
0
5
-1
-4
-2
0
2 10
4
Figure 20.4: EI and ET for the case where the electric field is tangent to the interface. Case
n′
n
< 1.
288 CHAPTER 20. WAVE PROPAGATION IN CONTINUUM MEDIA
0
10
1.0
0.5
0.0
-4 -0.5
-2
0
2 -1.0
4
Figure 20.5: EI and ET for the case where the electric field is tangent to the interface. Case
n′
n
> 1.
The other components of the electric field, namely those at the incidence plane, are zero.
Exercise: Consider the equations for these other components, show that they form a linear
homogeneous system and compute its determinant.
Exercise: Compute the total averaged Poynting vector along the interface normal, both, on
the left and on the right of the interface, and show that they are given by:
2
~ >T = 1 ℜkT z |kIz | |EI |2
n̂· < S
2π |kI + kT |2
Note: On each side compute the Poynting vector corresponding to the total electromagnetic
field.
~ ~
~ = −k ∧ B .
E
k2
~ ~
Thus, we expect to have the same formulas, but replacing ~k I by kn2I , and ~k T by kn′2T .
Thus, we get,
20.4. REFLECTION AND REFRACTION, THE GENERAL CASE. 289
′
BT y 2n′2 kIz ( nn )2 cos θI sin 2θI
= = 2 =
n kIz + n kT z sin (θI + θT ) cos (θI − θT )
q
BIy ′2 2 ′ ′
( nn )2 cos θI + ( nn )2 − sin2 θI
′
q
′
BRy n′2 kIz − n2 kT z ( nn )2 cos θI − ( nn )2 − sin2 θI tan (θI − θT )
= = = , (20.34)
n kIz + n kT z tan (θI + θT )
q
BIy ′2 2 ′ ′
( nn )2 cos θI + ( nn )2 − sin2 θI
As above, the second formulas are valid when the incidence medium is transparent, the third,
in terms of incidence and transmission angles, are valid only when both media are transparent.
cos θR |E ~⋆ |
~ R ||B ~ R |2
n cos θR |E ~ R |2
|E
R
R := ⋆ = =
~ I ||B
cos θI |E ~ | ~ I |2
n cos θI |E ~ I |2
|E
I
For transparent media they become,
sin2 (θT − θR )
R⊥ = (20.35)
sin2 (θT + θI )
tan2 (θT − θR )
R|| = (20.36)
tan2 (θT + θI )
Exercise: From this coefficients, and energy conservation find a bound for the power of the
transmitted power as a function of the incident power.
n′
!
= tan θI
n
Thus, there is always an incidence angle, called the Brewster angle ,
290 CHAPTER 20. WAVE PROPAGATION IN CONTINUUM MEDIA
n′
!
θb := arctan
n
such that the reflected wave is totally polarized in the direction tangent to the interface.
From the above expression we see that this incidence angle satisfies,
n
cos θb = sin θb = sin θT
n′
where we have used 20.29, and where θT is the transmission angle corresponding to the inci-
dence angle θb . Therefore, we see that θT = π2 − θb so the incidence direction is perpendicular
to the transmission direction. Notice that in this case the denominator in 20.36 for R|| tends
to infinity. So this coefficient indeed vanishes.
a − ib
, a and b real
a + ib
so they absolute value is unity. Thus, R⊥ = R|| = 1. Energy conservation then implies that
the transmitted flux in the incidence direction vanishes. This does not mean that the fields
inside vanish, they decay inside, but their Poynting vector is tangent to the interface. Note
that although the magnitude of the reflected field coincide with that of the incidence one,
there is a phase change between them. It is given by,
a − ib δ b
e−iδ = , ⇐⇒ tan =
a + ib 2 a
Exercise: Find the expression for the phase difference between the parallel and perpendicular
cases.
The general behavior of the reflection coefficients can be seen on the following figures
[20.6-20.7]. Notice that R|| ≤ R⊥ so that in general the reflected light is (partially) polarized
in the normal direction to the incidence angle.
0.5
0.4
0.3
0.2
0.1
Figure 20.6: Reflection coefficient as a function of the incidence angle for the case n′ > n.
1.0
0.8
0.6
0.4
0.2
Figure 20.7: Reflection coefficient as a function of the incidence for the case n′ < n.
292 CHAPTER 20. WAVE PROPAGATION IN CONTINUUM MEDIA
i4πσ
ε′ ≈ , |ε| >> 1,
ω
and so,
s
2πσ
n′ ≈ (1 + i),
ω
We shall now study the behavior of a plane wave traveling in a non-absorbing media when it
reaches a metallic surface. In this case we have two length scales: λ = 2π k
of a wave incident
in the metal surface, and the length scale built out of the metal resistivity, λσ = σc . We shall
consider situations in which λ >> λσ . In this case we have that,
c 1 c√ 1q √
r
δ := √ = λn = δσ λn << λ
2πσω 2π σ 2π
where n is the index of the transparent medium. Thus, using the continuity of the tangential
component of the wave vector at the media/metal
√
interface we conclude that inside the metal
we will√have a wave vector given by ~k ≈ c (1 + i)k̂ and so will decay along that direction
2πσω
2πσω
like e− c z . That is, at a distance, δ the field decreases to half its value. The fact that
we have decay does not mean there is absorption, it could mean reflection, or in general,
scattering. The skin effect is the fact that the electric fields penetrate a metal a length δ
called the metal skin thickness.
We have seen that with the conditions above the derivatives which are really important
in Maxwell’s equations inside a metal are just the normal derivatives. Thus, inside the metal
the solution would be approximately a plane wave perpendicular to the metal surface. That
is, defining ζ = n1′ ,
s
~T = µ′ ~ ~ T ∧ k̂.
E H T ∧ k̂ := ζ H (20.37)
ε′
For the case we are dealing with,
s √
ωµ′ 1 − i δσ
ζ = ζr + iζi = (1 − i) = √ |ζ| << 1,
8πσ 2 λn
where we are considering µ and σ (the resistivity) to be real.
Expression 20.37 can be considered as a boundary condition to the external fields. This is
so because they are tangential components of the fields and so are continuous across the metal
surface. Indeed, for the expert, this expression can be considered as a mixture of outgoing
and incoming modes for Maxwell’s equations. The condition is stable, and so allowed, if
the energy leaks from the interface towards the inside of the metal, that is, if the normal
component of Poynting vector is positive,
c
~ · k̂i := ~ ⋆ ) = c ζr |H
~T ∧ H ~ T |2 ≥ 0.
hS ℜ(E T
8π 8π
So 20.37 is a good boundary condition provided that,
20.6. ENERGY LOSSES IN CAVITIES AND GUIDES 293
0
10
1.0
0.5
0.0
-4 -0.5
-2
0
2 -1.0
4
Figure 20.8: EI and ET for a metal in the case where the incident electric field is tangent to
the interface.
ζr > 0.
Exercise: Find the frequency range (or/and wave length) for which cooper satisfies the
approximations assumed above. Compute δcooper .
dE
cζr
R
~ 2 dS
|H|
dt ∂V
= ,
E ~ 2 dV
2 V |H|
R
294 CHAPTER 20. WAVE PROPAGATION IN CONTINUUM MEDIA
where we have already used that the energy is equiparted among the electric and magnetic
fields. We are also assuming, for simplicity, that the interior of the cavity is empty. Using
(19.1, 19.2) we can transform the volume integral into a surface one, obtaining,
dE R
~ 2 dS
ωζr ∂V |H| −ω
dt
=R :=
E ~ 2 ~ 2 Q
∂V [|H| − |E| ]dV
The a-dimensional quantity Q is called the quality factor of the cavity, it depends on the
geometry of the cavity and on the conductivity of its walls. It counts the cavity’s attenuation
rate in periods of the mode that has been exited.
Similar expressions are valid for wave guides, in that case the relevant quantity is the
attenuation rate per unit energy flux across the guide section. In this case the final integral
is a line integral along the section contour.
Chapter 21
ωn(ω) kc
k= or ω(k) = ,
c n(k)
this last being the way we are going to use it. We shall assume n(ω) real, that is a transparent
medium, otherwise the wave dissipate and the effects get confused. This relation provides the
information from the evolution equation we are solving, and, in the case of electromagnetism,
information about the medium across which the wave moves. We shall consider now a wave
packet, that is a solution made out from contributions from many monochromatic plane waves.
For simplicity, we shall consider plane waves along a single direction, x̂, we shall comment on
the fact that when waves have finite extent this is no longer the case and further dispersion
occurs. A generic Cartesian component of the electric and magnetic field have then the form,
1 ∞
Z
u(t, ~x) = √ û(k)e−i(ω(k)t−kx) dk,
2π −∞
where û(k) is the number density of waves between wave number k and wave number k =
k + dk. For simplicity, we just integrate over the whole k-space although we are thinking in
adding contributions from a narrow frequency band, this will be evident from the form of
û(k), which we shall assume smooth and of compact support.
For Maxwell equations, and in fact for all second order systems, frequencies come into
pairs, or alternatively we can think that for each frequency there are two wave numbers for
each wave propagation direction,
±ωn(ω) ±kc
k= or ω(k) = ,
c n(k)
295
296CHAPTER 21. WAVE PACKETS, KRAMERS–KRONING RELATIONS AND CAUSALITY
1 Z∞
u(0, ~x) = √ {û(k)− + û(k)+ }eikx dk,
2π −∞
∂u(t, ~x) ic
Z ∞
|t=0 = √ k{−û(k)− + û(k)+ }eikx dk,
∂t 2π −∞
We shall consider now propagation of wave packets, that is where the initial data is of
compact support, initially spatially centered around a point, and such that they Fourier
coefficients are also centered around a given, dominant, frequency.
a2 k 2
ω(k) = ω0 + vk + ,
2
and take as initial data,
−x2
u(0, ~x) = e 2L2 cos(k0x)
∂u(t, ~x)
|t=0 = 0. (21.1)
∂t
So the wave has initially a large peak at x = 0 and oscillates at k = k0 . Since the time
derivative is zero we expect to have two wave packets moving initially in opposite directions,
indeed, since u(k)− = u(k)+ = u(k) we get,
1 ∞ −x2
Z
û(k) = √ e−ikx e 2L2 cos(k0 x) dx
2 2π −∞
1
Z ∞
−x2
= √ [e−i(k−k0 )x + e−i(k+k0 )x ]e 2L2 dx
4 2π −∞
1 Z ∞ −(k−k0 )2 L2 /2−[−i(k−k0 ) √L + √x ]2
= √ [e 2 L 2 + (k0 → −k0 )] dx (21.2)
4 2π −∞
where in the last line we have indicated the other term as obtained from the one written
by substituting k0 → −k0 . Taking as variable y := −i(k−k
√ 0 )L + √
2
x
L 2
, dy = √dx
2L
the integral
becomes,
√
2L ∞ −(k−k0 )2 L2 /2
Z
2 2 2
û(k) = √ [e + e−(k+k0 ) L /2 ]e−y dy
4 2π −∞
√ √
2L π −(k−k0 )2 L2 /2 2 2
= √ [e + e−(k+k0 ) L /2 ]
4 2π
L −(k−k0 )2 L2 /2 2 2
= [e + e−(k+k0 ) L /2 ] (21.3)
4
√
So û(k) has two peaks, one at k0 and the other at −k0 , both have with δk = L2 . We shall
compute the time evolution of this data. For simplicity, we shall concentrate in the term
corresponding to the peak at k = k0 , which we shall call u− (t, ~x), the other is obtained by
substituting k0 → −k0 on the expressions we shall derive.
We have,
L ∞ Z
2 2
u− (t, ~x) = √ e−(k−k0 ) L /2 e−i(ωt−kx) dk
4 2π −∞
L
Z ∞
a2 k 2 2 2
= √ e−i[(ω0 +vk+ 2 )t−kx]−(k−k0 ) L /2 dk
4 2π −∞
to compute this integral we also use the same trick as before and complete squares. Calling
R the exponent inside the integral we have,
298CHAPTER 21. WAVE PACKETS, KRAMERS–KRONING RELATIONS AND CAUSALITY
B2 B √
R = A + 2kB − k 2 C = (A + ) − ( √ + Ck)2
C C
which we use with
−i L2 ia2 t
A = −iω0 t − k02 L2 /2 B= (vt − x) + k0 L2 ) C= (1 + 2 ),
2 2 L
B
√ √
Defining l := √C + Ck, dl = Cdk, we get,
L 1 ∞
Z
2 2
u− (t, ~x) = √ √ eA+B /C e−l dl
4 2π C −∞
√
L π 2
= √ √ eA+B /C
4 2π C
(vt+a2 k0 t−x)2
− 4 2
2L2 (1+ a t )
e L4
= q e−iγ
ia2 t
1+ L2
vg t−x 2
−( 2L(t) )
e
= q e−iγ
ia2 t
1+ L2
(21.4)
q
dω a4 t2
where we have defined vg := |
dk k=k0
:= v + a2 k0 , L(t) := L 1 + L4
, and
a2 t 2a2 t2 ω0 (tv−x)2
(v + ω0 /k0 − x)k0 + 2L2
( L2 + k02 L2 + L2
)
γ := 4 2
1 + aLt4
• The peak moves as a whole with speed vg . This is called the group velocity of the
packet. In general is defined as vg := dω |
dk k=k0
, where k0 is some point considered as the
dominant wave number of the given packet. Note that the more spread is the initial
packet in physical space the more peaked is in Fourier space, and therefore the better
defined is the group velocity.
• The initial peak had width L and its Fourier transform one of width L1 . This is a general
effect and it is phrased saying that the product of both widths is constant, ∆x∆k ≥ 1.
dω 1 d2 ω
ω(k) = ω(k0) + |k=k0 (k − k0 ) + |k=k̃ (k − k0 )2
dk 2d k
2
where the last term represents the error, we can approximate the solution as,
1 ∞ 1 d2 ω
Z
dω 2
u(t, x) = √ û(k)e−i([ω(k0 )+ dk |k=k0 (k−k0 )+ 2 d2 k |k=k̃ (k−k0 ) ]t−kx) dk
2π −∞
−i(ω(k0 )− dω | k )t Z ∞
e dk k=k0 0 1 d2 ω 2
= √ û(k)e−i(vg t−x)k e−i( 2 d2 k |k=k̃ (k−k0 ) )t dk.
2π −∞
(21.5)
d2 ω 2
maxk∈[k0 −∆k0 ,k0 +∆k0 ] | ∆k0 t << 1,
d2 k
dω
u(t, x) = e−i(ω(k0 )− dk |k=k0 k0 )t u(x − vg t),
and we see that apart from a phase factor the peak propagates initially with the group speed
at the main frequency. At longer times scales we start to see dispersion as we did in the
2
example above. In it, 21 dd2ωk = a2 , ∆k0 = L1 and so the dispersion start to be important when
a2 t
L2
≈ 1.
ωp2
εω = 1 + (21.6)
ω02 − ω 2 − iγω
From this expression we can compute the actual temporal dependence between the fields
in time, as we already saw,
Z ∞
~ ~x) =
D(t, ~ − t̃, ~x) dt̃,
ε(t̃, ~x)E(t
−∞
300CHAPTER 21. WAVE PACKETS, KRAMERS–KRONING RELATIONS AND CAUSALITY
with
1 ∞ Z
ε(t̃) = √ ε(ω)e−iωt̃ dω
2π −∞
where we omit the spatial dependence from now on.
For the case at hand now
∞ e−iωt̃
Z
ε(t̃) = δ(t̃) + ωp2 dω,
−∞ (ω+ − ω)(ω− − ω)
with,
q
−iγ ± 4ω02 − γ 2 iγ q
ω± := = − ± ν0 ν0 := ω02 − γ 2 /4, ω02 > γ 2 /4 (21.7)
2
r 2
4ω02
1± 1− γ2
ω± := −iγ , ω02 < γ 2 /4 (21.8)
2
the two roots of the denominator in 21.6. Notice that both roots are complex (recall that
γ ≥ 0). If ω02 > γ 2 /4 both are equidistant from the ℜ[ω] = 0 axis and have ℑ[ω± ] = −iγ/2.
If ω02 < γ 2 /4, the discriminant, ν0 , becomes pure imaginary and both roots are in the line
ℜ[ω] = 0. In the limit ω0 → 0 one of them goes to −iγ while the other goes to zero.
Since ε(ω) is a meromorphic function on the complex plane we can compute the above inte-
gral using techniques from complex analysis: Since ℜ[−iω t̃] = ℑ[ω]t̃ we have that ℜ[−iω t̃] < 0
whenever t̃ < 0 and ℑ[ω] > 0. So for t̃ < 0 we can deform the integration along the ℑ[ω] = 0
axis into the path C+ (ρ) shown in the figure 21.1, since no pole is encounter along the defor-
mation the value of the integral along the deformed path does not change. But when ρ → ∞
the argument on the exponent has real part going to −∞ and so the integral goes to zero.
Thus, it must vanish at all paths and so at the original one.
C+ (ρ)
ω+ ω−
ρ
ω+ ω−
C− (ρ)
This time the path crossed the two points where the ε(ω) has poles, and so we pick these
contributions. Noticing that the limiting path (ρ → ∞) does not contribute we get,
1 e−iω+ t̃ e−iω− t̃
ε(t̃) = δ(t̃) + ωp2 (−i2π) √ [ + ]
2π ω− − ω+ ω+ − ω−
iπωp2 1
= δ(t̃) + √ [e−iω+ t̃ + e−iω− t̃ ] t̃ ≥ 0
ν0 2π
(21.9)
i2πωp2 1 γ t̃
ε(t̃) = δ(t̃) + √ θ(t̃)e− 2 sin(ν0 t̃),
ν0 2π
and so,
Z ∞
~ ~x) = E(t,
D(t, ~ ~x) + ~ − t̃, ~x) dt̃,
G(t̃, ~x)E(t
0
with,
i2ωp (~x)2 π 1 γ t̃
G(t̃, ~x) = √ θ(t̃)e− 2 sin(ν0 t̃).
ν0 2π
Thus, we not only see that the relation is causal, but also that it depends on the past
behavior on a time intervals of the order of γ −1 , which in normal materials is of the order of
10−10 seconds.
302CHAPTER 21. WAVE PACKETS, KRAMERS–KRONING RELATIONS AND CAUSALITY
In the case of metals, where the presence of free electrons imply in our simple model that
ω0 = 0, we obtain, in this limit ω− → −iγ while ω+ → 0, that is the case where we have a
simple pole at the origin. In terms of G(t) this indicates G(t) → 4πσ as t → ∞. Indeed, in
the case G(t) = 4πσ,
Z ∞
~
D(t) ~ + 4πσ
= E(t) ~ − t̃) dt̃,
E(t
0
and so
Z ∞
~
∂t D(t) ~ + 4πσ
= ∂t E(t) ~ − t̃) dt̃
∂t E(t
0
Z ∞
~ − 4πσ
= ∂t E(t) ~ − t̃) dt̃
∂t̃ E(t
0
~ − 4πσ[E(−∞)
= ∂t E(t) ~ ~
− E(t)] dt̃
~ + 4πσ E(t)
= ∂t E(t) ~
~
where in the last step we have assumed no initial electric field is present (E(−∞) = 0). We
see that this way recuperate Maxwell’s vacuum equations with an Ohmic current given by,
4πσ ~
J~ = E.
c
From the above calculation it is clear that always that the poles of ε(ω) are in the lower
~ and E.
half plane we shall have a causal relation between D, ~
We shall explore now the reverse, namely what are all the implications and restrictions
that the causality conditions imposes on ε(ω). They are called the Kramers-Kroning relations.
They appear in many places in physics and more generally in information theory, every time
we have a relation between two quantities which is causal, and linear,
Z ∞
~
D(t) ~ +
= E(t) ~ − t̃) dt̃,
G(t̃)E(t
0
with G(t̃) smooth and bounded. We shall study properties of this function. We shall follow
the simpler argument given in [3] which does not use Cauchy’s theorem and so does not
assume the involved functions to be meromorphic. Its Fourier transform is:
Z ∞
Ĝ(ω) := ε(ω) − 1 = G(t)eiωt dt. (21.10)
0
For simplicity in the argument we now extend G(t) to all of t and write is as,
with θ(t) the step function whose value is 1 for t > 0, and otherwise vanishes. For G± (t) we
mean an extension of G(t) to negative values of t as G± (t) = ±G± (|t|), that is an even and
an odd extension.
With this extension it is easy to compute Ĝ(ω) using the convolution formula for the
Fourier transform of a product in equation (21.11):
21.3. KRAMERS–KRONING RELATIONS 303
1 ∞
Z
Ĝ(ω) = Ĝ± (ω̃)θ̂(ω̃ − ω) dω̃.
2π −∞
Now, the Fourier transform of the step function is given by:
i
θ̂(ω) = πδ(ω) − .
ω
Thus, using it in the previous expression we arrive at the following relation:
1 −i Z ∞
Ĝ± (ω)
Ĝ(ω) = Ĝ± (ω) + PV dω̃.
2 2π −∞ ω̃ − ω
Consider first the case in of the even extension GE (t), notice that its Fourier transform is
real, so we get:
1 −1 ∞ Ĝ (ω) Z
ℜ[Ĝ(ω)] = Gˆ+ (ω)
+
ℑ[Ĝ(ω)] = PV dω̃.
2 2π −∞ ω̃ − ω
−1 ∞ ℜ[Ĝ(ω)]
Z
ℑ[Ĝ(ω)] = PV dω̃.
π −∞ ω̃ − ω
Exercise: : Using the odd extension get the other relation, namely,
1 Z ∞
ℑ[Ĝ(ω)]
ℜ[Ĝ(ω)] = PV dω̃.
π −∞ ω̃ − ω
So, we see that these relations assert that knowing the real part we can determine the
imaginary one and vice versa That is there is a relation between the dissipation part of the
relation (the imaginary part) and the propagation part (the real part). It is easy to see that,
∞ 1
Z
PV dx = 0,
0 x2 − ω 2
and so,
4πiσ
ε(ω) = ε0 +
ω
is a solution to the above relations.
Furthermore, one can check that if we choose any ℑ[ε(ω)] defined in ω ∈ [0, +∞], odd and
analytic, then the first relation defines a ℜ[ε(ω)] even and analytic. If we plug this function
in the second relation we get the original ℑ[ε(ω)], but corrected with a possible pole at the
origin, that is a different σ, and ε0 . We can think of these relations as one being the inverse
304CHAPTER 21. WAVE PACKETS, KRAMERS–KRONING RELATIONS AND CAUSALITY
of the other, except that each one of them have a non-zero kernel, and so they are not strict
inverses. So they are linear transforms in the same sense as the Fourier one.
From the experimental point of view these relations are very important for, for example
the knowledge of ℑ[ε(ω)] from absorption experiments (energy losses) allows to infer ℜ[ε(ω)],
and so ε(ω). Or vise-versa, some times it is difficult to perform absorption experiments
because the media is too opaque but reflection experiments are easy and so one obtains the
information that way.
Consider the extreme (limiting) case of a unique absorption line,
2K ω0
ℜ[ε(ω)] = 1 + + ℜ[ε̃(ω)]
π ω02 − ω 2
This effect can be easily seen in the following experimental result, 21.3
Figure 21.3: Measured extinction curves of F and M color centers in a KBr crystal at dif-
ferent temperatures (a); the corresponding changes of the refractive index are computed via
Kramers-Kronig relation (b). The curves correspond to 1) 250 K, 2) 170 K, 3) 100 K, and
4) 35 K, with peaks becoming sharper at low temperature. Adapted from Peiponen et al.
(1984).
21.4. EINSTEIN’S CAUSALITY 305
ω(~k) c
vp := = ,
k n(k)
the speed at which individual waves of wave number ~k moves, and the group velocity,
∂ω ∂ ck k ∂n
~vg := = ( ) = vp (k̂ − )
∂~k ∂~k n(k) n ∂~k
the velocity at which a peaked wave packet moves in first approximation. For many substances
n(~k) > 1 and | ∂n |, and both velocities are smaller than the speed of light. But there are cases
∂~
k
in which this is not so, nevertheless, under very generic conditions, which we spell in detail
below, and which have clear physical meaning, the systems remains Einstein causal, that is,
no propagation speed exceeds c. The scope of this proof is not so wide, as is stands only
works for wave packets where all the waves have the same propagation direction, but it is
simple and adds a lot of understanding on the subject.
Theorem 21.1 Let n(ω) be real analytic in ω, so it admits a unique extension to the complex
plane which we also denote with n(ω). Let such an extension satisfies:
Exercise: Show that the ε(ω) of our model satisfies this conditions.
e2 N 1
ε(ω) = 1 + 4π 2
, γ ≥ 0,
m ω0 − ω 2 − iγω
so it is real analytic. n−1 (ω) could have singularities only at points where ε(ω) vanishes, for
that we need the denominator to be real. That is,
At ℜ[ω] = 0 the denominator is positive (ℜ[ω02 − ω 2 − iγω] = ω02 − ℜ[ω]2 + ℑ[ω]2 + γℑ[ω]) and
so ε(ω) can not vanish. Therefore, we could only have poles at ℑ[ω] < 0.
We now prove the theorem. Let u(0, x) be a smooth and bounded, but otherwise arbitrary,
function with support on x < 0. We shall proof that it vanishes for x > ct. Thus, there is no
propagation faster than light. Its Fourier transform, is
1 ∞
Z
u(0, x)e−ikx dx.
û(k) :=
2π −∞
Since ℜ[−ikx] = ℑ[k]x < 0 the above function is analytic for ℑ[k] > 0.
The solution then is,
1 Z∞
u(t, x) = û(k)e−i(ω(k)t−kx) dk
2π −∞
1 ∞
Z
= û(k)e−ik(ct/n−x) dk
2π −∞
Since û(k) and n(k)−1 are analytic on the upper half of the plane 1
we could deform the
integration path as shown in the figure 21.4. Then,
Z
u(t, x) = lim û(k)e−ik(ct/n−x) dl.
ρ→∞ C+ (ρ)
But since n(k) → 1 as k → ∞, then, in the limit ρ → ∞ the argument of the exponential has
real negative part for ct − x < 0 and goes to −∞ as k → ∞. Thus, the integral vanishes and
we have Einstein Causality.
C+ (ρ)
1
The relation between ω and k is analytic also in this region, so n(k)−1 is analytic if n(ω)−1 is.
Chapter 22
22.1 Introduction
We want to address now the problem of determining the electromagnetic field produced by
a system of moving charges. The question in its complete generality does not have a simple
satisfactory solution since:
• The charges are affected by the radiation they themselves produce. In taking into
consideration this fact one is lead into considering also the evolution equations for the
sources. The resulting system is non-linear and so it is extremely difficult to split
fields into internal and charge generated ones. To overcome this problem we shall
assume the charge motion is given beforehand and so not affected by the radiation they
themselves produce. This is to be understood as a first step in solving a very complex
phenomenon. Once this step is taken one can consider the deviation of the source motion
this radiation produces and in some cases improve this first approximation. In other
cases this assumption is not justified and other methods/approximations must be used
to get anywhere
• Given a fixed charge distribution in space-time there are many, in fact infinitely many,
Maxwell’s solutions having them as sources, one for each initial data condition we can
give satisfying the constraint equations. Of course their radiation content is different,
so the question is: Which one to choose? The problem is only enhanced if we allow the
charges to self interact in a non-linear fashion as discussed above. If we fix the charge
motion, then we can also overcome this problem by requiring in the past the charges to
be stationary. In that case we can take initial data at those earlier times which are also
stationary. This condition fixes uniquely the initial data, and we have a problem we can
solve for. Notice that there is a catch on this: If we assume evolution equations also for
the sources, and the whole evolution systems has unique solutions for given data, then
the stationary data gives a stationary solution and no radiation!
To avoid many of the problems referred above we shall pretend now we have a given,
307
308CHAPTER 22. ELECTROMAGNETIC RADIATION PRODUCED BY CHARGE CURRENTS.
~ ·A
∂t φ + c∇ ~ = 0.
The equations to solve are then,
✷φ = 4πρ
4π ~
✷A~ =
c
J.
Notice that the gauge condition imposed allows to know φ(t, ~x) once we know A(t, ~ ~x) and
have initial data values for φ. Indeed, integrating in time (at fixed ~x) equation 22.1 we have,
Z t
φ(t, ~x) = φ0 (~x) − c ~ · A(
∇ ~ t̃, ~x) dt̃
t0
Since at t = t0 the sources, and the electromagnetic field, are summed to be stationary we
must have,
was,
∂
ψ(t, ~x) = [(t − t0 )Mc(t−t0 ) (ψ0 (~x))] + (t − t0 )Mc(t−t0 ) (ψ1 (~x))
∂t Z
t−t0
+ 4π t̃Mct̃ (f (t − t̃, ~x) dt̃,
0
with
c
Z
Mct (g(~x)) = g(~x + ctn̂) dΩ2 ,
4π S2
Since we are considering the case J~0 = 0 we can extend the integral to t0 = −∞ without
′
adding anything. Changing variables to ~x′ = ~x+ct̃n̂, we have, t̃ = |~x−~x | , (ct̃)2 d(ct̃)dΩ2 = d3~x′ ,
c
and the integral becomes,
Z ∞ Z
~ ~x) =
A(t, ct̃ ~ − t̃, ~x + ct̃n̂) dt̃ dΩ2
J(t
0 S2
1 1 ∞
Z Z
= ~ − t̃, ~x′ ) (ct̃)2 d(ct̃) dΩ2
J(t
c 0 ct̃ S 2
′
1 Z
J~(t − |~x−c~x | , ~x′ ) 3
= d ~x,
c lR3 |~x − ~x′ |
We consider now the general case, J(t,~ ~x) = J~0 (~x); t < t0 . We first notice that the
above expression is a solution to the wave equation for all values of t, if we define now
′
~ ~x) = J~0 (~x); t < t0 . Second we notice that if t < t0 then t − |~x−~x | < t0 and so,
J(t, c
4π ~~ x) =
J 0 (~x).
∆A(~
c
The stationary equation for the vector potential, so in fact this is the solution we wanted for
the general case!
310CHAPTER 22. ELECTROMAGNETIC RADIATION PRODUCED BY CHARGE CURRENTS.
(t, ~
x)
|~ x′ | = ct
x−~
t = t0
~ ~x)| ≈ O( 1
|A(t, ).
|~x|2
So, far away along spatial directions we see the electrostatic/magnetostatic fields of the far
in the past stationary distribution: A Coulombic electric field and a magnetostatic field, each
one with its own multipole distribution, obtained by expanding |~x − ~x′ | in the denominator
of Poisson’s formula.
′ ′
~ − |~x − ~x | , ~x′ ) = J(τ,
J(t ~ ~x′ )[ n̂ · ~x ] + O(|~x|2 )
~ ~x′ ) + ∂t J(τ,
c c
where n̂ = ~x . The second term will be important only if the first vanishes, or TRc ≈ 1, where
|~
x|
~
T is the characteristic time on which J~ varies, (∂t J~ ≈ JT ). Correspondingly T c would be
the wave length λ of the emitted radiation, for it should oscillate at the same characteristic
frequency as the source does. Thus, we see that the higher up terms are only important when
the wave length of the emitted radiation is of the order of the size of the source.
For longer wave length then,
1 |~x| ′ 3 ′
Z
~ ~x) ≈
A(t, J~(t − , ~x ) d ~x
c|~x| lR 3 c
−1 ~ ′ · J~(t − |~x| , ~x′ ) d3~x′
Z
≈ 3
~x′ ∇
c|~x| lR c
1 |~x| ′ 3 ′
Z
≈ 3
~x′ ∂t ρ(t − , ~x ) d ~x
c|~x| lR c
1 |~x|
≈ ∂t p~(t − )
c|~x| c
~ ~x) = ∇
B(t, ~ ∧ A(t,
~ ~x)
−1 |~x|
≈ 2 n̂ ∧ ∂t2 p~(t − ) (22.2)
c |~x| c
Since
~ ~x) = −1 ∂t A(t,
E(t, ~ ~x) − ∇φ(t,
~ ~x),
c
~
we must first compute ∇φ(t, ~x). To do that we use that it is also a solution to the wave
equation and so satisfies Poisson’s equation, but in this case we must keep terms up to second
order,
the first term is a constant, indeed it is just the charge at time τ = t− |~xc | , charge conservation
then implies that it does not depend on τ and therefore it does not depend on ~x, nor on t.
~ only to order O( 1 ) so we do not take it into account. The second
Thus, it contributes to ∇φ
|~
x|2
1
term gives a O( ) contribution,
|~
x|
312CHAPTER 22. ELECTROMAGNETIC RADIATION PRODUCED BY CHARGE CURRENTS.
1 |~x|
φ(t, ~x) ≈ n̂ · ∂t ~p(t − ).
c|~x| c
Thus,
~ ~x) ≈ −1 [∂ 2 p~(t − |~x| ) − n̂(n̂ · ∂ 2 p~(t − |~x| ))] = −1 (δ i j − n̂i n̂j )∂ 2 pj (t − |~x| )
E(t,
c2 |~x| t c t
c c2 |~x| t
c
where θ is the angle between the sight direction and the second time derivative of the dipole
vector. The energy flux by solid angle is obtained multiplying the norm of the Poynting vector
(which is in the n̂ direction) by the surface element, |~x|2 dΩ2 , so,
(t, ~
x)
~
x
|~
x|
x′
~ t=t− c
|~ x′ | = ct
x−~
Figure 22.2: First approximation, the integration region is at the surface t = t − |~xc | .
314CHAPTER 22. ELECTROMAGNETIC RADIATION PRODUCED BY CHARGE CURRENTS.
(t, ~
x)
|~
x| ′
n̂·~
x
t=t− c
− c
~
x
~
x’ |~
x|
t=t− c
|~ ~ ′ | = ct
x−x
Figure 22.3: Second approximation, the integration region is now along a null plane tangent
to the cone.
22.2. ASYMPTOTIC BEHAVIOR OF THE FIELDS 315
Since
Z Z 2π Z π
sin2 (θ) dΩ2 = sin3 (θ) dφ dθ
S2 0
Z0 π
= −2π sin2 (θ) d(cos(θ))
0
Z −1
= −2π (1 − x2 )dx
1
Z 1
= 2π (1 − x2 )dx
−1
1
= 2π(x − x3 )|1−1
3
8π
= .
3
1 |~x| ′
Z
~ ~x) ≈
A(t, ∂t J~(t − ), ~x )n̂ · ~x′ d3~x
2
c |~x| lR 3 c
1 −1 ~ ~ ′ 1
Z
≈ 2 [ ∂t (∇ · J)~x (n̂ · ~x′ ) − n̂ ∧ (~x′ ∧ ∂t J)] ~ d3~x′
3
c |~x| lR 2 2
|~
x|
1 (Q̈(t − c ), ~x′ ) · n̂ |~x| ′
≈ 2 [ − cn̂ ∧ m ~˙ (t − ), ~x )]
c |~x| 6 c
Exercise: Compute B ~ and E~ for this case and then the total power output. First make a
guess of the result (up to constant numbers) based on the previous result and on dimensional
analysis.
Exercise: A rugby player kicks the ball trying to make a conversion. The ball acquires a net
charge due to air friction. The kick in not very good and the ball spins along the three main
momentum axis. Estimate how much power gives away as radiation.
316CHAPTER 22. ELECTROMAGNETIC RADIATION PRODUCED BY CHARGE CURRENTS.
1 x′ 3 ′
′ n̂ · ~
Z
∂t
~
J (τ, ~
x )[ ] d ~x .
|~x| lR3 c
So these approximations are still valid, but we must now be careful in not dropping factors
~ and B.
higher than O(|~x|) in our calculations for E ~ Thus,
~ ~x) = 1 ∇
B(t, ~ ∧ ( 1 ~p˙(t − |~x| ))
c |~x| c
−1 |~x| 1 |~x|
= 2
(n̂ ∧ ~p˙(t − )) − 2 (n̂ ∧ ~p¨(t − ))
c|~x| c c |~x| c
−1 |~x| c ˙ |~x|
= 2 n̂ ∧ [~p¨(t − )+ ~p(t − )]
c |~x| c |~x| c
To compute the electric field we need the scalar potential, this time we shall compute it
using the gauge condition. Since we are in the Lorentz gauge we have,
~ · A.
∂t φ = −c∇ ~
Therefore,
~ · [ 1 ~p(t0 )] − ∇
φ(t, ~x) = φ0 (~x) + ∇ ~ · [ 1 p~(t − |~x| )],
|~x| |~x| c
where we have chosen the integration initial time so that all fields are static at that earlier
times. Thus, we have,
~ ~x) = −∇φ(t,
~ 1 ~
E(t, ~x) − ∂t A(t, ~x)
c
~ · [ 1 p~(t0 − |~x| )]) + ∇(
~ 0 (~x) + ∇
= −∇(φ ~ ∇~ · [ 1 ~p(t − |~x| )]) − 1 p~¨(t − |~x| ).
|~x| c |~x| c c2 |~x| c
~ ~x) = −∇[φ
E(t, ~ · ( 1 ~p0 )] + ∇(
~ 0 (~x) + ∇ ~ ∇~ · [ 1 ~p(t − |~x| )]) − ∆( 1 p~(t − |~x| ))
|~x| |~x| c |~x| c
~ · 1 p~0 ] + ∇
~ 0 (~x) + ∇
= −∇[φ ~ ∧ (∇~ ∧ ( 1 ~p(t − |~x| ))
|~x| |~x| c
~ ~x) = −∇(φ
E(t, ~ 0 (~x) + ∇~ · [ 1 p~0 ])
c|~x|
1 1 ˙ + 3n̂(n̂ · ~p˙)] − 1 [−p~¨ + n̂(n̂ · p~¨)]
+ [−~
p + 3n̂(n̂ · p
~ )] + [−p~
|~x|3 c|~x|2 c2 |~x|
Exercise: Check that when λ ≈ |~x| all terms above are of the same magnitude.
−i c |~
ω
x|
~ ω (~x) ≈ e
′
Z
A ~ω (~x′ )e−i ωc (n̂·~x ) d3~x′
J
c|~x| lR3
318CHAPTER 22. ELECTROMAGNETIC RADIATION PRODUCED BY CHARGE CURRENTS.
If ωc R = R
cT
= R
2πλ
<< 1, then
−i ωc (n̂·~
′
x)
∞
X(−i ωc n̂ · ~x′ )l
e =
l=0 l!
and only the first term will be important, it gives,
−i c |~
ω
x|
~ ω (~x) ≈ e ~ω (~x′ ) d3~x′ = −iω e−i ωc |~x| p~ω ,
Z
A J
c|~x| lR3 c|~x|
where we have used that ∇ ~ · J~ω = iωρω . From this expression we can compute the rest of the
fields and other multipoles as needed.
±iI0
.ρω (z) =
ωd
Thus, the momentum will point in the ẑ direction and will have magnitude,
d iI0 d
Z
pω = zρω (z) dz = ,
−d ω
22.4. LIENARD–WIECHERT POTENTIALS. 319
dPω I 2 d2 ω 2
= 0 3 sin2 (θ),
dΩ 8πc
and the total power,
I02 ωd 2 2πc
Pω = ( ) valid when λ := >> d.
3c c ω
where the source current becomes in the limit a distribution with support on a world–line. This
is a complicated limiting process: The charge distribution in this case is a four–dimensional
distribution, having support only on a world line, y µ (τ ), and the formula we have for com-
puting the potential is also a four–dimensional distribution, having support only in the past
light–cone of the observation point. We shall proceed in two ways, first we shall use an ar-
gument by Landau and Lifshitz to find the potential in a preferred frame and then get an
320CHAPTER 22. ELECTROMAGNETIC RADIATION PRODUCED BY CHARGE CURRENTS.
Figure 22.7: Quadrupolar Radiation, case a = 1, b = −1. Power radiated by solid angle.
22.4. LIENARD–WIECHERT POTENTIALS. 321
equivalent covariant expression valid in any system. After using this expression for finding the
Maxwell tensor y present a different argument leading to the same expression but performing
the distributional calculus needed.
The simplest way of doing this computation is as follows, we look into the past of our
observation point along null directions, those directions form the past like cone of this point.
Since its velocity is smaller than the velocity of light and is coming from the infinite past,
a particle world line would intersect the cone in just a single point. So the contribution to
the potential integral would come only from that point, (ct̃(τ0 ), ~y (τ0 )). Given the observer’s
point, (t, ~x), and the particle world line, y µ (τ ) we can find the intersection point of it with
the past like cone, and so we have, τ0 = τ0 (t, ~x), and correspondingly t̃ = t̃(τ0 (t, ~x)), and
~y = ~y (τ0 (t, ~x)). We can now go to a frame for which the particle at that past point is at rest,
so at that point, (ct̃(τ0 ), ~y (τ0 )), its four–velocity will be, uµ = (1, 0), and the corresponding
current vector j µ = euµ = (e, 0). At this point the proper time of the particle coincides with
0
its coordinate time, ct̃(τ0 ) = τ0 , and therefore at that instance, dx̃ dτ
= 1.
(t, ~
x)
|~
x−~x′ |
t= c
lµ
uµ
x′
x−~
~
xµ (τ )
In this frame the integral can be done easily and the four–potential is given by,
322CHAPTER 22. ELECTROMAGNETIC RADIATION PRODUCED BY CHARGE CURRENTS.
e ~ ~x) = 0
φ(t, ~x) = , A(t,
|~x − ~y |
so the field corresponds to the static field of the particle as if it continues at rest at the
position ~y it had at the past at time t̃ = t − |~x−c ~y | . Since the potential is a four-vector,
all we have to do now is to write this expression on a covariant way and then it would
be valid in any other frame. At our disposal we only have the four-velocity vector, uµ ,
which in this frame is uµ = (1, 0, 0, 0), the corresponding current j µ = euµ and the four-
vector connecting the observation point with the position of the charge along the light cone,
lµ = xµ − y µ = (c(t − t̃), ~x − ~y ). This last one is a null vector, thus,
µ jµ
A = ν ,
u lν
is the required expression, valid now on any frame we please. For a generic frame, tµ ,
~ in terms of three–dimensional quantities we have associated to it, namely using
uµ = γ(1, β),
coordinates where tµ = (1, 0, 0, 0),
e ~ ~x) = e~v
φ(t, ~x) = A(t, (22.4)
(|~x − ~y | − β~ · (~x − ~y )) c(|~x − ~y | − β~ · (~x − ~y ))
where ~v is the three velocity of the particle at point ~y and time t̃ = t − |~x−c ~y | .
We can now build the Maxwell tensor, Fµν := 2∂[µ Aν] . In taking time derivatives of the
potential we will have normal ones, on the dependence of the vector lµ connecting the emission
point and the observer, and on the dependence of lµ and uµ on τ . So first we compute the
space-time derivatives of τ .
∂ t̃
Differentiating the relation, c(t − t̃) = |~x − ~y | one gets, c(1 − ∂t ) = − (~x−~y )·~v ∂t
∂ t̃
. So,
|~
x−~y |
∂ t̃ 1 1
= = .
∂t 1− (~
x −~y )·~v 1 − n̂ · β~
|~
x−~y |c
Differentiating with respect to the space coordinates, the same relation, we get,
ni 1 ni 1
∂i t̃ = − =− .
c 1 − (~x−~y )·~v c 1 − n̂ · β~
|~
x−~y |c
Since dt̃
dτ
= γc , we have,
lµ
∂µ τ = , (22.5)
(uσ lσ )
22.4. LIENARD–WIECHERT POTENTIALS. 323
Something which it is simplest to see in the frame co-moving with the particle. Consequently,
lµ
∂µ uν = u̇ν ∂µ τ = aν ,
(uσ lσ )
similarly,
uσ lµ
∂µ lσ = δµ σ − uσ ∂µ τ = δµ σ − ,
uρ lρ
it follows that
F µν = 2∂ [µ Aν]
u̇[ν ∂ µ] τ u̇σ lσ u[ν ∂ µ] τ u[ν uσ ∂ µ] lσ
= 2e[ σ − − ]
u lσ (uσ lσ )2 (uσ lσ )2
2e [µ ν] aρ lρ ν] 1
= l [(a − σ u ) − σ uν] ]
(u lσ )
σ 2 u lσ u lσ
2e [µ ν] 1
=: l [s + σ uν] ],
(u lσ )
σ 2 u lσ
ρ
where, sν := aν − uaσρ llσ uν .
We easily see that,
F µν ⋆ Fµν = 0. (22.6)
So the magnetic and electric fields are always perpendicular to each other.
Notice that the Maxwell field is composed of two parts, one that depends on the vector
u and the other on the acceleration vector aµ . The term which depends only on the velocity
µ
vector corresponds to the Coulombic field of the particle as if it were following a straight line
with the velocity given by the value it had at time t̃. Indeed, in the frame where the particle
is momentarily at rest, namely the one defined by, uµ , the electric field will correspond to
that field,
e e ~x − ~y
F µν uν = (l µ
− (u ρ
lρ )u µ
) = (0, ),
(uρ lρ )3 |~x − ~y |2 |~x − ~y |
while the corresponding magnetic field would vanish. Notice that for that part of the field,
−2e2
F µν Fµν = < 0,
(lσ uσ )4
and so, since 22.6 also holds, there is a frame for which the field has only an electric part,
namely the one referred above.
We now analyze the term which depends on the acceleration. This term represents a pure
electromagnetic wave, indeed, note that,
aρ lρ ν
sν lν = (aν − u )lν = 0,
uσ lσ
and so, since lµ is a null vector, we have
324CHAPTER 22. ELECTROMAGNETIC RADIATION PRODUCED BY CHARGE CURRENTS.
F µν Fµν = 0.
Thus, both scalar invariants vanish, and we have a wave traveling along the vector lµ . Since
the vectors l and s appears in F , as the vectors k and v for plane waves, it is important
to see them in geometrical terms. The null vector lµ is such that lµ = u−1 ρ l ∂µ u, where u =
ρ
c(ttip − t) − |~xtip −~x| is the function whose zero level set is the past cone of the point (ttip , ~xtip ).
So as kµ was normal to the plane u = kµ xµ , here lµ is normal to the light cone. Since sρ lρ = 0
we see that s is the unique linear combination of the acceleration a and the velocity u which
is in the tangent plane of the cone. Since we can add to s any vector in the direction of
l without changing the value of F , and therefore of T , s represents an equivalent class of
tangent vectors at the cone. We can choose a particular one only when we choose a particular
time direction. These are the different electric fields.
As an example we compute now the radiation term of the electromagnetic field correspond-
ing to an arbitrary observer tµ . For this observer the four–velocity is given by, uµ = γ(tµ + β µ )
and the corresponding acceleration by,
1 duµ γ2 µ
aµ = = (ã + γ(β σ ãσ )uµ )
c2 dτ c2
where ãµ is the coordinate acceleration with null time-like component (ãµ tµ = 0). 1 The γ
factors appear from the difference between proper time and coordinate time, each derivative
contributing with one power.
The null vector lµ in the new frame has the same form, (since it is a null vector it remains
so)
lµ = R̃(tµ + nµ ),
where now nµ is a unit vector in the simultaneity surface perpendicular to tµ pointing in the
direction from the observer to the particle in the retarded position, and R̃ = lµ tµ the distance
between observer and particle in that simultaneity surface, as it was R = lµ uµ in the particle
rest frame. Contracting it with uµ we have,
uµ lµ = R = R̃(1 + nµ βµ )γ
Contracting it with aµ we get,
γ2 µ γ3 ρ µ
aµ lµ = R̃n ãµ + u lρ β ãµ .
c2 c2
aρ lρ µ
The vector sµ := aµ − uσ lσ
u becomes then,
γ2 µ nσ ãσ
sµ = (ã − uµ ).
c2 γ(1 + nρ βρ )
2 σ
Finally, using that sµ tµ = − γc2 1+n
n ãσ µ
ρ β , and l tµ = R̃, we have,
ρ
1 d dt d ~ γ2
This comes about because aµ := dτ (γ(tµ + β µ )) = dτ dt (γ(1, β)) = ρ
c2 (γβ ãρ u
µ
+ ãµ ). The vector ãµ is
not the space component of the acceleration vector aµ .
22.4. LIENARD–WIECHERT POTENTIALS. 325
Eµ := Fµν tν
e
= [lµ sν tν − sµ lν tν ]
(u lσ )2
σ
−e
= 2 [(nµ − βµ )nσ ãσ + ãµ (1 + nρ βρ )]
c R̃(1 + nρ βρ )3
(22.7)
which is the usual expression. Note that it is perpendicular to nµ . The magnetic field is
perpendicular to it and to nµ . So both fields can be written in a more compact vector
calculus expression, in which we have also added the Coulombic part.
~ ~x) = e v2 |~x − ~y |
E(t, [(1 − )(~x − ~y − ~v )
(|~x − ~y | − ~v ·(~xc−~y ) )3 c 2 c
1 |~x − ~y |
+ 2 (~x − ~y ) ∧ ((~x − ~y − ~v ) ∧ ~v˙ )] (22.8)
c c
~ ~x) = n̂ ∧ E(t,
B(t, ~ ~x), (22.9)
Once again we see that the magnetic and electric fields are perpendicular to each other, that
is, F µν ⋆ Fµν = 0.
The first term on the electric field decays at large distances (null or space–like) as O( 1 2 )
|~
x|
and depends only on the particle velocity. This is in fact a Coulomb field corresponding to
the particle as if it would be moving at constant speed from the past to the point where it
should be now. The second term decays as O( 1 ), and represents radiation coming out of
|~
x|
the particle due to its acceleration in the past.
Exercise: Check that the first term corresponds to the Coulomb field of a charged particle
moving at constant speed in the corresponding frame. Hint: realize that ~x − ~y − |~x−c ~y | ~v =
~x − ~y − (t − t̃)~v = ~x − (~y + (t − t̃)~v), is the vector connecting the observation point with the
point where the particle would have been, were it moving at constant velocity (~v).
326CHAPTER 22. ELECTROMAGNETIC RADIATION PRODUCED BY CHARGE CURRENTS.
−1 µσ ν e2
T µν = F F σ= l µ l ν sρ sρ
4π 4π(l uσ )
σ 4
ρ σ
where, sν = (aν − uaσρ llσ uν ), and therefore, since sρ sρ = aρ aρ − ( auρllσρ )2 ≥ 0 a space-like vector. 2
Using this formula we can compute Poynting’s vector and from it deduce the radiation
power as a function of emission angle and then the total power radiated away.
dP e2 c ρ e2 c ρ aρ lρ e2 e2
= s sρ = (a aρ − ( σ )2 ) = (~
a · ~
a − (~
a · n̂) 2
) = |~a|2 sin(θ)2
dΩ 4π 4π lσ u 4πc3 4πc3
where θ is the angle between ~a and n̂. Using now that,
8π
Z
sin(θ)2 dΩ =
S2 3
we can compute the total power output radiated by the particle,
2e2 2
P =
|~a| .
3c3
Where we have expressed it in terms of time units (not in terms of x0 = ct), to show that it
coincides with the dipole formula 22.3 if we consider the dipole to be a particle moving with
the given acceleration.
This is the power output the particle will see in its own frame. But this quantity is
independent of the frame chosen to describe it. This is clear from its definition, but it is
instructive to see the following argument: From the figure 22.9 we see what would be this
integral for two different observers. Using energy–momentum conservation we shall see that
2
To see this write the metric as η µν = −2l(µ k ν + q µν , where k µ = R̃1 (tµ − nµ ) and notice (by explicitly
writing it in Cartesian coordinates, taking, say nµ = (0, 1, 0, 0)) that q µν is the identity matrix when restricted
to the y, z coordinates. Since sµ lµ = 0, sµ sµ is just the square of the acceleration components in the direction
perpendicular to nmu .
22.4. LIENARD–WIECHERT POTENTIALS. 327
this integral is the same for both surfaces. First note that due to the particular form of the
energy–momentum tensor, −Tµν uµ nν = 41 Tµν k µ k ν , where k µ := uµ − nµ , is a unique null
vector, k µ kµ = 0, normal to the null surface ruled by lµ , and to the two-dimensional surface,
and normalized such that k µ lµ = 2R. Thus,
µ ν −1 µ ν −1 e2 c2
Tµν u n = Tµν k k = sρ sρ .
4 4 4π(l uσ )
σ 2
So this is the quantity to be integrated at a cut, C, on the outgoing light cone of the particle
at a given moment, obtained by intersecting any plane perpendicular to uµ .
I I
P (τ ) = Tµ ν uµ nν dS = j ν nν dS
C cut
where we have defined the current j ν = Tµ ν uµ . Notice that j µ lµ = 0, so this is indeed a vector
tangent to the light cone of the particle, and so we can use Gauss theorem to transform the
integral into a volume integral, between two different cuts, C and C. ˜
Thus,
I Z √ I
P (τ ) = j ν nν dS = ∂µ (j µ s) dΩ dr + j ν nν dS
C S C˜
But,
√ √ √ √ √
∂µ (j µ s) = ∂µ (Tµ ν uµ s) = ∂µ (Tµ ν )uµ sTµ ν ∂µ (uµ ) s + j µ ∂µ s = 0
where the first term vanishes because of Maxwell’s equations (energy conservation), the second
because we have extended the 4-velocity vector of the particle at time τ into a global frame,
and the final one because the T µν has only components in the lµ direction, and along that
direction the determinant of the cone metric is constant. Thus, we see that the power output
can be obtained by performing an integration on any 2-surface at the future light cone of the
particle which includes the tip of it in its interior.
It is instructive for applications to write the above form in different forms. Since,
γ2
0 = aµ uµ = (ãµ uµ − γ(β σ ãσ )),
c2
we have,
γ4 γ4
aµ aµ = (ãµ ãµ
+ γ 2
(β σ
ãσ ) 2
− 2γ(β σ
ã )ã
σ µ u µ
) = (ãµ ãµ − γ 2 (β σ ãσ )2 ),
c4 c4
and therefore,
328CHAPTER 22. ELECTROMAGNETIC RADIATION PRODUCED BY CHARGE CURRENTS.
uµ
C n̂
C˜
uµ
xµ (τ )
Figure 22.9: Different cuts where the flux integration can be carried out.
22.4. LIENARD–WIECHERT POTENTIALS. 329
−2e2 c ρ
P = a aρ
3
2e2 γ 4
= [~a · ~a + γ 2 (β~ · ~a)2 ]
3c 3
2e2 γ 6 2
= [|~a| (1 − β 2 ) + (β~ · ~a)2 ]
3c3
2e2 γ 6 2 ~v
= [|~a| − | ∧ ~a|2 ], (22.10)
3c 3 c
in terms of 3-velocities and 3-accelerations, or when the particles have a constant rest mass,
P m dE
dE = 0.5 × 10−14 [ ]
dt
MeV dx
while the typical energy delivered per meter in present machines is of the order of tens of
MeVs. Thus, radiation losses in these accelerators is negligible and the main limitation they
have is their length and the amount of energy delivered per unit length.
330CHAPTER 22. ELECTROMAGNETIC RADIATION PRODUCED BY CHARGE CURRENTS.
d~p
|
| ≈ γω|~p|,
dτ
v
where ω = R
is the orbital frequency, and R is the orbit radius. In this case we have,
2 e2 cβ 4 γ 4
P =
3 R2
In this case the relevant quantity is the energy lost per revolution compared with the energy
gain during the same trajectory. The first quantity is,
δE ≈ 300MeV.
For the LHC (Large Hadronic Collider), R = 4300m, and the proton beam reached energies
of 7 TeV. The energy loss per revolution in this case is, of about 10KeV. We see from the
form the energy loss scales with particle mass that it is more convenient to accelerate heavy
particles in a circular accelerator while it is more convenient to accelerate light particles in
linear accelerators. To have an idea of the total energy loss, one has to have into account
that the particles circle the ring about 11245 times per second, while the number of protons
at any given moment in the ring is about 3 × 1014 .
Since the all quantities in this expression depend on the retarded time, τ = t − |~x − ~x′ (τ )|/c
dt
it is necessary to express it in terms of that time, since dτ = (1 − β~ · n̂) the integral becomes,
Z τ2
E(τ2 , τ1 ) = ~ · n̂R2 (1 − β~ · n̂) dΩdτ,
S
τ1
22.5. ANGULAR DISTRIBUTION OF RADIATION 331
dP (τ ) ~ · n̂ R2 (1 − β~ · n̂)
=S
dΩ
which is the power radiated per unit solid angle and unit retarded time.
Using the expressions above 22.8, for the electric field, and noticing that the corresponding
magnetic field has the same magnitude and is perpendicular to it, we have,
dP (τ ) ~ ∧ β)|
e2 |n̂ ∧ ((n̂ − β) ~˙ 2
= .
dΩ 4πc (1 − β~ · n̂)5
The main feature of this distribution is the relativistic factor in the denominator, for
ultra–relativistic motion that term is very small and dominates the distribution.
dP (τ ) e2 a2 γ 8 43
(θmax ) ≈ ,
dΩ 4πc3 55
so it grows as the eighth power of γ!, see figures 22.10 and 22.11.
30
25
20
15
10
Figure 22.10: Power radiated by solid angle. Approximation for small angle for γ = 2.
332CHAPTER 22. ELECTROMAGNETIC RADIATION PRODUCED BY CHARGE CURRENTS.
1 ´ 107
8 ´ 106
6 ´ 106
4 ´ 106
2 ´ 106
Figure 22.11: Power radiated by solid angle. Approximation for small angle for γ = 10.
1 2 3
-1
-2
-3
Figure 22.12: Power radiated by solid angle in the co-linear case, β = 0.5.
22.5. ANGULAR DISTRIBUTION OF RADIATION 333
2000
1000
-1000
-2000
Figure 22.13: Power radiated by solid angle in the co-linear case, β = 0.9.
Exercise: Find the angle of maximal radiation for the ultra-relativistic limit by taking a
derivative of the exact expression with respect to the angle and equating to zero. Use the
approximations for β given above. Redo the calculation using the small angle approximation
expression and check that the angles coincide.
Figure 22.14: Power radiated by solid angle, velocity and acceleration co-linear, β = 0.4
Figure 22.15: Power radiated by solid angle, velocity and acceleration co-linear, β = 0.9.
dP (t′ ) e2 a2 1
= [(1 − β cos θ)2 a2 − (1 − β 2 )(~a · n̂)2 ]
dΩ 4πc3 (1 − β cos θ)5
e2 1
= [(1 − β cos θ)2 − (1 − β 2 ) sin2 θ cos2 φ]
4πc (1 − β cos θ)
3 5
2e2 a2 γ 6 a2 4γ 2 θ2 cos(φ)2
≈ [1 − ]
πc3 (1 + (γθ)2 )3 (1 + γ 2 θ2 )2
where the angles are shown in figure 22.16. One can see that there will be maximal radiation
in the ultra–relativistic case when the angle φ = 0, that is in the orbit plane. Furthermore,
the maximum is along the particle motion.
~
β
n̂
~a(τ )
y
~ (τ )
2 4 6 8
-1
-2
Figure 22.17: Power radiated by solid angle, case velocity perpendicular to acceleration,
β = 0.5, φ = 0 (blue) and φ = π/2 (violet).
336CHAPTER 22. ELECTROMAGNETIC RADIATION PRODUCED BY CHARGE CURRENTS.
Figure 22.18: Power radiated by solid angle, case velocity perpendicular to acceleration,
β = 0.92
Figure 22.19: Power radiated by solid angle, case velocity perpendicular to acceleration,
β = 0.92, cut to see the φ dependence.
22.5. ANGULAR DISTRIBUTION OF RADIATION 337
where y µ (s) is the particle trajectory and s is any parameter used for expressing it. As always
for vectorial integrals it is only valid when expressed in Cartesian coordinates.
Exercise: Check that this definition is independent on the choice of parameter s. Check
that it satisfies, ∂µ j µ = 0 in the distributional sense.
′
Z
j µ (t − |~x−c~x | , ~x′ ) 3 Z
j µ (t′ , ~x′ ) ′ |~x − ~x′ | 3
d ~
x = δ(t − t + ) d ~x dt. (22.12)
lR3 |~x − ~x′ | lR4 |~x − ~x′ | c
Inserting in the above expression the distributional form for the particle trajectory we get,
dy µ
e ∞ |~x − ~x′ | 3
Z Z
µ σ ds ′σ σ ′
A (x ) = δ(x − y (s))δ(t − t + ) d ~x d(ct) ds
c lR4 −∞ |~x − ~x′ | c
dy µ
eZ ∞ ds |~x − ~y (s)|
= δ(t̃(s) − t + ) ds
c −∞ |~x − ~y (s)| c
(22.13)
where ct̃(s) = y 0(s), and in the second equality we have performed the four dimensional
integral.
In order to perform the last integral we need to use, that
∞ g(xi )
Z X
g(x)δ(f (x)) dx = , ∀ {xi |f (xi ) = 0}
−∞ i |f ′(xi )|
but
dy µ
e ∞ |~x − ~y (s)|
Z
µ σ ds
A (x ) = δ(t̃(s) − t + ) ds
c −∞ |~x − ~y (s)| c
dy µ
dt̃ −1
e ds
( ds )
=
c |~x − ~y (s)| (1 − 1 d~y (s) · n̂)
c dt̃
dy µ
e dt̃
=
c |~x − ~y (s)|(1 − 1 d~
y (s)
c dt̃
· n̂)
dy µ
dτ
= e
1 d~
y (s)
γ|~x − ~y (s)|(1 − c dt̃
· n̂)
dy µ
= e dydτρ
l
dτ ρ
(22.14)
Exercise: Find F µν by first performing the derivatives in 22.13 and then performing the
space-time derivatives.
Chapter 23
23.1 Introduction
In many situations in electromagnetism, in particular when we want to study fields near
sources, we have to different time scales: the time that takes a wave to transverse the sources,
that is τc = L/c, where L is the source’s length, and some other time scales coming from the
sources. In particular there appears one which is related to the adjustment of the sources to
the effect of electromagnetic fields through Ohmic effects. If the first time scale is very short
compared with the second, then we can consider the fields as statics at each instant of time.
If they are comparable, then radiation effects must be taken into account. If we want to study
the limit in which τc → 0, so as to get the effect of the other time scale, we can look at the
equations in the limit of c → ∞. Defining ε = 1c . We have,
~ = 1∇
∂t E ~ ∧B~ − 4π J~
ε
~ = −1 ∇
∂t B ~ ∧E~
ε
~ ·E
∇ ~ = 4πρ
~ ·B
∇ ~ = 0
~ = E
E ~ 0 + ε2 E
~ 1 + ε4 E
~ R (ε)
~ = εB
B ~ 0 + ε3 B
~ 1 + ε5 B
~ R (ε)
we have omitted some terms because by re-scaling one can see that the solutions should depend
on even powers of ε, starting at some arbitrary power. It can be seen that the omitted term
are zero if appropriate boundary conditions are imposed. An important theorem for systems
with different time scales states that if the initial data is such that the time derivatives up to
some order are initially bounded then they remain so for subsequent times. So the requirement
that there would be solutions which are regular in the time variable is actually a restriction
339
340 CHAPTER 23. QUASI STATIC FIELDS, DIFFERENT TIME SCALES
on the set of possible initial data sets, we are restricting the solution not to have waves and so
to be quasi-static or quasi-Coulombic. So see what type of solutions we are getting we plug
the terms of the ansatz on the equation and to the lowest order we get,
~0 = ∇
∂t E ~ ∧B ~ 0 − 4π J~ (23.1)
~ 0 = −1 ∇
∂t B ~ ∧E ~0 + ∇ ~ ∧E
~2 (23.2)
ε2
~ ·E
∇ ~ 0 = 4πρ − ε2 ∇~ ·E
~2 (23.3)
~ ·B
∇ ~0 = 0 (23.4)
The smoothness requirement on the time derivatives implies,
~ ∧E
∇ ~0 = 0 (23.5)
imposing also that
~ ·E
∇ ~ 0 = 4πρ
we are back to the electrostatic equations, which now have to be thought as valid at each
instant of time. We are back to instantaneous influence. So, giving appropriate boundary
conditions we have a unique E ~ 0 and so equation 23.1 have to be thought as an equation for
~
B 0 . But actually it does not need to be solved in the present form. Taking its curl, using ,
and 23.4 we obtain,
~ ∧ J.
∆B0 = −4π ∇ ~ (23.6)
So we can solve this elliptic equation with appropriate boundary conditions and have a solution
at each time. We have recuperated the magneto-static equations. But since these equations
are now valid at each instant of time, and we can allow for source variations their field of
applications is much wider.
Consider for instance the case in which Ohm’s law is valid. Then,
J~ = σ E,
~
and we have,
∆B0 = −4π ∇ ~ ∧ J~
= −4πσ ∇~ ∧E ~
~ ∧E
= −4πε2 σ ∇ ~2
4πσ
= ∂t B0
c2
and we get a parabolic equation for the magnetic field to this order. Thus, the magnetic field
will diffuse itself in a time scale given by
4πσL2
τσ =
c2
23.1. INTRODUCTION 341
where L is the system size. This time scale is in general much larger than the radiation time
scale, related to the time the light spend in transversing the object, τc = L/c. For instance,
if we consider the earth core, then L ≈ 103 Km, σ ≈ 1017 s−1 and so, τσ ≈ 1012 s which is of
the order of tens of thousand of years.
Exercise: Find the length for which both time scales coincide for the case of cooper.
The relevant theorem for systems with different temporal scales, when applied to the case
of electromagnetism, is that if one considers smooth initial data satisfying the above equations,
then the resulting solution will be smooth (both in space and time) and will satisfy the same
equations for all times, in the sense that the error terms, E~ R (ε), and E
~ B (ε) are bounded.
Exercise: Find the equations the electromagnetic field would satisfy in a dielectric substance
(ε 6= 1) assuming the medium is homogeneous and isotropic.
Examination Questions 2
Problem 41 Why do we assert that if a symmetry of space-time is valid for Maxwell’s equa-
tions it is also valid for the wave equation?
Problem 42 Complete the argument after equation 13.7, that is the exercise that follows it.
Problem 43 Probe the Lemma in the variational principle for a particle in space-time. Do
the exercise below.
Problem 44 Find the expression for the time delay to first non-vanishing order between a
person on the longest path between two time-like events and a person along a nearby path.
Problem 47 Show that the sum of two future directed time-like vectors is also time-like and
future directed.
Problem 48 Show that the sum of a future directed time-like vector with a future directed
null vector is always time-like and future directed.
Problem 49 Under which considerations is total momentum conserved for a set of particles?
343
344 CHAPTER 24. EXAMINATION QUESTIONS 2
Problem 53 Show that if at some point of space-time Fµν F µν > 0 and Fµν ⋆ F µν = 0, that
is, if for any observer at that point the electric and magnetic fields are perpendicular and the
magnetic field is bigger in norm than the electric field, then there is an observer for which the
electric field vanishes.
Problem 54 Show that if for some time-like vector uµ , Fµν uν = 0 then ⋆ Fµν = 2b[µ uν] for
some space-like vector bµ . Show that it also follows that F µν ⋆ Fµν = 0 and F µν Fµν ≤ 0
~ and B
Problem 55 Check that pµ in terms of E ~ is given by,
−1
pµ := −etµ + P~ = ((E 2 + B 2 )tµ − 2tρ ερµσν Eσ Bν ),
8π
~ and B.
Problem 56 Compute pµ pµ in terms of E ~ Does it have a definite sign?
Problem 57 Show that ∂µ T µν = j µ Fµν if Maxwell’s equations are satisfied. Is the converse
true? If so, in which case?
Problem 60 Let Aµ be such that Fµν = 2∂[µ Aν] . Given a constant unit time-like vector t
define Aµ = φtµ + õ that is, φ = A0 = −A · t = −A0 and à · t = 0. Check that
B µ = ε̃µνσ ∂ν Aσ := tρ ερµνσ ∂ν Aσ ,
and
Eµ = −∂˜µ φ + ∂0 õ .
Check explicitly that both vectors are gauge invariant.
Problem 61 Describe the initial conditions problem for the vector potential in the Lorentz
gauge. Which fields can be given and which are specified and why.
Problem 62 Discuss for which type of problem one would preferably use the Lorentz gauge
and for which Coulomb’s one? In a radiation problem? In a quasi-stationary problem?
Problem 63 Deduce the equations of motion of a charged particle using the variational prin-
ciple.
Problem 64 Deduce Maxwell’s equations using a variational principle. Can you deduce this
way all equations or just some subset?
345
Problem 65 Study the plane waves in terms of the Maxwell tensor and also in terms of the
four-vector potential.
Problem 66 Deduce formula 18.11 for total reflection of a superconductor.
Problem 67 Find the Fourier amplitudes for the vector potential in terms of its initial data
in space for a monochromatic wave.
Problem 68 Check that given initial data (E ~ 0, B
~ 0 ) = (Pl,n,m E ~ 0lmn (x) for
~ 0lmn (x), Pl,n,m B
the electric and magnetic fields inside a rectangular cavity, we can construct a solution as
sum of the cavity modes. Find the explicit values for E ~ + and E ~− .
lmn lmn
Problem 69 Estate the problem of finding the solutions of waves in a wave guide in all three
cases. Assuming you have solved for all the modes, write the solution in terms of arbitrary
initial data on the wave guide.
Problem 70 Show that for a mode in a cavity,
Z Z
~ 2 dV =
|E| ~ 2 dV =
|B|
V V
ω2 ~ ~ ω = 0 and −iω B
~ω = ∇
~ ∧E
~ ω.
Hint: use E
c2 ω
+ ∆E
Problem 71 Find the amplitudes for the reflected and transmitted waves when the magnetic
vector is tangent to the interface, but doing the calculation with the boundary conditions for
the electric field.
Problem 72 Deduce the Kramers-Kroning relations.
Problem 73 Deduce the causality theorem.
Problem 74 Prove the assertion on the retarded Green function.
Problem 75 Describe in geometrical terms the nature of the approximation made in the
Green function integral regarding the power series expansion of the numerator.
Problem 76 Compute up to numerical factors the total output for quadrupole radiation.
Problem 77 A rugby player kicks the ball trying to make a conversion. The ball acquires a
net charge due to air friction. The kick in not very good and the ball spins along the three
main momentum axis. Estimate how much power gives away as radiation.
Problem 78 Deduce the Lienard–Wiechert potential from the argument of going to a pre-
ferred frame.
Problem 79 Deduce the Lienard–Wiechert potential from the argument using the retarded
Green function.
Problem 80 Find the Maxwell tensor corresponding to the Lienard–Wiechert potential. First
write the most general antisymmetric covariant tensor depending on four vectors, lµ , j µ , uµ ,
and the acceleration aµ . Recall that j µ , and uµ are parallel, and aµ is perpendicular to uµ .
346 CHAPTER 24. EXAMINATION QUESTIONS 2
Bibliography
[1] General Relativity from A to B, R. Geroch (University of Chicago Press, Chicago, 1978).
[3] Concepts in thermal physics Blundell SJ, Blundell KM. New York: Oxford University
Press 2010.
347