Angular Momentum: J. Broida UCSD Fall 2009 Phys 130B QM Ii
Angular Momentum: J. Broida UCSD Fall 2009 Phys 130B QM Ii
ijk
klm
=
il
jm
im
jl
.
Also recall the elementary commutator identities
[ab, c] = a[b, c] + [a, c]b and [a, bc] = b[a, c] + [a, b]c .
Using these results together with [x
i
, x
j
] = [p
i
, p
j
] = 0, we can evaluate the com-
mutator as follows:
[L
i
, L
j
] = [(r p)
i
, (r p)
j
] = [
ikl
x
k
p
l
,
jrs
x
r
p
s
]
=
ikl
jrs
[x
k
p
l
, x
r
p
s
] =
ikl
jrs
(x
k
[p
l
, x
r
p
s
] + [x
k
, x
r
p
s
]p
l
)
1
=
ikl
jrs
(x
k
[p
l
, x
r
]p
s
+x
r
[x
k
, p
s
]p
l
) =
ikl
jrs
(i
lr
x
k
p
s
+i
ks
x
r
p
l
)
= i
ikl
jls
x
k
p
s
+i
ikl
jrk
x
r
p
l
= +i
ikl
ljs
x
k
p
s
i
jrk
kil
x
r
p
l
= i(
ij
ks
is
jk
)x
k
p
s
i(
ji
rl
jl
ri
)x
r
p
l
= i(
ij
x
k
p
k
x
j
p
i
) i(
ij
x
l
p
l
x
i
p
j
)
= i(x
i
p
j
x
j
p
i
) .
But it is easy to see that
ijk
L
k
=
ijk
(r p)
k
=
ijk
krs
x
r
p
s
= (
ir
js
is
jr
)x
r
p
s
= x
i
p
j
x
j
p
i
and hence we have the fundamental angular momentum commutation relation
[L
i
, L
j
] = i
ijk
L
k
. (1.1a)
Written out, this says that
[L
x
, L
y
] = iL
z
[L
y
, L
z
] = iL
x
[L
z
, L
x
] = iL
y
.
Note that these are just cyclic permutations of the indices x y z x.
Now the total angular momentum squared is L
2
= L L = L
i
L
i
, and therefore
[L
2
, L
j
] = [L
i
L
i
, L
j
] = L
i
[L
i
, L
j
] + [L
i
, L
j
]L
i
= i
ijk
L
i
L
k
+ i
ijk
L
k
L
i
.
But
ijk
L
k
L
i
=
kji
L
i
L
k
=
ijk
L
i
L
k
where the rst step follows by relabeling i and k, and the second step follows by the
antisymmetry of the Levi-Civita symbol. This leaves us with the important relation
[L
2
, L
j
] = 0 . (1.1b)
Because of these commutation relations, we can simultaneously diagonalize L
2
and any one (and only one) of the components of L, which by convention is taken
to be L
3
= L
z
. The construction of these eigenfunctions by solving the dierential
equations is at least outined in almost every decent QM text. (The old book In-
troduction to Quantum Mechanics by Pauling and Wilson has an excellent detailed
description of the power series solution.) Here I will follow the algebraic approach
that is both simpler and lends itself to many more advanced applications. The
main reason for this is that many particles have an intrinsic angular momentum
(called spin) that is without a classical analogue, but nonetheless can be described
mathematically exactly the same way as the above orbital angular momentum.
2
In view of this generality, from now on we will denote a general (Hermitian)
angular momentum operator by J. All we know is that it obeys the commutation
relations
[J
i
, J
j
] = i
ijk
J
k
(1.2a)
and, as a consequence,
[J
2
, J
i
] = 0 . (1.2b)
Remarkably, this is all we need to compute the most useful properties of angular
momentum.
To begin with, let us dene the ladder (or raising and lowering) operators
J
+
= J
x
+iJ
y
J
= (J
+
)
= J
x
iJ
y
.
(1.3a)
Then we also have
J
x
=
1
2
(J
+
+J
) and J
y
=
1
2i
(J
+
J
) . (1.3b)
Because of (1.2b), it is clear that
[J
2
, J
] = 0 . (1.4)
In addtion, we have
[J
z
, J
] = [J
z
, J
x
] i[J
z
, J
y
] = iJ
y
J
x
so that
[J
z
, J
] = J
. (1.5a)
Furthermore,
[J
z
, J
2
] = J
[J
z
, J
] + [J
z
, J
]J
= 2J
2
] = kJ
k
. (1.5b)
It will also be useful to note
J
+
J
= (J
x
+iJ
y
)(J
x
iJ
y
) = J
2
x
+J
2
y
i[J
x
, J
y
]
= J
2
x
+J
2
y
+ J
z
and hence (since J
2
x
+J
2
y
= J
2
J
2
z
)
J
2
= J
+
J
+J
2
z
J
z
. (1.6a)
Similarly, it is easy to see that we also have
J
2
= J
J
+
+J
2
z
+ J
z
. (1.6b)
3
Because J
2
and J
z
commute they may be simultaneously diagonalized, and we
denote their (un-normalized) simultaneous eigenfunctions by Y
where
J
2
Y
=
2
Y
and J
z
Y
= Y
.
Since J
i
is Hermitian we have the general result
J
2
i
) = [J
2
i
) = J
i
[J
i
) = |J
i
|
2
0
and hence J
2
) J
2
z
) = J
2
x
) +J
2
y
) 0. But J
2
z
Y
=
2
2
Y
2
. (1.7)
Now we can investigate the eect of J
) = J
(J
2
Y
) =
2
(J
)
so that J
) = (J
J
z
J
)Y
= ( 1)J
and hence J
)
k
Y
) = (J
)
k
(J
z
Y
) k(J
)
k
Y
= ( k)(J
)
k
Y
)
k
Y
)
k
Y
= Y
k
(1.8)
where the normalization is again unspecied.
Thus, starting from a state Y
with a J
2
eigenvalue
2
and a J
z
eigenvalue ,
we can repeatedly apply J
+
to construct an ascending sequence of eigenstates with
J
z
eigenvalues , ( +1), ( +2), . . . , all of which have the same J
2
eigenvalue
2
. Similarly, we can apply J
=
u
Y
u
and J
z
Y
=
l
Y
(1.9a)
with
J
+
Y
u
= 0 and J
= 0 (1.9b)
and where, by (1.7), we must have
2
u
and
2
l
.
4
By construction, there must be an integral number n of steps from
l
to
u
, so
that
l
+
u
= n. (1.10)
(In other words, the eigenvalues of J
z
range over the n intervals
l
,
l
+ 1,
l
+ 2, . . . ,
l
+ (
l
+
u
) =
u
.)
Now, using (1.6b) we have
J
2
Y
u
= J
J
+
Y
u
+ (J
2
z
+ J
z
)Y
u
.
Then by (1.9b) and the denition of Y
, this becomes
2
Y
u
=
2
u
(
u
+ 1)Y
u
so that
=
u
(
u
+ 1) .
In a similar manner, using (1.6a) we have
J
2
Y
= J
+
J
+ (J
2
z
J
z
)Y
or
2
Y
=
2
l
(
l
+ 1)Y
so also
=
l
(
l
+ 1) .
Equating both of these equations for and recalling (1.10) we conclude that
u
=
l
=
n
2
:= j
where j is either integral or half-integral, depending on whether n is even or odd.
In either case, we nally arrive at
= j(j + 1) (1.11)
and the eigenvalues of J
z
range from j to j in integral steps of .
We can now label the eigenvalues of J
z
by m instead of , where the integer or
half-integer m ranges from j to j in integral steps. Thus our eigenvalue equations
may be written
J
2
Y
m
j
= j(j + 1)
2
Y
m
j
J
z
Y
m
j
= mY
m
j
.
(1.12)
We say that the states Y
m
j
are angular momentum eigenstates with angular mo-
mentum j and z-component of angular momentum m. Note that (1.9b) is now
written
J
+
Y
j
j
= 0 and J
Y
j
j
= 0 . (1.13)
5
Since (J
= J
Y
m
j
[J
Y
m
j
) = Y
m
j
[J
Y
m
j
) = Y
m
j
[(J
2
J
2
z
J
z
)Y
m
j
)
=
2
[j(j + 1) m
2
m]Y
m
j
[Y
m
j
)
=
2
[j(j + 1) m(m1)]Y
m
j
[Y
m
j
) .
We know that J
Y
m
j
is proportional to Y
m1
j
. So if we assume that the Y
m
j
are
normalized, then this equation implies that
J
Y
m
j
=
_
j(j + 1) m(m1) Y
m1
j
. (1.14)
If we start at the top state Y
j
j
, then by repeatedly applying J
Y
j
j
=
_
j(j + 1) j(j 1) Y
j1
j
=
_
2j Y
j1
j
or
Y
j1
j
=
1
1
2j
J
Y
j
j
.
Next we have
(J
)
2
Y
j
j
=
2
_
2j
_
j(j + 1) (j 1)(j 2) Y
j2
j
=
2
_
(2j)2(2j 1) Y
j2
j
or
Y
j2
j
=
2
1
_
(2j)(2j 1)2
(J
)
2
Y
j
j
.
And once more should do it:
(J
)
3
Y
j
j
=
3
_
(2j)(2j 1)2
_
j(j + 1) (j 2)(j 3) Y
j3
j
=
3
_
(2j)(2j 1)(2)(3)(2j 2) Y
j3
j
or
Y
j3
j
=
3
1
_
2j(2j 1)(2j 2)3!
(J
)
3
Y
j
j
.
Noting that m = j 3 so that 3! = (j m)! and 2j 3 = 2j (j m) = j +m, it
is easy to see we have shown that
Y
m
j
=
mj
(j +m)!
(2j)!(j m)!
(J
)
jm
Y
j
j
. (1.15a)
6
And an exactly analogous argument starting with Y
j
j
and applying J
+
repeatedly
shows that we could also write
Y
m
j
=
mj
(j m)!
(2j)!(j +m)!
(J
+
)
j+m
Y
j
j
. (1.15b)
It is extremely important to realize that everything we have done up to this
point depended only on the commutation relation (1.2a), and hence applies to both
integer and half-integer angular momenta. While we will return in a later section to
discuss spin (including the half-integer case), for the rest of this section we restrict
ourselves to integer values of angular momentum, and hence we will be discussing
orbital angular momentum.
The next thing we need to do is to actually construct the angular momentum
wave functions Y
m
l
(, ). (Since we are now dealing with orbital angular momen-
tum, we replace j by l.) To do this, we rst need to write L in spherical coordinates.
One way to do this is to start from L
i
= (rp)
i
=
ijk
x
j
p
k
where p
k
= i(/x
k
),
and then use the chain rule to convert from Cartesian coordinates x
i
to spherical
coordinates (r, , ). Using
x = r sin cos y = r sin sin z = r cos
so that
r = (x
2
+y
2
+z
2
)
1/2
= cos
1
z/r = tan
1
y/x
we have, for example,
x
=
r
x
r
+
x
+
x
=
x
r
r
+
xz
r
3
sin
y
x
2
cos
2
= sin cos
r
+
cos cos
r
sin
r sin
with similar expressions for /y and /z. Then using terms such as
L
x
= yp
z
zp
y
= i
_
y
z
z
y
_
we eventually arrive at
L
x
= i
_
sin
cot cos
_
(1.16a)
L
y
= i
_
cos
cot sin
_
(1.16b)
L
z
= i
. (1.16c)
7
However, another way is to start from the gradient in spherical coordinates (see
the section on vector calculus in the handout Supplementary Notes on Mathematics)
=r
r
+
1
r
1
r sin
.
Then L = r p = i r = i r (r ) so that (since r,
and
are
orthonormal)
L = ir
_
r r
r
+r
1
r
+r
1
r sin
_
= i
_
1
sin
_
If we write the unit vectors in terms of their Cartesian components (again, see the
handout on vector calculus)
= (sin , cos , 0)
then
L = i
_
x
_
sin
cot cos
_
+ y
_
cos
cot sin
_
+z
_
which is the same as we had in (1.16).
Using these results, it is now easy to write the ladder operators L
= L
x
iL
y
in spherical coordinates:
L
= e
i
_
i cot
_
. (1.17)
To nd the eigenfunctions Y
m
l
(, ), we start from the denition L
+
Y
l
l
= 0. This
yields the equation
Y
l
l
+i cot
Y
l
l
= 0 .
We can solve this by the usual approach of separation of variables if we write
Y
l
l
(, ) = T()F(). Substituting this and dividing by TF we obtain
1
T cot
T
= i
1
F
F
.
Following the standard argument, the left side of this is a function of only, and
the right side is a function of only. Since varying wont aect the right side,
and varying wont aect the left side, it must be that both sides are equal to a
constant, which I will call k. Now the equation becomes
dF
F
= ikd
8
which has the solution F() = e
ik
(up to normalization). But Y
l
l
is an eigenfunc-
tion of L
z
= i(/) with eigenvalue l, and hence so is F() (since T() just
cancels out). This means that
i
e
ik
= ke
ik
:= le
ik
and therefore we must have k = l, so that (up to normalization)
Y
l
l
= e
il
T() .
With k = l, the equation becomes
dT
T
= l cot d = l
cos
sin
d = l
d sin
sin
.
This is also easily integrated to yield (again, up to normalization)
T() = sin
l
.
Thus, we can write
Y
l
l
= c
l
l
(sin )
l
e
il
where c
l
l
is a normalization constant, xed by the requirement that
_
Y
l
l
2
d = 2
c
l
l
2
_
0
(sin )
2l
sin d = 1 . (1.18)
I will go through all the details involved in doing this integral. You are free to skip
down to the result if you wish (equation (1.21)), but this result is also used in other
physical applications.
First I want to prove the relation
_
sin
n
xdx =
1
n
sin
n1
xcos x +
n 1
n
_
sin
n2
xdx. (1.19)
This is done as an integration by parts (remember the formula
_
u dv = uv
_
v du) letting u = sin
n1
x and dv = sinxdx so that v = cos x and du =
(n 1) sin
n2
xcos xdx. Then (using cos
2
x = 1 sin
2
x in the third line)
_
sin
n
xdx =
_
sin
n1
xsin xdx
= sin
n1
xcos x + (n 1)
_
sin
n2
xcos
2
xdx
= sin
n1
xcos x + (n 1)
_
sin
n2
xdx (n 1)
_
sin
n
xdx.
Now move the last term on the right over to the left, divide by n, and the result is
(1.19).
9
We need to evaluate (1.19) for the case where n = 2l +1. To get the nal result
in the form we want, we will need the basically simple algebraic result
(2l + 1)!! =
(2l + 1)!
2
l
l!
l = 1, 2, 3, . . . (1.20)
where the double factorial is dened by
n!! = n(n 2)(n 4)(n 6) .
There is nothing fancy about the proof of this fact. Noting that n = 2l +1 is always
odd, we have
n!! = 1 3 5 7 9 (n 4) (n 2) n
=
1 2 3 4 5 6 7 8 9 (n 4) (n 3) (n 2) (n 1) n
2 4 6 8 (n 3) (n 1)
=
1 2 3 4 5 6 7 8 9 (n 4) (n 3) (n 2) (n 1) n
(2 1)(2 2)(2 3)(2 4) (2
n3
2
)(2
n1
2
)
=
n!
2
n1
2
(
n1
2
)!
.
Substituting n = 2l + 1 we arrive at (1.20).
Now we are ready to do the integral in (1.18). Since the limits of integration
are 0 and , the rst term on the right side of (1.19) always vanishes, and we can
ignore it. Then we have
_
0
(sin x)
2l+1
dx =
2l
2l + 1
_
0
(sin x)
2l1
dx
=
_
2l
2l + 1
__
2l 2
2l 1
__
0
(sin x)
2l3
dx
=
_
2l
2l + 1
__
2l 2
2l 1
__
2l 4
2l 3
__
0
(sin x)
2l5
dx
= =
_
2l
2l + 1
__
2l 2
2l 1
__
2l 4
2l 3
_
_
2l (2l 2)
2l (2l 3)
__
0
sin xdx
=
2
l
l(l 1)(l 2) (l (l 1))
(2l + 1)!!
2
= 2
2
l
l!
(2l + 1)!!
= 2
(2
l
l!)
2
(2l + 1)!
(1.21)
10
where we used
_
0
sin xdx = 2 and (1.20).
Using this result, (1.18) becomes
4
c
l
l
2 (2
l
l!)
2
(2l + 1)!
= 1
and hence
c
l
l
= (1)
l
_
(2l + 1)!
4
_
1/2
1
2
l
l!
(1.22)
where we included a conventional arbitrary phase factor (1)
l
. Putting this all
together, we have the top orbital angular momentum state
Y
l
l
(, ) = (1)
l
_
(2l + 1)!
4
_
1/2
1
2
l
l!
(sin )
l
e
il
. (1.23)
To construct the rest of the states Y
m
l
(, ), we repeatedly apply L
from equa-
tion (1.17) to nally obtain
Y
m
l
(, ) = (1)
l
_
(2l + 1)!
4
_
1/2
1
2
l
l!
_
(l +m)!
(2l)!(l m)!
_
1/2
e
im
(sin )
m
d
lm
d(cos )
lm
(sin )
2l
. (1.24)
Its just not worth going through this algebra also.
2 Spin
It is an experimental fact that many particles, and the electron in particular, have
an intrinsic angular momentum. This was originally deduced by Goudsmit and Uh-
lenbeck in their analysis of the famous sodium D line, which arises by the transition
from the 1s
2
2s
2
2p
6
3p excited state to the ground state. What initially appears as a
strong single line is slightly split in the presence of a magnetic eld into two closely
spaced lines (Zeeman eect). This (and other lines in the Na spectrum) indicates a
doubling of the number of states available to the valence electron.
To explain this ne structure of atomic spectra, Goudsmit and Uhlenbeck
proposed in 1925 that the electron possesses an intrinsic angular momentum in
addition to the orbital angular momentum due to its motion about the nucleus.
Since magnetic moments are the result of current loops, it was originally thought
that this was due to the electron spinning on its axis, and hence this intrinsic angular
momentum was called spin. However, a number of arguments can be put forth to
disprove that classical model, and the result is that we must assume that spin is a
purely quantum phenomena without a classical analogue.
As I show at the end of this section, the classical model says that the magnetic
moment of a particle of charge q and mass m moving in a circle is given by
=
q
2mc
L
11
where L is the angular momentum with magnitude L = mvr. Furthermore, the
energy of such a charged particle in a magnetic eld B is B. Goudsmit and
Uhlenbeck showed that the ratio of magnetic moment to angular momentum of the
electron was in fact twice as large as it would be for an orbital angular momentum.
(This factor of 2 is explained by the relativistic Dirac theory.) And since we know
that a state with angular momentum l is (2l + 1)-fold degenerate, the splitting
implies that the electron has an angular momentum /2.
From now on, we will assume that spin is described by the usual angular momen-
tum theory, and hence we postulate a spin operator S and corresponding eigenstates
[s m
s
) such that
[S
i
, S
j
] = i
ijk
S
k
(2.1a)
S
2
[s m
s
) = s(s + 1)
2
[s m
s
) (2.1b)
S
z
[s m
s
) = m
s
[s m
s
) (2.1c)
S
[s m
s
) =
_
s(s + 1) m
s
(m
s
1) [s m
s
1) . (2.1d)
(I am switching to the more abstract notation for the eigenstates because the states
themselves are a rather abstract concept.) We will sometimes drop the subscript s
on m
s
if there is no danger of confusing this with the eigenvalue of L
z
, which we
will also sometimes write as m
l
. Be sure to realize that particles can have a spin s
that is any integer multiple of 1/2, and there are particles in nature that have spin
0 (e.g., the pion), spin 1/2 (e.g., the electron, neutron, proton), spin 1 (e.g., the
photon, but this is a little bit subtle), spin 3/2 (the s), spin 2 (the hypothesized
graviton) and so forth.
Since the z component of electron spin can take only one of two values /2, we
will frequently denote the corresponding orthonormal eigenstates simply by [z ),
where [z +) is called the spin up state, and [z ) is called the spin down state.
An arbitrary spin state [) is of the form
[) = c
+
[z +) +c
[z ) . (2.2a)
If we wish to think in terms of explicit matrix representations, then we will write
this in the form
= c
+
(z)
+
+c
(z)
. (2.2b)
Be sure to remember that the normalized states [s m
s
) belong to distinct eigenvalues
of a Hermitian operator, and hence they are in fact orthonormal.
To construct the matrix representation of spin operators, we need to rst choose
a basis for the space of spin states. Since for the electron there are only two possible
states for the z component, we need to pick a basis for a two-dimensional space,
and the obvious choice is the standard basis
(z)
+
=
_
1
0
_
and
(z)
=
_
0
1
_
. (2.3)
12
With this choice of basis, we can now construct the 2 2 matrix representation of
the spin operator S.
Note that the existence of spin has now led us to describe the electron by a
multi-component state vector (in this case, two components), as opposed to the
scalar wave functions we used up to this point. These two-component states are
frequently called spinors.
The states
(z)
2
so that
S
z
=
2
_
1 0
0 1
_
. (2.4)
(We are being somewhat sloppy with notation and using the same symbol S
z
to
denote both the operator and its matrix representation.) That the vectors dened
in (2.3) are indeed eigenvectors of S
z
is easy to verify:
2
_
1 0
0 1
_ _
1
0
_
=
2
_
1
0
_
and
2
_
1 0
0 1
__
0
1
_
=
2
_
0
1
_
.
To nd the matrix representations of S
x
and S
y
, we use (2.1d) together with
S
= S
x
iS
y
so that S
x
= (S
+
+S
)/2 and S
y
= (S
+
S
)/2i. From
S
[z m
s
) =
_
3/4 m
s
(m
s
1) [z m
s
1)
we have
S
+
[z +) = 0 S
[z +) = [z )
S
+
[z ) = [z +) S
[z ) = 0 .
Therefore the only non-vanishing entry in S
+
is z + [S
+
[z ), and the only non-
vanishing entry in S
is z [S
=
_
0 0
1 0
_
.
Using these, it is easy to see that
S
x
=
2
_
0 1
1 0
_
and S
y
=
2
_
0 i
i 0
_
. (2.5)
13
And from (2.4) and (2.5) it is easy to calculate S
2
= S
2
x
+S
2
y
+S
2
z
to see that
S
2
=
3
4
2
_
1 0
0 1
_
which agrees with (2.1b).
It is conventional to write S in terms of the Pauli spin matrices dened by
S =
2
where
x
=
_
0 1
1 0
_
y
=
_
0 i
i 0
_
z
=
_
1 0
0 1
_
. (2.6)
Memorize these. The Pauli matrices obey several relations that I leave to you to
verify (recall that the anticommutator is dened by [a, b]
+
= ab +ba):
[
i
,
j
] = 2i
ijk
k
(2.7a)
j
= i
ijk
k
for i ,= j (2.7b)
[
i
,
j
]
+
= 2I
ij
(2.7c)
j
= I
ij
+i
ijk
k
(2.7d)
Given three-component vectors a and b, equation (2.7d) also leads to the extremely
useful result
(a )(b ) = (a b)I +i(a b) . (2.8)
We will use this later when we discuss rotations.
From the standpoint of physics, what equations (2.2) say is that if we have
an electron (or any spin one-half particle) in an arbitrary spin state , then the
probablility is [c
+
[
2
that a measurement of the z component of spin will result in
+/2, and the probability is [c
[
2
that the measurement will yield /2. We can
say this because (2.2) expresses as a linear superposition of eigenvectors of S
z
.
But we could equally well describe in terms of eigenvectors of S
x
. To do so,
we simply diagonalize S
x
and use its eigenvectors as a new set of basis vectors for
our two-dimensional spin space. Thus we must solve S
x
v = v for and then the
eigenvectors v. This is straightforward. The eigenvalue equation is (S
x
I)v = 0,
so in order to have a non-trivial solution we must have
det(S
x
I) =
/2
/2
=
2
2
/4 = 0
so that
+
= +/2 we solve
(S
x
+
I)v =
2
_
1 1
1 1
_ _
a
b
_
= 0
14
so that a = b and the normalized eigenvector is (we now write v =
(x)
(x)
+
=
1
2
_
1
1
_
. (2.9a)
For
= /2 we have
(S
x
I)v =
2
_
1 1
1 1
__
c
d
_
= 0
so that c = d and the normalized eigenvector is now (where we arbitrarily choose
c = +1)
(x)
=
1
2
_
1
1
_
. (2.9b)
To understand just what this means, suppose we have an arbitrary spin state
(normalized so that [[
2
+[[
2
= 1)
=
_
_
.
Be sure to understand that this is a vector in a two-dimensional space, and it exists
independently of any basis. In terms of the basis (2.3) we can write
=
_
1
0
_
+
_
0
1
_
=
(z)
+
+
(z)
_
=
a
2
_
1
1
_
+
b
2
_
1
1
_
=
_
a/
2 +b/
2
a/
2 b/
2
_
or
= a/
2 +b/
2 and = a/
2 b/
2 .
Solving for a and b in terms of and we obtain
a =
1
2
( +) and b =
1
2
( )
so that
=
_
+
2
_
(x)
+
+
_
2
_
(x)
. (2.10)
15
Thus the probability of measuring the x component of spin to be +/2 is [ +[
2
/2,
and the probability of measuring the value to be /2 is [ [
2
/2.
(Remark: What we just did was nothing more than the usual change of basis in
a vector space. We started with a basis
e
1
=
_
1
0
_
and e
2
=
_
0
1
_
(which we chose to be the eigenvectors of S
z
) and changed to a new basis
e
1
=
1
2
_
1
1
_
and e
2
=
1
2
_
1
1
_
(which were the eigenvectors of S
x
). Since e
1
= (e
1
+e
2
)/
2 and e
2
= (e
1
e
2
)/
2,
we see that this change of basis is described by the transition matrix dened by
e
i
= e
j
p
j
i
or
P =
1
2
_
1 1
1 1
_
= P
1
.
Then a vector
=
_
_
can be written in terms of either the basis e
i
as
=
_
1
0
_
+
_
0
1
_
=
1
e
1
+
2
e
2
or in terms of the basis e
i
as =
1
e
1
+
2
e
2
where
i
= (p
1
)
i
j
j
. This then
immediately yields
=
1
2
( +) e
1
+
1
2
( ) e
2
which is just (2.10).)
Now we ask how to incorporate spin into the general solution to the Schrodinger
equation for the hydrogen atom. Let us ignore terms that couple spin with the
orbital angular momentum of the electron. (This LS coupling is relatively small
compared to the electron binding energy, and can be ignored to rst order. We will,
however, take this into account when we discuss perturbation theory.) Under these
conditions, the Hamiltonian is still separable, and we can write the total stationary
state wave function as a product of a spatial part
nlm
l
times a spin part (m
s
).
Thus we can write the complete hydrogen atom wave function in the form
nlm
l
ms
=
nlm
l
(m
s
) .
Since the Hamiltonian is independent of spin, we have
H = H[
nlm
l
(m
s
)] = (m
s
)H
nlm
l
= E
n
[(m
s
)
nlm
l
] = E
n
16
so that the energies are unchanged. However, because of the spin function, we have
doubled the number of states corresponding to a given energy.
A more mathematically correct way to write these complete states is as the
tensor (or direct) product
[) = [
nlm
l
(m
s
)) := [
nlm
l
) [(m
s
)) .
In this case, the Hamiltonian is properly written as H I where H acts on the
vector space of spatial wave functions, and I is the identity operator on the vector
space of spin states. In other words,
(H I)([
nlm
l
) [(m
s
))) := H[
nlm
l
) I[(m
s
)) .
This notation is particularly useful when treating two-particle states, as we will see
when we discuss the addition of angular momentum.
(Remark: You may recall from linear algebra that given two vector spaces V
and V
V V
) V V
x
i
v
i
and v
y
j
v
j
then
v v
x
i
y
j
(v
i
v
j
). In particular, if V has basis e
i
and V
has basis e
j
,
then e
i
e
j
is a basis for V V
)
and called the direct (or tensor) product of V and V
by (A B)(v v
) := A(v) B(v
).)
2.1 Supplementary Topic: Magnetic Moments
Consider a particle of charge q moving in a circular orbit. It forms an eective
current
I =
q
t
=
q
2r/v
=
qv
2r
.
By denition, the magnetic moment has magnitude
=
I
c
area =
qv
2rc
r
2
=
qvr
2c
.
But the angular momentum of the particle is L = mvr so we conclude that the
magnetic moment due to orbital motion is
l
=
q
2mc
L. (2.11)
The ratio of to L is called the gyromagnetic ratio.
While the above derivation of (2.11) was purely classical, we know that the
electron also possesses an intrinsic spin angular momentum. Let us hypothesize
that the electron magnetic moment associated with this spin is of the form
s
= g
e
2mc
S.
17
The constant g is found by experiment to be very close to 2. (However, the rel-
ativistic Dirac equation predicts that g is exactly 2. Higher order corrections in
quantum electrodynamics predict a slightly dierent value, and the measurement
of g 2 is one of the most accurate experimental result in all of physics.)
Now we want to show is that the energy of a magnetic moment in a uniform
magnetic eld is given by B where for a loop of area A carrying current
I is dened to have magnitude IA and pointing perpendicular to the loop in the
direction of your thumb if the ngers of your right hand are along the direction of
the current. To see this, we simply calculate the work required to rotate a current
loop from its equilibrium position to the desired orientation.
Consider the gure shown below, where the current ows counterclockwise out
of the page at the bottom and into the page at the top.
B
B
B
FB
FB
a/2
a/2
Let the loop have length a on the sides and b across the top and bottom, so its area
is ab. The magnetic force on a current-carrying wire is
F
B
=
_
Idl B
and hence the forces on the opposite a sides of the loop cancel, and the force on
the top and bottom b sides is F
B
= IbB. The equilibrium position of the loop is
horizontal, so the potential energy of the loop is the work required to rotate it from
= 0 to some value . This work is given by W =
_
F dr where F is the force that
I must apply against the magnetic eld to rotate the loop.
Since the loop is rotating, the force I must apply at the top of the loop is in the
direction of and perpendicular to the loop, and hence has magnitude F
B
cos .
Then the work I do is (the factor of 2 takes into account both the top and bottom
sides)
W =
_
F dr = 2
_
F
B
cos (a/2)d = IabB
_
0
cos d = Bsin .
But note that B = Bcos(90 +) = Bsin , and therefore
W = B. (2.12)
18
In this derivation, I never explicitly mentioned the torque on the loop due to B.
However, we see that
|N| = |r F
B
| = 2(a/2)F
B
sin(90 +) = IabBsin(90 +)
= Bsin(90 +) = | B|
and therefore
N = B. (2.13)
Note that W =
_
|N| d. We also see that
dL
dt
=
d
dt
r p = r
dp
dt
= r F
where we used p = mv and r p = v p = 0. Therefore, as you should already
know,
dL
dt
= N. (2.14)
3 Mathematical Digression:
Rotations and Linear Transformations
Lets take a look at how the spatial rotation operator is dened. Note that there
are two ways to view symmetry operations such as translations and rotations. The
rst is to leave the coordinate system unchanged and instead move the physical
system. This is called an active transformation. Alternatively, we can leave the
physical system alone and change the coordinate system, for example by translation
or rotation. This is called a passive transformation. In the case of an active
transformation, we have the following situation:
x1
x2
r
r
Here the vector r is rotated by to give the vector r where, of course, |r| = |r| = r.
We dene a linear transformation T by r = T(r). (This is linear because it is easy
to see that rotating the sum of two vectors is the sum of the rotated vectors.) From
the diagram, the components of r are given by
x
1
= r cos( +) = r cos cos r sin sin
= (cos )x
1
(sin )x
2
19
x
2
= r sin( +) = r sin cos +r cos sin
= (sin )x
1
+ (cos )x
2
or
_
x
1
x
2
_
=
_
cos sin
sin cos
__
x
1
x
2
_
. (3.1)
Since T is a linear transformation, it is completely specied by dening its values
on a basis because
T(r) = T
_
i
x
i
e
i
_
=
i
x
i
Te
i
.
But Te
i
is just another vector in R
2
, and hence it can be expressed in terms of the
basis e
i
as
Te
i
=
j
e
j
a
ji
. (3.2)
Be sure to note which indices are summed over in this expression. The matrix (a
ji
)
is called the matrix representation of the linear transformation T with respect
to the basis e
i
. You will sometimes see this matrix written as [T]
e
.
It is very important to realize that a linear transformation T takes the ith basis
vector into the ith column of its matrix representation. This is easy to see if we
write out the components of (3.2). Simply note that with respect to the basis e
i
,
we have
e
1
=
_
1
0
_
and e
2
=
_
0
1
_
and therefore
Te
i
= e
1
a
1i
+e
2
a
2i
=
_
1
0
_
a
1i
+
_
0
1
_
a
2i
=
_
a
1i
a
2i
_
which is just the ith column of the matrix A = (a
ji
).
As an example, let V have a basis v
1
, v
2
, v
3
, and let T : V V be the linear
transformation dened by
Tv
1
= 3v
1
+v
3
Tv
2
= v
1
2v
2
v
3
Tv
3
= v
2
+v
3
Then the representation of T (relative to this basis) is
[T]
v
=
_
_
3 1 0
0 2 1
1 1 1
_
_
.
Now let V be an n-dimensional vector space, and let W be a subspace of V .
Let T be an operator on V , and suppose W has the property that Tw W for
20
every w W. Then we say that W is T-invariant (or simply invariant when the
operator is understood).
What can we say about the matrix representation of T under these circum-
stances? Well, let W have the basis w
1
, . . . , w
r
, and extend this to a basis
w
1
, . . . , w
r
, v
1
, . . . , v
nr
for V . Since Tw W for any w W, we must have
Tw
i
=
r
j=1
w
j
a
ji
for some set of scalars a
ji
. But for any v
i
there is no such restriction since all we
know is that Tv
i
V . So we have
Tv
i
=
r
j=1
w
j
b
ji
+
nr
k=1
v
k
c
ki
for scalars b
ji
and c
ki
. Then since T takes the ith basis vector to the ith column
of the matrix representation [T], we must have
[T] =
_
A B
0 C
_
where A is an r r matrix, B is an r (nr) matrix, and C is an (nr) (nr)
matrix. Such a matrix is said to be a block matrix, and [T] is in block triangular
form.
Now let W be an invariant subspace of V . If it so happens that the subspace
of V spanned by the rest of the vectors v
1
, . . . , v
nr
is also invariant, then the
matrix representation of T will be block diagonal (because all of the b
ji
in the above
expansion will be zero). As we shall see, this is in fact what happens when we add
two angular momenta J
1
and J
2
and look at the representation of the rotation
operator with respect to the total angular momentum states (where J = J
1
+ J
2
).
By choosing our states to be eigenstates of J
2
and J
z
rather than J
1z
and J
2z
, the
rotation operator becomes block diagonal rather than a big mess. This is because
rotations dont change j, so for xed j, the (2j + 1)-dimensional space spanned by
[j j), [j j 1), . . . , [j j) is an invariant subspace under rotations. This change
of basis is exactly what Clebsch-Gordan coecients accomplish.
Let us go back to our specic example of rotations. If we dene r = T(r), then
on the the one hand we have
r =
j
x
j
e
j
while on the other hand, we can write
r = T(r) =
i
x
i
T(e
i
) =
j
e
j
a
ji
x
i
.
21
Since the e
j
are a basis, they are linearly independent, and we can equate these last
two equations to conclude that
x
j
=
i
a
ji
x
i
. (3.3)
Note which indices are summed over in this equation.
Comparing (3.3) with (3.1), we see that the matrix in (3.1) is the matrix repre-
sentation of the linear transformation T dened by
( x
1
, x
2
) = T(x
1
, x
2
) = ((cos )x
1
(sin )x
2
, (sin )x
1
+ (cos )x
2
) .
Then the rst column of [T] is
T(e
1
) = T(1, 0) = (cos , sin )
and the second column is
T(e
2
) = T(0, 1) = (sin , cos )
so that
[T] =
_
cos sin
sin cos
_
as in (3.1).
Using (3.3) we can make another extremely important observation. Since the
length of a vector is unchanged under rotations, we must have
j
x
j
x
j
=
i
x
i
x
i
.
But from (3.3) we see that
j
x
j
x
j
=
i,k
a
ji
a
jk
x
i
x
k
and hence it follows that
j
a
ji
a
jk
=
j
a
T
ij
a
jk
=
ik
. (3.4a)
In matrix notation, this is
A
T
A = I .
Since we are in a nite-dimensional space, this implies that AA
T
= I also. In
other words, A
T
= A
1
. Such a matrix is said to be orthogonal. In terms of
components, this second condition is
j
a
ij
a
T
jk
=
j
a
ij
a
kj
=
ik
. (3.4b)
It is also quite useful to realize that an orthogonal matrix is one whose rows (or
columms) form an orthonormal set of vectors. If we let A
i
denote the ith row of A,
then from (3.4b) we have
A
i
A
k
=
j
a
ij
a
kj
=
ik
.
22
Similarly, if we let A
i
denote the ith column of A, then (3.4a) may be written
A
i
A
k
=
j
a
ji
a
jk
=
ik
.
Conversely, it is easy to see that any matrix whose rows (or columns) form an
orthonormal set must be an orthogonal matrix.
In the case of complex matrices, repeating the above arguments using |x|
2
=
i
x
i
x
i
, it is not hard to show that A
A = AA
= I where A
= A
T
. In this case,
the matrix A is said to be unitary. Thus a complex matrix is unitary if and only if
its rows (or columns) form an orthonormal set under the standard Hermitian inner
product.
To describe a passive transformation, consider a linear transformation P acting
on the basis vectors. In other words, we perform a change of basis dened by
e
i
= P(e
i
) =
j
e
j
p
ji
.
In this situation, the linear transformation P is called the transition matrix.
Suppose we have a linear operator A dened on our space. Then with respect to
a basis e
i
this operator has the matrix representation (a
ij
) dened by (dropping
the parenthesis for simplicity)
Ae
i
=
j
e
j
a
ji
.
And with respect to another basis e
i
it has the representation ( a
ij
) dened by
A e
i
=
j
e
j
a
ji
.
But e
i
= Pe
i
, so the left side of this equation may be written
A e
i
= APe
i
= A
_
j
e
j
p
ji
_
=
j
p
ji
Ae
j
=
j,k
p
ji
e
k
a
kj
=
j,k
e
k
a
kj
p
ji
while the right side is
j
e
j
a
ji
=
j
(Pe
j
) a
ji
=
j,k
e
k
p
kj
a
ji
.
Equating these two expressions and using the linear independence of the e
k
we have
j
p
kj
a
ji
=
j
a
kj
p
ji
which in matrix notation is just P
A = AP. Since each basis may be written in
terms of the other, the matrix P must be nonsingular so that P
1
exists, and hence
we have shown that
A = P
1
AP . (3.5)
23
This extremely important equation relates the matrix representation of an op-
erator A with respect to a basis e
i
to its representation
A with respect to a basis
e
i
dened by e
i
= Pe
i
. This is called a similarity transformation. In fact, this
is exactly what you do when you diagonalize a matrix. Starting with a matrix A
relative to a given basis, you rst nd its eigenvalues, and then use these to nd the
corresponding eigenvectors. Letting P be the matrix whose columns are precisely
these eigenvectors, we then have P
1
AP = D where D is a diagonal matrix with
diagonal entries that are just the eigenvalues of A.
4 Angular Momentum and Rotations
4.1 Angular Momentum as the Generator of Rotations
Before we begin with the physics, let me rst prove an extremely useful mathemat-
ical result. For any complex number (or matrix) x, let f
n
(x) = (1 + x/n)
n
, and
consider the limit
L = lim
n
f
n
(x) = lim
n
_
1 +
x
n
_
n
.
Since the logarithm is a continuous function, we can interchange limits and the log,
so we have
ln L = ln lim
n
_
1 +
x
n
_
n
= lim
n
ln
_
1 +
x
n
_
n
= lim
n
nln
_
1 +
x
n
_
= lim
n
ln(1 +x/n)
1/n
.
As n , both the numerator and denominator go to zero, so we apply
lHopitals rule and take the derivative of both numerator and denominator with
respect to n:
ln L = lim
n
(x/n
2
)/(1 +x/n)
1/n
2
= lim
n
x
1 +x/n
= x.
Exponentiating both sides, we have proved
lim
n
_
1 +
x
n
_
n
= e
x
. (4.1)
As I mentioned above, this result also applies if x is an n n complex matrix
A. This is a consequence of the fact (which I state without proof) that for any such
matrix, the series
n=0
A
n
n!
converges, and we take this as the denition of e
A
. (For a proof, see one of my
linear algebra books or almost any book on elementary real analysis or advanced
calculus.)
24
Now let us return to physics. First consider the translation of a wave function
(x) by an amount a.
x x + a
(x)
(x)
This results in a new function
(x +a) = (x), or
(x) = (x a) .
For innitesimal a, we expand this in a Taylor series to rst order in a:
(x a) = (x) a
d
dx
.
Using p = i(d/dx), we can write this in the form
(x a) =
_
1
ia
p
_
(x) .
For a nite translation, we consider a sequence of n innitesimal translations
a/n and let n . Applying (4.1) we have
(x) = (x a) = lim
n
_
1
iap/
n
_
n
(x) = e
iap/
(x) . (4.2)
Thus we see that translated states
= R()x where R() is the matrix that represents the rotation. In two
dimensions this is
_
x
_
=
_
cos sin
sin cos
__
x
y
_
.
If we have a scalar wavefunction (x), then under rotation we obtain a new wave-
function
R
(x), where (x) =
R
(R()x) =
R
(x
(x)
R(x)
Alternatively, we can write
R
(x) = (R
1
()x) .
Since R is an orthogonal transformation (it preserves the length of x) we know
that R
1
() = R
T
() (this really should be written as R()
1
= R()
T
, but it
rarely is), and in the case where 1 we then have
R
1
()x =
_
1
1
__
x
y
_
=
_
x +y
x +y
_
.
Expanding (R
1
()x) with these values for x and y we have (letting
i
= /x
i
for simplicity)
R
(x) = (x +y, y x) = (x) [x
y
y
x
](x)
26
or, using p
i
= i
i
this is
R
(x) = (x)
i
[xp
y
yp
x
](x) =
_
1
i
L
z
_
(x) .
Thus we see that angular momentum is indeed the generator of rotations.
For nite we exponentiate this to write
R
(x) = e
(i/)Lz
(x), and in the
case of an arbitrary angle in R
3
this becomes
R
(x) = e
(i/)L
(x) . (4.4)
In an abstract notation we write this as
[
R
) = U(R)[)
where U(R) = e
(i/)L
. For simplicity and clarity, we have written U(R) rather
than the more complete U(R()), which we continue to do unless the more complete
notation is needed.
What we just did was for orbital angular momentum. In the case of spin there
is no classical counterpart, so we dene the spin angular momentum operator S to
obey the usual commutation relations, and then the spin states will transform under
the rotation operator e
(i/)S
. (We will come back to justifying this at the end
of this section.) It is common to use the symbol J to stand for any type of angular
momentum operator, for example L, S or L +S, and this is what we shall do from
now on. In the case where J = L + S, J is called the total angular momentum
operator.
(The example above applied to a scalar wavefunction , which represents a
spinless particle. Particles with spin are described by vector wavefunctions , and in
this case the spin operator S serves to mix up the components of under rotations.)
The angular momentum operators J
2
= J J and J
z
commute, and hence they
have simultaneous eigenstates, denoted by [jm), with the property that (with = 1)
J
2
[jm) = j(j + 1)[jm) and J
z
[jm) = m[jm)
where m takes the 2j + 1 values j m j. Since the rotation operator is given
by U(R) = e
iJ
we see that [U(R), J
2
] = 0. Then
J
2
U(R)[jm) = U(R)J
2
[jm) = j(j + 1)U(R)[jm)
so that the magnitude of the angular momentum cant change under rotations.
However, [U(R), J
z
] ,= 0 so the rotated state will no longer be an eigenstate of J
z
with the same eigenvalue m.
Note that acting to the right we have the matrix element
j
[J
2
U(R)[jm) = j
[U(R)J
2
[jm) = j(j + 1)j
[U(R)[jm)
while acting to the left gives
j
[J
2
U(R)[jm) = j
(j
+ 1)j
[U(R)[jm)
27
and therefore
j
[U(R)[jm) = 0 unless j = j
. (4.5)
We also make note of the fact that acting with J
2
and J
z
in both directions yields
j
[J
2
[jm) = j
(j
+ 1)j
[jm)
and
jm
[J
z
[jm) = m
jm
[jm) = mjm
[jm)
so that (as you should have already known)
j
[jm) =
j
m
. (4.6)
In other words, the states [jm) form a complete orthonormal set, and the state
U(R)[jm) must be of the form
U(R)[jm) =
[jm
)jm
[U(R)[jm) =
[jm
)D
(j)
m
m
() (4.7)
where
D
(j)
m
m
() := jm
[U(R)[jm) = jm
[e
iJ
[jm) . (4.8)
(Notice the order of susbscripts in the sum in equation (4.7). This is the same as
the usual denition of the matrix representation [T]
e
= (a
ij
) of a linear operator
T : V V dened by Te
i
=
j
e
j
a
ji
.)
Since for each j there are 2j + 1 values of m, we have constructed a (2j +
1) (2j + 1) matrix D
(j)
() for each value of j. This matrix is referred to as the
jth irreducible representation of the rotation group. The word irreducible
means that there is no subset of the space of states [jj), [j, m 1), . . . , [j, j)
that transforms into itself under all rotations U(R()). Put in another way, a
representation is irreducible if the vector space on which it acts has no invariant
subspaces.
Now, it is a general result of the theory of group representations that any repre-
sentation of a nite group or compact Lie group is equivalent to a unitary represen-
tation, and any reducible unitary representation is completely reducible. Therefore,
any representation of a nite group or compact Lie group is either already irre-
ducible or else it is completely reducible (i.e., the space on which the operators act
can be put into block diagonal form where each block corresponds to an invariant
subspace). However, at this point we dont want to get into the general theory
of representations, so let us prove directly that the representations D
(j)
() of the
rotation group are irreducible.
Recall that the raising and lowering operators J
are dened by
J
[jm) = (J
x
iJ
y
)[jm) =
_
j(j + 1) m(m1) [j, m1) .
In particular, the operators J
[u) = (J
x
iJ
y
)[u) =
_
1
i
[1 U(R( x))] i
_
1
i
[1 U(R( y))]
__
[u)
=
1
_
[1 U(R( y))] i +iU(R( x))
_
[u)
V
by denition of
V and vector spaces. Since J
)
2
acting on [u) is again
some other linear combination of rotations acting on [u) and hence is also in
V . So
in general, we see that (J
)
n
[u) is again in
V .
By denition of V , we may write (since j is xed)
[u) =
m
[jm)jm[u) =
m
[m)m[u)
= [m)m[u) +[m + 1)m+ 1[u) + + [j)j[u)
where m is simply the smallest value of m for which m[u) , = 0 (and not all of
the terms up to j[u) are necessarily nonzero). Acting on this with J
+
we obtain
(leaving o the constant factors and noting that J
+
[j) = 0)
J
+
[u) [m+ 1)m[u) +[m + 2)m+ 1[u) + +[j)j 1[u)
V .
Since m[u) , = 0 by assumption, it follows that [m+ 1)
V .
We can continue to act on [u) with J
+
a total of j m times at which point we
will have shown that [m + j m) = [j) := [jj)
V . Now we can apply J
2j + 1
times to [jj) to conclude that the 2j + 1 vectors [jm) all belong to
V , and thus
)
2j+1
(J
+
)
jm
to [u), and each step along the way is just some vector in
V .)
As the last subject of this section, let us go back and show that e
(i/)S
really
represents a rotation in spin space. Let R denote the unitary rotation operator, and
29
let A be an arbitrary observable. Rotating our system, a state [) is transformed
into a rotated state [
R
) = R[). In the original system, a measurement of A
yields the result [A[). Under rotations, we measure a rotated observable A
R
in
the rotated states. Since the physical results of a measurement cant change just
because of our coordinate description, we must have
R
[A
R
[
R
) = [A[) .
But
R
[A
R
[
R
) = [R
A
R
R[)
and hence we have R
A
R
R = A or
A
R
= RAR
= RAR
1
. (4.9)
(Compare this to (3.5). Why do you think there is a dierence?)
Now, what about spin one-half particles? To say that a measurement of the spin
in a particular direction m can only yield one of two possible results means that the
operator S m has only the two eigenvalues /2. Let us denote the corresponding
states by [ m). In other words, we have
(S m)[ m) =
2
[ m) .
(This is just the generalization of (2.1c) for the case of spin one-half. It also applies
to arbitrary spin if we write (S m)[ mm
s
) = m
s
[ mm
s
).)
Let us rotate the unit vector m by to obtain another unit vector n, and
consider the operator
e
(i/)S
(S m)e
(i/)S
.
Acting on the state e
(i/)S
[ m) we have
[e
(i/)S
(S m)e
(i/)S
]e
(i/)S
[ m) = e
(i/)S
(S m)[ m)
=
2
e
(i/)S
[ m) .
Therefore, if we dene the rotated state
[ n) := e
(i/)S
[ m) (4.10a)
we see from (4.9) that we also have the rotated operator
S n := e
(i/)S
(S m)e
(i/)S
(4.10b)
with the property that
(S n)[ n) =
2
[ n) (4.10c)
as it should. This shows that e
(i/)S
does indeed act as the rotation operator on
the abstract spin states.
30
What we have shown then, is that starting from an eigenstate [ m) of the spin
operator S m, we rotate by an angle to obtain the rotated state [ n) that is an
eigenstate of S n with the same eigenvalues.
For spin one-half we have S = /2 (with = 1), and using (2.8), you can show
that the rotation operator becomes
e
i/2
=
m=0
(i /2)
n
n!
= I cos
2
i
sin
2
. (4.11)
Example 4.1. Let us derive equations (2.9) using our rotation formalism. We start
by letting m = z and n = x, so that = (/2) y. Then the rotation operator is
R() = e
iy/4
= I cos
4
i
y
sin
4
=
_
cos /4 sin/4
sin /4 cos /4
_
=
1
2
_
1 1
1 1
_
.
Acting on the state
[ m+) = [z +)
we have
R()[z +) =
1
2
_
1 1
1 1
__
1
0
_
=
1
2
_
1
1
_
= [ x+)
which is just (2.9a). Similarly, it is also easy to see that
R()[z ) =
1
2
_
1 1
1 1
__
0
1
_
=
1
2
_
1
1
_
= [ x)
which is the same as (2.9b) up to an arbitrary phase.
So far we have shown that the rotation operator R() = R((/2) y) indeed takes
the eigenstates of S
z
into the eigenstates of S
x
. Let us also show that it takes
S
z
=
z
/2 into S
x
=
x
/2. This is a straightforward computation:
e
iy/4
z
e
iy/4
=
1
2
_
1 1
1 1
_ _
1 0
0 1
_
1
2
_
1 1
1 1
_
=
1
2
_
0 2
2 0
_
=
_
0 1
1 0
_
=
x
as it should.
31
4.2 Spin Dynamics in a Magnetic Field
As we mentioned at the beginning of Section 2, a particle of charge q and mass m
moving in a circular orbit has a magnetic moment given classically by
l
=
q
2mc
L.
This relation is also true in quantum mechanics for orbital angular momentum, but
for spin angular momentum we must write
s
= g
q
2mc
S (4.12)
where the constant g is called a g-factor. For an electron, g is found by experiment
to be very close to 2. (The Dirac equation predicts that g = 2 exactly, and higher
order correction in QED show a slight deviation from this value.) And for a proton,
g = 5.59. Furthermore, electrically neutral particles such as the neutron also have
magnetic moments, which you can think of as arising from some sort of internal
charge distribution or current. (But dont think too hard.)
In the presence of a magnetic eld B(t), a magnetic moment feels an applied
torque B, and hence possesses a potential energy B. Because of this torque,
the magnetic moment (and hence the particle) will precess about the eld. (Recall
that torque is equal to the time rate of change of angular momentum, and hence a
non-zero torque means that the angular momentum vector of the particle must be
changing.) Let us see what we can learn about this precession.
We restrict consideration to particles essentially at rest (i.e., with no angular
momentum) in a uniform magnetic eld B(t), and write
s
= S where = gq/2mc
is called the gyromagnetic ratio. (If the eld isnt uniform, there will be a force
equal to the negative gradient of the potential energy, and the particle will move in
space. This is how the Stern-Gerlach experiment works, as we will see below.) In
the case of a spin one-half particle, the Hamiltonian given by
H
s
=
s
B(t) =
gq
2mc
S B(t) =
gq
4mc
B(t) . (4.13)
To understand how spin behaves, we look at this motion from the point of view of
the Heisenberg picture, which we now describe. (See Griths, Example 4.3 for a
somewhat dierent approach.)
The formulation of quantum mechanics that we have used so far is called the
Schr odinger picture (SP). In this formulation, the operators are independent of
time, but the states (wave functions) evolve in time. The stationary state solutions
to the Schrodinger equation
H[(t)) = i
t
[(t))
(where H is independent of time) are given by
[(t)) = e
iEt/
[) = e
iHt/
[) (4.14)
32
where H[) = E[) and [) is independent of time.
In the Heisenberg picture (HP), the states are independent of time, and the
operators evolve in time according to an equation of motion. To derive the equations
of motion for the operators in this picture, we freeze out the states at time t = 0:
[
H
) := [(0)) = [) = e
+iHt/
[(t)) . (4.15a)
In the SP, the expectation value of an observable O (possibly time dependent) is
given by
O) = (t)[O[(t)) = [e
iHt/
Oe
iHt/
[) .
In the HP, we want the same measurable result for O), so we have
O) =
H
[O
H
[
H
) = [O
H
[) .
Equating both versions of these expectation values, we conclude that
O
H
= e
iHt/
Oe
iHt/
(4.15b)
where O
H
= O
H
(t) is the representation of the observable O in the HP. Note in
particular that the Hamiltonian is the same in both pictures because
H
H
= e
iHt/
He
iHt/
= e
iHt/
e
iHt/
H = H .
Now take the time derivative of (4.15b) and use the fact that H commutes with
e
iHt/
:
dO
H
dt
=
i
He
iHt/
Oe
iHt/
e
iHt/
Oe
iHt/
H +e
iHt/
O
t
e
iHt/
=
i
[H, O
H
] +e
iHt/
O
t
e
iHt/
.
If O has no explicit time dependence (for example, the operator pt has explicit time
dependence), then this simplies to the Heisenberg equation of motion
dO
H
dt
=
i
[H, O
H
] . (4.16)
Also note that if H is independent of time, then so is H
H
= H, and (4.16) then
shows that H = const and energy is conserved.
Returning to our problem, we use the Hamiltonian (4.13) in the equation of
motion (4.16) to write
dS
i
dt
=
i
[H
s
, S
i
] =
i
gq
2mc
B
j
[S
j
, S
i
] =
i
gq
2mc
B
j
i
jik
S
k
= +
gq
2mc
ikj
S
k
B
j
=
gq
2mc
(S B)
i
or
dS(t)
dt
=
gq
2mc
S(t) B(t) =
s
(t) B(t) . (4.17)
33
This is just the operator version of the classical equation that says the time rate of
change of angular momentum is equal to the applied torque. From this equation we
see that the spin of a positively charged particle (q > 0) will precess in the negative
sense about B, while the spin of a negatively charged particle (q < 0) will precess
in the positive sense.
S
s
B
Figure 1: Precession for q < 0
Let us specialize to the case where B is both uniform and independent of time,
and write B = B
0
z. Then the three components of (4.17) become
dS
x
(t)
dt
=
gqB
0
2mc
S
y
(t)
dS
y
(t)
dt
=
gqB
0
2mc
S
x
(t)
dS
z
(t)
dt
= 0 . (4.18)
Dening
0
=
gqB
0
2mc
(4.19)
we can combine the S
x
and S
y
equations to obtain
d
2
S
x
(t)
dt
2
=
2
0
S
x
(t)
and exactly the same equation for S
y
(t). These equations have the solutions
S
x
(t) = a cos
0
t +b sin
0
t
S
y
(t) = c cos
0
t +d sin
0
t
S
z
(t) = S
z
(0) .
Clearly, S
x
(0) = a and S
y
(0) = c. Also, from the equations of motion (4.18) we
have
0
S
y
(0) = (dS
x
/dt)(0) =
0
b and
0
S
x
(0) = (dS
y
/dt)(0) =
0
d so that our
solutions are
S
x
(t) = S
x
(0) cos
0
t +S
y
(0) sin
0
t
S
y
(t) = S
y
(0) cos
0
t S
x
(0) sin
0
t
S
z
(t) = S
z
(0) .
(4.20)
34
These can be written in the form S
x
(t) = Acos(
0
t+
x
) and S
y
(t) = Asin(
0
t+
y
)
where A
2
= S
x
(0)
2
+S
y
(0)
2
. Since these are the parametric equations of a circle, we
see that S precesses about the B eld, with a constant projection along the z-axis.
For example, suppose the spin starts out at t = 0 in the xy-plane as an eigenstate
of S
x
with eigenvalue +/2. This means that
S
x
(0)) = x +[S
x
(0)[ x+) =
2
S
y
(0)) = S
z
(0)) = 0 .
Taking the expectation value of equations (4.20) we see that
S
x
(t)) =
2
cos
0
t S
y
(t)) =
2
sin
0
t S
z
(t)) = 0 . (4.21)
This clearly shows that the spin stays in the xy-plane, and precesses about the z-
axis. The direction of precession depends on the sign of
0
, which in turn depends
on the sign of q as dened by (4.19).
How does this look from the point of view of the SP? From (4.14), the spin state
evolves according to
[(t)) = e
iHt/
[(0)) = e
i(gqBt/2mc)Sz/
[(0)) = e
i0tSz/
[(0)) .
Comparing this with (4.10a), we see that in the SP the spin precesses about z
with angular velocity
0
.
Now let us consider the motion of an electron in an inhomogeneous magnetic
eld. Since for the electron we have q = e < 0, we see from (4.12) that
s
is anti-
parallel to the spin S. The potential energy is H
s
=
s
B, and as a consequence,
the force on the electron is
F = H
s
= (
s
B) .
Suppose we have a Stern-Gerlach apparatus set up with the inhomogeneous B
eld in the z direction (i.e., pointing up):
In our previous discussion with a uniform B eld, there would be no translational
force because B
i
= 0. But now this is no longer true, and as a consequence, a
particle with a magnetic moment will be deected. The original S-G experiment
was done with silver atoms that have a single valence electron, but we can consider
this as simply an isolated electron.
It is easy to see in general what is going to happen. Suppose that the electron is
in an eigenstate of S
z
with eigenvalue +/2. Since the spin points up, the magnetic
35
moment
s
is anti-parallel to B, and hence H
s
=
s
B > 0. Since the electron
wants to minimize its energy, it will travel towards a region of smaller B, which is
up. Obviously, if the spin points down, then
s
is parallel to B and H
s
< 0 so the
electron will decrease its energy by moving to a region of larger B, which is down.
Another way to look at this is from the force relation. If
s
is anti-parallel to
B (i.e., spin up), then since the gradient of B is negative (the B eld is decreasing
in the positive direction), the force on the electron will be in the positive direction,
and hence the electron will be deected upwards. Similarly, if the spin is down,
then the force will be negative and the electron is deected downwards.
It is worth pointing out that a particle with angular momentum l will split into
2l + 1 distinct components. So if the magnetic moment is due to orbital angular
momentum, then there will necessarily be an odd number of beams coming out of
a S-G apparatus. Therefore, the fact that an electron beam splits into two beams
shows that its angular momentum can not be due to orbital motion, and hence is a
good experimental verication of the existence of spin.
Suppose that the particle enters the S-G apparatus at t = 0 with its spin aligned
along x, and emerges at time t = T. If the magnetic eld were uniform, then the
expectation value of S at t = T would be given by equations (4.21) as
S)
unif
T
=
2
(cos
0
T) x
2
(sin
0
T) y. (4.22)
Interestingly, for the inhomogeneous eld this expectation value turns out to be
zero.
To see this, let the particle enter the apparatus with its spin along the x direction,
and let its initial spatial wave function be the normalized wave packet
0
(r, t). In
terms of the eigenstates of S
z
, the total wave function at t = 0 is then (from (2.9a))
(r, 0) =
1
2
_
0
(r, 0)
0
(r, 0)
_
. (4.23a)
When the particle emerges at t = T, it is a superposition of localized spin up wave
packets
+
(r, T) and spin down wave packets
(r, T):
(r, T) =
1
2
_
+
(r, T)
(r, T)
_
. (4.23b)
Since the spin up and spin down states are well separated, there is no overlap
between these wave packets, and hence we have
+
(r, T)
(r, T) 0 . (4.23c)
Since the B eld is essentially in the z direction there is no torque on the
particle in the z direction, and from (4.17) it follows that the z component of spin
is conserved. In other words, the total probability of nding the particle with spin
36
up at t = T is the same as it is at t = 0. The same applies to the spin down states,
so we have _
[
+
(r, T)[
2
d
3
r =
_
[
0
(r, 0)[
2
d
3
r
and
_
[
(r, T)[
2
d
3
r =
_
[
0
(r, 0)[
2
d
3
r
and therefore _
[
+
(r, T)[
2
d
3
r =
_
[
(r, T)[
2
d
3
r . (4.24)
From (4.23a), the expectation value of S at r and t = 0 is given by
S)
r,0
=
(r, 0)
2
(r, 0)
=
(r, 0)
2
(
x
x +
y
y +
z
z)(r, 0)
=
2
[
0
(r, 0)[
2
x
and hence the net expectation value at t = 0 is (since
0
is normalized)
S)
0
=
_
S)
r,0
d
3
r =
2
x.
(This should have been expected since the particle entered the apparatus with its
spin in the x direction.) And from (4.23b), using (4.23c) we have for the exiting
particle
S)
r,T
=
(r, T)
2
(r, T)
=
(r, T)
2
(
x
x +
y
y +
z
z)(r, T)
=
2
_
[
+
(r, T)[
2
[
(r, T)[
2
_
z .
Integrating over all space, we see from (4.24) that
S)
inhom
T
= 0 (4.25)
as claimed.
Why does the uniform eld give a non-zero value for S)
T
whereas the inhomo-
geneous eld gives zero? The answer lies with (4.19). Observe that the precessional
frequency
0
depends on the eld B. In the Stern-Gerlach apparatus, the B eld is
weaker up high and stronger down low. This means that particles lower in the beam
precess at a faster rate than particles higher in the beam. Since (4.25) is essentially
37
an average of (4.22), we see that the dierent phases due to the dierent values of
0
will cause the integral to vanish.
Of course, this assumes that there is sucient variation
0
in precessional
frequencies that the trigonometric functions average to zero. This will be true as
long as
(
0
)T 2 .
In fact, this places a constraint on the minimum amount of inhomogeneity that the
S-G apparatus must have in order to split the beam. From (4.19) we have
0
=
e
m
e
c
B
0
e
m
e
c
B
0
z
z
where z is the distance between the poles of the magnet. Therefore, in order for
the apparatus to work we must have
e
m
e
c
B
0
z
T
2
z
.
5 The Addition of Angular Momentum
The basic idea is simply to express the eigenstates of the total angular momentum
operator J of two particles in terms of the eigenstates of the individual angular
momentum operators J
1
and J
2
, where J = J
1
+ J
2
. (Alternatively, this could be
the addition of spin and orbital angular momenta of a single particle, etc.) Each
particles angular momentum operator satises (where the subscript a labels the
particle and where we take = 1)
[J
ai
, J
aj
] = i
3
k=1
ijk
J
ak
with corresponding eigenstate [j
a
m
a
). And since we assume that [J
1
, J
2
] = 0, it
follows that
[J
i
, J
j
] = [J
1i
+J
2i
, J
1j
+J
2j
] = [J
1i
, J
1j
] + [J
2i
, J
2j
]
= i
3
k=1
ijk
J
1k
+i
3
k=1
ijk
J
2k
= i
3
k=1
ijk
_
J
1k
+J
2k
_
= i
3
k=1
ijk
J
k
and hence J is just another angular momentum operator with eigenstates [jm).
Next we want to describe simultaneous eigenstates of the total angular momen-
tum. From [J
1
, J
2
] = 0, it is clear that [J
1
2
, J
2
2
] = 0, and since [J
a
2
, J
az
] = 0, we
see that we can choose our states to be simultaneous eigenstates of J
1
2
, J
2
2
, J
1z
and J
2z
. We denote these states by [j
1
j
2
m
1
m
2
).
38
Alternatively, we can choose our states to be simultaneous eigenstates of J
2
,
J
1
2
, J
2
2
and J
z
. That [J
2
, J
z
] = 0 follows directly because J is just an angular
momentum operator. And the fact that [J
a
2
, J
z
] = 0 follows because J
a
is an
angular momentum operator, J
z
= J
1z
+J
2z
and [J
1
, J
2
] = 0. Finally, to show that
[J
2
, J
a
2
] = 0, we simply observe that
J
2
=
_
J
1
+J
2
_
2
= J
1
2
+J
2
2
+ 2J
1
J
2
= J
1
2
+J
2
2
+ 2(J
1x
J
2x
+J
1y
J
2y
+J
1z
J
2z
) .
It is now easy to see that [J
2
, J
a
2
] = 0 because [J
1
, J
2
] = 0 and [J
a
2
, J
ai
] = 0 for
i = x, y, z. We denote these simultaneous eigenstates by [j
1
j
2
jm).
However, let me emphasize that even though [J
2
, J
z
] = 0, it is not true that
[J
2
, J
az
] = 0. This means that we can not specify J
2
in the states [j
1
j
2
m
1
m
2
), and
we can not specify either J
1z
or J
2z
in the states [j
1
j
2
jm).
We know that the angular momentum operators and states satisfy (with = 1)
J
2
[jm) = j(j + 1)[jm) (5.1a)
J
z
[jm) = m[jm) (5.1b)
J
[jm) =
_
j(j + 1) m(m1)[jm1) . (5.1c)
Let us denote the two individual angular momentum states by [j
1
m
1
) and [j
2
m
2
).
Then the two-particle basis states are denoted by the various forms (see Part I,
Section 1 in the handout Supplementary Notes on Mathematics)
[j
1
j
2
m
1
m
2
) = [j
1
m
1
)[j
2
m
2
) = [j
1
m
1
) [j
2
m
2
) .
When we write j
1
j
2
m
1
m
2
[j
1
j
2
m
1
m
2
) we really mean
(j
1
m
1
[ j
2
m
2
[)([j
1
m
1
) [j
2
m
2
)) = j
1
m
1
[j
1
m
1
)j
2
m
2
[j
2
m
2
) .
Since these two-particle states form a complete set, we can write the total com-
bined angular momentum state of both particles as
[j
1
j
2
jm) =
1
j
2
m1m2
[j
1
j
2
m
1
m
2
)j
1
j
2
m
1
m
2
[j
1
j
2
jm) . (5.2)
However, we now show that many of the matrix elements j
1
j
2
m
1
m
2
[j
1
j
2
jm) vanish,
and hence the sum will be greatly simplied.
For our two-particle operators we have
J
z
= J
1z
+J
2z
and J
= J
1
+J
2
.
39
While we wont really make any use of it, let me point out the correct way to write
operators when dealing with tensor (or direct) product states. In this case we should
properly write operators in the form
J = J
1
1
2
+ 1
1
J
2
.
Then acting on the two-particle states we have, for example,
J
z
([j
1
m
1
) [j
2
m
2
)) =
_
J
1z
1 + 1 J
2z
([j
1
m
1
) [j
2
m
2
))
= J
1z
[j
1
m
1
) [j
2
m
2
) +[j
1
m
1
) J
2z
[j
2
m
2
)
= (m
1
+m
2
)([j
1
m
1
) [j
2
m
2
)) .
As I said, while we wont generally write out operators and states in this form, you
should keep in mind what is really going on.
Now, for our two-particle states we have
J
a
2
[j
1
j
2
m
1
m
2
) = j
a
(j
a
+ 1)[j
1
j
2
m
1
m
2
)
J
az
[j
1
j
2
m
1
m
2
) = m
a
[j
1
j
2
m
1
m
2
)
where a = 1, 2. Taking the matrix element of J
1
2
acting to both the left and right
we see that
j
1
j
2
m
1
m
2
[J
1
2
[j
1
j
2
jm) = j
1
(j
1
+ 1)j
1
j
2
m
1
m
2
[j
1
j
2
jm)
= j
1
j
2
m
1
m
2
[j
1
j
2
jm)j
1
(j
1
+ 1)
or
[j
1
(j
1
+ 1) j
1
(j
1
+ 1)]j
1
j
2
m
1
m
2
[j
1
j
2
jm) = 0 .
Since this result clearly applies to J
2
2
as well, we must have
j
1
j
2
m
1
m
2
[j
1
j
2
jm) = 0 if j
1
,= j
1
or j
2
,= j
2
.
In other words, equation (5.2) has simplied to
[j
1
j
2
jm) =
m1m2
[j
1
j
2
m
1
m
2
)j
1
j
2
m
1
m
2
[j
1
j
2
jm) .
Next, from J
z
= J
1z
+J
2z
we can let J
1z
+J
2z
act to the left and J
z
act to the
right so that
j
1
j
2
m
1
m
2
[J
z
[j
1
j
2
jm) = (m
1
+m
2
)j
1
j
2
m
1
m
2
[j
1
j
2
jm)
= j
1
j
2
m
1
m
2
[j
1
j
2
jm)m.
This shows that
j
1
j
2
m
1
m
2
[j
1
j
2
jm) = 0 unless m = m
1
+m
2
40
and hence equation (5.2) has become
[j
1
j
2
jm) =
m1m2
m1+m2=m
[j
1
j
2
m
1
m
2
)j
1
j
2
m
1
m
2
[j
1
j
2
jm) .
Finally, we regard the values of j
1
and j
2
as xed and understood, and we simply
write the total angular momentum state as
[jm) =
m1m2
m1+m2=m
[m
1
m
2
)m
1
m
2
[jm) . (5.3)
The complex numbers m
1
m
2
[jm) are called Clebsch-Gordan coecients.
They are really nothing more than the elements of the unitary transition matrix that
takes us from the [m
1
m
2
) basis to the [jm) basis. (This is just the statement
from linear algebra that given two bases e
i
and e
i
of a vector space V , there is
a nonsingular transition matrix P such that e
i
=
j
e
j
p
ji
. Moreover, e
i
is just the
ith column of P. If both bases are orthonormal, then P is in fact a unitary matrix.
See Theorem 15 in Supplementary Notes on Mathematics.)
Since (normalized) eigenfunctions corresponding to distinct eigenvalues of a her-
mitian operator are orthonormal, we know that
j
[jm) =
jj
mm
and m
1
m
2
[m
1
m
2
) =
m1m
m2m
2
.
Therefore, taking the inner product of equation (5.3) with itself we see that
m1m2
m1+m2=m
[m
1
m
2
[jm)[
2
= 1 . (5.4)
Given j
1
and j
2
, this holds for any resulting values of j and m. (This is really
just another way of saying that the columns of the unitary transition matrix are
normalized.)
Now, we know that j
1
m
1
j
1
and j
2
m
2
j
2
. Therefore, from
m = m
1
+ m
2
, the maximum value of m must be j
1
+ j
2
. And since j m j,
it follows that the maximum value of j is j
max
= j
1
+ j
2
. Corresponding to this
maximum value j
1
+j
2
of j, we have a multiplet of 2(j
1
+j
2
) + 1 values of m, i.e.,
(j
1
+ j
2
) m j
1
+ j
2
. On the other hand, since there are 2j
1
+ 1 possible
values of m
1
, and 2j
2
+1 possible values of m
2
, the total number of (not necessarily
distinct) possible m
1
+m
2
= m values is (2j
1
+ 1)(2j
2
+ 1).
The next highest possible value of m is j
1
+j
2
1 so that there is a j state with
j = j
1
+j
2
1 and a corresponding multiplet of 2(j
1
+j
2
1) +1 = 2(j
1
+j
2
) 1
possible m values. We continue lowering the j values by one, and for each such
value j = j
1
+ j
2
k there are 2(j
1
+ j
2
k) + 1 possible m values. However, the
total number of m values in all multiplets must equal (2j
1
+ 1)(2j
2
+ 1).
41
Example 5.1. Consider the case j
1
= 1, j
2
= 2. Then the total number of possible
m values is
(2j
1
+ 1)(2j
2
+ 1) = 3 5 = 15 .
We can arrange these values of m = m
1
+m
2
in a table with m
1
across the top and
m
2
down the side:
m
2
m
1
1 0 1
2 3 2 1
1 2 1 0
0 1 0 1
1 0 1 2
2 1 2 3
The distribution of m values is as follows:
value of m = m
1
+m
2
: 3 2 1 0 1 2 3
number of occurances: 1 2 3 3 3 2 1
We have m
max
= 1 + 2 = 3 so that j
max
= j
1
+ j
2
= 3, and there are seven values
of m in this multiplet: 3, 2, 1, 0, 1, 2, 3. If we eliminate these, then the next
highest value of m is 2, so there must exist a j = 2 state with the ve m values
2, 1, 0, 1, 2. Eliminating these, the highest remaining m value is 1, so there is
a j = 1 state with the three m values 1, 0, 1. This exhausts all of the m values.
Note that summing the number of m values in each multiplet we have 7+5+3 = 15
as it must.
The point of this example is to illustrate the origin of the two ways of getting
an m value of j
1
+ j
2
1: one way is the multiplet corresponding to j = j
1
+ j
2
,
and the second way is the multiplet corresponding to j = j
1
+j
2
1. Similarly, the
case m = j
1
+ j
2
2 enters into three multiplets: j = j
1
+ j
2
, j = j
1
+ j
2
1 and
j = j
1
+j
2
2.
In any case, for each nonnegative integer k, the state with j = j
1
+j
2
k contains
2(j
1
+j
2
k) + 1 values of m, subject to the constraint
b
k=0
[2(j
1
+j
2
k) + 1] = (2j
1
+ 1)(2j
2
+ 1) . (5.5)
The minimum value j
min
= j
1
+ j
2
b is dened by the integer b which is to be
determined. Recalling the formula
m
k=n
k =
mn + 1
2
(m+n)
42
we have
b
k=0
[2(j
1
+j
2
k) + 1] = 2(j
1
+j
2
)(b + 1) 2
b
k=0
k + (b + 1)
= 2(j
1
+j
2
)(b + 1) (b + 1)b + (b + 1)
= (b + 1)[2(j
1
+j
2
) + 1 b] .
Then equation (5.5) becomes
(b + 1)[2(j
1
+j
2
) + 1 b] = (2j
1
+ 1)(2j
2
+ 1)
or
b
2
2(j
1
+j
2
)b + 4j
1
j
2
= 0 .
Solving for b using the quadratic formula yields
b = (j
1
+j
2
)
_
(j
1
+j
2
)
2
4j
1
j
2
= (j
1
+j
2
) (j
1
j
2
)
so that b = 2j
1
or b = 2j
2
. Since j
min
= j
1
+j
2
b, we see that either j
min
= j
2
j
1
or j
min
= j
1
j
2
. If j
1
,= j
2
then one of these is negative and must be rejected, and
hence we conclude that j
min
= [j
1
j
2
[. Thus we arrive at j
min
j j
max
or
[j
1
j
2
[ j j
1
+j
2
. (5.6)
Equations (5.1c), (5.3), (5.4) and (5.6) are all we need to calculate the Clebsch-
Gordan coecients. The procedure is best illustrated by a specic example.
Example 5.2. Let us consider the case j
1
= j
2
= 1/2 because you (should) already
know the answer. In this case we have m
1
= m
2
= 1/2, so according to equation
(5.6) we know that 0 j 1.
First we have j
max
= j
1
+j
2
= 1, and the maximum value of m is m
1
+m
2
= 1
also. From equation (5.3) we have the single term
1 1
_
=
1
2
1
2
__
1
2
1
2
1 1
_
and by equation (5.4) we must have
_
1
2
1
2
1 1
_
2
= 1 .
This species the topmost state up to a phase factor. Choosing this phase to be 1
(the Condon-Shortley convention), we are left with
_
1
2
1
2
1 1
_
= +1 .
43
Hence equation (5.3) becomes
1 1
_
=
1
2
1
2
_
(5.7a)
where the state on the left-hand side of this equation is the [jm) state, and the
state on the right-hand side is the [m
1
m
2
) state.
To construct the next lowest state, we act on the left-hand side of this last result
with J
= J
1
+ J
2
. Using equation (5.1c),
we obtain
1 0
_
=
_
3
4
+
1
4
_
1
2
1
2
_
+
1
2
1
2
_
_
or
1 0
_
=
1
2
_
1
2
1
2
_
+
1
2
1
2
_
_
. (5.7b)
Since (by equation (5.3))
1 0
_
=
1
2
1
2
__
1
2
1
2
1 0
_
+
1
2
1
2
__
1
2
1
2
1 0
_
we see that
_
1
2
1
2
1 0
_
=
_
1
2
1
2
1 0
_
=
1
2
.
We now act on equation (5.7b) with J
= J
1
+J
2
again to obtain
1 1
_
=
1
2
_
_
3
4
+
1
4
_
1
2
1
2
_
+
1
2
1
2
_
_
_
or
1 1
_
=
1
2
1
2
_
(5.7c)
and hence
_
1
2
1
2
1 1
_
= 1 .
This completes the j = 1 multiplet which, for obvious reasons, is called the triplet
state.
Now for the j = 0 state. We know from equation (5.3) that
0 0
_
=
1
2
1
2
__
1
2
1
2
0 0
_
+
1
2
1
2
__
1
2
1
2
0 0
_
:=
1
2
1
2
_
a +
1
2
1
2
_
b .
To solve for a and b, we rst see that from equation (5.4) we have a
2
+ b
2
= 1.
Next, we note that the [jm) = [1 0) state is orthogonal to the [jm) = [0 0) state,
and therefore
0 = 1 0[0 0) =
1
2
(a +b)
44
so that a = b. Therefore 2a
2
= 1 so that a = b = 1/
2 and we have
0 0
_
=
1
2
_
1
2
1
2
_
1
2
1
2
_
_
(5.7d)
which is called the singlet state.
We have thus constructed the well-known wavefunctions for two electrons:
triplet symmetric states:
_
1 1
_
=
1
2
1
2
_
1 0
_
=
1
2
_
1
2
1
2
_
+
1
2
1
2
__
1 1
_
=
1
2
1
2
_
singlet antisymmetric state:
0 0
_
=
1
2
_
1
2
1
2
_
1
2
1
2
_
_
.
While we will soon return to discuss the symmetry of wave functions in detail,
it should be fairly clear that exchanging electrons 1 and 2 in any of the triplet
states leaves the wave function unchanged, and hence the states are referred to as
symmetric. Similarly, exchanging electrons 1 and 2 in the singlet state changes
the overall sign of the state, and hence this is called an antisymmetric state.
45