0% found this document useful (0 votes)
14 views105 pages

Advanced Quantum Mechanics: 6ccm436a/7ccmms31

The document is a set of lecture notes on Advanced Quantum Mechanics, authored by Neil Lambert at King's College London, and is based on the works of Profs. N. Drukker and G. Watts. It covers various topics including one-dimensional quantum mechanics, angular momentum, perturbation theory, and the Feynman path integral, among others. The introduction discusses the revolutionary nature of quantum mechanics compared to classical physics and outlines the course structure and expectations for students.

Uploaded by

Adefope Owojori
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
14 views105 pages

Advanced Quantum Mechanics: 6ccm436a/7ccmms31

The document is a set of lecture notes on Advanced Quantum Mechanics, authored by Neil Lambert at King's College London, and is based on the works of Profs. N. Drukker and G. Watts. It covers various topics including one-dimensional quantum mechanics, angular momentum, perturbation theory, and the Feynman path integral, among others. The introduction discusses the revolutionary nature of quantum mechanics compared to classical physics and outlines the course structure and expectations for students.

Uploaded by

Adefope Owojori
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 105

Advanced Quantum Mechanics

6ccm436a/7ccmms31

Neil Lambert∗

Department of Mathematics, King’s College London


The Strand, London WC2R 2LS, UK

N.B. These notes are not particularly original but rather based on the lecture notes of
Profs. N. Drukker and G. Watts.


email: [email protected]
2
Contents

0.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
0.2 Plan . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6

1 Review Of One-Dimensional Quantum Mechanics 9


1.1 Hilbert Spaces . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
1.2 Observables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11
1.3 Canonical Quantisation . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14
1.4 Time Evolution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
1.5 The (non)-Collapse of the Wavefunction . . . . . . . . . . . . . . . . . . 17
1.6 The Quantum Harmonic Oscillator . . . . . . . . . . . . . . . . . . . . . 18

2 Angular Momentum 25
2.1 Symmetries . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25
2.2 A Spherically Symmetric Potential . . . . . . . . . . . . . . . . . . . . . 27
2.3 Angular Momentum Operators . . . . . . . . . . . . . . . . . . . . . . . 28

3 All About su(2) 31


3.1 Representations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
3.2 Groups vs Algebras . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
3.2.1 Construction of Irreducible Unitary Representations of su(2) . . . 36

4 Back to Angular Momentum 41


4.1 Differential Operators vs Irreducible Representations . . . . . . . . . . . 42
4.2 Addition of Angular Momentum . . . . . . . . . . . . . . . . . . . . . . . 45

5 The Hydrogen Atom 49


5.1 The Quantum 2-Body Problem . . . . . . . . . . . . . . . . . . . . . . . 49
5.2 The Hydrogen Atom . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 50

6 Perturbation Theory 55
6.1 A Simple Example . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 57
6.2 First Order Non-Degenerate Perturbation Theory . . . . . . . . . . . . . 58
6.3 Second Order Non-Degenerate Perturbation Theory . . . . . . . . . . . . 61
6.4 Third Order Non-Degenerate Perturbation Theory . . . . . . . . . . . . . 63
6.5 Degenerate Perturbation Theory . . . . . . . . . . . . . . . . . . . . . . . 63

3
4 CONTENTS

6.6 Time Dependent Perturbation Theory . . . . . . . . . . . . . . . . . . . 65

7 Semi-Classical Quantization: WKB 69


7.1 WKB . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 69
7.2 Particle in a Box . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 71
7.3 Turning Points, Airy and Bohr-Sommerfeld Quatization . . . . . . . . . . 71
7.4 Harmonic Oscillator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 74

8 Density Matrices, Entanglement and Thermal States 77


8.1 Density Matrices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 78
8.2 Thermal States . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 82
8.3 Entropy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 84
8.4 Entanglement . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 85

9 Relativistic Quantum Mechanics 89


9.1 Relativistic Wave Equations . . . . . . . . . . . . . . . . . . . . . . . . . 91
9.2 Relativistic Notation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 92
9.3 Spinors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 93
9.4 Back to Quantum Mechanics . . . . . . . . . . . . . . . . . . . . . . . . . 97

10 The Feynman Path Integral 99


10.1 Gaussian Path Integrals . . . . . . . . . . . . . . . . . . . . . . . . . . . 102
10.2 Computing Determinants . . . . . . . . . . . . . . . . . . . . . . . . . . . 104
0.1. INTRODUCTION 5

0.1 Introduction
Quantum Mechanics arguably presents our most profound and revolutionary under-
standing of objective reality. In the classical world, as studied for example “Classical
Dynamics” (5ccm231a) the physically observable world consists of knowing the posi-
tions and momenta (velocities) of every single particle at a moment in time. Then the
dynamical equations can be used to evolve the system so as to know the positions and
momenta in the future leading to a causal and highly predictive theory. You could
forgive the physicists of the 19th century for thinking that they were almost done. Ev-
erything seemed to work perfectly and there were just a few loose strings to get under
control.
It turned out that by pulling on those loose strings the whole structure of classical
physics and our understanding of objective relativity fell apart. We now have a more
predictive and precise theory in the form of Quantum Mechanics. But no one has got
to grips with what it truly means.
The classic experiment, although there are many, that lead to the unravelling of
classical mechanics relates to the question as to whether or not light is a particle or
a wave. If you consider a coherent light source and send a beam of light through two
small slits and see what happens you will find an interference pattern. This is what
you would expect if light was a wave and the two slits produced two wave sources that
could add constructively or cancel out deconstructively. Thus an interference pattern
is observed. Fine, light is a wave. Let’s do the same with electrons. We find the same
thing! Shooting an electron beam at a double slit also leads to an interference pattern.
So electrons, which otherwise appear as point like particles with fixed electric charges
also behave like waves (and similarly there are experiments such as the photoelectric
effect which show that light behaves like a particle). This interference pattern even
exists if you slow to beam down so that one electron at a time is released.
What has happened? Nothing that has a classical interpretation. We can think of
electrons as waves with profile ψ. After passing through the slits we have two wave-
functions ψhole1 and ψhole2 . Quantum mechanics tells us that the system is described by
ψhole1 + ψhole2 but in this case the two wave profiles can constructively or deconstruc-
tively interfere (i.e. add or cancel out). Classically there isn’t really an explanation but
one would expect two the electron beams coming out from each of the slits to behave
independently. Experiment tells us which is true (see figure 0.1.1). Needless to say there
have been countless experiments since which confirm the quantum picture and refute
any classical explanation (most notably in the Bell’s inequalities).
There are further revelations to that we won’t get to speak much of. Relativity
came and told us that time is not absolute. Combining this with Quantum Mechanics
means that there really aren’t any particles at all just local excitations of fields that
permeate spacetime. Why, after all, is every electron identical to every other? This is
the topic of Quantum Field Theory. However exploring Quantum Mechanics beyond an
introductory module such as “Introductory Quantum Theory” (6cmm332c) is crucial
6 CONTENTS

Figure 0.1.1: After 100, 200, 500 and 1000 electrons hit the screen the interference
picture is formed (on top). Compare this to the probability distribution computed from
the wave function of the electron (bottom). In particular the quantum probability (left)
is obtained from |ψhole1 + ψhole2 |2 vs. the classical sum of probabilities |ψhole1 |2 + |ψhole2 |2
(right).

for any physicist.


Furthermore now we have String and M-Theory which successfully combines Quan-
tum Mechanics and gravity and so we again think that we have everything wrapped-up
apart from a few loose ‘strings’. What could go wrong?

0.2 Plan
We hope to cover each chapter in a week - but this could fluctuate, its a course on
Quantum Mechanics after all. The main lectures, where new material is presented, are
on Mondays 4-6pm. There will also be in-person discussion sessions on Tuesdays from
11-12am. These are intended to discuss issues from the previous week’s lectures and
give further examples or calculations and are meant for the level 6 students (but anyone
can attend).
0.2. PLAN 7

In addition there will be weekly problem sets found on Keats. You are strongly
urged to do the problems - no one ever understood Quantum Mechanics based on their
personal experience or by only listening to lectures. There are weekly tutorials on
Thursday morning which will discuss the problem set from the previous week.
There is a reading week starting October 30 where there will not be any lectures,
discussion sessions or tutorials.
8 CONTENTS
Chapter 1

Review Of One-Dimensional
Quantum Mechanics

Let’s jump head first into the formulation of Quantum Mechanics. In this section we
will just consider one-dimensional systems, meaning one spatial dimension and time.
Conceptually the extension to higher dimensions is easy but more technically involved
and is the subject of much of this module. One can also consider quantum systems with
no spatial dimension. We will consider these in Chapter 8. It is expected that what is
said here is a review.
In the Hamiltonian formulation of Classical Mechanics the state of a system corre-
sponds to a point in phase space. One specifies the positions and momenta of all the
particles and then one knows everything there is to know about the system: the relevant
observable physical quantities, such as energy and pressure, are known functions of the
positions and momenta. Thus if you know the state of the system then you in principle
know all the observable physical quantities. Furthermore given the state of the system
at some initial time t = 0 one can use Hamilton’s equations to predict the state of the
system at a later time. In general Hamilton’s equations are non-linear so this is is com-
plicated problem but for suitable classes of the Hamiltonian it is well-posed. (One can
also use the Lagrangian description with momentum replace by velocity and Hamilton’s
equations by the Euler-Lagrange equations.) With this you can do experiments, build
fancy machines to fly you around the world or go to the moon. Fantastic. But as more
detailed experiments have shown over the past 100 years or more this is simply not true
fundamentally, on microscopic scales. Rather it is a coarse grained approximation to
reality.

1.1 Hilbert Spaces


In Quantum Mechanics the state of a system is given by an element |ψ〉 of unit norm in
a Hilbert space H.
Definition: A Hilbert space is a complete, complex vector space with positive definite

9
10 CHAPTER 1. REVIEW OF ONE-DIMENSIONAL QUANTUM MECHANICS

inner product

〈ψ1 |ψ2 〉 : H × H → C (1.1)

However in Quantum mechanics we also assume that H is “separable” (meaning that


given any two points one can find two disjoint open sets which contain each point).
Often Hilbert spaces are taken to be infinite dimensional as if it is finite dimensional
then it is isomorphic to CN for some N .1 The classic example of a Hilbert space is
L2 (R):
󰀝 󰁝 󰀞
L2 (R) = ψ : R → C | |ψ| dx < ∞
2

󰁝
〈ψ1 |ψ2 〉 = ψ1∗ ψ2 dx (1.2)

Showing this is a seperable Hilbert space is one of those “better leave it to our analysis
friends” types of question. In cases like these one often refers to the state |ψ〉 as the
wavefunction ψ(x) and often writes

ψ(x) = 〈x|ψ〉 (1.3)

although this isn’t really defined as 〈x| doesn’t exist, even by physicists standards. We
will often use the terms state and wavefunction interchangeably.
Formally Hilbert spaces are technically very involved and our treatment hides many
subtleties. Complete means that all Cauchy sequences converge. Separable means that
any two vectors admit two disjoint open sets such that each includes one vector but
not the other (it is ensured if there is a countable basis |e1 〉, |e2 〉, |e3 〉, ... which we will
assume to be the case). As far as a physicist is concerned a Hilbert space is an infinite
dimensional complex vector space that behaves like CN but with N = ∞. In this
module we will behave as physicists and assume that all definitions and theorems can
be rigorously stated and proved without affecting our use of them; trust me (or more
precisely trust our colleagues) they can.
The statement that the state of the system is an element of a Hilbert space is already
profoundly weird as vector spaces allow you to add vectors. Thus if |ψ1 〉 and |ψ2 〉 are
two orthogonal states then
1
|ψ〉 = √ (|ψ1 〉 + |ψ2 〉) (1.4)
2
is also a state. There is no classical analogue of this. It sounds like madness. For
example Schrödinger famously exploited this by taking the states of an alive and dead
cat thereby introducing a morass of philosophical confusion to the world.2
Definition The space of linear maps from H to C is called the dual space H∗ .
1
Hilbert himself did not know what a Hilbert space was and is reported to have entered a seminar
where it was being discussed saying ”Hilbert space? What is that?”
2
To quote Stephen Hawking “whenever I hear Schrödinger’s cat I reach for my gun”.
1.2. OBSERVABLES 11

Here we are using Dirac notation where we denote vectors by “kets” |ψ〉. Elements
of the dual space are denoted by “bras” 〈ψ| such that

〈ψ1 |(|ψ〉2 ) = 〈ψ1 |ψ2 〉 (1.5)

There is a theorem which states that H∗ is also a Hilbert space and is isomorphic to H.
This means that for each |ψ〉 there is a unique 〈ψ| and vice versa. Dirac was a hugely
important figure in 20th Century physics, second only to Einstein. His text book on
Quantum Mechanics was foundational and remains one of the best to this day. We will
see more from him later.3
Hilbert spaces admit linear maps (the analogue of matrices in finite dimensions)

O:H→H (1.6)

such that O(λ1 |ψ1 〉 + λ2 |ψ2 〉) = λ1 O|ψ1 〉 + λ2 O|ψ2 〉 where λ1 , λ2 ∈ C.


The innerproduct allows us to define the notion of an adjoint operator O† :
Definition The adjoint of a linear map O is a linear map O† : H → H that satisfies

〈O† ψ1 |ψ2 〉 = 〈ψ1 |Oψ2 〉 (1.7)

for any two vectors |ψ1 〉, |ψ2 〉 ∈ H.


In Dirac notation and in some basis |en 〉 we can write observables as
󰁛
O= Omn |em 〉〈en | (1.8)
mn

where Omn ∈ C. It is easy to see that the adjoint is then


󰁛
O† = †
Omn |em 〉〈en | (1.9)
mn

† ∗
where, just as in finite dimensions, Omn = Onm .

1.2 Observables
In addition to a Hilbert space of states Quantum Mechanics exploits the fact that there
exists a preferred class of linear maps which are self-adjoint (aka Hermitian) in the sense
that

〈ψ1 |Oψ2 〉 = 〈Oψ1 |ψ2 〉 ⇐⇒ O† = O

for all vectors |ψ1 〉 and |ψ2 〉. Such maps are call Observables and they lead to an
important

Theorem: A self-adjoint operator admits an othornormal basis of eigenvectors with


real eigenvalues.
3
Dirac was notoriously strange and was once asked at a College dinner what he was known for to
which he replied “I invented the bra”.
12 CHAPTER 1. REVIEW OF ONE-DIMENSIONAL QUANTUM MECHANICS

N.B. An eigenvector |ψ〉 ∈ H and corresponding eigenvalue λ ∈ C of an operator O


satisfy O|ψ〉 = λ|ψ〉
We won’t prove this theorem here: it has many subtleties in it. However we can
at least prove the following statement: eigenstates of self-adjoint operators have real
eigenvalues and eigenvectors with distinct eigenvalues are orthogonal. To see this we
consider two eigenvectors |ψ1 〉 and |ψ2 〉 with eigenvalues λ1 and λ2 respectively. Then
we can evaluate

〈ψ1 |Oψ2 〉 = λ2 〈ψ1 |ψ2 〉


〈Oψ1 |ψ2 〉 = λ∗1 〈ψ1 |ψ2 〉 (1.10)

If O is self-adjoint then these two expression are equal:

λ2 〈ψ1 |ψ2 〉 = λ∗1 〈ψ1 |ψ2 〉 (1.11)

First we suppose that |ψ1 〉 = |ψ2 〉 and hence λ2 = λ1 . Thus we have, since |ψ1 〉 must
have non-zero norm, λ1 = λ∗1 . Next we suppose that |ψ1 〉 ∕= |ψ2 〉. Since we now know
that λ1 and λ2 are real we have

(λ2 − λ1 )〈ψ1 |ψ2 〉 = 0 (1.12)

Thus if λ2 ∕= λ1 then 〈ψ1 |ψ2 〉 = 0. If there are degeneracies, meaning that there are
many eigenstates with the same eigenvalue then one can arrange to find an orthogonal
basis by a Gram-Schmidt process. But we can’t conclude that two such eigenstates are
orthogonal without extra work. Whereas we know immediately that if two eigenstates
of a self-adjoint operator have different eigenvalues then they are orthogonal.
If an operator is self-adjoint then there is a choice of basis |en 〉 so that (see problem
set 1)
󰁛
O= λn |en 〉〈en | (1.13)
n

with λn ∈ R. This is what a diagonal operator looks like in Dirac notation.


How do we extract numbers that we can compare with the so-called real world
through experiment? The key idea of Quantum Mechanics is that observables represent
a physical measurement and the eigenvalues are the possible outcomes of such measure-
ments. To this end we define the expectation value of an operator O in a state |ψ〉
to be

〈O〉 = 〈ψ|O|ψ〉 (1.14)

The above theorem tells us that we can find an orthonormal basis of eigenvectors of
an observable O. Thus any state |ψ〉 can be written as
󰁛
|ψ〉 = cn |ψn 〉 (1.15)
n
1.2. OBSERVABLES 13

where O|ψn 〉 = λn |ψn 〉. Intuitively the physical idea is that in order to measure an
observable one must probe the state you are observing, e.g. to find the position of an
electron you must look for it which means hitting it with light which will then impart
momentum and change it. Thus the act of measurement changes the system and hence
one doesn’t know what the system is doing after the measurement. Eigenstates are
special as they are, in some sense, unaffected by a particular measurement (but not
others).
Looking at |ψ〉 we can compute the expectation value of O to find
󰁛
〈ψ|Oψ〉 = c∗m cn 〈ψm |Oψn 〉
m,n
󰁛
= λn c∗m cn 〈ψm |ψn 〉
m,n
󰁛
= λn c∗n cn
n
󰁛
= λn p n (1.16)
n

where we think of pn = c∗n cn as a probability distribution since the unit norm of |ψ〉
implies that

1 = 〈ψ|ψ〉
󰁛
= c∗m cn 〈ψm |ψn 〉
m,n
󰁛
= c∗m cn 〈ψm |ψn 〉
m,n
󰁛
= c∗n cn
n
󰁛
= pn (1.17)
n

with 0 ≤ pn ≤ 1. The interpretation of this equation is that pn = |cn |2 is the probability


that a state will be observed to the have the value λn in the measurement associated to
O. The expectation value is therefore the average value that repeated experiments will
obtain for measuring the observable O on the state |ψ〉. It follows that if the state is an
eigenstate of an operator O then a measurement of O will produce the corresponding
eigenvalue with probability 1.
But this is not the general situation. Rather the system is not in an eigenstate
and so repeated measurements will produce different answers and hence there will be a
non-zero variance known as the uncertainty:

(∆O)2 = 〈O2 〉 − 〈O〉2 (1.18)

Note that the right hand side can be written as 〈(O − 〈O〉I)2 〉 and hence is positive.
Another theorem asserts that if two operators commute:

[O1 , O2 ] = O1 O2 − O2 O1 = 0 (1.19)
14 CHAPTER 1. REVIEW OF ONE-DIMENSIONAL QUANTUM MECHANICS

then one can find an orthonormal basis of states which are simultaneously eigenstates
of both operators.
On the other hand for pairs of operators that don’t commute we find the famous
Heisenberg uncertainty principle:
1
∆O1 ∆O2 ≥ |〈i[O1 , O2 ]〉| (1.20)
2
The classic example of this are the positions and momenta operators:


x̂ψ(x) = xψ(x) p̂ψ(x) = −i󰄁 ψ(x) (1.21)
∂x
so that [x̂, p̂] = i󰄁 and hence

∆x̂∆p̂ ≥ 󰄁/2 (1.22)

1.3 Canonical Quantisation


How does one construct such a Hilbert space and its corresponding obervables in a
physically meaningful way? There is a notion of canonical quantization whereby one
takes the physical quantities of a classical Hamiltonian system and replaces them by
self-adjoint operators on a Hilbert space where the Poisson bracket is replaced by the
commutator:
i
f (q, p), g(q, p) → Of , Og {f, g} → − [Of , Og ] (1.23)
󰄁
In particular the most fundamental quantities in a Hamiltonian system are the the
positions qi and conjugate momenta pi with i = 1, .., N which obey

{qi , qj } = {pi , pj } = 0 {qi , pj } = δij (1.24)

The ‘canonical’ way to quantise this system is consider the Hilbert space L2 (RN ) of
functions of qi and make the replacements4

qi →q̂i : q̂i ψ(q1 , ..., qN ) = qi ψ(q1 , ..., qN )



pi →p̂i : p̂i ψ(q1 , ..., qN ) = −i󰄁 ψ(q1 , ..., qN ) (1.25)
∂qi

and it is easy to see that indeed

[q̂i , q̂j ] = [p̂i , p̂j ] = 0 [q̂i , p̂j ] = i󰄁δij (1.26)

as required (and used above). Note that this corresponded to a particular choice of coor-
dinates qi and conjugate momentum pi . However Hamiltonian systems admit canonical
4
In general we will avoid putting hats on operators but we need to do so here as qi have very specific
meanings as real numbers.
1.4. TIME EVOLUTION 15

transformations that mix these and hence there are several ways to quantise correspond-
ing to choosing a ‘polarisation’, i.e. a choice of what is a q and what is a p.
However this is really putting the cart before the horse as the classical world emerges
from the quantum world and not the other way around. Indeed there are quantum
systems with no classical limit at all. They obey the rules of quantum mechanics that we
are outlining here but they need not come from “quantising” some classical Hamiltonian
system.

1.4 Time Evolution


The most important of all observables is Hamiltonian H and its eigenvalues are the
possible energies of the system. Under canonical quantisation the Hamiltonian equation

dO
= {O, H} (1.27)
dt
becomes
dOH i
= − [OH , H] (1.28)
dt 󰄁
Here I have introduced a subscript H on O to indicate that we are in the so-called
Heisenberg picture where operators evolve in time and states are constant. And 1.28 is
known as the Heisenberg equation.
A more familiar picture is the Schrödinger picture where

OS = e−iHt/󰄁 OH eiHt/󰄁
|ΨS 〉 = e−iHt/󰄁 |ΨH 〉 (1.29)

With this choice we see that


dOS
=0
dt
d i
|ΨS 〉 = − H|ΨS 〉 (1.30)
dt 󰄁
Furthermore the Hamiltonian governs the time evolution of the system through
Schrödinger’s equation


i󰄁 |ΨS 〉 = H|ΨS 〉 (1.31)
∂t

This follows by thinking of the Hamiltonian operator H as the energy operator Ê and
making the identification


Ê = i󰄁 (1.32)
∂t
In what follows we drop the subscript S and will always work in the Schrödinger picture.
16 CHAPTER 1. REVIEW OF ONE-DIMENSIONAL QUANTUM MECHANICS

One then sees that eigenstates of H with eigenvalue En evolve rather trivially (for
simplicity here we assume that H and hence its eigenvectors and eigenvalues are time
independent):

|Ψ(t)〉 = e−iEn t/󰄁 |ψn 〉 (1.33)

I will try to stick to a convention where the time dependent wavefunction is |Ψ〉 and
the time independent one, meaning it is an eigenstate of the Hamiltonian, |ψ〉. The
difference then is just the time-dependent phase.
More generally given any state |Ψ〉 we can expand it in an energy eigenstate basis:
󰁛
|Ψ〉 = cn (t)|ψn 〉 (1.34)
n

in which case substituting into the Schrödinger equation we find

cn (t) = cn (0)e−iEn t/󰄁 (1.35)

and hence we know the state at all times:


󰁛
|Ψ(t)〉 = cn (0)e−iEn t/󰄁 |ψn 〉 (1.36)
n

and the cn (0) can be found by expanding |ψ(0)〉 in the energy eigenstate basis. Of course
finding the eigenstates and eigenvalues of a typical Hamiltonian is highly non-trivial.
But in principle solving Quantum Mechanics is down to (infinite-dimensional) linear
algebra.
A central point is that time evolution is unitary, meaning that it preserves the
inner-product between two states. It is crucial for the consistency of the theory as
otherwise the probabilities we discussed above will fail to sum to unity.
For a particle in one-dimension subject to a potential V (x) the quantum Hamiltonian
is
1 2
H= p̂ + V (x̂)
2m
󰄁2 ∂ 2
=− + V (x) (1.37)
2m ∂x2
Following the canonical quantization described above leads to the most familiar form of
the Schrödinger equation:

∂Ψ 󰄁2 ∂ 2 Ψ
i󰄁 =− + V (x)Ψ (1.38)
∂t 2m ∂x2
Here we have taken the Hilbert space to be L2 (R) and x̂, p̂ as above. In principle one
can find a basis of eigenvectors of H which obey

󰄁2 ∂ 2 ψ n
− + V (x)ψn = En ψn . (1.39)
2m ∂x2
1.5. THE (NON)-COLLAPSE OF THE WAVEFUNCTION 17

This is some differential equation and one typically finds a discrete spectrum of eigenval-
ues En and wavefunctions ψn (x). This is the origin of quantum in Quantum Mechanics.
We can then compute various expectation values and probabilities.
For example to normalize the wavefunction we require that
󰁝
|ψ|2 dx = 1 (1.40)

Given a region R ⊂ R then interpretation of ψ is that the probability to find the particle
in R is
󰁝
P (R) = |ψ|2 dx (1.41)
R

On the other hand the expectation value of the position is


󰁝
〈x̂〉 = x|ψ|2 dx (1.42)

i.e. this is the average value of where you will observe the particle after many measure-
ments of its position.
This all sounds great and wonderful as we have reduced all the complexities of the
world down to linear algebra. Who knew?! But we seem to be no where near the classical
world we experience and you could be forgiven for thinking of Quantum Mechanics as
lying in the realm of crackpot science. But it’s not and it is in fact incredibly rich and
predictive with no known conflict with experiment.

1.5 The (non)-Collapse of the Wavefunction


One of the most important theorems is due to Ehrenfest:
Theorem: The expectation value of an observable obeys the classical equations of
motion. In particular for a particle in one-dimension one has
d 1 d ∂
〈x̂〉 = 〈p̂〉 , 〈p̂〉 = − 〈V (x̂)〉 (1.43)
dt m dt ∂x

So in a sense the classical world emerges as the most probable set of observations.
In particular there is the so-called correspondence principle which asserts that we
expect to find agreement with the classical theory in the limit of large quantum numbers
(i.e. large n in the case of equation 1.39). Okay so we are back in the so-called real world.
Whew.
The most controversial statement in all of physics is the infamous Copenhagen in-
terpretation which asserts that following a measurement corresponding to an eigenvalue
λ of an observable O the system then collapses to the associated eigenstate. This is
where indeterminacy enters as the system is evolving in time as some state |ψ(t)〉 as
dictated by the Schrödinger equation and then after the measurement it is suddenly
projected onto an eigenstate of O in a random and unpredictable way, not governed by
18 CHAPTER 1. REVIEW OF ONE-DIMENSIONAL QUANTUM MECHANICS

the Schrödinger equation but somehow chosen by god or the experimenter. In particular
it could be projected onto any of the eigenstates. It then evolves again according to the
Schrödinger equation until the next measurement.
No one believes this. However we all do it in practice as it works and the alternative
will keep you awake at night.
What people typically believe is that there is no classical world and no classical ob-
server who can interfere with the evolution of the state to make it suddenly collapse to
a particular eigenstate. Rather the world is a very large quantum system consisting of
a single wavefunction of many many variables and there are subtle correlations between
the various subsystems such that the wavefunction becomes strongly peaked around
certain eigenvalues and vanishingly small away from others. In particular it never col-
lapses to one eigenstate, there are no “measurements”, and it always evolves causally in
time according to the Schrödinger equation. The classical world appears because this
wavefunction is so strongly peaked around a particular set of classical outcomes that it
leads to an emergent classical history.
The disturbing thing about this is that there is no reason to think that the wave-
function will peak around a unique classical world. Rather there must be many, possibly
all, classical worlds that emerge from a single wavefunction. Different branches of the
wavefunction must surely exist and correspond to distinct classical histories and we just
get to live in one of these. Sleep well because in some branch of the wavefunction your
mirror self will have a very bad dream.

1.6 The Quantum Harmonic Oscillator


Let us review this most classic example of a quantum system corresponding to
1
V (x) = kx2 , (1.44)
2
hence the Schrödinger equation is

∂Ψ 󰄁2 ∂ 2 Ψ 1 2
i󰄁 =− + kx Ψ (1.45)
∂t 2m ∂x2 2
We could proceed by finding the general solution to this differential equation. However
this is a difficult task and a much better analysis can be done using algebra.
Again our Hilbert space is H = L2 (R). We can write the Schrödinger equation as
󰀕 󰀖
1 2 k 2
Ê|Ψ〉 = p̂ + x̂ |Ψ〉 (1.46)
2m 2

Next we introduce new operators â and â†


󰀕 󰀖 12 󰀣√ 󰀤 12 󰀕 󰀖 12 󰀣√ 󰀤 12
1 mk 1 mk
â = √ p̂ − i x̂ , ↠= √ p̂ + i x̂ (1.47)
2󰄁 mk 2󰄁 2󰄁 mk 2󰄁
1.6. THE QUANTUM HARMONIC OSCILLATOR 19

These satisfy
√ √
† 1 2 mk 2 i 1 2 mk 2 1
ââ = √ p̂ + x̂ − [x̂, p̂] = √ p̂ + x̂ +
2󰄁 mk 2󰄁 2󰄁 2󰄁 mk 2󰄁 2
√ √
1 mk 2 i 1 mk 2 1
↠â = √ p̂2 + x̂ + [x̂, p̂] = √ p̂2 + x̂ −
2󰄁 mk 2󰄁 2󰄁 2󰄁 mk 2󰄁 2
(1.48)

and hence

[â, ↠] = 1 (1.49)

Thus the Hamiltonian can be written as


󰁵 󰀖 󰀕
1 2 k 2 k󰄁2
1 †
H= p̂ + x̂ = â â + (1.50)
2m 2 m2
󰁳
At this point, if not before, one usually introduces ω = k/m so that
󰀕 󰀖
† 1
H = 󰄁ω â â + (1.51)
2

We know that the normalisable modes have energies which are bounded below by
zero. In particular there is a lowest energy state that we call the ground state |0〉. Let
us introduce another operator N̂ called the number operator

N̂ = ↠â (1.52)

i.e. H = 󰄁ω(N̂ + 12 ). Note that N̂ is Hermitian with postive eigenvalues

〈Ψ|N̂ Ψ〉 = 〈Ψ|↠âΨ〉 = 〈âΨ|âΨ〉 ≥ 0 (1.53)

Thus it follows that the lowest energy eigenstate |0〉 must satisfy

N̂ |0〉 = 0 (1.54)

which means that

â|0〉 = 0 (1.55)
󰁴
1 k󰄁2
hence Ĥ|0〉 = 2 m
|0〉, i.e. the ground state energy is

1
E0 = 󰄁ω (1.56)
2
We can create new states by acting with â†

|n〉 = Nn (↠)n |0〉 (1.57)


20 CHAPTER 1. REVIEW OF ONE-DIMENSIONAL QUANTUM MECHANICS

where Nn is a normalisation constant chosen so that 〈n|n〉 = 1. In particular it is


󰁵
1
Nn = (1.58)
n!
Next we prove the following lemma: N̂ |n〉 = n|n〉. To do this we proceed by in-
duction. By definition we have that N̂ |0〉 = 0. Next we consider, assuming that
N̂ |n − 1〉 = (n − 1)|n − 1〉,
Nn
N̂ |n〉 = N̂ ↠|n − 1〉
Nn−1
Nn † †
= â ââ |n − 1〉
Nn−1
Nn 󰀃 † 󰀄
= â [â, ↠] + ↠↠â |n − 1〉
Nn−1
Nn † 󰀓 󰀔
= â 1 + N̂ |n − 1〉
Nn−1
Nn †
=n â |n − 1〉
Nn−1
= n|n〉

This can also been seen from the following commutators

[N̂ , â] = [↠â, â]


= ↠ââ − â↠â
= ↠ââ − [â, ↠]â − ↠ââ
= −â (1.59)

and

[N̂ , ↠] = [↠â, ↠]


= ↠â↠− ↠↠â
= ↠â↠+ ↠[â, ↠] − ↠ââ†
= ↠(1.60)

Hence it follows that

N̂ â|n〉 = ([N̂ , â] + âN̂ )|n〉 = (−â + nâ)|n〉 = (n − 1)â|n〉 (1.61)

and

N̂ ↠|n〉 = ([N̂ , ↠] + ↠N̂ )|n〉 = (↠+ n↠)|n〉 = (n + 1)↠|n〉 (1.62)

The operators ↠and â are therefore known as raising and lowering operators respec-
tively. From this we see that
󰀕 󰀖 󰀕 󰀖
1 1
H|n〉 = 󰄁ω N̂ + |n〉 = 󰄁ω n + |n〉 (1.63)
2 2
1.6. THE QUANTUM HARMONIC OSCILLATOR 21

i.e. the energy of the state |n〉 is


󰀕 󰀖
1
En = 󰄁ω n + (1.64)
2

So we have completely solved the system without ever solving a differential equation.
Furthermore this approach also allows us to explicitly construct the corresponding
energy eigenstate wavefunctions. For example the ground state satisfies
󰀳 󰀣√ 󰀤 12 󰀴
󰀕 󰀖 12
1 mk
â|0〉 = 󰁃 √ p̂ − i x̂󰁄 |0〉 = 0 (1.65)
2󰄁 mk 2󰄁

Recalling the definitions of x̂ and p̂ we obtain a first order differential equation


󰀳 󰀣√ 󰀤 12 󰀴
󰀕 󰀖 12
󰁃−i󰄁 1 ∂ mk
√ −i x󰁄 Ψ 0 = 0 (1.66)
2󰄁 mk ∂x 2󰄁

where Ψ0 (x, t) = e−iEn t/󰄁 ψ(x) is the explicit element of L2 (R) that represents |0〉.
Rewriting this equation gives

dψ0 mk
=− xψ0 (1.67)
dx 󰄁

mkx2 /2󰄁
The solution to this differential equation is simply ψ ∝ e− and hence

itE0 mk 2
− − x
Ψ = N0 e 󰄁 e 2󰄁 (1.68)

where N0 is a normalisation and we have included the time-dependent phase factor


itE0
e− 󰄁 .
Furthermore we can obtain explicit solutions for the higher energy states Ψn (x, t) by
acting with â†
󰁵
1 − itEn † n
Ψn (x, t) = e 󰄁 (â ) ψ0 (x)
n!
󰁵 󰀳 󰀣 √ 󰀤 12 󰀴n
󰀕 󰀖 12
1 − itEn 󰁃 1 ∂ mk
= e 󰄁 −i󰄁 √ +i x󰁄 ψ0 (x)
n! 2󰄁 mk ∂x 2󰄁
󰁵 󰀳 󰀣 √ 󰀤 12 󰀴n
󰀕 󰀖 12 √
1 itEn 1 ∂ mk mk 2
= N0 e− 󰄁 󰁃−i󰄁 √ +i x󰁄 e− 2󰄁 x (1.69)
n! 2󰄁 mk ∂x 2󰄁


mk 2
These have the form of a polynomial times the Gaussian e− 2󰄁
x
. We can explicitly
22 CHAPTER 1. REVIEW OF ONE-DIMENSIONAL QUANTUM MECHANICS

construct the first few wavefunctions, for example,


󰀳 󰀣 √ 󰀤 12 󰀴
󰀕 󰀖 12 √
itE1
− 󰄁 󰁃 1 ∂ mk 󰁄 − 2󰄁mk 2
x
Ψ1 (x, t) = N0 e −i󰄁 √ +i x e ,
2󰄁 mk ∂x 2󰄁
󰀳 󰀣√ 󰀤1 󰀣 √ 󰀤 12 󰀴
2 √
itE1
− 󰄁 󰁃 mk mk 󰁄 mk 2
− 2󰄁 x
= N0 e +i x+i x e
2󰄁 2󰄁
󰀣√ 󰀤 12

mk itE1 mk 2
= 2iN0 e− 󰄁 xe− 2󰄁
x
2󰄁
(1.70)

It would have been very difficult to find all these solutions to the Schrödinger equation
by brute force and also to use them to calculate quantities such as 〈x̂〉 and 〈x̂2 〉.
We can also calculate various expectation values in this formalism. For example, in
the state |n〉, we have

〈x̂〉 = 〈n|x̂n〉
󰀕 󰀖 12
i 2󰄁
= √ 〈n|(â − ↠)n〉
2 mk
󰀕 󰀖 12
i 2󰄁 󰀃 󰀄
= √ 〈n|ân〉 − 〈n|↠n〉 (1.71)
2 mk

Now â|n〉 ∝ |n − 1〉 and ↠|n〉 ∝ |n + 1〉 so that

〈x̂〉 ∝ 〈n|n + 1〉 − 〈n|n − 1〉 = 0 (1.72)

since the |n〉 are orthonormal. On the other hand

〈x̂2 〉 = 〈n|x̂2 n〉
󰀕 󰀖
1 2󰄁
= − √ 〈n|(â − ↠)2 n〉
4 mk
󰀕 󰀖
1 2󰄁 󰀃 󰀄
= − √ 〈n|(â)2 n〉 + 〈n|(↠)2 n〉 − 〈n|â↠n〉 − 〈n|↠ân〉
4 mk
(1.73)

Now again (â)2 |n〉 ∝ |n − 2〉 and (↠)2 |n〉 ∝ |n + 2〉 so that the first two terms give zero.
However the last two terms give
󰀕 󰀖
1 2󰄁
2
〈x̂ 〉 = √ (〈↠n|↠n〉 + 〈ân|ân〉) (1.74)
4 mk
Now we have already seen that
󰁵
1
|n + 1〉 = ↠|n〉 , (1.75)
n+1
1.6. THE QUANTUM HARMONIC OSCILLATOR 23

which implies that


󰁵 󰁵 󰁵
1 1 1
â|n + 1〉 = â↠|n〉 = (N̂ + [â, ↠])|n〉 = (n + 1)|n〉 (1.76)
n+1 n+1 n+1
Thus
󰁵
1
|n〉 = â|n + 1〉 (1.77)
n+1
and finally
󰀕 󰀖
2 1 2󰄁
〈x̂ 〉 = √ (2n + 1) (1.78)
4 mk
You can also evaluate 〈p̂〉 and 〈p̂2 〉 this way and check that the Heisenberg uncertainty
relation is satisfied.
However it is instructive to see where these solutions come from. Let us look at the
time-independent Schrödinger equation
󰄁2 d 2 ψ 1 2
− + kx ψ = Eψ (1.79)
2m dx2 2
It is helpful to write
2
ψ = f (x)e−αx (1.80)

so that
󰀕 󰀖
dψ df 2 2
= − 2αxf e−α x
dx dx
󰀕 2 󰀖
d2 ψ df df 2 2

2
= 2
− 4αx − 2αf + 4α x f e−α x
2 2
(1.81)
dx dx dx
and
󰀕 󰀖
󰄁2 d2 f df 1
− 2
− 4αx − 2αf + 4α2 x2 f + kx2 f = Ef (1.82)
2m dx dx 2

Let us take α2 = km/4󰄁2 so that the x2 terms cancel:

d2 f df 2mE
− 4αx − 2αf = − 2 f (1.83)
dx 2 dx 󰄁
This equation will have polynomial solutions. Indeed if we write f = xn + . . . then the
leading order xn−2 term must cancel to give
2mE
−4αn − 2α = − (1.84)
󰄁2
From here we read off that
󰁵
󰄁2 α k
E=2 (n + 1/2) = 󰄁 (n + 1/2) (1.85)
m m
as before.
24 CHAPTER 1. REVIEW OF ONE-DIMENSIONAL QUANTUM MECHANICS
Chapter 2

Angular Momentum

Let us start our exploration of Quantum Mechanics by looking at more realistic three-
dimensional models. In Classical Dynamics the simplest, but also most common, sys-
tems have rotational symmetry which leads via Noether’s theorem to conserved angular
momentum. This in turn means that the system can typically be solved. The classic
example is the Kepler problem of a planet moving around the Sun.
It’s ridiculous to think of the quantum problem of something so macroscopic as a
planet moving around something even bigger such as the Sun. The correspondence
principle tells us that we should just reproduce the classical results with incredible
accuracy. But happily there is a suitable quantum analogue which is of a negatively
charged particle (an electron) moving around a heavier positively charged particle (a
nucleus) as found in atoms! But before we tackle this problem head on and see Quantum
Mechanics really working for us and matching experiment we should step back and think
of symmetries, and angular momentum in particular, in Quantum Mechanics.

2.1 Symmetries
In classical dynamics symmetries play an important role. Simply put these are transfor-
mations of the system that leave the dynamics unchanged. In a Hamiltonian formulation
these correspond to canonical transformations that leave the Hamiltonian invariant:

H(qi′ , p′i ) = H(qi , pi ) (2.1)

Recall that a canonical transformation means that

{qi′ , qj′ } = {p′i , p′j } = 0 {qi′ , p′j } = δij (2.2)

A particularly important class of transformations are infinitesimal ones: meaning that


one can expand

qi′ = qi + εTi + . . .
p′i = pi + εUi + . . . (2.3)

25
26 CHAPTER 2. ANGULAR MOMENTUM

here the ellipsis denote higher order powers of ε and Ti , Ui are some functions on phase
space. It is known that for this to be a canonical transformation there must exist a
function Q on phase space such that
∂Q ∂Q
Ti = Ui = − (2.4)
∂pi ∂qi
This in turn means that

qi′ = qi + ε{qi , Q} p′i = pi + ε{pi , Q} (2.5)

One then finds that


󰁛 󰀕 ∂H ∂H
󰀖
H(qi′ , p′i ) = H(qi , pi ) + ε Ti + Ui + . . .
∂qi ∂pi
󰁛 󰀕 ∂H ∂Q ∂H ∂Q 󰀖
= H(qi , pi ) + ε − + ...
∂qi ∂pi ∂pi ∂qi
= H(qi , pi ) + ε{H, Q} + . . . (2.6)

Thus an infinitesimal symmetry corresponds to function Q such that


dQ
{Q, H} = 0 ⇐⇒ =0 (2.7)
dt
where we have assumed that Q has no explicit time-dependence. This it the famous
Noether theorem: every infinitesimal symmetry gives rise to a conserved quantity
and vice-versa (although the converse is not true in a Lagrangian formulation).
What does this mean in a quantum theory? We say that a symmetry corresponds
to an observable Q that commutes with the Hamiltonian:

[H, Q] = 0 (2.8)

What does this mean? Well since they commute and are self-adjoint there is a basis of
eigenstates of both Q and H:

Q|en 〉 = qn |en 〉 H|en 〉 = En |en 〉 (2.9)

As we have seen a state that is an eigenstate of H will remain an eigenstate under time
evolution. This if it is also an eigenstate of Q it will remain an eigenstate of Q. However
a typically state will not be an eigenstate of either H or Q. Nevertheless it is also easy
to see that the expectation value of Q in any state will be time independent (assuming
that Q is time independent):
d d d
〈ψ|Q|ψ〉 = 〈 ψ|Q|ψ〉 + 〈ψ|Q| ψ〉
dt dt dt
i i
= 〈− Hψ|Q|ψ〉 + 〈ψ|Q| − Hψ〉
󰄁 󰄁
i
= 〈ψ|H † Q − QH|ψ〉
󰄁
i
= 〈ψ|[H, Q]|ψ〉
󰄁
=0 (2.10)
2.2. A SPHERICALLY SYMMETRIC POTENTIAL 27

2.2 A Spherically Symmetric Potential


In three-dimensions the Schrödinger equation takes the intimidating form
∂Ψ 󰄁2 2
i󰄁 =− ∇ Ψ + V (x, y, x)Ψ (2.11)
∂t 2m
Here we have assumed the potential is time-independent and
∂2 ∂2 ∂2
∇2 = + + . (2.12)
∂x2 ∂y 2 ∂z 2
This is in general too hard to solve but let us consider a potential with spherical sym-
󰁳
metry V = V (r) and r = x2 + y 2 + z 2 . It is also helpful to switch to spherical
coordinates:

x = r sin θ cos φ
y = r sin θ sin φ
z = r cos θ (2.13)

In which case (you can show this by a tedious calculation just using the chain rule, see
the problem sets, its easier if you know Riemannian geometry)
󰀕 󰀖 󰀕 󰀖
2 1 ∂ 2 ∂ 1 ∂ ∂ 1 ∂2
∇ = 2 r + 2 sin θ + 2 2 (2.14)
r ∂r ∂r r sin θ ∂θ ∂θ r sin θ ∂φ2
Furthermore we look for energy eigenstates so

Ψ = e−iEt/󰄁 ψ(r, θ, φ) (2.15)

which leads us to the following equation


󰀕 󰀕 󰀖 󰀕 󰀖 󰀖
󰄁2 1 ∂ 2 ∂ 1 ∂ ∂ 1 ∂2
− r + 2 sin θ + 2 2 ψ + V (r)ψ = Eψ
2m r2 ∂r ∂r r sin θ ∂θ ∂θ r sin θ ∂φ2
(2.16)
We can solve this equation by separation of variables. In particular we write

ψ(r, θ, φ) = u(r)Y (θ, φ) (2.17)

From here we find


󰀕 󰀖 󰀕 󰀕 󰀖 󰀖
󰄁2 d 2 du 󰄁2 u 1 ∂ ∂ 1 ∂2
− r Y − sin θ + Y + (V (r) − E)uY = 0
2mr2 dr dr 2mr2 sin θ ∂θ ∂θ sin2 θ ∂φ2
(2.18)
By the usual logic of separation of variables (divide by uY /r2 and see that the first and
third terms only depend on r where as the second is independent of r) splits into two
equations
󰀕 󰀖
󰄁2 d 2 du
− r + (V (r) − E)r2 = −󰂃
2mu dr dr
󰀕 󰀕 󰀖 󰀖
1 󰄁2 1 ∂ ∂ 1 ∂2
− sin θ + Y =󰂃 (2.19)
Y 2m sin θ ∂θ ∂θ sin2 θ ∂φ2
28 CHAPTER 2. ANGULAR MOMENTUM

where 󰂃 is a constant. In particular the second equation knows nothing about the
potential and hence nothing about the problem at hand:
󰀕 󰀕 󰀖 󰀖
1 ∂ ∂ 1 ∂2 2m󰂃
sin θ + + 2 Y =0 (2.20)
sin θ ∂θ ∂θ sin2 θ ∂φ2 󰄁
Indeed the solutions for Y (θ, φ) can be completely solved and the solutions are related
to spherical harmonics Yl,m (θ, φ), labelled by two integers l, m with l ≥ 0 and |m| ≤ l.
On the other hand the first equation is a single ordinary differential equation for
u(r) and we have some hope to solve it:
󰀕 󰀖
1 d 2 du 2m 󰀓 󰂃 󰀔
− 2 r + 2 V (r) + 2 − E u = 0 (2.21)
r dr dr 󰄁 r
This is the only place which is sensitive to the particular problem at hand through the
choice of potential V (r).
Thus we have reduced the problem to solving two independent differential equations.
And the only one specific to the particular problem is a second order ordinary linear
differential equation. One can proceed with brute force (the Y ′ s can be found by sep-
aration of variables once more). You can find them on Wikipedia or Mathematica so
we won’t say more about them here - come on Tuesday! Although we will say that one
finds
󰄁2 l(l + 1)
󰂃= (2.22)
2m
for some non-negative integer l. Later we will construct them in an insightful way. Indeed
spherical harmonics arise from some very important physics (angular momentum) and
mathematics (Lie algebras) underlying the system. So let us explore...

2.3 Angular Momentum Operators


In classical dynamics angular momentum is defined by L = x×p so the quantum version
is

L = x̂ × p̂ (2.23)

If we denote x = (x, y, z) = (x1 , x2 , x3 ) then we have, in components


2 󰁛
󰁛 3
Li = εijk x̂j p̂k
j=1 k=1

= εijk x̂j p̂k (2.24)

where i = 1, 2, 3 and we use the convention that repeated indices are summed over.
Therefore, following canonical quantisation, we find the operator

Li = −i󰄁εijk xj (2.25)
∂xk
2.3. ANGULAR MOMENTUM OPERATORS 29

for example
∂ 3 ∂
L1 = −i󰄁x2 + i󰄁x (2.26)
∂x3 ∂x2
To gain some intuition let us look at

[L1 , L2 ] = L1 L2 − (1 ↔ 2)
󰀕 󰀖󰀕 󰀖
2 k ∂ m ∂
= −󰄁 ε1kl x ε2mn x − (1 ↔ 2)
∂xl ∂xn
󰀕 󰀖
2 k ∂ k m ∂2
= −󰄁 ε1kl ε2mn x δlm n + x x − (1 ↔ 2)
∂x ∂xl ∂xn
∂ ∂2
= −󰄁2 ε1kl ε2ln xk n − 󰄁2 ε1kl ε2ln xk xm l n − (1 ↔ 2) (2.27)
∂x ∂x ∂x
Lets look at the second term. We see that it is symmetric under k ↔ m and l ↔ n.
Therefore it is symmetric under 1 ↔ 2 and is cancelled when we subtract 1 ↔ 2. Thus
we have

[L1 , L2 ] = −󰄁2 ε1kl ε2ln xk − (1 ↔ 2)
∂xn

= −󰄁2 ε123 ε231 x2 1 − (1 ↔ 2)
∂x

= −󰄁2 x2 1 − (1 ↔ 2)
∂x
= i󰄁L3 (2.28)

where we have used the fact that l ∕= 1 in ε1kl and l ∕= 2 in ε2ln so the only non-zero
term comes from l = 3 and hence k = 2. This is a general fact which follows from the
identity1

εikl εjlm − εjkl εilm = −εijl εlkm (2.29)

and so in general (see the problem sets)

[Li , Lj ] = i󰄁εijk Lk (2.30)

Does this look familiar? If not it will be the end of the semester. If we rescale Li = 󰄁Ji
then we find

[Ji , Jj ] = iεijk Jk (2.31)

which is the Lie-algebra of su(2) - the simplest example of a Lie-algebra. There will be
more on Lie algebras latter. This structure is powerful enough for us to deduce more
or less anything we need to know about states with angular momentum from purely
algebraic considerations.
1
Check it for your self or use Mathematica. I don’t know any “smart” way to do it, just think of
the cases and note that the left hand side is anti-symmetric in i, j.
30 CHAPTER 2. ANGULAR MOMENTUM
Chapter 3

All About su(2)

3.1 Representations
We have seen that the angular momentum operators (with factors of 󰄁 stripped off)
satisfy [Ji , Jj ] = i󰂃ijk Jk but it is also not hard to see that
1
Ji = τi (3.1)
2
where τi are the Pauli matrices also satisfy [Ji , Jj ] = i󰂃ijk Jk . So what else does?
The commutation relation satisfied by the J ′ s is an example of a Lie-algebra. Roughly
speaking (and there is a whole module devoted to this) a Lie algebra is what you get
when you look at infinitesimally small group transformations (more on this shortly). The
formal definition is that a Lie-algebra is a vector space G (often taken to be complex)
with a bi-linear map
[ · , · ]:G×G →G (3.2)
which is antisymmetric: [A, B] = −[B, A] and satisfies the Jacobi identity:
[A, [B, C]] + [B, [C, A]] + [C, [A, B]] = 0 (3.3)
This last relation is automatically obeyed by matrices if we take [A, B] = AB − BA
since matrix multiplication is associative (A(BC) = (AB)C). If we let Ta be a basis
for G where a = 1, .., dimV then it must be that there exist constants, called structure
constants, such that
[Ta , Tb ] = ifab c Tc (3.4)
where there is a sum over c and the Ta are known as generators of the Lie-algebra.
The factor of i requires some explanation. It is often not there in the mathematics
literature. However in Physics we like our generators to be Hermitian (just like the
angular momentum operators). If the Ta are Hermitian then
([Ta , Tb ])† = Tb† Ta† − Tb† Ta†
= Tb Ta − Ta Tb
= −[Ta , Tb ] (3.5)

31
32 CHAPTER 3. ALL ABOUT SU (2)

Thus the factor of i ensures that fab c are real. In Mathematics one often drops the factor
of i and takes Ta = −Ta† .
Next we wish to construct all finite dimensional unitary representations of su(2).
Formally a representation is a linear map

Π : G → End(V ) (3.6)

such that

Π([Ji , Jj ]) = [Π(Ji ), Π(Jj )] (3.7)

Here V is a vector space (typically complex) and End(V ) are the set of endomorphisms
of V (linear maps from V to V ). In physics language V = CN and End(V ) are N × N
matrices.
We are particularly interested in irreducible representations. These are represen-
tations with no non-trivial invariant subspaces. That is, there are no vector subspaces
of V that are mapped to itself by Π.
Let us suppose that we are given matrices Ji that satisfy [Ji , Jj ] = i󰂃ijk Jk . Since we
want a unitary representation we assume that Ji† = J i but we do not know anything else
yet and we certainly don’t assume that they are 2 × 2 matrices or differential operators.
Note that we have both the dimension of the Lie-algebra dimG and the dimension of
the representation N . The dimension of su(2) is three (the generators are J1 , J2 , J3 ) but
we will see that we can construct N × N Hermitian matrices Ji , i.e. N -dimensional
representations of su(2), for any N = 0, 1, 2, ....

3.2 Groups vs Algebras


An important observation is that we have been looking at representations of the Lie-
algebra and not the group. We can think of the Lie-algebra as infinitesimal group
elements near the identity:

g = I + iθi Ji + . . . (3.8)

with θ1 << 1 some arbitrarily small parameters. To obtain representations of the group
we simply exponentiate these matrices:
i i
g = eiθ J (3.9)

where now θi are not small.


Let us at look at what we get by exponentiating the Pauli matrices we found above:
i
g = e 2 θ·τ (3.10)

To this end let us take

θ = θn (3.11)
3.2. GROUPS VS ALGEBRAS 33

where n is a unit vector in R3 and θ = |θ|. We note that

(n · τ )2 = ni nj τi τj
= ni nj (δij I + iεijk τk )
= n · nI
=I (3.12)

This means that (in · τ )2 = −I. Thus we have (the proof is the same as for eiθ =
cos θ + i sin θ)
i
g = e 2 θ·τ
= cos(θ/2)I + in · τ sin(θ/2) (3.13)

and hence

g † g = (cos(θ/2)I − in · τ sin(θ/2)) (cos(θ/2)I + in · τ sin(θ/2))


= (cos2 (θ/2) + sin2 (θ/2))I
=I (3.14)
i
In other words g † = e− 2 θ·τ = g −1 so we have unitary matrices. We can also compute
i i
det g = det e 2 θ·τ = etr( 2 θ·τ ) = 1 (3.15)

Thus we have g ∈ SU (2) where SU (2) is the group

SU (2) = {2 × 2 complex matrices| det g = 1, g † = g −1 } (3.16)

In the previous section we constructed all finite-dimensional representations of su(2)


and by exponentiation we find representations of SU (2). One caveat: these are unitary
representations because we assumed that the Ji are Hermitian (so the group elements
i
eiθ Ji are Unitary). But we started off by looking at rotations, that is SO(3). What
happened?
Here encounter the fact that there are two groups with the same Lie-algebra: SU (2)
and SO(3). Indeed there is a well-known theorem:

SO(3) ∼
= SU (2)/Z2 (3.17)

where Z2 is generated by the centre of SU (2) - that is to say the set of all elements
in SU (2) that commute which each other. This means that they are multiples of the
identity and this leaves just ±I. In particular

g(θ = 0) = I g(θ = 2π) = −I (3.18)

So what about SO(3)? Here we first look at infinitesimal rotations in R3

R = I + θ · K + ... (3.19)
34 CHAPTER 3. ALL ABOUT SU (2)

with
󰀳 󰀴 󰀳 󰀴 󰀳 󰀴
0 0 0 0 0 1 0 1 0
󰁅 󰁆 󰁅 󰁆 󰁅 󰁆
K 1 = 󰁃0 0 1 󰁄 , K2 = 󰁃 0 0 0 󰁄 , K3 = 󰁃−1 0 0󰁄 (3.20)
0 −1 0 −1 0 0 0 0 0

Each of these generates an infinitesimal two-dimensional rotation in the x, y and z


planes respectively. For example take only θ3 small but non-zero:
󰀳 󰀴 󰀳 󰀴 󰀳 󰀴
x x x
󰁅 󰁆 󰁅 󰁆 󰁅 󰁆
R 󰁃 y 󰁄 = 󰁃 y 󰁄 + θ 3 K3 󰁃 y 󰁄
z z z
󰀳 󰀴
x + θ3 y
󰁅 󰁆
= 󰁃y − θ 3 x 󰁄 (3.21)
z

which is indeed an infinitesimal rotation around the z axis. One can see explicitly that

[Ki , Kj ] = εijk Kk (3.22)

which is almost su(2). We see that we have to make the identification

Ji = iKi (3.23)

then we find [Ji , Jj ] = iεijk Jk which is su(2). Note that the Ki ’s are not Hermitian so
we can’t diagonalise them but the Ji ’s are so we could (but the physical significance of
this is obscure!) and we’d find the l = 1 representation that we constructed above (it is
easy to see that the eigenvalues of J3 = iK3 are −1, 0, 1).
Exponentiating these gives a relatively complicated expression but we can just con-
sider rotations about z:
󰀳 󰀴 󰀳 󰀴 󰀳 󰀴
0 0 0 1 0 0 0 1 0
󰁅 󰁆 󰁅 󰁆 󰁅 󰁆
g = eθ3 K3 = 󰁃0 0 0󰁄 + cos θ3 󰁃0 1 0󰁄 + i sin θ3 󰁃−1 0 0󰁄
0 0 1 0 0 0 0 0 0
󰀳 󰀴
cos θ3 sin θ3 0
󰁅 󰁆
= 󰁃− sin θ3 cos θ3 0󰁄 (3.24)
0 0 1

Here we have used the fact that K3 , and all powers of K3 , splits into a non-trivial 2 × 2
bit that squares to minus the identity and a trivial part with only zeros. Thus we have
rotations with θ3 ∈ [0, 2π].
Thus the homomorphism Φ : SU (2) → SO(3) is
󰀕 󰀖
i
θ·τ
Φ e2 = eθ·K (3.25)

and indeed both θ = 0 and |θ| = 2π are mapped to the identity in R3 .


3.2. GROUPS VS ALGEBRAS 35

Along these lines we see that any representation of su(2) will exponentiate to give a
group element of SO(3) if

e2πiJ3 = I (3.26)

This corresponds to integer eigenvalues of J3 , i.e. l ∈ Z.


It’s also interesting to think about how SU (2) and SO(3) look as spaces (manifolds).
Rewriting equation 3.13 as

g = AI + iB · τ (3.27)

we see that A2 + B · B = 1 parameterises a 3-sphere in 4 dimensions: SU (2) ∼ = S 3.


What does SO(3) look like? To describe a rotation we choose an axis n, a unit vector,
and an angle θ ∈ [−π, π]. This allows us to construct θ = θn. But we can see that −n
and −θ parameterise the same rotation. As a result SO(3) ∼ = S 3 /Z2 and is not simply
connected. In particular the path
󰀳 󰀴
1
󰁅 󰁆
θ = πs , n = s 󰁃0 󰁄 , s ∈ [−1, 1] (3.28)
0

is closed in SO(3) but can’t be continuously deformed to a point. But curiously enough
going around this loop twice is deformable to a point, i.e. no loop, π1 (SO(3)) = Z2 .
More general representation arise by considering tensors Tµ1 ...µn over C2 for su(2)
or R3 for SO(3). The group elements act on each of the µi indices in the natural
way. In general this does not give an irreducible representation. For larger algebras
such as SU (N ) and SO(N ) taking Tµ1 ...µn to be totally anti-symmetric does lead to
an irreducible representation. So does totally symmetric and traceless on any pair of
indices.
What happens in Nature. We saw when constructing the spherical harmonics that
l was integer but we have also seen that representations with l half-integer exists. The
spherical harmonics arise because there is a fundamental SO(3) rotational symmetry of
space. Later we will see that actually the symmetry of space is SU (2) and that there
are particles, Fermions, which transform under SU (2) not just SO(3). Sometimes one
writes SU (2) = Spin(3) where Spin(d) is the simply connected cover of SO(d). But
this is critical as Fermions must satisfy the Pauli exclusion principle which means that
no two Fermions can be in the same state. This is ultimately what makes atoms and
matter stable enough for us to exist.
Furthermore with relativity the symmetry group of spacetime is enhanced to SO(1, 3).
But again there are Fermions and the actual symmetry group is Spin(1, 3) ∼ = SL(2, C) ∼
=
SU (2)×SU (2). By a happy coincidence everything is still described by SU (2). This is a
fluke of being in a relatively low dimension. In higher dimensions the spacetime groups
and algebras are more complicated than those of SU (2). However many physicists can
happily spend their lives only looking at SU (2).
36 CHAPTER 3. ALL ABOUT SU (2)

3.2.1 Construction of Irreducible Unitary Representations of


su(2)
Let us now return to our task: finding all N × N Hermitian matrices Ji that satisfy

[Ji , Jj ] = i󰂃ijk Jk (3.29)

We note that
J 2 = (J1 )2 + (J2 )2 + (J3 )2 , (3.30)
is a Casimir. That means it commutes with all the generators
󰁛
[J 2 , Ji ] = [Jj2 , Ji ]
j
󰁛
= Jj [Jj , Ji ] + [Jj , Ji ]Jj
j
󰁛
= Jj 󰂃jik Jk + 󰂃jik Jk Jj
jk
󰁛
= 󰂃jik (Jj Jk + Jk Jj )
jk

=0. (3.31)

There is a famous theorem known as Schur’s lemma which states that any such Casimir
must act as a multiple of the identity in an irreducible representation. This means that
J 2 = λI in any irreducible representation. In practical terms if J 2 commutes with all
other operators then nothing will change the eigenvalue of J 2 .
Since the Ji are Hermitian we can chose to diagonalise one, but only one since su(2)
has rank 1, say J3 . Thus the representation has a basis of states labelled by eigenvalues
of J3 and J 2 :
J3 |λ, m〉 = m|λ, m〉 J 2 |λ, m〉 = λ|λ, m〉. (3.32)
In analogy to the harmonic oscillator we swap J1 and J2 for operators

J± = J1 ± iJ2 J+† = J− . (3.33)

Notice that

[J3 , J± ] = [J3 , J1 ± iJ2 ]


= [J3 , J1 ] ± i[J3 , J2 ]
= iJ2 ± J1
= ±J± . (3.34)

We can therefore use J± to raise and lower the eigenvalue of J3 :

J3 (J± |λ, m〉) = ([J3 , J± ] + J± J3 )|λ, m〉


= (±J± + mJ± )|λ, m〉
= (m ± 1)J± |λ, m〉 (3.35)
3.2. GROUPS VS ALGEBRAS 37

Therefore we have

J+ |λ, m〉 = cm |λ, m + 1〉 J− |λ, m〉 = dm |λ, m − 1〉 , (3.36)

where the constants cm and dm are chosen to ensure that the states are normalized (we
are assuming for simplicity that the eigenspaces of J3 are one-dimensional - we will
return to this shortly).
To calculate cm we evaluate

|cm |2 〈m + 1|m + 1〉 = 〈λ, m|J+† J+ |λ, m〉


= 〈λ, m|J− J+ |λ, m〉
= 〈λ, m|(J1 − iJ2 )(J1 + iJ2 )|λ, m〉
= 〈λ, m|J12 + J22 + i[J1 , J2 ]|λ, m〉
= 〈λ, m|J 2 − J32 − J3 |λ, m〉
= (λ − m2 − m)〈λ, m|λ, m〉 (3.37)

Thus if 〈m|m〉 = 〈m + 1|m + 1〉 = 1 we find that



cm = λ − m2 − m . (3.38)

Similarly for dm :

|dm |2 〈λ, m − 1|λ, m − 1〉 = 〈λ, m|J−† J− |λ, m〉


= 〈λ, m|J+ J− |λ, m〉
= 〈λ, m|(J1 + iJ2 )(J1 − iJ2 )|λ, m〉
= 〈λ, m|J12 + J22 − i[J1 , J2 ]|λ, m〉
= 〈λ, m|J 2 − J32 + J3 |λ, m〉
= (λ − m2 + m)〈λ, m|λ, m〉 (3.39)

So that

dm = λ − m2 + m . (3.40)
Thus we see that any irrep of su(2) is labelled by λ and has states with J3 eigenvalues
m, m±1, m±2m, . . .. If we look for finite dimensional representations then there must be
a highest value of J3 -eigenvalue mh and lowest value ml . Furthermore the corresponding
states must satisfy

J+ |λ, mh 〉 = 0 J− |λ, ml 〉 = 0 (3.41)

This in turn requires that cmh = dml = 0:

λ − mh (mh + 1) = 0 and λ − ml (ml − 1) = 0 . (3.42)

This implies that


λ = mh (mh + 1) (3.43)
38 CHAPTER 3. ALL ABOUT SU (2)

and also that

mh (mh + 1) = ml (ml − 1) . (3.44)

This is a quadratic equation for ml as a function of mh and hence has two solutions.
Simple inspection tells us that

ml − −mh or ml = mh + 1 . (3.45)

The second solution is impossible since ml ≤ mh and hence the spectrum of J3 eigenvalues
is:
mh , mh − 1, ..., −mh + 1, −mh , (3.46)

with a single state assigned to each eigenvalue. Furthermore there are 2mh + 1 such
eigenvalues and hence the representation has dimension 2mh + 1. This must be an
integer so we learn that
2mh = 0, 1, 2, 3.... . (3.47)

We return to the issue about whether or not the eigenspaces |λ, m〉 can be more
than one-dimensional. If space of eigenvalues with m = mh is N -dimensional then
when we act with J− we obtain N -dimensional eigenspaces for each eigenvalue m. This
would lead to a reducible representation where one could simply take one-dimensional
subspaces of each eigenspace. Let us then suppose that there is only a one-dimensional
eigenspace for m = mh , spanned by |λ, mh 〉. It is then clear that acting with J− produces
all states and each eigenspace of J3 has only a one-dimensional subspace spanned by
|λ, m〉 ∝ (J− )n |λ, mh 〉 for some n = 0, 1, ..., 2λ + 1.
In summary, and changing notation slightly to match the norm, we have obtained
a (2l + 1)-dimensional unitary representation determined by any l = 0, 12 , 1, 32 , ... having
the Casimir J 2 = l(l + 1)I (in terms of what we had before l = mh ). The states can be
labelled by |l, m〉 where m = −l, −l + 1, ..., l − 1, l.
Let us look at some examples.
l = 0: Here we have just one state |0, 0〉 and the matrices Ji act trivially. This is the
trivial representation.
l = 1/2: Here we have 2 states:
󰀣 󰀤 󰀣 󰀤
1 0
| 12 , 12 〉 = | 12 , − 12 〉 = . (3.48)
0 1

By construction J3 is diagonal:
󰀣 󰀤
1/2 0
J3 = . (3.49)
0 −1/2

We can determine J+ through


󰁳
J+ | 12 , 12 〉 = 0 J+ | 12 , − 12 〉 = 3/4 − 1/4 + 1/2| 12 , 12 〉 = | 12 , 12 〉 (3.50)
3.2. GROUPS VS ALGEBRAS 39

so that 󰀣 󰀤
0 1
J+ = . (3.51)
0 0
And can determine J− through
󰁳
J− | 12 , 12 〉 = 3/4 − 1/4 + 1/2| 12 , − 12 〉 J− | 12 , − 12 〉 = 0 (3.52)

so that 󰀣 󰀤
0 0
J− = . (3.53)
1 0
Or alternatively
󰀣 󰀤
1 1 0 1
J1 = (J+ + J− ) =
2 2 1 0
󰀣 󰀤
1 1 0 −i
J2 = (J+ − J− ) = (3.54)
2i 2 i 0

Thus we have recovered the Pauli matrices.


40 CHAPTER 3. ALL ABOUT SU (2)
Chapter 4

Back to Angular Momentum

Let us return to thinking about angular momentum more concretely. Having constructed
all three operators

Li = −i󰄁󰂃ijk xj k (4.1)
∂x
is helpful to introduce the quadratic Casimir
L2 = L21 + L22 + L23 (4.2)
Please do not confuse L2 with L2 ! We have the result that
[L2 , Li ] = 0 (4.3)
Thus it follows that we can find states which are eigenstates of L2 and one of the
L′i s but no more. The usual choice by far is to take this to be L3 . So our states are
eigenstates of L2 (which we take to all have the same eigenvalue) and L3 . This is such
a well-studied system that the treatment is very standard. Let us introduce raising and
lowering operators
L+ = L1 + iL2 L− = L1 − iL2 (4.4)
which satisfy L− = L†+ . The commutation relations are now
[L3 , L+ ] = [L3 , L1 + iL2 ] = i󰄁L2 + 󰄁L1 = 󰄁L+
[L3 , L− ] = [L3 , L1 − iL2 ] = i󰄁L2 − 󰄁L1 = −󰄁L−
[L+ , L− ] = [L1 + iL2 , L1 − iL2 ] = −2i[L1 , L2 ] = 2󰄁L3 (4.5)
Finally we note that
󰀕 󰀖2 󰀕 󰀖2
2 L+ + L− L+ − L−
L = + + L23
2 2i
1󰀃 2 󰀄
= L+ + L2− + L+ L− + L− L+ − L2+ − L2− + L+ L− + L− L+ + L23
4
1
= (L+ L− + L− L+ ) + L23
2
1
= L− L+ + [L+ , L− ] + L23
2
= L− L+ + 󰄁L3 + L23 (4.6)

41
42 CHAPTER 4. BACK TO ANGULAR MOMENTUM

Everything we have done here is based on the simple algebraic fact that

[Ji , Jj ] = i󰄁εijk Jk (4.7)

where Li = 󰄁Ji Nothing has relied on the explicit differential form of the operators Li .
However now it might be helpful to see what the previous analysis looks like in terms
of these differential operators.

4.1 Differential Operators vs Irreducible Represen-


tations
Thus we have a representation of su(2) given by


Li = −i󰄁εijk xj (4.8)
∂xk
As written and interpreted the Li act on the space of differential functions on R3 . This
is an infinite dimensional vector space.
It’s not surprising then that this is a reducible representation. For example consider
the set of functions of the form ψ = f (x2 + y 2 + z 2 ) then

∂ψ df
Li ψ = −i󰄁εijk xj k
= −2i󰄁εijk xj xk =0 (4.9)
∂x dr
so these are an invariant subspace.
To continue it is wise to return to spherical coordinates

x = r sin θ cos φ
y = r sin θ sin φ
z = r cos θ (4.10)

To continue we need to invert the coordinate transformation to


󰁳
r= x2 + y 2 + z 2
θ = arccos(z/r)
φ = arctan(y/x) (4.11)

so that we can read off


󰀳 󰀴 󰀳 󰀴
∂r/∂x ∂r/∂y ∂r/∂z x/r y/r z/r
󰁅 󰁆 󰁅 󰁳 󰁳 󰁳 󰁆
󰁃 ∂θ/∂x ∂θ/∂y ∂θ/∂z 󰁄 = 󰁃zx/r3 1 − z 2 /r2 zy/r3 1 − z 2 /r2 −(1/r − z 2 /r3 )/ 1 − z 2 /r2 󰁄
∂φ/∂x ∂φ/∂y ∂φ/∂z −y/(x2 + y 2 ) x/(x2 + y 2 ) 0
󰀳 󰀴
sin θ cos φ sin θ sin φ cos θ
󰁅 󰁆
= 󰁃 cos θ cos φ/r cos θ sin φ/r − sin θ/r󰁄 (4.12)
− sin φ/r sin θ cos φ/r sin θ 0
4.1. DIFFERENTIAL OPERATORS VS IRREDUCIBLE REPRESENTATIONS 43

From here we find

∂ ∂r ∂ ∂θ ∂ ∂φ ∂
= + +
∂x ∂x ∂r ∂x ∂θ ∂x ∂φ
∂ cos θ cos φ ∂ sin φ ∂
= sin θ cos φ + −
∂r r ∂θ r sin θ ∂φ
∂ ∂r ∂ ∂θ ∂ ∂φ ∂
= + +
∂y ∂y ∂r ∂y ∂θ ∂y ∂φ
∂ cos θ sin φ ∂ cos φ ∂
= sin θ sin φ + +
∂r r ∂θ r sin θ ∂φ
∂ ∂r ∂ ∂θ ∂ ∂φ ∂
= + +
∂z ∂z ∂r ∂z ∂θ ∂z ∂φ
∂ sin θ ∂
= cos θ −
∂r r ∂θ
(4.13)

From which we find

∂ ∂
L1 = − i󰄁y + 󰄁z
∂z ∂y
󰀕 󰀖
∂ sin θ ∂
= − i󰄁r sin θ sin φ cos θ −
∂r r ∂θ
󰀕 󰀖
∂ cos θ sin φ ∂ cos φ ∂
+ i󰄁r cos θ sin θ sin φ + +
∂r r ∂θ r sin θ ∂φ
󰀕 󰀖
∂ ∂
= i󰄁 sin φ + cot θcos φ (4.14)
∂θ ∂φ

∂ ∂
L2 = −i󰄁z + i󰄁x
∂x 󰀕 ∂z 󰀖
∂ cos θ cos φ ∂ sin φ ∂
= −i󰄁rcos θ sin θ cos φ + −
∂r r ∂θ r sin θ ∂φ
󰀕 󰀖
∂ sin θ ∂
+ i󰄁r sin θ cos φ cos θ −
∂r r ∂θ
󰀕 󰀖
∂ ∂
= i󰄁 −cos φ + cot θsin φ (4.15)
∂θ ∂φ

and, nicest of all,

∂ ∂
L3 = −i󰄁x + 󰄁y
∂y ∂x
󰀕 󰀖
∂ cos θ sin φ ∂ cos φ ∂
= − i󰄁r sin θ cos φ sin θ sin φ + +
∂r r ∂θ r sin θ ∂φ
󰀕 󰀖
∂ cos θ cos φ ∂ sin φ ∂
+ i󰄁r sin θ sin φ sin θ cos φ + −
∂r r ∂θ r sin θ ∂φ

= −i󰄁 (4.16)
∂φ
44 CHAPTER 4. BACK TO ANGULAR MOMENTUM

We also find hence

L+ = L1 + iL2
󰀕 󰀖
iφ ∂ ∂
= i󰄁e −i + cot θ
∂θ ∂φ
L− = L1 − iL2
󰀕 󰀖
−iφ ∂ ∂
= ie i + cot θ (4.17)
∂θ ∂φ

From here we see that

L2 = L21 + L22 + L23


=L− L+ + 󰄁L3 + L23
󰀕 󰀖 󰀕 󰀖
−iφ ∂ ∂ iφ ∂ ∂
=i󰄁e i + cot θ i󰄁e −i + cot θ
∂θ ∂φ ∂θ ∂φ
2
∂ ∂
− i󰄁2 − 󰄁2 2
∂φ ∂φ
󰀕 2 󰀖
∂ i ∂ ∂2 ∂2 ∂2
=−󰄁 2
− + i cot θ − i cot θ 2
+ cot θ 2
∂θ2 sin2 θ ∂φ ∂θ∂φ ∂θ∂φ ∂φ
󰀕 2
󰀖
∂ ∂ ∂ ∂
− 󰄁2 cot θ + i cot2 θ +i + 2
∂θ ∂φ ∂φ ∂φ
󰀕 2 󰀖
∂ ∂ ∂2 ∂2
=−󰄁 2
+ cot θ 2
+ cot θ 2 + 2
∂θ2 ∂θ ∂φ ∂φ
󰀕 󰀕 󰀖 2
󰀖
1 ∂ ∂ 1 ∂
=−󰄁 2
sin θ + (4.18)
sin θ ∂θ ∂θ sin2 θ ∂φ2

We’ve seen these equations before: |l, m〉 = Yl,m are the spherical harmonics

L3 Yl,m = m󰄁Yl,m
L2 Yl,m = 󰄁2 l(l + 1)Yl,m (4.19)

The set {Yl,m |m = −l, ..., +l} therefore provides an irreducible, (2l + 1)-dimensional
representation of su(2) inside the space of all differentiable functions on S 2 (viewed as
the unit sphere in R3 ).
In particular we see that the 3-dimensional Hamiltonian with a spherically symmetry
potential can be written as
󰀕 󰀖
󰄁2 1 ∂ 2 ∂
H=− r + Vef f (r) (4.20)
2m r2 ∂r ∂r

with
󰄁2 2
Vef f (r) = V (r) + L
2m
󰄁2 l(l + 1)
= V (r) + (4.21)
2m
4.2. ADDITION OF ANGULAR MOMENTUM 45

where in the second line we have assumed that we are looking at angular momentum
eigenstates, i.e. eigenstates of L2 . It’s an easy excersize to see that L2 does indeed
commute with H.
Let us look for the associated eigenfunctions corresponding to the states |l, m〉. We
can also construct them relatively easily. These satisfy

L3 |l, m〉 = −i󰄁 |l, m〉 = m󰄁|l, m〉 (4.22)
∂φ
Thus

|l, m〉 = eimφ Θl,m (θ) (4.23)

for some function Θl,m (θ). We start with the highest weight state |l, l〉. We know that

0 = L+ |l, l〉
󰀕 󰀖
iφ ∂ ∂
= i󰄁e −i + cot θ eilφ Θl,l (θ)
∂θ ∂φ
󰀕 󰀖
d
0 = −i + il cot θ Θl,l (θ) (4.24)

This is easily solved:
dΘl,l
= l cot θdθ = ld ln sin θ
Θl,l
=⇒ Θl,l (θ) = Cl (sin θ)l (4.25)

We can then obtain

|l, m〉 = (L− )l−m |l, l〉


󰀕 󰀕 󰀖󰀖l−m
−iφ ∂ ∂
= Cl e − + i cot θ (sin θ)l eilφ
∂θ ∂φ
= Θl,m (θ)eimφ (4.26)

for some explicit function Θl,m . Note that we haven’t worried here about the normal-
isation but one can. But in practise these functions are given in books or built in to
Mathematica.

4.2 Addition of Angular Momentum


Let’s now imagine that we have two particles, each carrying angular momentum. The
combined system will have a single wavefunction

Ψ(t, x1 , x2 ) (4.27)

and we therefore have two sets of angular momentum operators:


(A) ∂
Li = −i󰄁εijk xjA (4.28)
∂xkA
46 CHAPTER 4. BACK TO ANGULAR MOMENTUM

with A = 1, 2. These satisfy


(A) (B) (B)
[Li , Lj ] = i󰄁εijk δ AB Lk (4.29)

i.e. each is a representation of su(2). It is easy to see that we have four commuting
observables:
(1) (2)
(L(1) )2 , L3 , (L(2) )2 , L3 (4.30)
(A) (A) (A)
where (L(A) )2 = (L1 )2 +(L2 )2 +(L3 )2 and hence we have four eigenvalues determined
by (l1 , m1 ; l2 , m2 ). Each of these has

−l1 ≤ m1 ≤ +l1
−l2 ≤ m2 ≤ +l2 (4.31)

so we have a total of (2l1 + 1)(2l2 + 1) states.


However what we are really interested in is the total angular momentum with oper-
ators
(T ) (1) (2)
Li = Li + Li (4.32)

From this point of view we would like to know the eigenvalues of


(T )
(L(T ) )2 , L3 , (L(1) )2 , (L(2) )2 (4.33)

Its easy to see that these all commute with each other and therefore we can find a basis
of eigenstates with eigenvalues determined by l, m, l1 , l2 . Here l and m3 relate to the
total angular momentum-squared and total angular momentum around the z-axis. This
amounts to a change of basis from |l1 , m1 ; l2 , m2 〉 to a basis labelled by |l, m, l1 , l2 〉 and
takes the form
l1
󰁛 l2
󰁛
|l, m, l1 , l2 〉 = Cml11,l,m
2 ,l
|l , m1 ; l2 , m2 〉
2 ,m 1
(4.34)
m1 =−l1 m2 =−l2

The constants Cml11,l,m2 ,l


2 ,m
are known as Clebsch-Gordan coefficients. There isn’t a very
nice formula for them (they can be found online, in books, or built into Mathematica).
And then can be derived using the methods we used before, starting with the mA = lA
states and working one’s way down with J identifying m = m1 + m2 .
Let us view all this from a more mathematical perspective. The 2-particle wavefunc-
tion Ψ(t, x1 , x2 ) lives in the tensor product of two 1-particle wavefunctions:

L2 (R6 ) = L2 (R3 ) ⊗ L2 (R3 ) (4.35)

Each of these factors carries a representation of su(2). It won’t be irreducible in general


but we can restrict to irreducible representations by looking at particular spherical har-
monics. Indeed let us restrict to H1 × H2 where HA are finite dimensional irreducible
4.2. ADDITION OF ANGULAR MOMENTUM 47

representations of su(2) (and spanned by spherical harmonics with fixed lA ) with di-
mensions 2l1 + 1 and 2l2 + 1 respectively. In such a finite dimensional Hilbert space
the angular momentum operators simply behave as matrices forming a representation
of su(2) (recall the definition of a representation):
(1) (2)
Π1 (Ji ) Π2 (Ji ) (4.36)
(A) (A)
Here we have switched from Li to Ji so we can drop ugly factors of 󰄁. We can lift
this to a representation on H1 × H2 by taking
(1) (2)
ΠT (Ji ) = Π1 (Ji ) ⊗ I + I ⊗ Π2 (Ji ) (4.37)

It’s easy to check that this is a representation. But it is generically reducible, i.e. not
irreducible. Rather it will be a direct sum of irreducible representations.
Without loss of generality we can take l1 ≥ l2 . It can be shown that it decomposes
into a direct sum over 2l2 + 1 irreducible representations with

l = l1 − l2 , l1 − l2 + 1, l1 − l2 + 2, ...., l1 + l2 (4.38)

Since representations can be labelled by their value of dimension 2l + 1 this is often


denoted as

(2l1 + 1) ⊗ (2l2 + 1) = (2(l1 − l2 ) + 1) ⊕ (2(l1 − l2 + 1) + 1) ⊕ . . . ⊕ (2(l1 + l2 ) + 1)


(4.39)

where the bold face means that we are talking about a whole representation not just a
number. As a check on this we can see that the dimensions agree. The dimension on
the left is (2l1 + 1)(2l2 + 1) whereas on the right we have
2l2
󰁛 2l2 (2l2 + 1)
(2(l1 − l2 + k) + 1) = (2(l1 − l2 ) + 1)(2l2 + 1) +
k=0
2
= (2l2 + 1)(2(l1 − l2 ) + 1 + 2l2 )
= (2l1 + 1)(2l2 + 1) (4.40)

as required.
Let us be more specific and look at the simplest non-trivial case of l1 = l2 = 1/2
(this doesn’t exactly correspond to angular momentum as it is non-integer but it does
correspond to the spin of an electron - more on this later). In the tensor product we
have the following basis of states (we drop the l1 = l2 = 1/2 to clean up the notation
since it is common to all states):

| 12 , 12 〉 = | 12 〉 ⊗ | 12 〉
| 12 , − 12 〉 = | 12 〉 ⊗ | − 12 〉
| − 12 , 12 〉 = | − 12 〉 ⊗ | 12 〉
| − 12 , − 12 〉 = | − 12 〉 ⊗ | − 12 〉 (4.41)
48 CHAPTER 4. BACK TO ANGULAR MOMENTUM

(1) (2)
These are eigenstates of J3 + J3
(1) (2) (1) (2)
(J3 + J3 )| 12 , 12 〉 = J3 | 12 〉 ⊗ | 12 〉 + | 12 〉 ⊗ J3 | 12 〉
= 12 | 12 〉 ⊗ | 12 〉 + 12 | 12 〉 ⊗ | 12 〉
= | 12 , 12 〉
(1) (2) (1) (2)
(J3 + J3 )| 12 , − 12 〉 = J3 | 12 〉 ⊗ | − 12 〉 + | 12 〉 ⊗ J3 | − 12 〉
= 12 | 12 〉 ⊗ | − 12 〉 − 12 | 12 〉 ⊗ | − 12 〉
=0
(1) (2) (1) (2)
(J3 + J3 )| − 12 , 12 〉 = J3 | − 12 〉 ⊗ | 12 〉 + | − 12 〉 ⊗ J3 | 12 〉
= − 12 | − 12 〉 ⊗ | 12 〉 + 12 | − 12 〉 ⊗ | 12 〉
=0
(1) (2) 1 (1) (2)
(J3 + J3 )| − 2
, − 12 〉 = J3 | − 12 〉 ⊗ | − 12 〉 + | − 12 〉 ⊗ J3 | − 12 〉
= − 12 | − 12 〉 ⊗ | − 12 〉 − 12 | − 12 〉 ⊗ | − 12 〉
= −| − 12 , − 12 〉 (4.42)
(T )
Thus we find the J3 eigenvalues (−1, 0, 0, +1). This doesn’t correspond to any irre-
ducible representation of su(2) but we can split it as 0 and (−1, 0, 1) which are the
eigenvalues of the l = 0 and l = 1 representations. Thus we expect to find

2⊗2=1⊕3 (4.43)

Indeed the correct basis is


1 1
|0, 0〉 = √ | 12 , − 12 〉 − √ | − 12 , 12 〉
2 2
|1, −1〉 = | − 12 , − 12 〉
1 1
|1, 0〉 = √ | 12 , − 12 〉 + √ | − 12 , 12 〉
2 2
1
|1, 1〉 = √ | 12 , 12 〉 (4.44)
2
You can check this for yourself.
Chapter 5

The Hydrogen Atom

Let us put what we learnt above to use in solving for an electron moving around a
positively charged nucleus. Here we can be a bit general and consider it as a two-
body problem where there is the position of the electron re and nucleus rn . Thus our
wavefunction is

Ψ(t, re , rn ) = e−iEt/󰄁 ψ(re , rn ) (5.1)

Furthermore the potential arising from the electrostatic force between them is
Ze2
V =− (5.2)
|rn − re |
where Z is the nucleon number and e the charge of an electron in appropriate units.
Thus the time dependent Schrödinger equation is
󰄁2 2 󰄁2 2 Ze2
Eψ = − ∇r e ψ − ∇r n ψ − ψ (5.3)
2me 2mn |rn − re |
where me and mn are the masses of the electron and nucleon respectively.

5.1 The Quantum 2-Body Problem


Thus we need to solve for a wavefunction of six coordinates. As in classical mechanics
we can exploit the fact that the potential only depends on rn − re to reduce the problem
to two one-body problems:

ψ(re , rn ) = ψCoM (R)ψrel (r12 ) (5.4)

for suitable choices of R and r12 namely:


me r e + mn r n
r12 = re − rn R= (5.5)
me + mn
In particular it is a small exercise (see the problem set) to show that
󰄁2 2 󰄁2 2 󰄁2 2 󰄁2 2
− ∇r e − ∇r n = − ∇R − ∇ (5.6)
2me 2mn 2M 2µ r12

49
50 CHAPTER 5. THE HYDROGEN ATOM

where M = me + mn and µ = me mn /M .
Thus our time-independent Schrödinger equation is

󰄁2 2 󰄁2 2 Ze2
Eψ = − ∇R ψ − ∇r12 ψ − ψ (5.7)
2M 2µ |r12 |
Now we can use seperation of variables and write

ψ(re , rn ) = ψCoM (R)ψrel (r12 ) (5.8)

and hence find two eigenvalue equations

󰄁2 2
󰂃CoM ψCoM = − ∇ ψCoM
2M R
󰄁2 Ze2
󰂃rel ψrel = − ∇2r12 ψrel − ψrel (5.9)
2µ |r12 |
with E = 󰂃CoM + 󰂃rel .

5.2 The Hydrogen Atom


The centre of mass part is trivial to solve:

ψCoM = eik·R χ(k) (5.10)

with 󰂃CoM = 󰄁2 |k|2 /2M . These just describe a free particle, the atom, in a basis where
the linear momentum is fixed. We can then construct wave packets through
󰁝
d3 k ik·R
ψCoM (R) = e χ(k) (5.11)
(2π)3
which will not be energy eigenstates but can be localized in position to an arbitrary
accuracy.
The more interesting part is to solve for ψrel (r12 ). Here we can switch from r12 to
spherical coordinates and use separation of variables yet again:

ψrel = ul (r)Yl,m (θ, φ) (5.12)

where r = |r12 | etc.. We already know what the Yl,m ’s are and we know that ul satisfies:
󰀕 󰀖 󰀕 󰀖
1 d 2 dul 2µ Ze2 󰄁2 l(l + 1)
− 2 r + 2 − + − 󰂃rel ul = 0 (5.13)
r dr dr 󰄁 r 2µr2

So solving the Hydrogen atom (which corresponds to Z = 1 but we can be more general
without causing any more pain) comes down to solving this equation and finding ul , E
and ψrel and ψ(re , rn ).
To continue we write

ul (r) = rl fl (r) (5.14)


5.2. THE HYDROGEN ATOM 51

to find
d2 fl 2(l + 1) dfl 2µZe2 2µ󰂃rel
− − − f l − fl = 0 (5.15)
dr2 r dr 󰄁2 r 󰄁2
This substitution is relatively standard in spherically symmetric examples as it removes
the l(l + 1) term from the equation.
To continue we look at the large r limit where only the first and last terms are
important. If 󰂃rel > 0 then we find oscillating solutions:
󰁳 󰁳
f ∼ C1 cos(− 2µ󰂃rel r/󰄁) + C2 sin( 2µ󰂃rel r/󰄁) (5.16)

which will not be normalisable. Solutions with 󰂃rel = 0 will also not be normaliable.
Thus we conclude that 󰂃rel < 0. In this case we expect solutions, at large r, to be of the
form
√ √
f ∼ C1 e − −2µ󰂃rel r/󰄁
+ C2 e −2µ󰂃rel r/󰄁
(5.17)

where now we must set C2 = 0 to find a normalizable solution.



Thus our last step is to write fl = Pl (r)e− −2µ󰂃rel r/󰄁 which leads to

d2 Pl 2(l + 1) dPl 2µZe2 2󰁳 dPl 2(l + 1) −2µ󰂃rel
− 2 − − 2 Pl + −2µ󰂃rel + Pl = 0 (5.18)
dr r dr 󰄁r 󰄁 dr 󰄁r
This probably doesn’t look much better but it is! Indeed it has polynomial solutions.
To see this let us assume we have a polynomial solution and look at the highest order
term Pl ∼ rn +. . .. The differential equation has a leading order rn−1 term with coefficient

2µZe2 󰁳 󰁳
− + 2 −2µ󰂃rel n + 2(l + 1) −2µ󰂃rel = 0 (5.19)
󰄁
which must vanish. We can rearrange this to give
󰀕 󰀖2
µZ 2 e4 1
󰂃rel =− (5.20)
2󰄁2 n+l+1

Let us return to the differential equation which we now write as

d2 Pl 2(l + 1) dPl 2µZe2 /󰄁2 µZe2 /󰄁2 dPl µZe2 /󰄁2 1


+ + P l − 2 − 2(l + 1) Pl = 0 (5.21)
dr2 r dr r n + l + 1 dr n+l+1r

next introduce ρ = 2µZe2 r/󰄁2 to find


󰀕 2 󰀖
d Pl 2(l + 1) dPl 1 dPl (l + 1)
(n + l + 1) 2
+ + Pl − − Pl = 0 (5.22)
dρ ρ dρ ρ dρ ρ

This is more amenable to solving through an expansion


n
󰁛
Pn,l = C k ρk (5.23)
k=0
52 CHAPTER 5. THE HYDROGEN ATOM

Substituting back in simply leads to a recursion relation obtained from setting the
coefficient of ρk−2 to zero:

(n + l + 1)(k(k − 1) + 2k(l + 1))Ck + ((n + l + 1) − (k − 1) − (l + 1)) Ck−1 = 0


(n + l + 1)(k(k − 1) + 2k(l + 1))
=⇒ Ck−1 = − Ck
n−k+1
(5.24)

We see that taking k = 0 leads to C−1 = 0 and hence the series indeed terminates giving
a polynomial. These polynomials are known as Laguerre Polynomials. We haven’t
shown here that these are the only normalizable solutions but this does turn out to be
the case.
In summary our solutions are of the form

ψ(re , rn ) = ψCoM (R)ψrel (re − rn )


µZe2 r

ψrel (re − rn ) = rl e (n+l+1)󰄁2 Pn,l (r)Θl (cos θ)eimφ (5.25)

and ψCoM is a generic free wave packet. It’s fun to plot |ψrel |2 for various choices of n, l
and m. For example look here https://en.wikipedia.org/wiki/Atomic_orbital#
/media/File:Hydrogen_Density_Plots.png On the other hand there are actual pic-
tures of atoms such as here: https://www.nature.com/articles/498009d
We can also reproduce the famous formula postulated by Bohr in the earliest days of
Quantum Mechanics (before all this formalism that we have discussed was formulated):

µZ 2 e4 1
EBohr = − (5.26)
2󰄁2 N 2
for some integer N = n + l + 1 = 1, 2, 3, ...,. In particular since a proton is 2000 times
more massive than an electron we have mn >> me and so µ = mn to very high accuracy.
Thus one finds (for Hydrogen where Z = 1)

13.6eV 2.18 × 10−18 J


EBohr ∼ − ∼ − (5.27)
N2 N2
This in turn also leads to the Rydberg formula for the energy spectrum of light coming
from excited atoms - known as spectral lines. In particular the energy of a photon (a
single particle of light) corresponds to the energy lost when an electron drops from a
higher orbital to a lower one (N1 > N2 ):

mp Z 2 e 4 1 mp Z 2 e 4 1
Ephoton = − +
2󰄁2 N12 2󰄁2 N22
󰀕 󰀖
1 1
=R 2
− 2 (5.28)
N2 N1

where we have computed the Rydberg constant R = mn Z 2 e2 /2󰄁4


We can also ask how many states there are at each energy level:

• N = 1: n = l = 0 1 state
5.2. THE HYDROGEN ATOM 53

• N = 2: n = 1, l = 0 or n = 0, l = 1 1+3=4 states

• N = 3: n = 2, l = 0 or n = 1, l = 1 or n = 0, l = 2 1+3+5 = 9 states

In fact we find N 2 states for each energy level (the sum of the first N odd numbers is
N 2 ).
However what are the lowest energy states with multiple electrons. In fact there
are twice as many as these as each electron can be spin up or spin down (more on this
later). This is just an additional quantum number (which only takes two values: up or
down) and corresponds to the fact that the correct rotation group of Nature isn’t SO(3)
but SU (2) = Spin(3). Note also that the terminology of spin up and spin down has
no physical interpretation in terms of up and down: if you take an electron of spin up
to Australia it does not become spin down. It has more to do with the fact that we
write vectors as column matrices and the up state are on top and then down state on
the bottom. We must know another fact: no two electrons can be in the same state
(the multi-electron wavefunction must be anti-symmetric - odd - under interchange of
any two electrons). Thus the degeneracies of low energy multi-electron states are (this
ignores inter electron interactions which become increasingly important)

• 2: n = l = 0

• 8: 2 states with n = 1, l = 0 and 6 from n = 0, l = 1

• 18: 2 from n = 2 6 from n = 1, l = 1 and 10 from n = 0, l = 2,

This pattern is evident in the periodic table whose rows have 2, 8 and 18 elements.
Which we have now predicted based on some crazy idea that states are vectors in a
Hilbert space and observables are self-adjoint linear maps.
54 CHAPTER 5. THE HYDROGEN ATOM
Chapter 6

Perturbation Theory

Next we must face the fact that solving the Schrödinger equation in general is too
complicated and we need to make approximations to make progress (for spherically
symmetric systems one is on better ground but still one has to solve a second order
ODE). The idea is to find a system you can solve exactly, for example a free, non-
interacting, system or a system you have solved exactly such as a single electron in an
atom, and then imagine that you perturb it (say by adding in another electron or putting
the atom in a small background magnetic field). This means by adding interaction whose
strength is controlled by a parameter g which we can make as small as we like such that
setting g = 0 reduces us the problem we can solve exactly. One then computes the
physically meaningful quantities in a power series:

E = E0 + gE1 + g 2 E2 + . . . (6.1)

The constants g are referred to as coupling constants - there can be several in any given
problem. The contributions at order g are called first-order and at g 2 second order etc..
Computing each term in this expansion is known as perturbation theory. Much, almost
all, of Physics is done this way.
Of course in Nature g is not a free parameter but some constant that you determine
from experiment. If g is small then we say the theory is weakly coupled whereas if g
is not small then it is strongly coupled. For example in electromagnetism the relevant
coupling constant is

e2 1
α= ∼ (6.2)
󰄁c 137
This is known as the fine structure constant. It is so named as it leads to corrections to
the Hydrogen atom spectrum that correspond to “fine”, e.g. small, corrections to what
we found above. This is indeed small. Nature was kind to Physicists as this meant
that accurate predictions from quantum electrodynamics (QED) could be made using
perturbation theory. Indeed in some cases theory and experiment agree to 13 decimal
place accuracy. This is like knowing the distance to the moon to within the width of a
human hair.

55
56 CHAPTER 6. PERTURBATION THEORY

On the other hand in Quantum Chromodynamics (QCD), the theory that describes
quarks inside protons and neutrons, the relevant coupling is
1
αs ∼ (6.3)
2
In fact neither α nor αs are constant. αs becomes larger at longer distances whereas α
gets smaller. The values I gave here are the approximate values of αs at the distance of
a proton and α at infinite distance. So we are not lucky with QCD (superconductivity
is another system that is strongly coupled and perturbation theory fails). At distances
around the size of a proton, QCD is strongly coupled and computations in perturbation
theory are next to useless (but we can use perturbation theory well at very short, sub
proton, distances where they behave in a weakly coupled way). This is asymptotic
freedom, meaning that at short distances the theory is weakly coupled. Pulling quarks
apart increases the force between them, somewhat like a spring, whereas the electric force
between two quarks gets weaker at large distances. Quarks are confined into protons
(and other baryons such as neutrons) but this is still poorly understood largely because
we can’t compute much.
It is important to ask what small means. It is meaningless to say that a quantity
that is dimensionful is small or large. By dimensionful we mean that it is measured in
some units. For example is a fighter jet fast? A typical speed for a fighter jet is 2000
kph and that is very fast compared to driving a car (say 100 kph):

vcar /vjet = 0.05 << 1 (6.4)

But the speed of light is 109 Kph so

vjet /vlight = 0.000002 << 1 (6.5)

So no in this sense a fighter jet is very very slow. Thus when we do perturbation theory
we need to identify a small dimensionless number. Small then means g 2 << g, so that
higher order terms in the expansion are indeed smaller (although their coefficients might
grow). Note that this requires us to compare g and g 2 so they must have the same units
and hence g must be dimensionless.
Lastly we note that even if we have a small, even tiny, dimensionless coupling con-
stant perturbation theory can still fail. This is because not all functions admit a Taylor
expansion about g = 0. The classic example is
󰀻 1
󰀿e− g 2 g ∕= 0
f (g) = (6.6)
󰀽0 g=0

This is an example of a non-analytic function. All derivatives of f (g) vanish at g = 0


and so a Taylor expansion about g = 0 gives
1

e g2 = 0 + 0 + 0... (6.7)
6.1. A SIMPLE EXAMPLE 57

i.e. perturbation theory misses all the information in f (g). This may seem like an esoteric
example but actually functions of this form arise all the time in quantum theories as
one can see in the path integral formulation.
In practice in quantum theories the perturbative series of the form 6.1 do not even
converge! They become more accurate as we include higher order terms for a while but
then they get worse if you include too many and then would ultimately diverge if one
could do infinitely many. In QED where g = α = 1/137 one expects the series to start
failing around the 137-th term. Such a series is known as an asymptotic series. The full
theory is not divergent and one expect the complete answer to take the form
2
E = (E0,0 + E0,1 g + E0,2 g 2 + . . .) + (E1,0 + E1,1 g + E1,2 g 2 + . . .)e−1/g + . . . (6.8)

6.1 A Simple Example


If we have a finite dimensional Hilbert space then solving the Schrödinger equations
boils down to diagonalising an N × N Hermitian matrix H. For illustration purposes
let us look at the simplest case of N = 2. Let’s look suppose that 󰂃 ∈ R and take
󰀣 󰀤
E1 󰂃
H= (6.9)
󰂃 E2
For 󰂃 = 0 we clearly have two eigenvalues E1 , E2 and eigenstates
󰀣 󰀤 󰀣 󰀤
1 0
|E1 〉 = |E2 〉 = (6.10)
0 1
This is the unperturbed system that we have solved exactly.
In this case the 󰂃 ∕= 0 system is simple enough that we can solve it exactly. So let’s
do that first and then compare with perturbation theory. It’s not hard to see that the
eigenvalues are
(󰂃) 1󰀓 󰁳
2 2
󰀔
E1/2 = E1 + E2 ± (E1 − E2 ) + 4󰂃 (6.11)
2
Here 1 refers to the + sign and 2 to the − sign. The square root is of a positive number
so we find real eigenvalues as we should. The eigenvectors are
󰀣 󰀤 󰀣 (󰂃) 󰀤
1 E2 −E2
(󰂃) (󰂃)
|E1 〉 = N1 E (󰂃) −E1 |E2 〉 = N2 󰂃
1
󰂃
1
󰀣 󰀤 󰀣 √ 󰀤
1 (E1 −E2 )− (E1 −E2 )2 +4󰂃2
= N1 −(E1 −E2 )+√(E1 −E2 )2 +4󰂃2 = N2 2󰂃 (6.12)
2󰂃
1
where the normalization constants are
󰁶
4󰂃2
N1/2 = 󰁳
󰂃2 + (E1 − E2 )2 ∓ 2(E1 − E2 ) (E1 − E2 )2 + 4󰂃2 + (E1 − E2 )2 + 4󰂃2
󰁶
4󰂃2
= 󰁳 (6.13)
5󰂃2 + 2(E1 − E2 )2 ∓ 2(E1 − E2 ) (E1 − E2 )2 + 4󰂃2
58 CHAPTER 6. PERTURBATION THEORY

Not nice but solved exactly! Note that for E1 ∕= E2 we can smoothly take 󰂃 → 0 as
󰀣 󰁶 󰀤
1 4󰂃 2
(󰂃)
E1/2 = E1 + E2 ± |E1 − E2 | 1 +
2 (E1 − E2 )2
󰀕 󰀖
1 2󰂃2
= E1 + E2 ± |E1 − E2 | ± + ...
2 |E1 − E2 |
󰂃2
= E1/2 ± + O(|󰂃|4 ) (6.14)
|E1 − E2 |

To first order in 󰂃 we find N1/2 = 1 + O(󰂃2 ) and


󰀣 󰀤 󰀣 󰀤
1 1
(󰂃)
|E1 〉 = √ = + O(󰂃2 )
−(E1 −E2 )+ (E1 −E2 )2 +4󰂃2 󰂃
2󰂃 E1 −E2
󰀣 √ 󰀤 󰀣 󰀤
(E1 −E2 )− (E1 −E2 )2 +4󰂃2 󰂃
(󰂃) − E1 −E
|E2 〉 = 2󰂃 = 2 + O(󰂃2 ) (6.15)
1 1

Thus we see that the small dimensionless parameter is g = 󰂃/|E1 − E2 |. Here we


can already identify a potential problem. What if E1 = E2 ? The problem is that even
at 󰂃 = 0 we could rotate |E1 〉 and |E2 〉 and still have energy eigenstates. So the limit
󰂃 → 0 need not return us to where we started. We will have to be more careful.

6.2 First Order Non-Degenerate Perturbation The-


ory
For now let us assume that E1 ∕= E2 . This is called non-degenerate perturbation
theory.
How would we handle this in perturbation theory? Let us switch notation a bit. At
zeroth order, i.e. 󰂃 = 0, we have (n = 1, 2)

H (0) |En(0) 〉 = Ea(0) |En(0) 〉 (6.16)

where
󰀣 󰀤 󰀣 󰀤 󰀣 󰀤
(0)
E1 0 (0) 1 (0) 0
H (0) = (0) |E1 〉 = |E2 〉 = (6.17)
0 E2 0 1

(0)
and we have relabelled En = En . Next we expand to lowest order in 󰂃. In principle
this involves an infinite number of terms but lets start with just the first order term

H = H (0) + 󰂃H (1) + . . .
En = En(0) + 󰂃En(1) + . . .
|En 〉 = |En(0) 〉 + 󰂃|En(1) 〉 + . . . (6.18)
6.2. FIRST ORDER NON-DEGENERATE PERTURBATION THEORY 59

We want to solve the following equations

H|En 〉 = En |En 〉 (6.19)

which to first order are

(H (0) + gH (1) + . . .)(|En(0) 〉 + 󰂃|En(1) 〉 + . . .) = (En(0) + 󰂃En(1) + . . .)(|En(0) 〉 + 󰂃|En(1) 〉 + . . .)


H (0) |En(0) 〉 + 󰂃H (1) |En(0) 〉 + 󰂃H (0) |En(1) 〉 + . . . = En(0) |En(0) 〉 + 󰂃En(1) |En(0) 〉 + 󰂃En(0) |En(1) 〉 . . .
(6.20)

The terms of order 󰂃0 cancel as we have solved the unperturbed problem. Thus, ignoring
higher orders in 󰂃, we find

H (1) |En(0) 〉 + H (0) |En(1) 〉 = En(1) |En(0) 〉 + En(0) |En(1) 〉 (6.21)


(0) (0) (1) (1)
Now we know H (0) , H (1) , En and |En 〉 so we need to solve for En and |En 〉. To this
(0)
end we use the fact that |En 〉 are an orthonormal basis. So
(0)
〈Em |H (1) |En(0) 〉 + 〈Em
(0)
|H (0) |En(1) 〉 = En(1) 〈Em
(0)
|En(0) 〉 + En(0) 〈Em
(0)
|En(1) 〉
= En(1) δnm + En(0) 〈Em
(0)
|En(1) 〉 (6.22)
(0) (0) (0)
However we also know that 〈Em |H (0) = Em 〈Em | and hence
(0)
〈Em |H (1) |En(0) 〉 = En(1) δnm + (En(0) − Em
(0) (0)
)〈Em |En(1) 〉 (6.23)

We can now take n = m to find

En(1) = 〈En(0) |H (1) |En(0) 〉 (6.24)

So we know the first order correction to the energies.


We can now write, for m ∕= n,

(0) 1
〈Em |En(1) 〉 = (0) (0)
(0)
〈Em |H (1) |En(0) 〉 (6.25)
En − Em
Here is where the non-degeneracy is important. On the other hand we want orthonormal
states |En 〉:

δnm = 〈En |Em 〉


= (〈En(0) | + 󰂃〈En(1) | + . . .)(|Em
(0) (1)
〉 + 󰂃|Em 〉 + . . .)
0 = 󰂃〈En(1) |Em
(0)
〉 + 󰂃〈En(0) |Em
(1)
〉 (6.26)

Thus we learn that

〈En(1) |Em
(0) (0)
〉 = (〈Em |En(1) 〉)∗ = −〈Em
(0)
|En(1) 〉 (6.27)

In particular we can take

〈En(1) |En(0) 〉 = 〈En(0) |En(1) 〉 = 0 (6.28)


60 CHAPTER 6. PERTURBATION THEORY

This tells us that the first order correction is orthogonal to the unperturbed eigenvalue.
So finally
󰁛
|En(1) 〉 = (0)
〈Em |En(1) 〉|Em
(0)

m
󰁛 1 (0)
= (0) (0)
〈Em |H (1) |En(0) 〉|Em
(0)
〉 (6.29)
m∕=n En − Em

We can then verify that


󰁛 1
〈Ep(0) |En(1) 〉 = (0) (0)
(0)
〈Em |H (1) |En(0) 〉〈Ep(0) |Em
(0)

m∕=n En − Em
1
= (0) (0)
〈Ep(0) |H (1) |En(0) 〉
En − Ep
󰀣 󰀤∗
1
=− (0) (0)
〈En(0) |H (1) |Ep(0) 〉
Ep − En
= −(〈En(0) |Ep(1) 〉)∗ (6.30)

as required for keep the eigenvectors orthonormal. Note that in this derivation, although
we started with a two-dimensional Hilbert space, we didn’t use this and our answer is
true more generally.
Let’s look at the simple example above where
󰀣 󰀤
0 1
H (1) = (6.31)
1 0

(1) (0) (0)


Here we see that En = 〈En |H (1) |En 〉 = 0 for n = 1, 2: e.g.
󰀣 󰀤󰀣 󰀤
󰀓 󰀔 0 1 1
(0) (0)
〈E1 |H (1) |E1 〉 = 1 0 =0
1 0 0
󰀣 󰀤󰀣 󰀤
󰀓 󰀔 0 1 0
(0) (0)
〈E2 |H (1) |E2 〉 = 0 1 =0 (6.32)
1 0 1

Thus the energies are not corrected to first order in 󰂃:

En = En(0) + O(󰂃2 ) (6.33)

This agrees with our exact result but will not be true in general since our H (1) is off-
diagonal.
(1)
However |En 〉 will be non-zero as:
󰀣 󰀤󰀣 󰀤
󰀓 󰀔 0 1 0
(0) (0)
〈E1 |H (1) |E2 〉 = 1 0 =1
1 0 1
󰀣 󰀤󰀣 󰀤
󰀓 󰀔 0 1 1
(0) (0)
〈E2 |H (1) |E1 〉 = 0 1 =1 (6.34)
1 0 0
6.3. SECOND ORDER NON-DEGENERATE PERTURBATION THEORY 61

From which we learn


(0) (0)
(0) 〈E2 |H (1) |E1 〉 (0)
|E1 〉 = |E1 〉 + 󰂃 (0) (0)
|E2 〉
− E1 E2
󰀣 󰀤 󰀣 󰀤
1 󰂃 0
= + (0) (0)
0 E1 − E2 1
󰀣 󰀤
1
= 󰂃 (6.35)
(0) (0)
E1 −E2

and
(0) (0)
(0) 〈E |H (1) |E2 〉 (0)
|E2 〉 = |E2 〉 + 󰂃 1 |E1 〉
E2 − E1
󰀣 󰀤 󰀣 󰀤
0 󰂃 1
= + (0) (0)
1 E2 − E1 0
󰀣 󰀤
− (0) 󰂃 (0)
= E1 −E2 (6.36)
1
Which agrees with what we found above.
However to see the shift in energy we will have to go to second order in perturbation
theory!

6.3 Second Order Non-Degenerate Perturbation The-


ory
To see the correction to the energy in our simple example we need to go one step further
and consider second order terms:

H = H (0) + 󰂃H (1) + 󰂃2 H (2) + . . .


En = En(0) + 󰂃En(1) + 󰂃2 En(2) + . . .
|En 〉 = |En(0) 〉 + 󰂃|En(1) 〉 + 󰂃2 |En(2) 〉 + . . . (6.37)

We still want to solve the following equations

H|En 〉 = En |En 〉 (6.38)

But now we have to expand to second order

(H (0) + 󰂃H (1) + 󰂃2 H (2) + . . .)(|En(0) 〉 + 󰂃|En(1) 〉 + 󰂃2 |En(2) 〉 + . . .)


= (En(0) + 󰂃En(1) + 󰂃2 En(2) + . . .)(|En(0) 〉 + 󰂃|En(1) 〉 + 󰂃2 |En(2) 〉 . . .) (6.39)

But we can assume that the zeroth and first order equations have been solved so we
find, at second order,

H (2) |En(0) 〉 + H (1) |En(1) 〉 + H (0) |En(2) 〉 = En(0) |En(2) 〉 + En(1) |En(1) 〉 + En(2) |En(0) 〉 (6.40)
62 CHAPTER 6. PERTURBATION THEORY

(2) (2)
Remember that the unknowns are En and |En 〉 everything else is known. So we take
matrix elements again:

(0)
〈Em |H (2) |En(0) 〉 + 〈Em
(0)
|H (1) |En(1) 〉 + 〈Em
(0)
|H (0) |En(2) 〉
= En(0) 〈Em
(0)
|En(2) 〉 + En(1) 〈Em
(0)
|En(1) 〉 + En(2) 〈Em
(0)
|En(0) 〉 (6.41)

which becomes

(0)
〈Em |H (2) |En(0) 〉 + 〈Em
(0)
|H (1) |En(1) 〉 + Em
(0) (0)
〈Em |En(2) 〉
= En(0) 〈Em
(0)
|En(2) 〉 + En(1) 〈Em
(0)
|En(1) 〉 + En(2) δnm (6.42)

Again we first look at m = n and find

〈En(0) |H (2) |En(0) 〉 + 〈En(0) |H (1) |En(1) 〉 + En(0) 〈En(0) |En(2) 〉 = En(0) 〈En(0) |En(2) 〉 + En(2) (6.43)
(0) (1) (0) (2)
where we have used 〈En |En 〉 = 0. The 〈En |En 〉 terms cancel so this tells us the
second order correction to En :

En(2) = 〈En(0) |H (2) |En(0) 〉 + 〈En(0) |H (1) |En(1) 〉


󰁛 〈Em
(0) (0)
|H (1) |En 〉
= 〈En(0) |H (2) |En(0) 〉 + 〈En(0) |H (1) (0) (0)
(0)
|Em 〉
m∕=n En − Em
󰁛 〈Em
(0) (0) (0) (0)
|H (1) |En 〉〈En |H (1) |Em 〉
= 〈En(0) |H (2) |En(0) 〉 + (0) (0)
(6.44)
m∕=n En − Em

(2)
Since we can compute everything on the right hand side we now know En . It is left
(2)
and an exercise to compute |En 〉 (just do it for H (2) = 0).
So let us try to compute the correction to the energy in our simple example. Here
we have H (2) = 0 and so (using the matrix elements we computed before 6.34)
(0) (0) (0) (0)
(2) 〈E2 |H (1) |E1 〉〈E1 |H (1) |E2 〉 1
E1 = (0) (0)
=
E1 − E2 E1 − E2
(0) (0) (0) (0)
(2) 〈E1 |H (1) |E2 〉〈E2 |H (1) |E1 〉 1
E2 = (0) (0)
=− (6.45)
E2 − E1 E1 − E2

which agrees with what we found above from the exact solution 6.14.
Let us summarise what we found and in a slightly cleaner notation. Typically (but
not always) the perturbation to the Hamiltonian comes from the potential so

H (1) = V (1) (6.46)


(0)
and for simplicity we take H (2) = 0. Let us denote the unperturbed system by |En 〉 =
|n〉 and introduce the matrix elements

〈m|V (1) |n〉 = Vmn (6.47)


6.4. THIRD ORDER NON-DEGENERATE PERTURBATION THEORY 63

If we looking at wavefunction on R3 then


󰁝
Vnm = ψn∗ (x)V (1) (x)ψm (x)d3 x (6.48)

where ψn (x) are a basis of normalised eigenfunctions of the unperturbed Hamiltonian:

󰄁2 2
En(0) ψn = − ∇ ψn + V (0) ψn (6.49)
2m
Then our formulae are (see the problem set)
󰁛 Vmn Vnm
En = En(0) + 󰂃Vnn + 󰂃2 (0) (0)
+ ...
m∕=n En − Em
󰁛 Vnm
|En 〉 = |n〉 + 󰂃 (0) (0)
|m〉
m∕=n En − Em
󰀥
󰁛󰁛 Vpm Vmn 󰁛 Vmm Vmn
+ 󰂃2 (0) (0) (0) (0)
|m〉 − (0) (0)
|m〉
− − 2
m∕=n p∕=n (En Em )(En Ep ) m∕=n (En − Em )
󰀦
1󰁛 Vmn Vmn
− |n〉 + . . . (6.50)
2 m∕=n (En(0) − Em
(0) 2
)

6.4 Third Order Non-Degenerate Perturbation The-


ory
You’re on your own here.

6.5 Degenerate Perturbation Theory


It’s clear from our previous discussion that if two or more energy eigenvalues are equal,
(0) (0)
so that En = Em for some pari m ∕= n then all Hell breaks loose. Lets go back to our
simple example and look at what happens when E1 = E2 = E:
󰀣 󰀤
E 󰂃
H= (6.51)
󰂃 E

Again it’s easy to solve for the energy


1󰀓 󰁳
2 2
󰀔
E1/2 = E1 + E2 ± (E1 − E2 ) + 4󰂃 = E ± |󰂃| (6.52)
2
and the eigenvectors are
󰀣 󰀤 󰀣 󰀤
1 1 1 −sgn(󰂃)
|E1 〉 = √ |E2 〉 = √ (6.53)
2 sgn(󰂃) 2 1

where sgn(󰂃) = 󰂃/|󰂃|. Here we see the problem: the answer is not analytic in 󰂃 and so a
naive Taylor series expansion will fail. In addition taking 󰂃 → 0± does not give us the
64 CHAPTER 6. PERTURBATION THEORY

unperturbed eigenvectors. This is fine as the eigenvectors at 󰂃 = 0 are degenerate and


hence we could just as well as started with
󰀣 󰀤 󰀣 󰀤
(0) 1 1 (0) 1 −1
|E1 〉 = √ |E2 〉 = √ (6.54)
2 1 2 1
or the other way around.
Thus we need to be more careful and treat degenerate eigenspaces as a special case.
The degeneracy maybe just a single pair, or more typically, every energy eigenstate has
some fixed degeneracy, e.g. two spin states of an electron.
How do we fix it? Well going back to the previous discussion for first order pertur-
bation theory we saw above:

H = H (0) + 󰂃H (1) + . . .
En = En(0) + 󰂃En(1) + . . .
|En 〉 = |En(0) 〉 + 󰂃|En(1) 〉 + . . . (6.55)

which led to
(0)
〈Em |H (1) |En(0) 〉 = En(1) δnm + (En(0) − Em
(0) (0)
)〈Em |En(1) 〉 (6.56)

Once again we can take m = n to read off that

En(1) = 〈En(0) |H (1) |En(0) 〉 (6.57)

So this formula remains the same. But now in the degenerate case we therefore have for
(0) (0)
some subset, labelled by {n′ , m′ }, of degenerate energy states En′ = Em′ ,
(0) (0) (1)
〈Em′ |H (1) |En′ 〉 = En′ δn′ m′ (6.58)
(0) (0)
This is a constraint on the choice of |En′ 〉. In particular we need to choose |En′ 〉 to be
(1)
eigenstates of H (1) with eigenvalue En′ . This is possible since H (1) is self-adjoint and
H (0) is proportional to the identity when acting on the degenerate subspace.
We then find that the degeneracy is lifted by the eigenvalues of H (1) in the degenerate
subspace
(0) (0) (0)
En′ = En′ + 󰂃〈En′ |H (1) |En′ 〉 + . . .
(0) (1)
= En′ + 󰂃En′ + . . . (6.59)
(0) (1) (0)
where H (1) |En′ 〉 = En′ |En′ 〉. Whereas for non-degenerate eigenvalues we find
(0) (0) (0)
En′′ = En′′ + 󰂃〈En′′ |H (1) |En′′ 〉 + . . . (6.60)

where n′′ labels the non-degenerate eigenstates. We then solve for the eigenvectors by

(1)
󰁛 〈Em
(0) (0)
|H (1) |En′′ 〉 (0)
|En′′ 〉 = (0) (0)
|Em 〉
m∕=n′′ En′′ − Em

(1)
󰁛 〈E (0)′′ |H (1) |E (0)
′ 〉 (0)
m n
|En′ 〉 = (0) (0)
|Em′′ 〉 (6.61)
m′′ E n′ − Em′′
6.6. TIME DEPENDENT PERTURBATION THEORY 65

Thus looking at our simple two-dimensional example we should start from the basis
󰀣 󰀤 󰀣 󰀤
1 1 1 −1
|E1 〉 = √ + O(󰂃2 ) |E2 〉 = √ + O(󰂃2 ) (6.62)
2 1 2 1

which are eigenstates of


󰀣 󰀤
0 1
H (1) = (6.63)
1 0

with eigenvalues 1 and −1 respectively and then we find

E1 = E + 󰂃 + O(󰂃2 ) E2 = E − 󰂃 + O(󰂃2 ) (6.64)

In fact in this simple case the perturbation ends at first order but this won’t be true in
general.
Of course it could be that there are still degeneracies but then one simply goes to
the next order. In quantum theories one expects that the only degeneracies that persist
to all orders in perturbation theory are those that are protected by a symmetry. That
is there exists an observable Q that commutes with the full Hamiltonian [Q, H] = 0
and hence the states |En 〉 and Q|En 〉 have the same energy to all orders in perturbation
theory:

HQ|En 〉 = QH|En 〉 = En Q|En 〉 (6.65)

One then works with a energy eigenstates that are also eigenstates of Q.

6.6 Time Dependent Perturbation Theory


Our next topic in perturbation theory is to allow for a small time dependence. So
far we have assumed that the Hamiltonian has no explicit time dependence. But this
need not be the case. So what happens if we consider a perturbation that includes
time-dependence:
∂|Ψ〉
i󰄁 = H (0) |Ψ〉 + gH (1) (t)|Ψ〉 (6.66)
∂t
(we assume the unperturbed Hamiltonian is time independent). Recall that in the
time-independent case we solve the Schröding equation by
󰁛
|Ψ(0) (t)〉 = cn e−i󰄁En t/󰄁 |ψn(0) 〉 (6.67)
n

(0)
where |ψn 〉 are a basis of eigenstates of H (0) :

H (0) |ψn(0) 〉 = En(0) |ψ (0) 〉 (6.68)

In this case the cn ’s are just constants that characterise the state at t = 0.
66 CHAPTER 6. PERTURBATION THEORY

To solve the perturbed Schrödinger equation we allow cn to be functions of time.


Substituting back we now find
󰁛 󰀕 dcn 󰀖 󰁛
i󰄁 + En cn e−i󰄁En t/󰄁 |ψn(0) 〉 = cn e−iEn t/󰄁 H (0) |ψn(0) 〉
n
dt n
󰁛
+g cn e−iEn t/󰄁 H (1) (t)|ψn(0) 〉 (6.69)
n

Next we use the fact that the ψn satisfy the unperturbed time-independent Schrödinger
equation so
󰁛 dcn 󰁛
i󰄁 e−iEn t/󰄁 |ψn(0) 〉 = g cn (t)e−iEn t/󰄁 H (1) (t)|ψn(0) 〉 (6.70)
n
dt n

(0)
Note that again we can still use the fact that the |ψn 〉 are an orthonormal basis of the
(0)
Hilbert space. Thus we can take the innerproduct of this equation with |ψm 〉 to obtain
dcm −iEm t/󰄁 󰁛
i󰄁 e =g cn e−iEn t/󰄁 〈ψm
(0)
|H (1) (t)|ψn(0) 〉
dt n
dcm ig 󰁛
=− cn ei(Em −En )t/󰄁 〈ψm(0)
|H (1) (t)|ψn(0) 〉 (6.71)
dt 󰄁 n
(0)
For the usual case where |ψn 〉 are realised by functions on R3 we have
󰁝
(1)
〈ψm |H (t)|ψn 〉 = (ψm (0)
(x))∗ H (1) (t)ψn(0) (x)d3 x (6.72)

Thus we obtain a first order differential equation for each cn (t) but it is coupled to all
other c’s. Thus cn (t) is determined by the original initial condition cn (0).
So far we haven’t made any approximation. But we have an infinite set of coupled
differential equations! At lowest order cn is constant so let us expand

cm (t) = cm (0) + gc(1) 2 (2)


m (t) + g cm (t) + . . . (6.73)
(k)
with the boundary condition that cm (0) = 0 for k = 1, 2, 3.... This sets up an infinite
(k+1) (k)
series of recursion relations between cm and cm but the first order one is
i󰁛
(1)
dcm
=− cn (0)ei(Em −En )t/󰄁 〈ψm
(0)
|H (1) (t)|ψn(0) 〉 (6.74)
dt 󰄁 n

which the solution


󰁝
i󰁛 t

c(1)
m (t) =− cn (0) ei(Em −En )t /󰄁 〈ψm
(0)
|H (1) (t′ )|ψn(0) 〉dt′ (6.75)
󰄁 n 0

It is natural to imagine a situation where the perturbation is turned on at t = 0


and then switched off at some later stage. That is to say H (1) (t) = 0 for t ≤ 0 and
H (1) (t) → 0 rapidly at late times. Further let us suppose that at t ≤ 0 the system is in
the k-th energy eigenstate:

cm (0) = δmk (6.76)


6.6. TIME DEPENDENT PERTURBATION THEORY 67

then we find that


󰁝 t
i ′ (0)
c(1)
m (t) =− ei(Em −Ek )t /󰄁 〈ψm
(0)
|H (1) (t′ )|ψk 〉dt′ (6.77)
󰄁 0

The interpretation is that |cm (t)|2 gives the probablity that after the perturbation the
(1)
system will lie in the m-th energy state (m ∕= k) at time t. cm (t) is called the first order
transition amplitude. Note that we don’t expect to find
󰁛
|cn (t)|2 = 1 (6.78)
n

The interpretation is that the perturbation has introduced (or taken away) energy into
the system which then gets redistributed.
So lets do an example! Consider a harmonic oscillator
p̂2 1
H (0) = + kx̂2 (6.79)
2m 2
with states |n〉, n = 0, 1, 2... and energies En = 󰄁ω(n + 12 ). Next we perturb it by
2 /τ 2
H (1) = −x̂e−t (6.80)

Classically this corresponds to an applied force F = g in the x-direction but only for a
period of time of the order 2τ . In this case we assume that the system is in the ground
state at t → −∞. Thus we need to evaluate (we have change the initial time to t = −∞)
󰁝
i t iωnt′
(1)
cn (t) = − e 〈n|H (1) (t′ )|0〉dt′
󰄁 −∞
󰁝 t
i ′ ′2 2
= − 〈n|x̂|0〉 eiωnt −t /τ dt′ (6.81)
󰄁 −∞

There is no closed form for this integral but we can look at late times so
󰁝 ∞
i ′ ′2 2
(1)
cn (∞) = − 〈n|x̂|0〉 eiωnt −t /τ dt′
󰄁 −∞

iτ π −τ 2 n2 ω2 /4
=− e 〈n|x̂|0〉 (6.82)
󰄁
where we have used to integral
󰁝 ∞
󰁵
−at′2 +bt′ π b2 /4a
e dt′ = e (6.83)
−∞ a
Lastly we need to compute
󰁵
󰄁
〈n|x̂|0〉 = 〈n|â + ↠|0〉
󰁵 2mω
󰄁
= 〈n|↠|0〉
󰁵 2mω
󰄁
= δn,1 (6.84)
2mω
68 CHAPTER 6. PERTURBATION THEORY

So it is only non-zero for the first excited state:


󰁵
π 2 2
c1 (∞) = −ig τ e−τ ω /4 (6.85)
2m󰄁ω

More generally its easy to see that if we started in |N 〉 then we could only jump to
|N + 1〉 with the same dependence on τ .
To make predictions it is necessary to see what happens to the ground state. The
equation for c0 is
(1)
c0 (t) = 1 + gc0 (t) + . . . (6.86)

with
󰁝 ∞
(1) i
c0 (∞) =− 〈0|H (1) (t′ )|0〉dt′
󰄁 −∞

=0 (6.87)

as 〈0|(a + a† )|0〉 = 0. Thus at late times our state is


󰁵
−iωt/2 π 2 2
|Ψ(t → ∞)〉 = e |0〉 − ig τ e−τ ω /4 e−3iωt/2 |1〉 (6.88)
2m󰄁ω
This is not normalized but one could normalize it and read off the relative probabilities
for finding the system in the ground state |0〉 or first excited state |1〉. To lowest order
in g we find
1
P0→0 = ∼ 1 − A2
1 + A2
A2
P0→1 = ∼ A2 (6.89)
1 + A2
where
πg 2 2 −ω2 τ 2 /2
A2 = τ e << 1 (6.90)
2m󰄁ω
Chapter 7

Semi-Classical Quantization: WKB

Let us now look at another type of approximation. We expect that for large systems,
whatever that means, classical results should emerge. Much of the quantum-ness of
quantum theory disappears. This disappearance arises as the wavefunctions typically
oscillate so quickly that only the quantum effects cancel out and only the dominant
classical configurations remain.

7.1 WKB
This motivates the idea that we should consider wave-functions of the form (let’s just
think of one dimension)
ψ(x) ∼ eiσ(x)/󰄁 (7.1)
where we aren’t worrying about normalizations at this time. We can fix that later. To
proceed we find the following (restricting to one dimension)
dψ i dσ iσ(x)/󰄁
= e
dx 󰄁 dx
󰀕 󰀖2
d2 ψ i d2 σ iσ(x)/󰄁 1 dσ
= e − 2 eiσ(x)/󰄁 (7.2)
dx2 󰄁 dx2 󰄁 dx
The time-independent Schrr̈odinger equation is therefore
d2 ψ
Eψ = − + V (x)ψ
dx
󰀕 󰀖2
i󰄁 d2 σ 1 dσ
E=− 2
+ + V (x) (7.3)
2m dx 2m dx
We are now in a position to make a semi-classical expansion:
σ = σ (0) + 󰄁σ (1) + . . . (7.4)
which gives at zeroth and first order (we assume there is no 󰄁 in V )
󰀕 (0) 󰀖2
1 dσ
E= + V (x)
2m dx
i d2 σ (0) dσ (0) dσ (1)
0=− + (7.5)
2 dx2 dx dx

69
70 CHAPTER 7. SEMI-CLASSICAL QUANTIZATION: WKB

The semi-classical limit is, roughly, “󰄁 → 0”. I put this in quotes as 󰄁 has units (and
can be set to one by a choice of units). What we mean is that the first term is small
compared to the second:
󰀕 󰀖2
d2 σ dσ
󰄁 2 << (7.6)
dx dx
To take the semi-classical limit we therefore ignore the first term which in a formal
sense corresponds to 󰄁 = 0. Thus at zeroth order we find a familiar formula:
dσ (0) 󰁳
= ± 2m(E − V (x)) (7.7)
dx
which is solved by
󰁝 x 󰁳
(0)
σ (x) = ± p(y)dy p(y) = 2m(E − V (y)) (7.8)
x0

Thus we have determined the wavefunction (in principle and assuming the condition
7.6). In other words our semi-classical wavefunction is
i
󰁕x i
󰁕x
ψsemiclassical = Ae 󰄁 p(y)dy
+ Be− 󰄁 p(y)dy
(7.9)

where we have allowed for the two possible choices of sign.


The integrand in 7.8 has a classical interpretation where p is the momentum of a
particle in a potential V (y):
p2
+ V (y)E= (7.10)
2m
When does our approximation 7.6 break down? We see that
󰁶 󰀕 󰀖2
2 (0)
dσ 1 2m dV dσ
2
=± = 2m(E − V ) (7.11)
dx 2 (E − V ) dx dx

Assuming that V is well behaved, i.e. dV /dx bounded, then we find the semi-classical
approximation breaks down at the ‘turning points’ where V = E. Here the momentum
p vanishes so nothing can be smaller than it.
Next we can easily solve for the first order term:
i d2 σ (0) dσ (0) dσ (1)
2
=
󰀕2 dx (0)
󰀖 dx dx
i d dσ dσ (1)
ln =
2 dx dx dx
i
ln p = σ (1) + const (7.12)
2
We can absorb the constant in the normalization cofficients A, B and hence, to first
order,
A i
󰁕x B i
󰁕x
ψW KB (x) = 󰁳 e󰄁 p(y)dy
+󰁳 e− 󰄁 p(y)dy
(7.13)
p(x) p(x)
This is known as the WKB approximation after Wentzel, Kramers and Brillouin. It
captures a surprisingly large amount of quantum information.
7.2. PARTICLE IN A BOX 71

7.2 Particle in a Box


Let us look at the particle in a box where V = 0 but we restrict the wavefunction to
vanish for x ≤ 0 and x ≥ L. Our wave function is therefore

A i 󰁕x B 󰁕
i x
ψ(x) = √ e 󰄁 pdy + √ e− 󰄁 pdy
p p
i i
= A′ e 󰄁 px + B ′ e− 󰄁 px (7.14)

where p = 2mE is constant and we have absorbed p into the coefficients A′ , B ′ . Next
we need to impose boundary conditions ψ(0) = ψ(L) = 0 which leads to

A′ + B ′ = 0 A′ eipL/󰄁 + B ′ e−ipL/󰄁 = 2A′ sin(pL/󰄁) = 0 (7.15)

So we find p = n󰄁π/L and E = n2 π 2 󰄁2 /2mL2 for n = 1, 2, .... This agrees perfectly with
the exact answer, for example see the discussion session in week 3 (because d2 σ (0) /dx2 =
dp/dx = 0). Next!

7.3 Turning Points, Airy and Bohr-Sommerfeld Qua-


tization
So you might think that the next easiest thing to consider is the harmonic oscillator -
that always works! Well no the Devil is in the detail and in the WKB approximation
the detail is the behaviour near a turning point. Near a turning point we can assume
the the potential is roughly linear. So let us look at this example first exactly before
using WKB.
Let us consider
󰄁2 d 2 ψ
Eψ = − + F xψ (7.16)
2m dx2
where F is a constant. Classically this is just a linear force in the x-direction. The
energy won’t be bounded from below but we can still find solutions. Cleaning this up
we first shift
󰄁2 d 2 ψ
x = x′ + E/F =⇒ − + x′ F ψ = 0 (7.17)
2m dx′2
and next we rescale

x′ = (󰄁2 /2mF )1/3 z (7.18)

to obtain the deceptively simple equation

d2 ψ
− zψ = 0 (7.19)
dz 2
72 CHAPTER 7. SEMI-CLASSICAL QUANTIZATION: WKB

The solution to this equation which decays at large z is known as the Airy function1 :

ψ = cAi(z) (7.20)

for some constant c. It is named after George Airy who was an Astronomer Royal at
the Greenwich Observatory. He features in their museum if you have ever been. The
Airy function is the poster-child of difficult functions to understand using perturbative
techniques due to its bi-polar characteristics of oscillating and decaying. But we love
it none the less as it is a thing of beauty with important and varied applications (and
with modern computer techniques it can be evaluated numerically with high precision).
It can be defined by
󰁝 ∞
1
Ai(z) = √ cos( 13 u3 + zu)du (7.21)
π 0
From here you can check that
󰁝 ∞
′ 1
Ai (z) = − √ u sin( 13 u3 + zu)du
π 0
󰁝 ∞
′′ 1
Ai (z) = − √ u2 cos( 13 u3 + zu)du (7.22)
π 0
so
󰁝 ∞
′′ 1
Ai (z) − zAi(z) = − √ (u2 + z) cos( 13 u3 + zu)du
π 0
󰁝 ∞
1 d
= −√ sin( 13 u3 + zu)du
π 0 du
=0 (7.23)

(one must be careful with the boundary term at infinity which is oscillating wildly). A
plot of Ai(z) is given below. It oscillates to the left of the turning point (where E > F x)
and then dies off exponentially to the right (where E < F x). Therefore, near a turning
point, this is what we expect the wavefunction to look like. In particular the asymptotic
form of Ai(z) is known:
󰀻
󰀿 1 e− 23 z3/2 z→∞
1/4
Ai(z) ∼ z (7.24)
󰀽 1 sin 2 |z|3/2 + π/4󰀄
󰀃
z → −∞
|z|1/4 3

It’s worth mentioning here the fact that Ai(z) and hence the wavefunction is not zero
to the right of the turning point where the potential energy V = F x is greater than
the total energy: V > E. Thus there is a non-zero probability to find the particle in a
region which is strictly forbidden in the classical world.
So let us try this with the WKB approximation. The idea is that WKB should
be good away from the turning points and near a turning point we can use the Airy
function. So it’s a question of patching together the various wavefunctions.
1
There is a second solution that grows at large z.
7.3. TURNING POINTS, AIRY AND BOHR-SOMMERFELD QUATIZATION 73

0.4

0.2

-20 -10 10 20

-0.2

-0.4

Figure 7.3.1: The Airy function Ai(z)

So lets do the WKB procedure for V = E + F (x − b). This is a potential that rises
to the right (F > 0) with the turning point at x = b:
󰁳
p(x) = 2m(F b − F x) (7.25)

and hence (we pick the lower bound of integration constant to make things simple as
the effect is just a constant that can be absorbed elsewhere)
󰁝 x󰁳
(0)
σ = 2m(F b − F y)dy
b
2√
=− 2mF (b − x)3/2 (7.26)
3
Thus our WKB wavefunction is
A i 2 √2mF (b−x)3/2 B −i 2 √2mF (b−x)3/2
ψ= e 3󰄁 + e 3󰄁 (7.27)
p(x) p(x)
but we don’t trust it near x = b and in particular p = 0 there. Rather we take the
following
󰀻 2 √ 2 √
󰁁
󰁁 √Ar
e −
3󰄁
2mF (x−b)3/2
+ √Br
e +
3󰄁
2mF (x−b)3/2
x>b
󰁁
󰁁
󰀿 p(x)󰀃 󰀄
p(x)

ψ = cAi (2mF/󰄁)1/3 (x − b) x∼b (7.28)


󰁁
󰁁 √ √
󰁁
󰁁
2 3/2 2 3/2
󰀽 √Al ei 3󰄁 2mF (b−x) + √Bl e−i 3󰄁 2mF (b−x) x<b
p(x) p(x)

Normalizability tells us that Br = 0. Looking at the asymptotic values of Ai for positive


z we find agreement if Ar = c. Looking at the large negative x − b we find agreement if
c iπ/4 c −iπ/4
Al = e Bl = − e (7.29)
2i 2i
74 CHAPTER 7. SEMI-CLASSICAL QUANTIZATION: WKB

Thus our solution is


󰀻 2 √
󰁁
󰁁 √ c
e −
3󰄁
2mF (x−b)3/2
x>b
󰁁
󰁁
󰀿 p(x)󰀃 󰀄
ψ = cAi (2mF/󰄁)1/3 (x − b) x∼b (7.30)
󰁁
󰁁 󰀓 √ 󰀔
󰁁
󰁁
󰀽 √ c sin 2 2mF (b − x)3/2 + π/4
3󰄁
x<b
p(x)

In other words the WKB solution agrees with the asymptotic regions of the exact Airy
function but not the region where it changes from oscillating to damping.
We can also imagine a turning point, rising to the left, located at x = a < b. The
analysis is the same but with x − b → a − x.
󰀻 󰀓 √ 󰀔
󰁁
󰁁 √ c
sin − 2
2mF (x − a) 3/2
− π/4 x>a
󰁁
󰁁 3󰄁
󰀿 p(x)󰀃 󰀄
ψ = cAi (2mF/󰄁)1/3 (a − x) x∼a (7.31)
󰁁
󰁁 √
󰁁
󰁁
2 3/2
󰀽 √ c e− 3󰄁 2mF (a−x) x<a
p(x)

Let us now consider a more general situation there there is a potential V with turning
points on the left and right. In the middle, a < x < b we need to match the two solutions
we find (this time assuming a general potential):
󰀕 󰁝 󰀖
c 1 x
ψb = 󰁳 sin − p(y)dy + π/4 x<b
p(x) 󰄁 b
󰀕 󰁝 󰀖
c 1 x
ψa = 󰁳 sin − p(y)dy − π/4 x>a (7.32)
p(x) 󰄁 a
These won’t agree unless
󰁝 󰁝
1 x 1 x
− p(y)dy + π/4 = − p(y)dy − π/4 + nπ (7.33)
󰄁 b 󰄁 a
for some integer n. Rearranging this gives the so-called Bohr-Sommerfeld quantization
rule:
󰁝 b
p(y)dy = π(n − 1/2) n∈Z (7.34)
a
where a and b are two turning points. This was more or less guessed by Bohr and
Sommerfeld in the early days of quantum mechanics where they conjectured that n =
1, 2, ....

7.4 Harmonic Oscillator


Having mastered turning points we can now tackle the Harmonic oscillator. We have

V = 12 kx2 so p(x) = 2mE − mkx2 and
󰁝 x󰁳
(0)
σ = 2mE − mky 2 dy
0
√ 󰁝 x󰁳
= 2mE 1 − ky 2 /2Edy
󰀕0 󰀖
√ 1 󰁳 2
1󰁳 󰁳
= 2mE x 1 − kx /2E + 2E/karcsin( k/2Ex) (7.35)
2 2
7.4. HARMONIC OSCILLATOR 75

The associated wavefunctions will not be the ones we found before: i.e. a polynomial
times an exponential suppression. But that’s okay, we don’t expect to land on the exact
answer in any non-trivial case.

Let us see what we can do. At large x we have σ (0) ∼ ±i mk x2 and hence we do
see an exponential suppression in ψ (for the right choice of sign - the must discard the
wrong choice due to normalizability). We can apply the Bohr-Sommerfeld quantization
condition:
󰁝 b 󰁝 √2E/k √
p(x)dx = √ 2mE − mkx2 dx
a − 2E/k
󰁳
= πE m/k
= πE/ω (7.36)

Setting this to π󰄁(n − 1/2) indeed gives us the correct spectrum (n = 1, 2, ...):

E = 󰄁ω(n − 1/2) (7.37)

Although more generally we would only expect agreement at large n.


76 CHAPTER 7. SEMI-CLASSICAL QUANTIZATION: WKB
Chapter 8

Density Matrices, Entanglement


and Thermal States

Let’s get back to the fun stuff. We have seen that, in the Copenhagen Interpretation,
there is an inherent probabilistic interpretation of quantum mechanics. There can also
be probabilistic predictions in classical mechanics. For example in thermal physics or
statistical physics, which describe systems with a large number of degrees of freedom.
Here one doesn’t expect to predict the motion of every particle. Rather one is after
the average but this will introduce probabilities: if one measures a given particle it will
only have the average properties “on average”, any particular measurement could give
a different result.
So are these the same uncertainties? Let us look at an example. Consider a two
state system (using yet another notation):

1
|ψ〉 = √ (| ↑〉 + | ↓〉) (8.1)
2

Let us compute the expectation value of τ3

1
〈ψ|τ3 |ψ〉 = (〈↑ | + 〈↓ |)τ3 (| ↑〉 + | ↓〉)
2
1
= (〈↑ | + 〈↓ |)(| ↑〉 − | ↓〉)
2
1
= (1 − 1)
2
=0 (8.2)

This means that if we make repeated measurements of τ3 in the state |ψ〉 we will find
+1 half the time and −1 the other half leading to an expected outcome of 0.
We can compare this to a more classical picture where we don’t know the underlying
system but we suppose that half the time the system is in the state is | ↑〉 half the time
| ↓〉. Therefore measuring τ3 still has an expectation value of zero.
But these are not the same. To probe the difference lets measure another operator

77
78CHAPTER 8. DENSITY MATRICES, ENTANGLEMENT AND THERMAL STATES

that doesn’t commute with τ3 . So let’s measure τ1 . If we look at the state |ψ〉 we find
1
τ1 |ψ〉 = √ (τ1 | ↑〉 + τ1 | ↓〉)
2
1
= √ (| ↓〉 + | ↑〉)
2
= |ψ〉 (8.3)

so 〈ψ|τ1 |ψ〉 = 1.
On the other hand lets measure τ1 in the classical model where half the time we are
measuring | ↑〉 and half the time | ↓〉. In the first case we have

〈↑ |τ1 | ↑〉 = 〈↑| ↓〉 = 0 (8.4)

and in the second case

〈↓ |τ1 | ↓〉 = 〈↓| ↑〉 = 0 (8.5)

as well. So a measurement of τ1 gives zero. Thus there is a difference!


This is the essence of the difference between classical and quantum probability. In
particular there are general bounds on the outcome of such classical probabilities, where
one assumes that the fundamental constituents of a system have definite values of some
observable but we don’t know what exactly makes up our system, and the quantum
answer where the system is genuinely in a superposition of possible measurement out-
comes. Such bounds are known as Bell’s in equalities and the Nobel prize this year went
to three experimenters who showed, in a variety of experiments, that Bell’s inequalities
are violated - meaning that the classical world is not the one we live in.

8.1 Density Matrices


How can we describe the ”classical” case in quantum theory. To do this we introduce
the notion of a density matrix where we trade the state |ψ〉 for the operator

ρ = |ψ〉〈ψ| (8.6)

Given any observable O we can compute the expectation value of |ψ〉 by computing a
trace:
󰁛
tr(Oρ) = 〈en |Oρ|en 〉
n
󰁛
= 〈en |O|ψ〉〈ψ|en 〉 (8.7)
n

Next we expand
󰁛
|ψ〉 = cm |em 〉 ⇐⇒ cm = 〈em |ψ〉 = (〈ψ|em 〉)∗ (8.8)
m
8.1. DENSITY MATRICES 79

so
󰁛
tr(Oρ) = cm 〈en |O|em 〉c∗n
n,m
󰁛
= cm λm 〈en |em 〉c∗n
n,m
󰁛
= λn |cn |2
n

= 〈ψ|O|ψ〉 (8.9)

where we have used an orthonormal basis |em 〉 of eigenstates of O with eigenvalues λm :


O|em 〉 = λm |em 〉.
Note that the normalization of |ψ〉 translates into the statement that (just take
O = 1)
󰁛
tr(ρ) = 〈en |ρ|en 〉
n
󰁛
= 〈en |ψ〉〈ψ|en 〉
n
󰁛
= cn c∗n
n

=1 (8.10)

Thus we can swap a state for a density matrix. Such a state is called a pure state,
meaning that it is equivalent to a single state in the more traditional formulation.
So what’s the point? We can consider more general density matrices:

ρ = p1 |ψ1 〉〈ψ1 | + p2 |ψ2 〉〈ψ2 | + . . . , tr(ρ) = 1 pi ≥ 0 (8.11)

Note that this is with respect to some basis. If we choose a different basis then ρ may
not take such a diagonal form. The second condition translates into
󰁛
pi = 1 (8.12)

In general these are called mixed states when they can’t be written as

ρ = |ψ〉〈ψ| (8.13)

for a single state |ψ〉. A mixed state allows us to introduce a statistical notion of
uncertainty, in the sense that we don’t know what the quantum state is (but maybe we
could if we did further experiments). For example we can compute

tr(Oρ) = p1 tr(O|ψ1 〉〈ψ1 |) + p2 tr(O|ψ2 〉〈ψ2 |) + ...


= p1 〈ψ1 |O|ψ1 〉 + p2 〈ψ2 |O|ψ2 〉 + . . . (8.14)

The expectation value then has the interpretation as a classical statistic average over
the individual quantum expectation values.
80CHAPTER 8. DENSITY MATRICES, ENTANGLEMENT AND THERMAL STATES

If we go back to the simple system we started talking about in the first, quantum
superposition, experiment we have

ρ1 = |ψ〉〈ψ|
1
= (| ↑〉 + | ↓〉)(〈↑ | + 〈↓ |)
2
1
= (| ↑〉〈↑ | + | ↑〉〈↓ | + | ↓〉〈↑ | + | ↓〉〈↓ |) (8.15)
2

whereas in the second, more classical, experiment we simply take

1
ρ2 = (| ↑〉〈↑ | + | ↓〉〈↓ |) (8.16)
2

Let us compute

tr(τ3 ρ1 ) = 〈↑ |τ3 ρ1 | ↑〉 + 〈↓ |τ3 ρ1 | ↓〉


1 󰀓 󰀔
= 〈↑ | τ3 | ↑〉〈↑ | + τ3 | ↑〉〈↓ | + τ3 | ↓〉〈↑ | + τ3 | ↓〉〈↓ | | ↑〉
2
1 󰀓 󰀔
+ 〈↓ | τ3 | ↑〉〈↑ | + τ3 | ↑〉〈↓ | + τ3 | ↓〉〈↑ | + τ3 | ↓〉〈↓ | | ↓〉
2
1 󰀓 󰀔 1 󰀓 󰀔
= 〈↑ | τ3 | ↑〉 + τ3 | ↓〉 + 〈↓ | τ3 | ↑〉 + τ3 | ↓〉
2 2
1 1
= − =0
2 2
tr(τ1 ρ1 ) = 〈↑ |τ1 ρ1 | ↑〉 + 〈↓ |τ1 ρ1 | ↓〉
1 󰀓 󰀔
= 〈↑ | τ1 | ↑〉〈↑ | + τ1 | ↑〉〈↓ | + τ1 | ↓〉〈↑ | + τ1 | ↓〉〈↓ | | ↑〉
2
1 󰀓 󰀔
+ 〈↓ | τ1 | ↑〉〈↑ | + τ1 | ↑〉〈↓ | + τ1 | ↓〉〈↑ | + τ1 | ↓〉〈↓ | | ↓〉
2
1 󰀓 󰀔 1 󰀓 󰀔
= 〈↑ | τ1 | ↑〉 + τ1 | ↓〉 + 〈↓ | τ1 | ↑〉 + τ1 | ↓〉
2 2
1 1
= + =1 (8.17)
2 2
8.1. DENSITY MATRICES 81

which agrees with what we saw before. On the other hand

tr(τ3 ρ2 ) = 〈↑ |τ3 ρ2 | ↑〉 + 〈↓ |τ3 ρ2 | ↓〉


1 󰀓 󰀔
= 〈↑ | τ3 | ↑〉〈↑ | + τ3 | ↓〉〈↓ | | ↑〉
2
1 󰀓 󰀔
+ 〈↓ | τ3 | ↑〉〈↑ | + τ3 | ↓〉〈↓ | | ↓〉
2
1 1
= 〈↑ |τ3 | ↑〉 + 〈↓ |τ3 | ↓〉
2 2
1 1
= − =0
2 2
tr(τ1 ρ2 ) = 〈↑ |τ1 ρ2 | ↑〉 + 〈↓ |τ1 ρ2 | ↓〉
1 󰀓 󰀔
= 〈↑ | τ1 | ↑〉〈↑ | + τ1 | ↓〉〈↓ | | ↑〉
2
1 󰀓 󰀔
+ 〈↓ | τ1 | ↑〉〈↑ | + τ1 | ↓〉〈↓ | | ↓〉
2
1 1
= 〈↑ |τ1 | ↑〉 + 〈↓ |τ1 | ↓〉
2 2
=0+0=0 (8.18)

which agrees with what we saw before but is different from ρ1 . Note that although the
density matrix ρ2 may look simpler than ρ1 it is a mixed state whereas ρ1 is pure.
How can we tell in general whether or not a density matrix corresponds to a pure or
mixed state? Well a pure state means ρ = |ψ〉〈ψ| for some state |ψ〉 and hence

ρ2 = |ψ〉〈ψ|ψ〉〈ψ| = |ψ〉〈ψ| = ρ (8.19)

So in particular tr(ρ2 ) = tr(ρ) = 1. But for a mixed state (we assume |ψn 〉 form an
orthonormal basis)
󰁛
ρ= pn |ψn 〉〈ψn |
n
󰁛
2
ρ = pn pm |ψn 〉〈ψn |ψm 〉〈ψm |
n,m
󰁛
= p2n |ψn 〉〈ψn | (8.20)
n

from which it follows that


󰁛 󰁛
tr(ρ2 ) = p2n ≤ pn = 1 (8.21)
n n

because 0 ≤ pn ≤ 1 we have p2n ≤ pn for each pn . Furthermore we can only have


tr(ρ2 ) = 1 if one, and therefore only one, pn = 1 with all others vanishing. Thus we
have a theorem:

tr(ρ2 ) ≤ 1 (8.22)

with equality iff ρ represents a pure state. For example in our classical density matrix
above we have p1 = p2 = 1/2 and so tr(ρ) = 1 as required but tr(ρ2 ) = 1/4 + 1/4 =
1/2 < 1 which tells us it is indeed a mixed state.
82CHAPTER 8. DENSITY MATRICES, ENTANGLEMENT AND THERMAL STATES

8.2 Thermal States


The most classic example of a mixed state is what we find at finite temperature. Here
we imagine that there are many particles with a variety of energies and we take

1 󰁛 −En /kB T
ρthermal = e |En 〉〈En | (8.23)
Z n

where |En 〉 are the energy eigenstates, T is the temperature and kB is Boltzman’s
constant, kB = 1.380649 × 10−23 J/o K that converts temperature into energy.1 Clearly
this is a mixed state and is known as the Boltzman distribution.
To determine the normalization Z we need to impose

1 = tr(ρ)
1 󰁛 −En /kB T
= e 〈Em |En 〉〈En |Em 〉
Z n,m
1 󰁛 −En /kB T
= e (8.24)
Z n

Thus
󰁛
Z= e−En /kB T (8.25)
n

This known as the Partition function and plays a very central role. It counts the number
of states available at each energy En :
󰁛
Z= d(n)e−βEn (8.26)
En

where d(n) counts the degeneracy of states at energy level En and and β = 1/kB T is
the inverse temperature.
At low temperatures, where T → 0, the density matrix will strongly peak around
the lowest energy state:

lim ρthermal = |E0 〉〈E0 | (8.27)


T →0

and hence becomes pure. However at high energy, T → ∞, all the energies states
contribute more or less equally and so

1 󰁛
lim ρthermal = |En 〉〈En | (8.28)
T →∞ dimH n

1
This value of kB is exact. It’s a definition.
8.2. THERMAL STATES 83

We can ask what is the expected energy of a thermal state:


󰁛
tr(Hρthermal ) = 〈Em |Hρthermal |Em 〉
m
1 󰁛 −En /kB T
= e 〈Em |H|En 〉〈En |Em 〉
Z n,m
1󰁛
= En e−En /kB T
Z n
1󰁛
= En e−βEn
Z n
1 ∂ 󰁛 −βEn
=− e
Z ∂β n

=− ln Z (8.29)
∂β
These formulae may (or may not) be familiar from statistical physics.
For example we can consider our friend the Harmonic oscillator. Here we have a
unique state at each level n with En = 󰄁ω(n + 1/2). Thus

󰁛 q 1/2
Z= q n+1/2 = (8.30)
n=0
1−q

where q = e−󰄁ωβ . Furthermore


1 󰁛 n+1/2
ρthermal = q |n〉〈n|
Z n
󰁛
= (1 − q) q n |n〉〈n| (8.31)
n

We can compute the expected energy to find


󰀕 1/2 󰀖
∂ q
tr(Hρthermal ) = − ln
∂β 1−q
󰀕 󰀖
1 1
=− + (−󰄁ω)q
2q 1 − q
1 1+q
= 󰄁ω (8.32)
2 1−q
We can re-write this as
1 q
tr(Hρthermal ) = 󰄁ω + 󰄁ω (8.33)
2 1−q
The first term is just the ground state energy and so we can subtract it off since it
doesn’t have a physical meaning to find
1
Ethermal = 󰄁ω
q −1−1
1
= 󰄁ω (8.34)
e󰄁ω/kB T − 1
84CHAPTER 8. DENSITY MATRICES, ENTANGLEMENT AND THERMAL STATES

Figure 8.2.1: The Planck Curve: N.B. The horizontal axis is wavelength λ ∼ 1/ω

This is the famous Planck formula for the spectrum of emitted light from a so-called
black body of temperature T , albeit in just one dimension. In three dimensions there
is an 4πω 3 in the numerator. The extra 4πω 2 comes from counting modes in three
dimensions; basically it arises from the growth of the size of a sphere as ω = |k| where k
is the spatial momentum. A plot of the three-dimensional curve is given in Fig. 8.2.1. In
particular it was observed that all bodies radiate with a certain spectrum that depends
on their temperature. Humans (living ones) radiate in the infra-red. If it is hotter you
can see it as visible light such as an oven hob (not an induction one!). Even the empty
space of the Universe emits radiation in the form of the cosmic microwave background
(CMB) corresponding to a temperature of about −270o C. Planck’s formula Ethermal
matches experiment but deriving it was one of those problems the eighteenth century
physicists thought was just a loose string. Rather Planck had to introduce the notion
of discrete energies and yes 󰄁 to derive it. And that was the beginning of the end for
Classical Physics.

8.3 Entropy
To continue we need to talk about the idea of entropy. Entropy is a measure of disorder.
Or more precisely it measures how many microscopic states lead to the same macroscopic
one. All your various charger cables are always tangled up as there are so many more
tangled possibilities than untangled ones. It’s not that life is against you its just that
being organised is an exponentially unlikely state of affairs: organisation requires some
8.4. ENTANGLEMENT 85

organiser and quite some effort2 . The most basic definition of entropy is it is the log of
the number of states with the same energy. So

S(En ) = ln d(n) (8.35)

where d(n) is given in 8.26.


But there are a variety of notions of entropy3 . To measure the entropy of a mixed
state we consider the von Neumman entropy:

SvN = −tr(ρ ln ρ) (8.36)

This requires a little motivation. Firstly to compute the Log of an operator the easiest
way is to diagonalise it so:
󰁛
SvN = − pn ln pn (8.37)
n

Thus the minus is needed as 0 ≤ pn ≤ 1 to ensure that SvN ≥ 0. We see that pn ln pn = 0


if pn = 1 (corresponding to a pure state) and also pn ln pn = 0 if pn = 0 (meaning that
states that don’t appear in ρ don’t contribute to the entropy). It follows that only mixed
states have SvN > 0. So this is another way to characterise mixed states.
In our simple mixed state ρ2 we have
1 1 1 1
SvN = − ln − ln
2 2 2 2
= ln 2 (8.38)

In fact there is a theorem that SvN ≤ ln(dim H). A mixed state with SvN = ln(dim H)
is said to be maximally mixed. For example a maximally mixed state arises when
pn = 1/N for all n, where N = dim H. Such a mixed state is equally mixed between all
pure states.

8.4 Entanglement
Perhaps the most mystical and celebrated feature of Quantum Mechanics is the existence
of entangled states. What do we mean precisely? First let’s look at the quintessential
example of two particles in the |0, 0〉 state that we saw in section 4.2:

1 󰀃 󰀄
|0, 0〉 = √ | 12 , 12 〉 ⊗ | 12 , − 12 〉 − | 12 , − 12 〉 ⊗ | 12 , 12 〉 (8.39)
2

This is a state that consists of two particles each of which carries ±󰄁/2 units of angular
momentum around the z-axis in such a way that the total angular momentum vanishes.
2
Of course life could still be against you; even though everyones charger cables get tangled, yours
are more tangled than others and always at the worst time.
3
One could say that the entropy of entropy definitions is non-zero.
86CHAPTER 8. DENSITY MATRICES, ENTANGLEMENT AND THERMAL STATES

But we don’t know if the first one is spinning clockwise and the second anti-clockwise
or vice-versa.
Now these two particles might be very far from each other. Perhaps they were created
in some experiment as a pair and then flew away across the universe, like endless rain
into a paper cup. Suppose sometime later we measure one of the particles, say the
first, and it is in the state | 12 , 12 〉. Then we know, we absolute certainty, that the second
particle, wherever it is in the Universe, is in the | 12 , − 12 〉 state. This is very counter
intuitive but it doesn’t contradict anything we know about the world (and indeed has
been verified experimentally). For example we can’t use it to send messages faster than
light as we can’t control how the first particle’s wavefunction ‘collapses’.
So how can we understand this more generally? The basic idea is that the total
Hilbert space of the system can be viewed as a direct product of two Hilbert spaces:

H = H1 ⊗ H2 (8.40)

Of course it could be a direct product of many separate Hilbert spaces but this is enough
for us. For example in the example above H1,2 are the Hilbert spaces of the individual
particles. Formally all these means is that a general state of the system can be written
as
󰁛
|Ψ〉 = cn′ n′′ |en′ 〉 ⊗ |en′′ 〉 (8.41)
n′ ,n′′

with |en′ 〉 a basis of H1 and |en′′ 〉 a basis of H2 . Note that pure state takes the form

ρ = |Ψ〉〈Ψ|
󰁛
= cn′ n′′ c∗m′ m′′ |en′ 〉 ⊗ |en′′ 〉〈em′ | ⊗ 〈em′′ |
n′ ,n′′ ,m′ ,m′′
󰁛
= cn′ n′′ c∗m′ m′′ |en′ 〉〈em′ | ⊗ |en′′ 〉〈em′′ | (8.42)
n′ ,n′′ ,m′ ,m′′

Next we want to we admit that we have no idea what is going on in the second Hilbert
space H2 . Thus we would construct a density matrix where we sum (trace) over all the
options for H2 :

ρreduced = trH2 (|Ψ〉〈Ψ|)


󰀣 󰀤
󰁛 󰁛
= 〈ep′′ | cn′ n′′ c∗m′ m′′ |en′ 〉〈em′ | ⊗ |en′′ 〉〈em′′ | |ep′′ 〉
p′′ n′ ,n′′ ,m′ ,m′′
󰀣 󰀤
󰁛 󰁛
= cn′ n′′ c∗m′ m′′ |en′ 〉〈em′ |
n′ ,m′ n′′ ,m′′
󰁛
≡ an′ m′ |en′ 〉〈em′ | (8.43)
n′ ,m′
8.4. ENTANGLEMENT 87

Note that an′ m′ defines a self-adjoint operator on H1 as


󰁛
a∗m′ n′ = c∗m′ n′′ cn′ m′′
n′′ ,m′′
󰁛
= cn′ n′′ c∗m′ m′′
m′′ ,n′′

= an′ m′ (8.44)

So we can find a new basis of H1 given by |e′n′ 〉 such that


󰁛
ρreduced = pn′ |e′n′ 〉〈e′n′ | (8.45)
n′

Furthermore the pn′ , which are the eigenvalues of an′ m′ , are real and in fact positive
(because, roughly, a ∼ cc† is positive definite) and less than one (because |Ψ〉 is nor-
malised).
Thus we find a density matrix, called the reduced density matrix, in the first Hilbert
space. Even though we started from a pure state, from the point of view of the first
Hilbert ρreduced is a mixed state in general. Furthermore we can evaluate it’s von-
Neumann entropy:

See = −trH1 (ρreduced ln ρreduced ) (8.46)

which is generally referred to as the entanglement entropy. Note that there is a theorem
that, had we traced over H1 and computed the entanglement entropy of the resulting
density matrix in H2 then we would find the same value for SEE . In other words the
state is equally entangled between the two sub-Hilbert spaces.
So lets go back to our two particle system and first consider the pure state

|Ψ〉 = | ↑〉 ⊗ | ↓〉 (8.47)

This leads to the density matrix

ρ = | ↑〉〈↑ | ⊗ | ↓〉〈↓ | (8.48)

If we trace over the second Hilbert space we find

ρreduced = | ↑〉〈↑ | ⊗ (〈↑| ↓〉〈↓| ↑〉 + 〈↓| ↓〉〈↓| ↓〉)


= | ↑〉〈↑ | (8.49)

which is pure and indeed See = 0. The reason is that the unknown part of the state aris-
ing from H2 is just one state. So averaging over it doesn’t really loose any information.
However suppose we start with
1
|Ψ〉 = √ (| ↑〉 ⊗ | ↓〉 + | ↓〉 ⊗ | ↑〉) (8.50)
2
88CHAPTER 8. DENSITY MATRICES, ENTANGLEMENT AND THERMAL STATES

The density matrix is therefore


1󰀓 󰀔
ρ= | ↑〉〈↑ | ⊗ | ↓〉〈↓ | + | ↑〉〈↓ | ⊗ | ↓〉〈↑ | + | ↓〉〈↑ | ⊗ | ↑〉〈↓ | + | ↓〉〈↓ | ⊗ | ↑〉〈↑ |
2
(8.51)

then tracing over the second Hilbert space the middle two terms drop out and we find
1󰀓 󰀔
ρreduced = | ↑〉〈↑ | + | ↓〉〈↓ | (8.52)
2
which agrees with ρ2 that we had above. We also see that
1 1 1 1
See = − ln − ln = ln 2 (8.53)
2 2 2 2
as before. The lesson is that the ‘Classical’ density matrix we considered above arises
in a quantum theory by tracing over hidden degrees of freedom. In other words the
quantum description is more refined, i.e. contains more information.
Chapter 9

Relativistic Quantum Mechanics

Let us consider the Schrödinger equation for a collection of N particles with positions
xa , a = 1, ..., N :

∂Ψ 󰄁2 2
i󰄁 =− ∇ Ψ + V (xa )Ψ (9.1)
∂t 2ma a
where Ψ = Ψ(t, x1 , ..., xN ). We don’t expect there to be a preferred point or direction
in space. So let us suppose that the potential is of the form

V (x1 , ..., xN ) = V (|xa − xb |) (9.2)

i.e. we assume that the potential only depends on distance between positions of the
particles. Schrödinger’s equation is then rotationally invariant:

xa → x′a = Rxa (9.3)

where R−1 = RT is an orthogonal matrix. It also has a notion of Galilean boosts:

xa → x′a = xa + vt (9.4)

Clearly V is invariant as xa − xb = x′a − x′b . And the Laplacian terms don’t change.
However if we write Ψ′ (t, xa ) = Ψ(t, x′a )
∂Ψ′ ∂Ψ 󰁛
= +v· ∇a Ψ (9.5)
∂t ∂t a

Thus we find an extra term on the left hand side and the Schrödinger equation is no
longer invariant. However this can be compensated for by taking instead
󰁓 󰀓 1
󰀔
′ − 󰄁i v· ma xa + ma vt
Ψ (t, xa ) = e a 2 Ψ(t, x′a ) (9.6)

i.e. we include an extra phase factor in the wavefunction. As a result we find (see the
problem set)

∂Ψ′ 󰄁2 ′ 2 ′
i󰄁 =− ∇ Ψ + V (x′a )Ψ′ (9.7)
∂t 2ma a

89
90 CHAPTER 9. RELATIVISTIC QUANTUM MECHANICS

Thus we have a symmetry under what are known as Galilean boosts, corresponding
to a notion of Galilean relativity. In particular in Galilean relativity time is absolute,
eveyone agrees on t and an infinitesimal tick of the clock dt. So also do they all agree
on spatial lengths: dx2 + dy 2 + dz 2 . Thus observers may move in reference frames which
are related by rotations and boosts of the form (9.4).
Thats great but as Einstein showed in 1905 Galilean boosts are not symmetries of
space and time. Rather Lorentzian boosts are and we need to consider Special Relativity!
In Galilean relativity the speed of light is not a constant because under a boost velocities
are shifted:

ẋ → ẋ + v (9.8)

But Einstein’s great insight (well one of them) was to insist that the speed of light is a
universal constant. I won’t go into why except to mention that Maxwell’s equations for
electromagnetism don’t allow you to go to a frame where c = 0. They are not invariant
under the Galilean boosts we saw above. They are instead invariant under Lorentz
transformations which include rotations and Lorentz boosts which we now describe.
So what is a Lorentz boost. In Special Relativity we have that the invariant notion
of length is, for an infinitesimal displacement,1

ds2 = c2 dt2 − dx2 − dy 2 − dz 2 (9.9)

Here c is the speed of light. Here we only require that ds2 is the same for all observers
(not dt and dx2 + dy 2 + dz 2 ). Thus we are allowed to consider a transformation of the
form

ct′ = γct + γβ · x x′ = γx + γβct


cdt′ = γcdt + γβ · dx dx′ = γdx + γβcdt (9.10)

However we require that

ds′2 = c2 dt′2 − dx′ · dx′


= γ 2 (cdt + β · dx)2 − γ 2 (dx + βcdt) · (dx + βcdt)
= γ 2 (1 − |β|2 )c2 dt2 − γ 2 (1 − |β|2 )dx · dx (9.11)

we need ds′2 = ds2 and hence (we could also take the other sign but that would flip the
direction of time)
1
γ=󰁴 (9.12)
1 − |β|2

What is β? Well consider a small β then to first order we have γ = 1 and

t′ = t + c−1 β · x x′ = x + βct (9.13)


1
A more common convention is ds2 = −c2 dt2 + dx2 + dy 2 + dz 2 but either will do - thats not true:
no sane person likes the convention we are using.
9.1. RELATIVISTIC WAVE EQUATIONS 91

Thus to recover Galilean boost x′ = x + vt we identify β = v/c. Small β means |v| << c.
Then we see that

t′ = t + c−2 v · x = t + O(|v|/c2 ) (9.14)

and we have recovered absolute time. These transformations are called Lorentzian boosts
and together with spatial rotations they for the Lorentz group.

9.1 Relativistic Wave Equations


Let’s get back to Quantum Mechanics where time plays a central role. If time is not
absolute then what happens to Quantum Mechanics? How does this fit with the fact
that the Schrödinger equation is first order in time derivative and second order in spatial
derivatives. In Special relativity time and space are interchangeble.
The first attempt to find a relativistic wave equations is to simply make time second
order and consider
󰄁2 ∂ 2 Ψ
− 󰄁2 ∇2 Ψ + m 2 c 2 Ψ = 0 (9.15)
c2 ∂t2
This is invariant under Special relativity if Ψ(x) → Ψ′ (x) = Ψ(x′ ) (see the problems).
It is known as the Klein-Gordon equation. But there are big problems:

1. If we look at energy and momentum eigenstates Ψ = e−iEt/󰄁+ip·x/󰄁 we find

E 2 = |p|2 c2 + m2 c4

(setting p = 0 immediately gives a famous formula). But this allows for both
positive and negative energy.

2. We would like to interpret Ψ∗ Ψ as a probability density but it is not conserved:


󰁝
d
Ψ∗ Ψd3 x ∕= 0
dt
One can find a conserved density: ρ = iΨ∗ ∂t Ψ − iΨ∂t Ψ∗ but this is not positive
definite (see the problems).

3. Knowing the state at time t = 0 is not enough information to determine it at later


times, one also needs ∂Ψ/∂t at t = 0 but what does that mean?

The other way to go, and one of those oh so clever moments of discovery, is to find an
equation that it first order in time and space. This is what Dirac did and it’s beautiful.
So let us try something like
󰀕 󰀖
i󰄁 0 ∂
γ − i󰄁γ · ∇ + mc Ψ = 0 (9.16)
c ∂t
92 CHAPTER 9. RELATIVISTIC QUANTUM MECHANICS

We will not be very specific yet about what γ 0 and γ i are. We want this equation to
imply the Klein-Gordon equation so we square it (and change the sign of i for good
measure):
󰀕 󰀖󰀕 󰀖
i󰄁 0 ∂ i󰄁 0 ∂
− γ + i󰄁γ · ∇ + mc γ − i󰄁γ · ∇ + mc Ψ = 0
c ∂t c ∂t
󰀕 2 󰀖
󰄁 0 2 ∂2 󰄁2 i j ∂2 󰄁2 0 i ∂ 2 2 2
(γ ) 2 + {γ , γ } i j − {γ , γ } +m c Ψ=0 (9.17)
c2 ∂t 2 ∂x ∂x c ∂t∂xi

Where {γ i , γ j } = γ i γ j + γ j γ i is known as the anti-commutator. Thus we will recover


the Klein-Gordon equation if

(γ 0 )2 = 1 {γ 0 , γ i } = 0 {γ i , γ j } = −2δ ij (9.18)

Clearly this can’t happen if the γ’s are just numbers. But maybe if they are Matrices?
Some trial and error gives the following (unique up to conjugation γ → U γU −1 ) solution:
󰀣 󰀤 󰀣 󰀤
0 I 0 τ i
γ0 = γi = (9.19)
I 0 −τi 0

where I is the 2 × 2 unit matrix and τi the Pauli matrices. Thus Ψ must be a complex
four-vector:
󰀳 󰀴
ψ1
󰁅ψ 󰁆
󰁅 2󰁆
Ψ=󰁅 󰁆 . (9.20)
󰁃ψ 3 󰁄
ψ4

And yet again we see the Pauli matrices.

9.2 Relativistic Notation


It’s time to adopt a better notation. We introduce the new coordinate xµ , µ = 0, 1, 2, 3
with

x0 = ct , x1 = x , x2 = y , x3 = z (9.21)

Note that the invariant length we introduced above is

ds2 = ηµν dxµ dxν (9.22)

where
󰀳 󰀴
1 0 0 0
󰁅0 −1 0 0󰁆
󰁅 󰁆
ηµν =󰁅 󰁆 (9.23)
󰁃0 0 −1 0 󰁄
0 0 0 −1
9.3. SPINORS 93

and repeated indices are summed over. Note that in such sums one index is always up
and one down. This ensures that any quantity that has no left-over indices is a Lorentz
invariant. It is also helpful to introduce the matrix inverse to ηµν which is denoted by
󰀳 󰀴
1 0 0 0
󰁅0 −1 0 0󰁆
−1 󰁅 󰁆
η µν = ηµν =󰁅 󰁆 (9.24)
󰁃0 0 −1 0 󰁄
0 0 0 −1

Note that numerically ηµν = η µν but this is really a coincidence.


Lastly we have γ µ = (γ 0 , γ 1 , γ 2 , γ 3 ) above. Then the Dirac equation is simply

−i󰄁γ µ ∂µ Ψ + mcΨ = 0 (9.25)

Beautiful.

9.3 Spinors
What about Ψ? It now has four components so we could introduce an index Ψα , α =
1, 2, 3, 4 in which case the γ-matrices also pick-up indices:

γ µ = (γ µ )α β (9.26)

The condition we found in (9.20) is

{γ µ , γ ν } = 2η µν I ⇐⇒ {γ µ , γ ν }α β = 2η µν δα β (9.27)

This is called a Clifford algebra2 . Ψ is called a spinor and the α, β indices spinor
indices. This is a new kind of object for us, so lets see the consequences.
Under a Lorentz transformation:

x′µ = Λµ ν xν (9.28)

we require that ds′2 = ds2 and hence

Λµ ρ Λν σ ηµν = ηρσ (9.29)

The coordinates xµ provide a representation of the Lorentz Group, which is sometimes


called SO(1, 3):

SO(1, 3) = {Λ|η = ΛT ηΛ, det(Λ) = 1} (9.30)

It is the spacetime version of the Euclidean rotation group in four-dimensions: SO(4)


Note that η = ΛT ηΛ is just the index-free, matrix version of (9.29). Here we suppressed
all the vector µ, ν indices and think of xµ as a vector in R4 and η, Λ are 4 × 4 real
matrices.
2
Clifford was a student at King’s in the 1800’s.
94 CHAPTER 9. RELATIVISTIC QUANTUM MECHANICS

Let us look at the infinitesimal form, i.e. the Lie-algebra:

Λµ ν = δνµ + ω µ ν ω µ ν << 1 (9.31)

then (9.29) becomes (to first order)

ω µ ρ ηµσ + ω ν σ ηρν = 0 (9.32)

In Relativity we adopt a notation where contracting an index with ηµν lowers it: ω µ ρ ηµσ =
ωρσ . Similarly one can raise an index by contracting with η µν . Thus the Lie-algebra
consists of anti-symmetric ωρσ :

ωσρ + ωρσ = 0 . (9.33)

A basis for so(1, 3) is therefore provided by the matrices

(Mµν )ρ σ = δµρ ηνσ − δνρ ηµσ (9.34)

These are just 4 × 4 matrices with a single ±1 somewhere in the upper triangle and a
∓1 in the corresponding lower triangle to make it anti-symmetric. For example:
󰀳 󰀴
0 0 0 0
󰁅0 0 −1 0󰁆
󰁅 󰁆
M12 = δ1ρ η2σ − δ2ρ η1σ =󰁅 󰁆 (9.35)
󰁃0 1 0 0󰁄
0 0 0 0

One finds the Lie-algebra satisfies (see the problem set)

[Mµν , Mλρ ] = Mµρ ηνλ − Mνρ ηµλ + Mνλ ηµρ − Mµλ ηνρ (9.36)

What about Ψ? Well it turns out that


1 1
Σµν = γµν = (γµ γν − γν γµ ) (9.37)
2 4
also forms a representation of so(1, 3), i.e. (9.36) (see the problem set). And it’s also
four-dimensional, because the spinor indices have the range α, β = 1, 2, 3, 4, but this time
four-complex dimensions and acts on the Ψ. This is called the spinor representation of
so(1, 3).3
As an aside we note that one can also consider spinors of so(3). These simply
correspond to the l = 1/2 representation of su(2) that we saw above and consist of
a complex 2-vector. Whereas the vector representation, as given by xi corresponds to
l = 1.
3
Things are a bit confusing in four-dimensions as many different indices are four-dimensional: xµ
and Ψα . This is another coincidence. For example in ten dimensions xµ would have µ = 0, 1, 2...., 9 but
α = 1, 2, 3, ...32.
9.3. SPINORS 95

Thus under an infinitesimal Lorentz transformation we have

Ψ(x) → Ψ′ (x) = SΨ(Λx)


󰀕 󰀖
1 µν
= 1 + ω Σµν Ψ(Λx) (9.38)
2

The extra factor of 1/2 comes from the over counting due to the anti-symmetry. To find
a finite transformation we exponentiate:

1 µν Σ
S = e2ω µν
(9.39)

Let us check that this is a symmetry of the Dirac equation:

−i󰄁γ µ ∂µ Ψ′ + mcΨ′ = −i󰄁γ µ ∂µ (SΨ(Λx)) + mcSΨ(Λx)


= −i󰄁γ µ SΛν µ ∂ν′ Ψ(Λx) + mcSΨ(Λx)
󰀃 󰀄
= S −i󰄁S −1 Λν µ γ µ S∂ν′ Ψ(x′ ) + mcΨ(x′ ) (9.40)

where we have used

∂x′ν ∂Ψ
∂µ Ψ = = Λν µ ∂ν′ Ψ (9.41)
∂xµ ∂x′ν

We want to show
󰀃 󰀄
−i󰄁γ µ ∂µ Ψ′ + mcΨ′ = S −i󰄁γ µ ∂µ′ Ψ(x′ ) + mcΨ(x′ ) = 0 (9.42)

Thus we need that, infinitesimally,

S −1 Λµ ν γ ν S = γ µ (9.43)

This becomes

1 1
− ω ρσ Σρσ γ µ + ω µ ν γ ν + γ µ ω ρσ Σρσ = 0
󰀕 2 2 󰀖
ρσ 1 µ 1 µ 1 µ 1 µ
ω − γρσ γ + δρ γσ − δσ γρ + γ γρσ = 0 (9.44)
4 2 2 4

which boils down to the identity

γρσ γ µ − γ µ γρσ = 2δσµ γρ − 2δρµ γσ (9.45)

This is in fact true as you can check. But for now just try some cases: if µ, ρ, σ are all
distinct then γ µ commutes with γρσ and the left hand side vanishes. So does the right
hand side. If µ = σ and ρ ∕= σ then the two terms on the left hand side will add to give
a factor 2 which agrees with the right hand side.
96 CHAPTER 9. RELATIVISTIC QUANTUM MECHANICS

As an example let’s consider a rotation in the x − y plane by an angle θ: ω12 = θ.


On the coordinates xµ we have
1 λρ M
x′µ = (e 2 ω λρ
) µ ν xν
= (eθM12 )µ ν xν
= (cos θI + M12 sin θ)µ ν xν
󰀳 ′󰀴 󰀳 󰀴󰀳 󰀴
ct 1 0 0 0 ct
󰁅 x′ 󰁆 󰁅0 cos θ − sin θ 0󰁆 󰁅 x 󰁆
󰁅 󰁆 󰁅 󰁆󰁅 󰁆
=⇒ 󰁅 ′ 󰁆 = 󰁅 󰁆󰁅 󰁆 (9.46)
󰁃 y 󰁄 󰁃0 sin θ cos θ 0󰁄 󰁃 y 󰁄
z′ 0 0 0 1 z

But what happens to a spinor?


1 µν Σ
Ψ′α = (e 2 ω µν
) α β Ψβ
1
= (e 2 θγ12 )α β Ψβ
= (cos(θ/2)I + γ12 sin(θ/2))α β Ψβ (9.47)

In particular for θ = 2π we find

Ψ′ = −Ψ . (9.48)

Thus under a rotation of 2π spinors come back with a minus sign. This is very important.
It turns out that this means particles described by spinors must have wavefunctions
which change sign under swapping any two particles. Such particles are call Fermions
(particles that don’t have such a sign, such as photons, are called Bosons). This in
turn implies the Pauli exclusion principle: no to spinor particles can be in the same
state. This is what makes matter hard: when you add electrons to a Hydrogen atom to
build heavier elements they must fill out different energy levels.
How can we construct a Lorentz scalar from a spinor? For a vector we form the
contraction: e.g. ds2 = ηµν dxµ dxν = dxµ dxµ . For a spinor we need to consider something
like

Ψ†1 CΨ2 (9.49)

for some spinor matrix C. Now under a Lorentz transformation δΨ1,2 = 14 ωµν γ µν Ψ1,2
and hence
1 1
δΨ†1 = Ψ†1 ωµν (γ µν )† = Ψ†1 ωµν (γ ν )† (γ µ )† (9.50)
4 4
Now it is easy to check that

(γ µ )† = γ 0 γ µ γ 0
=⇒ (γ ν )† (γ µ )† = γ 0 γ ν γ µ γ 0 (9.51)
9.4. BACK TO QUANTUM MECHANICS 97

thus
1 1
δΨ†1 = Ψ†1 ωµν γ 0 γ νµ γ 0 = − Ψ†1 ωµν γ 0 γ µν γ 0 (9.52)
4 4
What we need is

0 = δ(Ψ†1 CΨ2 )
= δΨ†1 CΨ2 + Ψ†1 CδΨ2
1
= Ψ†1 ωµν (−γ 0 γ µν γ 0 C + Cγ µν )Ψ2 (9.53)
4
Thus we can simply take C = γ 0 . This defines the Dirac conjugate:

Ψ̄ = Ψ† γ 0 (9.54)

and hence Ψ̄1 Ψ2 is a Lorentz scalar. Thus C = γ 0 plays same role for spinor indices
that η µν plays for spacetime indices. If we put indices on it we have C αβ = (γ 0 )α β .
A final comment is that Ψ is called a Dirac spinor. This is not an irreducible
representation of so(1, 3). One can impose the constraint

γ5 Ψ = γ 0 γ 1 γ 2 γ 2 Ψ = ±Ψ (9.55)

The eigenstates of γ5 are known as Weyl spinors and are often called left or right handed
spinors. Given the form of the γ-matrices above we have
󰀣 󰀤
I 0
γ5 = (9.56)
0 −I

and hence we can write


󰀣 󰀤
ΨL
Ψ= (9.57)
ΨR

where ΨL/R are the left and right handed spinors. These form irreducible representations
of so(1, 3) and each is a complex 2-vector in spinor space.

9.4 Back to Quantum Mechanics


Okay so now we have a wave equation that is relativistic and first order in time. In
particular we can re-write the Dirac equation in a form more compatible with quantum
mechanics:
∂Ψ
i󰄁 = HΨ H = −i󰄁cγ 0 γ · ∇ + mc2 γ 0 (9.58)
∂t
This H is self-adjoint but not positive definite. Indeed one can just look at constant Ψ
and observe that γ 0 has eigenvalues +1 and −1.
98 CHAPTER 9. RELATIVISTIC QUANTUM MECHANICS

So we still have negative energy states! But we do have a positive definite density.
Indeed we can identify a 4-current

J µ = Ψ̄γ µ Ψ (9.59)

It is easy to see that this is conserved:

∂µ J µ = ∂µ Ψ̄γ µ Ψ + Ψ̄γ µ ∂µ Ψ
mc mc
= i Ψ̄Ψ − i Ψ̄Ψ
󰄁 󰄁
=0 (9.60)

Here we have used the equation of motion

i󰄁∂µ Ψ̄γ µ + mcΨ̄ = 0 (9.61)

which you can show in the problem set. Along with showing Ψ̄γ µ Ψ a Lorentz vector.
This means that the time component can be used as a conserved quantity:
󰁝 󰁝
1d
J d x = ∂0 J 0 d3 x
0 3
c dt
󰁝
= − ∂i J i d3 x

=0 (9.62)

And furthermore

J 0 = Ψ̄γ 0 Ψ = Ψ† Ψ (9.63)

is positive definite.
So we’ve discovered something beautiful and new but haven’t solved all our problems.
We still have negative energy states and not all particles are described by the Dirac
equation: just Fermions such as electrons and quarks. Dirac’s solution to the negative
energy states was to use the Pauli exclusion principle to argue that the ground state
has all negative energy states filled. This however implies that you should be able to
knock a state with negative energy out of the vacuum. Such an excitation has the same
energy (mass) but opposite charge to a positive energy state. In so doing he predicted
anti-particles which were subsequently observed. We now know that the Higgs’ Boson
(as with other Bosons) satisfies a Klein-Gordon-like equation and that these too have
anti-particles. The full solution comes by allowing the number of particles to change,
i.e. one allows for the creation and annihilation of particles, and leads to quantum field
theory. But thats another module....
Chapter 10

The Feynman Path Integral

Our last topic is to present another way of formulating quantum mechanics known as
the path integral formulation. It is perhaps the most popular way as one can incorporate
symmetries such as Lorentz invariance from the outset. It also allows us to make contact
with the principle of least action and hence the Classical world.
Let us start by considering a general quantum mechanical system with coordinate q̂
and its conjugate momenta p̂:

[q̂, p̂] = i󰄁 (10.1)

Let us imagine that we have eigenstate of q̂:

q̂|a〉 = a|a〉 (10.2)

In a usual wavefunction representation this is slightly dodgy as

ψa (q) = 〈q|a〉 = δ(q − a) (10.3)

Similarly we can consider the momentum eigenstates

p̂|k〉 = k|k〉 (10.4)

which has a more reasonable, but still not normalizable, wavefunction

ψk (q) = 〈q|k〉 = eiqk/󰄁 (10.5)

which is clearly an eigenstate of p̂ = −i󰄁∂/∂q. The important point is that these


eigenstates are delta-function normalized:
󰁝
〈a1 |a2 〉 = ψa1 (q)∗ ψa2 (q)da

= δ(a2 − a1 )
󰁝
〈k1 |k2 〉 = ψk1 (q)∗ ψk2 (q)dq
󰁝
= ei(k2 −k1 )q/󰄁 dq

= 2π󰄁δ(k2 − k1 ) (10.6)

99
100 CHAPTER 10. THE FEYNMAN PATH INTEGRAL

In a suitable sense (don’t ask too many questions) the |a〉 form a basis. We can write
the identity operator as
󰁝
I = da|a〉〈a| (10.7)

Similarly so too do the |k〉:


󰁝
dk
I= |k〉〈k| (10.8)
2π󰄁
Let’s look at how |a〉 evolves in time (note that we need to put a hat on the Hamil-
tonian as we will later see):

|a; t〉 = eiĤt/󰄁 |a〉 (10.9)

Let is consider the overlap

K(a2 , t2 ; a1 , t1 ) = 〈a2 , t2 |a1 , t1 〉


= 〈a2 |eiĤ(t2 −t1 )/󰄁 |a1 〉 (10.10)

This is known as the propagator: it takes a particle at position a1 at time t1 and


computes the amplitude that it is found at a2 at time t2 . In addition if we start with a
more conventional wavefunction, i.e. one not localised in q:
󰁝
|ψ(t1 )〉 = da1 ψ(a1 , t1 )|a1 , t1 〉 (10.11)

then by linearity we have


󰁝
〈a2 , t2 |ψ(t1 )〉 = ψ(a1 , t1 )〈a2 , t2 |a1 , t1 〉da1
󰁝
= K(q, t; a1 , t1 )ψ(a1 , t1 )da1
󰁝
i.e. ψ(a2 , t2 ) = K(q2 , t2 ; a1 , t1 )ψ(a1 , t1 )da1 (10.12)

Thus the propagator evolves (propagates) wavefunctions in time. It is a Green’s function


for the Schrödinger equation.
We want to find an expression for K(a2 , t2 ; a1 , t1 ). Let us insert the identity matrix
written as an integral over momentum eigenstates:

K(a2 , t2 ; a1 , t1 ) = 〈a2 |IeiĤ(t2 −t1 )/󰄁 |a1 〉


󰁝
dk
= 〈a2 |k〉〈k|eiĤ(t2 −t1 )/󰄁 |a1 〉
2π󰄁
󰁝
dk ika2 /󰄁
= e 〈k|eiĤ(t2 −t1 )/󰄁 |a1 〉 (10.13)
2π󰄁
where we have used 〈a2 |k〉 = (〈k|a2 〉)∗ = eika2 /󰄁 . We will consider Hamiltonians of the
form
1 2
Ĥ = p̂ + V (q̂) (10.14)
2m
101

Next we consider an infinitesimal time step t2 − t1 = δt and expand:


󰁝
dk ika2 /󰄁
K(a2 , t1 + δt; a1 , t1 ) = e 〈k|I + 󰄁i Ĥδt|a1 〉
2π󰄁
󰁝 󰁝 󰀕 󰀖
dk ika2 /󰄁 i dk ika2 /󰄁 1 2
= e 〈k|a1 〉 + δt e 〈k| p̂ + V (q̂) |a1 〉
2π󰄁 󰄁 2π󰄁 2m
󰁝 󰁝 󰀕 󰀖
dk ika2 /󰄁 i dk ika2 /󰄁 1 2
= e 〈k|a1 〉 + δt e k + V (a1 ) 〈k|a1 〉
2π󰄁 󰄁 2π󰄁 2m
󰁝
dk ika2 /󰄁 i H(k,a1 )δt
= e e󰄁 〈k|a1 〉
2π󰄁
󰁝
dk i (ka2 +H(k,a1 )δt)
= e󰄁 〈k|a1 〉
2π󰄁
󰁝
dk i (k(a2 −a1 )+H(k,a1 )δt)
= e󰄁 (10.15)
2π󰄁
Note that now the Hamiltonian does not have a hat as we have evaluated it on position
and momentum eigenstates.
Now we want to add these all up to find K(a2 , t2 ; a1 , t1 ) for a finite time difference
t2 − t1 . To do this we split t2 − t1 = (t2 − tN +2 ) + (tN +2 − tN +1 ) + . . . + (t3 − t1 ) with
N → ∞ and N even:

K(a2 , t2 ; a1 , t1 ) = 〈a2 |IeiĤ(t−tN +2 )/󰄁 IeiĤ(tN +2 −tN +1 )/󰄁 IeiĤ(tN +1 −tN )/󰄁 . . . IeiĤ(t3 −t1 )/󰄁 |a1 〉
(10.16)
Next we replace each occurrence of I by (10.8) followed by (10.7) alternatively:
󰁝
dkN +2
K(a2 , t2 ; a1 , t1 ) = 〈a2 | |kN +2 〉〈kN +2 |eiĤ(tN −tN −1 )/󰄁
2π󰄁
󰁝
× daN +1 |aN +1 〉〈aN +1 |eiĤ(tN −1 −tN −2 )/󰄁 . . . |a1 〉
N
󰁜 +2 󰁝
dki
= dai 〈a2 |ki 〉〈ki |eiĤ(ti −ti−1 )/󰄁 |ai 〉 (10.17)
i=3
2π󰄁
We further suppose that each pair ti − ti−1 = (t2 − t1 )/(N + 1) = δt so that
N
󰁜 +2 󰀗󰁝 󰀘
dki i
(ki (ai+1 −ai )+H(ki ,ai )δt)
K(a2 , t2 ; a1 , t1 ) = dai e 󰄁
i=3
2π󰄁
󰁝 N󰁜 +2 i 󰁓
󰀓 a 󰀔
dki 󰄁 i ki
i+1 −ai
+H(ki ,ai ) δt
= dai e δt
(10.18)
i=3
2π󰄁
In the large N limit we can recognise the object appearing in the exponential as an
integral:
󰁛 󰀕 ai+1 − ai 󰀖 󰁝 t2
lim ki + H(ki , ai ) δt = k(t)ȧ(t) + H(k(t), a(t))dt (10.19)
N →∞
i
δt t1

where ai+1 − ai = ȧ(ti+1 − ti )/N = ȧδt with a dot denoting differentiation with respect
to time. Thus we find an example of a path-integral:
󰁝 󰀗 󰀘 󰁕
dk i
K(a2 , t2 ; a1 , t1 ) = [da]e 󰄁 (k(t)ȧ(t)+H(k(t),a(t)))dt (10.20)
2π󰄁
102 CHAPTER 10. THE FEYNMAN PATH INTEGRAL

where k(t) and a(t) are arbitrary paths such that a(t1 ) = a1 and a(t2 ) = a2 . This is an
infinite dimensional integral, we are integrating over all paths k(t) and a(t). It is not
well understood Mathematically but is very powerful in Physics.

10.1 Gaussian Path Integrals


The Path Integral formulation is conceptually very nice but of course impossible to
compute exactly for general systems. Formally it is defined (in so far as it is defined) as
a limit over a finer and finer net of finite-dimensional integrals. However there is one class
where we can have success: If the Lagrangian is quadratic then all the finite-dimensional
integrals that appear are just Gaussian integrals of the form
󰁝 󰁵
−Ak2 +Bk π B 2 /4A
dke = e (10.21)
A
The integral of a Gaussian gives another Gaussian.
In a particular we note that the integral over k(t) is a Gaussian. What exactly do
we mean by that. Let us look at one of the integrals in (10.18):
󰁝 󰀕
k2
󰀖 󰁵 󰀕
−m(ai+1 −ai )2
󰀖
dk 󰄁i k(ai+1 −ai )+ 2m δt+V (ai )δt 1 2mπ󰄁 󰄁i 2δt
+V (a i )δt
e = e
2π󰄁 2π󰄁 iδt
󰁵 󰀕
m(ai+1 −ai )2
󰀖
m − 󰄁i
2(δt) 2 −V (a1 ) δt
= e (10.22)
2π󰄁iδt
(We should be more careful and Wick rotate (t2 − t1 ) → −i(τ2 − τ1 ) to get a damped
exponential). We can do this for each of the ki integrals (for i ∕= 1, 2). Putting this back
into the finite t2 − t1 expression gives
󰁝 N󰁜 󰀕 󰀖
+2 i 󰁓 m(ai+1 −ai )2
− −V (a i ) δt
dai e 󰄁 2(δt)2
i
K(a2 , t2 ; a1 , t1 ) = N (10.23)
i=3

where N is an infinite but irrelevant constant that we can absorb. Again we recognise
the exponent as an integral
󰁛 󰀕 m(ai+1 − ai )2 󰀖 󰁝 t2
1 2
lim 2
− V (ai ) δt = mȧ − V (a)dt
N →∞
i
2(δt) t 1
2
= S[a] (10.24)

the classical action! Thus the propagator can be written as


󰁝 󰁕 1
K(a2 , t2 ; a1 , t1 ) = [da]e− 󰄁 ( 2 m(ȧ(t)) −V (a(t)))δt
i 2

󰁝
i
= [da]e− 󰄁 S[a(t)] (10.25)

An interesting and tractable case to consider is the free Schrödinger equation V = 0.


Going back a step we have
󰁝 N󰁜 󰀕 󰀖
+2 i 󰁓 ai+1 −ai ki2
dki k i + δt
dai e 󰄁
i δt 2m
KS (a2 , t2 ; a1 , t1 ) = (10.26)
i=3
2π󰄁
10.1. GAUSSIAN PATH INTEGRALS 103

The integrals over dai give δ-functions


󰁝
i
dai e 󰄁 (ki+1 (ai+1 −ai )+ki (ai −ki−1 )) = 2π󰄁δ(ki − ki+1 ) (10.27)

This leaves us with only one independent k:


󰁝 󰀓
−a1 󰁓 k2
󰀔
dk 󰄁i k a2δt + i 2m δt
KS (a2 , t2 ; a1 , t1 ) = e
2π󰄁
󰁝 󰀓 󰀔
dk 󰄁i k(a2 −a1 )+ 2m
k2
(t2 −t1 ) δt
= e (10.28)
2π󰄁
󰁓
where we have used that i δt = t2 − t1 . This leaves just one Gaussian integral which
evaluates to:
󰁵 im(a2 −a1 )2
m
KS (a2 , t2 ; a1 , t1 ) = e 2󰄁(t2 −t1 ) (10.29)
2π󰄁i(t2 − t1 )

This is known as the Schrödinger propagator. You can check that, for t2 ∕= t1

∂KS 󰄁 2 ∂ 2 Ks
i󰄁 =− . (10.30)
∂t2 2m ∂a22

A more careful treatment reveals that


∂KS 󰄁 2 ∂ 2 Ks
i󰄁 + = δ(t2 − t1 ) . (10.31)
∂t2 2m ∂t2
Thus we see that in the Path Integral formulation of Quantum Mechanics the ampli-
tude for a particle to go from q = a1 at t = t1 to q = a2 at t = t2 is given by an infinite
dimensional integral over all possible paths, classical or not, that start at a1 and end at
a2 . Each path is then given a weight e−iS/󰄁 .
Here we see the Lagrangian of classical mechanics appearing. As you may recall from
classical mechanics the action, as determined by the Lagrangian, was a mysterious object
with no physical interpretation yet it controlled everything (symmetries, equations of
motion,...). However here we see it arises as the logarithm of the weight that appears
in the path integral. Each path contributes a weight that is a pure phase so they all
contribute equally in magnitude. However there can be interference. The dominant
contributions then correspond to those where the interference is the weakest. These
corresponds to paths where δS = 0, i.e. extrema. Thus the classical paths corresponds
to the most dominant contribution to the path integral with other paths cancelling out
due to the rapid oscillatory nature of the path integral for ‘small’ 󰄁. This explains the
principle of least action.
Furthermore we see that we can marry Quantum Mechanics with Special Relativ-
ity by simply taking an action which is invariant under Lorentz transformations. For
example it is not too hard to see that the Klein-Gordon equation arises from the action
󰁝
m2 c3
SKG = c∂µ Ψ∗ ∂ µ Ψ − 2 Ψ∗ Ψ d4 x (10.32)
󰄁
104 CHAPTER 10. THE FEYNMAN PATH INTEGRAL

which is indeed invariant under Lorentz transformations. Therefore we can build a


relativistic quantum theory, in this case a quantum field theory, by considering
󰁝
i
Z = [dΨ]e− 󰄁 SKG [Ψ] (10.33)

This object is similar to the partition function that we saw above only in imaginary
temperature β = −i/󰄁 and weighted by the action not the energy. In fact the relation
between the partition function of a quantum theory, as defined by (10.33), and the
partition function we saw in our discussion of thermal statistical physics is quite deep.

10.2 Computing Determinants


1 2
Let us look at a simple example H = 2m
p̂ + 12 k q̂ 2 . The corresponding Lagrangian is of
course just
1 1
L = mq̇ 2 − kq 2 (10.34)
2 2
It is helpful to rewrite the action as
󰁝
1 d2
S=− qEqdt E =m 2 +k (10.35)
2 dt
where we view E as a differential operator. Let us expand q(t) in a real orthonormal
basis of functions
󰁛
q= qn en (t) (10.36)
with
󰁝
en (t)em (t)dt = δnm

Thus
󰁛 󰁝
1
S=− qn qm
en (t)Eem (t)dt (10.37)
2
Let us choose en (t) to be eigenstates of E with eigenvalues λn :
󰁛 󰁝
1
S=− qn qm en (t)λm em (t)dt
3
1󰁛
=− λn qn2 (10.38)
2 n
Thus
󰁝 󰁝 󰁜
i i 󰁓
λn q22
[dq]e− 󰄁 S[q] = dqn e 2󰄁 n

n
󰁜󰁝 i 2
= dqn e 2󰄁 λn qn
n
󰁵
󰁜 −πi󰄁
=
n
λn
N 󰁜
=√ , det E = λn (10.39)
det E n
10.2. COMPUTING DETERMINANTS 105

Here N is another normalization constant that we aren’t interested in.1


Its easy to see how this generalises: For a quadratic action with many coordinates
qa , a = 1, ..., n of the form
󰁝
1
S=− qa Eab qb dt (10.40)
2
we find
󰁝 󰁜
i N
[dq]e 󰄁 S[q1 ,...,qn ] = √ (10.41)
a
det Eab

where the determinant is the infinite product of eigenvalues of the operator Eab . More
generally still if one has

󰁝
1
S=− qa Eab qb + Ja qa dt (10.42)
2
we find
󰁝 󰁜
i N −1
[dqa ]e 󰄁 S[q1 ,...,qn ] = √ eJa Eab Jb (10.43)
a
det Eab

1
Some of you might be worrying that these integrals are not well defined and this is just hocus-pocus.
Don’t, just relax and don’t ask too many questions.

You might also like