Boettcher Thermodynamics
Boettcher Thermodynamics
Boettcher Thermodynamics
Igor Boettcher
University of Alberta
Contents
1 What is thermodynamics? 3
2 Thermodynamic systems 4
2.1 Thermodynamic variables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
2.2 States and equilibrium states . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
2.3 Equation of state . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11
2.4 Mixtures . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
2.5 Extensive and intensive variables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
2.6 Thermodynamic processes and work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14
5 Thermodynamics potentials 52
5.1 Free energy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52
5.2 Chemical potential . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54
5.3 Overview of thermodynamic potentials . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 57
5.4 Maxwell relations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 59
5.5 Phase transitions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 62
2
1 What is thermodynamics?
Thermodynamics is mainly concerned with the transformations of heat into mechanical work and the opposite
transformations of mechanical work into heat.
3
2 Thermodynamic systems
2.1 Thermodynamic variables
Typical systems we consider in thermodynamics are
a block of iron,
a bucket of water,
Systems with qualities (i) and (ii) are called macroscopic. This is in contrast to microscopic systems, which
are typically tiny and may contain only a few particles.
In each example system, we could try to describe the physics by solving Newton’s equations of motions
for each of the N particles in the system. Complete information then requires to determine the position
xi = (xi , yi , zi ) and velocity vi = (vx,i , vy,i , vz,i ) for each particle, i = 1, . . . , N , at each instance of time.
Exercise. Estimate the number of atoms N in a 1 cm3 block of iron, with mass density ρ = 7.87g/cm3
and atomic mass of m0 = 55.845 u. (The atomic mass unit is 1u = 1.66 × 10−27 kg.)
Answer. The block has a volume of V = 1cm3 . The total mass of the block is M = ρV . One atom
weighs on average m0 , so the total number of atoms is
M ρV
N= = (2.1)
m0 m0
7.87g × 1cm3
= (2.2)
cm3 × 55.845 × 1.66 × 10−27 kg
7.87 × 10−3 kg × 1cm3
= (2.3)
cm3 × 55.845 × 1.66 × 10−27 kg
7.87 × 10−3
= (2.4)
55.845 × 1.66 × 10−27
= 8.5 × 1022 . (2.5)
Solving that many Newtonian equations, or specifying the initial conditions, or even storing and sensibly
interpreting the final outcome, is practically impossible. However, it is also absolutely unnecessary:
Thermodynamic systems consist of a very, very large number of particles N . Experiments performed on
macroscopic length and time scales are dominated by the collective behavior of particles, not individual
particles. These systems are, therefore, described by a small set of thermodynamic variables.
We now discuss the most important thermodynamic variables and how they can be measured.
Volume V . The volume is simply the interior volume of the container that contains the thermodynamic
system. If the linear extensions of the system are Lx , Ly , Lz in x, y, z direction, then V = Lx Ly Lz . If Lz is
changed by an infinitesimal amount dLz , then this corresponds to a change in volume
V → V + dV (2.6)
4
with
dLz dLz
dV = Lx Ly dLz = Lx Ly Lz =V . (2.7)
Lz Lz
Note that on the right hand side it does not matter whether we insert the original volume (V ) or the new
volume (V + dV ), because the rules of differential calculus imply
dV dLz ≈ 0 (2.8)
Memorize that 1 liter of water almost exactly weighs 1 kg. 1 quart (= 1 pack of chicken stock) is roughly
the same as 1 liter. Another common measure is the gallon,
(
4.5 l (UK, Canada, ...)
1 gal = = 1 big milk bottle. (2.10)
3.8 l (US,...)
To measure the volume of a liquid, one pours it into a container of known linear extensions such as a
measuring jug. To measure the volume of a (perhaps weirdly shaped) solid object, one can suspend it into
water and then measure the amount of water that has been displaced. Measuring the volume of a gas is
more tricky, because gases are compressible, i.e. their volume can be changed upon exerting pressure, and
their volume stronger depends on temperature. For gas inside a container of known volume V , one always
needs to also measure pressure and temperature.
Particle number N . The particle number N remains constant in many elementary thermodynamic processes.
(Counter-examples are chemical reactions, or biological systems with osmotic walls which allow particles to
leave or enter the system.) Note that N is a dimensionless number, say, N = 1, 999, 045, 788. However,
chemists like to use a standard unit of particle number, which is the mole.
A mole consists of
particles (a dimensionless number!). Compare this to other such conventions: a pair consists of 2 entities,
a quartet consists of 4 entities, a dozen consists of 12 entities. In calculations, I recommend to use the
dimensionless unit
to keep track of the number of moles. For instance, the number of iron atoms determined earlier is
1 mol 1 mol
N = 8.5 × 1022 = 8.5 × 1022 = 8.5 × 1022 = 0.14 mol. (2.13)
NA 6.022 × 1023
The number of moles ν, i.e. the number that satisfies N = ν mol = νNA (ν = 0.14 in the example), is given
by
N
ν= . (2.14)
NA
The quantity ν is sometimes denoted n in other reference, but not in this lecture! Instead, we will use n to
denote the particle number density
N
n= . (2.15)
V
5
For physicists, the introduction of moles is not too useful, and I recommend to always work with N instead
of ν. Chemists might find ν more handy for expressing their results. Indeed, the unwieldy atomic mass unit
1 u = 1.66 × 10−27 kg corresponds to a mass per mole given by
NA 6.022 × 1023 g
1u=1u = 1.66 × 10−27 kg =1 , (2.16)
1 mol 1 mol mol
which is a decent quantity to work with. The hardly intuitive atomic mass of iron, m0 = 55.845 u, becomes
NA g
m0 = m0 = 55.845 , (2.17)
1 mol mol
and so on. For a thermodynamic system that consists of one chemical compounds with mass m0 , the total
mass of the system is
M = N m0 , (2.18)
which is the Earth’s average air pressure at sea level. Another popular unit is
so that
Other units are Torr or psi (pounds per square inch). Typical car tires have a pressure of
6
whereas bicycles have a wider range, and typically higher pressure.
Temperature T . Temperature is the most interesting variable here. In the kinetic picture, temperature is
defined as the average kinetic energy of a gas of particles, basically the formula
3
⟨E⟩ = N kB T, (2.26)
2
which, again, is a mechanistic definition. However, the property ”temperature” of a system can be defined
purely phenomenologically through the experimentally confirmed zeroth law of thermodynamics. For now,
we take for granted that there exist thermometers, which are devices which, when brought into contact with a
thermodynamic system, display an empirical temperature ϑ. For example, a mercury thermometer displays
ϑ through thermal expansion of a mercury column. A gas thermometer measures the temperature T via the
variation of volume or pressure of a gas. We call T the absolute temperature. Any two thermometers can
be calibrated with each other to determine the function ϑ(T ).
The absolute temperature T is measured in the SI-unit Kelvin. Note that we write (and say)
1 K, not 1◦ K. (2.27)
Temperature differences ∆T = T2 − T1 in K are identical when expressed in ◦ C. The freezing and boiling
point of water at 1 atm are
T = 0 K = −273.15◦ C. (2.30)
Temperatures expressed in degrees Fahrenheit can be converted to degrees Celsius through the rough rule
”minus 30, divide by two”, for instance
particle number N
pressure P
temperature T
The 6N positions and velocities of all particles in a system determine its dynamical state. The knowledge of
the thermodynamic variables of a macroscopic system are by no means sufficient to determine its dynamical
state. For any choice of admissible thermodynamic variables, there is an infinite number of states of atomic or
molecular motion that correspond to it. During macroscopic time scales, the system rapidly passes through
many of these corresponding dynamical states, while leaving the thermodynamic variables constant.
7
2.2 States and equilibrium states
The distinction between (general) thermodynamic states and (specific) equilibrium states of macroscopic
systems is a crucial aspect of thermodynamics. This important concept, for instance, needs to be mastered
to understand the difference between reversible and irreversible processes later.
A macroscopic system is in a thermodynamic state if it can be described by volume, V , and local particle
number density, pressure, and temperature, n(x), P (x), T (x), where x = (x, y, z) is any space point inside
the system. More local thermodynamic variables may be needed for more complex system.
What does, say, local T (x) mean? We only defined T for large macroscopic systems. While the concept of a
local temperature may be intuitive, we need to properly define it. For this note that the number of particles
N is usually so large that we can divide the volume V into many smaller compartments of volume V0 ≪ V ,
such that each volume V0 still contains a macroscopically large number of particles, and hence pressure and
temperature are well-defined thermodynamic variables. Furthermore, the compartments can be chosen small
enough so that pressure and temperature within each compartment are approximately constant. For any
individual compartment at position x0 inside the system, we then define
N0
n(x0 ) = , P (x0 ) = P0 , T (x0 ) = T0 , (2.34)
V0
where V0 , N0 , P0 , T0 are the volume, particle number, pressure, and temperature of that individual com-
partment. By considering all compartments at various locations, we obtain the functions n(x), P (x), T (x).
The thermodynamic variables of a system are usually controlled by external conditions. For instance, the
volume is determined by the geometry and hence the walls of the container. The temperature can be ma-
nipulated through the temperature of the environment. The particle number may be fixed by ensuring that
no particles can leave through leaks in the container walls.
An equilibrium state of a macroscopic system is a thermodynamic state whose local thermodynamic variables
do not vary as long as the external conditions remain unchanged.
The reference to time in this definition, of course, means macroscopic time scales relevant to experiment,
not time scales relevant for the microscopic motion of individual atoms or molecules. An equilibrium state
can be created in practice by
1. imposing certain external conditions,
2. waiting.
Over a sufficiently long time, the microscopic mechanisms in the system will make it equilibrate, i.e. reach
the equilibrium state.
In many cases, the external conditions do not impose any spatial variation of the thermodynamic variables
inside the system, and the equilibrium states are simply the homogeneous states, in which n(x), P (x),
T (x) are independent of x. Non-equilibrium states typically feature macroscopic currents as they appear in
hydrodynamics.
The following two examples of equilibrium and non-equilibrium states may be instructive.
Example 1. In the isothermal expansion, a gas is initially confined to the left one of two equal containers,
with a volume V , with constant n, P, T . This is an equilibrium state. The gas is coupled to a heat
8
equilibrium 2 non-equilibrium
1
3 non-equilibrium 4 equilibrium
These examples may give the impression that equilibrium states cannot describe thermodynamic systems
that ”move”, meaning that their thermodynamic variables vary in time. This would be bad, since, clearly,
a Diesel engine is only useful when it is running. Note, however, that our definition of equilibrium relies
on unchanged external conditions. In any motor or machine, pistons and the like are moving, so that the
external conditions themselves are time-varying.
We can now formulate some important definitions that specify the relation between a thermodynamic system
(the ”system”), assumed to be contained in a container, and the surrounding of the container (the ”environ-
9
environment
system
ment”). Consider the system as a compartment at position xs , with system thermodynamic variables
We say that the system is thermally insulated or thermally isolated, if the container walls do not allow heat
to be exchanged between the system and the environment. (For instance, a double wall filled with air, which
is a bad heat conductor.) In contrast, if the walls allow for heat to be transmitted, such as a thin sheet of
metal, we say that system and environment are in thermal contact or that they can exchange heat.1 When
a system and its environment are in thermal contact, we say that the system is in thermal equilibrium with
its environment if
Ts = Te . (2.37)
When a system and its environment are separated by a movable wall, then this wall will be shifted around if
the pressures from either side do not match. In this case, we say that they are in mechanical equilibrium if
Ps = Pe . (2.38)
Quite generally, if two systems with temperatures T1 and T2 , and pressures P1 and P2 , are in equilibrium
with each other, then
T1 = T2 , (2.39)
P1 = P2 . (2.40)
This equilibrium state can only be reached when the two systems are able to ”talk to each other” (transfer
heat or do work), hence we need thermal contact and movable walls to establish equilibrium.
In the following we will always assume that there is no exchange of particles between the system and the
environment. Let us discuss, however, how such an exchange would look like. Assume the system consists of
1 Good heat conductors are heat and stone, bad heat conductors are air, wood, clothes. As a rule of thumb, if something
feels cold to the touch, then it is a good heat conductor (heat leaves your hand), whereas if it feels warm, then it is a bad heat
conductor (heat stays in your hand).
10
water, with density n(wat) , and the environment contains oil, with density n(oil) , both separated by container
walls. Then a complete set of thermodynamic variables is (n(wat) , n(oil) , T, P ). We assume Ts = Te = T . If
the container walls are impenetrable for the molecules, then
On the other hand, if particles can pass through penetrable or porous walls, then we eventually reach a state
of chemical equilibrium with
Consider first a homogeneous system with thermodynamic variables n = N/V, P, T . There exists a relation
called the equation of state which relates these variables according to
ϕ(n, P, T ) = 0. (2.45)
Let us give some meaningful examples first:
Example 1. Almost all gases at low pressures and sufficiently far above the condensation temperature
are described by the ideal gas equation of state
P V = N kB T (2.46)
Example 2. When temperature and pressure are such that a real gas is near condensation, deviations
from the ideal gas law are observed. A good phenomenological model is then given by the van-der-
Waals equation of state
N kB T aN 2
P = − 2 (2.48)
V − bN V
nkB T
= − an2 (2.49)
1 − bn
where a, b are parameters specific to the gas under consideration. In this case,
We will discuss these two examples in more detail later. Some remarks are in order.
11
Thermodynamics is a set of general rules and formulas. The equation of state is the only point where
we specify what substance we are actually dealing with (water, air, iron, . . . ).
We always assume that we can solve the equation ϕ(n, P, T ) = 0 to express any of the three variables
n, P, T as a function of the remaining two variables. Only two variables of n, P, T are independent.
The equation of state has been measured for many substances and is stored in databases. Explicit
formulas for ϕ exist only in a few cases. Calculation of the equation of state is one of the subjects of
statistical mechanics or quantum field theory.
Often it is assumed that N is constant and the equation of state is then written as
ψ(P, V, T ) = 0. (2.51)
For instance, ψ(P, V, T ) = P V − N kB T for ideal gases. For fixed N , the thermodynamic state
is uniquely determined by any two of the variables P, V, T .
Consider now a macroscopic system in a general thermodynamic state, i.e. it is characterized by the ther-
modynamic variables V, n(x), P (x), T (x). Then at each point x there is a local equation of state
Crucially, the function ϕ(n, P, T ) is always the same, i.e. it does not depend itself on x. The x-dependence
is only through its arguments.
2.4 Mixtures
Mixtures of several chemical compounds are straightforward to treat within thermodynamics, but due to the
increased number of thermodynamic variables, they are a little unwieldy. Here we collect some basic facts.
For a homogeneous macroscopic system that is a mixtures of X different chemical compounds, we associate
particle numbers N (j) to each of the compounds labelled by an index j = 1, . . . , X. The concentration of
component j is given by
N (j)
c(j) = , (2.55)
N
with
X
X
N (j) = N, (2.56)
j=1
X
X
c(j) = 1. (2.57)
j=1
12
Example. The chemical composition of air is mostly nitrogen, oxygen, argon, and carbon dioxide,
hence X = 4 and j = N, O, Ar, CO2 , with concentrations
We have
X
c(j) = 100%. (2.62)
j
N (j)
n(j) = = c(j) n, (2.63)
V
with n = N/V the total particle number density. The temperature of the homogeneous system shall be
T . Each component j has an equation of state ϕ(j) (n(j) , P (j) , T ) = 0 that is independent of the remaining
chemical compounds. If we assume that the interactions between the chemical components j is approximately
absent (this need not be the case), then these equations of state remain valid and the partial pressure P (j)
of the jth compound is found from solving
Note also that in general each component j has a different mass, m(j) , and so ρ(j) = m(j) n(j) for the mass
density.
V ′ = 2V, (2.67)
′
N = 2N. (2.68)
The pressure and temperature, however, remain the same regardless of the spatial extent of the system.
Hence
T ′ = T, (2.69)
′
P = P. (2.70)
13
Similarly, the particle number density n = N/V remains the same,
n′ = n. (2.71)
A thermodynamic quantity that doubles during an imagined doubling of a homogeneous system is called
extensive. Examples are V and N . A quantity that remains invariant under an imagined doubling of a
homogeneous system is called intensive. Examples are P, T, n. The ratio of two extensive quantities is
intensive, as is the case for n = N/V .
Intensive quantities are usually local quantities. Typical examples are densities. For instance, the total mass
of a system, M = m0 N , is an extensive quantity, but the mass density
M
ρ= (2.72)
V
is an intensive quantity. The mass per particle,
M
m0 = , (2.73)
N
as a ratio of two extensive variables is, of course, also intensive.
If the left-hand-side of an equation is an extensive (intensive) quantity, then the right-hand side must be
extensive (intensive) quantity. For instance, for an ideal gas,
If you are uncomfortable with doubling a given system S, you may alternatively divide the system into
two macroscopic pieces, S = S1 ∪ S2 , by inserting an additional wall. Extensive quantities X then satisfy
X = X1 + X2 , whereas intensive quantities Y satisfy Y = Y1 = Y2 .
ψ(P, V, T ) = 0. (2.76)
Choose P, V as the independent variables. It is often convenient to represent them in a PV-diagram, with
P on the y-axis and V on the x-axis. Any point in the PV-diagram defines a state of the system. The
temperature T = T (P, V ) is a function of P and V . Points representing states of equal temperature lie on
a curve which is called an isothermal.
14
P P
non-equilibrium process
(e.g. inhomogeneous
intermediate states)
V V
Figure 2.3: Left. PV-diagram and three possible transformations connecting the states (V1 , P1 ) and (V2 , P2 ).
Right. A fourth transformation that utilizes non-equilibrium, inhomogeneous intermediate states and thus
cannot be drawn in a PV-diagram, because P and V are not defined during the intermediate steps.
We define a thermodynamic transformation or process as any transformation that brings a system from an
initial state to a final state through a continuous succession of intermediate states. If the initial and final
states are represented by two points (V1 , P1 ) and (V2 , P2 ) in a PV-diagram, then any curve connecting these
points corresponds to a possible transformation. They correspond to different ways of modifying the external
conditions to get from 1 to 2. However, not every transformation connecting 1 to 2 is represented by such a
curve.
A transformation is said to be reversible when the successive states of the process differ by infinitesimals
from equilibrium states. In particular, the initial and final states of a reversible transformation need to be
states of equilibrium.
If the intermediate states are non-equilibrium states, then they are typically inhomogeneous, and so cannot
be drawn in a PV-diagram.
We revisit here the central example of reversible vs. irreversible processes. We define a heat bath or heat
reservoir as an environment of temperature T that contains so many particles that its temperature does not
change when it is coupled to a (much smaller) system.
15
Isothermal expansion (irreversible run)
1. Consider a gas inside a container with immovable walls of volume 2V coupled to a heat bath of
temperature T , so that the system temperature remains constant.
2. Initially, at time t ≤ 0, a wall separates the left and right halves of the container. An ideal gas of N
molecules in the equilibrium state (Pi , V, T ) occupies the left half. Its pressure is Pi = N kB T /V .
5. After a sufficient long time teq , the gas reaches the final equilibrium state (Pf , 2V, T ) with pressure
Pf = N kB T /(2V ) = Pi /2.
This thermodynamic transformation from (Pi , V ) to (Pf , 2V ) at constant temperature T is irreversible, since
the intermediate states are far away from equilibrium states. Practically, it is impossible to start in the final
state of the gas occupying both halves of the container and make it move into the left half of the container
without any external influence. If you were shown a video recording of the time-interval t ∈ [0, teq ], you
could tell if the video ran forward or backward. The intermediate states cannot be plotted in a PV-diagram.
However, consider now the following modification.
1. As before. However, an externally controlled piston is inserted in the right half of the container,
initially touching the separating wall.
2. As before.
3. As before.
4. After the separating wall is removed, at t > 0, the piston very slowly inches to the right, by an amount
∆ℓ over a time-interval ∆t. The speed ∆ℓ/∆t is chosen slow enough so that the gas has enough time
to adjust itself to the altered external conditions through the usual molecular motion. No macroscopic
currents are generated. At each time t, the gas is infinitesimally close to an equilibrium state, with a
volume V < V (t) < 2V , temperature T , and pressure Pi > P (t) > Pf .
5. After a certain amount of time ttot , the piston opened up the whole right half of the container, and
the gas is in the final equilibrium state (Pf , 2V, T ).
This way of performing the isothermal expansion meets all criteria of the definition of a reversible process.
And, indeed, it can be reversed: Starting from the final gas occupying the volume 2V with pressure Pf , we
could slowly move the piston to the left with a negative velocity −∆ℓ/∆t, so that the gas is in equilibrium at
each time-step. After the amount of time ttot , the gas will be back in the initial state (Pi , V, T ). If you were
shown a video recording of the time-interval t ∈ [0, ttot ], you could not tell whether the video ran forward
or backward.
The fact that two equilibrium states can be connected by various sorts of processes, some reversible and
some irreversible, will be extremely important later on. In a sense, it is at the very heart of thermodynamics.
So make sure you understood this example of isothermal expansion.
16
isothermal expansion (irreversibe run)
During a transformation, the system can perform positive or negative work ; that is, the system can do work
on its surroundings or the surroundings can do work on the system. As an example, we consider a fluid
enclosed in a cylinder having a movable piston of area S at one end. If P is the pressure of the fluid against
the walls of the cylinder, then F = P S is the force exerted by the fluid on the piston. If the piston is shifted
by the gas by an infinitesimal distance dh, then an infinitesimal amount of ”work done”
dL = F dh = P S dh (2.79)
is performed, since the displacement is parallel to the force. But Sdh is equal to the increase in the volume
of the system, dV . Thus, we may write
dL = P dV. (2.80)
For a finite transformation, the work done by the system is obtained by integrating this equation,
V2
L= P dV. (2.81)
V1
Here, 1 and 2 are the initial and final states of the transformation.
Convention. The first law of thermodynamics will be about book-keeping the energy of the system—not
of the environment. Therefore, we want to denote every energy unit that increases the system’s energy as
positive, and every unit that decreases the system’s energy as negative. (Think of your bank account.) If a
system does work on its environment, it loses energy, so this is a negative contribution. This is logical, but
can be cumbersome in some arguments, as we may tend to prefer working with positive quantities.
17
P P
V V
Figure 2.5: Work done L = −W in the PV-diagram. The right figure shows a cyclic transformation.
We will, therefore, use two different symbols: ”work ” is denoted by the symbol
W. (2.82)
If the system does work, then W < 0, but if work is done on the system so that it gains energy, then W > 0.
We also say that W is the work performed by external forces. In addition, following Fermi, we define the
”work done”, meaning the work done by the system, as
L = −W. (2.83)
If the system does work, then L > 0. (L stands for lavoro, meaning work in Italian.)
In the reversible example above, the work done was dL = P dV , whereas dW = −P dV .
For a system whose state can be represented in a PV-diagram during the transformation, the work done
during a transformation from state (V1 , P1 ) to state (V2 , P2 ) has a simple geometric representation: the
process is represented by a curve P (V ) connecting 1 and 2, the shape of which depends on the type of
process considered. The work done during the process is given by the integral
V2
L1→2 = P (V ) dV. (2.84)
V1
This integral, and hence the work done, geometrically correspond to the area under the curve in the PV-
diagram. If the process can be run in the opposite direction, the work done is the negative of this result,
V1
L2→1 = P (V ) dV = −L1→2 . (2.85)
V2
The work performed in a transformation between two states 1 and 2 is typically different for reversible and
irreversible runs. For reversible runs we have
whereas for irreversible runs such a simple formula does not exist.
Consider again the example of isothermal expansion.
Work done in irreversible run. For the irreversible isothermal or ”free” expansion, no work is done:
L = 0. (2.87)
18
Indeed, after the wall has been removed in 3, while the gas expands into the void, the system volume 2V
remains constant and there are no movable walls ”to be pushed”. No work is done.
Work done in reversible run. Now compute the work done in the reversible isothermal expansion from an
initial volume V1 to a final volume V2 > V1 . At each instant of time, the system is homogeneous and can be
represented as a state in a PV-diagram. dL = P dV applies at each time. Since T is constant we have
N kB T
P = , (2.88)
V
hence
V2
L= P dV (2.89)
V1
V2
1
= N kB T dV (2.90)
V1 V
V2
= N kB T ln V (2.91)
V1
= N kB T ln V2 − ln V1 (2.92)
V2
= N kB T ln . (2.93)
V1
Since V2 > V1 we have L = L1→2 > 0. The system does work on the piston while expanding. In the reversed
run, the piston does work on the system when compressing the gas and W2→1 > 0.
Important transformations are those for which the initial and final states are the same. These are called
cyclic processes or cycles. If the state of the system can be represented on a PV-diagram, then a cycle can
be represented by a closed curve in this diagram. The work done during one run of the cycle equals the area
enclosed by the curve (or the negative thereof, if the cycle is run in the opposite direction).
Recall that processes with dT = 0 are called isothermal. Here we extend this list of names according to the
following processes:
dT = 0 or T = const: isothermal,
dP = 0 or P = const: isobaric,
dV = 0 or V = const: isochore.
19
3 First law of thermodynamics
3.1 Conservative and non-conservative forces
We start with a digression into Newtonian mechanics.
Conservative forces. Consider a particle at position x = (x, y, z) with potential energy V (x). The force
acting on the particle at x is
−∂x V (x)
F(x) = −∇V (x) = −∂y V (x) . (3.1)
−∂z V (x)
For example, Earth’s gravitational field is Vgr (x) = m0 gz, with m0 the mass of the particle, and
0
Fgr (x) = 0 . (3.2)
−m0 g
q1 q2
The electric potential between two electric charges q1 , q2 is Vel (x) = |x| , with x = x1 − x2 the relative
coordinate between the two charges, and
x
q1 q2
Fel (x) = y . (3.3)
|x|3
z
Forces Fgr and Fel are examples of conservative forces, because they can be written as the gradient of some
function −V (x). Note that F = 0 is also a conservative force.
Mechanical work. We say that F(x) is a force field, as it specifies the force on a particle at position x. In
order to move a particle from x to x + dx in a force field F(x), we need to perform the amount of work
This work could be supplied, for instance, mechanically (say, with our muscles) or electrically. To bring the
particle from position xA to xB , we can divide the path into K sufficiently small steps {∆x(1) , . . . , ∆x(K) }
according to
x(1) = xA , (3.5)
x(2) = xA + ∆x(1) , (3.6)
..
. (3.7)
(k+1) (k) (k)
x =x + ∆x , (3.8)
..
. (3.9)
(K+1)
x = xB . (3.10)
20
We could make the steps infinitesimal and write
xB xB
W = dW = F(x) · dx, (3.13)
xA xA
Non-conservative forces. Typical examples of non-conservative forces are friction, water drag, or air drag.
They cannot be written as the gradient of any function. For instance, a friction force acting on a particle
may have the form
Conservation of energy. A mechanical system is called conservative if all forces acting on it are conservative
forces. The total energy is then the sum of the kinetic and potential energy of all the particles at any given
time. It is determined by the dynamical state. The total energy of a conservative system is conserved.
Systems that feature non-conservative forces are called dissipative. Energy is still conserved, but only for the
total system including its environment. Mechanically, the heat that is created through friction corresponds
to disordered motion of particles. This case is captured by the 1st law of thermodynamics.
The first law of thermodynamics is essentially the statement of the principle of conservation of energy for
thermodynamical systems. It may be expressed by stating that the variation in energy of a system during
any transformation is equal to the amount of energy that the system receives from its environment.
Consider first a conservative mechanical system of N particles. The system energy is the sum of the potential
and the kinetic energy, hence a function of the dynamical state. Assume A and B are two successive states
of an isolated system with energies EA and EB . If no external forces act on the system, energy remains
constant, and
EB = EA . (3.16)
When external forces act on the system, this equality need no longer hold. If W is the work performed by the
external forces during a transformation from the initial state A to the final state B, then the conservation
of energy implies
EB = EA + W. (3.17)
21
Importantly, the work done during the transformation only depends on the initial and final states, A and
B, not on the particular way in which the transformation from A to B is performed. This is only true for
conservative systems.
EB = EA + W (3.18)
Assume that we find this property contradicted by an experiment performed on a particular thermodynamic
system. If we do not wish to discard the principle of conservation of energy, then we must admit the existence
of other methods, besides mechanical work, by means of which energy can be exchanged between the system
and its environment. Let us consider such an example.
The amount of work W in going from A to B depends on whether we go by means of the first or second way.
Assuming that the principle of conservation of energy holds for our system, then the energy transmitted to
the water in the form of mechanical work of rotating paddles in the second way must be transmitted to the
water in the first way in a non-mechanical form called heat. We are lead to the fact that heat and mechanical
work are equivalent in this example. They are two different aspects of the same thing, namely, energy.
First law of thermodynamics. We now put all this together into a quantitative formula. We first enclose
our system in a cylindrical container with a movable piston at one end. If the container walls are non-heat-
conducting, then the system is thermally isolated, and the exchange of energy between the system and the
environment can only occur in the form of mechanical work. The amount of work performed on the system,
W , when going from state A to state B only depends on the initial and final states of the transformation.
Denoting ∆E = EB − EA , the work satisfies
∆E = W. (3.19)
(First law of thermodynamics for thermally insulated systems)
If our system is not thermally insulated, this is generally not true, because there can be an exchange of energy
22
between the system and the environment in the form of heat. We then write the more general equation
∆E + L = Q, or (3.20)
∆E = W + Q, (3.21)
(First law of thermodynamics)
where Q is equal to zero for transformations performed on thermally insulated systems and otherwise, in
general, is different from zero.
The heat Q can be interpreted physically as the amount of energy that is received by the system in forms
other than work.
The first law of thermodynamics is a precise formulation of the equivalence of heat and work.
The work W appearing in the first law could also be electric or magnetic work, but we will mostly be dealing
with situations where W is mechanical work.
For a cyclic transformation, the first law takes on a very simple form. Since the initial and final states are
the same, we have ∆E = 0, thus
L = Q. (3.22)
(First law of thermodynamics for cyclic processes)
That is, the work done by a system during a cyclic transformation is equal to the heat absorbed by the system.
23
The energy E is an extensive quantity. The energy density E/V and energy per particle E/N are intensive.
all representing the same physical observable but through different variables.
x → x′ = x + dx, (3.29)
′
y → y = y + dy (3.30)
We write this as
f ′ = f + df (3.32)
with
∂f ∂f
df = dx + dy. (3.33)
∂x ∂y
So far so good.
In thermodynamics, we face the problem that we have three functions E(P, V ), E(P, T ), E(V, T ). So if we
quantify the (physically observable!) change E → E ′ = E + dE, there are three potential differentials
∂E ∂E
(i) dE = dP + dV, (3.34)
∂P ∂V
∂E ∂E
(ii) dE = dP + dT, (3.35)
∂P ∂T
∂E ∂E
(iii) dE = dV + dT. (3.36)
∂V ∂T
The differentials dE in (i)-(iii) are all identical ; they describe the same physical change in energy dE = E ′ −E,
but simply express dE in different variables. There is an ambiguity, however, if we simply write
∂E
, (3.37)
∂T
because we do not know if this expression corresponds to the one appearing in (ii) or the one in (iii). We
will later see that these two expressions are generally different.
24
To resolve the ambiguity, we enclose the partial derivative symbol in a parenthesis and place the variable
that is to be held constant at the foot of the parenthesis. Thus
∂E
(3.38)
∂T V
means the derivative of E with respect to T while keeping V constant, when T and V are taken as the
independent variables. This is in general different from
∂E
, (3.39)
∂T P
in which case the pressure is kept constant. The proper way to write equations (i)-(iii) is then
∂E ∂E
(i) dE = dP + dV, (3.40)
∂P V ∂V P
∂E ∂E
(ii) dE = dP + dT, (3.41)
∂P T ∂T P
∂E ∂E
(iii) dE = dV + dT. (3.42)
∂V T ∂T V
dF
F ′ (x) = (x). (3.43)
dx
A fundamental theorem of calculus states that
x2
F ′ (x)dx = F (x2 ) − F (x1 ). (3.44)
x1
F2
dF = F2 − F1 . (3.46)
F1
An expression dF that satisfies such an equation is called an exact differential, and F is called a function of
state in the thermodynamic context. Examples of functions of state are V, T, P, E, because they all satisfy
V2
dV = V2 − V1 , (3.47)
V1
T2
dT = T2 − T1 , (3.48)
T1
P2
dP = P2 − P1 , (3.49)
P1
E2
dE = E2 − E1 . (3.50)
E1
25
Here, for instance, E1 and E2 are the energies in state 1 and 2. In contrast, dW and dQ are not exact
differentials and
2
W and Q are not functions of state. Instead, dW and dQ are merely small quantities and
2
the integrals 1 dW and 1 dQ depend on the process or path we choose to connect two thermodynamic
states 1 and 2. We have
2
dW ̸= W2 − W1 , (3.51)
1
2
dQ ̸= Q2 − Q1 , (3.52)
1
because the object W1 or “work in state 1” does not exist, etc. Some authors, for this reason, introduce
other symbols like δW or δQ to emphasize this difference. We will keep using dW and dQ, always keeping
in mind what we just said.
Note also that
E2
dE = E2 − E1 (3.53)
E1
implies that for a cyclic process along a closed loop C in the PV-plane, with E2 = E1 , we have
dE = 0. (3.54)
C
In contrast,
Wcycl = dW (3.55)
C
is the total work performed during one cycle. This is generally not zero, otherwise there would be no useful
engines.
dE = dW + dQ = −P dV + dQ. (3.56)
(First law of thermodynamics for infinitesimal processes)
If we choose T and V as independent variables, E changes according to
∂E ∂E
dE = dT + dV. (3.57)
∂T V ∂V T
or
∂E h ∂E i
(I) dQ = dT + + P dV. (3.59)
∂T V ∂V T
26
Similarly, taking T and P as independent variables we generally have
∂E ∂E
dE = dT + dP, (3.60)
∂T P ∂P T
∂V ∂V
dV = dT + dP (3.61)
∂T P ∂P T
which, together with the first law, yields
∂E ∂E
(II) dQ = dT + dP + P dV (3.62)
∂T P ∂P T
∂E ∂E h ∂V ∂V i
= dT + dP + P dT + dP (3.63)
∂T P ∂P T ∂T P ∂P T
h ∂E ∂V i h ∂E ∂V i
= +P dT + +P dP. (3.64)
∂T P ∂T P ∂P T ∂P T
Finally, taking V and P as independent variables we obtain
∂E h ∂E i
(III) dQ = dP + + P dV. (3.65)
∂P V ∂V P
Note, again, that all three expressions for dQ are identical, just expressed in terms of different variables.
Heat capacity. The heat capacity, C, of a body is defined as the ratio C = dQ , i.e. the infinitesimal
dT
amount of heat dQ absorbed by the body as its temperature is increased by the infinitesimal amount dT .
In experiments, of course, finite amounts ∆Q and ∆T are used to determine ∆Q/∆T . The unit of C is
kB ∝ J/K.
In general, the heat capacity of a body will be different according to whether the body is heated at constant
volume or constant pressure. We denote this by a subscript according to
∂Q
CV = , (3.66)
∂T V
∂Q
CP = . (3.67)
∂T P
At this point, equations (I) and (II) come in handy. They need to be compared with
∂Q ∂Q
(I) dQ = dT +dV, (3.68)
∂T V ∂V T
∂Q ∂Q
(II) dQ = dT + dP, (3.69)
∂T P ∂P T
which yields,
∂Q (I)
∂E
CV = = , (3.70)
∂T V ∂T V
∂Q (II) ∂E ∂V
CP = = +P . (3.71)
∂T P ∂T P ∂T P
The second term in CP accounts for the fact that when temperature is increased at constant pressure, the
system typically expands and thus performs work. Hence we need to supply additional heat to the system
to enable this expansion. In CV , this term is not present, because the volume is kept constant and so no
expansion occurs. Note that the difference between CP and CV is largest for gases, whereas liquids and
solids expand only moderately when heated.
27
Specific heat capacity. The heat capacities CV and CP are extensive quantities. To get an intensive variable,
it is common to divide by the total mass M = m0 N of the body to obtain the specific heat capacities
CV CP
cV = , cP = . (3.72)
M M
While the unit of C is J/K, the unit of c is J/(kg K). The ratio
CP cP
κ= = (3.73)
CV cV
is called heat capacity ratio. cP and cV are examples of material constants, because they encode information
about a specific substance. They are measured and tabulated for most substances. We will meet more
material constants later.
Calorimeter. A calorimeter is a perfectly thermally isolated container containing two bodies in physical
contact with each other, whose individual temperatures are monitored. The device can be used to measure
heat in general or the specific heat capacity of one body if it is known for the other body. Typically we
have a solid object, initially at temperature T1 , immersed into a caloric liquid such as water, initially at
temperature T2 . As the two bodies reach thermal equilibrium, they acquire the same temperature T3 . Let
us assume that initially T1 > T2 . Then the water is heated during equilibration from T2 to T3 . If we neglect
the expansion of water upon heating, then the water absorbs the heat
(wat)
Q = Mwat cP (T3 − T2 ), (3.74)
(wat)
where Mwat is the total mass of water and cP = 4.182 J/(gK) is the specific heat capacity of water. To
(sol)
measure the specific heat capacity cP of the solid, we neglect the volume contraction of the solid upon
cooling, so that −Q is released as heat from the solid into the water. We have
(sol)
−Q = Msol cP (T3 − T1 ), (3.75)
hence
(sol) Mwat (T3 − T2 ) (wat)
cP = c . (3.76)
Msol (T1 − T3 ) P
When expressed in terms of the variables (V, T ) or (P, T ), the energy of an ideal gas is a function of T alone,
dE = CV dT. (3.80)
This property of ideal gases can be derived from statistical mechanics, and is approximately still valid for
real gases. Here we derive this result from an experimental observation.
28
T T
A B A B
Figure 3.1: Expansion of an ideal gas does not change the temperature of the water in the calorimeter.
The observation made in 3 implies that there had been practically no transfer of heat from the caloric water
to the chambers or vice versa. It is assumed that if this experiment could be performed with an ideal gas,
there would be no change at all.
We now apply the first law to the transformation. Since we observe Q = 0 between the gas and the water,
the gas enclosed in the chambers behaves like a thermally isolated system and we have
∆E + L = 0. (3.81)
Since the total volume of the chambers does not change (no movable walls), the gas cannot perform any
work, and so L = 0. Therefore,
∆E = 0. (3.82)
The energy of the gas does not change.
During the transformation, however, the volume of the gas changed from A to A + B. Since there was
no variation in energy during the process, we conclude that a variation in volume at constant temperature
produces no variation in energy,
∂E
= 0. (3.83)
∂V T
29
In order words, the energy of an ideal gas is a function of the temperature only and not a function of the
volume. Similarly, the experiment showed that the change in pressure during the isothermal expansion had
no effect on the energy, so
∂E
= 0. (3.84)
∂P T
This completes our proof.
Implication 1. We first determine the form of the function E(T ). We make use of the additional experimental
fact that CV of a gas only slightly depends on temperature. Integrating the equation
dE = CV dT (3.85)
we obtain
E(T ) = CV T. (3.86)
We have set the integration constant E0 , which would be the energy at T = 0, to zero.
∂E ∂ N kB T
CP = +P (3.88)
∂T V ∂T P
= CV + N kB . (3.89)
3
E(T ) = N kB T for a monoatomic gas, (3.92)
2
5
E(T ) = N kB T for a diatomic gas. (3.93)
2
Expressed in terms of (P, V ) this becomes
3
E(P, V ) = P V for a monoatomic gas, (3.94)
2
5
E(P, V ) = P V for a diatomic gas. (3.95)
2
We further have
5
CP = N kB for a monoatomic gas, (3.96)
2
7
CP = N kB for a diatomic gas, (3.97)
2
30
and
5
κ= for a monoatomic gas, (3.98)
3
7
κ = for a diatomic gas. (3.99)
5
2. The system is thermally insulated so that no heat can be exchanged between the system and the
environment.
Condition 2 implies that dQ = 0 during the transformation and that the first law reads ∆E + L = 0 for
adiabatic processes.
We can expand or compress a gas adiabatically by enclosing it in a cylinder with non-heat-conducting walls
(condition 2) and a movable piston on one end, and very slowly shifting the piston inward and outward.
In this case, the gas molecules have enough time to adjust to the small gradual volume changes during the
process, and we effectively are in an equilibrium state at any time, hence the transformation is reversible
(condition 1).
Now would be a good point to recall the reversible run of the isothermal expansion.
A gas that expands adiabatically does external work, so L is positive in the first law. Hence, ∆E must be
negative, and the energy of a gas decreases during an adiabatic expansion. Since the energy is related to the
temperature through E = CV T for an ideal gas (or a similar relation for real gases), a decrease in energy
means a decrease in temperature of the gas.
CV dT + P dV = 0. (3.100)
31
T V κ−1 = const. (3.105)
(Adiabatic transformation of an ideal gas)
Using P V = N kB T we readily find that
P V κ = const., (3.106)
κ 1−κ
T P = const. (3.107)
Note that in a PV-diagram, an isothermal transformation is given by P V = const. Hence, a curve corre-
sponding to an adiabatic transformation is steeper than an isotherm in the PV-diagram due to κ > 1.
32
4 Second law of thermodynamics
4.1 Perpetuum mobile
A perpetuum mobile (Latin for ”perpetual mover”) or perpetual motion machine is an impossible machine
that operates in cycles and, once started, would run forever and do work.
The first law of thermodynamics rules out the possibility of constructing a perpetuum mobile.
For one, in a thermally isolated system, the energy does not change in a cycle, and due to the 1st law the
work done is zero. When including the environment into the setting, any machine supplied with a finite
amount of energy at a given time will eventually stop because over time energy is lost to heat due to friction.
The 1st law forbids a machine that creates energy, which it could use to compensate this friction loss or do
work.
The first law does not limit the possibility of transforming one form of energy into the other, such as trans-
forming work into heat and heat into work, as long as the total amount of energy remains constant. Indeed,
we can always transform work into heat, for instance, by using friction to turn mechanical work into an equal
amount of heat, or using electric currents to heat up a resistance element, thus transforming electrical work
into heat.
Work can, in principle, to 100% be converted into heat.
However, one experimentally observes definite limitations to the possibilities of transforming heat into work.
If this were not the case, we could construct a machine that continuously absorbs heat from its environment
and uses this heat to perform work. Since the amount of thermal energy in the soil, water, atmosphere is
practically unlimited, such a machine would effectively be operating like a perpetuum mobile. Therefore, it
is called a perpetuum mobile of the second kind.
The second law of thermodynamics rules out the possibility of constructing a perpetuum mobile of the second
kind. It implies limits on what fraction of heat can (at best) be converted into work.
Some examples of perpetuum mobiles of the second kind: Drive a ship across the ocean only by converting
heat extracted from the water into work of the propellers. Light a light bulb or heat a pot by extracting
heat from the surrounding air.
In principle, the heat bath could exchange both heat Q and work W with the system. If the heat bath is
such that it cannot perform work on the system, it is called a heat source.
A system may be coupled to several reservoirs, or to different reservoirs at different times. The Carnot cycle
uses two heat sources, with temperatures T1 (lower) and T2 (higher). To memorize the lower and higher
temperatures in the following, use that 2 > 1. It is a reversible cycle built from two isothermal lines and
two adiabtic lines, intersecting at four points A, B, C, D. We use variables (V, T ) to label these states.
33
P V
V T
Figure 4.1: Carnot cycle, schematic, in the PV-diagram (left) and VT-diagram (right)
Carnot cycle
Consider a system whose initial state is A, with volume VA at temperature T2 . The Carnot cycle is
the following set of transformations that are supposed to be performed reversibly.
A → B. Isothermal expansion to volume VB at fixed temperature T2 .
B → C. Adiabatic expansion to volume VC . During that, temperature decreases to T1 .
C → D. Isothermal compression to volume VD at fixed temperature T1 .
Example of a Carnot cycle. Consider a fluid enclosed in a cylindrical container with isolating side wall and
a movable isolating piston on the top. Heat can only be transferred through the base of the cylinder, which
we take to be heat-conducting. There are two heat sources, with temperatures T1 and T2 , such that T2 > T1 .
1. Initially the fluid is assumed to have volume VA and temperature T2 (corresponding to some pressure
PA via the equation of state).
2. Place the system on the heat source T2 . No heat transfer will occur initially.
3. Keeping the system on the heat source, we slowly (reversibly) raise the piston until the volume reaches
VB . The heat source ensures that the fluid remains at temperature T2 .
4. Remove the system from heat source T2 and place it on an insulator. We continue to slowly raise the
piston until we reach volume VC . The fluid is thermally insulated and cannot absorb heat, hence cools
to temperature T1 during that process.
5. Now place the system on heat source T1 . No heat transfer will occur initially.
6. Slowly press down the piston until we reach volume VD . The heat source ensures that the temperature
of the fluid remains T1 .
7. Remove the system from heat source T1 and place it on an insulator. Continue to slowly compress
until volume VA is reached. Since the fluid is thermally insulated and cannot release heat, it heats up
to temperature T2 during the process.
34
insulator insulator
Heat and work. During the isothermal expansion A → B, the system absorbs an amount of heat Q2 > 0
from the source T2 . Similarly, during the isothermal compression C → D, it releases an amount of heat
Q1 > 0 to the source T1 (it absorbs an amount −Q1 from source T1 ). No heat is transferred in the adiabatic
sections. Hence the total amount of heat absorbed by the system during the cycle is Q2 − Q1 . The amount
of work done L during one cycle, given by the area enclosed in the PV-diagram, follows from the 1st law for
cycles to be
L = Q2 − Q1 . (4.1)
This equation means that only a fraction of the heat absorbed from the hot source at T2 is transferred into
work L in the Carnot cycle. The part Q1 , instead of being transformed into work, is surrendered to the
source at temperature T1 . We define the efficiency of the Carnot cycle as the ratio
L
η= . (4.2)
Q2
It represents the fraction of heat that is converted into work. We have
Q1
η =1− (4.3)
Q2
for the Carnot cycle, so the portion ∝ Q1 is not transformed into work.
Since the Carnot cycle is reversible, it can be run in the opposite direction, A → D → C → B → A. It then
absorbs the work L, and absorbs the amount of heat Q1 from source the colder T1 , whereas it gives up the
amount of heat Q2 to the hotter source T2 .
35
The word ”only” is important here. For one, (i) it is possible to completely transform heat from a source
at constant temperature into work, provided some other change in the state of the system is present at the
end of the process. Furthermore, (ii) if some heat is released to another body during the process, then the
conversion of heat to work is not the only final result.
Example for (i). In the reversible run of the isothermal expansion of an ideal gas, an amount of heat Q
is supplied by the heat source to keep the gas at T . Since E(V, T ) = E(T ) for a gas, its energy does not
change, ∆E = 0. Consequently L = Q. The heat is completely transformed into work. However, this is not
the only final result of the process, because at the end of the process the gas occupies a larger volume than
at the beginning.
The fact that the initial and final state differ in this (or any similar) example, makes it impossible to use
such a transformation to form a cycle that is run several times, which would then constitute a perpetuum
mobile of the second kind.
Example for (ii). The Carnot process is a cycle, so initial and final states are the same, which extracts
heat Q2 from a source and performs work L. However, it only converts a fraction of the heat into work,
L = Q2 − Q1 , while it dumps the amount of heat Q1 into another body. Hence the conversion of heat into
work is not the only final result, but there is another final result, namely the rejection of some heat Q1 .
The coefficient η between extracted heat and performed work in L = ηQ2 is precisely the efficiency. Hence
Kelvin’s statement implies η < 1 for all cyclic processes.
The word ”completely” is, strictly speaking, superfluous and implied by ”only”, see example (ii). If the
conversion of heat into work is not complete, then there must always be some rejected heat as an additional
final result of the process.
The wording is to suggest that a positive amount of heat Q > 0 is absorbed by the system and converted
into a positive work done L > 0. If L < 0, i.e. work W > 0 is performed on the system, then the system
could convert this work via friction into negative heat Q = −W < 0 that is released into the environment.
We then have a situation where the only final result is that work is completely converted into heat. No
constraint arises in this case, because work can always be completely converted into heat.
Clausius’s statement implies that some work must be performed in order to transfer heat from a colder body
to a hotter body. This work requires an external energy source, so that the supply of energy through work is
an additional final result of the process. Refrigerators, of course, operate in this way and are not forbidden
by the 2nd law of thermodynamics.
We prove the equivalence of Kelvin’s and Clausius’s statements. We use contraposition, i.e. to prove X ⇒ Y ,
we show not-Y ⇒ not-X.
not-Kelvin ⇒ not-Clausius. Assume Kelvin’s statement is wrong. We can then perform a transformation
which extracts an amount of heat from a source ”1” at temperature T1 , and whose only final result is to
completely transform this amount of heat into work. By means of friction, we could completely transform
this work into heat again, and use this heat to heat up a another body ”2” that is at any temperature T2 .
In particular, this would also work if T2 > T1 . The only final result of this whole process would then be to
36
transfer heat from a body ”1” at temperature T1 to a body ”2” at higher temperature T2 . Hence Clausius’s
statement is wrong.
not-Clausius ⇒ not-Kelvin. Assume Clausius’s statement is wrong. We can then transfer an amount of
heat Q2 from a heat source ”1” at temperature T1 to a body ”2” at higher temperature T2 . We next use a
Carnot process to absorb the same heat Q2 from source ”2” at the high temperature (during step A → B)
and transform it into work L. During the Carnot process, some amount of heat Q1 is dumped into ”1” at
the lower temperature. Since body ”2” absorbed and released the same amount of heat, there is no overall
change to body ”2” in the process. The net amount of heat extracted from ”1” is ∆Q = Q2 − Q1 , which
is completely transformed into work L = ∆Q. Hence we constructed a process which extracts an amount
of heat ∆Q from the source ”1” at constant temperature and transformed it completely into work. Thus
Kelvin’s statement is wrong.
Note that we have only shown the equivalence of both formulations of the 2nd law, we did not ”prove” it.
Its experimental justification comes from the fact that no-one ever accomplished to construct a perpetuum
mobile of the second kind, or observed heat to flow from a cold to a hot body by itself.
Crucially, the underlying cycle need not be reversible or a Carnot cycle. An engine using a reversible cycle
will be called reversible engine.
Note that η = L/Q2 is the fraction of heat converted into work. The closer η is to 100%, the better for
applications. Kelvin’s statement implies η < 1. In the following two sections, we show that the efficiency of
any reversible engine operating between T2 and T1 is identical, and satisfies
T1
ηrev = 1 − . (4.5)
T2
Typically the body at lower temperature, into which the amount of heat Q1 > 0 is dumped and lost, is the
environment of the engine, thus we usually cannot control T1 . In practice, it would be difficult to run an
engine with very low T1 → 0, since we continuously dump heat into ”1”. To get a large efficiency, we want
T2 to be as high as possible. Of course, any actual efficiency will generally be lower than ηrev , because all
actual heat engines are far from being reversible. In fact, we will show that
η ≤ ηrev (4.6)
in general.
We first show that if L > 0, then we have Q2 > 0 and Q1 > 0. Assume first that Q1 ≤ 0. This means the
engine absorbs the positive amount of heat |Q1 | = −Q1 from heat source ”1”. We could then use the hotter
source ”2” to heat up source ”1” to exactly compensate for the lost amount of heat |Q1 |. Indeed, this would
amount to heat flowing from a hot to a cold body, which is not forbidden by the 2nd law. During one cycle,
there would then be no change to source ”1”. Furthermore, the amount of heat effectively absorbed from
source ”2” by the engine is Q2 + |Q1 |, which is fully converted into work L = Q2 − Q1 = Q2 + |Q1 |. Since
37
Figure 4.3: Assuming Q1 ≤ 0 in heat engines leads to a contradiction.
this is the only result of the cycle, this is in contradiction to Kelvin’s formulation of the 2nd law. Hence
Q1 > 0. Since L > 0, we immediately get Q2 > 0. □
In the following, we consider two engines that operate between the same temperatures T1 and T2 . The
first engine shall be characterized by (L, Q1 , Q2 ), the second engine by (L′ , Q′1 , Q′2 ). We show the following
fundamental theorem:
A. If the first engine is reversible, then
Q2 Q′
≥ 2′ . (4.7)
Q1 Q1
Q2 Q′
= 2′ . (4.8)
Q1 Q1
Note that B implies that ηrev is identical for all reversible engines operating between T1 and T2 , whereas A
implies η ≤ ηrev . The second engine in A need not be reversible.
Statement A implies B, because if A is true and if the second engine is also reversible, then we obtain the
Q′2
relation with both ”≤” and ”≥, hence QQ1 = Q′ in this case.
2
1
L = Q2 − Q1 , (4.9)
′
L = Q′2 − Q′1 . (4.10)
Q2 N′
′ = , (∗) (4.11)
Q2 N
where N, N ′ are positive integers. Now consider a combined cycle that consists of N ′ runs of the second
engine and N runs of the reversed first engine. (This is permissible, since the first engine is reversible.)
When operated in the reverse, the first engine absorbs an amount of work L, gives up the amount of heat
38
Q2 to source ”2”, and receives the amount of heat Q1 from source ”1”. The total amount of work done by
the combined engine is
Ltotal = N ′ L′ − N L, (4.12)
the total amount of heat received from source ”2” is
Q2,total = N ′ Q′2 − N Q2 , (4.13)
and the total amount of heat given to source ”1” is
Q1,total = N ′ Q′1 − N Q1 . (4.14)
This yields
Ltotal = Q2,total − Q1,total . (4.15)
However, equation (∗) implies Q2,total = 0 and so
Ltotal = −Q1,total . (4.16)
Importantly, at this point we do not know the sign of Ltotal (which is the sign of −Q1,total ). The combined
cycle is constructed such that there is no change to source ”2” after the whole cycle. If Ltotal > 0, then the
only final result of the combined cycle is that the heat absorbed from source ”1” at constant temperature
T1 (which is −Q1,total ) is converted into the work Ltotal . This contradicts Kelvin’s statement. However, if
Ltotal ≤ 0, then work is performed on the engine, and the engine converts this work completely into the
amount of heat |Q1,total | that is dumped into source ”1”. Nothing limits the conversion of work into heat.
Hence
Ltotal ≤ 0 (4.17)
and
Q
2 ′
0 ≤ Q1,total = N ′ Q′1 − N Q1 = N Q − Q1 . (4.18)
Q′2 1
Hence
Q2 Q′
≥ 2′ , (4.19)
Q1 Q1
which proves A. □
A refrigerator is a heat engine run in the reverse direction, which extracts an amount of heat Q1 from a
source at low temperature T1 by absorbing work W > 0, and dumps an amount of heat Q2 to a source
at higher temperature T2 . (If the emphasis is on heating the hotter source, it is equivalently called a heat
pump.) The work W > 0 is required by Clausius’ formulation of the 2nd law. The coefficient of performance
is given by
Q1
COP = > 1. (4.20)
W
For a reversible refrigerator (such as a reverse Carnot cycle) we have
T1
COPrev = (4.21)
T2 − T1
For instance, for T2 = 20◦ C we have
T1 = 3◦ C ⇒ COPrev = 16, (4.22)
◦
T1 = −18 C (freezer) ⇒ COPrev = 6.7, (4.23)
whereas real household refrigerators have COP ∼ 2.
39
4.5 Zeroth law of thermodynamics
To formally define the concept of temperature, without resorting to kinetic theory or statistical mechanics,
one may use the zeroth law of thermodynamics. It is the following postulate that is confirmed by experiment:
When two thermodynamic systems are each in thermal equilibrium with a third system, then they are in
thermal equilibrium with each other.
Note the word ”thermal” here. This postulate implies that there is a property of thermodynamic systems
which is common to all systems in thermal equilibrium. This quantity serves as an empirical temperature
T̃ .
Let us call the systems A, B, C. For concreteness we may think of C as a thermometer. The volume and
pressure of each system are (VA , PA ), (VB , PB ) and (VC , PC ).
1. We assume that A and C are in thermal contact through a heat-conducting wall. For general values
of (VA , PA , VB , PB ), they would not be in thermal equilibrium. Instead, in thermal equilibrium there
will be a constraint of the form
f1 (VA , PA , VC , PC ) = 0 (4.24)
relating the four quantities through some function f1 . We want to show that this relation is actually
of the form
We could then define T̃ as temperature read off from the thermometer C. Furthermore, the relation
T̃ = T̃A (V, P ) constitutes the equation of state for system A, similarly for C.
2. To show Eq. (4.25), we also consider B and C, whose thermal equilibrium implies a relation of the
form
f2 (VB , PB , VC , PC ) = 0 (4.26)
f3 (VA , PA , VB , PB ) = 0 (4.27)
such that
(The functions f1,2 (◦, ◦, ◦, ◦) have four arguments, fˆ1,2 (◦, ◦, ◦) have only three arguments.) We can
then solve these two equations for T̃C to obtain both
40
and
where we introduced two new functions T̃A,B . This yields the desired formula
T̃A (VA , PA ) = T̃B (VB , PB ) ⇒ f3 (VA , PA , VB , PB ) := T̃A (VA , PA ) − T̃B (VB , PB ) = 0 (4.33)
4. The three systems in thermal equilibrium all have the same value of
This defines both the empiric temperature and the equations of state for systems A, B, C.
Comment 2. The derivation remains valid if the systems depend on more than the two variables V, P , say,
N1 , N2 , . . . for mixtures. The equation of state obtained would be of the form
Note that the particle number N that we tend to ignore falls into this category.
Comment 3. The zeroth law is proven fairly easily within the framework of statistical mechanics, where T
is defined through T −1 = ∂E
∂S
|V,N . Thermal equilibrium of A and C implies TA = TC and that of B with C
implies TB = TC . This implies TA = TB .
In this section, we show that the findings on heat engines imply the existence of an absolute thermody-
namic temperature scale, denoted ϑ, that is intrinsic to thermodynamics and thus independent of the special
properties of any thermodynamic substance. All thermodynamic laws take on a simple form when this tem-
perature scale is used. We will show that it coincides, up to an arbitrary constant factor a, with the ideal
gas temperature scale, i.e.
ϑ = aT. (4.39)
41
We then choose a = 1 by convention and arrive at
ϑ = T. (4.40)
We start from statement B that for all reversible engines operating between (empirical) temperatures T̃1 and
T̃2 , the ratio Q2 /Q1 is identical. Thus
Q
2
= f (T̃2 , T̃1 ) (4.41)
Q1 rev
for one cycle, with some fundamental function f (◦, ◦). Choose an arbitrary small temperature T̃0 such that
and three corresponding heat sources ”0”, ”1”, and ”2”. We consider two more reversible engines, En1 and
En2 . During one cycle, engine En1 extracts the amount of heat Q1 from ”1” and releases the amount of heat
Q0 into ”0”. We have
Q
1
= f (T̃1 , T̃0 ). (4.43)
Q0 rev
In contrast, engine En2 extracts the amount of heat Q2 from ”2” and releases Q0 into ”0”. We have
Q
2
= f (T̃2 , T̃0 ). (4.44)
Q0 rev
Taking the ratio of these two equations we conclude
f (T̃2 , T̃0 )
f (T̃2 , T̃1 ) = . (4.45)
f (T̃1 , T̃0 )
(More accurately, we would have to consider the combined process of running one cycle of engine En2
followed by one cycle of the reversed engine En1 . The outcome is the same.) Since T̃0 was arbitrary, we may
keep it constant in all our equations. Hence we regard f (T̃ , T̃0 ) as a function of T̃ alone. We define a new
temperature scale
Q ϑ(T̃2 ) ϑ2
2
= = . (4.47)
Q1 rev ϑ(T̃1 ) ϑ1
for all reversible cycles. The temperature scale defined by the variable ϑ is called the absolute thermodynamic
temperature scale. We are free to choose the unit of this scale, since only ratios appear in this equation. The
usual choice is made by setting the difference between the boiling temperature and freezing temperature of
water at atmospheric pressure to be 100 units (#).
We finally show that ϑ = aT by computing Q2 /Q1 for the Carnot process of an ideal gas, which is a reversible
process. The outcome is
Q T2
2
= ⇒ ϑ = aT, (4.48)
Q1 Carnot T1
with an arbitrary scale factor a. The way we have chosen the units for ϑ in (#), however, implies a = 1.
42
For the computation of Q2 /Q1 for the Carnot process we consider an ideal gas with heat capacity ratio κ.
The work done along the isothermal lines is
V
B
LA→B = N kB T2 ln > 0, (4.49)
VA
V
C
LC→D = −N kB T1 ln < 0. (4.50)
VD
No heat is absorbed along the adiabatic lines and we have
Now now use that the operating substance is an ideal gas and so the energy E(V, T ) = E(T ) only depends
on the temperature. Thus points along isotherms have the same energy,
EA = EB , (4.53)
EC = ED . (4.54)
This implies that work equals heat along the isothermal lines,
V
B
Q2 = LA→B = N kB T2 ln > 0, (4.55)
VA
V
C
Q1 = −LC→D = N kB T1 ln > 0. (4.56)
VD
To relate the volumes, we use that along the adiabatic lines we have T V κ−1 = const so that
T2 V κ−1 V κ−1 VC VB
C D
= = ⇒ = . (4.57)
T1 VB VA VD VA
For the heat ratio we eventually obtain
Q2 T2 ln( VVB
A
) T2
= V
= . (4.58)
Q1 T1 ln( VD )
C T1
We prove that
n
X Qi
≤ 0. (4.59)
i=1
Ti
(Clausius inequality)
43
Figure 4.4: Combined cycle used in the proof of Clausius’s inequality
Example. Let us consider the case of heat engines operating between two sources with T1 and T2 , i.e. n = 2.
We have shown that
T2 Q
2 Q2
= ≥ . (4.60)
T1 |Q1 | rev |Q1 |
Note that Q1 = −|Q1 | < 0 with the sign convention used here. This implies
T2 |Q1 | ≥ T1 Q2 ⇔ 0 ≥ T1 Q2 − T2 |Q1 | = T1 Q2 + T2 Q1 . (4.61)
Diving by T1 T2 we obtain
Q1 Q2
+ ≤0 (4.62)
T1 T2
as claimed. The same derivation also shows that equality holds for reversible cycles.
Proof. Consider an additional heat source ”0” at an arbitrary temperature T0 . Between the source at T0 and
each of the n heat sources at Ti we insert a reversible Carnot engine Ci . The Carnot engine Ci is such that if
the system receives (surrenders) the amount of heat |Qi | to the source ”i”, then Ci surrenders (receives) the
amount of heat |Qi | to the source ”i”. So if Qi is the heat absorbed by the system at Ti , then −Qi is the
heat absorbed by the Carnot engines at Ti . After running the combined cycle of S followed by C1 , . . . , Cn ,
the n heat sources remain unchanged.
Each Carnot engine Ci , schematically, is a heat engine that receives a positive amount of heat Q(a) > 0 at
T (a) and releases a positive amount of heat Q(b) > 0 at T (b) , with
Q(a) T (a)
= . (4.63)
Q(b) T (b)
44
In this case, |Qi | at Ti is one of the positive numbers. The other, |Qi,0 |, is the amount of heat exchanged
between Ci and the heat source ”0”. Thus
T0
|Qi,0 | = |Qi |. (4.64)
Ti
Denote by Qi,0 the amount of heat absorbed by Ci from ”0”. If Qi > 0 (system receives heat at Ti and Ci
releases heat at Ti ), then Qi,0 > 0. If Qi < 0 (system releases heat at Ti and Ci receives heat at Ti ), then
Qi,0 < 0. Thus Qi and Qi,0 have the same sign,
T0
Qi,0 = Qi . (4.65)
Ti
Now we run the combined cycle of S followed by C1 , . . . , Cn . The source at T0 surrenders a total amount of
heat
n n
X X Qi
Q0 = Qi,0 = T0 (4.66)
i=1
T
i=1 i
to the Carnot engines. Since the total system consisting of S, C1 , . . . , Cn returns to its initial state, the 1st
law for cyclic processes applies, and the amount of heat Q0 is converted into work done given by L = Q0 .
Since this is the only final result of the cycle (the heat sources at T1 , . . . , Tn remain unchanged), we conclude
that L ≤ 0, since otherwise Kelvin’s statement would be violated. Hence Q0 ≤ 0, or
n
X Qi
≤ 0. (4.67)
i=1
Ti
If the cycle is performed reversible, we can run it in the reversed direction, which means that all Qi change
sign (but Q0 does not). We then have
n
X Qi
− ≤ 0, (4.68)
i=1
Ti rev
which together with the Clausius inequality implies
n
X Q i
= 0. (4.69)
i=1
Ti rev
Comment. Note that T in the denominator is the temperature of the source. This coincides with the system
temperature for reversible cycles, where system and source are in equilibrium at each point. For irreversible
processes, however, heat can only flow from hot to cold bodies, so the source is hotter for dQ > 0 or colder
for dQ < 0.
45
4.8 Entropy
The Clausius (in)equality for reversible cycles reads
dQ
= 0, (4.72)
C T
where C is a reversible transformation, for instance a closed loop in the PV-diagram. Consider now two
equilibrium states A and B of a system and some reversible transformation Prev that takes the system from
A to B. The value of the integral
dQ
IA→B = (4.73)
Prev T
′
is independent of the reversible path chosen. Indeed, if Prev is another reversible path connecting A and B,
′−1
then we denote by Prev the reversed path connecting B and A. We have
′ dQ
IA→B = (4.74)
′
Prev T
and
′ dQ dQ ′
IB→A = =− = −IA→B . (4.75)
′−1
Prev T ′
Prev T
(This is true for any line-integral: Going along the path in one or the other direction gives an overall minus
′−1
sign.) Consider now the combined reversible transformation C = Prev ∪ Prev . This is a reversible cycle and
dQ
0=
T
C
dQ dQ
= +
Prev T Prev T
′−1
′
= IA→B + IB→A
′
= IA→B − IA→B . (4.76)
This proof our claim.
Entropy. The property just shown allows us to define a new function of state of the system, the entropy. It
is one of the most important quantities in thermodynamics and, in fact, most areas of physics. We choose
an arbitrary equilibrium state O of our system. For any other equilibrium state A we define the entropy of
state A as
A
dQ
SA = , (4.77)
O T
where the path connecting O and A is taken to be reversible. We have seen that the value of SA is indepen-
dent of the reversible path chosen. Since the path is reversible, its initial and final points must be equilibrium
states. Consequently, SA is only defined for equilibrium states A. If we choose another equilibrium state B,
then
B
dQ
SB = (4.78)
O T
along some reversible path connecting O and B. Since we have
A O
dQ dQ
=− , (4.79)
O T A T
46
this implies that the entropy difference between states A and B is given by
B A O B B
dQ dQ dQ dQ dQ
SB − SA = − = + = , (4.80)
O T O T A T O T A T
where any reversible path connecting A and B can be chosen. For an infinitesimal reversible transformation
we have
dQ = T dS (reversible). (4.81)
B
In the context of differentials introduced in Section 3.3, the heat dQ is not an exact differential and A
dQ
depends on the path. In contrast, the entropy
1
dS = dQ (reversible) (4.82)
T
is an exact differential and
B
dS = SB − SA (4.83)
A
is independent of the reversible path chosen to connect A and B. Similarly, the work dL is not an exact
B
differential and A dL depends on the path, but the volume
1
dV = dL (reversible) (4.84)
P
is an exact differential and
B
dV = VB − VA (4.85)
A
is independent of the reversible path chosen to connect A and B. One might say that the existence of the
function of state ”entropy” restores a symmetry between work and heat in these equations.
J
Comment 1. The natural unit of entropy is kB , because Q ∝ J, T ∝ K, so S ∝ K ∝ kB .
Some properties of the entropy. We show the following important properties of the entropy:
47
(1) The entropy of independent subsystems is additive. This implies that entropy is extensive.
(2) For any thermodynamic transformation connecting equilibrium states A and B we have
B
dQ
∆S = SB − SA ≥ , (4.86)
A T
SB ≥ SA (4.87)
thus
B
dQ dQ
SB − SA ≥ = (4.95)
P T A T
as claimed.
For (3), consider a thermally isolated system. Since it cannot exchange heat with its environment we have
48
dQ = 0 and thus SB − SA ≥ 0 as claimed.
For (4), use that ”=” applies in (3) for reversible transformations.
For (5), consider a thermally isolated system subject to some external constraints such as fixed volume
V . The system may be in one of many (possibly inhomogeneous) thermodynamic states that satisfy these
constraints.2 If the system is not in equilibrium, due to the microscopic motion of atoms or molecules, it
will, over time, undergo spontaneous thermodynamic processes that slightly change its thermodynamic state.
Any such transformation, however, needs to satisfy ∆S ≥ 0 according to (3). Consequently, if we wait long
enough for the time-independent equilibrium state to set in, then this state must be a state of maximum
entropy.
Comment 1. The entropy of a system can decrease in a thermodynamic transformation. This requires
that the system is not thermally isolated from its environment, which is often the case. However, if the
environment and the system together are thermally isolated from the rest of the universe, then the entropy
of the combined system cannot decrease.
Comment 2. From (4) and (5) it follows that the approach to equilibrium is an irreversible process. This is
intuitively clear.
Example 1. Reversible isothermal expansion of an ideal gas. Assume the volume increases from VA to VB .
Since T does not change and E = E(T ), we have ∆E = 0 and
VB
Q = L = N kB T ln . (4.96)
VA
Consequently, for the system,
B B
dQ 1 Q VB
∆Ssys = = dQ = = N kB ln . (4.97)
A T T A T VA
Example 2. Free isothermal expansion of an ideal gas. This is an irreversible process. As we have seen in
the section on heat capacity and calorimeters, no heat is supplied
by the environment and no work is done,
Q = W = 0. We may naively assume that Q = 0 implies ∆S ∼ dQ T = 0, but this is wrong. Indeed, we can
only conclude that
B
dQ
∆Ssys ≥ = 0, (4.100)
A T
2 The typical example is a gas that occupies only one half-space of a room: a clear non-equilibrium state, although consistent
49
T
S
Figure 4.5: Carnot cycle in the TS-diagram
which does not rule out ∆S ̸= 0. In fact, since entropy is a function of state and the initial and final states,
A and B, are the same equilibrium states for both the reversible and free expansion, we have
VB
∆Ssys = SB − SA = N kB ln (4.101)
VA
in both cases. The free expansion is thus an example of a process with dQ = 0, but dS ̸= 0. This is no
contradiction since we only have dQ = T dS for reversible processes. For the free expansion, however, we
have
∆Senv = 0 (4.102)
Carnot cycle in the TS-diagram. It is instructive to study the reversible Carnot cycle in the TS-diagram,
instead of the PV-diagram. During the adiabatic transformations B → C and D → A, we have dQ = 0,
and thus dS = 0. These correspond to vertical lines in the TS-diagram. The isothermal lines A → B and
C → D are horizontal lines. The net amount of heat extracted from the environment, due to dQ = T dS in
reversible processes, is given by the area under the curve. We have
Furthermore,
Q2 = T2 (S2 − S1 ), (4.105)
so that
L T1
η= =1− . (4.106)
Q2 T2
50
Differentials of energy and entropy. Consider a state in thermodynamic equilibrium, represented, for in-
stance, by a point in the PV-diagram. Choose any infinitesimally close equilibrium state and a reversible
transformation connecting them. The first law of thermodynamics then reads
dE = dQ + dW = T dS − P dV. (4.107)
Equivalently,
1 P
dS = dE + dV, (4.109)
T T
hence
1 ∂S P ∂S
= , = . (4.110)
T ∂E V T ∂V E
These relations are true for any thermodynamic equilibrium state.
51
5 Thermodynamics potentials
5.1 Free energy
In this section, we consider a system that is in thermal contact with a heat bath (environment) of constant
temperature T . Work W may be performed on the system. We also assume for simplicity that the system
is homogeneous.
Thermal equilibrium. Assume the system with energy E and entropy S is in thermal equilibrium with the
heat bath at T , so that the system also has temperature T . We define the free energy by
F = E − T S. (5.1)
The free energy is a function of state, since E, T, S are functions of state. Consider now a thermodynamic
transformation from state A to state B and define
∆F = FB − FA . (5.2)
We show that:
The work done during the transformation satisfies
L ≤ −∆F, (5.3)
Equality holds for reversible transformations. In words: For a system in thermal equilibrium with a heat
bath at constant temperature, the work done during a reversible transformation is equal to the decrease in
free energy of the system.
For the proof, recall that
B
dQ
≤ SB − SA , (5.5)
A T
where equality holds for reversible transformations. Since the temperature of the bath T is constant during
the transformation, we have
B
Q= dQ ≤ T (SB − SA ). (5.6)
A
From the first law of thermodynamics we have
L = −∆E + Q (5.7)
≤ EA − EB + T (SB − SA ) (5.8)
= [EA − T SA ] − [EB − T SB ] (5.9)
= FA − FB (5.10)
= −∆F. □ (5.11)
Thermal contact only. If the system is not in thermal equilibrium with the bath, but rather in a state A of
temperature TA , then the free energy of the system is
FA = E − TA S, (5.12)
52
or, more precise,
FA = EA − TA SA . (5.13)
For any thermodynamic transformation from A to B (with temperatures TA and TB ) when coupled to a
heat bath at temperature T , the above derivation yields
Consider now a system in thermal contact with a heat bath at constant temperature T such that the
environment cannot perform work on the system,
W = 0. (5.15)
For homogeneous systems, this reduces to constant volume, V . For any thermodynamic transformation
between states A and B of the system we have
FB ≤ FA , (5.16)
or,
∆F ≤ 0. (5.17)
Consequently, the free energy cannot increase during any transformation and the equilibrium state is the
state of minimum free energy. The value of the equilibrium free energy, Feq , is fully determined by the
thermodynamic variables of the system, here T and V , so that
Digression: Product rule for differentials. Consider two functions f (x) and g(x) of one variable. The product
rule of differentiation states that
d h i
f (x)g(x) = f ′ (x)g(x) + f (x)g ′ (x) (5.19)
dx
df dg
= g+f . (5.20)
dx dx
Multiplying by dx we obtain
dE = T dS − P dV. (5.22)
53
Consequently,
dF = d(E − T S) (5.23)
= dE − T dS − SdT (5.24)
= [T dS − P dV ] − T dS − SdT (5.25)
= −SdT − P dV. (5.26)
Analogous to our analysis of dE and dS, we conclude that for any thermodynamic equilibrium state we have
which implies
∂F ∂F
S=− , P =− . (5.28)
∂T V ∂V T
∂E
µ= . (5.29)
∂N S,V
dE = T dS − P dV + µdN. (5.30)
This replaces the infinitesimal form of the 1st law of thermodynamics if N is not constant during a thermo-
dynamic transformation. Since E and S are extensive, we find that µ is intensive.
We have the following three sets of extensive and intensive thermodynamic variables:
54
category extensive intensive
thermal S T
mechanical V P
chemical N µ
They always show up in these pairs in thermodynamic differentials. For instance, for the entropy S and free
energy F = E − T S we have
1 P µ
dS = dE + dV − dN, (5.31)
T T T
dF = −SdT − P dV + µdN. (5.32)
Exercise: Show these starting from dE.
Interpretation. What does the chemical potential mean or measure? The definition of µ implies that it is
measured in units of energy, say, Joule. It amounts to the energy costs ∆E when adding one particle to
the system at constant S and V , or the free energy cost ∆F when adding a particle at constant T and V .
Another perhaps more useful interpretation is to think of µ simply as some parameter which is such that if
µ increases, the particle number N increases. Similarly, the temperature T is some parameter such that if
T increases, the energy increases.
If two systems “1” and “2” are in thermal and mechanical equilibrium, then their temperatures T1,2 and
pressures P1,2 satisfy
Similarly, if they are in chemical equilibrium, their chemical potentials µ1,2 satisfy
E = E1 + E2 , (5.38)
V = V1 + V2 , (5.39)
N = N1 + N2 . (5.40)
55
In equilibrium, the numbers E1 , V1 , and N1 will be such that the entropy of the combined system is maximal.
(The numbers E2 = E − E1 , V2 = V − V1 , N2 = N − N1 are then automatically determined by the values
of E, V, N .) We denote these values by the subscript “eq” for equilibrium. This means that the function
S(E, V, N ) is maximal with respect to the free, independent parameters E1 , V1 , N1 . The first derivatives
vanish at the maximum, hence
∂S
! ∂S1 ∂S2 1 1
0= = − = − , (5.43)
∂E1 eq ∂E1 eq ∂E2 eq T1 T2 eq
! ∂S ∂S1 ∂S2
P P2
1
0= = − = − , (5.44)
∂V1 eq ∂V1 eq ∂V2 eq T1 T2 eq
! ∂S ∂S1 ∂S2 µ1 µ2
0= = − =− − . (5.45)
∂N1 eq ∂N1 eq ∂N2 eq T1 T2 eq
This completes the proof. □
Gibbs–Duhem relation. Consider a homogeneous system with entropy S(E, N, V ). If we double the system,
then we double the energy, volume, and particle number. Since entropy is extensive, we have
This can be seen as a special case of S(E, V, N ) = S1 (E1 , V1 , N1 ) + S2 (E2 , V2 , N2 ) for “1”=“2”. More
generally, resizing the system by a scale factor λ we have
E = T S − P V + µN (5.53)
Gibbs–Duhem relation. (5.54)
This is a very useful relation. It is true for every homogeneous thermodynamic system. By taking a differ-
ential of this relation we obtain
hence
56
S N
dP = dT + dµ (5.57)
V V
= sdT + ndµ. (5.58)
(Here s = S/V is the entropy density.) This equation is very important: It tells us that the three intensive
variables T, P, µ are not independent, meaning they cannot be varied independently of one another. This
thermodynamic constraint is true for every homogeneous system and independent of the equation of state.
23 − 1 = 7. (5.59)
However, only five of them are used in practice. Here is an overview of them:
The potential F (T, V, N ) is also called Helmholtz free energy, the potential G(T, P, N ) is also called Gibbs
free energy3 . The name “grand potential” of Φ(T, V, µ) derives from the fact that it naturally appears
in the so-called grand-canonical ensemble in statistical mechanics. Obviously, we did not use the variable
combinations (S, V, µ) and (S, P, µ).4
3 In Fermi’s book, G is called Φ.
4 The relations between the thermodynamic potentials are so-called Legendre transformations, see my lecture notes for PHYS
530, section 3.2.
57
The differentials of the thermodynamic potentials (valid for reversible transformations) read
dE = T dS − P dV + µdN, (5.60)
dF = −SdT − P dV + µdN, (5.61)
dH = T dS + V dP + µdN, (5.62)
dG = −SdT + V dP + µdN, (5.63)
dΦ = −SdT − P dV − N dµ. (5.64)
For instance,
dH = d(E + P V ) (5.65)
= dE + P dV + V dP (5.66)
= [T dS − P dV + µdN ] + P dV + V dP (5.67)
= T dS + V dP + µdN. (5.68)
Exercise: Show the remaining ones.
The enthalpy H(S, P, N ) and free enthalpy G(T, P, N ) find applications in chemistry, because chemical
reactions happening in a test tube are mostly performed at constant atmospheric pressure, not at fixed
volume.
Equation of state. Specifying the equation of state of a system is equivalent to specifying its corresponding
thermodynamic potential in the right variables, because all other thermodynamic parameters follow from
derivatives of the thermodynamic potential.
Example 1. Consider a system with thermal and chemical contact to a bath, that is mechanically isolated.
The equation of state is
Φ(T, V, µ) = −V P (T, µ). (5.75)
58
Indeed, we obtain all other thermodynamic functions from
∂Φ ∂Φ ∂Φ
S=− , P =− , N =− (5.76)
∂T V,µ ∂V T,µ ∂µ T,V
and
E = T S − P V + µN. (5.77)
To obtain E, we could also integrate dE = T dS − P dV + µdN , since we know everything on the right-hand
side.
Example 2. Consider a system in thermal contact with a bath, that is mechanically and chemically isolated.
The equation of state is
F (T, V, N ). (5.78)
E = T S − P V + µN. (5.80)
(1)
∂a ∂b
= (5.82)
∂y ∂x
and, hence,
∂R ∂R
a= , b= . (5.84)
∂x y ∂y x
59
simply states that second derivatives can be exchanged,
∂2R ∂2R
= . (5.86)
∂x∂y ∂y∂x
Property (2) is satisfied for all thermodynamic potentials, R = E, F, . . . , when considering reversible trans-
formations. The variables x, y can be chosen to be any of the three variables of the potential while keeping
the third one fixed. The equation (1) then yields quite nontrivial thermodynamic identities, called Maxwell
relations.
Example. Compressing a metal block. A typical application of Maxwell relations is to convert non-intuitive
thermodynamic derivatives into so-called material constants, yielding useful formulas for experiment and
engineering. As an example, consider a block of iron. Its specific properties are characterized by material
constants such as heat capacities,
∂S
CV = T , (5.92)
∂T V,N
∂S
CP = T , (5.93)
∂T P,N
isothermal and adiabatic compressiblities,
1 ∂V
κT = − , (5.94)
V ∂P T,N
1 ∂V
κS = − , (5.95)
V ∂P S,N
60
thermal expansion coefficient
1 ∂V
α= , (5.96)
V ∂T P,N
or thermal pressure coefficient
1 ∂P
β= . (5.97)
P ∂T V,N
(Careful: different authors use different labels.) These quantities can be measured and are tabulated for
many materials. Assume now we compress the block of iron reversibly and at constant temperature from a
pressure P1 to a higher pressure P2 > P1 . How much heat is released during this process? We have
dQ = T dS, (5.98)
We next use a Maxwell relation to rewrite the derivative. Since the variables T, P, N are involved, we might
want to look at
dQ = −T V αdP (5.103)
and
P2
Q = −T α V dP. (5.104)
P1
Since the volume of the metal block will not change much when compressed, we can assume V ≈ const and
so arrive at
Q = −T V α(P2 − P1 ). (5.105)
Comment. The material constants are not all independent. For instance, using thermodynamic (but not
Maxwell) relations, one can show that
CV κS
= (5.106)
CP κT
for any substance.
61
5.5 Phase transitions
Definition of a phase. When a system consists of only a single homogeneous substance, it is said to consist
of only one phase. If a heterogeneous system is composed of several parts, each of which is homogeneous in
itself, the system is said to consist of several phases.
Comment. The important property of each part that comprises a phase is the spatial homogeneity. They
can by themselves be mixtures of several chemical compounds, see the examples.
The thermodynamic properties of different phases are often quite distinctive, which allows us to tell them
apart macroscopically. For instance, n is much higher for solids than for liquids or gases, or whereas cP ≈ cV
for solids and liquids, we have cP > cV for gases.
Assume a system is in one phase. If this system changes its phase through an external change of thermody-
namic variables, it is said to undergo a phase transition. Some common phase transitions are:
liquid
va
po
g
co
tin
riz
nd
g
el
at
in
m
en
io
ez
n
sa
fre
tio
n
sublimation
solid gas
deposition
If a system consists of two or more phases, we say that these phases coexist. The coexistence of phases is
not generic in equilibrium, but rather requires special thermodynamic parameters; most systems are in one
phase. In fact, at a phase transition point between two phases 1 and 2, there is a coexistence of phases 1
and 2. A division of the PT-plane (or any other choice of variables) into regions of mutually distinct phases
and their coexistence regions is called the phase diagram of the system.
62
liquid
solid
gas
Figure 5.1: Left. Phase diagram of water: AB—coexistence curve of liquid and gas phases, AC—coexistence
curve of solid and gas phase, AD—coexistence curve of solid and liquid phase, A—triple point, f—freezing
point at P = 1 atm, b—boiling point at P = 1 atm. The slope of the melting curve AD represents the
anomaly of water ( dP
dT |melt < 0). Right. Phase diagram of a generic substance (without anomaly in the
melting curve).
Latent heat. Consider a homogeneous system of water in its solid phase. We add heat Q to the system to
increase its temperature at constant pressure P = 1 atm. The system undergoes a melting and vaporization
transition at temperatures Tf (P ) = 0◦ C and Tv (P ) = 100◦ C. The diagram of temperature vs. heat supplied,
looks like this:
T vapor
liquid
latent heat
ice of vaporization
latent heat
of melting
heat supplied Q
63
When the system turns from a solid to a liquid, the temperature remains constant and the heat supplied
is used within the substance to break some chemical bonds and modify its internal structure. This costs
energy that cannot be used to increase temperature. The amount of heat Qlat which does not lead to a
temperature increase is called latent heat. Similarly, when the system turns from a liquid to a gas, latent
heat of vaporization needs to be supplied.
First and second order transitions. A phase transition with nonzero latent heat is called a first-order phase
transition. Whereas the thermodynamic potentials such as F (T, V, N ) or G(T, P, N ) remain continuous at
the transition or “critical” temperature T0 , the entropy S = −( ∂F ∂G
∂T )V,N = −( ∂T )P,N jumps by an amount
Qlat
∆S = (5.109)
T0
upon increasing temperature through the transition. A phase transition where the entropy remains contin-
uous, and hence there is no latent heat, is called a second-order phase transition. Crucially, we see that
derivatives of thermodynamic potentials are not always continuous at a phase transition—they may have
kinks and jumps.
In water, all phase transitions are of first order, except for the so-called critical point, which is the end-point
of the condensation curve AB located at
Tc = 374◦ C, (5.110)
Pc = 218 atm. (5.111)
For P > Pc , pressures are so high that liquid water and vapor are no longer distinct phases. Exactly at
the critical point, water and vapor coexist and show the phenomenon of critical opalescence, which is an
enhanced scattering of light that makes the substance appear milky. This is an example of a larger class of
so-called critical phenomena, which universally appear at second-order phase transitions.
64
6 Kinetic theory of gases
6.1 Brief review of classical mechanics
This section is to review some notion of classical mechanics and fix our notation.
Notation. We consider systems of particles in three dimensions. Vector quantities, such as momentum and
position, will be indicated by an arrow,
p1 px x1 x
p⃗ = p2 = py , ⃗x = x2 = y . (6.1)
p3 pz x3 z
The set of all values of p⃗ and ⃗x is called momentum and position space, respectively. The three components
of a vector are labelled by indices i, j, · · · ∈ {1, 2, 3}, i.e. pi and xi . We denote
q
p := |⃗ p| = p2x + p2y + p2z , (6.2)
p
r := |⃗x| = x2 + y 2 + z 2 . (6.3)
⃗a · ⃗b = a1 b1 + a2 b2 + a3 b3 . (6.4)
If ⃗a = ⃗b, we write
65
where we introduce the volume element
d3 x := dxdydz. (6.12)
Similarly,
The integration boundaries for each of the integrals depend on the context and need to be specified. If we
are integrating over a volume V in position space, we also write
d3 x f (⃗x, p⃗). (6.14)
V
If no volume or boundaries are specified, it is usually assumed that we integrate over all possible values, i.e.
R3 ,
∞ ∞ ∞
d3 p f (⃗x, p⃗) = d3 p f (⃗x, p⃗) = dpx dpy dpz f (⃗x, p⃗). (6.16)
R3 −∞ −∞ −∞
For such an integral to be finite or meaningful, the function f (⃗x, p⃗) needs to decay sufficiently fast as each
of the components pi approaches ±∞. The three most important coordinate systems are
∞ ∞ ∞
x
Cartesian: ⃗x = y , d3 x f (⃗x) = dx dy dz f (⃗x), (6.17)
z −∞ −∞ −∞
∞ 2π ∞
r⊥ cos ϕ
cylindrical: ⃗x = r⊥ sin ϕ , d3 x f (⃗x) = dr⊥ dϕ dz r⊥ f (⃗x), (6.18)
z 0 0 −∞
∞ 2π π
r cos ϕ sin θ
spherical: ⃗x = r sin ϕ sin θ , d3 x f (⃗x) = dr dϕ dθ r2 sin θf (⃗x). (6.19)
r cos θ 0 0 0
Exercise: Compute the volumes of a cube, cylinder, and sphere using a suitable coordinate system.
Example for volume integral. Consider the so-called Gaussian distribution function
2
f (⃗x, p⃗) = e−A(⃗p−⃗p0 ) , (6.20)
66
Writing p⃗0 = (p0x , p0y , p0z ) we have
2
+(py −p0y )2 +(pz −p0z )2 ] 2 2 2
f (⃗x, p⃗) = e−A[(px −p0x ) = e−A(px −p0x ) e−A(py −p0y ) e−A(pz −p0z ) , (6.23)
and so
∞ ∞ ∞
−A(px −p0x )2 −A(py −p0y )2 2
I=V dpx e dpy e dpz e−A(pz −p0z ) . (6.24)
−∞ −∞ −∞
We introduce the shifted variables pi → p′i = pi − p0i and see that the integrals becomes independent of p⃗0 ,
∞
′ 2
∞
′ 2
∞
′ 2
I=V dp′x e−A(px ) dp′y e−A(py ) dp′z e−A(pz ) (6.25)
−∞ −∞ −∞
3
=: V J . (6.26)
Each of the three integrals has the same value J. It can be computed as follows:
∞
2
J= dx e−Ax (6.27)
−∞
∞ ∞
2
+y 2 )
J2 = dx dy e−A(x (6.28)
−∞ −∞
∞ 2π
2 x r cos ϕ
= dr r dϕ e−Ar , = (6.29)
0 0 y r sin ϕ
∞
2
= 2π dr re−Ar (6.30)
0
−Ar 2 ∞
e
= 2π (6.31)
−2A
r=0
π
= , (6.32)
A
so
∞
r
2 π
J= dx e−Ax = . (6.33)
−∞ A
N -particle systems. We consider a system of N particles in three dimensions. The particles are labelled by
the indices a, b, · · · ∈ {1, . . . , N }. The vectors corresponding to each of the N particles are p⃗a and ⃗xa . The
ith component of the momentum of the ath particle is indicated by a double index, pai . A dynamical state
of the system is specified by the positions and momenta of all particles. The total energy of a dynamical
state is given by the Hamiltonian function
67
with
N
X p⃗2a
Hkin = , (6.37)
a=1
2m
N N
X 1 X
Hpot = Vex (⃗xa ) + Vint (|⃗xa − ⃗xb |). (6.38)
a=1
2
a,b=1
Here Vex (⃗x) is an external potential and Vint (r) is the interaction energy between two particles at distance
r. For instance, Vex (⃗x) could be Earth’s gravitational field,
and Vint (r) could be a molecular Lennard–Jones type, electrostatic Coulomb, or gravitational Newtonian
potential,
σ τ
r12 − r6
1 q1 q2
Vint (r) = 4πε r
(6.40)
m1 m2
−G r
p⃗a
⃗va = = ⃗x˙ a , (6.42)
m
and Newton’s equation of motion reads
⃗
¨a = Fa .
⃗x (6.43)
m
A non-interacting (= ideal) system has Vint (r) = 0, and so
N
X
H= H1 (⃗xa , p⃗a ) (6.44)
a=1
p⃗2
H1 (⃗x, p⃗) = + Vex (⃗x). (6.45)
2m
For a thermodynamic system with fixed parameters such as (T, V, N ), we do not attempt to trace the
positions and momenta of all individual particles. Instead, we pick an arbitrary particle and denote by
68
a distribution function that is proportional to the probability that this particle is at position ⃗x with mo-
mentum p⃗ at time t. Obviously, if we know f (⃗x, p⃗, t), we know how the system behaves as a collection of
particles. More precisely, we choose a normalization such that
is the number of particles located inside a small cube of volume d3 x around the position ⃗x, and a small
cube d3 p around p⃗ in momentum space. One should think of d3 x and d3 p as containing a macroscopic
number of particles. For instance, for a gas at room temperature we have n ∼ 1019 cm−3 , so a volume of
d3 x = (1 µm)3 = 10−18 m3 = 10−12 cm3 would appear pointlike to us, although it still contains 107 particles.
The total number of particles is
N= dN = d3 x d3 p f (⃗x, p⃗, t) (6.48)
V
at any time t. The local number density of particles at position ⃗x (with any momentum p⃗) is
n(⃗x) = d3 p f (⃗x, p⃗, t), (6.49)
and
N= d3 x n(⃗x). (6.50)
V
The momentum distribution of particles with momentum p⃗ (at any position ⃗x) is
ñ(⃗
p) = d3 x f (⃗x, p⃗, t), (6.51)
V
with
N= d3 p ñ(⃗
p). (6.52)
Boltzmann equation. The Boltzmann equation is the equation of motion for f (⃗x, p⃗, t) in time. For this,
consider a particle with (⃗x, p⃗) at time t. At a later time t′ = t + δt, with δt infinitesimal, it will have
coordinates
p⃗
⃗x′ = ⃗x + ⃗v δt = ⃗x + δt, (6.53)
m
p⃗′ = p⃗ + F⃗ δt, (6.54)
where ⃗v is the velocity of the particle and F⃗ = −∇⃗x Vex the external force acting on the particle.5 If there
are no collisions (interactions) among the particles, we have
because the probability to find the particle ”moves” along with the particle’s trajectory, or
p⃗
f ⃗x + δt, p⃗ + F⃗ δt, t + δt − f (⃗x, p⃗, t) = 0. (6.56)
m
In the limit δt → 0 we obtain
5 In ⃗ that we will ignore in the following.
the presence of interactions, Vint ̸= 0, there is a mean-field contribution to F
69
p⃗
∂t + · ∇⃗x + F⃗ · ∇p⃗ f (⃗x, p⃗, t) = 0. (6.57)
m
Boltzmann equation for non-interacting particles (6.58)
Example. Consider non-interacting particles without external forces. The solution to the Boltzmann equation
∂t + ⃗v · ∇⃗x f (⃗x, p⃗, t) = 0 (6.59)
we have
because a particle that started at (⃗x, p⃗) at t might be scattered away and not end up at (⃗x′ , p⃗′ ) at t′ , or
scattered particles that did not start at (⃗x, p⃗) at t might appear at (⃗x′ , p⃗′ ) at t′ . We define the collision
integral as the difference,
hence
p⃗
∂t + · ∇⃗x + F⃗ · ∇p⃗ f (⃗x, p⃗, t) = Icoll . (6.65)
m
Boltzmann equation (6.66)
For simplicity assume that f (⃗x, p⃗, t) does not depend on ⃗x. Consider then a collision where our particle with
momentum p⃗ scatters elastically with a particle of momentum p⃗2 into two particles with final momentum p⃗3
and p⃗4 . Similarly, consider two particles with initial momenta p⃗3 and p⃗4 that scatter elastically into particles
with momenta p⃗ and p⃗1 :
The number of events of the first kind (which decrease f ) is proportional to f (⃗p, t)f (⃗
p2 , t), which is the
probability to find this initial configuration.6 Similarly, the number of events of the second kind (which
6 We assume molecular chaos here, where the momenta of the two particles are uncorrelated and the probability factorizes.
70
increase f ) is proportional to f (⃗
p3 , t)f (⃗
p4 , t). We have
h i
3 3
Icoll = d p2 d p3 d3 p4 M(⃗
p, p⃗2 ; p⃗3 , p⃗4 ) −f (⃗
p, t)f (⃗
p2 , t) + f (⃗
p3 , t)f (⃗
p4 , t) . (6.67)
In an elastic collision, energy and momentum are conserved, thus the momenta are constrained by
(
p⃗ + p⃗2 = p⃗3 + p⃗4 ,
(∗) ⃗2
p ⃗22
p ⃗23
p ⃗24
p . (6.68)
2m + 2m = 2m + 2m
The function M encodes the details of the interaction, and is not important for us here. A typical form is
p⃗2 p⃗2 p⃗2 p⃗2
M(⃗ p, p⃗2 ; p⃗3 , p⃗4 ) = Ṽint (⃗q)δ + 2 − 3 − 4 δ (3) (⃗
p + p⃗2 − p⃗3 − p⃗4 ), (6.69)
2m 2m 2m 2m
with momentum transfer ⃗q = p⃗3 − p⃗ and Ṽint (⃗q) the Fourier transform of the interaction potential Vint (r).
Icoll = 0. (6.73)
A sufficient (and, one can show, necessary) condition for this to hold is
f0 (⃗
p)f0 (⃗
p2 ) = f0 (⃗
p3 )f0 (⃗
p4 ), (6.74)
or, taking the logarithm
ln f0 (⃗
p) + ln f0 (⃗
p1 ) = ln f0 (⃗
p3 ) + ln f0 (⃗
p4 ). (6.75)
This has the form of a conservation law, since (⃗ p, p⃗2 ) and (⃗
p3 , p⃗4 ) can be any momenta that satisfy momentum
and energy conservation. If χ(⃗ p) is any quantity associated to a particle with momentum p⃗, so that χ(⃗ p1 ) +
χ(⃗p2 ) is conserved in an elastic collision between two particles with p⃗1 and p⃗2 , then the most general solution
is
ln f0 (⃗
p) = χ1 (⃗
p) + χ2 (⃗
p) + . . . , (6.76)
where χ1 (⃗
p), χ2 (⃗
p), . . . is a list of all independently conserved quantities. For point-particles, these are the
energy and momentum of the particle, see Eqs. (∗), and, of course, a constant. Hence the most general form
is
p⃗2 ⃗ ′ · p⃗ + C ′
p ) = A′
ln f0 (⃗ +B (6.77)
2m
= −A(⃗p − p⃗0 )2 + C (6.78)
71
⃗ ′ = 2A⃗
with suitable constants A′ = −2mA, B p0 , C ′ = C − A⃗
p20 . We conclude that the stable equilibrium
distribution is given by
ñ0 (⃗
p) 1 2
f0 (⃗
p) = = e−A(⃗p−⃗p0 ) (6.79)
V N
where N is a normalization constant. We define the expectation value of a function g = g(⃗x, p⃗) in equilibrium
as
3
⟨g⟩ = d x d3 p g(⃗x, p⃗)f0 (⃗
p). (6.80)
V
Note that ⟨g⟩ is extensive because of the volume integral V
d3 x.
thus
V π 3/2
N = . (6.82)
N A
The average momentum of a gas particle is
⟨⃗
ptot ⟩ 1
⟨⃗
p⟩ = = d3 x d3 p p⃗ f0 (⃗
p) (6.83)
N N V
1 1 2
= d3 x d3 p p⃗ e−A(⃗p−⃗p0 ) (6.84)
NN V
11 2 1 2
= d3 x d3 p p⃗0 e−A(⃗p−⃗p0 ) + p − p⃗0 ) e−A(⃗p−⃗p0 )
d3 x d3 p (⃗ (6.85)
N N V N
| V {z }
0
p⃗0 3 3
= d x d p f0 (⃗
p) (6.86)
N V
| {z }
N
= p⃗0 . (6.87)
∞x 0
∞ ∞
d3 x ⃗x η(⃗x) = dx dy dz y η(⃗x) = 0 . (6.88)
−∞ −∞ −∞ z 0
So p⃗0 is the average momentum of particles. For a system that does not move as a whole, we must therefore
choose
p⃗0 = 0. (6.89)
Indeed, for all particles with p⃗, there should be an equal number of particles with −⃗
p. At last we determine
A. We write
β
A= (6.90)
2m
72
and
β 3/2 2
f0 (⃗
p) = n e−βp /(2m) . (6.91)
2πm
We determine β from the pressure. Consider a container wall in the yz-plane and gas particles impinging on
it from the left with positive px > 0. The number of particles hitting an area A of the wall in a time-interval
∆t, i.e. the number of collisions, is given by
1 3 1 1
dNcoll = p) A∆x
f0 (⃗ p)Avx ∆td3 p =
| {z } d p = 2 f0 (⃗ p)Apx ∆td3 p.
f0 (⃗ (6.92)
2 2m
∆V
The factor 21 accounts for the fact that we only consider half the particles due to px > 0. The momentum
transfer onto the wall per collision is
1 A 3/2 1 π 3/2
= n × (6.98)
m π 2A A
n n
= = . (6.99)
2mA β
We used that
∞ ∞
π 1/2
r
2 ∂ 2 ∂ π
dx x2 e−Ax = − dx e−Ax = − = , (6.100)
−∞ ∂A −∞ ∂A A 2A3/2
which is another Gaussian integral. Since the absolute temperature is defined such that
P = nkB T (6.101)
for an ideal gas, which also holds for very weakly interacting gases, we conclude that
1
β= . (6.102)
kB T
The equilibrium distribution becomes
n 2
f0 (⃗
p) = 3/2
e−p /(2mkB T ) (6.103)
(2πmkB T )
Maxwell–Boltzmann distribution. (6.104)
73
p), which only depends on the magnitude p = |⃗
Note that the rotation-invariant or isotropic form of f0 (⃗ p|,
implies that
⟨p2x ⟩ = ⟨p2y ⟩ = ⟨p2z ⟩. (6.105)
Furthermore,
3
2
⟨p ⟩ = d x d3 p(p2x + p2y + p2z )f0 (⃗
p) = 3⟨p2x ⟩. (6.106)
V
3
E= P V, (6.112)
2
3
E = N kB T (6.113)
2
for point-like particles, i.e. monoatomic gases. The equilibrium distribution h0 (p) of magnitudes p (instead
of p⃗), defined from
∞ ∞
d3 p f0 (⃗
p) = 4π dp p2 f0 (⃗
p) =: dp h0 (⃗
p), (6.114)
0 0
is given by
8 3/2 2
h0 (p) = 4πp2 f0 (⃗
p) = n p2 e−p /(2mkB T ) . (6.115)
π 3 mkB T
For small momenta we have
h0 (p) ∼ p2 , (6.116)
for large momenta we have
2
h0 (p) ∼ p2 e−βp /(2m)
. (6.117)
Exercise: Show that the most probable momentum, p̄, defined as the momentum such that
h0 (p) = 4πp2 f0 (⃗
p) (6.118)
is maximal, is given by
p
p̄ = 2mkB T , (6.119)
74
whereas the root mean square momentum, defined through
⟨p2 ⟩ 1 3
p2rms = = d x d3 p p2 f0 (⃗
p), (6.120)
N N V
is given by
p
prms = 3mkB T . (6.121)
Example. For a gas of molecular nitrogen N2 (m = 2 × 14 u) at room temperature (T = 293 K), we have
r
2kB T m
v̄ = = 417 , (6.122)
m s
r
3kB T m
vrms = = 511 . (6.123)
m s
the three momentum components p⃗ = (px , py , pz ) are the only degrees of freedom of a particle. We define
⟨. . . ⟩
⟨. . . ⟩1 = (6.125)
N
as the average per particle. We have
⟨p2 ⟩1
⟨p2x ⟩1 = = mkB T (6.126)
3
D p2 E 1
x
⇒ = kB T (6.127)
2m 1 2
per particle, which yields
N D 2 E N D 2 E
X p X px 1
⟨E⟩ = =3 = 3N × kB T. (6.128)
a=1
2m 1
a=1
2m 1 2
If the gas consists of diatomic molecules, each particle has five degrees of freedom. Indeed, besides the
momentum p⃗ there are two rotational degrees of freedom, associated to rotations about the x- and y-axis,
where we assumed that the bond connecting both atoms is along the z-axis. This means that a rotation
about the z-axis leaves the molecule invariant and would therefore not amount to an actual physical motion
of the molecule. (Similarly, a pointlike particle has zero rotational degrees of freedom.) If I is the moment
of inertia of the molecule and ωi the angular velocity for rotation about the i-axis, then
N h
X 1 2 I i
H= pax + p2ay + p2az + ωx2 + ωy2 . (6.129)
a=1
2m 2
75
For such a diatomic gas we have
1
⟨E⟩ = 5N × kB T (6.130)
2
in equilibrium. More generally, for a gas of molecules with g degrees of freedom per molecule, we have
1
⟨E⟩ = g × N × kB T (6.131)
2
equipartition theorem (6.132)
in equilibrium. This implies, of course that
CV g
= kB . (6.133)
N 2
The equipartition theorem is only valid at high temperatures, where quantum statistics can be neglected.
where H1 (⃗x, p⃗) is the single-particle Hamiltonian and the interactions are assumed to be present, but weak.
An analysis of the Boltzmann equation and the vanishing of the collision integral then shows that ln f0 is a
sum of all conserved quantities. For a system at rest this yields
76
Inserting, for instance, A(⃗x, p⃗) = pi or A(⃗x, p⃗) = xi this yields
D ∂H E
1
pi = kB T, (6.143)
∂pi 1
D ∂H E
1
xi = kB T. (6.144)
∂xi 1
More generally, if the full H depends on any coordinate Z (position, momentum, angular velocity, . . . ), then
D ∂H E
Z = kB T. (6.145)
∂Z 1
Example. Dulong–Petit law. Consider a solid as a three-dimensional crystal lattice of atoms or ions. The
positions of the N ions are labelled by xai , their momenta by pai . The Hamiltonian reads
N
X p⃗2a
H= + Hpot (⃗x1 , . . . , ⃗xN ). (6.148)
a=1
2m
In the harmonic approximation, we assume that the positions ⃗xa only slightly deviate from their equilibrium
positions ⃗x0,a defined as the minima of Hpot . The Hamiltonian can then be approximated by
N 2
X p⃗a k
H= + ⃗u2a , (6.149)
a=1
2m 2
where ⃗ua = ⃗xa − ⃗x0,a is the displacement of the ath ion. This a three-dimensional harmonic oscillator with
g = 6 degrees of freedom per ion. We conclude that
CV = 3N kB (6.150)
Dulong–Petit law, (6.151)
which is a good approximation for CV ≈ CP of many solids at high temperatures.
H = Hkin (⃗
p1 , . . . , p⃗N ) + Hpot (⃗x1 , . . . , ⃗xN ). (6.152)
For
N
X p⃗2a
Hkin = (6.153)
a=1
2m
77
we have
N
∂Hkin ∂ X p⃗2b
p⃗a · = p⃗a · (6.154)
∂⃗
pa ∂⃗
pa 2m
b=1
N
X p⃗b p⃗2a
= p⃗a · δab = . (6.155)
m m
b=1
Consequently,
N D N
X ∂Hkin E XD p⃗2a E
p⃗a · = = 2⟨Hkin ⟩. (6.156)
a=1
∂⃗
pa a=1
m
Since
D ∂Hkin E D ∂Hpot E
p⃗a · = ⃗xa · (6.157)
∂⃗
pa ∂⃗xa
we conclude that
N D
X ∂Hpot E
2⟨Hkin ⟩ = ⃗xa · (6.158)
a=1
∂⃗xa
Virial theorem. (6.159)
More generally, one can show that if
then
N
X ∂Hpot
⃗xa · = nHpot . (6.162)
a=1
∂⃗xa
The kinetic energy is always homogeneous of degree two. For a potential of the form
N
X
Hpot (⃗x1 , . . . , ⃗xN ) = ca |⃗xa |n (6.163)
a=1
we then obtain
78