Quantum Gases Explained
Quantum Gases Explained
Quantum Gases
In this section we will discuss situations where quantum e↵ects are important.
We’ll still
restrict attention to gases — meaning a bunch of particles moving around and barely
interacting — but one of the first things we’ll see is how versatile the idea of a
gas
can be in the quantum world. We’ll use it to understand not just the traditional
gases
that we met in the previous section but also light and, ironically, certain
properties of
solids. In the latter part of this section, we will look at what happens to gases
at low
temperatures where their behaviour is dominated by quantum statistics.
3.1 Density of States
We start by introducing the important concept of the density of states. To
illustrate
this, we’ll return once again to the ideal gas trapped in a box with sides of
length
L and volume V = L3 . Viewed quantum mechanically, each particle is described
by a wavefunction. We’ll impose periodic boundary conditions on this wavefunction
(although none of the physics that we’ll discuss in this course will be sensitive
to the
choice of boundary condition). If there are no interactions between particles, the
energy
eigenstates are simply plane waves,
1 ~
= p eik·~x
V
Boundary conditions require that the wavevector ~k = (k1 , k2 , k3 ) is quantized
as
ki =
2⇡ni
with ni 2 Z
L
~2 k 2
4⇡ 2 ~2 2
=
(n + n22 + n23 )
2m
2mL2 1
with k = |~k|. The quantum mechanical single particle partition function (1.21) is
given
by the sum over all energy eigenstates,
X
Z1 =
e E~n
~
n
The question is: how do we do the sum? The simplest way is to approximate it by an
integral. Recall from the previous section that the thermal wavelength of the
particle
is defined to be
s
2⇡~2
=
mkB T
– 62 –
The exponents that appear in the sum are all of the form ⇠ 2 n2 /L2 , up to some
constant factors. For any macroscopic size box,
⌧ L (a serious understatement!
Actually ⌧⌧ L) which ensures that there are many states with E~n kB T all of
which
contribute to the sum. (There will be an exception to this at very low temperatures
which will be the focus of Section 3.5.3). We therefore lose very little by
approximating
the sum by an integral. We can write the measure of this integral as
Z
Z
Z 1
X
V
4⇡V
3
3
⇡
dn =
dk=
dk k 2
3
3
(2⇡)
(2⇡) 0
~
n
where, in the last equality, we have integrated over the angular directions to get
4⇡, the
area of the 2-sphere, leaving an integration over the magnitude k = |~k| and the
Jacobian
factor k 2 . For future applications, it will prove more useful to change
integration
variables at this stage. We work instead with the energy,
E=
~2 k 2
2m
dE =
~2 k
dk
m
(3.1)
where
V
g(E) = 2
4⇡
2m
~2
◆3/2
E 1/2
(3.2)
is the density of states: g(E)dE counts the number of states with energy between E
and E + dE. Notice that we haven’t actually done the integral over E in (3.1);
instead
this is to be viewed as a measure which we can integrate over any function f (E) of
our
choosing.
There is nothing particularly quantum mechanical about the density of states.
Indeed, in the derivation above we have replaced the quantum sum with an integral
over
momenta which actually looks rather classical. Nonetheless, as we encounter more
and
more di↵erent types of gases, we’ll see that the density of states appears in all
the
calculations and it is a useful quantity to have at our disposal.
3.1.1 Relativistic Systems
Relativistic particles moving in d = 3 + 1 spacetime dimensions have kinetic energy
p
E = ~2 k 2 c 2 + m 2 c 4
(3.3)
– 63 –
Repeating the steps above, we find the density of states is given by
VE p
g(E) = 2 3 3 E 2 m2 c4
2⇡ ~ c
In particular, for massless particles, the density of states is
g(E) =
V E2
2⇡ 2 ~3 c3
(3.4)
(3.5)
– 64 –
Photons have two polarization states (one for each dimension transverse to the
direction of propagation). To account for this, the density of states (3.5) should
be
multiplied by a factor of two. The number of states available to a single photon
with
energy between E and E + dE is therefore
g(E)dE =
V E2
dE
⇡ 2 ~3 c 3
V !2
d!
⇡ 2 c3
(3.6)
where we’ve indulged in a slight abuse of notation since g(!) is not the same
function
as g(E) but is instead defined by the equation above. It is also worth pointing out
an
easy mistake to make when performing these kinds of manipulations with densities of
states: you need to remember to rescale the interval dE to d!. This is most simply
achieved by writing g(E)dE = g(!)d! as we have above. If you miss this then you’ll
get g(!) wrong by a factor of ~.
The final fact that we need is important: photons are not conserved. If you put
six atoms in a box then they will still be there when you come back a month later.
This isn’t true for photons. There’s no reason that the walls of the box can’t
absorb
one photon and then emit two. The number of photons in the world is not fixed. To
demonstrate this, you simply need to turn o↵ the light.
Because photon number is not conserved, we’re unable to define a chemical potential
for photons. Indeed, even in the canonical ensemble we must already sum over states
with di↵erent numbers of photons because these are all “accessible states”. (It is
sometimes stated that we should work in the grand canonical ensemble at µ = 0 which
is basically the same thing). This means that we should consider states with any
number N of photons.
We’ll start by looking at photons with a definite frequency !. A state with N such
photons has energy E = N ~!. Summing over all N gives us the partition function for
photons at fixed frequency,
Z! = 1 + e
~!
+ e 2 ~! + . . . =
1
e
~!
(3.7)
We now need to sum over all possible frequencies. As we’ve seen a number of times,
independent partition functions multiply, which means that the logs add. We only
need
– 65 –
Figure 14: The Planck Distribution function (Source: E. Schubert, Light Emitting
Diodes).
to know how many photon states there are with some frequency !. But this is what
the density of states (3.6) tells us. We have
Z 1
Z 1
V
log Z =
d! g(w) log Z! =
d! ! 2 log 1 e ~!
(3.8)
2 c3
⇡
0
0
3.2.1 Planck Distribution
From the partition function (3.8) we can calculate all interesting quantities for a
gas of
light. For example, the energy density stored in the photon gas is
Z 1
@
V~
!3
E=
log Z = 2 3
d! ~!
(3.9)
@
⇡ c 0
e
1
However, before we do the integral over frequency, there’s some important
information
contained in the integrand itself: it tells us the amount of energy carried by
photons
with frequency between ! and ! + d!
V~
!3
d!
(3.10)
⇡ 2 c3 e ~! 1
This is the Planck distribution. It is plotted above for various temperatures. As
you
can see from the graph, for hot gases the maximum in the distribution occurs at a
lower
wavelength or, equivalently, at a higher frequency. We can easily determine where
this
maximum occurs by finding the solution to dE(!)/d! = 0. It is
E(!)d! =
kB T
~
⇣
where ⇣ ⇡ 2.822 solves 3 ⇣ = 3e . The equation above is often called Wien’s
displacement law. Roughly speaking, it tells you the colour of a hot object.
!max = ⇣
– 66 –
To compute the total energy in the gas of photons, we need to do the integration
in (3.9). To highlight how the total energy depends on temperature, it is useful to
perform the rescaling x = ~!, to get
Z
V (kB T )4 1 x3 dx
E= 2 3
⇡ c
~3
ex 1
0
R
The integral I = dx x3 /(ex 1) is tricky but doable. It turns out to be I = ⇡ 4
/15.
(We will e↵ectively prove this fact later in the course when we consider a more
general
class of integrals (3.27) which can be manipulated into the sum (3.28). The net
result of
this is to express the integral I above in terms of the Gamma function and the
Riemann
zeta function: I = (4)⇣(4) = ⇡ 4 /15). We learn that the energy density E = E/V in
a
gas of photons scales is proportional to T 4 ,
E=
4
⇡ 2 kB
T4
15~3 c3
Stefan-Boltzmann Law
The expression for the energy density above is closely related to the Stefan-
Boltzmann
law which describes the energy emitted by an object at temperature T . That energy
flux is defined as the rate of transfer of energy from the surface per unit area.
It is
given by
Energy Flux =
Ec
⌘ T4
4
(3.11)
where
=
4
⇡ 2 kB
= 5.67 ⇥ 10 8 Js 1 m 2 K 4
3
2
60~ c
– 67 –
Radiation Pressure and Other Stu↵
All other quantities of interest can be computed from the free energy,
F =
kB T log Z
Z
V kB T 1
= 2 3
d! ! 2 log 1
⇡ c
o
~!
Z 1
V~
! 3 e ~!
d!
3⇡ 2 c3 0
1 e ~!
Z 1
V~ 1
x3
dx
3⇡ 2 c3 4 ~4 0
ex 1
V ⇡2
(kB T )4
45~3 c3
@F
E
4 4
=
T
=
@V T
3V
3c
This is the equation of state for a gas of photons. The middle equation tells us
that the
pressure of photons is one third of the energy density — a fact which will be
important
in the Cosmology course.
We can also calculate the entropy S and heat capacity CV . They are both most
conveniently expressed in terms of the Stefan constant which hides most of the
annoying
factors,
S=
@F
16V
=
T3
@T V
3c
CV =
@E
16V
=
T3
@T V
c
– 68 –
This result is totally astonishing. The light has been traveling for 13.7 billion
years,
almost since the beginning of time itself. And yet we can understand it with
ridiculous
accuracy using such a simple calculation. If you’re not awed by this graph then you
have no soul. You can learn more about this in the lectures on cosmology.
3.2.3 The Birth of Quantum Mechanics
The key conceptual input in the derivation of Planck’s formula (3.10) is that light
of
frequency ! comes in packets of energy E = ~!. Historically, this was the first
time
that the idea of quanta arose in theoretical physics.
Let’s see what would happen in a classical world where light of frequency ! can
have
arbitrarily low intensity and, correspondingly, arbitrarily low energy. This is
e↵ectively
what happens in the regime h! ⌧ kB T of the Planck distribution where the minimum
energy ~! is completely swamped by the temperature. There we can approximate
1
e ~!
1
~!
V
! 2 kB T
⇡ 2 c3
Notice that all hints of the quantum ~ have vanished. This is the Rayleigh-Jeans
law for
the distribution of classical radiation. It has a serious problem if we try to
extrapolate it
R1
to high frequencies since the total energy, E = 0 E(!)d!, diverges. This was
referred
to as the ultra-violet catastrophe. In contrast, in Planck’s formula (3.10) there
is an
exponential suppression at high frequencies. This arises because when ~!
kB T , the
– 69 –
temperature is not high enough to create even a single photon. By imposing a
minimum
energy on these high frequency modes, quantum mechanics ensures that they become
frozen out.
3.2.4 Max Planck (1858-1947)
“A new scientific truth does not triumph by convincing its opponents and
making them see the light, but rather because its opponents eventually die”
Max Planck
Planck was educated in Munich and, after a brief spell in Kiel, moved to take a
professorship in Berlin. (The same position that Boltzmann had failed to turn up
for).
For much of his early career, Planck was adamantly against the idea of atoms. In
his view the laws of thermodynamics were inviolable. He could not accept that they
sat on a foundation of probability. In 1882, he wrote “atomic theory, despite its
great
successes, will ultimately have to be abandoned”.
Twenty years later, Planck had changed his mind. In 1900, he applied Boltzmann’s
statistical methods to photons to arrive at the result we derived above, providing
the
first hints of quantum mechanics. However, the key element of the derivation — that
light comes in quanta — was not emphasised by Planck. Later, when quantum theory
was developed, Planck refused once more to accept the idea of probability
underlying
physics. This time he did not change his mind.
3.3 Phonons
It is hard to imagine substances much more di↵erent than a gas
and a solid. It is therefore quite surprising that we can employ
our study of gases to accurately understand certain properties
of solids. Consider a crystal of atoms of the type shown in the
figure. The individual atoms are stuck fast in position: they
certainly don’t act like a gas. But the vibrations of the atoms
— in other words, sound waves — can be treated using the
same formalism that we introduced for photons.
Figure 15:
– 70 –
(3.12)
where cs is now the speed of sound rather than the speed of light.
The density of states for phonons is the same as that of photons (3.6) with two
exceptions: we must replace the speed of light c with the speed of sound cs ; and
phonons
have three polarization states rather than two. There are two transverse
polarizations
(like the photon) but also a longitudinal mode. The density of states is therefore
g(!)d! =
3V 2
! d!
2⇡ 2 c3s
There is one further important di↵erence between phonons and photons. While light
waves can have arbitrarily high frequency, sound waves cannot. This is because high
frequency waves have small wavelengths, = 2⇡cs /!. But there is a minimum
wavelength in the problem set by the spacing between atoms. it is not possible for
sound
waves to propagate through a solid with wavelength smaller than the atomic spacing
because there’s nothing in the middle there to shake.
We will denote the maximum allowed phonon frequency as !D . The minimum wavelength,
D , should be somewhere around the lattice spacing between atoms, which is
(V /N )1/3 , so we expect that !D ⇠ (N/V )1/3 cs . But how can we work out the
coefficient? There is a rather clever argument to determine !D due to Debye. (So
clever that
he gets his initial on the frequency and his name on the section heading). We start
by
counting the number of single phonon states,
Z !D
V !3
d! g(w) = 2 D3
2⇡ cs
0
The clever bit of the argument is to identify this with the number of degrees of
freedom
in the solid. This isn’t immediately obvious. The number of degrees of freedom in a
lattice of N atoms is 3N since each atom can move in three directions. But the
number
of single phonon states is counting collective vibrations of all the atoms
together. Why
should they be equal?
To really see what’s going on, one should compute the correct energy eigenstates of
the lattice and just count the number of single phonon modes with wavevectors
inside
the first Brillouin zone. (You can learn about Brillouin zones in the Applications
of
Quantum Mechanics lecture notes.) But to get the correct intuition, we can think in
the following way: in general the solid will have many phonons inside it and each
of
these phonons could sit in any one of the single-phonon states that we counted
above.
Suppose that there are three phonons sitting in the same state. We say that this
state
is occupied three times. In this language, each of the states above can be occupied
– 71 –
an arbitrary number of times, restricted only by the energy available. If you want
to
describe the total state of the system, you need to say how many phonons are in the
first state, and how many are in the second state and so on. The number of one-
phonon
states is then playing the role of the number of degrees of freedom: it is the
number of
things you can excite.
The net result of this argument is to equate
V !3
3N = 2 D3
2⇡ cs
!D =
6⇡ 2 N
V
◆1/3
cs
~!D
kB
+ e 2 ~! + . . . =
where the partition function Z! for a single phonon of frequency ! is the same as
that
of a photon (3.7). The total energy in sound waves is therefore
Z !D
Z !D
~!g(!)
3V ~
!3
E=
d! ~!
= 2 3
d! ~!
e
1
2⇡ cs 0
e
1
0
– 72 –
Figure 16: Experimental data for heat capacities. The solid line is the Debye
prediction.
(Source: D. Schroeder An Introduction to Thermal Physics)
Z TD /T
dx
x3
ex
4
@E
2⇡ 2 V kB
=
T3
@T
5~3 c3s
(T ⌧ TD )
(3.13)
12⇡ 4
CV = N k B
5
T
TD
◆3
(3.14)
In contrast, at temperatures T
TD we only integrate over small values of x, allowing
us to Taylor expand the integrand,
Z TD /T
0
dx
x3
ex
Z TD /T
0
1
dx x + . . . =
3
2
– 73 –
TD
T
◆3
+ ...
This ensures that the energy grows linearly with T and the heat capacity tends
towards
a constant value
CV =
4 3
V kB
TD
= 3N kB
2
2⇡ ~3 c3s
(T
TD )
(3.15)
This high-temperature behaviour has been known experimentally since the early
1800’s.
It is called the Dulong-Petit law. Debye’s argument for the value of !D was
basically
constructed to reproduce the coefficient 3N in the formula above. This was known
experimentally, but also from an earlier model of vibrations in a solid due to
Einstein.
(You met the Einstein model in the first problem sheet). Historically, the real
success of the Debye model was the correct prediction of the T 3 behaviour of CV at
low
temperatures.
In most materials the heat capacity is dominated by the phonon contribution. (In
metals there is an additional contribution from conduction electrons that we will
calculate in Section 3.6). The heat capacity of three materials is shown in Figure
16,
together with the predictions from the Debye model. As you can see, it works very
well! The deviation from theory at high temperatures is due to di↵erences between
CV
and Cp , the heat capacity at constant pressure.
What’s Wrong with the Debye Model?
As we’ve seen, the Debye model is remarkably accurate in capturing the heat
capacities
of solids. Nonetheless, it is a caricature of the physics. The most glaring problem
is our
starting point (3.12). The relationship E = ~! between energy and frequency is
fine;
the mistake is the relationship between frequency ! and wavevector (or momentum)
k, namely ! = kcs . Equations of this type, relating energy and momentum, are
called
dispersion relations. It turns out that that the dispersion relation for phonons is
a little
more complicated.
It is not hard to compute the dispersion relation for phonons.
(You will, in fact, do this calculation in the Applications of
Quantum Mechanics course). For simplicity, we’ll work with
a one dimensional periodic lattice of N atoms as shown in
the figure. The equilibrium position of each atom is xl = la
and we impose periodic boundary conditions by insisting that
xN +1 ⌘ x1 . Let ul be the deviation from equilibrium, ul =
xl la. If we approximate the bonds joining the atoms as har-
– 74 –
Figure 17:
monic oscillators, the Hamiltonian governing the vibrations is
H=
1 X 2 ↵X
u̇ +
(ui
2m i i
2 i
ui+1 )2
where ↵ is a parameter governing the strength of the bonds between atoms. The
equation of motion is
üi =
↵
(2ui
m
ui+1
ui 1 )
This is easily solved by the discrete Fourier transform. We make the ansatz
1 X
ul = p
ũk ei(kla !k t)
N k
Plugging this into the equation of motion gives the dispersion relation
r
✓ ◆
↵
ka
!k = 2
sin
m
2
To compute the partition function correctly in this model, we would have to revisit
the
density of states using the new dispersion relation E(k) = ~!k . The resulting
integrals
are messy. However, at low temperatures only the smallest frequency modes are
excited
and, for small ka, the sin function is approximately linear. This means that we get
back to the dispersion relationpthat we used in the Debye model, ! = kcs , with the
speed of sound given by cs = a ↵/m. Moreover, at very high temperatures it is
simple
to check that this model gives the Dulong-Petit law as expected. It deviates from
the
Debye model only at intermediate temperatures and, even here, this deviation is
mostly
negligible.
3.4 The Diatomic Gas Revisited
With a bit of quantum experience under our belt, we can look again at the diatomic
gas that we discussed in Section 2.4. Recall that the classical prediction for the
heat
capacity — CV = 72 N kB — only agrees with experiment at very high temperatures.
Instead, the data suggests that as the temperature is lowered, the vibrational
modes
and the rotational modes become frozen out. But this is exactly the kind of
behaviour
that we expect for a quantum system where there is a minimum energy necessary to
excite each degree of freedom. Indeed, this “freezing out” of modes saved us from
ultra-violet catastrophe in the case of blackbody radiation and gave rise to a
reduced
heat capacity at low temperatures for phonons.
– 75 –
Let’s start with the rotational modes, described by the Hamiltonian (2.18).
Treating
this as a quantum Hamiltonian, it has energy levels
E=
~2
j(j + 1)
2I
j = 0, 1, 2, . . .
1
X
(2j + 1)e
~2 j(j+1)/2I
j=0
When T
which agrees with our result for the classical partition function (2.19).
In contrast, for T ⌧ ~2 /2IkB all states apart from j = 0 e↵ectively decouple and
we have simply Zrot ⇡ 1. At these temperatures, the rotational modes are frozen at
temperatures accessible in experiment so only the translational modes contribute to
the heat capacity.
This analysis also explains why there is no rotational contribution to the heat
capacity
of a monatomic gas. One could try to argue this away by saying that atoms are point
particles and so can’t rotate. But this simply isn’t true. The correct argument is
that
the moment of inertia I of an atom is very small and the rotational modes are
frozen.
Similar remarks apply to rotation about the symmetry axis of a diatomic molecule.
The vibrational modes are described by the harmonic oscillator. You already
computed the partition function for this on the first examples sheet (and, in fact,
implicitly
in the photon and phonon calculations above). The energies are
1
E = ~!(n + )
2
and the partition function is
Zvib =
X
n
~!(n+ 12 )
=e
~!/2
~!n
e
1
~!/2
=
~!
1
2 sinh( ~!/2)
– 76 –
1, the partition function becomes Zvib ⇡ e ~!/2 . This is a contribution from the
zeropoint energy of the harmonic oscillator. It merely gives the expected additive
constant
to the energy per particle,
Evib =
@
~!
log Zvib ⇡
@
2
and doesn’t contribute the heat capacity. Once again, we see how quantum e↵ects
explain the observed behaviour of the heat capacity of the diatomic gas. The end
result is a graph that looks like that shown in Figure 11.
3.5 Bosons
For the final two topics of this section, we will return again to the simple
monatomic
ideal gas. The classical treatment that we described in Section 2.2 has
limitations. As
the temperature decreases, the thermal de Broglie wavelength,
s
2⇡~2
=
mkB T
gets larger. Eventually it becomes comparable to the inter-particle separation,
(V /N )1/3 .
At this point, quantum e↵ects become important. If the particles are non-
interacting,
there is really only one important e↵ect that we need to consider: quantum
statistics.
Recall that in quantum mechanics, particles come in two classes: bosons and
fermions.
Which class a given particle falls into is determined by its spin, courtesy of the
spinstatistics theorem. Integer spin particles are bosons. This means that any
wavefunction
must be symmetric under the exchange of two particles,
(~r1 , ~r2 ) = (~r2 , ~r1 )
Particles with 12 -integer spin are fermions. They have an anti-symmetrized
wavefunction,
(~r1 , ~r2 ) =
(~r2 , ~r1 )
At low temperatures, the behaviour of bosons and fermions is very di↵erent. All
familiar
fundamental particles such as the electron, proton and neutron are fermions. But an
atom that contains an even number of fermions acts as a boson as long as we do not
reach energies large enough to dislodge the constituent particles from their bound
state.
Similarly, an atom consisting of an odd number of electrons, protons and neutrons
will
be a fermion. (In fact, the proton and neutron themselves are not fundamental: they
– 77 –
are fermions because they contain three constituent quarks, each of which is a
fermion.
If the laws of physics were di↵erent so that four quarks formed a bound state
rather
than three, then both the proton and neutron would be bosons and, as we will see in
the next two sections, nuclei would not exist!).
We will begin by describing the properties of bosons and then turn to a discussion
of fermions to Section 3.6.
3.5.1 Bose-Einstein Distribution
We’ll change notation slightly from earlier sections and label the single particle
quantum
states of the system by |ri. (We used |ni previously, but n will be otherwise
occupied
for most of this section). The single particle energies are then Er and we’ll
assume that
our particles are non-interacting. In that case, you might think that to specify
the state
of the whole system, you would need to say which state particle 1 is in, and which
state
particle 2 is in, and so on. But this is actually too much information because
particle 1
and particle 2 are indistinguishable. To specify the state of the whole system, we
don’t
need to attach labels to each particle. Instead, it will suffice to say how many
particles
are in state 1 and how many particles are in state 2 and so on.
We’ll denote the number of particles in state |ri as nr . If we choose to work in
the
canonical ensemble, we must compute the partition function,
X
Z=
e n r Er
{nr }
where the sum is over all possible ways of partitioning N particles into sets {nr }
subject
P
to the constraint that r nr = N . Unfortunately, the need to impose this constraint
makes the sums tricky. This means that the canonical ensemble is rather awkward
when discussing indistinguishable particles. It turns out to be much easier to work
in
the grand canonical ensemble where we introduce a chemical potential µ and allow
the
total number of particles N to fluctuate.
Life is simplest if we think of each state |ri in turn. In the grand canonical
ensemble,
a given state can be populated by an arbitrary number of particles. The grand
partition
function for this state is
X
1
Zr =
e nr (Er µ) =
1 e (Er µ)
n
r
Notice that we’ve implicitly assumed that the sum above converges, which is true
only
if (Er µ) > 0. But this should be true for all states Er . We will set the ground
state
– 78 –
to have energy E0 = 0, so the grand partition function for a Bose gas only makes
sense
if
µ<0
(3.16)
Now we use the fact that all the occupation of one state is independent of any
other.
The full grand partition function is then
Z=
Y
r
1
1
(Er µ)
X
1 @
1
log Z =
(E
@µ
e r µ)
r
X
r
hnr i
Here hnr i denotes the average number of particles in the state |ri,
hnr i =
1
e (Er µ)
(3.17)
– 79 –
(3.18)
Ideal Bose Gas
Let’s look again at a gas of non-relativistic particles, now through the eyes of
quantum
mechanics. The energy is
E=
~2 k 2
2m
As explained in Section 3.1, we can replace the sum over discrete momenta with an
integral over energies as long as we correctly account for the density of states.
This
was computed in (3.2) and we reproduce the result below:
✓
◆3/2
V
2m
g(E) = 2
E 1/2
(3.19)
4⇡
~2
From this, together with the Bose-Einstein distribution, we can easily compute the
total number of particles in the gas,
Z
g(E)
N = dE
,
(3.20)
z 1e E 1
There is an obvious, but important, point to be made about this equation. If we can
do
the integral (and we will shortly) we will have an expression for the number of
particles
in terms of the chemical potential and temperature: N = N (µ, T ). That means that
if
we keep the chemical potential fixed and vary the temperature, then N will change.
But
in most experimental situations, N is fixed and we’re working in the grand
canonical
ensemble only because it is mathematically simpler. But nothing is for free. The
price
we pay is that we will have to invert equation (3.20) to get it in the form µ =
µ(N, T ).
Then when we change T , keeping N fixed, µ will change too. We have already seen an
example of this in the ideal gas where the chemical potential is given by (2.14).
The average energy of the Bose gas is,
Z
E = dE
Eg(E)
,
z 1e E 1
(3.21)
And, finally, we can compute the pressure. In the grand canonical ensemble this is,
Z
1
1
pV = log Z =
dE g(E) log 1 ze E
We can manipulate this last expression using an integration by parts. Because g(E)
⇠
E 1/2 , this becomes
Z
2
Eg(E)
2
pV =
dE
= E
(3.22)
1
E
3
z e
1
3
– 80 –
This is implicitly the equation of state. But we still have a bit of work to do.
Equation
(3.21) gives us the energy as a function of µ and T . And, by the time we have
inverted
(3.20), it gives us µ as a function of N and T . Substituting both of these into
(3.22)
will give the equation of state. We just need to do the integrals. . .
3.5.2 A High Temperature Quantum Gas is (Almost) Classical
Unfortunately, the integrals (3.20) and (3.21) look pretty fierce. Shortly, we will
start
to understand some of their properties, but first we look at a particularly simple
limit.
We will expand the integrals (3.20), (3.21) and (3.22) in the limit z = e µ ⌧ 1.
We’ll
figure out the meaning of this expansion shortly (although there’s a clue in the
title of
this section if you’re impatient). Let’s look at the particle density (3.20),
N
1
=
V
4⇡ 2
=
1
4⇡ 2
1
=
4⇡ 2
✓
✓
✓
2m
~2
2m
~2
2m
~2
◆3/2 Z 1
dE
E 1/2
z 1e E 1
dE
ze E E 1/2
1 ze E
◆3/2 Z 1
◆3/2
0
3/2
Z 1
dx
xe x (1 + ze x + . . .)
where we made the simple substitution x = E. The integrals are all of the Gaussian
type and can be easily evaluated by making one further substitution x = u2 . The
final
answer can be conveniently expressed in terms of the thermal wavelength ,
N
z
= 3
V
z
1 + p + ...
2 2
(3.23)
Now we’ve got the answer, we can ask what we’ve done! What kind of expansion is z ⌧
1? From the above expression, we see that the expansion is consistent only if 3 N/V
⌧
1, which means that the thermal wavelength is much less than the interparticle
spacing.
But this is true at high temperatures: the expansion z ⌧ 1 is a high temperature
expansion.
At first glance, it is surprising that z = e µ ⌧ 1 corresponds to high
temperatures.
When T ! 1, ! 0 so naively it looks as if z ! 1. But this is too naive. If we keep
the particle number N fixed in (3.20) then µ must vary as we change the
temperature,
and it turns out that µ ! 1 faster than ! 0. To see this, notice that to leading
order we need z/ 3 to be constant, so z ⇠ T 3/2 .
– 81 –
High Temperature Equation of State of a Bose Gas
We now wish to compute the equation of state. We know from (3.22) that pV = 23 E.
So we need to compute E using the same z ⌧ 1 expansion as above. From (3.21), the
energy density is
✓
◆3/2 Z 1
E
1
2m
E 3/2
=
dE
V
4⇡ 2 ~2
z 1e E 1
0
✓
◆3/2
Z 1
1
2m
z
=
dx x3/2 e x (1 + ze x + . . .)
5/2
4⇡ 2 ~2
0
✓
◆
3z
z
= 3
1 + p + ...
(3.24)
2
4 2
The next part’s a little fiddly. We want to eliminate z from the expression above
in
favour of N/V . To do this we invert (3.23), remembering that we’re working in the
limit z ⌧ 1 and 3 N/V ⌧ 1. This gives
✓
◆
3
N
1 3N
p
z=
1
+ ...
V
2 2 V
which we then substitute into (3.24) to get
✓
◆✓
◆
3N
1 3N
1 3N
p
E=
1
+ ...
1+ p
+ ...
2
2 2 V
4 2 V
and finally we substitute this into pV = 23 E to get the equation of state of an
ideal
Bose gas at high temperatures,
✓
◆
3
N
p
pV = N kB T 1
+ ...
(3.25)
4 2V
which reproduces the classical ideal gas, together with a term that we can identify
as the second virial coefficient in the expansion (2.21). However, this correction
to
the pressure hasn’t arisen from any interactions among the atoms: it is solely due
to
quantum statistics. We see that the e↵ect of bosonic statistics on the high
temperature
gas is to reduce the pressure.
3.5.3 Bose-Einstein Condensation
We now turn to the more interesting question: what happens to a Bose gas at low
temperatures? Recall that for convergence of the grand partition function we
require
µ < 0 so that z = e µ 2 (0, 1). Since the high temperature limit is z ! 0, we’ll
anticipate that quantum e↵ects and low temperatures come about as z ! 1.
– 82 –
Recall that the number density is given by (3.20) which we write as
N
1
= 2
V
4⇡
2mkB T
~2
◆3/2 Z 1
dx
x1/2
1
⌘
g (z)
3 3/2
z 1 ex 1
(3.26)
where the function g3/2 (z) is one of a class of functions which appear in the
story of
Bose gases,
Z 1
1
xn 1
gn (z) =
dx 1 x
(3.27)
(n) 0
z e
1
These functions are also known as polylogarithms and sometimes denoted as Lin (z) =
gn (z). The function g3/2 is relevant for the particle number and the function g5/2
appears in the calculation of the energy in (3.21). For reference, the gamma
function
p
has value (3/2) = ⇡/2. (Do not confuse these functions gn (z) with the density of
states g(E). They are not related; they just share a similar name).
In Section 3.5.2 we looked at (3.26) in the T ! 1 limit. There we saw that ! 0
but the function g3/2 (z) ! 0 in just the right way to compensate and keep N/V
fixed.
What now happens in the other limit, T ! 0 and ! 1. One might harbour the hope
that g3/2 (z) ! 1 and again everything balances nicely with N/V remaining constant.
We will now show that this can’t happen.
A couple of manipulations reveals that the integral (3.27) can be expressed in
terms
of a sum,
Z
1
zxn 1 e x
gn (z) =
dx
(n)
1 ze x
Z
1
X
1
=
z dx xn 1 e x
z m e mx
(n)
m=0
Z
1
X
1
=
z m dx xn 1 e mx
(n) m=1
Z
1
1 X zm
=
du un 1 e u
n
(n) m=1 m
But the integral that appears in the last line above is nothing but the definition
of the
gamma function (n). This means that we can write
gn (z) =
1
X
zm
m=1
– 83 –
mn
(3.28)
We see that gn (z) is a monotonically increasing function of z. Moreover, at z = 1,
it is
equal to the Riemann zeta function
gn (1) = ⇣(n)
For our particular problem it will be useful to know that ⇣(3/2) ⇡ 2.612.
Let’s now return to our story. As we decrease T in (3.26), keeping N/V fixed, z and
hence g3/2 (z) must both increase. But z can’t take values greater than 1. When we
reach z = 1, let’s denote the temperature as T = Tc . We can determine Tc by
setting
z = 1 in equation (3.26),
Tc =
2⇡~2
kB m
◆✓
1 N
⇣(3/2) V
◆2/3
(3.29)
What happens if we try to decrease the temperature below Tc ? Taken at face value,
equation (3.26) tells us that the number of particles should decrease. But that’s a
stupid conclusion! It says that we can make particles disappear by making them
cold.
Where did they go? We may be working in p
the grand canonical ensemble, but in
the thermodynamic limit we expect N ⇠ 1/ N which is too small to explain our
missing particles. Where did we misplace them? It looks as if we made a mistake in
the calculation.
In fact, we did make a mistake in the calculation. It happened right at the
beginning.
Recall that back in Section 3.1 we replaced the sum over states with an integral
over
energies,
X V (2m)3/2 Z
⇡
dE E 1/2
4⇡ 2 ~3
~k
p
Because of the weight E in the integrand, the ground state with E = 0 doesn’t
contribute to the integral. We can use the Bose-Einstein distribution (3.17) to
compute
the number of states that we expect to be missing because they sit in this E = 0
ground
state,
n0 =
1
z 1
1
(3.30)
For most values of z 2 (0, 1), there are just a handful of particles sitting in
this lowest
state and it doesn’t matter if we miss them. But as z gets very very close to 1
(meaning
z ⇡ 1 1/N ) then we get a macroscopic number of particles occupying the ground
state.
– 84 –
It is a simple matter to redo our calculation taking into account the particles in
the
ground state. Equation (3.26) is replaced by
N=
V
3
g3/2 (z) +
z
1
V
⇣(3/2) = 1
N 3
T
Tc
◆3/2
(3.31)
– 85 –
Figure 19: The velocity distribution of Rubidium atoms, taken from the Ketterle lab
at
MIT. The left-hand picture shows T > Tc , just before the condensate forms. The
middle and
right-hand pictures both reveal the existence of the condensate.
2E
kB T
= 3 g5/2 (z)
3V
(3.32)
Formally there is also a contribution from the ground state, but it is log(1 z)/V
which is a factor of N smaller than the term above and can be safely ignored. At
low
temperatures, T < Tc , we have z ⇡ 1 and
p=
kB T
3
⇣(5/2)
So at low temperatures, the equation of state of the ideal Bose gas is very
di↵erent from
the classical, high temperature, behaviour. The pressure scales as p ⇠ T 5/2
(recall that
there is a factor of T 3/2 lurking in the ). More surprisingly, the pressure is
independent
of the density of particles N/V .
3.5.4 Heat Capacity: Our First Look at a Phase Transition
Let’s try to understand in more detail what happens as we pass through the critical
temperature Tc . We will focus on how the heat capacity behaves on either side of
the
critical temperature.
– 86 –
We’ve already seen in (3.32) that we can express the energy in terms of the
function
g5/2 (z),
E
3 kB T
=
g5/2 (z)
V
2 3
so the heat capacity becomes
CV
1 dE
15kB
3 kB T dg5/2 dz
=
=
g5/2 (z) +
3
V
V dT
4
2 3 dz dT
(3.33)
The first term gives a contribution both for T < Tc and for T > Tc . However, the
second
term includes a factor of dz/dT and z is a very peculiar function of temperature:
for
T > Tc , it is fairly smooth, dropping o↵ at T = Tc . However, as T ! Tc , the
fugacity
rapidly levels o↵ to the value z ⇡ 1 1/N . For T < Tc , z doesn’t change very much
at
all. The net result of this is that the second term only contributes when T > Tc .
Our
goal here is to understand how this contribution behaves as we approach the
critical
temperature.
Let’s begin with the easy bit. Below the critical temperature, T < Tc , only the
first
term in (3.33) contributes and we may happily set z = 1. This gives the heat
capacity,
CV =
15V kB
⇣(5/2) ⇠ T 3/2
3
4
(3.34)
Now we turn to T > Tc . Here we have z < 1, so g5/2 (z) < g5/2 (1). We also have
dz/dT < 0. This means that the heat capacity decreases for T > Tc . But we know
that for T < Tc , CV ⇠ T 3/2 so the heat capacity must have a maximum at T = Tc .
Our goal in this section is to understand a little better what the function CV
looks like
in this region.
0
To compute the second term in (3.33) we need to understand both g5/2
and how z
changes with T as we approach Tc from above. The first calculation is easy if we
use
our expression (3.28),
gn (z) =
1
X
zm
m=1
mn
d
1
gn (z) = gn 1 (z)
dz
z
(3.35)
As T ! Tc from above, dg5/2 /dT ! ⇣(3/2), a constant. All the subtleties lie in the
remaining term, dz/dT . After all, this is the quantity which is e↵ectively
vanishing for
T < Tc . What’s it doing at T > Tc ? To figure this out is a little more involved.
We
start with our expression (3.26),
g3/2 (z) =
N 3
V
– 87 –
T > Tc
(3.36)
and we’ll ask what happens to the function g3/2 (z) as z ! 1, keeping N fixed. We
know that exactly at z = 1, g3/2 (1) = ⇣(3/2). But how does it approach this value?
To
answer this, it is actually simplest to look at the derivative dg3/2 /dz = g1/2 /z,
where
1
g1/2 (z) =
(1/2)
Z 1
dx
x 1/2
z 1 ex 1
The reason for doing this is that g1/2 diverges as z ! 1 and it is generally easier
to
isolate divergent parts of a function than some finite piece. Indeed, we can do
this
straightforwardly for g1/2 by looking at the integral very close to x = 0, we can
write
Z ✏
1
x 1/2
g1/2 (z) =
dx 1
+ finite
(1/2) 0
z (1 + x) 1
Z ✏
z
x 1/2
=
dx
+ ...
(1/2) 0
(1 z) + x
Z ✏
2z
1
1
= p
du
+ ...
1 + u2
1 z (1/2) 0
p
where, in the last line, we made the substution u = x/(1 z). So we learn that
as z ! 1, g1/2 (z) ! z(1 z) 1/2 . But this is enough information to tell us how
g3/2
approaches its value at z = 1: it must be
g3/2 (z) ⇡ ⇣(3/2)
A(1
z)1/2 + . . .
for some constant A. Inserting this into our equation (3.36) and rearranging, we
find
that as T ! Tc from above,
✓
◆2
1
N 3
z⇡1
⇣(3/2)
A2
V
!2
✓ ◆3/2
⇣(3/2)2
Tc
=1
1
A2
T
✓
◆2
T Tc
⇡1 B
Tc
where, in the second line, we used the expression of the critical temperature
(3.29).
In the final equality, we’ve Taylor expanded the temperature near Tc . (Write T /Tc
=
1 + (T Tc )/Tc and Taylor expand assuming that (T Tc )/Tc is small.) Here B is some
constant that we could figure out with a little more e↵ort, but it won’t be
important
for our story. From the expression above, we can now determine dz/dT as T ! Tc . We
see that it vanishes linearly at T = Tc .
– 88 –
Putting all this together, we can determine the expression for the heat capacity
(3.33) when T > Tc . We’re not interested in the coefficients, so we’ll package a
bunch
of numbers of order 1 into a constant b and the end result is
✓
◆
15V kB
T Tc
CV =
g5/2 (z) b
4 3
Tc
The first term above goes smoothly over to the
expression (3.34) for CV when T < Tc . But
the second term is only present for T > Tc .
Notice that it goes to zero as T ! Tc , which
ensures that the heat capacity is continuous at
this point. But the derivative is not continuous. A sketch of the heat capacity is
shown in
the figure.
CV
3NkB/2
T
TC
z=
1
1+
n0
◆ 1
1
1+
N1
1
(T /Tc )3/2
◆ 1
(T < Tc )
The di↵erence is that while dz/dT is of order one above Tc , it is of order 1/N
below
Tc . In the thermodynamic limit, N ! 1, this results in the discontinuity that we
saw
above. This is a general lesson: phase transitions with their associated
discontinuities
can only arise in strictly infinite systems. There are no phase transitions in
finite
systems.
Superfluid Helium-4
A similar, but much more pronounced, discontinuity is seen in Helium-4 as it
becomes a superfluid, a transition which occurs at 2.17 K. The atom contains two
protons, two neutrons and two electrons and is therefore a boson. (In contrast,
Helium3 contains just a single neutron and is a fermion). The experimental data for
the heat
capacity of Helium-4 is shown on the right. The successive graphs are zooming in
on the phase transistion: the scales are (from left to right) Kelvin, milliKelvin
and
microKelvin. The discontinuity is often called the lambda transition on account of
the
shape of this graph.
– 89 –
There is a close connection between BoseEinstein condensation described above and
superfluids: strictly speaking a non-interacting BoseEinstein condensate is not a
superfluid but superfluidity is a consequence of arbitrarily weak
repulsive interactions between the atoms. However, in He-4, the interactions
between atoms
are strong and the system cannot be described
using the simple techniques developed above.
(~r2 , ~r1 )
n(Er µ)
=1+e
n=0,1
– 90 –
(Er µ)
Q
So, the grand partition function for all states is Z = r Zr , from which we can
compute
the average number of particles in the system
X
X
1
N=
⌘
nr
e (Er µ) + 1
r
r
where the average number of particles in the state |ri is
nr =
(3.37)
e (Er µ) + 1
(3.38)
dE
Eg(E)
z 1e E + 1
dE g(E) log 1 + ze
– 91 –
2
= E
3
(3.39)
At high temperatures, it is simple to repeat the steps of Section 3.5.2. (This is
one of
the questions on the problem sheet). Only a few minus signs di↵er along the way and
one again finds that for z ⌧ 1, the equation of state reduces to that of a
classical gas,
✓
◆
3
N
pV = N kB T 1 + p
+ ...
(3.40)
4 2gs V
Notice that the minus signs filter down to the final answer: the first quantum
correction
to a Fermi gas increases the pressure.
3.6.2 Degenerate Fermi Gas and the Fermi Surface
In the extreme limit T ! 0, the Fermi-Dirac distribution becomes very simple: a
state
is either filled or empty.
(
1
for E < µ
1
!
(E
µ)
e
+1
0
for E > µ
It’s simple to see what’s going on here. Each fermion that we throw into the system
settles into the lowest available energy state. These are successively filled until
we run
out of particles. The energy of the last filled state is called the Fermi energy
and is
denoted as EF . Mathematically, it is the value of the chemical potential at T = 0,
µ(T = 0) = EF
(3.41)
Filling up energy states with fermions is just like throwing balls into a box. With
one
exception: the energy states of free particles are not localised in position space;
they
are localised in momentum space. This means that successive fermions sit in states
with ever-increasing momentum. In this way, the fermions fill out a ball in
momentum
space. The momentum of the final fermion is called the Fermi momentum and is
related
to the Fermi energy in the usual way: ~kF = (2mEF )1/2 . All states with wavevector
|~k| kF are filled and are said to form the Fermi sea or Fermi sphere. Those
states
with |~k| = kF lie on the edge of the Fermi sea. They are said to form the Fermi
surface. The concept of a Fermi surface is extremely important in later
applications to
condensed matter physics.
We can derive an expression for the Fermi energy in terms of the number of
particles
N in the system. To do this, we should appreciate that we’ve actually indulged in a
slight abuse of notation when writing (3.41). In the grand canonical ensemble, T
and
µ are independent variables: they’re not functions of each other! What this
equation
really means is that if we want to keep the average particle number N in the system
– 92 –
fixed (which we do) then as we vary T we will have to vary µ to compensate. So a
slightly clearer way of defining the Fermi energy is to write it directly in terms
of the
particle number
✓
◆3/2
Z EF
gs V 2m
3/2
N=
dE g(E) =
EF
(3.42)
2
2
6⇡
~
0
Or, inverting,
~2
EF =
2m
6⇡ 2 N
gs V
◆2/3
(3.43)
The Fermi energy sets the energy scale for the system. There is an equivalent
temperature scale, TF = EF /kB . The high temperature expansion that resulted in
the
equation of state (3.40) is valid at temperatures T > TF . In contrast,
temperatures
T < TF are considered “low” temperatures for systems of fermions. Typically, these
low temperatures do not have to be too low: for electrons in a metal, TF ⇠ 104 K;
for
electrons in a white dwarf, TF > 107 K.
While EF is the energy of the last occupied state, the average energy of the system
can be easily calculated. It is
Z EF
3
E=
dE Eg(E) = N EF
(3.44)
5
0
Similarly, the pressure of the degenerate Fermi gas can be computed using (3.39),
2
pV = N EF
5
(3.45)
Even at zero temperature, the gas has non-zero pressure, known as degeneracy
pressure.
It is a consequence of the Pauli exclusion principle and is important in the
astrophysics
of white dwarf stars and neutron stars. (We will describe this application in
Section
3.6.5). The existence of this residual pressure at T = 0 is in stark contrast to
both the
classical ideal gas (which, admittedly, isn’t valid at zero temperature) and the
bosonic
quantum gas.
3.6.3 The Fermi Gas at Low Temperature
We now turn to the low-temperature behaviour of the Fermi gas. As mentioned above,
“low” means T ⌧ TF , which needn’t be particularly low in everyday terms. The
number of particles N and the average energy E are given by,
Z 1
g(E)
N=
dE
(3.46)
1
z e E +1
0
– 93 –
n(E)
T=0
T << E F
E
EF
Figure 22: The Fermi-Dirac Distribution function at T = 0 and very small T . The
distribution di↵ers from the T = 0 ground state only for a range of energies kB T
around EF .
and
E=
Z 1
0
dE
Eg(E)
z 1e E + 1
(3.47)
Our goal is to firstly understand how the chemical potential µ, or equivalently the
fugacity z = e µ , changes with temperature when N is held fixed. From this we can
determine how E changes with temperature when N is held fixed.
There are two ways to proceed. The first is a direct approach on the problem by
Taylor expanding (3.46) and (3.47) for small T . But it turns out that this is a
little
delicate because the starting point at T = 0 involves the singular distribution
shown
on the left-hand side of Figure 22 and it’s easy for the physics to get lost in a
morass of
integrals. For this reason, we start by developing a heuristic understanding of how
the
integrals (3.46) and (3.47) behave at low temperatures which will be enough to
derive
the required results. We will then give the more rigorous expansion – sometimes
called
the Sommerfeld expansion – and confirm that our simpler derivation is indeed
correct.
The Fermi-Dirac distribution (3.37) at small temperatures is sketched on the
righthand side of Figure 22. The key point is that only states with energy within
kB T of
the Fermi surface are a↵ected by the temperature. We want to insist that as we vary
the temperature, the number of particles stays fixed which means that dN/dT = 0.
I claim that this holds if, to leading order, the chemical potential is independent
of
temperature, so
dµ
=0
dT T =0
– 94 –
(3.48)
Let’s see why this is the case. The change in particle number can be written as
Z 1
dN
d
g(E)
=
dE (E µ)
dT
dT
e
+1
✓
◆
Z 1 0
d
1
=
dE g(E)
dT e (E µ) + 1
0
✓
◆
Z 1
@
1
⇡ g(EF )
dE
@T e (E EF ) + 1
0
There are two things going on in the step from the second line to the third.
Firstly,
we are making use of the fact that, for kB T ⌧ EF , the Fermi-Dirac distribution
only changes significantly in the vicinity of EF as shown in the right-hand side of
Figure 22. This means that the integral in the middle equation above is only
receiving
contributions in the vicinity of EF and we have used this fact to approximate the
density
of states g(E) with its value at g(EF ). Secondly, we have used our claimed result
(3.48)
to replace the total derivative d/dT (which acts on the chemical potential) with
the
partial derivative @/@T (which doesn’t) and µ is replaced with its zero temperature
value EF .
Explicitly di↵erentiating the Fermi-Dirac distribution in the final line, we have
✓
◆
Z 1
dN
E EF
1
⇡ g(EF )
dE
⇡0
2
2
dT
kB T
4 cosh ( (E EF )/2)
0
This integral vanishes because (E EF ) is odd around EF while the cosh function
is even. (And, as already mentioned, the integral only receives contributions in
the
vicinity of EF ).
Let’s now move on to compute the change in energy with temperature. This, of
course, is the heat capacity. Employing the same approximations that we made above,
we can write
✓
◆
Z 1
@E
@
1
CV =
=
dE Eg(E)
@T N,V
@T e (E µ) + 1
0
✓
◆
Z 1
3
@
1
⇡
dE EF g(EF ) + g(EF )(E Ef )
2
@T e (E EF ) + 1
0
However, this time we have not just replaced Eg(E) by EF g(EF ), but instead Taylor
expanded to include a term linear in (E EF ). (The factor of 3/2 comes about
because
Eg(E) ⇠ E 3/2 . The first EF g(EF ) term in the square brackets vanishes by the
same
even/odd argument that we made before. But the (E EF ) term survives.
– 95 –
Writing x = (E
where we’ve extended the range of the integral from 1 to +1, safe in the knowledge
that only the region near EF (or x = 0) contributes anyway. More importantly,
however, this integral only gives an overall coefficient which we won’t keep track
of. The
final result for the heat capacity is
2
C V ⇠ kB
T g(EF )
✓
◆3/2 Z 1
E
gs 2m
E 3/2
=
dE
V
4⇡ 2 ~2
z 1e E + 1
0
3 gs
=
kB T f5/2 (z)
2 3
(3.51)
p
where = 2⇡~2 /mkB T is the familiar thermal wavelength and the functions fn (z)
are the fermionic analogs of gn defined in (3.27),
Z 1
1
xn 1
fn (z) =
dx 1 x
(3.52)
(n) 0
z e +1
– 97 –
p
p
where it is useful to recall (3/2) = ⇡/2 and (5/2) = 3 ⇡/4 (which follows, of
course, from the property (n + 1) = n (n)). This function is an example of a
polylogarithm and is sometimes written as Lin ( z) = fn (z).
Expanding fn (z)
We will now derive the large z expansion of fn (z), sometimes called the Sommerfeld
expansion. The derivation is a little long. We begin by splitting the dx integral
into
two parts,
Z µ
Z 1
xn 1
xn 1
(n)fn (z) =
dx 1 x
+
dx 1 x
z e +1
z e +1
0
µ
✓
◆ Z 1
Z µ
1
xn 1
=
dx xn 1 1
+
dx
1 + ze x
z 1 ex + 1
0
µ
Z µ
Z 1
(log z)n
xn 1
xn 1
=
dx
+
dx
n
1 + ze x
z 1 ex + 1
0
µ
We now make a simple change of variable, ⌘1 = µ x for the first integral and
⌘2 = x
µ for the second,
Z µ
Z 1
(log z)n
( µ ⌘1 ) n 1
( µ + ⌘2 ) n 1
(n)fn (z) =
d⌘1
+
d⌘2
n
1 + e ⌘1
1 + e ⌘2
0
0
So far we have only massaged the integral into a useful form. We now make use of
the
approximation µ
1. Firstly, we note that the first integral has a e⌘1 suppression
in the denominator. If we replace the upper limit of the integral by 1, the
expression
changes by a term of order e µ = z 1 which is a term we’re willing to neglect. With
this approximation, we can combine the two integrals into one,
Z 1
(log z)n
( µ + ⌘)n 1 ( µ ⌘)n 1
(n)fn (z) =
+
d⌘
n
1 + e⌘
0
We now Taylor expand the numerator,
( µ + ⌘)n 1
( µ
⇥
⇤
⌘)n 1 = ( µ)n 1 (1 + ⌘/ µ)n 1 (1 ⌘/ µ)n 1
⌘
⌘
n 1
= ( µ)
1 + (n 1)
+ . . . 1 + (n 1)
+ ...
µ
µ
= 2(n 1)( µ)n 2 ⌘ + . . .
1)(log z)
– 98 –
n 2
Z 1
0
d⌘
⌘
e⌘ + 1
We’re left with a fairly simple integral.
Z 1
Z 1
⌘
⌘e ⌘
d⌘ ⌘
=
d⌘
e +1
1+e ⌘
0
0
Z 1
1
X
=
d⌘ ⌘
e m⌘ ( 1)m+1
0
m=1
Z
1
X
( 1)m+1 1
=
du ue u
2
m
0
m=1
R1
The integral in the last line is very simple: 0 du ue u = 1. We have only to do the
sum. There are a number of ways to manipulate this. We chose to write
✓
◆
✓
◆
1
1
1
1
1
1
1
1
1
1
+
... = 1 + 2 + 2 + 2 + ...
2 2 + 2 + 2 + ...
22 32 42
2
3
4
2
4
6
✓
◆ ✓
◆
2
1
1
= 1
· 1 + 2 + 2 + ...
(3.53)
2
2
2
3
P
This final sum is well known. It is
1/m2 = ⇣(2) = ⇡ 2 /6. (The original proof of this,
by Euler, proceeds by Taylor expanding sin x/x, writing resulting sum as a product
of
roots, and then equating the x2 term).
After all this work, we finally have the result we want. The low temperature
expansion of fn (z) is an expansion in 1/ log z = 1/ µ,
✓
◆
(log z)n
⇡ 2 n(n 1)
fn (z) =
1+
+ ...
(3.54)
(n + 1)
6 (log z)2
Back to Physics: Heat Capacity of a Fermi Gas Again
The expansion (3.54) is all we need to determine the leading order e↵ects of
temperature
on the Fermi gas. The number of particles (3.50) can be expressed in terms of the
chemical potential as
!
✓
◆2
2
N
gs
⇡
k
T
B
= 2 3 (2mµ)3/2 1 +
+ ...
(3.55)
V
6⇡ ~
8
µ
But the number of particles determines the Fermi energy through (3.43). For fixed
particle number N , we can therefore express the chemical potential at finite
temperature
in terms of the Fermi energy. After a small amount of algebra, we find
!
✓
◆2
2
⇡
kB T
µ = EF 1
+ ...
12 EF
– 99 –
We see that the chemical potential is a maximum at T = 0 and decreases at higher
T . This is to be expected: recall that by the time we get to the classical gas,
the
chemical potential must be negative. Moreover, note that the leading order
correction
is quadratic in temperature rather than linear, in agreement with our earlier claim
(3.48).
The energy density can be computed from (3.51) and (3.54). It is
!
✓
◆2
E
gs
5⇡ 2 kB T
3/2 5/2
=
(2m) µ
1+
+ ...
V
10⇡ 2 ~3
8
µ
Our real interest is in the heat capacity. However, with fixed particle number N ,
the
chemical potential also varies with temperature at quadratic order. We can solve
this
problem by dividing by (3.55) to get
!
✓
◆2
E
3EF
5⇡ 2 kB T
=
1+
+ ...
N
5
12
EF
Now the only thing that depends on temperature on the right-hand side is T itself.
From this we learn that the heat capacity of a low temperature Fermi gas is linear
in
T,
CV =
@E
⇡2 T
= N kB
@T N,V
2 TF
(3.56)
But we now have the correct coefficient that completes our earlier result (3.49).
3.6.5 White Dwarfs and the Chandrasekhar limit
When stars exhaust their fuel, the temperature T ! 0 and they have to rely on the
Pauli
exclusion principle to support themselves through the degeneracy pressure (3.45).
Such
stars supported by electron degeneracy pressure are called white dwarfs. In
addition
to the kinetic energy Ekinetic of fermions given in (3.43) the system has
gravitational
energy. If we make the further assumption that the density in a star of radius R is
uniform, then
3 GN M 2
5 R
where GN is Newton’s constant. In the problem set, you will minimise Egrav +
Ekinetic
to find the relationship between the radius and mass of the star,
Egrav =
R ⇠ M 1/3
This is unusual. Normally if you throw some stu↵ on a pile, the pile gets bigger.
Not
so for stars supported by degeneracy pressure. The increased gravitational
attraction
wins and causes them to shrink.
– 100 –
As the star shrinks, the Fermi energy (3.43) grows. Eventually it becomes
comparable
to the electron mass me and our non-relativistic approximation breaks down. We can
redo our calculations using the relativistic formula for the density of states
(3.4) (which
needs to be multiplied by a factor of 2 to account for the spin degeneracy). In the
opposite regime of ultra-relativistic electrons, where E
me , then we can expand the
density of states as
✓
◆
V
m 2 c4
2
g(E) = 2 3 3 E
+ ...
⇡ ~c
2
from which we can determine the kinetic energy due to fermions, replacing the
nonrelativistic result (3.44) by
✓
◆
V
1 4 m 2 c4 2
Ekinetic = 2 3 3
E
EF + . . .
⇡ ~c 4 F
4
The Fermi energy can be expressed in terms of the particle number by
✓
◆
V
1 3 m2 c4
N= 2 3 3
E
EF + . . .
⇡ ~c 3 F
2
The total energy of the system is then given by,
"
✓
◆1/3
3~c 9⇡M 4
Ekinetic + Egrav =
4
4m4p
#
3
1
GN M 2
+ O(R)
5
R
~c
GN
◆3/2
1
m2p
(3.57)
This is roughly 1.5 times the mass of the Sun. Stars that exceed this bound will
not
end their life as white dwarfs. Their collapse may ultimately be stopped by
degeneracy
pressure of neutrons so that they form a neutron star. Or it may not be stopped at
all
in which case they form a black hole.
– 101 –
The factor in brackets in (3.57) is an interesting mix of fundamental constants
associated to quantum theory, relativity and gravity. In this combination, they
define a
mass scale called the Planck mass,
Mpl2 =
~c
GN
When energy densities reach the Planck scale, quantum gravity e↵ects are important
meaning that we have to think of quantising spacetime itself. We haven’t quite
reached
this limit in white dwarf stars. (We’re shy by about 25 orders of magnitude or
so!).
Nonetheless the presence of the Planck scale in the Chandra limit is telling us
that
both quantum e↵ects and gravitational e↵ects are needed to understand white dwarf
stars: the matter is quantum; the gravity is classical.
3.6.6 Pauli Paramagnetism
“With heavy heart, I have decided that Fermi-Dirac, not Einstein, is the
correct statistics, and I have decided to write a short note on paramagnetism”
Wolfgang Pauli, in a letter to Schrödinger, reluctantly admitting that electrons
are fermions
~ There are
We now describe a gas of electrons in a background magnetic field B.
two e↵ects that are important: the Lorentz force ~v ⇥ B on the motion of electrons
and the coupling of the electron spin to the magnetic field. We first discuss the
spin
coupling and postpone the Lorentz force for the next section.
An electron can have two spin states, “up” and “down”. We’ll denote the spin state
by a discrete label: s = 1 for spin up; s = 1 for spin down. In a background
magnetic
~ the kinetic energy picks up an extra term
field B,
Espin = µB Bs
(3.58)
where µB = |e|~/2m is the Bohr magneton. (It has nothing to do with the chemical
potential. Do not confuse them!)
Since spin up and spin down electrons have di↵erent energies, their occupation
numbers di↵er. Equation (3.50) is now replaced by
✓
◆3/2 Z 1
N"
1
2m
E 1/2
1
=
dE (E+µ B µ)
= 3 f3/2 (ze µB B )
2
2
B
V
4⇡
~
e
+1
0
✓
◆3/2 Z 1
1/2
N#
1
2m
E
1
=
dE
=
f (ze+ µB B )
(E µB B µ) + 1
3 3/2
V
4⇡ 2 ~2
e
0
– 102 –
Our interest is in computing the magnetization, which is a measure of how the
energy
of the system responds to a magnetic field,
@E
@B
M=
(3.59)
Since each spin has energy (3.58), the magnetization is simply the di↵erence in the
in
the number of up spins and down spins,
M=
µB (N"
N# ) =
µB V ⇥
3
f3/2 (ze
µB B
f3/2 (ze µB B )
At high temperatures, and suitably low magnetic fields, we can approximate f3/2 (z)
⇡ z
as z ! 0, so that
M⇡
2µB V z
3
sinh( µB B)
2V z
3
cosh( µB B)
@M
@B
(3.60)
Evaluating in the limit of vanishing magnetic field, we find that the magnetization
is inversely proportional to the temperature,
(B = 0) =
N µ2B
kB T
– 103 –
The above results hold in the high temperature limit. In contrast, at low
temperatures the e↵ects of the Fermi surface become important. We need only keep
the leading
order term in (3.54) to find
✓
◆3/2
✓
◆
⇥
⇤ µ2B V 2m 3/2 1/2
µB V 2m
3/2
3/2
M⇡
(EF + µB B)
(EF µB B)
⇡
EF B
6⇡ 2
~2
2⇡ 2
~2
where we’re written the final result in the limit µB B ⌧ EF and expanded to linear
order
in B. We could again replace some of the factors for the electron number N .
However,
we can emphasise the relevant physics slightly better if we write the result above
in
terms of the density of states (3.19). Including the factor of 2 for spin
degeneracy, we
have
M ⇡ µ2B g(EF )B
(3.61)
– 104 –
Let’s see how quantum e↵ects change this conclusion. Our first goal is to
understand
the one-particle states for an electron in a constant background magnetic field.
This is
a problem that you will also solve in the Applications of Quantum Mechanics course.
It is the problem of Landau levels.
Landau Levels
We will consider an electron moving in a constant magnetic field pointing in the ẑ
~ = (0, 0, B). There are many di↵erent ways to pick A
~ such that B
~ = r⇥A
~
direction: B
gives the right magnetic field. These are known as gauge choices. Each choice will
give
rise to the same physics but, usually, very di↵erent equations in the intervening
steps.
Here we make the following gauge choice,
~ = ( By, 0, 0)
A
(3.62)
Let’s place our particle in a cubic box with sides of length L. Solutions to the
Schrödinger equation Ĥ = E are simply plane waves in the x and z directions,
(~r) = ei(kx x+kz z) f (y)
But the wavefunction f (y) must satisfy,
~2 @ 2
1
+
(~kx
2
2m @y
2m
(3.63)
eB
m
y0 )2 f (y) = E 0 f (y)
where y0 = ~kx /eB. This is simply the Schrödinger equation for a harmonic
oscillator,
jiggling with frequency !c and situated at position y0 .
One might wonder why the harmonic oscillator appeared. The cyclotron frequency
!c is the frequency at which electrons undergo classical Larmor orbits, so that
kind of
makes sense. But why have these become harmonic oscillators sitting at positions y0
which depend on kx ? The answer is that there is no deep meaning to this at all! it
is
~ giving rise to
merely an artefact of our gauge choice (3.62). Had we picked a di↵erent A
– 105 –
~ we would find ourselves trying to answer a di↵erent question! The lesson
the same B,
here is that when dealing with gauge potentials, the intermediate steps of a
calculation
do not always have physical meaning. Only the end result is important. Let’s see
what
this is. The energy of the particle is
E = E0 +
~2 kz2
2m
where = L2 B is the total flux through the system and 0 = 2⇡~/e is known as the
flux quantum. (Note that the result above does not include a factor of 2 for the
spin
degeneracy of the electron).
Back to the Diamagnetic Story
We can now compute the grand partition function for non-interacting electrons in a
magnetic field. Including the factor of 2 from electron spin, we have
✓
◆
Z
1
X
L
2L2 B
~2 kz2
log Z =
dkz
log 1 + z exp
~!c (n + 1/2)
2⇡
2m
0
n=0
To perform the sum, we use the Euler summation formula which states that for any
function h(x),
1
X
n=0
h(n + 1/2) =
Z 1
h(x)dx +
1 0
h (0) + . . .
24
– 106 –
So the grand partition function becomes
1
VB X
log Z =
h(µ ~!c (n + 1/2))
⇡ 0 n=0
Z
VB 1
V B ~!c dh(µ)
=
h(µ ~!c x) dx
+ ...
⇡ 0 0
⇡ 0 24 dµ
Z µ
Z
Vm
(~!c )2 +1
=
h(y) dy
dk (~2 k2 /2m µ)
+ . . . (3.64)
2⇡ 2 ~2
24
e
+1
1
1
The first term above does not depend on B. In fact it is simply a rather perverse
way
of writing the partition function of a Fermi gas when B = 0. However, our interest
here is in the magnetization which is again defined as (3.59). In the grand
canonical
ensemble, this is simply
M=
1 @log Z
@B
The second term in (3.64) is proportional to B 2 (which is hiding in the !c2 term).
Higher order terms in the Euler summation formula will carry higher powers of B
and,
for small B, the expression above will suffice.
At T = 0, the integrand is simply 1 for |k| < kF and zero otherwise. To compare
to Pauli paramagnetism, we express the final result in terms of the Bohr magneton
µB = |e|~/2mc. We find
M=
µ2B
g(EF )B
3
This is comparable to the Pauli contribution (3.61). But it has opposite sign.
Substances whose magnetization is opposite to the magnetic field are called
diamagnetic:
they are repelled by applied magnetic fields. The e↵ect that we have derived above
is
known as Landau diamagnetism.
– 107 –