Notes Hoker PDF
Notes Hoker PDF
Eric DHoker
Department of Physics and Astronomy,
University of California, Los Angeles, CA 90095, USA
15 September 2012
1
Contents
1 Introduction 6
1.1 Brief history . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
1.2 Units and physical constants . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
2
4.5 Quantum rotational modes . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
3
9 Bose-Einstein statistics 81
9.1 Black body radiation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 81
9.2 Cosmic micro-wave background radiation . . . . . . . . . . . . . . . . . . . . 83
9.3 Thermodynamic variables for Bose-Einstein ideal gasses . . . . . . . . . . . . 84
9.4 Bose-Einstein condensation . . . . . . . . . . . . . . . . . . . . . . . . . . . . 86
9.5 Behavior of the specific heat . . . . . . . . . . . . . . . . . . . . . . . . . . . 86
4
Bibliography
Course textbook
Statistical Mechanics,
R.K. Pathria, Elsevier Publishing, (2011).
Classics
Thermodynamics and Statistical Mechanics,
A. Sommerfeld, Academic Press, 1956;
Thermodynamics and Statistics,
E. Fermi, Phoenix Science Series (1966);
Statistical Mechanics, An advanced course with problems and solutions,
R. Kubo, North Holland, 1971;
Statistical Physics, Part 1,
L.D. Landau and E.M. Lifschitz, Pergamon Press, Third edition (1982);
Statistical Mechanics, A set of Lectures,
R.P. Feynman, Benjamin Cummings Publishing (1982);
Statistical Thermodynamics,
E. Schrodinger, Dover Publications (1989).
Standard textbooks
Fundamentals of statistical and thermal physics,
F. Reif, McGraw Hill;
From Microphysics to Macrophysics,
Methods and Applications of Statistical Mechanics, Vol 1,
R. Balian. Springer (2007);
Introduction to Statistical Phyisics,
K. Huang, CRC Press, Second Edition (2010);
Statistical Physics of Particles,
M. Kardar, Cambridge University Press (2010);
Statistical Physics of Fields,
M. Kardar, Cambridge University Press (2007).
5
1 Introduction
The goal of statistical mechanics is to explain the physical properties of macroscopic systems
in terms of the dynamics of its microscopic constituents.
A macroscopic system is one that contains a large number of microscopic constituents;
for example, 12 grams of pure Carbon 12 C contains 1 mole or 6.022141 1023 Carbon atoms,
and constitutes a macroscopic system. As such, almost all of the matter that we encounter
in everyday life is macroscopic.
Microscopic constituents may be atoms, such as Carbon atoms in the above example.
They may, however, also be larger than atoms and consist of molecules each one of which
is composed of several or many atoms. It is also possible to consider the microscopic con-
stituents to be smaller than atoms, and work with nuclei and electrons, as would be appro-
priate for ionized gases, or even smaller constituents such as quarks and gluons in a quark
plasma environment. The choice of which particles to consider as microscopic constituents
is guided by the length-scales and time-scales on which the system is considered.
The dynamical laws which individual microscopic constituents obey are generally known.
If the microscopic constituents are taken to be electrons and nuclei, and the effects of
the weak, strong, and gravitational forces can be neglected, then the dynamics of electro-
magnetism remains as the only relevant interaction between the microscopic constituents.
Almost all matter encountered naturally on Earth should result, in one way or the other,
from the dynamics of the electro-magnetic interactions between these constituents.
The physical properties of the corresponding macroscopic system show, however, tremen-
dous qualitative and quantitative differences. The qualitative distinctions are usually referred
to as phases of matter. In gasses, the microscopic constituents interact only weakly. In both
solids and liquids the interactions are strong and produce various degrees of spatial ordering.
In crystals and quasi-crystals, only discrete translations and rotations survive as symme-
tries. Customary liquids are homogeneous and isotropic phases (with unbroken translation
and rotation symmetries). Liquid crystals share certain properties with liquids, others with
solids. For example, in a nematic phase molecules are aligned along a spatial direction in
a translationally invariant way, while in a smectic phase there is ordering along layers but
molecules need not be pointing along the same direction. Electric and magnetic properties
provide further and finer distinctions between possible phases, giving rise to conductors,
semi-conductors, super-conductors, insulators, paramagnets, ferromagnets, and so on.
The goal of statistical mechanics is to explain quantitatively the physical properties of
these phases in terms of the dynamics of the underlying microscopic constituents. The time-
scales and length-scales over which we seek this information are typically macroscopic, on
the order of seconds and centimeters. The degrees of freedom corresponding to time-scales
and length-scales on the order of atomic processes will be treated effectively, or statistically.
6
1.1 Brief history
The development of statistical mechanics was driven by attempts to understand thermody-
namics from a dynamical microscopic point of view. Thermodynamics is a phenomenological
(or purely macroscopic) theory used to describe the transfer of heat, work, and chemical
constituents in and out of a macroscopic system. The development of thermodynamics goes
back to the early 19-th century, at a time when the concept of energy, and its conservation,
were still ill-understood and controversial. Sadi Carnot (1796-1832) uncovered some of the
physical principles that underly the functioning of steam-engines, and was led to distinguish
between reversible and irreversible processes. James Joule (1818-1889) extended the concept
of energy by include heat. In doing so, he clarified conservation of energy, and laid the
foundation for the first law of thermodynamics. Rudolf Clausius (1822-1888) clarified the
role of energy conservation in the Carnot-cycle, introduced the concept of entropy, and first
stated the second law of thermodynamics. The third law of thermodynamics was proposed
in 1905 by Walther Nernst (1864-1941).
The development of statistical mechanics also grew out of the kinetic theory of gasses.
Daniel Bernoulli (1700-1782) accounted for the pressure exerted by a gas on the walls of its
container in terms of collisions of molecules against the wall. The concepts of mean free path
and collision rate were introduced by Clausius, and incorporated by James Clerk Maxwell
(1831-1879) into a precise theory for the statistical distribution of velocities obeyed by the
molecules of a gas in equilibrium, the so-called Maxwell distribution. Ludwig Boltzmann
(1844-1906) generalized the Maxwell distribution to include external forces, thereby replacing
the purely kinetic energy of the Maxwell distribution by the total energy in the famous
Boltzmann factor eE/kT . The Boltzmann transport equation describes the kinetic dynamics
of gasses out of equilibrium; it may be used to study how equilibrium is reached in time.
The range of applicability of these kinetic theories is limited by the fact that its microscopic
constituents are invariably assumed to be weakly (or non-) interacting.
The kinetic theories of gasses provide a statistical molecular interpretation of thermody-
namic quantities and phenomena. In particular, Boltzmanns famous relation between the
entropy S of a macro-state and the number of micro-states accessible to this macro-state,
S = k ln
provides one of the fundamental links between a thermodynamic quantity S, and a statis-
tical mechanical quantity . Understanding exactly what is being counted by requires
introducing the concept of an ensemble. The kinetic theory led Maxwell and Boltzmann to
introducing ensembles for gasses of non-interacting molecules.
Josiah Willard Gibbs (1839-1903) held the first professorship in theoretical physics in the
United States, at Yale University. Although his work on statistical mechanics (a term he
7
coined) and on vector calculus (which he invented) were carried out in relative isolation, its
influence on 20-th century physics has been profound. Amongst many other contributions,
Gibbs proposed to think of ensembles in an abstract but generally applicable formalism,
which allows for the counting of in a systematic statistical framework. Twenty-five years
before the discovery of quantum mechanics, Gibbs showed that counting molecules as dis-
tinguishable constituents leads to the Gibbs paradox. The paradox evaporates as soon as
constituents are treated quantum mechanically, and counted as indistinguishable. The Gibbs
approach to ensembles will be explained in detail and used extensively during this course.
The discovery of quantum mechanics through the analysis of the spectrum of black body
radiation by Max Planck (1858-1947) revolutionized physics in general and statistical me-
chanics in particular. Remarkably, the statistical counting formalism of Gibbs lent itself
perfectly to incorporating the changes brought about by the quantum theory. Theses fun-
damental changes are as follows,
1. Quantum states are discrete, countable entities;
2. Identical quantum constituents (they could be electrons, protons, atoms, or molecules)
are indistinguishable from one another and must be counted as such;
3. Integer spin particles obey Bose-Einstein statistics;
4. Half-integer spin particles obey Fermi-Dirac statistics.
Satyendra Bose (1894-1974) pioneered the quantum counting for photons, while Albert Ein-
stein (1879-1955) extended the method to (integer spin) atoms and molecules. Enrico Fermi
(1901-1954) and Paul Dirac (1902-1984) invented the modification needed to treat electrons,
protons, and other spin 1/2 particles. A systematic revision of the principles of statistical
mechanics, required by quantum mechanics, was pioneered by Lev Landau (1908-1968) and
John von Neumann (1903-1957), who introduced the tool of the density matrix. Landau
developed a systematic theory of second order phase transitions, and obtained general pre-
dictions for the critical behavior near the phase transition point. Landau also invented a
general theory of quantum liquids well-suited to describing liquids of bosonic or fermionic
excitations, and applied it successfully to explaining the superfluidity of liquid 4 He.
In the second half of the 20-th century, many developments of statistical mechanics went
hand in hand with progress in quantum field theory. The functional integral of Dirac and
Feynman provided a general starting point for the statistical mechanics of fields. Statistical
mechanical methods, such as block spin methods, were applied to the problem of renor-
malization in quantum field theory, while the renormalization group methods of field the-
ory provided systematic approximation methods in statistical mechanics. Kenneth Wilson
(1936-present) fully developed the interplay between quantum field theory and statistical me-
chanics to derive a physical understanding of renormalization, and to usher in a detailed and
calculable approach to the theory of critical phenomena and second order phase transitions.
8
In the 1970s, Jacob Bekenstein (1947-present) put forward a surprising and bold hy-
pothesis, namely that black holes, when considered quantum mechanically, must have a
definite entropy. Further evidence for this hypothesis was supplied by Stephen Hawking
(1942-present), who showed that the objects we refer to classically as black holes actually
emit radiation due to quantum effects. The spectrum of Hawking radiation is that of a black
body, which implies that a black hole must have temperature in addition to entropy, and
thus behaves as a thermodynamic objects. Intuitively, the assignment of an entropy makes
physical sense when one considers that micro-states living inside the black hole horizon are
shielded away from us and are thus accessible to us only in the form of a macro-state. From
a statistical mechanics point of view, the question thus arises as to what these micro-states
inside the event horizon are. For very special types of black holes, the answer is provided by
string theory, and the micro-states are found to be certain branes, namely spatially extended
objects of various dimensions that generalize membranes and strings. But the nature of the
micro-states for a regular black hole, like Schwarzschild is not understood to date.
9
1.2 Units and physical constants
The following values of physical constants may be useful.
A = 108 cm
1
1 erg = 1 g cm2 sec2
1 erg = 6.241 1011 eV
1 erg = 107 J
1 eV = 1.6022 1019 J
1 eV = 1.6022 1012 erg
1 eV = 1.1605 104 K
1 cal = 4.184 J
1 Btu = 1055 J
10
2 Statistical Physics and Thermodynamics
In this section, we shall introduce some of the basic methods of statistical mechanics and
their implications for thermodynamics in a relatively informal way. In the subsequent section,
more complete definitions and applications of statistical ensembles will be introduced and
applied. In as much as possible, we shall discuss the cases of classical mechanics and quantum
physics in parallel.
11
a function of E, and denoted (E). If the total volume V is used in addition, becomes a
functions of two variables (E, V ), and so on. The more thermodynamic variables we use,
the more precise our specification of the macro-state will be. In practice, macro-states will
be described by just a few thermodynamic variables. Thus, specifying a macro-state will
always amount to omitting a large amount of information about the system.
The key logical connection between statistical mechanics and thermodynamics is made
by Boltzmanns formula for entropy S,
S = k ln (2.1)
E = E1 + E2 (2.2)
12
System 1 System 2
E1 V1 E2 V2
and is fixed since S is assumed to be isolated. In general, the number of accessible micro-
states (E) of S at energy E (recall the volumes are fixed) depends on the interactions
between S1 and S2 and is not readily calculable. We shall consider here an idealized situation
where the interactions between S1 and S2 are neglected (which can be justified in part since
they occur through a surface and have small effects in the bulk). As a result, the system S
at energy E may be in any of the macro-states (E1 , E2 ) with E1 + E2 = E, with the following
corresponding number of micro-states,
We see that the partition of S into the subsystems S1 and S2 which are in mutual thermal
contact provides us with one macroscopic parameter which is not yet fixed, namely the energy
E1 of the subsystem S1 . The state of equilibrium will be achieved when 0 is maximal for
given E. In the process of evolving towards equilibrium, the free parameter E1 will adjust
to produce the maximum possible value for 0 (E1 , E2 ), so that we have,
1 S(E)
= (2.5)
T E
the equilibrium condition (2.4) may be re-expressed in terms of the temperature T of the
two systems,
T1 = T2 (2.6)
13
namely, at equilibrium, the temperatures of the two subsystems coincide.
Extending the above derivation, assume that the systems S1 and S2 are not only in
thermal contact, but that also the wall separating them is movable, so that the volumes V1
and V2 may vary, all the while keeping the total volume V of the system fixed,
V = V1 + V2 (2.7)
In addition, let us assume that one or several further conserved quantities N , such as mo-
mentum, angular momentum, or the conserved numbers of various species of particles, may
also be exchanged between the two systems, all the while keeping their total value fixed,
N = N1 + N2 (2.8)
By the same arguments as were used for energy exchange, equilibrium of the system S will
be characterized by the following equalities,
ln 1 (E1 , V1 , N1 ) ln 2 (E2 , V2 , N2 )
= 0
V1 V2
ln 1 (E1 , V1 , N1 ) ln 2 (E2 , V2 , N2 )
= 0 (2.9)
N1 N2
In terms of entropy S, temperature T , pressure P , and chemical potential , defined by,
S(E, V, N ) 1
=
E
V,N T
S(E, V, N ) P
=
V
E,N T
S(E, V, N )
= (2.10)
N
E,V T
the equilibrium conditions become,
T1 = T2
P 1 = P2
1 = 2 (2.11)
which constitute, of course, the well-known equations of thermodynamical equilibrium. Thus
equilibrium guarantees that T, P, are constant throughout the system.
These relations may be further generalized to the case of more than two subsystems in
thermal contact. If S may be subdivided into three subsystems S1 , S2 , S3 , which are all in
mutual thermal contact, then the equilibrium conditions between S1 , S2 on the one hand, and
S2 , S3 on the other hand, will imply that the subsystems S1 and S3 are also in equilibrium
with one another. This fact is referred to as the 0-th law of thermodynamics in Kardar.
14
2.3 The thermodynamic limit, extensive and intensive variables
Putting together multiple identical copies of a given macro-system allows us to scale the
energy E, volume V , and particle number N , as well as any other conserved quantities, by
a common scale factor ,
E E V V N N (2.12)
This scaling will be reliable in the limit where boundary and surface effects can be neglected,
an approximation which is expected to hold in the limit where . This limit is referred
to as the thermodynamic limit.
Thermodynamic variables which scale linearly with are referred to as extensive, while
those which are untransformed under are referred to as intensive. For example, the en-
ergy density E/V , and the number density N/V are intensive. Given that thermodynamic
equilibrium sets equal T , P , and across the bulk of a substance, these quantities should
be intensive. As a result of the defining relations in (2.10), the entropy is expected to be an
extensive thermodynamic variable.
15
All changes in the internal energy which are not work will be lumped together as heat,
and denoted by Q. Thus, heat transfer may be thought of as associated with changes in
the kinetic energy of the system.
dE = Q W (2.15)
The minus sign in front of work is a matter of convention, and results from the fact that we
have defined the work done *by* the system (as opposed to the work done *on* the system).
The notations Q and W are used here to stress the following important distinction with
dE, dS, dV etc. The thermodynamic variables V, E, N, S, are all thermodynamic state
functions, and depend only on the (equilibrium) macro-state under consideration. Thus, the
corresponding differentials dV, dE, dN, dS, are exact differentials, and their integration
depends only on the initial and final macro-states of the system. Heat and work are not, in
general, thermodynamic state functions, and integrals of the differentials Q, W do depend
on the path followed to go from one macro-state of the system to another. The differentials
Q and W are not closed (an thus not exact); the notation is used to indicate this (Clausius
introduced this distinction, and used the notation d rather than ).
When the thermodynamic changes are reversible, we may combine formulas (2.14) and
(2.15) which are generally valid, with formula (2.13) which is valid only for reversible pro-
cesses, to obtain a relation between heat and entropy,
Q = T dS (2.16)
In view of its very derivation, this relation will hold only for reversible processes. For
irreversible processes, the relation does not hold. Instead, we need to appeal to the second
law of thermodynamics to clarify the situation.
16
These statements may be translated into an inequality with the help of a theorem by Clau-
sius (proven, for example, in Kardar.) In a general thermodynamic process (reversible or
irreversible), the entropy change dS for given heat transfer Q and temperature will be larger
that Q/T , so that for any cyclic thermodynamic process we have,
I
Q
0 (2.17)
T
Here, Q is the heat increment supplied to the system at temperature T . For any reversible
processes we have equality, while for any irreversible process we have strict inequality.
dE = T dS P dV + dN (2.18)
(1) Applying the Euler equation for homogeneous functions of degree 1 to the entropy in the
thermodynamic limit, and using the defining equations of (2.10), we derive the relation,
E = T S P V + N (2.19)
Taking the total differential of this relation, and then eliminating dE with the help of (2.18)
gives the Gibbs-Duhem relation,
SdT V dP + N d = 0 (2.20)
(2) Next, we provide alternative formulas for T, P , and as derivatives of the internal energy
E rather that of the entropy S. The derivations are standard, and extend to many other
thermodynamic relations. It is immediate from (2.18) that we alternatively have,
! ! !
E E E
T = P = = (2.21)
S V,N
V S,N
N S,V
F = E TS (2.22)
dF = SdT P dV + dN (2.23)
17
It follows that S, P , and may be represented as derivatives of F with respect to the variables
respectively T, V , and N ,
! ! !
F F F
S= P = = (2.24)
T V,N
V T,N
N T,V
Again, alternative formulas are available in terms of derivatives of the internal energy, or
free energy. We quote the alternative formulas for CV and CP ,
! ! !
E E V
CV = CP = +P (2.26)
T V,N
T P,N
T P,N
(E, V, N ) V N
S(E, V, N ) N ln V (2.27)
P V = nRT = N kT (2.28)
18
2.7 The simplest ideal gas
Not all ideal gasses are the same. This is due to the fact that the number of degrees of
freedom of each microscopic constituent by itself may vary. Each microscopic constituent in
any ideal gas will have three translational degrees of freedom, namely the three components
of momentum. Clearly, the simplest ideal gas corresponds to the case where these are the
only degrees of freedom of each microscopic constituent. More generally, the atoms and
molecules in a general ideal gas will have also rotational and internal degrees of freedom,
with corresponding quantum numbers. In this section, we shall treat only the simplest ideal
gas, and postpone treatment of more complicated cases.
To also extract the dependence on energy and particle number, we need a more detailed
model, specifying in particular whether the gas is relativistic or not, and quantum mechanical
or classical. We shall work with the non-relativistic quantum version, for non-interacting
particles confined to a square box of size L and volume V = L3 , with Dirichlet boundary
conditions (or equivalently an infinite positive potential outside the box). The micro-states
of the system are specified by the quantum state of each particle i = 1, , N by three
non-negative integers ni = (nix , niy , niz ). The energy of a particle in this state is given by,
2 i 2
2h
(nix , niy , niz ) = (n x ) + (ni 2
y ) + (n i 2
z ) (2.29)
2mL2
The energy of the system in a micro-state specified by the ni is given by,
N
(nix , niy , niz )
X
E= (2.30)
i=1
We see that the counting of states depends on N , and on the combination EL2 = EV 2/3 but
not on E and V separately. As a result, the functional dependence of S can be simplified,
since it may be expressed in terms of a function s of only 2 variables,
Therefore, the derivatives of S with respect to E and V are related to one another,
S 3 S
E V =0 (2.33)
E V,N 2 V E,N
19
so that one derives the following formulas,
2E 3
P = E= N kT (2.34)
3V 2
The second relation reproduces the classic result that the average thermal energy per particle
is 23 kT , or the average energy per degree of freedom is 12 kT . Finally, combining the earlier
result of (2.27) with that of (2.32), we derive the complete dependence on E as well,
V E 3/2
!
S(E, V, N ) = N ln + s(N ) (2.35)
N 5/2
where s(N ) depends only on N . If S is to be an extensive quantity, then s(N ) should be a
linear function of N , given by s(N ) = s0 N for some constant s0 .
20
and converting the result of (2.38) to a formula for the entropy, we find,
" 3/2 #
V mE 3
S(E, V, N ) kN ln + kN + kN ln N (2.40)
N h2 N
3 2
Note that the argument of the logarithm is properly dimensionless, and forms an intensive
combination of E, V, N . The first two terms are extensive contributions, as one would expect
for the total entropy, and their dependence on E and V is in accord with the results derived
on general scaling grounds in equation (2.35). The last term is consistent with the form of
(2.35) as well, but fails to be properly extensive. It is useful to re-express E in terms of T
using the second equation in (2.34), and we find,
!3/2
V mkT 3
S(E, V, N ) = kN ln + kN + kN ln N (2.41)
N h2
2 2
The failure of the last term to be properly extensive is referred to as the Gibbs paradox.
21
are different for the two isolated systems S1 and S2 prior to thermal contact. Thus, upon
producing thermal contact, the total system will not in thermal equilibrium.
Let us explore the predictions of formula (2.41), in order to compare the entropies S of
the systems S for = 1, 2,
!3/2
1 mkT 3
S = kN ln + kN + ln N (2.42)
2h2 2
The mixing entropy increase S S1 S2 may be evaluated using the Stirling formula,
(N1 + N2 )!
S S1 S2 = k ln (2.44)
N1 ! N2 !
a result which clearly cries out for a combinatorial interpretation.
It was our initial treatment of the identical particles as distinguishable (namely each one
being labeled by (pi , qi )), that produced the paradox.
The heat capacities CV , CP for the ideal gas may be evaluated using nothing more than
the ideal gas laws of (2.34), and we find,
3 5
CV = N k CP = N k (2.46)
2 2
22
Evaluating the internal entropy as a function of entropy allows us to compute the chemical
potential for the ideal gas, and we find,
h2 N 5/3
3 2S 5
E = 2/3
exp
mV 3kN 3
!3/2
V mkT
= kT ln (2.47)
N 2h2
3 1 (2.50)
This is the condition for a Boltzmann (or sufficiently dilute) gas. For 3 1, we will need
to appeal to the full quantum statistics of Bose-Einstein or Fermi-Dirac statistic instead.
1. Quasi-static: if the external conditions change slowly enough to keep the system in
equilibrium throughout the transformation process.
23
2. Reversible: a transformation producing no increase in total entropy. (A reversible
process is quasi-static, but the converse doe snot hold.)
3. Adiabatic: a transformation in which only the external conditions on the system are
changed, but no heat is transferred: Q = 0. As a result, we have dE = W .
(Note that Landau and Lifshytz use terminology where the term stands for what we
understand to be reversible adiabatic; see next entry.)
4. Reversible adiabatic: combining the two properties, one shows that the entropy of the
system remains unchanged. (For the ideal gas, this means that V T 2/3 or equivalently
P V 5/3 are constant, for fixed particle number N .)
5. Isothermal: a process at constant temperature. (For an ideal gas, the internal energy
is then constant, and the product P V is constant, for fixed particle number N .)
24
3 Classical Statistical Ensembles
In the preceding section, we have succeeded in counting the number of micro-states for an
ideal gas of non-interacting non-relativistic particles confined to a box. As soon as interac-
tions are turned on, however, an exact counting of micro-states in the thermodynamic limit
can be carried out only very rarely. Therefore, a general formalism of the statistical approach
to arbitrary systems is needed which, in particular, will permit a systematic calculation of
perturbative approximations.
The general formalism of statistical mechanics is based on the theory of statistical en-
sembles, a concept which dates back to Maxwell and Boltzmann for gasses, and which was
articulated in wide generality by Gibbs. We shall begin with ensembles in classical mechan-
ics, where the formulation may be set and understood in more familiar terrain, and introduce
the density function. Ensembles in the quantum case will be introduced in the subsequent
section, where we shall also distinguish between pure and mixed ensembles, and introduce
the density matrix.
25
depicted schematically in Figure 2. Some of these leafs may be empty sets; this happened
for example for negative E with the harmonic oscillator Hamiltonian, which is non-negative
throughout P.
E4
E3
E2
E1
may offer a better prospect. Although the time-averaging approach may seem natural, it
has a number of drawbacks. The time-averaging of (3.3) may depend on the initial time
t, and on the initial conditions imposed on the time-evolution. Ideally, such data will be
washed out for large enough T provided we have strong enough interactions. Unfortunately,
characterizing the behavior of a system over long time intervals is very complicated, and the
question whether the initial conditions will be or are will not be washed out for large enough
T introduces a difficult dynamical complication. In particular, during its time evolution, the
system may or may not reach all of the allowed phase space, a problem that is the subject
of the ergodic theorem of dynamical systems. In fact, for free particles (say on a torus with
26
period boundary conditions), or for systems with maximal numbers of conserved quantities,
we know that not all of the phase space allowed by total energy will be reached.
2
To emphasize this distinction, we shall denote the number of particles or subsystems by N , but the
number of non-interacting systems collected in an ensemble by N .
27
(a) (b) (c)
Figure 3: Systems in phase space P: (a) a single system corresponds to a point, its time-
evolution gives a curve in P; (b) a cloud of points defines an ensemble; gives the time
evolution of an ensemble.
where dpdq stands for the canonical volume form on phase space. The interpretation is that
dw counts the number of points of the ensemble contained within an infinitesimal volume
dpdq surrounding the point (p, q) at time t. Equivalently, the number of phase space points
ND of the ensemble contained in a finite domain D, and the total number of points N in
the entire ensemble, are given by,
Z Z
ND = dpdq (p, q; t) N = dpdq (p, q; t) (3.5)
D P
28
The ensemble average f(t) of any function f (p, q) on phase space is defined by
1 Z
f(t) = dpdq (p, q; t) f (p, q) (3.6)
N P
Since all ensemble averages are unchanged under multiplication of by a positive constant,
we will throughout use the normalized distribution function (p, q; t)/N instead of itself,
and continue to use the notation (p, q; t) to signify the normalized distribution function. In
the standard language of probability theory, the normalized distribution function is then a
probability distribution.
The time-averaging procedure of (3.3) on a single system reaching equilibrium, may be
extended to a time-averaging procedure on the entire ensemble. To see how this works,
consider an infinitesimal volume of size dpdq surrounding a point (p, q) in phase space. At
time t, various individual points in the cloud of phase space points representing the ensemble
will be inside the volume dpdq. This number is given by N dpdq (p, q; t), for a normalized
distribution function . Thus, the time-average of (3.3) may be expressed as follows,
1 Z t+T 0 0 0 1 Z t+T 0 Z
lim dt f (p(t ), q(t )) = lim dt dpdq (p, q; t0 ) f (p, q) (3.7)
T T t T T t P
All dependence on T and t is now contained in the t0 -integral over the density function.
Clearly, equilibrium will be attained for large enough T provided the combination
1 Z t+T 0
dt (p, q; t0 ) (3.8)
T t
tends towards a finite, t-independent limit as T . This will be realized provided the
following equilibrium condition holds at all times t,
(p, q; t)
=0 (3.9)
t
Note that the general density function (p, q; t) introduced through the method of ensembles
need not correspond to a system in equilibrium. The added advantage of the ensemble
formulation is that we have a clear and clean characterization of equilibrium given in terms
of the density function by condition (3.9).
29
(p, q; t), so that the number of ensemble points inside an infinitesimal volume dpdq of phase
space is given by N dpdq (p, q; t). The number of ensemble points ND contained inside a
domain D (of dimension 2s), is given by the integral over the density function as in (3.5),
Z
ND (t) = N dpdq (p, q; t) (3.10)
D
Since depends explicitly on t, the number ND will in general depend on t as well. Conser-
vation of the number of points is expressed by the conservation equation, which relates the
density function to the current density (pi , qi ) on all of phase space,
s
!
X
+ (pi ) + (qi ) = 0 (3.11)
t i=1 pi qi
where the sum of partial derivatives gives the divergence of the current (pi , qi ). Using
Hamiltons equations to eliminate pi and qi , and using the relation,
H H
pi + qi = + =0 (3.12)
pi qi pi qi qi pi
we find,
s
!
X
+ pi + qi = 0 (3.13)
t i=1 pi qi
This conservation equation may be recast in two equivalent ways. First, given that is a
function of p, q, t, we see that (3.13) expresses its total time-independence,
d
=0 (3.14)
dt
as the Hamiltonian system evolves in time according to Hamiltons equations. This means
that the density does not change in time, so that the ensemble evolves as an incompressible
fluid does. Alternatively, (3.13) may be expressed with the help of Poisson brackets {, }. We
normalize the generalized coordinates canonically by {pi , pj } = {qi , qj } = 0 and {qi , pj } = ij .
Equations (3.13) and (3.14) are then equivalent to,
+ {, H} = 0 (3.15)
t
We stress that the above results hold for any distribution function, whether it describe an
ensemble in equilibrium or out of equilibrium.
30
3.6 Equilibrium ensembles and distribution functions
The construction of equilibrium ensembles, and their associated distribution function, is
guided by two key principles. The first is the condition for equilibrium in (3.9). The second
is Boltzmanns assumption of equal a priori probabilities. We explain each these below.
If a distribution function describes an ensemble in equilibrium, then by (3.9) and (3.15),
it must satisfy the following conditions which are equivalent to one another,
=0 {, H} = 0 (3.16)
t
Both relations provide a concrete characterization of equilibrium density functions, and thus
of ensembles in equilibrium. We shall concentrate on conservative systems, in which case H
has no explicit time dependence. The second equation in (3.16) then instructs us that is a
conserved quantity. Generically, a conservative Hamiltonian exhibits only a single conserved
quantity, namely the total mechanical energy E = H(p, q), so that will be a function of
the Hamiltonian only. When further conserved quantities, such as total momentum, total
angular momentum, particle number, electric charge, exist, then will also depend on the
corresponding conserved mechanical functions.
The principle of equal a priori probabilities postulates that, if a certain macro-state has
been completely specified by a certain number of thermodynamic variables, and is accessible
by a number of micro-states, then the probability for finding the system in any one of
these micro-states is the same for all micro-states, and thus equal to 1/. As this probability
distribution does not favor any one micro-state above another, it is referred to as an unbiased
probability distribution.
It is readily seen that these expressions will be finite and make physical sense only if the
volume of phase space is finite, for example when P is compact. For most systems of
physical interest, this will not be the case. Thus, in classical mechanics, this ensemble is not
that useful. The uniform ensemble will be useful later, however, when considering certain
quantum mechanical systems.
31
3.8 The micro-canonical ensemble
In the micro-canonical ensemble, the total (or internal) energy E is used to specify the macro-
states completely. All micro-states with total energy E have equal weight, in accord with the
Boltzmann prnciple of equal a priori probabilities. The motion of every system is restricted
to the subspace PE of energy E of the full phase space P, and the density function E for
energy E is supported on PE with uniform weight. As a result, the normalized distribution
function for the micro-canonical ensemble may be written down explicitly,
1 Z
E (p, q) = H(p, q) E N (E) = dpdq H(p, q) E (3.18)
N (E) P
The micro-canonical ensemble average of a phase space function f (p, q) is given by,
Z Z
f = dpdq E (p, q) f (p, q) dpdq E (p, q) = 1 (3.19)
P P
The need for the presence of a non-purely-mechanical normalization factor N0 in the entropy
formula was already made clear by considering the simplest ideal gas, which showed that N0
has a quantum component involving h . It is clear that this factor is needed also from the
consideration of dimensions: the total number of micro-states (E) is dimensionless, but
N (E), defined in (3.18) has dimensions. Note that (E) is defined in an intrinsic geometric
manner, and invariant under canonical transformations on p, q.
In practice, it will often be more convenient, and more physical, to define the micro-
canonical ensemble for a finite range of energies, instead of strictly at one value of energy.
This treatment will be especially valuable in the quantum case, where energy levels are
usually discrete. To do so, we define a shell of energies, in the range [E, E + ], with
1 if E H(p, q) E +
1
E, (p, q) = (3.21)
(E, )
0 otherwise
The ensemble averages are defined accordingly as in (3.19) with E replaced with E, .
32
The canonical ensemble may be obtained from the canonical ensemble by changing vari-
ables from fixed internal energy E to fixed absolute temperature T . The density function
for the canonical ensemble (p, q; T ) is the Laplace transform of the density function E (p, q)
for the micro-canonical ensemble, and is given by,
1 H(p,q)/kT
(p, q; T ) = e (3.22)
Z(T )
The normalization factor Z(T ) is referred to as the partition function, and is given by,
Z
Z(T ) = dpdq eH(p,q)/kT (3.23)
P
= kT 2 ln Z(T )
E=H (3.25)
T
To compare this result with formulas in thermodynamics, we express it in terms of the free
energy F , with the help of F = E T S and S = F/T to obtain,
F
E = T 2 (3.26)
T T
Comparison of (3.6) and (3.5) shows that the free energy is given by,
Z(T )
F (T ) = kT ln (3.27)
Z0
which is one of the most fundamental formulas of correspondence between statistical me-
chanics and thermodynamics, in the canonical ensemble. The constant Z0 is not determined
by classical statistical mechanics. Its presence shifts the entropy by a constant. We know
from the example of the simplest ideal gas that Z0 is quantum mechanical in origin. More
importantly, one needs the constant there to make the argument of the logarithm to be
properly dimensionless. Finally, we have already established in the ideal gas calculation
33
that the quantum indistinguishability of identical particles requires the presence of an extra
combinatorial factor N !. Thus, the correct constant is found to be,
Z0 = (2h)s N ! (3.28)
This result may be shown in all generality by first using the functional integral formulation
of the quantum problem and then taking its classical limit.
An alternative derivation makes use of the fact that the normalized density function is
a probability distribution. As appropriate for the canonical ensemble, we assume that the
density function only depends on energy, namely on the Hamiltonian of the system. Now
consider two independent systems (each system having its own phase space), with Hamilto-
nians H1 , H2 , and corresponding density functions (H1 ) and (H2 ). The Hamiltonian of the
total system is then H = H1 + H2 . The probability distributions (H1 ) and (H2 ) are prob-
abilistically independent. Therefore, as always with probability, the probability distribution
of the combined system (H) must be the product,
But since we have H = H1 + H2 , this implies that the dependence of (H) on H must be
exponential. This gives the Canonical distribution with the Boltzmann factor.
(E)0 (E 0 )
X
tot (Etot ) = (3.30)
E
where E and E 0 are the energies of the systems S and S 0 respectively, and the total energy
is given by Etot = E + E 0 . Labeling all the micro-states of S by the discrete label n, we have
34
E
E
where E,En is the Kronecker delta. It is important to stress that the sum over n is to run
over all distinct states, not just over all allowed energies; in particular, there may be several
states of S with energy En , and the degeneracy of the energy level En must be properly
included in the sum. Using (3.31) and (3.31), formula (3.30) may be cast in terms of a sum
over all micro-states of S,
1 0
0
X X
tot (Etot ) = (Etot En ) = exp S (Etot En ) (3.32)
n n k
Now, we use the fact that the heat bath system S is huge compared with S, so that we can
assume En /Etot 1, and expand in this quantity,
S 0 (Etot ) 1 2 2 S 0 (Etot )
S 0 (Etot En ) = S 0 (Etot ) En + En 2
+ O(En3 ) (3.33)
Etot 2 Etot
S 0 (Etot ) 1
= (3.34)
Etot T
gives the temperature T of the heat bath. Since S 0 (Etot is extensive, the second derivative
term is of order En /Etot and may be neglected compared to the first derivative term. Thus,
the total number of states becomes,
0
tot (Etot ) = eS (Etot )/k eEn /kT
X
(3.35)
n
35
Using now the principle of equal a priori probabilities, we know that every state n with
energy En is equally likely, so we find the number of states of the total system accessible to
micro-state n to be given by
0
eS (Etot )/k eEn /kT (3.36)
As a result, the probability of finding the whose system in a state such that S is in the
micro-state n is given by the ratio,
eEn /kT
eEn /kT
X
n = Z(T ) = (3.37)
Z(T ) n
Note that in these probabilities, all references to the heat bath has vanishes, except for its
temperature. In classical mechanics, the possible micro-states are labelled by the generalized
coordinates (p, q), the corresponding energy being given by the Hamiltonian H(p, q). The
probability becomes the normalized density function, so that we recover (3.22).
We shall assume that the generalized forces f on the system are kept fixed in time, as
is required for an equilibrium ensemble. The work done on the system by a generalized
36
displacement x is then linear in x, and given by W = f x. The corresponding change in the
Hamiltonian is given by H(p, q) H(p, q) f x(p, q). (If several generalized forces f are
P
applied, we will instead have a sum f x (p, q).) The distribution function is given by,
1
(p, q; T, f ) = e(H(p,q)f x(p,q))/kT (3.38)
Z(T, f )
The normalization factor Z(T, f ) is referred to as the grand canonical partition function,
Z
Z(T, f ) = dpdq e(H(p,q)f x(p,q))/kT (3.39)
P
In particular, the ensemble average of the generalized displacement variable x will be denoted
by X = x. The Gibbs free energy is defined by,
G(T, f ) = E T S f X (3.41)
When the exchange of particles is permitted, the corresponding Gibbs ensemble is sometimes
referred to as the grand-canonical ensemble.
37
4 Applications of Classical Statistical Ensembles
4.1 Application I: The Maxwell distribution
The density function in the canonical ensemble may be readily used to derive the Maxwell
distribution for the momenta or velocities in a classical ideal gas. The particles being non-
interacting, we can focus on the velocity distribution of a single particle, the case of N
particles being given by multiplying the probabilities and partition functions for single-
particles. The single-particle partition function Z1 is given by,
Z
d3 p d3 q H(p,q)/kT
Z1 (T ) = e (4.1)
(2 h)3
The Hamiltonian is that for a free non-relativistic particle H(p, q) = p2 /2m with no depen-
dence on q. Thus, the q-integral may be carried out and yields the space-volue V . Changing
variables from momentum to velocity, p = mv, we have,
m3 V Z 3 mv2 /2kT
Z1 (T ) = d ve (4.2)
(2h)3
It is straightforward to evaluate the integral, and we find,
!3/2
m3 V 2kT
Z1 (T ) = 3
(4.3)
(2h) m
This is the Maxwell distribution. It is readily checked that the average velocity hvi vanishes,
while the average velocity square, and thus the average energy, are given by,
Z
3kT 3
hv2 i = dv Max (v) v 2 = hHi = kT (4.5)
0 m 2
as expected from the equipartition theorem.
38
effect, at the same level as their spin. The magnetic coupling to electric charge is governed
by the minimally coupled Hamiltonian for charged particles,
N
X 1 2
H(p, q) = pi ei Ai (q) + U (q) (4.6)
i=1 2mi
Here, U (q) is any potential contribution which depends only on the generalized positions
q; Ai (q) is the vector potential acting on particle i; and ei , mi are respectively the electric
charge and mass of particle i. Note that the vector potential may include the effects of an
external magnetic field, as well as the effects of internal magnetic interactions. We begin by
computing the partition function,
N
1 Z Y
Z(T ) = d3 pi d3 qi eH(p,q)/kT (4.7)
h)3N N ! i=1
(2
All magnetic effects have vanished from the partition function. This result is referred to as the
Bohr - Van Leeuwen theorem. In particular, all thermodynamic functions are independent of
external magnetic fields. Thus, there is no diamagnetism in classical statistical mechanics. In
as far as intrinsic magnetic moments are purely quantum effects, there are strictly speaking
then also no paramagnetic effects classically.
39
quantum states, but these start to be relevant only at relatively high temperatures. For
example, the first excited state of Hydrogen is 13.6 eV above the ground state, corresponding
to a temperature of 105 K. Polyatomic molecules do, however, have low lying extra degrees
of freedom, such as rotational and vibrational, which are relevant at room temperatures.
Consider a simple model of an ideal gas composed of diatomic molecules. The atoms in
each molecule can rotate in the two directions perpendicular to the symmetry axis of the
molecule with moment of inertia I, while we will neglect the moment of inertia of rotations
about the symmetry axis. The atoms in each molecule can also vibrate by changing their
relative distance, and we shall model these vibrations by a harmonic oscillator of frequency
. The translational, rotational and vibrational degrees of freedom are all decoupled from
one another, so that their Hamiltonians add up, and their phase space measures factorize.
Thus, the partition function of an individual molecule (recall that the gas is ideal, so the
total partition function is given by the N -th power and a prefactor of 1/N !) factorizes,
p2 p2
Hrot = + (4.13)
2I 2I sin2
Carefully including the measure with the standard normalization, we find,
Z
d d dp dp Hrot 2IkT
Zrot = e = (4.14)
(2h)2 2
h
Using (3.16), we derive the contribution from the rotational energy per molecule,
Erot = kT (4.15)
41
The free energy and internal energy for the vibrational mode are given by,
1
Fvib = kT ln Zvib = h + kT ln 1 eh
2
1 1
Evib = ln Zvib = h
+ (4.23)
2 eh 1
Evib T2 eTvib /T
Cvib = = vib (4.24)
T T 2 (eTvib /T 1)2
The contribution to the total specific heat from the N vibrational modes is simply CV =
N Cvib . The behavior of this function may be inferred from its asymptotic behavior,
T Tvib CV kN
O with(plots): 2
p1:=plot(1/x^2*exp(1/x)/(exp(1/x)-1)^2, x=0..3, labels=[x,
vib Tvib /TT
T T
C_V/Nk], color=[blue],thickness=2):
vib
O Z:=x->sum((2*n+1)*exp(-n*(n+1)/x), V C kN
n=0..infinity); 2
e (4.25)
Ck:=x->diff(x^2*diff(ln(Z(x)),x),x);
p2:=plot(Ck(x), x=0..3, color=[red],thickness=2):
T
p3:=textplot([1.4, 0.8, "vibrational x=T/T_vib"], align =
{above, right}, color=[blue], font=[TIMES,BOLD,13]):
which is clearly being reflected in a numerical plot of the function given in Figure 5. The
p4:=textplot([1.4, 1.1, "rotational x=T/T_rot"], align = {above,
right}, color=[red], font=[TIMES,BOLD,13]):
display(p1,p2,p3,p4);
large T contribution corresponds to the classical behavior found in the previous section.
N
K
n nC1
Z := x/ > 2 n C 1 e x
Below Tvib , the vibrational contribution to the specific heat turns off quickly.
n=0
d d
Ck := x/ x2 ln Z x
dx dx
rotational x=T/T_rot
vibrational x=T/T_vib
0.8
0.6
C_V
Nk
0.4
0.2
0
0 1 2 3
x
42
4.5 Quantum rotational modes
A similar story applies to the rotational modes. Their contribution to the specific heat will
be cut off at a characteristic temperature Trot as well since the energy levels of the quantum
rotor are again discrete. The energy levels of a (simplified) spherical model are given by
` = `(` + 1)h/2I with degeneracy 2` + 1, so that the partition function is given by,
X
`(`+1)Trot /T 2
h
Zrot = (2` + 1) e Trot = (4.26)
`=0 2Ik
Analytically, this sum would be given by an elliptic function, but we can easily plot the
function numerically, as shown in Figure 5. It is useful to get some order of magnitude
estimated for the transition temperatures. For molecular Hydrogen, H2 , we have,
These order of magnitudes demonstrate that quantum effects are relevant at relatively low
as well as at relatively high temperatures.
43
5 Quantum Statistical Ensembles
In this section, ee shall define ensembles in quantum mechanics, distinguish between pure
and mixed ensembles, and introduce the density matrix. We have set up the classical case
in such a way that it will resemble most closely the treatment of the quantum case. The
uniform, micro-canonical and canonical ensembles are then introduced for the quantum case.
Understanding both of these cases will turn out to be of fundamental importance in quantum
statistical mechanics, and we shall begin by discussing them in some detail.
44
piece of apparatus used is usually a polarizer, responsible for filtering a definite polarization
out of an unpolarized beam. Particles in a polarized beam are pure quantum states. In
particular, a polarized beam of Silver atoms in the Stern-Gerlach experiment should be
viewed as an ensemble of spins which are all in the same quantum state. This states in the
beam are said to be coherent and to form a pure ensemble.
What we have not yet provided in quantum mechanics is a mathematical description for
a beam which is unpolarized, or is said to be in a mixed state. The defining property of an
unpolarized beam of spin 1/2 particles is that measurement of the spin operator n S along
any direction n, with n2 = 1, gives h/2 with equal probabilities, namely 50% each.
We begin by showing that an unpolarized beam cannot be described mathematically by
a vector in Hilbert space. Let us assume the contrary, and represent the particles in the
unpolarized beam by a vector |i. The probability Pn for observing the spins in a quantum
state |n, +i, which is the eigenstate of n S with eigenvalue +h/2, would then be given by,
Pn = |hn, +|i|2 . To compute this probability, we express the normalized states |n, +i and
|i in a fixed orthonormal basis |i of the Hilbert space, parametrized as follows,
Pn = cos2 cos2 + sin2 sin2 + 2 cos sin cos sin cos(2 2) (5.4)
45
not in H. Since we make no observations on system B, the only natural way to extend SAz
to an operator of H is by using the identity matrix IB in HB ,
The expectation value of SAz in the pure quantum state |i is then given as follows,
1
h|SAz |i = h+, A|SAz |+, Ai + h, A|SAz |, Ai = 0 (5.7)
2
using the orthonormality of the states |, Bi. In fact, by the same reasoning, one shows
that we also have h|n S|i = 0 for any n.
Remarkably, we have established that observing a subsystem of a total system which is
in a pure quantum state produces quantum expectation values which are identical to those
of a mixture of pure quantum states. Thus, the problems of representing mixtures of pure
quantum states and the problem of observing only a subsystem are closely related to one
another, and neither of them permits a description by vectors in Hilbert space alone.
A mixture is an incoherent superposition of pure states which means that all relative phase
information of the pure states must be lost in the mixture. This is achieved by superimposing,
not the states |i i in Hilbert space, but rather the projection operators |i ihi | associated
46
with each pure state. The density matrix for an ensemble of N pure states |i i incoherently
superimposed with population fractions wi is defined by
N
X
= |i iwi hi | (5.9)
i=1
Since the superpositions involved here are incoherent, the weights wi may be thought of as
classical probability weights assigned to each population. We stress that the various different
pure states |i i are not required to be orthogonal to one another, since it should certainly
be possible to superimpose pure states which are not orthogonal to one another.
The following properties of the density matrix immediately result,
1. Self-adjointness : = ;
2. Unit trace : Tr() = 1;
3. Non-negative : h||i 0 for all |i H;
4. A state corresponding to is pure (rank 1) if and only if is a projection operator.
Conversely, any operator in H satisfying properties 1, 2, 3 above is of the form of (5.9)
with normalization (5.8). The above three conditions guarantee that can be diagonalized
in an orthonormal basis |i i, with real non-negative eigenvalues pi ,
X X
= |i ipi hi | pi = 1 (5.10)
i i
Note that |i i need not be proportional to |i i, and pi need not coincide with wi . Thus,
a given density matrix will have several equivalent representations in terms of pure states.
To check property 4, we note that if corresponds to a pure state, it is rank 1, and thus
a projection operator in view of property 2. Conversely, substituting the expression (5.10)
into the projection operator relation 2 = requires pi (pi 1) = 0 for all i. The condition
is solved by either pi = 0 or pi = 1, but Tr() = 1 requires that pi equal 1 for a single i.
Using the above properties of density matrices, we construct the general density matrix
for a 2-state system for which the Hilbert space is 2-dimensional. The density matrix is then
a 2 2 matrix and may be expressed in terms of the Pauli matrices = ( 1 , 2 , 3 ),
= (I + a )/2 (5.11)
where a is a real 3-vector. Self-adjointness and normalization are guaranteed by (5.11),
while positivity requires |a| 1. Pure states precisely correspond to the boundary |a| = 1,
while any density matrix with |a| < 1 corresponds to a mixed state which is not pure. The
ensemble average of the spin operator n S in the direction n is readily calculated,
h
hn Si = Tr n S = n a (5.12)
2
47
The above formula confirms that the state is unpolarized for a = 0, partially polarized for
0 < |a| < 1, and pure for |a| = 1. This gives a nice geometrical representation of the space
of density matrices for the two-state system.
A = hAi =
X
wi hi |A|i i = Tr(A) (5.13)
i
It follows from the properties of the density matrix that the ensemble average of any self-
adjoint operator is real. Note that formula (5.13) holds whether corresponds to a pure
state or a mixture.
The time-evolution of the density matrix may be derived from the Schrodinger equation
for pure states. In the Schrodinger picture, the time evolution of a quantum state |(t)i is
governed by a self-adjoint Hamiltonian operator H on H via the Schrodinger equation (5.2).
Assuming that the population fractions wi of (5.9), or equivalently the pi of (5.10), do not
change in time during this evolution, the density matrix (t) will obey the following time
evolution equation,
d
ih (t) = [H, (t)] (5.14)
dt
To prove this formula, one first derives the time evolution equation for |(t)ih(t)| and
then takes the weighed average with the time-indepenent population fractions. This time
evolution may be solved for in terms of the unitary evolution operator U (t) = eitH/h , by
48
|j , Bi an orthonormal basis in HB with i = 1, , dim HA and j = 1, , dim HB . The
states |i , Ai |j , Bi then form an orthonormal basis in H = HA HB , with the above
ranges for i, j. A normalized pure quantum state |i in H may be expressed in this basis by,
|ij |2 = 1
X X
|i = ij |i , Ai |j , Bi (5.16)
i,j i,j
where ij are complex coefficients. Let us now compute the purely quantum mechanical
expectation value of an observable A of the subsystem HA in the pure state |i H.
Making no observations in subsystem B may be represented by extending the operator A
on HA to an operator A on H by letting,
A A = A IB (5.17)
where IB is the identity operator in HB . This is a generalization of a similar extension we
used for the two-state system in (5.6). Evaluating the expectation value of A in the pure
state |i of H, we find,
ij i0 j 0 hi , A| hj , B|A IB |i0 , Ai |j 0 , Bi
XX
h|A|i = (5.18)
i,j i0 ,j 0
Using the orthogonality relation hj , B|j 0 , Bi = j,j 0 , and introducing the linear operator A
on HA , defined by,
|i0 , Ai i0 j ij hi , A|
X
A = (5.19)
i,j,i0
valid for any operator in O on H = HA HB , as well as the relation h|A|i
= TrH (A|ih|).
Applying both of these relations gives (5.20) with,
A = TrHB |ih| (5.22)
It is readily checked that the right hand formula for A reproduces (5.19).
49
5.7 Statistical entropy of a density matrix
The density matrix for an ensemble of n pure (orthonormal) states |i i occurring with
probability pi in the ensemble is given by,
n
X n
X
= |i ipi hi | pi = 1 (5.23)
i=1 i=1
When all probabilities are equal and given by pi = 1/n, the density matrix corresponds to
the uniform ensemble, and the entropy is given by Boltzmanns formula S = k ln n. Let us
now compute the entropy when the probabilities are not all equal.
To do so, we follow Gibbs again. Instead of considering just one system, consider a very
large number N of identical systems, and let us compute the entropy not just for one system,
but for all N of them. By the definition of what probability means for large N , we then see
that of those N systems, Ni = pi N will be in state |i i, with
n
X
Ni = N Ni 0 (5.24)
i=1
The number of micro-states that can realize a macro-state with the Ni systems in state |i i
is then given by a standard combinatorial formula,
N!
N = (5.25)
N1 ! Nn !
This expression and the corresponding entropy N for the ensemble of N systems may be
recast in terms of the probabilities pi and the total number of copies N , and we find,
n
!
X
SN = k ln N = k ln N ! ln (pi N )! (5.26)
i=1
Since N is very large, we may use the Sterling formula m! m ln m m, and find,
n n
!
X o
SN = k N ln N N pi N ln (pi N ) pi N (5.27)
i=1
The term N ln N N cancels since the pi sum to 1, and we are left with,
n
X
SN = kN pi ln pi (5.28)
i=1
We find, that the entropy for the N systems of the ensemble is simply proportional to N ,
as should be expected. Thus, it makes sense to extract the entropy S for a single system by
dividing SN by N and setting S = SN /N , so that we find,
n
X
S(p1 , , pn ) = k pi ln pi (5.29)
i=1
50
or equivalently in terms of the density matrix directly, we have,
S() = k Tr( ln ) (5.30)
Setting all probabilities equal pi = 1/n for the uniform ensemble, we recover S = k ln n.
Some basic properties of the statistical entropy are as follows.
1. The above construction of the statistical entropy does not assume equilibrium.
2. Positivity, S(p1 , , pn ) 0;
3. The minimum S = 0 is attained when all probability assignments are 0, except for a
single entry pj = 1. The entropy vanishes if and only if corresponds to a pure state.
4. The maximum Smax = k ln n is attained when all probabilities are equal, pi = 1/n.
5. Invariance under conjugation of the density operator by a unitary transformation. In
particular, the entropy is invariant under time evolution, under the assumption that
the probabilities pi remain unchanged in time;
6. Additivity upon combination of two subsystems which are statistically uncorrelated.
Let the systems be described by Hilbert spaces Ha and Hb , with density operators
a and b respectively, then the full Hilbert space is Hab = Ha Hb and the density
matrix for the combined system is ab = a b . The entropy is then additive,
S(ab ) = S(a ) + S(b ) (5.31)
7. Subadditivity upon dividing a system with Hilbert space Hab and density opera-
tor ab into two subsystems with Hilbert spaces Ha and Hb , and density matrices
a = TrHb (ab ) and b = TrHa (ab ) which are statistically correlated. The full density
operator ab is not the tensor product of a and b , in view of the non-trivial statistical
correlations between the two subsystems. Instead, one only has a strict inequality,
S(ab ) S(a ) + S(b ) (5.32)
The proofs of these properties will be developed in problem sets.
51
5.8 The uniform and micro-canonical ensembles
The ensembles defined in classical statistical mechanics may be generalized to the quantum
case. In this section, we shall briefly discuss the uniform and micro-canonical ensembles, but
spend most of our attention on the canonical ensemble whose usefulness is the greatest.
In the uniform ensemble, no macro-state information is given. By the principle of a
priori equal probabilities, the density matrix is proportional to the identity operator in the
full Hilbert space H of the quantum system, = IH /, where is the total number of states
which is equal to the dimension of H. Clearly, the uniform density matrix is time-indpendent
and corresponds to an equilibrium ensemble. The corresponding entropy is S = k ln .
In the micro-canonical ensemble, the total energy of the system is fixed to be E. The
density matrix has support only on eigenstates |Ei , i i of the Hamiltonian with energy
Ei = E, the i denoting degeneracies of the energy level Ei . In view of the principle of equal
a priori probabilities, the weights of all the states |E, i for various values of are all the
same, and given by the total number of states (E) at energy E. Thus, the density matrix
may be expressed in terms of the projection operator PE onto states of energy E,
PE 1 X
= = |E, ihE, | (5.33)
(E) (E)
The normalization constraint has been made explicit here in order to be able to liberate
this condition on . Equilibrium will be achieved by maximizing the entropy, subject to
the above constraints. Extremization subject to constraints is carried out mathematically
with the help of Lagrange multipliers, one for each constraint. Thus, we shall extremize the
combination, S() Tr kTr(H), and set,
0 = S() Tr kTr(H)
= k Tr( ln ) Tr() k Tr( H) (5.35)
52
The first term is calculated with the help of the mathematical identity,
Tr( ln ) = Tr ln + (5.36)
The operators , , IH , and H are all self-adjoint as well. Satisfying equation for all
then requires that ln + H is proportional to the identity or,
eH
= (5.38)
Z()
The parameter must be identified with inverse temperature = 1/kT . The partition
function Z() is determined by the normalization condition Tr = 1, and is thus given by,
Z() = Tr eH (5.39)
where the trace extends over the full Hilbert space. The internal energy may now be com-
puted directly from Z() by,
1 ln Z()
E() = Tr(H) = Tr H eH = (5.40)
Z()
F () = kT ln Z() (5.41)
Finally, we show that the thermodynamic definition of the entropy coincides with its sta-
tistical definition. To this end, we start with the thermodynamic relation E = F T S to
obtain the entropy with the help of (5.40) and (5.41),
1
S= Tr(H) + k ln Z (5.42)
T
Taking the logarithm of the operator equation (5.38), we have ln = H ln Z. Using
this formula to eliminate H in (5.42), we find,
S = kTr( ln ) (5.43)
53
5.10 Generalized equilibrium ensembles
The derivation of the Boltzmann weights and associated density matrix corresponds to the
canonical ensemble, in which only the energy of the system is kept constant. In the grand
canonical ensemble, both the energy and the number of particles in the system is kept
constant. More generally, we consider an ensemble in which the ensemble average of a number
of commuting observables Ai , i = 1, , K is kept constant. To compute the associated
density operator of this ensemble, we extremize with respect to variations in the entropy,
under the constraint that the ensemble averages Tr(Ai ) are kept constant. Using again
Lagrange multipliers i , i = 1, , K, we extremize
K
X
Tr ln i Tr(Ai ) (5.44)
i=1
where N is the number operator and is the chemical potential. Other observables whose
ensemble averages are often kept fixed in this way are electric charge, baryon number, electron
number etc.
54
6 Applications of the canonical ensemble
We shall now illustrate the use of the canonical (and micro-canonical) distributions on some
simple, but physically important systems.
Quantum mechanically,3 the magnetic moment is proportional to the spin s of the particle
(or more generally to the total angular),
si h
e
~i = g = (6.2)
h
2me c
where is the Bohr magneton of the dipole, given in terms of its mass me of the electron,
and the basic unit electric charge e (namely the charge of the electron), and the Lande factor
g. For the electron, we have g 2. For a given type of particle, the total spin s will be fixed,
and given by the eigenvalue h 2 s(s + 1) of s2 . The eigenvalues of ~ B are then given by,
gB m m = s, s + 1, , s 1, s (6.3)
Since the particles are non-interacting, the partition function Z is given by the N -th power
of the single particle partition function Z1 , which in turn is given by,
s
Z = (Z1 )N egBm
X
Z1 = (6.4)
m=s
sinh((2s + 1)x) 1
Z1 = x = gB (6.5)
sinh x 2
3
The classical treatment is parallel to the classical treatment of the electric dipole in an electric field,
which was solved in problem set 3.
55
The corresponding free energy F , internal energy E, and magnetization M are given by,
F = kT N ln Z1
ln Z1
E = N
N ln Z1
M = (6.6)
B
For large x (corresponding to large B and/or small T ), the magnetization saturates at the
value sN g, while for small x (corresponding to small B and/or large T ), the magnetization
follows the Curie law,
1 B
M = s(s + 1)N g 2 2 (6.8)
3 kT
As expected, it behaves linearly in B for small B, and tends to zero for large T .
p2
H1 = Htransl + Hint Htransl = (6.9)
2M
Here, Htransl corresponds to the center of mass motion for total momentum p and total mass
M , while Hint corresponds to the internal degrees of freedom, in which we include rotational
and vibrational modes. These two Hamiltonians commute with one another, and may be
simultaneously diagonalized. Putting the system in a large cubic box of linear size L, and
volume V = L3 with Dirichlet boundary conditions, the energy levels are then given by,
2
h
Etransl = (n2 + n2y + n2z ) nx , ny , nz 0 (6.10)
2M L2 x
56
and Eint = n , where n are the 1-particle energy levels of a single particle. The levels n are
clearly independent of V and N . Thus, the partition function decomposes as follows,
Z = (Ztransl )N (Zint )N (6.11)
The translational part was already computed earlier, and we have,
N (T )3 h2
!
2 2
Ftransl = N kT + N kT ln = (6.12)
V M kT
where (T ) is the thermal wavelength introduced in (2.48). In all generality, all that we can
state about the internal part Zint is that it only depends on T , but not on V or N ,
en
X
Fint (T ) = kT ln Zint Zint = (6.13)
n
We conclude that a general non-relativistic Boltzmann ideal gas has the following free energy,
N (T )3
!
F = N Fint (T ) N kT + N kT ln (6.14)
V
Using the standard thermodynamic relations of (2.24), we find,
3
E = N kT N (Fint )
2
N (T )3
!
0 5
S = N Fint + N k N k ln
2 V
P V = N kT
N (T )3
!
= Fint + kT ln (6.15)
V
We observe as general properties of any ideal gas that,
the internal energy density E/N depends only on T , and not on the density N/V ;
57
Generally, U will be a function of all variables at once. In a mono-atomic gas, U will be a
function only of the positions of the atomd. In the approximation of a dilute real gas, how-
ever, one may assume that the interactions occur only between pairs of particles, neglecting
interactions between triples, quadruples etc. In the partition function, the momentum and
coordinate integrals factorize. The momentum integrals give the contribution for the ideal
gas, and may be factored out. Thus, we find,
Z = Zideal ZU (6.17)
58
This approximation is not very realistic for real gasses though. Instead, at room tempera-
tures, there is a core overlap of radius r0 where the interaction is very strong (compared to
T ), but negligible outside of this core. Thus, we may then use the following approximate
formula for B,
Z r0
2 2 Z 2
B(T ) = 2 r dr + r drU (6.24)
0 kT r0
The first term is independent of T , while the second is inversely proportional to it. For an
attractive potential U < 0, we have,
a
B(T ) = b a, b > 0 (6.25)
T
The corresponding free energy is found to be,
N (T )3 N2
!
F = N kT + N kT ln + (bkT a) (6.26)
V V
Computing the pressure, we find,
kN T N2
P = + 2 (bkT a) (6.27)
V V
or
N2 N2
!
kN T N kT
P +a 2 = + 2 bkT (6.28)
V V V V Nb
which gives the Van der Waals equation for a real gas,
N2
!
P + a 2 (V N b) = N kT (6.29)
V
under the assumption that V N b.
59
factors as follows,
N Z
1 Y P
Z = Zideal ZU ZU = N d3 ri e j<k U (rjk ) (6.31)
V i=1
where the ideal gas partition function was given in (??). Defining the following function,
The Mayer expansion is obtained by expansing the product in powers of f , and we have,
N Z
1 Y
d3 r i 1 +
X X
ZU = N fjk + fjk fmn + (6.34)
V i=1 j<k j<k, m<n
N2 Z 3
d rf (r) (6.35)
2V
and so on. The higher order terms admit a diagrammatic expansion.
60
7 Systems of indistinguishable quantum particles
Identical quantum particles (or systems of particles such as nuclei, atoms or molecules) are
not only indistinguishable; they obey specific quantum permutation symmetries. In 3 space-
dimensions (and higher), the only two types of permutation symmetries allowed by local
Poincare invariant quantum field theory correspond to either Bose-Einstein or Fermi-Dirac,
respectively for particles (or systems) either of integer spin or of half odd-integer spin. This
correspondence will be heavily used, but not be proven here, as it is properly the subject of
quantum field theory.
O | 1 , , A i = | 1 , , A i (7.1)
The Hilbert space HN of the N indistinguishable particles is then given by the tensor product
of N copies of H,
HN = H
|
H
{z
H} (7.2)
N
I} O I| {z
On = I| {z I} (7.3)
n1 N n1
The quantum numbers labeling the states of HN may be taken to be those of the maximal
set of commuting observables {On } of HN , with n = 1, , N and = 1, , A. Thus, the
basis states of HN may be labeled by,
|1 i |2 i |N i (7.4)
We stress that n here stands for the full array n = n1 , n2 , , nA . The action of the
operators Oi in this basis may be read off from their definition,
On |1 i |2 i |N i = n |1 i |2 i |N i (7.5)
61
The action of a permutation amongst these N particles is defined by its action on each
particle, and may be expressed in terms of the quantum numbers n . The quantum permu-
tation symmetry allowed for a physical state |1 , 2 , , N i HN of N identical particles
with quantum numbers n for n = 1, , N , is either one of the following,
+|1 , 2 , , N i Bose-Einstein
|(1) , (2) , , (N ) i = (7.6)
() |1 , 2 , , N i Fermi-Dirac
where
1 Bose-Einstein
= (7.8)
() Fermi-Dirac
We denote the eigenstates of H (1) by |i, and the corresponding eigenvalue by and use
as a collective label for all quantum numbers of the one-particle states ( was denoted by
62
in the preceding section). A macro-state with total number of particles N , total energy E,
and N particles in micro-state |i then satisfies,
X X
N = N N = E (7.10)
In the thermodynamic limit, the energy levels of H (1) will become closely spaced, and ap-
proach a continuum distribution.
To impose the quantum permutation symmetry on the micro-states, we appeal to the
coarse-graining procedure, and divide the (basis of states in the) Hilbert space H of all
states into discrete cells, labelled by an integer i = 1, 2, 3, . Physically, this may be done
by ordering the states according to the energy levels of a single particle H (1) and/or any
other conserved quantum number. The number of (distinct) states in cell i is denoted by
Gi . In the thermodynamic limit, the spectrum of H (1) becomes closely spaced, so that we
may assume Gi 1 for all i. We denote the number of particles in cell i by Ni , and assume
that Ni 1. Thus, each cell by itself may be viewed as a macroscopic subsystem of the
whole macro-state. The set-up is schematically represented in Figure 6, where the set of
basis states of H is divided into cells i (two cells are separated by a long vertical dash), each
cell having Gi quantum states (each state is indicated with a short vertical dash), and Ni
particles (each particle is indicated with a dot above an available micro-state).
N1 N2 N3
particles
states
G1 G2 G3
In practice, it will be assumed that the cells are small enough so that the energy and/or
other conserved quantum numbers remain constant throughout a given cell. Finally, it must
be stressed that any physically observable quantity should be independent of the precise
coarse graining mesh that has been applied to the system. One such physical quantity is the
mean occupation number ni for each cell, defined by,
Ni
ni = (7.11)
Gi
63
The number i of micro-states in each cell, and the associated contribution to the entropy
Si , are related to the total number of micro-states and the total entropy by,
Y X
= i S=k ln i (7.12)
i i
The object is to compute the numbers i of micro-states in each cell i, and the associated
total energy E and total particle number N , given by,
X X
N = Ni = Gi ni
i i
X X
E = Ni i = Gi ni i (7.13)
i i
Note that the sizes Gi of the cells should be viewed as artifacts of the coarse-graining, and
therefore should not enter into any physical quantities.
64
In the approximation of large Gi , the total entropy is given by,
X
S = k Gi ni ln ni + (1 ni ) ln(1 ni ) (7.17)
i
Note that this derivation of the entropy does not appeal to any equilibrium arguments, it is
simply based on coarse graining and quantum statistics counting.
To obtain the equilibrium distribution for the occupation numbers ni , we use again (7.14),
for fixed Gi , E and, where appropriate, fixed N ,
X
Gi ni ln ni + ln(1 ni ) i + = 0 (7.18)
i
Setting this variation to zero for each ni gives an equation for each ni ,
1
ni = (7.19)
e(i ) + 1
We see that Gi naturally has the interpretation as the degeneracy of the energy level i. The
partition function is readily deduced, and we find,
Y Gi
Z= 1 + e(i ) (7.21)
i
In the grand-canonical ensemble, the partition function is directly related to the Gibbs free
energy G by the relation G = kT ln Z of (3.42). For our case, this quantity is given by,
G = E T S N = P V (7.22)
which precisely corresponds to filling up each micro-state with at most one particle. This is
a very convenient relation, which gives directly the equation of state.
65
[black],thickness=2):
p2:=plot(1/(exp(x)+1), x=-1.5..3, color=[blue],thickness=2):
p3:=plot(1/(exp(x)-1), x=0.2..3, color=[red],thickness=2):
O
p4:=textplot([1.2, 4, "Bose-Einstein"], align = {above, right},
color=[red], font=[TIMES,BOLD,13]):
p5:=textplot([1.2, 3.5, "Boltzmann"], align = {above, right},
color=[black], font=[TIMES,BOLD,13]):
p6:=textplot([1.2, 3, "Fermi-Dirac"], align = {above, right},
color=[blue], font=[TIMES,BOLD,13]):
display(p1,p2,p3,p4,p5,p6);
Bose-Einstein
4
Boltzmann
Fermi-Dirac
3
n
2
K1 0 1 2 3
EK
kT
66
The free energy and partition function are readily deduced, and we find,
Gi ln 1 e(i )
X
F = kT
i
!Gi
n(i )
Y X
Z = e (7.28)
i n=0
n1 (7.29)
This was the limit in which the Sackur-Tetrode formula of the simplest ideal gas was valid.
Thus, we see that FD and BE statistics will produce modifications thereof when the occu-
pation numbers are not small.
67
8 Ideal Fermi-Dirac Gases
In this section, we shall discuss ideal gasses of particles obeying Fermi-Dirac statistics. Appli-
cations to the physics of electrons in metals, their magnetic properties, low energy behavior,
white dwarfs and neutron stars, will be treated in relative detail.
Recall the basic equations for an ideal FD gas from equations (7.20) and (7.23),
X 1
N =
i e(i ) +1
X i
E =
i e(i ) + 1
ln 1 + e(i )
X
G = P V = kT ln Z = kT (8.1)
i
where we take the sum over all states labelled by i, including their degeneracies, and corre-
sponding energy i . Here, Z stands for the grand canonical partition function. The Gibbs
free energy G, which is the natural thermodynamic potential in the grand canonical ensem-
ble, is related to these quantities by G = P V . As we are assuming that the gas is ideal,
no mutual interactions take place; the Hamiltonian is just that of the free relativistic or
non-relativistic particle, and i are the free particle energies.
where a degeneracy factor g has been included to account for the number of internal degrees
of freedom of each particle. For example, when this degree of freedom is spin s, we have
g = 2s + 1. The standard thermodynamic functions are then given as follows,
gV Z 3 1
N = 3
d p (p2 /2m)
(2h) e +1
2
gV Z
p
E = 3
d3 p (p2 /2m)
2m(2h) e +1
gV Z
3
(p2 /2m)
P V = kT d p ln 1 + e (8.3)
(2h)3
It is a property of the simplest ideal gas thermodynamics that temperature may be scaled
out of all these integrals. To do so, we express the quantities in terms of the composite
68
variable z, referred to as the fugacity, and we will continue to use the notation = (T ) for
the thermal wavelength,
!1
h2
2 2
z = e = (8.4)
mkT
Expressing in terms of z, and p in terms of x defined by p2 = 2mkT x, we have the following
alternative expressions,
N g 3 P g
= 3 f3/2 (z) E = PV = 3 f5/2 (z) (8.5)
V 2 kT
The Fermi-Dirac functions f (z) depend only upon the fugacity z, and are defined as follows,
1 Z x1 dx
f (z) (8.6)
() 0 z 1 ex + 1
From (8.4), we have 0 < z < . More generally, in the regime 0 < Re(z) < , the integral
is absolutely convergent for Re() > 0, and may be analytically continued to all C. The
functions have the following expansion in powers of z,
zn
()n+1
X
f (z) = (8.7)
n=1 n
The series is absolutely convergent only for |z| 1 and 1 < , and gives the small z behavior
of the functions, or equivalently for /kT 1. From the expansion in (8.7), it is immediate
that we have the following recursion relation,
z f (z) = f1 (z) (8.8)
z
which may also be derived directly from the integral representation, and holds for all z C.
The function f (1) is related to the Riemann -function (),
1 1
X
f (1) = 1 () () = (8.9)
21 n=1 n
69
> restart:
> with(plots):
f:=(nu,z)->evalf((1/GAMMA(nu))*int(x^(nu-1)/(exp(x)/z+1), x=0..
infinity)):
p1:=plot(f(1/2,z), z=0..30, labels=[z, f_nu(z) ], color=[blue],
thickness=2):
p2:=plot(f(3/2,z), z=0..30, labels=[z, f_nu(z) ], color=[black],
thickness=2):
p3:=plot(f(5/2,z), z=0..30, labels=[z, f_nu(z) ], color=[red],
thickness=2):
p4:=textplot([23, 1, "f_{1/2}(z)"], align = {above, right}, color=
[blue], font=[TIMES,BOLD,13]):
p5:=textplot([23, 3.5, "f_{3/2}(z)"], align = {above, right},
color=[black], font=[TIMES,BOLD,13]):
p6:=textplot([23, 7, "f_{5/2}(z)"], align = {above, right}, color=
[red], font=[TIMES,BOLD,13]):
display(p1,p2,p3,p4,p5,p6);
8
1.4
f_{5/2}(z)
7
6 1.3
f_nu z 5 PV
NkT
4 1.2
f_{3/2}(z)
3
2 1.1
f_{1/2}(z)
1
0 10 20 30 0 1 2 3 4 5 6 7 8
z z
> plot(f(5/2,z)/f(3/2,z), z=0..8, labels=[z, PV/NkT ], color=[blue],
thickness=2);
Figure 8: Left panel: numerical plot of the Fermi-Dirac functions f1/2 (z), f3/2 (z), f5/2 (z).
Right panel: the equation of state relation in terms of z.
70
we have z 1 (see Figure 8 for the numerical behavior of f3/2 (z)), and we recover the
Boltzmann gas with,
N 3
z f (z) z (8.16)
gV
In this limit the equation of state reduces to P V = N kT , and the internal energy, entropy
and specific heat are given by their Boltzmann expressions, with E = 3P V /2 as well as,
N 3
!
3 5 3
E = kT S = N k N k ln CV = N k (8.17)
2 2 gV 2
Using the expansion for small z of (8.7), we easily obtain the leading corrections to the
Boltzmann expressions,
N 3
!
P V = N kT 1 +
4 2gV
N 3 N 3
! !
5
S = Nk + N k ln
2 8 2gV gV
3
!
3 N
CV = Nk 1 (8.18)
2 8 2gV
The increase in the pressure, at fixed temperature, is the result of the exclusion principle.
1 1 for <
lim+ = (8.19)
T 0 e() +1 0 for >
p2
F = pF = |p| (8.20)
2m
71
is referred to as the Fermi surface. The magnitude pF of the momenta lying on the Fermi
surface is constant, and is referred to as the Fermi momentum. Strictly at T = 0, the various
thermodynamic functions may be evaluated in terms of the parameters m, V, F , and we find,
gV Z pF 2 4gV p3F
N = 4p dp =
h)3 0
(2 h)3
3(2
gV Z pF 2 p2 4gV p5F
E = 4p dp = (8.21)
h)3 0
(2 2m 10m(2 h)3
E 3 2N
= F P = F (8.22)
N 5 5V
Using the expression for the free energy F = E T S = P V + N , the entropy is found
to vanish. This is in accord with the fact that the system at T = 0 is expected to be in its
(unique) microscopic ground state.
Immediate applications of the low temperature thermodynamics of a gas of fermions is
to massive gravitational bodies in which the pressure generated by the exclusion principle is
compensated by the force of gravity at equilibrium. This is the case of neutron stars, and of
the hypothetical quark stars and strange stars. Later on, we shall discuss in some detail the
case of white dwarfs where electrons are the key players.
72
1
0.8
0.6
0.4
0.2
0
0 2 4 6 8 10
E
Figure 9: Low temperature behavior of the Fermi-Dirac occupation number for = 6, and
kT = 0.125, 0.25, and 1 respectively in blue, black, and red.
for any function (x) with polynomial growth in |x|. By separating the integration region
at the point x = , we readily obtain (after suitable shifts of variables in each region),
Z Z
( + x) ( x) Z (x)
f [] = dx (x) + dx x
+ dx x (8.24)
0 0 e +1 e +1
The first two terms produce power law dependences in large , while the last term is sup-
pressed by exponentials O(e ), and may be neglected compared to the first two. Expanding
( x) in a power series in x around x = 0, and using the formula, for the value = 2n + 2,
we have the following final formula,
Z
1
(2n + 2)(2n+1) () + O(e )
X
F [] = dx (x) + 2 1 (8.25)
0 n=0 22n+1
where (2n+1) () denotes the derivative of order 2n + 1 evaluated at the point . Using this
formula with (x) = x1 /(), we find,
2 7 4
!
f (z) = 1 + ( 1) 2 + ( 1)( 2)( 3) + + O(e ) (8.27)
( + 1) 6 360 4
1/2 2 7 4
!
f1/2 (z) = 1 + + O(e )
(3/2) 24 2 384 4
73
3/2 2 7 4
!
f3/2 (z) = 1+ 2 + + + O(e )
(5/2) 8 640 4
5/2 5 2 7 4
!
f5/2 (z) = 1+ 2 4
+ + O(e ) (8.28)
(7/2) 8 384
Recall that this expansion holds in = = /kT for large , i.e. for small T at fixed .
Corrections to various thermodynamic functions may now be evaluated with the help of
the general formulas (8.5), (8.11), and (8.13). We begin by expressing the chemical potential
at finite temperature in terms of the Fermi energy F , by comparing the formulas,
N 4g(2mF )3/2
=
V 3(2 h)3
2
!
N g g 3/2
= f3/2 (z) = 3 1+ 2 (8.29)
V 3 (5/2) kT 8
In terms of T and F , the energy per particle, entropy, and specific heat are given by,
!2
E 3 5 2 kT
= F 1 +
N 5 12 F
S CV 2 kT
= = (8.31)
Nk Nk 2F
p2
= B (8.32)
2m
where = eh/2mc is the Bohr magneton. Since the shifts B enter in combination with
the chemical potential for the spin 1/2 particles, the total Gibbs free energy is given by,,
74
where G0 is the Gibbs free energy of the simplest spinless ideal gas, and is given by,
kT V Z 3
(p2 /2m)
kT V
G0 (T, V, ) = 3
d p ln 1 + e = 3 f5/2 (e ) (8.34)
(2
h)
The paramagnetic susceptibility, per unit volume and at zero external magnetic field, is then
obtained as follows,
1 2 G 2 2 G0 2 N
para = = 2 = 2 (8.35)
V B 2 B=0 V 2 V T,V
At zero temperature, we may use the first relation in (8.21) for g = 1, and expressed in terms
of the chemical potential,
(2m)3/2
N =V (8.36)
3
3 2 h
and we find,
2 mpF 3 2 N
para = = (8.37)
2h3 2 F V
Finite temperature corrections may be deduced using the full expression for G0 . At high
temperatures, we recover, of course, the Curie-Weiss law already derived using Boltzmann
statistics in section 6.
p2 eB
h 1
(pz , n) = z + n+ (8.38)
2m mc 2
where pz is the momentum along the direction of the magnetic field. Particles with a magnetic
moment require an extra term B term, but at small magnetic fields, the case we are most
interested in here, this effect is paramagnetic and may be treated independently along the
lines of the preceding section.
Some care is needed in correctly normalizing the summation over Landau levels. For weak
magnetic fields, the Landau levels are closely spaced and degenerate into the continuous px , py
spectrum at B = 0. We may use this correspondence to normalize the sum, as follows. The
75
number of quantum states between two successive Landau energy levels (pz , n + 1) and
(pz , n) is eBh/mc; computing the same from the classical measure, we find,
Z (pz ,n+1)
dx dy dpx dpy Lx Ly eB
2
= (8.39)
(pz ,n) (2h) 2hc
where Lx and Ly are the linear dimensions of a square box in the plane perpendicular to
B. This factor provides the correct normalization for the calculation of the grand canonical
partition function, actually for all values of the field B,
Lx Ly eB Z dz dpz
ln 1 + e((pz ,n))
X
G = kT ln Z = kT (8.40)
n=0 2hc 2h
h/mc kT , we may use an approximate
Since we are interested in weak magnetic fields, eB
evaluation of the sum over n, by using the Euler-McLaurin formula to first order,
Z
1 1
dxf (x) + f 0 (0)
X
f n+ = (8.41)
n=0 2 0 24
up to higher derivative corrections (which will be accompanied by higher powers of B). The
contribution of the integral is proportional to
Z
2
eB dx ln 1 + e(pz /2m+eBh/mcx) (8.42)
0
By changing integration variables eBx x, we see that the integral is in fact independent of
B, and for our purposes of studying magnetic properties, will be immaterial. We abbreviate
its contribution to G as GB=0 . The remaining contribution is readily evaluated, and we find,
1 V e2 B 2 h
Z 1
G = GB=0 + 2 2
dpz (p2 /2m) (8.43)
24 (2
h) mc e z +1
Changing variables pz = 2mkT x and z = e , we may express the result in terms of a
Fermi-Dirac function, and the Bohr magneton = eh/2mc,
mV 1
2 2
G = GB=0 + 3 B (2mkT ) 2 f1/2 (z) (8.44)
24 2 h
The diamagnetic susceptibility per unit volume is defined by,
1 2 G
diam = (8.45)
V B 2 T,
and is found to be given by
m 1
2
diam = (2mkT ) 2 f1/2 (z) (8.46)
3
12 2 h
76
For the Boltzmann regime, with z 1, and z N 3 /V , we get,
N 2
diam (8.47)
V 3kT
In the low temperature regime where z 1, we have f1/2 (z) 2(/kT )1/2 , and we find,
N 2
diam = (8.48)
4V F
Assembling the contributions from the paramagnetic and diamagnetic parts, we find for large
temperature,
N 1 2N 2
= para + diam = 2 =
2
(8.49)
V kT 3 3V kT
Thus, at high temperatures, gasses tend to be paramagnetic. The same is generally true at
low temperatures.
M 1033 g
107 g/cm3
T 107 K 103 eV (8.50)
77
where M is the total mass of star, its mass density, and T its temperature. At this
temperature, all 4 He is completely ionized, so we have a plasma of N electrons and N/2
Helium nuclei. The relations between the total mass M and volume V are as follows,
N
M = N (m + 2mp ) 2mp N = (8.51)
V 2mp
The electron density N/V allows us to compute the Fermi momentum by,
N 2p3
= 2F 3 = (8.52)
V 3 h
2mp
Substituting the above data for the characteristics of a white dwarf, we get pF /mc 100,
so that the electrons are in fact relativistic. The Fermi energy and temperature TF = F /k
are respectively found to be F 106 eV, and TF 1010 K. Thus the actual temperature
of the white dwarfs is small compared to the Fermi temperature, and the electron gas is
fully degenerate. The electrons produce the largest contribution to the pressure against
gravitation, and so we shall neglect the contribution of the nuclei.
To study the equilibrium conditions quantitatively, we compute the pressure of this rel-
ativistic gas of electrons. Using the Gibbs ensemble, we have,
gkT Z 3
((p))
P = d p ln 1 + e (8.54)
h)3
(2
In the degenerate limit, is large and positive, and the logarithmic term vanishes when
(p) > . The contribution for (p) < is dominated by the exponential part of the
argument, so that we obtain,
g Z 3
P = d p ( (p)) ( (p)) (8.56)
h)3
(2
4
Equivalently, one could use for (p) only the relativistic kinetic energy; the difference amounts to a shift
of by the rest energy mc2 .
78
Carrying out the angular integration, and changing variables to
we have,
4gm4 c5 Z F
P = d cosh sinh2 (cosh F cosh ) (8.58)
h)3 0
(2
g m4 c5 pF
P = A (8.59)
h)3
6(2 mc
0.9
0.8
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0 2 4 6 8 10
x
79
Next, we need to balance the pressure of the electron gas against the gravitational pull
on them. In terms of the work done by each, we have,
EP = P 4R2 dR
GM 2
Eg = dR (8.61)
4 R2
Here, is a fudge factor of order unity introduced in order to account for the fact that the
mass distribution in a real star will not be uniform. At equilibrium, these two must balance
one another, so we must have,
GM 2
P = (8.62)
4 R4
We now relate the Fermi momentum to the radius, by using the formula for the mass density
= 3M/4R3 , and we find,
!1/3
pF R0 2
h 9M
x= = R0 = (8.63)
mc R mc 64 2 mp
Now, it is easy to see, for example from Figure 10, that the function A(x) satisfies,
for all 0 < x < . As a result, there is a maximum mass that a white dwarf can have, whose
value is given by the Chandrasekar limit, which to our approximation is given by,
1/2 !3/2
9 3 h
c
Mc = (8.67)
64m2p 3 G
In practice, the limiting mass is not much larger than the mass of the sun: Mc 1.44M .
We conclude that our sun will end its life as a white dwarf.
80
9 Bose-Einstein statistics
In this section, we shall discuss ideal gasses of particles obeying Bose-Einstein statistics.
Applications to Plancks black body radiation, microwave background radiation, an Bose-
Einstein condensation will be given.
Recall the basic equations for an ideal BR gas from equations (7.26), (7.27), and (7.28)
are given by,
X 1
N =
i e(i ) 1
X i
E = ( )
i e i 1
ln 1 e(i )
X
G = P V = kT ln Z = kT (9.1)
i
where we take the sum over all states labelled by i, including their degeneracies, and cor-
responding energy i . We shall normalize the ground state energy to zero, so that the
Bose-Einstein distribution requires,
0 (9.2)
Here, Z stands for the grand canonical partition function. The Gibbs free energy G is related
to these quantities by G = P V . As we are assuming that the gas is ideal, no mutual
interactions take place; the Hamiltonian is just that of the free relativistic or non-relativistic
particle, and i are the free particle energies.
81
where = |k|c. Since a photon has two degrees of polarization, we have g = 2, and obtain
the following formulas for energy and pressure,
2V Z 3 h
E = 3
d p h
(2
h) e 1
2V Z
3
h
PV = kT d p ln 1 e (9.5)
(2h)3
2V d3 p V 2 d
(9.6)
h)3
(2 2 c3
1.4
1.2
0.8
0.6
0.4
0.2
0
0 5 10 15
x
82
The temperature dependence may be easily extracted by changing integration variables,
h = x. The integrals may be performed by expanding the integrands in powers of the
exponentials ex , and we find,
E 2 (kT )4 Z
dx x3
= = 6 (4)
V 15 c3 h3 0 ex 1
(kT )4
2 Z
P = dx x2 ln 1 ex = 2 (4) (9.8)
45 c3 h3 0
where (z) is the Riemann zeta function, and we have (4) = 4 /90. It readily follows that
internal energy and pressure are relatedby,
E = 3P V (9.9)
83
Tcomb , nucleons and electrons form a plasma which interacts with the photons, and below
which ionized nuclei and electrons combine. The typical ionization energy of atoms and
molecules is on the order of Tcomb 1eV 104 K. At temperatures below Tcomb , photons
do not directly interact with charged particles any more, but rather with electrically neutral
atoms and molecules. As a result, below Tcomb , photons are not in equilibrium any more
with atoms and molecules, and decouple. So, below Tcomb , photons remain as an isolated
free gas.
The fact that we observe today CMB radiation at a much lower temperature is explained
by the expansion of the universe. Assuming that this expansion is adiabatic, namely entropy
is conserved, then we can read off the relation between temperature and volume of the
universe from formula (9.10). Thus, we know how the volume V of the present Universe is
related to the volume Vcomb of the Universe at the time of combination:
3
V Tcomb
= 3
(9.12)
Vcomb TCMB
Given the temperatures, this number evaluates to 5 1010 . The corresponding red-shift
factor is 3000, as may be deduced by comparing the average photon energies.
1 Z dx x1
g (z) (9.13)
() 0 z 1 ex 1
84
energy. This can happen when z is very close to 1, since then the occupation number of the
= 0 state can become large,
1 z
= =0 (9.15)
z 1 e 1 1z
The quantum states at finite energy are not affected by this phenomenon, and the customary
translation into continuum states may be applied there. For the internal energy, the zero
energy states do not contribute, and we have,
E 4 Z 3 p2 /2m
= d p (9.16)
V (2h)3 z 1 ep2 /2m 1
Isolating the = 0 contribution to N and P gives the following formulas,
N 1 z 4 Z 3 1
= + 3
d p 1 p2 /2m
V V 1 z (2 h) z e 1
kT 4kT Z
3
p2 /2m
P = ln(1 z) d p ln 1 ze (9.17)
V (2h)3
Next, we change variables, to exhibit the number N0 of particles in the ground state,
z N0 1
N0 = z= 1 (9.18)
1z N0 + 1 N0
Assuming that N0 /V tends to a finite value in the thermodynamic limit, we see that the
effects of N0 survives only in the above formulas for the particle number, but drops out of
the formula for the pressure, since we have, V 1 ln(1 z) V 1 ln N0 0 as V with
N0 /V fixed. The final formulas for the total number of particles and pressure are thus,
4V Z 3 1
N N0 = 3
d p 1 p2 /2m
(2h) z e 1
4kT Z 3
p2 /2m
P = d p ln 1 ze (9.19)
(2h)3
In terms of the BE functions, we obtain,
V 3 kT
N N0 = g3/2 (z) E = PV P = g5/2 (z) (9.20)
3 2 3
The equation of state is obtained by eliminating z between the relations
PV g5/2 (z)
= (9.21)
(N N0 )kT g3/2 (z)
and the first relation in (9.20).
85
9.4 Bose-Einstein condensation
The number Ne = N N0 of excited states, i.e. states with non-zero energy, is bounded
from above by the fact that,
g3/2 (z) g3/2 (1) = (3/2) 2.612 (9.22)
for all values of z in the physical range 0 < z 1, and we have,
!3/2
mkT
Ne = N N0 (3/2)V (9.23)
2h2
As T is lowered, at fixed volume and fixed total number of particles N , the number of excited
states Ne must decrease. At some critical temperature Tc , one is forced to populate the states
at zero entropy in a macroscopic way, namely with N0 /V finite in the large V limit. This
critical temperature is obtained by setting N0 = 0, and we find,
!2/3
2h2 N
Tc = (9.24)
mk (3/2)V
To leading approximation, namely by setting z = 1, the number of excited and zero energy
states are related as follows, for T < Tc ,
3/2
N0 Ne T
=1 =1 (9.25)
N N Tc
3 N 3
!
ln z
g1/2 (z) = (9.30)
T V
2 VT
Putting all together, the specific heat in the gaseous phase is found to be,
As z 1, we have g1/2 (z) , and so near the transition point, the second term cancels,
and we find that the specific heat in the gas phase and in the condensed phase match at the
transition point, since we have N 3 = V g3/2 (z). The slope at T = Tc is discontinuous across
the transition. This is a sign of a phase transition, as we shall study in detail in subsequent
sections.
C/Nk
3/2
T
0
T
Figure 12: Molecular specific heat curve across the Bose-Einstein condensation transition.
87
10 Phase coexistence: thermodynamics
A given system or material can exhibit different physically distinct phases, as a function
of the ranges of its macroscopic parameters. A common example is that of water, which
may exist in its solid phase of ice, its liquid phase, or its vapor phase. Different phases may
co-exist in equilibrium. The passage from one phase to another, as thermodynamic variables
are changed, is referred to as a phase transition. Given the equations of state in both phases,
thermodynamics gives an effective formulation of the conditions of equilibrium between dif-
ferent phases, and transitions between them. The underlying microscopic dynamics which
is responsible for the system existing in one phase or another, or transitioning between one
phase to another is, however, captured only by the application of statistical mechanics.
T2 = T1
P 2 = P1
2 = 1 (10.1)
Given these conditions, it will be most convenient to use not the variables E, V, N but
rather with the intensive parameters T, P, in terms of which the equilibrium conditions are
written. In either phase, these three variables are not independent, so we shall choose T, P
and express the chemical potential for each phase as a function of T, P ,
1 = 1 (T, P )
2 = 2 (T, P ) (10.2)
The chemical potentials 1 and 2 have different functional forms for the two phases. For
given T, P it may, or may not, be possible to set the chemical potentials equal. Thus we
conclude that
Generally, the relation 1 (T, P ) = 2 (T, P ) will hold on a line segment in the T, P plane, as
shown in figure 13.
88
P P
critical point critical point
P P
T V
Figure 13: Phase coexistence curve in (T, P ) and (V, P ) variables ending at a critical point.
solid liquid
P end point
gas
T
Figure 14: A gas-liquid-solid phase diagram exhibiting a triple point and an end point.
More than two phases can coexist pairwise at various temperatures and pressures. It is
also possible for three phases to coexist at a point (generically). In fact this is the case of the
standard gas-liquid-solid phase diagram, as show schematically in figure 14. At the triple
89
point, we have the relations,
gas (T, P ) = liquid (T, P ) = solid (T, P ) (10.4)
This phase diagram applies, for example, to water.
= E T S + P V = N (10.11)
d = SdT + V dP + dN (10.12)
but the last term not be important here as dN = 0. The relation may also be expressed in
terms of using (10.5) or (10.11), and for constant N we find,
where again s = S/N and v = V /N . The desired ingredients for (10.10) now immediately
follows, and we have,
= s =v (10.14)
T P P T
so that,
dP s2 s1
= (10.15)
dT v2 v1
This is the Clausius-Clapeyron equation.
P V 3 (bP + N kT )V 2 + aV ab = 0 (10.17)
The equation always has at least one real root, the additional two roots being either both real,
or complex conjugates of one another. For sufficiently large P and T , we may approximate
91
the equation by P V 3 N kT V 2 = 0 which reduces to the ideal gas equation of state, and the
solution is unique. Thus, for T, P sufficiently large, we have two complex conjugate solutions
in addition to the single real root. As T, P are lowered, all three real solutions will appear.
This will happen when the two roots of the derivative equation coincide, a point we denote
by Vc . But Vc must also satisfy the Van der Waals equation, so that the equation must be
equivalent to (V Vc )3 = 0. Multiplying by Pc and expanding gives,
Pc V 3 3Pc Vc V 2 + 3Pc Vc2 V Pc Vc3 = 0 (10.18)
Identifying coefficients with the equation of (10.17) gives,
bPc + N kTc = 3Pc Vc
a = 3Pc Vc2
ab = Pc Vc3 (10.19)
which is solved uniquely as follows,
a 8a
Vc = 3b Pc = N kTc = (10.20)
27b2 27b
Eliminating a, b from the Van der Waals equation in favor of Tc , Pc , Vc gives,
Vc2
!
P V T
+3 2 3 1 =8 (10.21)
Pc V Vc Tc
Note that this curve is universal, and has no extra free parameters in terms of the normalized
thermodynamic variables T /Tc , P/Pc , and V /Vc . A schematic plot in V, P variables of its
isothermal curves (lines of constant T ) are plotted in figure 15.
For P > Pc , the pressure is a monotonously decreasing function as V increases at constant
T . This is in accord with the physical dynamics of a gas. For P < Pc , however, the situation
is more complex. As one decreases the volume away from very large volume, the pressure
at first increases monotonically, as was the case for P > Pc . But as the volume is further
reduced, it would appear that there comes a point where the pressure now starts to decrease
(on the lowest curve in the left panel of figure 15). This behavior is not consistent with
the behavior of a gas. Rather, the tendency to decrease the pressure should be viewed as a
symptom of condensation into the liquid phase.
92
P P
1 2
V V V
Figure 15: The (V, P ) diagram of isotherms for the Van der Waals equation of state on the
left panel, and the Maxwell construction for phase equilibrium on the right panel.
Along an isotherm, we have dT = 0, so the first term drops out. The remaining terms, with
constant N , gives us a differential expression N d = V dP . Integrating this relation between
the two phases must reflect the equality of the chemical potential in the two phases, and
thus we must have,
Z 2
0 = N (2 1 ) = V dP (10.23)
1
This relation determines a horizontal line in the V, P diagram of the Van der Waals equation
of state which is such that the area in V, P space above equals the area below. This is
precisely the Maxwell construction for equilibrium between two phases.
93
11 Phase transitions: statistical mechanics
A statistical mechanics understanding of the existence and coexistence of various phases of
matter and the phase transitions between them is what we will pursue in this section. After
a brief classification of the different kinds of phase transitions in Nature, we will proceed to
studying the Ising model, which captures so many of the key features of phase transitions.
ei
X
Z() = (11.1)
i
94
If this sum were finite, then Z() is an analytic function of , and in particular all its
derivatives will be continuous, to all orders. So, in any system with a finite number of states,
no phase transitions can ever occur. Even if there are an infinite number of states available,
Z() will still often be analytic, as is the case for the harmonic oscillator for example. So,
one really needs an infinite number of degrees of freedom, usually corresponding to the
thermodynamic limit
N V (11.2)
In finite volume V , and with finite N , no phase transitions will occur. But in Nature, N
is very large, but not infinite. While this is certainly true, it becomes mathematically more
convenient to view a function as discontinuous rather than to consider a function whose slope
is N 1023 . So, to some extent, phase transitions provide a mathematically simplified
picture of a very very large system.
Here J is the coupling strength, and b is an exterior parameter such as a magnetic field; both
of these parameters are real, and could be positive or negative. The symbol hi, ji indicates
that the sum over the sites i, j is restricted in some way: for example to be nearest neighbor.
95
The Ising model captures binary degrees of freedom, which may capture the interactions
between magnetic moments of electrons at sites in the approximation where only the con-
tribution of one of their spatial components is retained (when all components are retained,
we have the Heisenberg model instead). It can also capture the degrees of freedom of a
binary mixture of substances A and B, where for si = +1 corresponds to an atom of A while
si = 1 corresponds to an atom of B.
To begin the study of the dynamics of the system, we set the external field b = 0. The
nature of the ground state of the system then depends on the sign of J.
The coupling J > 0 is referred to as ferro-magnetic because the two possible ground
states have all spins equal, either all up or all down. Note that the nature of these
ground states does not depend on the lattice structure. When thought of as magnetic
spins, the system would then be magnetized in either one of its ground state. Of course,
thermal fluctuations may wash out the magnetization as temperature is increased. One
can show that the spin waves around this ground state obey a non-relativistic dispersion
relation k2 .
The coupling J < 0 is referred to as anti-ferro-magnetic. The nature of the ground
state now depends to some degree on the structure of the lattice. The energy associated
with the coupling between two spins i, j included in the sum over hi, ji is minimized
when si and sj are opposite. But if the coupling is nearest neighbor on a triangular
lattice, then it is impossible to satisfy minimum energy for all three bonds on the
triangle. The system is said to be frustrated. On a d-dimensional square lattice, whose
sites are labelled by integers (i1 , 12 , , , id ), we do have a two-fold degenerate ground
state with absolutely minimal energy, given by a perfectly alternating spin assignment,
si1 ,12 ,,,id = (1)i1 +12 +,+id s0 s0 = 1 (11.4)
This state is called the Neel ground state of an anti-ferro-magnetic system. One can
show that the spin waves around this ground state obey a linear dispersion relation
|k|, which is akin to the relation for a massless relativistic particle.
96
1
Ei = exp Jsi si+1 + b (si + si+1 ) (11.6)
2
The problem now becomes one of 2 2 matrix multiplication. To see this, we work out the
matrix elements of Ei that enter here,
1
T0 , 0
= exp J + b( + 0 ) (11.7)
2
The partition function is then given by
T1 ,2 T2 ,3 TN ,1 = tr TN
X
Z= (11.8)
{i =1}
which is solved by
q
J
= e ch(b) e2J sh2 (b) + e2J (11.11)
This function is analytic in , so there are no phase transitions for any finite value of . In
other words, the system is in the same thermodynamic phase for all temperatures.
si = +1 for all i
si = 1 for all i (11.14)
These ground states are mapped into one another by the spin reversal symmetry R of the
Hamiltonian for b = 0. If both ground states contribute to the partition function, then the
total magnetization will get wiped out, and the system will remain in a disordered phase.
When N is finite this will always be the case. But when N , it is possible for the system
to get stuck in one ground state or the other. The reason this only happens for infinite N
is that it would then take an infinite number of spin flips to transition between the si = +1
and si = 1 states, and this may get energetically impossible. When the system gets stuck
in one of its ground states, then m(0) 6= 0 and we have spontaneous magnetization, familiar
from ferromagnetism below the Curie temperature. The operation of spin reversal, which is
a symmetry of the Hamiltonian for b = 0 is then NOT a symmetry of the physical system
any more, as a definite non-zero value of m(0) is not invariant under R. The symmetry R is
said to be spontaneously broken, and the system is then in an ordered phase, close to one
of its ground states. We have already shown that, for the 1-dimensional Ising model, this
phenomenon does not take place.
The 2-dimensional Ising model, however, does exhibit an ordered phase below a critical
temperature Tc . This is known since the model was solved exactly by Lars Onsager in 1944,
and the critical temperature is known analytically,
Note that as T Tc , the expression m(0) (Tc T )1/8 vanishes and joins continuously
with the T > Tc result m(0) = 0. The phase transition at T = Tc is actually second order.
98
The exponent 1/8 is referred to as a critical exponent, in this case of the magnetization
order parameter m. Critical exponents tend to be universal objects, whose dependence on
the detailed short-distance interactions is limited.
Whether the 3-dimensional Ising model allows for an exact solution is one of the great
outstanding problems of statistical mechanics. Proposals have been made that the model
behaves as a theory of free fermionic random surfaces, but the details have never been
conclusive. Numerical studies of the model have shown however that it also admits a phase
transition between ordered (low temperature) and disordered (high temperature) phases.
The key assumption of mean-field theory is that the statistical average of the first term
on the right side may be neglected. The remaining mean-field Hamiltonian is obtained by
summing over all pairs hi, ji. Reformulating this sum in terms of a sum over individual sites
may be done by noticing that on a d-dimensional square lattice, the number of bonds from
a site is 2d. The Hamiltonian is then given by,
Here, beff is the effective magnetic field felt by each spin si due to the mean value of the spins
that surround it. The partition function is now easily computed, leaving m as a parameter
which remains to be determined self-consistently. We find,
2
N
Z = eJN dm ebeff + ebeff (11.19)
Solving for m at b = 0, we see that there is always the solution m = 0. When 2dJ < 1,
namely for higher temperatures 2dJ < kT , then m = 0 is the only solution, and we have
99
> restart:
with(plots):
> n:=1:
> p1:=plot(x, x=0..n, color=[blue]):
p2:=plot(tanh(0.5*x), x=0..n, color=[red]):
p4:=plot(tanh(1*x), x=0..n, color=[red]):
p5:=plot(tanh(2*x), x=0..n, color=[red]):
> display(p1,p2,p4,p5);
1
0.8
0.6
0.4
0.2
0
0 0.2 0.4 0.6 0.8 1
x
Figure 17: Plot of the curves x in blue, and tanh(2x), tanh(x), and tanh(x/2); only the first
produces an intersection with x 6= 0.
no spontaneous magnetization. For 2dJ > kT on the other hand, we have one solution with
non-zero m, as may be seen by inspecting figure 17.
It is straightforward to solve in the approximation where 2dJ 1, so that m will
be small. One finds the following mean-field result for the magnetization near the phase
transition point,
q 1
m 6dJkc2 (Tc T ) 2 2dJc = 1 (11.21)
We see that mean-field theory gives a completely wrong answer for d = 1, where there is no
phase transition at all in view of our exact result. For d = 2, comparison with the Onsager
solution shows that c is off by a factor of 2, but more importantly, that the critical exponent
is off by a factor of 4. So, while mean field theory correctly predicts a phase transition for
d = 2, the details are not impressive. Numerical simulations show that in three dimensions,
we have
so this is closer to the mean-field exponent of 0.5 for d = 3. In fact, one can argue that
mean-field theory becomes exact in the limit of d .
100
12 Functional integral methods
Functional integral methods may be extended from quantum mechanics to quantum sta-
tistical mechanics, where they provide an extremely useful formalism. In this section, we
begin by deriving the functional integral for the partition function of a classical mechanical
system. We use the path integral representation to study the classical limit, low and high
temperature asymptotics, as well as perturbation theory.
101
This summarizes all that will be needed to derive the functional integral representation.
To evaluate the trace over the Boltzmann operator in (12.1), we divide into N intervals
of equal length such that = N , and use the formula,
N
eH = IH eH (12.6)
eH = IH H + O(2 ) (12.9)
The whole purpose of inserting the completeness relations in both |pi and |qi states is to
have a mixed representation in which matrix elements for both kinetic and potential terms
in the Hamiltonian may be evaluated explicitly. Consider, for example, a potential-type
Hamiltonian of the form H = P 2 /2m + U (Q). It is manifest that we have hp|H|qi =
hp|qiH(p, q), where H(p, q) is the classical Hamiltonian. Given a general Hamiltonian, H,
we can always order all P operators to the left of all Q operators, using [Q, P ] = i
h. It is in
this form that we can use it to define the classical Hamiltonian H(q, p) by,
The needed matrix elements are now readily evaluated, and we have,
hpn |eH |qn i = hpn |qn i 1 H(pn , qn ) + O(2 ) = hpn |qn i eH(pn ,qn ) (12.11)
hqn1 |pn ihpn |eH(P,Q) |qn i = ei(qn1 qn )pn /hH(pn ,qn ) (12.13)
102
The final step consists of a change of notation,
n = nh = h
n/N
qn = q(n )
pn = p(n ) (12.14)
and notice that the trace condition produces a periodicity requirement on the position co-
ordinate q(h) = q(0). Putting all together,
( )
1 Z h
Z
Z= Dp Dq exp d iq(
)p( ) + H(p( ), q( )) (12.17)
h
0
The momenta enter without derivatives in this expression. Without loss of generality, we
may impose periodic boundary conditions also on the variable p, so that we have,
p(
h) = p(0)
q(
h) = q(0) (12.18)
Although expression (12.17) involves a factor of i, we are assured from the outset that the
partition function must be real.
Generalization to s degrees of freedom pi , qi and i = 1, , s, with a Hamiltonian
H(p1 , , ps , q1 , , qs ) = H(p, q) is straightforward, and we find,
s Z s
! ( )
1 Z h
Y
X
Z= Dpi Dqi exp d i qi ( )pi ( ) + H(p( ), q( )) (12.19)
i=1 h
0 i=1
103
12.2 The classical = high temperature limit
In view of the periodicity conditions (12.18), we may expand the functions p( ) and q( ) in
an orthogonal basis of periodic functions on the interval [0, h],
an ei n + an ei n
X
q( ) = q +
n=1
bn ei n + bn ei n
X
p( ) = p + (12.20)
n=1
2n
n = (12.21)
h
1 Z h
d H(p( ), q( )) H(p, q) (12.22)
h
0
where only the zero modes p, q are retained on the right side in H(p, q). Separating also the
integration measure into their zero-mode and non-zero-mode contributions,
dp dq 0 0
Dp Dq = D pD q (12.23)
2h
and using the fact that the non-zero-modes precisely integrate to 1, namely
( )
Z
0 0 i Z h
D p D q exp d p( )q(
) =1 (12.24)
h
0
which is the properly normalized classical result we had already encountered earlier.
104
12.3 Integrating out momenta
Returning to the non-classical case for arbitrary h, we may seek to integrate out explicitly
the momentum variables, since they enter purely algebraically. When H is quadratic-linear
in p, this is always possible. Consider for example the case of a Hamiltonian for a charged
particle in a magnetic field and potential U (q),
(p eA(q))2
H= + U (q) (12.26)
2m
The integrand of the exponential may be rearranged as follows,
Z h Z h
1 1
d iq p + H(p, q) = d 2 + mq 2 + iq A + U (q)
(p eA + imq)
0 0 2m 2
The first term produces a Gaussian integral in p, which may be carried out and gives a factor
Z0 in the partition function which depends only on h , m, and , but not on the dynamics of
the system contained in A and U (q). Thus, we have,
( )
1 Z h 1
Z
Z = Z0 Dq exp d mq 2 + iq A + U (q) (12.27)
h
0 2
The object under the -integration is the Euclidean Lagrangian. Note the factor of i in front
of the electro-magnetic interaction !
105