You are on page 1of 840

A Modem Course in

Statistical Physics
2nd Edition

L. E. REICH;-=L=---

----,

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com

A Wiley-Interscience Publication
JOHN WILEY & SONS, INC.
New York . Chichester . Weinheim . Brisbane . Singapore . Toronto

This book is printed on acid-free paper. @)


Copyright 1998 by John Wiley & Sons, Inc. All rights reserved.
Published simultaneously in Canada.
No part of this publication may be reproduced, stored in a retrieval system or transmitted in any
form or by any means, electronic, mechanical, photocopying, recording, scanning or otherwise,
except as permitted under Sections 107 or 108 of the 1976 United States Copyright Act, without
either the prior written permission of the Publisher, or authorization through payment of the
appropriate per-copy fee to the Copyright Clearance Center, 222 Rosewood Drive, Danvers, MA
01923, (508) 750-8400, fax (508) 750-4744. Requests to the Publisher for permission should be
addressed to the Permissions Department, John Wiley & Sons, Inc., 605 Third Avenue, New York,
NY 10158-0012, (212) 850-6011, fax (212) 850-6008, E-Mail: PERMREQ @ WILEYCOM.
Library of Congress Cataloging-in-Publication Data:
Reichl, L. E.
A modem course in statistical physics/by L. E. Reichl. p.
ern.
Includes bibliographical references and index.
ISBN 0-471-59520-9 (cloth: alk. paper)
1. Statistical physics. 1. Title.
QCI74.8.R44 1997
530. 15'95--dc21
Printed in the United States of America
10987654

2nd ed.

97-13550
CIP

This book is dedicated to


Ilya Prigogine
for his encouragement and support
and because
he has changed our view of the world.

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com

CONTENTS

xix

Preface

1.. Introduction

Overview
Plan of Book
Use as a Textbook

l.A.
l.B.
l.C.

PART ONE

THERMODYNAMICS

2. Introduction to Thermodynamics
2.A.
2.B.
2.C.

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com

2.D.

2.E.
2.F.

1
2
5

01 s
2.C.5.
Elastic Wire or Rod
2.C.6.
Surface Tension
2.C.7.
Electric Polarization
2.C.8.
Curie's Law
The Laws of Thermodynamics
2.D.l.
Zeroth Law
2.D.2.
First Law
2.D.3.
Second Law
2.DA.
Third Law
Fundamental Equation of Thermodynamics
Thermodynamic Potentials
2.F.l.
Internal Energy
2.F.2.
Enthalpy
2.F.3.
Helmholz Free Energy
2.F.4.
Gibbs Free Energy
2.F.5.
Grand Potential

9
9
11

16
16
17
18
19
19
20
20
21
21
22
22
23
31
33
36
37
40
42
45
48
vii

viii

CONTENTS

2.0.

2.H.

S2.A.

S2.B.
S2.C.
S2.D.

Response Functions
2.0.1.
Thermal Response Functions (Heat Capacity)
2.0.2.
Mechanical Response Functions
Stability of the Equilibrium State
2.H.1.
Conditions for Local Equilibrium
in a PVT System
2.H.2.
Conditions for Local Stability in a PVT System
2.H.3.
Implications of the Stability Requirements for
the Free Energies
Cooling and Liquefactions of Oases
S2.A.1. The Joule Effect: Free Expansion
S2.A.2. The Joule-Kelvin Effect: Throttling
Entropy of Mixing and the Gibbs Paradox
Osmotic Pressure in Dilute Solutions
The Thermodynamics of Chemical Reactions
S2. "
rr"

Converted with

S2.
S2.E. The
References
Problems

STOI Converter

3. The Thermo

hDP://www.stdutililV.com

3.A.
3.B.
3.C.
3.D.

3.E.
3.F.

3.0.

trial version

Introductory Remarks
Coexistence of Phases: Gibbs Phase Rule
Classification of Phase Transitions
Pure PV T Systems
3.D.1.
Phase Diagrams
3.D.2.
Coexistence Curves: Clausius-Clapyron
Equation
3.D.3.
Liquid-Vapor Coexistence Region
3.D.4.
The van der Waals Equation
Superconductors
The Helium Liquids
3.F.1.
Liquid He4
3.F.2.
Liquid He3
3.F.3.
Liquid He3 _He4 Mixtures
Landau Theory
3.G.l.
Continuous Phase Transitions
3.0.2.
First-Order Transitions

50
50
53
55
55
57
63
66
66
68
72
74
78
78
82
86
89

90

96
96
98

100
103
103
105
110
115
118
123
123
124
126
128
128
134

ix

CONTENTS

3.H.

S3.A.
S3.B.
S3.C.
S3.D.
S3.E.

S3.F.
References
Problems

PART TW(
4.

Elemen
4.A.
4.B.
4.C.
4.D.

Critical Exponents
3.H.l.
Definition of Critical Exponents
3.H.2.
The Critical Exponents for Pure PVT Systems
Surface Tension
Thermomechanical Effect
The Critical Exponents for the Curie Point
Tricritical Points
Binary Mixtures
S3.E.l.
Stability Conditions
S3.E.2. Equilibrium Conditions
S3.E.3. Coexistence Curve
The Ginzburg-Landau Theory of Superconductors

Converted with

pRY

STOI Converter
trial version

hDP:llwww.stdutililV.com
""

135
136
137
142
146
149
151
153
154
155
160
162
166
167

Stochastic Variables and Probability


Distribution Functions
4.D.l.
Moments
4.D.2.
4.D.3.
Characteristic Functions
Jointly Distributed Stochastic Variables
4.D.4.
4.E.
Binomial Distributions
The Binomial Distribution
4.E.l.
The Gaussian (For Normal) Distribution
4.E.2.
4.E.3.
The Poisson Distribution
4.E.4.
Binomial Random Walk
4.F.
A Central Limit Theorem and Law of Large Numbers
4.F.1.
A Central Limit Theorem
4.F.2.
The Law of Large Numbers
S4.A. Lattice Random Walk
S4.A.1. One-Dimensional Lattice
S4.A.2. Random Walk in Higher Dimension

173
173
174
175
177
178
180
182
183
188
188
191
192
194
197
197
198
199
200
203

CONTENTS

S4.B. Infinitely Divisible Distributions


S4.B.l. Gaussian Distribution
S4.B.2. Poisson Distribution
S4.B.3. Cauchy Distribution
S4.BA.
Levy Distribution
S4.C. The Central Limit Theorem
S4.C.l. Useful Inequalities
S4.C.2. Convergence to a Gaussian
S4.D. Weierstrass Random Walk
S4.D.l. Discrete One-Dimensional Random Walk
S4.D.2. Continuum Limit of One-Dimensional Discrete
Random Walk
S4.D.3. Two-Dimensional Discrete Random Walk
(Levy Flight)
S4.E. General Form of Infinitely Divisible Distributions
S4 1
T _. .
1:'

S4
References
Problems

V'L'

T"'O

Converted with

STOI Converter

207
208
209
209
210
211
212
213
214
215
217
218
221
222
223
225
225

trial version
5.

Stochastic
5.A.
5.B.
5.C.

hDP://www.stdutilitv.com

Introduction
General Theory
Markov Chains
5.C.l.
Spectral Properties
5.C.2.
Random Walk
5.D. The Master Equation
5.D.l.
Derivation of the Master Equation
5.D.2.
Detailed Balance
5.D.3.
Mean First Passage Time
5.E.
Brownian Motion
5.E.l.
Langevin Equation
5.E.2.
The Spectral Density (Power Spectrum)
S5.A. Time Periodic Markov Chain
S5.B. Master Equation for Birth-Death Processes
S5.B.1. The Master Equation
S5.B.2. Linear Birth-Death Processes
S5.B.3. Nonlinear Birth-Death Processes

229
229
231
234
234
240
241
242
244
247
250
251
254
258
260
260
261
265

xi

CONTENTS

S5.C.

The Fokker-Planck Equation


S5.C.1. Probability Flow in Phase Space
S5.C.2. Probability Flow for Brownian Particle
S5.C.3. The Strong Friction Limit
S5.CA. Solution of Fokker-Planck Equations with
One Variable
S5.D. Approximations to the Master Equation
References
Problems

271
276
278
279

6.

285

The Foundations of Statistical Mechanics

6.A.
6.B.
6.C.
6.D.
S6.A.
S6.B.
S6.C.
S6.D.
S6.E.
S6.F.
References
Problems

Introduction
The Liouville Equation of Motion
Ergodic Theory and the Foundation of Statistical Mechanics
The Quantum Probability Density Operator
r...
_.,
Hierarchy
-no

.1

.1...,.

.1 ..

'1'

c1.

...,....,. ......-rT~r

Converted with

STOU Converter
trial ve rsion

hnp://www.stdutilitv.com

~----------------------------~

PART THREE

EQUILIBRIUM

7.D.

7.E.

285
286
296
303
310
314
319
321
326
334
335
336

STATISTICAL MECHANICS

7. Equilibrium Statistical Mechanics


7.A.
7.B.
7.C.

~bution

266
266
267
270

Introduction
The Microcanonical Ensemble
Einstein Fluctuation Theory
7.C.1.
General Discussion
7.C.2.
Fluid Systems
The Canonical Ensemble
7.D.1.
Probability Density Operator
7.D.2.
Systems of Indistinguishable Particles
Systems of Distinguishable Particles
7.D.3.
Heat Capacity of a Debye Solid

341
341
343
349
349
351
354
354
357
362
364

xii

CONTENTS

7.F.

Order-Disorder Transitions
7.F.1.
Exact Solution for a One-Dimensional Lattice
7.F.2.
Mean Field Theory for a d-Dimensional Lattice
7.G. The Grand Canonical Ensemble
7.H.
Ideal Quantum Gases
7.H.l.
Bose-Einstein Gases
7.H.2.
Fermi-Dirac Ideal Gases
S7.A. Heat Capacity of Lattice Vibrations on a OneDimensional Lattice-Exact Solution
S7.A.I. Exact Expression-Large N
S7.A.2. Continuum Approximation-Large N
S7.B. Momentum Condensation in an Interacting Fermi Fluid
S7.C. The Yang-Lee Theory of Phase Transitions
References
Problems

Converted with

8. Order-Di
8.A.
8.B.

STOI Converter

369
370
372
377
381
383
392
401
404
406
407
418
422
423
427

8.
8.C.
Sc
8.C.I.
Homogeneous Functions
8.C.2.
Widom Scaling
8.C.3.
Kadanoff Scaling
8.D. Microscopic Calculation of Critical Exponents
S8.A. Critical Exponents for the S4 Model
S8.B. Exact Solution of the Two-Dimensional Ising Model
S8.B.I. Partition Function
S8.B.2. Antisymmetric Matrices and Dimer Graphs
S8.B.3. Closed Graphs and Mixed Dimer Graphs
S8.B.4. Partition Function for Infinite Planar Lattice
References
Problems

427
428
429
431
433
433
434
437
440
448
462
462
466
469
475
485
486

9.

488

trial version

hDP://www.stdutilitv.com

Interacting Fluids
9.A.

9.B.

Introduction
Thermodynamics and the Radial Distribution Function

488
489

xiii

CONTENTS

9.C.

S9.A.

S9.B.
S9.C.

S9.D.
References
Problems

Virial Expansion of the Equation of State


Virial Expansions and Cluster Functions
9.C.1.
The Second Virial Coefficient
9.C.2.
Higher-Order Virial Coefficients
9.C.3.
The Pressure and Compressibility Equations
S9.A.1. The Pressure Equation
S9.A.2. The Compressibility Equation
Omstein-Zemicke Equation
Third Virial Coefficient
S9.C.1. Square-Well Potential
S9.C.2. Lennard-Jones 6-12 Potential
Virial Coefficients for Quantum Gases

PART FOUl
10. Hydrod;
10.A. I
1O.B. r
1

Converted with

492
493
500
506
507
508
509
510
513
514
515
517
526
527

CHANICS

STOI Converter
trial version

hDP://www.stdutilitv.com

1O.B.2. Entropy Source and Entropy Current


1O.B.3. Transport Coefficients
1O.C. Linearized Hydrodynamic Equations
10.C.l. Linearization of the Hydrodynamic Equations
1O.C.2. Transverse Hydrodynamic Modes
1O.C.3. Longitudinal Hydrodynamic Modes
1O.D. Dynamic Equilibrium Fluctuations and Transport Processes
10.D.I. Onsager's Relations
1O.D.2. Weiner-Khintchine Theorem
1O.E. Linear Response Theory and the Fluctuation-Dissipation
Theorem
10.E.1. The Response Matrix
10.E.2. Causality
1O.E.3. The Fluctuation-Dissipation Theorem
10.E.4. Power Absorption
10.P. Transport Properties of Mixtures
1O.P.I. Entropy Production in Multicomponent Systems

531
531
533
534
537
541
544
545
549
550
552
553
557
561
562
563
568
570
574
574

xiv

CONTENTS

1O.P.2. Fick's Law for Diffusion


10.P.3. Thermal Diffusion
10.P.4. Electrical Conductivity and Diffusion in Fluids
S 1O.A. Onsager's Relations When a Magnetic Field is Present
S1O.B. Microscopic Linear Response Theory
S10.C. Light Scattering
S1O.C.1. Scattered Electric Field
S 1O.C.2. Intensity of Scattered Light
SlO.D. Thermoelectricity
SIO.D.1. The Peltier Effect
S IO.D.2. The Seebeck Effect
S IO.D.3. Thomson Heat
S10.E. Entropy Production in Discontinuous Systems
S IO.E.l. Volume Flow Across a Membrane
S10.E.2. Ion Transport Across a Membrane
SlO.P. Stc
Converted with
SI
hctions
SI
SI
SIO.O. La
trial version
~
SI
SI
S10.0.3. Velocity Autocorrelation Function
S 1O.H. Superfluid Hydrodynamics
S10.H.l. Superfluid Hydrodynamic Equations
S10.H.2. Sound Modes
S10.1. General Definition of Hydrodynamic Modes
S 10.1.1. Projection Operators
S10.1.2. Conserved Quantities
S 10.1.3. Hydrodynamic Modes Due to Broken Symmetry
References
Problems

580
582
583
586
589
592
594
597
600
601
603
605
605
606
610
612
613
614
617
620
621
623
624
631
631
635
639
640
642
644
649
650

11. Transport Theory

656

STOI Converter
hDP://www.stdutilitv.com

II.A.
II.B.

Introduction
Elementary Transport Theory
II.B.l.
The Maxwell-Boltzmann Distribution
II.B.2.
The Mean Free Path

656
657
657
658

CONTENTS

xv

11.B.3.
11.B.4.
11.B.S.

The Collision Frequency


Self-Diffusion
The Coefficients of Viscosity and
Thermal Conductivity
11.B.6. The Rate of Reaction
11.C. The Boltzmann Equation
l1.C.1. Two-Body Scattering
11.C.2. Derivation of the Boltzmann Equation
11.C.3. Boltzmann's H Theorem
11.0. Linearized Boltzmann and Lorentz-Boltzmann Equations
11.0.1. Kinetic Equations for a Two-Component Gas
11.0.2. Collision Operators
11.E. Coefficient of Self-Diffusion
11.E.1. Derivation of the Diffusion Equation
11.E.2. Eigenfrequencies of the Lorentzn .,

659
661
664
666
670
671
679
680
682
682
684
688
688

11.G.2. Diffusion Coefficient


11.G.3. Thermal Conductivity
11.G.4. Shear Viscosity
S11.A. Beyond the Boltzmann Equation
References
Problems

690
691
692
697
700
701
702
703
704
708
710
717
718

12. Nonequilihriwn Phase Transitions

721

11.F.

(
1
1
1

l1.G.

(
1

Converted with

STOI Converter
trial version

ty
ons
~ation
ity

hDP://www.stdutilitv.com

12.A. Introduction
12.B. Nonequilibrium Stability Criteria
12.B.1. Stability Conditions Near Equilibrium
12.B.2. Stability Conditions Far From Equilibrium
12.C. The Schlogl Model
12.0. The Brusselator
12.0.1. The Brusselator-A Nonlinear
Chemical Model
12.0.2. Boundary Conditions

721
722
723
726
732
735
736
737

xvi

CONTENTS

12.D.3. Linear Stability Analysis


12.E. The Rayleigh-Benard Instability
12.E.l.
Hydrodynamic Equations and
Boundary Conditions
12.E.2. Linear Stability Analysis
SI2.A. Fluctuations Near a Nonequilibrium Phase Transition
SI2.A.1. Fluctuations in the Rayleigh-Benard System
SI2.A.2. Fluctuations in the Brusselator
SI2.A.3. The Time-Dependent Ginzburg-Landau Equation
References
Problems

739
742
743
747
753
753
760
764
765
767

APPENDICES
A. Balance

lJ{".ana.tJ.J~:..__

A.l.
G
A.2.
G
References
B.

Systems (
B.1.

B.2.

p(..,__

----,

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com
~------____J

B.1.1.
Free Particle
B.1.2.
Particle in a Box
Symmetrized N-Particle Position and
Momentum Eigenstates
B.2.1.
Symmetrized Momentum Eigenstates for
Bose-Einstein Particles
B.2.2.
Antisymmetrized Momentum Eigenstates
for Fermi-Dirac Particles
B.2.3.
Partition Functions and Expectation Values
The Number Representation
B.3.1.
The Number Representation for Bosons
B.3.2.
The Number Representation for Fermions
B.3.3.
Field Operators

768
768
771
773
774
774
775
776
777
778

References

779
780
781
782
785
788
790

C.

791

B.3.

Stability of Solutions to Nonlinear Equations


C.l.

Linear Stability Theory

791

CONTENTS

xvii

C.2.
Limit Cycles
C.3.
Liapounov Functions and Global Stability
References

795
796
798

Author Index

799

Subject Index

804

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com

PREFACE

In 1992 after finishing my book, "The Transition of Chaos," I realized that I


needed to write a new edition of "A Modem Course in Statistical Physics". I
wanted to adjust the material to better prepare students for what I believe are the
current directions of statistical physics. I wanted to place more emphasis on
nonequilibrium processes and on the thermodynamics underlying biological
processes. I also wanted to be more complete in the presentation of material. It
turned out to be a greater task than I had anticipated, and now five years later I
am finally finishing the second edition. One reason it has taken so long is that I
have created a detailed solution manual for the second edition and I have added
many worked out exercises to the text. In this way I hope I have made the second
edition much more student and instructor friendly than the first edition was.
There are two individuals who have had a particularly large influence on this
book and whom I want to thank, even though they took no part in writing the
book. (Any
. c
sponsibility.) The
biggest influ
Converted with
I have dedicated
this book to
~exas to join the
Physics fac
the Center for
Thermodyna
~s the Prigogine
Center for S
trial version
~ training was in
equilibrium
learned that the
focus of thi
f Texas, was on
nonequilibrium nonlinear phenomena, most of it far from equilibrium. I began
to work on nonequilibrium and nonlinear phenomena, but followed my own
path. The opportunity to teach and work in this marvelous research center and
to listen to the inspiring lectures of Ilya Prigogine and lectures of the many
visitors to the Center has opened new worlds to me, some of which I have tried
to bring to students through this book.
The other individual who has had a large influence on this book is Nico van
Kampen, a sometimes visitor to the University of Texas. His beautiful lectures on
stochastic processes were an inspiration and spurred my interest in the subject.
I want to thank the many students in my statistical mechanics classes who
helped me shape the material for this book and who also helped me correct the
manuscript.
This book covers a huge range of material. I could not reference all the work
by the individuals who have contributed in all these areas. I have referenced
work which most influenced my view of the subject and which could lead
students to other related work. I apologize to those whose work I have not been
able to include in this book.
L. E. Reichl
Austin, Texas
September 1997
.cC_

il

_]'.1

.1L

STOI Converter
hDP://www.stdutilitv.com

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com

1
INTRODUCTION

I.A. OVERVIEW
The field of statistical physics has expanded dramatically in recent years. New
results in ergodic theory, nonlinear chemical physics, stochastic theory,
quantum fluids, critical phenomena, hydrodynamics, transport theory, and
esults are rarely
biophysics i.
presented in
Converted with
und in statistical
physics can ;
ten in an effort to
incorporate t
~ysics. It includes
in a unified
tical physics and
develops frc
trial version
snd the concepts
underlying n
In the fiel
pas deepened our
understanding of the structure and dynamical behavior of a variety of nonlinear
systems and has made ergodic theory a modem field of research. Indeed, one of
the frontiers of science today is the study of the spectral properties of decay
processes in nature, based on the chaotic nature of the underlying dynamics of
those systems. Advances in this field have been aided by the development of
ever more powerful computers. In an effort to introduce this field to students, a
careful discussion is given of the behavior of probability flows in phase space,
including specific examples of ergodic and mixing flows.
Nonlinear chemical physics is still in its infancy, but it has already given a
conceptual framework within which we can understand the thermodynamic
origin of life processes. The discovery of dissipative structures (nonlinear
spatial and temporal structures) in nonlinear nonequilibrium chemical systems
has opened a new field in chemistry and biophysics. In this book, material has
been included on chemical thermodynamics, chemical hydrodynamics, and
nonequilibrium phase transitions in chemical and hydrodynamic systems.
The use of stochastic theory to study fluctuation phenomena in chemical and
hydrodynamic systems, along with its growing use in population dynamics and
complex systems theory, has brought new life to this field. The discovery of
scaling behavior at all levels of the physical world, along with the appearance of
Levy flights which often accompanies scaling behavior, has forced us to think

_1

_1

STOI Converter
hDP://www.stdutilitv.com

INTRODUCTION

beyond the limits of the Central Limit Theorem. In order to give students some
familiarity with modem concepts from the field of stochastic theory, we have
placed probability theory in a more general framework and discuss, within that
framework, classical random walks, Levy flights, and Brownian motion.
The theory of superfluids rarely appears in general textbooks on statistical
physics, but the theory of such systems is incorporated at appropriate places
throughout this book. We discuss the thermodynamic properties of superfluid
and superconducting systems, the Ginzburg-Landau theory of superconductors,
the BCS theory of superconductors, and superfluid hydrodynamics. Also
included in the book is an extensive discussion of properties of classical fluids
and their thermodynamic and hydrodynamic properties.
The theory of phase transitions has undergone a revolution in recent years. In
this book we define critical exponents and use renormalization theory to
compute them. We also derive an exact expression for the specific heat of the
two-dimensional Ising system, one of the simplest exactly solvable systems
which can exhibit a phase transition. At the end of the book we include an
introduction to the theory of nonequilibrium phase transitions.
Hydrodynami
rf
I
g-wavelength
and
biological
phenomena ~nc
Converted with
systems. ThIS
namics based
on the underlyi
r. We discuss
properties of
on-dissipation
theorem, the
odynamics in
trial version
terms of cons
lso include a
variety of appli
t essential for
biophysics.
Transport theory is discussed from many points of view. We derive Onsager's
relations for transport coefficients. We derive expressions for transport
coefficients based on simple "back of the envelope" mean free path arguments.
The Boltzmann and Lorentz-Boltzmann equations are derived and microscopic
expressions for transport coefficients are obtained, starting from spectral
properties of the Boltzmann and Lorentz-Boltzmann collision operators. The
difficulties in developing a convergent transport theory for dense gases are also
reviewed.
Concepts developed in statistical physics underlie all of physics. Once the
forces between microscopic particles are determined, statistical physics gives us
a picture of how microscopic particles act in the aggregate to form the
macroscopic world. As we see in this book, what happens on the macroscopic
scale is sometimes surprising.

STOI Converter
hDP://www.stdutilitv.com

I.B. PLAN OF BOOK


Thermodynamics is a consequence and a reflection of the symmetries of nature.
It is what remains after collisions between the many degrees of freedom of

PLAN OF BOOK

macroscopic systems randomize and destroy most of the coherent behavior. The
quantities which cannot be destroyed, due to underlying symmetries of nature
and their resulting conservation laws, give rise to the state variables upon which
the theory of thermodynamics is built. Thermodynamics is therefore a solid and
sure foundation upon which we can construct theories of matter out of
equilibrium. That is why we place heavy emphasis on it in this book.
The book is divided into four parts. Chapters 2 and 3 present the foundations
of thermodynamics and the thermodynamics of phase transitions. Chapters 4
through 6 present probability theory, stochastic theory, and the foundations of
statistical mechanics. Chapters 7 through 9 present equilibrium statistical
mechanics, with emphasis on phase transitions and the equilibrium theory of
classical fluids. Chapters 10 through 12 deal with nonequilibrium processes,
both on the microscopic and macroscopic scales, both near and far from
equilibrium. The first two parts of the book essentially lay the foundations for
the last two parts.
There seems to be a tendency in many books to focus on equilibrium
statistical mechanics and derive thermodynamics as a consequence. As a result,
students do
.
.
vast world of
thermodyna
Converted with
stems which are
too complica
begin the book

;:d:t~~:Oo STDOConverter

e:~::~:v~:

a large part
trial version
do not involve
phase transiti
is, and chemical
thermodyna
mics of phase
transitions an
e use 0
ermo ynarmc staoiu y eory In analyzing these
phase transitions. We discuss first-order phase transitions in liquid-vapor-solid
transitions, with particular emphasis on the liquid-vapor transition and its
critical point and critical exponents. We also introduce the Ginzburg-Landau
theory of continuous phase transitions and discuss a variety of transitions which
involve broken symmetries.
Having developed some intuition concerning the macroscopic behavior of
complex equilibrium systems, we then turn to microscopic foundations.
Chapters 4 through 6 are devoted to probability theory and the foundations of
statistical mechanics. Chapter 4 contains a review of basic concepts from
probability theory and then uses these concepts to describe classical random
walks and Levy flights. The Central Limit Theorem and the breakdown of the
Central Limit Theorem for scaling processes is described. In Chapter 5 we
study the dynamics of discrete stochastic variables based on the master
equation. We also introduce the theory of Brownian motion and the idea of
separation of time scales, which has proven so important in describing
nonequilibrium phase transitions. The theory developed in Chapter 5 has many
applications in chemical physics, laser physics, population dynamics, and
biophysics, and it prepares the way for more complicated topics in statistical
mechanics.

hDP://www.stdutilitv.com

INTRODUCTION

Chapter 6 lays the probabilistic foundations of statistical mechanics, starting


from ergodic theory. In recent years, there has been a tendency to sidestep this
aspect of statistical physics completely and to introduce statistical mechanics
using information theory. The student then misses one of the current frontiers of
modern physics, the study of the spectral behavior of decay processes in nature,
based on the chaotic nature of the underlying dynamics of those systems. While
we cannot go very far into this subject in this book, we at least discuss the
issues. We begin by deriving the Liouville equation, which is the equation of
motion for probability densities, both in classical mechanics and in quantum
mechanics. We look at the types of flow that can occur in mechanical systems
and introduce the concepts of ergodic and mixing flows, which appear to be
minimum requirements if a system is to decay to thermodynamic equilibrium.
Chapters 7-9 are devoted entirely to equilibrium statistical mechanics. In
Chapter 7 we derive the probability densities (the microcanonical, canonical,
and grand canonical ensembles) for both closed and opened systems and relate
them to thermodynamic quantities and the theory of fluctuations. We then use
them to derive the thermodynamic properties of a variety of model systems,
including h
.
'"
s, and superconductors.
In Chapter 8
f spin systems

STOI Converter

and show quali


tween fluctuations diverges
ce the idea of
scaling and us
trial version
xpressions for
the critical ex
Chapter 8 by
obtaining an ex
ensional Ising
lattice, and we compare our exac expressions 0 ose 0 mean field theory.
Chapter 9 is devoted to the equilibrium theory of classical fluids. In this
chapter we relate the thermodynamic properties of classical fluids to the
underlying radial distribution function, and we use the Ursell-Mayer cluster
expansion to obtain a virial expansion of the the equation of state of a classical
fluid. We also discuss how to include quantum corrections for nondegenerate
gases.
The last part of the book, Chapters 10-12, deals with nonequilibrium
processes. Chapter 10 is devoted to hydrodynamic processes for systems near
equilibrium. We begin by deriving the Navier-Stokes equations from the
symmetry properties of a fluid of point particles, and we use the derived
expression for entropy production to obtain the transport coefficients for the
system. We use the solutions of the linearized Navier-Stokes equations to
predict the outcome of light-scattering experiments. We go on to derive
Onsager's relations between transport coefficients, and we use causality to
derive the fluctuation-dissipation theorem. We also derive a general expression
for the entropy production in systems with mixtures of particles which can
undergo chemical reactions. We then use this theory to describe thermal and
chemical transport processes in mixtures, across membranes, and in electrical
circuits. The hydrodynamic equations describe the behavior of just a few slowly

hDP://www.stdutilitv.com

USE AS A TEXTBOOK

varying degrees of freedom in fluid systems. If we assume that the remainder of


the fluid can be treated as a background noise, we can use the fluctuationdissipation theorem to derive the correlation functions for this background
noise. In Chapter lOwe also consider hydrodynamic modes which result from
broken symmetries, and we derive hydrodynamic equations for superfluids and
consider the types of sound that can exist in such fluids.
In Chapter 11 we derive microscopic expressions for the coefficients of
diffusion, shear viscosity, and thermal conductivity, starting both from mean free
path arguments and from the Boltzmann and Lorentz-Boltzmann equations. In
deriving microscopic expressions for the transport coefficients from the
Boltzmann and Lorentz-Boltzmann equations, we use a very elegant method
which relies on use of the eigenvalues and eigenfunctions of the collision
operators associated with those equations. We obtain explicit microscopic
expressions for the transport coefficients of a hard sphere gas.
Finally, in Chapter 12 we conclude with the fascinating subject of
nonequilibrium phase transitions. We discuss thermodynamic stability theory
for systems far from equilibrium. We also show how nonlinearities in the rate
equations for
.
...
uilibrium phase
transitions wh
Converted with
.cal waves, and
spatially peri
the RayleighBenard hydro
shall also ex
point for thes
slowing down

STOI Converter
trial version

ction cells. We
d of the critical
d by a critical

hDP://www.stdutilitv.com

I.C. USE AS A TEXTBOOK


Even though this book contains a huge amount of material, it has been
designed to be used as a textbook. In each chapter the material has been divided
into core topics and special topics. The core topics provide key basic material in
each chapter, while special topics illustrate these core ideas with a variety of
applications. The instructor can select topics from the special topics sections,
according to the emphasis he/she wishes to give the course.
In many sections, we have included nontrivial demonstration exercises to
help the students understand the material and to help in solving homework
problems. Each chapter has a variety of problems at the end of the chapter that
can be used to help the students test their understanding.
Even if one covers only the core topics of each chapter, there may be too
much material to cover in a one-semester course. However, the book is designed
so that some chapters may be omitted completely. The choice of which chapters
to use depends on the interests of the instructor. Our suggestion for a basic
well-rounded one-semester course in statistical physics is to cover the core
topics in Chapters 2, 3, 4, 7, 10, and 11 (only Section l1.B if time is running
short).

INTRODUCTION

The book is intended to introduce the students to a variety of subjects and


resource materials which they can then pursue in greater depth if they wish. We
have tried to use standardized notation as much as possible. In writing a book
which surveys the entire field of statistical physics, it is impossible to include or
even to reference everyone's work. We have included references which were
especially pertinent to the points of view we take in this book and which will
lead students easily to other work in the same field.

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com

PART ONE
THERMODYNAMICS

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com

2
INTRODUCTION TO
THERMODYNAMICS

2.A. INTRODUCTORY

REMARKS

The science of thermod namics be an with the observation that matter in the
aggregate can
Converted with
0 not change in
time. These
ed by definite
mechanical p
substance bee
given equilib
back to the sa

STDI Converter

change as the
. However, any
trial version
g the substance
tate, all changes
some external
influence acts
ucibility of the
equilibrium states can be seen everywhere in the world around us.
Thermodynamics has been able to describe, with remarkable accuracy, the
macroscopic behavior of a huge variety of systems over the entire range of
experimentally accessible temperatures (10-4 K to 106 K). It provides a truly
universal theory of matter in the aggregate. And yet, the entire subject is based
on only four laws, which may be stated rather simply as follows: Zeroth Lawit is possible to build a thermometer; First Law-energy is conserved; Second
Law-not all heat energy can be converted into work; and Third Law-we can
never reach the coldest temperature using a finite set of reversible steps.
However, even though these laws sound rather simple, their implications are
vast and give us important tools for studying the behavior and stability of
systems in equilibrium and, in some cases, of systems far from equilibrium.
The core topics in this chapter focus on a review of various aspects of
thermodynamics that will be used throughout the remainder of the book. The
special topics at the end of this chapter give a more detailed discussion of some
applications of thermodynamics which do not involve phase transitions. Phase
transitions will be studied in Chapter 3.
We shall begin this chapter by introducing the variables which are used in
thermodynamics and the mathematics needed to calculate changes in the
thermodynamic state of a system. As we shall see, many different sets of

hDP://www.stdutilitv.com

10

INTRODUCTION TO THERMODYNAMICS

mechanical variables can be used to describe thermodynamic systems. In order


to become familiar with some of these mechanical variables, we shall write the
experimentally observed equations of state for a variety of thermodynamic
systems.
As we have mentioned above, thermodynamics is based on four laws. We
shall discuss the content of these laws in some detail, with particular emphasis
on the second law. The second law is extremely important both in equilibrium
and out of equilibrium because it gives us a criterion for testing the stability of
equilibrium systems and, in some cases, nonequilibrium systems.
There are a number of different thermodynamic potentials that can be used to
describe the behavior and stability of thermodynamic systems, depending on
the type of constraints imposed on the system. For a system which is isolated
from the world, the internal energy will be a minimum for the equilibrium state.
However, if we couple the system thermally, mechanically, or chemically to the
outside world, other thermodynamic potentials will be minimized. We will introduce the five most commonly used thermodynamic potentials (internal energy,
enthalpy, Helmholtz free energy, Gibbs free energy, and the grand potential),
and we will di
..
.. ized at equilibrium and wh
When expe
, the quantities
which are easi
lly, we change
one parameter
sponds to that
change, under
trial version
asures the way
in which the s
his chapter we
shall introduce
ctions and give
relations betwe'-e-n-t'--h-em-.
-----------------'

STOI Converter
hDP://www.stdutilitv.com

Isolated equilibrium systems are systems in a state of maximum entropy.


Any fluctuations which occur in such systems must cause a decrease in entropy
if the equilibrium state is to be stable. We can use this fact to find relations
between the intensive state variables for different parts of a system if those parts
are to be in mechanical, thermal, and chemical equilibrium. In addition, we can
find restrictions on the sign of the response functions which must be satisfied for
stable equilibrium. We shall find these conditions and discuss the restrictions
they place on the Helmholtz and Gibbs free energy.
Thermodynamics becomes most interesting when it is applied to real
systems. In order to demonstrate its versatility, in the section on special topics,
we shall apply it to a number of systems which have been selected for their
practical importance or conceptual importance.
We begin with a subject of great practical and historic importance, namely,
the cooling of gases. it is often necessary to cool substances below the
temperature of their surroundings. The refrigerator most commonly used for
this purpose is based on the Joule-Kelvin effect. There are two important ways
to cool gases. We can let them do work against their own intermolecular forces
by letting them expand freely (Joule effect); or we can force them through a
small constriction, thus causing cooling at low temperatures or heating at high

STATE VARIABLES AND EXACT DIFFERENTIALS

11

temperatures (Joule-Kelvin effect). The Joule-Kelvin effect is by far the more


effective of the two methods. We shall discuss both methods in this chapter and
use the van der Waals equation of state to obtain estimates of the coolling
effects for some real gases.
For reversible processes, changes in entropy content can be completely
accounted for in terms of changes in heat content. For irreversible processes,
this is no longer true. We can have entropy increase in an isolated system, even
though no heat has been added. Therefore, it is often useful to think of an
increase in entropy as being related to an increase in disorder in a system. One
of the most convincing illustrations of this is the entropy change which occurs
when two substances, which have the same temperature and pressure but
different identities, are mixed. Thermodynamics predicts that the entropy will
increase solely due to mixing of the substances.
When the entropy of a system changes due to mixing, so will other thermodynamic quantities. One of the most interesting examples of this is osmosis. We
can fill a container with water and separate it into two parts by a membrane
permeable to water but not salt, for example. If we put a small amount of salt
into one side,
.,
rease markedly
because of mi
Converted with
Chemical r
ay in terms of a

STDI Converter

thermodynarni
measure of the
distance of a
urn and will be
useful in later
trial version
of equilibrium.
We can obtai
conditions for
thermodynami
r 2, and at the
same time we can learn a number of interesting acts a out e thermodynamic
behavior of chemical reactions. A special example of a type of reaction
important to biological systems is found in electrolytes, which consist of salts
which can dissociate but maintain an electrically neutral solution.

hDP:llwww.stdutililV.com

2.B. STATE VARIABLES AND EXACT DIFFERENTIALS


Thermodynamics describes the behavior of systems with many degrees of
freedom after they have reached a state of thermal equilibrium-a
state in
which all past history is forgotten and all macroscopic quantities cease to
change in time. The amazing feature of such systems is that, even though they
contain many degrees of freedom (rv 1023) in chaotic motion, their thermodynamic state can be specified completely in terms of a few parameters=-called
state variables. In general, there are many state variables which can be used to
specify the thermodynamic state of a system, but only a few (usually two or
three) are independent. In practice, one chooses state variables which are
accessible to experiment and obtains relations between them. Then, the
"machinery" of thermodynamics enables one to obtain the values of any other
state variables of interest.

12

INTRODUCTION TO THERMODYNAMICS

State variables may be either extensive or intensive. Extensive variables


always change in value when the size (spatial extent and number of degrees of
freedom) of the system is changed, and intensive variables need not. Certain
pairs of intensive and extensive state variables often occur together because
they correspond to generalized forces and displacements which appear in
expressions for thermodynamic work. Some examples of such extensive and
intensive pairs are, respectively, volume, V, and pressure, P; magnetization, M,
and magnetic field strength, H; length, L, and tension, J; area, A, and surface
tension, IT; electric polarization, P, and electric field, E. The pair of state
variables related to heat content of thermodynamic system are the temperature,
T, which is intensive, and the entropy, S, which is extensive. There is also a pair
of state variables associated with "chemical" properties of a system. They are
the number of particles, N, which is extensive, and the chemical potential per
particle, J-L' , which is intensive. In this book we shall sometimes use the number
of moles, n, and the chemical potential per mole, J-L (molar chemical potential);
or the mass of a substance, M, and the chemical potential per unit mass, ii,
(specific chemical potential), as the chemical state variables. If there is more
than one type
..
.
le number and
chemical pote
Other state
behavior of a
y, C; compressystem are the
sibility, K:; ma
mic potentials,
ergy, A; Gibbs
such as the int
trial version
free energy,
e thoroughly
acquainted wi
If we change e t ermo ynarruc state 0 our system, t e amount by which
the state variables change must be independent of the path taken. If this were
not so, the state variables would contain information about the history of the
system. It is precisely this property of state variables which makes them so
useful in studying changes in the equilibrium state of various systems. Mathematically changes in state variables correspond to exact differentials [1];
therefore, before we begin our discussion of thermodynamics, it is useful to
review the theory of exact differentials. This will be the subject of the
remainder of this section.
Given a function F = F(XI' X2) depending on two independent variables Xl
and X2, the differential of F is defined as follows:

STOI Converter
hDP://www.stdutilitv.com

dF = (::.) x,dxt

+ (:~)

x,dx2,

(2.1 )

where (oF /OXI)x2 is the derivative of F with respect to Xl holding X2 fixed. If F


and its derivatives are continuous and
(2.2)

13

STATE VARIABLES AND EXACT DIFFERENTIALS

then dF is an exact differential.

If we denote

then the variables Cl and Xl and variables C2 and X2 are called


variables with respect to the function F.
The fact that dF is exact has the following consequences:

"conjugate"

(a) The value of the integral

is independent of the path taken between A and B and depends only on


the end points A and B.
(b) The integ/--U-'-....I...LI.-.LLL---LI..L.I'-'.LLLJL.L.L....<J.____..._..L.L3.L<.LI.......L.LLI.LLL...L:1--L.o.1
__
~

Converted with
(c) If one kn

If F depends

STOI Converter
trial version

n be found to

hDP://www.stdutilitv.com
,

generalize in a simple way: Let F


may be written

dF

= F(Xl,X2,

L (8F)8x.

given above
... ,xn), then the differential, dF,

i=l

(2.3)

dx..
{X#i}

The notation (8F j8Xi) {X#i} means that the derivative of F is taken with respect
to Xi holding all variables but Xi constant. For any pair of variables,
the
following relation holds:

An example

for the case of three independent

variables

is

14

INTRODUCTION TO THERMODYNAMICS

The Eq. (2.4) leads to the result

Differentials of all state variables are exact and have the above properties.
Given four state variables, x, y, z, and w, where w is a function of any two of the
variables x, y, or z, one can obtain the following useful relations along paths for
which F(x,y,z) = 0:
(2.5)

(2.6)

Converted with

STOI Converter

(2.7)

trial version

(2.8)

hDP://www.stdutilitv.com
It is a simple matter to derive Eqs. (2.5)-(2.8). We will first consider Eqs. (2.5)

and (2.6). Let us choose variables y and z to be independent, x = x(y, z), and
then choose x and z to be independent, y = y(x, z), and write the following
differentials;
dx = (8xjBy)zdy + (8xj8z)ydz
and dy = (Byj8x)zdx + (Byj
8z)xdz.
If we eliminate dy between these equations, we obtain

[(:),

(:) ,-l]dx + [(:), (:)

+ (:;)

dz = O.

Because dx and dz may be varied independently, their coefficients may be set


equal to zero separately. The result is Eqs. (2.5) and (2.6).
To derive Eq. (2.7) we let y and z be independent so that x = x(y, z) and write
the differential for dx. If we then divide by dw, we obtain

For constant z, dz

0 and we find Eq. (2.7)

15

STATE VARIABLES AND EXACT DIFFERENTIALS

Finally, to derive Eq. (2.8) we let x be a function of y and w,X = x(y, w). If
we write the differential of x, divide it by dy, and restrict the entire equation to
constant Z, we obtain Eq. (2.8).
When integrating exact differentials, one must be careful not to lose terms. In
Exercise 2.1, we illustrate two different methods for integrating
exact
differentials .

EXERCISE 2.1. Consider the differential dd: = (xl + y)dx + xdy. (a)
Show that it is an exact differential. (b) Integrate de/>between the points A
and B in the figure below, along the two different paths, 1 and 2. (c) Integrate
dd: between points A and B using indefinite integrals.
YB

- - -

Converted with

STOI Converter
Answer:
(a) From

trial version

hDP://www.stdutilitv.com

can

write

(8e/>/8x
= + y and
x = x. mce
(8e/>/8x)y]x =
[(8/8x)C8e/>/8yUy
= 1, the differential, dd; is exact.
(b) Let us first integrate the differential,
de, along path 1:
e/>B- e/>A= JXB (r

+ YA) dx + JYB

XA

1 3

XB dy

(1)

YA

= '3XB + XBYB -

1 3

'3XA - XAYA

Let us next integrate the differential,


dd; along path 2. Note that along
2, Y = YA + (~y/ &)(x
- XA), where ~y = YB - YA and
& = XB - XA If we substitute this into the expression for de, we
find
de/>= (xl + y)dx + xdy = [xl + YA + (~y/ &)(2x - xA)]dx.
Therefore

path

e/>B- e/>A=

XB

XA

='3X1 B 3

[
dx x2

+ YA + ~Y
& (2x -

+ XBYB-'3xA-1 XAYA'
3

XA)

(2)

Note that the change in e/> in going from point A to point B is


independent of the path taken. This is a property of exact differentials.

16

INTRODUCTION TO THERMODYNAMICS

(c) We now integrate the differential, de, in a different way. Let us first
do the indefinite integrals

J(::)

ydx =

J(x

+ y)dx

= thxy +Kt(Y),

(3)

where K, (y) is an unknown function of y. Next do the integral


(4)
where K2(X) is an unknown function of x. In order for Eqs. (3) and (4)
to be consistent, we must choose K2(x) = tx3 + K3 and K, (y) = K3,
where K3 is a constant. Therefore, c/> = x3 + xy + K3 and again,

".I,.

lpB -

".1,._13+

'PA -

'j"XB

XBYB

13

'j"XA -

XAYA

2.C. SOME

Converted with

An equation 0
system in equ
freedom neede
relates the the
variables for t

STDI Converter
trial version

hDP://www.stdutilitv.com

E
variables for a
ent degrees of
quation which
chemical state
tion about the

2.C.I. Ideal Gas Law


The best-known equation of state is the ideal gas law
PV= nRT,

(2.9)

where n is the number of moles, T is temperature in degrees Kelvin, P is the


pressure in Pascals, V is the volume in cubic meters, and R = 8.314 Jzmol- K is
the universal gas constant. The ideal gas law gives a good description of a gas
which is so dilute that the effect of interaction between particles can be
neglected.
If there are m different types of particles in the gas, then the ideal gas law
takes the form
m

PV = EniRT,
i=1

where n, is the number of moles of the ith constituent.

(2.10)

17

SOME MECHANICAL EQUATIONS OF STATE

2.C.2. Virial Expansion [2]


The virial expansion,
(2.11 )
expresses the equation of state of a gas as a density expansion. The quantities

B2(T) and B3(T) are called the second and third virial coefficients and are
functions of temperature only. As we shall see in Chapter 9, the virial
coefficients may be computed in terms of the interparticle potential.
Comparison between experimental and theoretical values for the virial
coefficients is an important method for obtaining the force constants for
various interparticle potentials. In Fig. 2.1 we have plotted the second virial
coefficient for helium and argon. The curves are typical of most gases. At low
temperatures, B2(T) is negative because the kinetic energy is small and the
attractive forces between particles reduce the pressure. At high temperatures the
attractive forces have little effect and corrections to the pressure become
positive. At hig
maximum.
For an ideal
Converted with
zero, but for an

STOI Converter

ideal quantum

boefficients are

trial version

hDP://www.stdutilitv.com

......

.0

0 .5 ~
0

-0 .5 :-1.0 -1.5 -

-2.0 -2.5

tP~

/:.

r-

helium

A argon

-3.0 r-3.5 r-4.0

10

20

50

T*

100

Fig. 2.1. A plot of the second virial coefficients for helium and argon in terms of the
dimensionless quantities, B* = Bzlbo and T* = kBT/e, where bo and e are constants,
ks is Boltzmann's constant, and T is the temperature. For helium, bo = 21.07 x
1O-6m3/mol
and e/kB = 10.22 K. For argon, bo = 49.8 x 1O-6m3/mol
and
e/kB = 119.8 K. (Based on Ref. 2.)

18

INTRODUCTION TO THERMODYNAMICS

nonzero. The "statistics" of quantum particles give rise to corrections to the


classical ideal gas equation of state.

2.C.3. Van der Waals Equation of State [3]


The van der Waals equation of state is of immense importance historically
because it was the first equation of state which applies to both the gas and liquid
phases and exhibits a phase transition. It contains most of the important
qualitative features of the gas and liquid phases although it becomes less
accurate as density increases. The van der Waals equation contains corrections
to the ideal gas equation of state, which take into account the form of the
interaction between real particles. The interaction potential between molecules
in a gas contains a strong repulsive core and a weaker attractive region
surrounding the repulsive core. For an ideal gas, as the pressure is increased, the
volume of the system can decrease without limit. For a real gas this cannot
happen because the repulsive core limits the close-packed density to some finite
value. Therefore, as pressure is increased, the volume tends to some minimum
value, V = Vm'

The ideal gas


equation of st
Converted with
istence of the
repulsive core

STOU Converter
trial ve rsion

hUP:llwww.stdutililV.com

The attracive r
reased slightly
relative to that
a "cohesion"
between molecules. The decrease in pressure will be proportional to the
probability that two molecules interact; this, in tum, is proportional to the
square of the density of particles (N /V)2. We therefore correct the pressure by a
factor proportional to the square of the density, which we write a(n2/V2). The
constant a is an experimental constant which depends on the type of molecule
being considered. The equation of state can now be written

an2)
+ V2

(V - nb)

= nRT.

(2.12)

In Table 2.1 we have given values of a and b for simple gases.


The second virial coefficient for a van der Waals gas is easily found to be
(2.13)
We see that B~VW) (T) will be negative at low temperatures and will become
positive at high temperatures but does not exhibit the maximum observed in real
gases. Thus, the van der Waals equation does not predict all the observed

19

SOME MECHANICAL EQUATIONS OF STATE

Table 2.1. Van der Waals Constants for Some


Simple Gases [4]

H2
He
CO2
H2O
02

N2

a(Pa m6/moI2)

b(m3/mol)

0.02476
0.003456
0.3639
0.5535
0.1378
0.1408

0.02661
0.02370
0.04267
0.03049
0.03183
0.03913

features of real gases. However, it describes enough of them to make it a


worthwhile equation to study. In subsequent chapters, we will repeatedly use
the van der Waals equation to study the thermodynamic properties of interacting
fluids.

2.C.4.Solidr--------------------,
Solids have tl

(l/v)( 8v / aT)!

where v = V/j
fairly low tern
series about z
following equs

Converted with

STOI Converter
trial version

pansion,

ap

l/v) (8v/aph,
e, for solids at

plid in a Taylor
and obtain

the

hDP://www.stdutilitv.com
(2.14)

where T is measured in Kelvins. TYpical values [5] of /'i,T are of the order of
10-10 /Pa or 1O-5/atm. For example, for solid Ag (silver) at room temperature,
/'i,T = 1.3 X 10-10 /Pa (for P = 0 Pa), and for diamond
at room temperature,
/'i,T = 1.6 X 10-10 /Pa (for P = 4.0 X 108 Pa - 1010Pa). TYpical values of ap
are of the order 1O-4/K. For example, for solid Na (sodium) at room
temperature we have ap = 2 x 1O-4/K, and for solid K (potassium) we have
Qp = 2 x 10-4/K.

2.C.S.Elastic
Wireor Rod
For a stretched wire or rod in the elastic limit, Hook's law applies and we can
write
J = A(T)(L

- La),

(2.15)

where J is the tension measured in Newtons per meter, A(T) is temperature


dependent coefficient, L is the length of the stretched wire or rod, and Lo is the

20

INTRODUCTION TO THERMODYNAMICS

length of the wire when J = O. The coefficient, A(T), may be written


= AD + Al T, where AD and Al are constants. The constant, AI, is negative
for most substances but may be positive for some substances (including rubber).

A(T)

2.C.6. Surface Tension [6]


Pure liquids in equilibrium with their vapor phase have a well-defined surface
layer at the interface between the liquid and vapor phases. The mechanical
properties of the surface layer can be described by thermodynamic state
variables. The origin of the surface layer is the unequal distribution of
intermolecular forces acting on the molecules at the surface. Molecules in the
interior of the liquid are surrounded by, and interact with, molecules on all
sides. Molecules at the surface interact primarily with molecules in the liquid,
since the vapor phase (away from the critical point) is far less dense than the
liquid. As a result, there is a strong tendency for the molecules at the surface to
be pulled back into the liquid and for the surface of the liquid to contract. The
molecular forces involved are huge. Because of this tendency for the surface to
contract, work'
. uid. When the
surface area is
rought to the
surface and the
ar forces. The
work per unit
tension of the Ii
on the area and

STDI Converter

d the surface

es not depend

trial version

hDP://www.stdutilitv.com

(2.16)

where t is the temperature in degrees Celius, 0"0 is the surface tension at


t = OC,t' is experimentally determined temperature within a few degrees of
the critical temperature, and n is an experimental constant which has a value
between one and two.

2.C.7. Electric Polarization [6-8]


When an electric field E is applied to a dielectric material, the particles
composing the dielectric will be distorted and an electric polarization field, P (P
is the induced electric dipole moment per unit volume), will be set up by the
material. The polarization is related to the electric field, E, and the electric
displacement, D, by the equation
D = coE +P

(2.17)

where cO is the permittivity constant, co = 8.854 X 1O-I2C2 IN m2 The


electric field, E, has units of Newtons per coulomb (N/C), and the electric
displacement and electric polarization have units of coulomb per square meter

21

THE LAWS OF THERMODYNAMICS

(C/m2). E results from both external and surface charges. The magnitude of the

polarization field, P, will depend on the temperature. A typical equation of state


for a homogeneous dielectric is
(2.18)
for temperatures not too low. Here a and b are experimental constants and Tis
temperature in degrees Kelvin.

2.C.S. Curie's Law [6-8]


If we consider a paramagnetic solid at constant pressure, the volume changes

very little as a function of temperature. We can then specify the state in terms of
applied magnetic field and induced magnetization. When the external field is
applied, the spins line up to produce a magnetization M (magnetic moment per
unit volume). The magnetic induction field, B (measured in units of teslas, 1
T=1 weber/rrr'), the magnetic field strength, H (measured in units of ampherel
meter), and th

Converted with

where J-lo is the


of state for sue
law:

STOI Converter
trial version

(2.19)
). The equation
ated by Curie's

hDP://www.stdutilitv.com
(2.20)

where n is the number of moles, D is an experimental constant dependent on the


type of material used, and the temperature, T, is measured in Kelvins.

2.D. THE LAWS OF THERMODYNAMICS

[6]

Thermodynamics is based upon four laws. Before we can discuss these laws in a
meaningful way, it is helpful to introduce some basic concepts.
A system is in thermodynamic equilibrium if the mechanical variables do not
change in time and if there are no macroscopic flow processes present. Two
systems are separated by a fixed insulating wall (a wall that prevents transfer of
matter heat and mechanical work between the systems) if the thermodynamic
state variables of one can be changed arbitrarily without causing changes in the
thermodynamic state variables of the other. Two systems are separated by a
conducting wall if arbitrary changes in the state variables of one cause changes
in the state variables of the other. A conducting wall allows transfer of heat. An
insulating wall prevents transfer of heat.

22

INTRODUCTION TO THERMODYNAMICS

It is useful to distinguish

among three types of thermodynamic

systems. An

isolated system is one which is surrounded by an insulating wall, so that no heat


or matter can be exchanged with the surrounding medium. A closed system is
one which is surrounded by a conducting wall so that heat can be exchanged but
matter cannot. An open system is one which allows both heat and matter
exchange with the surrounding medium.
It is possible to change from one equilibrium state to another. Such changes
can occur reversibly or irreversibly. A reversible change is one for which the
system remains infinitesimally close to the thermodynamic equilibrium-that
is, is performed quasi-statically. Such changes can always be reversed and the
system brought back to its original thermodynamic state without causing any
changes in the thermodynamic
state of the universe. For each step of a
reversible process, the state variables have a well-defined meaning.
An irreversible or spontaneous change from one equilibrium state to another
is one in which the system does not stay infinitesimally close to equilibrium
during each step. Such changes often occur rapidly and give rise to flows and
"friction" effects. After an irreversible change the system cannot be brought
back to its o' .
...
change in the
thermodynami
discuss the fo

2.D.I. Zerot
Equilibrium
Equilibrium

Converted with

STOI Converter
trial version

d, we an now

amic
amic

hDP:llwww.stdutililV.com

The zeroth law IS 0 un amen a impo ance 0 expenmenta


ermodynamics
because it enables us to introduce the concept of a thermometer and to measure
temperatures of various systems in a reproducible manner. If we place a
thermometer in contact with a given reference system, such as water at the
triple point (where ice, water, and vapor coexist), then the mechanical variables
describing the thermodynamic
state of the thermometer (e.g., the height of a
mercury column, the resistance of a resistor, or the pressure of a fixed volume
container of gas) always take on the same values. If we then place the
thermometer in contact with a third system and the mechanical variables do not
change, then we say that the third system, the thermometer, and water at the
triple point all have the same "temperature."
Changes in the mechanical
variables of the thermometer as it is cooled or heated are used as a measure of
temperature change.

2.D.2. First Law: Energy Is Conserved


The first law tells us that there is a store of energy in the system, called the
internal energy, U, which can be changed by causing the system to do work,
f/w, or by adding heat, f/Q, to the system. We use the notation, f/W, to indicate
that the differential is not exact.) The change in the internal energy which

23

THE LAWS OF THERMODYNAMICS

results from these two processes is given by


(2.21)
The work, ,tW, may be due to changes in any relevant extensive "mechanical"
or chemical variable. In general it can be written
,tW

PdV - JdL - adA - E dP - H dM - de - LJ-LjdNj,

(2.22)

where dU,dV,dL,dA,dP,dM,de,
and dNj are exact differentials, but ,tQ
and ,tW are not because they depend on the path taken (on the way in
which heat is added or work is done). The meaning of the first five terms
in Eq. (2.22) was discussed in Section (2.C). The term, -de, is the work
the system needs to do it is has an electric potential, , and increases its
charge by an amount, de. The last term, -J-LjdNj, is the chemical work
required for the system to add dNj neutral particles if it has chemical
potential, J-LJ.We may think of - P, J, a, E, H, ' and J-L~
as generalized forces,
and we may
generalized
displacements.
Converted with
hich denotes
It is useful t
placement,
X,
quantities such
,M,
and
e,
which denotes
trial
version
respectively. T

STOI Converter
hDP://www.stdutilitv.com

(2.23)

Note that J-Ljis a chemical force and dNj is a chemical displacement. Note also
that the pressure, P, has a different sign from the other generalized forces. If we
increase the pressure, the volume increases, whereas if we increase the force, Y,
for all other cases, the extensive variable, X, decreases.

2.D.3. Second Law: Heat Flows Spontaneously from High


Temperatures to Low Temperatures
There are a number of ways to state the second law, with the one given above
being the simplest. Three alternative versions are [6]:
(a) The spontaneous tendency of a system to go toward thermodynamic

equilibrium cannot be reversed without at the same time changing some


organized energy, work, into disorganized energy, heat.
(b) In a cyclic process, it is not possible to convert heat from a hot reservoir
into work without at the same time transferring some heat to a colder
reservoir.

24

INTRODUCTION TO THERMODYNAMICS

(c) The entropy change of any system and its surroundings, considered

together, is positive and approaches zero for any process which


approaches reversibility.
The second law is of immense importance from many points of view. From it
we can compute the maximum possible efficiency of an engine which
transforms heat into work. It also enables us to introduce a new state variable,
the entropy, S, which is conjugate to the temperature. The entropy gives us a
measure of the degree of disorder in a sysem and also gives us a means for
determining the stability of equilibrium states, and, in general, it forms an
important link between reversible and irreversible processes.
The second law is most easily discussed in terms of an ideal heat engine first
introduced by Carnot. The construction of all heat engines is based on the
observation that if heat is allowed to flow from a high temperature to a lower
temperature, part of the heat can be turned into work. Carnot observed that
temperature differences can disappear spontaneously without producing work.
Therefore, he proposed a very simple heat engine consisting only of reversible
steps, thereby eli minatin
steful h
fl
.ne consists of

Converted with

the four steps s

STOI Converter

(a) Isotherm

reservoir
than an i
(b) Adiabati
lower va

Q12 from a
a finite rather
).

trial version

hDP://www.stdutilitv.com

isothermal (Th)
.6.Q12

adiabatic
4

-,

.6.Q43 isothermal (Tc)


X
Fig. 2.2. A Carnot engine which runs on a substance with state variables, X and y. The
processes 1 ~ 2 and 3 ~ 4 occur isothermally at temperatures Th and r-. respectively.
The processes 2 ~ 3 and 4 ~ 1 occur adiabatically. The heat absorbed is ~Q12 and the
heat ejected is ~Q43. The shaded area is equal to the work done during the cycle. The
whole process takes place reversibly.

25

THE LAWS OF THERMODYNAMICS

(c) Isothermal expulsion of heat ~Q43 into a reservoir at temperature Te (the


process 3 -+ 4).
(d) Adiabatic return to the initial state at temperature Th (the process 4 -+ 1).
The work done by the engine during one complete cycle can be found by
integrating the differential element of work Y dX about the entire cycle. We see
that the total work ~ Wtot done by the engine is given by the shaded area in Fig.
2.2.
The total efficiency 'fJ of the heat engine is given by the ratio of the work
done to heat absorbed:
(2.24)
Since the internal energy U is state variable and independent of path, the total
change ~Utot for one complete cycle must be zero. The first law then enables us
to write
77

..

.,...

..

YY7

ro.

(2.25)

Converted with

and thus
~
If we combine

STOI Converter
trial version

(2.26)
~y in the form

hDP://www.stdutilitv.com
(2.27)

'fJ=I-~
~Q12

A 100% efficient engine is one which converts all the heat it absorbs into work.
However, as we shall see, no such engine can exist in nature.
The great beauty and utility of the Camot engine lies in the fact that it is the
most efficient of all heat engines operating between two heat reservoirs, each at
a (different) fixed temperature. This is a consequence of the second law. To
prove this let us consider two heat engines, A and B (cf. Fig. 2.3), which run
between the same two reservoirs Th and Te. Let us assume that engine A is a heat
engine with irreversible elements and B is a reversible Camot engine. We will
adjust the mechanical variables X and Y so that during one cycle both engines
perform the same amount of work (note that XA and YA need not be the same
mechanical variables as XB and YB):
(2.28)
Let us now assume that engine A is more efficient than engine B:
'fJA

> 'fJB

(2.29)

26

INTRODUCTION TO THERMODYNAMICS

Fig. 2.3. Two heat engines, A and B, work together. Engine B acts as a heat pump while
engine A acts as a heat engine with irreversible elements. Engine A cannot have a greater
efficiency than engine B without violating the second law.

and thus
(2.30)
or

Converted with

STOI Converter

We can use the


refrigerator. Si
efficiency whet
trial version
produced by A
the low-temper
~------------------------------~
extracted from reservoir Tc and delivered to reservoir Th is

hDP:llwww.stdutililV.com

(2.31 )
pt engine as a
ave the same
he work, ~ W,
pmp heat from
The net heat

(2.32)
If engine A is more efficient than engine B, then the combined system has
caused heat to flow from low temperature to high temperature without any work
being expended by an outside source. This violates the second law and therefore
engine A cannot be more efficient than the Carnot engine. If we now assume
that both engines are Carnot engines, we can show, by similar arguments, that
they both must have the same efficiency. Thus, we reach the following
conclusion: No engine can be more efficient than a Carnot engine, and all
Carnot engines have the same efficiency.
From the above discussion, we see that the efficiency of a Carnot engine is
completely independent of the choice of mechanical variables X and Y and
therefore can only depend on the temperatures Th and Tc of the two reservoirs.
This enables us to define an absolute temperature scale. From Eq. (2.27) we see
that
(2.33)

27

THE LAWS OF THERMODYNAMICS

x
Fig. 2.4. Two Carnot engines running between three reservoirs with temperatures
Th > l' > Tc have the same overall efficiency as one Carnot engine running between
reservoirs with temperatures Th > rc.

where f( Th, Te is some function of temperatures


f(Th, Te) has a
between three

Th

Converted with

and r-. The function


ngines running
~te

STOI Converter

(2.34)

trial version

hDP://www.stdutilitv.com
~Q43

"J \

ell

(2.35)

and

~Q65
~Q12 =!(Th,Te),

(2.36)

so that

!(Th,Te) =!(Th,T')f(T',

Te).

(2.37)

Thus, !(Th, Te) = g(Th)g-l (Te) where g(T) is some function of temperature.
One of the first temperature scales proposed but not widely used is due to
W. Thomson (Lord Kelvin) and is called the Thomson scale [9]. It has the form

~Q43
~Q12

eT;
= eTh

(2.38)

The Thomson scale is defined so that a given unit of heat ~Q12 flowing between
temperatures TO --+ (TO - 1) always produces the same amount of work,
regardless of the value of TO.

28

INTRODUCTION TO THERMODYNAMICS

A more practical scale, the Kelvin scale, was also introduced by Thomson. It
is defined as
~Q43
~Q12

Te
-

(2.39)

Th

As we will see below, the Kelvin scale is identical to the temperature used in the
ideal gas equation of state and is the temperature measured by a gas
thermometer. For this reason, the Kelvin scale is the internationally accepted
temperature scale at the present time.
The units of degrees Kelvin are the same as degrees Celsius. The ice point of
water at atmospheric pressure is defined as 0 C, and the boiling point is defined
as 100C. The triple point of water is 0.01 C. To obtain a relation between
degrees Kelvin and degrees Celsius, we can measure pressure of a real dilute
gas as a function of temperature at fixed volume. It is found experimentally
that the pressure varies linearly with temperature and goes to zero at
te = -273.15 C. Thus, from the ideal gas law, we see that degrees Kelvin,
T, are related to degrees Celsius, te, by the equation

Converted with

(2.40)

STOI Converter

The triple poin


In Exercise
which uses an
ideal gas as
trial version
~gines can be
constructed us
bles are left as
problems). Reo------l:l_h_n_P_:------,WWW.-----------.;;:o.S_t-d-ut-i-li~tv-.C-O-m-------Jgin
have the
same efficiency.

II

Exercise 2.2. Compute the efficiency of a Camot cycle (shown in the


figure below) which uses a monatomic ideal gas as an operating substance.

Answer: The equation of state of an ideal gas is PV = nRT, where P = - y


is the pressure and V = X is the volume, T is the temperature in Kelvins, and
n is the number of moles. The internal energy is U = (3/2)nRT. The Camot

29

THE LAWS OF THERMODYNAMICS

cycle for an ideal gas is shown in the figure below. The part 1 -+ 2 is an
isothermal expansion of the system, and the path 3 -+ 4 is an isothermal
contraction. It is clear from the equation of state that the temperature, Ti; of
path 1 -+ 2 is higher than the temperature, Tc, of path 3 -+ 4. The path
2 -+ 3 is an adiabatic expansion of the system, and the path 4 -+ 1 is an
adiabatic contraction. We shall assume that n is constant during each cycle.
Let us first consider the isothermal paths. Since the temperature is
constant along these paths, dU = ~nRdT = O. Thus, along the path
1 -+ 2, ~Q = ~W = nRTh(dV IV). The heat absorbed along the path
1 -+ 2 is
~Q12 =

sr,

V2

The heat absorbed along the path 3

-+

f.Q34

Since V2 >
V3 > V4, ~~
Let us n
~Q = 0 =d
of state, W(
-+

1, respe

(1)

VI

4 is

= st;

In(~:).

(2)
~1

Converted with

-+

2. Since

diabatic path,

STOI Converter

bf the equation
egrate

trial version

T3/2V = cor

= sr, In (V2) .

dV

VI

ths 2

to find
3 and

-+

hDP:llwww.stdutililV.com
- T. V2/3
T.C3V2/3 -h2

and

T.C4V2/3 -hI
- T. V2/3

(3)

For the entire cycle, we can write ~Utot = ~Qtot - ~Wtot = O. Thus
~Wtot = ~Qtot = ~Q12 + ~Q34. The efficiency of the Carnot cycle is
'fJ = ~Wtot

= 1 + ~Q34 = 1 _ Tc In(V3/V4)

~Q12

~Q12

ThIn(V2/VI)

= 1 _ Tc

(4)

Th'

since from Eq. (3) we have

We can use the Carnot engine to define a new state variable called the entropy.
All Carnot engines have an efficiency
(2.41)

30

INTRODUCTION TO THERMODYNAMICS

Fig. 2.5. An arbitrary reversible heat engine is composed of many infinitesimal


Carnot engines. The area enclosed by the
curve is equal to the work done by the heat
engine.

(cf. Exercise 2.2) regardless of operating substance. Using Eq. (2.41), we can
write the following relation for a Carnot cycle:
(2.42)
(note the chang
case of an arbi
engine as beirn
Fig. 2.5). Thus,

Converted with

STOI Converter

eralized to the
nsider such an
not cycles (cf.

trial version

hDP:llwww.stdutililV.com

(2.43)

The quantity

(2.44)
is an exact differential and the quantity S, called the entropy, may be considered
a new state variable since the integral of dS about a closed path gives zero.
No heat engine can be more efficient than a Carnot engine. Thus, an engine
which runs between the same two reservoirs but contains spontaneous or
irreversible processes in some part of the cycle will have a lower efficiency, and
we can write
(2.45)
and
~Q12

t;

_ ~Q43

i.

< O.

(2.46)

For an arbitrary heat engine which contains an irreversible part, Eq. (2.46) gives

31

THE LAWS OF THERMODYNAMICS

the very important relation

v~<o.

(2.47)

For an irreversible process, ~Q/T can no longer be considered an exact


differential.
A system may evolve between two thermodynamic states either by a
reversible path or by a spontaneous, irreversible path. For any process,
reversible or irreversible, the entropy change depends only on the initial and
final thermodynamic states of the system, since the entropy is a state variable. If
the system evolves between the initial and final states via a reversible path, then
we can compute the entropy change along that path. However, if the system
evolves between the initial and final states via an irreversible path, then we must
construct a hypothetical reversible path between the initial and final states in
order to use the equations of thermodynamics to compute the entropy change
during the spontaneous process. For the irreversible path, the heat absorbed by
the system will be less than that along the reversible path [cf. Eqs. (2.45)(2.47)]. Therefore,
Q T<
Q T. This means that/or an irreversible
process,
tropy change.
Some of it com
Converted with
esult is usually
written in the

.hrrev~

STOI Converter
trial version

(2.48)

hDP://www.stdutilitv.com

where d.S den


ocesses. For a
reversible proc
nge is entirely
due to a flow of heat into or out of the system. For a spontaneous (irreversible))
process, diS > 0.
For an isolated system we have ~ Q = 0, and we obtain the important relation
,I

(2.49)
where the equality holds for a reversible process and the inequality holds for a
spontaneous or irreversible process. Since the equilibrium state is, by definition,
a state which is stable against spontaneous changes, Eq. (2.49) tells us that the
equilibrium state is a state of maximum entropy. As we shall see, this fact gives
an important criterion for determining the stability of the equilibrium state for
an isolated system.

2.D.4. Third Law: The Difference in Entropy Between States


Connected by a Reversible Process Goes to Zero in the
Limit T ~ 0 K [9-11]
The third law was first proposed by Nemst in 1906 on the basis of experimental
observations and is a consequence of quantum mechanics. Roughly speaking, a

32

INTRODUCTION TO THERMODYNAMICS

So

Fig.2.6. The fact that curves Y = 0 and Y = Y1 must approach the same point (the third
law) makes it impossible to reach absolute zero by a finite number of reversible steps.

system at zero temperature drops into its lowest quantum state and in this sense
becomes completel
ordered. If entro
can be thou ht of as a measure of
disorder, then a
Converted with
An altemati
quence of the
above statemen
ite number of
steps if a reve
ent is easily
demonstrated b
have plotted
trial version
the curves as a
= Y1 for an
tic salt with
arbitrary syste
Y = H.) We c
e two states,
adiabatically and isothermally. From Eqs. (2.5) and (2.6), we write

STDI Converter
hDP://www.stdutilitv.com

(2.50)

As we shall show in Section 2.H, thermal stability requires that (as/aT) y ~ O.


Equation (2.50) tells us that if T decreases as Y increases isentropic ally, then S
must decrease as Y decreases isothermally, as shown in Fig. 2.6. For the process
1 ~ 2 we change from state Y = Yl to state Y = 0 isothermally, thus squeezing
out heat, and the entropy decreases. For process 2 ~ 3, we increase Y
adiabatically from Y = 0 to Y = Yl and thus decrease the temperature. We can
repeat these processes as many times as we wish. However, as we approach
T = 0 K, we know by the third law that the two curves must approach the same
point and must therefore begin to approach each other, thus making it
impossible to reach T = 0 K in a finite number of steps.
,
Another consequence of the third law is that certain derivatives of the
entropy must approach zero as T ~ 0 K. Let us consider a process at T = 0 K
such that Y ~ Y + dY and X ~ X + dX. Then the change in entropy if Y, T, and
,I

33

FUNDAMENTAL EQUATION OF THERMODYNAMICS

N are chosen as independent variables is (assume dN = 0)


dS = (8S)
8Y

(2.51)

dY,
N,T=O

or if X, T, and N are chosen as independent we obtain


dS

= (;~)

(2.52)

_ dX.
N,T-O

Thus, if the states (Y, T = 0 K) and (Y + dY, T = 0 K) or the states


(X, T = 0 K) and (X + dX, T = 0 K) are connected by a reversible process,
we must have dS = 0 (third law) and therefore
8S)
( 8Y

_ 0

(2.53)

N,T=O-

and

(2.54)

Converted with
Equations (2.5

STOI Converter

nces.

trial version
2.E. FUND

hDP://www.stdutilitv.com

THERMOD~~~~~------------~

The entropy plays a central role in both equilibrium and nonequilibrium


thermodynamics. It can be thought of as a measure of the disorder in a system.
As we shall see in Chapter 7, entropy is obtained microscopically by state
counting. The entropy of an isolated system is proportional to the logarithm of
the number of states available to the system. Thus, for example, a quantum
system in a definite quantum state (pure state) has zero entropy. However, if the
same system has finite probability of being in any of a number of quantum
states, its entropy will be nonzero and may be quite large.
The entropy is an extensive, additive quantity. If a system is composed of a
number of independent subsystems, then the entropy of the whole system will
be the sum of the entropies of the subsystems. This additive property of the
entropy is expressed mathematically by the relation
S()"U, AX, {ANi})

= )"S(U,X, {Ni}).

(2.55)

That is, the entropy is a first-order homogeneous function of the extensive state
variables of the system. If we increase all the extensive state variables by a
factor )..,then the entropy must also increase by a factor )...

INTRODUCTION TO THERMODYNAMICS

34

Differential changes in the entropy are related to differential changes in the


extensive state variables through the combined first and second laws of
thermodynamics:
TdS ~

#Q

= dU - YdX -

LMjdNj.

(2.56)

The equality holds if changes in the thermodynamic state are reversible. The
inequality holds if they are spontaneous or irreversible. Equations (2.55) and
(2.56) now enable us to define the Fundamental Equation of thermodynamics.
Let us take the derivative of AS with respect to A:

(2.57)

However, from

Converted with

STOI Converter

(2.58)

trial version

hDP://www.stdutilitv.com

(2.59)

and

(as)

M~
=_2.

aN
J
u,x,

{Nih}

(2.60)

Equations (2.58)-(2.60) are called the thermal, mechanical, and chemical


equations of state, respectively. The mechanical equation of state, Eq. (2.59), is
the one most commonly seen and is the one which is described in Section 2.C.
If we now combine Eqs (2.57)-(2.60), we obtain
TS= U-Xy-

LJ.LjNj.

(2.61)

Equation (2.61) is called the Fundamental Equation of thermodynamics (it is


also known as Euler's equation) because it contains all possible thermodynamic
information about the thermodynamic system. If we take the differential of Eq.
(2.61) and subtract Eq. (2.56) (we will take the reversible case), we obtain

35

FUNDAMENTAL EQUATION OF THERMODYNAMICS

another important equation,

SdT

+ XdY + ENjdl-Lj = 0,

(2.62)

which is called the Gibbs-Duhem equation. The Gibbs-Duhem equation relates


differentials of intensive state variables.
For a monatomic system, the above equations simplify somewhat if we
work with densities. As a change of pace, let us work with molar densities. For
single component system the Fundamental Equation can be written TS = UYX - I-Lnand the combined first and second laws (for reversible processes) can
be written TdS = dU - YdX -I-Ldn. Let us now introduce the molar entropy,
s = S/ n, the molar density, x = X/ n, and the molar internal energy, u = U / n.
Then the Fundamental Equation becomes

(2.63)

Ts = u - Yx -I-L,
and the combi

Converted with

processes)

(2.64)

STOI Converter
Therefore, (as
tion is simply

trial version

hDP://www.stdutilitv.com

-Duhem

equa-

(2.65)

and therefore the chemical potential has the form, I-L= I-L(T, Y), and is a
function only of the intensive variables, Tand Y. Note also that s = -(81-L/8T)y
and x = -(81-L/8Y)r. In Exercise 2.3, we use these equations to write the
Fundamental Equation for an ideal monatomic gas.

I
I

EXERCISE
2.3. The entropy of n moles of a monatomic ideal gas is
S= (5/2)nR+nRln[(V/Vo)(no/n)(T/To)3/2],
where Vo,no, and To are
constants (this is called the Sackur-Tetrode equation). The mechanical
equation of state is PV = nRT. (a) Compute the internal energy. (b)
~ompute the chemical potential. (c) Write the Fundamental Equation for an
Ideal monatomic ideal gas and show that it is a first-order homogeneous
function of the extensive state variables.

I Answer: It is easiest to work in terms of densities. The molar entropy can be


:, written s = (5/2)R + Rln[(v/vo)(T
/TO)3/2](v = V [n is the molar volume),
and the mechanical equation of state is Pv = RT.

36

INTRODUCTION TO THERMODYNAMICS

(a) The combined first and second law gives du = Tds - Pdv. If we
further note that ds = (8s/8T)vdT + (8s/av)rdv, then

dU=T(!;),dT+ [T(:t-+V=~RdT,

(1)

since (8s/8T)v = (3R/2T) and (8s/av)r = R], Therefore, the


molar internal energy is u = ~RT + Uo where Uo is a constant, and
the total internal energy is U = nu = ~nRT + U. where Uo = nus.
(b) Let us rewrite the molar entropy in terms of pressure instead of molar
volume. From the mechanical equation of state, v = (RT / P) and
Va = (RTo/Po). Therefore, s = ~R + R In[(Po/P)(T /To)512]. From the
Gibbs-Duhem
equation, (81-L/8T)p = -s = -GR+
Rln[(Po/P)
(T /TO)5/2]) and (81-L/8P)r = v = RT /P. If we integrate these we
obtain the following expression for the molar chemical potential:
J-l = -RTln

Po (T)5/2,
P To

'

(2)

Converted with
(c) Let us
numbe

STOI Converter
trial version

~,volume, and

(3)

hDP://www.stdutilitv.com
Equation (3) is the Fundamental Equation for an ideal monatomic gas.
It clearly is a first-order homogeneous function of the extensive
variables.
It is interesting to note that this classical ideal gas does not obey the
third law of thermodynamics and cannot be used to describe systems
at very low temperatures. At very low temperatures we must include
quantum corrections to the ideal gas equation of state.

2.F. THERMODYNAMIC POTENTIALS [11]


In conservative mechanical systems, such as a spring or a mass raised in a
gravitational field, work can be stored in the form of potential energy and
subsequently retrieved. Under certain circumstances the same is true for
thermodynamic systems. We can store energy in a thermodynamic system by
doing work on it through a reversible process, and we can eventually retrieve
that energy in the form of work. The energy which is stored and retrievable in
the form of work is called the free energy. There are as many different forms of
free energy in a thermodynamic system as there are combinations of

37

THERMODYNAMIC POTENTIALS

constraints. In this section, we shall discuss the five most common ones:
internal energy, U; the enthalpy, H; the Helmholtz free energy, A; the Gibbs free
energy, G; and the grand potential, f2. These quantities playa role analogous to
that of the potential energy in a spring, and for that reason they are also called
the thermodynamic potentials.

2.F.l. Internal Energy


From Eq. (2.61) the fundamental
written

equation

for the internal

= ST + YX + L MjNj,

energy

can be

(2.66)

where T, Y, and Mj are considered to be functions of S, X, and {Nj} [cf. Eqs.


(2.58)-(2.60)]. From Eq. (2.56), the total differential of the internal energy can
be written

Converted with

(2.67)

STOI Converter
The equality hol
which are spont

trial version

s for changes

hDP://www.stdutilitv.com
(2.68)

y (au)
ax
=

(2.69)
S,{Nj},

and

(2.70)

We can use the fact that dU is an exact differential to find relations between
derivatives of the intensive variables, T, Y, and
From Eq. (2.4) we know, for
example, that

M;.

38

INTRODUCTION TO THERMODYNAMICS

From Eqs. (2.68), (2.69), and (2.71), we obtain


(2.72)

(i

+ 1) additional

relations like Eq. (2.72) exist and lead to the identities

(2.73)

(2.74)
and

Converted with

(2.75)

STDI Converter

Equations (2.
oretically and
experimentally
of change of
trial version
seemingly div
For a substa
simplify if
we work with
ergy. Then the
Fundamental Equation can be written u = Ts + Yx + u, where s is the molar
entropy and x is a molar density. The combined first and second laws (for
reversible processes) are du = Tds + Ydx. Therefore we obtain the identities
(8u/8s)x = T and (8u/8x)s = Y. Maxwell relations reduce to (8T /8x)s =

hDP://www.stdutilitv.comions

(8Y/8s)x'
The internal energy is a thermodynamic potential or free energy because for
processes carried out reversibly in an isolated, closed system at fixed X and
{Nj }, the change in internal energy is equal to the maximum amount of work
that can be done on or by the system. As a specific example, let us consider a
PVT system (cf. Fig. 2.7). We shall enclose a gas in an insulated box with fixed
total volume and divide it into two parts by a movable conducting wall. We can
do work on the gas or have the gas do work by attaching a mass in a
gravitational field to the partition via a pulley and insulated string. To do work
reversibly, we assume that the mass is composed of infinitesimal pieces which
can be added or removed one by one. If PIA + mg > P2A, then work is done on
the gas by the mass, and if PIA + mg < P2A, the gas does work on the mass.
The first law can be written
(2.76)

39

THERMODYNAMIC POTENTIALS

..

. .'

.'. . .. ........

Fig. 2.7. For a reversible process in a closed,


insulated box of fixed size ( ll.S = 0, ll.V = 0,
ll.Nj = 0), the work done in lifting the weight
will be equal to the change in the internal
energy, (ll.U)S,V,N = -ll. Wfree'

where jj.U is the change in total internal energy of the gas, jj.Q is the heat flow
through the walls, and jj. W can be divided into work done due to change in size
of the box, PdV, and work done by the gas in raising the weight, jj. Wfree:

(2.77)
For a reversible
Fig. 2.7, jj.Q

process,

jj.Q = f TdS. For the reversible process pictured in


..,
ns take place).

Converted with

Therefore,

STOI Converter

(2.78)

trial version

e stored in the
Under these
form of inte
conditions, intOrrIDT"-.;;7II1~T""""",..;;7r:J~.,.,.---,~.,.....-g-""""P'O~n;ro;o;r--o;;!'rn;n7"'iT';-------'
For a spontaneous process, work can only be done at constant S, V, and {N;}
if we allow heat to leak through the walls. The first and second laws for a
spontaneous process take the form
Thus, for a rev

hDP://www.stdutilitv.com

dU = su

< T dS -

PdV - toWered

L J l-';dNj,

(2.79)

where the integrals are taken over a reversible path between initial and final
states and not the actual spontaneous path. We can do work on the gas
spontaneously by allowing the mass to drop very fast. Then part of the work
goes into stirring up the gas. In order for the process to occur at constant
entropy, some heat must leak out since jj.Q < T dS = O. Thus, for a
spontaneous process

(~U)s

V , {N}j

< -~Wfree.

(2.80)

Not all work is changed to internal energy and is retrievable. Some is wasted in
stirring the gas. (Note that for this process the entropy of the universe has
increased since heat has been added to the surrounding medium.)

40

INTRODUCTION TO THERMODYNAMICS

For processes involving mechanical variables Y and X we can write Eqs.


(2.78) and (2.80) in the form
(2.81)
where ~

Wfree

is any work done by the system other than that required to change

X. For a reversible process at constant S, X, and {Nj}, work can be stored as

internal energy and can be recovered completely.


If a process takes place in which no work is done on or by the system, then
Eq. (2.81) becomes
(2.82)
and the internal energy either does not change (reversible process) or decreases
(spontaneous process). Since a system in equilibrium cannot change its state
spontaneously, we see that an equilibrium state at fixed S, X, and {Nj} is a state
of minimum internal energy.

2.F.2. Enth

Converted with

STOI Converter

The internal e
sses carried out
at constant X,
ish to study the
thermodynam
{Nj}. Then it is
trial version
more conveni
The enthal
isolated and
closed but me
ined by adding
to the internal energy an additional energy due to the mechanical coupling:

hDP://www.stdutilitv.comly
B=

U-XY=ST+

LJ.LjNj.

(2.83)

The addition of the term - XY has the effect of changing the independent
variables from (S,X,Nj) to (S, Y, Nj) and is called a Legendre transformation. If
we take the differential of Eq. (2.83) and combine it with Eq. (2.67), we obtain
dB -5: T dS - X dY

+L

J.LjdNj

(2.84)

and, therefore,
(2.85)

(2.86)

41

THERMODYNAMIC POTENTIALS

and

, (8H)
8N

fL=
J

(2.87)

.
S,Y,{N/i'j}

Since dH is an exact differential, we can use Eq. (2.4) to obtain a new set of
Maxwell relations:
(2.88)

(2.89)

= -(~)

(2.90)

Converted with

and

STOI Converter
trial version

(2.91 )

hDP://www.stdutilitv.com

which relate s
For a substance with a single type of molecule, Eqs. (2.84)-(2.91) become
particularly simple if we work with densities. Let h = H / n denote the molar
enthalpy. Then the fundamental equation for the molar enthalpy can be written
h = u - xY = sT + u, The exact differential of the molar enthalpy is dh =
Tds - xdY (for reversible processes), which yields the identities (8h/8s)y = T
and (8h/8Y)s =x. Maxwell's relations reduce to (8T/8Y)s = -(8x/8s)y.
In
Exercise 2.4, we compute the enthalpy for a monatomic ideal gas in terms of its
natural variables.

EXERCISE 2.4. Compute the enthalpy for n moles of a monatomic


ideal gas and express it in terms of its natural variables. The mechanical
, equation of state is PV = nRT and the entropy is S = ~nR + nRln[(V /Vo)
(no/n)(T /TO)3/2J.
Answer: Let us write the molar entrop in terms of temperature and
" pressure. It is s = ~R + Rln[(Po/P)(T
/To) /2]. Also note that when P = Po
and T = To, s = So = ~R. Now since dh = Tds + vdP we have
I,

8h)

as

=
p

T = To

(!_)
Po

2/5 e(s-so)/SO

(1)

42

INTRODUCTION TO THERMODYNAMICS

and

(8h)
8P

= + RT.
s

(2)

If we integrate, we find h=~RTo(PIPo)2/5e(s-so)/so


=~RT. In terms of
temperature, the enthalpy is h = ~RT. There is an easier way to obtain these
results. From Exercise 2.3, the molar internal energy is u = ~RT. The
fundamental equation for the molar enthalpy is h = u + vP, where v = V In
is the molar volume. Since v = RT I P, we obtain h = ~RT and H = ~nRT.
For a YXT system, the enthalpy is a thermodynamic potential for reversible
processes carried out at constant Y. The discussion for the enthalpy is
completely analogous to that for the internal energy except that now we allow
the extensive variable X to change and maintain the system at constant Y. We
then find

(2.92)

Converted with
where the eq
spontaneous

STOI Converter
trial version

ality holds for a


efore,

(2.93)

hDP://www.stdutilitv.com

and we conclUrn:!LIrnr.---mra7el~~fl!1'JroCES'Sllrz;'iumfTll1lr..':lf;,
7., and {Nj}, work
can be stored as enthalpy and can be recovered completely.

If a process takes place at constant S, Y, and {Nj} in which no work is done


on or by the system, then

(2.94)
Since the equilibrium state cannot change spontaneously, we find that the
state at fixed S, Y, and {Nj} is a state of minimum enthalpy.

equilibrium

2.F.3. Helmholtz Free Energy


For processes carried out at constant T, X, and {lVj}, the Helmholtz free energy
corresponds to a thermodynamic potential. The Helmholtz free energy, A, is
useful for systems which are closed and thermally coupled to the outside world
but are mechanically isolated (held at constant X). We obtain the Helmholtz free
energy from the internal energy by adding a term due to the thermal coupling:
A

U - ST

= YX + LJ.LjNj
j

(2.95)

THERMODYNAMIC

43

POTENTIALS

The addition of -ST is a Legendre transformation which changes the


independent variables from (S, X, {~}) to (T, X, {~} ). If we take the
differential of Eq. (2.95) and use Eq. (2.67), we find
dA ::; -S dT

+ Y dX + EI1jd~.

(2.96)

Therefore,
(2.97)

(2.98)

and
(2.99)

Converted with
Again, from Eq

STOI Converter
trial version

hDP://www.stdutilitv.com

( 8N8S)
J

T ,x,{Nlh}

= -

(8111)
8T

X,{l\j}

'

(2.100)

(2.101)

(2.102)
and
(2.103)
for the system.
We can write the corresponding equations in terms of densities. Let us
consider a monatomic substance and let a = A/n denote the molar Helmholtz
free energy. Then the fundamental equations for the molar Helmholtz free
energy is a = u - sT = xY + J-L. The combined first and second laws (for
reversible processes) can be written da = -sdT + Ydx so that (8a/8T)x = -s

44

INTRODUCTION TO THERMODYNAMICS

and (8a/8xh = Y. Maxwells relations reduce to (8s/8xh = -(8Y /8T)x' In


Exercise 2.5, we compute that Helmholtz free energy for a monatomic ideal gas
in terms of its natural variables .
EXERCISE 2.S. Compute the Helmholtz free energy for n moles of a
monatomic ideal gas and express it in terms of its natural variables. The
mechanical
equation of state is PV = nRT and the entropy is
S = ~nR + nRln[(V /Vo)(no/n)(T /To)3/2].
Answer: Since da = -sdT - Pdv we have
(8a)
8T

=
v

-s = -~R _
2

Rln[~ (!_)3/2]

(1)

Vo To

and

(~~) =-P=--.RT

(2)

Converted with

If we integ

3/2]

and A

-nRT - nR~

STOI Converter

For a YXT

trial version

namic potential

for reversible
the thermodynlj
can be written

hDP:llwww.stdutililV.com
...-----__ --.......-----__.oltz

For a change in
free energy

s - J SdT + J YdX

- f.Wfree

L J J.tjdN;,

(2.104)

where the inequality holds for spontaneous processes and the equality holds for
reversible processes (AWfree is defined in Section 2.66). For a process carried
out at fixed T, X, and {Nj}, we find
(AAh,X{l\j}

(-AWfree),

(2.105)

and we conclude that for a reversible process at constant T, X, and {Nj}, work
can be stored as Helmholtz free energy and can be recovered completely.
If no work is done for a process occurring at fixed T, X, and {Nj}, Eq. (2.105)
becomes
(AA)r

,X,

{N}j

O.

(2.106)

Thus, an equilibrium state at fixed T, X, and {Nj} is a state of minimum


Helmholtz free energy.

4S

THERMODYNAMIC POTENTIALS

2.F.4. Gibbs Free Energy


For processes carried out at constant Y, T, and {Nj}, the Gibbs free energy
corresponds to the thermodynamic potential. Such a process is coupled both
thermally and mechanically to the outside world. We obtain the Gibbs free
energy, G, from the internal energy by adding terms due to the thermal and
mechanical coupling,
G

= U - TS-XY = LI1;Nj.

(2.107)

In this way we change from independent variables (S, X, {Ni}) to variables (T, Y,
{Ni}). If we use the differential of Eq. (2.106) in Eq. (2.67), we obtain

dG :::;-S dT - XdY

+ L l1;dNj,

(2.108)

so that
(2.109)

Converted with

STOI Converter

(2.110)

trial version

and

hDP://www.stdutilitv.com

(2.111)

The Maxwell relations obtained from the Gibbs free energy are
(2.112)

(8S)
8N

T,Y,{Nlij}

(8N8X)
J

T,Y,{NIFj}

=-

(8111)
8T

=-

(8~)
8Y

Y,{l\}}

T,{l\j}

'

(2.113)

(2.114)

and
(2.115)
and again relate seemingly

diverse partial derivatives.

46

INTRODUCTION TO THERMODYNAMICS

As we found in earlier sections, we can write the corresponding equations


in terms of densities. We will consider a monomolecular substance and let
g = G/ n denote the molar Gibbs free energy. Then the fundamental equation
for the molar Gibbs free energy is g = u - sT - xY = J-L and the molar
Gibbs free energy is equal to the chemical potential (for a monomolelcular
substance). The combined first and second laws (for reversible processes) can
be written dg = -sdT - xdY so that (8g/8T)y = -s and (8g/8Yh
= -x.
Maxwells relations reduce to (8s/8Yh
= +(8x/8T)y.
For a monatomic
substance, the molar Gibbs free energy is equal to the chemical potential.
For a YXT system, the Gibbs free energy is a thermodynamic potential for
reversible processes carried out at constant T, Y, and {Hj}. For a change in the
thermodynamic state of the system, the change in Gibbs free energy can be
written

fiG:'O -

SdT -

JX

dY - fiWfree

4: J ,,;dA'l,

(2.116)

where the equ


spontaneous p
fixed T, Y, and

Converted with

STOI Converter
trial version

Thus,for a rev
Gibbs free energy an

ality holds for


r processes at

(2.117)

hDP:llwww.stdutililV.com

n be stored as
e recovere comp e e y.
For a process at fixed T, Y, and {Hj} for which no work is done, we
can

obtain
(2.118)
and we conclude that an equilibrium
minimum

state at fixed T, Y, and

{Nj} is a state of

Gibbs free energy .

EXERCISE 2.6. Consider a system which has the capacity to do work,


!-W = -Y dX + !-W'. Assume that processes take place spontaneously so
that dS = (l/T)!-Q + d.S, where d.S is a differential element of entropy due
to the spontaneity of the process. Given the fundamental equation for the
Gibbs free energy, G = U - XY - TS, show that -(dG)YT = !-W' + Td.S.
Therefore, at fixed Yand T, all the Gibbs free energy is available to do work
for reversible processes. However, for spontaneous processes, the amount of
work that can be done is diminished because part of the Gibbs free energy is
used to produce entropy. This result is the starting point of nonequilibrium
thermodynamics.

THERMODYNAMIC

47

POTENTIALS

Answer: From the fundamental equation for the Gibbs free energy, we
know that dG = dU - X dY - Y dX - T dS - S dT. Also we know that
dU = (lQ + Y dX - (lW', so we can write dG = (lQ - (lW' - X dYTdS - SdT. For fixed Yand Twe have (dG)y T = (lQ - (lW' - TdS. Now
remember that dS = (I/T)(lQ + diS. Then we find (dG)y T = -sw' - Td.S.
Note that the fundamental equation, G = U - XY - TS contains the starting
point of nonequilibrium thermodynamics.

For mixtures held at constant temperature, T, and pressure, P, the Gibbs free
energy is a first-order homogeneous function of the particle numbers or particle
mole numbers and this allows us to introduce a "partial" free energy, a
"partial" volume, a "partial" enthalpy, and a "partial" entropy for each type of
particle. For example, the chemical potential of a particle of type i, when
written as /-li = (8G/8ni)r ,pIn
.. }, is a partial molar Gibbs free energy. The total
,t; ~#J
Gibbs free energy can be wntten
N

(2.119)

Converted with
The partial mol
the total volum

STOI Converter

.)T,P,{nYi}, and

trial version

hDP://www.stdutilitv.com
The partial molar entropy for particle of type i is s, = (8S / 8ni)r
total entropy can be written

(2.120)

,P , {no~FJ.},

and the

(2.121)
Because the enthalpy is defined, H = G + TS, we can also define a partial molar
enthalpy, hi = (8H/8ni)r ,P , {no~#J.} = /-li + TSj. Then the total enthalpy can be
written H = L:~1 nihi. These quantities are very useful when we describe the
properties of mixtures in later chapters.

EXERCISE 2.7. Consider a fluid with electric potential, cp, containing 1/


. different kinds of particles. Changes in the internal energy can be written,
dU = (lQ - PdV + cpde +
/-ljdnj. Find the amount of Gibbs free energy
needed to bring dn, moles of charged particles of type i into the system at
I fixed temperature, pressure, and particle number, nj(i -=I i), in a reversible
i

L:;

48

INTRODUCTION TO THERMODYNAMICS

manner. Assume that particles of type i have a valence, Zi. Note that the
amount of charge in one mole of protons is called a Faraday, F.
Answer: The fundamental equation for the Gibbs free energy,
G = U + PV - TS, yields dG = dU + P dV + V dP - T dS - S dT. Therefore, dG = ~Q + dde +
J-Ljdnj + V dP - TdS - S dT. For a reversible
process, dS = (l/T)~Q,
and dG = dde +
J-Ljdnj + V dP - SdT. Now
note that the charge carried by dn, moles of particles of type, i, is
de = ZiF 'dn.. Thus, the change in the Gibbs free energy can be written

L.:;

L.:;

1/

dG

= de + VdP

- SdT

+ EJ-Ljdnj
j
1/

= +V dP - S dT + (ZiF

+ J-li)dni + EJ-ljdnj.

(1)

1=Ii

=I

For fixed P, T, and nj (j

i), the change in the Gibbs free energy is

{A~\

From Exerci
of charged p.

(2)

Converted with

STOI Converter
trial version

is called the

add dn, moles


~.The quantity

(3)

hDP://www.stdutilitv.com

2.F.S. Grand Potential


A thermodynamic potential which is extremely useful for the study of quantum
systems is the grand potential. It is a thermodynamic potential energy for
processes carried out in open systems where particle number can vary but T, X,
and {J-lj} are kept fixed.
The grand potential, 0, can be obtained from the internal energy by adding
terms due to thermal and chemical coupling of the system to the outside
world:

= U - TS -

E J-ljH] = XY.

(2.122)

The Legendre transformation in Eq. (2.122) changes the independent variables


from (S,X, {Nt}) to (T,X, {J-lj}). If we add the differential of Eq. (2.122) to
Eq. (2.67), we obtain
dO. 5: -SdT

+ Y dX

- L~dJ-Lj,
j

(2.123)

49

THERMODYNAMIC POTENTIALS

and thus
(2.124)

(2.125)

and
(2.126)

The Maxwell relations obtained from the grand potential are


(2.127)

Converted with

STOI Converter

(2.128)

trial version

hDP://www.stdutilitv.com

(2.129)

and

(2.130)

and are very useful in treating open systems. The grand potential is a
thermodynamic potential energy for a reversible process carried out at constant
T, X, and {It]}. For a change in the thermodynamic state of the system, the
change in, the grand potential can be written

~fl ::; -

j SdT + j Y dX - ~Wftee - L j Njd"j,

(2.131)

where the equality holds for reversible changes and the inequality holds for
spontaneous changes (~Wfree is defined in Section 2.F.1). For a process at fixed

so

INTRODUCTION TO THERMODYNAMICS

T, X, and {I-j}, we obtain


(2.132)

(AO)r,x'{Il~} ::; (-AWfree).


J

Thus,for a r~versible process at constant T, X, and {I-j}, work can be stored as


grand potential and can be recovered completely.
For a process at fixed T, X, and {I-j} for which no work is done, we obtain
(AO)T ,x,{Ilj}
and we find that an equilibrium
minimum grand potential.

0,

::;

(2.133)

state at fixed T, X, and {I-j} is a state of

2.G. RESPONSE FUNCTIONS


The response functions are the thermodynamic quantities most accessible to
experiment. They give us information about how a specific state variable
changes as other inde endent state variables are chan ed under controlled
conditions. As
measure of the
Converted with
size of fiuctua
nctions can be
divided into
capacities, (b)
mechanical re
ceptibility, and
(c) chemical
thermal and
trial version
mechanical re

STOI Converter
hDP://www.stdutilitv.com

The heat capacity, C, is a measure of the amount of heat needed to raise the
temperature of a system by a given amount. In general, it is defined as the
derivative, C = ((lQ/dT). When we measure the heat capacity, we try to fix all
independent variables except the temperature. Thus, there are as many different
heat capacities as there are combinations of independent variables, and they
each contain different information about the system. We shall derive the heat
capacity at constant X and {1\j}, CX,{Nj}, and we shall derive the heat capacity at
constant Y and {1\j}, CY,{N
We will derive these heat capacities in two
different ways, first from the first law and then from the definition of the
entropy.
To obtain an expression of CX,{N we shall assume that X, T, and {1\j} are
independent variables. Then the first law can be written
j

}.

j},

so = dU -

Y dX -

,,[(8U)

+ L.."
j

2; I-jd1\j= (~~)

8Nj

-1-'
T,X,{NiiJ}

'J dNj.

'}

. dT
X,{l\'J}

+ [,(~~)
T,{Nj}

- yJl dX
(2.134)

51

RESPONSE FUNCTIONS

For constant X and {~}, we have [,IQ]X,{Nj} = Cx,{Nj}dT


CX,{"'i}

= (~~)

and we find

(2.135)

X,{N

j}

for the heat capacity at constant X and {~}.


To obtain an expression for CY,{Nj}, we shall assume that Y, T, and {~} are
independent variables. Then we can write
8T
(8X)

dX=

Y,{Nj}

ar ;

(8X)
8Y

T,{Nj}

dY+

E(8X)8N
J

T,y,{Niij}

(2.136)

d~.

If we substitute the expression for dX into Eq. (2.134), we obtain

,IQ

= {CX,{Nj}

+ [(~~) T,{Nj} -Y]

(~;)

Y,{Nj}

}dT

[(~U~_--",---,----,-____',------"",-,dY,,-----

Converted with

+~{ STOI Converter

j}

-l1j}d~.
(2.137)

trial version
For constant Y

find

hDP://www.stdutilitv.com

CY,{Nj} = CX,{Nj}

+ [(~~)

-Y]

T,{N

(2.138)

(~;)

Y,{Nj}

j}

for the heat capacity at constant Yand {~}.


For n moles of a monatomic substance, these equations simplify. Let us write
them in terms of molar quantities. We can write the heat capacity in the form
CX,n = (8U/8T)X,n
= n(8u/8T)x' where u = Ufn is the molar internal energy
and x = X/ n is a molar density of the mechanical extensive variable. The molar
heat capacity is then Cx = (8u/8T)x
so that CX,n = nc., Similarly, let us note
that (8X/8Th,n
= n(8x/8T)y
and (8U /8Xh n = (8u/8xh.
Therefore, the
molar heat capacity at constant Yis Cy = Cx + r(8u/8x)T - Y] (8x/8Th.
It is useful to rederive expressions for CX,{Nj} and CY,{"'i} from the entropy.
Let us first assume that T, X, and {~} are independent. Then for a reversible
process, we obtain
(iQ=TdS=T

(~)
8T

X,{Nj}

dT+T

(~)
8X
-

dX+L:T
T,{N

j}

(~)
8Nj

d~.
T,X,{Niij}
(2.139)

52

INTRODUCTION TO THERMODYNAMICS

For a processes which occurs at constant X and {Nj}, Eq. (2.139) becomes
(2.140)
and therefore
(2.141)
The second term comes from Eq. (2.97).
Let us now assume that T, Y and {Nj} are independent. For a reversible
process, we obtain

If we combine

(iQ =TdS

Converted with

STOI Converter
trial version

+T

hDP://www.stdutilitv.com
(2.143)

If we now compare Eqs. (2.142) and (2.143), we find

(2.144)
The last term in Eq. (2.144) comes from Eq. (2.109).
We can obtain some additional useful identities from the above equations. If
we compare Eqs. (2.100), (2.138), and (2.144), we obtain the identity
(2.145)

53

RESPONSE FUNCTIONS

Therefore,

Y)

2
8
(

8T2

(8CX,{lVj})

X,{lVj}

=-

8X

(2.146)
r,{lVj} ,

where we have used Eqs. (2.4) and (2.145).


For a monatomic substance, it is fairly easy to show that the molar heat
capacity at constant mechanical molar density, x, is Cx = T(8s/8T)x =
-T(82a/8T2t,
and the molar heat capacity at constant Y is
Cy = T(8s/8T)y
= -T(82a/8T2)y.
We also obtain the useful identities
(8s/8xh = (I/T)[(8u/8xh
- Y] = -(8Y /8T)x
and
(82y /8T2])x=

-(I/T)

(8cx/8xh.

2.G.2. Mechanical Response Functions


There are three mechanical response functions which are commonly used. They
are the isothermal susceptibility,

Converted with
the adiabatic s

(2.147)

STOI Converter
trial version

hDP://www.stdutilitv.com

(2.148)

and the thermal expansivity,


(2.149)
Using the identities in Section 2.B, the thermal and mechanical response
functions can be shown to satisfy the identities
Xr,{Nj}(CY,{Nj}

CY,{lVj}(Xr,{lVj}

CX,{Nj})

XS,{Nj})

T(ay,{Nj})2,

T(aY,{lVj})2,

(2.150)
(2.151)

and
CY,{Nj}
CX,{lVj}

= Xr,{Nj}
XS,{Nj}

The derivation of these identities is left as a homework problem.

(2.152)

54

INTRODUCTION TO THERMODYNAMICS

For PVT systems, the mechanical response functions have special names.
Quantities closely related to the isothermal and adiabatic susceptibilities are the
isothermal compressibility,

(2.153)

and adiabatic compressibility,

(2.154)

respectively. The thermal expansivity


from above. It is

for a PVT is defined slightly differently

(2.155)

Converted with
For a mona
even simpler i
compressibiliti
respectively, w
ap = (l/v)(8v

STDI Converter
trial version

hDP://www.stdutilitv.com

~ctions become
and adiabatic
1'I/v)(8v/8P)s'
expansivity is

~----------------------------~

EXERCISE 2.S. Compute the molar heat capacities, Cv and Cp, the
compressibilities, KT and Ks, and the thermal expansivity, ap, for a
monatomic ideal gas. Start from the fact that the molar entropy of the gas is
s = ~R + Rln[(v/vo)(T /TO)3/2] (v = V /N is the molar volume), and the
mechanical equation of state is Pv = RT.
Answer:
(a) The

molar heat capacity, Cv: The molar entropy is s = ~R+


Rln[(v/vo)(T /TO)3/2]. Therefore (8s/8T)v = (3R/2T)
and Cv =
T(8s/8T)v = 3R/2.
(b) The molar heat capacity,
The molar entropy can be written
s = ~R + Rln[(Po/P)(T /To) /2]. Then (8s/8T)p = 5R/2T and Cp =
T(8s/8T)p = 5R/2.
(c) The isothermal compressibility, KT: From the mechanical equation of
state, we have v = (RT/P). Therefore, (8v/8P)T = -(viP)
and
KT = -(l/v)(fJv/8P)r
= (l/P).

7/

55

STABILITY OF THE EQUILIBRIUM STATE

(d) The adiabatic compressibility, "'s: We must first write the molar
volume as a function of sand P. From the expressions for the molar
entropy and mechanical equation of state given above we find
v = vo(Po/p)3/5 exp[(2s/5R) - 1]. Then (av/8P)s = -(3v/5P) and

"'s = -(I/v)(8v/8P)s

(3/5P).

(e) Thermal Expansivity, Qp: Using the mechanical equation of state, we


find Qp = (l/v)(av/8T)p
= (I/T).

2.H. STABILITY OF THE EQUILIBRIUM STATE


The entropy of an isolated equilibrium system (cf. Section 2.D.3) must be a
maximum. However, for a system with a finite number of particles in
thermodynamic equilibrium, the thermodynamic quantities describe the average
behaviour of the system. If there are a finite number of particles, then there can
be spontaneous fluctuations away from this average behaviour. However,
fluctuations must cause the entropy to decrease. If this were not so, the system
could spontane
higher entropy
because of spot
Converted with
uilibrium state,
this, by definiti
We can use
~ain conditions
for local equili
terns. We will
trial version
restrict ourselv
also apply to
general YXT s

STOI Converter
hDP://www.stdutilitv.com

2.H.l. Conditions for Local Equilibrium in a PVT System


Let us consider a mixture of I types of particles in an isolated box of volume,
VT, divided into two parts, A and B, by a conducting porous wall which is free
to move and through which particles can pass (cf. Fig. 2.8). With this type of
dividing wall there is a free exchange of heat, mechanical energy, and particles
between A and B. One can think of A and B as two different parts of a fluid (gas
or liquid), or perhaps as a solid (part A) in contact with its vapor (part B). We
shall assume that no chemical reactions occur. Since the box is closed and
isolated, the total internal energy, UT, is
(2.156)

where Ua is the internal energy of compartment

Q.

The total volume, VT, is


(2.157)

56

INTRODUCTION TO THERMODYNAMICS

Fig. 2.S. An isolated, closed box containing


fluid separated into two parts by a movable
porous membrane.

where Va is the volume of compartment


of type j is

The total number of particles, Nj,T,

Q.

L Nj,a,

~,T=

(2.158)

a=A,B

where ~,a is the total number of particles of type j in compartment


entropy, ST, is ~----------------------------~

Q.

Converted with

The total

(2.159)

STOI Converter

where Sa is the
Let us now
volume, and p~

trial version

in the energy,
ints

hDP://www.stdutilitv.com
= dVT = d~,T =

dUT

(2.160)

(assume that no chemical reactions occur) so that dUA = -dUB, d VA =


-dVB, and ~,A
= -~,B'
The entropy change due to these spontaneous
fluctuations can be written

dST=

c:

""'" [(aSa)
au
a=A,B
a

V,.,{J\},,.}

dUa+

(as)

1
+ ""'"
~
l=1

dN.

_a

aN.

j.a

(aSa)
aV

U,.,v,.,{Nkyl<j,,.}

l,a

dVa
U,.,{J\},,.}

] + ...

(2.161)

From Eqs. (2.58)-(2.60) and (2.160), we can write Eq. (2.161) in the form

dST =

lIP
- - -

TA

TB

b.,UA +

(-

TA

- -PB

TB

d VA -

') ~,A
L1(' _
-_,
TA
TB
/-Lj,A

/-Lj B

+ ... ,

}=1

(2.162)

57

STABILITY OF THE EQUILIBRIUM STATE

where To. and Po. are the temperature and pressure, respectively, of the material
in compartment a, and J-Lj,o. is the chemical potential of particles of type j in
compartment a.
For a system in equilibrium, the entropy is a maximum. Therefore, any
spontaneous changes must cause the entropy to decrease. However, ~UA' ~ VA,
and ~,A
can be positive or negative. Thus, in order to ensure that ~ST ::; 0,
we must have
(2.163)
(2.164)
and

J-Lj,A

J-Lj,B'

j = 1, ... ,I.

(2.165)

Equations (2.163)-(2.165) give the conditions for local equilibrium in a system


in which no chemical reactions occur. Thus, if the interface between A and B
can transmit h ec
then the two
Converted with
systems must h~
~ potential for
each type of par
t o note that if
the partition c ar
~ = and we
p osition so no
can have J-L~A =1=
trial
ve
rsion
particles can p as
p we can have
PA =1= PB and J-Lj,
~

STOI Converter

hDP://www.stdutilitv.com

2.H.2. Conditions for Local Stability in a PVT System [12, 13]


Stability of the equilibrium state places certain conditions on the sign of the
response functions. To see this, let us consider a closed isolated box with
volume VT, total entropy ST, total internal energy UT, and a total number of
particles Nj,T of type j, where j = 1, ... ,I. We shall assume that the box is
divided into M cells which can exchange thermal energy, mechanical energy,
and particles. We shall denote the equilibrium volume, entropy, internal energy,
and number of particles of type j for the a th cell by ~, ~, ~, and Nj,o.'
respectively. The equilibrium pressure, temperature, and chemical potentials of
the various cells are pO,
and
respectively (they must be the same for all
the cells).
Because there are a finite number of particles in the box, there will be
spontaneous fluctuations of the thermodynamic
variables of each cell about
their respective equilibrium values. These spontaneous fluctuations must be
such that VT, and UT, and Nj,T remain fixed. However, since the equilibrium
state is stable, fluctuations must cause ST to decrease. If it did not decrease, the
equilibrium state would be unstable and spontaneous fluctuations would cause
the system to move to a more stable equilibrium state of higher entropy.

ro,

lj,

58

INTRODUCTION TO THERMODYNAMICS

We shall assume that fluctuations about the equilibrium state are small and
expand the entropy of the nth cell in a Taylor expansion about its equilibrium
value

In Eq. (2.166), we define


n

'I.

~Va

Converted with

STOI Converter
trial version

U,{Nj}

~.

hDP://www.stdutilitv.com
A

similar

expression
holds
we have

(2.167)

~(asa/aVa)ua,{Nj,a}'

for

For

~(asa/

a~ ,a)u a,Va,{N ki-},a


. },

aSa
aN

( j,Q)

U.,V.,{N"",.}

a
== ( au

as
aN

(J
(J

a
av

L
i-I

V,{NJ}

su;

as
aN

U,V,{N,,,,)

U,V,{N",,)

as

~Va
U,{NJ}
0

aN,.,( aN

U,v,{N",,)

~i,a.

(2.168)

U,V,{N,,,,}

In Eqs. (2.166)-(2.168),

the superscripts 0 on partial derivatives indicate that


they are evaluated at equilibrium. The fluctuations ~Ua, ~ Va, and ~,a
are
defined as ~Ua = Ua - ~, ~ Va = Va -~,
and ~~,a = ~,a a and
denote the deviation of the quantities Ua, Va, and Nj,a from their absolute
equilibrium values.

NJ

59

STABILITY OF THE EQUILIBRIUM STATE

We can obtain the total entropy by adding the contributions to the entropy
from each of the M cells. Because the entropy can only decrease due to
fluctuations, the terms which are first order in the fluctuations, ~Ua, ~ Va, and
Mj,a must add to zero (this is just the condition for local equilibrium obtained
in Section 2.H.l). Therefore, the change in the entropy due to fluctuations must
have the form

(2.169)

Equation (2.169) can be written in simpler form if we make use of Eqs. (2.58)(2.60). We than find

. ] + ...

Converted with

},a

STOI Converter

(2.170)

trial version

or

hDP://www.stdutilitv.com

J-L;~.

In Eq. (2.171), we have used the relation T~S = ~U + P~V- Lj==1


Equation (2.171) gives the entropy change, due to spontaneous fluctuations, in a
completely general form. We can now expand ~ST in terms of any set of
independent variables we choose.
Let us choose T, P, and Nj as the independent variables. Then we can
write

8S)
ASa = ( 8T

sr; + (8S)
8P

P,{J\j}

T,{J\j}

~Pa

I (8S)O
+L
8N.
j=l}

~,a,
T,P,{Nkij}

(2.172)

AVa= (8V)
8T

P,{J\j}

~Ta+ (BV)
8P

T,{J\j}

~Pa+L

I
}=1

(BV)
B~

0
T,P,{Nk>/-j}

j.a
(2.173)

60

INTRODUCTION TO THERMODYNAMICS

and

~J.L;,a=

(88;
1/)

sr; +

P,{Nj}

(88;
1/)

~P

+ LI

T,{N;}

(8J.L~')
8N,

;==1

~N;,a.

T,P,{Nkyfi}

(2.174)
If we now substitute Eqs. (2.172)-(2.174) into Eq. (2.171) and use the Maxwell
relations (2.112)-(2.115), the entropy change becomes
1 M [(8S)O
~ST = - 2TL
aT

- (8V)
8P

(~Ta)

- 2 (8V)
8T

P,{N;}

a=1

(~Pa)

T,{Nj}

~Ta~Pa

P,{N;}

+LLI

(8J.Li)O
8N

M;,a~Nj,a ] + ....

'T,P,{Nkyfi}

;=1 j=1

(2.175)
If we make use of Eqs. (2.6), (2.8) and (2.100), we can write

8S
( aT

Converted with
(2.176)

STOI Converter

If we plug thi

trial version

(2.177)

where
(2.178)
Because the fluctuations ~Tw ~P a, and ~,a
are independent, the requirement that ~ST :::;0 for a stable equilibrium state leads to the requirement that
CV,{Nj}

=T

and

(!:)

:2: 0,

~T,{N;}

= -

V,{N;}

I
~~

~~
;=1

j=1

(aJ.L')
is:
aN'T,P,{Nkyfi}

V1

(av)
8P

~NM>O.

'

1-

:2: 0,
T,{Nj}

(2.179)

61

STABILITY OF THE EQUILIBRIUM STATE

Conditions (2.179) are a realization of Le Chateliers's famous principle: If a


system is in stable equilibrium, then any spontaneous change in its parameters
must bring about processes which tend to restore the system to equilibrium.
The first condition in Eq. (2.179), CV,{N
0, is a condition for thermal
stability. If a small excess of heat is added to a volume element of fluid, then the
temperature of the volume element must increase relative to its surroundings so
that some of the heat will flow out again. This requires that the heat capacity be
positive. If the heat capacity were negative, the temperature would decrease and
even more heat would flow in, thus leading to an instability. From Eqs. (2.150)
and (2.179) we can also show that
j

(2.180)
The second condition in Eq. (2.179), ~T,{Nj} ~ 0, is a condition for
mechanical stability. If a small volume element of fluid spontaneously
increases, the pressure of the fluid inside the fluid element must decrease
relative to its surroundings so that the larger pressure of the surroundings will
stop the growth of the volume element. This re uires that the compressibility be
positive. If the
Converted with
d increase and
the volume ele
stability. From

STOU Converter

Eqs. (2.151) a

(2.181)

trial ve rsion

hnp://www.stdutilitv.com

The third c
0, where ~i
and ~
are artrnJrarv-varr.crmJIIS-;-rs.---m~UlJlamlUlI--rururreTJrrrr;ur.rlability.
We can
write the condition for chemical stability in matrix form:

,
,

('
,

(M1,M2,

,M1)

J-L2,1
;

P{N
,

il' ==

ki-I

~i'l) (Ml)
.
J-L2,l

~2

..

J-Ll,l

where J-LJ~i
== (aJ-LJ~/aNjh
,
Eq. (2.115)], the matrix

...
...

J-Ll,2
J-L2,2

J-Ll,l

J-L~,l

J-Ll,2

~O,

(2.182)

~l

.l: Because of the Maxwell relation J-LJ~i


= J-L~J', [cf.
,

C'l

,
,

J-Ll,2

J-L2,1

J-L2,2

J-Ll,l

...
...

~i,l)
J-L2,l

(2.183)
J-Ll,2

J-Ll,l

is symmetric. In addition, in order to satisfy the condition for chemical stability,


the matrix ii' must be a positive definite matrix. A symmetric matrix is

62

INTRODUCTION TO THERMODYNAMICS

positive definite if J-Lii > 0 (i = 1, ... , /) and if every principal minor is positive
or zero.

EXERCISE 2.9. A mixture of particles, A and B, has a Gibbs free


energy of the form
G = nAJ-L~(P, T)

+ nBJ-L~(P,

T)

+ RTnA

In(xA)

+ RTnB

In(xB)

+ A nAnB
n

where n = nA + nB,XA = nA/n, and XB = nB/n (n indicates mole number),


J-L~ and J-L~ are functions only of P and T. Plot the region of thermodynamic
instability in the XA - T plane.
For chemical stability, the matrix

Answer:

J-LA,A J-LA,B)
(
J-LB,A J-LB,B

mustbesY'~~~~~~~mtl~Dti~~~~~Uh~an
(8J-LB/8nB)

P T

chemical pot

STOI Converter

J-LA=

trial version
A condition

(2)

hDP://www.stdutilitv.com
(3)

x1 -

XA + RT /2A
T = (2A/R)(XA
-

or

> O. For

> A/2R, this is always satisfied. A plot of

x1) is given below.

critical point

0.5

1.0 XA

The shaded region corresponds to x~ - XA + RT /2A < 0 and is thermodynamically unstable. The unshaded region is thermodynamically stable. For
T < 'A/2R, two values of XA satisfy the condition ~ - XA + RT /2'A > 0 for

63

STABILITY OF THE EQUILIBRIUM STATE

each value of T. These two values of XA lie outside and on either side of the
shaded region and are the mole fractions of two coexisting phases of the
binary mixture, one rich in A and the other rich in B. For T > >"/2R, only
one value of XA satisfies the condition ~ - XA + RT /2>.. > 0, so only one
phase of the substance exists. (As we shall see in Chapter 3, a
thermodynamically stable state may not be a state of thermodynamic
equilibrium. For thermodynamic equilibrium we have the additional
condition that the free energy be minimum or the entropy be maximum. A
thermodynamically stable state which is not an equilibrium state is
sometimes called a metastable state. It can exist in nature but eventually
will decay to an absolute equilibrium state.)

2.H.3. Implications of the Stability Requirements for


the Free Energies
The stability conditions place restrictions on the derivatives of the thermodynamic potentials. Before we show this, it is useful to introduce the concept of
concave and co
r_

1 A'

Converted with
(a) A functi
Fig. 2.9)
lies abox

df(x)/tb.

STOI Converter
trial version

for all x (cf.

(xI) andf(x2)
1-1

< X < X2.

If

nt always lies

below th
(b) A functiL______"..----,---r-hD_P:_"_WWW
__
.S_td_u_t_il_itv_._C_Om_------"tion
-f(x)
convex.

is

We can now consider the effect of the stability requirements on the Helmholtz
and Gibbs free energies.

Xl

X2

Fig. 2.9. The function f{x) is a convex function of x.

64

INTRODUCTION TO THERMODYNAMICS

From Eq. (2.97) and the stability condition, Eq. (2.180), we can write

2
(aaT2A)

V,{Nj}

=-

(as)
aT

V,{i\'i}

CV,W}
T
< 0,

=-

(2.184)

and from Eq. (2.98) and the stability condition, Eq. (2.181), we can write
(2.185)
The Helmholtz free energy is a concave function of temperature and a convex
function of volume.
From Eq. (2.109) and the stability condition, Eq. (2.180), we can write

a( 2G)
aT2

= _ (as)
P,{i\'i}

aT

= _ Cp,{N <

P,{i\'i}

j}

'

(2.186)

Converted with

STOI Converter

trial version

hDP://www.stdutilitv.com
----~~----~T

V -_(8G)
8P

I
I
T,{Nj}'

s __

(8G)

8T

I
I

P,{Nj}:

I
I

vo-~

So
I

Po

(a)

-Y.
To

(b)

Fig. 2.10. (a) A plot of the Gibbs free energy and its slope as a function of pressure. (b)
A plot of the Gibbs free energy and its slope as a function of temperature. Both plots are
done in a region which does not include a phase transition.

65

STABILITY OF THE EQUILIBRIUM STATE

and from Eq. (2.110) and the stability condition, Eq. (2.181), we can write

(88p2G)
2

(8V)
8P

T,{~}

(2.187)

< 0,

-V~T,{~}

T,{~}

Thus, the Gibbs free energy is a concave function of temperature and a concave
function of pressure.
It is interesting to sketch the free energy and its slope as function of pressure
and temperature. A sketch of the Gibbs free energy, for a range of pressure and
temperature for which no phase transition occurs, is given in Fig. 2.10. The case
of phase transitions is given in Chapter 3.
The form of the Gibbs and Helmholtz free energies for a magnetic system is
not so easy to obtain. However, Griffiths [15] has shown that for system of
uncharged particles with spin, G(T, H) is a concave function of T and Hand
A(T, M) is a concave function of T and convex function of M. In Fig. 2.11, we
sketch the Gibbs free energy and its slope as a function of T and H for a
paramagnetic system.

Converted with

STOI Converter
trial version

G)

hDP://www.stdutilitv.com

---__..;:lloo,t-----~

'

G(To Ho)

slope

=(aG)

all

H
T,N

aT
,- , ~,

I.

HN

---!~

G(To,Ho)

.___-__...__--~

S _(8G)

To

=_(aG)

aH

er

H,N

I
I

T,N

I
I

soy

(a)

(b)

Fig.2.11. (a) A plot of the Gibbs free energy and its slope as a function of applied field.
(b) A plot of the Gibbs free energy and its slope as a function of temperature. Both plots
are done in a region which does not include a phase transition.

66

INTRODUCTION TO THERMODYNAMICS

.... SPECIAL TOPICS


.... S2.A. Cooling and Liquefaction of Gases [6]
All neutral gases (if we exclude gravitational effects) interact via a potential
which has a hard core and outside the core a short-ranged attractive region. If
such a gas is allowed to expand, it must do work against the attractive forces
and its temperature will decrease. This effect can be used to cool a gas, although
the amount of cooling that occurs via this mechanism alone is very small. We
shall study two different methods for cooling: one based solely on free
expansion and one which involves throttling of the gas through a porous plug or
constriction. The second method is the basis for gas liquefiers commonly used
in the laboratory .

.... S2.A.l. The Joule Effect: Free Expansion


r=~==~~=---,-==",-"<-=~~"""-,,,,-,,,----,,-,...____...._.._...._,..xpansion
were
on by Joule in
Converted with
. 2.12. The gas

STOI Converter

is initially con
ressure Pi and
temperature 1',
an insulated
evacuated cha
trial version
' no work will
be done and Ll
ulated, no heat
will be added
must
remain constan ,
,
free expansion
is a transfer of energy between the potential energy and kinetic energy of the
particles.
Because free expansion takes place spontaneously, the entropy of the gas will
increase even though no heat is added. During the expansion we cannot use
thermodynamics to describe the state of the system because it will not be in
equilibrium, even locally. However, after the system has settled down and

hDP:llwww.stdutililV.comlenergy

Fig. 2.12. The Joule effect. Free expansion of


a gas from an insulated chamber of volume,
VI, into an insulated evacuated chamber of
volume, V2, causes cooling.

67

SPECIAL TOPICS: COOLING AND LIQUEFACTION OF GASES

reached a final equilibrium state, we can use the thermodynamics to relate the
initial and final states by finding an imaginary reversible path between them.
During the expansion the internal energy does not change and the particle
number does not change. Thus, the internal energies and particle numbers of the
initial and final states must be the same, and for our imaginary reversible path
we can write

[dU]n = 0 = (~~)

dT
V,n

+ (~~)

dV,
T,n

(2.188)

where n is the number of moles of gas. From Eq. (2.188), we can write

dT = - (~~)T,ndV = (aT)
dV
(~~)v,n
av U,n '

(2.189)

where we have made use of the chain rule, Eq. (2.6). The quantity tor] aV) U,n
is called the louie coefficient.
Let us comp r--------------~___,.q.
(2.145), we
know that
Converted with

STDI Converter
trial version

For an ideal ga

(aU !aV)Tn =
(aT !aV)U'n =

hDP://www.stdutilitv.com

(2.190)
we find that
coefficient,
change during

free expansion.
For a van der Waals gas [cf. Eq. (2.12)], we have

=
(au)
aV T,n V2

an

(2.191)

For this case, there will be a change in internal energy due to interactions as
volume changes if temperature is held fixed. To obtain the Joule coefficient, we
must also find the heat capacity of the van der Waals gas. From Eq. (2.146), we
obtain

acv,n)
( aV

p)

T,n

T(a

aT2 V,n

(2.192)

Therefore, for a van der Waals gas the heat capacity, CV,n, is independent of
volume and can depend only on mole number and temperature:

CV,n = Cv,n(T, n).

(2.193)

68

INTRODUCTION TO THERMODYNAMICS

Since the heat capacity, CV,n, contains no volume corrections due to


interactions, it is thermodynamically consistent to choose its value to be equal
to that of an ideal gas (at least in the regime where the van der Waals equation
describes a gas). If we neglect internal degrees of freedom, we can write
CV,n = ~nR. The Joule coefficient for a van der Waals gas then becomes

(8T)
8V

U,n

2 an
-3RV2

(2.194)

Using Eq. (2.194), we can integrate Eq. (2.189) between initial and final states.
We find
(2.195)
The fractional change in temperature is therefore

ll.T=
If Vf > Vi, the
der Waals con
for some si
n = 1 mol, a
have Sr ~ conclude that

(2.196)

Converted with

STOI Converter
trial version

hDP:llwww.stdutililV.com

lues of the van


in temperature
3
, T, = 300K,
or oxygen we
.097. We must
to cool a gas.

~ S2.A.2. The Joule-Kelvin Effect: Throttling


Throttling of a gas through a porous plug or a small constriction provides a
much more efficient means of cooling than free expansion and is the basis of
most liquification machines. The throttling process in its simplest form is
depicted in Fig. 2.13. A gas initially at a pressure, Pi, temperature, Ti, and
volume, Vi, is forced through a porous plug into another chamber, maintained at
pressure, Pf < Pi All chambers and the plug are insulated so LlQ = 0 for the
process. The gas inside the plug is forced through narrow twisting chambers
irreversibly. Work must be done to force the gas through the plug. Even though
the entire process is irreversible, we can use thermodynamics to relate the initial
and final states.
The net work done by the gas is
Vf

LlW

I PfdV + I PidV = PfVf


o

VI

r.v..

(2.197)

SPECIAL TOPICS: COOLING AND LIQUEFACTION OF GASES

69

porous
Fig. 2.13. The Joule-Kelvin effect. Throttling of a gas through a porous plug can cause
cooling or heating.

From the first law, dU

-dW since dQ

== O. Thus,
(2.198)

or, in terms of enthalpy,


(2.199)
Thus, the throttl
Let us now c
enthalpy proces
have (assuming

Converted with

STDO Converter

enthalpy.
~ the constant
ible path, we

trial version

hnp://www.stdutilitv.com

(2.200)

We see that the increase in entropy due to the throttling process is accompanied
by a decrease in pressure. It is convenient to use temperature and pressure as
independent variables rather than entropy and pressure. We therefore expand
the entropy

(2.201 )

and obtain

[dHl. = 0 = cp,.dT +

[v -

T(~~)JdP,

(2.202)

In Eq. (2.202) we have used Eqs. (2.144) and (2.112). Equation (2.202) can be
rewritten in the form

dT = (~~)

dP,
H,n

(2.203)

70

INTRODUCTION TO THERMODYNAMICS

where (aT/ ap) H,n is the louie-Kelvin

/-LJK

coefficient and is defined

-v].

(aT)
= _ (:)T~
= _1 [T(aV)
ap H,n
(aT) P,n CP,n
et P,n

(2.204)

Let us now compute the Joule-Kelvin coefficient for various gases. For an ideal
gas, (aV/aT)Pn = (V/T) and therefore the Joule-Kelvin
coefficient,
(aT / ap) H n equals O. There will be no temperature change during the
throttling process for an ideal gas. Furthermore, since T, = Tf for ideal gases,
Pf Vf = Pi Vi and no net work will be done (Ll W = 0). If work had been done on
or by the gas, we would expect a temperature change since the process is
adiabatic.
For a van der Waals gas, assuming that CV,n = ~nR, we find

[2a (v -v b)2 -b] / [~_


~ (v -v b) 2] ,
2 RTv

= _!_
(aT)
ap H,n R RT

(2.205)

where v = V /n-._.J-'Lib..e..._...m.ll!u:........l.U::llI.J.:m.a._____I,i.fiU..l;Uu::m....JL.L.:....ll..Dl...l..!~~l
obtain from the
Converted with
(2.100). For an

ghtforward to
), (2.144), and
Joule-Kelvin

~O;t~c~:n~~a:

w densities so

SIDU Converter
trial version

hDP://www.stdutilitv.com

(2.206)

For low temperatures (aT/ ap) H n > 0, and gases cool in the throttling process,
but at high temperatures, we have isr] ap) H,n < 0, and they heat up. Two
effect determine the behaviour of the Joule-Kelvin coefficient. On the one hand,
the gas expand, which gives rise to cooling. On the other hand, work can be
done on or by the gas. If Pi Vi > PfVf, then net work is done on the gas, which
causes heating. If P;Vi < Pf Vj, then net work is done by the gas, which causes
cooling.
The inversion temperature (the temperature at which the sign of /-LJK
changes) for the Joule-Kelvin coefficient will be a function of pressure. Since
CP,n > 0, the condition for inversion [from Eq. (2.204)] is
(2.207)
or, for a van der Waals gas [cf. Eq. (2.205)],

2a (v - b)2= b.

RT

(2.208)

71

SPECIAL TOPICS: COOLING AND LIQUEFACTION OF GASES

We can use the van der Waals equation, (2.12), to write Eq. (2.208) in terms of
pressure and temperature. First solve Eq. (2.208) for v as a function R, T, a, and
b, and substitute into the van der Waals equation. This gives

fidfi

_ 2
P-bV~-b--

3RT
a
2b - b2

(2.209)

The inversion curve predicted by the van der Waals equation has the shape
of a parabola with a maximum at T!! = 8a/9bR.For CO2, T!! = 911 K
while the experimental value [15] is T!! = 1500 K. For H2, T!! = 99 K while
the experimental value [15] is T!! = 202 K. In Fig. 2.14, we plot the van der
Waals and the experimental
inversion curves for N2 The van der Waals
equation predicts an inversion curve which lies below the experimental curve
but qualitatively has the correct shape. For nitrogen at P = 105 Pa, /-lJK =
1.37 X 10-7 K/Pa at T = 573 K, J.lJK = 1.27 X 10-6 K/Pa at T = 373 K,
/-lJK = 6.40 X 10-6 K/Pa
at T = 173 K, and /-lJK = 2.36 X 10-5 K/Pa
at
T = 93 K. (For experimental value of /-lJK for other substances, see the
International Cr .
r,...
-.u
can be quite
large for throttli
A schematic
Kel vin effect is
inversion temper
causing it to coc
circulate about t

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com

es

the JouleA, below its


essel, B, thus
is allowed to
progressively

P(Pax10')

61-------------_
5
4
3
2

23 123 223 323 423 523 623


Fig. 2.14. A plot of the inversion temperature versus pressure for the Joule-Kelvin
coefficient of N2 The solid line is the experimental curve [6]. The dashed line is the
curve predicted by the van der Waals equation for a = 0.1408 Pa. M6/mo12 and
b == 0.03913m3/mol.

72

INTRODUCTION TO THERMODYNAMICS

cooler

compressor

~ liquid
Fig. 2.15. Schematic drawing of a gas liquefier using the louIe-Kelvin effect.

cooler before
and eventual I)
At times tht
highly compn
spontaneously
Kelvin heating

Converted with

pn continuously

STOI Converter

es. For example,

trial version

ure, can ignite


ause of Joule-

hDP://www.stdutilitv.com
.... S2.B. Entropy of Mixing and the Gibbs Paradox [6]
If entropy is a measure of disorder in a system, then we expect that an entropy
increase will be associated with the mixing of distinguishable substances, since
this causes an increase in disorder. As we shall now show, this does indeed
occur even for ideal gases.
Let us consider an ideal gas containing nl moles of atoms of type AI, n2
moles of atoms of type A2, ... , and nm moles of atoms of type Am in a box of
volume V and held at temperature T. The equation of state for the ideal gas
mixture can be written
m

P-- L...J
"p.J'

(2.210)

j=l

where Pj is the pressure of the j th component (partial pressure) and is defined as


follows:
(2.211 )

73

SPECIAL TOPICS: ENTROPY OF MIXING AND THE GffiBS PARADOX

The mole fraction of atoms of type Aj is defined as follows:


p.

J
m
2:i:::l ni

X.1-

(2.212)

_}_

The change in the Gibbs free energy for an arbitrary change in the variables,
P, T, nl, ... ,nm, is
m

dG = -SdT

+ L/-Ljdnj,

V dP

(2.213)

j=l

where /-Lj is the chemical potential and S, V, and /-Lj are given by Eqs. (2.109)(2.111), but with mole number replacing particle number.
Let us now find the entropy of mixing for a mixture of distinguishable
monatomic ideal gases. The Gibbs free energy of n moles of a monatomic ideal
gas is
1"'.

_c

l"'l ...

(2.214 )

Converted with
where G(O) is a c
a box held at te
and let us assum

STOI Converter

first consider
ompartments,
s of atoms of

trial version

hDP://www.stdutilitv.com

type Al at pressi
s n2 moles of
atoms of type A2 r
bompartments
are separated by walls that can transmit heat and mechanical energy, so the
pressure and temperature is uniform throughout the system. The Gibbs free
energy of the system is the sum of the free energies of each compartment and
can be written
T

G/(P, T,nl,'"

,nm)

~LJnjRTln

[T5/2]
-

(0)

+ G1

(2.215)

j=l

where G~O) is a constant. If we now remove the partitions and let the gases mix
so that the final temperature and pressure are T and P, the Gibbs free energy of
the mixture will be

= G/(P,

T, nI,"

., nm)

+ L njRTln

(Xj)

+ G~)

G)O) ,

j=I

(2.216)

74

INTRODUCTION TO THERMODYNAMICS

where G~) is a constant and we have used the relation Pj = PXj. The change in
the Gibbs free energy during the mixing process is therefore
m

flG

= GF

G] =

L:njRTln

(Xj)

+ G~)

- G)O).

(2.217)

j=l

From Eqs. (2.109) and (2.217), the increase in entropy due to mixing is
m

flSmix = - L:njRln

(Xj).

(2.218)

j=l

If Xj = 1 (one compartment and one substance), then flSmix == 0, as it should be.


If there are two compartments, each containing one mole, then Xl = (1/2) and
X2

(1/2) and flSmix

= 2Rln(2)

and the entropy increases during mixing.

It is important to note that Eq. (2.218) contains no explicit reference to the

type of particles in the various compartments. As long as they are different,


mixing increases the entropy. However, if they are identical, Eq. (2.218) tells us
that there will
..
s are removed,
even though
q. (2.218) does
not work for i
nd is called the
Gibbsparado
as we shall se
particles must
have different
able particles
Gibbs paradox .

STOI Converter
trial version

hDP://www.stdutilitv.com

tum mechanics,
ter 7). Identical
particles (they
nd distinguishesolution of the

.....S2.e. Osmotic Pressure in Dilute Solutions


Each spring, when the weather begins to warm up, sap rises in the trees, and the
yearly cycle of life starts again. The rising of sap is one of many examples in
biological systems of the phenomenon called osmosis. We can easily
demonstrate the effect in the laboratory. Let us take a beaker of water and
immerse a long tube (open at both ends) in it. The water levels of the tube and
of the beaker will be the same. Next, close off the bottom end of the tube with a
membrane which is permeable to water but not sugar. The water levels will still
be the same in the tube and the beaker. Now add a bit of sugar to the water in
the tube. Water will begin to enter the tube through the membrane, and the level
of the sugar solution will rise a distance h above the level of the water in the
beaker (cf. Fig. 2.16). The excess pressure created in the tube, 7r = Pshg, is
called the osmotic pressure (Ps is the density of the sugar solution and g is the
acceleration of gravity). After equilibrium is reached, the pressure on the side of
the membrane with sugar solution will be greater than that on the water side, by
a factor 7r. The membrane must sustain the unbalanced force.

SPECIAL TOPICS: OSMOTIC PRESSURE IN

murre

t~..... -

----+
add sugar to tube

...---

75

SOLUTIONS

h :::':"f-"'"
I ';.'.
t
.....--""'4
.::.::

sugar solution

~.:::.

semipermeable membrane
Fig. 2.16. The osmotic pressure of the sugar solution is 7r = p.hg, where Ps is the
density of the sugar solution and g is the acceleration of gravity.

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com

Fig. 2.17. A schematic representation of osmosis.

It is instructive to show the same phenomenon in another way (cf. Fig. 2.17).
We will consider a system consisting of pure water, separated by a permeable
(to water) membrane from a solution of sugar and water. The entire system is
kept at a fixed temperature T, and the membrane is rigid. At equilibrium, there
will be an imbalance in the pressures of the two sides. If Po is the pressure of
the pure water, than the sugar solution will have a pressure P = Po + 1f', where
7r is the osmotic pressure. This imbalance of pressures is possible because the
membrane is rigid and cannot transmit mechanical energy. Since the water is
free to move through the membrane, the chemical potential of the pure water
must be equal to the chemical potential of the water in the sugar solution.
Let us first write the thermodynamic relations for this system. First consider
the sugar solution. A differential change in the Gibbs free energy, G =

76

INTRODUCTION TO THERMODYNAMICS

G(P, T, nw, ns), of the sugar solution (with nw moles of water and n; moles of
sugar solution) can be written
(2.219)

where S = - (8G / 8T) P,n""ns is the entropy of the solution, V = (8G / 8P)T,n""ns
is the volume of the solution, and I-lw = (8G/8nw)p,T,ns and I-ls = (8G/8ns)p,T,n",
are the chemical potentials of the water and sugar, respectively, in the solution.
The chemical potentials are intensive and depend only on ratios ns/nw. It is
convenient to introduce mole fractions
(2.220)

Since Xw + Xs = 1, the chemical potentials can be written as a function of mole


fraction, Xs' Thus, I-lw = I-lw(P, T, xs) and I-ls = I-ls(P, T, xs).
At equilibrium the chemical potentials of the pure water and the water in the
sugar solution will
(0)
mical potential

Converted with

of the pure w

STOI Converter

(2.221 )

as the conditi
trial version
We want t
re in terms of
measurable qu
as much as
possible. We WI
a ns / nw 1 and
x, ~ ns/nw 1. We can construct a fairly simple model to describe the
solution. We write the Gibbs free energy of the solution in the form

hDP:llwww.stdutililV.COmion

G ( P, T, ns, nw ) -_ nwl-lw(0) ()P, T

+ nwRTln

+ nsl-ls(0) ( P,
(xw) + nsRTln

T) -

nsnw
n

A--

(2.222)

(xs).

The chemical potential I-l~) (1-l~0 contains contributions to the Gibbs free
energy due to the presence of water (sugar) molecules and due to selfinteractions. The term -A(nsnw/n)
gives the contribution to the free energy due
to interactions between sugar and water molecules. The last two terms on the
right give contributions to the free energy due to mixing [cf. Eq. 3.57]. The
chemical potential of the water in the solution can now be written
I-lw(P, T,xs)

= (:~)

e.r,

= J-l~)(P, T) -

.xx; + RTln

(1 - xs),

(2.223)

where J-l~) (P, T) is the chemical potential of pure water at pressure P and
temperature T. For a dilute solution, Xs = ns/n -e t and In(l - xs) =

77

SPECIAL TOPICS: OSMOTIC PRESSURE IN DILUTE SOLUTIONS


- Xs -

(1/2).x; - . . . . Thus, to lowest order in

Xs

ns / n,

we find
(2.224)

for the chemical potential of water in a dilute sugar solution.


We now can find an expression for the osmotic pressure, 7r = P - Po. Let us
note that water, as well as most liquids, is very incompressible. The
compressibility, liT, of water at OC is liT = 4.58 X 10-11 cm2/dyne. Therefore
the quantity (al-l~/ap)T n = (aVO/anw)r p == v~ (v~ is the partial molar
volume of water in the abs~nce of solute and VO is the volume of water in the
absence of solute) remains approximately constant for small changes in
pressure. With this observation we can integrate (aJ.L~/ap)r,n to find
w

1-l~(P, T) -1-l~(Po,

T) ~ v~(P - Po) =

V~7r.

(2.225)

Let us now assume that the change in the volume of water as we increase the
number of moles is proportional to the number of moles so that VO = nwv~.
Also, for very small concentrations of solute, we can assume that the change in
the volume of ,
.gible so that
yo ~ V, where
Converted with
bombine Eqs.
(2.221), (2.224),

STOI Converter
trial version

(2.226)

hDP://www.stdutilitv.com

Equation (2.226
ks very much
like the ideal gas law, although we are by no means dealing with a mixture of
ideal gases. Equation (2.226) is well verified for all dilute neutral solvent-solute
systems.
I EXERCISE 2.10. An experiment is performed in which the osmotic
pressure of a solution, containing nsuc moles of sucrose (C12H22011)and 1 kg
of water (H20), is found to have the following values [2]: (a) for
nsuc = 0.1, 7r = 2.53 X 105 Pa, (b) for nsuc = 0.2, 7r = 5.17 X 105 Pa, and (c)
for nsuc = 0.3, 7r = 7.81 X 105 Pa Compute the osmotic pressure of this
" system using van't Hoff's law. How do the computed values compare with
the measured values?
!
!

: Answer: The molecular weight of water (H20) is MH20 = 18 gr/mol.


: Therefore, 1kg of water contains 55.56 mol of water. The molar volume of
water is VH20 = 18 X 10-6 m3/mo!. The osmotic pressure of the solution,
, according to van't Hoff's law, is
7r

nsuc (8.317 J IK ) (303 K)


55.56 18 x 10-6 m3/mol .

78

INTRODUCTION TO THERMODYNAMICS

The computed values are as follows: (a) For nsuc = 0.1, 7r = 2.52 X 105 Pa,
(b) for nsuc = 0.2, 7r = 5.04 X 105 Pa, and (c) for nsuc = 0.3, 7r = 7.56x
105 Pa. The predictions of van't Hoff's law are good for a dilute solution of
sucrose in water, but begin to deviate as the mole fraction of sucrose
increases .

.....S2.D. The Thermodynamics


[16-18]

of Chemical Reactions

Chemical reactions occur in systems containing several species of molecules


(which we will call A, B, C, and D), which can transform into one another
through inelastic collisions. A typical case might be one in which molecules A
and B can collide inelastically to form molecules C and D. Conversely,
molecules C and D can collide inelastically to form molecules A and B. The
collisions occur at random and can be either elastic or inelastic. To be inelastic
and result in a reaction, the two molecules must have sufficient energy to
overcome any
exist, Chemical
equilibrium is
Converted with
hen the rate of

STOI Converter

production of
pletion through
chemical reac
r stop, even at
equilibrium.
In the early
trial version
er, found that it
was possible t
gle variable ~,
called the degr
np:
~etermine when
the Gibbs free energy has reached its minimum value (chemical reactions
usually take place in systems with fixed temperature and pressure) and therefore
when the chemical system reaches chemical equilibrium. It is important to
notice that the concept of degree of reaction assumes that we can generalize the
concept of Gibbs free energy to systems out of equilibrium .

II

d -1WWW.stutlltv.com

.....S2.D.l. The Affinity


Let us consider a chemical reaction of the form
(2.227)
The quantities VA, VB, vc. and VD are called stoichiometric coefficients; Vj is the
number of molecules of type j needed for the reaction to take place. By
convention, VA and VB are negative. The constant kl is the rate constant for the
forward reaction, and k2 is the rate constant for the backward reaction. The rate
constants give the probability per unit time that a chemical reaction takes place.
If the rate constants are known, then one can find the rate of change in the

SPECIAL TOPICS: THE THERMODYNAMICS OF CHEMICAL REACTIONS

79

number of each type of molecule involved in the chemical reaction. For


example, the rate of change in the number, NA, of molecules of type A is given
by
(2.228)
where IVAI denotes the absolute value of VA. Equation (2.228) reflects the fact
that IVAI molecules of A and IVBI molecules of B must collide to destroy A
molecules while I vc I molecules of C and IVD I molecules of D must collide to
create A molecules.
Let us now assume that initially there are nA = -VAnO moles of A,
ns = -vBnO + NB moles of B, nc = vcn~ moles of C, and no = vDn~ + ND
moles of D. The reaction to the right will be complete when

= vc(no + n~),

nc

We next define the degree of reaction by the equation

Converted with
As we have defi
moles of each s

STOI Converter

(2.229)
he number of

trial version

hDP://www.stdutilitv.com

(2.230)
(2.231)
(2.232)

and
(2.233)
Any changes in the concentrations

due to the reaction can therefore be written


(2.234)

or
dnA
vA

Equations

= dnB = dnc = dnD


VB

vc

VD

= d~.

(2.235)

(2.234) and (2.235) are very important because they tell us that any

80

INTRODUCTION TO THERMODYNAMICS

changes in the thermodynamic properties of a system due to a given reaction


can be characterized by a single variable.
From Eq. (2.108), differential changes in the Gibbs free energy may be
written
m

dG = -S dT

+ V dP + L /-ljdnj = -S dT + V dP + L /-ljVjde,
j=1

(2.236)

j=l

where the sum is over the species which participate in the reaction. Therefore
BG)
( Be

(2.237)

= ?:/-ljVj,

P,T

J=1

and the quantity


m

(2.238)

== L/-ljVj
j=1

is called the a
At chemical e

opposite sign).
mum,

Converted with

STOI Converter

(2.239)

trial version

hDP://www.stdutilitv.com

(the superscrip
cal equilibrium
the affinity mu
We can easily find the sign of the affinity as the system moves toward
chemical equilibrium from the left or right. At constant P and T, the Gibbs free
energy, G, must always decrease as the system moves toward chemical
equilibrium (at equilibrium G is a minimum). Therefore,
[dG]P,T = (~~)

de <
P,T

o.

(2.240)

If the reaction goes to the right, then de > 0 and A < O. If the reaction goes
to the left, then de < 0 and A > O. This decrease in the Gibbs free energy is
due to spontaneous entropy production resulting from the chemical reactions
(see Exercise 2.6).
If there are r chemical reactions in the system involving species, j, then there
will be r parameters,
needed to describe the rate of change of the number of
moles, nj:

eb

dnj =

L Vjkdek.
k=]

(2.241 )

SPECIAL TOPICS: THE THERMODYNAMICS OF CHEMICAL REACTIONS

81

Table 2.2. Values of the Chemical Potential, pO, for Some Molecules in the Gas
Phase at Pressure Po = 1 atm and Temperature To = 298 K
J..L0

Molecule

(kcal/mol)
0.00
0.31
4.63
0.00
12.39
-3.98
23.49

H2
HI
12
N2
N02
NH3
N204

The sum over ~


participate.
Using ideal ~
the gas phase. C
B, C, and D) wh
the ith constitue
written

I'i(Pi, T)

Converted with

ules of type j

STOI Converter
trial version

hDP://www.stdutilitv.com
= I'?(Po,

To) _ RTin [ (~)

5/2 (~:)

pr reactions in
molecules (A,
~alpressure of
ituent can be

(2.242)

] ,

where f..L?(Po, To) is the chemical potential of the ith constituent at pressure Po
and temperature To. Values of f..L?, with Po = 1 atm and To = 298 K, have been
tabulated for many kinds of molecules [20]. A selection is given in Table 2.2. If
we use Eq. (2.242), the Gibbs free energy can be written

G(T,p,e)

= ~nil'i

= ~nil'?(Po,

= ~nil'?(po,To)
+ RT In[x'!AABC~B

To) _ ~niRTJn

- ~niRTln[
xnc x':.D]
D

[ (~)

5/2 (~:)]

(~t(~)]
(2.243)

82

INTRODUCTION TO THERMODYNAMICS

and the affinity can be written

A(T,p,e)

= ~ViM?(PO'

To) _ ~ViRTln[

= ~ ViM?(Po, To) _ ~

+ RTln

(~)

viRTln

[ (~)

5/2 (~~)

5/2 (~)

(2.244)

[~~~X~I]'
XA

XB

where P = 2:i Pi is the pressure and T is the temperature at which the reaction
occurs.
For "ideal gas reactions" the equilibrium concentrations of the reactants can
be deduced from the condition that at equilibrium the affinity is zero, A 0 = o.
From Eq. (2.243) this gives the equilibrium condition

..
In

[ X"CX"D
C D
B

where Po = 1
action. As we ~
the degree of
equilibrium oc

.,
To),

Converted with

IliAIX Ivs I

XA

"','"

STOI Converter
trial version

(2.245)

~e law of mass
ute the value of
~hich chemical

hDP://www.stdutilitv.com

.... S2.D.2. Stability


Given the fact that the Gibbs free energy for fixed P and T is minimum at
equilibrium, we can deduce a number of interesting general properties of
chemical reactions. First, let us note that at equilibrium we have
(2.246)

and

(BBeG)
2

0
P,T

(BA)
= Be

0
P,T>

o.

(2.247)

Equations (2.246) and (2.247) are statements of the fact that the Gibbs free
energy, considered as a function of P, T, and is minimum at equilibrium for
fixed T and P.

e,

SPECIAL TOPICS: THE THERMODYNAMICS OF CHEMICAL REACTIONS

83

From the fundamental equation, H = G + TS, we obtain several important


relations. First, let us note that at equilibrium

8e

(8H)

8e

(8S)

P,T

(2.248)

P,T

[we have used Eq. (2.246)]. Thus, changes in enthalpy are proportional to the
changes in entropy. The left-hand side of Eq. (2.248) is called the heat of
reaction. It is the heat absorbed per unit reaction in the neighborhood of
equilibrium. For an exothermic reaction, (8H / 8e)~ T is negative. For an
endothermic reaction, (8H / 8e)~ T is positive. From Eq. '(2.109), Eq. (2.248) can
be written
'

(~7):,T

-r[; (:~)p,~L -r[:r (~~)p,TL= -r(:~):~.


=

(2.249)
For an "idea]
explicit express

Converted with

~) to obtain an

STOI Converter
trial version

hDP://www.stdutilitv.com
If the total number of particles changes during the reaction (~:::::iVi =1= 0), there
will be contributions to the heat of reaction from two sources: (1) There will be
a change in the heat capacity of the gas due to the change in particle number,
and (2) there will be a change in the entropy due to the change in the mixture of
the particles. If the total number of particles remains unchanged (L:i Vi = 0),
the only contribution to the heat of reaction will come from the change in the
mixture of particles (assuming we neglect changes to the heat capacity due to
changes in the internal structure of the molecules).
Let us now obtain some other general properties of chemical reactions. From
the chain rule [Eq. (2.6)] we can write

e)

8
( 8T

=_

P,A

(8A)
tfi
(8A)
8~

P,~

P,T

1
T

(8H)
8~
(8A)
8~

P,T

(2.251)

P,T

The denominator in Eq. (2.251) is always positive. Thus, at equilibrium any


small increase in temperature causes the reaction to shift in a direction in which
heat is absorbed.

84

INTRODUCTION TO THERMODYNAMICS

Let us next note the Maxwell relation


(2.252)
[cf. Eqs. (2.236) and (2.238)]. It enables us to write

(Be)
BP

=_

T,A

(BA)
7fP
(BA)
8~

T,~

=_

P,T

(BV)
(BA)
8~

P,T

8~

P,T

(2.253)

At equilibrium an increase in pressure at fixed temperature will cause the


reaction to shift in a direction which decreases the total volume .

EXERCISE 2.11. Consider the reaction

which occurs

Converted with

STOI Converter

N204 and no

N02 Assum
essure P. Use
ideal gas eq
and plot the
Gibbs free en
trial version
tion, for (i)
P = 1 atm a
(b) Compute
and plot the a
action, for
(i) P = 1 atm an
an 11
= an
c) What is the
degree of reaction, at chemical equilibrium for P = 1 atm and temperature
T = 298 K? How many moles of N204 and N02 are present at equilibrium?
(d) If initially the volume is Va, what is the volume at equilibrium for P = 1
atm and T = 298 K? (e) What is the heat of reaction for P = 1 atm and
T = 298K?

e,

e,

hDP:llwww.stdutililV.com

e,

Answer: The number of moles can be written


The mole fractions are

nN204

= 1-

e and

nN~

2e.
(1)

(a) The Gibbs free energy is

(2)

85

SPECIAL TOPICS: THE THERMODYNAMICS OF CHEMICAL REACTIONS

where i =(N204, N02). From Table 2.2, for Po = 1 atm and


To = 298 K, J.L~204 = 23.49 kcallmol and J.L~02 = 12.39 kcallmol.
Plots of G(T, P) are given below. Chemical equilibrium occurs for
the value of at the minimum of the curve.
A(kcal )

mol

P=Po

Converted with

(b) The af

STOI Converter

A(

(3)

trial version

hDP://www.stdutilitv.com
Plots of A(T, P) are given in the figures.
(c) Chemical equilibrium occurs for the value of

at which A = O. From
the plot of the affinity, the equilibrium value of the degree of reaction
is
0.166. Thus, at equilibrium nN204 = 0.834 and nN02 = 0.332.
At equilibrium the mole fractions are XN204 = (0.834/1.166) = 0.715
and XN02 = (0.332/1.166) = 0.285.

eeq ~

(d) Initially there are

nN204 = 1 mol of N204 and XN02 = 0 mol of N02


and a total of 1 mol of gas present. At chemical equilibrium, there are
nN204 = 0.834 mol of N204 and nN02 = 0.332 mol of N02 and a total
of 1.166 mol of gas present. The reaction occurs at temperature To and
pressure Po. Therefore, the initial volume is Vo = ((I)RTo/Po)
and
the final volume is V = ((1. 166)RTo/Po) = 1.166Vo.
(e) The heat of reaction for the reaction occurring at temperature To and
pressure Po is

(8H)
8e

0
P,T

= ~ RTo _ RTo In
2
= 4.68RTo.

[~02]
XN204

= ~RTo _ RTo In [(0.285)21

0.715

J
(4)

86

INTRODUCTION TO THERMODYNAMICS

.... S2.E. The Thermodynamics of Electrolytes [19-21]


A very important phenomenon in biological systems is the flow of charged
ions in electrically neutral solutions. Of particular interest is the behaviour
of dilute solutions of salt (the solutes), such as NaCI, KCI, or CaCI2, in water
(the solvent). If we denote the negative ion (the anion) as A-and the positive
ion (the cation) as C+, the dissociation of the salt into charged ions can be
denoted
A- C+
lIa

-->.

lIe"---

t/

A-

+ Vc C+

(2.254)

(e.g., CaCh ~ 2CI- -l-Ca"), where Va and Vc are the stoichiometric coefficients
for the dissociation. The condition for equilibrium is
(2.255)
where J-t~(J-t~) is the electrochemical of ion, A - (C+) and J-tac is the chemical
potential of the undissociated salt. Electrical ne
Ii
f
id requires that

Converted with

(2.256)

STOI Converter

e charge of an
where Zae(zce
electron. It is
h as NaCI, and
trial version
isting of NaCI
water must b
he numbers of
molecules an
anions and
. However, a
thermodynamic framework can be set up to deal with the ions as separate
entities, and that is what we will describe below.
The chemical potential of the salt in aqueous solution is extremely
complicated, but experiments show that it can be written in the form

hDP:llwww.stdutililV.com

J-tac(P, T,xac) = 1-L?u(P, T)

+ RTln

(aac) ,

(2.257)

where aac is called the activity and will be defined below, and J-t~c(P, T) is the
chemical potential of the salt in aqueous solution at temperature T and pressure
P in the limit of infinite dilution. J-t~AP, T) is proportional to the energy needed
to add one salt molecule to pure water. We now define a reference value for the
Gibbs free energy to be
(2.258)
where nac(nw) is the number of moles of salt (water) molecules in the system of
interest and J-t~(P, T) is the chemical potential of pure water. In principle, we
can obtain a numerical value for GO. The difference between this reference

87

SPECIAL TOPICS: THE THERMODYNAMICS OF ELECTROLYTES

Gibbs free energy and the actual Gibbs free energy can be written

G - GO = nae (J-tae - J-t~e) + nw (J-tw - J-t~) = naeRT In (ll!ae)

+ nw (J-tw -

J-t~).
(2.259)

We can now relate the activity, ll!ae, for the salt molecule
ions. We define

to activities for the

(2.260)
Then
(2.261)
The condition
electrochemical

for equilibrium,
Eq. (2.255),
potentials of the ions to be

is satisfied

if we define

the

(2.262)

Converted with

and

STOI Converter

(2.263)

trial version

where J-t0 = J-t0


ity. Here is
the electric pot~
ion (i = a, h).
The quantities ll!a an ll!e are e ne
0
e e ac IVI res 0
e anion and
cation, respectively. It is found experimentally
that in the limit of infinite
dilution, ll!a = laca and ll!e = lece, where Ca and c, are the concentrations
(moles/volume) of the anions and cations, respectively. The quantities Ie and Ic
are called activity coefficients. In the limit c, ~ 0, f; ~ 1 (i = a, c). Solutions
for which Ic = 1 and la = 1 are said to be ideal. For ideal solutions the
electrochemical potentials of the ions can be written

hDP://www.stdutilitv.com

J-t~(P, T)

+ RTln

(ca)

+ ZaF

(2.264)

J-t~ = J-t~(P, T)

+ RTln

(ce)

+ ZeF.

(2.265)

J-t:

and

Because of these relationships,


as

dG

-S dT

-SdT

we can write changes in the Gibbs free energy

+ V dP + J-tabdnab + J-twdnw
+ V dP + J-t:dna + J-t~dne + J-twdnw,

(2.266)

88

INTRODUCTION TO THERMODYNAMICS

where dna = vadnac and dn; = vcdnac. Therefore, changes in the Gibbs free
energy can be expressed either in terms of the salt or in terms of the ions. In
Exercise 2.12, we give some examples of how these quantities can be used .

EXERCISE 2.12. Consider a vessel held at constant temperature T and


pressure P, separated into two disjoint compartments, I and II, by a
membrane. In each compartment there is a well-stirred, dilute solution of a
solute and a solvent. Assume that the membrane is permeable to the solute,
but not permeable to the solvent. Compute the ratios of the concentrations of
solute in the two compartments for the following two cases. (a) The solute is
uncharged, but the solvents in the two compartments are different. (b) The
solute is charged and the fluids in the two compartments are maintained at
different electric potentials, but the solvents are the same.
G(kcal)

/
Converted with

STOI Converter
trial version

hDP:llwww.stdutililV.com

I "'-----

T = 2To

20~--~--~--~--~--~~~
0.2 0.4 0.6 0.8

1.0

Answer:
(a) Denote the chemical potential of the solute in compartment I (II) as
(/-Ls) I ( (J-Ls ) II ). Since the solute is uncharged, the chemical potential can
be written J-Ls = J-L~ + RTln (c.), where J-L~ is the chemical potential of
the solute in the solvent in the limit of infinite dilution, and cs is the
concentration of the solute. The condition for equilibrium of the
solute in the two compartments is
(1)
Thus, at equilibrium the ratio of the concentrations of solute in the
two compartments is

S = exp ((J-L~)

ciI

IIRT'- (J-L~) I)

(2)

89

REFERENCES AND NOTES

sf

The ratio (3s ==


Sf is called the partition coefficient. It is a measure
of the different solubility of solute in the two solvents.
(b) Since the solute is charged, the chemical potential can be written
I-Ls = I-L~+ RTln (cs) + zsF, where is the electric potential and z, is
the charge of the solute particles. Since the solvents are the same, the
condition for equilibrium of the solute in the two compartment is

(3)
Thus, at equilibrium the ratio of the concentrations of solute in the
two compartments is

s _exp (zsF6)RT

C~f -

'

(3)

where 6 = II - f. Conversely, the potential difference for a given


ratio of concentrations is

(4)
The pot
differen

Converted with

STOI Converter

ncentration

trial version

hDP://www.stdutilitv.com
REFERENCES AND NOTES
1. Most books on advanced calculus contain thorough discussions of exact differentials.
2. 1. O. Hirschfelder, C. F. Curtiss, and R. B. Byrd, Molecular Theory of Gases and
Liquids (John Wiley & Sons, New York, 1954).
3. J. B. Partington, An Advanced Treatise on Physical Chemistry, Vol. I (Longmans,
Green, and Co., London, 1949).
4. D. Hodgeman, ed., Handbook of Chemistry and Physics (Chemical Rubber Publishing Co., Cleveland, 1962).
5. International Critical Tables, ed. E. W. Washburn (McGraw-Hill, New York, 1957).
6. M. W. Zeman sky, Heat and Thermodynamics (McGraw-Hill, New York, 1957).
7. O. D. Jefimenko, Electricity and Magnetism (Appleton-Century-Crofts, New York,
1966).
8. P. M. Morse, Thermal Physics (w. A. Benjamin, New York, 1965).
9. J. S. Dugdale, Entropy and Low Temperature Physics (Hutchinson University
Library, London, 1966).
10. H. B. Callen, Thermodynamics (John Wiley & Sons, New York, 1960).
11. D. ter Haar and H. Wergeland, Elements of Thermodynamics (Addison-Wesley,
Reading, MA, 1969).

90

INTRODUCTION TO THERMODYNAMICS

and I. Prigogine, Thermodynamic Theory of Structure, Stability, and


Fluctuations (Wiley-Interscience,
New York, 1971).
13. D. Kondepudi and I. Prigogine, Thermodynamics: From Heat Engines to Dissipative
Structures (J. Wiley and Sons, New -York, 1998).
14. H. E. Stanley, Introduction to Phase Transitions and Critical Phenomena (Oxford
12. P. Glansdorff

University

Press, Oxford,

1971).

15. R. K. Griffiths, 1. Math. Phys. 5, 1215 (1964).


16. H. B. Callen, Thermodynamics (John Wiley & Sons, New York, 1960).
17. I. Prigogine and R. Defay, Chemical Thermodynamics (Longmans,
London, 1954).
18. J. Waser, Basic Chemical Thermodynamics (W. A. Benjamin,

Green and Co.,

New York, 1966).

19. H. S. Harned and B. B. Owen, The Physical Chemistry of Electrolytic Solutions


(Reinhold, New York, 1958).
20. S. G. Schultz, Basic Principles of Membrane Transport (Cambridge
Cambridge, 1980).

University Press,

21. A. Katchalsky and P. F. Curran, Nonequilibrium Thermodynamics in Biophysics


(Harvard University Press, Cambridge, MA, 1967).

Converted with

PROBLEM
Problem 2.1. 1
differential is e
(a) du.;

STOI Converter

= it

ases in which the

trial version

hDP://www.stdutilitv.com

(b) dui; = (y
(c) dUc=(2'r---~----~~--------------------

Problem 2.2. Consider the two differentials (1) du, = (2xy + .r)dx + .rdy and (2)
dU2= y(x - 2y)dx - .rdy. For both differentials, find the change in u(x,y) between two
points, (a, b) and (x, y). Compute the change in two different ways: (a) Integrate along
the path (a,b) ~ (x, b) ~ (x,y), and (b) integrate along the path (a,b) ~ (a,y) ->
(x,y). Discuss the meaning of your results.
Problem 2.3. Electromagnetic

radiation in an evacuated vessel of volume V at


equilibrium with the walls at temperature T (black body radiation) behaves like a gas of
photons having internal energy U = aVT4 and pressure P = (1/3)aT4, where a is
Stefan's constant. (a) Plot the closed curve in the P-V plane for a Carnot cycle using
blackbody radiation. (b) Derive explicitly the efficiency of a Carnot engine which uses
blackbody radiation as its working substance.

Problem 2.4. A Carnot engine uses a paramagnetic substance as its working substance.
The equation of state is M = (nDH IT), where M is the magnetization,
H is the
magnetic field, n is the number of moles, D is a constant determined by the type of
substance, and T is the temperature. (a) Show that the internal energy U, and therefore
the heat capacity CM, can only depend on the temperature and not the magnetization.
Let us assume that CM = C = constant. (b) Sketch a typical Carnot cycle in the M-H
plane. (c) Compute the total heat absorbed and the total work done by the Carnot
engine. (d) Compute the efficiency of the Carnot engine.

91

PROBLEMS

Fig.2.1S.

Problem 2.5. Find the efficiency of the engine shown in Fig. 2.18. Assume that the
operating substance is an ideal monatomic gas. Express your answer in terms of VI and
V2 (The process
1 and 2 -+ 3

Converted with
Problem 2.6. 0
20 atm. (a) How
average isotherm
and the average

STOI Converter
trial version

hDP:llwww.stdutililV.com

C from 1 atm to
Assume that the
0.5 x 1O-4/atm
= 2 x 1O-4;oC.

Problem 2.7. C
.19. The engine
uses a rubber ba
,
onstant, J is the
tension, L is the length per unit mass, and T is the temperature in Kelvins. The specific
heat (heat capacity per unit mass) is a constant, CL = C.

I
I
,
I
I

Fig. 2.19.

I
I

I
,

I
I
I

Lo

2Lo

92

INTRODUCTION TO THERMODYNAMICS

Problem 2.S. Experimentally one finds that for a rubber band

where J is the tension, a = 1.0 x 103 dyne/K, and Lo = 0.5 m is the length of the band
when no tension is applied. The mass of the rubber band is held fixed. (a) Compute
(8LI 8T) J and discuss its physical meaning. (b) Find the equation of state and show that
dJ is an exact differential. (c) Assume that the heat capacity at constant length is
CL = 1.0J/K. Find the work necessary to stretch the band reversibly and adiabatically to
a length of 1m. Assume that when no tension is applied, the temperature of the band is
T = 290 K. What is the change in temperature?
Problem 2.9. Blackbody radiation in a box of volume V and at temperature T has
internal energy U = aVT4 and pressure P = (1/3)aT4, where a is the Stefan-Boltzmann
constant. (a) What is the fundamental equation for blackbody radiation (the entropy)?
(b) Compute the chemical potential.
Problem 2.10. Two vessels, insulated from the outside world, one of volume VI and the
other of volume V2, contain equal numbers N of the same ideal gas. The gas in each
vessel is orgina
d and allowed to
reach equilibri
Converted with
sulated from the
outside world.
m work, LlWfree,

:~!~:
~~STOI Converter
~~,t

Problem 2.11.
order in the de

your answer in
rminated at first

trial version

hDP://www.stdutilitv.com
where B2 (T) is the second virial coefficient. The heat capacity will have corrections to
its ideal gas value. We can write it in the form
3
N2kB
CV,N = "2NkB ------y-F(T).

(a) Find the form that F(T) must have in order for the two equations to be
thermodynamically consistent. (b) Find Cp,N. (c) Find the entropy and internal energyProblem 2.12. Prove that

Cy, = (8H)
8T

and

Y,N

_ T (8X)
(8H)
8Y ,
8T
TN -

-x
YN
,

Problem 2.13. Compute the entropy, enthalpy, Helmholtz free energy, and Gibbs free
energy of a paramagnetic substance and write them explicitly in terms of their natural
variables when possible. Assume that the mechanical equation of state is m = (DH/T)
and that the molar heat capacity at constant magnetization is em = e, where m is the
molar magnetization, H is the magnetic field, D is a constant, e is a constant, and T is the
temperature.

93

PROBLEMS

Problem 2.14. Compute the Helmholtz free energy for a van der Waals gas. The
equation of state is (P + (Qn2/V2))(V - nb) = nRT, where a and b are constants which
depend on the type of gas and n is the number of moles. Assume that the heat capacity is
CV,n = (3/2)nR. Is this a reasonable choice for the heat capacity? Should it depend on
volume?
Problem 2.1S. Prove that (a) K,T(Cp

Cv) = TVQ~ and (b) (CP/Cv)

= (K,r/K,s).

problem 2.16. Show that Tds = cx(8T /8Y)xdY + cy(8T /8x)ydx, where x = X/n is
the amount of extensive variable, X, per mole, Cx is the heat capacity per mole at
constant x, and Cy is the heat capacity per mole at constant Y.
Problem 2.17. Compute the molar heat capacity Cp, the compressibilities K,T and K,s,
and the thermal expansivity Qp for a monatomic van der Waals gas. Start from the fact
that the mechanical equation of state is P = (RT /(v - b)) - (Q/v2) and the molar heat
capacity is Cv = 3R/2, where v = V [n is the molar volume.
Problem 2.1S. Compute the heat capacity at constant magnetic field CH,n, the
susceptibilities XT,n and XS,n, and the thermal expansivity QH,n for a magnetic system,
given that the mechanical equation of state is M = nDH/T and the heat capacity is
CM,n = nc, where M is the magnetization, H is the magnetic field, n is the number of
moles, D is a c
..
erature.

Converted with

Problem 2.19.
(a/RT2v) and
v = (V In) is t
Under what con

Qp

STOI Converter
trial version

Problem 2.20.
Which engine
efficiency of a _,_____

h D_P_: II_WWW
d -.__ .S_t_ut_I_ltv_.c_o_m
_

= (R/Pv)+

(b/P)), where
ion of state. (c)
ines in Fig. 2.20.
ot cycles. The
___J

Problem 2.21. It is found for a gas that K,T = Tvf(P) and Qp = (Rv/P) + (Av/T2),
where T is the temperature, v is the molar volume, P is the pressure, A is a constant, and
f(P) is an unknown function of P. (a) What is f(P)? (b) Find v = v(P, T).
Problem 2.22. A monomolecular liquid at volume VL and pressure PL is separated from
a gas of the same substance by a rigid wall which is permeable to the molecules, but
does not allow liquid to pass. The volume of the gas is held fixed at VG, but the volume

T2

Tl

I7

~b

81
Fig. 2.20.

(b)

(a)

82

94

INTRODUCTION TO THERMODYNAMICS

of the liquid can be varied by moving a piston. If the pressure of the liquid is increased
by pushing in on the piston, by how much does the pressure of the gas change? [Assume
the liquid is incompressible (its molar volume is independent of pressure) and describe
the gas by the ideal gas equation of state. The entire process occurs at fixed temperature

T.]
Problem S2.1. Consider a gas obeying the Dieterici equation of state,
P

= (V

nRT

- nb) exp

(na)
- V RT '

where a and b are constants. (a) Compute the Joule coefficient. (b) Compute the JouleKelvin coefficient. (c) For the throttling process, find an equation for the inversion curve
and sketch it. What is the maximum inversion temperature?
Problem S2.2. An insulated box is partitioned into two compartments, each containing
an ideal gas of different molecular species. Assume that each compartment has the same
temperature but different number of moles, different pressure, and different volume [the
thermodynamic variables of the ith box are (Pi, T, n., Vi)]. The partition is suddenly
removed and the system is allowed to reach equilibrium. (a) What are the final
temperature and pressure? (b) What is the change in the entropy?
Problem 82.3. 1
T.and pressure P
molecules of type
molecules of typ
mix so the final t
the entropy of mi
(c) What is the
results for (b) an

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com

~ at temperature
type a and Ns
f'type a and N2,b
p are allowed to
1). (a) Compute
~dNlb
=N2b
p. Dis~uss your

~-----------------------------------

Problem 82.4. An insulated box with fixed total volume V is partitioned into m
insulated compartments, each containing an ideal gas of a different molecular species.
Assume that each compartment has the same pressure but a different number of moles, a
different temperature, and a different volume. (The thermodynamic variables for the ith
compartment are (P, ni, Ti, Vi).) If all partitions are suddenly removed and the system is
allowed to reach equilibrium: (a) Find the final temperature and pressure, and the
entropy of mixing. (Assume that the particles are monatomic.) (b) For the special case
of m = 2 and parameters ni = 1mol, TI = 300 K, VI = 1 liter, nz = 3 mol, and V2 =
2 liters, obtain numerical values for all parameters in part (a).

V, P, T

V, P, T

2V, P, T

, = N1a + N2a
Nb
Fig. 2.21.

Nlb +N2b

95

PROBLEMS

Problem S2.5. A tiny sack made of membrane permeable to water but not NaCI
(sodium chloride) is filled with a 1% solution (by weight) of NaCI and water and is
immersed in an open beaker of pure water at 38 DC at a depth of 1 ft. (a) What osmotic
pressure is experienced by the sack? (b) What is the total pressure of the solution in the
sack (neglect surface tension)? Assume that the sack is small enough that the pressure of
the surrounding water can be assumed constant. (An example of such a sack is a human
blood cell.)
Problem S2.6. A solution of particles A and B has a Gibbs free energy
G(P, T, nA, ns) =nAgA (P, T)

+ nBgB(P,

+ AAB--nAnB
+ nARTln
n

T)

n2

n2

+ 2 AAA : + 2 ABB :

(XA)

+ nBRTln

(XB).

Initially, the solution has nA moles of A and ne moles of B. (a) If an amount, tl.nB, of B is
added keeping the pressure and temperature fixed, what is the change in the chemical
potential of A? (b) For the case AAA = ABB = AAB, does the chemical potential of A
increase or decrease?
Problem S2.7. Consider the reaction

which occurs in
12. Assume that
equations for
G(T, P, e), as a
and (ii) P = 1 a
~c~o:

~~~~~

Converted with

STOI Converter

I each of H2 and

trial version

hDP:llwww.stdutililV.comi~b~~~

Use ideal gas


bs free energy,
and T = 298 K
A(T, P, e), as a

70: ;t:

1 atm and temperature T = 298 K? How many moles of HI, H2, and 12 are present at
equilibrium? (d) If initially the volume is Yo, what is the volume at equilibrium for
P = 1 atm and T = 298 K? (e) What is the heat of reaction for P = 1 atm and
T = 298K?
Problem S2.8. Consider the reaction
2NH3 ~ N2 + 3H2,
which occurs in the gas phase. Start initially with 2 mol of NH3 and 0 mol each of H2
and N2. Assume that the reaction occurs at temperature T and pressure P. Use ideal gas
equations for the chemical potential. (a) Compute and plot the Gibbs free energy,
G(T, P, e), as a function of the degree of reaction,
for (i) P = 1 atm and T = 298 K
and (ii) P = 1 atm and T = 894K. (b) Compute and plot the affinity, A(T,p,e),
as a
function of the degree of reaction, for (i) P = 1 atm and T = 298 K and (ii) P = 1 atm
and T = 894 K. (c) What is the degree of reaction,
at chemical equilibrium for P =
1 atm and temperature T = 894 K? How many moles of HI, H2, and 12 are present at
equilibrium? (d) If initially the volume is Vo, what is the volume at equilibrium for
P = 1 atm and T = 894 K? (e) What is the heat of reaction for P = 1 atm and
T = 894K?

e,

e,

e,

3
THE THERMODYNAMICS OF
PHASE TRANSITIONS

3.A. INTRODUCTORY

REMARKS

A thermodynamic system can exist in a number of different phases whose


macroscopic behavior can differ dramatically. Generally, systems become more
ordered as ten
ion begin to
overcome them
Converted with
es into more
ordered states.
temperature,
although eviden
copic scale as
the critical temp
oncemed with
the thermodyna
trial version
~ion of phase
transitions in te
~e shall study
them from a mi
The first step in trying to understand the phase changes that can occur in a
system is to map out the phase diagram for the system. At a transition point, two
(or more) phases can coexist in equilibrium with each other. The condition for
equilibrium between phases is obtained from the equilibrium conditions derived
in Chapter 2. Since phases can exchange matter and energy, equilibrium occurs
when the chemical potentials of the phases become equal for given values of Y
and T. From the equilibrium condition, we can determine the maximum number
of phases that can coexist and, in principle, find equations for the regions of
coexistence (the Clausius-Clapeyron equation).
At a phase transition the chemical potentials of the phases, and therefore the
Gibbs free energy, must change continuously. However, phase transitions can be
divided into two classes according the behavior of derivatives of the Gibbs free
energy. Phase transitions which are accompanied by a discontinuous change of
state (discontinuous first derivatives of the Gibbs free energy with respect to
temperature and displacement) are called first-order phase transitions. Phase
transitions which are accompanied by a continuous change of state (but
discontinuous higher-order derivatives) are called continuous phase transitions.
We give examples of both in this chapter.
Classical fluids provide some of the most familiar examples of first-order
phase transitions. The vapor-liquid, vapor-solid, and liquid-solid transitions
1

STDU Converter
http://www.stdutilitv.com

96

INTRODUCTORY

REMARKS

97

are all first order. We shall discuss the phase transitions in classical fluids in
some detail. For the vapor-solid and vapor-liquid transitions, we can use the
Clausius-Clapeyron equation to find explicit approximate equations for the
coexistence curves. Since the vapor-liquid transition terminates in a critical
point, we will focus on it and compare the observed behavior of the vaporliquid coexistence region to that predicted by the van der Waals equation.
Superconductors and superfluids are especially interesting from the
standpoint of thermodynamics because they exhibit both first-order and
continuous phase transitions and they provide a test for the third law of
thermodynamics. In the absence of a magnetic field, the transition from a
normal to a superconducting state in a metal is a continuous phase transition. It
is a phase transition which is purely quantum mechanical in nature and results
from a macroscopic condensation of pairs of electrons into a single quantum
state. The superfluid transitions in liquid He3 and liquid He4 are of similar
quantum origin. The superfluid transitions in liquid He3 involve pairs of
"dressed" He3 atoms which condense, on a macroscopic scale, into a single
quantum state. In liquid He4 a macroscopic number of "dressed" He4 atoms
condense into
r u d state. When Ii uid He3 and li uid He4 are mixed
together, they
a continuous
superfluid phas
Converted with
ion.

STOU Converter

Most phase
int (the liquidsolid transition
ove which one
phase exists an
trial version
ars. When the
new phase ap
thermodynami
new phase. Fo
a connection
between the symmetries of the high-temperature phase and those of the lowtemperature phase. For continuous phase transitions, there is always a welldefined connection between the symmetry properties of the two phases.
Ginzburg and Landau developed a completely general theory of continuous
symmetry-breaking phase transitions which involves an analytic expansion of
the free energy in terms of the order parameter. We shall discuss the GinzburgLandau theory in this chapter and show how it can be applied to magnetic
systems at the Curie point and to superfluid systems.
The critical point plays a unique role in the theory of phase transitions. As a
system approaches its critical point from the high temperature side, it begins to
adjust itself on a microscopic level. Large fluctuations occur which signal the
emergence of a new order parameter which finally does appear at the critical
point itself. At the critical point, some thermodynamic variables can become
infinite. Critical points occur in a huge variety of systems, but regardless of the
particular substance or mechanical variable involved, there appears to be a great
similarity in the behaviour of all systems as they approach their critical points.
One of the best ways to characterize the behavior of systems as they approach
the critical point is by means of critical exponents. We shall define critical
exponents in this chapter and give explicit examples of some of them for the

http://www.stdutilitv.com

98

THE THERMODYNAMICS OF PHASE TRANSITIONS

liquid-vapor transition in simple fluids. At the end of the chapter in the special
topics section we define critical exponents for the Curie point.
The section on special topics contains further applications of thermodynamics at phase transitions and at the interface of different phases. In Chapter
2 we derived conditions for mechanical thermodynamic equilibrium between
two parts of a system in equilibrium. If the system consists of two phases, such
as a liquid and a gas phase, then the interface between the two phases may be
under tension due to an imbalance of molecular forces at the interface. If the
interface is under tension, the condition for mechanical equilibrium must be
modified to include mechanical effects due to the surface tension.
A rather spectacular phenomenon which occurs in superfluid systems is the
thermomechanical, or fountain, effect. As we shall see, it is possible to use a
heat source to drive a superfluid fountain. Even though superfluids are highly
degenerate quantum systems, the fountain effect surprisingly can be described
in terms of classical thermodynamics. All we need is a system composed of two
interpenetrating fluids, one of which carries no entropy. Then simple arguments
give us the fountain effect.
A binary mix
f mole ules in t e fluid state
of a first-orde
temperature w
each of which
Finally we
application of
Ginzburg-Lan
magnetic field.

STOU Converter
trial version

Most systems can exist in a number of different phases, each of which can
exhibit quite different macroscopic behavior. The particular phase that is
realized in nature for a given set of independent variables is the one with the
lowest free energy. For certain values of the independent variables, two or more
phases of a system can coexist. There is a simple rule, called the Gibbs phase
rule, which tells us the number of phases that can coexist. Generally, coexisting
phases are in thermal and mechanical equilibrium and can exchange matter.
Under these conditions, the temperature and chemical potentials of the phases
must be equal (cf. Section 2.H) and there will be another condition between
mechanical variables expressing mechanical equilibrium. For example, for a
simple PIT system, the pressures of the two phases may be equal (if surface
tension can be neglected).
For simplicity, let us first consider a YXT system which is pure (composed of
one kind of particle). For a pure system, two phases, I and II, can coexist at a
fixed value of Yand T if their respective chemical potentials are equal:

l (Y, T) = Jil (Y, T).

(3.1)

99

COEXISTENCE OF PHASES: GIBBS PHASE RULE

(The chemical potentials are functions only of intensive variables.) Equation


(3.1) gives a relation between the values of Yand T for which the phases can
coexist,

Y = Y(T),

(3.2)

and in the Y- T plane it defines a coexistence curve for the two phases. If the
pure system has three phases, I, II, and III, they can only coexist at a single
point in the Y-T plane (the triple point). Three coexisting phases must satisfy
the equations

l (Y, T)

J1-11 (Y,

T) = J1-1ll (Y, T).

(3.3)

Since we have two equations and two unknowns, the triple point is uniquely
determined. For a pure system, four phases cannot coexist, because we would
then have three equations and two unknowns and there would be no solution.
As an example of the Gibbs phase rule for pure substances, we show the
coexistence curves for various solid hases of water cf. Fi . 3.1). We see that
although wate
phases can co
For a mixt
coexist. To sh
then there w
(Y, T,Xl, ... ,
have several

Converted with

ore than three


2 phases can
in each phase,
hase, namely,
f type i. If we
given type of

STOI Converter
trial version

hDP://www.stdutilitv.com

40~----------------------~
VII

o~
-1000

_50

Fig. 3.1. Coexistence curves for the solid and liquid phases of water. In accordance with
the Gibbs phase rule, no more than three phases can coexist [1]. (Based on Ref. 2.)

100

THE THERMODYNAMICS OF PHASE TRANSITIONS

particle must have the same value in each of the various phases. Thus, if there
are r coexisting phases at a given value of Yand T, the condition for equilibrium
is
J-llI (Y , T ,xl I ,x2I,

= J-llII (Y "XlT

,X1I)
_I

=
J-l2I(y , T 'Xl""I

II 'X II , ...
2

p{(Y, T,x[, ...

II
= IL2I1(y , T 'Xl""

I)

'XI-1

= J-lHY, T,x[, ...


J-llI(y , T ,XI""I

T
= ILlI1(y "XI""

I)
'XI-1

= J-l!(Y,

II

T,x[, ...

= ...

)
,XI-II
1

(3.4)

,XI-I)'

= ...

(3.5)

II )
,Xl-I
= ...

(3.6)

II )

'Xl-I

,X~_I)'

,XI-I)'

Equations (3.4)-(3.6) give l(r - 1) equations to determine 2 + r(l- 1)


unknowns. For a solution, the number of equations cannot be greater than the
number of unknowns. Thus, we must have l(r - 1) ~2 + yU - 1) or r~l + 2.
The number 01
+ 2, where I is
the number of
Converted with
and r ~ 3 as we
~~~~~sb~!~r~~!

STOI Converter

t, four different

trial version
3.C. CLASS

hDP:llwww.stdutililV.com

As we change the independent intensive variables (Y, T, Xl, ... ,Xl) of a system,
we reach values of the variables for which a phase change can occur. At such
points the chemical potentials (which are functions only of intensive variables)
of the phases must be equal and the phases can coexist.
The fundamental equation for the Gibbs free energy, in a system with I
different kinds of molecules, is
I

= L:njJ-lj,

(3.7)

j=I

where nj is the number of moles of the jth constituent, and J-lj is its chemical
potential (cf. Section 2.F.4). For processes which occur at constant Yand T,
changes in the Gibbs free energy can be written
I

[dG]y,T

L:J-ljdnj.

(3.8)

j=1

Thus, at a phase transition, the derivatives ILl

(BG IBnj)

Y T {no }
, , 1#

must be equal

101

CLASSIFICATION OF PHASE TRANSmONS

,,
'" ,
"'~SlO'P

......
e~

I ~ ~ .......
-$//

,"'0

V-

I
I

Vll

(aG)
-

S=-

'",
' ,~
-,

:':l

~I

aT

T,{nj}

P,{nj}

'
VI

......

~----~------~~T

~--~--------~p
8G)
( - ~I

- ap

;?)~

r-

--I
~li

Sl/~__
51I

II

Fig. 3.2. Typical

---:

~~~~~6,'----~.~

~'~nL_

;:

Converted with

STOI Converter

ase transition.

and the Gibbs


trial version
However, no
restriction is
__
{nj} and S =
-(8G/8T)Y,{nj}
classify phase
transitions. If the derivatives (8G/8Y)r,{nj}
and (8Gj8T)Y,{nj}
are discontinuous at the transition point (that is, if the extensive variable X and the entropy S
have different values in the two phases), the transition is called "first-order." If
the derivatives (8Gj8T)y,{n.}
and (8G/8Y)r,{n}
are continuous at the transition
but higher-order derivativ~s are discontinuo~s, then the phase transition is
continuous. (The terminology "nth-order phase transition" was introduced by
Ehrenfest to indicate a phase transition for which the nth derivative of G was the
first discontinuous derivative. However, for some systems, higher-order
derivatives are infinite, and the theory proposed by Ehrenfest breaks down
for those cases.)
Let us now plot the Gibbs free energy for first-order and continuous
transitions in a PVT system. For such a system the Gibbs free energy must be a
concave function of P and T (cf. Section 2.H.3). The Gibbs free energy and its
first derivatives are plotted in Fig. 3.2 for a first-order phase transition. A
discontinuity in (8G/8P)r,{nj}
means that there is the discontinuity in the
volume of the two phases,

hDP://www.stdutllllV.com

AV=Vl_Vll=

8G)
( -8P

T,{n

J}

(80)
-

II

8P

r,{nj}'

(3.9)

102

THE THERMODYNAMICS OF PHASE TRANSITIONS

~--------~--------~T

5-- - (BG)
BT

I
P,{nj}

I
I
Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com
I

~--------~--------~T
Fig. 3.3. Typical behavior for the Gibbs free energy at a continuous phase transition.
The heat capacity can exhibit a peak at the transition.

and a discontinuity in (8G/8T)p,{nj}


entropy of the two phases,

means there is a discontinuity in the

(3.1 0)

103

PURE PVT SYSTEMS

Since the Gibbs free energy is the same for both phases at the transition, the
fundamental equation H = G + TS shows that the enthalpy of the two phases is
different,
~H

= HI

- HII

= T~S,

(3.11)

for a first-order phase transition. The enthalpy difference, ~H, is also called the
latent heat.
For a continuous phase transition, the Gibbs free energy is continuous but its
slope changes rapidly. This in tum leads to a peaking in the heat capacity at the
transition point. An example is given in Fig. 3.3. For a continuous transition,
there is no abrupt change in the entropy or the extensive variable (as a function
of Yand T) at the transition.
In the subsequent sections we shall give examples of first-order and
continuous phase transitions.

A pure PVT s

Converted with

STOI Converter

of molecule.

The molecules
nge attractive
region outside
: a gas phase,
a liquid phase
trial version
le of a pure
PVT system i
.es of phase
transitions in
ances in this
section, it is cOI'nviemem:-n:,.---rnescnneLnel[1lma:se-1l'3:nslTIUIlS"""InliermS
of molar
densities.

hDP://www.stdutilitv.com

3.D.I. Phase Diagrams


A typical set of coexistence curves for pure substances is given in Fig. 3.4 (Note
that Fig. 3.4 does not describe the isotopes of helium, He3 or He\ which have
superfluid phases, but it is typical of most other pure substances.) Point A on the
diagram is the triple point, the point at which the gas, liquid, and solid phases
can coexist. Point C is the critical point, the point at which the vaporization
curve terminates. The fact that the vaporization curve has a critical point means
that we can go continuously from a gas to a liquid without ever going through a
phase transition, if we choose the right path. The fusion curve does not have a
~ritical point (none has ever been found). We must go through a phase transition
I? going from the liquid to the solid state. This difference between the gashquid and liquid-solid transitions indicates that there is a much greater
fundamental difference between liquids and solids than between liquids and
gases, as one would expect. The major difference lies in their symmetry
properties. Solids exhibit spatial ordering, while liquids and gases do not. (We
will use "vapor" and "gas" interchangeably.)

104

THE THERMODYNAMICS OF PHASE TRANSmONS

T
Fig. 3.4. Coexistence curves for a typical pure PVF system. Point A is the triple point
and point C is the critical point. The dashed line is an example of a fusion curve with
negative slope.

J!\

.....

P
n

"Q

.....

""
Converted with
rc

STOI Converter
trial version

hDP://www.stdutilitv.com
_ .....

v
Fig. 3.S. A plot of the coexistence regions for a typical PVF system. All the phase
transitions here are first order. The dashed lines represent isotherms.

The transitions from gas to liquid phase, from liquid to solid phase, and from
gas to solid phase are all first-order transitions and are accompanied by a latent
heat and a change in volume. In Fig. 3.5, we have drawn the phase diagram in
the P-v plane. The dashed lines are lines of constant temperature. We notice
that the slope of the dashed lines is negative, (8P / 8v)y < O. This is a statement
of the stability condition, ""T > 0 (cf. Section 2.H.2). In the region of
coexistence of phases, the isotherms (dashed lines) are always flat, indicating
that in these regions the change in volume occurs for constant P and T.
It is interesting to plot the pressure as a function of molar volume and
temperature in a three-dimensional
figure. As we can see in Fig. 3.6, the result
is similar to that of a mountain with ski slopes. The height of the mountain at
any given value of v and Tis the pressure. Fig. 3.6 actually corresponds to a plot
of the equation of state for the pure system. The shaded region is the region of

lOS

PURE PVT SYSTEMS

v
Fig. 3.6. Three-dimensional sketch of the equation for the typical pure PVT system.

coexistence of more than one phase. Figure 3.4 is a projection of Fig. 3.6 on the
p.: T plane and'
.
'"
lane.

Converted with

STOI Converter

Gin, of two

coexisting pha
trial version
must be equal.
If we change t
es coexist (that
is, if we move
Gibbs free
energy of the
u
us, dt = dt1
along the coexistence curve. We can use this fact to find an equation for the
coexistence curve. We use Eq. (2.108) to write

hDP://www.stdutilitv.comlar

(3.12)
along the coexistence curve, where v is the molar volume and s is the molar
entropy. Thus,

(3.13)
along the coexistence curve, where l:l.s = sl - sII is the difference in the molar
entropy of the two phases and l:l.v = v! - V II is the difference in the molar
volume of the two phases. Equation (3.13) can also be written in terms of the
latent heat, l:l.h = T l:l.s [cf. Eq. (3.11)] so that

(~~)

coex

rti;v

(3.14)

106

THE THERMODYNAMICS OF PHASE TRANSITIONS

Equation (3.14) is called the Clausius-Clapeyron equation. The latent heat, f1h,
is the heat absorbed per mole in the transition from phase II to phase I. It is of
interest to discuss the Clausius-Clapeyron equation for the three coexistence
curves in Fig. 3.4.
Exercise 3.1. Prove that the latent heat must always be positive (heat is
absorbed) when making a transition from a low-temperature phase to a hightemperature phase.
G

Converted with

STOU Converter

Answer: Le
re phase and
phase II is
pressure and
temperature
trial version
s free energy,
we must ha
GI > Gil below the tran
s implies that
(8GJ/8T)P,{nj} <
II
P,{nj} 0
a ove an
e ow the transition
temperature. Therefore SI = -(8GJ/8T)p,{n.}
>
= -(8Gll/8T)P,{n-)
and f1S = Tf1H is always positive in going 1 from the 10w-temperattJe
phase to the high-temperature phase.

hnp://www.stdutilitv.com
s

3.D.2. a. Vaporization Curve


If we evacuate a chamber and partially fill it with a pure substance, then for the
temperatures and pressures along the coexistence curve (the vaporization curve)
from the triple point to the critical point (point A to point C in Fig. 3.4) the
vapor and liquid phases will coexist in the chamber. For a given temperature T,
the pressure of the vapor and liquid is completely determined and is called the
saturated vapor pressure. As we change the temperature of the system the vapor
pressure will also change. The Clausius-Clapeyron equation tells us how the
vapor pressure changes as a function of temperature along the coexistence
curve.
We can obtain a rather simple equation for the vaporization curve if we make
some approximations. Let us assume that changes in the molar volume of the
liquid may be neglected relative to changes in the molar volume of the vapor
(gas) as we move along the coexistence curve, and let us assume the vapor

107

PURE PVT SYSTEMS

obeys the ideal gas law. Then ~v ~ RT /P, and the Clausius-Clapeyron
equation for the vapor pressure curve takes the form
dP)
( dT

_ P~hlg
R'J'2 '

(3.15)

coex -

where ~hlg is the latent heat of vaporization. If we assume that the latent heat
of vaporization is roughly constant over the range of temperatures considered,
we can integrate Eq. (3.15) to obtain
(3.16)
Thus, as the temperature is increased, the vapor pressure increases
exponentially along the vaporization curve. Conversely, if we increase the
pressure, the temperature of coexistence (boiling point) increases.
Exercise 3.2. Compute the molar heat capacity of a vapor along the
vaporization CUL-.....__-------------------,

Converted with

Answer: Alon
variable, which
the vapor is a
vaporization cu
Clapeyron equ
can be written
Ccoex

ndependent
of
ut along the
e Clausiuspation curve

STDU Converter

! entropy

trial version

http://www.stdutilitv.com

T(as)aT = T(8S)aT
=
T(Bv)ot (8P)
et

+ T(8S)

coex

ap (ap)
aT
T

coex

(1)

Cp _

coex'

where we have used Eqs. (2.8) and (2.112). The molar heat capacity, Cp, is
the heat capacity of the vapor held at constant pressure as we approach the
coexistence curve. If we use the ideal gas equation of state to describe the
properties of the vapor phase and if we use the Clausius-Clapeyron equation,
We obtain the following expression for the molar heat capacity along the
coexistence curve,
~hlg

Ccoex

Cp -

T'

(2)

At low enough temperatures, it is possible for the heat capacity, Ccoex, to be


~e~ative. This would mean that if the temperature of the vapor is raised and
It IS maintained in equilibrium with liquid phase, the vapor would give off
heat.

108

THE THERMODYNAMICS OF PHASE TRANSITIONS

3.D.2.b. Fusion Curve


The fusion curve does not terminate at a critical point but can have either
positive or negative slope. In Fig. 3.4 we used a solid line for the fusion curve
with a positive slope and a dashed line for the case of negative slope. The
Clausius-Clapeyron equation for the liquid-solid transition is
dP)
Ll.hsl
( dT coex= T Ll.vsI '

(3.17)

where Ll. V sl is the change in molar volume in going from the solid to the liquid
phase and Ll.hsl is the latent heat of fusion. If the volume of the solid is greater
than that of the liquid, then Ll.vsI will be negative and the slope, (dP/dT)coex'
will be negative. For the case of a fusion curve with positive slope, if we
increase the pressure at a fixed temperature, we simply drive the system deeper
into the solid phase. However, if the fusion curve has a negative slope, then
increasing the pressure at fixed temperature can drive the system into the liquid
phase. Water is an example of a system whose fusion curve has negative slope.
The negative s
.
.ng possible. As
the skate blad
e skater floats
along on an

STOI Converter

If a solid is pl
and temperat
equilibrium w
sublimation curve

trial version

hDP://www.stdutilitv.com

some pressure
ill coexist in
uation for the

IS

dP)
Ll.hsg
( dT coex= T Ll.vsg ,

(3.18)

where Ll.vsg is the change in molar volume in going from the solid to the gas
phase and Ll.hsg is the latent heat of sublimation.
If we again assume that the gas
phases obeys the ideal gas equation of state, the Clausius-Clapeyron equation
takes the form
dP)
P Ll.hsg
( dT coex= RT2 .

(3.19)

If the vapor pressure is known over a small temperature interval, then the latent
heat of sublimation can be obtained from Eq. (3.19). We can rewrite Eq. (3.19)
in the form
d In(P)

Ll.hsg = -R delfT) .
Then

fl.hSg

is proportional to the slope of the curve, In(P) versus liT.

(3.20)

109

PURE PVT SYSTEMS

EXERCISE 3.3. In the neighborhood of the triple point of ammonia


(NH3) , the equation for the sublimation curve is In(P) = 27.79 - 3726/T
and the equation for the vaporization curve is In(P) = 24.10 - 3005/T,
where P is measured in pascals and T is measured in kelvins. (a) Compute
the temperature and pressure of the triple point. (b) What is the latent heat of
sublimation? What is the latent heat of vaporization?
Answer:
(a) At the triple point, the pressure and temperature of the vapor, liquid,
and solid are the same. Therefore, the equation for the triple point
temperature,
Tr; is 27.79 - 3726/Tt = 24.10 - 3005/Tt or T, =
195.4 K. The triple point pressure, Pr, is P, = 6.13 kPa.
(b) The slope of the sublimation

8P)
( 8T
Therefo

curve is

_ 3726 P ,....,P ~hsg


RT2 .

coex-~"""

(1)
ion curve is

Converted with

STOI Converter

(2)

trial version

Therefo

At moderatelJ--_h_D-,-P_:_"_WWW_____,,--.S_t_d_u_ti_li_tv_.-=C_Om_-.-------'tum
effects to
cause significant deviations from the classical gas equation of state), the vapor
pressure along the sublimation curve is very low. We can use this fact to obtain
a fairly simple .equation for the sublimation curve which includes temperature
variations in the latent heat and heat capacity of the vapor and solid. Let us
first note that infinitesimal changes in the molar enthalpy of the solid can be
written

dh = T ds

+ vdP

= cpdT

+ v(1

- Tap)dP,

(3.21 )

where ap is the thermal expansivity (cf. Eqs. (2.155), (2.112), (2.144)). We can
use Eq. (3.21) to find an approximate expression for the difference between the
enthalpy of the solid at two different points along the sublimation curve. We
restrict ourselves to regions where the vapor pressure along the sublimation
curve is small and we neglect pressure variations in Eq. (3.21). Note that
dhg - dh, = (cp - cf,) dT along the coexistence curve. We then can integrate
Eq. (3.21) and write
~T

~hsg = ~h~g

J (4 - Cp)dT,
To

(3.22)

110

THE THERMODYNAMICS OF PHASE TRANSmONS

where ~hsg = hg - hs is the latent heat of sublimation at temperature T,


~h~g = h~ - h~ is the latent heat of sublimation at temperature To, h, and hg (h~
and h~) are the enthalpies of the solid and gas phases, respectively, at
temperature T (To) and very low pressure, and
and c~ are the moler heat
capacities of the gas and solid, respectively. If we now integrate the ClausiusClapeyron equation, (3.21), using Eq. (3.22) for ~hsg, we obtain the following
equation for the sublimation curve:

(P) ~ho (1 1) + JT ~dT" JT"

In -

Po

=___!!

---

To

To

RT

dT'(4-c~).

(3.23)

To

Equation (3.23) can be useful for extending the results of experimental


measurements [3].

3.D.3. Liquid-Vapor Coexistence Region [4, 5]


The liquid-vapor coexistence region culminates in a critical point and will be of
special intere
.
.
0 examine the
coexistence r
Converted with
us redraw the
ition in the P-v
coexistence cu
es are indicated
plane (cf. Fig.
ith temperature
by the solid Ii
existence
curve
fixed at To <
trial version
essure remains
(point A). At
ssure begins to
fixed until all ~------------------------------~
rise again.

STOI Converter
hDP://www.stdutilitv.com

Po

To

To

< Tc

Tc

coexistence curve
VB

Fig. 3.7. The coexistence curve for the vapro-liquid coexistence region for a pure pVT
system.

111

PURE PVT SYSTEMS

The amounts of liquid and vapor which coexist are given by the lever rule.
Let us consider a system with temperature To < Tc, pressure Po, and total molar
volume VD. The system is then in a state in which the liquid and vapor phases
coexist. The total molar volume, VD, is given in terms of the molar volume of
the liquid at point B, VI, and the molar volume of vapor at point A, vg, by
(3.24)
where xi is the mole fraction of liquid at point D and Xg is the mole fraction of
gas at point D. If we multiply Eq. (3.24) by xi + Xg == 1, we find
xi
Xg

(Vg -

(VD -

VD)
VI) .

(3.25)

Equation (3.25) is called the lever rule. It tells us that the ratio of the mole
fractions of liquid to gas at point D is equal to the inverse ratio of the distance
between point D and points A and B.
As long as
table (cf. Eq.
(2.179)). If we
B (the dashed
line), we obtain
er correspond

STOI Converter

to a minimum
A correspond
to supercooled
e at point B
correspond to s
trial version
Ie and can be
produced in th
sible that the
superheated liq
ressure. Such
states can also De rea lZ 1
ry
1
aintain them.
Systems with negative pressure are under tension and pull in on walls rather
than push out against them. As we approach the critical temperature, the region
of metastable states becomes smaller, and at the critical temperature it
disappears. Thus, no metastable states can exist at the critical temperature.
Also, as we approach the critical temperature the molar volumes of the liquid
and vapor phases approach one another, and at the critical temperature they
become equal.
The actual shape of the coexistence curve in the T-p plane (p is the mass
density) has been given by Guggenheim [6] for a variety of pure substances and
is reproduced in Fig. 3.8. Guggenheim plots the coexistence curve in terms of
the reduced quantitites T [T; and p] Pc, where T; and Pc are the critical
temperature and density, respectively, of a given substance. The reduced
quantities T [T; and p] Pc give a measure the distance of the particular substance
from its critical point (T; and Pc are different for each substance). Most
substances, when plotted in terms of reduced temperature and density, lie in
approximately the same curve. This is an example of the so-called law of
corresponding states, which says that all pure classical fluids, when described
in terms of reduced quantities, obey the same equation of state [7]. We shall see
this again when we return to the van der Waals equation. The reduced densities

hDP://www.stdutilitv.com

112

THE THERMODYNAMICS OF PHASE TRANSITIONS

0.9

T
Tc
0.8

0.7

0.6

.----0__

=0.-=-4
_---=O_;_:.8:......__=1.=-2

()-",1=.6,--___:2=.0~------,;2.4

_.J.L..._

Converted with
Fig. 3.S. Exper
plot is of the re

of the liquid c:
equations [6]:

STOI Converter
trial version

f substances. The
Ref. 6.)

y the following

hDP://www.stdutilitv.com
(3.26)

and
PI - Pg = ~

Pc

(1 _ 2:.)
t;

1/3

(3.27)

These equations will be useful later.


It is possible to obtain expressions for response functions in the coexistence
region. As an example, we will consider the molar heat capacity, cv, for a liquid
and vapor coexisting at a fixed molar volume, VD (cf. Fig. 3.7). If we neglect any
effects of gravity, then the system will consist of droplets of liquid in
equilibrium with and floating in vapor. The internal energy per mole of the
liquid at point Dis UI(VB, To) and that of the vapor at point Dis Ug(VA' To) (the
thermodynamic properties of the liquid and the vapor at point D are the same as
on their respective sides of the coexistence curve). The total internal energy at
point Dis
(3.28)

113

PURE PVT SYSTEMS

where Vg = VA, VI = VB, and ng and ni are the number of moles of gas and liquid,
respectively, at point D. The total internal energy per mole at point D is
(3.29)
Let us now look at the variation of the internal energy with temperature along a
line of fixed molar volume at point D (the molar heat capacity at point D),
c - (au
--tot)
aT
V

- x (aug)
-

g aT

VD -

coex

+XI (au
- I)
aT

coex

+ ( UI - U ) (axI)
g

(3.30)

or

coex'

where we have used the fact that dxi = -dxg Equation (3.30) can be expressed
in terms of directly measurable quantities. There are several steps involved
which we itemize below.

Converted with
t B. Similarly,

STOI Converter

where c

(3.32)

trial version

hDP://www.stdutilitv.com

where c
t A.
(ii) Next consider the difference Au = ug - UI between the molar internal
energies of the gas and liquid. From the Clausius-Clapeyron equation
(3.14) and the fundamental equation for the enthalpy, Eq. (2.83), we can
write

dP)
( dT

Ah
coex

Au

A (Pv )

Au

(3.33)

= T A V = T A V + T A V = T Av + T '

where Ah = hg - hi and Av = Vg - VI (AP = 0 because the pressure of


the two coexisting phases are the same). Therefore,

flu = ug - UI = [(T(d~)

coex

-p) (vg - VI)]

(3.34)

coex

(iii) Finally, let us consider the quantity (8xL/ aT) coex: Since the total molar
volume at point D can be written VD = XgVg + XLV/' we can write

Ov
(OvD)
g

=
VD

(VI - vg) (8XI)


8T

+Xg (Bvg)
8T
coex

+XI (8Vl)
8T
coex

.
coex

(3.35)

114

THE THERMODYNAMICS OF PHASE TRANSITIONS

Here we have used the fact that as we vary the temperature along the line
VD =constant, the liquid and vapor vary along their respective sides of
the coexistence curve. We can rewrite Eq. (3.35) in the form
(axI)
aT

=
coex

g)[(av
Xg
(Vg - VI) aT
1

coex

+XI (Bvl)
-aT

coex

1.

(3.36)

We can now combine Eqs. (3.30), (3.32), (3.34), and (3.36) to obtain the
following expression for the heat capacity along the line VD =constant;

Cv = XgC + XICv1+ Xg (aUg)


aVg (avg)
aT coex +XI (aUI)
aVI (avI)
aT
- [T (dP)
-dT coex -P ] [(av
x g g)-aT coex +XI-(avI)
aT coex 1 .
Vg

coex

(3.37)

We now can make two final changes to Eq. (3.37). We can use the identity

Converted with

(3.38)

STOI Converter
and an analog
identity (2.8) t

we can use the

trial version

hDP://www.stdutilitv.com
(3.39)
coex

(ap aT)v!"

and an analogous expression for


I
If Eqs. (3.38) and (3.39) and
the analogous expressions for the liquid are substituted into Eq. (3.37), we
find

All quantities in Eq. (3.40) are measurable, and therefore a numerical value for
the heat capacity can be obtained without much difficulty. Equation (3.40) will
be useful later when we consider critical exponents.
The heat capacity at constant volume is finite in the coexistence region.
However, the heat capacity at constant pressure is infinite in the coexistence
region. If we add heat to a system with coexisting liquid and vapor phases
and keep the pressure fixed, liquid will tum to vapor but the temperature
will not change. Thus, Cp = 00 in the coexistence region, while Cv can remain
finite.
'

115

PURE PVT SYSTEMS

3.D.4. The van der Waals Equation


The van der Waals equation was first derived by van der Waals in his doctoral
dissertation in 1873. It was the first, and to this day is the simplest, equation of
state which exhibits many of the essential features of the liquid-vapor phase
transition. The van der Waals equation is cubic in the molar volume and can be
written in the form

v3 -

RT) v2 +pv-p=O.
a
ab
b+-p

(3.41 )

An isotherm of the van der Waals equation is plotted in Fig. 3.9. For small
values of Tand P, the cubic equation has three distinct real roots (three values of
v) for each value of P and T (this case is shown in Fig. 3.9). As T increases, the
roots coalesce at a critical temperature, Tc, and above T; two of the roots
become imaginary and therefore unphysical. As T ~ 00, Eq. (3.41) reduces to
the ideal gas equation of state, v = RT I P. The critical point is the point at
which the roots of Eq. (3.41) coalesce. It is also the point at which the critical
isotherm (T =
"
an inflection
2
point (8 P I 8v
Converted with
here the curve
r-v

~~a;::t

from C

STDOConverter
trial version

If one uses the


(3.42)

hDP://www.stdutilitv.com

Fig. 3.9. A sketch of a typical van der Waals isotherm. The line from D to F
corresponds to mechanical unstable states. The area, CDE, is labeled 2, and the area
EFG is labeled 1. The area under the curve, V = v(P), between any two points, is equal
to the difference in molar Gibbs free energy between the points.

116

THE THERMODYNAMICS OF PHASE TRANSmONS

at the critical point, then we obtain the following values for the temperature Tc,
pressure Pc, and molar volume vc, at the critical point:
a
P; = 27b2'

Vc

If we introduce reduced variables

Sa

3b,

= P Pc,

Tc

(3.43)

= 27bR'

= T lTc,

and

v = v lv.

then we

may write the van der Waals equation in the form

(- + v3)
P

(3v - 1)

ST.

(3.44)

It is important to note that Eq. (3.44) is independent of a and b. We are now


measuring pressure, volume, and temperature in terms of their distance from the
critical point. The values of vc, Tc, and P; will differ for different gases, but all
gases obey the same equation if they are the same distance from their respective
critical points-that
is, if they have the same values of P = PIPe, t = T tt..
and v = v [v.; Thus, we see again the law of corresponding states.

An unphysi
positive slope,
segment betw
mechanically
of the P- V cur
which we will

Converted with

ST 0 U C oover ter
trial version

hDP:llwww.stdutililV.com

From Eq. (2
molar Gibbs frl'--~........-:,.....,-~

s prediction of
s below Tc (the
to
nphysical parts
11construction,
corresponds

changes in the

......................
~.--------------'

dg = =sd T

+ vdP.

(3.45)

If we now restrict ourselves to one of the van der Waals isotherms so dT = 0,

we can determine how g varies with pressure along that isotherm. In Fig. 3.9 we
plot the molar volume as a function of pressure along a typical van der Waals
isotherm, and in Fig. 3.10 we plot the molar Gibbs free energy as a function of
pressure for the isotherm in Fig. 3.9. Along the isotherm the difference in molar
Gibbs free energy between any two points is equal to the area under the curve,
v = v(P), between those two points:
(3.46)

The Gibbs free energy increases and is concave between A and D. Between D
and F it decreases and is convex. Then between F and I it becomes concave
again and increases. We see that between D and F the states are mechanically
unstable since mechanical stability requires that g be concave (cf. Section
2.H.3). The regions from A to D and from F to I are both mechanically stable

117

PURE PVT SYSTEMS

p
Fig. 3.10. A plot of the molar Gibbs free energy as a function of pressure for the
isotherm in Fig. 3.9.

since g is concave. However, only the curve ACI in Fig. 3.10 corresponds to
states in therm
.
he Gibbs free
ilibrium states
energy is a mi
Converted with
es lying along
thus correspon
etween C and
the curve ACI.
n them, since
G we must dra
going from C
this is the only
trial version
tes) is the line
to G. The physi

STOI Converter
hDP://www.stdutilitv.com

ABCEGHlin F

Before we c
u
,
must decide
where C and G lie. For the points C and G, the molar Gibbs free energies are
equal. Thus

0=

PG

v(P) dP =

JPD

Pc

v(P) dP

Pc

PF

v(P) dP

PE

JPE

v(P)dP

PD

JPG

(3.47)

v(P) dP

PF

or, after rearranging,

~
JPc

v(P)dP -

J~
PE

v(P) dP =

J~

v(P) dP -

PF

J% v(P) dP.

(3.48)

PF

The left-hand side is equal to area 2 in Fig. 3.9 and the right-hand side is equal
to area 1. Thus, the line from C to G must be drawn so that the areas 1 and 2 are
equal:
Area 1 = Area 2.

(3.49)

118

THE THERMODYNAMICS OF PHASE TRANSITIONS

If this is done, the curve ACEGI then gives the equilibrium states of the system.
The condition given in Eq. (3.49) is called the Maxwell construction. Thus, with
the Maxwell construction, we obtain the equilibrium isotherms from the van der
Waals equation and the curves for metastable states.

3.E. SUPERCONDUCTORS

[8-10]

Superconductivity was first observed in 1911 by Kamerlingh annes. He found


that the resistance to current flow in mercury drops to zero at about 4.2 K (cf.
Fig. 3.11). At first this was interpreted as a transition to a state with infinite
conductivity. However, infinite conductivity imposes certain conditions on the
magnetic field which were not subsequently observed. The relation between the
electric current, J, and the applied electric field, E, in a metal is given by Ohm's
law,

(3.50)

J=aE,
where a is the
B by Faraday'

Converted with

STOI Converter
trial version

magnetic field

(3.51 )

hDP:llwww.stdutililV.com

If we substitut
te conductivity
a ~ 00 and
of the system
depends on its history. If we first cool the sample below the transition
temperature and then apply an external magnetic field, H, surface currents must
be created in the sample to keep any field from entering the sample, since B
must remain zero inside (cf. Fig. 3.12). However, if we place the sample in the
H-field before cooling, a B-field is created inside. Then, if we cool the sample,
the B-field must stay inside. Thus, the final states depend on how we prepare the
sample. With the hypothesis of infinite conductivity, the state below the

R (ohms)
0.15
0.10
0.05

T K

L--~-'"II::'___'_-"""___'_-~

4.22

4.26

4.30

Fig. 3.11. The resistance of mercury


drops to zero at about 4.2 K. (Based on
Ref. 10.)

119

SUPERCONDUCTORS
perfect diamagnetism

infinite conductivity

o ~

o ~
first cool,

then apply field

first cool,

then apply field

t
I

first apply field,


Fig. 3.12. A su
or perfect diama

transition tern

then cool

first apply field,

Converted with

then cool
erfect conductor

STOI Converter
trial version

hDP://www.stdutilitv.com

it depends on

history.
In 1933, M
tal of tin in a
magnetic field and found that the field inside the sample was expelled below the
transition point for tin. This is contrary to what is expected if the transition is to
a state with infinite conductivity; it instead implies a transition to a state of
perfect diamagnetism, B = O. It is now known that superconductors are perfect
diamagnets. When superconducting metals are cooled below their transition
point in the presence of a magnetic field, currents are set up on the surface of
the sample in such a way that the magnetic fields created by the currents cancel
any magnetic fields initially inside the medium. Thus B = 0 inside a
superconducting sample regardless of the history of its preparation.
No electric field is necessary to cause a current to flow in a superconductor.
A magnetic field is sufficient. In a normal conductor, an electric field causes
electrons to move at a constant average velocity because interaction with lattice
impurities acts as a friction which removes energy from the electron current. In
a superconductor, an electric field accelerates part of the electrons in the metal.
No significant frictional effects act to slow them down. This behavior is
reminiscent of the frictionless superftow observed in liquid He4 below 2.19 K
(cf. Section S3.B). Indeed, the superftuid flow in He4 and the supercurrents in
superconductors are related phenomena. The origin of the apparently
frictionless flow in both cases lies in quantum mechanics. It is now known

120

THE THERMODYNAMICS OF PHASE TRANSITIONS

that electrons in a superconducting metal can experience an effective attractive


interaction due to interaction with lattice phonons. Because of this attraction, a
fraction of the electrons (we never know which ones) can form "bound pairs."
The state of minimum free energy is the one in which the bound pairs all have
ths same quantum numbers. Thus, the bound pairs form a single macroscopically occupied quantum state which acts coherently and forms the
condensed phase. As we shall see, the order parameter of the condensed phase
behaves like an effective wave function of the pairs. Because the pairs in the
condensed phase act coherently (as one state), any friction effects due to lattice
impurities must act on the entire phase (which will contain many pairs and have
a large mass) and not on a single pair. Thus, when an electric field is applied,
the condensed phase moves as a whole and is not slowed significantly by
frictional effects.
The condensed phase flows at a steady rate when a magnetic field is present,
and it is accelerated when an electric field is applied. Just below the transition
temperature, only a small fraction of electrons in the metal are condensed and
participate in superflow. As temperature is lowered, thermal effects which tend
to destroy the
.
1 ger fraction of
the electrons c
If a superco
netic field, the

STOI Converter

superconducti
tion, B, versus
applied field,
ith a value less
than some te
trial version
e system is a
perfect diama
refore B = O.
However, for
B = J-LH. (For
normal metals m~:zo;-wrleI1~~SL1~petlIR~nnnJr1ne-Vc[ictIum.)
Thus,

hDP://www.stdutilitv.com
B=

{OJ-LoH

if H < Hcoex(T)
if H > Hcoex(T)

(3.52)

----/
~------~--------~H
Hcoex

Fig. 3.13. A plot of the magnetic induction, B, versus the applied magnetic field, H, in a
superconductor.

121

SUPERCONDUCTORS

~~~--~~-----4TK

t;

Fig. 3.14. The coexistence curve for normal and superconducting states.

The field, Hcoex (T) lies on the coexistence curve for the two phases. It has been
measured as a function of the temperature and has roughly the same behaviour
for most metals cf. Pi . 3.14 . The coexistence curve for the normal and
superconductin

Converted with

STOI Converter

(3.53)

trial version

where T; is th
present. The
slope (dH/dT)
c- The phase
diagram for a up
n uc in me a
as ana ogles 0
e vapor-liquid
transition in a PVT system, if we let Hcoex replace the specific volume. Inside
the coexistence curve, condensate begins to appear.
Along the coexistence curve, the chemical potentials of the superconducting
and normal phases must be equal and, therefore, any changes in the chemical
potentials must be equal. Thus, along the coexistence curve

hDP://www.stdutilitv.com

(3.54)
or

Sn -

s,

= -J..LoHcoex(T) (~~)

(3.55)

coex

Equation (3.55) is the Clausius-Clapeyron


equation for superconductors.
We
have used the fact that B, = 0 and Bn = J-loHcoex(T) on the coexistence curve.
Here sn(s) is the entropy per unit volume of the normal (superconducting) phase.
We see that the transition has a latent heat (is first order) for all temperatures
except T = T; where Hcoex = O. When no external magnetic fields are present,
the transition is continuous.

122

THE THERMODYNAMICS OF PHASE TRANSITIONS

The change in the heat capacity per unit volume at the transition is
3

_
H5 (T
3T )
( Cn - Cs) coex -_ [ T 8(sn 8- ss)]
- 2/-Lo- - -3
T
coex
t; t; t;

(3.56)

We have used Eq. (3.53) to evaluate the derivatives (dH/dT)coex.


At low
temperatures the heat capacity of the normal phase is higher than that of the
superconducting phase. At T = T; (the critical point) the heat capacity is higher
in the superconductor and has a finite jump, (c, - Cn)r=Tc = (4/-tO/Tc)H5. It is
worthwhile noting that as T ~ 0, (s, - sn) ~ 0 since (dH/dT)coex ~ 0 as
T ~ O. This is in agreement with the third law of thermodynamics.
It is useful to obtain the difference between the Gibbs free energies of the
normal and superconducting phases for H = O. The differential of the Gibbs
free energy per unit volume, g, is
dg = -sdT

- BdH.

(3.57)

If we integrate EQ. (3.57) at a fixed temperature. we can write

Converted with
For the norma

STOI Converter

(3.58)

trial version

hDP://www.stdutilitv.com
For the superconducting phase (H

(3.59)

< Hcoex), we have


(3.60)

since B = 0 inside the superconductor. If we next use the fact that


(3.61)
we obtain the desired result
(3.62)
Thus, the Gibbs free energy per unit volume of the superconductor in the
absence of a magnetic field is lower than that of the normal phase by a
factor J-lo/2H;oex(T) (the so-called condensation energy) at a temperature T.
Since the Gibbs free energy must be a minimum for fixed Hand T, the
condensed phase is a physically realized state. Note that gs(Tc, 0) = gn(Tc, 0) as
it should.

123

THE HELIUM LIQUIDS

3.F. THE HELIUM LIQUIDS


From the standpoint of statistical physics, helium has proven to be one of the
most unique and interesting elements in nature. Because of its small atomic
mass and weak attractive interaction, helium remains in the liquid state for a
wide range of pressures and for temperatures down to the lowest measured
values.
The helium atom occurs in nature in two stable isotopic forms, He3 and He4
He3, with nuclear spin (1/2), obeys Fermi-Dirac statistics; while He4, with
nuclear spin 0, obeys Bose-Einstein statistics. At very low temperatures, where
quantum effects become important, He3 and He4 provide two of the few
examples in nature of quantum liquids.
Chemically, He3 and He4 are virtually identical. The only difference between
them is a difference in mass. However, at low temperatures the two systems
exhibit very different behavior due to the difference in their statistics. Liquid
He4, which is a boson liquid, exhibits a rather straightforward transition to a
superfluid state at 2.19 K. This can be understood as a condensation of particles
into a single
..
3 al 0
n er oes a transition to a
superfluid stat
10-3 K). The
mechanism fo
Converted with
different from

STDI Converter

that of liquid H
particles) form
bound pairs w
m, I = 1. The
mechanism is
trial version
uperconductor
except that the
0 and angular
momentum I =
spherical with
no magnetic m
one axis, carry
angular momentum, and have a net magnetic moment. The fact that the bound
pairs in liquid He3 have structure leads to many fascinating effects never before
observed in any other physical system.
When He3 and He4 are combined to form a binary mixture, a new type of
phase point occurs (called a tricritical point) in which a ,x-line connects to the
critical point of a binary phase transition.
While we cannot discuss the theory of these systems at this point, it is worthWhileto look at their phase diagrams since they present such a contrast to those
of classical fluids and they tend to confirm the third law.

hDP://www.stdutilitv.com

3.F.1. Liquid He4 [12-14]


4

He was first liquefied in 1908 by Kamerlingh Onnes at a temperature of


4.215 K at a pressure of I atm. Unlike the classical liquids we described in
Section 3.0, it has two triple points. The coexistence curves for liquid He4 are
shown in Fig. 3.15 (compare them with the coexistence curve for a classical
liqUidin Fig. 3.4). He4 at low temperature has four phases. The solid phase only
appears for pressures above 25 atm, and the transition between the liquid and

124

THE THERMODYNAMICS OF PHASE TRANSITIONS


P (atm)
40

30
20
10

Fig. 3.15. The coexistence curves for He4

solid phases is first order. The liquid phase continues down to temperatures
approaching T =
phases. As the
normal liquid
Converted with
ccurs at about
T = 2 K (the
dicating that a

STOI Converter

continuous sy
line ..There is a
triple point at
roken is gauge
symmetry. Bel
trial version
liquid He(U) by
Keesom and
erties. The first
experimenters
e to leak out of
their container
ot leak through.
This apparently frictionless flow is a consequence of the fact that the condensed
phase is a highly coherent macroscopic quantum state. It is analogous to the
apparent frictionless flow of the condensed phase in superconductors. The order
parameter for the condensed phase in liquid He4 is a macroscopic "wave
function," and the Ginzburg-Landau theory for the condensed phase of He4 is
very similar to that of the condensed phase in superconductors, except that in
liquid He4 the particles are not charged. The specific heat of liquid He4 along
the A-line is shown in Fig. 3.16. We can see that it has the lambda shape
characteristic of a continuous phase transition.
The phase diagram of He4 provides a good example of the third law. The
vapor-liquid and solid-liquid coexistence curves approach the P-axis with zero
slope. From Eqs. (2.54) and (2.100), we see that this is a consequence of the
third law.

hDP:llwww.stdutililV.com

3.F.2. Liquid He3 [16-18]


The He3 atom is the rarer of the two helium isotopes. Its relative abundance in
natural helium gas is one part in a million. Therefore, in order to obtain it in
large quantities, it must be "grown" artificially from tritium solutions through

125

THE HELIUM LIQUIDS

3.0.----------:------,

C,)

1.0

0.0 .__--L----L_......__--'-_'----L-----L._.....___.
1.6
2.0
2.4
2.8
T (K)
Fig. 3.16. The specific heat of He4 at vapor pressure at the ,x-point [12].

,a-decay of the tritium atom. Thus, He3 was not obtainable in large enough
quantities to
fled in 1948 by
Sydoriack, G
Converted with
only (3/4) the

=t~~:3H:
pressure abou
The phase

SIDU Converter
trial version

::t:ta!~:.ri~:
:
e") is given in

hDP:llwww.stdutililV.com

Fig. 3.17. On
superfluid state.
There is, ho
,
curve. This is
attributed to the spin of the He3 atom. At low temperature the spin lattice of the
He3 solid has a higher entropy than the liquid. The entropy difference,

P (atm)

40

30
20

liquid

10

012
Fig. 3.17. Coexistence curves for

4
He3

[16].

T (K)

126

THE THERMODYNAMICS OF PHASE TRANSITIONS

P (atm)
solid

40

30

0.001

0.002

0.003

T (K)

Fig. 3.IS. Coexistence curves for superfluid phases of He3 when no magnetic field is
applied [16].

= Sliquid -

hes at about
e differences
ion dP/dT =
gative slope at
sfied, the slope
trial ve rsion
OK.
by Osheroff,
Superfluidit
Richardson, an
,
prize for this
work. The transition occurs at 2.7 x 10-3 K at a pressure of about 34 atm. The
phase diagram for a small temperature interval is shown in Fig. 3.18. There are,
in fact, several superfluid phases in liquid He3, depending on how the bound
pairs orient themselves. The so-called A-phase is an anisotropic phase. The
bound pairs (which have shape) all orient on the average in the same direction.
This defines a unique axis in the fluid and all macroscopic properties depend on
their orientation with respect to that axis. The B-phase is a more isotropic phase
and has many features in common with the superfluid phase of a superconductor. If we apply a magnetic field to liquid He3, a third superfluid phase
appears. The transition between the normal and superfluid phases appears to be
continuous, while that between the A and B superfluid phases appears to be first
order.
tiS

T = O.3K

and
remain virtual
ti V leads
low temperatur
of the liquid-s

ss/

Converted with

STOU Converter
hUP:llwww.stdutililV.com

3.F.3. Liquid He3 _He4 Mixtures


When He3 and He4 are mixed together and condensed to the liquid state, some
interesting phenomena occur. We will let X3 denote the mole fraction of He3. In
1949, Abraham, Weinstock, and Osborne [21] showed that He3-He4 mixtures
can undergo a transition to a superfluid state. In this early experiment, they

127

THE HELIUM LIQUIDS

T (K)
1.5

1.0

0.5
coexistence region

0.00

0.25

0.75

0.50

1.00

Z3

Fig. 3.19. The phase diagram for a liquid He3_He4 mixture plotted as a function of
temperature T a
.,

Converted with

STOI Converter

eout T = 1.56K
found that the
trial version
T = 0.87K for
for X3 = 0.282
X3 = 0.67 (cf.
In 1954, Pr.., ..,
p~ WWW.SUII
e existence of a
phase separation of liquid He3_He4 mixtures into an He3-rich phase and an He4rich phase. (The theory of binary phase separation in classical mixtures is
discussed in Sect. (S3.E).) This phase separation was found in 1956 by Walters
and Fairbank [23] using nuclear magnetic resonance techniques. They were
able to measure the magnetic susceptibility of the liquid in different vertical
layers of the mixture. In the coexistence region the magnetic susceptibility,
which is a measure of the concentration of He3 atoms varies with height. The
critical point for this binary phase transition lies at the end of the A-line at
T = 0.87 K and X3 = 0.67. The phase transition along the A-line is second
order. The binary phase separation is a first-order phase transition. In 1982, the
region of metastable states for this first-order phase transition was measured by
Alpern, Benda, and Leiderer [24].
The end point of the A-line is the critical point of the first-order binary
phase transition. It was first called a tricritical point by Griffiths [25]. He
pointed out that in a suitable space it is meeting point of three lines of
second-order phase transitions. One line is the A-line. The other two are lines of
critical points associated with the first-order phase transition. Thus, the
tricritical point is different from the triple points we have seen in classical
fluids.

hn .//

Id I-I-tv.~om

128

THE THERMODYNAMICS OF PHASE TRANSITIONS

3.G. GINZBURG-LANDAU THEORY [26]


In the late 1930s, Ginzburg and Landau proposed a mean field theory of
continuous phase transitions which relates the order parameter to the underlying
symmetries of the system. One of the features which distinguishes first-order
and continuous phase transitions is the behavior of the order parameter at the
transition point. In a first-order phase transition, the order parameter changes
discontinuously as one crosses the coexistence curve (except at the critical
point). Also, first-order phase transitions mayor may not involve the breaking
of a symmetry of the system. For example, in the liquid-solid and vapor-solid
transitions, the translational symmetry of the high-temperature phase (liquid or
vapor) is broken, but for the vapor-liquid transition no symmetry of the system
is broken. In the liquid and gas phases, the average particle density is
independent of position and therefore is invariant under all elements of the
translation group. The solid phase, however, has a periodic average density and
is translationally invariant only with respect to a subgroup of the translation
group. Solids may also exhibit first-order phase transitions in which the lattice
structure undergoes a sudden rearrangement from one symmetry to another and
the state of the
At a first-ord
Converted with
curve changes
discontinuously
mayor may no
free energy cur
such transitions

STOI Converter
trial version

hDP://www.stdutilitv.com

hd a symmetry
~e slope of the
ays broken. In
ter) appears in

the less symme


ar, a vector, a
tensor, a comp _
~ ~ ,_ ~
..
'". _
of the order
parameter is determined by the type of symmetry that is broken (an extensive
discussion of this point may be found in Ref. 26). For example, in the transition
from a paramagnetic to a ferromagnetic system, rotational symmetry is broken
because a spontaneous magnetization occurs which defines a unique direction in
space. The order parameter is a vector. In the transition from normal liquid He 4
to superfiuid liquid He4, gauge symmetry is broken. The order parameter is a
complex scalar. In a solid, the lattice might begin to undergo a gradual
reorientation as the temperature is lowered. The order parameter is the change
in the spatially varying number density. In continuous transitions, one phase
will always have a lower symmetry than the other. Usually the lower
temperature phase is less symmetric, but this need not always be the case.
All transitions which involve a broken symmetry and a continuous change in
the slope of the free energy curve can be described within the framework of a
mean field theory due to Ginzburg and Landau [26]. Ginzburg-Landau theory
does not describe all features of continuous phase transitions correctly, but it
does give us a good starting point for understanding such transitions. Near a
continuous phase transition the free energy is an analytic function of the order
parameter associated with the less symmetric phase. In real systems, the free
energy is not an analytic function of the order parameter near the critical

129

GINZBURG-LANDAU THEORY

point. Nevertheless, Ginzburg-Landau theory still describes enough qualitative


features of continuous phase transitions that it is well worth studying.
We will let 'T}denote the order parameter and let <p denote the free energy and
write
<p(T, Y,J)

= <Po(T, Y)

+ a2(T,

Y)'T}2

+ a3(T,

Y)'T}3

+ a4(T,

Y)'T}4

+ ... - /'T},
(3.63)

where Y is a generalized thermodynamic force and / is the force conjugate to the


order parameter (one can usually be found, even though it may only be a
mathematical construct and not realizable in the laboratory). There is no term in
Eq. (3.63) which is first order in 'T} because that would ensure a nonzero value
for the order parameter above the transition point. The molar free energy,
<Po(T, Y), describes the thermodynamic quantities not directly involved in the
transition and generally will depend on other state variables. For example, if we
are studying the transition from the paramagnetic to the ferromagnetic state of a
magnetic crystal, free energy <Po(T, Y) would contain all the nonmagnetic
contributions to the thermodynamic state of the crystal.
The actual v
.
in nature is the
one that gives
Converted with
tee energy as a

STOI Converter

function oi t; fc
df,itiseasyto
see under what
ccur. The types
of terms which
the symmetry
properties of a
trial version
quantity. If the
order paramete
hibinations can
appear in the
for a magnetic
crystal is the magnetization. The magnetization, which is a vector, changes sign
under time reversal. However, the free energy must be invariant under time
reversal, and therefore a cubic term in the magnetization is ruled out. There are
a number of different types of phase transition which can be described by the
free energy in Eq. (3.63). We shall now consider several of these.

hDP://www.stdutilitv.com

3.G.1. Continuous Phase Transitions


A continuous phase transition can occur if the cubic term in Eq. (3.63) cannot
appear and if no external field is present. The free energy then takes the form
(3.64)
The dependence of a2 (T, Y) on temperature is chosen so that for temperatures
above and at the critical temperature the free energy will only be minimized for
'T] = 0, while below the critical temperature it will be minimized for I'T}I > O.
In general, the free energy will be minimum if
(~<P)
'T}

=0
T,Y

and

(~2~) ~
TJ

T,Y

O.

(3.65)

130

THE THERMODYNAMICS OF PHASE TRANS mONS

Equations (3.65) give us conditions that must be satisfied for equilibrium states.
The right-hand equation in (3.65) is the condition for stability (not all stable
states are equilibrium states). In order to have global stability, we must have
CX4(T,

Y) > O.

(3.66)

This will ensure that as we increase TJ to very large values, the free energy will
continue to increase. The critical point occurs when cx2(T, Y) = O.This happens
at a temperature T = T; (Y). If the critical temperature is a function of another
variable, Y, then there will be a line of critical points in the (T, Y) plane. If we
choose cx2(T, Y) > 0 for T > Tc(Y) and cx2(T, Y) < 0 for T < Tc(Y), then the
free energy, <P, will have its minimum value at TJ = 0 when T > Tc(Y) and will
have its minimum value for TJ f. 0 when T < Tc(Y). Since the free energy must
vary continuously through the transition point, at T = T; (Y) we must have
cx2(Tc, Y) = O.We can combine all this information if we write cx2(T, Y), in the
neighborhood of the transition point, in the form

cx2(T, Y) = cxo(T, Y)(T - Tc(Y)),


where CXo is a s
In Fig. 3.20,
the free energy
The free energy
the free energy
one of these tw
point. The regie
region of unstat.;

(3.67)

Converted with

. In curve

STOI Converter

critical point.
. In curve (C),
ndomly select
pw the critical

trial version

hDP:llwww.stdutililV.com
--.:::r.r

a
(8)
TJ

= 2CX2TJ

+ 4CX4TJ 3 = 0

(A),

rresponds to a
---'

(3.68)

T,Y

Fig. 3.20. The behavior of the free energy, <P = a2rl + a4'T}4, for a continuous phase
transition for a4 = 4.0 and (A) a2 = 0.6, (B) a2 = 0.0, and (C) a2 = -0.6. In the
figure,

nc

= 2.7386.

131

GINZBURG-LANDAU THEORY

and therefore when


I)

or

I)

.f{!!!;

2':. tt; - T).

(3.69)

When T> Te, the minimum occurs for 7] = O. When T < Te, the minimum
occurs for 7] = y'(ao/2a4)(Te
- T). Thus, below the critical temperature, the
order parameter is nonzero and increases as y'Te - T. From the above
discussion, the free energy takes the following form:
for T > Te,
2
a (Te - T)2
= o(T, Y) 0 4
a4

(T, Y,

7]) =

(T, Y,

7])

o(T, Y)

for

T < Te,

(3.70)

where we have suppressed the dependence of Te on Yand the dependence of ao


and a4 on T and y.
The molar heat
capacity is
~~--~----------------------~

Converted with

STOI Converter
If we neglect de
temperature), w
critical point:

trial version

(3.71 )
~ slowly with
jump at the

hDP://www.stdutilitv.com
(3.72)

The jump in the heat capacity has the shape of a A, as shown in Fig. 3.21, and
therefore the critical point for a continuous phase transition is sometimes called
a A-point.
The transition from normal to superfluid in liquid He4 is an example of a
continuous phase transition. The order parameter, 7], is the macroscopic wave
function, W, for the condensed phase and the generalized force, Y, is just the
pressure, P (the force P is not conjugate to W). The free energy can be written
(T, P, w) = o(T, P)

+ a21wI2 + a41wI4 + ... ,

(3.73)

where a2(T, P) = ao(T, P)(T - Te) and ao(T, P) and a4(T, P) are slowly
varying functions of T and P. The order arameter, W = 0, above the critical
temperature, and W = eiO (ao/2a4)(Te - T) below the critical temperature.
The phase factor, 0, can be choosen to be zero as long as no currents flow in the
system. As we see in Fig. 3.15 there is in fact a line of continuous transition
points in the (P, T) plane. In Fig. 3.16 we showed the behavior of the heat

132

THE THERMODYNAMICS OF PHASE TRANSmONS

r----TcQ~
2Q4

~_
I

I
I
~----------------~T
t;
Fig. 3.21. The jump in the heat capacity at the critical point (lambda point) as predicted
by Landau theory.

capacity as we passed through the line of critical points.


figure, there is a finite lambda-sha
ed 'um in th h
If we tum
the continuous
the free energ

Converted with

STOI Converter
trial version

As we see from the


. of the liquid.
arameter, then
external force,

- /1],

(3.74)

hDP://www.stdutilitv.com

where a2 = a2
nonzero for all
temperatures.
4---."TTTT"""'-F"""""TT""""""''''''-':r--orIm:aT1rru---.-rr-Tl'TPr-nT'J''ll"l:!"FrrrT'"P""""T1iT""""':1--.n
is shown in
Fig. 3.22 for the same parameters
as Fig. 3.20.

Fig. 3.22. The behavior of the free energy, = o.2'f/2 + o.4'f/4 - f'f/, for a continuous
phase transition for 0.4 = 4.0, f = 0.06, and (A) 0.2 = 0.6, (B) 0.2 = 0.0, and (C)
0.2 = -0.6.
'f/A, 'f/B, and ttc locate the minima of the curves. In the figure 'f/A =
0.0485, ns = 0.1554, and nc = 0.2961.

GINZBURG-LANDAU

133

THEORY

From Eq. (3.74), we can obtain the susceptibility, X = (8",/8f)r,y


The equilibrium state is a solution of the equation

-(&q/ /8f2)r,y.

(3.75)
If we take the derivative ofEq. (3.75) with respect tofand solve for (8",/8f)r,y,
we obtain
(3.76)

In the limit, f ~ 0, ", = 0 for T > T; and ", = -a2/2a4 for T < T':
Therefore, in the limit f ~ 0 the susceptibility will be different above and
below the critical point. We find

Converted with
X = li

f--+

(3.77)

STOI Converter

Note that the s


trial version
The transitio
is one of the
simplest examp
hich exhibits
this behaviour is a magne IC so I ,suc as me e, w ose a Ice sites contain
atoms with a magnetic moment. The critical temperature is called the Curie
temperature. Above the Curie temperature, the magnetic moments are oriented
at random and there is no net magnetization. However, as the temperature is
lowered, magnetic interaction energy between lattice sites becomes more
important than randomizing thermal energy. Below the Curie temperature, the
magnetic moments became ordered on the average and a spontaneous
magnetization appears. The symmetry that is broken at the Curie point is
rotation symmetry. Above the Curie point, the paramagnetic system is
rotationally invariant, while below it the spontaneous magnetization selects a
preferred direction in space. The order parameter for this continuous phase
transition is the magnetization, M. The magnetization is a vector and changes
sign under time reversal. The free energy is a scalar and does not change sign
under time reversal. If a magnetic field, H, is applied to the system, the
Ginzburg-Landau free energy can be written in the form

hDP:llwww.stdutililV.com

(T, H)

= <Po(T)

- M .H

+ a2M . M + a4(M . M)2 + ... ,

(3.78)

where the coefficients a2 and a4 have the same properties as described above.
For the case when the applied field H equals 0, the magnetization M equals 0

134

THE THERMODYNAMICS OF PHASE TRANSITIONS

"......_

~6.3
"'t-:!

'-'"

0
.-4

4.2

C,)

Ie ..
I

I
I

I
I

2.1
173

373

573

773

T(K)

Fig. 3.23. The specific heat of nickel in the neighborhood of the Curie point. The
dashed line gives the Curie point [34]. (Based on Ref. 34.)

above the Curie temperature and M = J(<lO/2<l4)(Tc


- T))M below the
Curie temperature. M is a unit vector which gives the direction of the
magnetization vector. The actual direction of the magnetization vector, if
H = 0, will b
rand
fl ctuations or outside influences. The
heat capacity
ed peak. As an
example, the

STOU Converter
trial version

If the order p
de a third-order
term in the m
a second-order
tensor, there
sors to yield a
scalar and again the free energy can have a third-order term. Such systems
cannot exhibit a continuous transition. To see why, let us write the free energy
in the form

hUP:llwww.stdutililV.com

(3.79)

where <l2 = <l2(T, Y), <l3 = <l3(T, Y), and <l4 = <l4(T, Y).
The extrema of the free energy are given by the equation (8cfJ/8",)TY = 0,
which has solutions", = and", = ((-3<l3 J9<l~ - 32<l2<l4)/8<l4). As long
as 9<l~ - 32<l2<l4 < 0, the only minimum of the free energy occurs at '" =
because other values of '" will be complex and therefore unphysical. When
9<l~ - 32<l2<l4 > 0, two minima and one maximum can exist. A plot of the free
energy for <l2 > 0, <l4 > 0, and a range of values of <l3 is given in Fig. 3.24. For
curves A, B, and C the state of minimum free energy (the equilibrium state)
occurs for", = 0. Curve D shows the behavior of the free energy at the critical
temperature for this system. At this point the equilibrium state shifts from one
with order parameter", = 0 to one with order parameter", =
Therefore, this
is a discontinuous transition and is of the type one expects for a first-order
transition. The transition point for the first-order transition (curve D in Fig.

"'D.

135

CRITICAL EXPONENTS

Fig. 3.24. The behavior of the free energy, = a2'f/2 + a3'f/3 + a4'f/4, for a2 = 2.0,
a4 = 4.0, and (A) a3 = -4.5, (B) a3 = -5.333, (C) a3 = -5.5, (D) a3 = -5.6568,
and (E) a3 = -5.85. In the figure, ne = 0.7738.

3.24) is easil

(8cp/8"')r

=(
Therefore: the f
Y

before any cont


it occurs for a p
a negative valu

Converted with

STOI Converter
trial version

CPo = 0 and
= aV4a4.
efore it occurs
f a3 < 0, then
en it occurs for
f-

a2

hDP://www.stdutilitv.com
3.H. CRITICAL EXPONENTS [27-29]
The critical point is the point at which the order parameter of a new phase
begins to grow continuously from zero. We have seen many examples of critical
points. In the liquid-vapor transition, a critical point terminated the liquidvapor coexistence curve and was the one point for which the free energy slope
changed continuously. For the binary mixture, the critical point marked the
temperature at which phase separation could first take place as we lowered the
temperature. In the spin system the critical point for transition from a
paramagnetic state to ferromagnetic state was the Curie point. In the
superconductor, the critical point was the point at which the condensate first
appeared as we lowered the temperature in the absence of an external field. In
liquid He4, the condensed and normal phases were separated by a line of critical
points in the (P,T) plane.
Systems exhibit dramatically new behavior below the critical point. As we
approach the critical point from above (higher temperature), the system
anticipates its new behavior by making "adjustments" on a microscopic scale.
These "adjustments" appear in the form of fluctuations in density, magnetization, and so on, which become very large as the critical point is approached. Just

136

THE THERMODYNAMICS OF PHASE TRANSITIONS

below the critical point, the order parameter of the new phase first becomes
nonzero.

3.B.l. Definition of Critical Exponents


A subject of great interest is the way in which various systems approach the
critical point. For this purpose it is useful to introduce the idea of critical
exponents. As one approaches the critical point, various thermodynamic
functions may diverge or go to zero or even remain finite. It is therefore
convenient to introduce an expansion parameter
(3.80)
where T; is the critical temperature, which is a measure of the distance from the
critical point in terms of reduced variables.
Near the critical point all thermodynamic functions can be written in the
form

Converted with
where y

> O. "

STOI Converter

(3.81)
ned

trial version

hDP://www.stdutilitv.com

(3.82)

f( e) diverges at the critical point. If A is positive, f( e) goes


to zero at the critical point. In Fig. 3.25 we have plotted f(e) = eA
for A = -1, -3/2,1/2,
and 114. The case where A = 0 may correspond to
several different possibilities; for example, it may correspond to a logarithmic
divergence f(e) = Allnel + B or to a dependence on e of the form
f( e) = A + &1/2. For such cases a modified exponent is introduced. If j is
the smallest integer, such that dj f( e) / de! = f(j) (e) diverges, then
If A is negative,

A' = j

+ lim
e--+O

U)

In If (e)l.
In e

(3.83)

In Fig. 3.26 we have plottedf(e) as a function of e for some cases when A = O.


Although we have chosen to write e in terms of temperature, it is also of interest
to introduce exponents for the approach to the critical point involving quantities
other than temperature, such as pressure, density, magnetic field, and so on.
Thus, there are a number of different critical exponents that can be defined for a
system depending on how the critical point is approached. We shall give some
examples in the next section.

137

CRITICAL EXPONENTS

/()=>.

/()= >.

(a)

(b)

0.7

500
0.5
300
0.3
0.1
0

0.05

0.10

0.15

0.05

Fig. 3.25. Plots ofj(c) = c'>"for cases when A#- 0 (a) Plots for A
(b) Plots for A = and A =

!.

t
9 IL

Converted with

ir

STOI Converter

I~

IL
"1 I~

= -1 and A = -~.

(b)

trial version

I'

~~

0.15

>.'

/(e)=l-

f()=lln()1

0.10

L-

0.05

hDP://www.stdutilitv.com
~~
__~~------~--------~------~
0.10

Fig. 3.26. Plots ofj(c)


of j (c) = 1 - eX.

0.15

0.05

= c'>"for cases when A = 0, (a) Plot ofj(c) =

0.10
[In (c)

0.15

I. (b) Plot

3.H.2. The Critical Exponents for Pure PVT Systems


There are four critical exponents which are commonly used to describe the bulk
thermodynamic properties of PVT systems. We define them and give their
experimental values below.
(a) Degree of the Critical Isotherm. The deviation of the pressure (P - Pc)
from its critical value varies at least as the fourth power of (V - Vc) as the
critical point is approached along the critical isotherm. It is convenient to
express this fact by introducing a critical exponent, 8, such that
P_ P
!p _ Perc
To
= ACli_-li
sign(p-pc),
Pc
c

(3.84)

138

THE THERMODYNAMICS OF PHASE TRANS mONS

where P; is the critical pressure, Pc is the critical density, Ab is a constant, and


P cO is the pressure of an ideal gas at the critical density and temperature.
Experimentally it is found that 6 > 8 ~ 4. The exponent 8 is called the degree
of the critical isotherm.
(b) Degree of the Coexistence Curve. Guggenheim [6] has shown that the
deviation (T - Tc) varies approximately as the third power of (V - Vc) as the
critical point is approached along the coexistence curve from either direction.
One expresses this fact by introducing a critical exponent {3, such that
(3.85)
where PI is the density of liquid at temperature T < Te, Pe is the density of gas
at temperature T < Tc, each evaluated on the coexistence curve, and A,a is a
constant. The quantity PI - Ps is the order parameter of system. It is zero above
the critical point and nonzero below it. The exponent {3 is called the degree of
the coexistence curve and is found from experiment to have values {3 ~ 0.34.

Converted with

(c) Heat Cap


logarithmic di
critical expone

ears to have a
(V = Vc). The
as follows:

STOI Converter
trial version

(3.86)

hDP://www.stdutilitv.com
(d) Isothermal Compressibility. The isothermal
approximately as a simple pole:
T < Te,
T> Tc,

a are found experi-

compressibility

P = PI(T) or pg(T)
P = Pc,

diverges

(3.87)

where A~ and Ai' are constants. For T < T; one approaches the critical point
along the coexistence curve; for T > T; one approaches it along the critical
isochore. Typical experimental values of l' and I are I' ~ 1.2 and I ~ 1.3.
(e) Exponent Inequalities. It is possible to obtain inequalities between the
critical exponents using thermodynamic arguments. We shall give an example
here. Equation (3.40) can be rewritten in terms of the mass density as
Cv = XgC

Vg

xgT
+ XIC + fJgKT
g
V1

,,3

(8Pg)
aT

xlT
+
-3-1
coex
PI KT
2

(OPI)
8T

2
eoex

'

(3.88)

139

CRITICAL EXPONENTS

where Cv, C and C are now specific heats (heat capacity per kilogram), and r
is the isothermal compressibility. All terms on the right-hand side of Eq. (3.88)
are positive. Thus, we can write
Vg

V1

(3.89)
As the critical point is approached for fixed volume, Xg --+ (1/2), Ps --+ Pc (Pc
is the critical density), r diverges as tt; - T)-'Y' [cf. Eq. (3.87)], and
(8pg/8T)coex
diverges as tt, - T)f3-1 if we assume that [(1/2)(PI + pg) - Pc]
goes to zero more slowly than (PI - pg) [cf. Eqs. (3.26) and (3.27)]. Thus,
1 Te B(Tc - T)'Y'+2f3-2

cv~

'2

(3.90)

Pc

where B is a constant, and


(3.91 )

Incv~
If we next divid

Converted with

STOI Converter
The inequality i

trial version

d
(3.92)
If we choose

hDP:llwww.stdutililV.com

0/ = 0.1, {3 = (1
tion (3.92) is
called the Rushf.hT....,..,..,,,....,,.,...,,......,....,......-yo---------------'
The critical exponents can be computed fairly easily starting from mean field
theories such as the van der Waals equation (cf. Exercise (3.4 or GinzburgLandau theory. All mean field theories give similar results. The common feature
of these theories is that they can be derived assuming that the particles move in
a mean field due to all other particles. The mean field theories do not properly
take into account the effects of short-ranged correlations at the critical point and
do not give the correct results for the critical exponents. We shall return to this
point when we discuss Wilson renormalization theory of critical points. In
Section S3.C, we define the critical exponents for the Curie point in a magnetic
system.

EXERCISE 3.4. Compute the critical exponents, 0:, {3, 8, and, for a gas
whose equation of state is given by the van der Waals equation.
, ~swer:

The van der Waals equation in terms of reduced variables is


- 1) = 8T. In order to examine the neighborhood of the
critical point, we introduce expansion parameters c = (T /Tc) - 1,
w = (v/vc) - 1, and 7r = (P/Pe) - 1. In terms of these parameters, the

, (p + (3/v2))(3v

THE THERMODYNAMICS OF PHASE TRANSITIONS

140

van der Waals equation can be written

(I+7r)+

32][3(W+l)-IJ=S(I+).
(1 + w)

(1)

If we solve for n, we find

S + 16w + SW2 - 3w3


7r = -----....,....---2
3
2

(2)

+ 7w + Sw + 3w

(a) The degree a/the critical isotherm,


7r in powers of w. This gives

D. Let

= 0 in Eq. (2) and expand

(3)
Thus, the degree of the critical isotherm is 8 = 3.
(b) The ish.e.n:JJr.a.L.J'.:.l1.ll.w..rt~':2i.JijtlL.ma.ne.IJI.L:::f_..__U~s....c.cUIl1?ute(87r/ 8w ) E

and th

Converted with

STOI Converter
trial version

for w

(c) The d

(4)

hDP:llwww.stdutililV.com
__

critica.l.----r--~...,---_~_~

---r---=:....__ __

7r = 4 - 6w

--J;n

orhood of the

3
+ 9w 23-"2
w + ....

(5)

The values of w on either side of the coexistence curve can be found


from the conditions that along the isotherm,
(6)
where

P = P / P; is the reduced pressure and Vg = vg/vc

and

VI = vilv; are the reduced molar volumes of the gas and liquid,

respectively, on the coexistence curve. The condition that the pressure


and temperature of the liquid and gas in the coexistence region be
equal yields
-~wT

- 6wI

+ 9w;

= -~w!

- 6wg

If we note that WI is negative and write WI


Eq. (7) gives
-)
4(- WI + Wg

+ 6(-2
W1

-2)
Wg

+ 9w;.

-WI and Wg

-30=
+ WI-3 + Wg

(7)

= +wg, then
.

(8)

141

CRITICAL EXPONENTS

The fact that the molar Gibbs free energy of the liquid and gas phases
are equal means that JVrgVI vdP = 0 or that

[g' dw(l

+ 18we - ~J + .. -)

+w) ( -610

(9)
::::::-6e(WI - wg)

+ 6e(w;

- w;) - ~ (wi - w!)

+ ... = o.

or
- + Wg
-)
4 e (WI

- 2) + wI
- 3 + Wg
- 3 = 0.
+ 4 e (-wI2 - Wg

(10)

In order for Eqs. (8) and (10) to be consistent, we must have Wg = WI.
If we plug this into Eq. (8) or (10), we get WI = Wg = W. This gives
w2 ~ -4e. Note that e is negative.
Thus,
(11)

Converted with

(d) The he

obtaine

STOI Converter

acity can be
int along the
~ 1/2 and
y

trial version

cv~_h_n_p:_H_www~~.st~d~ut~ili~b~.c~o~m~
coex

aPI)

+ ( Bvl

(12)

(avI) 2 ]}
aT coex .

Along the coexistence curve (Ov/ar)coex = lel-1/2,


where the
minus sign applies to the liquid and the plus sign applies to the gas.
From Eqs. (5) and (11) we find

aPI)
(Ovl

== (ap- g)
T

Ovg

9 2 - 18wlel + ...
61el - -w

-121el

(13)

O(le 3/2).

If we note that (Pcvc/RTc) = (3/8), we find

cVc(Tc-) -cvc(T:)

=~R+O(e).

(14)

Thus, the van der Waals equation predicts a finite jump in the heat
capacity at the critical point and therefore it predicts a' = a = o.

142

THE THERMODYNAMICS OF PHASE TRANSmONS

... SPECIALTOPICS
... S3.A. Surface Tension
In Section 2.H.l, we showed that two parts of a thermodynamic system, at
equilibrium, must have the same pressure if the interface between them is free
to move and transmit mechanical energy. In deriving the condition for
mechanical equilibrium however, we neglected any contributions due to surface
tension. In this section we shall show that, if the surface moves in such a way
that its area changes, surface tension can affect the mechanical equilibrium
condition.
For simplicity, let us consider a monomolecular thermodynamic system
consisting of two phases, gas (vapor) and liquid in contact and in thermodynamic equilibrium. A typical example would be water droplets in contact
with water vapor. As we discussed in Section 2.C.6, very strong unbalanced
molecular forces at the liquid surface act to contract the surface. If the liquid
droplet grows and the surface area increases, then work must be done against
these surface
..
"
the mechanical
energy of the
In order t

Converted with

ical equilibrium

STDI Converter

due to surfac
et of radius, R,
floating in ga
ystem is held at
temperature
trial version
Vg + VI, where
VI = (4/3 )7rR
Vtot - VI is the
volume of th
f the droplet is
A = 47rR2 an 1 s su ace enslOn IS a.
e u
cules in the gas
phase is Ng and in the liquid phase is NI. The number of molecules comprising
the surface, Ns, generally will be very small compared to those in the liquid or

hDP:llwww.stdutililV.com

Vg,T

Fig. 3.27. A single liquid droplet in equilibrium with its vapor.

143

SPECIAL TOPICS: SURFACE TENSION

gas phases, so we take N, ~ O. Since the gas and liquid phases are free to
exchange particles and heat, at equilibrium the chemical potentials ilg and ill, of
the gas and liquid phases respectively, will be equal and the temperature will be
uniform throughout. However, the pressures of the gas and liquid phases need
not be equal if surface effects are included. Since we are dealing with a
monomolecular substances, the chemical potential will be a function only of
pressure and temperature. Therefore, at thermodynamic equilibrium we have
(3.93)
Because we are dealing with a system at fixed temperature, total volume, and
chemical potential, it is convenient to work with the grand potential (cf. Section
2.F.5). The grand potential for the entire system can be written

For a system at equilibrium with fixed temperature, total volume, and chemical
potential, the J rand potential must be a minimum. Therefore, we obtain the
condition for tl

Converted with

STOI Converter

(3.95)

trial version

or

hDP://www.stdutilitv.com
(3.96)
If the interface between the two parts of a thermodynamic system has a surface
tension and if the surface has curvature, then the condition for mechanical
equilibrium must be corrected to include the effects of surface tension. The
pressures in the two parts of the system need no longer be equal. The surface

Table 3.1. The Surface Tension for Various


Liquids in Contact with Air
Substance
Water
Water
Alcohol
Alcohol
Soap solution
Mercury

a (dynes/em)
72.0
67.9
21.8
19.8
25.0
487

25
50
25
50
25
15

144

THE THERMODYNAMICS OF PHASE TRANSITIONS

tension varies with temperature and generally decreases as the temperature is


raised. The addition of impurities to the liquid will reduce the surface tension
because it disrupts the tight network of "unbalanced" intermolecular forces at
the interface. Some values of the surface tension of several substances are given
in Table 3.1. Equation (3.96) was obtained by thermodynamic arguments. In
Exercise 3.5, we show that it can also be obtained by purely mechanical
arguments .
EXERCISE 3.5. Use purely mechanical arguments to derive Eq. (3.96)
for a spherical liquid droplet, with radius R, floating in a gas. Assume that
the liquid has pressure PI, the gas has pressure Pg, and interface has surface
tension a.
Answer: Consider the forces acting on an area element, dA =
rR2 sin( O)dOd, located in the interval 0 ~ 0 + dO and ~ + d on
the surface of the droplet. There will be a force outward (inward) along the
direction, r, due to the pressure of the liquid (gas). The net force on the area
element due

Converted with

STOI Converter

where r = s
exerts force
and perpend
element in
(work/area)=
rce en
dA, experiences a force

trial version

hDP://www.stdutilitv.com
dfl = -q,aR dO,

(1)
rface tension
to the surface
etch the area
e tension are
area element,
(2)

where R dO is the length of the side and the unit vector, q" can be written
q, = - sin()x + cos()y. The side, (0 ~ 0 + dO, + d), of the area
element, dA, experiences a force

(3)
where R dO is the length of the side and the unit vector, + d, can be
written q, + dq, == - sin( + d)x + cos( + d)Y. The side, (0, ~ +
d ), of the area element, dA, experiences a force
df3 = -OaRsin(O)d,

(4)

where R sin( O)d is the length of the side and the unit vector, 0, can be
written 0 = cos(O) cos()x + cos(O) sin()y + sin(O)i. The side, (0 + dO,
~ + d ), of the area element, dA, experiences a force
df4

(0 + dO)aRsin(O

+ dO)d,

(5)

145

SPECIAL TOPICS: SURFACE TENSION

where R sin( 0 + dO)d is the length of the side and the unit vector, 6 + d 6,
can be written 9 + d6 == cos(O + dO) cos()i + cos(O + dO) sin()y+
sin(O + dO)z. If the droplet is to be in equilibrium, these forces must add
to zero.
The condition for mechanical equilibrium takes its simplest form if we
integrate these forces over the top hemisphere of the droplet-that is, if we
integrate the forces over the interval 0 ~ 0 ~ 7r/2 and 0 ~ ~ 27r. [Before we
integrate, we must expand Eqs. (3) and (5) and only keep terms of order
I dOd]. The integration over causes all contributions in the i and y
directions to drop out and we get

"

[dFp + dfl + df2 + df3 + df4]

[(PI - Pg)7rR2 - 27raR]z. (6)

hemisphere

i
I

If the droplet is to be in equilibrium, these forces must add to zero.


Therefore,

(7)
I

and we obtai
obtained from

urn as was

Converted with

STOI Converter

It is interesti
the gas as a
function of the r
trial version
ble change in
the pressure of t
the system in
equilibrium and
aling with a
monomolecular su s ance, we
ow a revers 1 e c anges In the chemical
potential of the liquid and vapor can be written du' = vdP - sdT, where
V = V/ N is the volume per particle and s = S/ N is the entropy per particle. If
the radius of the droplet changes, keeping the droplet in equilibrium with the
vapor and keeping the temperature fixed (dT = 0), then d/-l~ = d/-l~ and

hDP://www.stdutilitv.com

(3.97)
where VI = VdNI and Vg = Vg/Ng. However, from Eqs. (3.96) and (3.97), we
can write dPI - dPg = d(2a/R) and
(3.98)
Let us now restrict ourselves to the case when the volume per particle of the
liquid is much smaller than that of the gas (far from the critical point) so that
~I vg Then we can approximate the volume per particle of the gas by the
Ideal gas value, Vg = (kBT/Pg), and Eq. (3.98) takes the form
kBT dPg
VI r,

= d(2a).
R

(3.99)

146

THE THERMODYNAMICS OF PHASE TRANSITIONS

peR)

P(Ro)

Ro

Fig. 3.28. A plot of the equilibrium vapor pressure of the droplet as a function of the
radius of the droplet.

From Eq. (3.99), we can now find how the equilibrium vapor pressure of the
gas varies as a function of the radius of the droplet. We will neglect changes in
VI as the pressure changes because liquids are very incompressible. Let us
integrate the ri ht-hand side of E . 3.99 from R = 00 to R and the left-hand
side from Pg (

Converted with

STOU Converter

(3.100)

trial ve rsion

s no curvature
be greater to
(i.e., is flat).
maintain the
n (3.100) tells
us that only droplets of a given size can exist in equilibrium with a vapor at
fixed P and T. Drops of different size will be unstable.
A plot of the equilibrium vapor pressure as a function of the radius of the
droplet is given in Fig. 3.28. For a given radius, R = Ro, the equilibrium vapor
pressure is Pg(Ro). If there happens to be a droplet with radius, R > Ro, then the
pressure, Pg(Ro), is too high and the droplet will absorb fluid in an effort to
reduce the pressure of the gas and eventually will condense out. If R < Ro, the
pressure of the gas is too low and the droplet will give up fluid in an effort to
increase the pressure and will disappear. As a result, the droplets tend to be a
uniform size.
The quantity P

hUP:llwww.stdutililV.com

~ Sl.B. Thermomechanical Effect [10]


The behavior of liquid He4 below 2.19 K can be described in terms of a model
which assumes that liquid He4 is composed of two interpenetrating fluids. One
fluid (superftuid) can flow through cracks too small for He4 gas to leak through,
and it appears to carry no entropy. The other fluid (normal fluid) behaves
normally. The fact that the superftuid carries no entropy leads to some very

SPECIAL TOPICS: THERMOMECHANICAL EFFECT

147

Fig.3.29. 1\\'0 vessels containing liquid He4 below 2.19 K and connected by a very fine
capillary. Only superfluid can pass between the two vessels.

interesting behavior, some of which can be described with classical thermodynamics. Let us consider two vessels, A and B, filled with liquid He4 at a
temperature below 2.19 K, and let us assume that they are connected by a
capillary so thin that only the superfluid can flow through it (cf. Fig. 3.29). Let
us further assume that the vessels are insulated from the outside world. This
means that the total entropy must remain constant if no irreversible processes
take place. Let us
f the system
remain constant.
Converted with
is a state of
minimum intern
essels if we
We can obtai
total internal
assume that matt
er kilogram
energy will be d
trial version
en given by
(specific internal

STOI Converter
hDP://www.stdutilitv.com

(3.101)

UT=
I=A,B

where Ml is the total mass of liquid He4 in vessel I. At equilibrium, the total
internal energy must be a minimum. Thus,

BUT = 0 =

L (u18MI + MI8ul).

(3.102)

I=A,B

Let us now assume that the total volume, VI, and the total entropy, Sl, of liquid
He4 in vessel I (for I = A, B) are constant (this is possible because only
superfluid can flow between the vessels and superfluid carries no entropy). The
entropy of liquid He4 in vessell can be written Sl = MISI, where Ml and Sl are
the total mass and specific entropy of liquid He4 in vessel I. Similarly,
VI :::: MIVI, where VI is the specific volume of liquid He4 in vessel I. Since
Sl and VI are constants, we can write 8S1 = Ml8s1 + sl8Ml = 0 and 8Vl =
Ml8vl + vl8Ml = O. Therefore,
(3.103)

148

THE THERMODYNAMICS OF PHASE TRANSmONS

and
(3.104)
Let us now expand the differential, 8UI, in Eq. (3.102) in terms of specific
entropy and specific volume. Equation (3.102) then takes the form

+ MI [(~:/)

(UI8MI
I=A,B

8s1 +

(:/)

VI

8VI])

= O.

(3.105)

SI

If we note that (8u/8s)v = T and (8u/Bv)s = -P [cf. Eqs. (2.68) and (2.69)1
and make use of Eqs. (3.103) and (3.104), we obtain

(UI - slTI
I=A,B
where ill is the
total mass is
equilibrium cor

+ vIPI)8MI =

ill8MI
I=A,B

= 0,

Converted with

STOI Converter

(3.106)

sel I. Since the


we obtain the

(3.107)

trial version
Since atoms Cal
He4 in the two
vessels must be
bd and volume
cannot change (no mechanical energy transfer), the pressure and temperature of
the two vessels need not be the same.
We can now vary the temperature and pressure in one of the vessels (vessel
A, for example) in such a way that the two vessels remain in equilibrium. The
change in chemical potential in vessel A is

hDP:llwww.stdutililV.com

(3.108)
where S = -(8fl18T)p
and V = (8fl18P)r. But to maintain equilibrium, we
must have ~ilA = 0 so the chemical potentials of the two vessels remain equal.
Therefore,
(3.109)
Thus, a change in temperature of vessel A must be accompanied by a change in
pressure of vessel A. If the temperature increases, the pressure will increase.
This is called the thermomechanical effect.
The thermomechanical effect is most dramatically demonstrated in terms of
the so-called fountain effect. Imagine a small elbow tube filled with very fine

SPECIAL TOPICS. THE CRITICAL EXPONENTS FOR THE CURIE POINT

149

~O>
light
Fig. 3.30. The fountain effect.

powder, with cotton stuffed in each end. Assume that a long, thin capillary tube
is put in one end and the elbow tube is immersed in liquid He4 at a temperature
below 2.19 K. If we now irradiate the elbow tube with a burst of light, the
pressure of heli
Converted with
11 spurt out of
the capillary tu
ect and is a

STOI Converter

consequence of
radiation, supe
potentials and i
trial version
fountain effect
vessel. Howeve
violate the seco'fn-r---r;n.,.,.-----------------

hDP://www.stdutilitv.com

is heated by
the chemical
ing that in the
oler to hotter
this does not

~ Sl.e. The Critical Exponents for the Curie Point


For magnetic systems, exponents 0, {3, 'Y, and 8 can be defined in analogy with
pure fluids. The phase diagrams for simple magnetic systems are given in
Figs. 3.31-3.33. In Fig. 3.31, we sketch the coexistence curve for a ferromagnetic system. Below some critical temperature the spins begin to order
spontaneously. The coexistence curve separates the two directions of
magnetization. In Fig. 3.32 we plot some isotherms of the magnetic system,
and in Fig. 3.33 we plot the magnetization as a function of temperature. It is
helpful to refer to these curves when defining the various exponents.
(a) Degree of the Critical Isotherm. The exponent 8 describes the variation of
magnetization with magnetic field along the critical isotherm

(3.110)

150

THE THERMODYNAMICS OF PHASE TRANS mONS

H=oo

H=

t t

= 0 t------ ___

-00 ---------~

t;

Fig. 3.31. Coexistence curve for a typical magnetic system. Below the Curie point the
magnetization occurs spontaneously. The curve H = 0 separates the two possible
orientations of the magnetization.

-r

-_l}_

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com

T <

r.>

-Mo

T=O

Fig. 3.32. A sketch of the isotherms for a ferromagnetic system.

where dj = kBT /mo, Mo(O) is the magnetization in zero field at zero


temperature, mo is the magnetic moment per spin, and Bo is a proportionality
constant. Experimentally, 8 has values 4 ~ 8 ~ 6 in agreement with the values of
8 for pure fluids.
(b) Magnetization Exponent. In a magnetic system, the exponent f3 describes
how the magnetization approaches its value at the critical point when no
external field is present. It is defined as follows:
Mo(T)
Mo(O)

= B (_ )fj
fj

(3.111)

where Bfj is a constant. For magnetic systems, j3 ~

!as it is for fluids.

151

SPECIAL TOPICS: TRICRITICAL POINTS

Fig. 3.33. A sketch of the magnetization


for a simple ferromagnetic system.

-u,

(c) The Heat Capacity. For magnetic systems, the coefficients


defined as follows:

and a' are

(3.112)

Converted with
where Bo. and B

STOI Converter

(d) The Magneti


the critical point

trial version

,
f"V

e vicinity of

hDP://www.stdutilitv.com
0-

XT

T>

Ryf.-'Y ,

r;

O.

(3.113)

H=O,

where B~ and B'Y are constants and X~ is the susceptibility of a noninteracting


system at the critical temperature. For real systems, 'Y has been found to be
'Y
1.3.
The striking feature about the critical exponents for fluids and for magnetic
systems is that the values are roughly the same. Indeed, there appears to be a
great similarity in the way in which many systems approach their critical
points.
f"V

~ S3.0. Tricritical Points


Ginzberg-Landau theory allows us to describe a tricritical point (a point where
a line of Apoints meets the critical point for a line of discontinuous transitions).
Let us consider a free energy of the form
1>(T, Y,

"1) = 1>o(T, Y)

+ Q2"12(T,

Y)

+ Q4"14(T,

Y)

+ Q6"16(T,

Y)

+ ... ,
(3.114)

152

THE THERMODYNAMICS OF PHASE TRANSITIONS

where 02 = 02(T, Y), 04 = 04(T, Y), and


free energy are given by

06

= 06(T,

Y).The extrema of the

(3.115)
This has solutions
(3.116)
Because there is no cubic or fifth-order term in the free energy, we can have a
regime in which there is a line discontinuous transitions and a regime in which
there is a line of continuous phase transitions. The conditions that must be
satisfied for a discontinuous transition to occur are - o = 0 and
(8/8",)r y = O. These two conditions give ",2 = -04/206 (the discontinuous
transition 'can only occur for 04 < 0 so ",2 is non-negative) and

Converted with
The line of disc
satisfy Eq. (3.1
condition

(3.117)

STOI Converter
trial version

rand Y which
given by the

hDP://www.stdutilitv.com
(3.118)

..........

y
Fig. 3.34. A plot of the neighborhood of the tricritical point in the (T, Y) plane. The
point t connecting the line of continuous phase transitions (dotted line) and the line of
first-order phase transitions (solid line) is a tricritical point. It is the meeting point of
three phases.

153

SPECIAL TOPICS: BINARY MIXTURES

The lines of discontinuous and continuous phase transitions meet at a point,


(Tt, Yt), determined by the conditions a2(Tt, Yt) = 0 and a4(Tt, Yt) = O. A
schematic picture is given in Fig. 3.34. The point, t, was called a tricritical
point by Griffiths [25]. He showed that in a suitable space it is the meeting point
of three lines of critical points (two of them associated with the discontinuous
transition).
A tricritical point occurs in He3-He4
mixtures (cf. Fig. 3.19). It is
simultaneously the critical point for the binary phase transition and the end
point of a line of A points associated with superfluid phase transition .

.... S3.E. Binary Mixtures [2, 31-33]


If we consider a fluid which is a mixture of several different types of interacting
particles, a phase transition can occur in which there is a physical separation of
the fluid into regions containing different concentrations of the various types of
particles. The simplest example of this type of phase transition occurs for binary
mixtures.
It is useful fir
applicable to
Converted with
all binary mixtur
posed of nl
moles of type 1

STOI Converter
trial version

and differential

(3.119)

hDP:llwww.stdutililV.com
(3.120)

The molar Gibbs free energy,

+ n2), is

g = G/n(n = nl

(3.121 )
where Xl and X2 are the mole fractions of particles of type 1 and 2, respectively.
From Eqs. (3.120) and (3.121), it is easy to show that

dg = -sdT

+ vdP + (I-ll

-1-l2)dxl,

(3.122)

so that g = geT, P, xt).


The chemical potential of type 1 particles is
I-ll =

(aG)
an}

=s+ (I-XI)(a )
P,T,n2

ax}

(3.123)
P,T

and the chemical potential of type 2 particles is

1-l2= (aG)
=g-x}an2 e.r,

(ag)
ax}

(3.124)
P,T'

154

THE THERMODYNAMICS OF PHASE TRANSmONS

where we have used the fact that


and

= _ Xl

(8XI)

8n2

(3.125)

nl

From Eqs. (3.122)-(3.124),


we see that the chemical potential depends on the
mole numbers nl and n2, only through its dependence on the mole fraction, XI.
We can also show that
8J-LI)
(- X

=(I-XI)P,T

(82g)
8xi

(3.126)
P,T

and

(3.127)

Converted with

~ S3.E.l. Sta
In Section 2.f
require that the

STOI Converter

~ell's

relations

trial version

hDP://www.stdutilitv.com
\J-L2,1 J-L2,2j
must be symmetric positive definite, where J-Li,j = (8J-L;/ 8nj) P,T,{nki'n }. This
requires that tsu > 0 (i = 1,2) and that every principal minor be pO~ltive or
zero. Thus, for a binary mixture we must have
det(J-LI,1 J-LI,2) 2: 0,
J-L2,1 J-L2,2

J-LI,I>O,

and

J-LI,2= J-L2,1'(3.128)

For processes that occur at constant T and P, we have some additional conditions. From the Gibbs-Duhem equation (2.62), for processes in which dT = 0
and dP = 0 we have
(3.129)
Furthermore,

we can write
(3.130)

155

SPECIAL TOPICS: BINARY MIXTURES

and
[dJ-L2]P,T = (aJ-L2)
an

P,T,n2

dn,

+ (aJ-L2)
an2

r.r,

(3.131)

dn-.

If we combine Eqs. (3.129)-(3.131) and use the fact that the differentials dn,
and dn2 are independent, we find that
(3.132)

and
nl (aJ-LI)
+n2 (aJ-L2)
=0.
an I P,T,n2
an I P,T,n2

(3.133)

Thus, if J-LI, I > 0, then J-L2,I < 0, then for binary systems the conditions for
chemical stabilir

(8~[)
anI

p,T,n2>

Converted with

STOI Converter
trial version

From Eq. (3.127

hDP://www.stdutilitv.com

(8

2g

-2 )

aXI

P,T

>0.

<0.

P,T,nl
(3.134)

ondition
(3.135)

Thus, for chemical stability, the molar Gibbs free energy must be a convex
junction of the mole fraction .

.....S3.E.2. Equilibrium Conditions


In Section 2.H.2, we found that for a binary mixture to be in equilibrium the
pressure P, the temperature T, and the chemical potential of each type of particle
must be equal (uniform) throughout the system. The chemical potentials of
different types of particles need not be equal.
If a phase separation occurs in a binary mixture, the chemical potential of
type 1 particles must be equal in the two phases and similarily for the chemical
potential of type 2 particles. This equality of chemical potentials between the
two phases gives us a condition for locating the coexistence curve. For type 1

156

THE THERMODYNAMICS OF PHASE TRANSITIONS

particles we have J.L{ = J.L{I or


(3.136)
and for type 2 particles we have J.L ~ = J.L

gl _ x~ ( ag) 1
Bx,

P,T

q or

gil _ x~1( ag ) II ,
ax!

(3.137)

P,T

where I and II denote the two phases. If we combine Eqs. (3.136) and (3.137),
we can write the conditions for equilibrium in the form

( ag)1
aX!

(ag)1I
P,T

aX!

(3.138)
P,T

and

Converted with

STOI Converter

(3.139)

Equations (3.1
ts in the (g, xI)
trial version
plane have eq
uations (3.138)
and (3.139) 10
~--~----~~--~~--~~~~~
In Fig. 3.35 we show a sketch of the molar Gibbs free energy which
illustrates these various properties. It shows a region where two phases can

hDP://www.stdutilitv.com

xF

Fig. 3.35. A sketch of the molar Gibbs free energy of a phase separated binary mixture.
The two points with common tangent are equilibrium states.

157

SPECIAL TOPICS: BINARY MIXTURES

coexist. The points that have a common tangent are the equilibrium points. The
concave region in the middle is unstable. In this region the two phases, one rich
type 1 particles and the other rich in type 2 particles, can coexist. In a test tube,
the heavier phase will sink to the bottom and the lighter one will float to the top.
As long as (a J.L2/ aXI ) P T < 0, the binary mixture will be stable and exist in
one phase. However, if (a~2/ aXI) P T > 0 the system has an unstable region and
phase separation occurs. The critical point for this phase separation is given by
(aJ.L2/aXt)~
T = O. The critical point is the point where the Xl first becomes a
double-valued function of J.LI or J.L2 as the temperature is changed. That is, two
different values of Xl give the same value of the chemical potential. Thus, in
analogy to the liquid-vapor critical point (with P -+ J.L2 and v -+ Xl), the critical
point is a point of inflection of the curve J.L2 = J.L2(T,P,XI)
for T and P
constant. Therefore, we have the additional condition that (a2 J.L2/ axD~ T = 0 at
the critical point.
'
A sketch of the coexistence curve, and the curve separating stable from
unstable states is given in Fig. 3.36. The region outside and above the coexistence curve corresponds to allowed single-phase equilibrium states. Below
the coexistence curve is a coexistence re ion in which two equilibrium states
with different
t at the same
Converted with
tes. These are
temperature.
rium but are
single-phase s
e are unstable
chemically sta
and cannot b
trial version
temperature, T
onsisting only
of type 2 partie
tration of type
1 particles inc
oint I. At this
point, the system separates into two phases, one in which type 1 particles have
concentration x{ and another in which type 1 particles have concentration x[I.
As we increase the number of type 1 particles relative to type 2 particles, the

STOI Converter
hDP://www.stdutilitv.com

metastable

region

coexistence curve

Fig. 3.36. The phase diagram for a binary mixture. The point C is the critical point.

158

THE THERMODYNAMICS OF PHASE TRANS mONS

one phase

0.5

C6H5N02

Fig. 3.37. The phase diagram


of n-hexane and nitrobenzene
at atmospheric pressure. The
the coexistence curve. (Based

for a mixture
(C6 Hs N02)
solid line is
on Ref. 2.)

amount of phase II increases and the amount of phase I decreases until we reach
the coexistence curve at point II. At point II, phase I has disappeared and we
again have a single-phase equilibrium state of concentration, xf.
We see that
hsition and the
Converted with
separation of 2
~ of a system
exhibiting this
itrobenzene at
atmospheric pre
n in Fig. 3.37.

STOI Converter
trial version

EXERCIS
whose Gibbs

hDP://www.stdutilitv.com

G = nllJ,?(P,

T)

+ n2J-L~(P,

T)

+ RTnlln(XI)

vpes 1 and 2,

+ RTn21n(x2) + AnxIX2,

where n = nl + n2, nl and n2 are the mole numbers, and Xl and X2 are the
mole fractions of particles of types 1 and 2, respectively. (a) Plot the molar
Gibbs free energy g(P, T, Xl) versus Xl for fixed P and T (for definiteness
assume that J-L?= 1.0, J-L~= 1.05, RT = 0.8, and A = 1). What are
and
x11? (b) Use conditions (aJ-LI/aXI)~T
=(&J-LI/aif.)~T
= 0 to locate the
critical point. (c) Find the condition for equilibrium between the two
phases. Show that the conditions J-L{ = J-L{I and J-L~= J-L and the condition
(ag/aXt)~,T
= (ag/aXt)~,T
are equivalent.

x1

Answer:
(a) The molar Gibbs free energy is
G
g = - = Xl J-L?+ X2J-L~ + RTx1ln(XI)

A plot of g versus
RT = 0.8,and A = 1.

Xl

+ RTx2In(x2)

+ AxIX2,

(1)

is given below for J-L?= 1.0, J-L~= 1.05,

159

SPECIAL TOPICS: BINARY MIXTURES

The equilibrium concentrations are

x1 = 0.169

and

x11= 0.873.

(b) The chemical potential, J.L2, is

Axi.

(2)

Converted with

STOI Converter

2/{)XI)~T =
/(I-XI')2+
e find that

trial version

hDP://www.stdutilitv.com

(c) There
urn between
the two phases. (1)
e con ition
g Xl P T = ({)g / {)xt)~,T'
together with Eq. (1), yields the equilibrium condition
A(l-

2xD +RTln(l

~~xJ

= A(l - 2x(/)

+RTlnC ~~(}
(3)

(ii) The conditions J.L1


conditions

= J.L11 and J.L~ = J.L yield the equilibrium

and
RTln(1 -

x1)

+ ..\(x1)2

= RTln(1 -

xn

+ ..\(xf)2.

(5)

If we subtract Eq. (5) from Eq. (4), we recover Eq. (3). Thus, the two
equilibrium conditions are equivalent.

160

THE THERMODYNAMICS OF PHASE TRANSITIONS

.... S3.E.3. Coexistence Curve


Before we write an equation for the coexistence curve, it is necessary to
introduce a new expression for the chemical potential. Let us first remember
that the Gibbs free energy can be written G = U - ST + PV = nIl},I + n2/-L2.
Therefore, the chemical potential for type 1 particles can be written

/-LI

= (OG)
Bn,

= (OU)
P,T,nz

Bn,

-T
P,T,nz

(OS)
-

Bn,

(OV)
Oni

+P P,T,nz

(oU lont)p

(3.140)

P,T,nz

==
, T , nz is a partial molar internal energy, SI ==
is a partial molar entropy, and VI ==
I) P,T nz is a partial
molar volume. Note that the chemical potential is a partial moiar Gibbs free
energy. Similar quantities exist for type 2 particles. In terms of these "partial"
quantities, the chemical potentials can be written
The quantity

(oSIon I) P,T,nz

UI

(oV I on

(3.141)
It is useful to
partial molar v

STOI Converter

In

and the

trial version

hDP://www.stdutilitv.com

(3.142)

The partial molar enthalpies of the type 1 and type 2 particles are

respectivel y.
If we are in a region of coexistence of the two different phases, then the
temperature, pressure, and chemical potentials of each type of particle must be
equal. As we move along the coexistence curve, changes in the chemical
potentials, temperature, and pressure of the two phases must be equal. Let us
now consider the differential of the quantity, /-Ll IT. It is

(3.144 )

161

SPECIAL TOPICS: BINARY MIXTURES

However,

[~8T (ILl)]
T Pn

, 1, Z

.:T2'

(3.145)

where we have used the fact that (8J.Lt/8T)Pn n = -(8S/8nt)PTn


and J.LI =
hI - sIT. If we make use of the fact that (8J.LI'/&P)T,n),nz = (8V /8nl)p,T,nz = VI
and (3.145), we can write Eq. (3.144) in the form
d(ILI)
T

+ VI dP + 2_ (8J.LI)

= _ ~dT

T2

dxi,

(3.146)

1 (8J.L2)
dxi;
T 8XI P,T

(3.147)

8XI

P,T

and we can write


d (J.L2)
T
where VI and V
curve, d(J.LI/T) I
dP, but
"# c
coexistence cun

dx1

_ ~hl dl
T

h2
V2
--dT+-dP+T2
T

e coexistence

Converted with

dpl = dplI

STOI Converter

==

ations for the

trial version
(3.148)

hDP://www.stdutilitv.com

and
(3.149)

v: - v:

1 for (i = 1,2).
where ~hi = h: - hf! and ~Vi =
Let us now obtain an equation for the coexistence curve in the (T,xt) plane
for constant pressure processes, dP = O. The equations for the coexistence
curve reduce to

(3.150)
and
_ ~h2
T

+ (8J.L2)1
8XI

(dx1)

P,T dT

_ (8J.L2)
P

8Xl

II

P,T

(dx{l) _ 0
dT

P-

(3.151)

162

THE THERMODYNAMICS OF PHASE TRANSmONS

We can solve Eqs. (3.150) and (3.151) for (dxUdT)p


Eqs. (3.126) and (3.127), we obtain

and (dx{l/dT)po If we Use

dx{)
x{1Ahl + (1 - xnAh2
( dT p = T(x{1 - xD(a2g/axD~,T

(3.152)

and
(3.153)
Thus, in general there is a latent heat associated with this phase transition and it
is a first-order transition.

~ S3.F. The Ginzburg-Landau

Theory of

Superconductrr....::....s--"-[9---L...-2_6~3__.:._5_.,_]
One of the

---,

Converted with

andau theory
in both normal
ust allow the
order paramete
trial version
mal magnetic
field, H, is ap
of a spatially
varying order
normal and
condensed regi
ing induction
field, B(r). The order parameter is treated in a manner similar to the wave
function of a quantum particle. Gradients in the order parameter (spatial
variations) will give contributions to the free energy of a form similar to the
kinetic energy of a quantum particle. Furthermore, if a local induction field B(r)
is present, the canonical momentum of the condensed phase will have a
contribution from the vector potential A(r) associated with B(r). The local
vector potential A(r) is related to the local induction field, B(r), through the
relation

STOI Converter
hDP://www.stdutilitv.com

B(r)

= Vr

x A(r).

(3.154)

The Helmholtz free energy as(r, B, T) per unit volume is given by

(3.155)

where e and m are the charge and mass, respectively, of the electron pairs, and
1i is Planck's constant. The quantity, (-itt'\! r - eA), is the canonical

SPECIAL TOPICS: THE GINZBURG-LANDAU THEORY OF SUPERCONDUCTORS

163

momentum operator associated with the condensate. It has exactly the same
form as the canonical momentum of a charged quantum particle in a magnetic
field. The Gibbs free energy per unit volume is given by a Legendre
transformation,

g(r,H, T) = a(r,B, T) - B H.

(3.156)

The total Gibbs free energy is found by integrating g(r, H, T) over the entire
volume. Thus,

(3.157)

If we now extremize Gs (H, T) with respect to variations in

that on the bou


momentum, (i1i~

\}I * (r)

and assume
the canonical

Converted with

STOI Converter

(3.158)

trial version
To obtain Eq. (3
Gs(H, T) with n ~----------------------------~

hDP://www.stdutilitv.com

ext extremize

(3.159)
To obtain Eq. (3.159), we have used the vector identities

dYC (Vr x A)

= fdS.

(A x C)

dVA Vr xC

(3.160)

and

(E x F) = fE . (F x tiS),

fdS.

(3.161 )

~here I dV is an integral over the total volume Vof the sample and f dS is an
Integral over the surface enclosing the volume V. We have assumed that
(B - J.LoH)x = 0, where
is the unit vector normal to the surface of the
sample. Thus, right on the surface the tangent component of the induction
field is the same as that of a normal metal. However, it will go to zero in the
sample.

164

THE THERMODYNAMICS OF PHASE TRANSmONS

The local supercurrent in a superconductor


induction field and is defined as

is driven by the local magnetic

J, (r) = -1 V'r x B (r) = -2en . (w * V',W - WV',W * ) - -e


~

ml

Alwl 2

(3.162)

as we would expect. It is a combination of the current induced by H and the


magnetization current. If we compare Eqs. (3.159) and (3.162) we see that the
supercurrent Js(r) has the same functional form as the probability current of a
free particle. Inside a sample, where B(r) = 0, there is no supercurrent. The
supercurrent is confined to the surface of the sample.
We can use Eqs. (3.158) and (3.159) to determine how the order parameter,
w(r), varies in space at the edge of the sample. Let us assume that A(r) == 0,
but that W can vary in space in the z direction W = w(z). Under these
conditions, we can assume that w(z) is a real function, and therefore there is no
supercurrent, J, = O. Let us next introduce a dimensionless function,

(3.163)

Converted with
Then Eq. (3.1:

STOI Converter
trial version

(3.164)

hDP://www.stdutilitv.com
where ~(T) is called the Ginzburg-Landau

coherence

{T

length and is defined as

(3.165)

~(T)=V~'

Equation (3.164) can be used to find how the order parameter varies in space on
the boundary of a superconducting sample.
Let us consider a sample that is infinitely large in the x and y directions but
extends only from z = 0 to z = 00 in the z direction. The region from z = - oo
to z = 0 is empty. We will assume that at z = 0 there is no condensate,
f(z = 0) = 0; but deep in the interior of the sample it takes its maximum value
f(z = 00) = 1 [that is, w(z) = (10:21/20:4)1/2 deep in the interior of the sample].
Equation 3.164 is a nonlinear equation for J, but it can be solved. To solve it we
must find its first integral. Let us multiply Eq. (3.164) by df [d ; and rearrange
and integrate terms. We then find

-t,2

(1)

2= f2

- ~f4

+ c,

(3.166)

SPECIAL TOPICS: THE GINZBURG-LANDAU THEORY OF SUPERCONDUCTORS

165

where C is an integration constant. We will choose C so that the boundary


conditions
df
- ~O
dz z-s-oc

and f---+l

z=-oo

are satisfied. This gives C = -(1/2)

e (ddzf)

and Eq. (3.166) takes the form


2

=!2 (1- f2)2.

(3.167)

We can now solve Eq. (3.167) and find

f(z)

= tanh.

wt

v2~

(3.168)

[f(z) is plotted in Fig. 3.38]. Most of the variation of the order parameter occurs
within a distance z = 2~(T) from the boundary. Thus, the order parameter is
zero on the surface but increases to its maximum size within a distance 2~(T) of
the surface. Nea
It is also usef
Converted with
~etic field and
determine the rru
at the surface
of the sample. Lc
hroughout the
superconductor (
is often a bad
trial version
assumption). Le
nted in the y
direction and car
.162) reduces
to

STOI Converter
hDP:llwww.stdutililV.com

(3.169)
where

A=

(3.170)

J(z)

z/V2e

Fig. 3.38. A plot of fez)


versus z/ (v'2e).

= tanh(z/ (v'2e))

166

THE THERMODYNAMICS OF PHASE TRANSmONS

is the penetration depth. Thus the vector potential drops off exponentially inside
the superconductor,
(3.171)
and we obtain the result that all supercurrents are confined to within a distance
A of the surface. Note that the penetration depth also becomes very large near
the critical point.
There are many more applications of the Ginzburg-Landau theory of
superconductors than can be presented here. The interested reader should see
Ref. 9.

REFERENCES
1. P. N. Bridgeman, J. Chem. Phys. 5, 964 (1937).
2. I. Prigogine and R. Defay, Chemical Thermodynamics (Longmans, Green, and Co.,
London, 1954).
3. M. W. Zem~
York, 1957).
Converted with
4. D. ter Haar
~ddison- Wesley,
Reading, M
5. L. D. Landar
s.Oxford, 1958).
6. E. A. Gugge
trial version
7. I. de Boer a

STOI Converter

hDP://www.stdutilitv.comity

8. F. London,
(Dover Publications, N" ~~~, ~ ~.f'
9. M. Tinkham, Introduction to Superconductivity (McGraw-Hill, New York, 1975).
10. H. K. annes, Leiden Comm. 122b, 124c (1911).
11. N. Meissner and R. Ochsenfeld, Naturwissenschaften 21, 787 (1933).
12. F. London, Superjluids II: Macroscopic Theory of Superjluid Helium (Dover Pub.
New York, 1964).
13. W. E. Keller, Helium-3 and Helium-4 (Plenum Press, New York, 1969).
14. I. Wilks and D. S. Betts, An Introduction to Liquid Helium (Clarendon Press, Oxford,
1987).
15. W. H. Keesom and M. Wolfke, Proc. R. Acad. Amsterdam 31, 90 (1928).
16. N. D. Mermin and D. M. Lee, Scientific American, Dec. 1976, p. 56.
17. A. I. Leggett, Rev. Mod. Phys. 47, 331 (1975).
18. I. Wheatley, Rev. Mod. Phys. 47, 415 (1975).
19. S. G. Sydoriack, E. R. Grllly, and E. F. Hammel, Phys. Rev. 75, 303 (1949).
20. D. D. Osheroff, R. C. Richardson, and D. M. Lee, Phys. Rev. Lett. 28, 885 (1972).
21. B. M. Abraham, B. Weinstock, and D. W. Osborne, Phys. Rev. 76, 864 (1949).
22. I. Prigogine, R. Bingen, and A. Bellemans, Physica 20, 633 (1954).
23. G. K. Walters and W. M. Fairbank, Phys. Rev. Lett. 103,262 (1956).
24. P. Alpern, Th. Benda, and P. Leiderer, Phys. Rev. Lett. 49, 1267 (1982).

167

PROBLEMS

25. R. B. Griffiths, Phys. Rev. Lett. 24, 715 (1970).


26. L. D. Landau and E. M. Lifshitz, Statistical Physics, 3rd edition, Part 1 (Pergamon
Press, Oxford, 1980).
27. H. E. Stanley, Introduction to Phase Transitions and Critical Phenomena (Oxford
University Press, Oxford, 1971).
28. M. E. Fisher, Repts. Prog. Phys. 30, 615 (1967).
29. P. Heller, Repts. Prog. Phys. 30, 731 (1967).
30. F. London, Superjluids II: Macroscopic Theory of Superjluid Helium (Dover Publications, New York, 1964).
31. 1. H. Hildebrand, Solubility of Non-electrolytes (Reinhold, New York, 1936).
32. J. S. Rowlinson, Liquids and Liquid Mixtures (Butterworth, London, 1969).
33. J. G. Kirkwood and I. Oppenheim, Chemical Thermodynamics (McGraw-Hill, New
York, 1961).
34. E. Lapp, Ann. Physique (10) 12, 442 (1929).
35. V. I. Ginzburg, Sov. Phys. - Solie State 2, 1824 (1961).

PROBLEMS~--------------------~
Converted with
Problem 3.1. A
(u + (a/v))5/2], w
Compute the mol
liquid and vapor p
R, and the liquid aJ
and Vg if you neec

STOI Converter
trial version

hDP://www.stdutilitv.com

R In[C(v - b)
pn of state. (b)
heat between
he gas constant
cit values of VI

Problem 3.2. Find the coefficient of thermal expansion, o.coex = (l/v) (av/8T)coex'
for a
gas maintained in equilibrium with its liquid phase. Find an approximate explicit
expression for o.coex, using the ideal gas equation of state. Discuss its behavior.
Problem 3.3. Prove that the slope of the sublimation curve of a pure substance at the
triple point must be greater than that of the vaporization curve at the triple point.
Problem 3.4. Consider a monatomic fluid along its liquid-gas coexistence curve.
Compute the rate of change of chemical potential along the coexistence curve,
(dJ..t/dT)coex' where J..t is the chemical potential and T is the temperature. Express your
answer in terms of SJ, Vz and Sg, vg, which are the molar entropy and molar volume of the
liquid and gas, respectively.
Problem 3.S. A system in its solid phase has a Helmholtz free energy per mole,
as == B/Tv3, and in its liquid phase it has a Helmholtz free energy per mole, a; = A /Tv2,
where A and B are constants, V is the volume per mole, and T is the temperature. (a)
Compute the molar Gibbs free energy density of the liquid and solid phases. (b) How are
the molar volumes, v, of the liquid and solid related at the liquid-solid phase transition?
(c) What is the slope of the coexistence curve in the P-T plane?
PrOblem 3.6. Deduce the Maxwell construction using stability properties of the
Helmholtz free energy rather than the Gibbs free energy.
Problem 3.7. For a van der Waals gas, plot the isotherms in the

P-V plane (p and V are

168

THE THERMODYNAMICS OF PHASE TRANSITIONS

the reduced pressure and volume) for reduced temperatures T = 0.5, T = 1.0, and
= 1.5. For T = 0.5, is P = 0.1 the equilibrium pressure of the liquid-gas coexistence
region?

Problem 3.8. Consider a binary mixture composed of two types of particles, A and B.
For this system the fundamental equation for the Gibbs free energy is O=nAJ.tA + nBJ.tB,
the combined first and second laws are dO = -S dT + V dP + J.tAdnA + J-lBdnB (S is the
total entropy and V is the total volume of the system), and the chemical potentials J.tA
and J.tB are intensive so that J-lA = J-lA(P, T,XA) and J-lB = J-lB(P, T,XA), where XA is the
mole fraction of A. Use these facts to derive the relations

xo.dJ.to. = 0
o.:=A,B

sdT - vdP+

(a)

and

L xo.(dJ.to.

o.=A,B
wheres
=
S/n,
v
=
V
[n,
n
= nA
.
with 0 = A,B and (3 = A,B.
Problem 3.9. C
with vapor mix
Clausius-Clape
phases when th

+ nB,

+ so.dT
So.

- vo.dP) = 0,

= (8S/8no.)PT ' ,nf3ia,and

(b)
Va

= (8V /8no.)PT ,

,nO!n

g in equilibrium
alization of the
liquid and vapor
s given by

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com

h
(8S
were So. =
0. P,T,n(3_i<>
0. 0. P,T,n/3ia
[Hint: Equation (b) of Problem (3.8) is useful.]

,B and (3 = A, 8.

Problem 3.10. A PVTsystem


has a line of continuous phase transitions (a lambda line)
separating two phases, I and II, of the system. The molar heat capacity Cp and the
thermal expansivity Op are different in the two phases. Compute the slope (dPldT)coex of
the A line in terms of the temperature T, the molar volume v, D..cp =
c~, and
D..op = cIp - c!f,.
Problem 3.11. Water has a latent heat of vaporization, D..h = 540 cal/ gr. One mole of
steam is kept at its condensation point under pressure at T = 373 K. The temperature is
then lowered to T = 336 K, keeping the volume fixed. What fraction of the steam
condenses into water? (Treat the steam as an ideal gas and neglect the volume of the
water.)
Problem 3.12. A liquid crystal is composed of molecules which are elongated (and
often have fiat segments). It behaves like a liquid because the locations of the center of
mass of the molecules have no long-range order. It behaves like a crystal because the
orientation of the molecules does have long range order. The order parameter for a
liquid crystal is given by the dyatic S = 17(nn - (1/3)1), where n is a unit vector (called
the director) which gives the average direction of alignment of the molecules. The free
energy of the liquid crystal can be written

4-

= o

+ ~ASijSij

~BSijSjkSki

+ ~C SijSijSklSkl

169

PROBLEMS

where A = Ao(T - T*), Ao,B and C are constants, I is the unit tensor so
Xi . I Xj = oij, Sij = Xl' S . Xj, and the summation is over repeated indices. The
quantities Xl are the unit vectors Xl = X, X2 = y, and X3 = Z. (a) Perform the summations
in the expression for 4> and write 4> in terms of TJ, A, B, C. (b) Compute the critical
temperature Te at which the transition from isotropic liquid to liquid crystal takes place,
and compute the magnitude of the order parameter TJ at the critical temperature. (c)
Compute the difference in entropy between the isotropic liquid (TJ = 0) and the liquid
crystal at the critical temperature.
Problem 3.13. The equation of state of a gas is given by the Berthelot equation
(P + a/Tv2)(v - b) = RT. (a) Find values of the critical temperature Te, the critical
molar volume Ve, and the critical pressure Pi; in terms of a, b, and R. (b) Does the
Berthelot equation satisfy the law of corresponding states? (c) Find the critical
exponents /3,0, and, from the Berthelot equation.
Problem S3.1. A boy blows a soap bubble of radius R which floats in the air a few
moments before breaking. What is the difference in pressure between the air inside the
bubble and the air outside the bubble when (a) R = 1em and (b) R = 1mm? The surface
tension of the soap solution is (7 = 25 dynes/em. (Note that soap bubbles have two
surfaces.
Problem S3.2. In
wire frame that ca
Assume that the \i
volume Vand kep

n = no + ns'

STOI Converter

wh,

is the surface area


system. Neglect c
droplet as it is str

pr, placed on a
hperature fixed.
in a fixed total
can be written
~ace tension, A
mainder of the
he of the water

Converted with

trial version
L_

hDP://www.stdutilitv.com
(7

(70

(1 _ .:)n
t' ,

where (70 = 75.5 dynes/em is the surface tension at temperature, t = ODC,n = 1.2, and
t' = 368 DC.(a) Compute the internal energy per unit area of the surface assuming that
the number of surface atoms, N, = O. (b) Plot the surface area and the surface internal
energy per unit area for the temperature interval t = 0 DCto t = i'.
Problem S3.3. Assume that two vessels of liquid He4, connected by a very narrow
capillary, are maintained at constant temperature; that is, vessel A is held at temperature
TA, and vessel B is held at temperature TB. If an amount of mass, D..M, is transferred
reversibly from vessel A to vessel B, how much heat must flow out of (into) each vessel?
Assume that TA > TB
Problem S3.4. Prove that at the tricritical point, the slope of the line of first-order phase
transitions is equal to that of the line of continuous phase transitions.
Problem S3.S. For a binary mixture of particles of type 1 and 2, the Gibbs free energy is
G = nl J.tl + n2J.t2 and differential changes in the Gibbs free energy are dG = -S dT +
V dP + J.tldnl + J.t2dn2. The Gibbs free energy of the mixture is assumed to be
G = nlJ.t?(P, T)

+ n2J.t~(P,

T)

+ RTniin

(Xl)

+ RTn21n

(X2)

+ AnxlX2,

170

THE THERMODYNAMICS OF PHASE TRANSmONS

where J-t? = J-t~are the chemical potentials of the pure substances. In the region in which
the binary mixture separates into two phases, I and II with concentrations x{ and X{l,
find the equation, (ax{ / aT) p for the coexistence curve. Write your answer in terms of
~ and T = T [T; where T; = A/2R.
Problem 83.6. Consider a mixture of molecules of type A and B to which a small
amount of type C molecules is added. Assume that the Gibbs free energy of the resulting
tertiary system is given by
G(P,T,nA,nB,nC)

=nAJ-t~ +nBJ-t~ +ncJ-t~ +RTnAln(xA)


+ RTnc In (xc)

+ RTnBln

(XB)

+ AnAnB/n + AlnAnc/n + AlnBnc/n,

where n = nA + nB + nc, nc nA, and nc


nB The quantities J-t~ = J-t~(p, T),
J-t~ = J-t~(P, T), and J-t~ = J-t~(P, T) are the chemical potentials of pure A, B, and C,
respectively, at pressure P and temperature T. For simplicity, assume that
J-t~ = J-t~ = J-t~. To lowest order in the mole fraction Xc, compute the shift in the
critical temperature and critical mole fraction of A due to the presence of C.

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com

PARTTWO
CONCEPTS FROM
PROBABILITY THEORY

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com

4
ELEMENTARY PROBABILITY
THEORY AND LIMIT THEOREMS

4.A. INTRODUCTION
Thermodynamics is a theory which relates the average values of physical
quantities, such
...
0 on, to one
another. Howev
Converted with
rring at the
microscopic leve
rmodynamic
~r:~~~::i;~:S:

STDU Converter

~~~:~ese;:::

of the element
trial version
ur again and
again throughou
We will begi
counting and
classifying large num ers 0 0 jec s, an we WI give an In U1 ive definition of
probability which will later be justified by the law of large numbers.
Stochastic variables are, by definition, variables whose values are determined by the outcome of experiments. The most we can know about a
stochastic variable is the probability that a particular value of it will be
realized in an experiment. We shall introduce the concept of probability
distributions for both discrete and continuous stochastic variables and
show how to obtain the moments of a stochastic variable by using characteristic functions. As examples we shall discuss in some detail the binomial distribution and its two limiting cases, the Gaussian and Poisson
distributions.
The simple concepts of this chapter can be carried a long way. In the section
on special topics we will use them to study random walks on simple lattices in
one, two, and three spatial dimensions. With these simple examples we can
explore the effects of spatial dimension on a stochastic process.
We have seen in Chapter 3 that there is a high degree of universality in the
thermodynamic behaviour of systems with many degrees of freedom. In
Particular, we found that many systems have the same critical exponents,
regardless of the microscopic structure of the system. This is an example of a
Limit Theorem [1] involving highly correlated stochastic quantities. In this

http://www.stdutiliJV.com

173

174

ELEMENTARY PROBABILITY THEORY AND LIMIT THEOREMS

chapter we will introduce


Limit Theorems
for uncorrelated
stochastic
quantities. We first introduce the concept of an infinitely divisible stochastic
variable and then derive the Central Limit Theorem, which describes the
limiting behavior of independent
stochastic variables with finite variance.
Finally, we study the limiting behavior of independent stochastic variables with
infinite variance which are governed by Levy distributions.
We use the
Weierstrass random walk to show that such systems can develop fractal-like
clustering in space.

4.B. PERMUTATIONS AND COMBINATIONS [2, 3]


When applying probability theory to real situations, we are often faced with
counting problems which are complex. On such occasions it is useful to keep in
mind two very important principles:
(a)

Addition principle: If two operations are mutually exclusive and the


first can be done in m wa s while the second can b done in n ways,
then 0

(b)

Multipl
and aft
perfo
operati

Converted with

STOI Converter

ed in n ways,
d operation is
s, then the two

trial version

hDP://www.stdutilitv.com

These two prin


remainder of
this chapter.
When dealing with large numbers of objects it is often necessary to find the
number of permutations and/or combinations of the objects. A permutation is
any arrangement of a set of N distinct objects in a definite order. The number of
different permutations of N distinct objects is N! The number of different
permutations of N objects taken R at a time is (N!j(N - R)!).
Proof: Let us assume that we have N ordered spaces and N distinct objects with which
to fill them. Then the first space can be filled N ways, and after it is filled, the second can
be filled in (N - 1) ways, etc. Thus the N spaces can be filled in
N(N - 1)(N - 2) x ... x 1 = NI

ways. To find the number of permutations, P~, of N distinct objects taken R at a time, let
us assume we have R ordered spaces to fill. Then the first can be filled in N ways, the
second in (N - 1) ways, ... , and the Rth in (N - R + 1) ways. The total number of ways
that R ordered spaces can be filled using N distinct objects is
~ = N(N - 1) x ... x (N - R

+ 1) = (N

NI
_ R)I.

175

DEFINITION OF PROBABILITY

A combination is a selection of N distinct objects without regard to order. The


number of different combinations of N objects taken R at a time is
(N!/(N

- R)!R!).

proof: R distinct objects have Rl permutations. If we let C~ denote the number of


combinations of N distinct
C~ = Nlj(N - R)IR!

objects taken R at a time, then R1C~ =

11

and

The number of permutations of a set of N objects which contains nt identical


elements of one kind, nz identical elements of another kind, , and nk identical
elements of a kth kind is N!/nt !n2! ... nk!, where nt + nz +
+ nk = N .
EXERCISE 4.1. Find the number of permutations of the letters in the
word, ENGINEERING. In how many ways are three E's together? In how
many ways are (only) two E's together.
Answer: The number of permutations is (11!/3!3!2!2!) = 277,200, since
there are 11 letters but two identical pairs (I and G) and two identical triplets
(E and N). Th
ether = the
number of p
5,120. The
number of wa
= 120, 960,
I

since there
permutations.

STOI Converter

G and its

trial version
4.C. DEFINI

hDP://www.stdutilitv.com

Probability is a quantization of our expectation of the outcome of an event or


experiment. Suppose that one possible outcome of an experiment is A. Then, the
probability of A occurring is P(A) if, out of N identical experiments, we expect
that NP(A) will result in the outcome A. As N becomes very large (N ---+ 00) we
expect that the fraction of experiments which result in A will approach P(A). An
important special case is one in which an experiment can result in any of n
different equally likely outcomes. If exactly m of these outcomes corresponds to
event A, then P(A) = min.
The concept of a sample space is often useful for obtaining relations between
probabilities and for analyzing experiments. A sample space of an experiment
is a set, S, of elements such that any outcome of the experiment corresponds to
one or more elements of the set. An event is a subset of a sample space S of an
experiment. The probability of an event A can be found by using the following
procedure:
(a) Set up a sample space S of all possible outcomes.
(b) Assign probabilities to the elements of the sample space (the sample
points). For the special case of a sample space of N equally likely
outcomes, assign a probability 1/ N to each point.

176

ELEMENTARY PROBABILITY THEORY AND LIMIT THEOREMS

Fig. 4.1. (a) The shaded area is the union of A and B, A U B; (b) the shaded area is the
intersection of A and B, An B; (c) when A and B are mutually exclusive there is no
overlap.

(c)

To obtain the probability of an event A, add the probabilities


elements of the subset of S that corresponds to A.

assigned to

In working ,..:-:w..::..;i
t=h~,-==-==.=...=.L.--=-=-=---=-=-=-==:""""::::'-=-=::""""="'::""::""""=':::""::::..::....L
union of two
t of all points
belonging to A
Converted with
two events is

STOI Converter

denoted An B.
A and B (cf.
Fig. 4.1b). If t
B = 0 where
is the empt
trial version
es of different
We can obt
events. We sha
e outcome of
an experimen
- ,
,
) denote the
probability that both events A and B occur as the result of an experiment;
and finally we shall let P(A U B) denote the probability that event A or event B
or both occur as the outcome of an experiment. Then the probability P(A U B)
may be written

hDP:llwww.stdutililV.com

P(A U B)

P(A)

+ P(B) - P(A n B).

(4.1 )

In wntmg P(A) + P(B), we take the region An B into account twice.


Therefore, we have to subtract a factor P(A n B).
If the two events A and B are mutually exclusive, then they have no points in
common and
P(A

U B) =

P(A)

+ P(B).

(4.2)

If events Al ,A2, ... ,Am are mutually exclusive and exhaustive, then
A I U A2 U ... U Am = S and the m events form a partition of the sample space
S into m subsets. If Al,A2, ... ,Am form a partition, then

P(At)

+ P(A2) + ... + P(Am) = 1.

We shall see Eq. (4.3) often in this book.

(4.3)

177

STOCHASTIC VARIABLES AND PROBABILITY

The events A and B are independent


P(A

n B)

if and

only

if
(4.4)

= P(A)P(B).

Note that since P(A n B) i= 0, A and B have some points in common. Therefore,
independent events are not mutually exclusive events. They are completely
different concepts. For mutually exclusive events, P(A n B) = O.
The conditional probability P(BIA) gives us the probability that event A
occurs as the result of an experiment if B also occurs. P(BIA) is defined by the
equation
P(BIA)

= P(A n B)

(4.5)

P(B)

Since P(A

n B)

= P(B

n A),

we find also that

P(A)P(AIB)

(4.6)

= P(B)P(BIA).

Converted with
The conditional
we use the set J.
I

I
I

STOI Converter

of event A if

trial version

hnp:llwww.s~~ut~~i1v.com

EXERCI~
and B such that P(A) = ~,P(B) =~, and P(A U B)
P(A n B), P(BIA), and P(AIB). Are A and B independent?

(4.7)

-0

hf events A
1. Compute

Answer: From Eq. (4.1), P(A n B) = P(A) + P(B) - P(A U B) = -ts. But
n B) i= P(A)P(B) so A and B are not independent. The conditional
probabilities are P(AIB) = P(A n B)/P(A) = ~ and P(BIA) = P(A n B)/
P(B) = ~.Thus, ~ of the points in A also belong to B and ~ of the points in B
i also belong to A.
i

P(A

4.D. STOCHASTIC VARIABLES AND PROBABILITY


In order to apply probability theory to the real world, we must introduce the
concept of a stochastic, or random, variable (the two words are interchangeable,
but we shall refer to them as stochastic variables). A quantity whose value is a
number determined by the outcome of an experiment is called a stochastic
variable. A stochastic variable, X, on a sample space, S, is a function which
maps elements of S into the set of real numbers, {R}, in such a way that the

178

ELEMENTARY PROBABILITY THEORY AND LIMIT THEOREMS

inverse mapping of every interval in {R} corresponds to an event in S (in other


words, a stochastic variable is a function which assigns a real number to each
sample point). It is useful to note that the statement select at random means that
all selections are equally probable.
In a given experiment, a stochastic variable may have anyone of a number of
values. Therefore, one must be careful to distinguish a stochastic variable
(usually denoted by a capital letter X) from its possible realizations, {Xi}. Some
examples of stochastic variables are (i) the number of heads which appear
each time three coins are tossed; (ii) the velocity of each molecule leaving a
furnace.
In this section we introduce the concept of a probability distribution
function, a probability density function, and a characteristic function which can
be associated with a stochastic variable and which contain all possible
information about the stochastic variable. We also generalize these quantities to
the case when several stochastic variables are necessary to fully describe a
system.

4.D.l. Distr

Converted with

4.D.1.la. Di
Let X be a sto
countable set 0
orn = 00). On
each realizatio
and must satis

STOI Converter
trial version

hDP://www.stdutilitv.com

e that X has a
a finite integer
obability, Pi, to
stribution on S
ned as

Px(x) = LPib(X

(4.8)

- Xi),

i=l

and a distribution

function,

Fx(x), defined as
x

Fx(x) =

-00

dyPx(Y) = ~Pi8(x

- xi).

(4.9)

where 8(x - Xi) is a Heaviside function and has values 8(x) = 0 for X < 0
and 8(x) = 1 for X > O. Note that the probability density function is just
the derivative of the distribution function, Px(x) = dFx(x)/dx.
The distribution function, Fx(x), is the probability that the stochastic variable, X, has
a realization in the interval (-00, x). In order that the probability density,
Px(x), always be positive, the distribution
function,
Fx(x), must be
monotonically increasing function of x. It has limiting values F x( -00) = 0
and Fx(+oo) = 1.

179

STOCHASTIC VARIABLES AND PROBABILITY

EXERCISE 4.3. Consider a weighted six-sided die. Let Xi = i


(i = 1,2, ... ,6) denote the realization that the ith side faces up when the
, die is thrown. Assume that Pi = ,P2 = ,P3 = ,P4 = ~,P5 = ,P6 =
in Eqs. (4.8) and (4.9). Plot Px(x) and Fx(x) for this system.

fz

fz'

Answer:
Px(x)

Fx(x)

0.3

1.0

0.2
0.1

0
I

l-

.1

I'

0.6

"

0.2
6

Note: we have represented the delta function, b(x height equal to Pi.

Xi),

as an arrow with

Converted with
4.D.l.lh.

Con

STOI Converter

Let X be a stoch
trial version
values, such
as an interval on
variable, we
know that an int
assume that
there exists a pi
, x ,
e probability
that X has a value in the interval {a ~ X ~ b} is given by the area under the
curve, Px(x) versus x, between X = a and X = b,

hDP://www.stdutilitv.com

Prob

(a,;;x';;b) =

dxPx{x).

(4.10)

Then X is a continuous stochastic variable, Px(x) is the probability density for


the stochastic variable, X, and Px(x)dx is the probability to find the stochastic
variable, X, in the interval x ~ x + dx. The probability density must satisfy the
conditions Px(x) ~O and
dxPx(x) = 1 (we have assumed that the range of
X is -oo~x~oo)
..
We can also define the distribution function, Fx(x) , for the continuous
stochastic variable, X. As before, it is given by

L':

Fx(x) =

[0

dy Px(y),

(4.11)

and it is the probability to find the stochastic variable, X, in the interval


{-00 ~x}.
The distribution function, FN(X), must be a monotonically

180
increasing

ELEMENTARY PROBABILITY THEORY AND LIMIT THEOREMS

function

of

x, and it has limiting

Fx( -00) = 0 and

values

Fx(+oo) = 1.
EXERCISE
4.4. Plot the probability density, Px(X) , and the distribution function, Fx(x), for the Gaussian probability
density, Px(x) =
(1 /2v'21r)e-x2 /8.

Answer:
Fx(x)

Px(x)

-6

-4

-2

6 X

Converted with
Often we wi
X, but for som
function of X.
defined as

STOI Converter
trial version

hDP://www.stdutilitv.com
where 8(y - H(x))

astic variable,
) is a known
variable, Y, is

(4.12)

is the Dirac delta function.

4.D.2. Moments
If we can determine the probability density, Px(x), for the stochastic variable,
X, then we have obtained all possible information about it. In practice, this
usually is difficult. However, if we cannot determine Px(x), we can often obtain
information about the moments of X. The nth moment of X is defined

(xn)

= [,,'

dxX'Px{x).

(4.13)

Some of the moments have special names. The moment, (x), is called the mean
value of X. The combination, (xl) - (X)2, is called the variance of X, and the
standard derivation of X, ax, is defined as
(4.14 )

181

STOCHASTIC VARIABLES AND PROBABILITY

The moments give us information about the spread and shape of the probability
density, Px(x). The most important moments are the lower-order ones since
they contain information about the overall behavior of the probability density.
We give some examples below.

4.D.2.1a. The First Moment


The first moment, (x), gives the position of the "center of mass" of the
probability density, Px(x). It is sometimes confused with two other quantities,
the most probable value, xP' and the median, xm The most probable value, xP'
locates the point of largest probability in Px(x). The median, xm, is the value of
x which divides the area under the curve, Px(x) versus x, into equal parts. In
other words, Fx(xm) =
In Exercise (4.4), because of the symmetric shape of the Gaussian
distributions shown there, the mean, (x), the most probable value, xP' and the
median, xm, all occur at the same point, x = O. In Exercise 4.5, we give an
example in which they all occur at different points.

!.

EXERCISIf---AI_.l;L___I____.c,._~---.L......L____.~....-...dL......,_~-.....l~--..L--'--'-I

i Px(x),

Converted with

and the

STOI Converter

trial version

hDP:llwww.stdutililV.com
-1.5

-1.0

-0.5

0.5

1.0

-1.5

I
I

Answer: The mean is (x)


= -1.0. The median is Xm

xp

= -0.5625.
= -0.862.

-1.0

-0.5

0.5

The most probable

1.0

1.5 X

value is

4.D.2.1h. The Second Moment


The second moment, (.r), gives the "moment of inertia" of the probability
about the origin. The standard deviation, ax [cf. Eq. 4.14] gives a measure of
how far the probability spreads away from the mean, (x). It is interesting to
consider some examples. In Exercise (4.4), (x2) = 4, (x) = 0, and ax = 2. In
Exercise 4.5, (.r) = 0.966, (x) = -0.5625, and ax = 0.806.

4.D.2.1c. The Third Moment


The third moment, (x3), picks up any skewness in the distribution of probability
about x = O. This can be seen in Exercise 4.4 and 4.5. For Exercise (4.4),
(x3) = 0 because the probability is symmetrically distributed about x = O.
However, for the probability distribution in Exercise 4.5, (r) = -0.844,
indicating that most of the probability lies to the left of x = o.

182

ELEMENTARY PROBABILITY THEORY AND LIMIT THEOREMS

4.D.3. Characteristic Functions


The characteristic
defined as

fx(k}

function,ix(k),

corresponding

= (eikx) = [,'

to the stochastic variable, X, is

dxeikxpx(x}

to

(ik}~?") .

(4.15)

The series expansion in Eq. (4.15) is meaningful only if the higher moments,
(X'), are small so that the series converges. From the series expansion in Eq.
(4.15) we see that it requires all the moments to completely determine the
probability density, Px(x). Characteristic functions are continuous functions of
k and have the property that ix(O) = 1, I ix(k) I ~ 1, and ix( -k) = ix*(k) (*
denotes complex conjugation). The product of two characteristic functions is
always a characteristic function.
If the characteristic function is known, then the probability density, Px(x), is
given by the inverse transform
(4.16)

Converted with
Furthermore, i
differentiating:

STOI Converter
trial version

in moments by

(4.17)

hDP:llwww.stdutililV.com
Equation (4.17) provides a simple way to obtain moments if we know ix(k).
It is often useful to write the characteristic function, ix(k), in terms of
cumulants, Cn(X), rather than expand it directly in terms of moments. The
cumulant expansion is defined

ix(k)

= exp

00

Ckt
7

Cn(X)

(4.18)

where Cn(X) is the nth-order cumulant. If we expand Eqs. (4.15) and (4.18) in
powers of k and equate terms of the same order in k, we find the following
expressions for the first four cumulants:
Cl

(X) = (x),

C2(X) = (x2) - (X)2,


C3(X) = (x3) - 3(x)(~)
and

(4.19)
(4.20)

+ 2(x)3,

(4.21 )

183

STOCHASTIC VARIABLES AND PROBABILITY

If higher-order

cumulants rapidly go to zero, we can often obtain a good


approximation tolx(k) by retaining only the first few cumulants in Eq. (4.18).
We see that CI (X) is just the mean value of X and C2(X) is the variance .

EXERCISE 4.6. Consider a system with stochastic variable, X, which


hj probability density, Px(x), given by the circular distribution; Px(x) =
~ 1 - x2 for Ixi ~ 1, and Px(x) = 0 for Ixi > 1. Find the characteristic
function and use it to find the first four moments and the first four cumulants.
Answer: The characteristic

function is

2fl

dxe' "kx.~v l - x2 = -kJ1(k),

Ix(k) = -

7r

-1

(cf. Gradshteyn and Ryzhik [8]) where JI


expand Ix (k) in powers of k

(k) is a Bessel function. Now

Converted with
, From Eq. (4.1
The cumulents

STOI Converter
trial version

4.D.4. Jointly

hDP://www.stdutilitv.com

The stochastic varraeres;: ~Ar]l-;-:,


A~:2Z;-,-:-. .: --;-,~Aln,;-aiarre~e:J}(JUlmlnmlu~:y{lffi
U1IsrniUlrImoU1Itilleoo:(
if they are
defined on the same sample space, S. The joint distribution function for the
stochastic variables, X1,X2, .. ,Xn, can be written

where {Xl < Xl, ... ,Xn < Xn} = {Xl < xI} n n {Xn < xn}. In other words,
Fx), ...x, (Xl, ... ,xn) is the probability
that simultaneously
the stochastic
Variables, Xi, have values in the intervals {-oo < Xi < Xi} for i = 1, ... , n.
The joint probability density, PX1, ...x; (Xl, ... ,xn) is defined as

Xl, ... ,x"

Xl,

,Xn

) _ anFx" ...,x" (Xl, ... ,xn)


Xl'"
Xn

'

(4.24)

so that
(4.25)
For simplicity, let us now consider the joint distribution function, FX,y(x, y),
and joint probability density, PX,y(x,y),
for two stochastic variables, X and y.

ELEMENTARY PROBABILITY THEORY AND LIMIT THEOREMS

184

distribution
function
satisfies
the
conditions
Fx,y( -oo,y)
Fx,Y(x, -00) = Fx,Y( -00, -00) = 0 and Fx,y( 00,00) = 1. Furthermore,

:::::

The

for

X2

>

Xl.

(4.26)
The probability density,
is normalized to one:

PX,y

(x, y), satisfies the condition 0 ~ PXY (x, y) ~ 1, and

[x,

dyPX,y{x,y)

dx ['"

If we want the reduced distribution

(4.27)

= I.

function, Fx(x),

for the stochastic variable,

X, it is defined as
Fx{x)
Similarly, the
is defined as

= Fx,Y{x,oo)

= ['" dx' ['"

dyPx,y{x,y).

(4.28)
stic variable, X,

Converted with

STOI Converter

(4.29)

trial version

hDP://www.stdutilitv.com

We can obtai
probability deTIJ."..,..wry_"..,---..---r-"rT.,I'---rr->
I~~~~~~~
The nth moment of the stochastic variable, X, is defined as

d the reduced

___J

(4.30)

(x") = ['" dx ['" dyx"PX,Y{x,y).


Joint moments of the stochastic

(x"'y")

= ['"

variables, X and Y, are defined as

There are two related joint moments


literature. They are the covariance,
Cov (X, Y)
and even more commonly

(4.31 )

dx ['" dyx"'y"Px,y{x,y).

((x - (x)(y

that are commonly

- (y))

used in the physics

= (xy) - (x)(y),

(4.32)

the correlation junction

Cor (X, Y) = ((x - (x) )(y - (y)


(jx(jy

) ,

(4.33)

185

STOCHASTIC VARIABLES AND PROBABILITY

where ax and ay are the standard deviations of the stochastic variables X and Y,
respectively. The correlation function, Cor (X, Y), is dimensionless and is a
measure of the degree of dependence of the stochastic variables X and Yon one
another (cf. Exercise 4.7). The correlation function has the following properties:
(i) Cor (X, Y) = Cor(Y, X).
(ii) -I ~Cor (X, Y) ~
(iii) Cor (X, X)

1.

= I,Cor(X,-X)

+ b, cY + d)

(iv) Cor (aX

=-1.

= Cor (X, Y) if a, c

-# 0.

The notion of joint correlation function can be extended to any number of


stochastic variables.
For two stochastic variables, X and Y, which are independent, the following
properties hold:
(i') PX,y(x,y)

Px(x)Py(y).

(ii') (XY) = (X)(Y).


(iii')

((X

+Y

~--------~--------~--------~

Converted with

(iv') Cor (X,


Note that the cc
does not always

EXERCIS

STOI Converter

(X, Y) = 0, it

trial version

hDP:llwww.stdutililV.com

(X, Y), is a

; measure of the degree to which X depends on Y.


: Answer: Let X = aY + b and choose a and b to minimize the mean square
i error,
e = ((x - ay - b)2). Set be = (8e/8a)8a + (8e/8b)8b = 0, and set
i the coefficients
of 8a and Bb separately to zero. This gives two equations,
-2(xy) + 2a(i) + 2b(y) = and -2(x) + 2a(y) + 2b = 0. Eliminate band
I solve for a to find a = -Cor
(X, Y)ax/ay. Thus when Cor (X, Y) = 0, a =
, and the random variables, X and Y, appear to be independent, at least when
viewed by this statistical measure.

When we deal with several stochastic variables, we often wish to find the
probability density for a new stochastic variable which is a function of the old
stochastic variables. For example, if we know the joint probability density,
PX,y(x,y), we may wish to find the probability density for a variable
Z
G(X, Y), where G(X, Y) is a known function of X and Y. The probability
density, Pz(z), for the stochastic variable, Z, is defined as

Pz(z) = ['" dx ['"

dy6(z - G(x,y))PX,y(x,y).

(4.34)

186

ELEMENTARY PROBABILITY THEORY AND LIMIT THEOREMS

From Eq. (4.34), the characteristic


easily found to be

fz(k) =

function

for the stochastic

variable, Z, is

J~oodx J~oo

dyeikG(x,y)pX,y(x,y).

(4.35)

EXERCISE 4.8. The stochastic variables, X and Y, are independent and


are Gaussian distributed with first moments, (x)
(y)
0, and standard
deviations,
(J'x (J'y 1. Find the joint distribution
function for the
stochastic variables, V
X + Y and W
X - Y. Are Vand W independent?

=
=

Answer: The probability densities for X and Yare Px(x)


(1/v'27r)e-(1/2)x2
and Py(y)
(11 v'27r)e-(1/2)y2. Since X and Yare independent, Eq. (4.34)

gives

Pv,w(v, w)

1 roo
roo dy8(v
= "'....
dx

- v'(x,y))8(w

- w'(x, y))e-(1/2)(xZ+y2) ,

Converted with

STOI Converter

where v'(x,y
i

"

(1)

8(v - v'(x,

- y'(v, w)),

trial ve rsion

=!

hDP://www.stdutilitv.com

=!

where x'(v, w)
(v + w), y'(v, w)
(v - w), and
Jacobian of the coordinate transformation. Thus,

P
Vand Ware independent

(2)

Ie"'"

J =!

w) -- _!_
e-(1/4)(v2+w2)
47r
.

(v

v,w,

... ,kN)

=Joo
-00

x
The joint moment,

(3)

since Pv,w(v, w) is factorizable.

We can also introduce characteristic


functions for jointly
stochastic variables. They are often useful for obtaining moments.
[x; ...,xN(kl,

is the

dx1

.Joo

dxNei(ktxt++kNXN)

-00

PXt, ...,XN(Xl,

...

,XN).

(Xi' .. Xn) (n ~N), is then given by

distributed
We define

(4.36)

187

STOCHASTIC VARIABLES AND PROBABILITY

One of the most important multivariant distributions, and one which we will use
repeatedly in this book, is the multivariant Gaussian distribution with zero
mean. Some of its properties are discussed in Exercise (4.9).

EXERCISE 4.9. The multivariant Gaussian distribution with zero mean


can be written

det(g)

--e

-(1/2)xT.g.x

(21ft

'

where g is a symmetric N x N positive definite matrix, x is a column vector,


and the transpose of x, xT = (Xl, ... ,XN), is a row vector. Thus,
xT . g. x = 2:~12:!1 gjjXjXj. (a) Show that PXl, ...,xN(Xl,
,XN) is normalized to one. (b) Compute the characteristic function.j-; ,xN(kl, ... ,kN). (c)
: Show that the moments, (Xj) = 0 and that all higher moments can be
expressed in terms of products of the moments, (xl) and (XjXj). This is the
simplest form of Wick's theorem used in field theory.
!
I

Converted with

Answer:

STOI Converter

(a) Since -

trial version

1. We can

hDP://www.stdutilitv.com
, ,
let ~ = o x = (al, ... ,aN). Since

rjj = 0 for
ic variables.
det (0) = 1, the
Jacobian of the transformation
is one. Thus, dxl dx; =
da, ... do, Furthermore, xT g . x = ~T t ...~.
Thus,

1-..J

dx, ...

tUNe-(1/2).T.g

-x

1-..J

dal

(21f)N
=

... daNe-(1/2)

'~"a~

since det(g) = det(t) = II ...In.


(b) The characteristic function is

OO

where kT

dx, ...

dxNeikT.Xe-(1/2)xT.g.x,

-00

(kl' ... ,kN). If we transform the integral into diagonal

188

ELEMENTARY PROBABILITY THEORY AND LIMIT THEOREMS

form, as we did in part (a), it is straightforward to show that


_

[x, ,...,XN (kl' ... , kN) - e

-(1/2)kT.

g-I

.k

(Xi) = limki~O ( - i) (8/ 8ki )!xl ,...,XN (kl,""


kN) =
0, since the derivative brings down a factor of kj. In fact, all odd
moments will be zero for the same reason. The second moment is
easily seen to be (XiXj) = (g-l )ij' Inspection shows that all higher
moments depend only on sums of products of factors of the form
(g-l )ij' and therefore only on sums of products of second moments.
More generally, the average of a product of 2n stochastic variables is
equal to the sum of all possible combinations of different pairwise
averages of the stochastic variables. For example,

(c) The first moment is

(XIX2X3X4)

= (XIX2) (X3X4)

+ (XIX3)

(X2X4)

+ (XIX4)

(X2X3).

(Note that (x;Xj) = (XjX;).) More generally, if we have 2n stochastic


variables, the number of terms in the expression for the 2nth moment,

is (2
( (2n)(
((2n
all the
will

STOI Converter
trial version

There are
at there are
so on. After
on rule, there
n) different

hDP://www.stdutilitv.com

because they
are di
,
tal number of
different terms in the expression for (XIX2 ... X2n) is (2n)!/n!2n.

4.E. BINOMIAL DISTRIBUTIONS


A commonly found application of probability theory is for the case of a large
number, N, of independent experiments, each having two possible outcomes.
The probability distribution for one of the outcomes is called the binomial
distribution. In the limit of large N, the binomial distribution can be
approximated by either the Gaussian or the Poisson distribution, depending
on the size of the probability of a given outcome during a single experiment. We
shall consider all three distributions in this section. We shall also use the
binomial distribution to find the probability density for a random walk in one
dimension.

4.E.l. The Binomial Distribution [2-5]


Let us carry out a sequence of N statistically independent trials and assume that
each trial can have only one of two outcomes, 0 or + 1. Let us denote the

189

BINOMIAL DISTRIBUTIONS

probability of outcome, 0, by q and the probability of outcome, + 1, by p so that


p + q = 1. In a given sequence of N trials, the outcome, 0, can occur no times
and the outcome, + 1 times, where N = no + nl. The probability for a given
permutation of no outcomes, 0, and nl outcomes, + 1, is qrlOpnlsince the N trials
are statistically independent.
The probability for any combination of no
outcomes, 0, and n I outcomes, + 1, is
(4.38)
since a combination of nz outcomes, 0, and nl outcomes, + 1, contains
(N!/no!nl!) permutations.
Equation
(4.38) is often called the binomial
distribution even though it is not a distribution function in the sence of
Section 4.D. From the binomial
theorem,
we have the normalization
condition
N

LPN(nl)
nl=O

N
N!
L---,pnl~-nl

(p+q)N

= 1.

(4.39)

Converted with
We can vie
variable descri
x = with pro
for the ith trial
the ith trial is

STOI Converter

trial version

the stochastic
o realizations:
ability density
tic function for

hDP://www.stdutilitv.com

t.

EXERCISE 4.10. The probability of an archer hitting his target is If he


shoots five times, what is the probability of hitting the target at least three
times?
I
: Answer:
j

t,

be the number of hits. Then N = no + nl = 5,p = and


of having
nl hits
in N = 5 trials
is
s nl
- The probability of at least three hits
= Ps(3) + Ps(4) + Ps(5) =
0.21.
Let

nl

; q =~. The probability


Ps(nt) = (5!/nd(5 - n] )!)

nt m
i23 ~

We will now let YN = Xl + X2 + ... + XN be the stochastic variable which


describes the additive outcome of N trials. Since the trials are independent, the
probability density for the stochastic variable, YN, is just
PYN(Y) =

J- .. Jdxl

... dxN8(Y-Xl-

... -XN)PXI(Xl)

x XPXN(XN).
(4.40)

ELEMENTARY PROBABILITY THEORY AND LIMIT THEOREMS

190

From Eq. (4.35), the characteristic


fyN(k)

=!xl(k)
We can expand
obtain

dx[

function for the random variable,

YN, is

x ... x PXN(XN)

dxNeik(x'+x'+"+XN)px,(xt)

(4.41 )

x x!xN(k) = (q+peik)N.

the characteristic

function

using the binomial

theorem.

We

(4.42)

The probability

density,

PyN(x), for the stochastic variable, YN, is

(4.43)

Converted with
where
trials.

PN(nt) i

STOI Converter

nt times in N

trial version

hDP://www.stdutilitv.com

ation or by

Also, it is easy to see that

(y) = lim(-i)aakfrN(k)
k_..O

=pN=

(nt).

(4.45)

In a similar manner, we obtain for the second moment


N

(i) = (ni) =

L niPN(nt)

= (Np)2 + Npq.

(4.46)

nl=O

The variance is given by

(ni) - (nt)2 = N pq, and the standard deviation is


(4.47)

191

BINOMIAL DISTRIBUTIONS

P10(nt)
0.3

0.2
0.1

nl

10

Fig. 4.2. The binomial distribution for p = ~and N = 10; for this case, (nl) = lj!.

The fractional deviation is


(4.48)

Converted with
trials with outc
trials. A small
N

STOI Converter
trial version

--+ 00, aN / (

case N

10

tion, nt/N, of
sequence of N
close to p. For
ution, for the

hDP://www.stdutilitv.com

4.E.2. The Gaussian (or Normal) Distribution


Let us make a change of stochastic variables, ZN = (YN - (y)) / ay =
((YN - PN)/ ..jjiijN). ZN is a stochastic variable whose average value, (z),
equals 0 when (y) = pN. The probability density for the stochastic variable, ZN,
is
(4.49)
The characteristic function is

where irN(k)

is defined in Eq. (4.41) and we make use of the fact that


--+ 00. We first expand the quantity in

q == 1 - p. We now will take the limit N

192

ELEMENTARY PROBABILITY THEORY AND LIMIT THEOREMS

brackets in powers of k and obtain


(4.51 )

where
(4.52)

As N ___.00, we have RN ___.0 and

where we have used the definition


lim

N__.oo

(1 + N~)N = e

(4.54)

Converted with

Thus, in the lin

STOI Converter

(4.55)

trial version
It is easy to sh
t the standard
deviation, oz. e._____hD_P:_"_WWW
__
For the case N 1 but N finite, we still have PZN(z) ~ (1/V2ir)e-z2/2.
Transform back to the stochastic variables, y = GyZ + (y). Then

.S_td_u_t_il_itv_._C_Om_------'

PYN(y) ~

oo

dz8(y -

GyZ -

(y) )Pz(z)

-00

=
Gy

1
J27i=ex
p

27r

((y

- (y)
2

)2)

(4.56)

Equation (4.56) is the Gaussian probability density for the number of outcomes,
+ 1, after many trials. It is also the most general form of a Gaussian probability
density. It is important to note that the Gaussian probability density is entirely
determined in terms of the first and second moments, (y) and (i) (since
Gy = J(y2) - (y)\
In Exercise 4.4, we have plotted a Gaussian probability
density for (y) = 0 and Gy = 2.

4.E.3. The Poisson Distribution


The Poisson distribution can be obtained from the binomial distribution in the
limit N ___.00 and p ___.0, such that Np = a N (a is a finite constant). Let us

193

BINOMIAL DISTRIBUTIONS

return to Eq. (4.41) and let p = (a/N).

Then

(4.57)
If we now take the limit, N
fy(k)

== lim frN(k)
N-+oo

am

= lim (1
N-+oo

L-, e
00

= e-a

--+ 00,

we obtain

+!:
(eik
N

1))N = exp(a(eik

1))

(4.58)

1mk

m=Om.

Thus the probability density is

(4.59)
The coefficient

Converted with

STOI Converter
trial version

is commonly c

(4.60)
for finding

hDP://www.stdutilitv.comtcome.+I.in

nl

outcomes, + 1,
a single trial),
,
, equals a. The
Poisson distribution depends only on the first moment, and therefore it is
sufficient to know only the first moment in order to find the probability density
for a Poisson process. In Fig. 4.3, we plot the Poisson distribution for a = 2.

0.2
0.1

I .
o

Fig. 4.3. The Poisson distribution for (n) = a = 2.

194

ELEMENTARY PROBABILITY THEORY AND LIMIT THEOREMS

EXERCISE 4.11. A thin sheet of gold foil (one atom thick) is fired upon
by a beam of neutrons. The neutrons are assumed equally likely to hit any
part of the foil but only "see" the gold nuclei. Assume that for a beam
containing many neutrons, the average number of hits is two. (a) What is the
probability that no hits occur? (b) What is the probability that two hits
occur?
Answer: Since the ratio (area nucleus/area atoms) ~ 10-12, the probability
of a hit is small. Since the number of trials is large, we can use the Poisson
distribution. Let nl denote the number hits. Then (nl) = 2 and the Poisson
distribution can be written P(nl) = e-22nt Ind.
(a) The probability that no hits occur is P(O) = e-22 /O!
(b) The probability that two hits occur is P(2) = e-222/2!

4.E.4.

BinoJ.I.!~1L!Utl.d..il~....t.&l..!!:aJ..IL

The problem

distribution to
move along the
the right and a
assume that th

= 0.135.
= 0.27.

--,

Converted with

f the binomial

trial version

constrained to
f length, A, to
the left. Let us
f each step is

STOI Converter
hDP:llwww.stdutililV.com

independent of
dent). For the
ith step, let the
, ,
"
for a step to
the right and x = - A for a step the left. The probability density for the ith step
is Pxi(x) = (8(x + A) + 8(x - A)). The characteristic function for the ith step
is IXi (k) = cos(kA).
The net displacement, YN (along the x axis), of the particle after N steps is
given by YN = Xl + ... + XN. The characteristic function for the stochastic
variable, YN, is

fr(k)=(cos(kA))
N

N = ( 1--+
k2A2
)N ~1---+
Nk2A2
2!
2!

(4.61 )

[cf. Eqs. (4.40) and (4.41)]. The first and second moments are Jy) = 0 and
(i) = NA2, respectively, and the standard deviation is uY = A.,fN. In Fig. 4.4
we show three realizations of this random walk for N = 2000 steps and A = 1.
There are 2N possible realizations for a given value of N.
We can find a differential equation for the characteristic function in the limit
when the step size, ~, and the time between steps, T, become infinitesimally
small. Let us write the characteristic function,fYN(k) =.[y(k, NT), where t = Nr
is the time at which the Nth step occurs. Also note that initially (N = 0) the
walker is at y = O. Thus, PYo(y) = 8(y) and Jy(k, 0) = 1. From Eq. (4.61), we
N

BINOMIAL DISTRIBUTIONS

195

-. , ...!
"'. ....
." .;

.
<Co

60

N(y)

40

20

.. .,-;..
...... .. ...
.~. ;..,...: .
:I~
~
I

60

N(y)
40
20

.....
.......
:\.. ,.
. ..~ ,.

60

N(y)
40

rl ..

It .....

')('1

Converted with

STOI Converter

Fig. 4.4. Three


q
step
times the walke r
walker starts at

= =!,

trial version

~ dom walk for


~ the number of
In each case the

hDP://www.stdutilitv.com

can write
fy(k, (N

+ l)r)

- fy(k,Nr)

(cos(k~) - l)fy(k,Nr)

12-~2
)
( --2-+'" fy(k,Nr),

(4.62)

where we have expanded cos(k~) in powers of k~. Now take the limits
N _, 00, r _, 0, and ~ _, 0 so that D = (~2 /2r) is finite and Nr = t. If we
note that
li fy(k, (N
11m
1m
N-oor-O

+ 1)r)

- fy(k, Nr)

afy(k, t)
= ...,;",,__;_,,;_~
(4.63)
at

and
lim lim lim

N-oor-OLl_O

(_12-2~2r + ...)fy(k,Nr)

= -D~fy(k,t),

(4.64)

196

ELEMENTARY PROBABILITY THEORY AND LIMIT THEOREMS

then combining these equations we obtain


afy(k, t) = -Dk2fy(k,
at

t).

(4.65)

For the case fy(k, 0) = 1, Eq. (4.65) has the solution


fy(k, t) = e-DJc2t

(4.66)

The probability distribution is

P(y, t) = -21
~

Joo

dke-lky e-Dk

-00

2) .

-4 exp ( - 4Y
~Dt
Dt

(4.67)

The probability de~


for the random walker is a Gaussian whose standard
deviation is a = y'2Dt. The probability density spreads with time.
We can derive Eq. (4.67) in a completely different way. If we make the
change of sto
.
.
characteristic
function, fZN(k)
Converted with

STOI Converter

(4.68)

trial version

[cf. Eq. (4.50)].

00

hDP://www.stdutilitv.com

and find
(4.69)

and therefore Pz(z) = (1/V27r)exp( -Z2 /2). For very large N, we can write
PZN(z) ~ Pz(z). The probability density for the net displacement, YN, then
becomes

PyN(y) ~

dz8(y -llvNz)pz(z)

OO

-00

1 2
J2~IlN

(I )

exp - 21l2N

(4.70)

If the particle takes n steps per unit time, then N = nt and the probability
density that the particle lies in the interval, y ~ y + dy, along the x axis, at time
t is

P(y, t) ~

OO

-00

dz8(y -llvNz)pz(z)

1
= ~exp
y4~Dt

(I )
- 4D

'

(4.71 )

where D = nll2 is the diffusion coefficient for the random walk. The
probability density as a function of time is plotted in Fig. 4.5. Initially, P(y, t)
has the form of a delta function, 8(y). As time passes, it spreads in such a way

197

A CENTRAL LIMIT THEOREM AND LAW OF LARGE NUMBERS

0.4

P(y,

t)

0.3

4--t= 1
0.2
0.1
-20

oy

-10

20

Fig. 4.5. A plot of the probability density, P (y, t) [cf. Eq. (4.71)], for D
t = 1,10,100.

= ~,

that the total area underneath the curve is equal to one, but the probability of
finding the particles at some distance along the x axis grows with time since the
standard deviati
t of the time.

Converted with

4.F. ACentr

STOI Converter

trial version
It can be sho
istribution of
outcomes of a I
aussian form
(provided the
np:
nts are finite).
This is called the Central Limit Theorem. The central limit theorem has been
proved in complete generality in Section S4.C. Below we give a slightly weaker
version, directly applicable to measurements. This result shows why Gaussian
distributions are so widely seen in nature. Another result of considerable
importance is the law of large numbers. The law of large numbers gives
quantitative justification to the use of probabilities.

II WWW.std utlltv.com
-1-

4.F.l. A Central Limit Theorem


The Central Limit Theorem (cf. Section S4.C) can be stated in a simplified form
which has direct relevance to the statistics of measurements. Let us consider a
stochastic variable, YN, which is the deviation from the average of N statistically
independent measurements of a stochastic variable, X. We write YN as
(4.72)
Where
Z,

= -(Xi
N

- (x))

198

ELEMENTARY PROBABILITY THEORY AND LIMIT THEOREMS

The characteristic function, fz(k; N),


(l/N)(Xi - (x)), can be written

fz(k;N)

= ['"

dxei(k/N)(x-(x))px(x)

for the stochastic

= 1 - ~~

variable, Zi

ai + ... ,

(4.73)

ai

where
= (.xl) - (x)2. For large N and finite variance, higher-order terms in
the expansion of the right-hand side can be neglected. (Note that the oscillatory
character of the integrand ensures that the integral goes to zero for large k.) The
characteristic function for YN is then

k2

1
Ir (k)= ( l---ai+
N
2~

)N -+exp (k2~)
__

mx

In Eq. (4.74), we have used the identity limN_oo(l

as

+ x/N)N

N -+

(4.74)

00.

= e. Thus

(4.75)
as N -+ 00. R
average of a I
be a Gaussian
the standard de

STOI Converter

moments, the
ents of X will
s 1/..jN times

trial version

hDP://www.stdutilitv.com

4.F.2. The Law of Large Numbers [3, 4]


The law of large numbers underlies the intuitive concept of probability that we
introduced in Section 4.C. Much of the content of the law of large numbers is
contained in the Central Limit Theorem as applied to measurements. We will
give a simple version of it here.
The law of large numbers applies to N independent experiments and may be
stated as follows: If an event, A, has a probability, p, of occurring, then the
fraction of outcomes, A, approaches p in the limit N -+ 00. The proof has two
steps. The first step involves the derivation of the Tchebycheff inequality. The
second step uses this inequality do derive the law of large numbers.
The Tchebycheff inequality establishes a relation between the variance and
the probability that a stochastic variable can deviate by an arbitrary amount, E (E
is positive), from its average value. The variance, az, of a stochastic variable, Z,
is written
(4.76)
If we now delete that range of the variable, z, for which

Iz -

(z)

I~

we can

199

SPECIAL TOPICS: LATIICE RANDOM WALK

write

(Z)-c
O"~ ~

dz(z - (Z))2 PZ(z)

(J

JOO

(z)-c

dz(z - (Z) )2PZ(Z).

(4.77)

(z)+c

-00

Since under the integral


(4.77) and write

0"~~c2

[z - (z) I ~ c2, we can replace

dzPz(z)

00

dzPz(z)

= c2p(lz

(z -

(Z))2 by c2 in Eq.

- (z)1 ~c),

(4.78)

(z)+c

-00

where P( [z - (z) I ~ c) is the probability that the stochastic variable, Z, deviates


from (z) by more than c. From Eq. (4.78), we obtain the Tchebycheff
inequality
0"2

P(lz - (z)1 ~c) ~

-t.
e

(4.79)

Converted with

Thus, for fixed

STOI Coover Ier

value by more
We now cor
measurements
outcomes, YN

trial version

om its average

N independent
n value of the
me of the ith

hDP:llwww.stdutililV.com

measurement.
ability that YN
deviates
from
v-t
I:>
,
the limN_oo
P( IYN - (z) I ~ c) = O. To prove this, let us first note that (YN) = (z). Since
we have independent events, the variance, O"YN' behaves as o}.N = ~/N. We
now use the Tchebycheff inequality to write
2

P(IYN - (z)1 ~c) ~

O"y

O"z

-T
= -2
e
Nc

(4.80)

Thus, we find

lim P(IYN - (z) I ~ c) = 0,

N-oo

provided that

az

(4.81)

is finite .

.....SPECIAL TOPICS
.....S4.A. Lattice Random Walk [7]
Random walks on lattices provide some of the simplest examples of the effect
of spatial dimension on physical processes. This type of random walk is so

200

ELEMENTARY PROBABILITY THEORY AND LIMIT THEOREMS

- - -.__-.-4. - .......- - .....-- .

-N

-N+1

-1

Fig. 4.6. A one-dimensional lattice with period 2 N


1= N, it enters from the left at site 1= -N.

N-1

+ 1. If the walker

steps right at site

simple that we can use the concepts of the previous sections to compute the
probability that the walker will reach any other part of the lattice. This section is
based on a discussion of lattice random walks by E. Montroll [7]. We will find
that random walks can have quite different properties in three dimensions than
in one or two dimensions .

.....S4.A.l. One-Dimensional Lattice


Let us consider a random walker on a periodic one-dimensional
lattice with
2N + 1 lattice sites (cf. Fig. 4.6). Let Ps(l) be the probability to find the walker
at site, I, at discrete time, s. Since the lattice is periodic, Ps(l) = Ps(l
[2N + 1]). If th
gain from the
left at site -N.
right with prob
probability

den

Converted with

one step to the


ity q = The

!-

STOI Converter

(4.82)

trial version

hDP://www.stdutilitv.com

s and that the

L bl,11+12+'+lsp(lI)p(12)

x ... x p(ls).

(4.83)

ls=-N

Since Ps(l) is a periodic function of I, we can expand it in a Fourier series

Ps(l) = 2N

+ 1 n~NJs(n)

exp

(27rinl
- 2N

+1

'

(4.84 )

where Js(n) is the Fourier amplitude and is also the characteristic function for
this stochastic process. We can revert this equation and write the Fourier
amplitude in terms of Ps(l),

Js(n) = l~

Ps(l) exp

(27rinl

+ 2N + 1)

(4.85)

201

SPECIAL TOPICS: LATTICE RANDOM WALK

We have used the identity

~Z:: exp (27rinl


-- ) = (2N
2N + 1

+ 1)8n,o.

I=-N

We can also Fourier transform the transition probability, p(l) (cf. Eq. (4.82)), to
obtain

~
(27rinl
)
( 27rn )
A(n) == If=NP(l) exp + 2N + 1 = cos 2N + 1 '
where
write

A(n) is the Fourier amplitude.

(4.86)

From Eqs. (4.83) and (4.85), we can

/s(n) = (A(n)Y
for the characteri
Let us now

(4.87)

Converted with

cp= (27rn/(2N

the variable

STOI Converter
trial version
and the probabi

hDP://www.stdutilitv.com
Ps(l)

="21 J1T
7r

dcp!s(cp)e-il

(4.88)

-1T

and
00

/s(cp) =

L Ps(l)e

i/

(4.89)

1=-00

The single-step

probability

becomes

(4.90)
From Eq. (4.87) we obtain

/s(cp) = (A(cp)y
(Note that if the walker starts at I = 0 at time s = 0, then fo(

(4.91)

cp) = 1.)

202

ELEMENTA.RY PROBABILITY THEORY AND LIMIT THEOREMS

It is now easy to compute Ps(l) for a random walk on a one-dimensional


lattice. If the walker starts from site, I = 0, we obtain

P,(l)

d</>(cos(</)'e-il

= 2~ [.

{?)' (S;l)~hl}

for s -I

even and s~l,

(4.92)

otherwise.
It is useful to ask another question about the walker. What is the probability
that the walker will escape from the origin? To formulate this in a simple way,
we first introduce a generating function, U (z, I), for P s (1):

To obtain the
have summed (
equation

e-il

1 J71'

00

U(z, I)

= ~rps(/)

= z;: -71' d</>Io(</ 1 - ZA(</>)"

Converted with

STOU Converter
trial ve rsion

(4.93)

(4.91) and we
~ U(z, I) by the

(4.94)

hnp://www.stdutilitv.com
EXERCISE 4.12. Compute the generating function, U(z,O), for the
probability to be at site 1 = 0, at time s, given that the walker was at site
1 = at time s = 0.

Answer: Since lo(</ = I, U(z, 0) is given by U(z, 0) = (1/27r) J_:r d(l/


[I - zcos())). We can make the change of variables, x = tan(</>/2). Then
cos(</ = ((I-x2)/(1
+x2))
and d</> = 2dx/(1 +x2)
and U(z, 0) =

(2/7r)Jooo (dx/[(I

- z) + x2(1 + z))) = (I -

Z2)-1/2.

We will now let Qs(l) denote the probability that the walker arrives at site I
for thefirst time after s steps. We can introduce a generating function v, l) for
Qs(/):
00

V(z, I) =

L rQs(I).

(4.95)

s=l

Thus, V (1, I) is the total probability that the walker reaches site 1 during the
random walk.

203

SPECIAL TOPICS: LATTICE RANDOM WALK

We can now relate the generating functions V(z, I) and U(z, I). The walker
can arrive at a lattice site 1 for the first time after s steps or for the first time after
one step and then again after s - 1 steps in any manner, and so on. Thus, if the
walker starts at 1 = 0,
(4.96)
and we can write
s

for

Ps(l) = L Qj(l)Ps-j(O)
j=l

s > O.

(4.97)

If we combine Eqs. (4.93), (4.96), and (4.97), we obtain


00

U(z,l) = 8[,0

00

+ LtPs(l)

8[,0

s=l

+ LLQj(I)Ps-j(O)t.

trial version
The total proba

(4.98)

s;l j=l

(4.99)

hDP://www.stdutilitv.com

~----------------------------~

V(I,O) = 1 - U(I,O)'

(4.100)

and the total probability that the walker escapes from the origin is
P escape

U ( 1, 0) .

(4.101)

For the one-dimensional walker, U ( 1, 0) = 00 and P escape = 0 (cf. Exercise


4.12). Thus, for random walk on a one-dimensional lattice, the walker will
return to the origin with probability equal to one if it starts at the origin. The
walker cannot escape.
It is very interesting to generalize the above discussion to random walks in
two or more dimensions, as we shall see below.
S4.A.2. Random Walk in Higher Dimension
To generalize the above results to d-dimensional lattices, we introduce unit
vectors, Xi (i = 1, 2, ... ,d) for the various orthogonal directions in space. If the

204

ELEMENTARY PROBABILITY THEORY AND LIMIT THEOREMS

lattice sites are a unit distance apart, we can denote their position by the vector,
+ ... + IdXd, where Ii are integers. The probability to find the walker at
site, I at time s is Ps(l) = Ps(/I, ... ,Id). We will let the lattice have 2N + 1 sites
in each spatial direction so that - N ~ Ii ~ N (i = 1, ... , d). We will assume that
the lattice is periodic in each direction. That is,

1= IIXi

P s (11 , la,

... , Id) = P s (11 [2N + 1], la,


= Ps(lI, 12 [2N
=

The probability,

P,(I)

where n = nIX}
is given by

Ps(lI,h,

, Id)
,Id) = ...

+ 1],

(4.102)

... ,Id [2N + 1]).

Ps(l), can be expanded in the Fourier series:


1
( 2N

+1

d N
n~N

+ ... + ndxd,

n,0./,(n) exp

2N;

21111 I )
1 '

(4.103)

and the Fourier amplitude, !s (n) = !s(nI, ... ,nd),

Converted with

STOI Converter
Let us no
<Pi= 27rni/(2N

(4.1 04)

trial version

the variable,
Then we find

hDP://www.stdutilitv.com
P,(I) =

(2~)

d [.

dcf>l ... [.

(4.105)

dcf>d[,(cfe-U'P,

and
00

!s(q) =

L
/1=-00

00

L P (l)e+
s

i1'P.

(4.106)

Id=-OO

It is of interest to consider a random walk on two- and three-dimensional


simple square lattices, where the walker steps to one of its nearest neighbors at
each step. Let us first consider the case, d = 2. Such a lattice is shown in
Fig. 4.7. If the lattice spacing is a = 1 and the walker is equally likely to step in
any direction, the single step probability can be written
(4.107)

where

= ilXI + 12"2.

The

characteristic

function

for

the

single-step

20S

SPECIAL TOPICS: LATTICE RANDOM WALK

(a)

12
-1--~--2~~-.-

m~~ttll
-4--~--+O

Fig. 4.7. (a) A simple square two-dimensional lattice. (b) Random walk for 1000 steps
on a square two-dimensional lattice. The walker starts at I = O.

probability is

Converted with

STOI Converter

A(<p) =

(4.108)

trial version

where

hDP://www.stdutilitv.com

If the walker starts at 1= 0, thenfo(<p) = 1 and the generating function, U(z,l),


is given by [cf. Eq. (4.93)]
1)

U(z, I) = ( 27r

7r

dl
-7r

J1T
e-u.'P
d2
.
-1T
1 - Hcos(t) + COS(2))

Let us compute U(z, 0). We make the change of variables,

Xi

(4.109)

= cos( i). Then

(4.110)

We first integrate over

X2

and obtain
(4.111)

206

-:

ELEMENTARY PROBABILITY THEORY AND LIMIT THEOREMS

./'

./

./
./

/'

./

./

./

./

./

/
/1

./

V
./

Fig. 4.8. A simple cubic lattice with lattice


spacing, a.

(cf. Ref. 8). We next integrate over

Xl

U(I,O)

where K(k) is t
When k --+ I,}
probability for
P escape = 0 [cf.
of a random wa
1000 steps are
Let us now
spacing, a = 1

and find
2
7r

(4.112)

= -K(z),

the modulus).
us, the escape
lattice is zero,
bne realization
from 1= 0 and

Converted with

STOI Converter
trial version

ce with lattice

'---_h_n_p_:/_/www
__ .S_t_d_u_ti_li_tv_.C_O_m_------'~ility
is

(4.113)

and the characteristic

function is

A(<p) = ~(cos(</>t) + COS(</>2)


+ COS(</>3)).
If the walker starts at I
1

U(z, 0) = ( 27r

)3

0, the generating function, U(z, 0), is

J1I"

1I"

-11"

d</>l

-11"

(4.114)

J1I"
d2

-11"

d</>3 1 - ~ (cos( </>1)+ cos( </>2)+ cos( 3)) .


(4.115)

Let us now compute the escape probability,

Pescape

I/U(I,

0). We may write


(4.116)

SPECIAL TOPICS: INFINITELY DIVISIBLE DISTRIBUTIONS

With a change of variables, Xi = COS(<Pi),

We first integrate over

X3

7r

JI
-1

U(I,O) takes the form

and obtain

(cf. Ref. 8). We next integrate over


U(I,O) =2"6

207

X2

dxi

(3 -xt)Jl-xi

and find
(
2 ) ~1.516
K-3 -Xl

(4.119)

(cf. Ref. 7), wh


t kind (k is the
Converted with
modulus). Fro
merically. An
for a random
analytic expres
walker on a si
Thus, for this
three dimensio
that a random
trial version
walker, who st
he walker can
escape.
The results 0
IS sec IOn are specia cases 0 a more general result due to
Polya [10] where he shows that random walkers (whose step size distribution
has finite variance) on one- and two-dimensional lattices can never escape their
origin. They eventually will return to their starting point. But in three
dimensions, there is a finite probability that a walker can escape forever from its
starting point. Random walks for which the escape probability is zero are said
to persistent. Random walks with finite escape probability are said to be

STOI Converter
hDP://www.stdutilitv.com

transient.

~ S4.B. Infinitely Divisible Distributions

[6, 11]

The concept of an infinitely divisible distribution function is important to all


limit theorems in statistical physics. The limiting distribution can only belong
to this class of distributions. A stochastic variable, Y, is infinitely divisible if for
any integer N, it can be represented by a sum
(4.120)
of identically distributed stochastic variables, Xi (i = 1, ... , N). The distribution

208

ELEMENTARY PROBABILITY THEORY AND LIMIT THEOREMS

functions, F y (y), for infinitely divisible stochastic variables, Y, are called


infinitely divisible distribution functions. The distribution function, Fy(y), is
infinitely divisible if and only if, for any N, its characteristic function fy(k) is
the Nth power of some characteristic function, fx (k, 1/ N). Thus, fy (k) =
(fx(k, l/N)t
or

fx(k, l/N) = (fy(k))l/N

(4.121)

(we take the principal branch of the Nth root). Infinitely divisible distributions
may have either finite or infinite variance. We give some examples of infinitely
divisible distributions below.
EXERCISE 4.13. Show that the characteristic function for an infinitely
divisible distribution has no real zeros.
Answer: Let g(k) = limN--.oo(f(k))I/N.
since g(k) = 0 whenf(k)
= 0 and g(k)
and f(k) is'
.,
g (k) = 1 in
characteristic
I
I

to g(k) = 0 at
all k. The con
no real zeros,

~ 54.B.l. Gaussian

Then either g(k)

= 1 whenf(k)

STOU Converter

= 0 or g(k) = 1
O. Sincef(O) = 1
origin and
) is also a
lue g(k) = 1

f(k) # 0 for
function has

trial version

hnp://www.stdutilitv.com
IS

Ion

The Gaussian distribution is infinitely divisible. To see this, let us consider the
most general form of the Gaussian probability density for the stochastic
variable, Y. The probability density is
py(y)

((y_a)2)

(1

V ~exp

2aj,

(4.122)

with average (y) = a and standard deviation O"y. The characteristic function is

fy(k) = exp

(iak -1T2).

(4.123)

The Nth root iSfx(k; l/N) = (fy(k))I/N = exp (i(a/N)/k - (aj,/2N)/2-). Thus,
the probability density for each of the identically distributed stochastic
variables, Xi is
1

Px(x; l/N)

27r(aj,/N) exp

(x - a/N)2)
2 (O"}/N)

(4.124 )

SPECIAL TOPICS: INFINITELY DIVISIBLE DISTRIBUTIONS

209

and the average value and standard deviation of the stochastic variable Xi are
(x) = (a/N) and ax = (ur/VN), respectively.
~ 54.B.2. Poisson Distribution
The Poisson distribution is infinitely divisible. The Poisson probability density
for a stochastic variable, Y, is given by
00

"Ame->'

Py(y)= L-,-8(y
m=O

- a - mh)

(4.125)

m.

where (y) = a +"Ah is the average value of Yand Uy = h.JX is its standard
deviation. The characteristic function for stochastic variable Y is
fr(k)

+ "A(eikh -

exp (iak

1)).

(4.126)

The Nth root is fx(k; I/N) = (fy(k))l/N


= exp (i(a/N)k
+ "A/N(eikh - 1)).
Thus, the probability density for each of the N identically distributed stochastic
variables Xi is

Converted with

STOI Converter
and the Poisso
standard devia
respectively.

trial version

hDP://www.stdutilitv.com

(4.127)
~ of X and its
~= (uy/VN),

~ 54.B.3. Cauchy Distribution


The Cauchy distribution function,

Fy(y)

=Hi+tan-t:b))

(4.128)

(a > 0 and b are finite constants) for stochastic variable Y, is an example of an


infinitely divisible distribution which has an infinite variance. The probability
density for a stochastic variable, Y, is
Pr

(y)

dFy

dy

= 1
7r

a2

+ (y

_ b )2 .

(4.129)

The first moment is (y) = b, but the variance and the standard deviation are
infinite. The characteristic function for stochastic variable Y is
fy(k)

= exp (ikb - Ikla).

(4.130)

210

ELEMENTARY PROBABILITY THEORY AND LIMIT THEOREMS

-15

-10

10

15

Fig. 4.9. A plot of the Cauchy probability density for b = 0 and a =

liN) = (fy(k))l/N = exp(ik(bIN)

TheNthrootisfx(k,

tt,

-lkl(aIN)).

Cauchy distributions is infinitely divisible. The probability


the N identically distributed stochastic variables Xi is

Thus, the
density for each of

Converted with
(4.131 )

STOI Converter
From Eq. (4.1
Fig. 4.9.

p is plotted in

trial version

hDP://www.stdutilitv.com
~ 54.B.4. Levy Distribution
The Cauchy distribution is a special case of a more general distribution, called
the Levy distribution. The Levy distribution has infinite variance and has a
characteristic function of the form

fy(k) = exp (-clkI

(4.132)

),

where 0 < a < 2. We can only obtain an analytic expression for Py(y), for a
few special cases. We shall discuss the properties of these distributions in more
detail in Section (S4.D).

EXERCISE
4.14. Consider
the characteristic
function
(1 - b)/(1 - beik)(0 < b < 1). Show that it is infinitely divisible.
Answer:

First note that


In f (k)

In (1 - b) - In (1 - be' )

= ~b"'km
L..J m=l

(e'

- 1).

f(k)

SPECIAL TOPICS: THE CENTRAL LIMIT THEOREM

211

Thus

f(k) =

II
00

e(b"'/m)(eikm-l).

m=l

Since f(k)
divisible.

is a product of Poisson characteristic functions, it is infinitely

~ 54.C. The Central Limit Theorem


We wish to consider the sum, YN, of N independent identically distributed
stochastic variables Xi such that
(4.133)
We will let FyN(Y) denote the distribution function, andfyN(k)
function, of sto
..
.
tion function, a
XfN). The char

fyN(k) =

= (Ix

the characteristic
te the distribu-

STOU Converter
trial version

hnp://www.stdutilitv.com

(4.134)

[cf. Eq. (4.35)], where Px(x; liN) = dFx(x; IIN)ldx is the probability density
of Xi.
The Central Limit Theorem describes the limiting behavior (N --+ 00) of
the stochastic variable YN for the case where the stochastic variables Xi have
finite variance. We will also assume that Xi has zero mean, (x) = 0, since this
causes no loss of generality. More specifically, we consider system for which
lim fx(k; liN)

N-+oo

= I

(4.135)

C,

(4.136)

and
lim N (Xl)
N-+oo

where C is a finite constant. For such systems, the limiting distribution function,
Fy(y) = limN-+oo FyN(Y) is infinitely divisible even though YN and Xi might not
infinitely divisible. Furthermore, the limiting distribution is a Gaussian. This is
the content of the Central Limit Theorem.

212

ELEMENTARY PROBABILITY THEORY AND LIMIT THEOREMS

Before we take the limit, N ~ 00, and prove the Central Limit Theorem, it
useful in the next section to obtain some inequalities.
~ S4.C.l. Useful Inequalities
Let us first define
t1fx,N =fx(k; liN) - I =
=

I(e

ikx

I(e

ikx

l)dFx(x; liN)
(4.137)

I - ikx)dFx(x, liN).

where we have used the fact that (x)

0 in the last inequality. Next note that


(4.138)

(we have plotted the left- and right-hand sides ofEq. (4.138) in Fig. 4.10. If we
combine Eqs. (
.

Converted with
(4.139)

STOI Converter

This is our first


Let us now
assume that ~

trial version

e N, we can

hDP://www.stdutilitv.com

Iln[fx(k; liN)] - ~fX,NI = 11n[1


001

- ~fX,NI

100

001

= L;(~fx,Nt

+ Llfx,N]

~ L;I~fx,Nlm~2LI~fx,Nlm
m=2
m=2

(4.140)

m=2
1 I~fx NI2
2
= 21 -1~fx,NI
~ l~fx,NI .

10

I(x) I

-10

-5

o X

10

Fig. 4.10. Plot of f(x):::


(solid line) and of f(x)
line) for k = 1.

leikx

:::!r

1 - ikx I
(dashed

213

SPECIAL TOPICS: THE CENTRAL LIMIT THEOREM

We can now combine the above inequalities

Iln(!YN) - N~fx,NI

= Nlln(!x)

to obtain

~fX,NI ~NI~fx,NI2

~!Nk(~)I~fx,NI,
(4.141)

~fX,N is defined in Eq. (4.137). We will use these inequalities below .

where

.....S4.C.2. Convergence to a Gaussian


If we combine Eqs. (4.136) and (4.141), we find
lim IlnfyN(k)
N__.oo

N__.oofx (k, 1IN)

since lim

fy(k)

- NJ(eikx
--+

1-

ikx)dFx(x;

11N)1 = 0,

(4.142)

ikx)dFx),

(4.143)

1. Therefore,

= N__.oo
lim fyN(k) = lim exp
N__.oo

(NJ(eikx

1-

Converted with

where fy(k) is
It is useful t

STOI Converter
trial version

(4.144)

hDP:llwww.stdutililV.com

Then KN( -00


(u)
decreassing bounded function. Using Eq. (4.144), we can write
N

J(eilu -1- ikx)dFx(x; liN) = J(e'kx -1- ikx) x~dKN(X).


function,fy(k),

Thus, the limiting characteristic

fy(k)

= lim exp

N__.oo

[J(eikx

1-

is a non-

(4.145)

takes the form

ikx) x12dKN(X)].

(4.146)

Equation (4.146) is the Kolmogorov formula (cf. Section S4.E). In Section S4.E
we give a general proof that since K(u) is a bounded monotonically increasing
function,fy(k)
is infinitely divisible. Below we show this for the special case of
variables, Xi, with finite variance.
Let us introduce a new stochastic variable, Zi, which is independent of N,
such that

Zi - (z)
Xi = y'NO'z .

(4.147)

214

ELEMENTARY PROBABILITY THEORY AND LIMIT THEOREMS

Here CTZ is the standard deviation of Zi and is finite. We assume that the
stochastic variables Zi are identically distributed. We see that (x) = 0 and
(r) = ((Z2) - (z)2)IN~.
Thus,

N('?-) = (Z2) ~ (Z)2 = I

(4.148)

z
and Eq. (4.136) is satisfied. Furthermore,
lim
Nr+cx:

/x(k; liN)

lim

N-.oo

we have

e-ik(Z)/v'Naz/

z (_k_)
yIN az

(4.149)

(see Eq. 4.50) and Eq. (4.135) is satisfied.


Let us look at the limiting behavior of KN(U) in Eq. (4.144). First note that

KN(u) = N

[>0 dx'?- Px(x; liN)


(4.150)

Converted with

STOI Converter

so that

trial version

hDP://www.stdutilitv.com

(4.151)

It is easy to show that this is just the condition that the limiting distribution

be a

, Gaussian, We can write Eq. (4.151) in the form limN-.oo KN(U) = 8(u), where
8(u) is the Heaviside function. Therefore, limN-.oo dKN(u) = b(u) duo If we
substitute this into Eq. (4.146), we obtain
(4.152)
From

Py(y)

Eq.

(4.55), this corresponds


to a Gaussian
probability
density,
centered at (y) = 0 with standard deviation equal to one. This

= el/2,

result is the Central Limit Theorem .

.... S4.D. Weierstrass Random Walk [12, 13J


One of the simplest random walks which does not satisfy the conditions of
the Central Limit Theorem is the Weierstrass random walk. In the continuum
limit, the Weierstrass random walk is governed by a Levy distribution and not
a Gaussian. We will first consider the discrete case and then the continuum
limit.

215

SPECIAL TOPICS: WEIERSTRASS RANDOM WALK

.....S4.D.l.

Discrete One-Dimensional

Random Walk

Let us consider a walker constrained to move along the x axis. At discrete time
intervals, T, the walker takes a step of length b" Ll to the left or to the right with
probability (a - 1)/2an+1 (n can be any integer, n = 0, 1, ... ). We assume that
each step is independent of all the others. The displacement of the walker after
N steps is YN = Xl + ... + XN, where Xi is the displacement at the ith step. The
probability density for the displacement Xi is
(4.153)
where a~ 1 and b~ 1. It is easy to show that (x) = 0 and (x2) =
(a - I)Ll2/(a - b2). Thus for b2 < a we have (xl) < 00, and we expect the
limiting distribution to be a Gaussian. For b2 = a we have (x2) = 00, and for
b2 > a, (x2) is undefined. Therefore, for b2 ~ a the conditions of the Central
Limit Theorem are not satisfied, and as N --t 00 the probability density for YN
need not approach a Gaussian.
The characn

Converted with

STOI Converter

~).

(4.154)

trial version
which is knOWI
k) for the case
a = 4 and va
(x) = Ll2 and
f(k) = cos (kLl). Thus, for this case the random walk reduces to a binomial
random walk. In Fig. 4.11, we plot f(k) for the cases (a = 4, b = 3) and
(a = 4, b = 5). The characteristic function, f(k),
has the very interesting
property that for b > a it has finite derivatives at no value of k [14]. In fact,
except for special cases, the characteristic function has structure to as small a
scale as we wish to look.

hDP:llwww.stdutililV.com

1.0
I

a=4
b=3

(a)

f(k)

a=4!
b=5

-1.0
-2

-2

Fig. 4.11. The Weierstrass characteristic function, f(k),


for (a) a
(J1. = In (a)j In (b) = 1.26) (b) a = 4, b = 5 (J1. = In (a)j In (b) = 0.86).

= 4,

=3

216

ELEMENTARY PROBABILITY THEORY AND LIMIT THEOREMS


12

(a)

N(y)
8

._ ..

-2000

-1000

1000

(b)

N(y)
20 10-

. -_-.-...
.... . -- ..
o ::;.._.:. -_...;:-:~;::- .... .:..::
-::
-200 -100

100

200

(c)

Converted with

STOI Converter
-2(

~ of the Weier= 2000, ~ = 1,


b = 3, and (c)

trial version

hDP://www.stdutilitv.com
The characteristic function for the displacement YN

= Xl + ... + XN is
(4.155)

wheref(k) is defined in Eq. (4.154). In Fig. 4.12 we show one realization of the
Weierstrass random walk for N = 2000 and for cases (a = 4, b = I),
(a = 4, b = 3), and (a = 4, b = 5). For the case (a = 4, b = 5), the random
walk shows clustering with a spread that is growing dramatically with N. For
the case (a = 4, b = 3) the random walk no longer clusters but still is spreading
rapidly with N. For the case (a = 4, b = 1) the random walk reduces to a
binomial random walk which spreads as yIN.
It has been shown in Ref. 15 that this phenomenon of clustering persists as
N ~ 00 for the parameter range 0 < J-L = In(a)/ln(b) < 1. In the limit N ~ 00,
there will be clusters on all scales down to size ~ and the probability will have
a fractal-like structure in space. In the parameter range 0 < J-L < 1, it has been
shown that the walker has a finite probability of escaping the origin. The
random walk is transient. Therefore the walker does not return to smooth out
the clusters. For 1 < J.L < 2, the random walk is persistent and the clusters get
smoothed out.

217

SPECIAL TOPICS: WEIERSTRASS RANDOM WALK

0.2

0.06

(a)

(b)

0.04
0.1

f~(Ic)

0.02

0
-0.02

-0.1 ;

-0.4

-0.2

o Ie

0.2

0.4

-0.05

o k

0.05

Fig. 4.13. Self-similarity embedded in the Weierstrass function. We plot


f~(k) == (l/am+l) f(bn+1 k) for a = 4, b = 5, and (a) n = 0 and (b) n = 1. In going
from (a) to (b) we are enlarging the central peak.

The characteristic function,f(k),


has embedded in it a self-similar structure.
To see this let us rewrite f(k) in the form

Converted with
or, more gener

(4.156)

STOI Converter
trial version

(4.157)

hDP://www.stdutilitv.com

We show this be aVlor 1 Ig..


(l/a2)f(b2k)
for (a = 4, b = 5). As we move to a smaller scale (increasing n) we focus on the
central peak of the previous scale. The figure is self-similar to as small a scale
(in k) as we wish to examine.
It is important to note thatf(k) is not infinitely divisible becausef(k) can go
to zero for some values of k (cf Exercise 4.13 and Fig. 4.11).
~ S4.D.2. Continuum Limit of One-Dimensional
Discrete Random Walk
Let us now take the continuum limit of the Weierstrass random walk. We
proceed as in Section 4.E.4. Let us assume that the time between steps is T, and
rewrite the characteristic function,fyN(k) =fy(k, NT). We assume that initially
the walker is at y = 0 so that PYo(y) = 8(y) and the characteristic function,
fy(k,O), equals 1. We can write

Jy(k, (N + 1)T) - Jy(k, NT) = (!(k) - 1)Jy(k, NT),


where f(k)

(4.158)

is defined in Eqs. (4.154). We will now let a = 1 + ad and


the limits N -+ 00, d -+ 0, and T -+ 0 so that

b = 1 + {3d and take

218

ELEMENTARY PROBABILITY THEORY AND LIMIT THEOREMS

J-L = In(a)/ln(b)

~ a/(3,NT

~ t, and fl.1l-/T ~ 8. It can be shown that [13]


(4.159)

where r (- J-L) is the Gamma function.


For 0 < J-L < 1, I'( - J-L) is negative and cos(WTr/2) is positive. For
1 < J-L < 2, I'( -J-L) is positive and cos(J-L7r/2)is negative. From Eq. (4.159),
we can write
af(k, t)
at

= -DLlklll-f(k,

t),

(4.160)

where
DL

= J-L8

(4.161)

cosC;)q-J-L)
is

Converted with
This is the
displacement

STOI Converter
0

trial version

(4.162)
y,

t), for the

en by

hDP://www.stdutilitv.com

(4.163)

-00

Bochner [16] has shown that P(y, t) is only a nonnegative function for
o < J-L:::;;;2 and therefore can only be interpreted as a probability for this range of
J-L. It is easy to check that P(y, t) is normalized to one.
There are only a few values of J-L for which Eq. (4.163) can be integrated and
obtained in closed form. The case J-L = 1 gives the Cauchy distribution [cf. Eq.
(4.163)], while the case J-L = 2 gives the Gaussian distribution. In Fig. 4.14 we
show plots of P(y, t) for t = I/DL and for J-L = 0.86 and J-L = 1.26, the cases
considered in Fig. 4.11. P(y, t) has a very long tail indicating that there is no
cutoff in the structure for small k (long wavelength) .

.... S4.D.3. Two-Dimensional

Discrete Random Walk (Levy Flight)

The Levy flight is a special case of the more general Rayleigh-Pearson random
walk [17, 18] in two and three dimensions. We will first set up the RayleighPearson random walk and then consider the case of a discrete Levy flight.
Let us consider a random walk in the (x, y) plane such that at the ith step the
particle takes a step of length a, along a path that makes an angle OJ with respect

219

SPECIAL TOPICS: WEIERSTRASS RANDOM WALK

P(y, t)

0.12

I' \

0.04

-100

-50

50

100

Fig. 4.14. Plots of P(y, t) for J..t = 0.86 (solid line) and J..t = 1.26 (dashed line) for time
t = l/DL.

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com
Fig. 4.15. The coordinates for a two-dimensional ran om walk.

to the x-axis (cf. Fig. 4.15). We will assume that tt and OJ are independent
stochastic variables, that aj has a probability density P(rj), and that OJ is
uniformly distributed over the interval 0 -+ 21l'so that the probability density
P(Oj) = (1/21l'). If the walker is at (x = O,y = 0) at the initial time, then after
N steps it will be at (x = XN,Y = YN),
where XN = rlcos(Ol)+
r2cos(02) + ... + rNcos(ON) and YN = rlsin(Ot) + r2sin(02) + ... + rNsin(ON)'
If we assume that the ith step is independent of the (i + l)th step for all i,
then the probability density to find the particle at x -+ x + dx and Y -+ Y + dy
after N steps is
PN(X,y)

=(2~) J:K
N

00

dO[.

J:K dON 10

x 8(x - rl cos (01)


x 8(y -

Tt

OO

dr[

r2 cos (02) -

sin (01) - r2 sin (02) -

drNP(r[)

x ... x P(rN)

- rN cos (ON))
- rN sin (ON )).

(4.164)

220

ELEMENTARY PROBABILITY THEORY AND LIMIT THEOREMS

x
Fig. 4.16. One realization of a Rayleigh-Pearson random walk for the case when all
steps are of unit length (a = 1). The total number of steps is N = 2000. The walker
starts at (x = 0,)

Converted with

STOI Converter
In Fig. 4.16
trial version
ndom walk for
the case when t
.stributed with
probability den
~alker always
takes steps of unit length, although the direction of the steps is completely
random.
In Fig. 4.17 we show a Rayleigh-Pearson random walk for the case when the
step lengths r are variable and probability density P(r) is given by a Weierstrass
probability density:

hDP:llwww.stdutililV.com

a_lO

P(r)

= -'""
a

~an

[b(r - bnLl))].

(4.165)

n=O

Note that the step lengths are all positive. With this distribution of step
lengths, there is a small chance that at a given step the walker will take a
very long step. Therefore the behavior of this random walk is completely
different from that of a binomial random or simple lattice random walk.
Figure 4.17 is an example of a discrete Levy flight. For 0 < J-L < 2 and in
both two and three spatial dimensions the random walk is transient [15].
Therefore, as N ~ 00 the random walk will form fractal-like clusters on
all scales down to scale size Ll. In the continuum limit, it will form clusters
on all scales.

SPECIAL TOPICS: GENERAL FORM OF INFINITELY DIVISIBLE DISTRIBUTIONS

221

o
y
-1000

-2000

-3000
-200

200

600

50
400

o
o

Converted with

STOI Converter
Fig. 4.17. One
trial version
e case when the
distribution of
ensity. The total
number of step
he sequence of
angles is the sa.nlcClS""TITl~:-"'I':"Tu-(CIT""Tll~OIII]pICl~n:IIl~rTTTrrzq;nnn;dtion
of the upper
section of (a). (c) Magnification of hatched box in (b).

hDP://www.stdutilitv.com

It is also possible to obtain a Levy-type random walk by haviing the walker


move on a lattice which has self-similar structure. Such types of random walks
are discussed in Ref. [19].

~ S4.E. General Form of Infinitely Divisible


Distributions
The definition of an infinitely divisible distribution function is given in Section
S4.B. For the case of infinitely divisible distributions with finite variance, A. N.
Kolmogorov [6, 20] has given a general form of the characteristic function,
called Kolmogorov's formula, which is also unique. P. Levy [21, 22] and A.
Khintchine [23] have generalized Kolmogorov's result to the case of infinitely
divisible distributions with infinite variance. Their formula is called the LevyKhintchine formula. In Sections S4.E.I-S4.E.2 below, we describe these two
formulas.

222

ELEMENTARY PROBABILITY THEORY AND LIMIT THEOREMS

.... S4.E.l. Levy-Khintchine

Formula [22, 23]

The most general form of an infinitely divisible characteristic


form

function has the

[. L ( ,

f(k) = exp ika +

1 +.xl
e'kx - 1 - 1 ikx
+ x2 ) -;r-dG(x)

-00

1,

where a is a constant and G(x) is a real, bounded, nondecreasing


such that G( -00) = O. The integrand at x = 0 is

eikx_l_~)I+.xl}
{(

1 + x2

(4.166)
function of

=_k2.
x2

x=O

(4.167)

Both a and G(x) are uniquely determined by f(k). Equation (4.166) is called the
Levy-Khintchine formula.
It is fairly easy to see that the Levy-Khintchine
formula is infinitely
divisible. We can write Eq. (4.166) in the form

Converted with
where

a2

I(k) = lim
=

STOI Converter

G((

e-+O

(4.168)

trial version

Ix

hDP://www.stdutilitv.com

-"'- ( 'kxl
lim lim
e, - 1 e-+O m-+oo

1=1

ikxi ) 1 + x2
--:2
--2 _I (G(XI) - G(XI-I)).
1 + Xi
XI

(4.169)

Thus f (k) is the product of a Gaussian characteristic function and the limit of a
product of Poisson characteristic functions, all of which are infinitely divisible.
Therefore,J(k) is infinitely divisible.
The quantities a and G(x) can be obtained if we know FN(x), the distribution
function associated with (!(k))l/N. Let us write

Inf(k) = lim N(fN(k) - 1)


N-+oo

=
But

lim

N-+oo

(ikaN

I(

N(fN(k) - 1) = N r:(eikx

e''kx - 1 -

ikx ) 1 +.xl

--2

1+x

--2

-dGN(X ) .

(4.170)

- l)dFN(x), so that
(4.171)

SPECIAL TOPICS: GENERAL FORM OF INFINITELY DIVISIBLE DISTRIBUTIONS

EXERCISE 4.15. Show that the characteristic

[(1 - b)/(1
divisible.

+ a)][(1 + aeik)/(1

- be-ik)](0

function, f(k)

223

< a~b < 1) is not infinitely

Answer: Note that


In(f(k)) = In(1 - b) + In(1 + aeik) -In(1 + a) -In(1 - be-ik)
n
=
[b (e -ink _ 1) + (-1 t+ 1 an (eink - 1)]
~1
n
n

f:

== ink +

1 + u2
(e-1ku - 1) -2-dG(u)
-00
u
00.

- ik

00
-00

1
-dG(u).
u

, We can satisfy the above equation if we let


00 [ nb"
G(u) = '" --8(u 2
L...J
1+ n
n=l

+ n)

- (-It--8(u

nan]

- n)

1 + n2

and

Converted with

STOI Converter
But G(u) is r
not infinitely

refore,f (k) is

trial version

hDP://www.stdutilitv.com
It is easy to see that for the Gaussian characteristic function, Eq. (4.123),
Q

= a and G(x) = 0'~8(x). For the Poisson characteristic function, Eq. (4.126),

+ )"h/(1 + h2)

and G(x) = )..h2/(1 + h2)8(x - h). For the Cauchy


characteristic function, Eq. (4.130), o = band G(x) = (a/,rr) tan"! (x) + (a/2) .
Q

=a

.... 54.E.2. Kolmogorov Formula [6, 20]


Kolmogorov was the first to find a general formula for infinitely divisible
distributions, although his formula only applies to distributions with finite
variance. He obtained

Jy(k) = exp [hk +

[0

{,flo - 1 -

iku}

:2 dK(U)]

(4.172)

where "y is a constant and the function K(u) [whose differential, dK(u), appears
in Eq. (4.172)] is a nondecreasing function with bounded variation. Note that
limk~o(df/dk) = i"( and lim/Ho(tPf /dt2) = -y K( 00 ). Thus the first moment is (y) = "y and the standard deviation 0' = K(oo). Equation (4.172) is
called Kolmogorov's formula and it is unique.

224

ELEMENTARY PROBABILITY THEORY AND LIMIT THEOREMS

(a)
o

u
K(u)

(b)
o

Fig. 4.18. Plots of K(u) for (a) the


Gaussian distribution and (b) the Poisson
distribution.

The Kolmogorov formula can be obtained from the Levy-Khintchine


formula if we let K(x) = S~oo(1 + i)dG(y)
and, = a + S_O:ydG(y). If we
compare Eq. (4.172 to the characteris ic function ~ r
.
istribution [cf.
Eq. (4.126)], th
n function of
spacings betwe
ings between
discrete jumps

K(u) = Ah28(u

Eq. (4.123), th
realizations) an
distributions in
Cauchy distribu

STOI Converter
trial version

hDP:llwww.stdutililV.com

distribution,
n distribution,
cing between
and Poisson
apply to the
.ance .

EXERCISE
4.16. Consider the characteristic function, f(k) =
(1 - b)/(1 - bik) (0 < b < 1) (cf, Exercise 4.15). Find the Kolmogorov
function, K(u).
Answer: First note that
b"
_(ikm
m=l m
b
= ik (1 _ b)
00

lnf(k)

=L

00

- 1) = ik Lbm
m=l
00

bm

+Lm

00

b"

+L -+
m=l

00

L
m=l

b"
_(ikm
m

- 1- ikm)

(e'km

1 - ikm).

m=l

By comparison with Eq. (4.172), we find dK(u) = L:':l mbm8(u - m) and


K(u) = L:~=1mbm8(u - m). Thus, ,= bl(1 - b) and a'l = L:~=1mb" =
b I (1 - b) 2 Furthermore, K (u) is a nondecreasing function of u and has
bounded variation. Thus, conditions of Kolmogorov's formula are satisfied
andf(k) is an infinitely divisible characteristic function.

225

PROBLEMS

REFERENCES
II Nuovo Cimento 26B, 99 (1975).

1. G. Jona-Lasinio,

2. F. Mosteller, R. E. K. Rourke, and G. B. Thomas,


(Addison-Wesley,
Reading, MA, 1967).
3. S. Lipschutz,
1965).

Probability, Schaum's

Outline

Series,

Probability and Statistics


(McGraw-Hill,

New York,

4. W. Feller, An Introduction to Probability Theory and Its Applications, Vol. I (John


Wiley & Sons, New York, 1968).
5. F. Reif, Fundamentals of Statistical and Thermal Physics (McGraw-Hill,
1965).

New York,

6. B. V. Gnedenko and A. N. Kolmogorov, Limit Distributions for Sums of Independent


Random Variables (Addison-Wesley, Reading, 1954).
7. E. Montroll,

Proc. Symp. Appl. Math. 16, 193 (1964).


Table of Integrals, Series, and Products

8. I. S. Gradshteyn
and I. M. Ryzhik,
(Academic Press, New York, 1980).

9. G. N. Watson, Quart. 1. Math., Oxford Ser., 10, 266 (1939).


10. G. Polya,

Converted with

11. E. Lukacs,
12. B. West, 1.

STOI Converter

13. E. W. Mon
and 1. L. L

trial version

14 . G . H. Hard

hDP:llwww.stdutililV.com

. Sci. USA, 78,


3287
(1981J---------------------'
S. Bochner, Duke Math. 1. 3, 726 (1937).
K. Pearson, Nature 77, 294 (1905).
L. Rayleigh, Nature 72, 318 (1905).
C. van den Broeck, Phys. Rev. Lett. 62, 1421 (1989); Phys. Rev. A40, 7334 (1989).
A. N. Kolmogorov, Atti. Acad. Naz. Lincei. Rend. Cl. Sci. Fis. Mat. Nat. (6) 15, 805-

15. B. D. Hugh
16.
17.
18.
19.
20.

808, 866-869

(1932).

21. P. Levy, Theorie de L'addition des Variables Aleatoires (Gauthier-Villars,


1937).

Paris,

22. P. Levy, Annali R. Scuola Norm. Sup. Pisa (2) 3, 337 (1934); 4, 217 (1935).
23. A. Ya. Khintchine,
(1937).

Bull. Univ. d'Etat. Moskau. Ser. Interact, Section A.1, No.1,

PROBLEMS
Problem 4.1. A bus has nine seats facing forward and eight seats facing backward. In
how many ways can seven passengers be seated if two refuse to ride facing forward and
three refuse to ride facing backward?

226

ELEMENTARY PROBABILITY THEORY AND LIMIT THEOREMS

Problem 4.2. Find the number of ways in which eight persons can be assigned to two
rooms (A and B) if each room must have at least three persons in it.
Problem 4.3. Find the number of permutations of the letters in the word
MONOTONOUS. In how many ways are four O's together? In how many ways are
(only) 30's together?
Problem 4.4. In how many ways can five red balls, four blue balls, and four white balls
be placed in a row so that the balls at the ends of the row are the same color?
Problem 4.5. Three coins are tossed. (a) Find the probability of getting no heads. (b)
Find the probability of getting at least one head. (c) Show that the event "heads on the
first coin" and the event "tails on the last two coins" are independent. (d) Show that the
event "only two coins heads" and the event "three coins heads" are dependent and
mutually exclusive.
Problem 4.6. Various six digit numbers can be formed by permuting the digits 666655.
All arrangements are equally likely. Given that a number is even, what is the probability
that two fives are together? (Hint: You must find a conditional probability.)
Problem 4.7. Fifteen boys go hiking. Five get lost, eight get sunburned, and six return
home without pr
oy got lost? (b)
What is the prob

Converted with

Problem 4.8. A
variable Y can
PX,y(x,y)
= Ei=
the following tw
Pl,4 = P3,2 =

i-

STOI Converter
trial version

3. A stochastic
bability density
of X and Y for
= P3,4 = 0 and

hDP://www.stdutilitv.com

Problem 4.9. T
and Gaussian
distributed with first moment (x) = (y) = 0 and standard deviations ax = Uy = 1. Find
the characteristic function for the random variable Z = X2 + y2, and compute the
moments (z), (Z2), and (Z3). Find the first three cumulants.
Problem 4.10. A die is loaded so that even numbers occur three times as often as odd
numbers. (a) If the die is thrown N = 12 times, what is the probability that odd numbers
occur three times? If it is thrown N = 120 times, what is the probability that odd
numbers occur thirty times? Use the binomial distribution. (b) Compute the same
quantities as in part (a) but use the Gaussian distribution. [Note: For parts (a) and (b)
compute your answers to four places.] (c) Compare answers for (a) and (b). Plot the
binomial and Gaussian distributions for the case N = 12.
Problem 4.11. A book with 700 misprints contains 1400 pages. (a) What is the
probability that one page contains 0 misprints? (b) What is the probability that one page
contains 2 misprints?
Problem 4.12. Three old batteries and a resistor, R, are used to construct a circuit. Each
battery has a probability P to generate voltage V = Vo and has a probability 1 - P to
generate voltage V = O. Neglect any internal resistance in the batteries. Find the average
power, (V2) / R, dissipated in the resistor if (a) the batteries are connected in series and
(b) the batteries are connected in parallel. In cases (a) and (b), what would be the
average power dissipated if all batteries were certain to generate voltage V = vo? (c)
How would you realize the conditions and results of this problem in a laboratory?

227

PROBLEMS

Problem 4.13. Consider a random walk in one dimension. In a single step the
probability of a displacement between x and x + dx is given by
P(x)dx

= .~

v 27rl12

exp (_ (x;
(J'

:)2).

After N steps the displacement of the walker is S = Xl + ... + XN, where Xi is the
displacement during the ith step. Assume the steps are independent of one another. After
N steps, (a) what is the probability density for the displacement, S, of the walker and (b)
what are the average displacement, (S), and the standard deviation of the walker?
Problem 4.14. Consider a random walk in one dimension for which the walker at each
step is equally likely to take a step with displacement anywhere in the interval
d - a ~ x ~ d + a, where a < d. Each step is independent of the others. After N steps the
walker's displacement is S = Xl + ... + XN, where Xi is the displacement during the ith
step. After N steps, (a) what is the average displacement, (S), and (b) what is his
standard deviation?
Problem 4.15. Consider a random walk for which the probability of taking of step of
length, x ---+ x + dx, is given by P(x)dx = ~(a/(r + a2))dx. Find the probability density
for the displacement of the walker after N ste s. Does it satisf the Central Limit
Theorem? Shoul
Problem S4.1.

Converted with

STOI Converter

where the walke


V(z, I). (b) Co
Compute the pro
trial version
after s = 4 steps
I = 1for the firs
of steps needed w--,,...,...,.,OI;OTT""l:7rI:'I...,....---r-r-""""rxrq.,---.:r...,.......'""""'~xr--oro:nll'Lo"""""l<=q::>r<O:Jrn-'

hDP://www.stdutilitv.com

I (d = 1) lattice
tions U(z, I) and
dom walk. (c)
1 after s = 3 and
s I = 0 and site
average number
your result.

Problem S4.2. Consider a random walk along nearest neighbors on an infinite, periodic
face centered two-dimensional (d = 2) square lattice. A unit cell is shown in Fig. 4.19.
Assume that the lattice spacing is a = 2 and that the walker starts at site 11 = 0,12 = o.
(a) By counting paths (draw them), find the probability, P4(O), that the walker returns to
the origin after four steps. Find the probability, Q4(O), that the walker returns to the
origin for the first time after four steps. (b) Compute the generating function, U(z, 0).
Use this result to compute P4(O) and Q4(O). (c) Compute the escape probability.
Problem S4.3. Consider a random walk along nearest neighbors on an infinite, periodic
body-centered three-dimensional (d = 3) cubic lattice. A unit cell is shown in Fig. 4.20.
Assume that the lattice spacing is a = 2 and that the walker starts at site,

Fig. 4.19. Unit cell for a face-centered square lattice.

228

ELEMENTARY PROBABILITY THEORY AND LIMIT THEOREMS

Fig. 4.20. Unit cell for a body-centered cubic lattice.

(/1 = 0,12 = 0,13 = 0). Compute the generating function. U(I,O), and the escape
probability. [Hint: Make a change of variables, Xi = tan(4);/2). Then change to
spherical coordinates,
Xl = rsin(0)cos(4,x2
= rsin(O)sin(4, and Xl = rcos(O).
Replace the integration over r by an integration over t, where t =
rsin(O)

!cos(O)sin(~). The three integrations over t,O, and 4> then separate.]

Problem S4.4. Consider a Rayleigh-Pearson random walk in which the walker has a
probability P(r)
r. If the walker
starts at the origi
Converted with
ithin a circle of
radius Rafter N
Problem S4.S.
(a) Compute the
Kolmogorov's fo
variable X infinit

STOI Converter
trial version

) = 1/( 1 + k2).
ance. (b) Write
s the stochastic

hDP://www.stdutilitv.com

F(x) = (~)n/2_1_
rx dyy(n/2)-le-y/2
2
r(n/2)Jo

for

x> 0

and F(x) = 0 for X < 0 (n is an integer). (a) Find the characteristic function for this
distribution. (b) Write the Levy-Khintchine formula for the characteristic function. That
is, find a and G(x). Is X infinitely divisible? [Hint: First find FN(X) and use it to find
GN(x) and take the limit.]

5
STOCHASTIC DYNAMICS AND
JJROWNIAN MOTION

S.A. INTRODUCTION
Now that we have reviewed some of the basic ideas of probability theory (cf.
Chapter 4), w
ions in which
probability can
Converted with
bd some simple
random walks 1
repeated applic
this chapter we
probability dist

ST 0 U C oover Ier
trial version

hDP://www.stdutilitv.com

~
be built from
ne intervals. In

he evolution of
~ch we discuss

this time evolu


cern ourselves
with the relatic ~
._..
_ .~___
hat will be the
subject of Chapter 6. Also, in this chapter we will limit ourselves to Markov
processes, which are stochastic processes with a very limited memory of
previous events.
The equation which governs the stochastic dynamics of Markov processes is
the master equation. It is one of the most important equations in statistical
physics because of its almost universal applicability. It has been applied to
problems in chemistry, biology, population dynamics, laser physics, Brownian
motion, fluids, and semiconductors, to name a few cases. As a system of
stochastic variables evolves in time, transitions occur between various
realizations of the stochastic variables. Because of these transitions, the
probability of finding the system in a given state changes until the system
reaches a final steady state in which transitions cannot cause further changes in
the probability distribution (it can happen that the system never reaches a steady
state, but we will be most concerned with cases for which it can). To derive the
master equation, we must assume that the probability of each transition depends
only on the preceding step and not on any previous history. This assumption
applies to many different types of physical system found in nature.
One of the simplest types of Markov processes that we will consider is the
Markov chain. Markov chains are processes in which transitions Occur between
realizations of discrete stochastic variables at discrete times, as in the case for

230

STOCHASTIC DYNAMICS AND BROWNIAN MOTION

the random walks we treated in Chapter 4 using simple probability theory. In


this chapter we will formulate such problems in a more elegant manner. We will
get a very good picture of how stochastic systems can decay to a unique steady
state for the case in which the transition matrix is "regular" and we can
introduce the concept of ergodicity, which we shall come back to in Chapter 6.
The dynamics of Markov chains is totally governed by a "transition matrix"
which is real and generally is nonsymmetric. We shall show how the stochastic
dynamics can be expressed in terms of a spectral decomposition of the
probability in terms of the left and right eigenvectors of the transition matrix.
If the time interval between events can vary in a continuous manner, then we
must introduce the master equation which is a differential difference equation
governing the time evolution of the probability. For the very special case when
the transition rate between realizations exhibits detailed balance, solutions of
the master equation can also be expressed in terms of a spectral decomposition.
Detailed balance holds for many types of systems which are near thermal
equilibrium or do not have internal probability currents as they reach the long
time state. Some examples for which detailed balance holds can include
pn~~~aw~~~~~rnru~~MM~ruru~~somerandom

Converted with
reactions and p

STOI Converter

lude chemical
special topics.

One often e
ut for which a
few of the deg
scale than the
others. An ex
trial version
tively massive
particle, such a
as water. The
grain of polle
motion. Browman motion, w IC was rst ma e popu ar y the work of
biologist Robert Brown, was used by Einstein as evidence of the atomic nature
of matter. Indeed, it only occurs because of the discrete particulate nature of
matter. A phenomenological theory of this motion can be obtained by writing
Newton's equation of motion for the massive particle and including in it a
systematic friction force and a random force to mimic the effects of the many
degrees of freedom of the fluid in which the massive particle is immersed. This
gives rise to the Langevin equation of motion for the massive particle.
Given a Langevin equation for a Brownian motion process, we can obtain an
equation for time evolution of the probability distribution in phase space of the
Brownian particle, called the Fokker-Planck equation. In the sections devoted
to special topics we will derive the Fokker-Planck equation and we will solve it
for Brownian motion with one spatial degree of freedom in the presence of
strong friction. We shall also mention the interesting connection of FokkerPlanck dynamics to chaos theory when one considers more than one spatial
degree of freedom. For cases in which the master equation cannot be solved
exactly, it can sometimes be approximated by a Fokker-Planck equation. We
will discuss some of those cases in this chapter. In later chapters we will use the
concepts developed here to describe other types of Brownian motion in many
body systems.

hDP://www.stdutilitv.COmdom.agitated

GENERAL

231

THEORY

S.D. GENERAL THEORY [1-4]


Let us consider a system whose properties can be described in terms of a single
stochastic variable Y. Y could denote the velocity of a Brownian particle, the
number of particles in a box, or the number of people in a queue, to name a few
of many possibilities.
We will use the following notation for the probability density for the
stochastic variable Y:
PI (Yl, tl)

== (the probability density that the stochastic


variable Y has value Yl at time ri):

P2(Yl,

tl;Y2, t2)

(5.1)

== (the joint probability density that the


stochastic variable Y has value Yl at

(5.2)

time tl and Y2 at time t2);


the

Converted with

STOI Converter

at

IDe tn'

(5.3)

trial version

The joint prob

hDP://www.stdutilitv.com

(5.4)

they can be reduced:


r

jPn(YI,

tljY2, t2,jjYn,tn)dYn

=Pn-1(Yl,tl;Y2,

t2j ... jYn-l,tn-l);

(5.5)

and they are normalized:


(5.6)
In Eqs. (5.5) and (5.6) we have assumed that Y is a continuous stochastic
variable. However, if Y is discrete we simply replace the integrations by
summations. We can introduce time-dependent moments of the stochastic
variables, (Yl(tI)Y2(t2)
x ... X Yn(tn)). They are defined as

and give the correlation between values of the stochastic variable at different

232

STOCHASTIC DYNAMICS AND BROWNIAN MOTION

times. A process is called stationary

for all nand

T.

if

Thus, for a stationary process


(5.9)

and (Yl(tt)Y2(t2))
depends only on It 1 - t21-the absolute value of the
difference in times. All physical processes in equilibrium are stationary.
We shall also introduce a conditional probability:
Pili ( Yl, tl I Y2, t2) = (the conditional probability density for the

stochastic variable Y to have value Y2 at time

(5.10)

tz given that it had value YI at time tl)


It is defined by the identity

(5.11)

Converted with
Combining Eq~
probability den

STOI Converter

bn between the

trial version

(5.12)

hDP://www.stdutilitv.com
where the conditional probability PIll (Yl, tllY2, t2) has the property
(5. ]3)
as can be demonstrated easily.
We can also introduce a joint conditional probability density as follows:
Pkll(Yl, tl; ;Yk, tk IYk+l, tk+l,;. ;Yk+l, tk+l)
= (the joint conditional probability density that the stochastic variable
Y has values (Yk+l, tk+l; ... ;Yk+l, tk+l) given that (Yl, tl; ... ;Yk, tk)

are fixed).

(5. 14)

The joint conditional probability density is defined as


PkI1(YI, tl; ... ;Yk,tkIYk+l,tk+l;
... ;Yk+l,tk+l)
Pk+l (Yl, tl; ... ;Yk, tk; Yk+l, tHl; ... ;Yk+l, tk+l) .
=
Pk(Yl, tl; ;Yb tk)

(5.15)

233

GENERAL THEORY

The joint probability densities are important when there are correlations
between values of the stochastic variable at different times-that is, if the
stochastic variable has some memory of its past. However, if the stochastic
variable has memory only of its immediate past, then the expressions for the
joint probability densities and the joint probability densities simplify
considerably.
If the stochastic variable has memory only of its immediate past, the joint
conditional probability density Pn-1 jl (Y1, t1; ... ;Yn-1, tn-llYn, tn), where
tl < t2 < ... < tn, must have the form
(5.16)
That is, the conditional probability density for Yn at t is fully determined by the
value of Yn-I at tn-I and is not affected by any knowledge of the stochastic
variable Y at earlier times. The conditional probability density
PIll (YI, tIl Y2, t2) is called the transition probability. A process for which
Eq (5.16) is satisfied is called a Markov process. A Markov process is fully
determined by'
). The whole
hierarchy of pr
Converted with
For example,

STDI Converter
trial version
If we integrate

P2(YI,

3, t3)
,t2IY3,t3)'
(5.17)

""'"'r--~h~n.....
P,.........,:/~/www~_.S~t---:d~U-rti~li,.........tv~.C~0-::7"'m,............

tl;Y3, t3)

= PI(YI, ti)

J PIII(YI,

tl!Y2, t2)PIIl(Y2,

t2!Y3, t3)dy2.
(5.18)

If we now divide Eq. (5.18) by PI(YI,


Pill (YI, tl!Y3, t3)

t1), we obtain

Pill (YI, tl ! Y2, t2)P1I1 (Y2, t2!Y3, t3)dy2.

(5.19)

Equation (5.19) is called the Chapman-Kolmogorov


equation. Notice that we
have broken the probability of transition from YI, t1 to Y3, t3 into a process
involving two successive steps, first from Y1, tl, to Y2, ti and then from Y2, tz to
Y3, t3. The Markov character is exhibited by the fact that the probability of the
two successive steps is the product of the probability of the individual steps.
The successive steps are statistically independent. The probability of the
transition Y2, t2 ~ Y3, t3 is not affected by the fact that it was preceded by a
transition YI, tl ~ Y2, tzIn the next section we will use these equations to study some of the simplest
Markov processes, namely those of Markov chains.

234

STOCHASTIC DYNAMICS AND BROWNIAN MOTION

s.c.

MARKOV CHAINS [2-6]

One of the simplest examples of a Markov process is that of a Markov chain.


This involves transitions, at discrete times, between values of a discrete
stochastic variable, Y. Let us assume that Y has realizations {y(n)}, where
n = 1,2,
,M, and that transitions
occur at times t = ST, where
s = 0, 1, ,00. We let P(n, s) denote the probability that Y has realization,
y(n), at "time" t = s. We let PII1(nl, Sljn2, S2) denote the conditional
probability that Y has realization y(n2) at time S2, given that it had realization
y(nI) at time Sl. The two quantities P(n, s) and PIll (nl' sll n2, S2) completely
determine the evolution of a Markov chain.
We can write Eq. (5.12) for the probability P(n, s) in the form
M

P(n, s+ 1) = LP(m,

s)PIII(m,

sin, s+ 1)

(5.20)

m=l

and from the Chapman-Kolmogorov


conditional pr

equation, (5.19), we can write the

Converted with
PIll

(no, So

STOU Converter

The quantity
trial ve rsion
conditional pro
state n in the n
basic transition mec arusm III e system.
Let us now introduce the transition matrix Q(s), whose (m,
the transition probability

hnp://www.stdutilitv.com

Qm,n(s) == PIII(m, sin, s

+ 1).

1).

(5.21 )

bility. It is the
it will jump to
ation about the
n)th element is
(5.22)

In this section we will consider Markov chains for which the transition matrix is
independent of time, Q(s) = Q. In Section S5.A, we consider the case when the
transition matrix depends periodically on time Q(s) = Q(s + N).

S.C.I. Spectral Properties


For the case of a time-independent transition matrix, we have
Qm,n

= P111(m, O], 1) = Pll1(m, sin, s+ 1).

(5.23)

The formal solution ofEq. (5.21) for the conditional probability can be obtained
by iteration and is given by
(5.24 )

235

MARKOV CHAINS

where the right-hand side denotes the (m, n)th element of the matrix
the s - So power. The probability P( n, s) is given by

Q raised to

P(n, s) = LP(m,

O)(QS)m,n'

(5.25)

m=l

It is useful at this point to introduce Dirac vector notation. We will let


P(n, s) == (p(s) In), and PIll (m, So I n, s) == (m I P(so Is) I n). Here (p(s) I is the
probability vector and P(so Is) is the conditional probability matrix. The left and
right states, (n I and I n) respectively, denote the possible realizations of the
stochastic variable Yand are assumed to be complete,
In)(nl = 1, and
orthonormal, (min) = 8m,n. The probability P(n, s) == (p(s) In) can be thought
of as the nth component of a row vector. We shall now express P(so I s) and
(p(s) I in terms of eigenstates of the transition matrix, Q.
The transition matrix, Q, in general is not a symmetric matrix. Therefore, the
right and left eigenvectors
of Q will be different. The eigenvalues,
Ai
(i = 1, ... ,M), of Q are iven b values of A which satisf the condition that
the determinant

I:~=1

Converted with

(I is the unit
which mayor

STOI Converter
trial version

(5.26)
eigenvalues,
nvalue, there

hDP://www.stdutilitv.com

will be a left ei
eft eigenstate
satisfies the eig:O"""""..._,...,...---.......-.....,........,.,..,--""T7{""1"T""'"""<lr-\7{'""TT""iiii~-r----M

Xi(n)Ai = L

Xi(m)Qm,n,

(5.27)

m=l

where Xi (n) == (xd n) and Qm,n == (m I Q In). The right eigenstate


eigenvalue equation, Ad~M = QI~i)' or

satisfies the

Ai~i(n) = L

Qn,m~i(m),

(5.28)

m=l

where ~i(n) == (n I ~i)'


We can prove that the left and right eigenvectors of Q are complete and
orthogonal. Let us first prove orthogonality. Consider the eigenvalue equations
~ I ~j) = Q I ~j) and (Xi I Ai = (Xi I Q. Multiply the first equation by (Xi I and
the second equation by I ~j)' and subtract the first equation from the second.
We get

236

STOCHASTIC DYNAMICS AND BROWNIAN MOTION

Thus, if Ai =1= Aj, then (Xi 1 1M = O. If x, = Aj, then (Xi 1 'i) can be nonzero. We
shall always normalize the left and right eigenvectors so that
(5.30)
Let us next consider completeness. Let us expand a probability vector (p I, in
terms of left eigenvectors so that (p 1 =
Qi(Xi I. We then multiply on the
right by I 'j) and use the orthonormality
condition, Eq. (5.30). We find
Qj = (p 1 'j) so that

I:~1

(pi = L(pl'i)(xd
i=l
This expresses the completeness

condition

(5.3] )

I Ai II'i(n) I ~
Let us assume that

'i(n)

1 ~

o.; I'i(m)

(5.32)

C for all n, where C is a positive constant. Then

LQn,ml'i(m)!
m=l
Assume that for n = no,
IAdC~C and

L
m=l

'i(no)

~CLQn,m
m=l
1

C.

= C. Then Eqs. (5.32)

(5.33)

and (5.33) give

(5.34)
Thus, all eigenvalues of the transition matrix have a magnitude less than or
equal to one.
.
Now that we have established that ! Ai! ~ 1, we can also prove that Ai = 1 IS
always an eigenvalue. Let us first note that the particular choice, 'i(n) = 1 for

237

MARKOV CHAINS

all n, is a right eigenvector with eigenvalue, Ai = 1 [cf. Eq. (5.28) and note that
Qnm = 1]. The corresponding left eigenvector, (Xi I, with eigenvalue,
Ai = 1, must satisfy the eigenvalue equation

l:~=1

Xi(n) = Lx;(m)Qm,n.

(5.35)

m=l

Equation

(5.35)

is the equation
for a stationary
probability
vector,
is, one that does not change with time.
We can expand the transition matrix, Q, in terms of its left and right
eigenvectors. Consider the left eigenvalue equation, (Xi Ix, = (Xi I Q. Multiply
from the left by the right eigenstate, I 'i) , and sum over i. If we use the
completeness relation, Eq. (5.31), we obtain Q =
Ail'i)(xil or

Xi(n) == PsT(n)-that

I:~l

Qm,n

Ai'i(m)x;(n).

(5.36)

i=l

If we now com

Converted with

and right eigen

ality of the left

STOI Converter

(5.37)

trial version
Thus, we have

hDP://www.stdutilitv.com

kov process in

terms of the eigenvalues and left and right eigenvectors of the transition matrix
Q.
The detailed behavior of the conditional probability depends on the structure
of the transition matrix. There is one case that we are particularly interested in,
and that is the case when the transition matrix is regular. The transition
matrix Q is called regular if all elements of some power QN (N an integer) are
nonzero. If Q is regular, the probability P(n, s) tends to a unique stationary
state PsT(n) after a long time. This means that there will be only one eigenvalue
of Q with eigenvalue A = 1. Let us denote it Al = 1. Then Eq. (5.37) can be
written
M

PIll (m, So In, s)

= 'I

(m)PST(n) + L

A:-SO'i(m)Xi(n),

(5.38)

i=2

where 'l (m) = 1 for all m. Since


and obtain
lim

s-oo

PIll(m,

Ai < 1 for i

soln,s) = 'I(m)PST(n)

=1=

1, we can take the limit

with 'I (m) = 1 for all m.

s ---; 00

(5.39)

238

STOCHASTIC DYNAMICS AND BROWNIAN MarION

Thus, for regular transition matrices, the conditional


unique value for long times. Similarly,

P(n, s)

lim

s-+oo

probability

tends to a

Psr(n).

(5.40)

In Exercise 5.1, we shall illustrate these ideas with an example.


Markov chains which are governed by regular transition matrices are
ergodic. That is, it is possible, starting from anyone realization of the stochastic
variable, to reach every other realization through transitions during the
evolution of the system. This is no longer true if, for example, the transition
matrix is block-diagonal. Then some realizations will be decoupled from others.
When the transition matrix is block-diagonal,
the system can have multiple
long-time states .

EXERCISE
5.1. Consider two pots, A and B, with three red balls and
two white balls distributed between them so that A always has two balls and

B always has

LI_

_J~~.

s for the pots,

~.

Converted with

as shown in

STDI Converter

configuration.
random and i
the condition
PI (2, 0) = 0,

trial version

out of B at
hiatrix Q and
Pi(l, 0) = 1,
s) (n = 1, 2,

hDP://WWW .stdutililV.com

Pfo~t~h~::~~~

::~~n~~
tt~;
initial conditions

these three

as in part (b).

B,

B,

A,

y,

Y3

Answer:
(a) Inspection shows that we can make the following transitions from Yi
to Yj with transition probability, Qib such that Qll = 0, Q12 = 1,
Q13 = 0, Q21 =~, Q22 =!, Q23 = Q3I = 0, Q32 = ~, and Q33 =
The transition matrix and its square are given by

i,

t)
o ~

= (~

~
3

respectively.

!
3

The transition

and

2_

Q -

1 623' 1 I)
12310

( 129363'36
1

matrix is regular

20

36

so this system has a

239

MARKOV CHAINS

unique long-time stationary state. The eigenvalues of Q are Al = 1,


A2 = ~,and A3 = -~. The right eigenstates can be written

The left eigenstates can be written

(xII

(to

fo

10),

(x31 = (~ - ~ ~).
We can now construct the conditional probability matrix

P(sols)

= tg-sol'lj;;)(x;1

= (~

;=1

TO TO

~)
3

TO

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com
Note that a the pro a iliy IS carne
y e ong-time state.
(b) Denote the probability vector as (p(s) I = (P(I, s), P(2, s), P(3, s)).
Then (p(s) I = (p(O) IP(O Is). If we assume an initial condition
(p(O) I = (1,0,0), then the probability vector at time s is

(c) The first moment is given by (y(s)) = I:!=1 nPI(n, s). From the
expression for (p(s) I we obtain
11 6 (I)S
=5-5
6 .

(y(s))

Finally, we compute the autocorrelation function,


3

(y(O)y(s))

=L

L mnl', (m, 0)P

m=l n=l

I11

(m,

0 I n, s).

240

STOCHASTIC DYNAMICS AND BROWNIAN MOTION

We obtain
(y(O)y(s))

66 4 (l)S
5
+5 6 .

-1

Note that the eigenvalue A3 =


does not contribute to (y(s)) and
(y(O)y(s)).
However, if we choose the dependence of y(n) on n
differently, it will, in general, contribute to these.

S.C.2. Random Walk


The problem of random walk on a lattice which was considered in Sections 4.E
and 4.F can be formulated as a Markov chain of transition probabilities. For
simplicity, let us consider random walk on an infinite one dimensional lattice
with lattice spacing, ~, along the x axis, and let us assume that the time
between steps is T. Let PI (n~, ST) be the probability to find the particle at point
x = n~ after S r"'-"""' ..............--L..LJ"""'"""-----------------,

Converted with
PI(n~,(S+

STOU Converter

l)T),

(5.41)

trial ve rsion
o go from site
where PIll (m
x = m ~ to sit
As a specifi
,
andom walker
has an equal probability to go one lattice site to the left or right during each
step. Then the transition probability is

hUP:llwww.stdutililV.com

and Eq. (5.41) takes the form


PI (n ~, (S + l)T)

!PI ((n + 1)~,

ST)

+ !PI ( (n - 1)~, ST) .

(5.43)

We can obtain a differential equation for the probability PI (n~, ST) in the
continuum limit. Let us subtract PI (n~, ST) from both sides of Eq. (5.43) and
divide by T. We then can write
PI (n~,

ST + T) - PI (n~, ST)
T

~2

2T

!PI(n~

+ ~,ST) + PI (n~ - ~,ST) - 2PI(n~,sT)1


~2

(5.44 )

J.

241

THE MASTER EQUATION

If we now let x = n~, t = S7", and take the limit ~ ~ 0 and 7"~ 0 so that
D == ~ 2/27" is finite and x and t are finite, we obtain the following differential
equation for P(x, t) :
aPI (x,

at

t) = D a2PI (x, t)

(5.45)

ax2.

(cf. Egs. (4.62)-(4.64)). Equation (5.45) is a diffusion equation for the


probability density PI (x, t).
Let us solve Eq. (5.45) for the case PI (x, 0) = 8(x). We first introduce the
Fourier transform of PI (x, t) :

i\(k, t)

t)e

= [,,' dxPl(X,

ikx
,

Then Eq. (5.45) takes the form


(5.47)

Converted with
We can solve E

STOI Converter
trial version
(5.48)

hDP://www.stdutilitv.com
where we have used the fact that P(k, 0)
now take the inverse transform to obtain

1 since PI (x, 0)

= 8(x).

We can

(5.49)
It is easy to show that (x(t)) = 0 and (x2(t))

S.D. THE MASTER EQUATION

2Dt.

[3, 4]

We often need toconsider processes for which the time interval between events
can vary in a continuous and random manner, but as in Section 6.C the
realizations of the stochastic variable Y are discrete. For such processes, we
need to obtain a differential equation for the time dependence of the probability
PI (n, t). In this section we consider processes for which the elementary events
or transitions occur over a very short time interval compared to the time during
which the stochastic process evolves. The time evolution of such processes is
governed by the master equation.

242

STOCHASTIC DYNAMICS AND BROWNIAN MOTION

S.D.I. Derivation of the Master Equation


To begin, let us rewrite Eq. (5.12) in the form
M

PI(n, t+~t)

LPI(m,

t)PIII(m,

tin, t+~t).

(5.50)

m=l

The differential equation for PI (n, t) can be constructed from Eq. (5.50) if we
note that

aPI (n, t) == lim (PI (n, t + ~t) - PI (n, t))


at
~HO
~t
1 M
= lim A LPI(m,

~t-+Out m=l

(5.51 )

t)(PIII(m,

tin, t+~t)

-8m,n)'

Since we will take the limit ~t ~ 0, we can expand the transition probability
tin, t'
. .
y the lowestorder term. In
Converted with
most general
form is
PII1(m,

STOI Converter
trial version

+ ... ,

(5.52)

hDP://www.stdutilitv.com

where wm,n(t) i
, m,n(t)~t is the
probability of a transition from state m to state n during the time interval
t ~ t + ~t. Similarly, [1 - L:~l Wm,l(t)~t] is the probability that no transition
occurs during the time interval t ~ t + ~t. If we now substitute Eq. (5.52) into
Eq. (5.51), we obtain

aP1(n, t) ~
at
= L)P1 (m, t)Wm,n(t) - PI (n, t)Wn,m(t))].

(5.53)

m=l

Equation (5.53) is called the master equation. The master equation gives the
rate of change of the probability PI (n, t) due to transitions into the state n from
all others states (first term on the right) and due to transitions out of state n into
all others states (second term on the right).
The conditional probability PIll (no, 0 In, t) also satisfies a master equation

aP111(nO, Oln, t) ~
at
= L.)P111(no,

Olm, t)Wm,n t) -P1II(no,

())]
Oln, t)Wn,m t ,

m=l

(5.54 )

243

THE MASTER EQUATION

where PIll (no, 0 I n, t) is the probability to find the system in the state n at time
t, given that it was in the state no at time t = O. The conditional probability
satisfies an initial condition PIll (no, 0 I n, 0) = 8n
The master equation can be written in a more concise form if we introduce
the transition matrix
,1I{J'

Wm,n(t) = wm,n(t) - 8m,n

L wn,n,(t).

(5.55)

n'=l

The master equation then takes the form


aPI(n,

at

t)

= ~PI(m,

t)Wm,n(t).

(5.56)

m=l

From Eq. (5.55), we see that the transition matrix must satisfy the conditions

Converted with
Thus, the entri
We can wri
notation. We le

(5.57)

STOI Converter
trial version

Introduce Dirac
lity vector, and
PIII(no, toln,.
nal probability
operator. Similarly we let Wm,n(t) = (m I W(t) I n). The master equation for the
probability vector becomes

hDP:llwww.stdutililV.com

a (~;t) I = (p(t) I W(t),

(5.58)

and the master equation for the conditional probability operator becomes

ap~~lt)

P(Olt)W(t),

(5.59)

The transition matrix, Wm,n(t), in general is not symmetric so its left and right
eigenvectors will be different. However, one can often use the method of
Section 5.C.1 to obtain a spectral decomposition of W(t) and therefore of
(p(t) I and P(O I t). However, care must be used. There may be cases where the
eigenvectors of W(t) do not span the solution space [7]. Then the spectral
decomposition cannot be used. There is one type of system for which a spectral
decomposition can always be done, and that is the case for which the transition
rates wm,n(t) satisfy detailed balance.

244

STOCHASTIC DYNAMICS AND BROWNIAN MOTION

S.D.2. Detailed Balance [3, 4, 8]


The transition

rates satisfy

detailed balance if
(5.60)

where pS(n) == limHooPI (n, t) is the long-time stationary probability of the


system (we assume the transition rates, wm,n, are independent of time) and is
independent of time. ps](n) is the left eigenvector of the transition matrix, W, so
that (pS IW = 0 where pS(n) = (pS In). Equation (5.60) tells us that at equilibrium, the flow of probability into level n from level m is equal to the flow of
probability from level m to level n. It is useful to note that the state, pS(n), can
be obtained
from Eq. (5.60) by iterating.
For
example,
PS(2) =
PS(1)(wI,dw2,
I)' PS(3) = PS(2)(W2,3/W3,2) = PS(1)(wI,dw2,
1) (W2,3/W3,2)
and so on. P' (1) can then be found by requiring that the probability be
normalized to one , ~M
ps](n)=l.
L_m=1
Given Eq. (5.60), we can show that the dynamical evolution of the master
equation is governed bv a symmetric matrix. Let us define

Converted with
Vn,m =

STOI Converter

(5.61 )

trial version

hDP://www.stdutilitv.com
where

we have

P(n, t)

use~

(5.60).

If we now introduce

a new function,

PI (n, t)/ yI PS(n), the master equation takes the form

ap(n, t)
at

~-

= ~P(m,

t)Vm,n'

(5.62)

m=1

Let us again use Dirac notation to write (p(t) In) = P(n,


(n I Vim). Then the solution of the master equation becomes

(p(t)

I = (p(O) leVt.

t) and Vn,m =

(5.63)

Since Vn m is a symmetric matrix, it has a complete orthonormal set of


eigenvectors, We shall denote the eigenvalues of V by Aj and left and right
eigenvectors by ('Ij;j I and I 'lj;j), respectively, where i = 0, ... ,M - 1. The left
and right eigenvectors are the same. Thus, ('Ij;j IV = ('Ij;j I Aj and VI 'lj;j) = Ai I 'ljJ;).
Since the eigenvectors
form a complete orthonormal
set, we can wri te
o~<:~ormality
condition as ('Ij;i I 'lj;j) = bj,j and the completeness condition as
E j=O I'lj;j) ('Ij;j I = 1.

245

THE MASTER EQUATION

The solution to the master equation can now be written in terms of a spectral
decomposition
M-I

(p(t)

L (p(O) I '!f;i)/'i

('!f;i I

(5.64)

i=O

In terms of the probability PI (n, t), it takes the form


M-I

P1(n, t)

= ~~

ps(n)
ps(m) PI(m,

AI

O)(ml'!f;i)e

('!f;iln).

(5.65)

The eigenvalues Ai must be negative or zero. Let i = 0 denote the zero


eigenvalue, AO = O. Then
pS(n) = limPI(n,
1-+00

t) =

In order to be c
since
PI (m

O)(ml'!f;o)('!f;oln).

(5.66)

~(m), for all m,

Converted with

Lm

PI(n, t) =

ps((n)) PI(m,
ps m

m=I

STOI Converter
trial version

In).

(5.67)

hDP:llwww.stdutililV.com

In Exercise 5.2
obeys detailed
balance and in Exercise 5.3 we give an example of one that does not obey
detailed balance.

EXERCISE 5.2 Consider an asymmetric random walk on an openended lattice with four lattice sites. The transition
rates are
WI,2 = W4,3 = 1, W2,3 = W3,4 = W2,I= W3,2 =!, and Wi,j = 0 for all
; other transitions. (a) Write the transition matrix, W, and show that this
system obeys detailed balance. (b) Compute V and find its eigenvalues and
eigenvectors. (c) Write PI (n, t) for the case PI (n, 0) = 8n,I. What is
PI (2, t)?
I

i,

I
I

Answer:
(a) The transition matrix, W, is given by

w=

(I 11

~I

~
-1

).

246

STOCHASTIC DYNAMICS AND BROWNIAN MOTION

4,

AO = 0, Al = -1 +
A2 = -2,
and A3 =
-1 The left eigenvector of W with eigenvalue, AO = 0, is the
stationary probability distribution, (pS I = (-k, f3 ' -& '
This system
satisfies detailed balance because pS(n)wn,n+1 = pS(n + l)wn+l,n. For
example, PS(I)wI,2
= (-k)(I)
= and PS(2)W2, I = (f3)(~) =
(b) The matrix V is given by

It has eigenvalues

4.

f6).

-k.

-k

V=

(11 ~ ~ ~).
o

Y1
2

-1

4,

It also has eigenvalues AO = 0, Al = -1 +


-1 The orthonormalized eigenstates are

4.

{, 1&,~) ",

(ol = ( ~,
(~I

= -2,

and

A3

(0.196,0.392,0.679,0.588),

f"'-'---L'L...I....lIo.._LL.L...L.J~..____L..I........I..........__---L...L-<o..l..L.L..l'---

__

-------,

Converted with

(~21
and (~
(c) For th

Y2

STOU Converter

e written

trial ve rsion

hnp://www.stdutilitv.com

n).

Using the numbers from (b) we find


PI (2, t) ~ 0.154
Note that PI (2, 0)

+ (0.799)e-0.567t - (0.154)e-2t - (0.799)e-1.433t.


=

0 as it should .

EXERCISE 5.3 Consider an asymmetric random walk on a periodic


lattice with four lattice sites. The transition rates are WI, 2 = W2,3 =
W3 4 = W4 1 =~, W2 I = W3 2 = W4 3 = WI 4 =
and Wj j = 0 for all other
tra~sition~. Write the transi'tion matrix,
and show th~t this system does
not obey detailed balance.

W,

1,

Answer:
The transition matrix, W, is given by
W _
-

(!1~l i ).
0
~

-1
~

~
-1

247

THE MASTER EQUATION

The left eigenvector of W with eigenvalue, AO = 0, is the stationary


probability distribution, (pS I =
Therefore, detailed balance
cannot be satisfied and we cannot find a symmetric matrix which governs
the dynamics of this system. It is interesting that the long-time state is one
for which it is equally likely to find the random walker at any lattice site. Yet
because detailed balance is not satisfied, there is also a probability current
around the lattice.

(!,!,!,!).

S.D.3. Mean First Passage Time [9]


One of the useful quantities we can compute is a mean first passage time. If we
consider the random walk problem, the mean first passage time is the average
time for the walker to reach some site, np' for the first time assuming that the
walker started at site no at time t = o. Computing this quantity involves a trick.
We will assume that PI (np, t) == 0 for all time. This way, if the walker steps on
site n p' it can never return to the random walk. This is called an absorbing
boundary condi
To be clear a
Converted with
he conditional
probabili ty to fi
site no at time
t = 0 (we suppr
0) = 8n,no and
Qnp (t) = 0 (the
of motion for
trial version
Qn(t) are the sa
ept that we set
Qn (t) = O. Thl

STOI Converter

hDP://www.stdutilitv.com
M

Qm(t)wm,n(t)

- Qn(t)

L wm,n(t))
m=I

(5.68)
Note that we set Qnp(t) = 0 but wnp,n -=I 0 in Eq. (5.68). In Dirac notation we
write Eq. (5.68) as
aQ(t)
at

= QM

'

(5.69)

Qn(t) == (no I Q(t) In) and (m 1M In) = wm,n - 8m,n I:~=I wn,n' with
The matrix M is not symmetric in general, and therefore its left and
right eigenvectors will not be the same. We can use the methods of Section 5.C
to expand the solution of Eq. (5.69) in terms of left and right eigenvectors of M
and its eigenvalues (see Exericse 5.4). However, when using this approach one
must check that the eigenvectors are complete. Since Qn(t) must decay, at least
for a finite lattice (the walker eventually steps on site, np ), all the eigenvalues of
M must be negative.
where

m =1= np.

248

STOCHASTIC DYNAMICS AND BROWNIAN MOTION

The probability that the walker is still "alive" (i.e., has not stepped on the
site, n = np) at time tis
M

Pno(t)

==

Qn(t).

(5.70)

n(=lnp)=l
We now define

== {the probability to reach the site,

Ino,np(t)

the time interval t

--+

np, during
(5.71 )

+ dt}

We then note that to be "alive" at time t, the probability must equal the sum of
the probabilities to either have "died" during the time t --+ t + dt or stilI be
alive at time t + dt. Thus, we can write
(5.72)
If we now not
write

., then we can

Converted with

STOI Converter

(5.73)

trial version
From Eqs. (5j
written

Ino,np(t)

, can also be

hDP://www.stdutilitv.com

=-

L dQn(t)
~
M

n(=lnp)=l
M

L L

Qm(t)wm,n,(t)

n'(=lnp)=l m(=lnp)=l

L
n'(=lnp)=l

Qn,(t)

L Wm,n,(t)
m=l

=-

Qn,(t)wnp,n,(t)).

n'(=lnp)=l
(5.74)

Thus Ino, n (t) can be expressed in terms of the probability to find the walker on
neighbors of the site np and the transition rate between n and its neighbors.
The average time it takes the walker to reach the site np, given that it starts at
site no at time t = 0, is
(t)

= Joo
o

dt tfno,np(t)

= Joo
0

dtPno(t),

(5.75)

249

THE MASTER EQUATION

where we have used Eq. (5.72) and have integrated by parts. (t) is the mean first
passage time .

EXERCISE 5.4 Consider an asymmetric random walk on a lattice with


five lattice sites. Assume that the fifth site, P, absorbs the walker. The

1
i transition
rates Wl,2 = Wl,3 = Wl,4 = Wl,P = 4' W2,1 = W2,3 = W2,P =
1 -3,31W
-W 32--W 34--3'
_1
W41--W 43--W 4P--3'
_1
We.i =
-w P2- -w P4- 1

'

and wi] = 0 for all other transitions. (a) Write the transition matrix, M,
and compute its eigenvalues and left and right eigenvectors. (b) If the walker
starts at site n = 3 at time t = 0, compute the mean first passage time.

~,

Answer:
(a) The transition matrix, M, is given by

Converted with

STOI Converter
trial version
It has e

-1.5.

hDP://www.stdutilitv.com

tI ~

(x

(x ~ I

(0.286, 0.204, 0.265, 0.204),

= (0, -0.5,0,0.5),

I ~ (0.614, -0.054, -0.415,


(xi I = (-0.2,0.3, -0.3,0.3).
(X~

-0.054),

The right eigenvectors are

I Xl

) ~

1.0 )
0.950
1.234 '

Ixf)

(+),

0.950
R

IX3) ~

1.0 )
-0.117
-0.901
'
-0.117

1.0
-05)

Ix:)

~~~o .
1.0

250

STOCHASTIC DYNAMICS AND BROWNIAN MOTION

(b) The spectral decomposition of the matrix Q(t) can be written


4

Q(t)

Q(O) I>'\jtlxf)(xfl.

(1)

i=l

If we take matrix elements and note that (no IQ(O)

In) = 8n,no'

we find

(noIQ(t)

In) = l::>'\jt(nolxf)(xfln).

(2)

i=l

The mean first passage time is

(t)

oo

= J0

dtPno(t)
4

= ~~(nolxf)
If we

(t) ~

=~

JOO
0

dt(no IQ(t) In)

(-1)
>:: (xfln).

Converted with

(3)

(a), we find

STOI Converter
S.E. BROW

trial version

hDP://www.stdutilitv.com

Brownian moti
dence, on the
"macroscopic" scale, for the discrete or atomic nature of matter on the "microscopic" scale. The discreteness of matter causes fluctuations in the density of
matter, which, in tum, causes observable effects on the motion of the Brownian
particle. This can be seen if one immerses a large particle (usually about one
micron in diameter) in a fluid with the same density as. the particle. When
viewed under a microscope, the large particle (the Brownian particle) appears to
be in a state of agitation, undergoing rapid and random movements. Early in the
nineteenth century, the biologist Robert Brown wrote a paper on this
phenomenon [10] which received wide attention, and as a result it has been
named for him.
The modem era in the theory of Brownian motion began with Albert
Einstein, who, initially unaware of the widely observed phenomenon of
Brownian motion, was looking for a way to confirm the atomic nature of matter.
Einstein obtained a relation between the macroscopic diffusion coefficient, D,
and the atomic properties of matter. This relation is D = RT / NA 67r'fJa, where R
is the gas constant, NA = 6.02 X 1023 mor-' is Avogadro's number, T is the
temperature in kelvins, 'fJ is the viscosity, and a is the radius of the Brownian
particle [11, 12]. It has since been confirmed by many experiments on Brownian
motion [13].

251

BROWNIAN MOTION

In this section we derive the theory of Brownian motion starting from the
Langevin equations for a Brownian particle. That is, we focus on a large particle
(the Brownian particle) immersed in a fluid of much smaller atoms. The
agitated motion of the large particle is much slower than that of the atoms and is
the result of random and rapid kicks due to density fluctuations in the fluid.
Since the time scales of the Brownian motion and the atomic motions are vastly
different, we can separate them and focus on the behavior of the Brownian
particle. The effect of the fluid on the Brownian particle can be reduced to that
of a random force and a systematic friction acting on the Brownian particle.
The Langevin theory of Brownian motion provides a paradigm theory for
treating many-body systems in which a separation of time scales can be
identified between some of the degrees of freedom. For this reason we consider
it in some detail here.

5.E.l. Langevin Equation [3, 4]


Consider a particle in a fluid undergoing Brownian motion. For simplicity we
will consider motion in one dimension. The results can easilY be generalized to
three dimensio
n the fluid but
that the effect c
Converted with
is proportional
~~et~u~:.lc;~~~

STOI Converter
trial version

hDP://www.stdutilitv.com
....,..,.,
= v(t)
dt
'

fluctuations in
(5.76)
(5.77)

where v(t) and x(t) are the velocity and position, respectively, of the particle at
time t, m is the mass of the particle, and 'Y is the friction coefficient. Equations
(5.76) and (5.77) are the Langevin equations of motion for the Brownian
particle.
The random force, e(t), is a stochastic variable giving the effect of
background noise, due to the fluid, on the Brownian particle. We will assume
that e(t) is a Gaussian white noise process with zero mean so that (e(t))e = O.
The noise is Markovian and stationary and the average, () e' is an average with
respect to the probability distribution of realizations of the stochastic variable
~(t). We will not write the probability distribution explicitly. The assumption
that the noise is white means that the noise is delta-correlated,
(5.78)
and therefore (as we shall show in Section 5.E.2 its power spectrum contains all
frequency components. The weighting factor, g, is a measure of the strength of
the noise. Such a correlation function indicates that the noise is not correlated

252

STOCHASTIC DYNAMICS AND BROWNIAN MOTION

from one instant to the next, and therefore it is impossible to represent a single
realization of ~(t) in terms of a continuously drawn line. The fact that the noise
is Gaussian with zero mean (cf. Exercise 4.9 means that correlation functions
containing an odd number of terms, ~(t), are zero and that correlation functions
containing an even number of terms, ~(t), can be expressed in terms of sums of
products of the pairwise correlation function, (~(tt)~(t2))e'
For example,

(~(tl )~(t2)~(t3 )~(t4))

e = (~(tl )~(t2)) e(~( t3)~(t4)) e


+ (~(tt)~(t3))e(~(t2)~(t4))e

+ (~(tl)~(t4))e(~(t2)~(t3))~.

Equations (5.76) and (5.77) can be solved fairly easily. Let us assume that at
time t = 0, the velocity and position of the Brownian particle are v(O) = vo and
x(O) = xo, respectively. Then the solution the Eqs (5.76) and (5.77) is

v(t) = voe-b/m)t

+ -I Jt dse-b/m)(t-s)~(s)
m

(5.79)

and

Converted with

x(t) = Xo

s).

SIDU Converter

Equations (5.7'

(5.80)

zation of @.

Since ~(t) is a
trial version
astic variables
whose properti
(subject to the
condition that v
~isplacement is
((x(t) - xo))e = !!!. (1 - e 'm"')' )vo.
We can also obtain correlation functions from Eqs. (5.78), (5.79), and (5.80).
If we make use of the fact that (vo~(t))e = 0, then we can write

hDP:llwww.stdutililV.com

(V(t2)V(tt))e

v6e-b/m)(Hti)
(5.81 )

The integral is easily done to yield

We can also obtain the variance in the displacement. If we use Eqs. (5.78)
and (5.80) and the fact that (xo~(t))e = 0, we can write

_!_)

2
((x(t) - xo)2)c = m (v6 _
(1 - e-(r/m)t)2 +.!.
",2
2m,

,2,

'lrt - ~

(1 _

e-(r/m)t)] .
(5.83)

253

BROWNIAN MOTION

Thus, after a long time the variance goes as ((X(t2) - xo)2)e = (g/,,/)t
(neglecting some constant terms). This is the same behavior that we saw for
random walks in Chapter 4, if we choose a diffusion coefficient, D = g 12-y2.
We can use a simple trick to determine the value of g for a Brownian particle
in equilibrium with a fluid. Let us assume that the Brownian particle is in
equilibrium with the fluid and average over all possible initial velocities, vo- We
denote this "thermal" average by OT' By the equipartition theorem, for a
particle in equilibrium the average kinetic energy is !kBT for each degree of
freedom, m(v6h = kBT, where ke is Boltzmann's constant and T is the
temperature in kelvins. If the Brownian particle is in equilbrium, its velocity
correlation function must be stationary. Therefore, we must have v6 = g 12m,
so the first term on the right in Eq. (5.82) is removed. If we now take the
thermal average of Eq. (5.82), we see that we must have g = 2,kBT. The
correlation function can be written

((v(t2)v(tt))eh
The absolute
decay as the ti
the Brownian
For the cast:
variance of the

k T
m

__!_e-(T/m)lt2-ttl.

(5.84)

always
itial velocity of

elations

Converted with

STOI Converter

th the fluid, the

trial version

hDP://www.stdutilitv.com
,
,

(5.85)

where we have assumed that (xoh = (voh = 0 and that Xo and Vo are
statistically independent so that (xovo) = O. Thus, after a long time,
(((x(t) - xo)2)eh = (2kBT lI)t and the diffusion coefficient becomes D =
kBT II. The friction coefficient, " can also be determined from properties of the
fluid and and hydrodynamics. For large spherical Brownian particles, we can
assume that the fluid sticks to the surface. The friction coefficient is then the
Stokes friction, , = 61r'T]R, where 'T] is the shear viscosity of the fluid and R is
the radius of the Brownian particle (see Sect. (SIO.0)). For very small
Brownian particles, stick boundary conditions might not apply and the friction
coefficient, " might be different.

EXERCISE 5.5. Consider a Brownian particle of mass m which is


attached to a harmonic spring with force constant k and is constrained to
move in one dimension. The Langevin equations are

dx
-=v

dt

'

254

STOCHASTIC DYNAMICS AND BROWNIAN MOTION

where Wo = /kfiii. Let Xo and vo be the initial position and velocity,


respectively, of the Brownian particle and assume that it is initially in
equilibrium with the fluid. Then by the equipartition theorem, the average
kinetic energy is !m(v6)T = !kBT and average vibrational potential energy is
!w6(x6h = !kBT. We also assume that Xo and vo are statistically
independent so (xovoh = O. (a) Show that a condition for the process to
be stationary is that the noise strength is g = 4,kBT. (b) Compute the
velocity correlation function, ((V(t2)V(tt) )eh.
Answer: The Langevin equations can be solved and give the following
expression for the velocity at time t:
2

v(t)

voe-rtC(t)

w
- _Qxoe-rtsinh(~t)

+ -1 Jt dt'~(t')e-r(t-t')C(t
m

- t'),

W5.

where C(t) = cosh(~t) - (T I~) sinh(~t), r = ,1m, and ~ = yir2 we use the fact that (xovoh = 0 and assume that tz > tl, the velocity
correlation function can be written

lf

((V(t2)

Converted with

STOI Converter
trial version
If we choose
obtain

~o---~h_n_p_:_II_www
__ .S_t_d_U---,ti.----li_tv_.C_o_m__

~t)
t).
algebra we

A similar calculation for tl > t: yields the same answer but with tl ~
Thus,

ti-

S.E.2. The Spectral Density (Power Spectrum)


A quantity of great interest in studying stationary stochastic processes is the
spectral density (or power spectrum). If one has experimental data on a
stochastic variable "p(t)-that is, if one has a time series for "p(t)-then it is
useful to compute the power spectrum because this contains a great deal of
information about the stochastic process, as we shall see below.

255

BROWNIAN MOTION

In practice, in dealing with data from experiments, one has a finite length of
time series to work with. Let us assume that we have a time series of length T.
We denote it 1jJ(t; T), where 1jJ(t; T) = 1jJ(t) for - T /2 ~ t ~ T /2 and 1jJ(t; T) = 0
otherwise. The limT-H)()'l/J(t; T) = 1jJ(t). The Fourier transform of 1jJ(t; T) is

{b( W; T) = [", dt1/J(t; T)eiwt

(5.86)

Because 1jJ(t; T) is real, we have ;j;(w; T) = ;j;*( -w; T), where


complex conjugation.
The spectral density is defined as
1 -

s.,. tjJ(w) == T-+oo


lim -'l/J*(w;
T

T)'l/J(w; T).

denotes

(5.87)

'f',

If we substitute Eq. (5.86) into (5.87), we find

Converted with

;T)

STOI Converter

(5.88)

trial version
where (1jJ(tl

(1jJ(tl

hDP:llwww.stdutililV.com
+ r)1jJ(tl))i ==

lim -IJOO
T-+oo

dtl1jJ(tl

+ r;

T)1jJ(tl; T).

(5.89)

-00

Equation (5.89) denotes the average of the product of numbers,


1jJ(tt; T)1jJ(tl + r; T), over each point of the time series, 1jJ(t). Since we assume
that the stochastic process is stationary, the stochastic properties of the time
series don't change as we shift along the time series. Therefore we expect that
the time average, (1jJ(tl + r)1jJ(tt));:, should be equal to the statistical average,
(1jJ(tl + r)1jJ(tl))tjJ, in which we average the product, 1jJ(tl + r)1jJ(tl), at a given
point, t = tl, but over an infinite number of realizations of the time series, 1jJ(t),
at time t. Therefore, for a stationary process, we expect that

If we now combine Eqs. (5.89) and (5.90), we obtain the following expression
for the spectral density:

256

STOCHASTIC DYNAMICS AND BROWNIAN MOTION

(a)

-i'C~~(T)
g

o
(b)

S~~(w)

~---g
....
....
....

..
....
....
..
....
.... .
..
....
....
....
.... ..
..
....
.......................
....
....
..
....
....
....
...
....
........................
.... .

0,

'"

'"

"

"

'"

'"

'"

..
..
.....
..
...
....
.
.....
..

"

Fig. 5.1. (a) The correlation function,

Converted with

STOI Converter
trial version
a

hDP://www.stdutilitv.com

o
(b)

Svv(w)

o
Fig. 5.2. (a) The correlation function, Cvv(t).

w
(b) The spectral density, Svv(w).

C{{(t).

257

BROWNIAN MOTION

It is interesting to compute the spectral densities for the Brownian motion


described in Section 5.B.l. Let us consider the case when the Brownian particle
is in thermal equilibrium with the fluid and the process is stationary. From
Eq. (5.78), the spectral density for the white noise is

Thus, a white noise process contains all frequencies with equal weight. The
correlation function, Cee(t), and spectral density, See(w), are shown in Fig. 5.1.
The spectral density for the velocity time series is

Sv, v

() J
W

oo
-00

dre

-iWT( (

Plots of the velocity correlation


particle are giVrr"'-'-'-...I..I
.L.....J:::!..I..IJ.

v tl

) ())

2k8T"(

(5.93)

+ r v tl e = m2w2 + "(2 .

function and spectral density for the Brownian

L_____._:J_L_

Converted with

STOI Converter

EXERC
harmonically
velocity corr
case WQ >

trial version

(W), for the


5.5. Plot the

,v(w) for the


article).

hDP://www.stdutilitv.com

Answer: Th~ern~~rnm~mmrn~rnm~nr~WTI~~~

where T

= ~ and 8 = y'w5 - r2 =

oo

S" ,(wi =

-ill. The spectral density can be written

dr cos (wr)e-rITI [~ cos (61T I) -

4"(k8Tw2

- m(w5 - 28w

+ w2)(w6 + 28w + w2)

~ sin 61 TI)]

Both Cv,v(r) and Sv,v(w) are plotted in the accompanying graphs.


Because the Brownian particle is underdamped, the velocity correlation
: function, Cv, v (T), oscillates as it is damped out exponentially. The spectral
: density, Sv,v(w), has peaks near W = wQ and the peaks have an
approximate width, ~.
I

When the noise is not "white", the spectral density still affords a means of
making contact with experiment [14].

258

STOCHASTIC DYNAMICS AND BROWNIAN MOTION

.... SPECIAL TOPICS


.... S5.A. Time Periodic Markov Chain
Let us now consider the case in which the transition probability is timedependent but periodic in time with period N. That is, Qn,m(s) ==
PIll (n, slm, s + 1) = PIll (n, s + Nlm, s + N + 1). The probability vector
after one period of the transition probability can be written

(P(N) I = (P(O)IQ(O)Q(I) x ... x Q(N - 1) == (P(O)IU,

(5.94)

where U = Q(O)Q(I) x ... x Q(N - 1) is the transition probability that


takes the system from the initial state, (P(O) I, to the state after one period,
(P(N) I. More generally, the probability vector after I periods of the transition
matrix is

I = (P(O) lUi.

(5.95)

Converted with

~eft and right


alue of U, and
vectors of U,
~nd (Xa IU =
is an M x M

(P(IN)
We can expand
eigenvectors of
let (Xa I and 11
respectively, w
(XaIAa. We th
matrix, then its

STOI Converter
trial version

hDP://www.stdutilitv.com
LVI

U =L

s; l7/Ja)(Xa I

(5.96)

a=1
The probability vector after I periods of the transition matrix is given by
M

(P(IN)I

LA~(P(O)I7/Ja)(xal.
a=1

(5.97)

The probability for the nth realization of the stochastic variable Yat time s
is given by
M

P(n, IN)

= IN

L L A~P(m, O)7/Ja(m)Xa(n).
m=1a=1

(5.98)

We will demonstrate the use of these equations in Exercise 5.7.


EXERCISE 5.7. Let us consider a stochastic variable Y with three
realizations, y(I), y(2), and y(3). Let us assume that the transition

259

SPECIAL TOPICS: TIME PERIODIC MARKOV CHAIN

probabilities
between these states are Ql, 1 (s) = Q2,2(S) = Q3,3(S) = 0,
Q1,2(S) = Q2,3(S) = Q3, 1 (S) = cos2(271"s/3),
and
Q1, 3(S) = Q2, 1 (s) =
Q3,2(S) = sin2(27I"s/3). If initially the system is in the state y(I), what is
the probability to find it in the state y( 2) after I periods of the transition
matrix?
The transition matrix has a period N = 3. In general it is given by

Answer:

0
Q(s) =

cos

sin2

e~S)
(
2
cos e~S)

e~S)
0

sin

2 e~S)

sin

cos2

e~S) )
e~S) .

Therefore,

Q(O) =

G D
1
0
0

and the trans

and

Q(l) = Q(2) =

4
0
3

i}

Converted with

STOI Converter

trial version

hDP:llwww.stdutililV.com

The eige~vah
~e+iO ,where
0 = tan- (3\~::J""/--'-l-"-::J'J'--.------------------'
:
The right eigenstates can be written
I

I
I

The left eigenstates

can be written

, and

(x31 = (-

t e+i7r /3, - t e-i7r /3, t) .

, We can now compute the probability P(2, 31). The initial probability
: is P(n, 0) =
1. From Eq. (5.98), the probability P(2, 31) is

s;

P(2, 31)

= Ai7Pl(I)Xl(2) + A~7P2(I)X2(2) + A~7P3(I)X3(2)


= "31 + (-1)"312

( 7 )

16 cos [0 -

271")
3"
.

density

260

STOCHASTIC DYNAMICS AND BROWNIAN MOTION

.... S5.B. Master Equation for Birth-Death Processes [3, 4]


One of the most common applications of the master equation is the description
of birth-death
processes
such as one finds in chemistry and population
dynamics. Birth-death processes are processes in which transitions can only
take place between nearest-neighbor states. For example, if one considers a
population in which only one individual is produced at each birth and one
individual dies at each death, then we have a typical birth-death process .

.... S5.B.l. The Master Equation


For concreteness let us consider a population of bacteria. We assume that at
time t, there are n bacteria and we make the following assumptions:

(1) The probability of a bacterium dying during the time t ___..t + ~t is given
by Wn,n-I (t)~t = dn(t)~t.
(2) The prob::th1.lJ..b..Lrl:...!.l._]3!l.l:!.~u.I.D:l.....bl;;Un.J:1...l:!.I~:a.d..._.1.n..JI:.u:n.GLL.........
+ ~t is given

by Wn,n+I
(3) The prob
t ___..
t+

(4) The pro


zero.

lation in time
+ dm(t))~t].
t ___..
t + ~t is

STOI Converter
trial version

.S_td_u_t_il_itv_._C_o_m_-----'

The transition p..,_____h-..-D_P_:/_/www


__
PIlI (m, tin, t

+ ~t) =[1 - (bm(t) + dm(t) )~t]8m,n


+ (bm(t)8n,m+I + dm(t)8n,m-I~t

+ ...

(5.99)

and the master equation takes the form


aPI (n, t)
at
=bn-I (t)PI (n - 1, t)
- (bn(t)

+ dn(t))P(n,

+ dn+I (t)P1 (n + 1, t)

(5.100)

t).

Note that we have allowed the birth and death rates, bn(t) and dn(t) respectively,
to depend on time. In most applications of Eq. (5.100) that we will consider
here, they will be independent of time.
Let us consider the case in which b and d; are independent of time. Then the
master equation will have at least one stationary state, P~ == P(n, 00), which is
independent of time and satisfies the condition ap~/at = o. (If the transition
matrix, W, is made up of uncoupled blocks, then there may be more than one
stationary state.) The master equation for the stationary state can be written in

SPECIAL TOPICS: MASTER EQUATION FOR BIRTH-DEATH PROCESSES

261

the form
(S.101)
Note that since dn+IP~+1 - bnP~ = dnP~ - bn-IP~_t,
the quantity J ==
bn-IP~-t - dnP~ must be independent of n. The quantity J is just the net
probability current between pairs of sites.
The case where no probability flows, J = 0, is equivalent to detailed balance
since
(S.102)
Systems obeying detailed balance are very similar to systems in thermodynamic
equilibrium, since no net probability current flows between microscopic states
of those systems. For systems obeying detailed balance, we can iterate
Eq. (S.102) to obtain an expression for P~ in terms of Po:
(S.103)

Converted with

STOI Converter

The value of
probability be
normalized to
The full rna
special cases.
trial version
In the followin
g it exactly for
some of the fe
e consider the
case when the
ulation number.
In Section SS.B.3 we consider the case when the birth and death rates depend
nonlinearly on the population number. We will discuss some approximation
schemes for solving the master equation in Section SS.D.

hDP://www.stdutilitv.com

.... S5.B.2. Linear Birth-Death Processes


Let us assume that the probability of a birth or death is proportional to the
number of bacteria present and is independent of time. Then b = (3n and
dn = ,n, where (3 and, are constants, and the master equation takes the form
apt (n, t)

at

= (3(n - I)Pt (n - 1,

t) + ,(n + l)Pt (n + 1, t) - ((3n + ,n)P(n, t).


(S.104)

Equation (S.104) describes a linear birth-death process because the


coefficients of PIon the right-hand side of the equation depend linearly on
n. Note that n is the number of bacteria and therefore n~O. Thus, Eq. (5.104)
must never allow probability to flow into regions where n < O. We see that

262

STOCHASTIC DYNAMICS AND BROWNIAN MOTION

Eq. (5.104) satisfies that condition. If n = -1, the coefficient of PI (n + 1, t) is


zero so flow can never occur from positive to negative values of n. Equation
(5.104) is said to have a natural boundary at n = O.This may not always be the
case and one must be careful when using such master equations.
Linear master equations depending on discrete stochastic variables are often
most easily solved by means of a generating function, F(z, t). We will use this
method to solve Eq. (5.104). The generating function is defined as
00

L t'Pl(n,

F(z, t) =

(5.105)

t).

n=-oo

Since we have a natural boundary at n = 0, it does not matter that the


summation in Eq. (5.105) extends to n = -00. Various moments of the
population number, n, are obtained by taking derivatives of F(z, t) with respect
of z and then allowing z ~ 1. For example,
00

8F(z, t)

(5.106)

Converted with

STOI Converter

and

trial version

(5.107)

hDP://www.stdutilitv.com
Higher moments may be obtained in an analogous manner.
We can now obtain a differential equation for the generating function,
F(z, t). Let us multiply Eq. (5.104) by
and sum over n. Note, for example,
that

zn

L t'(n-1)P (n-1,t)=z2_
00

n=-oo

L t'-lP (n-1,t)=z2
00

dF(z, t)
d

n=-oo

and

L
00

t'(n

+ 1)P1(n + 1, t)

L t'+lP (n
00

+ 1, t) =

Z n=-oo

n=-oo

dF(z t)
d'
.
Z

We obtain the following differential equation for F(z, t):

8F

8t

8F
Bz

= (z - t)(f3z - 1)-'

(5.108)

If we substitute Eq. (5.105) into Eq. (5.l08) and equate like powers of z, we

263

SPECIAL TOPICS: MASTER EQUATION FOR BIRTH-DEATH PROCESSES

retrieve Eq. (5.104). Equation (5.108) is a first-order linear partial differential


equation and may be solved using the method of characteristics.
Method of Characteristics [15]. Let us consider a first order linear
differential equation of the form

8F(z, t)
8t
I

+ g ( z ) 8F(z,Bz

t)

+ h( z )F( z,

) =0

(1)

where g(z) and h(z) are arbitrary functions of z. Write F(z, t) in the form

F(z, t)

[ Sz~]

= e-

g(z)

<I>(z,t),

(2)

where <I>(z,t) is an unknown function of z and t. If we substitute Eq. (2) into


Eq. (1), we find a partial differential equation for <I>(z,t):

8<I>(z, t)
8t
We now find
along whic

dt = dz/g(z)
show that an
we write the
takes the fo

( ) 8<I>(z, t) _ 0
8z
-.

(3)

+g z

Converted with

STOI Converter
trial version

the z-t plane


he equation
It is easy to
3). Generally!
ion to Eq. (1)

hDP:llwww.stdutililV.com
F(z, t)

(4)

=e

If we use the method of characteristics to solve Eq. (5.108), we find


(5.109)
Although we know that F(z, t) is a rather complicated function of 13, " and t,
we still do not know the exact form of F(z, t). This we obtain from the initial
conditions. Let us assume that at time t = 0 there are exactly m bacteria in the
population. Thus, P(n, 0) = Dn,m and F(z, 0) = z'". From Eq. (5.109) we find

1))

F(f3(Z f3z - ,
Ifwe now let u
so

= z".

(5.110)

= f3(z - 1)/(f3z - ,), then we can write z = (,u - f3)/f3(u - 1),


,u - f3)m
F(u) = ( f3(u - 1)

(5.111)

264

STOCHASTIC DYNAMICS AND BROWNIAN MOTION

and

1'(Z - l)eCB-,)t
F(z, t) = (
{3(z - 1)eU3-,)t

+ 1')m

(3z

{3z + l'

(5.112)

Equation (5.112) is exact and contains all possible information about the linear
birth-death process. We can now obtain PI (n, t) and all the moments of n from
Eq. (5.112). In order to obtain PI (n, t) we must expand Eq. (5.112) in a Taylor
series in z. PI (n, t) is the coefficient of t'. The first moment is

(n(t)) =

(OF)
OZ

me(f3-,)t.

(5.113)

z=I

The variance is

(n2(t)) - (n(t))2

f!._2:J..

e(f3-,)t(1-

e(f3-,)t)

(5.114)

Converted with

If the birth ra
on will grow
exponentially
birth rate, the
population will
oversimplified
description of
eft out factors
trial version
involving food
good starting
point. For an e
equations to
population dynaifiics, see e.
Birth-death equations can also be applied to chemical reaction kinetics. For
example, in Problem S5.2 we consider the reaction

STOI Converter
hDP://www.stdutilitv.com

where the number of A molecules is kept constant but the number of X


molecules and Y molecules can vary. kI (k2) is the probability per unit time that
a reaction takes place if an X molecule (Y molecule) collides with an A
molecule. The reaction in the forward direction requires nA A molecules, nx + J
X molecules, and ny - 1 Y molecules to produce a system containing nA A
molecules, nx X molecules, and ny Y molecules. Similarly, the backward
reaction requires nA A molecules, nx - 1 X molecules, and ny + 1 Y molecules
to produce a system containing nA A molecules, nx X molecules, and ny Y
molecules. The transition rate for the forward reaction depends on the number
of binary collisions, nA(nx + 1), between A molecules and X molecules.
Similarly, the transition rate for the backward reaction depends on the product,
nA(ny + 1). Thus,
the transition
rate for the forward
reaction
is
WI = kInA(nX + 1), and for the backward
reaction it is W2 = k2nA(ny + 1).

265

SPECIAL TOPICS: MASTER EQUATION FOR BIRTH-DEATH PROCESSES

If we assume that the total number of X molecules and Y molecules


constant and equal to N, then the master equation can be written

aPI (n, t)

at

= k2(N - n

- (kIn

+ I)PI(n -

+ k2(N

1,

- n))P(n,

t) + kI(n + I)P1(n + 1, t)

is

(5.115)

t).

where n is the number of X molecules and N - n is the number of Y molecules.


For simplicity we have absorbed nA into ki and k2. We leave it as a homework
problem to solve this master equation .

.... S5.B.3. Nonlinear Birth-Death

Processes [3, 4, 17]

Some of the simplest examples of nonlinear birth-death processes come from


population dynamics and chemical reaction kinetics. A nonlinear birth-death
process is one for.whi
th of he transition rates b (t) and dn(t),

Converted with

depend non line


processes usual

STOI Converter

However, for bi
because the p
solvable in term
An example
reaction [18]

birth-death
g-time state.
olved exactly
ion might be

trial version

hDP://www.stdutilitv.com
k

2X-+B.
The transition rate for this reaction is proportional both to the reaction rate, k,
and to the number of different ways, n x (n x-I),
to form pairs of X molecules
if there are n X molecules in the system. If the system has nx + 2 X molecules
and ne - 1 B molecules before the reaction, then it will have nx X molecules
and ne B molecules
after
the reaction.
The
transition
rate
is
w = ~(nx + 2)(nx + 1). The master equation can be written

ap(n, t)

at

2" (n + 2)(n + I)P(n + 2, t) - 2"n(n - l)P(n, t).

(5.116)

This equation has a natural boundary at n = 0 since probability cannot flow


into P( -1, t) from above. The equation for the generating function takes the
form
(5.117)

266

STOCHASTIC DYNAMICS AND BROWNIAN MOTION

The generating function, F(z, t), in terms of Gegenbauer polynomials can be


written

LAne-(k/2)n(n-l)tc;1/2(z).
00

F(z, t) =

(5.118)

n=O

The coefficient An def;ends on initial conditions. The first few Gegenbauer


pol~nomials are c~ /2(Z) = I, C~1/2(Z) = -z, C~1/2(Z) = (I - Z2), and
C3"/\z) = !z(1 - Z2). We shall not attempt to compute An here. Solutions
can be found in Ref. 17.
It is important to note that most nonlinear master equations cannot be solved
exactly. It is then often necessary to resort to approximation schemes or
numerical methods for solutions .

.....S5.C. The Fokker-Planck Equation [3, 4, 19]


The Fokker-Pl
probability den

Converted with

STOI Converter

volution of the
er differential

equation and i
the Brownian
particle is Gau
lanck equation
is a two-step pr
trial version
the probability
density p(x, v, t
~ x + dx and
v ~ v + dv at
~(t). We then
obtain an equa
"-,,
{,
p(x, v, t) over
many realizations of the random force ~(t). The probability density P(x, v, t) is
the macroscopically observed probability density for the Brownian particle. Its
dynamical evolution is governed by the Fokker-Planck equation .

hDP:llwww.stdutililV.com

.....S5.C.l. Probability Flow in Phase Space


Let us obtain the probability to find the Brownian particle in the interval
x ~ x + dx and v ~ v + dv at time t. We will consider the space of coordinates,
X = (x, v) (x and v the displacement and velocity, respectively, of the Brownian
particle), where -00 < x < 00 and -00 < v < 00. The Brownian particle is
located in the infinitesimal area, dxdv, with probability p(x, v, t)dxdv. We
may view the probability as a fluid whose density at point (x, v) is given by
p(x, v, t). The speed of the fluid at point (x, v) is given by X = (x, v). Since
the Brownian particle must lie somewhere in this space, we have the
condition
'OO

-00

dx

Joo
-00

dvp(x, v, t) = 1.

(5.119)

SPECIAL TOPICS: THE FOKKER-PLANCK

::]. i:i! !:)J[ :!!! i!i!! i!!

267

EQUATION

JI:;~,!llk:d~Oi:) \

.:::::::::::::::::
:::::::::::::::::Lo:::::::::::::::::::::::
.
. .. ... ... . ...".,,". .. ... .... .. ... ...
."
".""

""

""
"
"""
"
."."
" ""
" .. ,,"

""
"."."
" " ..
" . ".,,""

.. """"
"

",,"

..

""

"

"

""""
"
""""
""
""".""

""."

"
"
"

""""
"""",,
" "
"""""."
" " "."
"
"
" "
"".
"
.. "."""."""".,,",,,,

.. ...

"
""

"

"
"

"
"

".,,

,,.,,"

.. ....
".""".,,

"

".

"

"

""

...:::::::::::::::::::::::::::::::::::::::::::::::::::.X::
......
... ". ...". .
. .
.....

"

"""
",,"
"""

""
".""""
"".. "."."""
"

"

"

..

"

"

"

"

".

"

""."""""",,.,,
""" "".,, "

.......""

""

.""
" ""

"

..

"

"

Fig. 5.3. A finite area in Brownian particle phase space.

Let us now consider a fixed finite area, Ao, in this space (cf. Fig. 5.3). The
probability to find the Brownian particle in this area at time t is P(Ao) =
fA o dxdvp(x, v, t). Since the Brownian particle cannot be destroyed, and change
in the probability contained in Ao must be due to a flow of probability through
the sides of Ao ..............
'T'hll1.....
I~~

---,

Converted with
(5.120)

STOI Converter

where dSo deno


trial version
~f area Ao, pX
is the probability
__
hd the edge of
area element A(
the surface
integral into an area integral, fLo p(x, v, nx dSo = fAo dxdvV'x . (Xp(x, v, t)),
where V'x denotes the gradient, V'x = (ajax, ajfJv). We find

hDP://www.stdutIIIlV.COm

: JJ
t

dxdvp(x, v, t) = -

Ao

dxdvV'x

(Xp(x, v, t)).

(5.121 )

Ao

Since the area Ao, is fixed, we can take the time derivative inside the integral.
Since the area Ao, is arbitrary, we can equate integrands of the two integrals in
Eq. (5.121). Then we find that (ap(t)jat)
= -V'x (Xp(t)), or
ap(t) = -V'x . (Xp(t))
at

= _ a(xp(t))
ax

_ a(vp(t))
fJv

(5.122)

where we have let p(t) = p(x, v, t).


~ S5.C.2. Probability Flow for Brownian Particle
In order to write Eq. (5.122) explicitly for a Brownian particle, we first must
know the Langevin equation governing the evolution of the Brownian particle.

268

STOCHASTIC DYNAMICS AND BROWNIAN MOTION

Let us assume that Brownian particle moves in the presence of a potential, V(x).
The Langevin equations are then

dv(t)

-d

'Y
m

--v(t)

1
m

1
m

+-F(x)

+-~(t)

where the force, F(x) equals - (dV(x)/dx).


into Eq. (5.122), we find

op(t)
ot

= _ o(vp(t)) + 1. o(vp(t))
=

"
La

= vox

Since ~(t) i
different for e
Brownian parti
Therefore, we
Brownian part'
observable pro

'Y
m

'Y 0
m 8v

- - - -v-

La

-d

= v(t),

(5.123)

If we substitute these equations

_ _!_F(x) op(t) _ _!_~(t) op(t)


m
OV
m
ov

ox
m Bv
-Lop(t) - t, (t)p(t),

where the differential operators

dx(t)

and

and

1 are
0

+ -F(x)
m
ov

(5.124)

defined as

and

,,1
0
Ll = -m~(t) l..'
uv

(5.125)

(x, v, t) will be

Converted with

STOI Converter
trial version

erve an actual
om force on it.
dv, to find the
We define this

hDP://www.stdutilitv.com
~.

(5.126)

We now must find the equation of motion of P(x, v, t).


Since the random force, ~(t), has zero mean and is a Gaussian white noise,
the derivation of P(x, v, t) is straightforward and very instructive. It only takes a
bit of algebra. We first introduce a new probability density, u(t), such that
(5.127)

Using Eqs. (5.124), (5.126), and (5.127), it is easy to show that a(t) obeys the
equation of motion:

o~~t) = _ V(t)a(t),
where V(t)

= e+Lot1 (t)e-Lot

(5.128)

Equation (5.128) has the formal solution

17(1) = exp [-

I:

dt'V(t')] 17(0).

(5.129)

SPECIAL TOPICS: THE FOKKER-PLANCK

269

EQUATION

Let us now expand the exponential in Eq. (5.129) in a power series. Using
the identity tr = L::o(x" In!), we obtain

(J>t'V(t') r]17(0).

(5.130)

u(t) = [~( ~~)"

We now can take the average, 0(, of Eq. (5.130). Because the noise, ~(t), has
zero mean and is Gaussian, only even values of n will remain (cf. Exercise 4.9).
Thus, we find
(5.131)

We can use some results from Exercise 4.9 to simplify Eq. (5.131). The
average, (U; dt'V(t'))2n)~, will decompose into (2n!jn!2n) idel}tical terms, each
containing a product of n pairwise averages,
dtjV(tj)
dtjV(tj))~. Thus, Eq.
(5.131) takes

(J;

S;

U~--'-'-LI...LL.L-

Converted with
((1(t))~

STOI Converter

(5.132)

trial version

We can now su

hDP://www.stdutilitv.com
(5.133)
Let us compute the integral in Eq. (5._133),

21 JI dti JI dt, (V(t2)V(tt))~


0

g
=2m2
= _LJI
2m2

I dtz JI dt18(t2
0

- t})e+LoI2 _e-Lo(12-1t)
8v

_e-LoI1
8v

(5.134)

dt e+LOI1 8 e-LoI1
0
1
8v2

If we substitute Eq. (5.134) into Eq. (5.133) and take the derivative of Eq.
(5.133) with respect to time t, we find the following equation of motion for
(a(t)~,
(5.135)
With this result, we can obtain the equation of motion of P(x, v, t) =(p(x, v, t)e.

270

STOCHASTIC DYNAMICS AND BROWNIAN MOTION

Let us note that (p(t))~ = D(t)(a(t))~, where D(t) = e-iot, and take the
derivative of (p(t))~ with respect to time t. We then obtain

a(p(t))~ = -L ( ()) + D( ) a(a(t))e = -L ( ()) + g a(p(t))e


at
0 p t ~
t
at
0 p t ~
2m2 8v2 .
(5.136)
If we combine Eqs. (5.125), (5.126), and (5.136), the equation for the

observable probability density, P(x, v, t), becomes

ap
at =

ap

a [(,

-v ax + 8v

;v -

;F(x)

)]

g ap
P + 2m2 8v2 '

(5.137)

where P = P(x, v, t). Equation (5.134) is the Fokker-Planck equation for the
probability P(x, v, t)dxdv to find the Brownian particle in the interval
x -+ x + dx and v -+ v + dv at time t.
It is important to note that the Fokker-Planck equation conserves probability.
We can write it in the form of a continuity equation

Converted with
where V = i(8
and J is the pre

STOI Converter

(5.138)
~) phase space

trial version

hDP://www.stdutilitv.com

(5.139)

in the (x, v) phase space. By the same arguments used in Eqs. (5.120) and
(5.121), we see that any change in the probability contained in a given area of
the (x, v) phase space must be due to flow of probability through the sides of the
area, and therefore the probability is a conserved quantity. It cannot be created
or destroyed locally.
In this section we have derived the Fokker-Planck equation for a Brownian
particle which is free to move in one spatial dimension. The Fokker-Planck can
be generalized easily to three spatial dimensions. However, when the force F(x)
couples the degrees of freedom, little is known about the details of its
dynamical evolution. Below we consider Brownian motion in the limit of very
large friction. For this case, detailed balance holds, the force can be expressed
in terms of a potential, and we can begin to understand some of the complex
phenomena governing the dynamics of the Fokker-Planck equation.
.....S5.C.3. The Strong Friction Limit
Let us now consider a Brownian particle moving in one dimension in a potential
well, V(x), and assume that the friction coefficient, " is very large so that the

SPECIAL TOPICS: THE FOKKER-PLANCK

271

EQUATION

velocity of the Brownian particle relaxes to its stationary state very rapidly.
Then we can neglect time variations in the velocity and in the Langevin
equations [Eq. (5.123)], we assume that (dv/dt) ~ O. The Langevin equations
then reduce to

dx(t)
1
1
= -F(x) + -~(t),
dt
'"'(
'"'(

(5.140)

where F(x) = -(dV(x)/dx).


We now can use the method of Section S5.C.2 to
find the probability P(x, t)dx to find the Brownian particle in the interval
x ~ x + dx at time t. The probability density, P(x, t), is defined as the average,
P(x, t) = (p(x, t))~, where the equation of motion for the density, p(x, t), is
given by

ap(t) = _ a(xp) = _ 1 a(F(x)p) _ ~~(t) ap


at
ax
'"'( ax
'"'( ax

(5.141 )

= -L

The differentia

Converted with

STOI Converter

Lo

(5.142)

trial version

hDP://www.stdutilitv.com
ap(x, t) = ~~ (dV P(x, t) + g ap(x, t)) = _ aJ ,
at
,",(ax dx
2'"'(
ax
ax

(5.143)

where J = -(I/,",()(dV /dx)P + (g/2'"'(2)(dP/dx) is the probability current.


Equation (5.143) is now a Fokker-Planck equation for the probability density
P(x, t) to find the Brownian particle in the interval x ~ x + dx at time t.
Because Eq. (5.143) has the form of a continuity equation, the probability is
conserved.

~ SS.C.4. Solution of Fokker-Planck Equations with One Variable


For the case of a "free" Brownian particle, one for which V(x) = 0, the
Fokker-Planck equation (5.143) reduces to the diffusion equation

ap(x, t)
at

g a2p(x, t)
2'"'(2

Bx?

= D (}2p(x,
{}x2

t)

(5.144)

272

STOCHASTIC DYNAMICS AND BROWNIAN MOTION

As we showed in Section 4.E, this has a solution

P(x,t) =

~exp(~~}

(5.145)

Note that (I/D) = (2'"'// g) = ('"'(


/k8T). Thus for large friction coefficient, ,",(,the
spatial relaxation is very slow.
For the case when Vex) =1= 0, we can obtain a spectral decomposition of the
probability density P(x, t). Let us first introduce a rescaled time T = t hand
write the Fokker-Planck equation (5.143) as

oP(x, T)
d2V
OT
= dx2 P

dV oP
ox

+ dx

g 02p

"

+ 2'"'( ox2 = -LFPP(x,

T).

(5.146)

The operator, LLP = (d2V /~)


+ (dV /dx)(%x)
+ (g/2'"'() (02 /ox2), is a nonself-adjoint operator because of its dependence on the first-order partial
derivative. However, it is possible to rewrite the Fokker-Planck
equation in
terms of a self-adjoint operator via a simple transformation. Then the solutions
become more i..-o-_.___-----------------,

Converted with

Let us write

where \}I(x, T)
(5.146) we obt

o\}l(x, T)
=
OT

STOI Converter

(5.147)
titute into Eq.

trial version

hDP://www.stdutilitv.com

-(!

The operator, HFP =


(d2V /dx2) - ('"Y/2g)(dV /dx)2) - (g/2'"'() (02 /ox2), is
a self-adjoint operator and we can use the many well established techniques for
dealing with such operators. We will let rPn(x) and An denote the nth eigenvector
and eigenvalue,
respectively,
of HFP so that HFPrPn(x) = AnrPn(X). The
eigenvectors are complete and can be made orthonormal so that

['"

dx</>n,(x)</>n(x) = on',n.

(5.149)

Furthermore, the eigenvalues are real and must have zero or positive values in
order that the probability remains finite. We can expand \}I(x, t) in terms of the
eigenvectors and eigenvalues of HFP:
00

w(x, T)

= Lane-AnT
n=O

rPn(x).

(5.150)

SPECIAL TOPICS: THE FOKKER-PLANCK

273

EQUATION

It is interesting to note that HFP has at least one zero eigenvalue, which we
denote >'0 = 0, and a corresponding eigenvector, cf>o(x), which satisfies the
equation

d2V 2g,(dV)2)
dx
( 2 dx
1

8 cf>0(x)
cf>o(x) + 2, 8x2
g

2 -

(5.151)

o.

Equation (5.150) has the solution


(5.152)
where C is a normalization constant. This is just the transformation used in Eq.
(5.146). Therefore we can now combine Eqs. (5.147), (5.150), and (5.152) and
write the probability as
00

P(x, r)
In this form,
orthonormality

cf>~(x)

+ L ane->."r

cf>O(x)cf>n(x).

Converted with

STOI Converter

(5.153)
due to the
3), we obtain

trial version

(5.154 )

hDP://www.stdutilitv.com
00

P(x, 0) = cf>~(x)

+ L ancf>o(x)

cf>n(x).

(5.155)

n=l

If we now divide through by cf>o(x), multiply by cf>no(x), and integrate over x, we


obtain

ano =

00
-00

dx cf>no(x) P(

cf>o(x)

x,

0)

(5.156)

After a long time, the probability approaches the stationary state:

P(x, 00) = cf>5(x).

(5.157)

There are several examples of Fokker-Planck equations with one variable


Whichcan be solved analytically. We will consider one of them in Exercise (5.8)
and leave the others as homework problems.

274

STOCHASTIC DYNAMICS AND BROWNIAN MOTION

This method can also be extended to Fokker-Planck equations with two or


more spatial degrees of freedom when a transformation analogous to Eq. (5.146)
can be found which allows us to write the Fokker-Planck equation in terms of a
self-adjoint operator. For such cases, it is possible that the dynamics governed
by the self-adjoint operator can undergo a transition to chaos. Examples of such
cases have been studied in Refs. 22-24 .
EXERCISE 5.8. Consider the "short-time" relaxation of a free
Brownian particle. The Langevin equation for the velocity is m(dv/dt) =
+ ~(t). (a) Find the Fokker-Planck equation for the probability
P(v, t)dv to find the Brownian particle with velocity v ~ v + dv at time t.
(b) Solve the Fokker-Planck equation, assuming that at time t = 0 the
velocity is v = yo.

-,v

Answer:
(a) To obtain the Fokker-Planck equation, we will follow the method of

Section
realizat

or a specifie

Converted with

STOI Converter
trial version

From E

(1)
We can now

hDP://www.stdutilitv.com

plug in
ation for the
probabil.-..-.-r_J~~"""_J"""""-"""""""""'---'-'.......----r"""""-----------'
8P
8t

g 82P
2m 8v2

,8(vP)
m 8v

-=---+--.2

(2)

(b) To solve the Fokker-Planck equation, we follow the method of


Section S5.C.3. Make a transformation, P(v, t) = e-m-yv2j2gw(v, t). If
we plug this into Eq. (2), we obtain the following equation for w(v, t),
~:

G - 2 !2}V = Hw.
4~ v +A

(3)

=! -

where A
(g/2m,). The operator il
(1/4A)v2 +A(82/8v2)
is
self-adjoint and has eigenfunctions cf>n(v) (n = 0, 1,2, ... ,00) which
can be expressed in terms of Hermite polynomials [20]. The nth
normalized eigenfunction of iI is
cf>n(V) =

v2nn!v'21rA

Hn(

~)e-y2j4A,

(4)

v 2A

where H; (y) is the nth-order Hermite polynomial and can be written


Hn(Y) = (-lrel(dn/dyn)e-y2
The operator il satisfies the eigen-

SPECIAL TOPICS: THE FOKKER-PLANCK

275

EQUATION

value equation ifc/>n(v) = -nc/>n(v), so the nth eigenvalue is An = -no


The eigenfunctions satisfy the orthonormality condition

(5)

dv</>., (v)</>n(v) = on',.

['"

If we redefine the time to be 'T = (,Im)t, then we obtain the following


spectral decomposition of W(v, t),
00

L ane-nr

w(v, t) =

c/>n(v).

(6)

c/>O(v)c/>n(v).

(7)

n=O

The probability P(v, 'T) is


00

P(v, 'T) =

L ane-nr
n=O

The initial probability distribution is P( v, 0) = 8(v - vo). This gives


an = c/>nVo c/>oVo and we obtain

Converted with

P(v,

STOI Converter

(9)

trial version
We no
1
~exp
v 1 - z-

hDP://www.stdutilitv.com

(-(r

+1-2xyz))
=e
1_ 2
Z

-:x2-l~(~)H
n ( )H n ( )
n,
n=O n.
LJ 2

(10)
(see Ref. 21, page 786). Using this identity, the probability can be
written

(11)
Thus, the probability density has the form of a Gaussian. The average
velocity decays as (v( t)) = voe-r, and the velocity distribution has a
standard deviation, a, = A(1 - eIn the limit of "long time"
the probability density takes the form

2r).

vi

P(v,t)

vi27r,kBT

1m2

exp

(-m2v2)
--

2,kBT

(12)

Thus, for large friction coefficient, " the relaxation time for the

276

STOCHASTIC DYNAMICS AND BROWNIAN MOTION

velocity distribution goes as


spatial distribution goes as , .

,-1

while the relaxation time for the

.....55.0. Approximations to the Master Equation [3, 4]


The master equation, under some conditions, may be reduced to a Fokker-.
Planck equation. For example, if the step size shrinks and the transition rate
grows as a small parameter decreases to zero, then we can use the KramersMoyal expansion to obtain a Fokker-Planck equation from the master equation.
We have already used this fact in Chapter 4 when we obtained the continuum
limit for discrete random walks.
Let us consider a random walk in which the step size is given by Ll. The
master equation can be written
BPI (nLl, t)
Bt

~
= ...
~""

Converted with

where PI (nLl, t
Let us choose t

(5.158)

[PI (mLl, t)wm,n(Ll) - PI (nLl, t)wn,m(Ll)],

nLl at time t.

STDI Converter
trial version

(5.159)

Then the maste1-----..r~h~n~P_:/~/www~-.S-td-u-t-il-itv-.-C-o-m


_

___J

BPI (nLl, t)
1
Bt
= Ll2 [P 1 ( (n

+ 1) Ll, t) + PI ( (n

- 1) Ll, t) - 2P 1 (nLl, t)].

(5.160 )

We now let x = nLl in Eq. (5.159) and let Ll ~ O. To determine what becomes
of the left-hand side in this limit, we expand it in a Taylor series in the small
parameter, Ll keeping x = nLl. Then
BPI (x, t)
.
1
B
= hm A 2 [P 1 (x
t
~--+O u

+ Ll, t) + PI (x

- Ll, t) - 2P 1 (x, t)]


2

. 2 1 [ PI (x, t)
hm
~--+O Ll

+ PI (x, t)
=

a2pI
ax2

+ (BPI)
-

- (BPI)
Bx

Bx

Ll

+ -1 (B--2PI)
2

~=O

Bx

~
~=O

PI)
+ -1 (B--2
2

Bx

Ll 2

+ ...

~=O
~2

+ ... -

2PI (x, t )]

~=O

(5.161)

Thus, in the limit ~ ~ 0 the random walk is described by a simple Fokker-:

277

SPECIAL TOPICS: APPROXIMATIONS TO THE MASTER EQUATION

Planck equation. Higher-order derivatives in the Taylor series do not contribute


because their coefficients go to zero. Lower-order terms in the Taylor series
cancel. We see that the condition to obtain a Fokker-Planck equation is that the
step size, ~, decreases and the transition rate, wm,n(~)' increases as 1/ ~2. This
is a simple example of a Kramers-Moyal-type expansion [25, 26].
It is useful to consider a more general example. Let us write the master
equation for the continuous stochastic variable x:
oPI (x, t)
Joo
ot
= -00 dx'[Pl (x', t)w(x'lx)

(5.162)

- PI (x, t)w(xlx')],

where w(x'lx) is the transition rate. Let y = x' - x denote the change in
stochastic variable x at a transition, and introduce the notation r(x, y) =
w(xlx + y) so that r(x - y,y) = w(x - YIY). In Eq. (5.162), make the change of
variables x' = x + y, and in the first term under the integral let y -+ -Yo Then
the master equation can be rewritten
OPI (x, tJr--___._oo
ot

---,

Converted with

STOI Converter

We can now

(5.163)

in y. This, of
d about y = O.

trial version
aPI (x, t) =

ot

hDP://www.stdutilitv.com

r(x,y)]

-00

oo

-00

00

dy ~

(-yr

an

---;;r- oxn

(5.164)

(PI(X,t)T(X,y)).

Thus,
OPI(x,t)=~(-lr~(
L...J
t
n=I

n.

where an(x) is the nth moment,

","(x) = ['" dyy"r(x,y)

))
x, t ,

(5.165)

dyy" w(xlx + y).

(5.166)

Xn

J:

()P(
an x

Equation (5.165) is the Kramers-Moyal expansion of the master equation. It


only has meaning for those special forms of w(xlx + y) = T(X, y) for which the
infinite series of higher-order derivatives truncates. We shall now give an
example of such a case.

278

STOCHASTIC DYNAMICS AND BROWNIAN MOTION

Let us choose w(xlx

w(xlx

+ y)

r(x, y) to be of the form


1

+ y) = r(x,y) = v:rr~3 exp

((y- - A(X)~2)2)
~2
.

(5.167)

Then

0:1

= ['" dyyw(xlx+y)

=A(x),

(5.168)

and
(5.169)
Higher-order odd moments are identically zero, and higher-order even moments
are proportional to powers of ~. Thus, for the choice of w(xlx + y) given in Eq.
(5.167), as ~ -

Converted with

STOI Converter

(5.170)

For this proces


trial version
the Gaussian
shrinks, and tf
occurs in the
particular mannc.,., __ D_P_:______..,WWW_--rr.S_t_u_tl_l_tv_.C_O_m_--=r- expansion has
meaning.
For the case when the step size in a master equation cannot be made
arbitrarily small, as is the case for chemical reactions or population dynamics,
the Kramers-Moyal expansion of the master equation may not give a good
approximation. Then one must use some other approximation scheme or solve
the equation numerically. An alternative approximation scheme has been
introduced by van Kampen. He has shown that it is still possible to approximate
the master equation by a Fokker-Planck equation if the system has a large
parameter, such as a volume or total particle number, and if the transition rates
depend on that large parameter in a certain way. A full discussion of this
method may be found in van Kampen's book [3].

II

d -1-

REFERENCES
1. N. U. Prabhu, Stochastic Processes (Macmillan, New York, 1965).
2. O. E. Uhlenbeck, Unpublished Lecture Notes, University of Colorado, 1965.
3. N. O. van Kampen, Stochastic Processes in Physics and Chemistry, revised edition
(North-Holland, Amsterdam, 1992).

279

PROBLEMS
4. C. W. Gardiner,
1983).

Handbook of Mathematical Methods (Springer-Verlag,

5. A. Scheerer, Probability on Discrete Sample Spaces (International


Scranton, PA, 1969).
6. S. Lipschutz,

Probability, Schaums' Outline Series (McGraw-Hill,

Berlin,

Textbook

Co.,

New York, 1965).

7. B. Friedman, Principles and Techniques of Applied Mathematics (John Wiley &


Sons, New York, 1956).
8. H. Haken, Synergetics, 3rd edition (Springer-Verlag,

Berlin,

1983).

9. G. H. Weiss, "First Passage Time Problems in Chemical Physics,"


XIII, 1 (1967).

Adv. Chem. Phys.

10. R. Brown, Philosophical Magazine N. S., 4, 161 (1828).


11. A. Einstein, Ann. der Physik, 17, 549 (1905).
12. A. Einstein, Investigations on the Theory of Brownian Movement (Methuen and Co.
Ltd., London, 1926). This book contains a history of early work on Brownian motion.
13. M. J. Perrin, Brownian Movement and Molecular Reality (Taylor
London, 1910).

and Francis,

14. M. Dykman and K. Lindenberg in Contemporary Problems in Statistical Physics,


edited by G'
..

Converted with

15. I. N. Snedd
1957).

-Hill, New York,

STOI Converter

16. N. S. Goel
York, 1974)
17. G. Nicolis
Wiley & So
18 . D. A. McQ '---

emic Press, New

trial version

Systems (John

hDP://www.stdutilitv.com

Phys. 40, 2914

(1964).
19. H. Risken, The Fokker-Planck Equation (Springer-Verlag,

Berlin,

20. G. Arfken, Mathematical Methodsfor Physicists (Academic

Press, New York, 1985).

21. P. Morse and H. Feshbach,


York, 1953).
22. M. Milonas

1984).

Methods of Theoretical Physics (McGraw-Hill,

New

and L. E. Reichl, Phys. Rev. Lett. 68, 3125 (1992).

23. P. Alpatov and L. E. Reichl, Phys. Rev. E49, 2630 (1994).


24. Sukkeun

Kim and L. E. Reichl, Phys. Rev. E53, 3088 (1996).

Physica 7, 284 (1940).


26. J. E. Moyal, J. R. Stat. Soc. Bll, 150 (1949).
25. H. A. Kramers,

PROBLEMS
Problem 5.1. Urn A initially has one white and one red marble, and urn B initially has
one white and three red marbles. The marbles are repeatedly interchanged. In each step
of the process one marble is selected from each urn at random and the two marbles
selected are interchanged. Let the stochastic variable Y denote "configuration
of the
urns." Three configurations are possible: (1) Urn A-2 white balls, Urn B-4 red balls;

280

STOCHASTIC DYNAMICS AND BROWNIAN MOTION

(2) Urn A-one white and one red ball, Urn B-one white and three red balls; (3) Urn
A-two red balls, Urn B-two white and two red balls. We shall denote these three
realizations as y(l), y(2), and y(3), respectively. (a) Compute the transition matrix, Q,
and the conditional probability matrix, P(sols). (b) Compute the probability vector,
(P(s)l, at time s, given the initial condition stated above. What is the probability that
there are 2 red marbles in urn A after 2 steps? After many steps? (c) Assume that the
realization, y(n), equals n2 Compute the first moment, (y(s)), and the autocorrelation
function, (y(O)y(s)), for the same initial conditions as in part (b).
Problem 5.2. Three boys, A, B, and C, stand in a circle and play catch (B stands to the
right of A). Before throwing the ball, each boy flips a coin to decide whether to throw to
the boy on his right or left. If "heads" comes up, the boy throws to his right. If "trials"
comes up, he throws to his left. The coin of boy A is "fair" (50% heads and 50% tails),
the coin of boy B has heads on both side, and the coin of boy C is weighted (75% heads
and 25% tails). (a) Compute the transition matrix, its eigenvalues, and its left and right
eigenvectors. (b) If the ball is thrown at regular intervals, approximately what fraction of
time does each boy have the ball (assuming they throw the ball many times)? (c) If boy
A has the ball to begin with, what is the chance he will have it after two throws? What is
the chance he will have it after s throws?
Problem 5.3. A

doors of the roor


room." There are
"mouse in room
the transition rna

1,

...J

regular intervals
changes rooms. ,

~;::::::te~;C!

.L

.L

.L

,"".

Converted with

STOI Converter
trial version

hnp:llwww.stdU~ililV.com

I:

A bell rings at

rings, the mouse


rough any of the
e in a particular
in room B," and
~ly. (a) Compute
(b) Compute the

~~)~ss~~~

th:~
the realization, y(n), equals n. Compute the first moment, (y(s)), and the autocorrelation
function, (y(O)y(s)), for the same initial conditions as in part (b).
Problem 5.4. The doors of the mouse's house in Fig. 5.4 are fixed so that they
periodically get larger and smaller. This causes the mouse's transition probability
between rooms to become time periodic. Let the stochastic variable Y have the same
meaning as in Problem 5.3. The transition matrix is now given by
Qt,t (s) = Q2,2(S) = Q3,3(S) = 0,
Qt,2(S) = cos2(7rs/2), QI,3(S) = sin 2 (Jrs/2),
Q2,t (s) =
+ ~sin 2(7rs/2) , Q2,3(S) = + ~cos2(7rs/2), Q3,t (s) = ~cos2(7rs/2), and
Q3,2(S) = ~+ ~sin 2(7rs/2). (a) If initially the mouse is in room A, what is the

c
Fig. 5.4. Mouse's house.

281

PROBLEMS

probability to find it in room A after 2s room changes? In room B? (b) If initially the
mouse is in room B, what is the probability to find it in room A after 2s room changes?
In room B?
problem 5.5. Consider a discrete random walk on a one-dimensional periodic lattice
with 2N + 1 lattice sites (label the sites from -N to N). Assume that the walker is
equally likely to move one lattice site to the left or right at each step. Treat this problem
as a Markov chain. (a) Compute the transition matrix, Q, and the conditional probability
matrix, P (sols). (b) Compute the probability PI (n, s) at time s, given the walker starts at
site n = O. (c) If the lattice has five lattice sites (N = 2), compute the probability to find
the walker on each site after s = 2 steps and after s = 00 steps. Assume that the walker
starts at site n = O.
Problem 5.6. At time t, a radioactive sample contains n identical undecayed nuclei,
each with a probability per unit time, A, of decaying. The probability of a decay during
the time t --+ t + t:1t is Ant:1t. Assume that at time t = 0 there are no undecayed nuclei
present. (a) Write down and solve the master equation for this process [find PI (n, t)]. (b)
Compute the mean number of undecayed nuclei and the variance as a function of time.
(c) What is the half-life of this decay process?

trial version

Problem 5.8. C
absorbs the wal

5.6. The site, P,

hDP://www.stdutilitv.com

2,3

= W2,P

= ~,

matrix, M, and
compute its eigenvalues and and left and right eigenvectors. (b) If the walker starts at
site n = 1 at time t = 0, compute the mean first passage time.
W3,I

W3,2

W,

3'

Problem 5.9. Let us consider on RL electric circuit with resistance, R, and inductance,
L, connected in series. Even though no average electromotive force (EMF) exists across
the resistor, because of the discrete character of the electrons in the circuit and their
random motion, a fluctuating EMF, ~(t), exists whose strength is determined by the
temperature, T. This, in tum, induces a fluctuating current, l(t), in the circuit. The
Langevin equation of motion for the current is

dl(t)
Ld(

2~'
4

+ RI(t) = ~(t),

Fig. 5.5.

27'
p

Fig. 5.6.

282

STOCHASTIC DYNAMICS AND BROWNIAN MOTION

If the EMF is delta-correlated, (e(t2)e(tt)){ = g8(t2 - tt) and (e(t)){ = 0, compute, g,


and the current correlation function, ((/(t2)/(tt)){h.
Assume that the average magnetic
energy in the inductor is !L(/5h = !kBT and (/O)T = 0, where 1(0) = 10.
Problem 5.10. Due to the random motion and discrete nature of electrons, and LRC
series circuit experiences a random electromotive from (EMF), e(t). This, in tum,
induces a random varying charge, Q(t), on the capacitor plates and a random current,
I(t) = (dQ(t)/dt), through the resistor and inductor. The random charge, Q(t), satisfies
the Langevin equation

L d~~2(t)

+ Rd~;t) + Qg) = e(t).

Assume that the EMF is delta-correlated, (e(t2)e(tt)){ = g8(t2 - tt), and (e(t)){ = O.
Assume that the circuit is at temperature T and that the average magnetic energy in the
inductor and average electric energy in the capacitor satisfy the equipartition theorem,
!L(/5h = !kBT and
(Q6)t = !kBT, where Q(O) = Qo and 1(0) = 10. Assume that
(Qo)r = {/oh = (Qolo)r = 0. (a) Compute the current correlation function,
((/(t2)/(tt)){h.
(b) Compute
the variance of the charge distribution,

2b

(((Q(t) - QO)2){)T.
Problem 5.11. C
presence of a COl
force constant ,

(e(t2)e(tt)){

= gl

of the Brownian
velocity correlati

Converted with

STOI Converter
trial version

Ilimension in the
) in a fluid with
e e(t) such that
hd displacement
~) Compute the

x(t) - xo)2)e.

hDP:llwww.stdutililV.com

~:::;:~
s:a:~l~
c~::i~~!t~o~~:
birth and death of individuals in the population. Assume that at time t the population
has n members. Let at:1t be the probability that one individual enters the society due
to immigration in time t --t t + t:1t. Let /3nt:1t(,nt:1t) be the probability of a birth (death)
in time t --t t + t:1t. Assume that at time t = the population has n = m members.
(a) Write the master equation and compute the generating function, F(z, t) =
E:-oo Z'P1 (n, t), for this process. (b) Compute the first moment (n(t)), and the
variance, (n2(t)) - (n(t))2, for this process.

Problem SS.2. Consider the chemical reaction


kl

A+X~A+Y,
k2

where molecule A is a catalyst whose concentration is maintained constant. Assume that


the total number of molecules X and Y is constant and equal to N. kl (k2) is the
probability per unit time that molecule X (Y) interacts with a molecule A to produce a
molecule Y (X). (a) Find the equation of motion for the probability PI (n, t) to find n X
molecules in the system at time t, and find the generating function, F(z, t) :=;;
E:-oo z"P1 (n, t). (b) Find the average number of X molecules in the system at time
t, and find the variance in the distribution of X molecules. In the limit, N -t ()(J,
how does the variance for the particle distribution compare with that of a Gaussian
distribution?

283

PROBLEMS

. .. '
.o..
. n-'.:
..... . .'.. :. .
: o~~.' " ' p . '
"

......

"

'.~.,.... :.".

. on :'. . :.'.'

'.

:>. :;_.'
A

Fig. 5.7. Diffusion through a small hole.

Problem S5,3. Consider a box (A) of volume n, connected to another box (B) of infinite
volume via a small hole (cf. Fig. 5.7). Assume that the probability that a particle moves
from box A to box B in time ~t is (n/n)~t
and that the probability that a particle
moves from box
sumed constant)
of particles in b
Converted with
bability PI (n, t)
to find n partie
F(z,t)=E:_o
(b) Compute PI
number of partie
time.

ST 0 U C ODVerIer
trial ve rsion

rating function,
Aattimet=O.
ind the average
as a function of

http://www.stdutiliJV.com

Problem S5.4. (
t N = 3, kI = 2,
and k2 = 1. (a) Write the transition matrix, Wand compute its eigenvalues and left and
right eigenvectors. (b) If initially there are zero X-molecules in the system, what is the
probability of finding three X-molecules at time t. What is the probability of finding
three X-molecules at time t = oo?
Problem S5,5. Consider the following chemical reaction,
k}

A + M--+X

ka

+ M and 2X--+E + D,

where molecules A, M, E, and D are obtained from large reservoirs and can be assumed
constant. (a) Find the probability to have n X molecules in the system after a very long
time, t -. 00. (b) Find the average number of X molecules after a very long time. [Some
hints: Use the boundary conditions F(I) = 1 and F(-I) = 0 for the generating
function, F(z). The generating function can be found in terms of modified Bessel
functions. The transformation F(z) = v'SG(s), where s = (1 + z)/2, might be helpful.]
Problem S5.6. The motion of an harmonically bound Brownian particle moving in one
dimension is governed by the Langevin equations,
m dv(t)
~

-,),v(t)

- mw~x(t)

+ ~(t)

and

dxd(t) = v(t),
t

where v(t) and x(t) are the velocity and displacement of the particle at time t, m is the

284

STOCHASTIC DYNAMICS AND BROWNIAN MOTION

mass, , is the friction coefficient, Wo is the natural frequency of the harmonic oscillator,
and ~(t) is a delta-correlated random force. If the particle at time t = 0 is in equilibrium
with the fluid, compute the variance, (((x(t) - xo)2)~h. [Note that for this case,
(~(t2)~(tt))~ = 4,kBT8(t2 - t1), and by the equipartition theorem, (~h = kBT /mw6
and (v~h = kBT [m. Also assume (voh = (xoh = (xovoh = 0.]
Problem S5.7. Consider a Brownian rotor with moment of inertia, I, constrained to
rotate through angle, (), about the z axis. The Langevin equations of motion for the rotor
are I(dw/dt) = -rw + ~(t) and (d()/dt) = w, where w is the angular velocity of the
rotor, r is the friction coefficient, and ~(t) is a Gaussian white noise torque. The torque
is delta-correlated, (~(t')~(t))~ = G8(t' - t), where G is the noise strength. (a) For the
case of large friction coefficient, I', write the Fokker-Planck equation for the probability
density, P((), t), to find the Brownian rotor in the interval () -+ ()+ d() at time, t. (b) Solve
the Fokker-Planck equation assuming that at time t = 0 the rotor is at () = ()o. (c)
Compute the probability current at time t.
Problem S5.8. A Brownian particle of mass m moves in one dimension in the presence
of a harmonic potential V(x) = 4k_x2, where k is the force constant. The Langevin
equations are given by m[dv(t)/dt] = -,v(t) - dV(x)/dx + ~(t) and dx(t)/dt = v(t),
where, is the friction coefficient and ~(t) is a Gaussian white noise force. The noise is
delta-correlated,
.
.
. (a) Write the
Fokker-Planck e
of large friction
coefficient. (b) S
ion, P(x, t), for
arbitrary times.
an approximate
Problem S5.9. C
transition rate

w(n.6ln'~)

STOI Converter

= Xo (c) Write

trial version

tep size .6 and

hDP://www.stdutilitv.com
=

8n',n-l,

where -00 ~ n ~ 00 and -00 ~ n' ~ 00. Use the Kramers-Moyal expansion to obtain a
Fokker-Planck equation for this random walk in the limit ~ -+ 0 with n~ = x.

6
THE FOUNDATIONS OF
Sl~TISTICAL MECHANICS

6.A. INTRODUCTION
In Chapter 5 we studied the time evolution of probability distributions in the
Markov approximation, where the dynamics of the process was determined in
terms of a si
ability itself is
usually dete
Converted with
we derived, the
master equati
considered as

STDO Converter

phenomenolo
y the type of
behavior we
of systems to a
unique equili
trial version
achinery for a
microscopic p
al and quantum
mechanical s
for statistical
mechanics an we WI earn ow ermo ynamics an Irreversible processes
are thought to arise from the reversible laws of dynamics.
We want to describe the behavior of systems with a large number of degrees
of freedom, such as N interacting particles in a box or N interacting objects on a
lattice. The motion of such objects is governed by Newton's laws or,
equivalently, by Hamiltonian dynamics. In three dimensions, such a system has
3N degrees of freedom (if we neglect internal degrees of freedom) and
classically is specified by 6N independent position and momentum coordinates
whose motion is uniquely determined from Hamiltonian dynamics. If we set up
a 6N -dimensional phase space, whose 6N coordinates consist of the 3N
momentum and 3N-position variables of the particles, then the state of the
system is given by a single point in the phase space, which moves according to
Hamiltonian dynamics as the state of the system changes. If we are given a real
N-particle system, we never know exactly what its state is. We only know with a
certain probability that it is one of the points in the phase space. Thus, the state
point can be regarded as a stochastic variable and we can assign a probability
distribution to the points in phase space in accordance with Our knowledge of
the state of the system. We then can view the phase space as a probability fluid
which flows according to Hamiltonian dynamics. In this way, we obtain a
connection between the mechanical description and a probabilistic description

http://www.stdutilitv.com

285

286

THE FOUNDATIONS OF STATISTICAL MECHANICS

of the system. The problem of finding an equation of motion for the probability
density reduces to a problem in fluid dynamics. In this chapter, we shall also lay
the foundations for equilibrium statistical mechanics. For classical dynamical
systems which are ergodic, an equilibrium probability distribution can be
constructed which gives excellent agreement with experiment.
When we deal with quantum systems the phase space variables no longer
commute and it is often useful to use representations other than the coordinate
representation to describe the state of the system. Thus, we introduce the idea of
a probability density operator (a positive definite Hermitian operator), which
can be used to find the probability distribution in any desired representation. We
then can use the Schrodinger equation to find the equation of motion for the
probability density operator.
The N-body probability density for a classical system contains more
information about the system than we need. In practice, the main use of the
probability density is to find expectation values or correlation functions for
various observables, since those are what we measure experimentally and what
we deal with in thermodynamics. The observables we deal with in physics are
generally one- or two-bod 0 erators and to find their ex ctation values we
only need redu
ot the full Nbody probabilit
Converted with
find that the

STOI Converter

equations of m
hierarchy of
equations calle
ubov, Green,
Kirkwood, and
trial version
impossible to
solve without so
. This, in fact,
is a general fea
In quantum s
,
hich specify
both the position and momentum of the particle, because these quantities do not
commute. However, we can introduce quantities which are formally analogous,
namely, the Wigner functions. The Wigner functions are not probability
densities, because they can become negative. However, they can be used to
obtain expectation values in a manner formally analogous to that of classical
systems, and the reduced Wigner functions form a hierarchy which in the
classical limit reduces to the classical BBGKY hierarchy.
Finally, in the last sections of the special topics section, we shall describe
conditions under which systems which are governed by the reversible laws of
Newtonian dynamics display irreversible behavior.

hDP:llwww.stdutililV.com

6.B. THE CLASSICAL PROBABILITY DENSITY [1-3]


If we view the flow of points in the phase space of a classical N-body
Hamiltonian system as a fluid whose dynamics is governed by Hamilton's
equations, then we can derive the equation of motion for the probability density
of the classical system in phase space. This equation of motion is called the
Liouville equation. Because Hamilton's equations preserve volume in phase

THE CLASSICAL PROBABILITY DENSITY

287

space, the probability fluid flow is incompressible. It is also nondissipative,


unlike the probability flow governed by the Fokker-Planck equation.
Let us consider a closed classical system with 3N degrees of freedom (for
example, N particles in a three-dimensional box). The state of such a system is
completely specified in terms of a set of 6N independent real variables (pH, qN)
(pN and qN denote the set of vectors pN = (PI' P2' ... , PN) and qN =
(qI' q2, ... ,~),
respectively; PI and ql are the momentum and position of
the lth particle. If the state vector XN = XN (pV ,qN) is known at one time, then
it is completely determined for any other time from Newton's laws.
If we know the Hamiltonian, H (XN , t), for the system, then the time
evolution of the quantities PI and ql (1 = 1, ... ,N) is given by Hamilton's
equations,
(6.1)
and

Converted with
If the Hamilt(
constant of th

STOI Converter

(6.2)
en it is a global

trial version

hDP://www.stdutilitv.com

(6.3)

where the constant, E, is the total energy of the system. In this case the system
is called conservative.
Let us now associate to the system a 6N-dimensional phase space, r. The
state vector XN (pN, qN) then specifies a point in the phase space. As the system
evolves in time and its state changes, the system point XN traces out a trajectory
in r -space (cf. Fig. 6.1). Since the subsequent motion of a classical system is
uniquely determined from the initial conditions, it follows that no two
trajectories in phase space can cross. If they could, one could not uniquely
determine the subsequent motion of the trajectory.
When we deal with real physical systems, we can never specify exactly the
state of the system. There will always be some uncertainty in the initial
conditions. Therefore, it is useful to consider XN as a stochastic variable and to
introduce a probability density p(XN, t) on the phase space, where p(XN, t )dXN
is the probability that the state point, XN, lies in the volume element
X" ~ X" + dXN at time t. (Here dXN = dql X .. X d~dpl
X ... x dPN.) In
so doing we introduce a picture of phase space filled with a continuum (or fluid)
of state points. If the fluid were composed of discrete points, then each point
would be assigned a probability in accordance with our initial knowledge of the
system and would carry this probability for all time (probability is conserved).

288

THE FOUNDATIONS OF STATISTICAL MECHANICS

Pm
Fig. 6.1. Movement of a system point, XN(t), in a 6N-dimensional phase space
(t2 > t}). We show only four of the 6N coodinate directions.

Converted with
The change in 0
by the way in w

STOI Converter

is determined
e points form
on the phase

hDP:llwww.stdutililV.com

pace, we have

a continuum, w
trial version
space.
Because state
the normalizatioou~u:rn~rurr-------------

(6.4)
where the integration is taken over the entire phase space. If we want the
probability of finding the state point in a small finite region R of I' space at time
t, then we simply integrate the probability density over that region. If we let
P(R) denote the probability of finding the system in region R, then
(6.5)

If at some time there is only a small uncertainty in the state of the system, the
probability density will be sharply peaked in the region where the state is
known to be located, and zero elsewhere. As time passes, the probability density
may remain sharply peaked (although the peaked region can move through
phase space) and we do not lose any knowledge about the state of the system.
On the other hand, it might spread and become rather uniformly distributed, in
which case all knowledge of the state of the system becomes lost.

289

THE CLASSICAL PROBABILITY DENSITY

Probability behaves like a fluid in phase space. We can therefore use


arguments from fluid mechanics to obtain the equation of motion for the
probability density (cf. Section SIO.A). We will let iN = (qN, pH) denote the
velocity of a state point, and we will consider a small volume element, Vo, at a
fixed point in phase space. Since probability is conserved, the total decrease in
the amount of probability in Vo per unit time is entirely due to the flow of
probability through the surface of Vo. Thus,

where So denotes the surface of volume element Vo, and dSN is a differential
area element normal to So. If we use Gauss's theorem and change the surface
integral to a volume integral, we find

where
'VXN

'VXN

((8j8(j

the derivative
arguments of t

Converted with

STOI Converter

pace variables
). We can take
. If we equate

trial version

hDP://www.stdutilitv.com

(6.8)

Equation (6.8) is the balance equation for the probability density in the 6Ndimensional phase space.
We can use Hamilton's equations to show that the probability behaves like an
incompressible fluid. A volume element in phase space changes in time
according to the equation

(6.9)
where fN (t, to), the Jacobian of the transformation, is the determinant of a
6N x 6N -dimensional matrix which we write symbolically as
8p~

f N() t, to = det ~aqr

(6.10)

The Jacobian fN (t, to) can be shown to satisfy the relation


(6.11)

290

THE FOUNDATIONS OF STATISTICAL MECHANICS

if we remember that the product of the determinant of two matrices is equal to


the determinant of the products.
Let us now assume that the system evolves for a short time interval
~t = t - to. Then the coordinates of a state point can be written
(6.12)
and
(6.13)
[Again Eqs. (6.12) and (6.13) have been written symbolically to denote the set
of 6N equations for the components of XN.] If we combine Eqs. (6.10), (6.12),
and (6.13), we find
~~t

1+~~t

IN (t, to)

8q~

8p~

= det

8'N

8qN

=1+

Converted with

STOI Converter

(cf. Appendix
we obtain

(6.14)
6.1) and (6.2)]

trial version

hDP://www.stdutilitv.com

(6.15)

and, therefore,
(6.16)
From Eq. (6.11) we can write

and the time derivative of the Jacobian becomes

dJN =
dt

lim
6.HO

IN (to + ~t,

0) -

IN (to, 0) = o.

(6.18)

~t

Thus, for a system whose dynamics is determined by Hamilton's equations, the


Jacobian does not change in time and
(6.19)

291

THE CLASSICAL PROBABILITY DENSITY

Equation (6.19) is extremely important for several reasons. First, it tells us that
volume elements in phase space do not change in size during the flow (although
they can change in shape):

dxf
Second, it tells us that the probability

(6.20)

=dX~.
behaves like an incompressible

fluid since
(6.21)

[cf. Eq. (6.15)]. If we combine Eqs. (6.8) and (6.21) the equation of motion for
the probability density takes the form
8p(XN,t)
at

_ -X' N.
-

r7

VXNP

(XN)
,t .

(6.22)

Note that Eq. (6.22) gives the time rate of change of p(XN, t) at a fixed point in
phase space. If we want the time rate of change as seen by an observer moving
with the prob'
.
the total time

Converted with

derivative of p

s defined as

STOI Converter

(6.23)

trial version
[cf. Appendix

hDP://www.stdutilitv.com

obtain

(6.24)
Thus, the probability density remains constant in the neighborhood of a point
moving with the probability fluid.
If we use Hamilton's equations, we can write Eq. (6.22) in the form
ap(XN,t)=_irN
at
where the differential

operator,

iIN,

(XN

)
,t ,

(6.25)

is just the Poisson bracket

(6.26)

(we put a hat on irN to indicate that it is differential


is often written in the form

i 8p(XN, t)
at

= LNp(XN

operator). Equation (6.25)

t)
'

(6.27)

292

THE FOUNDATIONS OF STATISTICAL MECHANICS

where iN = -iifN.
Equation (6.27) is called the Liouville equation and the
differential operator, L N, is called the Liouville operator. The Liouville operator
is a Hermitian differential operator.
If we know the probability density, p(XN,O), at time, t = 0, then we may
solve Eq. (6.27) to find the probability density, p(XN, t), at time t. The formal
solution is
(6.28)

A probability density, Ps(XN),


condition

which remains constant in time must satisfy the

(6.29)

and is called a stationary solution of the Liouville equation. The Liouville


equation is particularly simple to solve explicitly if the mechanical system is
integrable and one can make a canonical transformation from phase variables,
(PI' ,PN,ql'
..
1, ,ON) [4,
5]. We show h
Converted with
The Liouvill
he probability
density of a cl
an important
respect from th
hapter 5. The
Liouville opera
trial version
erator is not.
Thus, the soluti
ot decay to a
unique equilibri
q. (6.28), we
do not change the equation of motion for the probability density since the
Liouville operator changes sign under time reversal. This is different from the
Fokker-Planck equation, which changes into a different equation under time
reversal. Thus, Eq. (6.27) does not admit an irreversible decay of the system to a
unique equilibrium state and thus cannot describe the decay to equilibrium that
we observe so commonly in nature. And yet, if we believe that the dynamics of
systems is governed by Newton's laws, it is all we have. The problem of
obtaining irreversible decay from the Liouville equation is one of the central
problems of statistical physics and is one that we shall say more about in
subsequent sections.

STOI Converter
hDP:llwww.stdutililV.com

EXERCISE 6.1. Consider a particle which bounces elastically and


vertically off the floor under the influence of gravity (assume no friction
acts). At time t = 0, the particle is located at z = 0 and has upward
momentum, p = Po. It rises to a maximum height, z = h. Solve the Liouville
equation to find the probability density, p(p, z, t), at time t.
Answer:
(p2/2m)

The

Hamiltonian for the particle can be written H =


where V(z) = mgz for z ~ 0 and V(z) = 00 for

+ V(z) = E,

293

THE CLASSICAL PROBABILITY DENSITY

I
I

z < O. The turning point (the point where the momentum goes to zero) of the
orbit is at z = h, the energy is E = mgh, and the initial momentum is
Po = mV2iJi. Hamilton's
equations are jJ = - (oH / oz) = -mg and
Z = (p/m). Hamiltonian's equations may be solved to give

- ~ t2

z(t) = y'fiht

p(t) = my'fih

and

- mgt

for

O~t~T,
(1)

where T = 2V(2h/g)
is the period of the motion. Both z(t) and p(t) are
periodic functions of time with period T.
We can also describe this system in terms of action-angle variables. The
action is
J

= _!_ipdz
27T J

v'2mJh
7T

dZVE - mgz = _2_ f%_ E3/2

37rg

(2)

V;;;

so the energy, as a function of action, is

(3)

Converted with
From Hamilt
constant of

STOI Converter
trial version

iJ

e action is a
equation

hDP://www.stdutilitv.com

(4)

i We see again that the period of the motion is T = (27T/W) = 2V(2h/g). If


: the angle variable at time t = 0 is 8 = 0, then at time t it is 8 = wt. We can
now use Eqs. (1), (3), and (4) and the fact that t = 8/w to write the canonical
transformation from phase space variables (p, z) to phase space variables
(J, 8). We find
z(

J, 8) = _1_ (37TgJ) 2/3 (7T8 _ ~ (2)


g7T2

and

(3 7TJ)
g

p(J, 8) = -; -;;-

I Because the transformation


! transformation

(5)
1/3

(7T - 8)

(5) is canonical,

J = det

!!I!.8J !!I!.)
oe
( 1&
1&
8J
8e

'

for

0 ~ 8 ~ 27T.

the Jacobian

of the

(6)

I is equal to one as can be easily shown. Thus, dpd; = dJd8 and

294

THE FOUNDATIONS OF STATISTICAL MECHANICS

8(p - Po)8(z - zo) = 8(J - Jo)8(e - eO), where Po = p(Jo, eo) and Zo =
z(Jo, eo). We can write the Liouville equation in terms of the canonical
variables (p, z),
8p .Bp
8t + p ap

.Bp

+ z 8z

(7)

= 0,

or in terms of canonical variables (J, e),

ap'

.ap'

7ft + e

8e

(8)

=0

where pep, z, t) = p' (J, e, t) are the probability densities. The initial conditions are given by pep, ~
= 8(p - Po)8(z) and p'(J, e, 0) = 8(J - Jo)8(e),
where Jo = (2m/37T}J2gh3 Because the probability density, p'(J, e, t), is a
periodic function of we can expand it in a Fourier series

e,

p'(J, e, t) = 2~

f: Pn(J, t)einO.

(9)

n=-oo

If we plug E

~:-oo
e ==
Fourier coeffi

the fact that


find that the

Converted with

inO

STOI Converter

(to)

trial version

hDP://www.stdutilitv.com

where iJ = w(

p' (J, e, t)

2~

f: Pn(J, O)ein(O-w(J)t).

(11 )

n=-oo

If we now make use of the initial conditions, we obtain

p'(J, e, t) = 8(J - Jo)8(e - w(J)t)

(12)

and

p(p,z, t) = 8(p - p(Jo,wt))8(z

- z(Jo,wt)),

(13)

where p(Jo, wt) and z(Jo, wt) are given by Eq. (5).
We shall often be interested in obtaining the expectation value of phase
functions, oN (XN). The expectation value of oN (XN) at time t is given by
(O(t))

=
=

I
I

dX1
dXt

dXNoN(XN)p(XN,t)
(6.30)

dXNoN (XN)e-iiftp(XN

, 0).

295

THE CLASSICAL PROBABILITY DENSITY

We have written the expectation value in the "Schrodinger" picture. That is, we
have allowed the state function, p(XN, r), to evolve in time and we have kept the
phase function, oN (XN), fixed in time. We could equally well allow the phase
function to vary in time and keep the state function fixed. We note that the
Liouville operator, LN, contains derivatives with respect to pN and qN. If we
expand the exponential in a power series and integrate by parts, there will be a
change of sign for each partial derivative and we find

J
J

(O(t)) = dXt
= dXt

J
J

dXNd'(XN,t)p(XN,O)
(6.31 )
dXNP(XN,O)e+LN'd'(XN,O).

(we assumed that p(XN, 0) ----.0 for large XN). Thus, we obtain the classical
version of the "Heisenberg" picture. We see that phase functions and state
functions evolve according to different laws. The equation of motion of a phase
function is given by

Converted with
where Eq. (6.-=
space.

STOI Converter

(6.32)
point in phase

trial version

hDP://www.stdutilitv.com

EXERC
Hamiltonian
H = L:~1 P; 12m + L:~J.r-l)72 V(lqj - qjl). The phase function which gives
the particle density at position
R in configuration
space is
i n(qN,R) = L:~18(qj - R). Write the equation of motion of n(qN,R).
!
i

Answer: The equation of motion of n(qN, R) is given by Eq. (6.32). Since


; n(qN,R) does not depend on momentum, Eq. (6.32) reduces to
!
!

an
at

-=

N.

Lqi
;=1

8
-8(qi
8qj

-R).

(1)

: If we now replace the differentiation with respect to qi by a differentiation

, with respect to R, we obtain

(2)
(since

qi = (pJm))

or
(3)

296

THE FOUNDATIONS OF STATISTICAL MECHANICS

where J(~,qN;R) = L~1(p;/m)8(qi


- R) is the particle current phase
function. Equation (3) is a balance equation which reflects the conservation
of particle number on the microscopic level.

It is interesting to note that the probability density, p(XN, t), is often


interpreted in terms of an "ensemble" of systems. This was the view originally
taken by W. Gibbs. Let us consider an ensemble of"l identical systems ("l very
large). If we look at each system at a given time, it will be represented by a
point in the 6N-dimensional phase space. The distribution of points representing
our ensemble of systems will be proportional to p(XN, t). That is, the density of
system points in phase space will be given by "lP(XN, t).

6.C. ERGODIC THEORY AND THE FOUNDATIONS OF


STATISTICAL MECHANICS [6-13]
Thesu~ectof~nd~~~~~mrnrurr~nY~wrunn~~~
the advent of
Converted with
even more im
nee in such
diverse fields
system) and
chemistry (stab
asks questions
which lie at th
trial version
As we shall
a very special
type. There ar
storically, two
types of probabi I y ow ave een Important m un erstan mg e behaviour of
phase space, namely, ergodic flow and mixing flow. For systems with ergodic
flow, we obtain a unique stationary probability density (a constant on the energy
surface) which characterizes systems with a fixed energy at equilibrium.
However, a system with ergodic flow cannot necessarily reach this equilibrium
state if it does not start out there. For decay to equilibrium, we must have at
least the additional property of mixing. Mixing systems are ergodic (the
converse is not always true, however) and can exhibit random behavior. In
addition, reduced distribution functions can be defined which decay to an
equilibrium state. We give examples of mixing flow in the special topics
Section (S6.D).
Ergodic and mixing behavior for real systems is difficult to establish in
general. It has been done only for a few model systems. However, there is a
large class of conservative systems, the anharmonic oscillators, which are of great
importance in mechanics, chemistry, and the theory of solids. These systems are
neither ergodic nor mixing but exhibit behavior reminiscent of both in local
regions of their phase space. They have been studied extensively with computers
in recent years and give great insight into the behavior of flows in phase
space and the possible mechanism behind the irreversibility we observe in
nature. We briefly discuss such systems in the special topics in Section (S6.E).

STOI Converter
hDP:llwww.stdutililV.com

ERGODIC THEORY AND THE FOUNDATIONS OF STATISTICAL MECHANICS

297

Let us now define ergodic flow. Consider a Hamiltonian system with 3N


degrees of freedom with Hamiltonian H(pN, qN) = E. If we relabel the
momentum coordinates so PI = Px,I, pz = Py,I, P3 = Pz,b P4 = Px,z,,
P3N = Pz,N (with similar relabeling for the position coordinates), then
Hamilton's equations can be written

(6.33)

Equation (6.33) provides us with 6N - 1 equations between phase space


coordinates which, when solved, give us 6N - 1 constants, or integrals, of the
motion,
(6.34)
where i = 1, 2,
integrals of the
Converted with
motion can b
ting. Isolating
integrals defin
rtant in ergodic
theory, while n
e unimportant
[6, 14]. One 0
ine
how many
trial version
isolating integ
integral is the
only isolating
total energy,
integral (at leas or ar sp eres .
Let us consider a system for which the only isolating integral of the motion
is the total energy and assume that the system has total energy, E. Then
trajectories in I' space (the 6N-dimensional phase space) which have
energy, E, will be restricted to the energy surface, SE. The energy surface,
SE, is a (6N - 1)-dimensional "surface" in phase space which exists
because of the global integral of the motion, H (P I,... ,P3N, q I, ... ,q3N) = E.
The flow of state points on the energy surface is defined to be ergodic
if almost all points, X(PI,'"
,P3N,qI, ... ,Q3N), on the surface move in
such a way that they pass through every small finite neighborhood, RE, on
the energy surface. Or, in other words, each point samples small neighborhoods over the entire surface during the course of its motion (a given
point, X(PI,""
P3N, QI, ... , Q3N) cannot pass through every point on the
surface, because a line which cannot intersect itself cannot fill a surface of
two or more dimensions). Note that not all points need sample the surface,
only "almost all." We can exclude a set of measure zero from this
requirement.
A criterion for determining if a system is ergodic was established by
Birkhoff [15] and is called the ergodic theorem. Let us consider an integrable
phase function j'(X") of the state point XN. We may define a phase average of

STOI Converter
hDP://www.stdutilitv.com

298

THE FOUNDATIONS OF STATISTICAL MECHANICS

the function

(f)s

f(XN)

1
= 2:(E)

on the energy surface by the equation

s/(X N)dSE

Jr 8(H N(XN) - E)f(X N)dX N,

1
= 2:(E)

(6.35)

where dS E is an area element of the energy surface which is invariant (does not
change size) during the evolution of the system and 2:(E) is the area of the
energy surface and is defined as

(we are using the notation of Section 6.B). We may define a time average of the
function f(XN) by the equation

(fh
for all trajector
time average ir
interest (that is
It terms of a
is ergodic if for
almost all XN (

lim -I

T-oo

I"

f(XN (t))dt

(6.37)

to

~- howed that the


bns of physical

Converted with

STOI Converter
trial version

II
hDP: www.stdutiliJV.com

the phase aver;


To find the

lows: A system
(fh, exists for
ts it is equal to
first write an

expression for the volume of phase space, n(E), with energy less than E-that
is, the region of phase space for which 0 < HN (XN) < E. We shall assume that
the phase space can be divided into layers, each with different energy, and that
the layers can be arranged in the order of increasing energy. (This is possible for
all systems that we will consider.) The volume, n(E), can then be written

n(E)

dXN
O<HN(XN)

<E

o <HN(XN)

dAHdnH,

(6.38)

<E

where dAH is an area element on a surface of constant energy and dnn is normal
to that surface. Since V xHN is a vector perpendicular
to the surface
HN(XN) =constant, we can write dHN = IVxHNldnH and the volume becomes

(6.39)
where
(6.40)

299

ERGODIC THEORY AND THE FOUNDATIONS OF STATISTICAL MECHANICS

is a function of HN and is an invariant measure of the surface area for a given


value of HN. If we take the derivative of nee), we find

dn(E) = 2)E)
dE

J
SE

dAE.

(6.41)

IVXHNIH=E

The area, L:(E), is called the structure function. By the same argument, if we
wish to take the average value of a function f(XN)
over the surface, we can
write

Thus, the differential

dAE

(6.43)

dSE=-~""""""'-IVXHNIHN=E
is the invariant surface area element.

Converted with

EXERCI~

STOI Converter

noninteracting
total energy E
Answer:

a gas of N
system has a

trial version

The

hDP://www.stdutilitv.com
HN = 2;.!:!_ = E.
2m

(1)

i=1

: The volume of phase space with energy less than E is


I
I

\ fleE)

Iv dql ... Iv dllN J dPI ... J dPN

for

pi

+ ... + p~ :::;2mE.
(2)

!
I
I

I This can be written

nee) = vNnp, where

flp

dPI ...

dPN8(Jil-

pi - ... - ~).

(3)

\ is the volume enclosed in momentum space and R2 = 2mE. The volume in


, momentum space, np, has the form
= A3NR3N. Let us find A3N' This can
: be done by a trick. First do the integral

np

, ['"

dpI.x ... ['"

dP3N.zexp[-(Pi.x+

.+pjN.')]

= (['" dPe-rl-) 3N=~N/2.


(4)

300

THE FOUNDATIONS OF STATISTICAL MECHANICS

= I dPl

Next note that (dnp/dE)

Joo

dnp _R2
Joo
dR dR e
=

I dpN8(R2

...

dPl,x' . .

JOO

-00

pi - ... - p~) so that


2

dP3N,zexp[-(Pl,x

2
+ ... + P3N,z)]

-00

(5)

00

= 3NA3N

where

10

dR~N~le~R2

= ~NA3NrGN),

r(x) is the gamma function. If we equate Eq. (4) to Eq. (5), we find

2i3N/2

(6)

A3N = 3Nr(3N /2) .

Thus, the volume of the region of phase space with energy less than E is

2VN 7T3N
/2 R3N
VN (27TmE)3N/2
n(E) = 3Nr(3N /2) = r(3N /2 + 1) ,
The structure function,

L:(E),

(7)

equals (dn(E)/dE).

Converted with
If a system
in a given regi
surface SE oc

STOU Converter
trial ve rsion

hUP:llwww.stdutililV.com
lim

TRE

= L:(RE)

L:(E)

fraction of the
such' that
is easy to see

RE)

(RE) = 1 whe
that, for an erg""~""""""""""""","""'_"'~-----------_____j

T-oo

~N, qN), spends

(6.44 )

where TRE is the time the trajectory spends in RE and L:(RE) is the area
occupied by RE
A system can exhibit ergodic flow on the energy surface only if there are
no other isolating integrals of the motion which prevent trajectories from
moving freely on the energy surface. If no other isolating integrals exist, the
system is said to be metrically transitive (trajectories move freely on the energy
surface). If a system is ergodic, it will spend equal times in equal areas of the
energy surface. If we perform measurements to decide where on the surface
the system point is, we should find that result. We can also ask for the
probability of finding the system in a given region RE of the energy surface.
Since we have nothing to distinguish one region from another, the best
choice we can make is to assume that the probability P(RE) of finding
the system in RE is equal to the fraction of the energy surface occupied by

ERGODIC THEORY AND THE FOUNDATIONS OF STATISTICAL MECHANICS

301

RE. Thus,

P(RE)

= "L(E)

RE

"L(RE)
dSE = "L(E) .

(6.45)

From Eq. (6.45) it is a simple matter to write down a normalized probability


density for the energy surface, namely,
(6.46)
Equation (6.46) is called the fundamental distribution law by Khintchine and
called the microcanonical ensemble by Gibbs. Since it is a function only of the
energy, it is a stationary state of the Liouville equation (6.27). It says that all
states on the energy surface are equally probable. Equation (6.46) forms the
foundation upon which all of equilibrium and most of nonequilibrium statistical
mechanics are built. Its importance cannot be overemphasized. In Exercise 6.2,

However, it is
systems. In fa
and only if the
of course, th
Hamiltonian.

Converted with

STDI Converter
trial version

ssical systems.
s for quantum
m is ergodic if
]. This means,
ute with the

hDP://www.stdutilitv.com
EXERCISE 6.4. Consider a dynamical flow on the two-dimensional unit
square, 0 5:p 5: 1 and 05: q 5: 1, given by the equations of motion,
(dpldt) = a and (dqldt) = l. Assume that the system has periodic
boundary conditions. (a) Show that this flow is ergodic. (b) If the initial
probability density at time, t = 0, is pep, q, 0), compute the probability
density at time, t.
Answer:
(a) The equations of motion are easily solved to give
p(t)

= Po

+ at

and

q(t) = qo + t,

(1)

where Po and qo are the initial momentum and position, respectively.


If we eliminate the time t, we obtain the phase space trajectory,

p = Po + a(q - qo), on the square surface. If a is a rational number,


a = (min) (m and n integers), then the trajectory will be periodic and
repeat itself after a period, n. If a is irrational, the trajectory will be
dense on the unit square (but will not fill it. A trajectory is shown in
the accompanying figure.

302

THE FOUNDATIONS OF STATISTICAL MECHANICS

I
I
I I I
I I

Po

--}T'-I I I
I I I

I I
I
1

When a is irrational, the system is ergodic. Let us show this


explicitly. Since the phase space is periodic, any integrable function,
j(p, q), can be expanded in a Fourier series,

Converted with
We wi
functio
by

(2)

STOI Converter
trial version

verage of the
rage is given

hDP://www.stdutilitv.com

~---=~=-~-------;;W--~-~-----X-=~IllT"T""JT"l"'1I(Po+at)]

=A

00,00,

+ T-+oo
lim 0,0
T

'""
~

'"" A e211'i[m(qo+to)+n(po+ato)]
~
m,n

m=-oo n=-oo

(e

11'i(m+n)T

27T'i(m + an)

1).

(3)

The primes on the summations indicate that the values m = 0 and


n = 0 are excluded from the summation. For irrational values of a,
the denominator can never equal zero. Therefore

(4)

(f)T = Ao,o.
Similarly, we can show that

(f)s

f: f:

dpdqf(p,q)

= Ao,o.

(5)

Hence, the system is ergodic (note that dpdq = dpodqo, so area is


preserved).

THE QUANTUM PROBABILITY DENSITY OPERATOR

303

(b) The probability density satisfies periodic boundary conditions, so we


can write

L L
00

p(P, q, t) =

00

Pm,n(t)e27rimqe27rinp.

(6)

m=-oon=-oo

Itl is eas,y to show that Pm,n(t) = pm,n(0)e27ri(m+an)t. Next note that


Io dqe2mmq = 8m,o and L:~=-oo e2mmq = 8(q). Using these relations,
we find

p(P, q, t) = p(p

+ at, q + t, 0).

(7)

From Eq. (7) we see that ergodicity is not sufficient to cause a system
which initially has a nonstationary distribution (localized on the
energy surface) to approach a stationary state (spread throughout the
energy surface). The probability density in Eq. (7) does not change
shape with time, but simply wanders intact through the phase
space. In order to approach a stationary state the system must be
'mixin'
ond'rons for mixin flow are discussed in special topics
Sectio

Converted with

6.D. THE
OPERATO

STOU Converter
trial ve rsion

hnp://www.stdutilitv.com

For quantum
,
mmute so we
cannot introduce a probability density function directly on the phase space.
Because of the noncommutivity of phase space coordinates, we cannot
simultaneously know the values of all the phase space coordinates. Instead we
will introduce a probability
density operator or density operator as it is
commonly called. The density operator contains all possible information about
the state of the quantum system. If we wish we can use it to construct the
Wigner distribution, which is a function that reduces to the classical probability
density in the limit where Planck's constant goes to zero.
The probability density operator p(t) (we shall call it the density operator),
contains all possible information about the state of the quantum system. It is a
positive definite Hermitian operator. Given that we know the density operator,
p(t), for a system, we can use it to obtain the expectation value of any
observable 0 at time t. The expectation value is defined as
(O(t))

Tr Op(t) ,

(6.47)

where Tr denotes the trace. The density operator is normalized so that


Tr p(t) = 1.

(6.48)

304

THE FOUNDATIONS OF STATISTICAL MECHANICS

In Eqs. (6.47) and (6.48), the trace can be evaluated using any convenient
complete set of states. For example, we could use the eigenstates of the
operator, 0, or any other Hermitian operator, A, which mayor may not
commute with 0. We will let {Io;)} and {Ia;)} denote the complete orthonormal
sets of eigenstates of the operators, o and A,,, respectively, and le~ {o;} and {ai}
be the corresponding sets of eigenvalues (010;) = 0;10;) and (Ala;) = a;la;).
For simplicity we use Dirac notation (cf. Appendix B). The trace can be
evaluated in either of these basis. Thus, we can write
(O(t))

L o;(o;I,o(t)

10;)

L ~)a;IOlaj)(ajl,o(t)la;),
i

(6.49)

0; = (0;1010;) and
L:; la;) (a; I = 1, where i is

where

we have used the completeness


relation,
the unit operator. The diagonal matrix element,
(o;jp(t) 10;)( (a;jp(t)la;)),
gives the probability to find the system in the state
10;)(la;)),
at time t. The set of numbers, (ajl,o(t)lai),
forms a matrix
representation of the density operator (called the density matrix) with respect
to the basis sta
nite Hermitian
matrix. The 0
Converted with
~ j, cannot be
interpreted as a
The introduc
~escription of a
see, it can also
quantum systen
trial ve rsion
f a many-body
be used to desc
s" and "mixed
system. To see
lves according
states." Consid
,/I
to the Schrodinger equation,

STOU Converter
http://www.stdutilitv.com

-:J.

01

iii a I ~~t)) = iI I'l/J(t) ) ,

(6.50)

where iI is the Hamiltonian operator, and Ii is Planck's constant. The density


operator which describes this "pure state" is simply
,o(t) = 1'l/J(t))('l/J(t)l

(6.51 )

A "mixed state" is an incoherent mixture of states I'l/Ji(t)):


,o(t)

LPil'l/Ji(t))

('l/Ji(t)l,

(6.52)

where Pi is the probability to be in the state I'l/Ji(t)), and the states I'l/Ji(t)) each
satisfy the Schrodinger equation. Equilibrium and near-equilibrium states of
many-body systems are of this type.

305

THE QUANTUM PROBABILITY DENSITY OPERATOR

Using the Schrodinger equation, the equation of motion of the density


operator is easily found to be

.8p(t) = ![iI A(t)] = L A(t)


8t
n ,P
P,

(6.53)

where [iI, p(t)] is the commutator of the Hamiltonian, iI, with p(t), and the
operator L == [iI], is p~oportional to the commutator of iI with everything on
its right. The operator L is the quantum version of the Liouville operator and
is a Hermitian operator. Equation (6.53) is called the Liouville equation
and gives the evolution of the state of the system (in the Schrodinger picture).
If the density operator is known at time t = 0, then its value at time t is
given by

p(t) = e-;Lt p(O)

e-(i/n)i1t p(O)e+(i/n)i1t

(6.54)

If we substitute Eq. (6.54) into Eq. (6.47) and use the invariance of the trace
under cyclic r
.
.L

I'"

T:'.

.1,

..4"7\

.L

& .

111

Converted with

STDI Converter

where

(6.55)

trial version

hDP://www.stdutilitv.com
Thus the operator,

0, obeys

a different equation of motion,

80(t)
1
---at
= h [H, O(t)]
A

-i

(6.56)

= LO(t) ,

(6.57)

which is different from that of the density matrix. Equation (6.57) gives the
evolution of the system in the "Heisenberg" picture.
It is often convenient to expand the density operator in terms of a complete
orthonormal set of eigenstates {IE;)} of the Hamiltonian, iI, where E; is the
eigenvalue corresponding to eigenstate IE;). If we note the completeness
relation L; IE;) (E;I = 1, then Eq. (6.54) takes the form

p(t) = L

L(E;jp(O)
;

IEj)e-(;/n)(Ei-Ej)tIE;) (EJI.

(6.58)

From Eq. (6.58), we see that a stationary state, Ps, occurs when all off-diagonal
matrix elements (E;lp(O)IEj) with i i=j, of p(O) vanish of E; i= Ej Thus, for a
state with no degenerate energy levels, the stationary state, Ps, must be
diagonal in the energy basis. This can only happen if Ps is a function of the

306

THE FOUNDATIONS OF STATISTICAL MECHANICS

Hamiltonian,

Ps = f(iI).

(6.59)

For a system with degenerate energy levels, one may still diagonalize both p
and iI simultaniously by introducing additional invariants of the motion, 1
which commute with each other and with iI. Thus, in general, a stationary state
will be a function of all mutually commuting operators, iI,il, ...

.t;

(6.60)
For systems which approach thermodynamic equilibrium, the stationary state
may be an equilibrium state.

Answer:

hDP:llwww.stdutililV.com

(a) The probability to find the particle in the interval x ~ x + dx at


time t = 0 is (xlp(O) Ix)dx, where Ix) is an eigenstate of the

position operator
Then

x.

We will use the notation Px/,x(O)

==

(x'lp(O)lx).

to:

where we have used the completeness relation,


dplp)(PI = 1, for
momentum eigenstates and the conventions of Appendix B. The
probability to find the particle in the interval x ~ x + dx is Px,x(O)dx,
where
Px,x(O)

fa e -ux2 .
y;

(2)

THE QUANTUM PROBABnJTY

307

DENSITY OPERATOR

(b) The Liouville equation in the position basis is

ili8p~x(t)
t

= -2

1
m

00

dx"((x'ljJ21x")Px",x(t)
- x2)PX/,x(t)

+ ~mw2(x'2

_1_1 1
00

00

dp

4m7r1i

-00

dx"p2(ei(P/h)(X-xl)

PX",x(t)

+ ~mw2(x'2

- ~)pX/,X(t)

-00

- Px',xl(t)ei(p/h)(xl-x)

- PX/,x"(t) (x"ljJ2Ix))

-00

4-"': [,
- p x',(t)
I

x,

dX"(

dp [,

(a~~2et(P/n)(x'-x"))Px',x(t)

_!!_ei(P/h)(X1-X))
8x"2

Converted with
If we
can in

(3)

STOI Converter

00,

then we

trial version

iii 8Px.:.n------_hD.....,."P,---.-:'_,---.-.www.....-=--_.S_td_U..,........ti,.,,-li_tv--.---.C_o--.---m_---'2-x2

)Pxl ,x(t).
(4)

(c) To find Px,x(t), let us first solve the I:iouville equation in the basis of
eigenstates of the Hamiltonian, H = (1/2m)jJ2 + !mw2 2.
From
Exercise 5.8, we see that the Hamiltonian has eigenvalues
En = liw(n +!) and eigenstates In), (Hln) = Enln)), which in the
position basis are

_
(xln) -

y~fl(mw)

7r1i

1/4

H,

((iiiW)
yr:x

mwx2/2h

(5)

where HneJ(mw/li)x)
is a Hermite polynomial. The Liouville
equation in the basis of eigenstates In) is

In. 8pn'at n(t)


where Pn/,n(t) = (n'lp(t)ln).

= 1iw

('
)
n - n Pn/,n(t),

(6)

The solution to Eq. (6) is

Pnl,n(t) = e-iw(n'-n)tpn/,n(O).

(7)

308

THE FOUNDATIONS OF STATISTICAL MECHANICS

L:dx~ L:dxo(n/lx'o)px~,xo(O)(xoln).

Let us now note that Pn',n(O) =


Then the density matrix, Px',x(t),

can be written

tWo

prAt)=~~[

(8)
x ['"

dxo (x' In') (nix)

If we use the initial condition


Px,x(t), can be written
1

Px,x(t) =

Let us
5.8. W'
00

ra

00

2V;~~

roo

00

(.xo In) (n'lx'o)e-lw(n' -n)1Px;,xo (0).


in part (a), the probability

Joo
-00

density,

tWo

dxo(xln') (nix) (xoln) (n'Ix;,)e-iw(n'-n)1

Converted with

(9)

of Exercise

STOI Converter
trial version

E(xol
n=O

'----_h_n_p:/_I_www-----y---r-._st--..=-dU."..,ti---r-<li_tv~.c---rT""0m_____._:r__"iwt) ) ,

(10)

where /(t) = ie-iwt /2sin(wt). If we use Eq. (10) to perform the


summations in Eq. (9), then we are left integrals over x'o and xo which
can be performed explicitly. After considerable algebra, we find
ux2IB
P x,x (t) =@_Re[_l_eyB(i)(t)]

V -;

(11 )

where Re denotes the real part and

B(t) = cos2(wt)

2
+ bmw
~ 2 sin2(wt) + i a\os(wt)sin(wt).
mw

(12)

i)

EXERCISE
6.6. An ensemble of silver atoms (each with spill
is
prepared so that 60% of the atoms are in the Sz =: +~eigenstate of S, and
40% of the atoms are in the S; = - ~eigenstate of S, (Sx and Sz are the x and
z components of the spin angular momentum operator). (a) Compute the
density matrix at time t = 0 in the basis of eigenstates of
(b) Assume that

s;

309

THE QUANTUM PROBABILITY DENSITY OPERATOR

the silver atoms sit in a magnetic field, B = BoY, and have a magnetic
Hamiltonian, iI = J-lS . B, where J-lis the magnetic moment of a silver atom.
Compute the density matrix at time t (in the basis of eigenstates of Sz). (c)
Compute (Sz(t)) at time t = 0 and at time t.
Answer:

li) denote the eigenstates of Sk(k = x, y, z) with eigenvalues


(n/2) so (Sklk) = (n/2)lk)). The density operator at time t = 0

(a) Let

is the mixed state


jJ(O) = 1~ Iz+)(z+I

+ 1~ Ix-)(x-I

(1)

Now note that the matrix representation of the components


of eigenstates of
are

s,

s, in the basis

s., s; and

Converted with
The ei

of

STOI Converter

~J

trial version

hDP://www.stdutilitv.com

respect
are

G~tD

= (~)

and

s, are

nstates of

G~tD

= (~).

(3)

s,
(4)

Using these results we find the initial density matrix in the Iz)
basis,

p(O) =

(!! k!).

(5)

(b) The Hamiltonian is iI = !J-lnBo(ly+)(y+I-ly-)(y-1)


(this is its
spectral decomposition). If we let Py++ == (y+ljJly+),
then the
Liouville equation for various matrix elements of the density matrix
in basis Iy) is given by
,.apy++(t)
at

= 0,

.apy+_(t)
at

_ B
(t)
.apy_+(t)
- J-l OPy+, I
at
.apy __ (t)
= -J-lBOPy-+ (t ) an d I at
= 0,
I

(6)

310

THE FOUNDATIONS OF STATISTICAL MECHANICS

Now note that the initial density matrix in the basis Iy) is given by

p(O) = (

J:Ui)
~

1
3~2i
10

(7)

Using Eq. (7), we can solve Eqs. (6) and write the density matrix at
time t in the basis Iy). We find

p(t) =

J:Ui
e-iJ.LBot)
10

1
(

3-2i
10

2
+iJ.LBot

(8)

We now can transform Eq. (8) to the basis Iz) to obtain

- (!+foCOS(Ji,Bot) +! sin (Ji,Bot)


3'
1
(
10
sm ( J-LBot) -scos
/-lBot )

fosin(Ji,Bot)-!cos(Ji,Bot)

p(t ) -

1 3
(
)
2-1OCOS /-lBot

1 . (
-ssm
/-lBot ).

(9)
(c) The av.re.t:aLJ.Ie.....:z...c.:.amt.nllln.entL.o:t.....l.Ilin._..a.ru:m..l.!ar._tD.llln.e.tl..tllJ~t
time, t = 0
is

Converted with

STOI Converter
trial version

t time tis

hDP://www.stdutilitv.com
(11)

.....SPECIAL TOPICS
.....S6.A. Reduced Probability Densities and the BBGKY
Hierarchy [2, 5, 17]

The N-particle probability density, p(XN, t), contains much more information
than we would ever need or want. Most quantities we measure experimentally
can be expressed in terms of one-body or two-body phase functions. One-body
phase functions are usually written in the form
N

oft)(XN)

:E O(X

i),

i=1

(6.61 )

311

SPECIAL TOPICS: REDUCED PROBABILITY DENSITIES

and two-body phase functions are written in the form


N(N-l)/2

o(;)(XN)

O(X;,Xj),

(6.62)

;<j

L:~lPf

An example of a one-body phase function is the kinetic energy,


12m, of
an N-particle system. An example of a two-body phase function is the potential
1
energy,
)/2 V(lq; - qjl), of an N-particle system. To find the expectation
value of a one-body phase function, we only need to know the one-body
reduced probability density. Similarly, to find the expectation value of a twobody phase function, we only need to know the two-body reduced probability
density.
The one-body reduced probability density is given by

L:fS-

.AT

where p(Xl,.
given by
Ps(X1,

pility density is

Converted with

STOI Converter

t).

(6.64)

trial version
If the probab
expectation va
(O(l)(t))

me t, then the

hDP://www.stdutilitv.com

t f- .. J

dXI ... dXN O(X;)p(XN, t) = N

iven by

dXIO(XI,

t)PI (Xl, t).


(6.65)

Similarly, the expectation value of the two-body phase function at time t is


N(N-l)/2

(0(2)(t)) =

~<.

= N(N

In Eqs. (6.65) and


is symmetric under
symmetric.
The equation of
Liouville equation.

J... J

dXI ... dXN O(X"Xj)p(XN,

1)

t)
(6.66)

J J dXldX20(XI,X2,t)P2(XI,X2,t).

(6.66), we have assumed that the probability density


interchange of particle labels if the Hamiltonian is
motion of Ps(X1"",
Xs, t) can be obtained from the
It is convenient to first introduce another quantity,

312

THE FOUNDATIONS OF STATISTICAL MECHANICS

Fs(Xl' ... ,Xs, t), defined as

and
(6.68)
Let us assume that the evolution of the system is governed by a Hamiltonian of
the form

(6.69)
where (Iqi - qjl) is a two-body spherically symmetric interaction potential
between particl'
.
.,
.

Converted with

STOI Converter

(6.70)

trial version
where

hDP://www.stdutilitv.com
(6.71 )

and ij = (Iqi - qjl). If we integrate the Liouville equation, (6.27), over


Xs+1, . ,XN and multiply by V we obtain
S

a;, + tirr, f- ..J


=V'

dXH1

...

dXN

X{_tPi.~+
i=s+l m 8qi
X

pN (X 1,

...

, XN,

i::;s;s+l ::;j::;N

8ij+

Ski}

(6.72)

s+I ::;k<l

t).

If we assume that p(Xl" .. ,XN, t) --+ 0 for large values of Xi, then the first
and third terms on the right-hand side of Eqs. (6.72) go to zero. One can see this
by using Gauss's theorem and changing the volume integration to surface
integration. For a large system the contribution from p(X 1, ... , XN, t) on the
surface goes to zero. The second term on the right-hand side can be written in

313

SPECIAL TOPICS: REDUCED PROBABILITY DENSITIES

the form

v'

I- ..J

= V'(N - s)
=

dXH, dXN

tJ

(N-S)~J ~
V

; $.

dX,+,8",+,
"

8ijr!'(X"

... ,XN,t)

s;s+;$.i $. N

J ... J

dXs+18;,s+lFs+l(Xl,

dXH2" .dXNr!'(X"

... ,XN,t)

... ,XS+l,t).

;=1

Equation (6.72) then becomes


(6.73)
For a fixed values of S we may take the limit N ~ 00, V -4 00, . such that
v V IN remains constant (this is called the thermodynamic limit) and Eq.
(6.73) becomes

Converted with

er,

-+
at

STOI Converter

(6.74)

trial version
Equation (6.74
~r the reduced
probability den
hierarchy after
authors Bogoliubov [17], Born and Green [18], Kirkwood [19], and Yvon [20].
The most useful equations in the hierarchy are those for F 1 (X 1 , t) and
F2(Xl, X2, t):

hDP:llwww.stdutililV.com

(6.75)
and

Notice that the equation of motion of FI depends on F2, the equation of motion
for F 2 depends on F 3, and so on. This makes the equations of the hierarchy
impossible to solve unless some way can be found to truncate it. For example, if
We could find some way to write F2(XI, X2, t) in terms of F, (Xl, t) and
Fl (X2, t), then we could in principle solve Eq. (6.75) for the reduced
probability density F, (Xl, t). Equation (6.75) is called the kinetic equation.

314

THE FOUNDATIONS OF STATISTICAL MECHANICS

.... S6.B. Reduced Density Matrices and the Wigner


Distribution [16, 21-25]
For quantum mechanical system, the phase space coordinates of particles do not
commute, and therefore it is impossible to specify simultaneously the position
and momentum of the particles. As a result, it is also not possible to define a
distribution function on the phase space which can be interpreted as a
probability density. However, Wigner [16] was first to show that it is possible to
introduce a function which formally analogous to the classical probability
density and which reduces to it in the classical limit.
Before we introduce the Wigner function, it is useful to introduce the idea of
one- and two-body reduced density matrices. In quantum mechanics, as in
classical mechanics, we generally deal with one-body operators,
N

00) = L O(p;, I;),

(6.77)

;=1

such as the N_...........


operators,

..ld.....z..._ ...........
l.

~~...-.......~'---""'-"'~~~_...,,_A~~._1

........... ~A .............

+iwith two-body

Converted with

STOI Converter

(6.78)

trial version

hDP://www.stdutilitv.com

such as the in
position basis, can oe

(O(1)(t))

perator, in the

written

= Tr 00/*)
=

tJ

tUI ...

x (X~,

=N
==

tUI

J tUN J ttx; .,. J diN(XI,'"

,x'N)

,X~ IjJ( t) lXI, ... ,XN)

J tUN J ttx;

J J ttx; (xdOdX;)(X;
tUI

,xNIOilX;,

(XI I(h

1X;)(X;,X2, .,. , XNl.o(t)IXI, ... ,XN)

1,o(1)(t)lxI),
(6.79)

is the one-body reduced density matrix. (We use the notation of Appendix B.)

315

SPECIAL TOPICS: REDUCED DENSITY MATRICES

The two-body reduced density matrix is defined in an analogous manner. The


trace of a two-body operator in the position basis can be written

(O(2)(t))

= Tr

O~)p(t)

N(N-l)/2

L.: J

J J dx\ ... J

dxi -:

dx

dx'N(Xl, ... ,XNIO"jlX;,

,x'N)

I<.J

!XI, ... , XN)

x (x;, ... , x~!p(t)

= N(N

==

J J dx, J

1) dx1

(x;,X~,X3,'"

dxN

,XN!p(t)!XI,

H J J dx\ J
dx, dx2

~(xJ,x2101,21X;,xi)

... ,XN)

~(Xl,X2101,21X;,x'2)(X;,xilp(2)(t)lxl,X2)'
(6.81 )

where

o., = 0 ~. ~

and

Converted with

(X;,~!

STOI Converter

(6.82)

trial version
is the two-bod
We can no
~~--~=---~~--~~~~----~
The one-particle reduced Wigner function is defined as

hDP:llwww.stdutililV.com

gner functions.

(6.83)

and the two-particle

/2 (kl , k2,

reduced Wigner function is defined as

Rl , R2; t) =

JJ
x

d rl d r2eik,.r, e'k,r,

rl
\Rl + 2";R
2

~
+2"!P(2)(t)!RI

rl~)
2",R
2

2" .
(6.84)

Higher-order Wigner functions


In analogy to the classical
function obeys the relations

J (27r)\/1
d

can be defined in a similar manner.


distribution function, the one-particle

(k, R, t) = (Rlp(l) (t) IR)

= n(R,

t),

Wigner

(6.85)

316

THE FOUNDATIONS OF STATISTICAL MECHANICS

where n(R, t) is the average number of particles at point R and time, t, and

J dRJi (k,

R, t)

= (klp(I)(t)lk) = n(k, t),

(6.86)

and n(k, t) is the average number of particles with wavevector, k, at time t. The
Wigner function can be used to take phase space averages in the same way as
the classical distribution functions. For example, the average current is defined
as
(j(R, t)) =

J (27T")
dk 3/ikII (k, R, t),

(6.87)

where 1i is Planck's constant. However, the Wigner function can become


negative and therefore cannot always be interpreted as a probability density.
We can derive the equation of motion for 11 (k, R, t) in the following way.
For a system with a Hamiltonian

Converted with
the equation of

STOI Converter
trial version

(6.88)

px

is

(rllp(1)(t)I'-----r-_h_n....-y::L~:=-:..----II---y-www--.".S-,--t~d-ut-r--il-itv~.c~o~m---.-r-<"-r-lr2)

-H

dr'[V(r2

r') - V(r, - r')](r, , r'lp(2)(t)lr2,r'),


(6.89)

where V(rl - r2) = (rl,r21V(<<Ii - Ij)lrl,r2).


Let us now change to relative and center-of-mass coordinates, r = rl - r2
and R = (r. + r2), respectively. We then multiply by eik.r and integrate over
dr and find

(6.90)

317

SPECIAL TOPICS: REDUCED DENSITY MATRICES

In Eq. (6.90) we can integrate the first term on the right by parts to remove the
derivative with respect to r. We can also introduce dummy variables into the
second term on the right and make use of definitions in Eqs. (6.83) and (6.84) to
obtain

-at /I(k,R,t)

=--k\7R/l(k,R,t)
m

-- i

J --d k' J dr ,[ V ( R-r / ---2i1 aka)

n (27r)3

a) 1

- V R-r ,1+--

2i ak

x h (k, R; k', r'; t).


(6.91)
Equation (6.91) is the quantum kinetic equation for the one-particle reduced
Wigner function. If we rewrite it in terms of momenta p = nk and p' = nk', it
takes the form

ata 11'( p, R,

=-H(

Converted with

(6.92)

trial version

x/~(p
where

a)]
.ap

STOI Converter
hDP://www.stdutilitv.com

(6.93)
We can now expand the potential on the right-hand side in powers of n and take
the limit n ~ O. We then retrieve the classical kinetic equation.
The Wigner function can be used to take the average value of a large class of
ordinary functions of momentum and position but in some cases it will give the
wrong answer. The average value of any quantity which is only a function of
position or only a function of momentum can always be taken (this is easily
seen from Eqs. (6.85) and (6.86)). However, only those functions which involve
both position and momentum can be used for which the Weyl correspondence
between the quantum and classical version of the operators holds. To see this,
let us consider the classical function O(p, q) of phase space variables p and q.
We can find the quantum version of this function as follows. We first introduce
its Fourier transform, O(tI, ,,), with the equation

O(p, q)

J J d tid"

O(a, ")ei(tlop+,,oq)

(6.94)

318

THE FOUNDATIONS OF STATISTICAL MECHANICS

Matrix elements of the quantum operator corresponding to O(p, q) can be


obtained from the Fourier transform, O( a, 'I), via the equation

(r'[O[rll)
where

J J d ad '10(0', '1) (r'[e

i(6.';- . ij)

[r"),

(6.95)

p and q are momentum

and position operators. If the classical function


0 are related by the
above procedure, then the expectation value of 0 is given by

O(p, q) and matrix elements of the corresponding operator

(O(t)) =

J I dpd

(6.96)

rO(p, rlf:(p, r, t).

There are some cases for which the Weyl procedure does not give the correct
correspondence between classical and quantum operators (such as the
commutator [p, f]_, the square of the Hamiltonian iI2, the square of the
angular momentum L2, etc.) and the Wigner function gives the wrong result.
Then It IS necessary to introduce more general auantum phase space
distributions wb

Converted with

STOI Converter
trial version

EXERCIS
density operate

stern with a

hDP://www.stdutilitv.com

Answer: First compute the matrix element of the density operator in the
position basis,
(1)
Now let Xl = X + (x/2) and
by (after some algebra)

f(k,X)

[x,

ih
(

dxe

Xl -

(x/2). The Wigner function is then given

+ ~ [piX -~)

2
2
2
2 2
4ab1i
(-aX )
( -bk 1i )
(2ab1i
1 + ab1i2 exp 1 + ab1i2 exp 1 + ab1i2 cos 1 + ab1i2

kX)

(2)
A plot of the Wigner function is shown here for a = 1 and b1i2 = 2. Note that
there are some regions where it becomes negative, indicating that it is not a
probability density.

319

SPECIAL TOPICS: MICROSCOPIC BALANCE EQUATIONS

.... S6.C. Microscopic Balance Equations [26]

For quantum
inhomogeneit
density, mome
derivation for
momentum 01

[Pi,pj

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com
LA,

L A,

AJJ

The commutator of the momentum operator


coordinate operators is

A JJ

ong-wavelength
for the particle
nalogous to the
Ute position and

(6.97)

Pi with an arbitrary function of

(6.98)
while the commutator
G(Pt, ... , PN) is

of

Ii

with an arbitrary

function

of momenta

(6.99)
Let us assume that the dynamics of the system is governed by a Hamiltonian of
the form

(6.100)

320

THE FOUNDATIONS OF STATISTICAL MECHANICS

Then from Eq. (6.57) the equation of motion for the operator

qi is
(6.101)

The equation of motion of

Pi

is given by
(6.102)

These equations have the same form as the classical equations.


For quantum systems, the microscopic expressions for the densities must be
Hermitian in order to be observable. Thus, we must have symmetrized
expressions for operators which involve both momentum and position. Using
Eq. (6.57) and the above equations, we can show that the balance equation for
the particle number density is given by
(6.103)
where the parti

Converted with

STOI Converter
trial version

and the particle

(6.104)

hDP://www.stdutilitv.com
(6.105)

As usual, we let pN denote the set of momenta pN = (PI, ... , PN) and we let qN
denote the set of positions qN = (ql, ... , IN).
The balance equation for the momentum density takes the form
m :tjn(pN,

qN; R) = -VR . jp(pN, qN; R),

(6.106)

where the momentum current tensor, jP(pN, qN; R), is defined as

(6.107)

321

SPECIAL TOPICS: MIXING FLOW

In Eqs. (6.106) and (6.107) the notation


(6.108)
has been used.
Finally, the balance equation for the energy density can be written

ata h" ("N


p,

"N R) = q;

t"7
V R

J"h( p,
"N q;
"N R) ,

(6.109)

where the energy density is defined as


(6.110)
with
I"

12

Converted with
and the energ

(6.111)

STOI Converter
trial version

J"h("N
P ,q".11

hDP://www.stdutilitv.com
(6.112)

To obtain Eq. (6.112), one must use the fact that the center-of-mass coordinates
commute with the relative coordinates .

.... S6.D. Mixing Flow [7-9]


Ergodicity is not a sufficient condition on a dynamical flow to ensure that a
probability distribution that initially is localized on an energy surface will
spread, in a course grained manner, throughout the energy surface. Spreading
throughout the energy surface is the type of behavior that we need for a

322

THE FOUNDATIONS OF STATISTICAL MECHANICS

Newtonian system to approach a state that might be used as an equilibrium


state. A type of flow that does have this feature, and therefore might exhibit
irreversibility in a course-grained sense (because of the unstable nature of its
dynamics!) is mixing flow. Mixing flow is chaotic and causes any initial
probability distribution to spread throughout the energy surface. Mixing flow is
ergodic, but ergodic flows are not always mixing (the exceptions, however, are
rare).
A system is mixing if, for all square integrable functions, f(XN) and g(XN),
on the energy surface, SE, we obtain

where .L:(E) is the structure function defined in Section 6.C.


Equation (6.113) ensures that the average value of a dynamical function
f(XN)
will approach
a stationary
value in the limit t -4 oo. Let
g(XN) = p(XN
N
sity. Then

Converted with
(f(t)) f(t)
1
[.L:(E)r

Thus,

app

STDI Converter
trial version

hDP:llwww.stdutililV.com

(6.114)

state Ps

It is importa
ned and not a
fine-grained approach to a stationary state. The average of the probabi1ity
density becomes uniform, but the probability density itself cannot because of
Eq. (6.24). The probability density does not change in a neighborhood of a
moving phase point, but it can change at a given point in space. We can
visualize this if we consider a beaker containing oil and water. We add the oil
and water carefully so that they are initially separated, and we assume that they
cannot diffuse into one another. We then stir them together (Fig. 6.2). The local
density and the total volume of the oil remain constant, but the oil will get
stretched into filaments throughout the water. Therefore, on the average, the
density of the oil will become uniform throughout the beaker. If we are careful
enough, we can also stir the oil back into its original shape. Therefore, while we
get an approach to uniformity, the whole process can be reversed. However, as
we shall see, mixing does lead to the appearance of random behavior in
deterministic systems and coarse-grained irreversibility.
The meaning of Eq. (6.113) may therefore be summarized as follows. Let A
and B be two finite arbitrary regions on the surface SE. Let us assume that al1
phase points initially lie in A. If the system is mixing, and we let it evolve in
time, the fraction of points which lie in A or B in the limit t -4 oo will equal
the fraction of area SE occupied by A or B, respectively.

323

SPECIAL TOPICS: MIXING FLOW

Fig. 6.2. Stirring oil and water together leads to uniformity, on the average.

An example of a discrete dynamical system which is mixing is the Baker's


map. The dynamics evolves on a unit square, 0 ~ p ~ 1 and 0 ~ q ~ 1. The state
points evolve in discrete time steps and change value in a discontinuous manner.
The dynamical evolution is governed by an "alphabet" with two letters, 0 and
1, and the set, {S}, of all possible infinite sequences of letters. Each infinite
sequence has the form S = (... ,S-2,S-l,SO,S1,S2'"
.), where Sk =(0 or 1)
and k = (0,
. t in the twodimensional p
uare according
to the rules,

STOI Converter

p=

trial version

(6.115)

hDP://www.stdutilitv.com
Thus, if all Sk

m
O~q~
L~1

= 1.

0, then p = q = 0, and if all Sk = 1, then p = q = 1 since


All other cases also lie on the unit square, 0 ~ p ~ 1 and

1.

The dynamics is introduced into the phase space by means of


shift, U, which is defined so that USk = Sk+ 1. That is, the operator,
a sequence, S, shifts each element to the right one place. The shift
sequence, S, is equivalent to the following transformation
transformation) on the unit square:

U(p

) _ { (2p,!q),
,q -

The inverse transformations

(2

- 1 1
P
'2q

O~p~!

+ 1)
2'

!~p

~ 1.

the Bernoulli
U, acting on
acting on the
(the Baker's

(6.116)

is

(6.117)

324

THE FOUNDATIONS OF STATISTICAL MECHANICS

The Jacobian of the transformation is equal to one. Therefore, the mapping, U,


is area-preserving. Let us now look at how the probability density evolves under
the action of the shift, U. Let Po(P, q) be the initial probability density and let us
assume that it is continuous and smooth. After we allow U to act n times, we
obtain
(6.118)
More explicitly,

(6.119)
The time evolution operator, U, is an example of a Frobenius-Perron operator
[27] and [28].
The effect of the Baker's transformation is to stretch an initial area element
into filaments throughout the unit square, much as a baker does in kneading
dough. Note th
~e sequence, S,
will lie to the
Converted with
right of p =
Therefore, poi
random in the I
random. If init
0:::; q:::; 1 and

!.

STDI Converter

distributed at
P = by U at
0 p:::;
and
~e probability

trial version

Ilwww

density at time
hDP :
Sid U I-I-tv
As we have

II
the position of
a point in the p direction. Let us look at the reduced probability density, (P), in
the p direction. The quantity n (P)dp is the probability of finding a point in
the interval p ~ dp at 'time' n. It is defined as n ==
dqPn(P, q). Using
Eq. (6.119), n+l (P) becomes

.com

f~

(6.120)

The reduced probability evolves in a Markovian manner. We can show that


limn-toon(p) = 1 and, therefore, that the reduced or coarse-grained probability

Fig. 6.3. Evolution of an initial probability under the action of the Baker's map.

325

SPECIAL TOPICS: MIXING FLOW

density approaches

a constant.

If we iterate Eq. (6.120), we obtain

(6.121)

For a continuous

lim 4>n(P)

n-oo

4>0(p ), we obtain

and smooth function,

lim

1-1/2n

+ y) =
n

dy 4>0(-2

n-oo 0

11
0

dY4>o(y) =

1,

(6.122)

where we have let y = k'l:", Notice that for n -+ -oo,4>n(P)


will not approach
a constant. However, a reduced probability density, defined in terms of the
variable q, does. Therefore, the Baker's transformation exhibits irreversibility in
a course-grained sense.

Converted with
Answer:

if(P,

Let
q) are f

STOI Converter

trial version
hDP:llwww.stdutililV.com
e' ....._.........
~.

pn(p,

(1)

IJ).

q) and

ten

Before we
,,"n ........
_~
-0 p ..vp ....
u. of the Dlfac
delta function. Let us consider a function,J(x),
which has zeros at points XO,k
(where k = 1,2, ... ,M), so thatJ(xO,k)
= O. Then
u~

\U"

......

_ ~
8(f(x))

= {;;{

8(x - XO,k)
If'(xO,k)I

(2)

'

whereJ'(x)
= (dJ/dx).
We now can evaluate the trace of
Notice that the delta functions will
give contributions when r" = p and if = q. That is, for all 2n nth-order
periodic points of the map, tr, First note that (dPn/dpo) = 2n and
(dqn/dqo) = (1/2n). Thus,

tr.

Tr

o =

IdPn _
Period n points dpo
1)

= ( 1 - 2n

-2

00

II- dqn

dqo

= ~ (m +

11- = 2n(2n
1)

If1

(1 _~)-1
2

1) 2m

(cf. Ref. 28). The trace of U" is equal to the sum of the eigenvalues

(3)
of U",

326

THE FOUNDATIONS OF STATISTICAL MECHANICS

~ S6.E. Anharmonic Oscillator Systems [4]


The study of anharmonic
oscillator systems has long been important in
statistical mechanics in connection with the theory of heat transport in solids
[29]. One would like to know the mechanism of heat conduction. If heat is
added to a harmonic lattice and divided unequally between the normal modes,
there is no way for the system to reach equilibrium because the normal modes
do not interact. However, if slight anharmonicities
exist in the lattice, it was
expected that equipartition of energy would occur and that the system would
thus reach equilibrium.
In 1955, Fermi, Pasta, and Ulam [30] conducted a computer experiment
intending to show this. They studied a system of 64 oscillators with cubic and
broken linear coupling. They found that when energy was added to a few of the
lower modes there was no tendency for the energy to spread to the other modes.
This behavior is quite different from what one would expect if the anharmonic
oscillator system were an ergodic system. Then one expects the system to reach
a stationary state in which all states with the same energy would be equally
probable, and one ex ects
sharin amon th modes.
The type of
ow fairly well
understood in
old [32], and
~~t~~ [~~(c~
KAM theorem)
and the syste

STOI Converter
trial version

tates that for a


f invariant tori
to that of an

hDP://www.stdutilitv.com

unperturbed h
will not be
metrically tran
,
, the invariant
regions of phase space break down and at some point one expects to see a sharp
transition to chaotic behavior and something similar to equipartition of energy
between the modes. (True equipartition requires ergodicity, and it is not clear
that anharmonic oscillator systems are ergodic above the transition energy.)
There is now a variety of nonlinear oscillator systems which have been
studied and which exhibit a transition from stable to chaotic behavior as certain
parameters are changed [4]. Henon and Heiles [34] studied the bounded motion
of orbits for a system with two degrees of freedom governed by the Hamiltonian
(6.123)
The trajectories move in a four-dimensional phase space but are restricted to a
three-dimensional
surface because the total energy is a constant of the motion.
It is possible to study a two-dimensional cross section of the three-dimensional
energy surface. For example, we can consider the surface q2 = 0 and look at a
trajectory each time it passes through the surface with positive velocity P2 > O.
It is then possible to plot successive points of the trajectory (q2 = 0,P2 > 0) in
the PI, ql plane. If the only constant of motion is the total energy, E, then the
points should be free to wander through the region of the PI, qi plane

SPECIAL TOPICS: ANHARMONIC OSCILLATOR SYSTEMS

327

0.4

0.2

-0.2

-0.4
-0.4

-0.2

Fig. 6.4. Henon-Heiles result for E

0.2

= 0.08333.

0.4

(Based on Ref 34.)

Converted with
ear to be quite
corresponding
e motion, the
similar to ergo
points will lie
trial version
Henon and
(6.123) for a v
energy, E = O.
,
eating that to
computer accuracy there was an additional constant of the motion. Each closed
curve in Fig. 6.4 corresponds to one trajectory. The three points of intersection
of lines are hyperbolic fixed points, and the four points surrounded by curves
are elliptic fixed points [4]. However, at an energy E = 0.12500, the picture
begins to break down (cf. Fig. 6.5). Each closed curve in Fig. 6.5 corresponds to
one trajectory. The five islands correspond to one trajectory, and the random
dots outside the closed curve correspond to one trajectory. At an energy of
E = 0.16667, almost no stable motion remains (cf. Fig. 6.6). A single trajectory
is free to wander over almost the entire energy surface. In a very small energy
range the system has undergone a transition from stable to chaotic behavior.
Additional studies of the Henon-Helies system [35, 36] have shown that
trajectory points move apart linearly in the stable regions, whereas they move
apart exponentially in the chaotic regions.
The change from stable to chaotic behavior sets in rather abruptly. This has
been understood in terms of an overlapping of resonances in the system. The
Hamiltonian for a general anharmonic system with two degrees of freedom can
be written in terms of action angle variables in the form

STOI Converter
hDP:llwww.stdutililV.com

(6.124)

328

THE FOUNDATIONS OF STATISTICAL MECHANICS

0.4

'c;>.

"' ..., ..

<:::::)

0.2
;

~.

..\...... :..-:-..

.::......

. ..:o:~...
.~
,.. ~

...:.

.. :

..

. ".'.

c.-::a

....

1::-

..:

..~
~ ..
!:
.

~~

...

~:.~. oJ.

-0.2

e.

-0.4
-0.4

-0.2

0.2

0.4

Fig. 6.5. Henon-Heiles result for E = 0.12500. (Based on Ref 34.)

Converted with

Pl

STOI Converter

0.4

trial version
0.2

hDP://www.stdutilitv.com
:.:.. .. .. "'..
....
. .... -l"'~

.':

.:

-0.2

::

~..

6.~~""

~~

:JI":'4fIit

0.2

~.
o~
",

'_ ,... _.
: ~

z :

. ... .~:,..:.;..~:... : :
......

. .

-0.4

-0.4

-0.2

Fig. 6.6. Henon-Heiles result for E

= 0.16667.

0.4

0.6

0.8

(Based on Ref 34.)

by means of the transformation


Pi

= _(2mwJi)I/2sinq)i

and

q,

= (~)

1/2COSq)i'

The function HO(J1,h) has a polynomial dependence on the action variables J,


and h (not merely a linear dependence as would be the case for a harmonic
system) and no angle dependence, while V(J1, Jz, CP1, CP2) depends on both

329

SPECIAL TOPICS: ANHARMONIC OSCILLATOR SYSTEMS

action and angle variables and is a periodic function of the angles. When ,\ = 0
the action variables will be constants of the motion, and the angles CP1 and CP2
will change in time according to the equations
(6.125)
for (i

1,2) where
(6.126)

For the anharmonic case, the frequencies Wi(J1, h) will be continuous because
they depend on the action variables, even for two degrees of freedom. This
continuous dependence on the action variables is quite different from a
harmonic oscillator system where the frequencies, Wi, are constant.
For systems which satisfy the conditions of the KAM theorem (namely,
small ,\ and nonzero Hessian, det IEPHo/8Jj8Jjl '# 0 for (i,j) = 1 and 2), only a
very small region of hase s ace the resonance re ions will exhibit chaotic
behavior. The
Converted with
motion. If one
tries to const
of the motion
when ,\ '# 0 b
space, except
expansions for
the action vari
the form
H = Ho(J1,h)

STOI Converter

t of the phase
s, perturbation
nsion [37] for
amiltonian of

trial version

hDP://www.stdutilitv.com
+,\

L Vn),n2(J1,h)cos(n1CP1

n),n2

+ n2CP2)

(6.127)

and let us introduce the generator

F(f

1,

2,

CP1, CP2) = f 1CP1 + f

CP2 +

L Bn),n2sin(n1

n),n2

CP1 + n2CP2)

of a canonical transformation from variables Jt, CPi to variables


the variables fi are constants of the motion. Then

(6.128)

fi, q>i,such that

(6.129)
for i = 1, 2 and

of

q>j= afi

(6.130)

330

THE FOUNDATIONS OF STATISTICAL MECHANICS

for i = 1,2. If we substitute Eq. (6.129) into Eq. (6.127) and keep terms to
lowest order in ,\ (this requires a Taylor series expansion of Ho), we obtain

(6.131 )
To lowest order in '\,

/1 and /2 will
B

_
n),n2 -

be constants of motion if we choose


-,\Vn),n2

nlWI

+ n2W2)

(6.132)

Then
H = HO(/1 /2)

+ 0(,\2)

(6.133)

and
/1 =J
Note, however,
of it and h fc
perturbation ex

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com
nlwl

+ n2w21:s

,\Vn),n2'

(6.134)

here are values


zero, and the
~
(6.135)

the perturbation expansion will diverge and /i is not a well-behaved invariant.


This region of phase space is called the resonance zone and (nlwl + n2w2) = 0
is the resonance condition. It is in the resonance zones that one observes chaotic
behavior.
If the regions of phase space which contain resonances, and a small region
around each resonance, are excluded from the expansion for /1' then one can
have a well-behaved expression for /1. Thus, one can exclude regions which
satisfy the condition

For smooth potentials, Vn),n2 decreases rapidly for increasing nl and n2. Thus
for increasing nl and nz, ever smaller regions of the phase space are excluded.
Kolmogorov, Arnold, and Moser proved that as ,\ -+ 0 the amount of
excluded phase space approaches zero. The idea behind their proof is easily
seen in terms of a simple example [38]. Consider the unit line (a line of length
one). It contains an infinite number of rational fractions, but they form a set of

331

SPECIAL TOPICS: ANHARMONIC OSCILLATOR SYSTEMS

measure zero on the line. If we exclude a region

around each rational fraction, the total length of the unit line that is excluded is
n

00

~~

(2c)
-3 =2c
n

00

1 c,,(l
-=----40.
3

~n2

e--+O

Thus, for small '\, we can exclude the resonance regions in the expansion of "1
and still have a large part of the phase space in which "1 is well-defined and
invariant tori can exist.
Walker and Ford [37] give a simple exactly soluble example of the type of
distortion that a periodic potential can create in phase space. It is worth
repeating here. They consider a Hamiltonian of the type

Converted with

where

STOI Converter
For this model,
H=Eand

trial version

(6.137)
e total energy

hDP://www.stdutilitv.com
1= J1

+h.

(6.138)

Therefore, we do not expect to see any chaotic behavior for this system.
However, the unperturbed phase space will still be distorted when ,\ =1= O. The
frequencies Wi for this model are given by
aHO

WI

aJ = 1 -

=-

2lt - 3h

(6.139)

and
aHO

W2

= -a =
h

1 - 3lt

+ 2h.

(6.140)

If we want the frequencies to remain positive, we must choose 0 ~ J1 ~ -& and


E~
Let us rIot trajectories for the Walker-Ford case (q2 = 0,P2 > 0) (note that
qi = (21i) /2cOScPi and Pi = -(21i)1/2sincPi). We find that for ,\ = 0 the
trajectories trace out concentric circles in the PI, q1 plane. When the

o ~ h s -h and, therefore,

n.

THE FOUNDATIONS OF STATISTICAL MECHANICS

332

PI

Fig. 6.7. Cross-section of the energy surface for the Hamiltonian, H = JI +


3J1h + Ji + )Jlh cos (2>] - 2>2) = E. There is no chaotic behavior.
(Based on Ref. 37.)

h-

Jr -

perturbation is
we set cP2 = ~1l
(6.136), we ob

(3 + Ace
They are sketcl

Converted with

ly distorted. If
~.138) into Eq.
el curves:

STOI Converter
trial version

~~P:llwww.st~utili~.COm

(6.141)
'J

ghtly distorted

from the unperturbed case. However, there is a region which is highly distorted
and in which two elliptic fixed points (surrounded by orbits) and two hyperbolic
fixed points .appe~ [4]. The fixed points occur for values of Jt and cPi such that
Jl + Jz = (cPl - cP2) = O. If we use the fact that Ji = -8H/8cPi
and
1>i = 8H/8Ji and condition cP2 = 311"/2, we find that the hyperbolic orbits
occur when
(6.142)

while the elliptic orbits occur for


(6.143)

The

first-order resonance
condition for this model [cf. Eq. (6.135)] is
= 0 or, from Eqs. (6.139) and (6.140), Jl = 5h. Therefore, from
Eqs. (6.142) and (6.143) we see that the distorted region of phase space lies in
the resonance zone.

2Wl - 2W2

333

SPECIAL TOPICS: ANHARMONIC OSCILLATOR SYSTEMS

In general, for a Hamiltonian of the form


(6.144)
there will be no chaotic behavior because there is always an extra constant of
motion,
(6.145)
However, when the Hamiltonian is of the more general form given in Eq.
(6.127), the extra constant of motion is destroyed and the resonance zones
become more complicated and begin to overlap. When this occurs one begins to
see chaotic behavior.
Walker and Ford study the example

where an extra cosine term has been added to Eq. (6.136). For this model there
is no longer a
ary resonances
which grow as
Converted with
heir results. For
low energies tl
the resonance 2
overlap beeom
dots correspon

STOI Converter

v). However, as

p the regions of
In Fig. 6.8 the

trial version

hDP://www.stdutilitv.com
Pl

E=0.20000000

Pl

E =0.20950000

Fig. 6.S. Cross section of the energy surface for the Hamiltonian, H = I} + Jz
-If - 3I}I2 + Ii + >'lI}lz cos(2c/>}- 2c/>2) + )..2I}lz cos(2c/>}- 3c/>2) = E. (a) Phase
space, trajectories below the energy of primary resonance overlap. (b) Phase space
trajectories above the energy of primary resonance overlap. When primary resonances
overlap, large-scale chaos occurs in their neighborhood. (Based on Ref. 37.)

334

THE FOUNDATIONS OF STATISTICAL MECHANICS

Thus, from these simple examples we see that the chaotic, or ergodiclike,
behavior of phase space for the anharmonic oscillator system appears to be
caused by the overlapping of resonances. If the energy surface is filled with
resonance zones, as is often the case, then we expect chaotic behavior to set in
at very low energy.
Anharmonic oscillator systems are a rather special type of system and
their ergodicity has never been established, for obvious reasons. A completely
different type of system is a system of hard spheres. For systems of hard
spheres, ergodicity and mixing behavior have been established [39]. A
proof that systems with Lennard-Jones types of potential are ergodic has
never been given. However, when the number of degrees of freedom becomes
large, the "regular" regions of the phase space appear to become relatively
less important than the chaotic regions and statistical mechanics, which is
built on the assumption that ergodicity appears to work perfectly for those
systems.
The chaotic behavior illustrated in this section is indicative of unstable flow
in phase space. Orbits in the chaotic region which initially neighbor one another
move apart ex onentiall and rna move to com Ie I .
nt parts of the
energy surface.
egion of phase
space and assi
Converted with
ity distribution
:;:~~p:~:!

~~

exhibiting de
probability dis
energy surface.

STOI Converter
trial version

~ai;O:~~~~lt~~
ally localized
se, can fill the

hDP://www.stdutililV.com

.... S6.F. Newtonian Dynamics and Irreversibility [40, 41]


The instability and chaos that we have described in the baker map, Eq. (6.116),
and that we have illustrated in the Henon-Heiles system appears to be a source
of the irreversibility seen in nature. One of the great paradoxes of physics is the
fact that Newton's equations are reversible, but much of nature evolves in an
irreversible manner: Nature appears to have an "arrow of time." There is a new
field of statistical physics which finally is resolving this paradox [40-44]. The
resolution of the paradox is most easily seen in the spectral properties of chaotic
maps such as the baker map. Individual trajectories in chaotic systems move
apart exponentially and become impossible to compute even after a fairly short
time. However, in such systems, smooth initial probability distributions
generally relax to a smooth final distribution after some time. There are now
several "reversible" chaotic maps for which a spectral decomposition can be
obtained in terms of the decay rates and their associated eigenstates [28, 42,
43]. The decay rates are related to the Lyopounov exponents for the underlying
chaos, and determine the physically observable decay properties of such

335

REFERENCES

systems. The spectral theory of these systems can be formulated outside of


Hilbert space.
Considerable progress has also been made in understanding the emergence
of irreversible behavior in unstable Hamiltonian systems, at least for the case
when the dynamical phase space contains dense sets of resonances. For
such systems a spectral theory can also be formulated outside the Hilbert space
[45-46]. We don't have space to say more about this beautiful new area of
statistical physics, but the cited references should give interested readers a fairly
readable entrance to the field. Ref. 41 gives a historial overview.

REFERENCES
1. H. Goldstein, Classical Mechanics (Addison-Wesley,
Reading, MA, 1950).
2. R. L. Liboff, Introduction to the Theory of Kinetic Equations (John Wiley & Sons,
New York, 1969).

3. I. Prigogine, Nonequilibrium Statistical Mechanics (Wiley-Interscience,

New York,

1962).
4. L. E. Reichl

STOI Converter

(Prentice-H
6. I. E. Farquh
York, 1964).

tic Descriptions
terscience,

trial version

New

hDP:llwww.stdutililV.com

7. J. L. Lebow
8. v. I. Arnol
Benjamin,

stems: Quantum

Converted with

Manifestati
5. R. L. Libo

echanics (W. A.

an
vez,
New York, 1968).

9. I. E. Farquhar, in Irreversibility in the Many-Body Problem, edited by J. Biel and J.


Rae (Plenum Press, New York, 1972).
10. P. R. Halmos,
1955).

Lectures on Ergodic Theory (Chelsea Publishing

11. A. I. Khintchine,
Mathematical
Publications, New York, 1949).

Co., New York,

Foundations of Statistical Mechanics (Dover

12. D. S. Ornstein, Ergodic Theory, Randomness,


University Press, New Havan, CT, 1974).

and Dynamical Systems (Yale

13. O. Penrose, Foundations of Statistical Mechanics (Pergamon

Press, Oxford,

1970).

Physica 53, 98 (1971).


15. G. D. Birkhoff, Proc. Natl. Acad. (U.S.) 17, 656 (1931).
16. E. Wigner, Phys. Rev. 40, 749 (1932).
14. N. G. van Kampen,

17. W. N. Bogoliubov,

"Problems

of a Dynamical

Theory

in Statistical

Physics," in
(North-

Studies in Statistical Mechanics, edited by J. de Boer and G. E. Uhlenbeck


Holland,

Amsterdam,

1962).

18. M. Born and H. S. Green,


University Press, Cambridge,

A General Kinetic Theory of Liquids, Cambridge


1949).

336

THE FOUNDATIONS OF STATISTICAL MECHANICS

19. J. G. Kirkwood,

1. Chem. Phys. 14, 180 (errata 14, 347); 15, 72 (1947).

20. 1. Yvon, La Theorie Statistique des Fluides et l'Equations d'Etat (Hermann


Paris, 1935).
21. 1. H. Irving and R. W. Zwanzig,
22. J. Ross and Kirkwood,

et Cie,

1. Chem. Phys. 19, 1173 (1951).

1. Chem. Phys. 22, 1094 (1956).

23. J. E. Moyal, Proc. Cambridge Phil. Soc. 45, 99 (1949).


24. T. Takabayasi,

Progr. Theor. Phys. (Kyoto) 11, 341 (1954).

25. A. O. Barut, Phys. Rev. 108, 565 (1957).


26. H. Mori, Phys. Rev. 112, 1829 (1958).
27. A. Lasota and M. Mackey, Probabilistic Properties of Deterministic Systems (Cambridge University Press, Cambridge, 1985).
28. H. H. Haswgawa

and W. C. Saphir, Phys. Rev. 46,7401

(1992).

29. H. Wergeland in Irreversibility in the Many-Body Problem, edited by 1. Biel and 1.


Rae (Plenum Press, New York, 1972).
30. E. Fermi: Collected Papers, Vol. II (University
p.978.
31. A. N. Kolmogorov,
Benjamin,

in R. Abraham,

of Chicago

Press, Chicago,

1965),

Foundations of Mechanics, Appendix D (W. A.

Converted with

32. V. I. Arnol
33. J. Moser,

STOI Converter

34. M. Henon
35. G. H. Luns
36. G. Benettin
37. C. H. Walk

trial version

962).

8 (1976).

hDP://www.stdutilitv.com
,edited by E. O. D.

39. Ya. G. Sinai, in The Boltmann Equation, edited by E. G. D. Cohen and W. Thirring
(Springer- Verlag, Vienna, 1973).
40. I. Prigogine,
41.
42.
43.
44.
45.
46.

Int. 1. Quantum Chemistry 53, 105 (1995).


I. Prigogine, The End of Certainty (The Free Press, New York, 1997).
I. Antoniou and S. Tasaki, Int. 1. Quantum Chemistry 46, 425 (1993).
H. Hasagawa and D. J. Driebe, Phys. Rev. E50, 1781 (1994).
P. Gaspard, Chaos, 3, 427 (1993).
T. Petrosky and I. Prigogine, Proc. Natl. Acad. Sci USA 90, 9393 (1993).
T. Petrosky and I. Prigogine, Chaos, Solitons, and Fractals 4, 311 (1994).

PROBLEMS
Problem 6.1. Consider a system of N uncoupled harmonic oscillators with Hamiltonian,
H
12mj + kjq; 12). Assume that the system initially has a probability density
p(pN,qN,O) =
b(Pi -Pio)b(qj
-qiO). Compute the probability density p(pN,qN,t)
at time t, where pN = (PI, ... ,PN) and qN = (qt, ... ,qN).

= E~I (p;

n~I

337

PROBLEMS

Problem 6.2. Consider a particle which bounces vertically in a gravitational field, as


discussed in Exercise 6.1. Assume an initial probability distribution, p(P, z, 0) =
1j8(z)8(1.0
- p)8(P - 0.1)(8(x)
is the Heaviside function; 8(x) = 1 for x> 0 and
8(x) = 0 for x < 0). What is p'(J, (), O)? Sketch p(P, z, t) and p'(J, (), t) for t = 0.4, mass
m = 1, and gravitational acceleration g = 1.
Problem 6.3. Consider a particle with mass m = 1 moving in an infinite square well
potential, V(x) = 0 for -1 < x < 1 and V(x) = 00 otherwise. Assume that initially the
particle lies at x = -1 with momentum, P = Po for 0.1 :::;Po :::;1.0 in the positive x
direction. (a) Find the solution of the Liouville equation in action-angle space at time t.
(b) At what time does the initial distribution of points begin to break apart in (P,x)
space?
Problem 6.4. For a noninteracting gas of N particles in a cubic box of volume V = L3,
where L is the length of the side of box, find the solution, p(p3N, q3N, t), of the Liouville
equation at time t, where p3N = (PI' ... ,PN) and q3N = (qI, ... ,~)
with
Pi = (Pix,Piy,Piz) and qi = (qix, qiy, qiz) Assume periodic boundary conditions, and
assume that the probability density at time t = 0 is given by
3N N
p(p3N . q3N, 0) =

II II

7r

(T-)

e-PTa sin

for

0:::;qia

Converted with

Problem 6.S.
governed by a
energy. Assume
for p(p, q, t). (
involves elliptic

STOI Converter

Problem 6.6.

hDP://www.stdutilitv.com

s L.

ose dynamics is
ere E is the total
iouville equation
les. The solution

trial version

where, for example, HI,2

1,1

1,2

H2,I

H2,2

= (1IifI2). The density matrix at time


PI,I(O)
( P2,I(0)

PI,2(0))
P2,2(0)

= (1
0

t =

0 is

0)
0 .

(a) Find the density matrix


PI,I (t)
( P2,I(t)

PI,2(t))
P2,2(t)

at time t. (b) What is the probability to be in the state 11) at time t = O? At time t? For
simplicity, assume that Ii = 1.
Problem 6.7. An atom with spin 1 has a Hamiltonian if = AS2Z + B(82X - 82Y ), where
Sx, Sy, and S, are the x, y, and z components of the spin angular momentum operator. In
the basis of eigenstates of the operator, s; these three operators have the matrix
representations
A

s, = n (10

0 0)
0 0
0 -1

S,

Ii
Pi

y2

(0 1 0)
1 0 1
0 1 0

and

1i

(0

s, =- ( -1
iV2 \ 0

1 0)

-1 0

338

THE FOUNDATIONS OF STATISTICAL MECHANICS

(a) Write the density matrix (in the basis of eigenstates of Sz) at time t = 0 for two
different cases: (i) The atom is initially in an eigenstate of s, with eigenvalue +1i; (ii)
the atom is initially in an eigenstate o~ with eigenvalue +1i. (b) Compute the density
matrix (in the basis of eigenstates of Sz) at time t for each of the two cases in (a). (c)
Compute the average z component of spin at time t for the two cases in (a).

s,

Problem 6.8. Consider a harmonic oscillator with Hamiltonian iI = (112m )jJ2 +


~mw2 _xl. Assume that at time t = 0 the oscillator is in an eigenstate of the momentum
operator, p(O) = IPo)(Pol. (a) Write the Liouville equation in the momentum basis. (b)
Compute the density matrix (p'lp(t))!P),
at time t.
Problem S6.1. Locate all period-3 points of the Baker map in the (p, q) plane.

Converted with

STDI Converter
trial version

hDP://www.stdutilitv.com

PART THREE
EQUILIBRIUM STATISTICAL
MECHANICS

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com

7
EQUILIBRIUM STATISTICAL
MECHANICS

7.A. INTRODUCTION
Most systems in nature, if they are isolated (i.e., do not exchange energy or
matter with
ependent state
(thermodyna
Converted with
ay be described
in terms of on

STOI Converter

consistent wi

trial version

:~~~:~~~,
i~:

roscopic states
have no way of

hDP:llwww.stdutililV.com

distinguishing
have a way of
assigning pro
seen in Chapter
6, for the special case of ergodic systems, we can assign a probability to various
microscopic states based on the mechanical properties of the system. If a system
is ergodic, it is equally probable to find the state of the system in different
regions of the energy surface if those regions are equal in size. This leads
naturally to the following probability distribution on the energy surface:

(7.1)
otherwise,
where 2:(E) is the area of the energy surface (the structure function). This
choice of probability density for an isolated system forms the foundation upon
which statistical mechanics is built.
Once the distribution function is given, it is a straightforward matter to
compute the expectation values of various quantities, such as energy,
magnetization, and so on. However, there is one quantity which still eludes
us, and that is the entropy. We know that the entropy must be additive and
positive and must have a maximum value at equilibrium, An entropy of the
form S = f dXN [p(XN)fnCN,
where n is some positive integer and CN is a
constant, is one type that could be used. However, the form that we shall use in

342

EQUILIBRIUM STATISTICAL MECHANICS

this book is the Gibbs entropy


(7.2)
where ke = 1.38 X 10-23 J/K is the Boltzmann constant. The constant, CN, is
inserted to give the correct units, but also has important physical meaning. It is
determined from quantum mechanics. The form of entropy in Eq. (7.2) was
chosen by Gibbs [1] because it gives the correct expression for the temperature
of systems which are closed but not isolated. For quantum systems, the Gibbs
entropy can be written
S = -kBTr[jJ

In(jJ)],

(7.3)

where jJ is the density operator and the trace is taken over any complete
orthonormal set of basis states.
As we have seen in Chapter 2, we must have the ability to describe the
behavior of systems under a variety of external constraints. Closed isolated
systems have fi
systems have
fixed particle n
age energy is
specified. Open
ergy. A closed

STOU Converter

isolated system
icrocanonical
ensemble) of th
y densities for
closed and op
trial version
al ensembles,
respectively)
probability
density for a s
ermodynamic
quantities is straig orwar.
In this chapter we have selected a variety of fairly simple examples to
illustrate how the equilibrium probability densities can be applied to real
systems. We have selected the examples for their historical and practical
importance or because they illustrate important concepts that we shall need
later. The models and systems we will study in this chapter are all exactly
soluble. In subsequent chapters, we will introduce a variety of approximation
schemes for systems which cannot be solved exactly.
This chapter is divided more or less into three sections. We begin by deriving
the thermodynamic properties of two closed isolated systems (fixed energy and
particle number), namely, an ideal gas and an Einstein solid using the
microcanonical ensemble. Following a method due to Einstein, we can also use
the microcanonical ensemble to derive a probability distribution for fluctuations
of thermodynamic quantities about the equilibrium state. The Einstein method
is based on a Taylor series expansion of the entropy about absolute equilibrium,
and therefore it breaks down near the critical point where fluctuations can
become very large. However, it still gives us valuable insight concerning the
behavior of fluctuations near a critical point. In this chapter we will apply
Einstein fluctuation theory to fluid systems.
We next consider closed systems in which the temperature and particle

hnp://www.stdutilitv.com

343

THE MICROCANONICAL ENSEMBLE

number is fixed but the energy can fluctuate. Such systems are described by the
canonical ensemble. As examples of closed systems, we consider in some detail
the effect of lattice vibrations on the thermodynamic properties of solids. We
also compute the thermodynamic properties of an Ising spin lattice. We can
obtain exact expressions for the thermodynamic properties of one-dimensional
spin lattices, and we use a mean field model to obtain approximate expressions
for the thermodynamic properties of higher-dimensional spin lattices. Mean
field theory predicts that a phase transition from a disordered state to an ordered
state occurs on the spin lattice.
Finally we consider open systems in which both the energy and particle
number can fluctuate but the temperature and chemical potential are fixed. Such
systems are described by the grand canonical ensemble. The grand canonical
ensemble is especially suitable for describing systems in which phase transitions
occur which break gauge symmetry because particle number need not be conserved. We will use the grand canonical ensemble to compute the thermodynamic
properties of ideal quantum gases, both Bose-Einstein and Fermi-Dirac.
An ideal Bose-Einstein gas is composed of identical bosons and at very low
temperatures c n
articles do not
interact) in w
e into a single
momentum st
n ideal Fermi-

STOI Converter

Dirac gas on
but, because of
the Pauli excl
own. No two
fermions can
trial version
refore, at low
temperature th
tum states and
even at T = 0
Thus a Fermi
gas, even at T
For an ideal Fermi gas, there is no condensation into a single momentum
state. However, if we allow an attraction between fermions, then they can form
bound pairs which can condense in momentum space. This is what happens to
electrons in a superconductor. In a superconducting solid, electrons interact
with lattice phonons and with one another through a phonon-mediated
interaction which is attractive in the neighborhood of the Fermi surface. The
fermion pairs condense in momentum space and act coherently, thus giving rise
to the unusual superconducting properties observed in such systems.

hDP://www.stdutilitv.com

~==-rI~----n:rnr--n-!:J""(nS"""""!:r"lr!l1"'lrnr-1"H"l'rt!1!'IrT1"i5-----_J

7.B. THE MICRO CANONICAL ENSEMBLE [1-3]


For an ergodic mechanical system with fixed particle number (closed) and fixed
energy (isolated), all states on the energy surface are equally probable. This fact
forms the basis upon which equilibrium statistical mechanics is built and is the
origin of the microcanonical ensemble. As we shall now show, this distribution
extremizes the Gibbs entropy and therefore allows us to make contact with
thermodynamics. Let us first consider a quantum system. As we saw in Chapter
6, the equilibrium probability density, in order to be a stationary state, must be a

344

EQUILIBRIUM STATISTICAL MECHANICS

function of the Hamiltonian, p = p(H). Let IE, n) denote a set of states of


energy E with respect to which the density operator is diagonal. The integer, n,
takes values n = 1, ... ,N(E), where N(E)is the total number of states with
energy E. The probability to find the system in state IE, n) is P; = (E, nlplE, n),
and the entropy can be written
N(E)

S = =ko Tr[ p In(p)] = -kB

L r; In{P

n).

(7.4)

n=l

We must determine the set of probabilities, {P n}, which extremize the entropy
subject to the constraint, Tr(p) = I:~~~)
P; = 1. The simplest way to do this is
to use Lagrange multipliers. Since we have one constraint, we need one
Lagrange multiplier, which we call Qo. We then require the following variation
to be zero:
(7.5)
Since the variat

Converted with

STOI Converter
The Lagrange

Tr{p) = I:~~~)

trial version

hDP://www.stdutilitv.com
1
P; = N{E)"

In{Pn) = 0 or
(7.6)
ion condition,
p by
(7.7)

Thus, the probability distribution which extremizes the Gibbs entropy is the one
for which all states of the same energy are equally probable. This is called the
microcanonical ensemble. If we substitute Eq. (7.7) into (7.4), we find that the
entropy is given by
S = ks In{N{E)).

(7.8)

The entropy is proportional to the logarithm of the number of states with energy
E.

If we are given a thermodynamic system with a fixed mechanical energy E


and if we know the number of microscopic states with energy E, then we can
use Eq. (7.8) to compute the thermodynamic properties of the system. The
internal energy of the system is just U=E. Therefore, the temperature can be
found from the thermodynamic relation (8S18E)N x = liT. The generalized
force is Y = {8E/8X)N,S. The chemical potential is J-L' = (8E/8N)s,x.
As an
illustration, in Exercise 7.1, we compute the thermodynamic properties of an
Einstein solid.

345

THE MICROCANONICAL ENSEMBLE

EXERCISE 7.1. An Einstein solid consists of a lattice in threedimensional space with N lattice sites. Each lattice site contains three
harmonic oscillators (one for each direction in space), each of which has
frequency w. Neither the lattice sites nor the harmonic oscillators are coupled
to one another. The total energy of the lattice is H = Iu 2: nj+
~Nhio = E, where n, is the number of quanta of energy on the ith harmonic
oscillator. The total number of quanta on the lattice is fixed at M = 2: n..
Assume that M and N are very large. (a) What is the total number of
microscopic states with energy E? (b) Compute the entropy as a function of
temperature, T, and number of lattice sites, N. (c) Compute the heat capacity
of this solid.

i~l
i~l

Answer: Because the harmonic oscillators are spatially separated on


different lattice sites and at a given lattice site are associated with different
spatial directions, they are distinguishable. The quanta are indistinguishable.
(a) The number of ways to assign M identical quanta to 3N distinguish-

able oscillators is

Converted with

1)!

)I'

STOI Converter

(1)

An eas
~N - 1 black
dots m'
trial version
t quanta. The
3N - 1
3N "pots"
(harmo
~l number of
different states is simply the number of different permutations of the
black and white dots.
(b) The entropy is

hDP://www.stdutilitv.com

(3N + M - I)!)
S = ks In ( M! (3N _ I)! .

(2)

For M and N large we can simplify this since by Stirlings formula,


In(N!) ~ N In(N) - N. Using this formula we find

S = ke In

M)M+3N)
(( 1 + 3N

(3)

(M/3N)M
Now note that l/T

=
(8S)
8M

(8S/8E)N
=
N

liw
T

(1/liw)(8S/8M)N

= kB 1n(3N

M'

1)

and
(4)

346

EQUILIBRIUM STATISTICAL MECHANICS

Solving for M, we find

M =
where

f3

= l/kBT.

S=

3N
e(3/iw - 1 '

(5)

The entropy is

3Nk

In( 1 - e-f3Jiw) +

3N~~:';:'~1iw

(6)

The internal energy is

3Nnwe-(3/iw

= (1 _ e(3/iw) + 2N tu.

(7)

The heat capacity is

It is useful

3Nli2w2
= kBT2

e -(3hw

(8)

(1 _ e-(3/iw)2 .

Converted with

STOI Converter

ystem. Let us

lume, V, and a
energy shell,
E ~ E + t:t.E.
trial version
st the energy
surface becaus
ysics. We can
make t:t.E as
ergy shell is
nLl.E(E, V, N) =
"
,were
"
IS
e structure function,
Eq. (6.41) (the area of the energy surface). To obtain the equilibrium probability
density we must find an extremum of the Gibbs entropy subject to the
normalization condition
consider a clos
fixed number

hDP:llwww.stdutililV.com

(7.9)

where the integration is restricted to the energy shell. We again use the method
of Lagrange multipliers and write

[J

dXN (QOp(XN) - kBP(XN) In[CNP(XN)])]


(7.10)

E<H(E}<E+Ll.E

dXN(QO - ke - ke In[CNP(XN)])8p(XN)

= O.

E<H(E) <E+Ll.E

Since 8p(XN) is arbitrary, we must have p(XN) = C;;leao/ks-1 = constant.


The Lagrange multiplier, Qo, is determined from the normalization condition,

347

THE MICROCANONICAL ENSEMBLE

Eq. (7.9). We find

p(XN)

= {

fl<)'E(~ V, N)

for E

< H(E) < E

+ Dt.E,

(7.11)

otherwise.
If we substitute

Eq. (7.11) into Eq. (7.2) for the entropy, we find


S = kn

In(fl<l.E(~~V,N}

(7.12)

Comparison with Eq. (7.8) shows that n6.E(E, V,N)/CN


is just the number of
states in the energy shell, E ~ E + Dt.E. The constant, CN, can be determined
from quantum mechanics. The volume of a single quantum state in a 6Ndimensional phase space is h3N, where h 6.63 x 10-34 JS is Planck's constant.
Therefore, the constant, CN, is chosen as follows. For distinguishable particles,
eN = h3N so nD.E/ CN is just the number of quantum states confined to the
energy surface. For indistinguishable particles, CN = Nlh3N The factor of N! is
included to av .
r
_1.
..1.
.1
~ indistinguish_:J

able. It is pu

~::~a1:h;~~

pt be justified

trial version

of a system of
ce, n(E, V, N),

SIDU Converter

In practice,
energy E, it is (

trzz:

Converted with

hDP://www.stdutililV.com

we shall show

;rg;u~~ellili~~

IV

N
= constant), both quantities give the same entropy. To see this, let us
divide the volume enclosed by the surface of energy E into E I Dt.E shells each of
width Dt.E. These shells will range in energy from 0 to E. Shells with low energy
will have a smaller volume than shells with higher energy. The total volume can
be written as a sum of the volumes of all the shells.

n(E, V, N) =

EjD.E
nD.E(Ei, V, N).

:E

(7.13)

i=1

The shell with the largest volume will be that with energy E ~ E
we can write

E
nfj.E(E, V,N) ~n(E, V,N) ~ Dt.EnD.E(E, V,N).
If we take the logarithm

+ Dt.E. Thus,
(7.14)

of each term in Eq. (7.14), we find

In(fl<l.E) .;;; In(fl) .;;; In(fl<)'E) - In

(~E).

(7.15)

348

EQUILIBRIUM STATISTICAL MECHANICS

But In(O) rv Nand


In(E/ flE) rv In(N). Therefore, in the limit N ~ 00,
In(O) = In(O~E)' because we can neglect terms of order In (N) relative to
terms of order N. Thus, in the thermodynamic limit the entropy can be written
(7.16)
where CN = h3N for distinguishable particles and CN = N! h3N for indistinguishable particles. In Exercise 7.2, we compute the thermodynamic
properties of an ideal classical gas using Eq. (7.16) and the phase space
volume, O(E, V, N), computed in Exercise 6.2.
EXERCISE 7.2. Use the microcanonical ensemble to find the entropy
and equation of state of an ideal gas of N identical particles of mass m
confined to a box of volume V. Assume that Nand Vare very large.
Answer: From Exercise 6.2, the volume of phase space with energy less than
E is

Converted with

STOI Converter

(1)

trial version

hDP://www.stdutilitv.com

(2)

Let us note that for large N, we have In(N!) ~ N In(N) - Nand


r(3N/2 + 1) = (3N/2)! ~ (3N/2e)3N/2. We can rewrite the entropy in the
form
3NkB
S=-2-+NkBln
5N ke
=-2-+NkB

47rmE 3/2]
[ V ( 3h2N )

[v (47rmE)
In N 3h2N

+NkB-NkBln(N)
(3)

3/2]

In the middle term we have explicitly separated off the contribution


(+N ke - N ke In(N)) coming from the Gibbs counting factor. We see that
without it, the entropy is not a homogeneous function. In the last term, the
Gibbs counting factor is combined into the rest of the terms, and we see that
the entropy is just the expected Sackur- Tetrode entropy and is a
homogeneous function of the extensive variables.
If we remember that the internal energy, U, equals E, then Eq. (3) is the
fundamental equation for the entropy. The pressure is given by

349

EINSTEIN FLUCTUATION THEORY

P = -(8U/8V)S,N.
If we take the derivative of Eq. (3) with respect to V
holding S and N fixed, we obtain an equation for P = -(8U /8V)S,N which
we can solve to give P = 2U /3V. Next note that T = (8U /8S)v N. If we
take the derivative of U with respect to S holding Vand N fixed, we obtain an
equation for (8U/8S)V,N which we can solve to yield U = ~NkBT. If we
combine these two results, we obtain the ideal gas equation of state
PV = NkBT.

7.C. EINSTEIN

FLUCTUATION

THEORY

[4,5]

There is a very simple method due to Einstein [5] for obtaining the probability
distribution for fluctuations about the equilibrium state. This theory uses the
microcanonical ensemble. We first derive the general formalism and then apply
it to thermodynamic systems

7.C.I. Gener---___.y-".
......
.___------------------,

Converted with

Let us consider
E~

E+D..E.

microscopic st
number of mic
the system is g

STOU Converter
trial ve rsion

hnp://www.stdutilitv.com
T

'/J.

~y in the range
~at all possible
E) denote the
the entropy of
(7.17)

Often it is useful to have a more detailed microscopic description of the system


than that given by Eq. (7.17). For example, we might want to subdivide a
system into cells and to specify its state according the thermodynamic behavior
of each cell (each cell would be an open system). We can do this if we assume
that, in addition to the energy, the macroscopic state of the system is describable
in terms of n independent macroscopically measurable parameters (state
variables) Ai(i = 1,2, ... ,n), where Ai could denote quantities such as energy,
pressure, temperature, and so on of the ith cell.
Let I'( E, A I, ,An) denote the number of microstates with energy E and
parameters AI,
,An. Then the probability that the system is in a macroscopic
state described by parameters E, A I , ... ,An, is given by
A ) _ r(E, Al ... ,An)

P(E A
,

1 , .

r (E)

(7.18)

The entropy of the system in a state with parameters (E, AI, ... ,An) is given by
S(E, AI, ... ,An)

= ke

In[r(E, AI, ... ,An)].

(7.19)

350

EQUILIBRIUM STATISTICAL MECHANICS

Hence
(7.20)
The entropy will be a maximum when the system is in an equilibrium sate,

A?, ... ,A~. Any fluctuation about the equilibrium state must cause the entropy
to decrease. If we let a, denote the fluctuations
(7.21)
then we can expand the entropy about its equilibrium value to obtain
1

S(E, AI, ... ,An) = S(E, A?, ... ,A~) -

2L

i=I

j=I

gij ai aj

+ ... ,

(7.22)

where

Converted with

(7.23)

STOI Converter

Einstein fiuctua
trial version
uations about
the equilibrium
expansion of
the entropy at
near a phase
transition wher
gij is posmve
definite since the entropy must decrease and is symmetric since the quantities,
Ai, are state variables. (A matrix is positive definite if every principal minor is
positive.) Equation (7.22) contains no terms which are first order in Aj to ensure
that spontaneous fluctuations about the equilibrium do not cause a decrease in
the entropy. Equilibrium is a state of maximum entropy.
We can now substitute Eq. (7.22) into Eq. (7.20) and obtain the following
expression for the probability distribution of fluctuations about the equilibrium
state (cf. Exercise 4.9):

hDP://www.stdutilitv.com

P(a.) =
(7.24)

=
The vector a.T is the row vector a.T = (aI, ,... ,an) and is the transpose of the
column vector a.. The quantity gij is the ijth element of the n x n matrix, g. The
quantity det Ig I is the determinant of the matrix g. The probability P( a.) is

351

EINSTEIN FLUCTUATION THEORY

normalized to one:
Drl.P{rr.) =' ['"

['"

do;

X .

x ['"

do..P{rr.)

=1

(7.25)

Since only small fluctuations are assumed to be probable, there is no difficulty


in extending the limits of integration in Eq. (7.25) from -00 to +00.
We often want to find expectation values of various moments of the
fluctuations. To do this, it is convenient to introduce a more general integral,

J(h)

==

det

I g I Joo

(27rkB

-00

D7i exp [__ 1_ a.T

g. a. + hT . a.] = e!kBhT.g-t.h

2kB

'

(7.26)
where I(h) is a characteristic function (cf. Exercise 4.9). The moment, (Qi Qj),
is defined as

Converted with
Since the prot
the first mome
the moments

STDI Converter
trial version

(7.27)
~ith zero mean,
ssed in terms of

hDP://www.stdutilitv.com
7.e.2. Fluid Systems
Let us consider a closed isolated box with volume, VT, which contains a fluid
with total entropy, ST, total energy, ET, and total number of particles, NT, and
total mass, MT. We shall assume that the box is divided into 1/ cells of equal size
and that at equilibrium the volume, entropy, internal energy, and mass of
particles in the ith cell are Vi = Vo, Si, Ui, and Mi, respectively (cf, Fig. 7.1).
The total volume of the box is then VT = 1/Vo. The equilibrium pressure,
temperature, and chemical potential of the whole system are denoted Po, To, and
fio, respectively.
From Eq. (2.171), the entropy change due to deviations in various
thermodynamic quantities from their equilibrium state will be given by

(7.28)
If we now use the expressions obtained in Section 7.C.l, we can write an
expression for the probability of a given set of fluctuations. To obtain an explicit
expression, we must choose a particular set of independent variables.

352

EQUILIBRIUM STATISTICAL MECHANICS

ET, VT, MT, Po, To, {to


I

---+----I---L-I
I

I
I

I
I

---r--T---f--I
I Vi, s, I
I
, Ui,MiI
---T--,---I--,
I
Fig. 7.1. A system with fixed total energy, ET, total volume, VT, total mass, MT,
equilibrium pressure, Po, equilibrium temperature, To, and equilibrium chemical
potential, flo. To describe local fluctuations about absolute equilibrium, we divide the
system into cells. Cell i has volume, entropy, internal energy, and mass given by
Vi = VO,Si, Vi, a

Converted with

STOI Converter
trial version

Schematically,

hDP://www.stdutilitv.com
P(flST)

[flTjflSj

-(2___;_k--')3-=-11
exp --2k-];-.
~

- flPjflVj

+ fljLjflMd].

1=1

(7.29)
where the matrix g is determined once the independent variables are decided
upon.
Let us first pull the volume dependence out of Eq. (7.28). We will write
S = Vs and M = V p, where s is entropy/volume and p is mass/volume. Let us
next note that the Gibbs-Duhem equation, when written in terms of mass
density, p, takes the form pdjL = -sdT + dP. Therefore, flT flS - flPfl V +
fljLflM = V(flTfls + fljLflp). The differential of the Helmholtz free energy/
volume, a, satisfies the equation da = -sdT + [uip, so (8iL/8T)p = -(8s/8ph
If we now choose temperature, T, and mass density, p, as independent variables,
we find
(7.30)
where cp is the equilibrium specific heat. The row vector,

(17, can be written

353

EINSTEIN FLUCTUATION THEORY

~= (Y ~, ... , Y 1/, p~, ... , PI/) and the matrix, g, has matrix elements

if i = j and 1 ~ i ~ v;

T2

gij

r(~)

if i = j and v

8p

+ 1~

(7.31 )

i ~ 2v;

otherwise.

The probability of a fluctuation,


therefore

~T

= (D,.T1, ... , D,.T 1/, D,.P1, ... , D,.PI/) , is

(~r(~(~)Tr

P( {D,.Ti' D,.Pi}) =

(27rkB)21/
X exp

- -1 ~
~
2kB i=1

[vc

-2

(D,.Ti)

2 + -V (8[1,)
T

Bp

(D,.Pi)

2]]

(7.32)

Converted with
We can now
fluctuations in

STOI Converter

ations between

trial version

hDP://www.stdutilitv.com
A

l_AI'

-rl)

_Or}

_.Af{

~,

(7.33)
(7.34)

(7.35)
and
(7.36)
Thus, we have found that fluctuations in temperature and density are
statistically independent. One can also show that pressure and entropy
fluctuations are statistically independent. However, fluctuations in most other
pairs of thermodynamic variables are not independent.
It is important to note that in Eq. (7.33) we found no correlation between
various cells. This result was built into the theory because Eq. (7.33) contains
no information about coupling between cells. In real systems, there is coupling
between cells. This can be included in the theory by expressing /),.Si and /)"Pi in

354

EQUILIBRIUM STATISTICAL MECHANICS

terms of temperature and volume variations in other cells and not just those of
cell i. The more general expression will then contain coupling constants which
reflect the strength of the coupling between the cells.
In this section, we have analyzed fluid systems by dividing them into discrete
cells. This, of course, is a rather artificial way to proceed, but it is conceptually
very simple and gives us good intuition about what thermodynamic quantities
govern the behavior of fluctuations about the equilibrium state. It is a simple
matter to change the summations over discrete cells to integrations over
continuously
varying densities, provided that the spatial variations have
sufficiently long wavelengths (vary slow enough). We shall look at the spatial
dependence of fluctuations in later chapters.

7.D. THE CANONICAL ENSEMBLE


A closed system can exchange heat with its surroundings and as a consequence
will have a fluctuating total energy. We therefore need to find a probability
density which
systems. This
probability dis
Section 7.D.1
we derive the
m system, and
in Sections 7.
properties for
distinguishable

STOI Converter

ermodynamic
r systems of

trial version

In order to obtain the probability density operator for a closed system we must
extremize the Gibbs entropy subject to two constraints. We require that the
probability density operator, p, be normalized.
(7.37)
and we require that the average energy be fixed to some value (E),
(7.38)

If we introduce the Lagrange multipliers, ao and aE, we can find the probability
density operator, p, which extremizes the entropy subject to the constraints,
Eqs. (7.37) and (7.38). The extremization condition is

+ aEHNP - kBP In(p)]}


= TrN[{(ao - kB)i + aEHN - kB In(p)}8p] =

8{TrN[aop

where

1is

(7.39)
0,

the unit operator. Since 8p is arbitrary, we must have

(7.40)

355

THE CANONICAL ENSEMBLE

and therefore
(7.41)
The two Lagrange multipliers, ao and aE, can be determined from the
constraints Eqs. (7.37) and (7.38). From Eqs. (7.37) and (7.41) we find
(7.42)
The quantity ZN(T) is called the partition function. We next determine the
Lagrange multiplier, aE. Let us multiply Eq. (7.40) by P and take the trace. We
obtain
TrN[(ao - kB)p

+ aEHNP

In(p)] = -kB In[ZN(aE)]

- kBP

+ aE(E) + S = 0,
(7.43)

where we hav
definitions of
From Section
be written A we can make t
energy is

Converted with

STOI Converter
trial version

(7.38), and the


tropy Eq. (7.3).
free energy can
s U = (E), then
Helmholtz free

hDP://www.stdutilitv.com
A = -kBT

In[ZN(T)).

(7.44)

The partition function, ZN (T), takes the form


(7.45)
where

!3 =

1/ kBT. The probability density operator can be written

= e-f3(HrA)

e-f3HN

,
TrN(e-f3HN)

(7.46)

Equation (7.46) is the probability density for the canonical ensemble. The trace
is evaluated using any convenient complete orthonormal set of N-particle basis
states. However, as we shall show below, we must make a distinction between
systems of N distinguishable particles and systems of N indistinguishable
particles.
It is important to note that Eq. (7.46) was obtained by extremizing the Gibbs
entropy. Historically, the Gibbs entropy was chosen because it gave this result
[1, 2] for the probability density of a system at temperature T. At high

356

EQUILIBRIUM STATISTICAL MECHANICS

temperature, where equipartition of energy occurs, Eq. (7.46) gives the correct
relation between the temperature and the average energy.
Equation (7.44) is the Fundamental Equation for a closed system. From it we
can obtain all thermodynamic quantities. For example, the entropy is given by
S = -(8A/8T)x N The generalized force is given by Y = (8A/8X)r N. The
chemical potential is given by /-l' = (8A/8N)r X. Another useful relation for the
internal energy is U = (8{3A/8{3)x N.
'
In the canonical ensemble the' temperature, T, is fixed and the average
energy, U = (E), is fixed. However, because there can be a flow of energy in
and out of this system, it is important to know how large the fluctuations in the
energy about the average, (E) , will be. Let us therefore compute the variance of
the energy fluctuations. We first write the normalization condition
(7.47)
If we differentiate Eq. (7.47) twice with respect to {3, we find

Converted with
This gives

STOI Converter
trial version

hDP://www.stdutilitv.com

(7.49)

X,N

where CX,N is the heat capacity at constant X and N. The heat capacity, CX,N,
and average energy, (E), are each proportional to N. Therefore, the fractiona1
deviation behaves as

(7.50)
and goes to zero as the number of degrees of freedom becomes infinite. This
means that the fluctuations in energy become very small relative to the
magnitude of the energy itself. In the thermodynamic limit, most microstates
will have an energy approximately equal to the average energy, U = (E), and
the canonical ensemble becomes equivalent to the microcanonical ensemble.
In evaluating the trace in Eq. (7.45) we must distinguish between systems of
indistinguishable particles and systems of distinguishable particles. A system of
N indistinguishable particles, by definition, has the property that the
Hamiltonian and all other physical observable remain invariant under permutation of the particles. We will consider both cases below.

357

THE CANONICAL ENSEMBLE

7.D.2. Systems of Indistinguishable Particles


As we show in Appendix B, for systems of identical particles we can evaluate
the trace either in terms of complete sets of symmetrized or antisymmetrized Nbody position or momentum eigenstates, or in terms of the "number"
representation. However, the fact that in the canonical ensemble the number of
particles is restricted to N makes the particle number representation unwieldy
when using the canonical ensemble. As we have shown in Appendix B [Eqs.
(B.30) and (B.32)], the trace of the probability density for a set of N identical
particles may be written
(7.51)
where the states Ik1, .. , kN) denote "momentum" eigenstates for the N-body
system. The symmetrized state Ik1, ... , kN) (+) must be used for bosons. The
antisymmetrized state Ik1, .. , kN) (-) must be used for fermions. The
"momentum" eigenstates may be written as the direct product of the
"momentum"

Converted with

(7.52)

STOI Converter

he "momenta"
o right from 1
"momentum"
to N. Thus, p
kb, ... , particl
. ,
omentum" we
include spin and any other internal degrees of freedom of a particle that
might affect the thermodynamic properties of the system.) The state Ik~) 1 is
the "momentum" eigenstate for particle 1. The state Ikl, ... , kN) l) is
defined as
On the left-har
of the particle

trial version

hDP://www.stdutilitv.com

.J

Ik1, .. ,kN)()

2)1(Plk1,

,kN),

(7.53)

where L:p P denotes the sum over all permutations of the momentum in
the state Ikl, ... , kN) (cf. Appendix B). In Exercise 7.3 we evaluate the
partition function for the (unrealistic) case of a three-body system of identical
particles .

EXERCISE 7.3. Compute the partition function, Z3 (T), for an ideal gas
of three identical particles (bosons or fermions) in a cubic box of volume
V = L3. Assume the walls of the box are infinitely hard. For symplicity,
neglect any spin or other internal degrees of freedom. What approximations
can be made for a high-temperature and low-density gas?

358

EQUILffiRIUM STATISTICAL MECHANICS

Answer: The Hamiltonian for _!histhree-body ideal gas is ih = PI/2m+


where Pi = liki is the momentum operator for the ith
particle and k; is the wave vector of the ith particle.
It is instructive first to work out the partition function for a single particle,
Z1 (T) = Trdexp( _,Bp2 /2m)] = Lk(klexp( -,Bli2k2 /2m)lk), where [k) is
orthonormal eigenstate of the wavevector operator, k = pili and
(k I k') = 8k,k" Since the particle is confined to a cubic box of volume
V = L3 and the walls of the box are infinitely hard, the kinetic energy
eigenstates can be written

pV2m + pV2m,

where nx, ny, and nz are integers 1,2, ... ,00, and A is a normalization
constant. The allowed wavevectors are
(2)
where i, j, an
The single pa
Z

respectively.

Converted with

STOI Converter

(3)

trial version

hDP://www.stdutilitv.com

If the volume
h change the
summation to an integration in Eq. (3). Let us note that
=
and
make a change of variables, Px = nxTililL. For very large L, liTi/L ~ 0 and
we find

::3) tOO dpx r

ZI (T) = (,..
= (;)

dp;

dp, exp [-

-00

-00

ItOdn-,

!(p; + p; + p;)]

JOO dp, JOO dp; JOO dpz exp [-,:


-00

L~=1

(p;

+ p; + p;)]

=~ ,
T

(4)
where
(5)
is the thermal wavelength and has units of length. As we shall show below,
the thermal wavelength is a measure of the distance over which the particles
maintain phase coherence in the gas.

359

THE CANONICAL ENSEMBLE

The three-body partition function can be written

Z3(T) = Tr3(e-PH) = ;,

(kl' k2, k3, Iplk1, k2, k3) (),

(6)

. k"k2,k3

where
Ikl' k2, k3) () = Ikl' k2, k3) Ikl' k3, k2) Ik2' kl' k3)

(7)

Ik3, k2, k1) + Ik3, kl, k2) + Ik2' k3, kj ).


If we
I

note

8ka,k.J8kb,k~8kc,kf

the orthonormality
condition,
then Eq. (6) becomes

Z3(T) = ;, [(Z[(T))3 3Z[(T)Z[

(ka, kb' k,

I~, Ice, kf)

m m]
+ 2Z[

(8)

Converted with

STOU Converter

trial ve rsion
and/or low
The semicla
densities (lar
can neglect
terms propo'~1IIUl~~T'1"T--r---.::nn'----'~-.::ro:::T~r--o;;]~""""""TDE"'I>~ical
partition
function for a gas of three identical particles:

hUP:llwww.stdutililV.com

(9)
For a gas of N identical particles, the semiclassical
approximate partition function:

ZN(T)~,
!:

1
N.

(V)N
3"
AT

limit gives an

(10)

The factor N! in Eq. (10) is exactly the counting factor used in the
It resolves the
Gibbs paradox.

! microcanonical ensemble for indistinguishable particles.

As we have shown in Exercise 7.3, symmetrized and anti symmetrized states


must be used when computing the partition function for systems of identical
particles at low temperatures and/or high densities where quantum effects

360

EQUILIBRIUM STATISTICAL MECHANICS

strongly influence the translational degrees of freedom. However, for


moderately high temperatures and/or moderate densities the partition function
takes the approximate form
(7.54)

where we now neglect the contributions from permuted states. In doing so, we
are neglecting terms of order, (N IV)A~, where AT = hi y'27rmkBT is the
thermal wavelength. Equation (7.54) gives a good description of a semiclassical
gas of N identical particles. Quantum effects still must be taken into account in
computing any contributions to the partition function due to internal degrees of
freedom. To see this, let us consider a gas of noninteracting molecules which
have internal degrees of freedom which can absorb energy. Such internal
degrees of freedom might include the rotational and vibrational motions of the
molecule and electronic or nuclear excitations. For such a system, the
Hamiltonian of t~e ith molecule c~n be writte~ Hi = I~m + Hi (rot) +
Hi(Vib) + Hi(elec)
Hi(nuc) denote
the rotational,
Converted with
es of freedom,
respectively. In
assumed that
these various in
another. For a
gas of N uncou
unction can be
written
trial version

fir

STOU Converter

hnp://www.stdutilitv.com

ZN(V,

1 ~-.--~~~~~~--~~--~~--~r
) =N!

i(nuc)

+. ..)

])

(7.55)
The partition function takes a particularly simple form for this system
Hamiltonians in Eq. (7.55) commute with one another. Then we find
ZN(T,

V)

1
= N!
(ZI(tr)

ZI(rot)

Zlvib ZI(elec) ZI(nuc)

x ...

)N

if the

(7.56)

where ZI (tr) = Tr, (e-/3p2 12m), ZI (rot) = Tr, (e-/3HI (rot)), and so on. The trace, Tr,
is taken over a complete set of single particle states appropriate for the
Hamiltonian appearing under it.
For a semiclassical ideal gas with no internal degrees of freedom, the
partition function is (cf. Exercise 7.3)
ZN(T, V) = ~! (~)

N~ (~~~)

N,

(7.57)

where we have made use of Stirling's approximation, N! ~ (N let. The Helmho]tz

361

THE CANONICAL ENSEMBLE

free energy is

= -kBTln(ZN)

= -NkBT

- NkBTln

2 ) -3 2]
vN ( 27rmkBT
h

(7.58)

The entropy is

S=-

(aA)
aT

5
"2NkB +NkB

V,N =

[vN ( 27rmkBT
h -3/2] '
2

In

(7.59)

which is just the Sackur- Tetrode equation. In Exercise 7.4 we compute the
thermodynamic properties for a semiclassical gas of N identical spin-~ particles
in a magnetic field.

EXERCIS.I.L__:__t_..lJ....__4_.,.._._"........._""O""':';-.........L

............."""---"-............,"'-'---'----.........._

V = L3, conta

s =~, and m
system. (a) C
internal energ

STOI Converter

.....................,,1 of volume,
ich has spin,
plied to the
ompute the
iation?

trial version
Answer:

hDP://www.stdutilitv.com

(a) The partition function takes the form


(1)
where

is the translational contribution for a single molecule and


is the magnetic contribution for a single molecule. The single
particle translational partition function can be written ZI(tr) = VIA},
where AT = hi J27rmkBT is the thermal wavelength.
Let us now compute the partition function due to the magnetic
degrees of freedom. Each atom will have magnetic energy
E(s) = -~s/lB, where s = l. The magnetic partition function for
a single atom is
ZI(tr)

Zl(mag)

ZI(mag)

L e-

(J/-LB/2

2COSh(f3~B).

(2)

s=1

The partition function for the gas is


1

ZN = N!

2V N
N
f3/lB )
A~ ) cosh ( -2-'

(3)

1
I

362

EQUILffiRIUM STATISTICAL MECHANICS

(b) The internal energy is given by


U

3
= -NkBT
2

= - (8InZN)
--8f3

I
- -N/-lBtanh
2

(f3/-lB)
-2

(4)

The heat capacity is

CV,N=

8T
(8U)

3
V,N,B='2NkB+NkB

B
(f3/-lB)
2
2 (f3/-l )
-2- sech -2- .

(5)

-(8ip/8Bh

(c) The magnetization

is given by M =
N,where ip is the
free energy for this problem (the translational part is like a Helmholtz
free energy, a function of T 1 V, and N, and the magnetic part is like a
Gibbs free energy, a function of T, B, and N). The free energy of the
combined system doesn't have a name so we call it ip. Then
4> = -KBTln (ZN) and

M=-

8ip

(6)

Converted with
7.D.3. Syste

STOI Converter
trial version

For systems of
using a compl
"momentum"
eigenstates

hDP://www.stdutilitv.com

n be evaluated
ple, if we use

to evaluate the trace, we have

TrN(p) =

E (k

11 ...

,kNlplkl,'"

,kN)'

(7.60)

kl, ..,kN

One of the best examples of such a system is a semiclassical solid composed of


N lattice sites. Even if the atoms at each lattice site are identical they do not
have translational kinetic energy. They are located at different points in space
and they can be distinguished by their position in space. An Einstein solid is an
example of a semiclassical solid in which the atoms can be considered to be
distinguishable because they remain at well-defined lattice sites. In Exercise 7.5
we compute the thermodynamic
properties of an Einstein solid using the
canonical ensemble.

EXERCISE 7.5 Use the canonical

internal energy,
Exercise 7.1.

and heat

capacity

ensemble to compute the entropy,


of the Einstein solid described in

363

THE CANONICAL ENSEMBLE

Answer: The Hamiltonian for an Einstein solid with N lattice sites and 3N
harmonic oscillators with frequency w is

(1)
where nj is the number operator for energy quanta in the ith harmonic
oscillator. The number operators have eigenvalues which are integers,
" n, = 0, 1,2, ... ,00, and eigenvectors Inj), so that njlnj) = njlnj). The partiI' tion function is given by
I

II

ZN(T) = TrN

[exp( -,6nw ~(iii

+!)) l

(2)

We can evaluate the trace in Eq. (2) using the basis states Inl' n2, ... ,
n3N) = Inl) x ... x In3N). Then
oo~~~----------~----~----~~

ZN(T)

, ... ,n3N)

Converted with

STOI Converter
(3)
(Note that the number of particles is fixed at N. The number operator, nj
counts the number of vibrational quanta in the ith harmonic oscillator. Even
though the number of particles is fixed at N, the number of vibrational
quanta has no restriction.)
The Helmholtz free energy is

The entropy is

(5)
The internal energy is
8f3A)
U = ( 8f3

N=

2: Nnw + 3Nnw

e-(3nw
(1 _ e-(3nw) .

(6)

364

EQUILIBRIUM STATISTICAL MECHANICS

The heat capacity is


(7)

7.E. HEAT CAPACITY OF A DEBYE SOLID [6, 7]


The intermolecular cohesive forces which bind the simplest molecular crystals
are the same as the forces which give the correction terms in the van der Waals
equation of state (Section 2.C.3). The short-range repulsive forces between
atoms make crystals virtually incompressible. The longer-range attractive force
holds the atoms of the crystal in place at some fixed lattice spacing, although
the atoms can undergo oscillations about their lattice positions. To first
approximation, a crystal may be viewed as an ordered array of atoms, each one
fixed to a latti
.
.
.
ite. While the
potential whic
st approximaConverted with
tion, in derivin
In this section
we shall com
harmonicall y
coupled lattice
first proposed
by Debye and
7.A we obtain
trial version
an exact expr
ns for a oneresult for that
dimensional la
case.)
Let us consider the contribution to the heat capacity of the sound modes in a
simple three-dimensional harmonic lattice. Consider a rectangular lattice with
sides of lengths Lx, Ly, and Lz. We will assume that all sides are pinned so that
the sound waves form standing waves with wavelengths 2Lx/lx,2Ly/ly,
and
2Lz/lz in the x,y, and z directions, respectively, where lx, Iy, and Iz are integers.
The ath sound mode will have a dispersion relation of the form

STOI Converter
hDP:llwww.stdutililV.com

(7.61 )

The Hamiltonian operator takes the form


(7.62)

where no: = alao: is the number operator for energy quanta (phonons) in the O'th
normal mode. Let Ino:) be the eigenvector of the number operator, no:. Then

365

HEAT CAPACITY OF A DEBYE SOLID

nalna} = nalna). The partition function can be written

(7.63)

The average energy is

m,

Since the average can also be written (E) = (~=~=l1iwa(na +


comparison
with Eq. (7.64) shows that the average number of quanta in the ath phonon
mode is

Converted with

(7.65)

STOI Converter

which is Plane
The allowe
trial version
ints in a threedimensional fr
x direction is
7re/Lx, in the
7rc/Lz. The
volume per poi
u
re 7rC
,w re V = LxLyLz is
the volume of the crystal. The number points per unit volume in frequency
space is V/ (7rc)3. The total number of allowed values of Wa less than some
value W is given by (47rw3 /3)(V /(7rc)\
where (47rw3/3) is the volume of
of a sphere of radius W (the phonon frequency is positive). Thus, the number of
allowed frequencies in the range W ~ W + dw is given by

hDP:llwww.stdutililV.comis

dl

= -27rC
2 3w

dw.

(7.66)

In general, there will be two transverse sound modes and one longitudinal sound
mode since crystals can sustain both longitudinal and transverse sound modes (a
fluid can only sustain longitudinal sound modes). The transverse and
longitudinal sound modes in a crystal propagate at different velocities which
we shall denote as c, and C[, respectively. If the three different sound modes are
taken into account the number of allowed frequencies, dn, in the interval,
w --t W + dw, is
(7.67)

366

EQUILIBRIUM STATISTICAL MECHANICS

Since there is a minimum wavelength allowed on the lattice due to the finite
spacing of the atoms, there will be a cutoff frequency, WD (the Debye
frequency). If there are N atoms on the three-dimensional
crystal, there will be
3N sound modes. We can then determine the Debye frequency by relating it to
the total number of modes,

3N=

3N

dn=- V

27T'2

If we solve for

wb, we

(2-+-1) 1
ct3

3
clOt

WD

Jdw=-

V
27T'2

find that the Debye frequency

(2-+-1)
c3

c3I

J2..
3

(7.68)

is given by

(7.69)
The density of states can be written
0('''\ _

dn _

vw2 (2

J_

1_, _ 9Nw2

(7.70)

Converted with
Once we are

STOI Converter
trial version

We then find

rgy is given by
(7.71 )

hDP://www.stdutilitv.com
(E) = 9NliwD + 9N
8
w1

fWD

liw3dw .
e{jnw - 1

(7.72)

The heat capacity is

(7.73)
where the Debye temperature, TD, is defined TD = liwD/kB.ln the limit T ~ 0,
the heat capacity becomes approximately
(7.74)

These results are in good qualitative agreement with the experimentally


observed heat capacity of many solids. We give two examples in Fig. 7.2
where we have plotted the prediction of Debye theory and have compared

367

HEAT CAPACITY OF A DEBYE SOLID

3R

1.0

0.6
0.2

TD

Fig. 7.2. The solid line is the heat capacity curve predicted by Debye theory. The
dashed line is that predicted for Einstein solid. The circles and triangles are
experimental values for the heat capacity of Al (TD = 390 K) and Cu (TD = 315 K),
respectively. (Based on Ref. 8.)

it to experime
Debye temper
high tempera
3NkB = 3nR,
However, as th
as T3.

Converted with

STOI Converter
trial version

hDP:llwww.stdutililV.com

e appropriate
e finds that at
t with eN ~
gas constant.
ty goes to zero

ehavior for the


heat capacity 0
1
onically coupled
lattice with N atoms has a heat capacity, eN = 3N kB' which is independent of
temperature. The first theory to give the correct qualitative behavior of solids at
all but very low temperatures was due to Einstein. He assumed that the solid
was composed of N uncoupled harmonic oscillators. The thermodynamic
properties of the Einstein solid have been computed in Exercises 7.1 and 7.5.
The heat capacity of an Einstein solid goes to zero as T ---t 0 K and therefore
gives a vast improvement over classical theory. It accounts for the fact that only
the longest-wavelength phonons can be excited at very low temperature,
whereas classical theory allows them all to be excited. However, Einstein theory
gives a vastly oversimplified density of states. It assumes that all phonon modes
have a single frequency and therefore that the density of states is
g(w) = 3N8(w - wo), where Wo is the frequency of all the phonon modes. In
real solids, the phonon modes have a distribution of frequencies, and this is
more accurately taken into account by the Debye (continuum) theory. The heat
capacity for the Einstein solid goes to zero exponentially with temperature and
therefore does not give the correct low-temperature behavior for a solid. In Fig.
7.3 we compare the density of states for the Einstein solid and the Debye solid.
The actual density of states for some solids has been measured using neutron
scattering techniques. Neutrons interact with the nuclei of atoms in the solid

368

EQUILIBRIUM STATISTICAL MECHANICS

3N

Einstein

~--~------------~------~----~~
1
WD

Fig. 7.3. Plots of the density of states of Einstein and Debye solids.

0.5

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com

(1013 rad

5-1)

Fig. 7.4. The density of states for aluminum, obtained from neutron scattering
experiments. Reprinted, by permission, from R. Stedman, et al., Phys. Rev. 162, 549
(1967).

and are scattered by phonons, which are simply the normal modes of the atomic
oscillations. In Fig. 7.4 we show the density of states for aluminum. The lowfrequency contribution looks somewhat like the Debye result, but at high
frequencies the Debye density of states is completely inadequate. It is
interesting to note in Fig. 7.2 that the heat capacity of aluminum at high
temperatures is not sensitive to the density of states. This is a result of the
equipartition principle. At high temperatures the average energy of an oscillator
is ke T and is independent of frequency of the oscillator.

369

ORDER-DISORDER TRANSITIONS

7.F. ORDER-DISORDER

TRANSITIONS

An important class of systems that can be studied using methods of equilibrium


statistical mechanics are those which exhibit a transition from a disordered to an
ordered state. One of the simplest examples of such a system consists of a
lattice of N spin-! objects with magnetic moment u. Each lattice site can
interact with the magnetic fields of their nearest neighbors and with any
external applied magnetic fields that might be present. The Hamiltonian for
such a system can be written
N

H=

LE;jS;Sj

-/-lBL

{;j}

S;,

(7.75)

;=1

where 2:;. denotes the sum over nearest-neighbor pairs ij (one must be careful
to count ~ach pair only once), E;j is the magnetic interaction energy between
nearest neighbors i and j, S; is the z component of spin at the ith lattice site, and
B is the external magnetic field. For spin-! obects, S; = + 1 -1 if the spin of
site i is oriented
kinetic energy
Converted with
in Eq. (7.75). 1'.
in orientation
or B = 0 the
~~~c~h~i~f~!~
ve spin up or
all the lattice
trial version
probable. If
B =f. 0, then the
with spin up
will be energeti
the configuration in which 'In;;:q;:;nruu:rrn:~prIlr.:r--.;II'V"-unrvII'~C"'Uf7pU~~T-u:rn:;--unother
will be
favored.
The partition function for this spin lattice can be written

STOI Converter
hDP:llwww.stdutililV.com

ZN(T) =

L
all config.

exp

[-13L

EijS;Sj+ 13/-lB
{ij}

s;] ,

(7.76)

;=1

where ~allconfig. denotes the sum over all 2N possible different configurations of
spin on the lattice. The partition function introduces an additional influence,
that of thermal energy, kBT. While the magnetic interaction energy, H, will
cause the spins on the lattice to become ordered, the thermal energy, kBT, will
tend to randomize the spins on the lattice. It is these two competing influences
which lead to an order-disorder phase transition on the spin lattice. At low
temperature, the lattice will be ordered. As we raise the temperature, at some
point the order dissappears and the spins become randomly oriented. The
system, described by the partition function, Eq. (7.76), is called the Ising model
and was originally used by Ising [10] as a model for ferromagnetism. However,
the model also has been used to describe the behavior of lattice gases, binary
alloys "melting" of DNA, and so on.

370

EQUILIBRIUM STATISTICAL MECHANICS

The Ising model can be solved exactly, and analytic expressions for its
thermodynamic properties can be found in one and two dimensions, but no one
has ever succeeded in solving it analytically in three dimensions. In one
dimension it does not exhibit a phase transition at finite temperature, but in two
dimensions it does. In one dimension the lattice does not have enough nearest
neighbors for the ordering effects of the interaction energy to compete
effectively with the disordering thermal energy. However, for two or more
spatial dimensions it does. The Ising model was first solved in two dimensions
by Onsager [11-13]. It is one of the few exactly soluble models which exhibit a
phase transition.
In this section we shall consider two simple versions of the Ising model. We
will first obtain exact expressions for the thermodynamic properties of a onedimensional Ising lattice. The one-dimensional lattice does not exhibit a phase
transition at finite temperature. However, the method of solution contains some
ideas which are used to solve the much more difficult case of a two-dimensional
lattice. We will solve the two-dimensional case in Chapter 8. In this section we
will also compute the thermodynamic properties of an Ising lattice with d
dimensions in t
.
.
roximation of
the d-dimensio
Converted with
mperature.

7.F.t. Exact

STOU Converter

Let us consider
trial version
evenly spaced,
est neighbors
have the same i
on with lowest
energy is one
odicity of the
lattice is imposed by assuming that Sj+N = si. The total energy for a given
configuration, {Sj}, is

hnp://www.stdutilitv.com

E{sj} =

-f

LSjSi+l
;=1

-I-lB LS;'

(7.77)

;=1

N-1

3
N-2

Fig. 7.5. A one-dimensional periodic lattice with N


lattice sites.

371

ORDER-DISORDER TRANSITIONS

The partition function can be written


ZN(T,B)

= S~I

S~I

exp

[,6

(fSiSi+1

+ ~JLB(SI +

SHI) ],

(7.78)

where we have used the fact that 2::1 Sj = ~2::1 (S; + SHt) for a periodic
lattice. It is now convenient to introduce a 2 x 2 matrix,
(7.79)
whose matrix elements are defined as
(7.80)
The partition function may then be written

STOI Converter
trial version

(7.81)

hDP://www.stdutilitv.com

e convention
where A are
A+ > A_. The eiIrenvamIf:s-orr'"""al'e-easIllVlUUJrRrlrnt:re---____j

In the limit N ~ 00, only the largest eigenvalue, A+, contributes to the
thermodynamic quantities. This is easily seen if we note that the Gibbs free
energy per site is
g(T, B)

= N-+oo
lim N_!_GN(T,

B)

= -kBT

lim N_!_In[ZN(T, B)]


N-+oo

(7.83)

= -kBTln[A+].

In Eq. (7.83), we have used the fact that limN-+oo(A_/A+)N


Gibbs free energy per site is

g( T, B) = -f - ke In [COSh(,6JLB) +

= O. Thus the

cosh2(,6JLB) - 2e-2i1< sinh(2j3f) ].


(7.84)

372

EQUILIBRIUM STATISTICAL MECHANICS

The order parameter is given by


(7.85)

From Eq. (7.85) we see that the one-dimensional Ising model cannot exhibit a
phase transition because when B -+ 0 the order parameter also goes to zero.
Hence, no spontaneous nonzero value of the order parameter is possible. The
exact solution to the two-dimensional Ising model is givin in Chapter 8, and it
does allow a phase transition.

7.F.2. Mean Field Theory for a d-Dimensional Lattice


We can obtain approximate analytic expressions for the thermodynamic
properties of a d-dimensional Ising lattice using the mean field approximation first introduced by Weiss [14]. In the mean field approximation, the
Hamiltonian of
can be written

Converted with

STOI Converter

(7.86)

trial version
where E is the c
hbor spins, and
E( E, B) = IH(j
ount the same
pair of spins twice. The quantity (s) = (s;) is the average spin per site. The
quantity VE(S)S; is an average magnetic interaction energy between site i and its
nearest neighbours, assuming that the neigbors all have spin (s). As we shall
show below, the average spin per site, (s) , must be determined in a selfconsistent manner.
The partition function can be written

hDP:llwww.stdutililV.com

ZN

(L

efJESi)

= (2cosh({3E))N.

(7.87)

= -kBTln[2cosh({3E)].

(7.88)

sl=l

The Gibbs free energy per site is

g(E,B)

= -kBT lim

N-+oo

(2. (ZN))
N

ln

The probability P(s;) that site i has spin

ESj=l

is
efJEsl

efJESj

P(Si)

S;

efJEsl

= 2 cosh({3E) .

(7.89)

373

ORDER-DISORDER TRANSITIONS

Note that the probability P(s;)has an especially simple form in the mean field
approximation because coupling to nearest neighbours is taken into account in
such a simple manner.
The average magnetization of the lattice is given by

(M) = Np,(s) ,

(7.90)

where

L:

se(3Es
(s) = ii=1
I (3Esj = tanh[,BE] = tanh
Sj=l e

(1

,B 2"lJE(S)

+ p,B

)] .

(7.91 )

The magnetization is the order parameter for the spin lattice. If B = 0, the
magnetization will be zero for the high-temperature paramagnetic phase of the
lattice (randomly ordered spins) and it will be nonzero at lower temperatures
where the spins have spontaneously aligned.
We can determine the critical temperature, Tc, at which the lattice starts
to become oru

.
.. .,....
int) from the
expression for
Converted with
~ for the case
B=O;

STOI Converter
(7.92)

trial version

hDP:llwww.stdutililV.com

We must solve
by plotting
ly
/1 == (s) versus (s) and h == tanh(a(s)) versus (s) on the same graph. The
solution to Eq. (7.92) is given by those points where the two curves crossthat is, where /1 = h. In Figs. 7.6a and 7.6b, we plot /1 and h versus (s)
for a < 1 and a> 1, respectivel y. For the case a < 1, there is only one
crossing point and it is at (s) = O. For a> 1, there are three crossing
points, at (s) = 0 and at (s) = sQ. The free energy per site for these various
cases is
if(s) = 0,
if(s) = sQ.

(7.93)

Thus, the values, (s) = sQ [when they are solutions to Eq. (7.92)] describe
possible states of thermodynamic equilibrium since they minimize the free
energy. The transition point (critical point) occurs at a = 1 in Fig. 7.6 and
therefore when lJE/2kBT = 1. Thus the critical temperature in the mean field
approximation is]' = T; = lJE/2kB. In Fig. 7.7, we plot the order parameter (s),
versus T ITc.
We see that mean field theory predicts a phase transition at a finite
temperature for a d-dimensional lattice. This does not agree with our exact

374

EQUILmRIUM STATISTICAL MECHANICS

It ={s}

(a)
(8)
0'<1
h=(s}

(b)

!2 = tanh(a{s})
L----

Converted with
Fig. 7.6. Plots
0>1.

STOI Converter
trial version

(a) 0

< 1. (b)

hDP://www.stdutilitv.com
1-1----

(s)

O+------------~~

TITe

Fig. 7.7. A plot of the order parameter, (s),


versus the reduced temperature, TITc.

result for the case d = I (cf. Section 7.F.1) where we found no phase transition
at finite temperature. As we shall see in Chapter 8, mean field theory gives too
high an estimate of the critical temperature for spatial dimensions, d ~ 3. It
gives good estimates for d ~ 4 which is not of physical interest but is of
mathematical interest.

375

ORDER-DISORDER TRANSITIONS

Let us next examine the behavior of the heat capacity in the neighborhood of
the critical temperature. The internal energy for B = 0 is
U

= ----

aZN

ZN aj3

= --NVE(S)
2

(7.94)

The heat capacity is


(7.95)
But

so
(7.97)

Converted with
The heat capac

STOI Converter
trial version

CN=

(7.98)

hDP://www.stdutilitv.com
~----~~~~~~~~~~~~~

In Fig. 7.8 we plot the heat capacity as a function of temperature. We see that
the heat capacity has a finite jump at the critical temperature. It reaches a
maximum value of 3NkB at T = Te.

3NkB

O+---~----------~-----o

t;

Fig. 7.S. A plot of the heat capacity, Cu, as a function of temperature in the
neighborhood of the critical point.

376

EQUILIBRIUM STATISTICAL MECHANICS

The final quantity we wish to compute is the magnetic susceptibility,


The magnetic susceptibility is defined as

XT,N(B).

XT,N(B)

8(M))
( 8B

T,N

= NJ-L(8(S))
8B

T,N

(7.99)

From Eq. (7.91) we can write

(7.100)
or

8(S))
8B

T,N = 2cosh2

[(3 ~E

The magnetic susceptibility, XT,N(B),

2(3J-L
(s) + (3J-LB] - (3vE .

(7.101)

is then given by

Converted with

(7.102)

STOI Converter
The magnetic s

trial version

hDP://www.stdutilitv.com
XT,N(O) = 2cosh2 [liE]
(32: (s) - (3VE

ve cosh 2())(s Tc/T

- Tc/T

,
(7.103)

Fig. 7.9. A plot of the magnetic susceptibility. X(O)T,N'


the neighborhood of the critical point.

as a function of temperature in

377

THE GRAND CANONICAL ENSEMBLE

with (s) given by Eq. (7.92). Examination of Eq. (7.103) shows that the
magnetic susceptibility has an infinite jump at the critical point. In Fig. 7.9, we
plot XT,N(O) as a function of temperature.
The results we have obtained in this section for the thermodynamic
properties of a spin lattice in the neighborhood of the phase transition are
qualitatively similar to the results we obtained in Chapter 3 using mean field
theories. Using statistical mechanics to obtain these results allows us to
compute various constants that appear in the thermodynamic
expressions
in terms of microscopic
interaction energies and magnetic moments. A
generalization of mean field methods has been used by Suzuki to study critical
phenomena [15]. As we shall see in Chapter 8, where we give some exact
results for the two-dimensional
spin lattice and use renormalization
theory,
mean field theory gives a rough qualitative picture of the phase transition, but it
is not quantitatively correct.

7.G. THE GRAND CANONICAL ENSEMBLE


An open syste
therefore, both
equilibrium, w
number, (N). ~
environment, w
entropy subjec
normalization t

Converted with

STDI Converter
trial version

undings and,
pen system in
rage particle
a fluctuating
ze the Gibbs
uire that the

hDP:llwww.stdutililV.com
Tr(p) = 1.

We require that the average energy have a fixed value,

(7.104)

(E), so that

Tr(Hp) = (E).

(7.105)

And finally, we require that the average particle number have a fixed value, (N),
so that

Tr(Np) = (N),

(7.106)

where N is the total particle number operator.


We can now find the probability density operator, p, which extremizes the
Gibbs entropy subject to the constraints, Eqs. (7.104)-(7.106).
The extremization condition is

+ QEHp + QNNp - kBpln{p)]}


= Tr[{(Qo - kB)l + QEil + aNN - kB In {p)}8p] = 0,

8{Tr[Qop

(7.107)

378

EQUILffiRIUM STATISTICAL MECHANICS

where

ao, aE,

and

aN

8p is arbitrary,

are Lagrange multipliers. Since

we have
(7.108)

We can now use Eq. (7.108) and the three constraints, Eqs. (7.104)-(7.106) to
determine the Lagrange multipliers. The normalization condition, Eq. (7.104),
allows us to introduce the grand partition function,
(7.109)
and relates ao to aE and aN. To determine
(7.108) by P and take the trace. We find

aE

and

aN,

let us multiply Eq.

(7.110)
If we compare Eq. (7.110) to the fundamental equation for the grand potential,
n = u - TS - ,.
identifications,
aE = -1/T an

Converted with

STOI Converter
Equation (7.11
written in term

trial version

(7.111)
otential. It is
d as

hDP://www.stdutilitv.com
(7.112)
with

f3

= l/kBT.

The probability density operator can be written


A

p-e

-(3(iI-Ji/N-n)

e-(3(iI-Ji/N)

".
Tr( e-(3(H-Ji/ N))

(7.113)

Equation (7.113) is the probability density operator for the grand canonical
ensemble. Equation (7.112) is the Fundamental Equation for an open system.
While we have not written it explicitly, the grand potential will generally
depend on a generalized displacement, X, whose form depends on the
mechanical properties of the system being considered. For example, for a gas,
X= V is the volume and for a magnetic system, X = M is the magnetization.
Once we know the grand partition function, we can compute the grand
potential, and from it we can obtain all thermodynamic quantities. The entropy
is given by S = -(8n/8T)x
I.The generalized force is given by Y:::::
(8n/8X)r"i/' The average parti~le number is given by (N) = -(8n/8J-t'h,x
In exercise 7.6 we use the grand canonical ensemble to compute the
thermodynamic properties of a gas of photons.

379

THE GRAND CANONICAL ENSEMBLE

EXERCISE 7.6 A cubic box of volume V = L3 contains electromagnetic


energy (photons) in equilibrium with the walls at temperature T (black-body
variation). The allowed photon energies are determined by the standing
waves formed by the electromagnetic field in the box. The photon energies
are nwj = hck., where kj is the wavevector of the ith standing wave. Compute
the pressure of this photon gas. (Note that since photons have no mass, the
gas has chemical potential, J-L' = 0.)

Answer: Since there are in infinite number of standing waves allowed, the
grand partition function is

Zp,(T)

00

=~

00

... n~o

00

exp -(3 ~ njnwj

Doo(

1_

e-{31iw;

(1)

The grand potential is


00

n = -PV

= kBT

L In(1 -

e-{31iw;).

Converted with

We can chang
standing wav
These corresp

STDI Converter
trial version

(2)
ng way. The
+ (nz7r/L)k.

(3)

hDP:llwww.stdutililV.com
If we imagine a atnce

0 a owe
requencies, W = c, e spacing per point
in W space is cnfl: (see the figure below). The volume per point is (C7r/L)3.
The number of points per unit volume is (L/C7r)3 .

The number,

lJ

of allowed frequencies less than some value w is therefore

lJ=w"JK~r=~:~2.
I

(4)

For each allowed frequency of the electromagnetic field, there are two

380

EQUILIBRIUM STATISTICAL MECHANICS

transverse modes. Thus,


therefore becomes

L::l

= 2 du =

(L3 /~c3)

Iooo uldw.

The pressure

(5)
If we integrate by parts, we find
P=--

n Joo

3C37r2

w3dw

e-i31iw

7r2k4T4 1
=_B_=_a~
45c3li3

(6)

'

where a = 7r2kU15c3li3 is the Stefan-Boltzmann constant.


In the grand canonical ensemble, the temperature, T, and chemical potential,

J-L',are fixed and the average energy, U = (E), and average particle number,
(N), are fixed. However, because there can be a flow of energy and particles in
and out of this system, it is important to know how large the fluctuations are.
The derivation of the variance in energy fluctuations is similar to the derivation
given in Section 7.D.l. Let us derive the variance in oarticle number
fluctuations. V\J

If we different

Converted with

brm
(7.114)

STOI Converter

obtain

trial version

(7.115)

hDP://www.stdutilitv.com
Thus, the fractional deviation behaves as

(7.116)
As the average number of particles increases, the size of the fluctuations in
particle number comes small compared to the size of the average particle
number. Most microstates will have a particle number approximately equal to
the average number and we retrieve the canonical ensemble.
It is useful to write Eq. (7.115) in terms of the isothermal compressibility.
FromEq. (2.6), (8N/8{l')r,v
= -(8N/8V)T,/1/
(8V/8J-L'h,N' FromEq. (2.102),
(8J-L'/8V)r,N = -(8P/8Nh,v'
From Eq. (2.6), (8P/8Nh,v
= -(8P/8V)T,N
(8V / 8N)r,p. If we now combine these equations and note that (8N / 8V)r"t' =
(8N/8Vh,p
= (N)/V since J-L'= J-L'(T,P), we find

= -k T (N)2 (8(V))
(N2) _ (N)2 = k BT(8(N))
8'J-L r x
B V2
8P

= kBT(N)2
T,N

r;,

(7.117)

381

IDEAL QUANTUM GASES

Thus, the variance in particle number fluctuations is proportional to the


isothermal compressibility. Near a critical point, the compressibility can
become infinite and, therefore, fluctuations in the particle number (or density)
become very large.

7.H. IDEAL QUANTUM GASES


When a collection of identical particles forms a gas, the particles must be
regarded as indistinguishable. Therefore, at low temperatures, where the
thermal wavelength of particles becomes large and overlapping of wave
functions occurs, the statistics of the particles plays a crucial role in
determining the thermodynamic behavior of the gas. All particles which are
known today obey either Bose-Einstein or Fermi-Dirac statistics. At very low
temperatures, Bose-Einstein and Fermi-Dirac gases behave in completely
different ways because their different statistics introduce different kinds of
correlations between particles even though they do not interact via Newtonian
forces. At hi h tern eratures the statistics la no role and all ideal gases,
regardless of
Converted with
n easiest to use
When wor

STOI Converter

the grand co
to conserve
particle numb
ber of particles
in the system
trial version
. This becomes
especially us
dergo a phase
transition to a
gas of identical
particles each 'ur"TIIl:ll)"l)~:-mlC");naJ[JU"""UaJrutirun~[JR;lamT"t;cnrUC"--written

DP:llwww.stdutililV.com

(7.118)
where flo is the kinetic energy operator and total particle number operator N is
the number operator.
For an ideal gas, it is most convenient to evaluate the trace in the number
representation since both the energy operator and the number operator are
diagonal in that representation (cf. Appendix B). We shall assume that our gas
is contained in a "box" with periodic boundary conditions. We will let the
volume of the "box" be V = LxLyLz, where Lx, Ly, and Lz are the lengths of the
sides of the box in the x, y, and z directions, respectively. The momentum
operator for a single particle can be written p = lik, where k is the wavevector of the particle. For a "box" with periodic boundary conditions, the
momentum operator has eigenvalues

p,=

27r'ix

27r'iy

27r'i'l.

Ii ( -I+-J+t;
t;
i,

where lx, ly, and lz are integers each ranging from

k)

-00

(7.119)

'
to

00.

Let us denote the

382

EQUILffiRIUM STATISTICAL MECHANICS

set of integers, I = (lx, ly, lz). Each different choice of integers lx, ly, t,
corresponds to a possible state of the system. The kinetic energy operator for
a single particle is jJ2/2m and has eigenvalues
(7.120)
Let ni == nlx,ly,lz denote the number of particles with momentum PI. Then the
kinetic energy operator can be written Ho = Ll n'l, and the number operator
can be written No = Ll n,. The summation, Ll' ranges over all possible values
of the integers lx, ly, and lz.
The difference between Bose-Einstein and Fermi-Dirac particles lies in the
numbers of particles that can occupy a given momentum (and therefore kinetic
energy) eigenstate. For a gas of identical Bose-Einstein particles there is no
restriction on the number of particles that can have a given set of quantum
numbers, lx, ly, and lz. Therefore, the trace in Eq. (7.118) must include the
possibility that any numbr of particles, nl, ranging from 0 to 00, can occupy the
momentum stat,
m state. Thus,
for a Bose-Eim
Converted with

-t_

STDU Converter

(7.121 )

trial ve rsion
In Eq. (7.121),

http://www.stdutiliJV.com

citly as

(7.122)
In Eq. (7.121) we have not explicitly included the possibility that the particles
have (integer) spin or other internal degrees of freedom. However, it is
straightforward to include them.
For a gas of identical Fermi-Dirac particles, again, each different set of
quantum numbers, lx, ly, and lz, corresponds to a state of the system. However,
for Fermi-Dirac particles the Pauli exclusion principle restricts the number of
particles, nJ, which can occupy a given state, I, to nl = 0 or nl = 1. Thus, for a
Fermi-Dirac gas the grand partition function becomes
(7.123)

In writing the partition function for the Fermi-Dirac gas, we have not explicitly

383

IDEAL QUANTUM GASES

included the (half-integer) spin or other internal degrees of the freedom of the
particles. We will do this below in Section 7.H.2, where we describe the
behavior of Fermi-Dirac gases in more detail.

7.H.1. Bose-Einstein Gases


The grand partition function for a Bose-Einstein ideal gas is given by
Eq. (7.121). It implicitly assumes that the bosons have spin s = 0, so spin
degrees of freedom do not have to be taken into account. If we expand the
exponential into a product of exponentials, each depending on a different value
of I, then we can rearrange terms in Eq. (7.121) and write the grand partition
function in the form
ZBE(T , V , r://) =

IT (~e-f3nl(CI-Il))
L...J
I

~=o

IT( 1 - e-f3(cl-{L')
1
) .

(7.124)

From Eq. (8.B.52), the grand potential can be written

Converted with

STOI Converter
The average n

(7.125)

trial version

hDP://www.stdutilitv.com
(7.126)

where (nl) is the average number of particles in the state 1 and is defined
(7.127)
The quantity z = ef3{L' is called the fugacity.
The ideal Bose-Einstein gas is particularly interesting because it can
undergo a phase transition. We get our first indication of this by looking at the
way particles are distributed among the allowed states. Let us first note that
since the exponential ef3c1 can only have a range of values 1 ::; ef3c1 :::; 00, the
fugacity must have values
z ~ 1. Otherwise, (nl) could become negative,
and that would be unphysical since (nl) is an average particle number. Thus, for
a Bose-Einstein gas the chemical potential, J-L' = kBT In(z), must be negative or
zero. This means that it is "easy" to add additional particles to the gas. Let us
notice something else about the average particle number, (n,). For the state with
quantum numbers 1 = 0 == (Lx = 0, ly = 0, lz = 0), the energy, cO = 0, and the

:::;

384

EQUILIBRIUM STATISTICAL MECHANICS

average particle number, (no), can become very large as z

(no) =

-+

1, since

C ~J

(7.128)

The fact that limz_l (no) = 00 means that the state with zero energy can become
macroscopically occupied as z -+ 1. This is precisely what happens at the phase
transition.
Let us now compute some thermodynamic quantities for the Bose-Einstein
gas. The summations in Eqs. (7.125) and (7.126) cannot be done explicitly, but
for large enough volume, V, we can change the summation, L:l' to an
integration. For simplicity, let us assume the gas is in a cubic box so that
L; = L, = L, = Land V = L3. We first compute the average particle number,
(N), which from Eqs. (7.120) and (7.126) takes the form
(7.129)

where Px = 2nJz
summations in
Ix = 0, ly = 0, a
removed from tl
values of the II
py = 27rnly/L, a
Px = O,py = 0,
we obtain

'J'~

Converted with

STOU Converter
trial ve rsion

hnp://www.stdutilitv.com
dl

V
= --3
(27rn)

J'

dp

47rV
= --3
(27rn)

JOO

to change the
the sum with
pre it must be
~.The allowed
Vx = 27rnlx/ L,
~e the point at
~lcoordinates,

2
pdp.

(7.130)

27rn/L

The summation, L::, and integration, f'dl, exclude the point at Ix = 0, Iy = 0,


and l, = 0. The integration,
dp, excludes a sphere of radius p < 27rn/ L about
the origin. We can now rewrite the average particle number as

(7.131 )
Similarly, we can write the grand potential in the form
f2BE(T, V, 11) = kBTln(1 - z)

+ 47rVkBTJOO
3
(27rn)

P 2dpln[1 - zexp( -f3p

2/ 2m )] ,

27rh/L

(7.132)
where we again have separated out the point at the origin.

385

IDEAL QUANTUM GASES

It is useful to introduce dimensionless variables. If we let:x2 = {3p2/2m, the


grand potential becomes

(7.133)
and the average particle number becomes
(N)

Z
4V JOO
=--+-3-

I- Z

ATvffl

Ar.;:ff/L

x'2 dx (

-2-

eX

Z
-

z '

(7.134)

where

(7.135)
is the thermal wavelength. The thermal wavelength is the deBroglie wavelength
of a quantum
e of the spread
of a wavefunct
Converted with
an the average

STOI Converter

interparticle s
significant inft
when AT beco
thermodynami

ill not have a


gas. However,
le spacing, the
lyon whether

trial version

hDP://www.stdutilitv.com

the&~i:~~:

e variables and

examine the volume dependence of various terms. From Eq. (2.122), the
pressure can be written
(7.136)
where
gS/2(Z)

=-

V 7r

Joo ~dx
0

In[1 - ze-x2] =

L a~/2
00

(7.137)

0=1

and
(7.138)
In Eq. (7.137), the summation is obtained by expanding the integrand in powers
of z and integrating each term.

386

EQUILIBRIUM STATISTICAL MECHANICS

We can also rearrange Eq. (7.134) to obtain an equation for the average
particle density,
(n)

(N)

V = V 1 _ Z + A~ g3/2(Z)

)q.,fi) ,

(
- In Z, -L-

(7.139)

where
(7.140)

and

In(z,a) = . ~\3

V 7r AT

g5/2(1)

~dx(-2-Z-),

to show lima-+oIp(l,a)

It is straightforward

Therefore, in c
term I = 0 had
The quantitie
are well-behave
g3/2(Z) remain

Ja

Z ~

= O. and

lima-+oIn(1, a) = O.
if the singular
~e been lost.
p) and (7.139),
th g5/2(Z) and

STOI Converter
trial version

((&.-70,_h_D_P_:/_/WWW
__

1, where

(7.141)

- Z

Converted with

._St_d_Ut_il_itv----.-.=-->CO_m__

(7.142)

,-/

as

eX

(m and (m are Riemann zeta functions.

2.612

~,

1.342

~'!I\'"

9o/Z(Z)

o~~----------~------------~~
o
z
0.5

Fig. 7.10. Plots of gS/2(Z)

and g3/2(Z)

versus z,

1.0

387

IDEAL QUANTUM GASES

1 z
V}::";"

0.5

Fig. 7.11. Plots of (l/V)[z/(l


V = 1000.

1.0

- z)] versus z for (A) V

10, (B) V = 100, (C)

Finall y, let lJ-'!---'lLE..!1.l:la.i.n..QL..t.h.o;L..ti.u~::u::tJ::!...__LJ_l_jL_.L.l.n.1_l__--'Ll._!:l.!1.1_~


which appear i
(n) and Tappe
the neighborh
constant, so z
that

V)[ z/ (1 - z)]
nsive variables
and z ~ 1. In

STOI Converter

where no is a
is easy to see

trial version

hDP://www.stdutilitv.com

lim - -11'nn---=-)"IVITlr::::II----anu-----rnrnl-=-====-T---===
V-+oo (
V

no.

(7.143)

Thus, in the limit V ~ 00, we can neglect the contribution to pressure due to
the term -(I/V)ln(1
- z), but we must retain the contribution to the average
particle density due to the term (I/V)[z/(1
- z)]. Let us now examine the
behavior of the function (I/V)[z/(1
- z)]. In Fig. 7.11 we plot (I/V)
[z/(1 - z)], as a function of z for increasing values of the volume, V. Note
that for any given value V, it always becomes infinite as z ~ 1. The larger the
volume, the closer z must be to z = 1 before the term (I/V)[z/(1
- z)] becomes
significant.
Let us now take the thermodynamic limit in Eqs. (7.136) and (7.139). We let
V ~ 00 and (N) ~ 00 so that (n) = (N) /V remains finite. The pressure then
takes the form

<

1,

z=

1.

if z

(7.144)
if

388

EQUILmRIUM

STATISTICAL MECHANICS

The average particle density takes the form


if z < 1,
(7.145)
if

z = 1.

A somewhat exaggerated plot of (n)).~, no'\~, and g3/2(Z) as a function of z is


shown in Fig. 7.12. The plot is actually done at large but finite volume, so the
growth of no'\~ can be seen more clearly. The main contribution to the growth
of the quantity (n)'\~ comes from g3/2(Z) until z approaches z = 1. Then as
z ~ 1, g3/2(Z) approaches a constant and no).~ determines the further growth of
(n)'\~. This is called Bose-Einstein condensation. What we are seeing as z ~ 1
is the macroscopic condensation of particles into lowest energy state of the gas.
The number density, no, of particles in the zero energy state, co, becomes
macroscopic in size.
In the thermodynamic limit, Bose-Einstein condensation begins to occur
when the fugaq'
I Jl(T, (n)) =
,1

.1.

..L

,1.

.1

Converted with

kBTln(z) = 0).
the temperature

STOI Converter

occur when

>D

(7.146)

trial version
[cf. Eq. (7.145
density, (n)c (thL-

hDp ..//www.stdutl-II-IV.COmnticaIParticle
---" as a function

(n}A}
2.612

o~~----~-=~~------------~~
0.5
1.0
o
z
Fig. 7.12. Plots of (n).\.}, g3/2(Z), and no.\.} versus z. (The contribution of no.\.} for
z < 1 has been exaggerated by taking V = 100 rather than V = 00.)

389

IDEAL QUANTUM GASES

of temperature:

(n) = _1_ = g3/2(1) ~ 2.612 (mkBT)2


c
(v) c).~
27r1i '

3/2

(7.147)

where (v) c is the critical volume per particle. The critical temperature, T; (the
temperature at which condensation occurs), as a function of particle density is
given by
2

or

27r1i )
T; = ( mkB

(n) )

g3/2(1)

2/3

(27r1i )
~
mkB

(n) )

2/3

2.612
(7.148)

The order parameter, Tj, for Bose-Einstein condensation is the fraction of


particles in the condensed phase, tt = no / (n). From Eqs. (7.145) and (7.148),
we can write

Converted with

Tj=

A plot of the 0
Equation (7
the "normal"
(7.144), we se

(7.149)

STOI Converter
trial version

hDP://www.stdutilitv.com
, ,
..
, , c"

en in Fig. 7.13.
curve between
From Equation
, , c), the pressure

coexistence
region

O+-------------------~~
o

Fig. 7.13. A plot of the order parameter,


Einstein condensation.

1
'fJ

no/ (n),

versus temperature, T, for Bose-

390

EQUILIBRIUM STATISTICAL MECHANICS

........

_- ------v

Fig.7.14. A plot of the coexistence curve (dashed line) and some isotherms in the P - v
plane for the phase transition in a Bose-Einstein gas.

becomes independent
write the critic

(v)c:

of particle

density. If we now use Eq. (7.147), we can


.
e per particle,

Converted with

STOI Converter

(7.150)

trial version

hDP:llwww.stdutililV.com

A plot of the c
e isotherms in
the P-v plane, .
d curve is the
coexistence region, the region where both condensed particles and noncondensed particles can coexist.
Another quantity of great interest in the neighborhood of a phase transition is
the heat capacity. From Eq. (2.127), the entropy per unit volume is
s = (8S/8V)r,IJ' = (8P/8T)v,IJ'(s = (8S/8V)r,IJ' only if the gas is composed
of single type of particle). Therefore, from Eq. (7.144) we can compute the
entropy and we obtain

where we have made


density, we can now
en = T(8s/8T)n' It is
fixed and not J-L'. The

if z

<

1,

if z

1,

(7.151)

use of Eq. (7.145). Given Eq. (7.151) for the entropy


compute the heat capacity/volume
at constant density,
important to note that in order to compute en, we hold n
computation of en requires the following quantity,

Of3J-L')
( er

= _ 2_ g3/2(Z) .
2T gl/2(Z)

(7.152)

391

IDEAL QUANTUM GASES

Fig.7.1S. A plot of the heat capacity per unit volume for a Bose-Einstein ideal gas as a
function of the temperature. The temperature, Tc, is the critical temperature for the onset
of Bose-Einstein condensation.

Equation (7.152) is obtained by differentiating Eq. (7.145) with respect


to T holding
traightforward.
We find

Converted with

STOI Converter

1,

trial version

1.

(7.153)

hDP://www.stdutilitv.com

In Fig. 7.15
Bose-Einstein
gas. The location of the critical point is clear in the plot. Also, the BoseEinstein gas clearly obeys the third law of thermodynamics. In the limit
T -+ OK, the entropy approaches zero with temperature dependence, T3/2. In
the high-temperature limit, the heat capacity approaches a constant value as we
would expect for a classical ideal gas. In the high-temperature limit the effect of
statistics becomes negligible. The phase transition in an ideal Bose-Einstein gas
in entirely the result of statistics. As we shall see, the Fermi-Dirac gas exhibits
no such transition.
The high-temperature behavior of the Bose-Einstein gas is readily obtained.
At high temperature, Z -+ 0 and gS/2(Z) ~ g3/2(Z) ~ gl/2(Z) = z. From Eq.
(7.145) we obtain
(7.154)
for the particle density, and from Eq. (7.144) we obtain
(7.155)

392

EQUILIBRIUM STATISTICAL MECHANICS

for the pressure. From Eq. (7.153) we obtain

15kBz

9kB(N)

3 (N)kB
V

=------=--4,x3T
4V

(7.156)

for the heat capacity per unit volume. Thus, at high temperature
Einstein gas behaves like an ideal classical gas .

the Bose-

EXERCISE
7.7. Compute the variance, ((N - (N) )2), in the number of
particles for an ideal boson gas (below the critical temperature) in the
neighborhood of T = 0 K.

Answer: The variance in the particle number is given by


((N _ (N) )2) =

.!.
f3

(8(~)) .
8J-L

Converted with

The average
temperature is

(1)

TV

the

STOI Converter
trial version

where g3/2

critical

(2)

hDP://www.stdutilitv.com

(z)
2

((N - (N)) )

= 1 _z z +

(Z)2

1_ z

V
2
+ ,x~
gl/2(Z) ~ (N) + (N) ,

(3)

For an ideal boson gas at low temperature, the particle number distribution
has a huge variance so it is not possible to give good predictions about how
many particles are in the system at any given instant.

7.H.2. Fermi-Dirac Ideal Gases


We shall now examine the thermodynamic behavior of a gas of identical, noninteracting spin s = particles of mass m. For simplicity, we shall assume the
gas is in a cubic box so L, =Ly = L; = L and we shall assume periodic
boundary conditions. When we include the effects of spin we must generalize
the expression for the grand partition function given in Eq. (7.123). Spin -~
particles can exist in two spin states, Sz =
Ii. Therefore, each momentum
state can have two particles, one particle for each of the two spin states, and sti 11
not violate the Pauli exclusion principle. We will let nl,O' denote the number of
particles with quantum numbers I = (lx, ly, lz) and spin (J, where (J =l (1) for

393

IDEAL QUANTUM GASES


Sz

+!liC -! Ii).

The grand partition function takes the form

(7.157)

The power of 2 is due to the fact that there are two possible spin states for each
set of quantum numbers, I. If we are dealing with a gas of spin-s fermions, then
there will be g = 2s + 1 spin states for each value of 1 and the partition function
is given by
ZFD(T,

V, /-l)

II (1 + e-!1(c tl
l-

))g.

(7.15S)

The grand potential is given y [cf. Eq. (S.B.52)]

Converted with

STOU Converter
where g = 2:[i

trial ve rsion

hnp://www.stdutilitv.com

(N)=~-~==~r-==~~e!1=(c='_=~=')=+=I~==~~r

(7.159)
es in the gas is
(7.160)

where (nl), the average number of particles with quantum numbers I, is defined
as
(7.161)
In Eq. (7.161) the quantity z = e!1~'
is the fugacity. For Fermi-Dirac particles,
the average particle number has no possibility of diverging. The fugacity can
take on the entire range of values 0 ::.;z ::.;00, and the average particle number
can take on a range of value 0 :S (nl) ::.;g. In Fig. 7.16 we plot (nl) as a function
of CI at low temperature (solid line) and at T = OK (dashed line). We see that at
low temperature the particles completely fill all the states with lowest energy.
Only those states at higher energy are partly occupied. At zero temperature, all
states below a cutoff energy, Cl = cf = /-lo, is called the Fermi energy. The
momentum, Pf = ~,
is called the Fermi momentum. The distribution of
Particles in momentum space at low temperature is like a "sea" with all the
lower states filled with particles. Only particles in states near the "top" of the

394

EQUILIBRIUM STATISTICAL MECHANICS

-1-------___=_- -----j
I
I
I
I

:T=OK
I
I
I
I
I
I
I

O+-----------------~~~~

Fig. 7.16. Plots of the average occupation number, (nl), as a function of energy, c), at
very low temperature. The solid line is a plot for T > 0 K, and the dashed line is a plot
for T = 0 K. J.L is the chemical potential.

"sea" can change their state. For this reason this distribution of particles is
called the Fermi sea.
Let us now pI3lr:n.nu..t.e_j":hh,.""Q_jJ,t~I:lO.l..d._:Llln.!u:nJI..!....1~~b.o.J"-a:l""~+L
..tIo~lh~_n~rmi_Dirac
gas.
For large enor
integration,

Converted with

uion, E" to an

STOI Converter
trial version

(7.162)

hDP:llwww.stdutililV.com

[cf. Eq. (7.130


erms to remove
from the summation before we approximate the sum, EJ, by an integral.
Therefore, the magnitude of the momentum, p, in Eq. (7.162) ranges from 0 to
00. The grand potential takes the form
nFD(T, V, Jl) = -PV = -

47rkBTV
3
(27r1i)

JOO P2 dp ln] l + ze f3.P212m ].

(7.163)

Similarly, the average particle number takes the form

47rg V
- (27r1i)3

N) -

00

2d

'P

z
ef3p2/2m

+z .

(7.164)

Let us now make the change of variables, xl = {3p2/2m, in Eqs. (7.163) and
(7.164). Then the pressure of the Fermi-Dirac gas can be written
(7.165)
where

)q

is the thermal wavelength [cf. Eq. (7.135)] and the functionf5/2(z)

is

395

IDEAL QUANTUM GASES

defined as
(7.166)
The average particle density can be written

(N)
g
(n) = V = .\3 h/2(Z),

(7.167)

where

In Fig. 7.17 we plotfs/2(z) andf3/2(Z) as a function of z.


It is interestin
.
ideal classical g
Converted with
the average volu

ith that of an
plot P versus
of the same
1
so we are
temperature for
comparing only
instein gas is
dramatically 10
trial version
mall v. This
happens because
I volume per
particle) a macr
0 momentum
state and can no longer contribute to the pressure. The pressure of the FermiDirac gas, on the other hand, always lies a little above that of the classical gas.
This happens because for a Fermi-Dirac gas only one or zero particles can
occupy a given momentum state, whereas in a classical gas any number can
occupy a given momentum state. As a result, when comparing a Fermi-Dirac
ideal gas to a classical ideal gas with the same particle density, (n) = v-1, the

STOI Converter
hDP://www.stdutilitv.com

Fig. 7.17. Plots offs/2(z)

andf3/2(z)

versus z.

EQUILffiRIUM STATISTICAL MECHANICS

396

O+-------------------------~
v
o
Fig. 7.1S. Plots of the pressure of a Fermi-Dirac (FD), a Bose-Einstein (BE), and a
classical (CI) ideal gas as a function of volume per particle assuming the particles have
the same mass and neglecting spin. One isotherm for each gas is shown. The
temperatures of the three isotherms are the same.

Fermi-Dirac gas will contain more particles at higher momentum. The Fermi-

dominant gro
trial version
Dirac gas the
product, f3J.l/,
owth in {n).A~
occurs when f3
ee below when
we focus on be
of T = OK, the
chemical potential for a Fermi-Dirac gas approaches a positive finite constant as
T ---+ 0 K. This is not evident in Fig. 7.19. We can revert the series expansion of
{n)A~/g [cf. Eqs. (7.167) and (7.168)] and obtain a series expansion for the
fugacity, z = e{3/i', which is convergent for sufficiently low density and high

hUP:llwww.stdutililV.com

(n)'\}

{3/-L

Fig. 7.19. Plots of (n)A~ versus f3J-L for a Bose-Einstein (BE) and Fermi-Dirac
gas.

(FD)

397

IDEAL QUANTUM GASES

temperature. We obtain

z = e{3/1'= (n)A~ + _1_ ((n)A~)


g

23/2

(2.__
22

1_) ((n)A~)
33/2
g

+. ..

(7.169)

The coefficients of various powers of the quantity, (n)A~/g, will always be


positive. Thus, as T -+ 00, Z -+ 0 and the product, 13/1,', must be large and
negative. Since 13 -+ 0 the chemical potential J-l ---t -00 at high temperature.
For low temperatures, Z ---t 00 and f3J-l' -+ 00. Since 13 ---t 00, in the limit T ---t 0
the chemical potential can remain finite and indeed it does.
Let us now compute the thermodynamic properties of the ideal Fermi-Dirac
gas at low temperatures. Let us first examine the behavior of 13/2 (z) at low
temperatures. We may write it in the form

(7.170)

Converted with

STOI Converter

obtain the last

where we have
integral we ha
which appears
trial version
number (nl), a
the strongest
integration in
,
y
we then let t = (y - v), we can write /3/2 (z) as

-v

hDP://www.stdutilitv.com

f3/2(Z)

= --4

3y11r

JOO

dt

(1

-v

+ e )2

v3/2

3 I/2t + _v3 I/2t2 + ....)


+ _v
2

~(y, v)
0.25

f!
v
Fig. 7.20. A plot of ~(y, v) == eY-v /[1

/[1

+ ey-v]2

y
versus y.

+ ey-vf,

the occupation
v = 13J-l' where
o perform the
about y = t/, If

(7.171)

398

EQUILIBRIUM STATISTICAL MECHANICS

The contribution from the lower limit in the integral will be of order e-f3J.L. At
low temperatures we can neglect it and extend the lower limit to -00 so that
f3/2(Z)=-- 4
3fo

00

dt

-00

e'
(3/2
v
(1 + et)2

3 1/2t+-v3 1/2t+.
2
)
+-v
2
8

(7.172)

To evaluate Eq. (7.172), we must evaluate integrals of the form


In =

Joo
-00

dt

met

(1 + e')

2.

(7.173)

The result is In = 0 for n odd, 10 = 1, and In = (n - 1)!(2n)(1 - 21-n)((n)


for
n even, where ((n) is a Riemann zeta function [((2) =
((4) =
(( 6) = ;;5]' etc.
We can use the above results to obtain an expansion for the quantity (n)>'~/ g
which is valid at low temperature. From Eqs. (7.167), (7.172) and (7.173), we
find

f,

Converted with
If we take the

dependent expr

STOI Converter
trial version

to,

(7.174)
wing density-

hDP://www.stdutilitv.com
(7.175)
The chemical potential, 1-0 == er. at T = 0 K is also called the Fermi energy,
because at T = 0 K it is the maximum energy that a particle in the gas can have
(cf. Fig. 7.16). At very low temperatures, only particles within a distance, kBT,
of the Fermi surface can participate in physical processes in the gas, because
they can change their momentum state. Particles lower down in the Fermi sea
have no empty momentum states available for them to jump to and do not
contribute to changes in the thermodynamic properties. Equation (7.174) may
be reverted to find the chemical potential a a function of temperature and
density. The result is
(7.176)

Thus, the chemical potential approaches a finite constant as T ---+ 0 K. In


Fig. 7.21 we plot the chemical potential of an ideal Fermi-Dirac gas a function
of temperature for fixed particle density.

399

IDEAL QUANTUM GASES

Fig. 7.21. A plot of the chemical potential of a


Fermi-Dirac gas as a function of temperature for
fixed particle density.
The internal energy, U
manner. At low temperature,
U

= (if) = I:, c,n"

can be computed

in a similar

it is given by

= -3 {N)cF [ 1 + -5 (kBT)2
+ ...] .
5

12

(7.177)

CF

From Eq. (7 .1,-L'-L-''____~''''''''''''~'''''''''''''''''''''-_---'''''~~''''''''''''''''''__''''''''''''''''~~~_'_')irac


gas in the
limit T

---+

Converted with

STDU Converter

(7.178)

trial version
~~;~r~~r~e;t

http://www .stdutiliJV.com

li~:~%o~:a~:

with the third law. It is important to note, however, that particles in an ideal
Fermi-Dirac gas can have a large zero-point momentum and, therefore, a large
pressure and energy even at T = 0 K. This is a result of the Pauli exchange
principle.
It is a simple matter to show that at high temperatures,
all quantities
approach values expected for an ideal classical gas .
EXERCISE 7.S. Compute the variance in particle number,
for a Fermi-Dirac gas for temperatures near T = 0 K.
Answer:

First note the thermodynamic

(N - (N) )2),

identity,

(1)
Near T

=0

K, we can write
(2)

400

EQUILIBRIUM STATISTICAL MECHANICS

where g is a multiplicity factor due to the degeneracy of spin states, and


2
AT = J27rn /mkBT
is the thermal wavelength. From Eq. (2), we find
4gV

(mp,l)

3/2

(3)

(N) = 3y17r 27rn2


If we now take the derivative of Eq. (3), we obtain

("~(N))
=
o'
/-L

T,V

2gV p,ll/2m3/2

J1r(27rn2)

(4)

3/2 .

Let us now solve Eq. (3) for /-L' as a function of (N) and substitute into Eq.
(4). We find

(''.O/-L'
l(N))

T,V

= Vm2
n

Ct(N)t

(5)

4n4V

The variance then becomes

Converted with
The variance
much smaller

STOI Converter
trial version

(6)
emperature is

hDP://www.stdutilitv.com
I

EXERCISE 7.9. Compute the density of states at the Fermi surface for

I an ideal Fermi-Dirac gas confined to a box of volume V.

Answer: For simplicity we assume the box is cubic with volume V


The momentum is

= L3.
(1)

where lx, ly, and lz are integers each ranging from -00 to 00. Let us denote
the set of integers 1 = lx, lylz. Each different choice of integers lx, ly, and l,
corresponds to a possible state of the system. The allowed values of the
momentum, p, form a grid of points located atpx = 27rnlx/L,py = 27rnly/L,
and pz = 27rnlz/L in p space. The volume per point in p space is (27rn/L)3.
The number of points per unit volume is (L/27rn)3 . The number of points
(states), V, inside a spherical volume of p space with radius less than p is
V

4 7rp3 ( L
=3
21rn

)3 .

(2)

401

SPECIAL TOPICS: HEAT CAPACITY OF LATTICE VffiRATIONS

The energy of a particle with momentum p = lik is


number of states in the interval v -+ v + du is
m3/2V..j4.

du

du

= -dck

dCk

M
y

27r21i

Ck

= li2e 12m. The


(3)

dex.

The density of states at the Fermi surface is


N (0)

dV)

= (-

dCk

32

m/

k=kj

V..jlj

= ------____:_..,,-2 3
V27r li

(4)

'

where cf is the Fermi energy. It should be noted that Eq. (4) is the density of
state for a single spin component.

.....SPECIAL TOPICS
of Lattice Vibrations on a One-Dimensional

Converted with
lattice of h
mass m coupl
constant, K. F

ne-dimensional
of N atoms of

STOU Converter
trial version

are coupled to

hUP:llwww.stdutililV.com

~~!!~~:~el~

F!g.p~.2~d T~~

respectively (the displacement, qj, is measured relative to the left wall). When
the lattice is completely at rest, the springs have length a. The distance between
the walls is L = (N + 1)a. The Hamiltonian operator is
N

"2

"
""' P]
H=
L...J-+i= 1

2m

N-I
"
"
L...J(qj+I
-q;-al)

K ""'

"2

2 i=1

K"
+-(qI
2

K"
-al)"2 +-(qN

-Nal)"

(7.179)
L = (N

~-"'

+ l)a
j

...ve-"' "'!~
qj

Fig. 7.22. A one-dimensional lattice of N masses, m, coupled by harmonic forces with


force constant K-. The lattice is attached to rigid, infinitely massive walls at each end. {/j
measures the displacement of the jth mass from the left wall. The distance between
walls is L = (N + 1)a, where a is the spacing between masses when the lattice is at
"rest" (for a quantum lattice, there will always be zero point motion).

402

EQUILIBRIUM STATISTICAL MECHANICS

where i is the unit operator. The position and momentum operators obey the
commutation relations [qj,h] = thoU, Ih,p},] = 0, and [qj, qj'] = O.
We can also measure the displacement, lj, of the jth atom relative to its
equilibrium position. If we let qj = jal + Uj, the Hamiltonian operator takes the
simpler form
(7.180)
Let us now introduce the interaction matrix,

V=-

2
-1

'"

-1
2

0
-1

0
0

0
0
(7.181)

0
0

0
0

0
0

-1
2

2
-1

Then the Hamiltonian onerator takes the form

Converted with

STOI Converter

(7.182)

where p = (PI,
~trices containtrial version
ing the momer
N atoms. The
quantities pT a
hd I is the unit
matrix.
The interaction matrix, ii, is symmetric and therefore can be diagonalized by
an orthogonal matrix, X. Since X is orthogonal it satisfies the conditions,
XT . X = X . XT = I. Because v is an N x N matrix it has N eigenvalues which
we denote by ~, where a = 1, 2, ... N. Then

hDP://www.stdutilitv.com

-T

X . VX=A,

where

(7.183)

A is the diagonal matrix

wi
_

A=

w~
(7.184)

One can easily check that the eigenvalues are


2
WQ

= ",%sm

2(

7ra

2(N

+ 1)

(7.185)

403

SPECIAL TOPICS: HEAT CAPACITY OF LATTICE VIBRATIONS

where Wo = ~
is the natural frequency of a harmonic oscillator with force
constant K, and mass m. Matrix elements of the orthogonal matrix are

(7.186)
We can now use the orthogonal matrix to transform to normal mode
coordinates, (Pa) Qa). Let itj = L~=l Xj,aQa and Pi = L~=l Xj,aPa. Then the
Hamiltonian operator becomes

(7.187)

Thus, the normal modes consist of a collection of uncoupled harmonic


oscillators, each with different frequency, Wa. They correspond to standing
sound waves on the lattice. Since the orthogonal transformation is also
canonical, the
'" ) Pa'] = i1i8a a' ,
[P a) P a'] = 0
Converted with
'
We now h
ian operator in
particularly si
erator al, and
annihilation 0
sound mode),

STOU Converter
trial version

hDP://www.stdutilitv.com

(7.188)

o:

Note that
and P a are Hermition operators.
Using the commutation relations for the normal mode coordinates, we can
show that the creation and annihilation operators satisfy the commutation
relations, [aa) a~,] = 18a,a'. The Hamiltonian operator then takes the form
(7.189)
where na = alaa is the number operator for energy quanta (phonons) in the nth
normal mode, Ina) is its eigenstate and nalna) = nalna).
The partition function can be written in terms of the Hamiltonian, Eq.
(7.180) or (7.189), since inside the trace we can use the orthogonal matrix to
transform between them. We simply insert the identity, X . XT = I, into the
trace and obtain
TrN

(e-PH) = TrN (X . XT . e-PH) =

TrN (XT .

e-PH . X) = TrN (e-pxT,fI.X).


(7.190)

404

EQUILIBRIUM STATISTICAL MECHANICS

Therefore, the partition function can be written


(7.191)
The average energy is
(E)

= _ 8ln(ZN) = ~ /iwa + ~
8(3

/iwa
ef31iwo -

.
1

(7.192)

The average number of quanta in the oth phonon mode is


(7.193)
which is Planck's formula. The heat capacity is
(7.194)

Converted with

STOI Converter

an change the
The summatio
e heat capacity
summation to
trial version
as we shall sh
It is easy t
es the correct
classical expre
00, (3/iwa 1
and ef31iwo ~ 1 + (3/iwa+ . Then the lowest order term in an expansion in powers
of (3/iwa is C = NkB, which is the correct expression for the heat capacity of N
classical harmonic oscillators on a one-dimensional lattice .

hDP://www.stdutilitv.com

.....S7.A.l. Exact Expression-Large

If the number of atoms in the chain is very large, then in Eqs. (7.192) and

(7.194) we can change the summation into an integration. The normal mode
frequency, w, as a function of mode number, a, is
(7.195).
where 2wo is the maximum allowed phonon (sound) frequency. Let us denote
this "cutoff" frequency by WL = 2wo. For N 1 we can write

LN ~IN
a= 1

do. ~

JWL
0

2N
g(w)dw ~ 1r

JWL
0

vVrdw
L -

w2

N,

(7.196)

SPECIAL TOPICS: HEAT CAPACITY OF LAmCE

405

VIBRATIONS

where

g(w) == da =
2N
dw 7rVw'i - w2

(7.197)

is the density of states (obtained by differentiating Eq. (7.195) with respect to

w) and g(w)dw is the number of phonon modes (not quanta) in the frequency
interval w -t W + dw. The average energy can be written
1 JWL

(E) = -

nwg(w)dw +

JWL

nwn(w)g(w)dw

= Nnw
__ L + jWL nwn(w)g(w)dw,
7r

(7.198)

where n( w) = (e/31iw - 1
is the average number of quanta in the phonon
mode with frequency w. The quantity NnwL/ 7r is the zero point energy of the
lattice. In the limit T -t 0, n(w) -+ 0 and (E) -+ NnwL/7r. Thus, the quantum
lattice always has some motion even at absolute zero kelvin. This is a
consequence of the Heisenberg uncertainty relation, which does not allow the
momentum of
ed to remain in
the neighborhr
Converted with
\Prom Eqs.
tten

STOI Converter
trial version

hDP://www.stdutilitv.com

~ e/31iw

f/w'i - w2
(7.199)

It is useful, however, to make one last change of variables. Let x = {3nw and
define a lattice temperature, TL = nwL/kB This is the temperature the lattice
must have so the thermal energy, ke T, is equal to the largest allowed quantum of
energy, nwL. The heat capacity then takes the form

(7.200)

In the limit T

-+

0, the heat capacity becomes approximately


(7.201)

where we have used the fact that


(7.202)

406

EQUILIBRIUM STATISTICAL MECHANICS

(expand the integrand in powers of e-x and integrate). For special values of n
the sum can be performed. For example, Iz = 1r/3 and 14 = 41r4/15. Thus, at
very low temperatures the heat capacity of this one-dimensional lattice goes to
zero linearly with the temperature. Notice also that at very low temperatures the
high-frequency (short-wavelength) phonon modes give almost no contribution
to the heat capacity of the lattice because there is not enough thermal energy to
excite them .
.....S7.A.2. Continuum Approximation-Large

We will assume that we do not know the exact dispersion relation [Eq. (7.185)]
for the phonon modes, and we will use the information we do have. We know
that there are N atoms on the lattice, that it has length L, and that it is pinned on
the ends. The allowed wavelengths for phonons on an elastic lattice of length L
which is pinned at its ends are given by Aa = 2L/ a, where a = 1,2, ... ,N. The
allowed phonon wavevectors are k., = 21r/ Aa = 1ra / L. The minimum allowed
wavelengths are of the order of the lattice spacing, and therefore there is a
cutoff frequen
.
The dispersion
relation for the
peed of sound.
The Debye fre
onon modes is
equal to the n

STOI Converter

N=L

a=1

trial version

(7.203)

hDP://www.stdutilitv.com

Therefore, the Debye frequency is WD = NC1r / L and the density of states is


do

gD(W)=-=-=-.
dw

C1r

WD

(7.204)

The density of states in the continuum approximation is a constant, whereas the


exact solution gives a frequency-dependent density of states. In Fig. 7.23 we
plot the exact density of states and the density of states obtained using the
continuum approximation. The continuum approximation dramatically underestimates the density of states near the cutoff frequency.
The average energy is given by Eq. (7.198) if we simply replace the lattice
density of states, g(w), by the continuum density of states, gD(W). We then find
(E)

NnwD
N JWD
=--+nwn(w)dw.
4

WD

(7.205)

The heat capacity is


_ NkB fWD ({3nw)2ef31iwdw _ NkBTJTDIT
2 T
0
(e/31iw - 1)
D
0

eN _- WD

dx:x2~
(eX - 1)

2'

(7.206)

407

SPECIAL TOPICS: MOMENTUM CONDENSATION

1fg(w)

1.0

continuum
4---------~,...c;..--

0.6 +-e-x-a-c-t-----

~--------------------------~--~~
1.0

0.0

We

Fig. 7.23. Plots of the exact density of states and the density of states obtained using the
continuum approximation, both for the one-dimensional lattice. Here we take
We

= WL = WD

where in the s
and have defi
In the n
becomes

Converted with

STOI Converter

iables, x = {3liw,
approximation

trial version

c = ~~~~=-~~----------~D~----~
hDP://www.stdutilitv.com

(7.207)

At very low temperatures the continuum approximation


also gives a heat
capacity for the one-dimensional
lattice which goes to zero linearly with the
temperature. The coefficient differs slightly from the exact result in Eq. (7.201) .

.... 57.B. Momentum Condensation


[15-17]

in an Interacting Fermi Fluid

An ideal Bose-Einstein
gas can condense in momentum space and thereby
undergo a phase transition, but an ideal Fermi-Dirac
gas is prevented from
doing so because the Pauli exclusion principle does not allow more than one
fermion to occupy a given quantum state. Electrons in a conducting solid are
free to wander through the lattice and form a Fermi fluid. At low temperatures
the electrons form a Fermi sea and only those near the Fermi surface affect the
thermodynamic properties of the electron fluid (cf. Section 7.H). The electrons
experience a mutual Coulomb repulsion which is screened by lattice ions.
However, as first noted by Frohlich [18], those electrons in the neighborhood of
the Fermi surface also experience a lattice-phonon-mediated
effective attraction

408

EQUILIBRIUM STATISTICAL MECHANICS

(two electrons may in effect be attracted to one another because they are both
attracted to the same lattice ion). Cooper [15] showed that this effective
attraction at the Fermi surface could cause bound pairs of electrons to form, and
these pairs could then condense in momentum space, giving rise to a phase
transition in the interacting Fermi fluid. Bardeen, Schrieffer and Cooper, [16]
showed that this momentum space condensation of Cooper pairs is the origin of
superconductivity in materials. In 1972, they received the Nobel Prize for this
work.
We shall now derive the thermodynamic properties of a Fermi fluid which
can form Cooper pairs. It is found experimentally that Cooper pairs have zero
total angular momentum and zero total spin. If the pairs are not undergoing a
net translation through the fluid (no supercurrent), then we can assume that only
those electrons at the Fermi surface with equal and opposite momentum and
opposite spin components are attracted to one another. We shall assume that all
other electrons behave like an ideal gas. With these assumptions, we can write
the Hamiltonian of the electron fluid in the form
(7.208)

STOU Converter

where ek = li
and takes valu
The operators,
momentum lik

trial version

'

given electron
respectively).
electron with
atisfy fermion

hDP:llwww.stdutililV.com

anticommutati
destroys a pair
of electrons wi
onents, and it
creates a pair of electrons with momenta lik and -lik and opposite spin
components. Since the electrons only experience an attraction at the Fermi
surface, the interaction energy, Vk,h can be written

,=

Vkl

Ill' -

ekl ~ Ae and

-Vo

if

otherwise,

Ill' -

ell ~ Ae,

(7.209)

where Vo is a positive constant, IL' is the chemical potential of the fermi fluid,
and Ae is a small energy of order ke T.
In order to simplify our calculations, we shall compute the thermodynamic
properties of this interacting Fermi fluid in the mean field approximation. We
write the Hamiltonian in the form
(7.210)
where

Ak _ {A0

if III - ekl ~ Ae,


otherwise

(7.211)

409

SPECIAL TOPICS: MOMENTUM CONDENSATION

and
(7.212)
The prime on the summation, L:~,means that the summation is restricted to a
distance, ~c, on either side of the Fermi surface. The average, (a-k,!ak,j), is
defined as
(7.213)
where the density operator,

p, is defined as
(7.214)

The average, (at.ja~k,!), is similarly defined. The number operator,


defined as

N,

is

Converted with

STOI Converter
The quantity
thermodynam
the Cooper I

(7.215)

trial version

complex. It is a
ng energy of all
airs form, then
(at,ja~k,!) ~ yin; and (aba~k,!) ~ yin; where nc is the average number of
Cooper pairs in the fluid. ~ is the order parameter for this transition.
It is important to notice that the Hamiltonian, Hmf, does not commute with
the number operator, N, if ~ =f. O. This means that if a macroscopic number of
Cooper pairs form, the system does not conserve the particle (electron) number
and the gauge symmetry is broken. The formation of a macroscopic number of
Cooper pairs is a phase transition somewhat analogous to Bose-Einstein
condensation (cf. Section 7.H). In both cases, gauge symmetry is broken. Since
we are working in the grand canonical ensemble and only specify the average
particle number, the fact that gauge symmetry is broken is not a problem. If a
macroscopic number of Cooper pairs form, the total energy of the system is
lowered. The transition to the condensed phase occurs when the thermal energy,
kBT, which tends to break Cooper pairs apart, becomes less important than the
phonon-mediated attraction between electrons.
It is useful now to introduce an effective Hamiltonian
__

hDP://www.stdutIIIlV.COm

(7.216)

EQUILIBRIUM STATISTICAL MECHANICS

410

where
(7.217)
and we have made use of the fermion anticommutation relations. The effective
Hamiltonian, K, differs from Hmj only by a constant term. Therefore the
density operator can also be written

tdv

"

ef3K

e-f3(HmrpN)

P = Tr [e-f3(HmrpN)]

The effective Hamiltonian,

= Tr [e-f3K]

(7.218)

K, can be written in matrix form:

K=

LlltKkllk)

(7.219)

where

Converted with

STOI Converter

(7.220)

trial version
nian, K, can be
As was firs
es the fermion
diagonalized
anticommuta
,
ian for effective
excitations (called bogolons) of the system. To diagonalize the effective
Hamiltonian, we introduce a 2 x 2 unitary matrix,

hDP://www.stdutilitv.com

(7.221 )
Since [J~[Jk= [J.tJ~ =
introduce the vectors

I (unitarity), we must have IUkl2 + IVkl2 = 1. We also

rk

(ik,O)
"t)

rt k = ("t
" ))
Tk,O ')'k,l

(7.222)

Tk.l

which are related to the vectors, llk, via the unitary transformation
(7.223)
The physical significance of the vectors, rk, will become clear below. It is easy
to show that since ilL and ilk,A obey fermion anticommutation relations, the

411

SPECIAL TOPICS: MOMENTUM CONDENSATION

operators, it,1 and ik,1 (i = 0, 1), must also obey fermion anticommutation
relations
A]
[A')'k,j, ,)'k'
,I' +

[At

At]

= Tk,I' Tk',I' +

0
= .

(7.224)

If we revert Eq. (7.223), we see that ik,o decreases the momentum of the system
by Tikand lowers the spin by Ti (it destroys a particle with quantum numbers,
(k, j), and creates one with quantum numbers, (-k, 1), whereas ik,l increases
the momentum of the system by Tikand raises the spin by h.
We now require that the unitary matrix,
diagonalize the effective
Hamiltonian, KK. That is,

o;

with Ek = (EkO'O
We find that Ek,O = Ek and Ek,l

= -Ek

0)

Ek,l

(7.225)

with
(7.226)

Converted with

STOI Converter

With this tr
gas of electr
operators, th

teracting Fermi
s of bogolon

trial version

hDP://www.stdutilitv.com
k

L:(Ek,O it,o ik,o - Ek,l it,l ik,l + Ek,l)'

(7.227)

The effective Hamiltonian, when written in terms of bogolon operators, looks


like that of an ideal Fermi gas. The bogolons are collective modes of the system
and play a role analogous to that of phonons in a Debye solid, although their
dispersion relation is quite different.
We can now obtain a self-consistent equation for the gap function, ~. First
note that

_ "21 [1 -

A ) _
(At"Yk,O
')'k,o - (1+ e (3Ek,O) -

tanh

Ek,O)]
(f3-2-

(7.228)

k,l )]
({3E
-2-

(7.229)

and

(rk,lTk,l)

- (1+
_

e-(3Ek,l)

1 [ 1 + tanh
2

412

EQUILIBRIUM STATISTICAL MECHANICS

Then

(7.230)
Let us now e

Converted with

STOI Converter
trial version

If we multip

(7.212), we (

(7.231)
Eqs. (7.211) and

hDP://www.stdutilitv.com
1 = Vo

L:' -2Ek1 tanh (f3Ek/2).

(7.232)

It is useful to note that under the primed summation the bogolon energy can be

J ~~ 1~12.

written Ek =
+
Equation (7.232) is the equation for the gap function.
It has been obtained from the grand canonical ensemble. Therefore, the
solutions of the gap equation correspond to extrema of the free energy. The
solution at a given temperature which corresponds to the stable thermodynamic
state is the one which minimizes the free energy. Since the energy, Ek, depends
on the gap, Eq. (7.232) is rather complicated.
Let us now determine some properties of the gap function from Eq. (7.232).
If we assume that the system is contained in a large volume, V, we can change
the summation to an integration [cf. Eq. (7.162)]. Note that
(7.233)
where we have Eg. (7.217). The summation,

L:~, which

is restricted to the

413

SPECIAL TOPICS: MOMENTUM CONDENSATION

neighborhood of the Fermi surface, can be written


(7.234)
where we have set J-l ~ cf (cf is the Fermi energy) and N(O) = mVktl2~li2 is
the density of states at the Fermi surface for a single spin state (cf. Exercise
7.9). We can now write Eq. (7.232) as

(7.235)

Equation (7.235) determines the temperature dependence of the gap, ~(T), and
can be used to find the transition temperature.
The energy of bogolons (measured from the Fermi surface) with momentum

Jc

lik is Ek =
their momentu
temperature, r;
of an ideal Fe
(7.235). It is
critical tempera

Converted with

STOI Converter

, regardless of
At the critical
reduces to that
ned from Eq.
. Thus, at the

trial version

hDP://www.stdutilitv.com
1 = VoN(O)

= N(O) Voln

where f3c = (kBTcr1,

ek

(7.236)

[~f3c~c] ,
a = 2.26773, and we have used the fact that
b

Jo

tanh(x) dx
x

In(ab),

(7.237)

for b > 100. Thus, Eq. (7.236) holds when f3c~c/2 > 100. This means that
N(O)Vo < 0.184 and therefore use of Eq. (7.236) restricts us to fairly weakly
coupled systems. From Eqs. (7.236) and (7.237) we obtain
k T. = ~ ~ce-l/N(O)Vo
B c

'

(7.238)

for f3c~c/2 > 100. Thus, the critical temperature, Tc, varies exponentially with
the strength of the attractive interaction.

414

EQUlLffiRIUM STATISTICAL MECHANICS

We can also use Eq. (7.235) to find the gap, ~(O)


tanh (00) = 1, we can write
1 = VoN(O)

6C

dek

J e2 1+ ~o 2 = VoN(O)sinh

== ~o,

-I

at T = 0 K. Since

(~c)
"6" '

(7.239)

or
A

~c

'-10 -

sinh (l/VoN(O))

'" 2~

'"

-l/N(O)Vo

ee

(7.240)

The rightmost expression for ~o applies for weakly coupled systems when

N(O)Vo < 0.184. Comparing Eqs. (7.238) and (7.240), we obtain the following
relation between the critical temperature and the zero temperature gap for
weakly coupled systems:

(7.241 )

Converted with

Equation (7.2
superconduct
of the gap as
the case (sud
show the beh
Since bogc

STOI Converter
trial version

hDP://www.stdutilitv.com
S

= -2kB ~)nkln(nk)

+ (1 -

s of this ratio for


to obtain a plot
real function for
ht is present. We
~stems.
en in the form

nk)ln(l - nk)],

(7.242)

where nk = (1

+ e,BEkrl

(cf. Problem 7.23). The heat capacity,

CV,N,

is easy to

a(T)
~

Fig. 7.24. A plot of the ratio D.(T)I D.o


versus the reduced temperature, T
for
a weakly coupled system.

rr;

415

SPECIAL TOPICS: MOMENTUM CONDENSATION

find from Eq. (7.242). let us first note that for a Fermi gas at very low temperature we have J-l r.::::: cf, where cf is the Fermi energy, and (8J-l/8T)v,(N)
r.::::: O.Thus,

CV,N = T

(8S)

V,(N) r.::::: 2{3kB

8T

= -2ak
fJ

8nk

~
B ~

8E

~8nk
( nk )
~
8{3 In 1 - nk

(E2 +!2fJa81~kI2)
8a
k

fJ

(7.243)
.

We can now examine the heat capacity, both at the critical temperature and in
the limit T ~ 0 K.
Let us first look at the neighborhood of the critical point. The first term in Eq.
(7.243) is continuous at T = Tc, but the second term is not since 81~k12/8{3 has
a finite value for T < r; but is zero for T > Tc. Near T = r.; we may let
Ek ~ lekl. Then the heat capacity just below the critical temperature is

(7.244)

Converted with
and just abo

STOI Converter
trial version

(7.245)

hDP://www.stdutilitv.com
The discontinuity in the heat capacity at the critical temperature is

(7.246)

Thus, the heat capacity has a finite discontinuity at the critical temperature, as
we would expect for a mean field theory.
Let us now compute the heat capacity in the limit T ~ O. As we can see
from Fig. 7.24, the gap function, ~, approaches a finite value, ~o, as T ~ 0 and
f}A/ 8T ~ 0 as T ~ O. As a result the heat capacity takes a fairly simple form
III the limit T ~ O. If we assume that J-l' r.::::: cf and ~ r.::::: ~o in Eq. (7.243), then
the heat capacity takes the form
(7.247)

416

EQUlLffiRIUM STATISTICAL MECHANICS

~a.

JeZ

where Ek =
+ In order to change Eq. (7.247) it is useful to introduce
the bogolon density of states. We can write

For momenta, k ~ kf' the density of states is singular. Therefore, the dominant
contribution to the integral comes from the neighborhood of the Fermi surface
and we can write

:E ~ N(O)

Joo

60

EkdEk
JE~ -

~a

Let us next note that in the limit T ~ 0 we can write


Thus, the heat capacity takes the form

e(3Ek /

(7.249)

(1 + e(3Ek)2

e-(3Ek

(7.250)

Converted with
The integral

STOI Converter
trial version

te that
(7.251)

hDP://www.stdutilitv.com
(7.252)
Thus, the heat capacity takes the form
CV,N

= 2.1 {32 kBN(0)~o[3Kl3

((3~o)

+ K3({3~O)].

(7.253)

If we now make use of the asymptotic form of the modified Bessel functions,
~ J7r/2{3~o e-(360, the heat capacity takes the form

Kn({3~o)

(7.254)
in the limit T ~ O.Thus, the heat capacity of the condensed Fermi fluid goes to
zero exponentially with temperature rather than linearly as in the case for an
ideal Fermi gas. In Fig. 7.25 we show a sketch of the heat capacity of the
interacting Fermi fluid (superconductor). The solid line is the Fermi fluid, and
the dashed line is an ideal Fermi gas.

417

SPECIAL TOPICS: MOMENTUM CONDENSATION

CV,N

TITe

Fig. 7.25. A sketch of the heat capacity for a superconductor. The straight dashed line
gives the heat capacity in the absence of interaction (ideal Fermi gas). The solid line
shows the jump in the heat capacity at the critical point and the exponential decay for
temperatures below the critical point.

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com
~
f-

0.5

0-

r-

~
....
r

r>
I

0
0

0.2

0.6

0.4

0.8

1.0

T/Tc
Fig. 7.26. Variation of 61kBTc with reduced temperature, T lTc, for tin. The data points
are obtained from ultrasonic acoustic attenuation measurements
[20] for two different
frequencies. The solid line is BCS theory. Reprinted, by permission, from R. W. Morse
and H. V. Bohm, Phys. Rev. 108, 1094 (1954).

The mean field theory gives a surprisingly good description of the behavior
of real superconductors. In Fig. 7.26 we show experimental measurements of
the gap function, ~, as a function of temperature for tin. The solid line is the
mean field theory of Bardeen, Cooper, and Schrieffer. The experimental points,
Which are obtained from ultrasonic accoustic attenuation measurements [21],
fit it very well.

418

EQUlLffiRIUM STATISTICAL MECHANICS

.... S7.C. The Yang-Lee Theory of Phase Transitions [22, 23]


Yang and Lee have used very simple arguments involving the grand canonical
ensemble to arrive at a mechanism by which a phase transition can take place in
a classical fluid. We present it here because it gives valuable insight into the
structure of the grand partition function.
Let us consider a classical system of particles interacting via a potential,
00

V(lqijl) =

-cij,
{

if I'Iv I < a
if a~ I'Iv I ~b,
if b < I'Iv I,

(7.255)

where q ij - qi - qj Because the particles have an infinite hard core, there is a


maximum number of particles, M, which can be fitted into a box of volume V.
Therefore, the grand partition function must have the form

STOI Converter

trial version
where AT is
integral and

(7.256)

hDP://www.stdutilitv.com

the configuration

IS

QN(T, V) = dqNexp { -(3

N(N-l)/2
i~1

(7.257)

V([fluJ)

In the last term in Eq. (7.256), we have performed the momentum integrations.
Since we are now dealing with classical particles, the phase space coordinates
commute and the momentum integrations are trivial. Information about
deviations from ideal gas behavior is contained in the configuration integral.
For N > M, QN(T, V) = 0 because the hard core in the potential prevents more
than M particles from occupying a box of volume V.
Let us now introduce a new variable, y == e(3p, / A}. Then, Eq. (7.256) can be
written in the form
(7.258)
We see that for finite volume, V, the grand partition function, Z (T, V), is a
polynomial of order M in the parameter y. The coefficients of
are positive

SPECIAL TOPICS: THE YANG-LEE THEORY OF PHASE TRANSITIONS

419

and real for all N. Since ZJ-L{T, V) is a polynomial of order M, we can rewrite it
in the form
ZJ-L{T, V)

ft(l- ~).
i=l

(7.259)

YI

where the quantities Yi, are the M roots of the equation ZJ-L{T, V) = O. Because
the coefficients QN / N! are all real and positive, none of the roots Yi can be real
and positive for finite M if we wish to satisfy the equation ZJ-L{T, V) = O.
Therefore, for finite M all roots of Y must be either real and negative or
complex. If they are complex, then they must occur in complex conjugate pairs
since ZJ-L{T, V) is real. As the volume is increased, the number of roots, M, will
increase and move around in the complex plane. In the limit V ~ 00, it can
happen that some of the roots will touch the positive real axis (cf. Fig. 7.27).
When this happens, a phase transition occurs because the system can have
different behavior for Y < Yo and Y > Yo, where Yo is the value of the root on the
real axis. In general, the pressure, P, will be continuous across the point Yo, but
the density a
.
..
~ ...
..
. ous (we give an
example late]
Converted with
re given by
The pressi

STOI Converter
trial version

(7.260)

hDP://www.stdutilitv.com
(a)

Fig. 7.27. A schematic plot of the roots of

ZJ-L{T, V) = 0 in the complex y plane. (a)


For finite V, no roots lie on the positive
real axis. (b) For V = 00, roots can touch

the positive real axis and separate regions,


A and B, with different phase.

420

EQUILIBRIUM STATISTICAL MECHANICS

and
1

(N)

{) 1

- = hm = hm y ~.-ln[ZJl(T,
v
V-+ooV
V-+oo v.r V

V)]

(7.261)

In general, the two operations Iimv-+ooand y( 8/ By) cannot be interchanged


freely. However, Yang and Lee proved that for the type of interaction
considered in Eq. (7.255), the limit in Eq. (7.261) exist and the operations
limv-+oo and y(8/8y) can be interchanged. The results of Yang and Lee are
contained in the following theorems (proofs can be found in Ref. 22).
Theorem I. For all positive real values of y, (l/V) In[ZJl(T, V)] approaches, as V ~ 00,
a limit which is independent of the shape of the volume V. Furthermore, the limit is a
continuous monotonically increasing function of y.
Theorem II. If in the complex y plane a region R containing a segment of the positive
real axis is always free of roots, then in this region as V ~ 00 the quantities
1

1,2, ... ,00,

Vln[ZJl(

Converted with
approach Ii
operations y

STDI Converter

Furthermore, the

trial version
I
v

hDP://www.stdutilitv.com

,vlJ).

Theorems I and II, together with Eqs. (7.260) and (7.261), enable us to
obtain the following relation between v and P:
(7.262)
They tell us that the pressure must be continuous for all y but that the
derivatives of the pressure need only be continuous in regions of the positive
real y axis where roots of ZJl(T, V) = 0 do not touch the real axis. At points
where roots touch, the derivatives of the pressure can be discontinuous. In
general, if 8P / By is discontinuous, then the system undergoes a first-order
phase transition. If a higher-order derivative is the first to be discontinuous, then
the phase transition is continuous.
If v is discontinuous at a point Yo (where Yo is a root of ZJl(T, V) = 0), then it
will decrease with y in the direction of increasing y. This can be proved as
follows. Note that

421

SPECIAL TOPICS: THE YANG-LEE THEORY OF pHASE TRANSITIONS

The quantity, ((N - (N) )2), is always greater than zero. Therefore, l/v always
increases and v always decreases with increasing y.
Yang and Lee applied this theory of phase transitions to the two-dimensional
Ising model [21, 23]. They found that for the Ising model the roots of
Zp,(T, V) = 0 all lie on the unit circle and close onto the positive real y axis in
the limit of an infinite system. The point at which the roots touch the real axis
gives the value of y for which the system undergoes a phase transition.
It is of interest to consider an explicit example [24]. We will consider a
system with the following grand partition function:
(7.263)

where V is an int,eger. Zp, has V real roots at y = -1 and it has complex roots of
the form y = e2mk/V, where k = 1, 2, ... , V - 1. As the value of V increases, the
density of roots on the unit circle increases and the roots get closer to the point
y = I (cf. Fig. 7.28).
The functi~...LL.I...J-...!:J:!.__----L.:.-----L..l.___LLju...L-....L.!:Z..__L_L-.......L:...t.:.L:.---""~
niting values for

Converted with

y < 1 and y

STOI Converter

if y < 1
if y

trial version

>

1.

(7.264)

hDP://www.stdutilitv.com

Note that the pressure is continuous at y = 1. The volume per particle, v, is


easily found from Eqs. (7.262) and (7.264):

y
(l+y)

v = y ay (kBT)

2y

if y < 1,
(7.265)

+I

(1 + y) ,

ify>

1.

Im(y)

J Re(y)

Fig. 7.2S. An example where the roots of the


grand partition function lie on the unit circle in the
complex y plane.

422

EQUILIBRIUM STATISTICAL MECHANICS

P
kBT
1

o~----------~~------~
o
1

Fig. 7.29. A plot of P /kBT

ZJ.I = (I

versus v for a system with grand partition function,

+ y) v (1 - y) v / (1 - y).

Thus the volume per particle is discontinuous at y = 1 and we have a first-order


phase transition. If we combine Eqs. (7.264) and (7.265), it is straightforward to
show that the system has the following equation of state.

Converted with

STOI Converter

(7.266)

trial version

hDP://www.stdutilitv.com
In Fig. 7.29 we plot P / ks T as a function of volume per particle, v. We see that it
has the behavior expected for a first-order phase transition.

REFERENCES
1. J. W. Gibbs, Elementary Principles in Statistical Mechanics (Dover Publications,
New York, 1960).
2. P. and T. Ehrenfest, The Conceptual Foundations of the Statistical Approach to
Mechanics (Cornell University Press, Ithaca, NY, 1959).
3. A. I. Khintchine, Mathematical Foundations of Statistical Mechanics (Dover Publications, New York, 1949).
4. L. D. Landau and E. M. Lifshitz, Statistical Physics (Pergamon Press, Oxford, 1958).
5. A. Einstein, Investigation on the Theory of Brownian Movement (Methuen and Co.,
London, 1926).
6. A. J. Dekker, Solid State Physics (Prentice-Hall, Englewood Cliffs, NJ, 1962).
7. N. W. Ashcroft andN. D. Mermin, Solid State Physics (W. B. Saunders, Philadelphia,
1976).
8. P. Debye, Ann. Physik, 39, 789 (1912).
9. R. Stedman, L. Almquist, and G. Nilsson, Phys. Rev. 162, 549 (1967).

423

PROBLEMS

10.
11.
12.
13.
14.
15.
16.
17.
18.
19.
20.
21.
22.
23.
24.
25.

E. Ising, Z. Phys. 31, 253 (1925).


L. Onsager, Phys. Rev. 65, 117 (1944).
B. Kaufman, Phys. Rev. 76, 1232 (1949).
B. Kaufman and L. Onsager, Phys. Rev. 776, 1244 (1949).
P. Weiss, J. Phys. Radium 6, 661 (1907).
M. Suzuki, 1. Phys. Soc. (Japan) 55, 4205 (1986)
L. N. Cooper, Phys. Rev. 104, 1189 (1956).
J. Bardeen, J. R. Schrieffer, and L. N. Cooper, Phys. Rev. 108, 1175 (1957).
M. Tinkham, Introduction to Superconductivity (McGraw-Hill, New York, 1975).
H. Frohlich, Phys. Rev. 79, 845 (1950).
N. N. Bogoliubov, JET P, 41, 51 (1958).
I.S. Gradshteyn and I.M. Ryzhik, Table of Integrals, Series and Products (Academic
Press, New York, (1980.
R. W. Morse and H. V. Bohm, Phys. Rev. 108, 1094 (1954).
C. N. Yang, Phys. Rev. 85, 809 (1952).
C. N. Yang and T. D. Lee, Phys. Rev. 87, 404, 410 (1952).
G. E. UbI
hanics (American

Converted with
PROBLE

STOI Converter
trial version

Problem 7.1.
onic oscillators,
each with freq
this
structure funct:lur.rarIQOrrR7""""IID~n7anu:l"In7l:n:-1;;nm;;rrr1~~:nTl"]pu-t~rn:T"1l:1Dtropy
and the heat
capacity of the system.

hDP://www.stdutilitv.comrgyE.Using

Problem 7.2. A system consists of N noninteracting, distinguishable two-level atoms.


Each atom can exist in one of two energy states, Eo = 0 or E, = c. The number of atoms
in energy level, Eo, is no and the number of atoms in energy level, E 1, is n 1. The internal
energy of this system is U = noEo + niEI. (a) Compute the entropy of this system as a
function of internal energy. (b) Compute the temperature of this system. Under what
conditions can it be negative? (c) Compute the heat capacity for a fixed number of
atoms, N.
Problem 7.3. A lattice contains N normal lattice sites and N interstitial lattice sites. The
lattice sites are all distinguishable. N identical atoms sit on the lattice, M on the
Interstitial sites, and N - M on the normal sites (N)> M 1). If an atom occupies a
normal site, it has energy E = O. If an atom occupies an interstitial site, it has energy
E :::: c. Compute the internal energy and heat capacity as a function of temperature for
this lattice.
Problem 7.4. Consider a lattice with N spin-I atoms with magnetic moment 1-. Each
atom can be in one of three spin states, S, = -1,0, + 1. Let n-l, no, and nl denote the
respective number of atoms in each of those spin states. Find the total entropy and the
Configuration which maximizes the total entropy. What is the maximum entropy?
(Assume that no magnetic field is present, so all atoms have the same energy. Also
a~sume that atoms on different lattice sites cannot be exchanged, so they are
dIstinguishable ).

424

EQUILIBRIUM STATISTICAL MECHANICS

Problem 7.5. A system has three distinguishable molecules at rest, each with a
quantized magnetic moment which can have z components + ~J-t or - ~JL Find an
expression for the distribution function, /; (i denotes the ith configuration), which
maximizes entropy subject to the conditions Ei/; = 1 and Ei Mi,z Ji = 'YJ-t, where Mi,z is
the magnetic moment of the system in the ith configuration. For the case 'Y = ~,compute
the entropy and compute /;.
Problem 7.6. A fluid in equilibrium is contained in an insulated box of volume V. The
fluid is divided (conceptually) into m cells. Compute the variance of enthalpy
fluctuations, ((!:l.Hi), in the ith cell (For simplicity assume the fluctuations occur at
fixed particle number, Ni). (Hint: Use P and S as independent variables.)
Problem 7.7. A fluid in equilibrium is contained in an insulated box of volume V. The
fluid is divided (conceptually) into m cells. Compute the variance of internal energy
fluctuations, ((!:l.Ui)2), in the ith cell (For simplicity assume the fluctuations occur at
fixed particle number, Ni). What happens to the internal energy fluctuations near a
critical point?
Problem 7.S. What is the partition function for a van der Waals gas with N particles?
Note that the result is phenomenological and might involve some guessing. It is useful to
compare it to the artition
.
at the particles are
indistinquish
Use this part

P(N, T, V),
Assume that
has either ze
c

> O. Assu

Converted with

STOI Converter
trial version

rt

a counting factor.

T, V), the pressure,

ttice with N, sites.


hat each lattice site

hDP:llwww.stdutililV.com

(a) If the s
,
tial of the adsorbed
atoms as a function of T, e, and Na/Ns (use the canonical ensemble).
(b) If the surface is in equilibrium with an ideal gas of similar atoms at temperature T,
compute the ratio Na/Ns as a function of pressure, P, of the gas. Assume the gas
has number density n. (Hint: Equate the chemical potentials of the adsorbed
surface atoms and the gas.)

Problem 7.10. Consider a two-dimensional lattice in the x-y plane with sides of length
L; and Ly which contains N atoms (N very large) coupled by nearest-neighbor harmonic

forces. (a) Compute the Debye frequency for this lattice. (b) In the limit T
the heat capacity?

--+

0, what is

Problem 7.11. A cubic box (with infinitely hard walls) of volume V = L3 contains an
ideal gas of N rigid HCI molecules (assume that the effective distance between the H
atom and the CI atom is d = 1.3A. (a) If L = 1.0 ern, what is the spacing between
translational energy levels? (b) Write the partition function for this system (include both
translation and rotational contributions). At what temperature do rotational degrees of
freedom become important? (c) Write expressions for the Helmholtz free energy, the
entropy, and the heat capacity of this system for temperatures where the rotational
degrees of freedom make a significant contribution.
Problem 7.12. An ideal gas is composed of N "red" atoms of mass m, N "blue" atoms
of mass m, and N "green" atoms of mass m. Atoms of the same color are

425

PROBLEMS

indistinguishable. Atoms of different color are distinguishable. (a) Use the canonical
ensemble to compute the entropy of this gas. (b) Compute the entropy of an ideal gas of
3N "red" atoms of mass m. Does it differ from that of the mixture? If so, by how much?
Problem 7.13. An ideal gas consists of a mixture of "green" and "red" spin-~ particles.
All particles have mass m. A magnetic field, B, is applied to the system. The "green"
particles have magnetic moment "YG, and the "red" particles have magnetic moment "YR,
where "YR < "YG Assume the temperature is high enough that Fermi statistics can be
neglected. The system will be in equilibrium if the chemical potentials of the "red" and
"green" gases are equal. Compute the ratio NRING, where NR is the number of "red"
particles and NG is the number of "green" particles. Use the canonical ensemble (no
other ensemble will be accepted).
Problem 7.14. Consider a one-dimensional lattice with N lattice sites and assume that
the ith lattice site has spin Sj = 1. the Hamiltonian describing this lattice is
H = - L:~1 SjSi+l Assume periodic boundary conditions, so SN+l == SI. Compute the
correlation function, (SIS2). How does it behave at very high temperature and at very
low temperature?
Problem 7.15. In the mean field approximation to the Ising lattice, the order parameter,
(s), satisfies the e uation S = tanh s!.&.. where T. = v 2k
the strength of
the coupling b
ors. (a) Show that
(s) has the fol
T if T rv 0 K, and
(ii) (s)::::::

3
T = Tc. (c) C
T = T; for bo

Problem 7.16.

STOI Converter
trial version

heat capacity at
neighborhood of
r both cases?
s

hDP://www.stdutilitv.com
o

if E < 0,

where a is a constant. Compute the critical temperature for Bose-Einstein condensation.


Problem 7.17. An ideal Bose-Einstein gas consists of noninteracting bosons of mass m
which have an internal degree of freedom which can be described by assuming, that the
bosons are two-level atoms. Bosons in the ground state have energy Eo = p2/2m, while
bosons in the excited state have energy E, = p2/2m + ~, where p is the momentum and
~ is the excitation energy. Assume that ~kB T. Compute the Bose-Einstein
condensation temperature, Teo for this gas of two-level bosons. Does the existence of the
internal degree of freedom raise or lower the condensation temperature?
Problem 7.1S. Compute the Clausius-Clapyron equation for an ideal Bose-Einstein gas
and sketch the coexistence curve. Show that the line of transition points in the P-v plane
obeys the equation

Problem 7.19. Show that the pressure, P, of an ideal Bose-Einstein gas can be written in
the form P = au, where u is the internal energy per unit volume and a is a constant.
(a) What is u? (b) What is a?

426

EQUILIBRIUM STATISTICAL MECHANICS

Problem 7.20. Electrons in a piece of copper metal can be assumed to behave like an
ideal Fermi gas. Copper metal in the solid state has a mass density of 9gr / em". Assume
that each copper atom donates one electron to the Fermi gas. Assume the system is at
T = 0 K. (a) Compute the Fermi energy, CF, of the electron gas. (b) Compute the Fermi
"temperature," TF = erlkeProblem 7.21. The density of states of an ideal Fermi-Dirac gas is
g(E) =

{g

if E
if E

> 0,
< 0,

where D is a constant. (a) Compute the Fermi energy. (b) Compute the heat capacity at
very low temperature.
Problem 7.22. Compute the magnetization of an ideal gas of spin-! fermions in
the presence of a magnetic field. Assume that the fermions each have magnetic moment
/-te. Find an expression for the magnetization in the limit of weak magnetic field and
T-+OK.
Problem 7.23. Show that the entropy for an ideal Fermi-Dirac ideal gas (neglecting
spin) can be written in the form

},

Converted with
where(nl) =
Problem 7.2
isothermal c
that the ferrn
now conside

STOI Converter

in the pressure and


ermion gas. Assume
hless, (Note: You are

trial version

hDP://www.stdutilitv.com

~-----------------------------------

Problem S7.1. Show that near the critical temperature the gap function, ~(T), in a
weakly coupled, condensed Fermi fluid (superconductor) in the mean field approximation has temperature dependence

~(T)

~(O) = 1.74

(T)

1/2

1 - Tc

'

where T; is the critical temperature and ~(O) is the gap function at T = 0 K.


Problem

S7.2. The unitary matrix,

Hamiltonian

Kk == (~~

Uk

~k).
-ck

U. '"

Compute

(Uk, ""),
Vk ;;

Uk~k

diagonalizes the effective

8
ORDER-DISORDER TRANSITIONS
AND RENORMALIZATION THEORY

S.A. INTRODUCTION
In previous chapters we have used mean field theory to construct a variety of
models of eq ., ..
ase transitions.
In this chapte
niversal theory
of critical phe
order-disorder
transitions.
an equilibrium
A useful
system is lin
be equilibrium
trial version
systems with
thermodynamic
quantities of i
the magnetization and allow
em. The way in
which a system responds to an external field is determined by the type of
fluctuations which occur in it. Indeed, as we shall see, the response function can
be expressed directly in terms of the correlation functions for equilibrium
fluctuations. In this chapter we shall use time-independent linear response
theory to obtain a relation between the long-wavelength part of the equilibrium
correlation functions and the static response functions. We will then use mean
field theory to show that near critical points, fluctuations become correlated
over a wide region of space, indicating that long-range order has set in.
If we are to describe the thermodynamic behavior of systems as they
approach a critical point, we must have a systematic way of treating
thermodynamic functions in the neighborhood of the critical point. Such a
m_ethodexists and is called scaling. We can write the singular part (the part
affected by the phase transition) of thermodynamic functions near a critical
point in terms of distance from the critical point. Widom was first to point out
that as the distance from the critical point is varied, thermodynamic functions
change their scale but not their functional form. The idea of scaling can be
eXpressed mathematically by saying that the thermodynamic functions are
homogeneous functions of their distance from the critical point. As we shall see,
the idea of scaling underlies all theories of critical phenomena and enables us to
obtain new equalities between various critical exponents. The scaling behavior

STOI Converter
hDP://www.stdutilitv.com

428

ORDER-DISORDER TRANSITIONS AND RENORMALIZATION THEORY

of thermodynamic functions near a critical point has been verified experimentally.


Kadanoff was able to apply the idea of scaling in a very clever way to the
Ising model and in so doing opened the way for the modern theory of critical
phenomena introduced by Wilson. The idea behind Kadanoff scaling is the
following. As the correlation length increases, we can rescale (increase) the size
of the interacting units on the lattice. That is, instead of describing the lattice in
terms of interacting spins, we describe it in terms of interacting blocks of spin.
We take the average spin of each block and consider it as the basic unit. As the
system approaches the critical point, the correlation length gets larger and the
block size gets larger in such a way that the thermodynamic functions do not
change their form, but they do rescale. With this picture, Kadanoff was able to
find a relation between the critical exponents associated to the correlation
length of fluctuations and the critical exponents which were introduced in
Section 3.H.
Wilson carried Kadanoff's idea a step farther and introduced a means of
computing critical exponents microscopically. Wilson's approach is based on a
systematic ~~lirul....OlLt.tlt.e....f~:c.tt~HaID.1.l1rntJ.ian.___yil.b..lJc.h_t1.ellQibes
a system near
the critical
critical point, one
repeatedly
tions and requires
that the Ha

STOU Converter

leads to nonlinear
on different length
f these recursion
trial version
relations.
earized about the
ressed
in terms of
fixed point
the critical exponen s. ere ore, 1 we can n
e eigenva ues, the problem is
solved. In this chapter we will show how Wilson's theory can be applied to
some simple models, such as the triangular planar spin lattice, and in the special
topics section section we will apply it to the S4 model. We will leave some
examples for homework problems.
Finally, as a special topic, we shall derive Onsager's exact expression for
the heat capacity of a two-dimensional square planar Ising spin lattice, and
we shall show that this system does indeed have a critical point. The method
we use to derive Onsager's original result is not the same as that used by
Onsager. Instead we follow a procedure developed by Kasteleyn and by Fisher
using dimer graphs. The derivation of the Onsager result is then fairly
straightforward.

http://www.stdutilitv.com

S.D. STATIC CORRELATION FUNCTIONS AND


RESPONSE FUNCTIONS [1, 2]
In this section we will investigate some general properties of static correlation
functions and response functions in an equilibrium system. These will prove
useful when we discuss scaling properties of equilibrium systems in subsequent

429

STATIC CORRELATION FUNCTIONS AND RESPONSE FUNCTIONS

sections. We will first give some general relations and then use mean field
theory to help build intuition about their qualitative behavior.

S.B.l. General Relations


Let us consider a system of N particles in a container of volume V at a fixed
temperature T. We will assume that the ith particle has spin Si, magnetic
moment u, momentum operator Pi' and position operator 1;. The magnetization
density operator is defined as
N

m(r)

= J-l

L S; 0(4; -

(8.1)

r).

i=1

The total magnetization operator is

1\1 =
If a magnetic
written

Jv drm(r)

= J-l

(8.2)

Si.

i=1

Converted with

~ltonian may be

STOI Converter

(8.3)

trial version
where flo is

hDP://www.stdutilitv.com
(8.4)

In Eq. (8.4), the first term on the right is the kinetic energy. The second term on
the right is the interaction energy between particles, and the summation is over
all pairs of particles.
Let us now assume that the applied magnetic field, B, is constant throughout
the volume, V. The average magnetization in the presence of this field is

(8.5)
If we let!VIo:denote the ath component of the magnetization operator,
ex = x, y, Z, then we can write for the static magnetic susceptibility

Xo:,o:l = (8~MO:)B)
'B0:1

= f3((Mo:MO:/) - (Mo:)(Mo:/))
T,N,B=O

f3((Mo: - (Mo:))( Mn' - (Mo:/) )).

M, where

(8.6)

430

ORDER-DISORDER TRANSITIONS AND RENORMALIZATION THEORY

Note that the susceptibility as we have defined it is independent of the applied


magnetic field. It contains information about the thermodynamic properties of
the unperturbed system.
Let us now introduce the magnetization density fluctuation,
(r) =
mo(r) - (mo(r)). Then the static susceptibility can be written

omo

Xa,a'

= f3

Iv dr, Iv dr2(Oma(rl)Oma,(r2))

== a

Iv dr, Iv dr2Ca,a,(rl,r2),
(8.7)

where

(8.8)
is the static spatial correlation function between magnetization density
fluctuations at points rl and r2 in the system.
For systems with very large volume, V, we can neglect boundary effects if
there are no spatially varying external forces present, and the static spatial
correlation
ment, r = rl - r2,
of the two
Converted with

STOI Converter

(8.9)

trial version

The static s

hDP://www.stdutilitv.com
1'\.0,0

tJ

J v ...........

0,0 \ .. l

(8.10)

It is useful to introduce yet another quantity, the static structure factor,


(8.11 )
This can also be written

Go,o/(k)

=.!.J v dr, Jv dr2e,'}((rt-rl)(omo(rl)8mo/(r2))


V

(8.12)

= ~ (8mo(k)8mo/(-k)).

In Eq. (8.12) we have made use of the Fourier decomposition of the spatially
varying fluctuations
(8.13)

STATIC CORRELATION FUNCTIONS AND RESPONSE FUNCTIONS

431

If we now compare Eqs. (8.7) and (8.12), we see that the static susceptibility
can be written in terms of the infinite wavelength component of the static
sDnlcturefactor
Xo,o'

= {3VGo,o,(k = 0).

(8.14)

Thus, we see that the way in which a system responds to a constant external
field is completely determined by the long-wavelength equilibrium fluctuations.
In Chapter 3 we found that the static susceptibility becomes infinite as we
approach the critical point. Thus, Eq. (8.14) tells us that near the critical point
the correlation function will have a large long-wavelength component,
indicating that long-range order has begun to occur.

8.B.2. Application to the Ising Lattice


Consider now a three-dimensional Ising lattice with volume V. Assume that
the lattice has N lattice sites. Let us break the lattice into blocks of volume ~,
so each block has n = ~ V N s ins. Let us assume the lattice has a
temperature,
1 magnetization
Converted with
is zero, (M)
agnetization of
the Zth block .
values ranging
from mi =on is zero, we
must have
that n is large
trial version
and each blo
Iso assume that

L:

STOI Converter
hDP:llwww.stdutililV.com

(ml) = O.
Let us no
ean field theory
(cf. Section 3.G). We can write a phenomenological expression for the partition
function
ZN(T) =

(8.15)

e-V{m/},

{m/}

where {ml} is the free energy density of the system, e-V{m/} is the probability
of finding the lattice in a configuration {ml}, and the summation is over all
possible configurations of the lattice. Since we require that (ml) = 0, {ml}
must be an even function of mi.
If we assume that the lattice is very large, then we can let the discrete spatial
Variation of the local magnetization density become continuous, ma:~ om(r).
For small fluctuations away from equilibrium we can write
{om(r)}

= {3a(T) +!C1(T)

+ !C2(T)
2

J dr(om(r))2
V

Jv dr(\lom(r))

(8.16)
(\lom(r))

+"',

432

ORDER-DISORDER TRANSITIONS AND RENORMALIZATION THEORY

where a( T) is the nonmagnetic free energy density. Let us next note that

Jv dr(8m(r))2

..!..
V

2: 8m(k)8m(

-k),

(8.17)

where 8m( -k) = 8m* (k) and

Jv dr(V8m(r)).

(V8m(r))

= ..!..

2: k28m(k)8m(

-k).

(8.18)

The free energy can then be written

{8m(k} = {3a(T)

+ -2

2:(C (T) + k2C2(T) )8m(k)8m(


I

-k)

+ .. '.

(8.19)

We can use this free energy to obtain the probability for a fluctuation, 8m(k), to
occur. It is

Converted with
where Cis
we can co

STOI Converter
trial version

(8.20)
sity in Eq. (8.20),

hDP://www.stdutilitv.com
(8.21 )

The static susceptibility is given by

{3V
X = {3VG(k = 0) = Ct
Near a phase transition, the susceptibility behaves as X ~ (T - Tefl
Section S3.C). Therefore, CI ~ (T - Te).
The static correlation function is given by

(8.22)
(cf.

The correlation function has a length, ~ JC2/CI. Since CI ~ (T - Tc) near a


critical point, the range , ~ JC2/(T - Tc) goes to infinity as (T - Te)-1/2 as

433

SCALING

we approach a critical point. Therefore, at the critical point the spatial


correlations between fluctuations extend across the entire system.

s.c.

SCALING [1, 3]

As we approach the critical point, the distance over which fluctuations are
correlated approaches infinity and all effects of the finite lattice spacing are
wiped out. There are no natural length scales left. Thus we might expect that in
the neighborhood of the critical point, as we change the distance from the
critical point (for example, by changing the temperature), we do not change the
form of the free energy but only its scale.
The idea of scaling underlies all critical exponent calculations. To understand scaling, we must first introduce the concept of a homogeneous function.

S.C.I. Homogeneous Functions


A function F(

Ax)

is homoe:eneous if for all values of A. we obtain

Converted with
The general

(8.24)

STOI Converter

~e first note that

trial version
so that

(8.25)

hDP://www.stdutilitv.com
g(AJi-)

= g(A)g(Ji-).

If we take the derivative with respect to

(8.26)

u; we find
(8.27)

where g'(Ji-)

= dg(Ji-)/dJi-.

We next set Ji- = 1 and g'(l)


Ag' (A)

= p.

= pg(A).

Then
(8.28)

If we integrate from 1 to A and note that g( 1) = 1, we find

(8.29)
Thus,
F(Ax) = VF(x)

(8.30)

434

ORDER-DISORDER TRANSITIONS AND RENORMALIZATION THEORY

and F(x) is said to be a homogeneous function of degree p. A homogeneous


function has a very special form. In Eq. (8. 30), if we let A = x-I, we obtain

F(x) = F(1)xp.

(8.31)

Thus, the homogeneous function F(x) has power-law dependence on its


arguments.
Let us now consider a homogeneous function of two variables f(x, y). Such a
function can be written in the form
(8.32)
and is characterized by two parameters, p and q. It is convenient to write f(x, y)
in another form. We will let A = y-I/q. Then
(8.33)
and we se
through th
these ideas

8.C.2. W

Converted with

STOI Converter

s on x and y only
e can now apply
int.

trial version

hDP://www.stdutilitv.com

As we hav
sition occurs in a
system, p
er which leads to
singular behavior in some of the thermodynamic response functions. If we
assume that the "singular" part of the free energy scales, then we can find a
relation between various critical exponents which agrees with experiment. We
will consider magnetic systems since they give a simple picture, and we shall
assume that a magnetic induction field, B, is present. Let us write the free
energy per lattice site in terms of a regular part, gr(T, B), which does not
change in any significant way as we approach the critical point, and a singular
part, gs(c, B), which contains the important singular behavior of the system in
the neighborhood of the critical point. Then

g(T, B)

gr(T, B)

+ gs(c,

B),

(8.34)

where e = (T - Tc)jTc and T; is the critical temperature.


We shall assume that the singular part of the free energy is generalized
homogeneous function of its parameters,
(8.35)
We now write the free energy as a function of the magnitude of B. For

435

SCALING

the systems we consider here, its direction does not play an important
role. The critical exponents in Section 3H can all be determined in terms
of p and q. Let us first find an expression for (3, which is defined [cf.
Eq. (3.85)] as

M(c:,B = 0)

rv

(-c:) fJ.

(8.36)

If we differentiate Eq. (8.35) with respect to B, we obtain


(8.37)
If we next let A

= (-c:r1/p

and set B

= 0,

we obtain

M(c:,O) = (-c:/1-q)/PM(-1,0).

(8.38)

Thus,
I-a

(8.39)

Converted with
and we obtai
Let us nex
which is defii

STOI Converter

ritical isotherm),

trial version

hDP://www.stdutilitv.com

(8.40)

If we set e = 0 and A = B-1/q in Eq. (8.35), we can differentiate with respect to


B and obtain

M(O,B) = B(I-q)/qM(O, 1).

(8.41 )

q
8=-1-q

(8.42)

Thus,

and we obtain our second relation.


The magnetic susceptibility is obtained from the thermodynamic relation
(8.43)
By differentiating Eq. (8.37) with respect to B, we can write
(8.44)

436

ORDER-DISORDER TRANSITIONS AND RENORMALIZATION THEORY

If we now set B = 0 and let A =

(crl/p, we find

X(c,O) = c(I-2q)/PX(I,

0).

(8.45)

Thus, the critical exponent for the susceptibility is


2q - 1
p

,=--,

(8.46)

and we obtain our third relation between p and q and the critical exponents. By
a similar calculation, we find that , = , '.
The heat capacity at constant B is given by
(8.47)
[cf. Eq. (3.86)]. From Eq. (8.35), we obtain

Converted with
If we set B

STOI Converter
(8.49)

trial version
and therefo

(8.48)

hDP://www.stdutilitv.com
a=2--

1
p

(8.50)

is our fourth relation. By a similar calculation we find a = a'.


In Eqs. (8.39), (8.42), (8.46), and (8.50), we have obtained the four critical
exponents, a, (3", and 8, in terms of the two parameters p and q. If we combine
Eqs. (8.39), (8.42), and (8.46), we find

,'=,={3(8-1).

(8.51)

From Eqs. (8.39), (8.42), and (8.50) we find


a + (3( 8 + 1) = 2.

(8.52)

Thus, the Widom scaling assumption allows us to obtain exact relations


between the critical exponents. These relations agree with mean field theory
(a = 0, (3 =!, 8 = 3, , = 1) as one can easily check. They also agree with
experimentally obtained values of the critical exponents which generally differ
from mean field results (cf. Table 8.1).

437

SCALING
For later reference,
exponents. We find

it is useful to express p and q in terms of the critical

(8.53)

P=fj(8+1)
and

(8.54)

The scaling property for systems near the critical point has been verified
experimentally for fluids [5] and magnetic systems [1].

8.C.3. Kadanoff Scaling [6]


Kadanoff has shown how to apply the idea of scaling to the Ising model. Let us
consider a d-dimensional
Ising system with nearest-neighbor
coupling (I'
nearest neigh......,.....,__,_____.__.,'-"'--"...t..I.JI:o:u...u""'"""'o..a..n_....1J:I.._-------~

Converted with

STOI Converter

(8.55)

trial version

where N is th
e into blocks of
length La, w
We choose L so
that La were
IS e corre a Ion eng 0 spin uc ua Ions on the lattice
[cf. Eq. (8.23)]. The total number of spins in each block is tr. The total number
of blocks is NL -d. The total spin in block I is

hDP:llwww.stdutililV.com

S]' =

LSj.

(8.56)

ta

e,

Since L is chosen so that La


the spins in each block will be highly
correlated and it is likely that they will be aligned to some extent. In view of
this, it is useful to define a new spin variable, S], through the relation
(8.57)
where S] = 1 and Z = LY .
. Spins interact with nearest-neighbor
spins, so blocks should also interact
With nearest-neighbor
blocks. Thus, the block Hamiltonian will be of the form
rNL-d

H{SL}

= -KL

L
(IJ)

/2

NL-d

SIS] - BL

L S]
]=1

(8.58)

438

ORDER-DISORDER TRANSITIONS AND RENORMALIZATION THEORY

,
-,--,- -- ..
..
--------t-------.. ', .... ,, ..
-,

,..

-I

.,. - . -I

.1 __

.1.

'.. .a

_:.

-- ~ '; -. -_- :-r:-.- ;- :-,-:-. - -

.: - - .:. .:. I4a


.,.

_~

_:I~ ~ _._
.,.

,-

-_l ~ ~:....:

.,-

.'

-I

Fig. 8.1. Decomposition of a square lattice into square blocks whose sides have length
= 4a.

La

ghbor blocks. The


n except that all
nergy per block,
y per site, g(c, B).

,1b.f~~e.ttf~~Ln.tfrrru:non.lle.rnLe.eltL.rue.ane.s1::lli4

Converted with

quantities
g(cL' BL), t

Since there

STOI Converter
trial version

(8.59)

hDP://www.stdutilitv.com

If we reseal
ther than sites, we
reduce the effective correlation length (measured in units of La) and therefore
we move farther away from the critical point. Thus, the correlation length will
behave as
(8.60)

Since rescaling moves us away from the critical point, the temperature
magnetic field B must also rescale. We assume that

and

(8.61)

where x is positive. Similarly,


N
BLSj
i=1

NL-d
=BLLSi
1=1

NL-d
=BLSI'
iel

1=1

NL-d
= BZLS[,

(8.62)

1=1

so that
BL=BZ=UB.

(8.63)

439

SCALING

Equation (8.58) now becomes

g(Ve,UB)

= Ldg(e,B).

If we compare Eq. (8.64) with Eq. (8.35), we find x


q

(8.64)

= pd

and

<1

= qd. Thus,
(8.65)

in agreement with experiment


The Kadanoff view of scaling allows us to introduce two new critical
exponents which are associated with the spatial correlations of spin fluctuations
in the system. The block correlation function is defined
(8.66)
where ri. is the distance between blocks I and J in units of La. We can write
Eq. (8.66) as
C(rL' c

Converted with
(8.67)

STOI Converter

e),

trial version

hDP:llwww.stdutililV.com

The distances ri

where ris the


and r are rel.l1""'_~-;r-"_~r----~~---------

(8.68)
and we write
(8.69)
If we choose L = r/ a, the correlation function takes the form

( r)

C(r,e) = ~

2 (y-d)

C a,e

(r~ )X)

(8.70)

We can now introduce two new exponents for the correlation function. We first
define a critical exponent, t/, for the correlation length as

(8.71)

t.

For ~ean field theories V = From Eq. (8.23) we see that the correlation
function away from the critical point depends on r and e in the combination
r/~ :::::
rev. In Eq. (8.70) the correlation function depends on r and e in the

440

combination

ORDER-DISORDER TRANSITIONS AND RENORMALIZATION THEORY

r.

Thus,
x = pd = y-l.

At the critical point, e

(8.72)

0 and the correlation function varies as


C(r,O)

r-..J

(r)2(y-d).

(8.73)

In three dimensions, we expect the correlation function at the critical point to


behave as
C(r,O)

r-..J

(1)
-;.

1+1]

(8.74 )

where 1] is another new exponent. For mean field theories,


dimensions, C(r,O) varies as
,..,_L

_.ill

1]

j_1 \ (d-2+1])
~

= O. In d

(8.75)

Converted with
and we cat

STOI Converter

(8.76)

trial version
Thus, the e
exponents ""'---_h_D_P_:------.,WWW
__
Eqs. (8.50) and (8.72) we find

II

-1-

d
...,.--SI_U_I_I_ltv_.c_o_m_------:7V

ten in terms of the


considered. From

2-0:
d

y=--

(8.77)

and from Eqs. (8.54) and (8.76) we find,

~=2-dG~:) =2-2::1'

(8.78)

Thus, Kadanoff scaling allows us to obtain two new critical exponents and new
identities between all the exponents.

S.D. MICROSCOPIC CALCULATION OF CRITICAL


EXPONENTS [3, 7-11]
The Kadanoff picture of scaling was given firm mathematical foundation by
Wilson [12], who developed a technique for computing the critical exponents
microscopically. We shall outline Wilson's procedure for the case of spin
systems, and then in Exercise 8.1 we apply it to a triangular planar lattice.

441

MICROSCOPIC CALCULATION OF CRITICAL EXPONENTS

Let us consider a system described by the partition function

Z(K,N)

(8.79)

{Si},N)].

= Lexp[-Jr(K,
{Si}

The effective Hamiltonian, Jr(K, {Si},N) (which includes temperature), can be


written in the form
(1)

Jr(K, {Si},N) = Ko

+ x, LSi + K2 LSiSj

(2)

+ K3 LS;Sj

(i,j)

(i,j)

(8.80)

(1)

+ K4

SjSjSk
(;,j,k)

+ ... ,

where K is an infinite-dimensional vector containing all coupling constants, and


the summation 2:(;) means that only (ith) nearest neighbors are included. The
coupling co'
e Ising model,
Kl = -{3B,
Converted with
between spins,
and K3 = K4
We can in
ach block. Thus,

STOI Converter
trial version

Z(K,

._____---r-h-----.:n77'<""l:p:....--II--.:rr:www~.,..........,.S....-r-td....,......U..,......"ti-litv~.-cz:rr_0m~........"...J

d).

(8.81 )

Since the new partition function has the same functional form as the old one, we
can write the following expression for the free energy density per site:
1

g(K) = lim -N InZ(K,N) = lim -lnZ(KL,NL-d)


N-+oo

Nr-+o:

= L-dg(KL).

(8.82)

The coupling constant vectors, K and KL, of the site spin and block spin
systems, respectively, will be related by a transformation,

KL = T(K),

(8.83)

where KL will be a vector whose elements are nonlinear functions of the


components of K. Since our new Hamiltonian is identical in form to the old
one, Wecan repeat the process and transform to even larger blocks nL. After n
transformations we find
KnL

= T(K(n-l)L)'

(8.84)

442

ORDER-DISORDER TRANSITIONS AND RENORMALIZATION THEORY

If the system is not critical, there will be a finite correlation range. Thus, when
we transform to larger blocks the effective correlation range appears to shrink
and we move away from the critical point. However, when the system is critical,
the correlation range is infinite and we reach a fixed point of the transformation.
At the fixed point, the transformation T can no longer change the vector K.
Thus, the critical point occurs for values of the coupling constant vectors, K*,
which satisfy the condition
K* = T(K*).

(8.85)

The sequence of transformations T is called the renormalization group


(although T only has properties of a semigroup). The motion of the vector K
in coupling constant space is similar to the motion of the vector describing a
state point in phase space, except that the motion of K occurs in discrete jumps
rather than continuously.
It is useful to illustrate the possible motions of K for the case of a twodimensional vector, K = (KI' K2)' Equation (8.84) can be thought to describe
the motion of K in K space (the space of components of K) as we change block
size. To loc
..
fixed points of Eq.
(8.84). The
Converted with
classify the fixed
.1

points of cl
transformat
how the vee
Eq. (8.84) ~
for small 8

_.YZ.

_.1.

STOI Converter
trial version

L1

fixed point of the


fVillwant to know
We must linearize
t= (K - K*). Then

hDP://www.stdutilitv.com
(8.86)

where

(8.87)

We next find the eigenvalues and eigenvectors of the matrix A. Since A, in


general, will be nonsymmetric, we must use the method of Section S.C. The
eigenvectors can be written
(8.88)
where

A is the matrix of eigenvalues,


(8.89)

MICROSCOPIC CALCULATION OF CRITICAL EXPONENTS

443

Fig. 8.2. A hyperbolic fixed point with its eigencurves and the flow of points in the
neighborhood of the fixed point.

and 8u is the

Converted with

STOI Converter

(8.90)

trial version
The eigen
the behavior of
trajectories in
e have drawn the
case of a hyperbolic fixed point and its eigencurves. Points along the
eigencurves move as

hDP://www.stdutilitv.com

8UnL,1 = (Alt8ul,
8unL,2 = (A2t8u2.

(8.91)
(8.92)

Thus, for A > 1 the point moves away from the fixed point under the
transformation, and for A < 1 it moves toward the fixed point. The dashed lines
represent the trajectories of points which do not lie on the eigencurves. For a
hyperbolic fixed point, they will always move away from the fixed point after
many transformations. All systems with vectors K lying on an eigencurve with
eigenvalue A < 1 are critical, since with enough transformations they will come
arbitrarily close to the fixed point. Such systems are said to exhibit
"universality." The behavior of a point along an eigencurve with A > 1 is
reminiscent of the actual behavior of noncritical systems. As we increase the
block size, we move away from the critical point. Thus, an eigenvalue A > 1 is
called relevant and its eigenvector is identified as one of the physical quantities
(s or B, for example) which measure the distance of the system from the critical
Point.

444

ORDER-DISORDER TRANSITIONS AND RENORMALIZATION THEORY

In general we write the singular part of the free energy density in terms of the
eigenvectors 8Ui and eigenvalues Ai as follows:

[cf. Eq. (8.82)]. This looks very much like Widom scaling. Indeed, for the case
of an Ising system for which there are two relevant physical parameters which
measure the distance of the system from the critical point, we expect that two of
the eigenvalues will be relevant, let us say Al > 1 and A2 > 1. If we compare
Eq. (8.93) with Widom scaling of the Ising model,
(8.94)
we can make the identification

BUt

=E

and

8U2

= B. Thus,
(8.95)

In .Al

(8.96)

Converted with
and

STOI Converter
trial version

If we now use

(8.97)

hDP://www.stdutilitv.com

......,
/
,'.
tical exponents
in terms of the relevant eigenvalues.
In Exercise 8.1 we use Wilson's theory to obtain approximate expressions for
the critical exponents of a triangular planar lattice [13].
n

EXERCISE 8.1. Compute the critical exponents for 3-spin blocks on the
triangular planar lattice for the two dimensional nearest-neighbor Ising
model. Retain terms to lowest order in (V).

(a)

445

MICROSCOPIC CALCULATION OF CRITICAL EXPONENTS

Answer: The Hamiltonian

can be written

(1)

H= -KLSiSj-BLsi.
i

ij

We assign to a single block (block I in the figure below) a spin given by


Sf = sign(s1 + s~ + sD. Therefore, Sf = + 1 for the following configurations:

(0: = 1; 1ii)(o: = 2; i 1i)(o: = 3; ii 1)(0: = 4; iii)


The internal spin degrees of freedom are defined,
For configurations defined above we have: a} = a;

aJ = Is1 + s~ + s~lo:
= aj = 1, and aj = 3.

~,
1~2
\

Converted with

STOI Converter
trial version
The partitioi

hDP://www.stdutilitv.com

~--------------------------~

Z(KL,N) = Lexp[-H(KL,{Sf})]
{Sf}

= LLexp[-H(K,{Sf,af})].
{Sf}

{CTi}

(2)
Now define

where

Ho(K,{S[,af})

(4)

= -KLLLSiSj

fief

jef

and
(5)

446

ORDER-DISORDER TRANSITIONS AND RENORMALIZATl ~

We next define the expectation

(A( {SI } ))

THEORY

value

{SI, a/} )exp[-Ho(K, {SI, a/})]


L:CTI exp[-Ho(K, {SI, a/})]

L:CTI A(
=

')

(6)

N ow note that

is], <TI} )]) (e~V)

= (~exp[-HO(K,

exp[-H(KL, {S]})]

(7)

(e(V)+![(V")~(V)'1 + ... ).

'" (~e~Ho)

We can perform the summations

[Zo(K)]M =

in the unperturbed

L exp[-Ho(K,

part to get

(8)

{SI, a/})],

CT

Converted with

where M
single blo

on function of a

STOI Converter

3 e -K .

trial version

hDP:llwww.stdutililV.com

The inter'actJrorr~TTf-=-=n:TY""ty;;-=~,-y;;-..---r"'lIe"""--avt~~alue
spin index is

(s~)

Zo(Krl

L s~exp[K(s~s~

of a single

+ s~s~ + s~sD]
(10)

CTI
=

Zo(K)-l

If we take the logarithm

H(KL' BL, {SI })

S/(e3K

(9)

e-K).

of Eq. (7), we obtain

= MIn

(Zo(K)) + (V) + 2:1 [(V 2 ) - (V 2 ) ] ...

= MIn (Zo(K)) - 2K
- 3

L(
I

e3K
e

3K

3K
( ~K
J e

+ -K)2
; -K
SISJ
+ e

(11 )

+ e-K
3 -K SIB + ... ,
+ e

where we have retained terms up to (V). Thus the Hamiltonian

for the blocks

447

MICROSCOPIC CALCULATION OF CRITICAL EXPONENTS


r

has been written in the form

H(KL,BL,{S/})

= -KLLS1SJ
IJ

-BLLSI.

(12)

: If we now compare Eqs. (1) and (12), we find


:

K -2K
L -

3K

( e3K

+ e _K)2
+ 3e-K

(13)

i and
I

(14)

Fixed points occur for (K* = 0, B* = 0) and for (K* = Kf, B* = 0) where Kf
is the solution to the equation

Converted with
(15)

STOI Converter

I and positive.
e fixed points

trial version

hDP://www.stdutilitv.com

8KL = 0.58K

and

8BL = 1.58B.

(16)

This fixed point corresponds to temperature, T = ooK.


Let us next consider the fixed point, (K* = 0.3356, B* = 0). If we let
K = 0.3356 + 8K and B = 8B and linearize Eqs. (13) and (14) in 8K and
8B, we obtain
8KL

1.6238K

and

8BL

= 2.1218B,

(17)

I so the eigenvalues are AK = 1.623 and AB = 2.121. The Widom scaling

exponents become
= In(AK)
p

2In (V3)

= 0.441

and

In (AB)
q = 2ln (V3) = 0.684.

(18)

448

ORDER-DISORDER TRANSITIONS AND RENORMALIZATION THEORY

The critical exponents are


1
a = 2 - - = -0.27,
p
2q -1
,=--=0.83,
p

(3 = 1- q = 0.72,
p

and

q
8=--=2.2

(19)

1-q

The exact solution of this two-dimensional Ising model yields (AK) exact = 1.73
and (AB)exact = 2.80. Thus, we are close for AK but our calculation of AB is
not very good. It is possible to carry the calculation to higher orders in (vn).
In so doing, more elements of the vector, K, are introduced and better
agreement with the exact results is obtained .

.... SPECIAL TOPICS


.... S8.A. Critical Exponents for the 54 Model [3, 7, 81
Let us com
partition fu

Converted with

lattice sites. The

STOI Converter
trial version

(8.98)

hDP:llwww.stdutililV.com
where K = (31 is the effective interaction and e is the vector indicating the
positions of various nearest neighbors of the site, D. The summation is over all
possible configurations of the lattice. We can change the summation to an
integration if we introduce a weighting factor, W(Sn) = 8(S! - 1). Then Eq .
(8.98)can be written

The partition function is now in a form which allows some generalization.


Let us choose the weighting function in a slightly different manner. We will
let
(8.100)
If we choose b = -4u, the W(Sn) takes the simple form

W(Sn) = e-u(S~-1)2,

(8.101)

449

SPECIAL TOPICS: CRITICAL EXPONENTS FOR THE S4 MODEL

Fig. 8.3. The weighting function,


W(Sn) = e-u(S;-1)2, for u = 1.

and gives a fairly good approximation to the Ising partition function, Eq. (8.98)
since it is peaked at So = l (cf. Fig. 8.3). With this weighting factor, the
partition function takes the form
Z(K, b, u)

Converted with

STOI Converter
The syste

s~ +

uS:)

(8.102)

trial version

hDP://www.stdutilitv.com
(8.103)

where the sum 2:~is restricted to positive values of e and 2:~= d for a cubic
lattice. The system described by Eq. (8.103) is called the S4 model of the Ising
system.
It is useful now to rewrite the Hamiltonian and partition function in terms
of continuous variables. To do this we first introduce a spin field. Let So denote
the spin at point x = na (we assume an infinitely large cubic lattice with
spacing, a, between lattice sites). Then we introduce the Fourier amplitude,
S(k),
S(k)

= ad

L Soe

ik

na.

(8.104)

The components, ki (i = 1, ... , d), of the wavevectors, k, can only take on


values in the range - ~ < ki < ~ since wavelengths smaller than the lattice

450

ORDER-DISORDER TRANSITIONS AND RENORMALIZATION THEORY

spacing are not allowed. It is also useful to note the identities


(8.105)

and
ad

~::">ik.na

8(k).

(8.106)

From Eq. (8.104), we can define a continuous spin field


(8.107)

The spin field, S(x), is dimensionless. Note also that S(x


need not be zero between lattice sites.
Let us nc
that
Converted with

Id, S(k). First note

STOI Converter

LL
o

= na) = So, but S(x)

trial version

hDP://www.stdutilitv.com

and

(8.109)
X

+ k2 + k3 + ka).

S(kl)S(k2)S(k3)S(kt)8(kl

The Hamiltonion then takes the form


H

=K

(_1 )d

2 27ra Jdk

IS(k)12Ieik.ea _

J
(2~rJ J J J

+ ~ (~+;

2d) (2!a)
dk,

dk2

S(kl)S(k2)S(k3)S(kt)8(kl

112

dklS(kW
dk3

(8.110)

dk.

+ k2 + k3 + ka).

Near the critical point, spin fluctuations begin to exhibit long-range order.
Therefore we only need to be concerned with long-wavelength (small-

451

SPECIAL TOPICS: CRITICAL EXPONENTS FOR THE S4 MODEL

wavevector) phenomena. If we note that leik.eo - 112~ k2a2 for small k (note
that
[k- el2 = k2), then the Hamiltonian can be written

E~

=~

U/I.) J dk(CI +
d

+;
X

C2k2)IS(k)1

(8.111)

(2~rJ J J J dI<.
dkl

dk2

dk3

S(kl)S(k2)S(k3)S(kt)8(kl

+ k2 + k3 + ka),

where
(8.112)
It is useful to rescale the spin fields. We will let S'(k) = S(k) (Ka2-d) 1/2,
r = (b/K - 2d)a-2 and u' = u(Ka2-d)-2.
The Hamiltonian then becomes

Converted with
H(r, u',

is'

STOI Converter

(8.113)

trial version

hDP://www.stdutilitv.com

+ ka),

The partition function is given by

Z(r, u', {S'}) = DS' eH("u',{S'}),

(8.114)

where DS' now denotes a functional integral over values of the spin variables,
S' (k). Equation (8.113) is the starting point for the renormalization theory, for
the S4 model. When u' = 0, the model is called the "Gaussian model". In order
to find a scaled Hamiltonian for the system, we follow a procedure similar to
that for the triangular lattice in Exercise 8.1, but now we work in wavevector
space and not real space.
. Our first step is obtain a block Hamiltonian. In order to do this, we will
llltroduce a scaling parameter. L > 1, and divide the momentum intervals
o <: kj <: 7r / a into short-wavelength intervals 7r / La <: kj < 7r / a and longwavelength intervals
< kj < 7r / La. We then separate the spin fields into
their long-wavelength and short-wavelength parts:

for

< < :a
kj

(i

1, ... , d)

(8.115)

452

ORDER-DISORDER TRANSmONS

AND RENORMALIZATION THEORY

and
for ~
La

< ki < ~ (i
a

(8.116)

= 1, ... , d).

From Eq. (8.107), we introduce the corresponding spin fields .9L(x) and O"L(X).
The spin field .9L(x) is defined as
.9(X)

(__!_) d ...

27r

dk.9L(k)eik-x

7r/La

(8.117)

-7r/La

and is slowly varying over a region the size of the block, while the spin field
O"L(x) is defined as
O"L(X) = ( -1
27r

S:

S:

S~:

)d

...

7r a
/

dkO"L(k)eikox,

(8.118)

-7r/La

where
dk ==
dk +
dk and O"L(X) is rapidly varying inside a block.
We now decompose the quadratic part of the effective Hamiltonian into its
long-wavelene:th and short-wavelene:th Darts and define the quantities

Converted with
*'o({

STOI Converter
trial version

-k),

(8.119)

),

(8.120)

hDP://www.stdutilitv.com

and

v({YL, 17L},

U ')

= u'
X

(2~) J... r;
4d

dk( dk2dk3dk.,

S'(k1)S'(k2)S'(k3)S'(kt)8(kl

+ k2 + k3 + ka),
(8.121)

where .9L(k) and O"L(k) are defined in Eqs. (8.115) and (8.116). We now can
introduce an expectation value defined as
(8.122)

and write the partition function in the form


Z(r, u ')

[J
= [J
=

DI7Le-Jt"o({

DI7Le -Jt"o(

Ud)] J
{Ud)] J

(e-V({.9"L,uL))

D9'

Le-Jt"o({.9"L)

D9'

Le-Jt"o( {.YL} )-(V)+![(V')-(V)'j

+ ... ,
(8.123)

453

SPECIAL TOPICS: CRITICAL EXPONENTS FOR THE S4 MODEL

where we have introduced a cumulant expansion are before. We are only


concerned with those parts of the partition function which contribute to the
singular part of the free energy. Thus, the term in brackets, which comes from
short-wavelength variations, can be neglected and the effective Hamiltonian
takes the form
1( 1)

ru u D = 2 27T'

Jf( {YL},

1f La
/
-1f/La dk(1C

+ r)YL(k)YL(

-k)
(8.124)

(V) - -1 [2(V ) - (V) 2]


2

+ ...

After a scale transformation, we seek a new Hamiltonian of the form


JIl'L( {.'I'd, TL, uL)

=~

(2~r

J::a

dkL(ki

+ TL)9'L(kL)9'L(

-kL)

4d

~----~~~~~~~~~~~~~d~L

(8.125)

Converted with

STOI Converter
trial version

If we can wri
the new coupl
obtain
(V)

hDP://www.stdutilitv.com

[J

DULe-JI"O({UL})]

{u J
1

-I

DULe-JI"o({uL)

n we can obtain
value, (V), and

(2~)

4d

... J1f/La dkldk2dk3d~


-1f/La

+ k2 + k3 +~)
1f La
1f a
+ 6u I J / dk1 J1f/La dk2 [J1f/a dk3 J / d~Y L(k1)Y L(k2)
-1f/La
-1f/La
1f/La
1f/La
x YL(kl)YL(k2)YL(k3)YL(~)8(kl

uL(k3)UL(I<.)6(kl

+u
X

J ...

6(kl

+ k2 + k3 + 1<.)]

1f a
J / dkl dk2dk3d~O'L(ki)O'L(k2)O'L(k3)O'L(~)
1f/La

+ k2 + k3 + 1<.)}.

(8.126)

The factor 6 enters because there are six terms with two factors of YL(k)

and

454

ORDER-DISORDER

TRANSITIONS AND RENORMALIZATION THEORY

two factors of UL(k). They each give identical contributions. The last term can
be neglected since it contains only short-wavelength contributions and
contributes to the regular part of the free energy. To evaluate the second term,
we use Eq. (7.27) to obtain
(uL(k3)UL(kt))

(27T/

+ kt).

= -k2
8(k3
3+r

(8.127)

Then the singular part of (V) takes the form

(8.128)

Converted with
onian, but it yields
uartic part can be

STOU Converter

We see tha
no correcti

obtained fr
trial version
sion.
Many t
1 [(V2) - (V)2]. To
obtain expl
orem (see Ex. 4.9).
We will n
ill summarize the
results. As Wilson and Kogut [8] have shown, the only term of interest in
[(V2) - (V)2] is the correction to the quartic term in the effective Hamiltonian.
They have shown that corrections to the quadratic terms can be neglected.
Furthermore, contributions to the second-order cumulant coming from pairwise
averages of spins which are both located on the same potential, V, give no
contribution since for those terms (V2) = (V)2 and they cancel out. Thus, we
only retain terms which have at least one pairwise average between spins on
different factors, V. The only term which involves four spins, g'L(k), and which
contributes to the second-order cumulant is

hUP:llwww.stdutililV.com

~ (V2)quartic = ~

[J
{ J J

DULe-Jf'O'L})r

DULe-Jt'o({.L})

(2~fi

J7r1a - dk3dktdksdk6
-7rILa
7rILa
X g'L(kl)g'L(k2)UL(k3)UL(kt)uL(k5)aL(~)g'L(k7)g'L(ks)
X

36u/2

+ ... }.

...

7rILa

dkl dk2dk7kS

...

(8.129)

455

SPECIAL TOPICS: CRITICAL EXPONENTS FOR THE S4 MODEL

The origin of the numerical factor, 36, can be understood if we note that there is
a factor,
in front of (V2)Quartic, there are six terms in each factor V with two
!I L 's and two ai, 's, and there are two different ways of taking pairwise
averages for each term in V2(~ x 6 x 6 x 2 = 36).
From the results of Exercise 4.9 and Section 7.C.l, we obtain

!,

7r/

J [l:

dk

( -2-1

) (

+r

(k3

+~ -

+r

k)

+ other

terms.
(8.130)

If we combine

contributions

from Eqs. (8.128) and (8.130), the Hamiltonian

,Jt( {YL}, ri- UL) in Eq. (8.124) takes the following form:

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com
\l7r J
X

J-7r/La

YL(kl)YL(k2)YL(k3)YL(~)
,

x [U

36u,2
(27r)d

8(kl

+ k2 + k3 +~)

7r a

/
7r/La

dk

k2

+r

) (

(k3

+ ~ _ k)2 + r

)]

+ ...
(8.131)
Where we have neglected some contributions (terms not containing block spins),
so~e second-order quadratic terms, and terms containing six or eight block
SPIllS. Also, we neglected all contributions
from higher-order cumulants.
To obtain Eq (8.131) in a form similar to the original effective Hamiltonian,
We must make a change of scale. We let
(8.132)
and
(8.133)

456

ORDER-DISORDER TRANSITIONS AND RENORMALIZATION THEORY

Then we find

(8.134)

Equation (8.134) is the transformed Hamiltonian that we have been looking for.
If we let 2
on for rt:

Converted with

STOI Converter
trial version

The expres
which is iJ
approximat-es-..

(8.135)
n expression for UL
we must make an

hDP://www.stdutilitv.com
--rr--.-.-.:;r---r<l:71l,..",..,..J,---,-.,...."<t,.,------.".-,.-----.-.-...----c>"O""""""nnr---____J

36u'2

=1'

UL

U' ---d

(27r)

(1)+

7r a
/

dk -k2

n[Ia

2]
r

(8.136)

'

where & = 4 - d. This approximation is made every time we change the scale
of the Hamiltonian. Eqs. (8.135) and (8.136) are the recursion relations we have
been looking for. They are correct as long as 1&1 1 (this will become clearer
below).
To find the fixed points of Eqs. (8.135) and (8.136), we will turn them into
differential equations. We first write the recursion relation for a transformation
from block L to block sL,
ra. =

s2

ri.

+ ~12u

(27r)

J ... J

7r a
/

dk

7r/sa

1]

+ rL

'

(8.137)

2] ,

(8.138)

and
UsL

S"

UL -~

J
(27r)
36u2

.. ,

7r a
/

[sa

dk

( -2-1

+ rt.

457

SPECIAL TOPICS: CRITICAL EXPONENTS FOR THE S4 MODEL

where s is now considered to be a continuous parameter. Let s = 1 + h for


h 1. Then, Eqs. (8.137) and (8.138) can be approximated by

r(1+h)L

(1 + 2h)

rL

12uL
+ --d

7r a
/

(27T-)

(
dk
-k2 1

+ ri.

(7r/a)(l-h)

)]

(8.139)

and

dk (1)
-k
+ rt.

7r a

U(1+h)L

(1 + tCh) [ UL

36u2 J /
--Ja
(27r) (7r/a)(l-h)

2]

(8.140)

We next evaluate the integrals. We can write for small h


_1 [7r/a
(27r)d J (7\"/a)(1-h)

dk(_I_)n
k2

= _2

+ ri.

(~)2 +rz

(27T/
/

)nJ

dk

7r a
/
(7r/a)(l-h)

\ n

Converted with

STOI Converter
where A is a (

(to first order in

trial version

h)

hDP://www.stdutilitv.com
(8.141)

and
U(1+h)L -

UL

= 8 hUL

3ui_Ah
2.

(8.142)

( (~\rL)
If Wenext divide by hL and take the limit h
dri,

LdL =2rd

---+

0, we find

AUL

((~\rL)

(8.143)

and
(8.144)

458
Now let t

ORDER-DISORDER TRANS mONS

== InL

AND RENORMALIZATION THEORY

and choose the units of a such that 'Tr/a = 1. Then

dr
-=2r+--

Au

(8.145)

1+ r

dt

and

du = 8u _

3Au

(8.146)

(1 + r)2

dt

Higher-order terms can be neglected as long as


Eqs. (8.145) and (8.146) occur when

181 1. The

dr"
Au*
-=2r*+--=0

1 + r*

dt

fixed points of

(8.147)

and

Converted with
There are
r" = -8/f
transforma

STOI Converter
trial version

(8.148)
and (u* = 8/3A,
pen the linearized

hDP://www.stdutilitv.com
d8r)

dt

( dbu

(2

-Au*
0

A(1 - r*)) (8r)


8 - 6Au*
Su

(8.149)

dt
and we obtain the eigenvalues Ar = 2 - Au* and Au = 8 - 6Au* for small u*
and r", Thus, Ar is a relevant eigenvalue. For the fixed point (u* = 0 and t" = 0)
the eigenvalues are Ar = 2 and Au = 8 (this is just the Gaussian fixed point),
and for the fixed point (u* = 8/3A and t" = -8/6) the eigenvalues are
Ar = 2 - 8/3 and Au = -8. For d > 4,8 < 0 and the second fixed point is
repulsive and therefore unphysical since both eigencurves are directed away
from the fixed point and it can never be reached. Thus, for d > 4, the physics is
governed by the Gaussian fixed point. For d < 4, the Gaussian fixed point is
repulsive and for this case the fixed point (u* = 8/3A, r: = -8/6) is the
physical one. For d = 4, the fixed points coalesce. In Fig. 8.4 we sketch the
trajectories in the neighborhood of the two fixed points.
We can now obtain expressions for the critical exponents a and v, but it
requires a bit of algebra. To obtain the equations for the eigencurves of the 54
model, we must rewrite the differential equations in Eq. (8.149) in terms of L.

SPECIAL TOPICS: CRITICAL EXPONENTS FOR THE S4 MODEL

459

(b)

(a)

Fig. 8.4. A sketch of trajectories in the neighborhood of the physical fixed points for the
~ model: (a) f(
(b) for d < 4 the
fixed point (u*
Converted with

STOI Converter
We first write
trial version
envalues. (Note
that the left a
matrix on the
right-hand sid~
not
necessary to find explicit expressions for the eigenvectors. If we let 8Ul (I)
denote the relevant eigenvector, we can write

hDP:llwww.stdutililV.COmrposes.itis

d8 U I (I)
dt
=

(2 - A U *) OUt
1:

()

(8.150)

The solution to Eq. (8.150) can be written


(8.151 )
If We now remember that t = InL, we find

(8.152)
so that Al = exp[(2 - Au*) In L].
obtain

We then make use of Eq. (9.E.18) to

_ (2

p-

-AU*)
d

(8.153)

460

ORDER-DISORDER TRANSITIONS AND RENORMALIZATION THEORY

Using Eqs. (8.50) and (8.153) we obtain


d
0'=2---2 -Au*

(8.154)

and using Eqs. (8.77) and (8.154) we obtain

1/=

2 -Au*

(8.155)

Since the fixed points which characterize the system are different for d
d < 4, the critical exponents will also differ for those two cases.
Let us first consider the case d > 4. Then u: = 0 and we find
d
8
a=2--=-

> 4 and

(8.156)

2'

where we have used the fact that 8 = 4 - d. We also find

Converted with
When d

<

(8.157)

STOI Converter
trial version
(8.158)

hDP://www.stdutilitv.com
and

12

1/~-+The magnetic

critical exponent can also be computed

(8.159)
and the result is
(8.160)

as for the case d > 4. The other exponents can be obtained from the identities in
Eqs. (8.51), (8.52), and (8.78). In Table 8.1 we compare the results of the S4
model with the experimental results, with mean field results, and with the
results due to exact calculations for the Ising model. The first thing to note is
that, for d = 4, the mean field theories and the S4 model give the same results.
For d = 3, the S4 model gives very good agreement with the experimental and
Ising values of the exponents (the exact results for the three-dimensional
Ising
are obtained from a numerical calculation). However, there is really no reason
why it should. We have retained only the lowest-order approximation for the
exponents in the 8 expansion. If we take 8 = 1 (as it is for d = 3), then the
expansion need not converge. In fact, higher order terms in the series give large

SPECIAL TOPICS: CRITICAL EXPONENTS FOR THE S4 MODEL

['~
-('f"l

['00

O-IN('f"l

_-INO

-I/")

oo~~oo

~
~~
~

'<t

c~e
J
E-4
rI>

:go~
~A

~
C"I~w

............
I

+ _-INO

W~('f"l

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com

C"I~

~['

OOI/")~O_

6~J~~o
o ~o

oci

461

462

ORDER-DISORDER

TRANS mONS

AND RENORMALIZATION THEORY

contributions and ruin the agreement. For this reason the S expansion is thought
to be an asymptotic expansion-that is, one for which the first few terms give a
result close to the result obtained by summing the entire series.
One may well ask the meaning of an expansion about d = 4 and in terms of
noninteger d, when real physical systems correspond to d = 1,2, or 3. However
consideration of these "unphysical" values of d provides us with a theoretical
tool for understanding the role of the dimension of space in physical
phenomena, and it has led to a major advance in our understanding of critical
phenomena .

.... S8.B.EXACTSOLUTION OF THE TWO-DIMENSIONAL


ISING MODEL [14, 15]
The two-dimensional Ising model is one of the simplest systems that exhibit a
phase transition and one of the few that can be solved exactly. The phase
transition in this system is an order-disorder transition. It has served as the
paradigm system for describin order-disorder transitions in a variety of
contexts. In p
Converted with
sition from the
paramagnetic
gnetic crystals.
However, it h
to model le

STOI Converter
trial version

changes due t
DNA [19]. It

t has been used


al networks. In
conformational
denaturation of

hDP://www.stdutilitv.comsolation[20].In

this section whilJajrrtrse-~rrr[[S""""PIIysIics-cmne:ITl[UJmoUe11U1uer-disorder


in a
spin system.
An analytic expression for the partition function of an Ising system can be
obtained in one and two dimensions, but no one has succeeded in solving it
analytically in three dimensions. In one dimension it does not have a phase
transition. In two dimensions, it does have a phase transition. In Section 7.F we
considered the one-dimensional Ising model and the mean field approximation
to the Ising model. In this section we will consider the case of a planar Ising
spin lattice and show how an analytic expression for the partition function may
be obtained and how the phase transition appears [21].

.... S8.B.l. Partition Function


Let us consider a planar lattice of N spin-! objects with magnetic moment, /-1. A
planar lattice is one which can be laid out flat in a two-dimensional plane
without any bonds intersecting. For example, a lattice with periodic boundary
conditions is not planar. Each lattice site can interact with the magnetic fields
of its nearest neighbors. Some examples of planar lattices are shown in
Fig. 8.5.

463

EXACT SOLUTION OF THE TWO-DIMENSIONAL ISING MODEL

(c)

(b)

(a)

Fig. 8.5. Three types of planar lattice. (a) Square lattice. (b) Triangular lattice. (c)
Hexagonal lattice.

The Hamiltonian

for a two-dimensional

planar lattice can be written


- 1),

H = - LJ(SiSj
{ij}

(8.161 )

where ~{ij} de'~~....I.!A.!o"---..loU.u.y,_....!o.!...L.!oa.._"""""''''''''''''''l..I<.....AJ'''''''''''~~,"----,,--..>Y...IJ''''--L


the magnetic
interaction ene
component of
spin at the ith I
shifts the zero

STOI Converter

point of energ
objects, Sj = +
direction. If J
trial version
sites have spin
neighboring sp
The partitio'mum:;rn:JITlrurLmrs-tmurar1aIn:cc~m-oe-wrIIIe:n-

s. For spin-i
ve (negative) z
all the lattice
tion in which

hDP://www.stdutilitv.com

ZN(T) = Lexp (LK(SiSj


a.c.
(ij)
= e-NnnK L

.,.

sl=l

1))

II eKs;Sj,

L
sN=l

(8.162)

(jj)

where K = (3J, ~a.c. denotes


the sum over all 2N possible
different
configurations of spin on the lattice, and Nnn is the number of nearest neighbor
pa~rs. The sum, ~(ij) and the product, I1(ij) , are taken only over nearest-neighbor
parrs. Let us now note that since s, = 1, we can write
eKS;Sj= cosh(K)
Therefore,
ZN(T)

the partition

+ sjsjsinh(K)
function

= e-NnnK(cosh(K)tnn

= cosh(K) [1 + tanh(K)sjsiJ.

(8.163)

can be written in the form


L
sl=1

...

L
sN=l

II[1 + tanh(K)siSj].
(ij)

(8.164)

464

ORDER-DISORDER TRANSITIONS AND RENORMALIZATION THEORY

To evaluate the partition function, we must expand the product and then sum
over each spin variable, s.. There are several properties of the spin that should
be noticed and that simplify the expression for the partition function. First note
that

L s;n = 2

and

L s;n+l = 0,

(8.165)

i=l

i=l

where n is an integer (including zero). Thus, only even powers of the spin
variable can appear in the partition function. Also, only nearest-neighbor pairs
can appear and a given nearest-neighbor
pair can appear only once. Each
summation over a spin variable yields a factor of 2. Expansion of the product in
Eq. (8.164) yields a sum of terms depending on varying powers of tanh(K).
After summation over spins is performed, each term in the sum can be
represented by a picture which we call a closed graph. We then can write the
partition function in the form
(8.166)

Converted with

The close
(1) Lay 0
lattice
sites.

STOI Converter
trial version

original planer
between lattice

hDP:llwww.stdutililV.com

(2) Find a
between lattice
sites so ey orm c ose
oops any num er 0 c ose loops including
zero). We only require that no nearest-neighbor bond appears more than
once. There may be more than one closed loop on a graph, and they may
intersect, as long as they have no bonds in common (cf. Fig. 8.6) for an
example of an allowed and forbidden graph).

1
(a)

(b)

Fig. 8.6. Some closed graphs on a 3 x 3 square lattice. (a) An allowed graph. (b) A
forbidden graph because two closed loops share a bond.

EXACT SOLUTION OF THE TWO-DIMENSIONAL ISING MODEL

465

(3) A given graph contributes a factor, tanhr(K), where r is the number of


bonds in the graph.
An example of the use of these rules is given in Exercise 8.2
In writing Eq. (8.166), we have made progress toward obtaining an
analytic expression for the partition function, but we still have a ways to go.
In the next section, we will change the subject slightly and show how to
represent an anti symmetric matrix in terms of a directed graph. We will
also show how to represent the determinant of an antisymmetric matrix
(actually its square root) in terms of dimer graphs. Then in Section S8.B.3
we will make a connection between the closed graphs described above and
dimer graphs.

EXERCISE 8.2 Consider the 3 x 3 spin lattice shown in the figure.


Assign a coupling constant, Jx(Jy) to horizontal (vertical) bonds. Obtain an
analytic expr ,~~,~"",+"...+1-.4

c.
...

Converted with

STDU Converter
trial version

http://www.stdutiliJV.com
2

Answer: First note that the lattice has N = 9 lattice sites and Nnn = 12
nearest-neighbor pairs. The partition function can be written

~ =

e-6Kxe-6Ky

L '"L
s(

Where K,

=l

s9=1

eKyS4S7 eKXS4S5 eKxS5S6

= {3Jx

and K;

Next we note that

= {3Jy

eKxs(szeKyS1S4eKxsZS3eKyszs'!JeKyS3S6

eKyS5S8 eKyS6S9 eKXS7S8 eKxS8S9

We may write this as

(1)

466

ORDER-DISORDER TRANSITIONS AND RENORMALIZATION THEORY

L (all closed graphs)

...

.0.

x2y2

. . 0
0 . .
+

x2y2

x2y2

.0+
..
(3)

We therefore obtain the following analytic expression for ~:

I
I
I

Converted with
where x::

STOI Converter
trial version

~ S8.B.2.

hDP://www.stdutilitv.com

22]

Let us now return to the planar lattice whose partition function we wish to
compute. We again draw a picture of the planar lattice, but now associate a
direction to each bond. The resulting picture we call a directed graph. We can
associate an antisymmetric matrix, A with a directed graph by assigning to the
directed bond between lattice sites, i and}, a matrix element, aij. Only nearestneighbor lattice sites give rise to nonzero matrix elements. We can associate the
following matrix elements with the directed graph:
for bond directed i ~ i,
for bond directed if-}, .
if i and} have no bond

(8.167)

Here Q = tanh(K). It is possible to generalize Eq. (8.167) and assign different


weights, Q, to different bonds. For the case of a square directed graph, for
example, for the horizontal bonds we let Q = X = tanh(Kx) and for the vertical
bonds we let Q = Y = tanh(Ky) (such a case is considered in Exercise 8.3)
If we want all the terms in the determinant of the antisymmetric matrix, A, to
be positive, then the choice of directions given the nearest-neighbor bonds must
be carefully made. Kasteleyn [22] showed that/or graphs with an even number

467

EXACT SOLUTION OF THE TWO-DIMENSIONAL ISING MODEL

t
J'
l'

-,

...

j..

l-

I'

~I

...

-,

'"
,j..

(c)

(b)

(a)

Fig. 8.7. Three types of directed planar lattice. (a) Directed square lattice. (b) Directed
triangular lattice. (c) Directed hexagonal lattice

of lattice sites, directions can be chosen in such a way that all terms in the
determinant of the antisymmetric matrix, A, are positive. This is important when
we later connect the antisymmetric matrix generated by a directed graph to the
partition func
nd any even loop
on the direc
Converted with
ers only an odd

STOI Converter

number of cl
an even nun
zero) of Iattii
trial
twice. Corre
An example
Exercise 8.3. '--------------------------'

version

hDP://www.stdutilitv.com

J path

containing
mber (including
same lattice site
rwn in Fig. 8.7.
mer is given in

EXERCISE 8.3. Consider the directed planar lattice shown in the


accompanying figure. Write the 12 x 12 antisymmetric matrix, A, associated
with this lattice, and compute the determinant of the matrix, A.

10

'1'

12

'V

...

'V

-I'

11

-"

-r,
2

61
1

t1
....

468

ORDER-DISORDER

TRANSITIONS AND RENORMALIZATION THEORY

Answer: The antisymmetric matrix associated with the accompanying graph


is

A=

-x

-y

0
0

0
0
0

0
0
0

-y
0
0
0
0
0
0

-x

-y

0
0
0
0
0
0
0
0

y
0
0
0
0
0
0
0

0
0
0

-x

-x

-y

0
0
0

0
0
0
0
0

0
0
0

-y
0
0
0

0
0
0
0

0
0
0
0

-y

0
0
0
0
0

0
0
0
0
0
0

-x

-x

-y

0
0
0

0
0

0
0
0

-y

0
0
0
0
0
0
0

-y

0
0
0
0
0
0
0
0

-x

-x

(1)
where x

= te

Converted with

STOI Converter

is matrix is
12

(2)

trial version

hDP://www.stdutilitv.com

The determinant, det(A), of an antisymmetric matrix, A, can always be


written as the square of a polynomial containing the weighting factors, a (or
whatever weighting factors are assigned to the graph). Furthermore, it is
possible to express
det(A) (A obtained from a directed graph) in terms of a
sum of graphs which we shall call dimer graphs. A dimer graph is obtained in
the following way:

(1) Layout

the grid of lattice sites (but do not yet draw bonds) as they
appear in the original directed graph.
(2) Connect all pairs of lattice sites by bonds so that each lattice site is
attached to only one bond and all lattice sites are members of a bond.
The lattice consists of a set of unconnected bonds (we call them dimer
bonds).

The quantity,
det( A), can be represented by the set of all different dimer
graphs which may be obtained from the original planar lattice. Thus,
Jdet(A)

= L(all

different dimer graphs).

(8.168)

469

EXACT SOLUTION OF THE TWO-DIMENSIONAL ISING MODEL

Once a dimer graph is drawn, an analytic expression can be associated with it.
If a graph has s dimer bonds, then we can assign a factor, tanh'{X), to the
graph. As we show in Exercise 8.4, we can also generalize this to give
different weights to different bonds. For example, let us consider a square
dimer graph with different bond strengths, Kx, for horizontal and, Ky, verticle
bonds. If the graph has r horizontal dimer bonds and s verticle dimer bonds, we
associate a factor, tanh'(Kx)tanh'(Ky) with ,e graph. In Exercise 8.4, we
construct the dimer graphs that represent
det(A) for the antisymmetric
matrix, A, obtained in Exercise 8.3.

EXERCISE 8.4. Consider the directed lattice shown in Exercise 8.3 and
the antisymmetric matrix, A, generated by that directed lattice. Draw all
dimer graphs that represent
det( A) and show that they do indeed
: reproduce the analytic expression obtained in Exercise 8.3.
Answer: For the jiS;tmmetriC matrix, A, generated in Exercise 8.3, we
want to show that det(A) = L: [all dimer graphs]. Note that

v'

!
I

Converted with

STDI Converter

=r
IIr

trial version

hDP://www.stdutilitv.com
+

.-1

1.-

_11_

.-

1.

+=r+r:'_+1

~r+r=+r=

II

~l
(2)

Thus, if we add the contributions from the various graphs, we find


l)all

dimer graph]

4x4i

+ 6_x2l + l,

in agreement with Exercise 8.3.

.....S8.B.3. Closed Graphs and Mixed Dimer Graphs


Kasteleyn [22] and Fisher [23] have found a way to write the partition function
in terms of dimer graphs and thereby have found a way to obtain an expression
for the partition function in terms of the determinant of an antisymmetric

470

ORDER-DISORDER TRANSITIONS AND RENORMALIZATION THEORY

matrix. The method is based on the following observation. Consider a lattice


with two different coupling constants, J and J*, for example. A bond with
coupling constant, J*, will contribute factor, 2e-K cosh(K*) and tanh(K*),
where K* = {3J*, to the partition function. [Note that the factor -1 in Eq.
(8.161)] gives rise to this property.] However, note that
lim 2e-K cosh(K*)
r-oo

= 1 and

lim tanh(K*)
r-oo

= 1.

(8.169)

The next thing to notice is that the partition function of a given planar lattice
depends on a sum of terms depending on different powers of the weighting
factor, tanh(K). However, the dimer graphs which are generated from a directed
graph all depend on the same power of the weighting factor, tanh(K). Therefore
we need to construct a directed graph which can generate dimer graphs
depending on different powers of the weighting factors. The limiting behavior
in Eq. (8.169) enables us to do this in a unique and unambiguous way for planar
graphs.
The expa
Converted with
phs requires two
important pi
the theorem of

STOI Converter

Kasteleyn
ed graph which
generates an
ositive terms. On
the other ha
trial version
planar lattice so
that a direct
mixed directed
graph) whic
r graphs) whose
analytic expressions are In one- o-one correspon ence WI
e closed graphs
for that planar lattice. Kasteleyn [21] and Fisher [22] had different ways to
accomplish this. In this section, we use Fisher's approach [23].
We can construct mixed directed graphs according to the following
procedure.

hDP:llwww.stdutililV.com

(1) Draw the planar lattice of interest. (Figure 8.5 gives some examples of
planar lattices.) The planar lattice will have N vertices and Nnn nearestneighbor bonds. On the planar lattice we shall call the bonds primary
bonds. Represent primary bonds by solid lines.
(2) Replace each vertex of degree q( q 2: 4) by a cluster of q - 2 vertices
connected by q - 3 internal bonds. (A vertex of degree, q, is one at
which q bonds meet.) Represent internal bonds by dashed lines. (See
Fig. 8.8 for an example of this replacement.) With this replacement, all
vertices are now either of degree q = 2 or 3.
(3) Replace each vertex of degree, q = 2, by two vertices connected by
one internal bond. Replace each vertex of degree, q = 3, by a cluster of
three vertices connected by three internal bonds. (See Fig. 8.9 for
examples.)

EXACT SOLUTION OF THE TWO-DIMENSIONAL ISING MODEL

+
'*

_Jr

471

(a)

>J.<b)

Fig. S.S. Replace vertices of degree, q(q~4), by a cluster of q - 2 vertices connected


by q - 3 internal bonds. (a) q = 4. (b) q = 5.

_....,..

___

.~

(a)

1
Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com
___.~...-.:.----- (d)

____

-.-

____

'.

-----.::.---::.----- (e)

Fig. 8.9. (a) Replace vertex of degree 2 with two vertices and an internal bond. (b)-(e)
Replace vertex of degree 3 with three vertices and three internal bonds.

(4) Assign directions to all the bonds (primary and internal) so that
counterclockwise circulation around any even loop yields an odd number
of clockwise directed bonds. (According to the theorem of Kasteleyn,
this can always be done.)
The mixed directed graph which results
directed solid and dashed lines connecting
Once a mixed directed graph has been
can associate an antisymmetric matrix, Am,

from steps 1-4 is a collection of


vertices of degree 3 or smaller.
associated with a planar lattice, we
with the directed mixed graph in the

472

ORDER-DISORDER

TRANSITIONS AND RENORMALIZATION THEORY

following manner. The antisymmetric matrix, Am, is constructed by assigning to


matrix elements, aij, the following values:

aij

1/0:
-1/0:
1
-1

for a solid bond directed i _. j,


for a solid bond directed i ~ j,
for dashed bond directed i _. j,
for a dashed bond directed i ~ j,
if i and j have no bond.

(8.170)

As we found in Section S8.B.2, we can associate a collection of dimer


graphs (we shall call them mixed dimer graphs) with each directed mixed
graph. Dimers can now be solid or dashed depending on whether they Occur
at the position of primary or internal bonds, respectively, on the mixed
directed graph which generated them. No more than one dimer bond attaches
to any given vertex. To each dimer bond we assign a factor 1/0:, if the
dimer bond is a solid line and assign a factor 1 if the dimer bond is a
dashed line.
n assocated with
it if it has
(dashed) dimer
bonds.
mer graphs that
generates the

STDI Converter
trial version

hDP://www.stdutilitv.com

(8.171)

The mixed dimer graphs that result from a given planar lattice are in one-to-one
correspondence with the closed graphs which are generated by that same planar
lattice. The correspondence is "inverted," however. The primary (solid) dimer
bonds on the mixed dimer graph correspond to the lack of bonds on the closed
graph.
We can now relate Vdet(Am) to the partition function associated with the
planar lattice. We can write
ZN(T) = 2Ne-NMK(cosh(K)tMo:sVdet(Am)
= 2Ne-NMK(cosh( K)

)NM o:s

E (all mixed dimer graphs),

(8.172)

where s is the maximum number of primary (solid) dimer bonds that can appear
on any of the mixed dimer graphs. With this definition, the dimer graph in
which all dimers are primary (solid) bonds actually corresponds to the unit
graph. In Exercise 8.5 we show these steps for the planar lattice considered in
Exercise 8.2.

473

EXACT SOLUTION OF THE TWO-DIMENSIONAL ISING MODEL

EXERCISE 8.5. Consider the planar lattice shown


: in the accompanying figure. (a) Construct a mixed
directed graph for this lattice and construct the
associated antisymmetric matrix, Am. (b) Compute
the partition function from this antisymmetric matrix,
Am. (c) Draw the mixed dimer graphs for this lattice
and write an analytic expression for the partition
function.

B
7
----.-.--

..
9

4__
--5 .....
--6 ...

Answer:
(a) We first obtain the directed mixed lattice. We replace the vertex of

degree 4 by two vertices of degree 3 and an internal line. We then


replace each vertex of degree 2 by two vertices of degree 2 connected
by an internal line. We replace each vertex of degree 3 by two vertices
of degree 3 connected by three internal bonds. The results is a graph
with 26 vertices.
Thisr""'-"""""","""~:hu.1"",---"-"'____"_,
l.........,.,.LLU!OL.--".,...,.____j_4-:.b.o._l..--"'LO!,_.............
,~c...l.l"""'~lLL
..
t.ou:,n,__J ........

<---_-----,

Converted with

STOI Converter
trial version

We next give directions to all the lines in the mixed graph so that
for any even loop, clockwise circulation encounters an odd number of
clockwise directed bonds. Such a choice is shown below. We have
also numbered the vertices.

474

ORDER-DISORDER TRANSITIONS AND RENORMALIZATION THEORY

(b) Given the numbered mixed directed graph above, we can use it to
construct an anti symmetric matrix, Am. Choose the matrix elements
according to the directions of the bonds and their type. Assign a value,
1/ x, to horizont~l bonds and assign a factor, ; to vertical bonds.
The nonzero matnx elements are al,2 = +1, a2,1 = -1, a2,3 = +1/x,
a3,2 = -1/x,
a3,S = +1, as,3 = -1,
as,12 = +1/y, al2,S = -1/y,
al,3 = a3,1 = 0, and so on. The matrix, AL is too big to write here,
but it is easy to construct given the directed graph above. Once it has
been constructed we can find its determinant. It is

The partition function is


ZN(T)
r--

= 2ge-6KXe-6Kycosh6(Kx)cosh6(Ky)x6l
7 x4y4)

----""'---'----_1 ___.__.----'--=------"---..L........<..<.....___.._----'-~~___._,

__

Converted with

STOI Converter
trial version

Let u

~~~~

).

__hn_p_:_H_www ._st_d_u_til_ib
__ .c_o_m
I

I-

(I)

- -

:
I
(ill '-1 I
'. "

i_

'.-

fI',

('"I

(1/x6)(1/y6)

.'

.-

i_I
. I

I-

(Iv)

(1/x4)(1/y2)

H
i

(vi)

.(1/ x2)(I/y2)

I(VII)I-1_,

...

(2)

graph for each


some of them

EXACT SOLUTION OF THE TWO-DIMENSIONAL ISING MODEL

475

The analytic expression associated with each mixed dimer graph is


shown below it. Graph (i) above corresponds to the first closed graph
in Exercise 8.2, Eq. (3). Four mixed dimer graphs of the type (ii)
above correspond to closed graphs, 2-5 in Exercise 8.2, Eq. (3). The
two mixed dimer graphs of the type (iii) above correspond to closed
graphs 6 and 7, and the two mixed dimer graphs of the type (iv) above
correspond to closed graphs 8 and 9 in Exercise 8.2, Eq. (3). Two ,
mixed dimer graphs of the type (v) above correspond to closed graphs
10 and 11 in Exercise 8.2, Eq. (3). Four mixed dimer graphs of the
type (vi) above correspond to closed graphs 12-15 in Exercise 8.2,
Eq. (3). And finally, graph (vii) corresponds to the last graph in
Exercise 8.2, Eq. (3). Thus there are 16 possible mixed dimer graphs.
They each consist of 26 vertices and a mixture of solid and dashed
dimer bonds. The partition function is
ZN(t)

2ge-KXe-6Kycosh6(Kx)cosh6(Ky)
x x6l

L (all mixed dimer graphs).

(3)

Converted with
~ S8.8.4. Pal

STOI Converter

We now have
ession for the
trial version
partition functi
will consider a
square planar I
ment from the
infinite lattice, and then take the limit, n -4 00. In order to demonstrate the
procedure, we will first extract a 3 x 3 segment from the infinite planer lattice

hDP://www.stdutilitv.com

Fig. 8.10. A 3 x 3 segment of an infinite square directed graph corresponds to a


54 x 54 mixed directed graph.

476

ORDER-DISORDER TRANSITIONS AND RENORMALIZATION THEORY

and describe its properties in detail. Then we will go on to write partition


function for an n X n segment and take the limit, n --t 00 .
......S8.B.4.1. A 3 x 3 Segment of the Infinite Square Planar Lattice
Let us consider a 3 x 3 segment of an infinite square planar lattice. We first
construct a 54 x 54 mixed directed graph as shown in Fig 8.10 using the rules
of Section S8.B.3. We can then assign an anti symmetric matrix, A54, to the
directed mixed graph. We obtain

_XT

o
AS4

_XT

0
X

_yT

_yT

o
o
o

0
0

roo

_yT

0
0
r
X
0
Y
0 _XT
r
X
0
y
0
_XT
0
0
y
0
0
0
0
0
y
0
o
0
0
0
0
y
~----------------~------~--~

o
o
o
o

_yT

0
0

_yT

o
o
o

_yT

_XT

_XT

r
(8.173)

Converted with
where "0" i
zero:

o
o
o
o
o

~ith all elements

STOI Converter
trial version

hDP://www.stdutilitv.com
o
o
o

o
o
o
o
o
o
o

-1

-1

-1

-1 0

(8.174)

1/x 0 0 0 0 0

o
o
o
o

1/y

000000

o
o

'

477

EXACT SOLUTION OF THE TWO-DIMENSIONAL IS1NG MODEL

18

12

Fig. 8.11. The 54 x 54 directed lattice with bonds connecting opposite sides added.

Converted with

matrices. This
The subscript 6
aph, Fig. 8.10,
notation will be
irected graph,
we have left ou
and the antisy
trial version
Let us now
t by giving it
periodic bound
ne another by
adding bonds an
en
e con
u Ions ro
bonds to the
antisymmetric matrix. It turns out that as we consider larger and larger
segments of the infinite planar lattice, these additional terms contribute less and
less to the determinant of the anti symmetric matrix, and in the limit of an
infinite lattice they do not affect the result. In Fig. 8.11 we redraw the 54 x 54
directed lattice so that bonds on opposite sides are connected. We then no
longer have a planar lattice. Note the directions of these added bonds are
~pposite to normal connections to the infinite lattice. The antisymmetric matrix,
AS4, that results from adding these new bonds to A54 is given by

STOI Converter
hDP://www.stdutilitv.com

XT

_yT

_XT

_yT

-X
Y

_XT

o
o
yT

o
o

0
Y
0
0

roo

yT

0
0
Y
0
0

yT

_XT

-X
Y
0
0

_XT

0
Y
0

0
0

-y

_yT

0
0

xr

_yT

0
0

_yT

-y

roo
0
0
Y

o
o
-y

o
o

_yT

XT

_XT

-X

_XT

r
(8.175)

478

ORDER-DISORDER TRANSITIONS AND RENORMALIZATION THEORY

The anti symmetric matrix, ..454, is now in a form that is easy to deal with.
First note that we can write ..454 as the tensor product of smaller matrices (tensor
products are defined in the subsection below). We can write

D1s
..454

-/3
D1s

I) Y6

I)

yr

yr

-I) Y6 )

D1s

I) Y6

= I) DIS + E3

yr

-I)

(I)

yr) - E3

(I) Y6),

(8.176)

where

(8.177)

E,j~

~1~11 andl,j~

~ ~),

(8.178)

Converted with

STOU Converter

Tensor

trial ve rsion

matrices, A
order tensor

product of two
is a fourth-

hnp://www.stdutilitv.com
(1)

where AI] and Bkl are the (ij)th and (kl)th elements of matrices A and 8,
respectively. The product of two fourth-order tensors, Q and IR, is given by

(QIR)ijkl

= L QjmJenlRmjnl,

(2)

m,n

where
Q and
1 is

QjmJen

and

IRmjnl

IR, respectively.

are the (imkn)th and (mjnl)th elements of the tensors


The tensor product of a matrix A with the unit matrix

(3)
where
write

8k,1

is the Kronecker

delta function.

(A lJ)(CD)

If lJ and C commute,

= (AC) (lJD).

then we can

(4)

479

EXACT SOLUTION OF THE TWO-DIMENSIONAL ISING MODEL

The matrix

DIS

can also be written as a tensor product:


DIS =1) s r, -E3
X6 +E31 Xr

(8.179)

Therefore, the anti symmetric matrix (A)54 takes the form


A54 = h (I) r6) 1

- E3

h (E3 X6) + t, (E31

Xn + E3 (I) yn

(h Y6),

(8.180)

where we have used the fact that E~ = E31 Let us next introduce the matrix P3,
which diagonalizes E3. That is,
P"31E3P3 =

(~1o ~2 ~),

(8.181)

A3

where Ak is the kth eigenvalue (k = 1,2,3) of the matrix E3 and is given by


Ak = exp(27rik/3). Then we find that
A~4 == (P"31 (P31 I6))A54(P3 (P3 I6))

STOI Converter

We next note,

(8.182)

trial version

hDP://www.stdutilitv.com

(8.183)

and

( AtX6
(P"31E3P3) x6 =
~

o ) .

A2X6

(8.184)

A3X6

Then one can readily show we have obtained an expression for A ~4 which is
block diagonal. It can be written

A;4 =

n(3)
1,1

n(3)
1,2

n(3)
1,3

n(3)
2,1

n(3)
2,2

0
0

0
0

0
0

0
0

D(3)
2,3

0
0

0
0
0

n(3)
3,1

0
0
0

0
0

0
0
0

0
0
0

0
0

n(3)
3,2

n(3)
3,3

(8.185)

480

ORDER-DISORDER TRANSmONS

AND RENORMALIZATION THEORY

where
(3)

Di,j

TTl

= r6 - AiX6 + Ai X6 + AjY6

-1

-1

-1

0
-XAi
0

The determinant of

Aj Y6
0

X/Ai
0

YAj

-1

-1

-Y/ Aj

-1

-1

(8.186)

D;'~ is given by

Converted with

STOI Converter

(8.187)

trial version
Let us now
determinant a
product of determinants of those matrices, we obtain

hDP://www.stdutilitv.com

det[As4l = det[A ~l =

trices under the


es is equal to the

In] [1+ G) '. G) ', Gr Gr -2 G)cos (2;i)

- 2G)cos C;i) +2G) G) cosC;i) +2G) G)cosC;i)].


2

(8.188)

These results are easily generalized to an n x n segment of the infinite planar


lattice.
~ S8.B.4.2. An n x n Segment of an Infinite Square Planar Lattice
Let us now consider an n x n segment of the infinite square spin lattice. The
mixed directed lattice that we construct will have 6n2 lattice sites, and therefore
the anti symmetric matrix that we associate with the directed lattice will be
6n2 x 6n2-dimensional. We proceed exactly as before. We introduce the n x ndimensional matrix, En, which consists of all zeros except for (i) matrix
elements -1 along the nearest upper off-diagonal and (ii) the matrix element 1

481

EXACT SOLUTION OF THE TWO-DIMENSIONAL ISING MODEL

in the bottom left comer. We shall restrict outselves of odd values of n. Then the
eigenvalues of En are given by Ak = exp(21rik/n), where k = 1,2, ... , n. The
determinant of the antisymmetric matrix, A6n2, obtained by connecting opposite
bonds of the 6n2 x 6n2-dimensional directed lattice will be given by

det[A6n'] =

Ull [1 + G) \G) \Gr Gr-2G)cos(Z:i)


-2 G) cosC:k) +2G) G)cosC:k) +2G) G)cosC:i)]
2

(8.189)
The partition function can be written

Zn(T) = 2n2e-Kxn(n-l)e-Kyn(n-l)coshn(n-l)

(Kx)coshn(n-l)

(Ky)

(8.190)

Converted with
If we remen
e-Ksinh(K)
=
can write the p

hd note that
~q. (8.189) we

STOI Converter
trial version

hDP://www.stdutilitv.com
,
,
,

Zn(T)

'"

(8.191)

~ J...;I "-=1

The logarithm of the partition function becomes


In[Zn(T)] = n2In(2)

+ n(n

- I)ln(_x_)

l+x

+ n(n

- I)ln(-Y-)

l+y

Inn

""
+ 2" L_..
L_..ln(Dij (n))

(8.192)

i=l j=l

We now can take the thermodynamic limit. It is useful to note that


(l/n)cos(21ri/n)
= (1/21r) 5;11" dcpcos(cp).
Then for the free
energy density (free energy per lattice site), a(T), we find
limn-+oo

L:7=1

-:~;. = Iim,,~oo

[>rz,,(T)]] 1n2+ InC :x) + InC: y)


=

1 I 1211"
1211"
[
1
1
1
+ -2-4
2
dCPI dcp2ln 1 + '2 + '2 + 22
1r 0
0
X
Y
xy

+ ~COS(CP2)
+ xy
_';cos(CPt)].
xy

2
2
- -COS(CPI) - -COS(CP2)
x
Y
(8.193)

482

ORDER-DISORDER

We can
x

simplify

TRANSITIONS AND RENORMALIZATION THEORY

subsequent

calculations

if we let Jx

= Jy Then

= y = tanh(K) and the free energy density takes the form


_ a(T) = In(2) + 2ln(_X_)
kBT

1 +x
1 1 J 27r
J 27r
[2
1
+ -dCPI
dc/J2ln 1 + -2 + 2 41r2 0
0
x
x4

- (~- ~) (cos(t) + cos(2))]


1 1 J27r

= In(2) - 21n(1 +x) +-2-2


41r
27r
X

dcp2ln[1

+ ~

dCPI

+ x4 + 2x(1 - x2)(cos(cpt) + COS(CP2))]'


(8.194)

We now can ......1.~

.~

~~

..t

'-1

Converted with

(2K)/cosh4(K).

STOI Converter
trial version

Then the free

a(T)
- -

kBT

(8.195)

= In(.~) -

hDP://www.stdutilitv.com
L.A

1 1 J27r
J27r
+241r2
0 dCPI 0 dcp2In[cosh2(2K)

+ sinh(2K)

(cos (cpt) +COS(4)2))]

Equation (8.196) can be simplified still further. Note that COS(CPI)+ COS(CP2) =
2cosa (CPt+ CP2)]COS[i(CPI
- CP2)]. If we make the change of variables,
fh = i(CPI - CP2) and fh = i(CPI + CP2), and let /'i,= 2sinh(2K)/cosh2(2K),
then the free energy density becomes
- ~~~= In(2) - 2K

+ In [cosh (2K)]
(8.197)

+ -21 J7r dOl J27r d02ln[1


41r -7r
0

+ /'i,cos(OI) COS(02)]'

We can integrate Eq. (8.197) if we note that [24]


27r
o dxln(a

+ bcos(x))

= 27r In[(a

+ Va2

b2)/2].

(8.198)

483

EXACT SOLUTION OF THE TWO-DIMENSIONAL ISING MODEL

Then we obtain

- a(T)
kBT

= In(2) -

+~

2K

+ In[cosh

(2K)]
(8.199)

d61n[~ (1 +..j 1- 2Sin2(6))].


1<

The internal energy density is easily obtained from Eq. (8.199). We find

u(T) = J ~

8K

(a(T))
kBT

= 2J -

+ (2 tanh2(2K)

Jcoth(2K)[1

- 1) ~K(K,)],
7r

(8.200)
where K( K,) is the complete elliptic integral of the first kind [25] and is defined
as
71"/2

dO

(8.201)

Converted with
The specific he

STOI Converter
(8.202)

The complete elliptic integral of the first kind, K(K,), has a singularity for K, = 1.
The phase transition occurs at this point. The temperature at which the phase
transition occurs is then given by the condition
2 sinh (2Kc)
K,c = cosh2(2Kc)

= .

(8.203)

This gives a critical temperature, T; = 2.269J [ke- In Fig 8.12 we plot the
Internal energy per site and the specific heat as a function of K, which is
proportional to the inverse temperature.
Ising-like phase transitions have been measured in the two-dimensional
I~ing-like antiferromagnets,
K2CoF4 and Rb2CoF4. These substances behave
lIke two-dimensional
spin systems because they consist of strongly coupled
anti ferromagnetic COF2 planes separated by weakly coupled planes containing
the remaining molecules. In Fig 8.13 we show two measurements of the heat
capacity of Rb2CoF4 We see the characteristic Ising-like singularity in the heat
capacity.

484

ORDER-DISORDER

TRANSITIONS

AND RENORMALIZATION

THEORY

2.0

U/J
1.0

0.2

0.4

0.6

0.8

J{

(b)
2.0

___.__.""'-"-.........._~............_~~.............,mal
energy per
(b) The specific
K. The phase
c = 1/2.269 =

.--------L___.__

1.0

STOI Converter

0.2

trial version

hDP://www.stdutilitv.com

~ ..
"

s :

t
\
t

:
~

f
;

..

....~ ...... . ...,.'.


:"':~~

o
Q

100

105

110

Fig. 8.13. The heat capacity of Rb2CoF 4


(given in arbitrary units) as a function of
temperature, T, measured in Kelvin. Two
sets of experimental
data are shown.
Reprinted, by permission, from H. Ikeda,
1. Hatta, and M. Tanaka, J. Phys. Soc.
(Japan) 40, 334 (1976).

485

REFERENCES

In the section we have obtained analytic expressions for the free energy, the
internal energy, and the heat capacity of a two-dimensional square planar Ising
lattice. From Eq. (8.202) we can obtain the critical exponent, G. The type of
singularity that occurs in the heat capacity is easy to find from properties of the
comp~te elliptic integrals. Let us note that as K, ~ 1, E(K,) ~ 1 but K(K,) ~
In(4/ 1 - K,2). Thus, the singularity in the heat capacity is logarithmic and
the critical exponent, G, for the two-dimensional Ising lattice is G = O.
It is interesting to compare Fig. 8.12, the exact heat capacity of the twodimensional Ising lattice, to that of the mean field approximation given in Fig.
7.8. Mean field theory predicts a finite discontinuity, while the exact solution
gives a logarithmic singularity. The two theories also give different critical
temperatures.

REFERENCES
1. H. E. Stanley, Introduction to Phase Transitions and Critical Phenomena (Oxford
University Press, Oxford, 1971).
rrelation Func2. D. Forster,
Converted with
tions (w. A.
75-1976, Nor3. F. Ravendal,
ptionally clear
dita, Blegda
which we have
presentation
trial version
followed clo
4. B. Widom, J.
5. F.C.Nixan&'~nro~~?rev..mm~~~rn)TT~~----~
6. L. P. Kadanoff, Physics 2, 263 (1996).
7. S. Ma, Modem Theory of Critical Phenomena (W. A. Benjamin, New York, 1976).
8. K. Wilson and J. Kogut, Phys. Rep. 12C, 75 (1974).
9. M. E. Fisher, Rev. Mod. Phys. 47, 773 (1975).
10. M. N. Barber, Phys. Rep. 29, 1 (1977).
11. K. Wilson, Rev. Mod. Phys. 47,773 (1975).
12. K. Wilson, Phys. Rev. B4, 3174 (1971).
13. Th. Niemeijer and J. M. 1. van Leeuwen, Phys. Rev. 31, 1411 (1973).
14. L. Onsager, Phys. Rev. 65, 117 (1944).
15. H. S. Robertson, Statistical Thermophysics (Prentice-Hall, Englewood Cliffs, NJ,
1993). (A really excellent discussion of the one- and two-dimensional Ising model
and its history.)
16. C. Davis, 1. ACM 38, 106 (1993).
17. S. Chamois et aI., J. ACM 44, 213 (1994).
18. M. Gray, MicroBio. Today 76, 89 (1993).
19. T. Dauxois, et. aI., Phys. Rev. E47, 684 (1993).
20. C. Johanson, Psychol. Today 58, 46 (1994)
21. The author thanks Dan Robb for careful reading of this section.

STOI Converter
hDP://www.stdutilitv.com

486

ORDER-DISORDER TRANS mONS

AND RENORMALIZATION THEORY

22. P. W. Kasteleyn, J. Math. Phys. 4,287 (1963).


23. M. E. Fisher, J. Math. Phys. 7, 1776 (1966).
24. M. R. Spiegel. Mathematical Handbook (Schaum Outline Series), (Me GrawHill,
New York, 1968), # 15.104.
25. P. F. Byrd and M. D. Friedman, Handbook of Elliptic Integrals for Engineers and
Scientists (Springer-Verlag, Berlin, 1971).
26. H. Ikeda, I. Hatta, and M. Tanaka, J. Phys. Soc. (Japan) 40, 334 (1976).

PROBLEMS
Problem 8.1. A one-dimensional lattice of spin-! lattice sites can be decomposed into
blocks of three spins each. Use renormalization theory to determine whether or not a
phase transition can occur on this lattice. If a phase transition does occur, what are its
critical exponents? Retain terms in the block Hamiltonian to order (V), where V is the
coupling between blocks.
Problem 8.2. Find the critical exponents for five spin blocks on a square lattice for the
two-dimensional nearest-neighbor Ising model. Retain terms to lowest order in (V).
where V is the

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com

Fig. 8.14.
Problem S8.1. Consider the directed lattice shown in the figure. Assume that horizontal
bonds have a coupling constant, Jx, and that tilted vertical bonds have coupling constant,
Jy (a) Find the anti symmetric matrix, A, associated with this lattice. Compute its
determinant. (b) Draw all disconnected dimer graphs for this lattice. Show that they
represent JA.

487

PROBLEMS

Problem S8.2. Consider the directed lattice shown in the Figure. Assume that horizontal
bonds have coupling constant, lx, and that vertical bonds have a coupling constant, Iy
(a) Find the anti symmetric matrix, A, associated with this lattice. Compute its
determinant. (b) Draw all disconnected dimer graphs for this lattice. Show that they
represent JA.

t:!:

f: f
3

Problem S8.3. Consider the planar triangular lattice shown in the figure. Assign a
coupling constant, I, to each bond. (a) Compute the partition function and draw all
closed graphs for this lattice. (b) Draw the directed mixed graph for this lattice and
obtain the antisymmetric matrix it generates. Show that the determinant of the
antisymmetric yields the correct expression for the partition function.

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com

9
INTERACTING FLUIDS

9.A. INTRODUCTION
One of the great successes of equilibrium statistical mechanics has been the
theory of interacting fluids. For real fluids, the interparticle potential has a hard
core and a weak attractive region. Consequently, the usual perturbation theories
do not work. For dilute or moderatel dense fluids the onl small parameter is
~~et~::i~

SP~:ri~h:~yCh:

Converted with

STOU Converter

n be formulated
f~~~~~~~;i~~

describing su
trial version
ch is a measure
of the probab
any particle in
the fluid. The
rectly by X-ray
scattering ex
cular dynamics
experiments. We shall begin this chapter by deriving an expression for the
internal energy in terms of the radial distribution function, and in the special
topics section we will derive expressions for both the pressure and
compressibility in terms of the radial distribution function, using both the
canonical and grand canonical ensembles.
In the chapters on thermodynamics, we have already been introduced to the
virial expansion of the equation of state. We shall now derive an expression for
the virial coefficients using microscopic theory. The method involves writing
the grand partition function in terms of a cumulant expansion and then
expressing the grand potential and pressure in terms of the cumulants or cluster
functions. The techniques for doing this were first developed by Ursell, whose
work was later extended by Mayer. The analysis is greatly facilitated by the use
of graph theory which enables one to give a simple physical interpretation to
various terms in the virial expansion. The first few virial coefficients have been
computed for a variety of interparticle potentials. We shall discuss the results
for the hard-core potential, the square-well potential, and the Lennard-Jones
6-12 potential and compare them with experiment.
For very light molecules at low temperatures, quantum effects introduce
significant deviation from classical behavior. In the special topics section we
will construct the virial expansion for interacting degenerate quantum fluids.

hUP:llwww.stdutililV.com

489

THERMODYNAMICS AND THE RADIAL DISTRIBUTION FUNCTION

9.B. THERMODYNAMICS AND THE RADIAL


DISTRIBUTION FUNCTION [1-5]
In this chapter we will compute the equilibrium properties of a classical fluid
consisting of N particles with mass m whose dynamics is determined by a
Hamiltonian of the form

(9.1)
where qjj = qj - qj is the relative displacement of particles, i andj (we use the
notation of Section 6.B). It is useful to introduce the one-body phase function,
Of(qN) = 2:~1 01 (qj) (the kinetic energy is an example of a one-body phase
function), and the two-body phase function, d{ (qN) = 2:~~:~1)/202( qj - qj)
(the potential energy is an example of a two-body phase fun~tion). The average
values of these phase functions can be written
(d() =

r, T)

Converted with

(9.2)

STOI Converter

and

trial version

(9.3)

hDP://www.stdutilitv.com
where nf(ql;V,T)
and n~(ql,q2;V,T)
are the one-body and two-body
reduced distribution functions, respectively, in the canonical ensemble. They
are defined

N(
. '
_
1
n[ ql, ... ,q[,V,T)-QN(V,T)(N_I)!

N!

dql+ld~e

-(3vN

(9.4)

L~X~=~1)/2

=
V(qjj) is the potential energy. In Eq. (9.4) the quantity
QN(V, T) is called the configuration integral and is defined as

where VN

(9.5)
The reduced distribution functions have the property that
(9.6)
One of the simplest quantities to compute is the internal energy, U(V, T, N).
The internal energy is just the average value of the Hamiltonian and can be

490

INTERACTING FLUIDS

written

U(V, T,N)

3
= (H) = iNkBT

= ~NkBT +

if J

N
e-(3v
dqNe-(3VN

IdqNVN

(9.7)
dq1dfb V(qI2)n~(q" fb; V, T).

The term ~NkB T is the kinetic contribution to the internal energy, and
the remaining term is the contribution due to the interaction between the
particles.
The expression for the internal energy simplifies considerably for systems
whose particles interact via spherically symmetric potentials, V(qij) = V(qij),
where qij = Iqi - qj I is the magnitude of the relative displacement. Then the
two-particle reduced probability density takes the form n~ (ql , q2; V, T) =
n~(q12; V, T). Furthermore, it will depend only on the magnitude, qij, of the
relative displacement. If we change the integration variables to relative, qij' and
center-of-mas
_
n~(q12; V, T) =
Converted With
ass coordinates
(li)2...N(q
v S2 Il'... V ,
and find

STOU Converter

trial version

, T),

(9.8)

hnp://www.stdutilitv.com
where q = q12. The function t{ (q; V, T) is called the radial distribution
function. The radial distribution function is extremely important because it
completely characterizes the behavior of a classical fluid of spherically
symmetric particles. It also has a direct physical interpretation. The quantity
(N /V)g(q)47rq2dq is the average number of particles in a spherical shell of
width q -+ q + dq at a distance q from any particle in the fluid.
The form of the radial distribution function can be determined from neutron
scattering experiments in which slow neutrons scatter from atomic nuclei in a
liquid. The angular dependence of the scattered neutrons is measured, and
this information can be used to construct the static structure factor, Snn (k)
(cf. Exercise 9.1), of the liquid. The structure function can be expressed in
terms of the radial distribution function. In Fig. 9.1 we show the structure factor
and the radial distribution function obtained from neutron scattering
experiments on liquid 36Ar at 85 K. The radial distribution function goes to
zero at about the hard-core radius, indicating that no particles can penetrate the
hard core. It has a maximum at about the distance of the minimum of the
attractive potential between nearest-neighbor argon atoms in the liquid. The
next peak is due to the high probability of finding next nearest neighbors at that
position, and so on.

THERMODYNAMICS

AND THE RADIAL DISTRIBUTION

491

FUNCTION

(a)

S(k)
2

10

12

g(q)
2

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com

Fig. 9.1. (a) The


,nn
,
iquid 36 Ar at 85
K obtained from neutron scattering experiments. The dots are data poins. The solid line
is a best fit to the data. (b) The radial distribution function, g(q), versus q (in A) obtained
from the data in (a). Reprinted, by permission, from I. L. Yarnell, et. al., Phys. Rev. A7,
2130 (1973).

,
,
:
:
I

EXERCISE 9.1. The static density correlation function, Cnn(r), for an


equilibrium fluid can be written Cnn(r) = (liN) I dr'(n(r' + r)n(r')),
where the density phase function is defined n(r, qN) = L:~l b(q; - r).
The static structure factor for the fluid is defined as Snn(k) = I dreik-r Cnn(r).
Show that in the limit N ~ 00 and V ~ 00 with n = N IV = constant, the
static structure factor when written in terms of the pair correlation function
has the form
411"

Snn(k) = 1 + n8(k) + T

Joo0

qdqsin(kq)(g2(Q)

- 1).

492

INTERACTING FLUIDS

Answer: The static structure factor may be written


S.. (k) =
=

kf
kf f

dre,"r f dr'(n(r' + r)n(r'))

dr'dre'krtt(O(qi

- r' - r)o(qj - r'))

= - LL(eik(qi-qj)).
N

(1)

;=1 j=l

If we evaluate the average value in the canonical ensemble, we find

Converted with

STOI Converter

(2)

trial version
The minimu
the size of the
box. In the 1
ant, the length
of the wavevector can be zero. We can separate this contribution from the
remainder of the integral. We find

hDP:llwww.stdutililV.com

S.. (k) = 1 + nO(k)

+n

f dq[2eik.q,,(~

(ql2; V, T) - 1).

(3)

The integration over angles in Eq. (3) can be performed to finally give
Snn(k) = 1 + nO(k)

+ 4;n

qdqsin(kq)(~

iq; V, T) - 1).

(4)

The structure factor can be measured in neutron scattering experiments, and


the term n8(k) is the contribution due to coherent forward scattering.

9.C. VIRIAL EXPANSION OF THE EQUATION OF


STATE [1-3, 7, 8]
The first step in obtaining tractable microscopic expressions for the thermodynamic properties of fluids is to find a small parameter that can be used as an

493

VIRIAL EXPANSION OF THE EQUATION OF STATE

expansion parameter. For dilute or moderately dense gases, such a parameter is


the density. In this section we shall first obtain a microscopic expression for the
virial expansion of the equation of state (density expansion) for a classical fluid,
and then we shall compare the predictions of the microscopic
theory to
experimental results.

9.C.1. Virial Expansion and Cluster Functions


Let us consider a classical gas of identical particles of mass m which interact via
two-body short-ranged forces. We shall assume that the potential has a large
repulsive core and short-ranged attraction (cf. Fig. 9.2). For such a system the
grand partition function can be written in the form

_ ~
Zp,(T, V) -

1
:0 N!1 AWe

where the momentum


. t
1 QN (TV ~, .
In egra,
If the inter
separation bet
convenient to

integrations

I"

(3Np,'
QN(V, T),

have been performed


'r.

(9.9)

and the configuration

I!!'''\

Converted with

STOI Converter
trial version

large values of
~V ij) = 1. It is

(9.10)

hDP://www.stdutilitv.com

The function f
n. Furthermore,
in the region of the hard core, where V ij = 00, the function hj = -1. Thus hj is
a much better expansion parameter than V ij (cf. Fig. 9.2). In terms of the

:.r- ,8v(x)

I
I

0.5
\

-1

.I'

x=q/a

\,.1

Fig. 9.2. The dashed line is a plot of a typical interparticle potential, v(x) (LennardJones 6-12), the solid line is a plot ofJ(x) = e-{3v(x) - 1, and a is a measure of the hardcore radius (v(l) = 0). The plots are given for a fixed temperature.

494

INTERACTING FLUIDS

function

/0, the configuration integral can be written as


QN(V, T)

N{N-l}/2

J... Jdq

1 ...

IT (1 +/0)

d~

(9.11)

(O)

In Eq. (9.11) the product is taken over all pairs of particles (ij). There are
1)/2 such pairs.
It was shown by Ursell [9] that the partition function can be written in terms
of a cumulant expansion. We first write the configuration integral in the form

N(N -

so that
...
We now defi
function:

dql d~WN(ql'

Converted with

STOI Converter
trial version

Zp,(T, V)

hDP://www.stdutilitv.com

,~).

(9.13)

grand partition

,q+

(9.14)

where Ut(ql, ... , qt) is called a cluster function of Ursell function. In terms of
the cluster functions, the grand potential takes a simple form:

n(v,

T, J-L) = -kBTln[Zp,(T,

= -kBT

V)]

f ~ J... Jd
t=1

~e{jtP,'
l. AT

q1

...

dqtUt(ql,

., qt)

(9.15)

If we know the function WN(q, ... , ~), then we can easily find the cluster
functions UN(ql' ... , ~). We simply expand Eqs. (9.13) and (9.14) in powers
of (A'T3exp({3J-L'))and equate coefficients. We then obtain the following
hierarchy of equations:

u. (ql)
U2(ql,q2)

= WI (q.),
= W2(ql,q2)

(9.16)
- WI(ql)Wl(q2),

U3(ql, Q2,Q3) = W3(ql, Q2,Q3) - WI (Ql)W2(Q2,Q3)


- W1(Q2)W2(Ql,Q3)- W1(Q3)W2(Q.,Q2)
+ 2WI (ql )WJ (Q2)W1 (Q3),

(9.17)

(9.18)

495

VIRIAL EXPANSION OF THE EQUATION OF STATE

and so on. We can revert Eqs. (9.16)-(9.18)

and find
(9.19)

WI (qt) = UI (ql),
W2(ql,q2)
W3(ql,q2,q3)

+ UI(ql)UI(q2)'
= U3(ql,~,q3)
+ Ul(Qt)U2(q2,q3)
+ UI(~)U2(QI,Q2) + UI (Qt)UI

(9.20)

= U2(ql,q2)

and so on. The cumulant

expansion

+ Ul(Q2)U2(Ql,Q3)
(Q2)U1(Q3),

we have just introduced

(9.21)
is independent

of

the form of WN(QI,, 'IN).


For the systems we consider in this chapter, we can write

N{N-l)/2
WN(ql,''IN)=

IT

(1+10)

(9.22)

(ij)

Thus, the first few terms are

WI(
W2(QI,
W3(QI,Q2,
W4(QI,Q2,Q3,

Converted with

STOI Converter
trial version

(9.23)
(9.24)

(9.25)

(1 +134),
(9.26)

hDP://www.stdutilitv.com

and so on. If we expan ou


e expressions or N Ql, ... , 'IN ' we see that the
case N = 1 contains one term, the case N = 2 contains two terms, the case
N = 3 contains eight terms, and the case N = 4 contains 64 terms. From
Eqs. (9.16)-(9.18) we can find the first few cluster functions. They are

UI(QI)

1,

U2(ql, Q2) = 112,

(9.27)
(9.28)

and

The expression for U4 (Ql , q2, Q3, Q4) is too long to write down here, so we will
leave it as an exercise. It contains 38 terms. It is easy to see that the functions
~ N ( Ql , ... , 'IN) and UN ( ql , ... , 'IN) rapidly become very complicated
as N
Increases.
Since the grand partition function contains terms with arbitrarily large values
of N, it is necessary to find a systematic procedure for categorizing and sorting
various terms in the expressions for WN(ql, ,'IN) and UN(ql' .. ,'IN). Such

496

INTERACTING FLUIDS

a procedure is called graph theory. We can represent WN( ql" .. ,~)


of graphs in the following way:
WN(ql, .. ,~)

L (all different N-particle

graphs).

in terms

(9.30)

An N-particle graph is obtained by drawing N-numbered circles and connecting


them together with any number of lines (including zero) subject only to the
condition that any two circles cannot be connected by more than one line. Two
N-particle graphs are different if the numbered circles are connected differently.
An algebraic expression can be associated with an N-particle graph by assigning
a factor Ii) to each pair of numbered circles that are connected by a line. For
example, the six-particle graph is expressed as
2

=/34/25/26,
~4

(9.31 )

and the five-p

Converted with

STOI Converter

(9.32)

trial version
The graphical

hDP:llwww.stdutililV.com

where we have grouped together all graphs with the same topological structure
before labels are attached.
From Eqs. (9.16)-(9.18), it is not difficult to see that the cluster functions can
also be represented by graphs, but of a special form. We find that
Ul( ql, ... ,ql) =

L (all different I-clusters).

(9.34)

An I-cluster is an I-particle graph in which all numbered circles are


interconnected. Thus, in an I-cluster it must be possible to reach all numbered
circles by continuously following the lines connecting the circles. Graphical
expressions for the first few cluster functions are given by
Ul(qI)

=.,

U2(ql,q2) = ......
U3(ql, q2, q3) =

31.. + A,

(9.35)
(9.36)
(9.37)

497

VIRlAL EXPANSION OF THE EQUATION OF STATE

and
U4(Ql,Q2,Q3,q4)

=12

+4~

+ 12t2:

+30 +6t2l+M

(9.38)

Again, we have grouped together all graphs which have the same topological
structure be~re lables are attached.
We see from graph theory that the cluster functions have a special meaning.
An N-particle cluster function UN( Ql, ... , ~) is nonzero only if all N-particles
interact. Thus, they contain information about the way in which clusters occur
among interacting particles in the gas. By writing the partition function in terms
of a cumulant expansion, we have isolated the quantities which have physical
content.
There is also another reason for writing Zp,(T, V) as a cumulant expansion,
which does not become apparent until we try to integrate over the functions
WN(Ql,'"
,~).
Let us consider some examples and restrict ourselves to fourparticle graphs.
~------------------------------~

(a) Each u

Converted with

pIe,

STOI Converter
trial version

and

hDP://www.stdutilitv.com
~--~~~~~--~~~--~~~~~r~h4
= v J J J dq,_ d(b dq4 !24 /34.
(b) If two parts of a graph are not connected by a line, the corresponding
algebraic expression factors into a product. For example,

I-

+qt"

dq4

(I I) I- +qt" dq4!td34
(J J dqtdq,_!t2)(JJ
=
=

d(bdq

4/34)'

(c) Each completely connected part of a graph (cluster) is proportional to


the volume (this comes from integrating over the center of mass of a
cluster).
For example,

498

INTERACTING FLUIDS

and

J- +ql

.. dq4

(II:) = v JJJ dqldll2dll3


2

=V

J J dq31 dll2tf(

fld23 fl3

1113Ii) f( Iq211) f( 11131- 11211),

where q31 = q3 - ql and q21 = q2 - q.. In the second example, we have


integrated over the position of particle 1 in the cluster. This gives a factor V
since the cluster can be anywhere in the box. The positions of particles 2 and 3
are then specified relative to particle 1 (it is useful to draw a picture). Thus, the
configuration integral QN(V, T) contains terms which have volume dependence
of the form
where a can range anywhere from a = 1 to a = N. The
integrals over cluster functions, however, are always proportional to the volume.
We can write

va,

(9.39)
where bl(V,
Thus, the ex
proportional
We are no
state. From

Converted with

s on the density.
n Eq. (9.15) is

STOI Converter
trial version

the equation
tential as

hDP://www.stdutilitv.com

of

(9.40)

The pressure then becomes


__
P -

n(V, T,

J-L') _ k
-

~ bl(T, V)ef3IJL'
BT L...J
31
1=1

(9.41 )

AT

and the particle density is given by


(N) __

_!_
V

(8n) _ ~
8J-L'

lbl(V, T)ef3IJL'

-L...J

V,T

1=1

(9.42)

AT

The virial expansion of the equation of state is an expansion in powers of the


density,

PV

00

(N)kBT = ~BI(T)

((N))

I-I

(9.43)

499

VIRIAL EXPANSION OF THE EQUATION OF STATE

To determine the lth virial coefficient, B" let us for simplicity consider all
equations
in the thermodynamic
limit (V --t 00, (N) --t 00, such that
(N)/V =constant). Then bl(T, V) ---t limitsv-+cxlil(T). If we combine Eqs.
(9.41), (9.42), and (9.43), we obtain

(t

OI(T~~{3/~')
1=1

(tn1;n~1ne{Jnp,) tBf (t nl1;n,~;:,t!n'~)


-1=

AT

n=1

/'-1

/'=1

n'=1

(9.44)
If we now expand both sides of Eq. (9.44) and equate coefficients of equal
powers of A':r3exp({3p/), we obtain the following expressions for the first four
virial coefficients:
B 1 (T)

= 01(T) =

(9.45)

1,

(9.46)

B2(T) = -02(T),

(9.47)
(9.48)

Converted with
and so on.
larger and lar
The functi
show that

STOI Converter
trial version

determined

by

quently, we can

hDP://www.stdutilitv.com

Similarly,

and

These identities enable us to simplify the expressions


and write them as

for the virial coefficients

(9.49)

500
where

INTERACTING FLUIDS

Vn(ql, ... ,~)

is called a star function and

Vn(ql, ... ,~)

= L(all

different n-particle

star graphs).

(9.50)

An n-particle star graph is a connected graph which is so tightly connected that


if we remove any point and all lines that connect to it the remaining graph will
still be connected.
In terms of graphs, the first few star functions are given by

V2(Ql,q2) = ..... ,

A,

V3(Ql,Q2,Q3) =

(9.51 )
(9.52)

and
(9.53)
The numeric a
that many waj
different grapl
coefficient is (
We are no
experimental (

Converted with

STOI Converter
trial version

~te that there are


in topologically
nth-order virial
icle graphs.
this theory to

hDP://www.stdutilitv.com
9.C.2. The Second Virial Coefficient [1]
The second virial coefficient gives the correction to the ideal gas equation of
state due to two-body clustering. For very dilute gases, two-body clusters give
by far the dominant contribution from interactions in the fluid and it is sufficient
to terminate the virial expansion at second order.
From Eqs. (9.10), (9.28), (9.39), and (9.46), the second virial coefficient can
be written

where we have changed to center of mass and relative coordinates and have
integrated over the center of mass. The behavior of the second virial coefficient
has been studied for a variety of interparticle potentials. For very simple
potentials it can be computed analytically, and for realistic potentials it must
be computed numerically. We shall focus on two potentials which historically
have been important in understanding the behavior of the virial coefficients.
They are the square-well potential and the Lennard-Jones
6-12 potential
(cf. Fig. 9.3).

501

VIRIAL EXPANSION OF THE EQUATION OF STATE

V(q)

V(q)

(a)

(b)

Ra

Fig. 9.3. Sketch of (a) the square-well potential and (b) the Lennard-Jones 6-12
potential.

9.C.2.1. Squa
The square-we

Converted with

orm

STOI Converter

(9.55)

trial version

hDP://www.stdutilitv.com

The square-we
uare attractive
region of depth e and width (R - 1)0'. The second virial coefficient can be
computed analytically and has the form
B2(T)sw

27T'a2
= -3[I -

(R - I)(e

{3e

- I) .

(9.56)

Note that B2(T)sw differs from B2(T)HC by a temperature-dependent term (see


Ex. 9.2). At low temperatures it is negative and at high temperatures it becomes
positive. At low temperatures the attractive interaction energy due to the square
well can compete with the thermal energy, kBT, and causes a lowering of the
pressure relative to the ideal gas value. At high temperature the hard core
becomes dominant and the pressure increases relative to that of an ideal gas.
We can write B2(T)sw
in a reduced form if we let B*(T)sw =
B(T)swlbo(bo
= 27T'0'3/3) and T* = kBT [e. Then we find
Bi(T)sw

= [I -

(R3 - 1)(e1/T

I)J.

(9.57)

Equation (9.57) will be useful when we compare the square-well results to


experiment.

502

INTERACTING FLUIDS

9.C.2.2. Lennard-Janes 6-12 Potential


A potential which gives a very good approximation to the interaction between
atoms is the Lennard-Jones 6-12 potential,

W 12_W 6]

Vu(q) -.

(9.58)

(cf. Fig. 9.3b). The Lennard-Jones potential has a gradually sloping hard core,
which takes account of the fact that particles with high energy can, to some
extent, penetrate the hard core. When q = a we obtain Vu(u) = O.Thus, q = a
is the radius of the hard core when the potential changes from positive to
negative. The minimum of the Lennard-Jones potential occurs when
q = (2)1/6u
and the value of the potential at the minimum is
VU((2)1/6u) = +E, Thus, e is the depth of the Lennard-Jones potential.
The second virial coefficient for the Lennard-Jones potential can be found
analytically in the form of a series expansion. If we integrate Eq. (9.54) by parts
and introduce the notation x = q G', T* = k8T c, and B* T
= B2(T)u/bo,
we find

Converted with

B~(T)u

STOI Converter

6J}.

(9.59)

trial version
If we expand
be computed

f the series can


n for Bi(T)u:

hDP:llwww.stdutililV.com
,

00

B2(T)u

~:::>~n
n=O

1)
T*

2n+1)/4)

'

(9.60)

where the coefficients a are defined in terms of gamma functions as

= _ 2_
4n!

(2n - 1).
4

(9.61)

Values of the coefficients an for n = 0,


,40 are given in Ref. 1, p. 1119. In
Table 9.1 we give all values for n = 0,
,12. The expansion for Bi(T)u
converges rapidly for T* > 4, but more slowly for lower values of T*. Values of
Bi(T)u for T* ranging from 0.3 to 400 are given in Table 9.2.
In Fig. 9.4 we plot Bi (T) versus T* for both the square-well potential and the
Lennard-Jones potential. We also give experimental values of Bi (T) for a
variety of substances. The Lennard-Jones potential gives values of BiCT) in
good agreement with experimental results. At high temperatures, Bi (T)u and
the experimental points for He gas exhibit a maximum, while Bi(T)sw does
not. The maximum occurs because at high temperatures, particles can penetrate

503

VIRlAL EXPANSION OF THE EQUATION OF STATE

Table 9.1. Coefficients for the Expansion of the


Second Virial Coefficient for Lennard-Jones 6-12
Potential"
n
0
1
2
3
4
5
6

an

an
+ 1.7330010
-2.5636934
-0.8665005
-0.4272822
-0.2166251
-0.1068205
-0.0505458

7
8
9
10
11
12

-0.0228901
-0.0099286
-0.0041329
-0.0016547
-0.0006387
-0.0002381

aBased on Ref. 1.

Table 9.2 Values of the Reduced Second Virial


Coefficient versus the Reduced Temperature for
the Lennard-Jones Potential"

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com
2.00
2.50
3.00
3.50

-0.6276
-0.3126
-0.1152
+0.0190

40.00
50.00
100.00
400.00

+0.5186
+0.5084
+0.4641
+0.3583

aBased on Ref. 1.

into the hard core and lower the amount of excluded volume. The square-well
potential has a hard core with infinite slope and cannot account for this effect,
while the Lennard-Jones potential, with a sloping hard core, can account for it.
The data points for He deviate from the classical results at low temperature.
These deviations are due to quantum effects and will be discussed in
Sec. (S9.D). The second virial coefficients for all classical gases, when plotted
in terms of reduced quantities, are identical. This is an example of the law of
corresponding states.
In Table 9.3 we have given a list of values for the parameters c/kB and a for
various substances, assuming they interact via square-well or Lennard-Jones
potentials. These parameters are obtained from experimental values for the

S04

INTERACTING FLUIDS

B*

o N2

.A
A Ne

A He

Fig. 9.4. The reduced second virial coefficient. The solid line is the calculated curve for
the Lennard-Jones 6-12 potential. The dashed line is the calculated curve for the squarewell potential
r
rT'O,'

.1
.1.
c.
the gases listed.
n

second virial
gases provide
effective inter
found if we ta

'"

Converted with

(Based on Ref.

STDOConverter
trial version

hDP://www.stdutilitv.com

~------------------------------~

~fficients of real
he form of the
ues of e can be
nt temperatures,

Table 9.3 Values of the Parameters a and c for the Square-Well (S-W) and
Lennard Jones (L-J) 6-12 Potentials Taken from Experimental Data on the Second
Virial Coefflcienr"

Argon (A)
190C -+ 600C
Krypton (Kr)
160C -+ 600 C
Nitrogen (N2)
150C -+ 400 C
Carbon dioxide (CO2)
DoC -+ 600C
Xenon (Xe)
OC -+ 700C
Methane (CH4)
DoC -+ 350C
aBased on Ref. 10.

Potential

a (A)

S-W

1.70

3.067
3.504
3.278
3.827
3.277
3.745
3.571
4.328
3.593
4.099
3.355
3.783

L-J

S-W

1.68

L-J

S-W

1.58

L-J

S-w

1.44

L-J

S-w

1.64

L-J

S-W
L-J

1.60

c/kB

(K)

93.3
117.7
136.5
164.0
95.2
95.2
283.6
198.2
198.5
222.3
142.5
148.9

505

VIRIAL EXPANSION OF THE EQUATION OF STATE

say T, and T2, equate the two ratios

B2(T2))
( Bl (Tl)

(Bi(c, T2)theor)
B2( C, Tt)theor '

exper =

(9.62)

and solve for 6. In this ratio, all dependence on a is eliminated. Once we find e,
we can find a if we note that

B2(T)expe, = 2~~ Bi(e:, T)th"",.

(9.63)

The only unknown in Eq. (9.63) is a,


The Lennard-Jones 6-12 potential is perhaps the most widely used
interparticle potential, but there are many other forms of potential that may
be used to compute the virial coefficients, and some of them give better
agreement with experimental results over a wider range of temperature than
does the Lennard-Jones potential.
Computation of the third virial coefficient is more challenging but still can be
done for some simple potentials. However, new effects can enter, such as three
body forces WI .
~t.
.
II. here. Some of
these effects ar
Converted with
~the third virial
coefficients for
e-well potential
and by the LeI
_1.

.1 .

.1

.1.

STOI Converter

I
I

trial version

EXERCU

sphere gas of
equation of st'-- __

T), for a hard

hDP:llwww.stdutililV.com
--. ....
...---_~
..--.L

'

V. Write the
__._n
the density.

: Answer: The second virial coefficient, B2(T), is defined as


!

(1)
where q21 = q2 - ql and!(q2t) = (e-!3V(q2d - 1). The interaction potential,
i V(q21), for
hard spheres is given by V(q2t) = 00 for q21 < Rand
, V(q21) = 0 for q21 > R. We can make a change of variables and let
q = q2 - ql and Q = (ql + q2) Then

B2(T) =

-~~J J

= +27r
I
I

dQ

27rR3

Jo

dqf(q)

dq q"

= --

-U

dq(e-PV(q)

- 1)

(2)

== boo

The equation of state, to first order in the density, is

L_

_P_V_=l+boN+

N._kB_T

---V--

....
~

506

INTERACTING FLUIDS

9.C.3. Higher-Order Viral Coefficients [2]


As we increase the density or decrease the temperature of a gas, at some point it
will undergo a transition to a liquid or a solid phase. Past this point, and well
before we reach it, the virial expansion ceases to give good agreement with the
observed equation of state. It is possible to obtain some idea of the limits of the
virial expansion for a hard-sphere gas. Although hard-sphere gases do not exist
in nature, they can be constructed on a computer by means of molecular
dynamics calculations. In these calculations, the motion of a finite number of
particles is followed by solving the equations of motion of the system. The
phase space coordinates of all the particles can be obtained at any given time,
and the phase functions of interest can be computed. The thermodynamic
properties are obtained by time averaging of the phase functions of interest.
The second through sixth virial coefficients for systems of hard spheres have
been computed
[11] and the results
are B2(T) = o, B3(T) =
B4(T) = O.29b6, Bs (T) = O.llbri, and B6(T) = O.04bg, where bo = 2m? /3
and CT is the hard-core radius. The virial expansion for the equation of state is
plotted in Fig. 9.5 and compared to "experimental"
results from molecular
dynamics cal
.al expansion up

ib6,

Converted with
16
14

s-

STOI Converter
trial version

hDP:llwww.stdutililV.com

12
18

I:x:I

...sc 10

16

s
c,

14

12
6
4

10
B
molecular
dynamics

0
0.2 0.3 0.4 0.5 0.6 0.7

2
0.4 0.5 0.6 0.7 0.8 0.9

Vo/V

Vo/V

Fig. 9.5. Equation of state for hard-sphere and hard-disk fluids. (a) Hard-sphere fluid:
Line 1 includes all virial coefficients up to Bs; line 3 gives the prediction of the Pade
approximate; molecular dynamics results are indicated by the dots. (b) Hard-disk fluid:
All quantities have meaning similar to part (a). (Based on Ref. 11.)

507

SPECIAL TOPICS: THE PRESSURE AND COMPRESSIBILITY EQUATIONS

to and including Bs(T) and the virial expansion up to and including B6(T). The
agreement gets better as we add more terms, but it is not extremely good at
higher densities. Better agreement occurs if we construct the so-called Pade
approximate to the virial expansion. The Pade approximate is obtained as
follows. If we are given a finite expansion for some function F(x),
F(x) = ao

+ alX + ... + a.x",

(9.64)

we can construct the Pade approximate to F(x) by writing it in the form


F{x) = bo +b1x+b2
Co

+ ... +bmX".

+ CIX + ... + c[xl

(9.65)

If we expand Eq. (9.65) in a power series, we can find the coefficients b, and Cj
from the coefficients an. If we know a finite number of coefficients, an, then we
can find a finite number of coefficients b, and Cj. By doing this, we replace a
finite series expansion by an infinite series expansion.
Ree and Ho
'
to the equation
of state for a
Converted with
PV

STOI Converter

NkBT

trial version

(9.66)

hDP://www.stdutilitv.com

The Pade app


g. 9.5. It gives
much better results than the simple virial expansion. It is worth noting that
the hard-sphere gas exhibits a phase transition from a fluid state to a
crystalline state as the density increases. While we will not give the details
here, Fig. 9.5 shows the results of an analogous calculation for a system of
hard disks. The drop in the pressure at about Vo/V = 0.75 indicates a transition
to a crystalline state. The Pade approximate gives poor agreement for the crystal
phase.
Expressions for higher virial coefficients for square-well potentials have
been obtained, but, as we can see from the results for hard spheres, a simple
virial expansion does not give very good agreement for dense systems .

.....SPECIAL TOPICS
.....S9.A. The Pressure and Compressibility Equations
It is possible to express other thermodynamic quantities in terms of the radial
distribution function. Below we derive expressions for the pressure and the
compressibility in terms of the radial distribution function.

508

INTERACTING FLUIDS

~ S9.A.l. The Pressure Equation


We can obtain an expression for the pressure by using a simple trick. Consider a
box of volume V = L3 (L is the length of a side) which contains N particles. We
will assume the box is large enough that we can neglect the effects of the walls
(in the thermodynamic
limit this is rigorously true). From Eqs. (2.98) and
(7.44), we can write

(9.67)

The configuration

QNCV, T)

integral can be written

1: 1:

= VN

dq3Ne-fJV(qN)

dql ...

J:...J:

dxl

... dx3Ne-PV(L>"),

(9.68)

Converted with
where we hav
QN(V, T) with

STOI Converter

of

trial version

8QN(V, T
(

~ derivative

hDP://www.stdutilitv.com

8L

- f3VN

Jo J dx, ... dx
1

...

3N

(8V(LX
8L

))

e-f3v(uN).
T,N

(9.69)
If we now note

we can combine Eqs. (9.67), (9.69), and (9.70) to give


(9.71)
For spherically

symmetric

_!_
kBT

potentials,

'!.. _ 47rn2f3
V

(')0

Jo

Eq. (9.71) takes the form


3d aV(q) fl( . V T)
q q 8q 82 q, , .

(9.72)

509

SPECIAL TOPICS: THE PRESSURE AND COMPRESSIBILITY EQUATIONS

Thus we find that the pressure depends only on the radial distribution function.
Eq. (9.72) is called the pressure equation.
~ S9.A.2. The Compressibility

Equation

As a final example, let us obtain an expression for the isothermal


compressibility, KT,N' For this, we need the variance of the particle number,
as was shown in Eq. (7.117), and therefore we must now describe the system in
the grand canonical ensemble. The average values of one- and two-body phase
functions can be written
(01)

dqlOI (qt)nj(ql;

dql

dq202(qt> Cb)ni(ql' q2; V, T)

(9.73)

V, T)

and

(02)

where ni( ql ; \I
distribution fui
defined

Converted with

STOI Converter
trial version

nr(ql,

,q/;

hDP://www.stdutilitv.com

(9.74)
b-body reduced
mble. They are

I.

~e

_(3(HN_JlN)

(9.75)
The one-body and two-body reduced distribution functions, in the grand
canonical ensemble, have the property that
(9.76)
and
(9.77)
If we now combine Eqs. (7.117), (9.76), and (9.77), we can write
(N)kBTKT
V

1
-1 = (N)

J J
dql

J-t
J-t
dq2 [JL(
n2 Ql,Q2, . V,T ) -n1(Ql;V,T)n
1(Q2;V,T)].

(9.78)

510

INTERACTING FLUIDS

For a system whose particles interact via a spherically symmetric potential, it is


useful again to let n~(q12; V, T) = (N /V)2g~(qij; V, T). Then Eq. (9.78) takes
the form

nkBTKT = kBT(;;t

[1 + (n) LOO dq41fi[~(q;

v,

T) -

1J],

(9.79)

IV.

where (n) = (N)


Equation (9.79) is called the compressibility equation.
The compressibility
equation can be written in terms of the structure
function, h(q) == g~(q) - 1, and takes the form

M(;l

= 1 + (n)

(9.80)

h(q)dq.

The structure function contains all possible information about correlations


between particles
a distance q apart. Near a phase transition,
where
00, it must become very long-ranged so the integral will diverge.

(an/aph ~

~ S9.B. Orr
Ornstein and
function into
structure func

Converted with

STOI Converter

ng the structure
They wrote the

trial version

hDP://www.stdutilitv.com
J

(9.81 )

Equation (9.81) is called the Omstein-Zemicke


equation. The first term, C(q12)
(called the direct correlation function), contains the effects of short-ranged
correlations, and the second contains the effect of long-ranged correlations and
allows for interactions between the particles 1 and 2 which first propagate
through the medium. If we introduce the Fourier transform

h(k) = dq12eik'Q"h(qI2)

(9.82)

and a similar one for C(q12), we can write Eq. (9.81) in the form

h(k) = C(k)

+ (n)C(k)h(k),

(9.83)

where C(k) is the Fourier transform of the direct correlation function, C(q12)'
In terms of the quantity h(k), the compressibility equation takes the form

kBT (an)
8P

= 1 + (n)h(k = 0).
T

(9.84)

SPECIAL TOPICS: ORNSTEIN-ZERNICKE

511

EQUATION

S(k)
2

12

10

Fig. 9.6. The structure factor, S(k) = Snn(k), versus k (in A -1 ) for liquid 36Ar obtained
from neutron scattering experiments (dots), and from molecular dynamics (solid line)
using the Lennard-Jones potential with parameters fitted to Argon. Reprinted, by
permission, from J. L. Yarnell, et. al., Phys. Rev. A7, 2130 (1973).

If we take the
1 (8P)

kBT

8n

Converted with

STOI Converter

r.) dq C(q).

trial version

hDP:llwww.stdutililV.com

(9.85)

Since (8Pj8n
short-ranged.
The direct correlation function has been obtained from molecular dynamics
simulation for a classical fluid of 864 atoms which interact via a Lennard-Jones
potential, V(r) = 46 [(O"/r)12 - (0"/r)6]. In Fig. 9.6 we first make a comparison
between the structure factor, Snn(k), obtained from neutron scattering
experiments on argon and that obtained from molecular dynamics using the
Lennard-Jones potential. The agreement is extremely good. The direct
correlation function obtained from the molecular dynamics using the
Lennard-Jones potential is shown in Fig. 9.7. The direct correlation function
is negative in the region of the hard core and then rises sharply to a positive
value outside the hard core.
The Ornstein-Zernicke equation, in general, must be solved numerically.
However, Percus and Yevick [14] introduced an approximation which gives
quite good agreement with the experimentally measured radial distribution
function; and for the case of a hard sphere gas can be solved analytically. Percus
and Yevick chose to write the direct correlation function in the form
Cpy(q)

= gpy(q)(1

- ef3V(q)).

(9.86)

When this is substituted into the Omstein-Zernicke equation, one obtains the

512

INTERACTING FLUIDS

qC(q)

o
-2

-4

q/(J

Fig. 9.7. The direct correlation function, qC(q), versus q] (a is the hard-core radius)
for a Lennard-Jones fluid, obtained from a molecular dynamics simulation. Reprinted,
by permission, from L. Verlet, Phys. Rev. 165, 201 (1968)

Percus- Yevick equation

Converted with

~)hPY(q23),

(9.87)

STOI Converter

where brv
Watts [
pute the equation of
state for a
trial version
Jones potential. In
Fig 9 8 ~
The Percus- Yevick
equ~ti~n d
vapor phase. There
is a large region where the Percus- Yevick equation has no solution. This region
separates different branches of each isotherm of the equation of state and
corresponds to the coexistence region.
Wertheim [16] and Thiele [17] independently showed that the Percus-Yevick
equation for a hard-sphere fluid can be solved analytically. Since the radial
distribution function obtained from the Percus- Yevick equation is approximate,
we do not expect the pressure and compressibility equations to give the same
results for the equation of state. Thiele obtained the following equation of state
from the pressure equation:

hDP://www.stdutilitv.com

(9.88)

where x = n. From the compressibility equation, he obtained


(9.89)
In both equations the pressure becomes infinite at x = 1, which is a density

513

SPECIAL TOPICS: THIRD VIRIAL COEFFICIENT

P
kBT
0.12

0.10

0.08
,
1 ,:

.,

0.06
2

(n*)-l

Fig. 9.S. Isotherms of the equation of state for a Lennard-Jones fluid obtained by
solving the Pe
kBT Ie, where (1)
r: = 1.2, (2)
Converted with
s are experimental
points for arg
Yevick equati

STOI Converter
trial version

line, the Percus-

greater than
unphysical re
good agreemem~rm-nre:re"SID'CS{~ffO]:eclIIID---mnramTI~~eti'

hDP:llwww.stdutililV.com

e equations give
nsities they give
ments [18].

.... S9.C. Third Virial Coefficient [1, 3]


The third virial coefficient contains the contribution
the gas and may be written in the form

83(T)

=-

3~

JJJ

dql d'l2d'b U3(qt>q2,'b)

from three-body

+~

(JJ

clusters in

dql d'12 U2(Ql,'I2))

(9.90)
Equation (9.90) is completely general and does not depend on the form of the
interparticle potential. In deriving microscopic expressions for the cluster
functions, we have assumed that the N-particle potential was additive-that
is,
that it could be written as the sum of strictly two-body interactions:
(1/2)N(N-l)
yN(qN)

L:
(ij)

Yij(qij)'

(9.91 )

514

INTERACTING FLUIDS

In fact, in a real gas this is not sufficient. For example, if three bodies in a gas
interact simultaneously, they become polarized and an additional three-body
polarization interaction occurs. This polarization interaction has a significant
effect on the third virial coefficient at low temperatures and must be included
[19]. Let us write the total three-body interaction in the form
(9.92)
where ~ V123 is the three-body polarization interaction. The polarization
interaction has been computed by several methods [20,21] and has the form
~V123 =

0:(1 + 3COS(,1)COS(,2)COS(,3))

(9.93)

q12q13q23

,i

where qij are the lengths of the sides of the atomic triangle,
are the internal
angles, and 0: is a parameter that depends on the polarizability and properties of
the two-body interaction. Thus, the polarization interaction is repulsive.
If we inc
.
. .
can write the third
virial coeffi
Converted with

where

STOI Converter

(9.94)

trial version

hDP:llwww.stdutililV.com

(9.95)

and

The third virial coefficients for additive potentials and the corrections due to
nonadditive polarization effects have been obtained for a number of potentials.
We shall discuss some of these results below.
.... S9.C.l. Square-Well Potential
An expression for the additive third virial coefficient for the square-well
potential was first obtained analytically by Kihara [1, 21]. For R ~ 2, it has the
form
B3(T)~~

= kbMs

+ 32R3 - lS)x
+ 32R3 + 18R2 - 16)x2
+ 18R2 - 6)x3],
4

(R6 - 18R

- (2R6 - 36R4
- (6R6 - 18R4

(9.97)

515

SPECIAL TOPICS: THIRD VIRIAL COEFFICIENT

B;

B;,add
(a)

r\.".sw

,,

r ' \

(R-2)

\
\

LJ: ,\ \ ...
'" I I ,
0.4

"

, I
I I

'f-

sw

(R-1.6)

o J..-f+--i'----I
- 0.4 ~...._

.........
-......-.....-~

0.5 1

5 10

T*

o '--...L-.-'-_....::::.:L-~
0.5 1

10

T*

Fig. 9.9. The third virial coefficient. (a) The additive contribution as calculated for the
Lennard-Jones 6-12 potential and the square-well potential for R = 1.6 and R = 2. (b)
The polarizatio
..
. 1 in a. Based on Ref. 10.)

Converted with
while for R~

STOI Converter
trial version

(9.98)

hDP://www.stdutilitv.com

where x = [e(3c - 1]. In computing the third virial coefficient, the values of R, e,
and (J obtained from the second virial coefficient are generally used. Sherwood
and Prausnitz have plotted values of B3 (T)~~ and the correction AB3 (T)sw for
a variety of values of R. Their results are shown in Fig. 9.9. At low temperature,
the contribution from polarization effects can have a large effect. The corrected
values give better agreement with experimental results than does the additive
part alone. In Fig. 9.10 we compare the additive and corrected results with
experimental values of B3(T) for argon.

~ S9.C.2. Lennard-lones 6-12 Potential


The additive third virial coefficient for the Lennard-Jones potential can be
obtained analytically in the form of a series expansion in a manner similar to
that used for the second virial coefficient. However, it is much more
complicated. The series expansion takes the form
00

B3(T)u

= b~ ~

1)

{3n p

-(n+l)/2

(9.99)

516

INTERACTING FLUIDS

B3(T)

(cc/mole)2

Fig. 9.10. Comparison of observed and calculated third virial coefficient for argon. The
solid lines include the polarization effects. The dashed lines give only the additive
contribution. (Based on Ref. 10.)

Converted with

A table of val
them here. C
very similar
curves for
Lennard-Jon
experimental
insufficient curva ure

we will not give


ig. 9.9. They are
e Lennard-Jones
Fig. 9.10. The
ement with the
ential well has

STOI Converter
trial version

hDP://www.stdutilitv.com

EXERCISE 9.3. Compute the third virial coefficient, B3(T), for a hard
sphere gas of atoms of radius, R, confined to a box of volume, V. Use a
geometrical method to find B3(T). Write the equation of state for this gas as
a virial expansion to second order in the density.
Answer: The third virial coefficient is written

B3(T) = - 3~

III

dq[ d'lzd'l3f('lz[V('l3[V('l32)

(1)

First we change from coordinates (q1, q2' q3) to coordinates (q1, q21, ~1)'
where q21 = q2 - q1 and q31 = q3 - q1 The Jacobian of this transformations is one. We can integrate over q1 to obtain

B3(T) = -~

J J d~1dq3J(q21)f(q31)f(q31

- q21)

(2)

Next make the change of variables, q = q31 - q21 and Q = ~(q31 + q21)
The Jacobian of this transformation is one. The third virial coefficient then

517

SPECIAL TOPICS: VIRIAL COEFFICIENTS FOR QUANTUM GASES

I takes the form


B3(T)=-WdqdQt(Q-Ht(Q+Ht(q).

(3)

I
I

The integration
over
accompanying figure.

Q can be reduced

to geometry.

Converted with

; The circle 0
sphere on th
i Q in Eq. (3)
twice the v

h=R-!q.
integration

STOI Converter
trial version

hDP://www.stdutilitv.com

Consider

the

!QI $. R.

The
tegration over
is volume is
R) of height
h). Thus, the

(4)
(there are two spherical caps). Integration

over q gives

B (T) = 57r R6 = ~b2


3
18
8 o
I

(5)

The equation of state, to second order in the density, is


PV
--=
NkBT

2 (N)2

5b

1 +bo-+_o
V
8

+ ... .

(6)

... S9.D. Virial Coefficients for Quantum Gases [1, 22]


For fluids composed of molecules of small mass, such as He, the classical
expressions for the virial coefficients do not give very good results at lower
temperatures
(cf. Fig. 9.4). For such particles, the thermal wavelength

518

INTERACTING FLUIDS

>I.T = (27r1i2/mkBT)1/2 will be relatively large and quantum corrections must be


taken into account. There are two kinds of quantum effects which must be
considered: diffraction effects, which are important when the thermal
wavelength is the size of the radius of the molecules, and the effects of
statistics, which are important when the thermal wavelength is the size of the
average distance between particles.
To find general expressions for the virial coefficients, we proceed along lines
similar to those for a classical fluid except that, for the quantum case,
momentum variables no longer commute with position variables and cannot be
eliminated immediately. The grand partition function for a quantum system can
be written in the form
(9.100)
where >I.T is the thermal wavelength,
(9.101 )
and iIN is
position ope
partition fun

Converted with

momentum and
xpand the grand

STOI Converter
trial version

(9.102)

hDP://www.stdutilitv.com

where UI (13) depends on the momentum and position operators for I particles. If
we equate coefficients of equal powers of the parameter e(3/J.' / A~ in Eqs. (9.100)
and (9.102), we obtain
Trt[Ul (f3)J = Trt[Wl (f3)J,

(9.103)

Tr2 [U 2(13)] = Tr2 [W2 (13)] - (Trt[Wl (13))) 2,

(9.104 )

Tr3[{h(f3)] = Tr3[W3(f3)]

- 3(Trt[Wl (13)]) (Tr2 [W2(f3)])

+ 2(Tri[Wl (13)])3,
(9.105)

and so on. From Eq. (9.102) we can write the grand potential in the form
(9.106)
and the average particle density takes the form
(n)

1(80,)
J.l

V,T

le(3/J.'1

= L ~bl(V,
00

= - V 8'

1=1

T),

(9.107)

519

SPECIAL TOPICS: VIRIAL COEFFICIENTS FOR QUANTUM GASES

where

1
"
Trl [UI(,6)J.
I.V

bl(V, T) == -,
The virial expansion

(9.108)

for the equation of state may be written


(9.109)

where the virial coefficients, PAI(T), are related to the quantities bl(V, T)
through Eqs. (9.45)-(9.48). Thus, PAl = bi, PA2= -b2, PA3= 4b~ - 2b3, and so
on. In Exercises 9.4 and 9.5 we give some examples of quantum corrections to
the classical virial coefficients.

Bf(T),

EXERCISE 9.4. Compute the second virial coefficient,


gas of spin, S = particles.

ideal Fermi-Dirac

!,

for an

[ Answer: Th~""'-"""-"'..J"'____"_"'__
.....................
...,.,._.,.c!.....,
1....."

B~(T)

---.L._,._.__",,1....,.___,,_,,~...,.___

__

Converted with

= -

STOI Converter

wg(,

:, where
! kinetic energ
: can write

rt[W ?(,6)])2,
(1)

Pf 12m

trial version

is the
Eq. (B.30) we

hDP:llwww.stdutililV.com

2 2

(,61i
)
- k2ml

" Trt[e-.8TI]=LL(kl,slle-.8Tllkl,sl)=2Lexp
A

kl

sl=1

=2,\3V

kl

(2)
and
I

Tr2[e-.8(T1+T2)]= ~L

kl

SI

=l

k2 s2=1

[(k I , Sl ., k 2, S2Ie -.8(TI+T2)Ik I, Sl ,. k 2, S2)


- (kl, Sl; k2, S21e-.8(T1+T2)
Ik2, S2; kl, Sl)],
= ~ (L

(kl,Slle-.8hlkl,SI))

kl sl=1

-~L

L
kl sl=l

14V2

= 2:

>..6 T

(L

(k2,S2Ie-.8T2Ik2,S2))

k2 s2=1

(kl,Slle-.8Tllk2,S2)(k2,S2Ie-.8T2Ikl,SI)

k2 s2=1

V
225/2,\3

.
T

(3)

520

INTERACTING FLUIDS

If we combine the above equations, the second virial coefficient becomes


o

.x}

B 2 = 23/2

(4)

Thus, Fermi-Dirac statistics causes an effective repulsion between particles


in the gas.
The only case that we can hope to treat analytically for an interacting
quantum system is the second virial coefficient; because the Schrodinger
equation can sometimes be solved analytically for two-body interactions. Below
we show how that may be done. The second virial coefficient may be written
(9.110)
where
W2(,6) = 2!.x~e-,B(tl+t2+VI2), WI (,6) = ~}e-,B\1'i =pf/2m
is the
kinetic energy operator for the ith particle, and V 12 is the interaction potential
between p .
. raction potential
depends on
Converted with
icles and it may
or may not
V( q, SI, S2). It is
useful to se
a relative part. If
P2' and relative
momentum,
trial version
r can be written
1'1 + 1'2 =
-of-mass kinetic
energy, and
with the relative
motion of the two particles. Let us consider the difference between the second
virial coefficient for the interacting and noninteracting gas. We can write

STOI Converter
hDP:llwww.stdutililV.com

(9.111)
The trace must be taken using symmetrized states for a Bose-Einstein gas and
antisymmetrized states for a Fermi-Dirac gas.
If we consider a spin zero boson gas, the symmetrized states are simply
(9.112)

and we can write

521

SPECIAL TOPICS: VIRIAL COEFFICIENTS FOR QUANTUM GASES

If the bosons have spin, then the states will also depend on spin and the
expression for Tr2 [W2 ((3)] becomes more complicated.
Let us now consider a spin-! fermion gas. We will let j denote Sz = +! and
let 1 denote Sz =
For each value of the momenta, kl and k2' there are four
spin states that must be considered. The total spin of the two particles can be
S = or S = 1. We let Ik1,k2;S,Sz)(a) denote an anti symmetric state with spin
Sand z component of spin Sz. Then the trace must be taken with respect to the
states

-!-

Ikl, k2; 0, 0) (a) = !(lk1, k2) + Ik2, k1)


Ik1,k2; 1, 1)(a)

= ~(lkl,k2)

-lk2,k1))1

Ik1,k2; 1,0)(a)

= !(lk1,k2)

-lk2,k1))(1

Ik1,k2; 1, _1)(a)

= ~(lkl,k2)

-lk2,k1))1

)(1 j, 1) - I L j)),
j, j),
(9.114)

j, 1) + 11, i)),

L 1)

With these states, we can write

Converted with

D..B2,jd = -

STOI Converter

>

, k2; S, Sz)(a).

trial version

(9.115)

hDP://www.stdutilitv.com
..
.

If t~e inter
is, if V 12 = V (g)-then
Eq. (9.115) reduces to

llB2,Jd = - ;~
(a)

Y on spin-that
over spin can be performed easily and

E E[(') (kl; k 1[e-{lt, (e-fJ(t",+v,,) 2

kl

+ 3

the summation

e-{lt~I)llkl' kd')

k2

(kl, k21 [e-.BTcm e-.B(Trel+VI2)

e-.BTrel)] Ik1, k2) (a)],

(9.116)

where
(9.117)
The momentum eigenstates, Ik1, k2), can be written in terms of center-ofmass momentum,
P = iiI{, and relative momentum, p = lik. That is,
Ik1, k2) = IK) Ik). The symmetrized and anti symmetrized states become

Ik1,k2)(')

= ~IK)(lk)

+ 1- k))

and

Ik1,k2)(a)

= ~IK)(lk)

-1-

k)),

(9.118)

522

INTERACTING FLUIDS

respectively. If we substitute Eqs. (9.118) into Eqs. (9.113) and (9.116) and
perform the sum over the center-of-mass momentum (first change it to an
integral), we find for the Bose-Einstein gas
(9.119)
and for the Fermi-Dirac gas
l::iB2,Jd =

_23/2,\3

T L[(s)

(k] (e-,B(Trel+VI2)

e-,BTr"l)

Ik) (s)
(9.120)

+ 3 (a) (kl (e-,B(T

re1+VI2)

_ e-,BTr"l)

Ik) (a)].

Before we can proceed further, we must compute the energy eigenstates,


lEn), of the relative Hamiltonian, fIrel = p2/m + V(q). Let us denote the
position eigenstates by Ir) so that qlr = rlr). Then the eigenvalue problem,
fIrellEn) = E

Converted with

STOI Converter
trial version

where 'lj;n(r)

~::::~;~~t~
hDP://www.stdutililV.com

(9.121)
ete set which we

~~~~g;
~~;::~

'Frel. If the interaction potential is attractive, then the spectrum of fIrel may
contain contributions from bound states, in addition to states which extend over
the size of the container. In the limit of infinite volume, the spectrum may
consist of bound states in addition to a continuum of states.
It is useful now to change to the position basis. If we also insert a complete
set of energy eigenstates, we find

and for the Fermi-Dirac gas we have


t:ill2,{d = _23/2A~

I dr [ (~I'IjJ~)(r)12e-~E'

- ~

1'IjJ~?(r)12e-~E")

+ 3 ( ~ 1'IjJ~a)(r)12e-flE"- ~ l'ljJt)(r)12e-~E., ) ] ,
where 'lj;~)(r) = (s)(rIEn) and 'lj;~a)(r)

=(a)

(rJEn).

(9.123)

SPECIAL TOPICS: VIRIAL COEFFICIENTS FOR QUANTUM GASES

Let us now restrict

ourselves

to spherically

symmetric

523

potentials,

v(r) = V(lrl). The energy eigenstates may be written in the form


1

00

V;n(r) =

L: L: Rn,/(r)YI,m(O, </J),

(9.124)

1=0m=-I

where r = (r, 0, </J) denotes the relative displacement in spherical coordinates,


and YI,m(O,</J) is a spherical harmonic. The coefficient, Rn,/(r), satisfies the
Schrodinger equation
2

2
+ 1(1+ l)ti
2 R

d (rRn,/(r))
2
mr
dr
ti

mr'

n.l

+ V()R
r

()r

n.l

()r -- E nl R n.l ()r .


"

(9.125)

If we note that
00

V;n(-r) =

L: L: (-I)/Rn,/(r)YI,m(O,

</J),

(9.126)

1=0m=-I

then

Converted with

STOI Converter

and

(9.127)

trial version
v;~a)

hDP://www.stdutilitv.com

~--~V~4~1~=~0-m-=-~I~----------------~

(9.128)

We can now substitute Eqs. (9.127) and (9.128) into Eqs. (9.122) and (9.123). If
we use the orthonormality of the spherical harmonics
271"

Jo

11 d(cos(O))Yl~,m'((}'

d</J

</J)Y1,m(O,</J) = 81',18m',m,

(9.129)

-1

and use the fact that the energy eigenstates must be normalized to one so that
I;' r2dr IRn,l(r)12 = 1, we then find for the spin-O Bose-Einstein gas
AB2,be =
and for the spin
AB2,Jd =

_23/2A~

E(41
1=0

-! Fermi-Dirac

_23/2A~

E
1=0

+ 3(41 + 3)

+ 1) [L:e-fJE",21

L:e-fJE"o,2/]

(9.130)

no

gas

[(41 + 1)

(L:e-fJE",21
n

L:e-fJEno,21)
no

(~>-~E",uH - ~,>-~E...,uH)J.

(9.131)

524

INTERACTING FLUIDS

For the spin-O Bose-Einstein gas, only even angular momentum states
contribute to the second virial coefficient. For the spin-! Fermi-Dirac gas,
only even angular momentum states contribute for pairs of particles with total
spin S = 0, and only odd angular momentum states contribute for pairs of
particles with total spin S = 1.
We can simplify the calculation of the second virial coefficients still
further with a simple trick. Let us assume that the system is contained in
a large spherical box of radius R and that the surface of the sphere has
infinitely hard walls. The most general form of solution of Eq. (9.125) can be
written
Rn,l(r)

An,I[COS( 8n,I)jz(kn,lr)

- sin( 8n,l)nl(kn,zr)],

(9.132)

where kn,z = y'iiiE,;i/1i and 8n,1 is the phase shift of the wavefunction relative to
its value for the noninteracting system. For the noninteracting system, 8n,z = O.
The Bessel functions have the property that for large argument,
limx-+ocJz(x) -t (1/x)sin[x -17r/2] and limx-+oonz(x) -t -(1/x)cos[x
- 17r/2].
Thus, near
.

Converted with

STOI Converter
At the outer
we must ha

trial version

(9.133)
e interacting gas

hDP://www.stdutilitv.com
kn,zR - 217r+ 8n,z

nnt,

(9.134)

and for the noninteracting gas we must have

k zR - -2117r= no z7r,
0,

(9.135)

where ni and no,l are integers which label the allowed energy levels in the
spherical container. If we take the derivative of Eq. (9.133) [Eq, (9.134)] with
respect to kn,l(kno,z) and then take the difference of the resulting equations, we
find
(9.136)

This is useful if we note that in the limit as R -t 00, the energy levels in the
container form a continuum and we can write En = f dk(dn/dk),
where dnf dk
is the density of states.
Let us again consider the virial coefficient for the spin-O Bose-Einstein gas
[Eq. (9.130)]. If we explicitly separate out the bound state contributions (and

525

SPECIAL TOPICS: VIRIAL COEFFICIENTS FOR QUANTUM GASES

suppress their dependence on I), we can write


~B2,be

= - 23/2

A~ :l::e-,BE",bd
n,bd

(9.137)

Similarly, we find for the spin-! Fermi-Dirac gas


A D

UD2,/d

23/2

\ 3 "'"
AT L....t

e-,BE",bd

n,bd

(9.138)
We cannot sa
specify the sha
an exercise.

EXERCI
I hard sphere

ient, unless we
~e leave this for

Converted with

STOI Converter
trial version

a dilute gas of

hDP://www.stdutilitv.com

~----------------------------~
For hard spheres of radius R = a, in an infinitely

Answer:
: scattering phase shift, 81(k), is given by
!

jz(ka)
tan(8z(k))

large box, the

(1)

= nz(ka)

, For a dilute gas of bosons at fairly low temperature, the momenta, k, of the
particles will be fairly low. Therefore, we can approximate the Bessel
, functions in Eq. (1) by their asymptotic limits for small argument, and we
find
I

tan(8 (k)) ~
z

(k )2/+1
----'_a~ __
~
(21 + 1)(1 . 3 ... (21 _ 1)) 2

(2)
.

Thus for small ka, the dominant term comes from the partial wave, 1 = 0,
and 8o(k) ~ -ka. Hard spheres have no bound states. Therefore, the second
virial coefficient can be written
(3)

526

INTERACTING FLUIDS

The second virial coefficient for the noninteracting gas is B~ = - A~/25/2


Thus, we find

(4)
While Bose-Einstein statistics tends to lower the pressure of the boson gas, .
the hard core tends to increase it.

REFERENCES
1. J. O. Hirschfelder, C. F. Curtiss, and R. B. Bird, Molecular Theory of Gases and
Liquids (John Wiley & Sons, New York, 1954).
2. T. K. Hill, Statistical Mechanics (McGraw-Hill, New York, 1956).
3. D. A. McQuarrie, Statistical Mechanics (Harper & Row, New York, 1976).
4. D. Chand
d University Press,
Oxford, 1
Converted with
, Englewood Cliffs,
5. H. L. Frie
NJ, 1985)
6. J. L. Yarr
~ys. Rev. A7, 2130
(1973).
trial ve rsion
& Sons, New York,
7. J. E. Maye
1941).
8. G. E. Uhlenbeck and G. W. Ford, in Studies in Statistical Mechanics, Vol. 1, edited by
J. de Boer and G. E. Uhlenbeck (North-Holland, Amsterdam, 1962).
9. H. D. Ursell, Proc. Cambridge Philos. Soc. 23, 685 (1927).
10. A. E. Sherwood and J. M. Prausnitz, J. Chem. Phys. 41, 413, 429 (1964).
11. F. H. Ree and W. G. Hoover, J. Chem. Phys. 40, 939 (1964).
12. Reprinted in The Equilibrium Theory of Classical Fluids, edited by H. L. Frisch and
J. L. Lebowitz (w. A. Benjamin, New York, 1964).
13. L. Verlet, Phys. Rev. 165, 201 (1968).
14. J. K. Percus and G. J. Yevick, Phys. Rev. 110, 1 (1958).
15. R. O. Watts, J. Chem. Phys. 48, 50 (1968).
16. M. S. Wertheim, Phys. Rev. Lett. 10, 321 (1963).
17. E. Thiele, J. Chem. Phys. 39, 474 (1963).
18. N. W. Ashcroft and J. Lekner, Phys. Rev. 145, 83 (1966).
19. H. W. Graben and R. D. Present, Phys. Rev. Lett. 9, 247 (1962).
20. B. M. Axilord, J. Chem. Phys. 19, 719 (1951).
21. Y. Midzuno and T. Kihara, J. Phys. Soc. (Japan) 11, 1045 (1956).
22. K. Huang, Statistical Mechanics (John Wiley & Sons, New York, 1987).

STOU Converter

hnp://www.stdutilitv.com

527

PROBLEMS

PROBLEMS
Problem 9.1. Compute the second virial coefficient for a gas which interacts via the
potential
00

V(q) =
{

R(>.c_I)

(q - >..R)

if q < R,
if R~q~>..R,
if q > >..R.

Problem 9.2. Consider a classical gas in a box of volume V. Compute the second virial
coefficient, B2, for the Gaussian model,f(qij) = e-lU/t, where a-I/2 VI/3 Sketch the
effective potential, f3V(qij).
Problem 9.3. Compute the second coefficient for the weakly coupled particles with
potential V(q) = Vo for q :::;Rand V(q) = 0 for q > R.
Problem 89.1. Assume at low density the direct correlation function can be
approximated by C(q) = (e-,BV(q) - 1). For a low-density hard-sphere gas (radius R),
use the Omstein-Zemicke equation to obtain the first-order correction (in density) to the
ideal gas expression for the isothermal compressibility. Show that your result agrees
with the result
.
.
nkB T = 1 + nB2,
where n = N /V
Converted with
Problem 89.2. (
coefficient, B3,

STOI Converter

Problem 89.3 .:

trial version

e the third virial

K VI/3.

hDP://www.stdutilitv.com
wheref(q) = (e-,BV(q) - 1) and V(q) is the interaction potential between particles with
relative position, q. The Fourier. transform of f(q) is j(k) = f eikClf(q). The inverse
transform isf(q) = (1/2'lli!e-,~q flk). (aj Show that the third _virial coefficient can
be written B3 =
(1/2'71/ dkf(k) f( -k) f( -k). (b) Compute j'(k) for a gas of hard
spheres with radius R. (c) Use part (b) to compute the third virial coefficient for a hard
sphere gas.

-1

Problem 89.4. Compute the third virial coefficient for weakly coupled particles with
potential V(q) = Vo for q :::;Rand V(q) = 0 for q 2:: R.
Problem 89.5. Compute the third virial coefficient for an ideal Fermi-Dirac gas of spin,
s = ~,particles.
Problem 89.6. Compute the second virial coefficient for a dilute gas of soft-sphere spinO bosons at fairly low temperature. Assume that the spheres have radius r = a and a
potential V(r) = +Vo for r:::; a and V(r) = 0 for r > a. Assume also that

k2 mVo/1i2

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com

PART FOUR
NONEQUILIBRIUM
lViECHANICS

STATISTICAL

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com

10
HYDRODYNAMIC PROCESSES NEAR
EQUILIBRIUM

to.A. INTRODUCTION
When a system is disturbed from its equilibrium state, quantities which are not
conserved du
meters for an
underlying br
Converted with
values. After
~rfe;d~~ll!~:

STOI Converter

s ~; ~~~::i~:d

quantities an
trial version
nonequilibrium
behavior of t
otion for the
densities of
ters are called
the hydrodyn
ed quantities
may include particle number, momentum, and energy. Examples of order
parameters may include average magnetization, or a complex function
characterizing a superfluid state. If there are inhomogeneities in the densities
of conserved quantities, then particles, momentum, or kinetic energy must be
transported from one part of the fluid to another to achieve equilibrium.
Therefore, very-long-wavelength disturbances will take a long time relax,
whereas short-wavelength disturbances relax more quickly. This dependence of
relaxation time on wavelength is a feature that characterizes hydrodynamic
processes.
Hydrodynamic equations describe the long-wavelength, low-frequency
phenomena in a large variety of systems, including dilute gases, liquids, solids,
liquid crystals, superfluids, and chemically reacting systems. For complicated
systems, transport processes are often coupled. For example, in a multicomponent system, it is possible to have a temperature gradient drive a particle
current and a concentration gradient drive a heat current. Some complicated
systems can have as many as 10 or 20 (or more) transport coefficients to
describe the decay to equilibrium from the hydrodynamic regime.
In 1932, Onsager showed that the reversibility of the dynamical laws on the
microscopic level requires that certain relations exist between transport
coefficients describing coupled transport processes. Onsager's relations are of

hDP://www.stdutilitv.com

532

HYDRODYNAMIC PROCESSES NEAR EQUILIBRIUM

immense importance because they enable us to link seemingly independent


transport processes and thereby reduce the number of experiments that must be
performed in order to measure all the transport coefficients. In this chapter we
will derive Onsager's relations and apply them to transport processes in reacting
multicomponent systems and superfluid systems. At the same time we will
show how the hydrodynamic equations for such complicated systems can be
derived from a knowledge of the thermodynamics and symmetry properties of a
system.
Fluctuations about the equilibrium state decay on the average according to
the same linear macroscopic laws (hydrodynamic equations) which describe the
decay of the system from a nonequilibrium state to the equilibrium state. If we
can probe the equilibrium fluctuations, we have a means of probing the
transport processes in the system. The jiuctuation-dissipation theorem shows
that it is possible to probe the equilibrium fluctuations by applying a weak
external field which couples to particles in the medium but yet is too weak to
affect the medium. The system will respond to the field and absorb energy from
it in a manner which depends entirely on the spectrum of the equilibrium
fluctuations. Accordin to the fluctuation-dissi ation theorem the spectrum of
equilibrium
Converted wl-th
rom the external
field can be
is related to the
correlation
In this ch
orem and apply
steps. We first
trial version
elate the spectral
ation matrix for
fluctuations i
eory and use the
assumption of causality to obtain a relation between the real and imaginary
parts of the response matrix.
One of the simplest applications of linear response theory involves a
harmonically bound Brownian particle immersed in a medium. If the Brownian
particle is pulled away from its equilibrium position, it will decay back to
equilibrium and dissipate energy into the fluid. We will obtain an expression for
the linear response function of the Brownian particle and in terms of it derive an
expression for the correlation function for equilibrium fluctuations in the
position for the oscillator.
In the special topics section we discuss important applications of the
fluctuation-dissipation theory-for example, the scattering of light from a fluid.
The electric field of the incident light wave polarizes the particles in the
medium, thus allowing the light to couple to the medium. The light will be
scattered by density fluctuations, and by measuring the spectrum of the
scattered light we can measure the spectrum of the density fluctuations. We will
find that the density fluctuations are of two types: thermal density fluctuations
due to fluctuations in the local entropy and mechanical density fluctuations due
to damped sound waves. For low-frequency and long-wavelength fluctuations,
the spectrum of scattered light can be obtained from the linearized hydro-

STOU Converter

hnp://www.stdutilitv.com

NAVIER-STOKES HYDRODYNAMIC EQUATIONS

533

dynamic equations, and, therefore, light scattering experiments give us a means


of measuring transport coefficients in the fluid.
In the special topics section we shall apply hydrodynamic theory to transport
processes in electric circuits composed of different metals coupled together and
held at nonuniform temperature. We shall also discuss the transport of mixtures
across membranes.
The hydrodynamic equations describe the long time behavior of a few
hydrodynamic modes in a system with 1023 degrees of freedom. The nonhydrodynamic degrees of freedom decay on a much faster time scale than the
hydrodynamic modes and provide background noise. We can use the
fluctuation-dissipation theorem to find the correlation functions for this
background noise as we will show in the special topics section.
It is also possible to derive correlation functions for microscopic Brownian
particles using the hydrodynamic equations. The hydrodynamic flow of the
medium around a Brownian particle creates memory effects which cause its
velocity autocorrelation to decay with a slow long time tail.
The poles of the spectral density matrix give us information about the
spectrum of flu
.
.
rresponding to
very-low-frequ
hydrodynamic
modes in the
, it is possible
to introduce pr
space orthogo
hydrodynamic
hydrodynamic
trial version
modes arise
from broken
symmetries wh

SIDU Converter

~;n!~~~
:

hDP://www.stdutilitv.com

~----------------------------------

lO.B. NAVIER-STOKES
EQUATIONS [1-3]

HYDRODYNAMIC

The Navier-Stokes equations describe the macroscopic behavior of an isotropic


fluid of point particles out of quilibrium. They are essentially the macroscopic
"balance equations" (cf. Appendix A) for the quantities that are conserved
during collision processes on the microscopic scale. The conserved quantities
for an isotropic fluid of point particles include the particle number, the
momentum, and the energy. The balance equations for the conserved quantities
Contain no source terms because the conserved quantities cannot be created or
destroyed on the microscopic scale. In addition to the balance equations for
conserved quantities, it is essential to write the balance equation for the entropy
density of the fluid. Entropy is not a conserved quantity. For a fluid in which
irreversible processes can occur, there will be an entropy source term. The
entropy source term in a fluid is the hydrodynamic equivalent of the Joule
heating which occurs in an electric circuit which has resistance. The entropy
source term enables us to identify generalized forces and currents in the fluid.

S34

HYDRODYNAMIC PROCESSES NEAR EQUILmRIUM

The conductance in a fluid (the proportioality constant between force and


resulting current) is called a transport coefficient. Once the transport
coefficients for the fluid have been identified, we can write the Navier-Stokes
equations.

IO.B.I. Balance Equations


In this section we will derive the balance equations for the mass density,
momentum density, energy density, and entropy density for an isotropic fluid of
point particles. The balance equation for the mass density is especially simple
since particles can't be created and there can be no dissipative particle currents.
The balance equation for the momentum density is based on Newton's second
law. Momentum can only be created in a fluid of external unbalanced forces act
on the fluid, and therefore there can only be a momentum source term if
external forces are present. The balance equations for the energy density and the
entropy density can be written in a general form, and thermodynamic relations
can be used to relate them.
lO.B.1.1.
In the absenc
leaving a collis
total mass of

Converted with

s entering and
processes, the
will also be
uid, V(t) (with
mount of mass
, t) denote the
al mass in the

STOU Converter
trial ve rsion

a given set of
inside this vol
mass density (m ss e
volume, V(t). Then

hnp://www.stdutilitv.com

dM
-d =-dd
t
t

pdV=
V(t)

V(t)

P
(d-d
+PV'rv
t

) dV=O,

(10.1)

where v = v(r, t) is the average velocity of the fluid at point r and time t, and
we have used Eq. (A.16) in Appendix A. Since the volume element, V(t),is
arbitrary, the integrand must be zero and we find
(10.2)

If we note that the convective derivative is given by d/dt


we can also write

= a/at + v . V'r, then


(10.3)

The quantity, J

== pv

is the mass current or mass flux and has units mass/

S3S

NAVIER-STOKES HYDRODYNAMIC EQUATIONS

area- time. It is also a momentum density. The derivative, dp [dt, gives the time
rate of change of the mass density for an observer moving with the fluid. The
derivative, 8p/8t, gives the time rate of change of the mass density for an
observer at a fixed point in the fluid (cf. Appendix A). Equation (10.3) is
sometimes called the continuity equation. It is a direct consequence of the
conservation of mass in the fluid.

lO.B.l.2. Momentum Balance Equation


The total momentum, P(t) = IV(t) pvdV, of the volume element, V(t), evolves
acording to Newton's law. The time rate of change of the momentum, pet), must
be equal to the sum of the forces acting on the volume element, V (t). Therefore
we can write the equation of motion of the fluid element in the form
dP(t) = -d
-dt

dt

pvdV =

V(t)

pFdV

V(t)

fdS,

(10.4)

S(t)

where F is an external force per unit mass which couples to the particles inside
the volume elem
.
..
ample), f is a
force per unit ar
Converted with
S(t) denotes
the surface of th
e to the fluid
surrounding
th
component
perpendicular to
(a fluid with
dissipation) it w
trial version
If we write
dS, directed
outward perpend
P, where P is
the pressure tensor, f = D P, and n is a unit vector
irected outward
perpendicular to the surface. The pressure tensor has nine components. In
Cartesian coordinates it can be written P = p
+P
+ ... + P zzzz, where
X, y, and Z are unit vectors in the x, y, and z directions, respectively. The unit
vector, 0, can be written 0 = nxx + nyY + n,z, where nx, ny, and n, are
components of 0 in the x, y, and z directions, respectively. Note that the ith
~omponent of the vector, f, can be written ji = Lj njPjj, where i = x, y, z and
J
x,y,Z. If we use Gauss's theorem, we can write

STDI Converter
hDP://www.stdutilitv.com
xxxx

xyxy

dS . P =

S(t)

av

r .

(10.5)

V(t)

Here '\l r . P is a vector those ith component


is ('\l r . P)i = Lj 8jPjj =
8Pxi/8x + 8Py;/By + 8P,i/8z and 8, = 8/8x,8y = 8/By, and 8, = 8/8z. Then
the argument of Eq. (l0.4) must satisfy the equation
dpv
dt

+ pV('\l r V

[cf. Appendix A, Eq. (A.16)].

P F + '\l r . P

(10.6)

536

HYDRODYNAMIC PROCESSES NEAR EQUILIBRIUM

For an ideal fluid (no dissipation) the only force on the walls of V(t) is due to
the hydrostatic pressure, P = P(r, t), which is always perpendicular to the walls
and pointed inward. Thus, for an ideal fluid we have f = - P nand P = - PU(U
is the unit tensor, (J = xx + yy + zz). For a nonideal fluid, there is also a
dissipative contribution, n, to the pressure tensor and Eq. (10.6) takes the form
dpv
dt

+ PV(V'r

. v)

pF - V'rP - V'r . n

(10.7)

since V'r . (PU) = V'rP. The tensor fi is called the stress tensor. If we make use
of the convective derivative, df d: = a/at + v . V', we can also write Eq. (10.7)
in the form
apv
at

+ V'r

. (PU

+ pvv + Il)

pF.

(10.8)

The term pvv is the momentum flux. It is a nine-component dyatic tensor,


pvv = pvxvxxx + PVxVyxy + ... + pVzYz;ZZ. Equations (10.7) and (10.8) are
alternative vc
an isotropic fluid
of point pal
Converted with
e of change of
momentum d
describes the
a fixed point

~:';~~~y~

STOU Converter

id, and Eq. (10.8)


by an observer at

trial ve rsion

~P:II~.stdUdlitv.com

wn. We willie! ,
denote the energy per unit mass, and let pe denote the energy per unit volume of
the fluid. For the case when the external force has the form F = - V' r</J(</J is a
potential energy per unit mass), the energy density can be written pe =
pu + pv2 + p</J, where u is the internal energy per unit mass. Thus the energy
density has contributions from thermodynamic processes in the fluid, the kinetic
energy of the fluid, and the external potential energy.
Since the energy inside the volume element cannot be created or destroyed
by collisions, the only way the energy inside the volume element can change is
by flow through the walls of V(t). The balance equation for the total energy can
be written

(10.9)
where J~ is the reactive (nondissipative) energy current and J~ is the dissipative
energy current. Both J~ and J~ have yet to be determined.
The entropy balance equation can be written in an analogous fashion.
However, now we must take account of the fact that entropy can be created
inside the volume element due to spontaneous processes inside of V (t). Let s
denote the entropy per unit mass (the specific entropy) and ps denote the

537

NAVIER-STOKES HYDRODYNAMIC EQUATIONS

entropy per unit volume. The entropy balance equation then becomes
(10.10)

J:

J?

where
is the reactive entropy current,
is the dissipative entropy current,
and (1s is an entropy source term due to disspative processes inside of V(t). The
explicit forms of
and (1s will be determined in the next section.

J:, J?,

to.B.2.

Entropy Source and Entropy Current

Let us now assume that spatial variations of deviations from thermodynamic


equilibrium occur on very large length scales compared to microscopic
distances. Then locally the system will be in equilibrium, but the actual values
of thermodynamic quantities vary slowly in space and time. The fundamental
equation for the internal energy is pu = T ps - P + pii, where jj is the chemical
potential per unit mass [cf. Eq. (2.66)]. If we take the differential of this
equation and use the Gibbs-Duhem e uation, d - + sdT - dP = 0, we find
d(pu) = Td(ps

Converted with

STOI Converter

(10.11)

trial version

hDP://www.stdutilitv.com

t us next note

v Vr(pu) - Tv Vr(ps) - j1vVr(p)

= Vr . [(pu -

Tps - pi1)v] + psv VrT + pvVrj1+ PVr . V

= V

+ pVrj1-

(psVrT

VrP)

(10.12)

0,

where we have used the fundamental equation, pu = T ps - P + pii, and the last
term is zero because of the Gibbs-Duhem equation, pdj1 + psdT - dP = o.
Therefore
(10.13)
Let us now assume, for simplicity, that all external forces are zero so that
F ~ O. Then the total energy per unit volume is pe = pu + pv2, and we can
Wnte

ape
Bou
-=-+v'---v
at
at

apv
at

1
2

ap
-.
at

(10.14)

538

HYDRODYNAMIC PROCESSES NEAR EQUILmRIUM

If we combine Eqs. (10.13) and (10.14), we find

= 8pe

T 8ps
8t

8t

_ v . 8pv
8t

(!v

M) 8p .

2 _

(10.15)

8t

We can now use the balance equations to obtain an expression for the
entropy current and source. If we combine Eqs. (10.3), (10.8), (10.9), and
(10.15), we obtain
Bps
T{it = -Vr

(JRe + JDe) + v[Vr (PU- + pvv + II)]


-

(12"v

M) Vr (pv).
(10.16)

This equation can be simplified further. Note that

v [Vr . (pvv)] = L

L
i

ViVj(PVjVi)

=L

+ V~Vj(pVj)].

L[pViVj(VjVi)
i

(10.17)

Converted with

Note also the

STOI Converter

(10.18)

trial version
If we combir

hDP://www.stdutilitv.com

~------------------------------

(10.19)
Let us now substitute Eq. (10.19) into Eq. (10.16) and make use of the GibbsDuhem relation, psVrT + PVrM - VrP = O. Then after some rearrangement
we find
Bps
Tat = -Vr'

(Je +Je

"

PM v - II v)

+ psv-

VrT - II: Vrv,

(10.20)

n:

where
Vrv == L:i Lj n, ViVj and M" = M+ v2 We can now write Eq.
(10.20) in the form of a balance equation. If we divide through by the
temperature, we can rearrange terms to obtain
8ps =
8t

-v. .

(J: + J~ - T

+ (J: + J~-

PM"V -

fi

PI""V - psTv -

v)
(10.21)

fi v) V'r (}) - ~fi: V'rv.

539

NAVIER-STOKES HYDRODYNAMIC EQUATIONS

We can simplify Eq. (10.21) still further if we first write its form for a
nondissipative fluid.
Consider a fluid in which no dissipative processes occur. By definition, the
stress tensor, fi, equals 0 and the dissipative energy current, J~, equals 0 for
such a fluid. Also there will be no entropy source term. Therefore, from
Eq. (10.21) the reactive energy current must be given by
(10.22)
where h is the enthalpy per unit mass. If we substitute Eq. (10.22) into
Eq.(1O.21) and set
= 0 and
= 0, we obtain the entropy balance equation
for a nondissipative fluid,

J?

Bps

8t = -V'r

. (psv).

(10.23)

The entropy current in a nondissipative fluid is psv.


If we substi
.. .,....
_.._
...
.
(10.20), we 0
dissipative fluic
Bps
-=

STOI Converter
trial version

8t

where the dissi

J:,

t,
into Eq.
quation for a

Converted with

(10.24)

hDP://www.stdutilitv.com
D

JS =-(J
T e - n v)

(10.25)

The entropy source in a dissipative fluid is


(10.26)

J? .

n:

One can easily check that the quantity


V'rT +
V'rv has units (energy/
volume time). It is the hydrodynamic version of Joule heating. In an electric
circuit, Joule heating is given by J . E, where E is the electric field and J is the
electric current. This is the (energy/volume- time) dissipated in the electric
circuit. We see that the hydrodynamic entropy production has the same form if
We identify
and ii as the generalized currents and V'rT and V'rv as the
generalized forces driving those currents.
In an electric circuit the electric current and the electric field (driving force)
are related by Ohm's law, J = O"E, where 0" is the conductivity of the electric
rnedium. Hydrodynamic media also have a version of Ohm's law in which the
driving forces are linearly related to the currents. In hydrodynamic systems the

J?

540

HYDRODYNAMIC PROCESSES NEAR EQUILIBRIUM

"conductivities" are called "transport coefficients." In the next section We


will obtain the transport coefficients for an isotropic fluid of point particles,
and in so doing we will be able to write the Navier-Stokes equations for the
fluid.

EXERCISE (10.1) (a) Write the dyatic tensor V'rv in orthogonal


curvilinear coordinates. Assume that the coordinates have unit vectors
and
coordinates Ul,U2, and U3, and scale factors h1,h2, and h3. (b) Write
the dyatic tensor V'r v for the special case of cylindrical coordinates
UI = r, U2 = <p, and U3 = z with scale factors hI = 1, h2 = r, and h3 = 1 and
unit vectors
=
=
and
=

el, e21

e3,

r, e2 J,

el

e3 z.

Answer
(a) In orthogonal curvilinear cordinates the gradient is

(1)

Converted with

STOI Converter

Let x, denote
. Unit vectors
a = 1,2,3.
trial version
rdinates with
is useful to
note
0:
ith respect to
the generalized coordinates UI, U2, and U3 are zero. That is,
axo:/8Ui = O. The unit vectors ei are related to the Cartesian unit
vectors through a transformation

hDP://www.stdutilitv.comit

ei = LRi,o:(Ul,

U2, U3)Xo:

Xo: = .L:R~,~(UI' U2, u3)ei.

and

(2)

0:

We can now make use of the above relations to write the dyatic tensor,
V'r v in the form

V'rV

.L:.L:eiej8vj + .L:.L:.L:.L:eiekVj
i

h, Bu,

(b) In cylindrical coordinates we have

0:

8Rj,o: R~~.
hi

Bu,

(3)

541

NAVIER-STOKES HYDRODYNAMIC EQUATIONS

Therefore, we find

10.B.3. Transport Coefficients


Transport coefficients are the generalized conductivities of a hydrodynamic
system. Before we can determine how many transport coefficients are needed
for our isotropic fluid of point particles, we must determine how many
independent driving forces and currents there are in the fluid. In the last section
we found that the generalized d~ving forces are '\J_rT and V'rv and the
.ne component
generalized cu
. For example,
tensors, and ea
Converted with
athematicswe can write
Tensors", belo

STOI Converter
trial version

(10.27)

hDP://www.stdutilitv.com

In Cartesian cOlonnnareS"";'we--na:ve----------------'

Similarly, we can write the stress tensor as

n = t [Tr(fi)]fJ + n- + n-.

(10.28)

For an isotropic fluid of point particles or particles which interact via a


spherically symmetric potential, the antisymmetric part of the stress tensor is
identically zero. The antisymmetric "force," [V'rv]a = V'r X v, is the curl of the
velocity. It is the contribution from vortex motion in the fluid. For a fluid of
particles which interact via spherically symmetric forces, there is no way to
couple rotational motion of the particles to vortex motion of the fluid.
Therefore, there is no way to dissipate energy from the fluid flow by
transfOrming angular momentum of the fluid into angular momentum of the
Particles. Thus, for a fluid of point particles,
== O. For a fluid of particles
Which interact via nonspherically symmetric forces, there will be coupling
between angular momentum and translational flow [4].

n-

542

HYDRODYNAMIC PROCESSES NEAR EQUILffiRIUM

Useful Mathematics -Tensors


An arbitrary tensor, T, can be decomposed into three orthogonal
components. In cartesian coordinates we can write the nine component
tensor in the form

(1)
The nine terms in Eq. (1) can be regrouped into three orthogonal quantities.
We may write T as
T

xx

= !Tr(T)fJ + TS + Ta,

(2)

zz

where fJ _=
+ yy + is the unit tensor and Tr(T) = Txx + Tyy+ T zz- In
Eq. (2), TS is a symmetric tensor with zero trace and is defined as

(3)
where TT is
the form

. t""",,,.,, ..

~J..""

m"n

gl, 0

Converted with

STDI Converter
where T~ =
Ta is an ant

be written in

trial version

hDP://www.stdutilitv.com

(4)
on. In Eq. (2),
as

(5)
The antisymmetric tensor may be written in the form
(6)

=! tr; -

where T~
0, T~
Tyx),_ and so on.
If we introduce a second tensor, V, then it also can be decomposed in the
manner described above. The scalar product of the two tensors can be
written
T : V = ![Tr(T)][Tr(V)]

+ fS : VS + fa : v-.

(7)

One can also show that


(8)
Thus, the tensor, T, can be decomposed into three independent orthogonal
components.

NAVIER-STOKES

HYDRODYNAMIC

543

EQUATIONS

In order to write "Ohm's law" for the fluid, we shall make use of Curie's
principle, which states that in an isotropic fluid, a given force cannot drive a
current of a different tensor character. Let us note that there are four forces in
the fluid, each with a different tensor character. They are \l rT, a polar vector;
[Vrv]S, a symmetric tensor; \lr' V, a scalar: and [\lrV]Q= \lr X v, an axial
vector. (Polar vectors and axial vectors behave differently under inversion
through the origin of coordinates. Polar vectors change their sign. Axial vectors
do not.) Since the antisymmetric part of the stress tensor is zero, the axial vector
cannot drive a current. Thus, an isotropic fluid of point particles has three
generalized forces. They =,\lrT, \lr_' v, and [\lrvt They drive the generalized currents,
II == Trffl) , and ns. Since they are all of different tensor
character, we find the following generalization of Ohm's law for the fluid:

J?,

(10.29)
where K is the coefficient of thermal conductivity,
(10.30)

Converted with
where 1]is the

STOI Converter

(10.31)

trial version

Wh;:~~ :s~~'''r---..---h_n-,-p-.-:_II_WWW----rr-_.S_t_d---,Ur-::rti----.-li_tv_._C_om_----,
the form
(10.32)
where l\lrTl2 = (\lrT) . (\lrT) and 1[\lrvrI2 = [\lrvr : [\lrvr. The mass
balance, momentum balance, and entropy balance equations take the form
(10.33)
apv

at + \lr

. (pvv) = -\lrP

+ 1]\lr' v + (( + t1])\lr(\lr'

v).

(10.34)

and

respectively. The derivation of the specific form of Eq. (10.34) is given in


Exercise (10.2). Equations (10.33)-(10.35) collectively are called the Navier-

544

HYDRODYNAMIC PROCESSES NEAR EQUILIBRIUM

Stokes equations or hydrodynamic equations. They describe the hydrodynamic


behavior of an isotropic fluid of point particles. In the next section we describe
how to solve the linearized version of these equations. The linearized
hydrodynamic equations describe the behavior of the fluid if it is only slightly
displaced from absolute equilibrium. They are also useful in describing the
behavior of fluctuations for a fluid in equilibrium .
EXERCISE 10.2. Prove that

\lr . fi
iffis = -21][\lrv]S
the unit tensor.

and Il

-1]\l;v - (( + tTJ)\lr(\lr . v)

= -((\lr'

Answer: The stress tensor is

v) with fi =

n = -21][\lrvr

rm + fis.

Note that U is

- (U(\lr . v), where


(1)

The ijth com

Converted with

STOU Converter
Then

(2)

trial ve rsion

hnp://www.stdutilitv.com
(3)

Thus

(4)

iuc,

LINEARIZED HYDRODYNAMIC EQUATIONS [5]

The hydrodynamic equations, (10.33)-(10.35), depend nonlinearly on the


thermodynamic variables and the average velocity. This makes them very
difficult to solve and one generally must resort to numerical methods for solving
them. However, if we restrict ourselves to the neighborhood of absolute
equilibrium, then we can look at the behavior of small deviations from the
equilibrium state, and it is possible to linearize the hydrodynamic equations.

LINEARIZED

to.c.t.

HYDRODYNAMIC

545

EQUATIONS

Linearization of the Hydrodynamic Equations

Let po, To, So, and Po, denote the equilibrium mass density, temperature, the
specific entropy, and pressure, respectively. Then close to the equilibrium state,
we can write p(r, t) = Po + Ap(r, t), T = To + AT(r, t), s(r, t) = So + As(r, t),
and P(r, t) = Po + AP(r, t), respectively. The average velocity, v(r, t), is zero
at equilibrium, so it is already first order in deviation from equilibrium. If we
now substitute these expansions into the hydrodynamic equations (10.33)(10.35), and only retain terms to first order in deviations from equilibrium, we
find

8Ap
---at+PoVrV=O,

(10.36)

(10.37)

and

Converted with

STOI Converter

(10.38)

trial version
~) to eliminate
To obtain the fi
two terms and '
The momentum balance equation, (10.37), is actually three equations since
the velocity has three components. Thus, we have five equations but we have
seven unknowns, Ap, As, AT, AP, v vy, and Vz. In order to solve these
equations, we can use thermodynamics to reduce the number of unknowns to
five since the quantities Ap, As, AT, and AP are related by thermodynamic
equations. We can choose two of them to be independent and expand the other
two in terms of them. The choice of the two independent variables depends on
convenience for the problem of interest.
Let us choose Ap and AT to be independent. Then we can write

hDP://www.stdutilitv.com
T

"

X,

&(r, r) = (;;): dp(r,t) + (;;):

stt. t)

(10.39)

8p b..p(r, t) + (8P)
8T AT(r, t).
(8P)

(10.40)

and

AP(r, t) =

546

HYDRODYNAMIC PROCESSES NEAR EQUILffiRIUM

The linearized hydrodynamic equations take the form


8flp

7it+

av

Po 8t = -

(8P)
8p

0
T Vrflp

(8P)
8T

POVr

(10.41 )

= 0,

+ 1]V;V + (( + 11])Vr(Vr,

0
p VrflT

v),
(10.42)

and

(10.43)
Equations (10.41)-(10.43) form a set of coupled equations which describe the
behavior of small deviations from absolute equilibrium.
It is useful
.
namic equations
__t_

for an ideal 1
(= o. Then E

._I_

_l_

_l_

,.

Converted with

F=

0,1] = 0, and

STOI Converter
(10.44)

trial version

hDP://www.stdutilitv.com

and Eq. (1O.4'i;;.,~;uu'~;---';


..;U--'V------------

av

Po at = -

(8P)
op

0
T Vrflp

(8P)
8T

0
p VrflT.

(10.45)

If we now take the time derivative of Eq. (10.45) and use Eqs. (10.41) and
(10.44), we obtain
(10.46)
where we have used the fact that (8P/8p)~ = (8P /8p)~ + (8P /8T)~(8T /8p)~.
Equation (10.46) is the wave equation for longitudinal velocity variations. It

vi

describes the behavior of sound waves with speed, c = (8P / 8p)~, in the ideal
fluid. Thus, an ideal fluid supports sound waves whose speed is determined by
the adiabatic compressibility of the fluid.
In a dissipative fluid, sound modes can be damped. We will now determine
the dispersion relations for normal modes of the dissipative fluid. As a first
step, let us Fourier transform the space dependence of the linearized

LINEARIZED

HYDRODYNAMIC

547

EQUATIONS

hydrodynamic equaions. We let dp(r, t) = (1/2nl


dkPk(t)e-ik.r so that
ik
r
(Jk(t) = drAp(r, t)e . . Similarly, v(r, t) = (1/271l dkvk(t)e-lk-r
and
T(r, t) = (1/271l dkTk(t)e-ik.r. If we substitute these expressions into Eq.
(10.41)-(10.43), we find

aPk .
at - lpok . Vk = 0,

(10.47)

and

Po(as)OaPk
ap T at
where Pk = Pk(t),
We can sim .
into longitudin

Tk

= Tk(t),

+ Po(as)OaTk
aT

and

Vk

at

= _ K k2Tk,

(10.49)

To

= Vk(t).
.

.de the velocity


) components:

STOI Converter

(10.50)

where k v~(t
trial version
kv~(t) is the
amplitude of a
gth A = 27r/k.
These are vel
is a
two-component vector which describes velocity variations with wavelength
A = 27r[k, but transverse (perpendicular) to the direction of k.
The transverse and longitudinal components of the average velocity decouple
in the linearized hydrodynamic equations. They evolve according to different
equations of motion. From Eqs. (10.47)--(10.49) and (10.50), we can write

hDP:llwww.stdutililV.COmtor.vi(t).

aPk _
at

ipokv" = 0

(10.51 )

k'

(10.52)

(10.53)
and

8v
Poar

kJ..

= -ryk

1.
Vk

(10.54)

548

HYDRODYNAMIC PROCESSES NEAR EQUILIBRIUM

Equations (10.51)-(10.53) govern the evolution of the coupled heat and sound
modes (sound consists of longitudinal oscillations of the fluid). Equation
(10.54) governs the evolution of shear modes (transverse velocity oscillations).
We want to solve the hydrodynamic equations, given some set of initial
conditions fJk(O) , Tk(O), and Vk(O). It is useful to introduce the Laplace
transform

/Jk(z) =

OO
dtfJk(t)e-zt
fo

and

f8+iOO
7rl
8-ioo
1

fJk(t) = -2 .

dZ/Jk(z)ezt,

(10.55)

where 8 is a positive number. The Laplace transform of the time derivative is


(10.56)
Similar Laplace transforms can be written for the other variables.
Let us now Laplace transform Eqs. (10.51)-(10.54). We find

Converted with

(10.57)

STOI Converter
trial version

hDP://www.stdutilitv.com

(10.58)

(10.59)
and
Po zV~ (z)

+ 1] /{l v~ (z)

Pov~ (0) .

(10.60)

Let us now simplify these equations. First we note the following


thermodynamic identities

8P)
( 8p

=
T

c5
T

and

8P) = poc5ap = Po Cp - cp
( 8T p
T
To ap ,

(10.61 )

where cp = To(8s/8T)p
and cp = To(8s/8T)p
are the specific heats at constant
density and pressure, respectively, Co = y'C8P/8p)s
is the speed of sound,
ap = -(1/ po)C8p/8T)p is the thermal expansivity, and T = cp/cpo Note the

LINEARIZED

HYDRODYNAMIC

549

EQUATIONS

Maxwell relation, (8s/8p)r = -(I/P5)(8P/8T)p.


the longitudinal kinetic viscosity, 1/1,the transverse
thermal diffusivity, X, which are defined
1/t

"7
=-,

and

Also, let us now introduce


kinetic viscosity, 1/" and the

K
X=-_-,

(10.62)

poCp

Po
respectively, Then Eqs. (10.57)-(10.60) take the form

zfJk(z) - ipokV~(z) = {Jk(0),

(z + 1/lk2)v~(Z) - ik
2

_ Z coap iJk(z)
po,

c5
,Po

+ cp (z + ,Xk2)Tk(Z)
To

(10.63)

iJk(z) - ik c5 p Tk(z) = v~(O),


,
2

= - coap {Jk(0) +
Po,

(10.64)

Tk(O),

(10.65)

1.0

and

Converted with
Solving the
doing some alg
longitudinal me

STOI Converter
trial version

(10.66)
ly a matter of
Duplefrom the

hDP://www.stdutilitv.com
to.e.2. Transverse Hydrodynamic Modes
Let us assume that the Fourier components of the transverse velocity at time
t = 0 are known to be vi-(O). Then from Eq. (10.66) we have
(10.67)
The Fourier components of the transverse velocity at time t are then
(10.68)
In order to obtain the final result in Eq. (10.68), we have changed the integration
along the line z = 8 + iy (with y varying from -00 to +00) to a contour
integration with the contour shown in Fig. 10.1. We see that any transverse
velocity variations in the fluid decay in time and cannot propagate. The shortwavelength disturbances decay faster than the long-wavelength disturbances.
This wavelength dependence of the decay time is the signature of a hydrodYnamic mode.

550

HYDRODYNAMIC PROCESSES NEAR EQUILIBRIUM

Re(z)

Fig. 10.1. The contour, C, used to integrate


Eq. (10.68).

Converted with

lO.C.3. Lo

STOI Converter

The time evo


(10.63)-(10.6

ined from Eqs.


from:

trial version

hDP://www.stdutilitv.com
(

m
-z~

Po7

5!
[z + ""Xk2]
To
I

~z
Tk(z)
(10.69)

1 0 0) (A{Jk(O))

~
-

~Po7

1 0
0 :..e.
To

v~(O).
ATk(O)

If we multiply by the inverse of the matrix on the left in Eq. (10.69), we obtain

(10.70)

where D(k, z) is the determinant of the 3 x 3 matrix on the left in Eq. (10.69),
and M/D(k, z) is its inverse. The determinant D(k, z), can be written
D(k, z) = ~

[Z3

+ r(v/ + 'YX)k2 + z(rxv/k4 + c~~) + c~Xk4].

(10.71)

551

LINEARIZED

HYDRODYNAMIC

EQUATIONS

The matrix,

M, which appears in Eq. (10.70) can be written


2

-(JoapClok IT
ikzapc
Z2 + ZVlk2 + k2 ~

6h

(10.72)
Before we use Eqs.(1O.70) and (10.72) to solve the equations of motion, it is
useful to look more closely at the determinant, Eq. (10.71).
To second order in k, the three roots of the equation D(k, z) = 0, are given by
Zl = -X~
and z = icok - ~k2[vl + x(T - 1)]. Therefore, for long-wavelength disturbances (small k), D(k, z) is approximately given by
D(k,

z) ~ cp (z + x~) (z
To

+ ikco + rk2)

(z - ikci,

+ r~),

(10.73)

Converted with
her order in k.
where r = ~[VI
) = 0 must be
This approxima
roots must be
used with care.
kept.
The dissipat
trial version
es. The time
evolution of the
ariations each
contain contribi
p:
One normal
mode is primarily a heat mode and has a decay rate given by Zl. The other two
normal modes are primarily damped sound modes. The sound modes can
propagate but eventually will be damped out by dissipative processes in the
fluid, both viscous and thermal.
The time evolution of the density, longitudinal velocity, and temperature
variations can be obtained by performing the following integration:

STOI Converter
hn II

Pk(t))
v~(t)
Tk(t)

Id I-I-tv
WWW.Sull.com

1
= -.

Je+ioo

2m

e-IOO

dze"

(Pk(Z))

v~(z)
Tk(Z)

(10.74)

In Exercise 10.3 we obtain the time dependence of Fourier components of the


density under special conditions .

EXERCISE 10.3. Compute

Pk(t)

assuming that at time

t = 0 we have

Pk(O) ::j:. 0, v~(O) = 0, and Tk(O) = O. Write the amplitude of the evolution to
lowest order in k.

552

HYDRODYNAMIC PROCESSES NEAR EQUILIBRIUM

Answer: From Section lO.e.3 we can write

(1)
The density at time t is
fJk(t)

1 J8+ioo

= -2 .
7rl

8-ioo

dziJk(z)ezt

(2)

If we retain terms in the amplitudes to zeroth order in k, Eq. (2) reduces to

Thus, the initial disturbance in the density is eventually damped out. The
long-wavelen th com onents are the last to o.

Converted with

re,n

DYN

TRANSPO

STOI Converter
trial version

SAND

Because matt
rgo fluctuations
about their
ns govern the
dynamics of e
ow- requency components of those
fluctuations, and they determine the behavior of dynamic equilibrium
correlation functions and dynamic structure factors. Onsager was able to use
this fact, along with the reversibility of dynamical laws on the microscopic
level, to show that certain relations exist between transport coefficients describing coupled transport processes. Hydrodynamic equations (equations which
describe the long-wavelength, low-frequency dynamics) can be obtained for a
large variety of systems, including dilute gases, liquids, solids, liquid crystals,
superfluids, and chemically reacting systems. For complicated systems,
transport processes are often coupled. For example, in a multicomponent
system, it is possible to have a temperature gradient drive a particle current and
a concentration gradient drive a heat current. Some complicated systems can
have as many as 10 or 20 (or more) transport coefficients to describe the decay
to equilibrium from the hydrodynamic regime. Onsagar's relations are of
immense importance because they enable us to link seemingly independent
transport processes and thereby reduce the number of experiments that must be
performed in order to measure all the transport coefficients.
Once an expression has been obtained for the dynamic correlation function
of an equillibrium system, we can use the Weiner-Khintchine theorem to derive
the dynamic structure function which gives the power spectrum of the

hDP://www.stdutilitv.com

DYNAMIC EQUILIBRIUM FLUCTUATIONS AND TRANSPORT PROCESSES

equilibrium fluctuations. The power spectrum


often measured in experiments.

io.o.t.

is the quantity

553

which is most

Onsager's Relations [6, 7]

Systems which are out of equilibrium generally return to the equilibrium state
through a variety of transport processes which mayor may not be coupled to
one another. Onsagar, without any reference to a particular physical system, was
able to derive relations between the transport coefficients for coupled transport
processes. These relations are a general consequence of the invariance under
time reversal of Hamiltonian dynamic. There are essentially two aspects of the
proof: (a) It uses the time-reversal
invariance of mechanical equations of
motion at the microscopic level (Newtons's equations) to determine properties
of time-dependent correlation functions, and (b) it assumes that fluctuations
about the equilibrium state decay, on the average, according to the same laws
that govern the decay of macroscopic deviations from equilibrium-that
is, the
hydrodynamic equations.

to.tu.t.

Mi(

The first step ir


variance of HaJ
functions of mc

Converted with

STDI Converter
trial version

The quantities

ime-reversal in
ent correlation
~y the relations

(10.75)

hDP:llwww.stdutililV.COmions

about the

equilibrium values of the state vari~bles Ai arid Aj (cf Section 7.C). Equation
(10.75) tells us that the correlation between a fluctuation O'.i at time t = 0 and a
fluctuation O'.j at time t = T is the same as that of a fluctuation O'.j at time t = 0
and a fluctuation O'.i at time t = T. The quantities a, and O'.j can correspond to
fluctuations in the same state variables at different points in space. Thus, Eq.
(10.75) can also be turned into an equation relating correlations between spaceand time-dependent fluctuations.
To establish Eq. (10.75), we note that the correlation matrix (exex(T)) can be
written

(<XCI(r

==

J J drzdrz'rzriJ(rz)P(rzlri,

r)

(10.76)

where P( exlex' , T) is the conditional probability that the fluctuation has value ex'
at time t = T, given that it had value ex at time t = 0 (we are now using the
notation of Section 5.B).
For a closed isolated system, f( ex) can be written

f( ex) =

_
(

(27rkB)

) 1/2
n

e-g:a.a./2k

(10.77)

554

HYDRODYNAMIC PROCESSES NEAR EQUll..IBRIUM

[cf. Eq. (7.24)], and the change in entropy which results from these fluctuations
is
(10.78)
It is useful to introduce

a generalized

force, 'X, which is defined as

(10.79)
and a generalized

current, " defined as


q

= da.

(10.80)

dt .

Then the time rate of change of the entropy due to fluctuations

is

d~S

-=-1'x,
dt
For a resistor I
is the applied
energy is dissi
We must m
value of a. the
relate the joint
the microscopi

(10.81)
.c current and 1.
~ rate at which

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.COming

~------------------------------

f(<l)P(<lI<X',r)

JJ

= n~(E)
X

JJ

Thus, for each


ystem. We can
ions a. and a.' to
way:

drfdpN

(CHll+dll)
(E_E+ll.E)

(10.82)

dq'Ndp/N P(p", qN IpIN,

,r).

(a.'-a.'+da.')

In Eq. (10.82) we have used the fact that p(~, qN) = nM(E)-l for a closed
isolated system (n~E(E) is the volume of the energy shell). The phase space
integrations are restricted to the energy shell and to trajectories with values of a.
and a.' appearing in the left-hand side ofEq. (10.82); P(~, qNlplN, qlN, r) is the
conditional probability that a system be in a state (p'N, q'N) at time t = r, given
that it was in the state (~, qN) at time t = O. Since classical systems are
completely deterministic, we must have
p(pN,qNlplN,q'N,r)

= 8(q'N _ qN _ ~qN(pN,qN,r))
x 8(p'N _ pN _ ~pN(~,qN,r)),

where ~qN and d~

are uniquely determined

from Hamilton's

(10.83)
equations.

555

DYNAMIC EQUILIBRIUM FLUCTUATIONS AND TRANSPORT PROCESSES

Because Hamilton's equations are causal and time-reversal invariant, reversal


of all momenta in the system will cause the system to retrace its steps. This
implies that
(10.84)
Since (X and (x' are even functions of momentum, we can combine Eqs. (10.82)(10.84) to obtain microscopic detailed balance
(10.85)
From Eq. (10.85), Eq (10.75) follows easily.

lO.D.l.2. Regression of Fluctuations


Equation (10.75) is important because we can use it to find relations between
various transport coefficients. Let us first introduce the conditional average
(at(t))o, which is the average value of (X at time t, given that the initial value of
at was oto. We
y..u....LL...L..u..L.L..--""-J...LL

....__----------~

Converted with

STOI Converter

(10.86)

for the conditi


trial version
Onsagar ass
y according to
the same line
the decay to
equilibrium of Svs:renrrs-l~rcn--are:macrosc:opjrcaJ~om~-errmnuIium. Thus, the
average fluctuation, ((X(t))!Xo' obeys an equation of the form

hDP://www.stdutilitv.com

(10.87)
Equation (10.87) has the solution
(10.88)
The time derivative in Eq. (10.87) must be used with caution. It is defined in the
following sense:

d((X(t))!Xo
dt
where

_ ((X(t + T))!Xo - ((X(t))!Xo


=

(10.89)

is a small time interval whose values are bounded by inequalities,

To T T.

(10.90)

556

HYDRODYNAMIC PROCESSES NEAR EQUILIBRIUM

To is the time between collisions, and T is the time it takes the fluctuation to
decay to equilibrium. The limitation in Eq. (10.90) rules out fluctuations which
are too small-that is, fluctuations which decay to equilibrium in a few collision
times. Similarly, Eq. (10.90) is not valid when the fluctuation has just been
created. It takes a few collision times for it to settle down to a hydrodynamic
decay.
Equation (10.75) imposes a condition on the matrix M. If we expand Eq.
(10.88) for short times,

((I(t))!Xo = (XQ

tM . (XQ

+ O(r2),

(10.91)

and substitute it into Eq. (10.75), we obtain


(10.92)
If we now use the fact that

M . (I = (IT . MT and use Eq. (7.27) for the variance

in the fluctuations, we obtain

Converted with
where T denote

STOI Converter
trial version

then Eq. (l 0.9~

(10.93)

(10.94 )

hDP://www.stdutilitv.com
-

= L-T

or

Lij

= Lji.

(10.95)

Equations (10.95) are called Onsagar's relations. If we make use of the


generalized force 1. [cf. Eq. (10.79)], the time rate of change of the fluctuation
can be written
(10.96)
Equation (10.96) has the same form as the linearized hydrodynamic equations
(cf. Section 10.C). Eq. (10.95) is so useful that Onsagar received a Nobel prize
for deriving it. The matrix L is a matrix of transport coefficients. Equation
(10.96) tells us that a force resulting from a fluctuation Cti can cause a flux of Ai'
and a force arising from a fluctuation Ctj can cause a flux of Ai. Eq. (10.95) tells
us that the transport coefficients for the two processes are the same. For
example, a particle concentration gradient can derive a heat current, and a
temperature gradient can drive a particle current. The transport coefficients for
the two processes are the same although the processes physically appear to be
very different.

DYNAMIC EQUILIBRIUM FLUCTUATIONS AND TRANSPORT PROCESSES

io.nz.

Weiner-Khintchine

557

Theorem [3, 8, 9]

The Wiener-Khintchine theorem enables us to obtain a relation between the


correlation matrix for time-dependent fluctuations and the spectral density
matrix of fluctuations for ergodic systems with stationary distribution functions.
We shall first derive some properties of time-dependent correlation matrices and
then derive the Wiener-Khintchine theorem.
Let us consider the time-dependent correlation matrix, ((I(r)(I(O)),
for a
system governed by a stationary distribution function (such as a system in
equilibrium). The correlation matrix has the property

where we have let t = -r and T denotes the transpose of the correlation matrix.
From the condition of microscopic reversibility, we know that
(~(r)~) = (~~(r)) and, therefore,
-

Ccn(r) = Ccn(r) .
Furthermore, f]

Converted with

STOI Converter
where g-l del
(10.88) and (l(

(10.98)

trial version

(10.99)

ps. From Eqs.

hDP://www.stdutilitv.com
(10.100)

since M is a self-adjount matrix (Irl indicates the absolute value of r).


We now will introduce the spectral density matrix of fluctuations and show
that it is the Fourier transform of the correlation matrix. Let us introduce a
slight modification of the state variable (I(t) as follows:

It I < T,
It I > T

(X(t; T) '" { ~(t)

(10.101 )

such that
lim (I(t; T) = (I(t).

T-oo

(10.102)

We next introduce the Fourier transform of (I( t; T):

(X(W; T) =

J~oodt(X(t; T)e

VN

'

fr

dt(X(t; TJeiw',

(10.103)

558

HYDRODYNAMIC PROCESSES NEAR EQUILIBRIUM

Since the fluctuations, tX(t), are real, we find


tX*

(w; T) = tX( -w; T)

(10.104)

(* denotes complex conjugate).

The spectral density matrix is defined as


-

Scxcx(w) = lim -tX*(w; T)tX(w; T)


T-ooT

(10.105)

(cf. Section 5.E.2). Combining Eqs. (10.102) and (10.104) we can write

Scxcx(w)
= Joo. drei" lim -1 Joo dttX(t; T)tX(t + 7; T).
T -00

-00

(10.106)

-00

If we now invoke the ergodic theorem (cf. Section 6.C), we can equate the time
average in Eq. (10.106) to the phase average of the fluctuations:

Converted with

STOI Converter

(10.107)

trial version

hDP://www.stdutilitv.com

Then Eqs. (10.

S.. (w)

= ['"

drtF'(a.a.(r))

= ['"

drei""C .. (r).

(10.108)

Thus, the spectral density matrix is the Fourier transform of the correlation
matrix. Equation (10.108) is called the Wiener-Khintchine
theorem.
We can now derive some general properties for the spectral density matrix.
From Eq. (10.97) we find that the spectral density matrix is Hermitian,
(10.109)
where we have let
real, we have

7~

S:.(w)
It is useful to divide

-7

= [,,'

to obtain the last term. Furthermore, since Ccxcx( 7) is

dre-iwTC .. (-r)

= S.. (-w),

(10.110)

8cxcx(w) into its real and imaginary parts, Rcxcx(w) and iCICI(w),

DYNAMIC EQUILIBRIUM FLUCTUATIONS AND TRANSPORT PROCESSES

559

respectively:
(10.111)
Then from Eqs. (10.109) and (10.110) we find that Rarx(w) is a real, symmetric
T
matrix and an even functin of w (Rrxrx (w) = Rrxrx (w) = ~ (-w) ), while Iarx (w)
is _ a re~l, antisymmetric
matrix and an odd function
of w (Ifrx (w) =
- Irxrx(w) = - Iarx ( -w) ). Furthermore, we find that Srxrx(w) = Srxrx (w ) . Thus,
Lrx (w) = -Irxrx ( w) = 0, and the spectral density matrix is a real, symmetric
matrix and an even function of w. We therefore can write the correlation matrix
in the form
Ct%t%(r)
= -21
~

Joo

. -

dwe-IWTSrxrx( -w)

-00

= -1 Joo dwRrxrx(w)cos(wr).
~

(10.112)

If a magnetic field is present or the system is rotating, Eq. (10.112) must be


generalized to take into account the behavior of the magnetic field or angular
velocity under time reversal [3]. This is done in Section SlO.A. In Exercise
in an isotropic
lOA, we com
Converted
with
fluid.

STDU Converter
EXERCI
a box of voh

trial version

1for

a fluid in

http://www.stdutiliJV.com

Cnn(p, r) = ~ dr{n(r

+ p, r)n(r,

0),

where N is the total number of particles. The dynamic


given by

structure

factor is

Compute both Cnn(p, r) and Snn(k,O)


for a fluid of point particles in
equilibrium. For such a fluid, the average, 0, is taken over the equilibrium
distribution of fluctuations. Assume that fluctuations in the average velocity
are independent of temperature and density fluctuations.
Answer:
(a) Let us first write the density variations as n(r, t) = no + An(r, t),
where no = (n(r, t)) = N IV, is the equilibrium density and An(r, t)
describes fluctuations about equilibrium such that (An(r, t)) = 0

560

HYDRODYNAMIC PROCESSES NEAR EQUILIBRIUM

If we substitute into the expression for the correlation


obtain

Cnn(p, r) = ~ dr(n(r

= no + ~
Let us introduce

An(r, t)

+ p, t)n(r, 0)),
(1)

dr(t.n(r

+ p, t)t.n(r,

the Fourier series expansion

= _!_ I>-ik.rnk(t)
V k

and of the correlation

function, we

with

0)).

of the density

nk(t) = Jdreik.r An(r, t),

(2)

function

(3)
To obtr:___....__--1-_.__.-__;,-.....J::2.--..L2.),,___~....J--~L......o..:..................,d
the Fourier

Converted with

expans
identit

ave used the

STDI Converter

(4)

trial version

hDP:llwww.stdutililV.comin

where
fluctuation
theory
ec Ion.
we
ow
a empera ure an
nsity fluctuations are statistically independent. Therefore, (Ank(O)ALk(O))
= O.
Also, we will assume that velocity fluctuations are statistically
independent of density fluctuations. Therefore, any contributions to
the density at time t, due to fluctuations in velocity or temperature at
time t = 0, will disappear when the average, (), is taken. As a result,
we can use the solution for the density, Pk (t) == mnk (t), obtained in
Exercise 10.3. to determine the time evolution of the density
correlation function. We can write

Cnn(k, t) = N8kO,
x

+ -2mN

[(1-~)

(nk(O)n-k(O))

(5)
e-xk'itl

+ ~e-rk'ltlcOS(Cokt)]

=!

where v
cp/cp, Co is the speed of sound, r
[VI + X('Y - 1)], V/ is
the longitudinal kinetic viscosity, and X is the thermal diffusivity. The
absolute value of the time appears in Eq. (5) because it is the
correlation function for a stationary process. We must remember that

561

LINEAR RESPONSE THEORY

Eq. (5) is an approximate expression for the density correlation


function. Some terms proportional to the wave vector have been
neglected. We shall leave it as a homework problem to determine the
size of these neglected terms.
(b) Let us now consider the spectral density function (also called the
dynamic structure factor),

If we substitute Eq. (1) into Eq. (6) and made use of Eqs. (3) and (5),
we find

Snn(k,O)

(1-~)

= 211N6(0)4.,o + m!N (nk(O)n_k(O))


[

----;;---20
where
from
indep
critica
(8.21)
terms

STOI Converter
trial version

022:~20

(7)

8(0). Away
O)n-k(O)), is
lity. Near the
- Tc) [cf. Eq.
uid has three

hDP://www.stdutilitv.com
(8)

where j'(Ir) is a Lorentzian centered at frequency, 0', with half-width


at one-half maximum given by ~. Therefore, the dynamic structure
factor, as a function of frequency, has three peaks. One is centered at
o = 0 with width Xk2 and is due to the thermal hydrodynamic mode.
This is called the Rayleigh peak. The other two are centered at
0= cok with width, rk2. They are called the Brillouin peaks.

lO.E LINEAR RESPONSE THEORY AND THE


FLUCTUATION-DISSIPATION THEOREM [3]
Fluctuations in an equilibrium system decay, on the average, according to the
same linear macroscopic laws (in a fluid system the hydrodynamic equations)
that describe the decay of the system from a nonequilibrium state to the
equilibrium state. If we can probe the equilibrium fluctuations, we have a means

562

HYDRODYNAMIC PROCESSES NEAR EQUILIBRIUM

of probing the transport processes in the system. The fluctuation-dissipation


theorem shows that it is possible to probe the equilibrium fluctuations by
applying a weak external field which couples to particles in the medium but yet
is too weak to affect the medium. The system will respond to the field and
absorb energy from the field in a manner which depends entirely on the
spectrum of the equilibrium fluctuations. According to the fluctuationdissipation theorem, the spectrum of the equilibrium fluctuations and the rate
of absorption of energy from the external field can be expressed in terms of a
response matrix which is related to the correlation matrix for equilibrium
fluctuations.
In this section we derive the fluctuation-dissipation theorem. We first
introduce linear response theory and use the assumption of causality to obtain a
relation between the real and imaginary parts of the dynamic susceptibility
matrix, which is the Fourier transform of the response matrix. We then obtain a
relation between the dynamic susceptibility matrix and the correlation matrix,
and we obtain an expression for power absorption in terms of the dynamic
susceptibility matrix. This gives us a relation between the fluctuations in an
equilibrium system and ener absorbed b that s stem when an external field
is applied.

Converted with

IO.E.I. The
Let us assume
and that these

STOU Converter
trial ve rsion

hnp://www.stdutilitv.com

to deviate fro
from equilibri
, , , ... ,
applied forces (linear response). Then we can write

(lZ(t))F = ['" dlK(t

- t') . F(I) = ['" K(r)

n,

ied to a system
, causing them
the deviations
linearly on the

F(t - r)dr.

(10.113)

The matrix K(t - I) is real and is called the response matrix. Since the
response must be causal (the response cannot precede the force which causes
it), K(t - t') must satisfy the causality condition,
K(t - t') = 0,

t - t' < 0,

(10.114)

Since Eq. (10.113) is linear in the force, its Fourier transform has a very simple
form. If we note that
(10.115)
and use similar expressions relating F(r) to F(w) and K(r)
(Ci(w))p = X(w) . F(w),

to X(w), we obtain
(10.116)

563

LINEAR RESPONSE THEORY

where

(10.117)
is the dynamic susceptibility. We have also used the definition for the delta
function, 8(t) = (1/27T)
dwe-iwt
Thus, a force of a given frequency can only excite a response of the same
frequency. This will not be true if the response function depends on the force
(nonlinear response).

L':,

to.E.2. Causality
Let us now study the restrictions imposed on the response matrix by causality.
We shall assume that the response matrix relaxes fast enough that the integral

[Xl K(t) dt <


is finite. Physi
response. It is (
complex plane

(10.118)

00

Converted with

STOI Converter

rise to a finite

bf X(w) into the

trial version

hDP:llwww.stdutililV.com

(10.119)

where z = w + ie ( positive). If X(w) is well-behaved (nonsingular), then X(z)


will be well-behaved since the extra factor e-et introduced simply makes the
integral fall off faster. The function X(z) is not singular (has no poles) for z in
the upper-half complex plane, but we can say nothing about its behavior in the
lower-half complex plane. Note also that X(z) --+ 0 as --+ 00.
The matrix X(w) is complex (X*(w) = X( -w)), but causality enables us to
obtain a relation between its real and imaginary parts. We do this by
introducting the following trick. Define a new matrix

f(z)

X(z)

(10.120)

z-u

and integrate it over a contour C' (cf. Fig. 10.2) so that C' encloses no poles of

f(z). (Note that u is real.) Then

f
c

Since x(z)

--+

0 as

--+ 00

f(z)dz = 0 =

fc,z-u

-X(z)

dz.

there will be no contribution

(10.121 )
from the semicircle

at

564

HYDRODYNAMIC PROCESSES NEAR EQUILIBRIUM

Im{z)
complex z - plane

z=u
Fig. 10.2. Integration contour used to obtain the Kramers-Kronig relations.

infinity. Thus,

Converted with

(1O.122)

STOI Converter
trial version

It is useful to i

hn~:llwww.stdutilitv.com

00

Equation
P

dw--=hm
-00

w-

dw--+

r-O

W -

-00

dw-U

u+r

W -

(10.123)

(10.122) then gives

Joo
-00

dw i(w) = -lim
w-

JO

r-O

i dX (u + r ei) = i7rX(u).

(10.124)

7r

or

i(u) = ~ P
7rl

Equation (10.125) is a consequence


part, i'(w), and the imaginary part,
Let us write

Joo
-00

i(w)dw
w- u

(10.125)

of causality and allows us to relate the real

i"(w), of the dynamic susceptibility matrix.

x(w) = X'(w) + iX"(w)


and make us of Eq. (10.125). We then obtain the following

(10.126)
relations between

LINEAR RESPONSE THEORY

565

x'(w) and X"(w):

Joo

i'(u)=-P 1

71"

i"(w)
--dw
W -

-00

(10.127)

and

i"(u) = -~ P
71"

Joo

i'(w) dw.
W -

-00

(10.128)

Equations (10.127) and (10.128) are called the Kramers-Kronig relations and
enable us to compute the real part of i (w) if we know the imaginary part and
vice versa. As we shall see, the imaginary part of i (w) can be obtained from
experiment.
The Kramers-Kronig relations give us a relation between the real and
imaginary parts of the dynamic susceptibility matrix. Therefore, it is only
necessary to find one or the other to determine the response, ((Z(t)). We can
show this expli .
.
. s allow us to

Converted with

rewrite the dyn

STOI Converter

(10.129)

trial version
where we have

hDP://www.stdutilitv.com

lim

17"--+0 W

1 .

=F

1_)

= P (_,

lTJ

i71"8(w' _ w).

(10.130)

The response ((Z(t)) then becomes

J dwe-iwt i(w) . F(w)

((Z(t)) = _!_
271"

1]

Joo

= l~

--2
0

(271")

dt'

-00

Joo

dw

-00

Joo

dw',

-00

e-iw(t-t')
_.

w -

i"(w').

F(t').

lTJ

(10.131)
In Eq. 10.131, we have used Eq. (10.129) and have Fourier transformed F(w). If
We now make a change of variables w" = w - w' and introduce the following
spectral representation of the Heaviside function,

OO

6 t-t'
(

= -lim
)

1]->0

-00

dw e-iw(t-t')
--

(271"i)

+ iTJ

'

(10.132)

566

HYDRODYNAMIC PROCESSES NEAR EQUlLmRIUM

we obtain

(OI(t)) = 2i

[,0

dt'K" (t - I) . F(I),

(10.133)

where K" (t - I) is the Fourier transform of K" (w). Thus, we have succeeded in
writing the response entirely in terms of the imaginary part of the dynamic
susceptibility matrix. The expression for the linear response given in Eq.
(10.133) is the one most commonly seen in the literature.
Everything we have done to this point is completely general. Let us now
obtain an explicit expression for the response for the case of a constant force
which acts for an infinite length of time and then is abruptly shut off. The force
we consider has the form

F(t) =

for t < 0
for t > O.

(10.134)

Converted with
F(w)

STOI Converter

(10.135)

trial version

hDP://www.stdutilitv.com
If we now use the following definitions for the Cauchy principal part

(1)
-

w
= Im----+O

(10.136)

w2 + 2

and the Dirac delta function


8(w)

lim w2

--+O

(10.137)

2'
+

we obtain
(10.138)
From Eqs. (10.113) and (10.116) the response can be written in the form
(ot(t))F

= -1

27r

Joo
-00

dwe-lWtx(w) . F(w).

(10.139)

LINEAR RESPONSE

THEORY

567

It is useful to write the response in a slightly different form. To do this we


must generalize Eq. (10.125). Let us introduce the matrices

eiztX(z)
f+(z, t)= --,
z-u

t>O

(10.140)

and

f_(z, t) =

e-iztx(z)
,
z-u

<0

(10.141)

With our choice of sign in the exponential factors, the functions


still go to zero as c ---+ 00. From Eq. (10.125) we obtain
e

Joo

iut- ()

1
'l u =-;-P
l7r

eiwtx(w)dw

w- u

-00

f(z, t) will

(10.142)

and

Converted with

STOI Converter
Then, for

<0

(10.143)

trial version

hDP://www.stdutilitv.com
... J
w
.11,.\'"'/

UIoV

l7r

(10.144)

-00

and, for t > 0, Eq. (10.142) yields the expression

X (0)

= -;-P

Joo

it:

dw

eiwtx(w)

(10.145)

-00

If we combine Eqs. (10.138), (10.139), and (10.145), the response for


is
((J(t))F = X(O) . F
and for

<0

(10.146)

>0
1
((J(t))F = -;-P
l7r

Joo
-00

dw

X(w) . F
W

cos (wt).

(10.147)

Thus, while the force is turned on, the response is constant. When it is turned
off, the response becomes time dependent. The variables AI, ... ,An begin to
decay back to their equilibrium values.

568

HYDRODYNAMIC PROCESSES NEAR EQUILIBRIUM

10.E.3. The Fluctuation-Dissipation

Theorem

The fluctuation-dissipation theorem is extremely important because it relates


the response matrix to the correlation matrix for equilibrium fluctuations. As a
result, the external field can be used as a probe of equilibrium fluctuations.
To derive the fluctuation-dissipation theorem, let us consider a system to
which a constant force is applied from t = -00 to t = 0 and switched off at
t = O. We first write the response, (IX(t))F' for times t2:0 in terms of the
conditional average (IX(t))~,
for

2: 0,

(10.148)

where /(IXo, F) is the distribution for fluctuations IXo at time t = 0 in the


presence of a constant external field, F. For times t > 0, the field is no longer
turned on and we can write

(IX(t))~ = e-M1 . IXo

for t2:0

(10.149)

btain

Converted with

[cf. Eq. (10.88

STOI Converter
(10.150)

trial version

hDP://www.stdutilitv.com

where we have used Eqs. (10.146) and (10.147). Thus, we find


e-M1 . X(O) = ~ P

foo

tt:

dwcos (wt) X(w) .

(10.151)

-00

If we remember that
(10.152)

[cf. Eq. (10.100)] we may combine Eqs. (10.150) and (10.151) to obtain

Cczcz(t)

= ~B P
tt:

foo

dwcos (wt) X(w) . X-I (0) . g-I

(10.153)

-00

for t > O. Thus, we have obtained a relation between the dynamic susceptibility
matrix, X(w), and the equilibrium correlation function for fluctuations. In
Exercise 10.5 we show that X(O) =
IT. Therefore, we finally obtain

r'

kBT

Cczcz(t) = -.- P
l7r

Joo
-00

X(w)
dw -cos (wt).
W

(10.154)

LINEAR RESPONSE

569

THEORY

Equation (10.154) is the famous jluctuation-dissipation


theorem. It gives a
relation between the linear response function and the correlation function for
equilibrium fluctuations.

i EXERCISE 10.5. Prove that X(O) = g-l IT, where g is the matrix whose
; matrix element is s = (82S18oi8oj)u.
i

: Answer: The external field, F, does work on the system and increases its
, internal energy by an amount

(1)

dU = F da.

: We can expand the differential of the entropy dS and use internal energy and
state variables a. as independent variables,

Converted with
But (8SI8U)~
I as
!

STOI Converter
trial version

(2)
~nrewrite dS

hDP://www.stdutilitv.com
(3)
" For a constant force, (a.) = X(O) . F is the expectation value of (a.) rather
than zero, and the entropy will have its maximum for a. = X(O) . F. Thus,
i from Eq. (3) we have
i

(4)
I

and the condition that entropy be maximum at a. = X(O) . F yields


8S)
(-

8a.

1
= -F
~=X(O).F

__
- g X(O) . F = 0

(5)

: or
i(O) = 2_ g-l.
T

(6)

570

HYDRODYNAMIC PROCESSES NEAR EQUILmRIUM

lO.E.4. Power Absorption


The work done by an external force F to change

IX

by an amount dIX is

,rW = -F dIX

(10.155)

(this is work done on the medium). The average rate at which work is done on
the medium is just the power P(t) absorbed by the medium:

/,rW)

P(t) = \ dt

F=

-F(t)

. (li(t))F = -F(t)

d Joo

. dt

-00

- - t') . F(t').
dt'K(t
(10.156)

If we write the right-hand side in terms of Fourier transforms X(w) and F(w), we
obtain

i(.2_)2Joo

P(t) =

27r

dw

-00

Joo

dw' w'F(w). X(w') F(w')e-i(w+w')t.

(10.157)

-00

We can now (
various types (

y absorbed for

Converted with

STOI Converter

10.E.4.1. De
Let us assume

ed. Then,

trial version

(10.158)

hDP://www.stdutilitv.com
Substituting into Eq. (10.157), we obtain

P(t) = i(__!__)
27r

JOO dw JOO dw' w'X(w') : FF e-(w+w')t.


-00

(10.159)

-00

(Note: F X(w) . F = X(w) : FF.) We can find the total energy absorbed by
integrating over all times:

abs = Joo

P(t)dt

-00

-(f-)
7r

Joo

dwwX"(w):

FF,

(10.160)

-00

where X"(w) is the imaginary part of the dynamic susceptibility matrix. Since
the total energy absorbed must be a real quantity, only the imaginary part of
X(w) contributes.

1O.E.4.2. Oscillating Force


Now let us consider a monochromatic oscillating force of the form
(10.161)

LINEAR RESPONSE

571

THEORY

Then

F(w) = 7rF(<5(w + wo) + <5(w- wo)).

(10.162)

From Eqs. (10.157) and (10.162), the power absorbed can be written

p(t) = - i[( -iwo)

+ 1 )X(wo) + (iwo) (ei2wot + 1 )X( -wo)] :

(e-i2wot

FF

(10.163)
As we can see, the instantaneous power absorption oscillates in time. We can
find the average power absorbed by taking the time average of Eq. (10.163) over
one period of oscillation:

(P(t))

~J~

= -;

dtP(t)

="4

[X(wo) - X( -wo)] : FF

= TX"(wo)

: FF,
(10.164)

where X"(wo) is the imaginary part of X(wo). Again we see that the average
power absorbe
.
.
se matrix. In
principle, the a
Converted with
refore
(wo)
can be measure
us to obtain

x"

:~o~~~
~:t~SIDU Converter n~~i~~k:~
trial version

fluctuations an

S.(w), of equili
system, we can

nsity matrix,
mal field to a

hDP://www.stdutilitv.com

: EXERCISE 10.6. A Brownian particle of mass m is attached to a


I harmonic spring with force constant, k, and is driven by an external force,
I F(t). The particle is constrained to move in one dimension. The Langevin
i equation is

d2x(t)

m-dt2

dx(t)

+ ,- dt + mw20x(t)

e(t) + F(t)

w5

:, where
= (kim), , is the friction constant, and e(t) is a Gaussian white
! noise with zero mean. Then equation of motion of the average position,
~ (X(t))F' in the presence of the external force, F(t), is
I
:

i
: (a) Compute

d2~~))F

+,

d(X~~))F

+mw6(X(t))F

F(t).

and plot the linear response function, K(t), where


(b) Compute the total energy absorbed for
the case of a driving force,
= Fob(t) [10].

(X(t))F =

to: K(t - t')F(t')dt'.F(t)

573

LINEAR RESPONSE THEORY

The response function is given by

Joo

K(t) = -2

7r

1
dwX(w) e-lwt = -2

-00

7r

Joo

dw

-00

-m

w2

e-iwt

rnu.?n

0 -

lfW

.
(9)

poles of X(w) occur at w = -i(r/2m)


which is in the lower half-plane. X(w) has no
poles in the upper half-plane. For t > 0 we can write Eq. (9) as a
contour integral, but we must close the contour in the lower half-plane
so that the integral will be convergent. There are two poles in the
lower half plane, so we get a nonzero result. For t < 0 we must close
the contour in the upper half-plane, but since there are no poles there,
we get zero. Therefore we find
Notice

JW5 -

that

the

(r2/4m2),

K(t) _ B(t)e-'Yt/2m sin [Jw5 - ,2/4m2 t]


m
- 24m2
'
where
for t

Converted with

STOI Converter
trial version

hDP://www.stdutilitv.com

(10)

and B(t) = 0
ow for the (1)
)=l,m=l)
/2m(wo = 2,
with a delta
lowly decays

K(t)

t
(b) The total energy absorbed by the harmonic oscillator for the case of
the delta function force can be obtained from Eq. (10.160) and Eq. (8)
above. We find
(8)

575

TRANSPORT PROPERTIES OF MIXTURES

L.f=1

Let us now return to the first law, dU = itQ - PdV +


jljdmj, which
keeps track of energy exchange. We can divide the heat transfer, itQ, into a
contribution from pure heat exchange and a contribution due to transport of
matter. This is easy to see for reversible processes where we can write
N

itQ

== TdS

= T[dS]{mj}

L sjdmj.

+T

(10.166)

j=1

where sJ' = (8S /8mj)T P {m.. } is a partial specific entropy. If we further note that
"

'F}

SjT = hj - {lj, where hj = (8H/8mj)r,P,{miij} is a partial specific enthalpy and


{lj is a partial specific Gibbs free energy (the chemical potential of species, j),
then Eq. (10.166) takes the form
N

= TdS = T[dShmj} + L (hj

itQ

- jlj)dmj.

(10.167)

j=1

The first term,


mechanical an
not include the
system. For a
ilQ = itQ' + ~
the environmen
terms of itQ',

Converted with

STOI Converter
trial version

system due to
nment. It does
to or out of the
we can write
exchange with
~hen written in

hDP://www.stdutilitv.com
N

dU = itQ' - PdV

L hjdmj.

(10.168)

j=1
We will now write Eq. (10.168) for densities. We again define s = SlY,
u = U IV, and Pj = mj/V, but we also introduce the heat exchange per unit
volume, itq' = ilQ' IV. The Eq. (10.168) can be written

(dU -

ift/ -

t hidPi) =

H
. ~wev~r,
IS

dV

J=1

N u + P = H/V and Lj=1


hjpj
Identically zero and we obtain

(-U -

= H/V.

t hiP!) .

(10.169)

J=1

Therefore, the term on the right

du

= itq' +

L:
hjdpj
j=1

== dq,

(10.170)

where dq is the total amount of heat transfered per unit volume. If we substitute
Eq. (10.170) into Eq. (10.165), we can write the following expression for the

576

HYDRODYNAMIC PROCESSES NEAR EQUlLmRIUM

entropy change in our open system


N

Tds = dq -

L {ljdpj.

(10.171)

j=1

We can now use Eq. (10.171) to compute the entropy production in open
systems.
We will neglect convection of fluid in our small volume, V. That is, we
assume that the baryocentric velocity of the _fluid, V = :2:~1 pjVj! P = 0, where
P = :2:;1 Pj is the total mass density. If Jj = PjVj is the mass flux (grams/
area second) of particles of type j, this gives the additional condition,
:2:;1 Jj = O. The time rate of change of thermodynamic quantities inside the
volume are related through Eq. (10.171). Since V = 0, we can write
T as _ aq _ ~
at - at
~

J=1

_. apj
J.lJ at .

(10.172)

Changes insidp,...i:lb.A....lI.L.a..lJ..J..D:>""--lL....o::UI!.Q.....<:b..1.OL....L:......t:L:>.11.U:>......<L<~o
............
~rv and sources
inside the volu
ance equations
for the entrop

STOI Converter

(10.173)

trial version
the mass densi

hDP://www.stdutilitv.com
apj
8t
= -

r7
V r'

J-

~
+ L.....t

- J':
lIj,a a'

(10.174)

a=1

and heat density,


(10.175)
In Eq. (10.173), (J" is the entropy source term and is the quantity we wish to find.
In Eq. (10.174), Jj is the mass flux of particles oftypej across the boundaries,
and Vj,a = ~lIj,a, where lIj,a is the stoichiometric coefficient for the jth particle
in the oth chemical reaction. The oth chemical reaction acts as a source (or
sink) of particles of type j, and the "source" term, J~ = (I!V)(dea!dt),
represents this contribution (ea is the degree of reaction for the oth chemical
reaction). In Eq. (10.175), Jq is the total heat flow across the boundaries. If we
substitute these equations into Eq. (10.173), we find

-V'r' J5 +

<7

= -~ v. . r, -

t~
[-Vr. + t
jj

J=l

a=1

iij,aJ~].

(10.176)

TRANSPORT PROPERTIES OF MIXTURES

S77

If we now rearrange terms we obtain

(10.177)

where Aa = :2:,7=1 Vj,ajlj is the affinity of the oth chemical reaction. Thus, the
entropy current is
(10.178)
and the entropy source is

Converted with

a=

STOI Converter

(10.179)

trial version

hDP://www.stdutilitv.com

The entropy so
and an affinity
or force (the two words are often used interchangeably). Gradients in the
temperature and chemical potentials of the various constituent particles act as
forces driving the heat current, Jq, and the particle mass current, ij. The
chemical affinity, Aa, is the force driving the oth chemical reaction.
It is sometimes useful to write Eq. (10.179) in a slightly different form. Let
usnotethatd(jlj/T) = [djlj]r/T + [(8/8T) (jlj/T)]P,{mi} . But [(8/8T) (jlj/T)]P,{mi}

= -(1/T2)(jlj-sjT)

= -hj/T2,

We finally obtain d(jjj/T)


tion Vr(jjj/T)
= [Vrjljlr/T
we obtain

wheresj= (8S/8mj)r,P,{mi;ii} = -(8jjj/8T)P,{mi}'


- (hT2)dT. This in turn leads to the rela- (hj/T2)VrT.
If we substitute into Eq. (10.179),

= [djljlr/T

(10.180)
",:here J~ = Jq - :2:j hjlj is the heat current minus the contribution due to
dIffusion. Notice that the entropy production is again written in terms of
products of affinities (forces) and currents. The affinity driving the heat current
is a temperature gradient. The affinity driving a mass current of particles of type

579

TRANSPORT PROPERTIES OF MIXTURES

i Therefore, the generalized

Ohm's law takes the form

: Now we must find explicit expressions


for the On sager coefficients,
Lll,LI2,~I,
and ~2.
From the rate of change of mole numbers, we can write the currents as
I

h = k2cB - k_2Cc,

and

h = k3CC

k-3CA,

(5)
! where c, = n;/V is the molar concentration
of molecules of type i. At
: equilibrium these currents are zero. If we denote the equilibrium molar
i densities as CA, CB, and cc, then we find the equilibrium conditions
I
i

Cc
CB
We can also

nvvJ.-."'----'u.'L~

..........
L'~

......
........

and

k-2'

(6)

Cc

ee energy of

~T.J...........jo............,~~~"""'-""-"~.L........__.r<L.!.o..Io,:L

.....
L""'__"'r.r.,

Converted with

the system by
O(P,

and C,

STOI Converter

(7)

trial version
where

0(0)

is

hnp:II~.stdutililV.com

The chemical potentials

~es of type i.

are partial Gibbs free energies and are given by

J-lA = -RT

TS/2)
In ( p

+ RT

In(xA),

J-lB = -RTln

TS/2)
(p

+ RTln(xB),

J-lc = -RTln

TS/2)
(p

+ RTln(xc).

and

(8)

~ The affinities are therefore

-RTln(::) -RTln(::),
= -RTIn(::}
-RTln(;;).

AI =

A2

A3 =

and

(9)

580

HYDRODYNAMIC PROCESSES NEAR EQUILIBRIUM

We now want to approximate the currents and affinities by their values close
to equilibrium. Let us expand the molar densities about their equilibrium
values, CA = CA + QA, CB = CB + QB, and Cc = Cc + Qc, where QA, QB, and
Qc are small deviations from equilibrium. Then we find

(10)
Thus, we finally obtain

J
I -

klcA

+ k3CC

RT

A
I

k3CC A
RT
2

(11)

and
(12)
We therefore find Lll
cyclic set of

(I/RT)(klcA

+ k3Cc),

LI2

= LzI =

Lz2 = (1/ R

(k3Cc!RT),
and
tisfied for this

Converted with

STOI Converter
trial version

lO.F.2. Fie

hDP://www.stdutilitv.com

In order to de
eful to consider
Brownian mo
wave
a very 1 u e mixture of a
solvent (which we will choose to be water) and a solute (which we will leave
variable). Let Pw (Ps) be the density of water (solute) and let p = Ps + Pw be the
total density. We will assume that the mass density of solute is much smaller
than the mass density of water, so that x = (ps / Pw) 1. Also, we shall assume
that the pressure, P, and temperature, T, are uniform throughout the fluid. From
Eq. (10.180) the entropy production, a, is given by

(10.181)
From the Gibbs-Duhem equation (2.62), for constant temperature and pressure,
we find Ps[VrJl,s]TP + Pw[VrJl,w]rp = O. Also,
+jw = 0 since the baryocentric velocity is assumed to be zero. Therefore, the entropy production can be
written

is

(10.182)
The

current

that

l? = ls - (Ps/Pw)jw

appears

in

Eq.

= js(l + Ps/Pw),

(10.182)
is the diffusion
current,
of solute relative to water. This is easy

581

TRANSPORT PROPERTIES OF MIXTURES

to see. Since is == Psvs is the current of solute relative to the walls and
is the current of water relative to the walls,
= is - (Psi Pw),
lw = Ps(vs - vw) is the current of solute relative to water. It is this current that
must be nonzero to have diffusion of solute relative to water. In terms of the
diffusion current, the entropy production takes the form

j?

r, == Pwvw

(10.183)
We can write a generalized Ohm's law for this process. Since for a binary
mixture, /-ls = /-ls(T, P,xs), where Xs is the mole fraction of solute, we have
(10.184)
If we treat the mixture as an ideal mixture, then jls = jl~ + (RT IMs) In(xs) and
(8fisI8xs)PT
= RTlxsMs.
Also note that Vrxs = (cwlc2)Vcs - (cslc2)Vcw,
where c, = n;/V is the concentration (moles/volume) of molecules of type i.
However, since Cs cw, to lowest order in cslcw, we can write the particle flux,
(moles/area

i?

Converted with

STOI Converter

(10.185)

J!

The equation
trial version
Lst IM;cs is
the diffusion c
nt, D, to the
Einstein diffusi
rvA67r1]a), then
we find is = csMs2INA67r1]a, where 1] is the coefficient of viscosity for water, a
is the radius of the solute molecule, and NA is Avogadro's number. We can
obtain a numerical estimate for these quantities. At T = 25C = 298.15 K, the
viscosity of water is 1] ~ 0.9 X 1O-3P sec (Pascal-seconds). For solute
molecules with a radius of 4;\, the Einstein diffusion coefficient is
D ~ 0.61 x 10-5 cm2/sec. We can compare this with actual measurements of
the diffusion coefficient, given in Table 10.1, for very dilute aqueous solutions.
The value given by the Einstein diffusion coefficient is of the correct order of
magnitude. It is also interesting to see how the diffusion coefficient depends on

hDP:llwww.stdutililV.com

TABLE 10.1. Diffusion Coefficients for Some Typical Substances in the


Limit of Infinitely Dilute Aqueous Solution (T = 25C) [14]
Solute

Formula

D (10-5 cm2 sec-I)

Acetone
Benzene
Ethanol
Methane
Ethylbenzene

C3H60
C6H6
C2H6O
CH4
CsHIO

1.28
1.02
1.24
1.49
0.8

583

TRANSPORT PROPERTIES OF MIXTURES

The diffusion coefficient is then D = (LssIMs)(8jjsI8csh,p.


The coefficient of
thermal conductivity is K = Lqq. In experiments it is found that the coefficient
Lsq is linearly proportional to the solute concentration. Therefore, a coefficient
of thermal diffusion, DT,
introduced and is defined so Lsq = MscsDT. The
solute particle flux, J~ = J~ IMs, then takes the form

is

(10.189)
The Soret coefficient, ST, is defined as ST = DT ID ~ Lsql(MscsD). The Soret
coefficient is a measure of the concentration gradient that is set up as the result
of a temperature gradient for a system in its steady state. The steady state is
defined so that J~ = 0 and V'rT =constant. Therefore, if we set J~ = 0, the
Soret coefficient can also be written ST = -V'rln(cs)/V'rT.
The effect of
thermal diffusion is generally small with DT ~ 10-2 D or DT ~ 10-3 D.
However, it has been used as a basis to separate isotopes (Kr84 from Kr86 and
HCl35 from HC137) in the gaseous phase [13].

10.F.4. Elect
Electrical cond

Converted with

STOI Converter

s [12]
usion of those

particles are cl
the process is
governed by h
them. Let us
imagine a fluid
trial version
articles (ions),
with charge zse,
and create an
electric field,
radient in the
concentration so u e parnc es. n er sue con I Ions ere WI e a transport of
solute particles from one region of the fluid to another. For simplicity we shall
assume the temperature, T, and pressure, P, are uniform thoughout the fluid.
However, we shall assume that there are gradients in the chemical potentials of
the ions and the water and an electric potential gradient in the fluid. The entropy
production, (7, is given by

hDP://www.stdutilitv.com

(10.190)
where jj~ is the electrochemical potential of the ions. From the Gibbs-Duheim
equation we know that for constant T and P, Ps V'rjj~ + Pw V'r[lw = O. Therefore,
Wecan rewrite the entropy production as
(10.191)
where J~ = Js - (Psi Pw)Jw is the ion current (relative to the water) induced by
the electrochemical potential gradient. For a very dilute solution, Ps Pw and
Js The generalized Ohm's law for this system can be written

J? ~

(10.192)

585

TRANSPORT PROPERTIES OF MIXTURES

diffusion coefficients for some commonly found ions in the limit of infinitely
dilute aqueous solution.
I EXERCISE 10.S. Consider a very dilute solution of salt, which we
denote A~ C~, which dissociates as
!:

where Va and Vc are the stoichiometric coefficients for the dissociation.


Compute the diffusion coefficient for the salt in terms of the diffusion
coefficients of the ions in the limit of an infinitely dilute solution. (Assume
uniform temperature and pressure throughout the system).
I:

Answer: Let us assume that there are no external electric fields present, so

I the mass current is driven only by gradients in the chemical potentials of the
; ions. The entropy production, a, is given by
I
I

Converted with
;

I
I
I

I where

J is the

I Duhem equat
: Pw"Vriiw = O.
II

STOI Converter

(1)

n the Gibbs~+ Pc "V ri),c+


~ as

trial version

hDP://www.stdutilitv.com

if

(2)

where
= i; - (p;/ Pw)Jw(i = a, c) is the diffusion mass current of ion i. In
, the limit of infinitely dilute solutions, we can neglect the effect contributions
: to the current of cations due to gradients in the chemical potential of anions,
I and vice versa. Therefore, in this limit the generalized Ohm's law for this
I system can be written in the simple form
I
I

(3)
\

~We can restrict ourselves to the case when the total electric current is zero so
that

(4)
I

where Ji(i

= a, c)

is the flux of ions of type i. From Eq. (4) we obtain


(5)

586

HYDRODYNAMIC PROCESSES NEAR EQUILIBRIUM

The condition for chemical equilibrium of the dissociation of salt into ions is
given by
(6)
If we combine Eqs. (5) and (6), we can write
(7)
and

(8)
The particle flux of salt molecules is given by

(9)

Converted with

STOI Converter

We can now
pte the particle
flux of sal
M;ca/ RT and
trial version
L;
= DcMc'c2
ions of type i.
Let us also
If
we make UStr--.:;~~~_D1A"_='V""._~-:n::n;J["Q'I:ID"i,;--n-~_AAnrgAA
.....
".,n,.
V-1"1r~::;Ss" AAr:d~re

hDP:llwww.stdutililV.coml/Mscs)\lrCs.
;)r-------,~l::>'Ts

"Tt"'Il'

Ds = (zc + IZal)DaDc
zcDc + IZaIDa
is the diffusion
molecules .

coefficient

for a very dilute aqueous

(10)
solution

of salt

.....SPECIAL TOPICS
.....S10.A. Onsager's Relations When a Magnetic Field is Present [3]

When a magrietic field, B, is present or when we consider a system in a rotating


container where Coriolis forces are present, the Onsager relations must be
generalized. When we reverse the time and velocities in the equations of
motion, we must also reverse the magnetic field, B, in order to keep the
equations of motion invariant, since the magnetic force on particles with
velocity v and charge q is F = qv x B. If we have a rotating container, we must

587

SPECIAL TOPICS: ONSAGER'S RELATIONS WHEN A MAGNETIC FIELD

reverse both the velocity of the particles and the angular velocity m of the
container to keep the Coriolis force F = 2mv x m from changing sign.
In the presence of a magnetic field and fluctuations /3 (the fluctuations /3
change sign under time reversal), the distribution function must satisfy the
relation

f(rz, /l, B) = f(rz, -/l, -B)


under time reversal [here rz = (a
easily understood if we write f(rz,

f(a.,P,B)

(10.197)

1, a2, ... , an) and /l = (/31, ... , /3m)]. This is


/l, B) in terms of microscopic states:

= O.c.;(E)

J ... J dr"dP".

(10.198)

GI-Gl+dGl
,-,+d,
E-E+AE

For both classical and quantum systems, energy is an even function of the
momentum and the magnetic field. Therefore, if we reverse time and the
magnetic field .
~ Eq. (10.197)
follows immedi
Converted with
When a rna]
can write the
distribution fun

STOI Converter

f(rz,/l,B)

trial version

),

(10.199)

hDP://www.stdutilitv.com
where C is the normalization
constant and the tensors iii and 0 can be
represented by m x n rectangular matrices. The tensors g, iii,0, and ii may now
depend on the magnetic field. In order to satisfy Eq. (10.197), they must have
the following behavior when the magnetic field is reversed:

g(B)

g( -B),

ii(B)

= ii( -B)

(10.200)

o(B)

= -o(

(10.201 )

and

IiI(B)
Note that when B
transpose of 0:

=0

-1iI(-B),

-B).

we have iii = ii = O. The matrix

iii is simply

the

(10.202)
The change in the entropy due to fluctuations in the presence of a magnetic field
lS

(10.203)

588

HYDRODYNAMIC PROCESSES NEAR EQUILIBRIUM

and we must introduce

two generalized

tl.S
1
x = -- aa~
= g.
~ + -fJ
2

forces:
12

m + -n fJ

= g.
~ + n-

fJ

(10.204)

and

atl.s

y = - afJ = h . fJ + iii .~.

(10.205)

[Note that in Eqs. (10.204) and (10.205) we have used Eq. (10.202)].
various time-independent
correlation functions take the form

The

h--l m
- )-1 ,
( ~~) = k Bg-n
(- h- -1 m
- )-1 - h- -1
( ~fJ) =-Bg-n
k (-n
,

(10.207)

(fJ~) =- k B (h--mg

(10.208)

and

--1

-)-1

-n

--1

mg,

(10.206)

Converted with

STOI Converter
In the present

to the form

trial version

hDP://www.stdutilitv.com
and the joint probability

density for fluctuations

f(~, fJ, B)P(~, fJl~', fJ', r, B) = f(~',

(10.209)

(10.210)

is generalized

-s, -B)P(~',

to the form

-fJ'I~, -fJ, r, -B).


(10.211)

From Eq. (10.210) it follows that the time-dependent


correlation functions
given by expressions analogous to Eq. (10.76) satisfy the following relations:

(~~(r);B)

= (~(r)~;

(~fJ( r); B)

= - (~(r)fJ;

-B),

(10.213)

(p~(r);B)

= -(fJ(r)~;

-B),

(10.214)

-B),

(10.212)

and

(fJp(r);B)
under time reversal.

(fJ(r)Jl; -B)

(10.215)

589

SPECIAL TOPICS: MICROSCOPIC LINEAR RESPONSE THEORY

The equations describing the regression of fluctuations can be written as in


Eq. (10.87) except that now the matrix fW1I depends on the magnetic field. Thus,

!!_ ( (ot(t))oJJo)
dt

(fJ(t)) oJJo

_ -M(B)
-

. ((ot(t))oJJo)
(fJ( t)) oJJo '

(10.216)

where

_ (MOtOt(B) MOtJJ(B))
MJJOt (B) MJJJJ (B)

MB -

(10.217)

or

!!_ ( (ot(t))oJJo)
dt

(fJ(t))oJJo

_ -L(B)
-

. ((X(t))oJJo)
(Y(t))oJJo
.

(10.218)

If we 'use arguments similar to those in Section 10.0, Onsager's relations take


the form

Converted with

STOI Converter
and

(10.219)
(10.220)
(10.221 )

trial version

hDP://www.stdutilitv.com

(10.222)

Equations (10.219)-(10.222) give the general form of Onsager's relations. If we


were considering a rotating system, then we would replace the magnetic field,
B, by the angular velocity, w .

... S10.B. Microscopic Linear Response Theory [15]

Kubo was the first to show that it is possible to derive the linear response matrix
directly from microscopic theory. Let us consider a system to which we apply
an external field, FAr, t), which couples to microscopic densities, tlj(r). The
total Hamiltonian of the system in the presence of the external field is

H(t) = Ho + M(t),

(10.223)

where Ho is the Hamiltonian of the system in the absence of the field and

tlH(t) = -

J dra(r) . F(r, t)

(10.224)

590

HYDRODYNAMIC PROCESSES NEAR EQUILIBRIUM

is the contribution to the Hamiltonian due to the external field. We assume that
the external field is turned on at time t = -00. The total density operator, p(t),
in the presence of the field satisfies the equation of motion

in 8~~t)

[H(t), p(t)],

(10.225)

Let us write the total density operator in the form

p(t) = Peq + Ap(t),


where Ap(t)

(10.226)

is the change in the state of the system due to the perturbation and

Peq = e-flHo jTr [e-flHo] is the equilibrium density operator. Since [Ho, Peq] = 0,
we find

8p(t)
inEit

"

"

"

= [AH(t), Peq] + [Ho, Ap(t)] + [AH(t), Ap(t)].

To obtain an
neglect the nc

(10.227)
ed field, we will

Converted with

STOI Converter

(10.228)

trial version

hDP://www.stdutililV.com

It is now poss

The solutic\.A___----r -----;.-------",~---~~~

and plug this expression into Eq. (10.228). Then the equation for Ap/(t)
it. 8Ap/(t)

rt

8t

_ [eiHot/TiM(t)e-iHot/Ti
-

p"

(10.229)

, eq

We can integrate Eq. (10.229). If we assume Ap/( -00)

is

0, we find
(10.230)

The equation for Ap( t) is then

Ap(t) = .;_Jt

In

dt'[e-iHo(t-t')/Ti M(t')eiHo(t-t')/Ti,

Peq].

(10.231)

-00

We can use Eqs. (10.226) and (10.231) to find the average value of any desired
quantity.

591

SPECIAL TOPICS: MICROSCOPIC LINEAR RESPONSE THEORY

Let us first compute the average value of the operator, ai(r), at time t. Let us
assume that Tr[ai(r)Peq] = 0, although this not always the case. Then we find
(ili(r, t))

= Tr [ili(r).:lp(t)

=-

i~

Lo J
dt'

f
t), ilj(r', t )]) eqFj(r', tf),

dt' ([ili(r,

(10.232)
where
(10.233)
and
(10.234)
To obtain Eq. (10.232), we have cyclically permuted operators under the trace
and have assumed that (ai(r))eq = O.If we compare Eqs. (10.133) and (10.232),
we can write th.-------~~----------------,

Converted with

STOI Converter

'),

(10.235)

trial version

where we use

hDP://www.stdutilitv.com

eq

(10.236)

Thus, from microscopic linear response theory we obtain the result that the
dissipative part of the response matrix is given by the equilibrium average of the
commutator of the relevant operators at different times and positions in space.
We can also obtain the fluctuation-dissipation theorem from microscopic
linear response theory. The position-dependent correlation function in a
translationally invariant stationary system can be written
(10.237)
Wherewe have assumed that (ai(r))eq
Ga"a,

(k, r)

= O. Write

the Fourier transform is

d(r - rf)e-,k.(r-r) (ili(r, r)ilj(r'))

=~

J J dr dr' e-ik.(r-r) (ili(r, r)ilj(r'))

=~

(ai( -k, r)aj(k))eq

=~

eq

eq

(a7(k, r)aAk))eq.

(10.238)

592

HYDRODYNAMIC PROCESSES NEAR EQUILIBRIUM

In Eq. (10.238), we have used the definition, aj(k) = dre+ik.r aj(r). From Eq.
(10.236), the momentum-dependent response matrix can be written
(10.239)
However,
("ajai" ( T )) eq
where ai( -ih(3)

= ("ai ( T

in.1;;(3)"aj ) eq = e -ifJh(8/8r)(" a, ( T )")aj

== efJHoaie-fJHo.

eqs

(10.240)

Therefore

fJ
Kifa, ,aj (k ,T ) -- ~
2h (1 - e -i h(8/8r)G

ai.a]

(k ,T.)

(10.241)

The Fourier transform of Eq. (10.241) takes the form


(10.242)

Converted with

STOI Converter

where Saj,aj(
(10.242) wer
dissipation t
see that the i

s (10.240) and
the ftuctuationonse theory. We
directly in terms

trial version

hDP://www.stdutilitv.com

of the equili
expressed dir'--k

Xaj,a' ( , w
J

ptibility can be

-----.-,~_~--~-~_-

=.

lim

"1-+0

dw'

2.1;;

7rn

(1 - e-fJhw')Saj,aj(k,w')
,
.
W

W -

'''l

(10.243)

where we have made use of the Kramers-Kronig relations .

.....S10.C. Light Scattering [16-19]


The use of light scattering as a probe of the structure of matter began a little
more than 100 years ago. The incident light polarizes a medium, and the
medium then reemits the light in a manner which depends on the dynamics of
the medium. Light scattering as a tool to study the internal dynamics of systems
has developed in three stages. From the late 1800s until 1928, light scattering
experiments involved measurement of the intensity of scattered light as a
function of angle with respect to the incident direction. However, in 1928,
Raman measured the spectrum of light inelastically scattered from gases, and he
found frequency shifts in the scattered light due to internal degrees of freedom
of the molecules. From then on, most light scattering experiments focused on

You might also like