Pottier 2010 NonEquilPhysics
Pottier 2010 NonEquilPhysics
Pottier 2010 NonEquilPhysics
Noëlle Pottier
Laboratoire Matière et Systèmes Complexes, CNRS, and
Université Paris Diderot – Paris 7
1
3
Great Clarendon Street, Oxford OX2 6DP
Oxford University Press is a department of the University of Oxford.
It furthers the University’s objective of excellence in research, scholarship,
and education by publishing worldwide in
Oxford New York
Auckland Cape Town Dar es Salaam Hong Kong Karachi
Kuala Lumpur Madrid Melbourne Mexico City Nairobi
New Delhi Shanghai Taipei Toronto
With offices in
Argentina Austria Brazil Chile Czech Republic France Greece
Guatemala Hungary Italy Japan Poland Portugal Singapore
South Korea Switzerland Thailand Turkey Ukraine Vietnam
Oxford is a registered trade mark of Oxford University Press
in the UK and in certain other countries
Published in the United States
by Oxford University Press Inc., New York
c Noëlle Pottier, 2010
⃝
The moral rights of the author have been asserted
Database right Oxford University Press (maker)
First Published 2010
All rights reserved. No part of this publication may be reproduced,
stored in a retrieval system, or transmitted, in any form or by any means,
without the prior permission in writing of Oxford University Press,
or as expressly permitted by law, or under terms agreed with the appropriate
reprographics rights organization. Enquiries concerning reproduction
outside the scope of the above should be sent to the Rights Department,
Oxford University Press, at the address above
You must not circulate this book in any other binding or cover
and you must impose the same condition on any acquirer
British Library Cataloguing in Publication Data
Data available
Library of Congress Cataloging in Publication Data
Data available
Printed in Great Britain
on acid-free paper by
CPI, Chippenham, Wiltshire
1 3 5 7 9 10 8 6 4 2
Preface
The subjects treated in this book belong to the extremely vast field of nonequilibrium
statistical physics. We encounter, in all domains of physics, a very large variety of
situations and phenomena involving systems which are not in thermodynamic equi-
librium, and this, at all scales. One of the difficulties of the statistical physics of such
systems lies in the fact that, contrary to equilibrium where we have a unified approach
allowing us to explain the macroscopic properties of matter in terms of the micro-
scopic interactions (J.W. Gibbs), out of equilibrium we have only a limited number
of results with a general scope. The approaches used to describe the passage from the
microscopic description to the macroscopic one in the case of nonequilibrium systems
are varied, and they may depend on the particular system under study. It is however
possible to classify these approaches into two large groups, using, in the first one,
kinetic equations (L. Boltzmann), and, in the second one, the linear response theory
(R. Kubo). In spite of their diversity, these methods have in common several essential
fundamental features on which they rely.
One of the aims of this book is to clarify some central ideas common to these
di↵erent approaches, taking various physical systems as examples. Due to the vastness
of the subject, only near-equilibrium situations, in which the irreversible processes
involved may be qualified as linear, are taken into consideration.
Although extremely diverse, out-of-equilibrium phenomena display very generally
the essential role played by the existence in the systems under study of well-separated
time scales. Most of these time scales, very short, are associated with the microscopic
degrees of freedom, while other ones, in small number and much longer, are macro-
scopic and characterize the slow variables. The book attempts in particular to stress,
for each of the approaches, the importance of the role played by the separation of time
scales.
A central property common, in the linear framework, to the di↵erent approaches,
is the fluctuation-dissipation theorem, which expresses the relation between the re-
sponse of a system in a slightly out-of-equilibrium situation and the equilibrium fluc-
tuations of the dynamical variables concerned. This result constitutes the cornerstone
common to the di↵erent methods of nonequilibrium statistical physics in the linear
range.
The elements prerequisite for the reading of this book are rather limited. It is however
necessary to master the basic notions of quantum mechanics and equilibrium statistical
physics. As for the mathematical techniques used in the book, they are standard.
Generally speaking, the notions necessary to the understanding of each chapter are
vi Preface
included in it or have been provided in the preceding chapters, and the calculations
are explained in detail. At the end of each chapter, a list of textbooks on the subject
(in alphabetic order by author) is provided. It is followed, where possible, by a list of
original papers (in chronological order).
Basic notions
In statistical physics, each macroscopic variable is a statistical average of the corre-
sponding microscopic quantities. The notions of mean and of fluctuations (and, more
generally, the useful definitions and results concerning random variables as well as ran-
dom processes) are drawn together in Chapter 1. The most important results in view
of the study of fluctuations are the central limit theorem which underpins the central
role played in physics by Gaussian laws, and the Wiener–Khintchine theorem which
relates the autocorrelation function and the spectral density of a stationary random
process.
Kinetic approaches
Chapters 5 to 9 are devoted to the description of transport phenomena by means of
irreversible kinetic equations, mainly the Boltzmann equation.
To begin with, in Chapter 5, we consider the ideal classical gas of molecules
undergoing binary collisions (historically the first system to have been studied by
means of a kinetic equation). The kinetic approach relies in this case on the molecular
chaos hypothesis, that is, on the hypothesis of the absence of correlations between the
velocities of two molecules about to collide. This assumption leads to the (irreversible)
Boltzmann equation for the distribution function.
In Chapter 6, we show how the Boltzmann equation allows us, through convenient
approximations, to determine the transport coefficients of the gas. In Chapter 7, we
derive the hydrodynamic equations from the Boltzmann equation. Chapter 8 is devoted
to the application of the Boltzmann equation to solid-state physics, where it is widely
used to determine transport coefficients in semiconductors and in metals (semiclassical
Bloch–Boltzmann theory of transport).
Very generally, the validity of the kinetic approaches relies on the existence in the
system under study of two well-separated time scales. In dilute gases, the shorter time
scale is the duration of the collision, while the longer one is the mean time interval
separating two successive collisions of a given molecule. In Chapter 9, this type of
approach is generalized to systems in which the interactions may be considered as
local and instantaneous. Under decorrelation assumptions analogous to the molecular
chaos hypothesis, the out-of-equilibrium evolution of such systems may be described
by irreversible kinetic equations, generically called master equations.
Brownian motion
Chapters 10 and 11 deal with Brownian motion, that is, with the erratic motion of a
particle immersed in a fluid of much lighter molecules.
Brownian motion is one of the paradigmatic problems of nonequilibrium statistical
physics. It is generally studied by means of the Langevin equation, which describes the
evolution of the Brownian particle velocity over time intervals intermediate between
a short time scale, namely the correlation time of the random force exerted on the
particle, and a long time scale, namely the relaxation time of the particle’s mean
velocity. We recover in this framework the Einstein relation between the mobility and
the di↵usion coefficient.
The theory of Brownian motion plays a role all the more important as the Brown-
ian particle may be not a true particle, but the representation of a collective property
of a macroscopic system.
We present in addition a microscopic model of the Brownian motion of a particle
coupled with a thermal bath made up of an infinite ensemble of harmonic oscillators
viii Preface
in thermal equilibrium. This model is widely used to describe the dissipative dynamics
of various classical or quantum systems (A.O. Caldeira and A.J. Leggett).
Transport coefficients
In Chapters 15 and 16, we show how the linear response theory allows us to ob-
tain microscopic expressions for the transport coefficients in terms of the equilibrium
correlation functions of the appropriate currents. These expressions constitute the
Green–Kubo formulas.
In Chapter 15, we establish the microscopic expression for the electrical conduc-
tivity tensor in terms of the correlation functions of the relevant components of the
electrical current. In the homogeneous case, we deduce from this expression the real
part of the conductivity of a non-interacting electron gas in terms of matrix elements of
one-particle currents (Kubo–Greenwood formula). The corrections with respect to the
semiclassical conductivity deduced from the Boltzmann equation are due to quantum
interference e↵ects. We also briefly discuss the Landauer’s approach of the conductance
of mesoscopic systems and the one-dimensional localization phenomenon.
Chapter 16 deals with ‘thermal’ transport coefficients, such as the thermal conduc-
tivity or the di↵usion coefficient, which cannot be directly calculated using the Kubo
theory, as well as with the way they can be determined through di↵erent experiments
(for instance, light scattering experiments).
To conclude this preface, I would like to emphasize that this book has not been the
work of a person alone, but that it is based on teaching prepared and executed in team.
Preface ix
Indeed, this book stems from the lecture notes of a graduate course given during several
years in the ‘DEA de Physique des Solides’ of the Paris/Orsay Universities. Most of
the chapter supplements originate from the guided work and the associated homework
problems. I thus wish to first express my thanks to Jean Klein, Sylvie Rousset, and
Frédérick Bernardot, with whom I closely collaborated for several years in the context
of this teaching. I am also especially grateful to Frédérick Bernardot for his careful
reading of the manuscript, as well as for his precise and judicious remarks.
My thanks also go to the numerous students and readers of the lecture notes,
who, by their remarks, contributed to the improvement of the manuscript.
Finally, it is a pleasure to thank Michèle Leduc, director of the series ‘Savoirs
actuels’ (EDP Sciences/CNRS Éditions), whose constant encouragement enabled the
successful production of the French version of this book, as well as Sönke Adlung, from
Oxford University Press, for his steady support.
This page intentionally left blank
Contents
Chapter 1
Random variables and random processes 1
1 Random variables, moments, and characteristic function 2
2 Multivariate distributions 4
3 Addition of random variables 6
4 Gaussian distributions 7
5 The central limit theorem 9
6 Random processes 12
7 Stationarity and ergodicity 14
8 Random processes in physics: the example of Brownian motion 16
9 Harmonic analysis of stationary random processes 17
10 The Wiener–Khintchine theorem 19
Appendix
1A An alternative derivation of the Wiener–Khintchine theorem 23
Bibliography 25
References 25
Chapter 2
Linear thermodynamics of irreversible processes 27
1 A few reminders of equilibrium thermodynamics 28
2 Description of irreversible processes: affinities and fluxes 29
3 The local equilibrium hypothesis 32
4 Affinities and fluxes in a continuous medium in local equilibrium 34
5 Linear response 37
6 A few simple examples of transport coefficients 38
7 Curie’s principle 42
8 The reciprocity relations 43
9 Justification of the reciprocity relations 45
10 The minimum entropy production theorem 48
Bibliography 50
References 50
xii Contents
Supplement 2A
Thermodynamic fluctuations 51
1 The fluctuations 51
2 Consequences of the maximum entropy principle 52
3 Probability of a fluctuation: the Einstein formula 53
4 Equilibrium fluctuations in a fluid of N molecules 54
Bibliography 58
References 58
Supplement 2B
Thermoelectric e↵ects 59
1 Introduction 59
2 The entropy source 60
3 Isothermal electrical conduction 61
4 Open-circuit thermal conduction 62
5 The Seebeck e↵ect 62
6 The Peltier e↵ect 63
7 The Thomson e↵ect 65
8 An illustration of the minimum entropy production theorem 66
Bibliography 67
Supplement 2C
Thermodi↵usion in a fluid mixture 68
1 Introduction 68
2 Di↵usive fluxes in a binary mixture 68
3 The entropy source 69
4 Linear relations between fluxes and affinities 70
5 The Soret and Dufour e↵ects 72
Bibliography 73
References 73
Chapter 3
Statistical description of out-of-equilibrium systems 75
1 The phase space distribution function 76
2 The density operator 80
3 Systems at equilibrium 83
4 Evolution of the macroscopic variables: classical case 84
5 Evolution of the macroscopic variables: quantum case 86
Bibliography 88
Chapter 4
Classical systems: reduced distribution functions 89
1 Systems of classical particles with pair interactions 90
2 The Liouville equation 91
3 Reduced distribution functions: the BBGKY hierarchy 93
4 The Vlasov equation 96
5 Gauge invariance 97
Contents xiii
Appendices
4A Pair interaction potentials 99
4B Hamilton’s equations for a charged particle 100
4C Gauge invariance of the Liouville equation 102
Bibliography 104
Chapter 5
The Boltzmann equation 105
1 Statistical description of dilute classical gases 106
2 Time and length scales 107
3 Notations and definitions 108
4 Evolution of the distribution function 109
5 Binary collisions 110
6 The Boltzmann equation 113
7 Irreversibility 116
8 The H-theorem 117
9 Equilibrium distributions 120
10 Global equilibrium 121
11 Local equilibrium 123
Bibliography 125
References 125
Supplement 5A
The Lorentz gas 126
1 Gas in the presence of fixed scattering centers 126
2 Time scales 126
3 Collisions with the fixed scatterers 127
4 Kinetic equation of the Lorentz gas 127
Bibliography 130
References 130
Supplement 5B
The irreversibility paradoxes 131
1 The paradoxes 131
2 The time-reversal paradox 131
3 The recurrence paradox 132
Bibliography 133
References 133
Chapter 6
Transport coefficients 135
1 The relaxation time approximation 136
2 Linearization with respect to the external perturbations 138
3 Kinetic coefficients of a Lorentz gas 138
4 Electrical conductivity 142
5 Di↵usion coefficient 144
Bibliography 147
References 147
xiv Contents
Supplement 6A
Landau damping 148
1 Weakly coupled plasma 148
2 The Vlasov equations for a collisionless plasma 148
3 Conductivity and electrical permittivity of a collisionless plasma 151
4 Longitudinal waves in a Maxwellian plasma 154
Bibliography 157
Chapter 7
From the Boltzmann equation to the hydrodynamic equations 159
1 The hydrodynamic regime 160
2 Local balance equations 161
3 The Chapman–Enskog expansion 165
4 The zeroth-order approximation 168
5 The first-order approximation 169
Appendices
7A A property of the collision integral 175
7B Newton’s law and viscosity coefficient 176
Bibliography 180
Chapter 8
The Bloch–Boltzmann theory of electronic transport 181
1 The Boltzmann equation for the electron gas 182
2 The Boltzmann equation’s collision integral 184
3 Detailed balance 187
4 The linearized Boltzmann equation 188
5 Electrical conductivity 189
6 Semiclassical transport in the presence of a magnetic field 192
7 Validity limits of the Bloch–Boltzmann theory 198
Bibliography 200
References 200
Supplement 8A
Collision processes 201
1 Introduction 201
2 Electron–impurity scattering 201
3 Electron–phonon scattering 207
Bibliography 211
References 211
Supplement 8B
Thermoelectric coefficients 212
1 Particle and heat fluxes 212
2 General expression for the kinetic coefficients 213
3 Thermal conductivity 213
4 The Seebeck and Peltier coefficients 215
Bibliography 217
Contents xv
Chapter 9
Master equations 219
1 Markov processes: the Chapman–Kolmogorov equation 220
2 Master equation for a Markovian random process 223
3 The Pauli master equation 226
4 The generalized master equation 228
5 From the generalized master equation to the Pauli master equation 229
6 Discussion 231
Bibliography 233
References 233
Chapter 10
Brownian motion: the Langevin model 235
1 The Langevin model 236
2 Response and relaxation 238
3 Equilibrium velocity fluctuations 243
4 Harmonic analysis of the Langevin model 247
5 Time scales 249
Bibliography 251
References 251
Supplement 10A
The generalized Langevin model 253
1 The generalized Langevin equation 253
2 Complex admittance 255
3 Harmonic analysis of the generalized Langevin model 255
4 An analytical model 257
Bibliography 259
References 259
Supplement 10B
Brownian motion in a bath of oscillators 260
1 The Caldeira–Leggett model 260
2 Dynamics of the Ohmic free particle 265
3 The quantum Langevin equation 267
Bibliography 269
References 269
Supplement 10C
The Nyquist theorem 270
1 Thermal noise in an electrical circuit 270
2 The Nyquist theorem 270
Bibliography 275
References 275
xvi Contents
Chapter 11
Brownian motion: the Fokker-Planck equation 277
1 Evolution of the velocity distribution function 278
2 The Kramers–Moyal expansion 279
3 The Fokker–Planck equation 282
4 Brownian motion and Markov processes 285
Bibliography 288
References 288
Supplement 11A
Random walk 290
1 The drunken walker 290
2 Di↵usion of a drunken walker on a lattice 291
3 The di↵usion equation 292
Bibliography 293
References 293
Supplement 11B
Brownian motion: Gaussian processes 294
1 Harmonic analysis of stationary Gaussian processes 294
2 Gaussian Markov stationary processes 295
3 Application to Brownian motion 297
Bibliography 300
References 300
Chapter 12
Linear responses and equilibrium correlations 301
1 Linear response functions 302
2 Generalized susceptibilities 303
3 The Kramers–Kronig relations 306
4 Dissipation 307
5 Non-uniform phenomena 308
6 Equilibrium correlation functions 310
7 Properties of the equilibrium autocorrelation functions 314
Appendix
12A An alternative derivation of the Kramers–Kronig relations 319
Bibliography 321
References 321
Supplement 12A
Linear response of a damped oscillator 322
1 General interest of the study 322
2 The undamped oscillator 322
3 Oscillator damped by viscous friction 323
4 Generalized susceptibility 324
5 The displacement response function 327
Bibliography 328
Contents xvii
Supplement 12B
Electronic polarization 329
1 Semiclassical model 329
2 Polarization response function 330
3 Generalized susceptibility 331
4 Comparison with the Lorentz model 331
Bibliography 334
Supplement 12C
Some examples of dynamical structure factors 335
1 The examples 335
2 Free atom 335
3 Atom in a harmonic potential 337
Bibliography 340
Chapter 13
General linear response theory 341
1 The object of linear response theory 342
2 First-order evolution of the density operator 342
3 The linear response function 345
4 Relation with the canonical correlation function 347
5 Generalized susceptibility 348
6 Spectral function 350
7 Relaxation 352
8 Symmetries of the response and correlation functions 357
9 Non-uniform phenomena 359
Appendices
13A Classical linear response 361
13B Static susceptibility of an isolated system and isothermal susceptibility 363
Bibliography 367
References 367
Supplement 13A
Dielectric relaxation 368
1 Dielectric permittivity and polarizability 368
2 Microscopic polarization mechanisms 371
3 The Debye theory of dielectric relaxation 371
4 A microscopic model of orientational polarization 374
Bibliography 378
References 378
Supplement 13B
Magnetic resonance 379
1 Formulation of the problem 379
2 Phenomenological theory 380
3 A microscopic model 383
Bibliography 388
xviii Contents
Chapter 14
The fluctuation-dissipation theorem 389
1 Dissipation 390
2 Equilibrium fluctuations 393
3 The fluctuation-dissipation theorem 395
4 Positivity of ! 00AA (!) 398
5 Static susceptibility 398
6 Sum rules 400
Bibliography 403
References 403
Supplement 14A
Dissipative dynamics of a harmonic oscillator 404
1 Oscillator coupled with a thermal bath 404
2 Dynamics of the uncoupled oscillator 404
3 Response functions and susceptibilities of the coupled oscillator 407
4 Analysis of xx (!) 409
5 Dynamics of the weakly coupled oscillator 415
Bibliography 417
References 417
Chapter 15
Quantum theory of electronic transport 419
1 The Kubo–Nakano formula 420
2 The Kubo–Greenwood formula 423
3 Conductivity of an electron gas in the presence of impurities 427
Bibliography 431
References 431
Supplement 15A
Conductivity of a weakly disordered metal 433
1 Introduction 433
2 The Kubo–Greenwood formula 433
3 Conductivity of a macroscopic system 436
4 Conductance of a mesoscopic system: Landauer’s approach 438
5 Addition of quantum resistances in series: localization 440
Bibliography 445
References 445
Chapter 16
Thermal transport coefficients 447
1 The indirect Kubo method 448
2 The source of entropy and the equivalent ‘Hamiltonian’ 452
Bibliography 457
References 457
Contents xix
Supplement 16A
Di↵usive light waves 458
1 Di↵usive light transport 458
2 Di↵usion coefficient of light intensity 459
3 Di↵usive wave spectroscopy 462
Bibliography 467
References 467
Supplement 16B
Light scattering by a fluid 468
1 Introduction 468
2 Linearized hydrodynamic equations 468
3 Transverse fluctuations 470
4 Longitudinal fluctuations 472
5 Dynamical structure factor 478
Bibliography 480
References 480
Index 481
This page intentionally left blank
Chapter 1
Random variables
and random processes
In statistical physics, we are led to consider the physical quantities characterizing the
macroscopic state of a system consisting of a large number of particles as statistical
averages of the corresponding microscopic quantities. The macroscopic variables, thus
defined as averages, are accompanied by fluctuations due to the thermal agitation
of the associated microscopic degrees of freedom. When the system under study is
out of equilibrium, the temporal evolutions of the means and of the fluctuations have
to be taken into account in the description and the modelization of the phenomena
concerning it. Random variables and random processes are thus essential tools of
nonequilibrium statistical physics.
Some fundamental notions on these topics are, for this reason, gathered in this first
chapter. To begin with, we introduce the probability distributions and the moments
of random variables in one or in several dimensions. We then study the distribution
of the sum of two or of several independent random variables, and the central limit
theorem concerning the distribution of the sum of N independent random variables in
the limit N ! 1. This theorem is of crucial importance in statistical physics, since it
is the basis of the prominent role played there by the Gaussian laws.
Then, about random processes, we introduce the notion of stationarity, and we
briefly discuss the ergodicity properties, namely, the equivalence between temporal
averages and statistical averages. We consider more specifically the stationary ran-
dom processes, and we present the outline of their harmonic analysis, a method well
suited to the study of processes governed by linear di↵erential equations. We derive in
particular the Wiener–Khintchine theorem relating the noise spectral density and the
autocorrelation function of a stationary random process.
2 Random variables and random processes
1.2. Moments
The average or expectation value hf (X)i of any function f (X) defined over the con-
sidered state space is:
Z 1
⌦ ↵
f (X) = f (x)p(x) dx, (1.1.5)
1
1
We can equally say stochastic variable, or also variate: these three expressions are synonyms.
2
A realization, or possible value, of X is denoted here by x. The upper case letter thus denotes
the random variable, and the lower case letter one of its realizations. When no confusion will be
possible between these two notions, we will employ a unique notation.
3
For more clarity, we will sometimes denote it as pX (x).
Random variables, moments, and characteristic function 3
provided that the integral on the right-hand side of equation (1.1.5) does exist.
In particular, µm = hX m i is the moment of order m of X. The first moment
µ1 = hXi is the mean value of X. When its mean value hXi vanishes, the random
variable X is said to be centered. The quantity:
⌦ 2↵
2
= (X hXi) = µ2 µ21 (1.1.6)
Z 1
⌦ ↵
G(k) = eikX = eikx p(x) dx. (1.1.8)
1
The characteristic function is also the moment generating function, in the sense
that the coefficients of its Taylor expansion in powers of k are the moments µm :
X1 m
(ik)
G(k) = µm . (1.1.11)
m=0
m!
Relation between G(k) and the connected moments. @27.01.2018
4 Random variables and random processes
The derivatives of G(k) at k = 0 thus exist up to the same order as the moments.
However, the generating function G(k) exists even when the moments µm are not de-
fined. For instance, the Cauchy law (1.1.7) has no finite moments, but its characteristic
function is: Z
a 1 eikx
G(k) = 2 dx = e a|k|+ikx0 . (1.1.12)
⇡ 1 (x x0 ) + a2
The function e a|k|+ikx0 is not di↵erentiable at k = 0, in accordance with the fact
that the moments of the Cauchy law do not exist.
2. Multivariate distributions
2.1. Joint densities, marginal densities, and conditional densities
When several random variables come into play, which is for instance the case when
we consider a multidimensional random variable, it is necessary to introduce several
types of probability distributions.
• Joint probability density
Let X be an n-dimensional random variable, of components X1 , . . . , Xn . Its proba-
bility density pn (x1 , . . . , xn ) is called the joint probability density of the n variables
X1 , . . . , Xn .
• Marginal probability density
Let us consider a subset of s < n relevant variables X1 , . . . , Xs . The probability density
of these s variables, independently of the values taken by the irrelevant variables
Xs+1 , . . . , Xn , is obtained by integrating over these latter variables:
Z
ps (x1 , . . . , xs ) = pn (x1 , . . . , xs , xs+1 , . . . , xn ) dxs+1 . . . dxn . (1.2.1)
The density (1.2.1) is called the marginal probability density of the s relevant variables.
• Conditional probability density
We can attribute fixed values xs+1 , . . . , xn to the n s variables Xs+1 , . . . , Xn , and
consider the joint probability distribution of the s variables X1 , . . . , Xs . This latter
distribution is called the conditional probability density of X1 , . . . , Xs . It is denoted
by ps|n s (x1 , . . . , xs |xs+1 , . . . , xn ).
The diagonal elements of the covariance matrix are the previously defined variances,
and are thus positive, while the o↵-diagonal elements, called the covariances, are of
any sign.
We can show, using the Schwarz inequality, that:
2
|⌫ij | 2 2
i j, (1.2.11)
6 Random variables and random processes
bounded by 1 and +1, is called the correlation coefficient of the variables Xi and Xj .
Two random variables are said to be non-correlated when their covariance vanishes
(no hypothesis is made concerning the higher-order moments). Non-correlation is a
weaker property than statistical independence.
We deduce from equation (1.3.1) the expression for the density pY (y):
ZZ Z
pY (y) = (x1 + x2 y)pX (x1 , x2 ) dx1 dx2 = pX (x1 , y x1 ) dx1 . (1.3.2)
Z
pY (y) = pX1 (x1 )pX2 (y x1 ) dx1 . (1.3.3)
The probability density of the sum of two independent random variables is thus the
convolution product of their individual probability densities.
Gaussian distributions 7
We can equally derive this result by remarking that, if the variables X1 and X2
are independent, the characteristic function of Y factorizes:
⌦ ↵ ⌦ ↵⌦ ↵
GY (k) = eik(X1 +X2 ) = eikX1 eikX2 = GX1 (k) GX2 (k). (1.3.4)
The result (1.3.3) follows from formula (1.3.4) by inverse Fourier transformation.
4. Gaussian distributions
4.1. The one-variate Gaussian distribution
The most general form of the one-variate Gaussian distribution is:
1 2
p(x) = Ce 2 Ax Bx
. (1.4.1)
The Gaussian distribution is also called the normal distribution. The parameter A is
a positive constant determining the width of the Gaussian. The parameter B fixes the
position of its maximum. The normalization constant C is expressed in terms of A
and B as:
⇣ A ⌘1/2 2
C= e B /2A . (1.4.2)
2⇡
It is often preferable in practice to express the parameters A and B in terms of the
mean µ1 = B/A and the variance 2 = 1/A. Accordingly, we write the Gaussian
distribution as:
2
1 (x µ1 )
p(x) = p exp · (1.4.3)
2⇡ 2 2
All the moments of the Gaussian law are finite, in accordance with the fact that
1 2 2
the function eiµ1 k 2 k is infinitely di↵erentiable at k = 0. They can be expressed
in terms of the two first moments µ1 and µ2 , or in terms of the mean µ1 and the
variance 2 .
8 Random variables and random processes
where the matrix A of elements Aij is a positive definite symmetric matrix of dimen-
sions n ⇥ n. In vectorial notations, we write:
✓ ◆
1
pn (x) = C exp x.A.x B.x , (1.4.6)
2
⌦ ↵
⌫ij = (Xi hXi i)(Xj hXj i) = Mij . (1.4.10)
The covariance matrix of the Gaussian distribution (1.4.6) is M = A 1
.
A multivariate Gaussian distribution is thus fully determined by the means and
the covariance matrix of the variables. If the variables are uncorrelated, the matrices
A and M = A 1 are diagonal, and so the variables are independent. Thus, although
in general non-correlation is a weaker property than statistical independence, they
constitute, in the Gaussian case, equivalent properties.
The central limit theorem 9
of determinant:
1
Det A = · (1.4.13)
1 2 (1 ⇢212 )
2 2
The two-variate Gaussian distribution (of centered variates) may thus be written as:
✓ 2 ◆
1 1 x1 2⇢12 x1 x2 x22
p2 (x1 , x2 ) = exp + 2 · (1.4.14)
2⇡ 1 2 (1 ⇢212 )
1/2 2(1 ⇢212 ) 2
1 1 2 2
Formula (1.4.14) displays the fact that, when two Gaussian random variables are
uncorrelated (⇢12 = 0), their joint probability distribution factorizes into the product
of their individual densities. We can verify in this particular case that uncorrelated
Gaussian random variables are statistically independent.
where the summation extends to all possible subdivisions in pairs of the indexes
i, j, k, l, . . ., while the odd-order moments all vanish.
X1 + · · · + XN
YN = (1.5.1)
N 1/2
4
For convenience, the variables Xi are taken here as centered, the generalization to non-centered
variables being straightforward.
10 Random variables and random processes
PN
is, too, a Gaussian variable of zero mean and variance hYN2 i = N 1 i=1 hXi2 i = 2 .
The probability distribution of YN being the same as that of the original variables, the
Gaussian law is said to be stable with respect to the addition of random variables.5
We therefore have: ✓ ◆
k
log GYN (k) = N log GX · (1.5.4)
N 1/2
We can write, expanding7 in powers of k/N 1/2 ,
✓ ◆ ◆ ✓
k 1 2k
2
k3
log GX = +O , (1.5.5)
N 1/2 2 N N 3/2
As N ! 1, the term O(k 3 /N 1/2 ) approaches zero because of the factor N 1/2 in the
2 2
denominator. Therefore, GYN (k) tends towards e k /2
, which is the characteristic
function of the Gaussian distribution of zero mean and variance 2 .
5.3. Discussion
In this discussion, we will continue to limit ourselves to the case of identically dis-
tributed variables Xi . Even in this case, the central limit theorem is actually valid
under more general hypotheses than the ones we made (namely, of independent indi-
vidual random variables having a distribution of finite variance).
First of all, the condition of statistical independence of the individual random
variables, if it is a sufficient condition, is not a necessary one. Indeed, for the theorem
to apply, these variables must not exhibit long-range correlations (that is, correlations
between variables Xi and Xj with |i j| 1).9 Short-range correlations may however
be present without a↵ecting the result. It is clearly more difficult to treat the case of
strongly correlated individual random variables.
Secondly, the condition of finite variance of the distribution pX (x) is also a suffi-
cient, but not necessary, condition of convergence towards the normal law. However,
the distribution of the individual random variables which are summed must not be
too ‘broad’.10 We identify precisely the functions pX (x) which belong to the attraction
domain of the normal law by the criterion:
R
|x|>y
pX (x) dx
lim y 2 R = 0. (1.5.7)
y!1
|x|<y
x2 pX (x) dx
3
For instance, a distribution which decreases as |x| for |x| ! 1 belongs to the
attraction domain of the normal law, in spite of the fact that its variance is infinite.
3
All distributions decreasing faster than |x| for |x| ! 1 also belong to the domain
of attraction of the normal law, which is thus extremely vast.
This is the reason why the Gaussian law is omnipresent in physical situations,
the exceptions to this law (qualified as ‘anomalous behaviors’) being comparatively
much rarer. The anomalous behaviors correspond to the fact that other laws than
the Gaussian one possess the stability property with respect to the summation of
individual random variables.11 The stable laws were studied and classified by P. Lévy
and A. Khintchine in 1936. They, too, possess attraction domains, which may be
characterized by proper generalizations of the central limit theorem, and to which
belong the individual so-called broad laws.
9
This is quite apparent if we think for instance of the case where the N variables are identical.
10
An example of broad law is provided by the Cauchy law (1.1.7), whose moments are all infinite.
The sum of N independent variables distributed according a Cauchy law, is, too, distributed according
a Cauchy law, which we can check by using the corresponding characteristic functions. However large
N may be, there is no tendency towards the normal law.
11
The sum of the individual random variables must, for each stable law, be multiplied by an
appropriate scaling factor cN . For instance, in the case of the Cauchy law, we have cN = N 1 .
12 Random variables and random processes
6. Random processes
When a set of random variables X1 , X2 , . . . is not countable, we cannot label the
di↵erent variables using a discrete index. We therefore introduce, for this purpose, a
continuous parameter t. The quantity X(t) is then a random function of t. When t is
the time, which we assume here to be the case, X(t) is a random process (or stochastic
process).
Taking at each instant, for the random variable X, one of its possible realizations
x, we obtain a realization x(t) of the process X(t). Such a realization is also called a
sample. A sample does not depend on time in a deterministic way.
To define the average, we can also consider the set of all realizations or samples
{x(r) (t)} of the process X(t). The average of X at time t may be obtained by averaging
over this ensemble of realizations:
N
⌦ ↵ 1 X (r)
X(t) = lim x (t). (1.6.3)
N !1 N
r=1
Equations (1.6.2) and (1.6.3) are equivalent definitions of the ensemble average
or statistical average of X(t).
which specifies that the result of the integration over x2 must not depend on t2 . The
two-time density allows us to calculate two-time averages such as:
Z
⌦ ↵
X(t1 )X(t2 ) = x1 x2 p2 (x1 , t1 ; x2 , t2 ) dx1 dx2 . (1.6.5)
A quantity which can be calculated with the help of one- and two-time densities is the
autocorrelation function (t1 , t2 ) of the process X(t), as defined by:
⌦⇥ ⌦ ↵⇤⇥ ⌦ ↵⇤↵ ⌦ ↵ ⌦ ↵⌦ ↵
(t1 , t2 ) = X(t1 ) X(t1 ) X(t2 ) X(t2 ) = X(t1 )X(t2 ) X(t1 ) X(t2 ) .
(1.6.6)
The autocorrelation function allows us in particular to estimate the time range of the
correlations.
In the same way, in order to calculate quantities such as hX(t1 ) . . . X(tn )i, we
have to know the n-time density pn (x1 , t1 ; . . . ; xn , tn ). For any integer s < n, we have
the consistency condition:
Z
pn (x1 , t1 ; . . . ; xs , ts ; xs+1 , ts+1 ; . . . ; xn , tn ) dxs+1 . . . dxn = ps (x1 , t1 ; . . . ; xs , ts ).
(1.6.7)
The diagonal elements of the correlation matrix represent autocorrelations, and the
o↵-diagonal ones crossed correlations.
12
However, the processes whose present is not at all, or only weakly, influenced by their past his-
tory, may be characterized more simply. These processes are called Markov processes (see Chapter 9).
14 Random variables and random processes
where the summation extends to all possible subdivisions in pairs of the indexes
i, j, k, l, . . ., while the averages involving an odd number of times vanish.
Gaussian processes, which are especially simple to treat, are frequently used in
physics for an approximate description of random phenomena.
The average of a centered stochastic process X(t) does not depend on time. It is
often convenient to deduce from X(t) this constant average hXi, and to work with the
centered process X(t) hXi.
The autocorrelation function (t1 , t2 ) of a stationary stochastic process only de-
pends on ⌧ = t1 t2 . In the real case, the autocorrelation function (⌧ ) is an even
Stationarity and ergodicity 15
7.2. Ergodicity
Let us consider a stationary random process X(t), possibly complex, whose di↵erent
realizations are denoted by x(r) (t). It frequently happens in practice that any given
realization of the process contains all the statistical information available about it.
The stationary random process X(t) is then said to be ergodic.
To make this notion precise, we shall first define the temporal average of the
stationary process X(t). For this purpose, we consider a particular realization x(r) (t),
simply denoted by x(t), and a finite time interval (t T2 , t + T2 ), of width T . The
temporal average of the process over this time interval is, by definition:
Z T
T 1 t+ 2
X(t) = x(t0 ) dt0 . (1.7.5)
T t T2
It depends both on t and on the interval width T , as well as on the particular considered
T
realization. In the limit in which T tends towards infinity, X(t) becomes the temporal
average of the process, denoted by X:
Z T
T 1 t+ 2
X = lim X(t) = lim x(t0 ) dt0 . (1.7.6)
T !1 T !1 T t T
2
The temporal average X depends neither on t nor on T any longer, but it may a priori
still depend on the considered realization. If this is not the case, it then coincides with
the ensemble average or statistical average hXi, and the process X(t) is said to be
ergodic in the mean.13
More generally, a random stationary process X(t) is said to be ergodic in the
full sense if there is an identity between the temporal average and the statistical
average, not only as far as X(t) is concerned, but also for all the products of the type
X(t1 )X ⇤ (t2 ) . . . which serve to define the higher-order correlation functions.14
13
It can be shown that a sufficient condition for a stationary stochastic process to be ergodic in
the mean is that the correlations decrease sufficiently rapidly at large times for the autocorrelation
function to be integrable. Then every particular realization of the process contains enough statistical
information for the temporal average to coincide with the ensemble average.
14
Criteria may also be established for the ergodicity of higher-order correlation functions.
16 Random variables and random processes
obtained by taking the limit of very large N , and intervals between the observation
times getting smaller and smaller.
Inversely, we have: Z 1
1
xT (t) = x(!)e i!t
d!. (1.9.2)
2⇡ 1
and, inversely: Z 1
1
X(t) = X(!)e i!t
d!. (1.9.4)
2⇡ 1
17
An illustration of the efficiency of harmonic analysis in linear problems will be given in Chap-
ter 10, with the harmonic analysis of the Langevin equation for Brownian motion.
18
For the sake of simplicity, we use the same notation x(.) for the function x(t) and its Fourier
transform x(!), as well as the same notation X(.) for the process X(t) and its Fourier transform
X(!).
18 Random variables and random processes
The angular frequencies !n and the Fourier coefficients an are given by the usual
formulas:
Z
2⇡n 1 T
!n = , an = x(t)ei!n t dt, n = 0, ±1, ±2, . . . . (1.9.6)
T T 0
we have: 8⌦ ↵
> A = 0, n 6= 0
< n
Z T (1.9.11)
: A0 = 1
> ⌦ ↵ ⌦ ↵ ⌦ ↵
X(t) dt = X .
T 0
A realization a0 of A0 is the temporal average of a realization x(t) of the process X(t)
over the interval (0, T ):
Z
T 1 T
a0 = x(t) = x(t) dt. (1.9.12)
T 0
If the process X(t) is ergodic in the mean, which we assume to be the case,19 then:
T ⌦ ↵
lim x(t) = X . (1.9.13)
T !1
• Two-time averages
For a stationary process, the two-time averages only depend on the di↵erence of the
two times involved. The autocorrelation function (⌧ ) = hX ⇤ (t)X(t+⌧ )i of the process
X(t), assumed to be centered, may be written, using the series expansion (1.9.7), as:
1
X 1
X ⌦ ↵
(⌧ ) = An A⇤n0 e i(!n !n0 )t
e i!n ⌧
. (1.9.15)
n= 1 n0 = 1
Thus, there is no correlation between two Fourier coefficients of unequal angular fre-
quencies.
where A0n and A00n denote the random variables of respective realizations a0n and a00n .
When a convenient filter is used to select the angular frequencies belonging to the
interval (!, ! + !), the mean observable intensity is:
X ⌦ 2↵
(!) ! = |An | . (1.10.3)
!n in (!,!+ !)
The right-hand side of equation (1.10.3) involves a sum over all angular frequencies
included in the considered band of width !. The number of modes of this type is:
! T
= !. (1.10.4)
2⇡/T 2⇡
2
In the limit T ! 1, we can write, provided that h|An | i is a continuous function of
the angular frequency:
T ⌦ 2↵
(!) = lim |An | . (1.10.5)
T !1 2⇡
called the spectral density or the noise spectrum20 of the process X(t). Using the
relation (1.9.9), we can also express S(!) as a function of the squared modulus of the
Fourier transform X(!):
1⌦ 2↵
S(!) = lim |X(!)| . (1.10.7)
T !1 T
The introduction of the spectral density allows us to make explicit the continuous
limit of equation (1.9.16) displaying the fact that there is no correlation between
Fourier coefficients of unequal angular frequencies. Indeed, in this limit:
⌦ ↵
X(!)X ⇤ (! 0 ) = 2⇡ (! ! 0 )S(!). (1.10.8)
Formula (1.10.8) has been established assuming that S(!) is a continuous function
of !. This relation may also be viewed as defining the spectral density.
20
It is equally called the power spectrum.
The Wiener–Khintchine theorem 21
The autocorrelation function (⌧ ) thus appears as the Fourier transform of the spectral
density S(!):
Z 1
1
(⌧ ) = S(!)e i!⌧ d!. (1.10.11)
2⇡ 1
Inversely, we have:
Z 1
S(!) = (⌧ )ei!⌧ d⌧. (1.10.12)
1
10.4. An example
Let us consider again the example (1.7.2) of an autocorrelation function (⌧ ) decreas-
ing
R 1 exponentially with a time constant ⌧c , the parameter C being chosen such that
1
(⌧ ) d⌧ = 1:
1
(⌧ ) = e |⌧ |/⌧c . (1.10.13)
2⌧c
This cusped autocorrelation function, non-di↵erentiable at the origin, corresponds by
Fourier transformation to a Lorentzian spectral density:
!c2
S(!) = , ! c = ⌧c 1 . (1.10.14)
!c2 + !2
Appendix
The autocorrelation function appearing in the integrand depends only on the time
di↵erence ⌧ = t t0 . The integration over t and t0 takes place in the square 0 t T ,
0 t0 T (Fig. 1.1).
When t > t0 , we take as new integration variables22 t t0 = ⌧ and t0 . The
2
integration over t0 is done from 0 to T ⌧ , which yields a contribution to h|An | i equal
to: Z T
1
(T ⌧ )(⌧ )ei!n ⌧ d⌧. (1A.2)
T2 0
2
In the same way, the integration of the part t < t0 yields a contribution to h|An | i
equal to:
Z T
1
(T ⌧ )( ⌧ )e i!n ⌧ d⌧, (1A.3)
T2 0
that is, to: Z
1 0
(T + ⌧ )(⌧ )ei!n ⌧ d⌧. (1A.4)
T2 T
t'
T
T−τ
τ
t=
t'−
τ
=
t'
τ
t−
0 τ T t
2
Adding the contributions (1A.2) and (1A.4) to h|An | i, gives:
Z ⇣
⌦ 2↵ 1 T
|⌧ | ⌘
|An | = 1 (⌧ )ei!n ⌧ d⌧. (1A.5)
T T T
Importing the result (1A.5) into the expression (1.10.6) for the spectral density,
gives:
Z T⇣
|⌧ | ⌘
S(!) = lim 1 (⌧ )ei!⌧ d⌧. (1A.6)
T !1 T T
In the limit T ! 1, we get:
Z 1
S(!) = (⌧ )ei!⌧ d⌧. (1A.7)
1
Z 1
1
(⌧ ) = S(!)e i!⌧
d!. (1A.8)
2⇡ 1
The couple of Fourier relations (1A.7) and (1A.8) constitute the Wiener–Khintchine
theorem.
Given the inequality:
Z ⇣ Z
T
|⌧ | ⌘ T
1 |(⌧ )| d⌧ |(⌧ )| d⌧, (1A.9)
T T T
a
R1sufficient condition for the Wiener–Khintchine theorem to be valid is that the integral
1
|(⌧ )| d⌧ be convergent.23
23
We recover here a sufficient condition for the process to be ergodic in the mean.
Bibliography 25
Bibliography
References
S = S(Xi ). (2.1.1)
This function possesses the additivity (or extensivity) property: the entropy of a
system made up of several subsystems is the sum of the entropies of the constituent
subsystems. In an isolated composite system, the lifting of one or of several constraints
allows for exchanges of extensive quantities between di↵erent subsystems. These ex-
changes result in modifications of the Xi variables relative to the subsystems. These
modifications must be compatible with the remaining constraints, as well as with the
conservation laws. The values taken by the extensive variables in the equilibrium state
reached after the lifting of one or of several internal constraints correspond to the
maximum of the entropy compatible with the remaining constraints.”
(Gibbs relation). The Fi are the intensive variables conjugate to the conserved quan-
tities Xi . The relations:
@S
Fi = (2.1.4)
@Xi
are the equations of state of the system. For a fluid mixture at equilibrium in the
absence of chemical reactions, formula (2.1.1) takes the form:
S = S(E, V, Ni ). (2.1.5)
1 P X µi
dS = dE + dV dNi , (2.1.6)
T T i
T
where T and P denote the temperature and the pressure, and µi the chemical potential
per molecule of the constituent i. The intensive variables conjugate to E, V , and Ni
are respectively 1/T , P/T , and µi /T .
by the Xi ’s. We then defines an instantaneous entropy S(Xi ) at each step of the relax-
ation of the Xi ’s. This separation of the variables into two groups implies a separation
of time scales between the microscopic rapid variables and the slow ones. The existence
of well-separated time scales may schematically be represented by the inequality:
⌧c ⌧ ⌧r , (2.2.1)
For the global system, both Xi and Xi0 are internal variables. The total entropy S (0)
is the sum of the entropies of each of the subsystems:
The equilibrium values of Xi and Xi0 are determined by the maximum of the total
entropy (2.2.3), that is, by the condition:
@S (0) @(S + S 0 ) @S @S 0
= = = Fi Fi0 = 0. (2.2.4)
@Xi (0)
Xi @Xi (0)
Xi @Xi @Xi0
Fi = Fi Fi0 (2.2.5)
T 0 are the temperatures of the subsystems. If this affinity vanishes, no heat exchange
can take place across the diathermal wall. Conversely, if this affinity is non-zero, a heat
exchange will take place between the two subsystems. Similarly, if the wall is movable
and if Xi represents the volume, its conjugate affinity is (P/T ) (P 0 /T 0 ), where P
and P 0 are the pressures of the subsystems. If this affinity is non-zero, there will be
a volume exchange between the two subsystems. If the wall is permeable and if Xi
represents a number of molecules, its conjugate affinity is (µ0 /T 0 ) (µ/T ), where µ
and µ0 are the chemical potentials per molecule of the subsystems. If this affinity is
non-zero, an exchange of molecules will take place between the two subsystems.
We characterize the response of the system to the applied force by the rate of
change of the extensive quantity Xi . The flux Ji is defined by:
dXi
Ji = · (2.2.6)
dt
In the above example, each flux vanishes if its conjugate affinity vanishes. Inversely,
a non-zero affinity leads to a non-zero conjugate flux. Generally speaking, it is the
relation between fluxes and affinities which characterizes the changes arising in the
course of irreversible processes.
In the discrete case described here, affinities as well as fluxes have a scalar char-
acter.
that is:
dS (0) X
= Fi Ji . (2.2.8)
dt i
The entropy production has a bilinear structure: it is the sum of the products of each
flux and its conjugate affinity.
This property of the entropy production is not restricted to the case of a discrete
system, but it generalizes to a continuous medium.2
2
See Subsection 4.4.
32 Linear thermodynamics of irreversible processes
According to the local equilibrium hypothesis, the local entropy is, at any time,
the same function of the local thermodynamic quantities as is the equilibrium entropy
of the global ones. In other words, according to this hypothesis, the Gibbs relation is
equally valid locally. We write accordingly:
Z X
dS = dr Fi (r)d⇠i (r), (2.3.2)
i
and: X
d(⇢s) = Fi d(⇢xi ). (2.3.3)
i
The local intensive variables Fi (r) are thus defined at any time as functional derivatives
of S:
@ (r) S
Fi (r) = = · (2.3.4)
@⇠i (r) ⇠i (r)
It is in this way, for instance, that we define in a fluid mixture a local temperature
T (r), a local pressure P(r), and a local chemical potential µi (r) for the molecules of
chemical species i. The relations (2.3.4) are the local equations of state. They do not
3
We adopt here the Eulerian description, in which the cells are fixed volume elements. Each of
these cells constitutes an open system, exchanging, in particular, particles with the surrounding cells.
4
The dependence with respect to the space coordinates being here explicit, the index i refers only
to the nature of the considered variable.
The local equilibrium hypothesis 33
depend on the presence of gradients: indeed, the local Gibbs relation (2.3.2) having
the same form as the equilibrium Gibbs relation (2.1.3), the local equations of state
(2.3.4) have the same form as the equilibrium equations of state (2.1.4).
In the case of a fluid, the mechanism ensuring the tendency towards an equilibrium
can be described in terms of collisions between molecules. Local equilibrium can be
achieved only if these collisions are frequent enough. Conversely, local equilibrium may
be not realized in a very rarefied gas in which collisions are not sufficiently numerous
to thermalize the molecules.5
⇧ ⇧eq
< ⌧ 1. (2.3.5)
⇧ ⇧
5
Such a gas is called a Knudsen gas.
6
See Chapters 5 and 7.
7
For instance, in the case of argon at room temperature and at atmospheric pressure, the density
is n ⇠ 3 ⇥ 1019 cm 3 and the mean free path ` ⇠ 10 5 cm, so that the number of molecules in a local
equilibrium cell is N = n`3 ⇠ 3 ⇥ 104 . This number is sufficiently large for allowing us to neglect
1/2
fluctuations and to use thermodynamics: the relative fluctuations N /N are of order N (see
Supplement 2A), that is, 5 ⇥ 10 3 .
8
Let us examine for instance the criterion (2.3.5), as applied to the temperature, in the case
of argon at room temperature and at atmospheric pressure. It yields the order of magnitude of the
maximum acceptable temperature gradient for the gas to remain in local equilibrium:
Taking again the mean free path value as quoted in Note 7, we find that the local equilibrium
hypothesis for the temperature remains valid as long as the temperature gradient does not exceed
105 K.cm 1 (which in fact confers a large domain of validity to this hypothesis).
34 Linear thermodynamics of irreversible processes
where Ji is the flux of the extensive conserved quantity Xi . The first term involved in
the formula (2.4.8) for S is, according to equation (2.3.3):
@(⇢s) X @(⇢xi )
= Fi · (2.4.10)
@t i
@t
13
The environment is a thermostat or a reservoir with which the system of interest is in contact.
It is considered as being very little perturbated by the processes in which it is involved. In other
words, all the processes concerning the environment are supposed to be reversible, without entropy
production.
36 Linear thermodynamics of irreversible processes
According to the continuity equations (2.4.5) for the extensive conserved variables,
the first and third terms of the expression (2.4.12) for S cancel each other. We are
thus left with: X
S = rFi .Ji . (2.4.13)
i
Fi = rFi , (2.4.14)
P
S = i Fi .Ji . (2.4.15)
In the same way as in a discrete system, the entropy source has the bilinear structure of
a sum of the products of each flux and its conjugate affinity. The fluxes are the current
densities Ji of the conserved extensive variables (and not their time derivatives as it
is the case in a discrete system). As for the affinities Fi , they are the gradients of the
local intensive variables (and not, as in the discrete case, the di↵erences between the
values taken by these variables in two neighboring subsystems). More precisely, we
have: X
1 µi
S = JE .r( ) + JNi .r( ), (2.4.16)
T i
T
where JE denotes the energy flux and JNi the particle flux of the constituent i. The
expression for S is often rewritten with the aid of the entropy flux. This flux is related
to JE and to the JNi ’s by formula (2.4.9), which can be written more explicitly as:
1 X µi
JS = JE JNi · (2.4.17)
T i
T
This gives:14
1 1X
S = JS .rT JNi .rµi · (2.4.18)
T T i
14
This displays the fact that affinities and fluxes are not defined in a unique way.
Linear response 37
Let us now come back to the formula (2.4.15) for S . Established in the partic-
ular case of a fluid in which the only fluxes coming into play are the energy flux and
the particle fluxes of each constituent, this expression does not constitute the most
general form of the entropy source in a continuous medium. Indeed, various types of
irreversible processes, of various tensorial characters, may occur in such a medium.
Chemical reactions are scalar processes, heat and matter transport are vectorial pro-
cesses, whereas the viscous transport is a tensorial process of order 2. The bilinear
structure of the source of entropy is however fully general.
5. Linear response
In a continuous medium in local equilibrium, the fluxes at a given point of space and
at a given time are determined by the values of the affinities at this point and at this
time:15
⇥ ⇤
Ji (r, t) = Ji F1 (r, t), F2 (r, t), . . . . (2.5.1)
The response of the system is local and instantaneous, which implies that the quantities
concerned vary sufficiently slowly in space and in time. Besides, formula (2.5.1) also
expresses the fact that each flux may depend, not only on its conjugate affinity (‘direct
e↵ect’), but also on the other affinities (‘indirect e↵ects’).
Near equilibrium, this formula may be written in the form of a Taylor series
expansion. Since the fluxes vanish if the affinities vanish, this expansion possesses no
zeroth-order term. When the deviations from equilibrium are small,16 the expansion
may be limited to first order:
P
Ji (r, t) = k Lik Fk (r, t). (2.5.2)
The quantities:
@Ji
Lik = , (2.5.3)
@Fk
called the kinetic coefficients, are determined by the equilibrium values of the intensive
variables Fi (such as the temperature or the pressure), and they do not depend on the
constraints maintaining the system out of equilibrium (such as pressure or temperature
gradients):
Lik = Lik (F1 , F2 , . . .). (2.5.4)
The matrix of kinetic coefficients L of elements Lik characterizes the linear response
of the system.
15
In this section, for the sake of simplicity, the symbols Ji and Fi are used to represent scalar
fluxes and affinities, and components of vectorial and/or tensorial ones as well. The index i thus
denotes here both the nature of the conserved variable and, when necessary, the relevant component.
16
The domain of validity of the linearity hypothesis must be studied in each particular situation.
It is not necessarily identical to the domain of validity of local equilibrium.
38 Linear thermodynamics of irreversible processes
In the linear regime, taking into account the form (2.5.2) of the fluxes, we write
the entropy source, which is the sum of the products of each flux by its conjugate
affinity, as: X
S = Lik Fi Fk . (2.5.5)
ik
In the case of e↵ects implying rapid variations in space and time, local equilibrium
may be not realized. Then, the linear response relations do not take the form (2.5.2),
but involve non-local and retarded kinetic coefficients (which signifies that the fluxes
at a given point of space and at a given time depend on the affinities at other points of
space and at times before the time of interest). Such e↵ects lie beyond the framework of
the linear thermodynamics of irreversible processes. They will be studied in statistical
physics using the linear response theory.17
neutral (the carrier density is thus uniform) and maintained at uniform temperature.
In these conditions, the medium is subjected to an electric current, whose density
J depends linearly on the field E provided that this field is not too intense. This is
Ohm’s law,
J = .E, (2.6.1)
JN = D.rn, (2.6.5)
@µ
rµ = rn. (2.6.8)
@n T
1 @µ
D= LN N . (2.6.9)
T @n T
D n @µ
= · (2.6.12)
µD q @n T
The existence of such a relation leads us to think that, in the vicinity of equilibrium, the
linear response to an external perturbation (as characterized here by the mobility),
on the one hand, and the equilibrium fluctuations (characterized by the di↵usion
coefficient), on the other hand, involve the same microscopic mechanisms. We observe
here, in a particular case, a general behavior of near-equilibrium systems.22
22
See Chapters 13 and 14.
A few simple examples of transport coefficients 41
D kT
= · (2.6.14)
µD q
D 2 "F
= , (2.6.16)
µD 3 q
D kTF
= · (2.6.17)
µD q
23
The Einstein relation will be studied anew in several di↵erent contexts later in the book (see
Chapter 6, Chapter 10 devoted to Brownian motion, and Chapter 16).
24
In three dimensions, the Fermi temperature is defined by "F = 32 kTF . Note that relation (2.6.17)
is formally similar to the Einstein relation (2.6.14), provided that T be replaced by TF .
42 Linear thermodynamics of irreversible processes
JE = .rT, (2.6.18)
JE = rT, (2.6.19)
1
JE = LEE r( )· (2.6.20)
T
The comparison between formulas (2.6.19) and (2.6.20) allows us to relate to LEE :
1
= LEE . (2.6.21)
T2
7. Curie’s principle
In each of the above examples, a given affinity solely produces a flux of its conjugate
quantity; this is a direct e↵ect. In other cases an affinity produces additional fluxes
of quantities which are not conjugate to it; such e↵ects are qualified as indirect. In
complex systems, transport phenomena are thus frequently coupled. For instance, in a
fluid mixture, a temperature gradient produces a heat flux and a di↵usion flux as well.
This phenomenon is called the thermodi↵usion.26 Coupled transport phenomena are
also involved in thermoelectric e↵ects, which occur in metals and in semiconductors.27
In most systems, the components of the di↵erent fluxes do not actually depend on
all the affinities. This is a consequence of the symmetry principle or Curie’s principle
(P. Curie, 1894). This principle is traditionally stated as follows: “When some causes
25
The situation is di↵erent in a conductor, where thermal conduction is mainly due to the charge
carriers (see Supplement 2B).
26
See Supplement 2C.
27
See Supplement 2B.
The reciprocity relations 43
produce some e↵ects, the symmetry elements of the causes must also be found in the
e↵ects they produce. When some e↵ects display some dissymmetry, this dissymmetry
must also be found in the causes which produced them.” We mean here by ‘cause’
a physical system with its environment (external fields, constraint fields . . .), and by
‘e↵ect’ a physical property.
However, Curie’s principle applies in this simple form only if the considered e↵ect
is unique. If this is not the case, an e↵ect may have fewer symmetry elements than
its cause (in other words, a given solution of a problem may have fewer symmetry
elements than the data of this problem). In the linear range, the simple formulation
of Curie’s principle is sufficient.
Curie’s principle has several consequences for the linear thermodynamics of irre-
versible processes, in isotropic media and in anisotropic ones as well.
• Isotropic media
In an isotropic medium, fluxes and affinities whose tensorial order di↵er by one unit
(for instance, scalar and vector) cannot be coupled. Indeed, if in the coupling relation
(2.5.2) the di↵erence of tensorial orders between fluxes and affinities were odd, the
kinetic coefficients28 L↵
ik would be the components of an odd-order tensor. Such a
tensor cannot remain invariant with respect to rotations. Isotropic systems may there-
fore be characterized only by kinetic coefficients either scalar or of even tensorial order.
In an isotropic medium, the transportation of scalar quantities such as the particle
number or the energy involve tensors such as L↵ ↵
N N or LEE , proportional to the unit
matrix: L↵ ↵
N N = LN N ↵ , LEE = LEE ↵ . The same is true for the electrical conduc-
tivity, di↵usion, and thermal conductivity tensors, which are respectively of the form
↵ = ↵, , D↵ = D ↵, , and ↵ = ↵, .
• Anisotropic media
In an anisotropic medium, for instance a crystalline medium invariant with respect to
some transformations of a symmetry group, the consideration of symmetry properties
allows us to reduce the number of independent kinetic coefficients. Thus, in a cubic
crystal, the tensors L↵ ↵
N N and LEE are proportional to the unit matrix, as in an
isotropic medium. The same is true of the tensors of electrical conductivity, di↵usion,
and thermal conductivity.
of the microscopic equations of motion, that is, from their invariance under time-
reversal29 (provided however that, if the case arises, we also change the sign of the
magnetic field and/or of the rotation vector).
In order to be able to formulate the reciprocity relations resulting from this in-
variance, we first have to make sure that the chosen affinities and fluxes possess a
well-defined parity under time-reversal. Generally speaking, the entropy being inde-
pendent of the direction of time, the entropy source S is an odd function of t. From
the general structure of the expression for S (formula (2.4.15)), it results that, if
affinities and fluxes possess a well-defined parity, each affinity has a parity opposite to
that of its conjugate flux. Quantities such as the energy density or the particle density
are invariant under time-reversal. They are said to be even, or that their signature
under time-reversal is ✏i = +1. The fluxes of these quantities are odd. Conversely, a
quantity such as the momentum density changes sign under time-reversal. It is an odd
quantity, of signature ✏i = 1. Its flux (tensorial) is even.
Let us consider the example of a conductor with one type of charge carrier, sub-
mitted to a temperature gradient and an electrochemical potential gradient. These
gradients give rise to a particle flux and an energy flux. The general formulas (2.5.2)
take the form:30 8
> µ 1
< JN = LN N r( T ) + LN E r( T )
>
(2.8.2)
: JE = LEN r( µ ) + LEE r( 1 )·
>
>
T T
Both particle and energy densities being invariant under time-reversal, the correspond-
ing Onsager reciprocity relation reads:31
LEN = LN E . (2.8.3)
29
In classical mechanics, time-reversal changes the signs of the velocities, without modifying the
positions. In quantum mechanics, time-reversal associates with a wave function its complex conjugate
quantity.
30
For the sake of simplicity, the conductor is assumed to be isotropic.
31
The relation (2.8.3) was established as early as 1854 by W. Thomson (later Lord Kelvin) in the
framework of the study of the thermoelectric e↵ects related to the simultaneous presence of a particle
flux and an energy flux in conductors (see Supplement 2B).
Justification of the reciprocity relations 45
According to the sign of the product ✏i ✏k , the Onsager–Casimir relations express the
symmetry or the antisymmetry of the kinetic coefficients Lik and Lki . In the case
✏i ✏k = 1, the coefficient Lik corresponds to a coupling of processes with the same parity,
and the reciprocity relation is a symmetry relation (Lik = Lki ). When ✏i ✏k = 1, Lik
corresponds to a coupling between processes with di↵erent parities, and the reciprocity
relation is an antisymmetry relation (Lik = Lki ).
8.3. Generalization
When irreversible processes take place in the presence of a magnetic field H and/or
in a system rotating at the angular frequency ⌦, the Onsager–Casimir relations read:
They relate the kinetic coefficients of two physical systems which di↵er from one an-
other by the change of sign of the parameters H and/or ⌦. If there is neither magnetic
field nor rotation, the Onsager–Casimir relations, which are then given by formula
(2.8.4), simply express the symmetry or the antisymmetry of the kinetic coefficients
of a given system.
values may be obtained with the aid of the theory of thermodynamic fluctuations,33
developed in particular by A. Einstein in 1910.
To calculate the probability of a fluctuation in an isolated system at equilibrium,
Einstein proposed to make use of the fundamental Boltzmann formula for entropy.
The entropy of an isolated system is proportional to the logarithm of the number
of accessible microscopic states (the proportionality relation involves the Boltzmann
constant k). By ‘inverting’ this formula, we get the probability of a fluctuation. If
the system is not isolated, the entropy variation to consider is that of the global
isolated ensemble composed of the system and its environment, that is, the entropy
variation Sint related to the changes internal to the system. The probability of such
a fluctuation is thus given by the Einstein formula:
✓ ◆
Sint
w ⇠ exp · (2.9.1)
k
The entropy variation related to the changes internal to the system is of the
form SintP = S Sexch , where S is the variation of the entropy of the system and
Sexch = i Fi Xi , the entropy exchange between the system and its environment.
The quantity Fi is the equilibrium value, defined by the environment properties, of
the intensive variable conjugate of Xi . The probability of a fluctuation characterized
by the variations S and Xi of the entropy and of the other extensive variables thus
reads:
✓ ◆
S 1X
w ⇠ exp Fi X i . (2.9.2)
k k i
⌦ ↵
Xi Fj = k ij , (2.9.3)
where the generalized forces Fj = Fj are the fluctuations of the intensive variables
Fj . In formula (2.9.3), the symbol h. . .i denotes the average calculated using the
distribution w.
We get from formula (2.9.2):
@w 1
= w Xi . (2.9.4)
@Fi k
33
See Supplement 2A.
Justification of the reciprocity relations 47
whose right-hand side vanishes, since it corresponds to the derivative of the mean value
of a fluctuation, a quantity null by definition. Since Fi represents the equilibrium value
of the intensive variable conjugate of Xi , we have @ Fj /@Fi = ij . This gives:
Z Y
@w
Fj d( Xl ) = ij . (2.9.8)
@Fi
l
Importing this result into formula (2.9.6), we demonstrate the property (2.9.3).
Taking the derivative with respect to ⌧ of both sides of equation (2.9.10), gives:
⌦ ↵ ⌦ ↵
Xi Ẋk = Ẋi Xk . (2.9.11)
34
This hypothesis constitutes one of the first statements of the fluctuation-dissipation theorem (see
Chapter 14). It will be verified in Chapter 10 for the particular case of the equilibrium fluctuations
of a Brownian particle’s velocity.
48 Linear thermodynamics of irreversible processes
In virtue of the property (2.9.3), the identity (2.9.13) reduces to the Onsager
relation (2.8.1) relative to Lik and Lki . It is a symmetry relation, in accordance with
the fact that Xi and Xk have been assumed to be invariant under time-reversal.
9.4. Discussion
The previous arguments rely upon the fluctuation regression hypothesis (formula
(2.9.12)). However, to show that the fluctuations relax according to the same laws as
the averages, we have to study them at the microscopic level. In a strictly thermody-
namic framework, it is therefore impossible to demonstrate the regression hypothesis.
This is the reason why the arguments presented in this chapter do not constitute a
genuine proof. They nevertheless have the interest of shedding light on the funda-
mental link between the reciprocity relations and the reversibility of the microscopic
equations of motion.
dS dSexch dSint
= + = 0. (2.10.1)
dt dt dt
In formula (2.10.1), dSexch /dt represents the rate of change of the entropy exchange
between the system and its environment, and PS = dSint /dt, the entropy production
related to changes internal to the system. The entropy production is non-negative:
Z
dSint
= S dV 0. (2.10.2)
dt V
dSexch
0. (2.10.3)
dt
The minimum entropy production theorem 49
dPS
0. (2.10.4)
dt
The entropy production PS being non-negative, the inequality (2.10.4) shows that the
rate of change of the entropy production always eventually vanishes. The system then
finds itself in a stationary state characterized by the smallest entropy production com-
patible with the imposed boundary conditions. This is the statement of the minimum
entropy production theorem.35
35
We do not provide here a general demonstration of this theorem. An illustration of it will be
provided in connection with the thermoelectric e↵ects (see Supplement 2B).
50 Linear thermodynamics of irreversible processes
Bibliography
References
1. The fluctuations
The theory of fluctuations at thermodynamic equilibrium was established by
J.W. Gibbs in 1902 and A. Einstein in 1910. The fluctuations are viewed as ran-
dom variables, and we have to determine their probability distribution in order to be
able to calculate their di↵erent moments. To begin with, it will be shown that the
definition of the equilibrium fluctuations cannot be carried out in the same way, de-
pending on whether we are concerned with the fluctuations of extensive quantities or
of intensive ones.
Formula (2A.1.1) is formally similar to the Gibbs relation between the di↵erential of
the entropy and the di↵erentials of the Xi variables.
where Sexch comes out from the exchange of entropy between the system and its
environment, and Sint from the entropy variation related to changes internal to the
system. Interestingly, the fluctuation Sint = S Sexch represents the fluctuation of
the entropy of the global isolated ensemble composed of the system and its environ-
ment. In virtue of the maximum entropy principle applied to this isolated ensemble at
equilibrium, we have:
Sint 0. (2A.2.2)
The inequality (2A.2.2) applies to any thermodynamic system, whatever the equilib-
rium conditions under which it is placed. It includes all the situations for which the
second principle of thermodynamics is usually expressed in terms of thermodynamic
potentials. Let us now check this latter statement using a few examples.
• Isolated system
For an isolated system, we have Sexch = 0, since there is no interaction with an
environment. We thus have S = Sint . The second principle for an isolated system
implies S 0, that is, Sint 0.
1
It is for instance the case in the study of the validity of the local equilibrium criterion for an
intensive quantity ⇧ such as the temperature or the pressure: the variation ⇧ produced inside
a local equilibrium cell by an imposed gradient must indeed remain smaller than the equilibrium
fluctuation ⇧eq .
Probability of a fluctuation: the Einstein formula 53
The form (2A.3.2) of the distribution w is fully general. It allows us to compute the
di↵erent moments of the fluctuations of the thermodynamic quantities around their
equilibrium values. Using this expression for w, we can in particular demonstrate the
identity: ⌦ ↵
Xi Fj = k ij , (2A.3.3)
which expresses the absence of correlations between the fluctuations of an extensive
quantity and those of the intensive quantities other than its conjugate variable.
54 Thermodynamic fluctuations
we define the fluctuations of these intensive quantities as resulting from those of the
extensive quantities S and V :
✓ 2 ◆
@2E @2E @ E @2E
T = S + V , P = S + V · (2A.4.5)
@S 2 @S@V @S@V @V 2
✓ ◆
P V T S
wN ⇠ exp · (2A.4.7)
2kT
Equilibrium fluctuations in a fluid of N molecules 55
N cv @P @P @P
S= T+ V, P= T+ V, (2A.4.8)
T @T V @T V @V T
where cv denotes the specific heat at constant volume per molecule. Importing the
expressions (2A.4.8) for S and P into formula (2A.4.7), gives:
N cv 2 1 @P 2
wN ⇠ exp ( T) + ( V) · (2A.4.9)
2kT 2 2kT @V T
The mean square fluctuations of temperature and volume are given by:
⌦ 2↵ kT 2 ⌦ 2↵
( T) = , ( V) = kT V T. (2A.4.11)
N cv
2
If the number of molecules of the fluid is allowed to vary, we can write an analogous approximate
formula for the probability w of a fluctuation,
✓ ◆
P V T S µ N
w ⇠ exp ,
2kT
where µ denotes the chemical potential per molecule. The physical conditions being di↵erent, the
moments computed with the aid of w are not necessarily identical to those obtained with the aid
of wN .
3
In the expression for S, we have taken into account the Maxwell relation (@S/@V )T = (@P/@T )V .
4
This approximate form of wN allows us to compute exactly the second-order moments of the
fluctuations, but not their higher-order moments.
56 Thermodynamic fluctuations
In practice, rather than in the volume fluctuations, we are interested in the fluctuations
of the density n = N/V . The number of molecules being fixed, we have, using neq to
denote the average equilibrium fluctuation of the density:
✓ ◆1/2
neq Veq kT T
= = · (2A.4.13)
n V V
We deduce from equation (2A.4.13) the relative fluctuation of the number of molecules
N = nV in a local equilibrium cell of fixed volume V within a fluid in local equilib-
rium: ✓ ◆1/2
N nkT T
= · (2A.4.14)
N N
• Variables P and S
We express the fluctuations of volume and temperature in terms of those of pressure
and entropy. We thus write:6
@T @V T @T
V = S+ P, T = S+ P, (2A.4.15)
@P S @P S N cp @P S
where cp denotes the specific heat at constant pressure per molecule. Importing the
expressions (2A.4.15) for V and T into formula (2A.4.7), gives:
1 @V 2 1 2
wN ⇠ exp ( P) ( S) · (2A.4.16)
2kT @P S 2kN cp
⌦ ↵
P S = 0. (2A.4.17)
5
The root-mean-square deviations or average equilibrium fluctuations Teq and Veq are defined
by the formulas:
( Teq )2 = h( T )2 i, ( Veq )2 = h( V )2 i.
6
In the expression for V , we have taken into account the Maxwell relation (@V /@S)T = (@T /@P)S .
Equilibrium fluctuations in a fluid of N molecules 57
The mean square fluctuations of pressure and entropy are given by:7
⌦ 2↵ kT cp ⌦ 2↵
( P) = , ( S) = kN cp . (2A.4.18)
V cv T
Formulas (2A.4.12) and (2A.4.19) show that, in a fluid of N molecules, the average
relative fluctuations of temperature and pressure are of order N 1/2 . This result plays
an important role in the discussion of the local equilibrium criteria for temperature
and pressure.
7
We have made use of the relation between the isentropic and isothermal compressibilities:
1 @V cv
= T.
V @P S cp
8
The root-mean-square deviation or average equilibrium fluctuation Peq is defined by the for-
mula:
( Peq )2 = h( P)2 i.
58 Thermodynamic fluctuations
Bibliography
References
1. Introduction
The thermoelectric e↵ects are phenomena associated with the simultaneous presence of
an electric current and a heat flux¡ in a system, in practice a metal or a semiconductor.
These e↵ects were studied in 1854 by W. Thomson (later Lord Kelvin).
We assume, for the sake of simplicity, that there exists in the conductor only one
type of carrier, with charge q. The system is assumed to be macroscopically neutral (the
carrier density is thus uniform). The temperature T (r) and the electrostatic potential
(r) may vary from one point to another inside the sample, but they remain constant
over the course of time. Since the density is uniform, the local chemical potential
depends on r solely via the local temperature: it is accordingly denoted by µ[T (r)].
The local electrochemical potential is:
⇥ ⇤
µ(r) = µ T (r) + q (r). (2B.1.1)
and, inversely: 8
>
> LN N = L11
>
<
LN E = L12 + µL11 , LEN = L21 + µL11 (2B.2.6)
>
>
>
:
LEE = L22 + µ(L12 + L21 ) + µ2 L11 .
2 ⇤ and J
The relation (2B.2.2) between the fluxes JQ Q will be established in Subsection 6.3.
Isothermal electrical conduction 61
dQ 1
= J 2. (2B.3.6)
dt Joule
The divergence of the entropy flux vanishes (r.JS = 0), and the entropy source
(2B.2.3) reads:
1 2
S = J . (2B.3.7)
T
The local balance equation for the entropy is thus:
@(⇢s) 1 2
= J . (2B.3.8)
@t T
This entropy production is at the root of the Joule e↵ect, and it expresses the irre-
versibility of the charge transport. It is characterized by the electrical conductivity ,
a positive quantity. The electrical conductivity is a transport coefficient or dissipative
coefficient (and not a thermodynamic equilibrium quantity).
62 Thermoelectric e↵ects
1 L12
rµ = rT. (2B.4.1)
T L11
1 LEE LN N LN E LEN
= · (2B.4.4)
T2 LN N
1
rµ = ⌘rT, (2B.5.1)
q
where ⌘ denotes the thermoelectric power or Seebeck coefficient of the material. Com-
paring equations (2B.4.1) and (2B.5.1) gives:
1 L12
⌘= · (2B.5.2)
qT L11
The Seebeck e↵ect exists only if L12 is non-zero. In that sense it is an indirect e↵ect.
We can also write: ✓ ◆
1 LN E
⌘= µ . (2B.5.3)
qT LN N
3
In a conductor, thermal conduction is mainly due to charge carriers, that is, to particles whose
number is conserved. The principal mechanism coming into play is thus di↵erent from the one at
work in an insulating solid (in this latter case, heat transport is due solely to phonons, whose number
is not conserved, and we have = LEE /T 2 ).
The Peltier e↵ect 63
Metal M
A1 A2
Metal M'
B1 B2
To observe the Seebeck e↵ect, we realize the thermocouple represented in Fig. 2B.1.
It is a circuit made of two di↵erent conductors, namely, two wires of thermopowers ⌘
and ⌘ 0 , welded together at points A1 and A2 . The welds are brought up to unequal
temperatures T1 and T2 . The circuit being open, we measure between points B1 and
B2 the potential di↵erence:
Z
1 B2
2 1 = rµ.dl, (2B.5.4)
q B1
the integral being taken along the circuit B1 A1 A2 B2 . Taking into account formula
(2B.5.1), we get:
Z T2
2 1 = (⌘ 0 ⌘) dT. (2B.5.5)
T1
6.1. Description
At uniform temperature, an electric current of density J = qJN is accompanied by a
heat flux:
JQ⇤
= ⇡J , (2B.6.1)
4
This potential di↵erence is measured with the aid of a highly resistive voltmeter, so that the
current in the circuit is negligible. Once calibrated, the device allows us to measure the temperature
of A2 , that of A1 being fixed at a reference temperature T1 . The measurement is independent of the
ambient temperature.
64 Thermoelectric e↵ects
dQ
= (⇡ ⇡ 0 )J. (2B.6.4)
dt Peltier
The Peltier e↵ect is linear in J ; depending on the direction of the electric current, an
absorption or a release of heat5 takes place at the junction.
⇡ = ⌘T. (2B.6.5)
Thus, the four kinetic coefficients associated with the thermoelectric e↵ects may
be expressed with the aid of three independent experimentally measurable dissipative
coefficients, such as , , and ⌘.
JE = J Q
⇤
+ µJN = rT + (µ + ⌘qT )JN . (2B.7.1)
The power exchanged with the outside per unit volume, given by the energy balance:
dQ ⇥ ⇤
= r.JE = r.(rT ) r. (µ + ⌘qT )JN , (2B.7.2)
dt
reads, since the flux JN is uniform:
dQ
= r.(rT ) JN .r(µ + ⌘qT ). (2B.7.3)
dt
The expression for rµ is obtained from the first of equations (2B.2.4), and formulas
(2B.3.2) and (2B.5.2):
q2
rµ = ⌘qrT JN . (2B.7.4)
We thus have:
dQ q2 2
= r.(rT ) qT JN .r⌘ + JN , (2B.7.5)
dt
that is, the thermopower being a function of the local temperature:
dQ d⌘ 1
= r.(rT ) T J .rT + J 2 . (2B.7.6)
dt dT
In the expression (2B.7.6) for dQ/dt, the first term represents the power exchanged
with the outside due to thermal conduction. The third term represents the power dis-
sipated by the Joule e↵ect. As for the second term, it corresponds to the Thomson
e↵ect: it represents the power exchanged with the external medium when an electric
current of density J goes through the temperature gradient rT . The Thomson coef-
ficient ↵ is defined as the Thomson power per unit electric current density and unit
temperature:
dQ
= ↵J .rT. (2B.7.7)
dt Thomson
In contrast to the Joule e↵ect, the Thomson e↵ect is linear in J ; depending on the
current direction, there is an absorption or a release of heat. According to formula
(2B.7.6), the Thomson coefficient is related to the Seebeck coefficient:
d⌘
↵=T · (2B.7.8)
dT
66 Thermoelectric e↵ects
d⇡
↵= ⌘. (2B.7.9)
dT
Formula (2B.7.9) is the first Kelvin relation. It can also be established directly with
the aid of the first law of thermodynamics.
Purely thermodynamic considerations do not allow us to go further. To obtain
explicit expressions for the three dissipative coefficients , , and ⌘, we have to turn
to a microscopic theory such as the Boltzmann transport equation6 or the Kubo linear
response theory.7
Bibliography
1. Introduction
When put in the presence of a temperature gradient, the constituents of a uniform
mixture manifest a tendency towards separation. This indirect phenomenon is called
the thermodi↵usion, or the Soret e↵ect. It takes place in gases and in liquids, and also
in solids. The inverse e↵ect of the Soret e↵ect is called the Dufour e↵ect. This relates
to the fact that a concentration gradient imposed to a mixture, besides the di↵usion
current which tends to re-establish homogeneity, also produces a heat flux.
We shall describe here these e↵ects in the simple case of a binary fluid mixture
in the absence of external forces, viscous phenomena, and chemical reactions.
nA nA
c= = · (2C.2.1)
nA + nB n
The concentration varies over the course of time and also from one point of space to
another, because of di↵usion and of the fluid macroscopic motion. We denote by uA
and uB the local mean velocities of the constituents A and B, and by u the local
barycentric velocity defined by:
We also introduce the di↵usive fluxes of A and B with respect to the local barycentric
velocity:
JAdi↵
= nc(uA u), di↵
JB = n(1 c)(uB u). (2C.2.3)
The entropy source 69
If the fluid consists of molecules all of the same species, there is no di↵usion flux. If
for instance c = 0, there are solely B molecules in the mixture. Then we have uB = u.
The di↵usive fluxes thus vanish: JA di↵
= JB di↵
= 0. A similar reasoning holds in the
case c = 1.
1 µA µB
S = JE .r + JA .r + JB .r , (2C.3.1)
T T T
where JE is the energy flux and Jk the flux of the constituent k (k = A or k = B). In
formula (2C.3.1), T denotes the local temperature and µk the local chemical potential
per molecule of the constituent k.
The flux of A is the sum of a convective flux and the di↵usive flux JA
di↵
:
JA = ncu + JA
di↵
. (2C.3.2)
µk 1 1
r = rµk + hk r , (2C.3.4)
T T T T
where rµk |T denotes the isothermal gradient of µk , and hk the enthalpy per molecule
of the constituent k, we obtain:
✓ ◆
1 ⇥ ⇤ 1
S = J E .r nu. crµA T + (1 c)rµ B T + ch A + (1 c)h B r
T T
µA µB
+ JA
di↵
.r + JB
di↵
.r ·
T T
(2C.3.5)
The pressure being uniform, the Gibbs–Duhem relation for the mixture reads:
crµA T
+ (1 c)rµB T
= 0. (2C.3.6)
We thus get:
1 µA µB
S = JE nhu .r + JA
di↵
.r + JB
di↵
.r , (2C.3.7)
T T T
70 Thermodi↵usion in a fluid mixture
where:
h = chA + (1 c)hB (2C.3.8)
is the enthalpy per molecule of the mixture. Introducing the heat flux JQ = JE nhu,1
⇤
JQ = JE µA J A µB JB .
cond , is defined by:
The heat flux due to thermal conduction, denoted here by JQ
cond ⇤
JQ = JQ T sA JA T sB JB ,
where sk is the entropy per molecule of the constituent k. We have the relation:
cond di↵ di↵
JQ = JQ hA JA hB J B .
The di↵erence between JQ and JQ cond represents that part of the heat transport which is due to
matter di↵usion.
2
In a pure fluid, the di↵usive fluxes vanish, and the entropy source reduces to its thermal part
(we do not take into account here the viscous phenomena):
⇣1⌘
S = JQ .r ·
T
This expression for S will be encountered in Chapter 16, in which we calculate microscopically the
thermal conductivity of a fluid.
Linear relations between fluxes and affinities 71
where the kinetic coefficients verify the Onsager symmetry relation L12 = L21 .
When the di↵usion flux vanishes (J di↵ = 0), there is solely thermoconduction (JQ =
rT ). This allows us to deduce the expression for the thermal conductivity of the
fluid mixture in terms of the kinetic coefficients:
1 L11 L22 L12 L21
= · (2C.4.3)
T2 L11
@µ @µ
rµ = rT + rc. (2C.4.4)
@T p,c @c p,T
Introducing the expression (2C.4.4) for rµ into equations (2C.4.1) and taking into
account the Onsager symmetry relation, we obtain for the di↵usion flux and the heat
flux the expressions:
8 ⇣ ⌘
> kT
>
> J di↵ = nD rc + rT
< T
⇣ ⌘ (2C.4.5)
>
> @µ @µ
>
: Q J = µ + k T T J di↵
rT.
@c p,T @T p,c
L11 1 @µ
D= · (2C.4.6)
n T @c p,T
T ⇣ L11 @µ L12 ⌘
DT = + · (2C.4.7)
n T @T p,c T2
kT = ↵c(1 c) (2C.5.2)
(an expression consistent with the fact that no thermodi↵usion can take place in a
pure fluid). Making the approximation:
we obtain for the concentration di↵erence between the top and the bottom of the box
the expression:
Ttop
ctop cbottom = ↵c0 (1 c0 ) log · (2C.5.4)
Tbottom
A substance with a positive (resp. negative) Soret coefficient kT tends to di↵use to-
wards the coldest (resp. hottest) region and to accumulate there.3
The inverse e↵ect of the Soret e↵ect, that is, the existence of a heat current due
to a concentration gradient, was studied in 1872 by L. Dufour.
3
The Soret e↵ect is thus the basis of a method of separation of isotopic gaseous mixtures (however,
in liquid phase, the order of magnitude of the e↵ect is too small to allow us to consider its practical
application).
Bibliography 73
Bibliography
References
L. Dufour, Über die Di↵usion der Gase durch poröse Wände und die sie begleitenden
Temperaturveränderungen, Annalen der Physik 28, 490 (1873).
C. Soret, Propagation de la chaleur dans les cristaux, Comptes rendus des séances
de la Société de Physique et d’Histoire Naturelle de Genève 10, 15 (1893); Archives
des Sciences Physiques et Naturelles (Genève) 29, 4 (1893).
This page intentionally left blank
Chapter 3
Statistical description
of out-of-equilibrium systems
In most problems discussed in this book, the material systems under consideration,
be they gaseous, liquid, or solid, are composed of a very large number of particles.
When such a system is in equilibrium (or in local equilibrium), its macroscopic state
may be parametrized by a limited number of extensive variables (or by their local
densities). On the other hand, its microscopic state is not known exactly: the number of
microscopic degrees of freedom coming into play is indeed so huge that it is completely
impossible to have at one’s disposal a full knowledge of the system at this level.
Such a detailed knowledge is in fact not necessary to describe the system in
a macroscopic way. In statistical physics, we define the variables parametrizing the
macroscopic state of a system of a very large number of particles as statistical averages
of the variables associated with the microscopic states compatible with the macroscopic
state under consideration. These microscopic variables are functions of the coordinates
and momenta of all the particles of the system. In order to determine their averages,
we introduce, for a classical system, the phase space distribution function or, for a
quantum system, the density operator.
Out of equilibrium, the averages defining the macroscopic variables are functions
of time. One of the most used procedures to determine their temporal dependence
consists in studying the evolution of the phase space distribution function or of the
density operator. The notions of phase space distribution function and density opera-
tor, as well as their temporal evolution, are therefore the subject of this introductory
chapter to out-of-equilibrium statistical physics.
76 Statistical description of out-of-equilibrium systems
is the probability for the representative point of the microscopic state of the fluid to
be found at time t in the volume element dq1 . . . dq3N dp1 . . . dp3N centered at the point
of coordinates q1 , . . . , q3N , p1 , . . . , p3N . This probability may also be denoted for short
by f (q, p, t) dqdp. The phase space distribution function is a non-negative quantity. A
priori, it must be normalized in such a way that:
Z
f (q, p, t) dqdp = 1, (3.1.2)
the integration being carried out over the whole phase space.
Once the distribution function is known, we calculate the ensemble average hA(t)i
of a microscopic variable A(q, p) as the phase space integral of A(q, p) weighted by
f (q, p, t):
R
⌦ ↵ A(q, p)f (q, p, t) dqdp
A(t) = R · (3.1.3)
f (q, p, t) dqdp
where u represents the 6N -components vector q̇1 , . . . , q̇3N , ṗ1 , . . . , ṗ3N , and ⌫ the unit
vector normal to ⌃ and oriented towards the exterior of V. The surface integral on the
right-hand side of equation (3.1.8) may be transformed into a volume integral:
Z
dn
= N r.(f u) dqdp. (3.1.9)
dt V
The expressions (3.1.7) and (3.1.9) for dn/dt must be identical whatever the volume V.
We deduce from this identity the local conservation equation of the number of repre-
sentative points in the phase space:
@f
+ r.(f u) = 0. (3.1.10)
@t
In order to take into account the microscopic equations of motion, we make explicit
r.(f u):
3N ✓
X ◆ X3N ✓ ◆
@f @f @ q̇i @ ṗi
r.(f u) = q̇i + ṗi + + f. (3.1.11)
i=1
@qi @pi i=1
@qi @pi
The second sum appearing on the right-hand side of equation (3.1.11) vanishes as a
consequence of equations (3.1.5). We are thus left with:
3N ✓
X ◆
@f @f
r.(f u) = q̇i + ṗi . (3.1.12)
i=1
@qi @pi
6
Here, we use the normalization (3.1.2).
The phase space distribution function 79
Using once again Hamilton’s equations, we can write the local conservation equation
(3.1.10) in the form of the Liouville equation:
3N ✓ ◆
@f X @H @f @H @f
+ = 0. (3.1.13)
@t i=1
@pi @qi @qi @pi
The sum appearing on the left-hand side of equation (3.1.13) is the Poisson bracket of
H and f , which we denote by7 {H, f }. We thus have:
@f
+ {H, f } = 0. (3.1.14)
@t
@f
= iLf. (3.1.15)
@t
where the symbol • stands for any function of the generalized coordinates and mo-
menta. The Hamiltonian H being time-independent, the formal integration of the
evolution equation (3.1.15) yields, for the distribution function at time t, the expres-
sion:
f (q, p, t) = e iLt f (q, p). (3.1.17)
In formula (3.1.17), f (q, p) represents the distribution function at a given time chosen
as origin.
@f
+ u.rf + f r.u = 0. (3.1.18)
@t
7
The sign convention adopted in the definition of the Poisson bracket is not universal. Here, we
use the convention:
3N ⇣
X @↵ @ @↵ @ ⌘
{↵, } = ·
i=1
@pi @qi @qi @pi
80 Statistical description of out-of-equilibrium systems
d @
= + u.r, (3.1.19)
dt @t
we obtain once again, from equation (3.1.18), the Liouville equation, now written in
the form:
df
= 0. (3.1.20)
dt
Equation (3.1.20) can be interpreted as the fact that the set of the representative
points in the phase space behaves like an incompressible fluid (that is, like a fluid
whose density does not change when we follow the flow over the course of time).
where the {| n i} denote the eigenstates of a complete orthonormal base. The complex
numbers cn are the coefficients of the expansion of the pure state | i over the states
| n i. The expectation value of a microscopic observable A in the pure state | i is:
X
hAi = h |A| i = h ⇤
n |A| m icn cm . (3.2.2)
n,m
8
Formula (3.1.19) establishes the link, as far the derivation with respect to time is concerned,
between the Lagrangian description, in which the derivative d/dt takes the fluid motion into account,
and the Eulerian description, in which the derivative @/@t is calculated at a fixed point in space.
The density operator 81
with each of these states. This is the reason why a macroscopic state is also called a
statistical mixture. The probabilities pi are non-negative and normalized:
X
pi 0, pi = 1. (3.2.3)
i
The expectation value of a physical quantity A in the considered macroscopic state is:
X
hAi = pi h i |A| i i. (3.2.4)
i
P
⇢= i | i ipi h i |. (3.2.7)
The density operator is Hermitean and positive.9 Over the orthonormal base {| n i},
its matrix elements are:
X
⇢mn = h m |⇢| n i = pi c⇤ni cmi . (3.2.8)
i
All the information about the macroscopic state of the system P is contained in the
2
density operator. The diagonal element ⇢nn = h n |⇢| n i = i pi |cni | represents the
average probability of finding the system in the state | n i. This is the reason why the
diagonal elements of the density matrix are called the populations. The o↵-diagonal
elements are called the coherences.
9
The positivity of ⇢ signifies that, for any state | i, we have h |⇢| i 0.
82 Statistical description of out-of-equilibrium systems
that is:
hAi = Tr(⇢A). (3.2.11)
d⇢
= iL⇢. (3.2.14)
dt
The symbol L denotes here the quantum Liouville operator,10 defined by:
1
L• = [H, •], (3.2.15)
h̄
where the symbol • stands for any operator.
The Hamiltonian being time-independent, the formal integration of the evolution
equation (3.2.14) yields, for the density operator at time t, the expression:
⇢(t) = e iLt
⇢. (3.2.16)
In formula (3.2.16), ⇢ represents the density operator at a given time chosen as origin.11
We can also write, equivalently:
⇢(t) = e iHt/h̄
⇢eiHt/h̄ . (3.2.17)
If the {| n i} are the eigenstates of H, of energies "n , we have:
(
⇢nn (t) = ⇢nn
(3.2.18)
⇢mn (t) = ⇢mn e i!mn t
, m 6= n,
with !mn = ("m "n )/h̄. The populations do not depend on time and the coherences
oscillate at the Bohr angular frequencies !mn .
10
The Liouville operator does not act in the state space, but in the operator space: this is why it
is sometimes referred to as a superoperator.
11
As displayed by formulas (3.2.16) and (3.1.17), the use of the Liouville operator leads to formally
analogous expressions for ⇢(t) and f (q, p, t).
Systems at equilibrium 83
3. Systems at equilibrium
The most often encountered equilibrium situations are either that of a closed system
likely to exchange heat with a thermostat (canonical equilibrium) or that of an open
system likely to exchange heat and particles with a reservoir (grand canonical equi-
librium). Let us make precise the form of the density operator (resp. of the classical
distribution function) in both situations.
1
⇢ (resp. f ) = e H
, = (kT )
1
. (3.3.1)
Z
The partition function Z is expressed as a discrete sum over the states of the system
(resp. as an integral over the phase space of the system).
• Quantum case
We have:
Z = Tr(e H
), (3.3.2)
where the trace concerns all the states of the system.
• Classical case
The trace is replaced by a phase space integral. The partition function of a classical
system of N particles in a three-dimensional space reads for instance:
Z
1
ZN = e H
dqdp. (3.3.3)
N !h3N
1
⇢ (resp. f ) = e (H µN )
, (3.3.4)
⌅
• Quantum case
For a quantum system, we have:
⌅ = Tr[e (H µN )
], (3.3.5)
where the trace concerns all states with all possible values of the number N of particles.
• Classical case
In order to obtain the grand partition function of a classical system, we can for instance
use the general relation:
1
X
⌅= e µN ZN , (3.3.6)
N =0
Z
⌦ ↵ 1
A(t) = A(q, p)e iLt
f (q, p) dqdp. (3.4.2)
N !h3N
• Second approach
This relies on the fact that the generalized coordinates and momenta evolve with
time according to Hamilton’s equations (3.1.5). We thus consider that the dynamical
variable A[q(t), p(t)] = A(t) itself evolves with time. We have:
3N ✓ ◆
dA X @A @A
= q̇i + ṗi , (3.4.3)
dt i=1
@qi @pi
Evolution of the macroscopic variables: classical case 85
3N ✓ ◆
dA X @A @H @A @H
= · (3.4.4)
dt i=1
@qi @pi @pi @qi
The sum on the right-hand side of equation (3.4.4) is the Poisson bracket of H and A.
Equation (3.4.4) thus reads:
dA
= {H, A}, (3.4.5)
dt
that is, in terms of the Liouville operator (3.1.16):
dA
= iLA. (3.4.6)
dt
Z
⌦ ↵ 1
A(t) = eiLt A(q, p)f (q, p) dqdp. (3.4.9)
N !h3N
Formula (3.4.13) can be proved using integrations by parts and the fact that the
distribution function f vanishes for qi = ±1 or pi = ±1. It follows from property
(3.4.12) that the expressions (3.4.10) and (3.4.11) for dhA(t)i/dt are identical, hence
the identity of the expressions (3.4.2) and (3.4.9) for hA(t)i. This latter property is
expressed through the equality:
Z Z
A(q, p)e iLt
f (q, p) dqdp = eiLt A(q, p)f (q, p) dqdp, (3.4.14)
which allows us to transfer the time dependence of the distribution function to the
dynamical variables, and vice versa.
⌦ ↵
A(t) = Tr e iHt/h̄
⇢ eiHt/h̄ A . (3.5.2)
Evolution of the macroscopic variables: quantum case 87
dA
= iLA. (3.5.4)
dt
⌦ ↵
A(t) = Tr ⇢ eiHt/h̄ A e iHt/h̄
. (3.5.8)
13
Once again, the use of the Liouville operator leads to a formal analogy between the quantum
formula (3.5.5) and the classical formula (3.4.7).
88 Statistical description of out-of-equilibrium systems
Bibliography
XN N
X
p2i
HN = + u |ri rj | . (4.1.1)
i=1
2m i,j=1
i<j
In formula (4.1.1), each pair interaction term is assumed to depend only on the distance
between the two particles i and j of the considered pair.2 Interactions involving more
than two particles are not taken into account.
XN N
X X N
p2i
HN = + (ri ) + u |ri rj | . (4.1.2)
i=1
2m i=1 i,j=1
i<j
• Particles with pair interactions in the presence of a scalar potential and of a vector
potential
If the particles carry a charge q and are submitted to an external electromagnetic field
deriving from a scalar potential (r, t) and a vector potential A(r, t), the Hamiltonian
becomes:
N ⇥
X
⇤2
XN X N
pi qA(ri , t)/c
HN = +q (ri , t) + u |ri rj | , (4.1.3)
i=1
2m i=1 i,j=1
i<j
1
The pair interactions are interactions between the particles taken two by two.
2
Some most used pair interaction potentials are described in an appendix at the end of this
chapter.
The Liouville equation 91
In equation (4.2.3), rri (resp. rpi ) denotes the gradient with respect to the coordinate
(resp. momentum) variable of the particle i. While ṙi is clearly just the velocity vi , it
is on the other hand necessary to make explicit ṗi . We have, in a general way,
dpi
= Fi , (4.2.4)
dt
where Fi represents the force exerted on the particle i. This force is the sum of the
force Xi due to the scalar and vector potentials possibly present,3 and of the forces Xij
due to the other particles:
XN
Fi = Xi + Xij . (4.2.5)
j=1
j6=i
3
The calculation of the force Xi exerted on a particle in the presence of a scalar potential and
of a vector potential describing an external electromagnetic field is carried out in an appendix at the
end of this chapter.
92 Classical systems: reduced distribution functions
The force Xij derives from the corresponding pair interaction potential:
We will now introduce the expression for ṗi in equation (4.2.3), first in the case
without interactions (Fi = Xi ), then in the case of pair interactions (Fi = Xi +
PN
j=1 Xij ).
j6=i
N N
@f X X
+ vi .rri f + Xi .rpi f = 0. (4.2.7)
@t i=1 i=1
N N
@f X X
+ vi .rri f + Fi .rpi f = 0. (4.2.8)
@t i=1 i=1
The similarity between equations (4.2.7) and (4.2.8) is only an apparent one. Indeed,
in contrast to Xi , the force Fi involves the parameters of all particles j interacting
with the particle i.
For a system with N 1 particles, it is an impossible task to directly tackle
the resolution of such an equation. This is the reason why we establish a formalism
relying on a hierarchy of coupled equations for reduced distribution functions, that is,
for distribution functions concerning only a limited number of particles. This approach
makes it possible, through some approximations, to eliminate the irrelevant informa-
tion in a simpler way than by working directly on the Liouville equation (4.2.8) obeyed
by the full distribution function.
Reduced distribution functions: the BBGKY hierarchy 93
with:
Z XN
⌦ ↵ 1 p2i
Ec (t) + Epgrav (t) = + (ri ) f (r1 , p1 , . . . , rN , pN , t) d⌧N (4.3.2)
N !h3N i=1
2m
and:
Z X
N
⌦ ↵ 1
Epint (t) = u |ri rj | f (r1 , p1 , . . . , rN , pN , t) d⌧N . (4.3.3)
N !h3N i,j=1
i<j
In formulas (4.3.2) and (4.3.3), hEc (t)i, hEpgrav (t)i, and hEpint (t)i denote respectively
the expectation values of the kinetic energy, the gravitational potential energy, and
the interaction potential energy. Also, we have introduced the condensed notation:
N
Y
d⌧N = d⌦i , d⌦i = dri dpi . (4.3.4)
i=1
4
The total kinetic energy is the sum of the individual kinetic energies p2i /2m (one-particle dy-
namical variables). In the same way, the potential energy in a gravitational field described by the
potential (r) is the sum of the individual potential energies (ri ). The interaction potential energy
is the sum of the pair interaction potential energies (two-particle dynamical variables).
94 Classical systems: reduced distribution functions
distribution functions, respectively denoted by f (1) (r1 , p1 , t) and f (2) (r1 , p1 , r2 , p2 , t),
and defined by:
8 Z
> 1
>
>
< f (r1 , p1 , t) =
(1)
f (r1 , p1 , . . . , rN , pN , t) d⌦2 . . . d⌦N
(N 1)!h3N
> Z
>
> 1
: f (r1 , p1 , r2 , p2 , t) =
(2)
f (r1 , p1 , . . . , rN , pN , t) d⌦3 . . . d⌦N .
(N 2)!h3N
(4.3.5)
The functions f (1) and f (2) obey the relations:
Z Z Z
f d⌦1 = N,
(1)
f d⌦1 d⌦2 = N (N 1),
(2)
f (2) d⌦2 = (N 1)f (1) . (4.3.6)
In the same way, and more generally, we define reduced distribution functions
concerning any number n (1 n < N ) of particles:
Z
1
f (n) (r1 , p1 , . . . , rn , pn , t) = f (r1 , p1 , . . . , rN , pN , t) d⌦n+1 . . . d⌦N .
(N n)!h3N
(4.3.8)
The n-particle reduced distribution function is proportional to the probability den-
sity of finding n particles at r1 , p1 , . . . , rn , pn at time t, whatever the positions and
momenta of the (N n) other particles.
The reduced distribution function f (1) may be interpreted as the distribution func-
tion in a one-particle (and six-dimensional) phase space. Indeed, the average number
of particles in a volume element drdp around the point (r, p) of this phase space is
given by: DX E
(r ri ) (p pi ) drdp = f (1) (r, p) drdp. (4.3.9)
i
Among the reduced distribution functions, the most useful in practice are the
one-particle and two-particle distribution functions. We now want to establish their
evolution equations in the presence of pair interactions. To begin with, we will address
the simpler case of non-interacting particles.
evolution equation for f (1) . The calculation relies on the fact that the distribution
function f (q, p, t) vanishes for qi = ±1 or pi = ±1. All terms in which the presence
of a derivative allows us to reduce by one the number of successive integrations thus
lead to a vanishing result.5 We show in this way that, in the absence of interactions
between the particles, f (1) obeys a closed evolution equation:
@f (1) (4.3.10)
+ v1 .rr1 f (1) + X1 .rp1 f (1) = 0.
@t
The term v1 .rr1 f (1) is called the drift term, and the term X1 .rp1 f (1) the driving
term. Equation (4.3.10) is a simple conservation equation. It is identical to the Liouville
equation for the distribution function of a system which would reduce to a unique
particle evolving in a six-dimensional phase space.6
Z
@f (1)
+ v1 .rr1 f (1) + X1 .rp1 f (1) + X12 .rp1 f (2) d⌦2 = 0. (4.3.13)
@t
5
The calculation is fairly easy to do in the case considered here, in which, in the absence of a
vector potential, ṙi is independent of ri and ṗi is independent of pi . The result may be extended to
the case where a vector potential is present. It is then necessary to take into account the appropriate
Hamilton’s equations (these equations are given in an appendix at the end of this chapter).
6
Equation (4.3.10) for the one-particle distribution function was introduced in stellar dynamics by
J.H. Jeans in 1915 under the name of the collisionless Boltzmann equation or the Liouville equation.
96 Classical systems: reduced distribution functions
The drift term and the driving term are the same as in equation (4.3.10) for the case
without interactions. The existence of pair interactions leads to the presence of an
extra term, whose expression involves the two-particle distribution function.
We can, in a similar way, derive the evolution equation of f (2) . In the presence of
pair interactions, this equation involves f (3) :
@f (2)
+ v1 .rr1 f (2) + v2 .rr2 f (2) + (X1 + X12 ).rp1 f (2) + (X2 + X21 ).rp2 f (2)
@t
Z
+ (X13 .rp1 f (3) + X23 .rp2 f (3) ) d⌦3 = 0.
(4.3.14)
In the same way, the evolution equation of f (n) involves f (n+1) . We thus get, step by
step, a chain of coupled equations, the last of which, for n = N , is nothing but the
Liouville equation for the full distribution function f (N ) = f . None of these equations,
except for the last one, is closed: we have indeed to know f (n+1) in order to be able
to determine the evolution of f (n) (for n < N ).
This set of equations, the index n varying from 1 to N , constitutes the BBGKY
hierarchy, from the names of the di↵erent physicists who established it independently:
N.N. Bogoliubov (1946), M. Born (1946), M.S. Green (1946), J.G. Kirkwood (1946),
and J. Yvon (1935). This system of chain equations, of a hierarchical structure, involv-
ing distribution functions with a reduced but increasing number of particles, is exact.7
Considered as a whole, it is equivalent to the Liouville equation for the full distribu-
tion function. As long as no approximation is made, the BBGKY hierarchy is valid
whatever the particle density (provided, however, that we remain in the framework
of pair interactions). Its major interest is to allow us to proceed to approximations
leading to truncating the hierarchy at some level (in practice, after f (1) ).
An example of such an approximation is the mean field approximation, in which
the correlations between particles are neglected. This approximation allows us to ob-
tain an evolution equation for the one-particle distribution function applicable to the
plasmas (ionized gases): this is the Vlasov equation.
Z
@f (1)
+ v1 .rr1 f + X1 + X12 f (1) (r2 , p2 , t) d⌦2 .rp1 f (1) = 0.
(1)
(4.4.2)
@t
7
It is valid even in the presence of scalar and vector potentials corresponding to applied external
fields.
Gauge invariance 97
5. Gauge invariance
Consider a system of charged particles in the presence of a scalar potential and a
vector potential describing an external electromagnetic field. The Liouville equation
and the BBGKY hierarchy have been obtained in the framework of the Hamiltonian
8
A substance whose temperature is elevated successively passes through liquid, then gaseous,
phases. When the temperature further increases, the translational kinetic energy of each particle be-
comes higher than the ionization energy of its constituent atoms. The collisions between the molecules
then put the electrons in free states and leave positively charged ions, the gas remaining globally neu-
tral. This state of matter is called a plasma. A plasma is thus composed of positively and negatively
charged particles in non-bounded states. These particles interact via Coulomb interactions. To de-
scribe a plasma, we have to introduce a couple of one-particle distribution functions, one for the
electrons, the other for the ions. We can write, for these two distribution functions, Vlasov equations
generalizing equation (4.4.2). Some phenomena occurring in out-of-equilibrium plasmas are properly
described by these equations. We can quote in particular the Landau damping of plasma waves (see
Supplement 6A).
9
In the presence of a vector potential, we also have to change the sign of the latter.
98 Classical systems: reduced distribution functions
formalism in which the dynamical variables are expressed in terms of the generalized
coordinates and momenta, and where we make use of the potentials, and not of the
fields. As a consequence, these evolution equations are not gauge-invariant.
To display the gauge invariance, it is necessary to work with the aid of gauge-
independent quantities, the dynamical variables being expressed in terms of the coor-
dinates and the velocities, and the use of the electric and magnetic fields replacing that
of the potentials. We will illustrate this point on the simple example of the Liouville
equation for the one-particle distribution function10 (the case without interactions).
This equation reads, in the Hamiltonian formalism,
@f (r, p, t)
+ v.rr f (r, p, t) + X.rp f (r, p, t) = 0, (4.5.1)
@t
where X denotes the force due to the scalar and vector potentials.11 In order to respect
gauge invariance, we must, in place of equation (4.5.1), write the evolution equation
of the one-particle distribution function F (r, v, t) of gauge-independent arguments r
and v. This latter function is defined through the identity:
We can deduce,12 from equation (4.5.1) for f (r, p, t), the evolution equation of
F (r, v, t):
@F (r, v, t) 1 ⇥ 1 ⇤
+ v.rr F (r, v, t) + q E(r, t) + v ⇥ H(r, t) .rv F (r, v, t) = 0. (4.5.4)
@t m c
⇥
In the gauge-invariant
⇤ equation (4.5.4), the driving force is the Lorentz force q E(r, t)+
v ⇥ H(r, t)/c .
This type of analysis may be extended to the evolution in the presence of pair
interactions.13
10
For the sake of simplicity, this distribution function is denoted here by f (instead of f (1) ).
11
The expression for X in terms of the scalar and vector potentials is given in an appendix at the
end of this chapter.
12
The details of the calculation are given in the appendix.
13
The gauge-invariant evolution equation, extended to the case with pair interactions, plays an
important role in electronic transport in the presence of a magnetic field (see Chapters 6 and 8).
Appendices 99
Appendices
q2
uC (r) = · (4A.1)
r
As with any function decreasing as a power law, the Coulomb potential lacks a pa-
rameter characterizing its range, which may therefore be considered as infinite.
• Screened Coulomb potential
In a gas of charged particles, the Coulomb interaction between two particles is actually
screened by the other particles. At equilibrium, the resulting e↵ective interaction is
described by the screened Coulomb potential15 uC,scr (r):
q2 e k0 r
uC,scr (r) = · (4A.2)
r
The range of uC,scr (r) is characterized by the screening length k0 1 , whose expression
depends on the character, degenerate or not, of the gas.
In a classical gas of particles of density n in equilibrium at temperature T , the
1/2
screening length is the Debye length: k0 1 = (kT /4⇡nq 2 ) . In a quantum gas at
T = 0, for instance a gas of electrons with charge q = e, the screening length
1/2
is the Thomas–Fermi length: k0 1 = (kTF /4⇡ne2 ) , where TF denotes the Fermi
temperature.16
14
We are using Gauss units.
15
This potential also appears in nuclear physics, where it is known as the Yukawa potential.
16
Note that the expression for the Thomas–Fermi length is formally similar to that for the Debye
length, the temperature T being replaced by the Fermi temperature TF .
100 Classical systems: reduced distribution functions
where (r, t) is any function of r and t, describe the same electromagnetic field as
(r, t) and A(r, t).
The evolution equation of the particle’s position operator r is the Hamilton’s
equation dr/dt = rp H1 , which reads, in virtue of formula (4B.1):
dr p qA(r, t)/c
= · (4B.4)
dt m
Thus, in the presence of a vector potential A(r, t), the relation between the momen-
tum p and the velocity v = dr/dt of a particle of mass m and charge q is:
qA(r, t)
p = mv + · (4B.5)
c
The kinetic momentum mv is a true, gauge-independent, physical quantity. Equation
(4B.5) shows that on the other hand the momentum p depends on the gauge choice.
In the presence of electric and magnetic fields, we have:
dv ⇥ 1 ⇤
m = q E(r, t) + v ⇥ H(r, t) . (4B.6)
dt c
The expression for the Lorentz force on the right-hand side of equation (4B.6) involves
the fields E(r, t) and H(r, t). The Lorentz force is thus gauge-invariant. The evolution
equation of the momentum operator p is the Hamilton’s equation dp/dt = rr H1 ,
from which we get, for instance, using formula (4B.1), the evolution equation of px :17
dpx @ (r, t) q @A(r, t)
= q + v. · (4B.7)
dt @x c @x
We finally write:
dp
= X, (4B.8)
dt
where: ⇥ ⇤ q
q
X= qr (r, t) + v ⇥ r ⇥ A(r, t) + (v.r)A(r, t) (4B.9)
c c
represents the force due to the scalar and vector potentials.
17
Formula (4B.7) can also be derived in the following way. We deduce from equation (4B.5) the
equality:
dp dv q dA(r, t)
=m + ,
dt dt c dt
which also reads:
dp dv q @A(r, t) q
=m + + (v.r)A(r, t).
dt dt c @t c
Making use of the expression (4B.6) for mdv/dt, we get:
dp h 1 i q @A(r, t) q
= q E(r, t) + v ⇥ H(r, t) + + (v.r)A(r, t),
dt c c @t c
that is: h i q
dp q
= qr (r, t) + v ⇥ r ⇥ A(r, t) + (v.r)A(r, t).
dt c c
We thus recover expression (4B.7) for dpx /dt.
102 Classical systems: reduced distribution functions
@f (r, p, t)
m3 + m3 v.rr f (r, p, t) + m3 ṗ.rp f (r, p, t) = 0. (4C.1)
@t
We then successively compute the three terms on the left-hand side of equation (4C.1).
• Term m3 @f /@t
From the equality:
p qA(r, t)/c
m3 f (r, p, t) = F r , ,t , (4C.2)
m
we deduce:
@f @F 1 q @A(r, t)
m3 = · rv F. (4C.3)
@t @t mc @t
• Term m3 v.rr f
We have the relation:
✓ ◆
@f @F 1 q @F @Ax @F @Ay @F @Az
m3 = + + , (4C.4)
@x @x mc @vx @x @vy @x @vz @x
as well as analogous formulas for @f /@y and @f /@z. The quantity m3 v.rr f can thus
be expressed in terms of F :
⇢ ✓ ◆
1 q @F @Ax @F @Ay @F @Az
v.rr F vx + +
mc @vx @x @vy @x @vz @x
✓ ◆
@F @Ax @F @Ay @F @Az
+ vy + +
@vx @y @vy @y @vz @y
✓ ◆
@F @Ax @F @Ay @F @Az
+ vz + + ·
@vx @z @vy @z @vz @z
(4C.5)
Bibliography
The statistical evolution of a classical system of N particles with pair interactions can
in principle be studied by means of the BBGKY hierarchy for the reduced distribu-
tion functions. If no approximation is made, the evolution equation of the one-particle
distribution function involves the two-particle distribution function, and so on. The
study of the evolution of the one-particle distribution function is thus tricky. How-
ever, in some cases, it is possible, through proper approximations, to obtain a closed
evolution equation for this distribution function. If the e↵ect of collisions leading to
an irreversible evolution is taken into account by these approximations, the evolution
equation thus obtained for the one-particle distribution function belongs to the class
of kinetic equations. There are several types of such equations, each of them relating
to a particular physical system placed in given conditions.
Historically, the first systems to be studied by means of a kinetic equation were
the dilute classical gases of molecules undergoing binary collisions. In this case, the
kinetic approach relies on the ‘molecular chaos hypothesis’, according to which, in the
description of a collision, the possible correlations between the velocities of the two
colliding molecules prior to their collision may be neglected. The evolution of the one-
particle distribution function is then governed by the Boltzmann equation (L. Boltz-
mann, 1872). A theorem expressing the irreversibility of the evolution, namely the
H-theorem, is closely associated with this equation. This approach of the irreversible
evolution of a dilute classical gas turned out to be extremely fruitful. It rendered
possible in particular the computation of the thermal conductivity and the viscosity
coefficient of a gas in terms of microscopic quantities such as the collision time.
As a general rule, the validity of the kinetic approaches relies on the existence
within the system of two well-separated time scales. In the case of dilute classical gases
to which the Boltzmann equation applies, the shorter time scale is the duration of a
collision, whereas the longer one is the mean time interval separating two successive
collisions of a given molecule.
106 The Boltzmann equation
r0 ⌧ d, (5.1.1)
where r0 denotes the range of the intermolecular forces and d ⇠ n 1/3 the mean
distance between the molecules (n = N/V is the gas density). Thus, the molecules in
a dilute gas are most of the time free and independent. However, the collisions, which
redistribute energy between the molecules, play an essential role in the evolution of the
gas towards equilibrium. We will study this e↵ect, taking into account binary collisions
only.1
Besides, the temperature is assumed sufficiently high and the gas density low
enough for the molecules to be conveniently represented by localized wave packets,
1/2
whose dimensions, as measured by the thermal wavelength = h(2⇡mkT ) , are
small compared to the mean intermolecular distance:
⌧ d. (5.1.2)
Any molecule of the gas may then be considered as a classical particle with well-defined
position and momentum. The molecules are treated as indistinguishable particles.
1
Collisions involving more than two molecules will be neglected, which is justified in a dilute gas.
Time and length scales 107
⌧0 ⌧ t ⌧ ⌧r , (5.2.1)
1 d3
`⇠ 2 ⇠ 2· (5.2.2)
nr0 r0
The gas being dilute, we have r0 ⌧ d and, consequently, d ⌧ `. We have thus also
r0 ⌧ `. Finally, we have to take into account a final length scale, macroscopic and
denoted L, characteristic of the linear dimensions of the box containing the gas.
2
However, we often take for the relaxation time the estimate ⌧r ⇠ ⌧ (see Chapter 6 and Chap-
ter 7).
3
The kinetic stage is followed by a hydrodynamic stage, corresponding to the evolution of f over
a much longer time interval t (⌧r ⌧ t ⌧ ⌧eq ). In the case of a dilute gas in local equilibrium, the
hydrodynamic equations can be derived from the Boltzmann equation (see Chapter 7).
4
See Chapter 7.
108 The Boltzmann equation
The distances l which come into play in the Boltzmann equation5 are interme-
diate between r0 and `:
r0 ⌧ l ⌧ `. (5.2.3)
1
f (r, p, t) = F (r, v, t), (5.3.1)
m3
with6 p = mv. The quantity f (r, p, t) drdp represents the mean number of molecules7
which, at time t, are in the phase space volume element drdp about the point (r, p).
Over a scale t, the collisions are considered as instantaneous. As for the linear
dimensions l of the spatial volume element dr, they are considered as much larger
than r0 . The collisions which take place in this volume element modify the kinetic
momentum of the molecules concerned, but they leave them inside this volume element.
Over a scale l, the collisions are thus treated as local.
Z
n(r, t) = f (r, p, t) dp. (5.3.2)
5
For longer distances l (` ⌧ l ⌧ L), the evolution of the gas is described by the hydrodynamic
equations (see Chapter 7).
6
Note that p designates from now on, not the momentum, but the kinetic momentum.
7
This definition implies that there must be a sufficient number of molecules in the spatial volume
element dr or, otherwise stated, that the inequality d ⌧ l (more restrictive than the inequality
r0 ⌧ l) be verified.
Evolution of the distribution function 109
The six-dimensional space spanned by the vector (r, p) is traditionally called the µ-
space.8 A point in this space represents a state of a molecule of the gas. At any time,
the state of a gas of N molecules is represented by N points in the µ-space. We have:
Z Z
N = n(r, t) dr = f (r, p, t) drdp. (5.3.3)
R Z
f (r, p, t)v dp 1
u(r, t) = hvi = R = f (r, p, t)v dp. (5.3.4)
f (r, p, t) dp n(r, t)
8
This terminology is due to P. and T. Ehrenfest (1911), who proposed to call -space the 6N -
dimensional phase space associated with the gas of N molecules and µ-space the six-dimensional
phase space corresponding to an individual molecule.
9
The only interactions considered here being pair interactions, it is also possible to derive the
Boltzmann equation from the BBGKY hierarchy by truncating it in an appropriate way.
110 The Boltzmann equation
where the right-hand side or collision term (@f /@t)coll represents the rate of change
of f due to collisions.
We usually write the collision term in a form displaying its balance structure,
✓ ◆ ✓ ◆(+) ✓ ◆( )
@f @f @f
= , (5.4.3)
@t coll @t coll @t coll
(+) ( )
where (@f /@t)coll is called the entering collision term and (@f /@t)coll the leaving col-
( ) (+)
lision term. The quantity (@f /@t)coll drdpdt (resp. (@f /@t)coll drdpdt) represents the
mean number of molecules undergoing a collision between times t and t + dt, one
of these two molecules finding itself, before (resp. after) the collision, in the volume
element drdp around the point (r, p) of the µ-space.
We now have to specify the form of the entering and leaving collision terms in the
case of a dilute gas of molecules undergoing binary collisions.
5. Binary collisions
The collision term (@f /@t)coll contains all the physics of the collisions. We will make
its expression precise under the following assumptions:
(i) Only binary collisions are taken into account, which is justified provided that the
gas is dilute enough.
(ii) The molecules are assumed to be devoid of internal structure. In other words, they
are supposed to be monoatomic (this implies the elasticity of collisions, any energy
transfer towards internal degrees of freedom being precluded).
(iii) The collisions are considered as local and instantaneous (in particular, any possible
e↵ect of the external force F on the collisions is neglected).
(iv) Finally, in the description of a collision between two molecules, any possible
correlations between their velocities prior to the collision are neglected. This latter
approximation is traditionally called the molecular chaos hypothesis. It was formulated
by L. Boltzmann in 1872 under the name of Stosszahlansatz. This assumption, which
is justified provided that the gas density is sufficiently low, plays a crucial role in the
derivation of an irreversible kinetic equation.
Introducing the total kinetic momentum before the collision ⇧ = p + p1 and the
relative10 kinetic momentum before the collision ⇡ = 12 (p1 p), as well as the cor-
responding quantities ⇧ 0 and ⇡ 0 after the collision, we can rewrite the conservation
equations (5.5.1) and (5.5.2) in the equivalent form:
⇧ = ⇧0 (5.5.3)
and:
|⇡| = |⇡ 0 |. (5.5.4)
As it is elastic, the collision produces a rotation of ⇡ which brings it on ⇡ 0 without
changing its modulus. The collision is fully determined by the knowledge of ⇧ and ⇡,
as well as of the angles (✓, ), called the scattering angles, of ⇡ 0 with respect to ⇡.
The problem is equivalent to that of the scattering of a molecule by a fictitious
fixed force center, represented by the point O in Fig. 5.1. The molecule comes near O
with a kinetic momentum ⇡ and an impact parameter b. As |⇡ 0 | = |⇡|, the final state is
made precise by the knowledge of both scattering angles ✓ and , collectively denoted
by ⌦.
π'
dφ
π π'
b φ
θ
O
db
The sole knowledge of the initial kinetic momenta p and p1 does not suffice to
fully determine the collision, since the impact parameter is not known precisely. The
knowledge of p and p1 defines a class of collisions with various impact parameters,
and thus various scattering angles. We generally describe this class of collisions by
imagining a beam of particles with initial kinetic momentum ⇡, uniformly spread in
space, incident on the force center O. The incident flux is defined as the number of
molecules per unit time crossing the unit area perpendicular to the incident beam. By
definition of the di↵erential scattering cross-section (⌦), the number of molecules per
unit time deflected in a direction pertaining to the solid angle element d⌦ is equal to
the product of the incident flux and (⌦)d⌦.
10
The mass associated with the ‘relative particle’ is the reduced mass, which here equals m/2
since both particles are identical.
112 The Boltzmann equation
The di↵erential cross-section (⌦) can be measured directly. It can also be computed
provided that the pair interaction potential is known.11
The interactions between the molecules being of electromagnetic origin, the micro-
scopic equations of motion and, consequently, the scattering cross-section possess the
following invariance properties:
(i) time-reversal invariance (t ! t):
In the following, we will be interested in the inverse collision {p0 , p01 } ! {p, p1 },
obtained from the original collision {p, p1 } ! {p0 , p01 } by interchanging the initial and
final states. The cross-section for the inverse collision is (p0 , p01 |p, p1 ). Making suc-
cessive use of the time-reversal invariance (formula (5.5.7)) and of the space inversion
invariance (formula (5.5.8)), we obtain the microreversibility relation:
11
In contrast to the notion of impact parameter, the notion of scattering cross-section keeps a
sense in quantum mechanics. The calculation of (⌦) for binary collisions in a dilute gas must be
carried out quantum-mechanically (and not by using the classical formula (5.5.5)). Indeed, although
the molecules may be considered as classical particles between collisions (the thermal wavelength
being very small compared to the mean distance d between molecules), this is not the case over spatial
scales comparable to the range of the scattering potential (we cannot say ⌧ r0 ). Over such spatial
scales, the notion of trajectory loses its significance and the analysis in terms of impact parameter is
not correct.
The Boltzmann equation 113
We are looking for an explicit expression for the entering and leaving collision
( ) (+)
terms (@f /@t)coll and (@f /@t)coll , the molecular chaos hypothesis being taken into
account. According to this assumption, in a spatial volume element dr centered at r,
the mean number of pairs of molecules with kinetic momenta in the elements dp
centered at p and dp1 centered at p1 can be written, at time t, in the factorized form:
⇥ ⇤ ⇥ ⇤
f (r, p, t) drdp ⇥ f (r, p1 , t) drdp1 . (5.6.2)
The number of collisions of the type {p, p1 } ! {p0 , p01 } taking place on the molecule
of interest between times t and t + dt is:
where (⌦) stands for (p, p1 |p0 , p01 ). We get, taking into account all the molecules of
the considered type (in number f (r, p, t) drdp), and integrating over p1 :
✓ ◆( ) Z Z
@f
drdpdt = f (r, p, t) drdp dp1 d⌦ (⌦) dt |v v1 |f (r, p1 , t). (5.6.5)
@t coll
The number of such collisions taking place on the considered molecule between times
t and t + dt is:
f (r, p01 , t) dp01 |v 0 v10 | 0 (⌦) d⌦dt, (5.6.8)
(+)
where 0 (⌦) stands for (p0 , p01 |p, p1 ). The rate of change (@f /@t)coll thus verifies the
equality:
✓ ◆(+) Z Z
@f
drdpdt = dp01 d⌦ 0
(⌦) dt |v 0 v10 |f (r, p0 , t) drdp0 f (r, p01 , t), (5.6.9)
@t coll
obtained by taking into account all the molecules of the considered type (in number
f (r, p0 , t) drdp0 ), and integrating over p01 . From equation (5.6.9), it follows:
✓ ◆(+) Z Z
@f
dp = dp01 d⌦ 0
(⌦)|v 0 v10 |f (r, p0 , t) dp0 f (r, p01 , t). (5.6.10)
@t coll
The di↵erential cross-sections (⌦) and 0 (⌦), which refer to collisions inverse of
each other, are equal. Since we also have the equalities:
|v v1 | = |v 0 v10 |, (5.6.11)
(+)
we get from formula (5.6.10) the expression for (@f /@t)coll :
✓ ◆(+) Z Z
@f
= dp1 d⌦ (⌦)|v v1 |f (r, p0 , t)f (r, p01 , t). (5.6.13)
@t coll
In formula (5.6.14), (⌦) represents the di↵erential scattering cross-section for the
collision {p, p1 } ! {p0 , p01 } as well as for the inverse collision {p0 , p01 } ! {p, p1 }. We
have introduced the following condensed notations:
8
>
> f = f (r, p, t)
>
>
>
>
< f1 = f (r, p1 , t)
(5.6.15)
>
> f 0 = f (r, p0 , t)
>
>
>
>
: 0
f1 = f (r, p01 , t).
Importing this form of the collision term into the evolution equation (5.6.1), we
obtain the Boltzmann equation (1872):
Z Z
@f
+ v.rr f + F .rp f = dp1 d⌦ (⌦)|v v1 | f 0 f10 f f1 . (5.6.16)
@t
In equation (5.6.17), the quantity W (p0 , p01 |p, p1 ) = 4 (⌦)/m2 is proportional to the
probability per unit time for two molecules of kinetic momenta p and p1 having,
after their collision, kinetic momenta p0 and p01 (the factor 1/2 takes into account the
indistinguishability of the molecules with kinetic momenta p0 and p01 ).
to the collisions aa, bb, and ab (or ba). Arguments similar to those previously developed
lead to Boltzmann equations for each of the distribution functions fa and fb . These
equations may be written, using condensed notations of the type of those of formula
(5.6.15),
Z Z
@fa
+ v.rr fa + F .rp fa = dp1 d⌦ aa (⌦)|v v1 | fa0 fa1 0
fa fa1
@t
Z Z
+ dp1 d⌦ ab (⌦)|v v1 | fa0 fb1 0
fa fb1 ,
(5.6.18)
and:
Z Z
@fb
+ v.rr fb + F .rp fb = dp1 d⌦ bb (⌦)|v v1 | fb0 fb1
0
fb fb1
@t
Z Z
+ dp1 d⌦ ab (⌦)|v v1 | fb0 fa1
0
fb fa1 .
(5.6.19)
In the particular case in which the b molecules are much heavier than the a
molecules (mb ma ) and where the collisions involving only a molecules may be
neglected, the system reduces to a gas of non-interacting particles undergoing collisions
with scattering centers which may be considered as fixed. Such a system is known as
a Lorentz gas.13
7. Irreversibility
The Boltzmann equation is not time-reversal invariant. If we change t into t in
equation (5.6.16), we must simultaneously change the sign of the velocities. The left-
hand side then changes sign, whereas the collision integral on the right-hand side is
not modified. Thus, f (r, p, t) does not obey the same equation as f (r, p, t). In
other words, the dynamics described by the Boltzmann equation is irreversible: the
Boltzmann equation belongs to the class of kinetic equations.
The determinant assumption in this respect in the derivation of the Boltzmann
equation is the molecular chaos hypothesis. This assumption, which amounts to ne-
glecting any correlation between the velocities of two molecules preparing to collide,
means that no molecule undergoing a collision carries information about its previous
encounters. The memory of the dynamic correlations due to the foregoing collisions is
thus lost before a new one takes place. This implies in particular that there are two
well-separated time scales, namely the duration of a collision, on the one hand, and
13
The evolution equation of the one-particle distribution function of the Lorentz gas is derived in
Supplement 5A via a direct calculation. It is also possible to deduce it from the Boltzmann equation
(5.6.18) for fa by a limiting procedure: the collision term in aa must be dropped, and the term in
ab evaluated by setting fb (r, p, t) drdp = ni dr, ni being the spatial density of the fixed scattering
centers.
The H-theorem 117
the mean time interval separating two successive collisions of a given molecule, on the
other hand. This separation of time scales is expressed through the inequality:
⌧0 ⌧ ⌧. (5.7.1)
The molecular chaos hypothesis introduces a clear-cut distinction between the event
‘before a collision’ and the event ‘after a collision’. This distinction is at the source of
irreversibility in the Boltzmann equation.14
It is worthwhile to underline the fundamental di↵erence between the Boltzmann
equation for dilute classical gases and the Vlasov equation for plasmas. The Vlasov
equation is applicable to a transitory stage of the system’s evolution: it is reversible
and does not belong to the class of kinetic equations (it is sometimes qualified as
‘collisionless’). As for the Boltzmann equation, it describes a statistical evolution in
which the collisions are taken into account: it is an irreversible evolution equation. The
information about the two-particle dynamics intervenes in the Boltzmann equation via
a quantity of a statistical nature, namely the scattering cross-section.
8. The H-theorem
8.1. The Boltzmann’s H-functional
The Boltzmann’s H-theorem concerns the evolution of the entropy of an out-of-
equilibrium dilute classical gas, such as it can be deduced from the Boltzmann equa-
tion. The functional H(t) was defined by Boltzmann as:
Z
H(t) = f (r, p, t) log f (r, p, t) drdp, (5.8.1)
where the one-particle distribution function f (r, p, t) obeys the Boltzmann equation.
We associate with H(t) an entropy SB (t), the Boltzmann entropy, defined by:
Z h i
SB (t) = k f (r, p, t) 1 log h3 f (r, p, t) drdp. (5.8.2)
The Boltzmann entropy is associated with a state of the gas characterized by the dis-
tribution function f (r, p, t) with f (r, p, t) a solution of the Boltzmann equation. It
generally di↵ers from the thermodynamic equilibrium entropy S and from the entropy
S(t) in a local equilibrium state (except when the gas finds itself either in thermody-
namic equilibrium or in a local equilibrium state).
The Boltzmann entropy and the H(t) functional are equal up to an additive
constant:
SB (t) = kH(t) + Cste . (5.8.3)
14
Some questions linked to the understanding of the way in which irreversibility is taken into
account in the Boltzmann equation were the subject of strong controversies in the nineteenth century.
The objections to Boltzmann’s ideas were formulated in two paradoxes, which are presented and
discussed in Supplement 5B.
118 The Boltzmann equation
We have: Z
@⌘ @
= f log f dp, (5.8.5)
@t @t
that is: Z
@⌘ @f
= 1 + log f dp. (5.8.6)
@t @t
To evaluate expression (5.8.6) for @⌘/@t, multiply both sides of the Boltzmann equation
(5.6.16) by (1 + log f ), and integrate over p. The derivative @⌘/@t appears as the sum
of three contributions, respectively due to the drift term, the driving term, and the
collision term.
The contribution of the drift term to @⌘/@t is given by:
Z Z
p
1 + log f v.rr f dp = rr . f log f dp. (5.8.7)
m
The distribution function vanishes when the kinetic momentum becomes infinite: the
contribution (5.8.9) thus vanishes.15
We can therefore write the following local balance equation:16
@⌘
+ r.JH = H, (5.8.10)
@t
15
This result can be extended to the case where there is an external magnetic field.
16
Since there is no ambiguity, rr is simply denoted by r.
The H-theorem 119
is the contribution of the collision term to @⌘/@t. The associate global balance equation
reads: Z Z
dH(t)
+ (r.JH ) dr = H dr. (5.8.12)
dt V V
To study H , we first note that this quantity remains invariant under the permu-
tation of the kinetic momenta p and p1 (or, accordingly, of the velocities v and v1 ).
We can therefore write:
Z
H = 1 + log f1 f 0 f10 f f1 (⌦)|v v1 | dpdp1 d⌦. (5.8.13)
We can take for H the half-sum of the expressions (5.8.11) and (5.8.13), that is:
Z
1 h i
H = 2 + log f f1 f 0 f10 f f1 (⌦)|v v1 | dpdp1 d⌦. (5.8.14)
2
We then carry out the change of variables (p, p1 ) ! (p0 , p01 ), which amounts to con-
sidering the collision {p0 , p01 } ! {p, p1 }, inverse of the collision {p, p1 } ! {p0 , p01 }.
The cross-sections relative to both collisions are identical. Also, we have the equalities:
|v v1 | = |v 0 v10 |, (5.8.15)
We therefore have:
Z
1 h i
H = 2 + log f 0 f10 f f1 f 0 f10 (⌦)|v v1 | dpdp1 d⌦. (5.8.17)
2
Taking the half-sum of the expressions (5.8.14) and (5.8.17) for H , we finally obtain:
Z ✓ ◆
1 f f1
H = log f 0 f10 f f1 (⌦)|v v1 | dpdp1 d⌦. (5.8.18)
4 f 0 f10
The quantity log(f f1 /f 0 f10 ) f 0 f10 f f1 being only negative or vanishing, the same
property holds for H .
Let us now come back to the globalR balance equation (5.8.12). For a gas contained
in
R a box of volume V , the integral V
(r.JH )dr vanishes and dH(t)/dt reduces to
V H dr. From there comes the H-theorem, established in 1872 by L. Boltzmann:
dH(t)
0. (5.8.19)
dt
120 The Boltzmann equation
dSB (t)
0. (5.8.20)
dt
The Boltzmann entropy cannot decrease over the course of time. It can only either
increase or remain stationary.
8.3. Discussion
The physical origin of the increase of the Boltzmann entropy has to be found in
the collisions between the gas molecules. The increase of SB (t) results from the way
in which the colllisions are taken into account in the Boltzmann equation via the
molecular chaos hypothesis.
The role of collisions as a source of entropy finds its origin in molecular chaos,
which constantly destroys information. After a collision between two molecules, their
positions and velocities are actually correlated.17 However, each molecule will then
undergo collisions with other molecules. After a number of collisions, the correlations
between the two molecules first considered will fade away. This introduces an irre-
versibility in time.
9. Equilibrium distributions
The definition of SB (t) allows us to show that it is a bounded quantity. Besides, we
have dSB (t)/dt 0. The Boltzmann entropy must thus tend to a limit as t ! 1. In
this limit, we have:
dSB (t)
= 0. (5.9.1)
dt
Equation (5.9.1) admits several types of solutions, the global macroscopic equilibrium
distribution, on the one hand, and local equilibrium distributions, on the other hand.
• Global equilibrium distributions
The global equilibrium distribution f0 is the time-independent solution of the Boltz-
mann equation (@f0 /@t = 0). The global macroscopic equilibrium is the thermody-
namic equilibrium. The gas at equilibrium possesses well-defined values of the density n
of molecules, of their mean velocity u, and of the temperature T .
• Local equilibrium distributions
Equation (5.9.1) also admits solutions which do not obey the Boltzmann equation.
These distributions, denoted f (0) (r, p, t), are local equilibrium distributions. They
correspond to well-defined values of the local density n(r, t) of molecules, of their
local mean velocity u(r, t), and of the local temperature T (r, t).
17
Indeed, were their velocities reversed, both molecules would have to undergo a collision.
Global equilibrium 121
The time scales necessary for the establishment of these two types of equilib-
rium are not the same. The collisions begin by locally uniformizing the density, the
mean velocity, and the temperature, but inhomogeneities continue to be present on a
larger scale. In the absence of external constraints, these inhomogeneities eventually
disappear. The collision term alone controls the rapid relaxation towards a local equi-
librium, whereas the combined e↵ect of the collision term, on the one hand, and of
the drift and driving terms, on the other hand, controls the slow relaxation towards
global equilibrium.
that is:
log f0 (p) + log f0 (p1 ) = log f0 (p0 ) + log f0 (p01 ). (5.10.3)
Equation (5.10.3) is a conservation law relative to the collision {p, p1 } ! {p0 , p01 }. It
admits solutions of the form:
log f0 (p) = (p), (5.10.4)
where (p) is a collisional invariant.18 Equation (5.10.3) being linear, its most general
solution is a linear combination of the independent collisional invariants (namely, the
mass, the three components of the kinetic momentum, and the kinetic energy), which
can be formulated in the following way:
.
122 The Boltzmann equation
3/2
the density of molecules is n = C(⇡/A) .
• Mean kinetic momentum of a molecule
From the formula: Z
1
hpi = f0 (p)p dp, (5.10.8)
n
we deduce hpi = mu. In the absence of any global translation motion of the gas, we
thus have u = 0.
• Mean kinetic energy
The mean kinetic energy of a molecule is:19
Z
1 p2
h"i = f0 (p) dp. (5.10.9)
n 2m
1
We get, for u = 0, h"i = 3(4Am) .
The expressions for A and C in terms of n and h"i being taken into account, the
equilibrium distribution (5.10.6) reads:
⇣ 3 ⌘3/2 ⇣ 3 ⌘
f0 (p) = n exp p2 · (5.10.10)
4⇡h"im 4h"im
The global equilibrium described here is simply the thermodynamic equilibrium. The
mean kinetic energy of a molecule is thus equal to its equipartition value 3kT /2. We
therefore have:
⇣ p2 ⌘
f0 (p) = n(2⇡mkT )
3/2
exp · (5.10.11)
2mkT
The distribution f0 (p), called the Maxwell–Boltzmann distribution, not only makes
the collision integral (5.6.14) vanish, but furthermore is a solution of the Boltzmann
19
In formula (5.10.9), we set p = |p|.
Local equilibrium 123
where f0 (p) is given by formula (5.10.11) (provided that there is no global translational
motion). The distribution (5.10.13) actually makes the integrand in the expression
(5.8.18) for H vanish: indeed, log f0 (r, p) is a collisional invariant, since log f0 (p)
is such an invariant and the collisions are considered as local. This distribution also
makes the left-hand side of the Boltzmann equation (formulas (5.6.16) or (5.6.17))
vanish. It is thus actually the global equilibrium distribution.
deduced from the general formula (5.8.2) for SB (t). From the expression (5.10.11) for
f0 (p), we get: ✓ ◆
V 5
S = N k log + · (5.10.15)
N 3 2
Formula (5.10.15) is the Sackur–Tetrode formula for the equilibrium entropy of an
ideal classical gas.
2
⇥ ⇤ 3/2 |p mu(r, t)|
f (0)
(r, p, t) = n(r, t) 2⇡mkT (r, t) exp · (5.11.1)
2mkT (r, t)
124 The Boltzmann equation
but: ✓ ◆
@
+ v.rr + F .rp f (0) (r, p, t) 6= 0. (5.11.3)
@t
The entropy S(t) is solely defined in a local equilibrium regime, that is, for times t
much longer than the relaxation time ⌧r . Since the distribution f (0) is not a solution of
the Boltzmann equation, there is no ‘H-theorem’ which would state that S(t) should
increase over the course of time. This is nevertheless the case for times t ⌧r , since
the entropy production within the gas in local equilibrium is necessarily non-negative.
It can be shown that the Boltzmann entropy, the local equilibrium entropy, and
the thermodynamic equilibrium entropy obey the inequalities:
which express the fact that the three descriptions of the gas, first, in terms of the
one-particle distribution function (Boltzmann description, entropy SB (t)), then in
terms of the hydrodynamic variables20 (local equilibrium description, entropy S(t)),
and, finally, in terms of the thermodynamic variables (global equilibrium description,
entropy S) are less and less detailed.
20
These variables are the local density n(r, t), the local mean velocity u(r, t), and the local tem-
perature T (r, t).
Bibliography 125
Bibliography
References
2. Time scales
The quantity f (r, p, t) drdp represents the mean number of particles of the Lorentz gas
which, at time t, are in the phase space volume element drdp about the point (r, p).
The discussion about the time and length characteristic scales in the Lorentz
gas is similar to that concerning dilute classical gases of particles undergoing binary
collisions. In particular, the duration ⌧0 of any collision is assumed to be much smaller
than the collision time ⌧ . Like the Boltzmann equation, the kinetic equation of the
Lorentz gas describes the evolution of the one-particle distribution function over a
time interval t intermediate between the duration ⌧0 of a collision and the relaxation
time ⌧r ⇠ ⌧ towards a local equilibrium.
1
The Lorentz gas model also applies to holes. Electrons and holes constitute approximately two
independent gases of charge carriers. We will only be interested here in the electron distribution
function (similar considerations can be made for holes).
Kinetic equation of the Lorentz gas 127
✓ ◆
@f @f
+ v.rr f + F .rp f = , (5A.4.1)
@t @t coll
where F denotes a possibly applied external force. As in the case of the Boltzmann
equation, the collision term (@f /@t)coll , which represents the rate of change of f due
to collisions, may conveniently be written in a form displaying its balance structure:
✓ ◆ ✓ ◆(+) ✓ ◆( )
@f @f @f
= · (5A.4.2)
@t coll @t coll @t coll
( ) (+)
The quantity (@f /@t)coll drdpdt (resp. (@f /@t)coll drdpdt) represents the mean num-
ber of particles undergoing a collision between times t and t+dt and finding themselves,
before (resp. after) the collision, in the volume element drdp around the point (r, p)
of the µ-space.
2
More generally, any function of the kinetic momentum modulus is conserved in the collision.
The mass and the kinetic energy are collisional invariants of the Lorentz model. On the other hand,
the kinetic momentum itself is not conserved.
128 The Lorentz gas
We are looking for an explicit expression for the collision term (5A.4.2), on account
of the hypotheses made about the collisions, and in particular of the fact that the
scattering centers are assumed to be randomly distributed in space. Their spatial
density is denoted by ni .
The number of collisions of the type {p} ! {p0 } taking place on this scattering center
between times t and t + dt is:
The number of collisions of this type taking place on the considered scatterer between
times t and t + dt is:
f (r, p0 , t) dp0 |v 0 | (p0 |p) d⌦dt, (5A.4.8)
where d⌦ is the solid angle element around p.
We can write:
dp = p2 dpd⌦, dp0 = p02 dp0 d⌦0 . (5A.4.9)
Kinetic equation of the Lorentz gas 129
(+)
We obtain the rate (@f /@t)coll by summing the contributions of the ni dr scat-
tering centers present in the volume element dr and integrating over the direction
of p0 :
✓ ◆(+) Z
@f
drdpdt = ni dr |v| dpdt d⌦0 (p0 |p)f (r, p0 , t). (5A.4.12)
@t coll
Importing this form of the collision term into the evolution equation (5A.4.1),
gives the kinetic equation of the Lorentz gas:
Z
@f ⇥ ⇤
+ v.rr f + F .rp f = ni |v| d⌦0 (p|p0 ) f (r, p0 , t) f (r, p, t) . (5A.4.15)
@t
Equation (5A.4.15) is a linear partial di↵erential equation for the one-particle distri-
bution function. Its linearity comes from the fact that, since the scattering centers
are considered as fixed and randomly distributed in space, they come into play in the
evolution equation solely via their density ni . Equation (5A.4.15) is not time-reversal
invariant. It thus describes an irreversible dynamics.3
The kinetic equation of the Lorentz gas is commonly used in the semiclassical
description of electronic transport in solids.4
3
We can derive an H-theorem for the Lorentz gas.
4
See Chapter 8 and Supplement 8A.
130 The Lorentz gas
Bibliography
References
H.A. Lorentz, Arch. Neerl. 10, 336 (1905). Reprinted in Collected Papers, Martinus
Nijho↵, Vol. 3, The Hague, 1936.
Supplement 5B
The irreversibility paradoxes
1. The paradoxes
The H-theorem led at the end of the nineteenth century to a somewhat confused situ-
ation. In spite of the fact that the Boltzmann equation had successfully been applied
to the description of numerous physical phenomena, Boltzmann’s ideas encountered
strong opposition, emanating from physicists as well as from mathematicians. Their ob-
jections were formulated as two paradoxes, the time-reversal paradox (or Loschmidt’s
paradox) and the recurrence paradox (or Zermelo’s paradox).
Both paradoxes are enunciated starting from an insufficiently precise formulation
of the H-theorem, according to which the derivative dH/dt would be negative or
vanishing at any time.1
Bibliography
References
Solving the kinetic equations is a tricky problem. The main technical difficulties of this
resolution arise from the complicated form of the collision integral. This is particularly
apparent in the case of the Boltzmann equation for dilute classical gases, whose col-
lision integral is quadratic with respect to the distribution function. Even the kinetic
equation of the Lorentz gas, in which the collision integral is linear, cannot in general
be solved in a simple way.
We are thus led to devise approximate resolution methods. One of the most com-
monly used procedures relies on the relaxation time approximation, which allows us to
have to solve, instead of an integro-di↵erential equation (if the case arises, non-linear),
a linear partial di↵erential equation. The relaxation time approximation is based on
the idea that the main e↵ect of the collisions is to produce a relaxation of the distribu-
tion towards a local equilibrium distribution, over a time of the order of the collision
time. When the amplitudes of the external forces or of the applied gradients are not
too important, the solution of the kinetic equation does not depart very much from
the local equilibrium distribution towards which it relaxes. It then becomes possible
to solve the kinetic equation at first perturbation order.
This resolution method thus relies on two successive linearizations, the relaxation
time approximation (which leads to a linearization with respect to the distribution
function), and the linearization with respect to the perturbations. This procedure
allows us to get an explicit analytic expression for the solution of the kinetic equation.
Making use of this solution, we can justify from a microscopic point of view the linear
phenomenological laws of transport and compute the transport coefficients. In this
chapter, this procedure is applied to the determination of the electrical conductivity
and the di↵usion coefficient of a Lorentz gas.
136 Transport coefficients
Amongst the approximate resolution methods of equations of the type (6.1.1), one
of the simplest and most commonly used relies on the relaxation time approximation.
This approximation is based on the physical idea that the main e↵ect of the collisions as
described by the term (@f /@t)coll is to induce a relaxation of the distribution function
towards a local equilibrium distribution f (0) (r, p, t) appropriate to the physics of the
problem.
• Lorentz gas
The kinetic momentum is not a collisional invariant: only its modulus is such an
invariant.
Accordingly, the collision integral vanishes for local Maxwell–Boltzmann distri-
butions with no mean velocity parameter2 (u = 0):
⇥ ⇤ 3/2 p2
f (0) (r, p, t) = n(r, t) 2⇡mkT (r, t) exp · (6.1.5)
2mkT (r, t)
@f f f (0)
+ v.rr f + F .rp f = · (6.1.6)
@t ⌧ (v)
The transport equation (6.1.6) is designated under the generic name of Boltzmann
equation in the relaxation time approximation. It can be applied to a gas of molecules
undergoing binary collisions,4 or to a gas of particles carrying out collisions with fixed
scattering centers (Lorentz gas).
The remainder of this chapter is devoted to the determination of transport co-
efficients using equation (6.1.6). Other applications of this equation will be brought
into play in the context of the semiclassical theory of electronic transport in solids5
(the wave vector or the energy dependence of the relaxation time is then related to
the details of the electronic collisions mechanisms6 ).
2
The collision integral of the Lorentz model even vanishes, more generally, for any local distribu-
tion depending only on the modulus of the kinetic momentum. The thermalization of the gas indeed
would imply an exchange of energy between the particles and the scatterers, which the Lorentz model
does not account for.
3
However, the collision integrals (6.1.2) and (6.1.3) vanish for any local equilibrium distribu-
tion (from the very definition of a local equilibrium distribution), whereas the approximate form
(f f (0) )/⌧ (v) of the collision term only vanishes for the particular local equilibrium distribution
f (0) under consideration.
4
See Chapter 7.
5
The semiclassical theory of electronic transport, known as the Bloch–Boltzmann theory of trans-
port, will be presented in Chapter 8.
6
See Supplement 8A.
138 Transport coefficients
Since there is no applied external force, the Boltzmann equation (6.1.6) reads:
@f f f (0)
+ v.rr f = · (6.3.4)
@t ⌧ (v)
@f (1) f (1)
+ v.rr f (0) = , (6.3.6)
@t ⌧ (v)
The distribution f (0) does not contribute to the fluxes (the corresponding integrals
vanish for symmetry reasons). To determine the contribution of f (1) , whose expression
involves the scalar product v.rr f (0) (formula (6.3.7)), we make use of the relation8
(deduced from the expression (6.3.3) for f (0) ):
1 (0) µ 1
rr f (0) = f r( ) + "r( ) . (6.3.9)
k T T
8
Since there is no ambiguity as far as they are concerned, the spatial gradients of the intensive
variables 1/T and µ/T are simply denoted by r(1/T ) and r( µ/T ) (and not by rr (1/T ) and
rr ( µ/T )).
140 Transport coefficients
which express the linear response of the particle and energy fluxes to the affinities
r( µ/T ) and r(1/T ).
1.3 1/2 ⇣ 2kT ⌘5/2 1.3.5 1/2 ⇣ 2kT ⌘7/2 1.3.5.7 1/2 ⇣ 2kT ⌘9/2
I4 = ⇡ , I6 = ⇡ , I8 = ⇡ ·
23 m 24 m 25 m
Kinetic coefficients of a Lorentz gas 141
made, we get the following microscopic expressions for the kinetic coefficients:
8
> n⌧ T
>
>
> LN N =
>
> m
>
<
5 n⌧
L = LN E = kT 2 (6.3.13)
> EN
> 2 m
>
>
>
> 35 n⌧ 2 3
>
: LEE = k T .
4 m
• Di↵usion coefficient
In the same way, the relation:
1 @µ
D= LN N (6.3.16)
T @n T
yields the expression for the di↵usion coefficient of the Lorentz gas:
kT ⌧
D= · (6.3.17)
m
• Thermal conductivity
The number of particles is conserved. The formula relating the thermal conductivity
to the kinetic coefficients is thus:
1 LEE LN N LN E LEN
= · (6.3.18)
T2 LN N
From formulas (6.3.13), we get the expression for the thermal conductivity of the
Lorentz gas:
5 nk 2 T ⌧
= · (6.3.19)
2 m
Let us now come back in more detail to the electrical conductivity and the di↵usion
coefficient, which we will directly compute by looking for the solution of the Boltzmann
equation (6.1.6) in the presence either of an electric field or of a density gradient.
142 Transport coefficients
4. Electrical conductivity
Consider a gas of particles of mass m and charge q in the presence of an applied
electric field E spatially uniform and time-independent. The medium is supposed to
be macroscopically neutral (the particle density n is thus uniform), and at a uniform
temperature T .
The considered particles are classical. These particles may be, for instance, elec-
trons in a non-degenerate semiconductor. The electrons are scattered mainly via their
collisions with impurities or other types of defects. The fixed scatterers are exterior
to the system of particles. This latter system may thus be treated as a Lorentz gas
(insofar as the collisions of the particles between themselves may be neglected with
respect to their collisions with the scattering centers).
nq 2 ⌧
= · (6.4.4)
m
However, to simplify the calculations, we will treat here the Boltzmann equation
under the hypothesis ⌧ (v) = ⌧ . The carrier density and the temperature being uniform,
the relevant local equilibrium distribution function f (0) is the Maxwell–Boltzmann
distribution characterized by the density n and the temperature T . This distribution
is just the thermodynamic equilibrium distribution f0 (p):
⇣ p2 ⌘
3/2
f0 (p) = n(2⇡mkT ) exp · (6.4.5)
2mkT
On account of the hypothesis ⌧ (v) = ⌧ , the Boltzmann equation (6.1.6) reads:
@f f f (0)
+ v.rr f + qE.rp f = · (6.4.6)
@t ⌧
@f (0)
+ v.rr f (0) = 0. (6.4.7)
@t
Equation (6.4.7) is actually verified by the distribution (6.4.5), since the density and
the temperature are uniform and time-independent. At next order, we get a linear
partial di↵erential equation for f (1) ,
@f (1) f (1)
+ v.rr f (1) + qE.rp f0 = , (6.4.8)
@t ⌧
whose uniform and stationary solution is:
comes: Z Z
J =q f0 v dp 2
q ⌧ v(E.rp )f0 dp. (6.4.11)
The first term of the expression (6.4.11) for J vanishes for symmetry reasons (there
is no current at equilibrium). The second term may be computed with the help of the
formula:
1
rp f0 = vf0 , (6.4.12)
kT
144 Transport coefficients
giving: Z
q2 ⌧
J= v(v.E)f0 dp. (6.4.13)
kT
Formula (6.4.13) may be identified with Ohm’s law J = .E. The conductivity
tensor here is proportional to the unit matrix: ↵ = ↵ . The electrical conduc-
tivity is given by:
nq 2 ⌧ 2
= hv↵ i, (6.4.14)
kT
that is, with the help of the equipartition relation hv↵2 i = kT /m:
nq 2 ⌧
= · (6.4.15)
m
Formula (6.4.15), deduced from the Boltzmann equation in the case of a velocity-
independent relaxation time, is identical to the Drude–Lorentz formula (6.4.4). How-
ever, interestingly enough, the determination of the conductivity via the Boltzmann
equation can be generalized to other laws ⌧ (v).
5. Di↵usion coefficient
In the same way, we aim here to establish directly a microscopic expression for the
di↵usion coefficient of a Lorentz gas.12
We consider a Lorentz gas at a uniform temperature T , in which a density gradient
is maintained, and we assume that there is no applied external force.
@f f f (0)
+ v.rr f = · (6.5.2)
@t ⌧
12
The following calculation equally applies to the di↵usion coefficient of an ideal classical gas
of molecules undergoing binary collisions. In this latter case, we are computing the self-di↵usion
coefficient, that is, the di↵usion coefficient of given recognizable molecules, which are sometimes
designated as tagged particles, within a gas consisting of molecules identical, except for the tagging,
to those under study. The tagged particles are proportionally small in number. We thus neglect
the collisions that they carry out between themselves. The Boltzmann equation (6.1.6) in this case
concerns the distribution function of the tagged particles.
Di↵usion coefficient 145
@f (1) f (1)
+ v.rr f (0) = , (6.5.4)
@t ⌧
whose stationary solution is:
Since the temperature is uniform, f (0) depends on r only via the density n(r). This
gives:13
@f (0)
f (1) = ⌧ v.rn. (6.5.6)
@n
This gives: Z Z
@f (0)
JN (r) = f (0)
(r, p)v dp ⌧ v (v.rn) dp. (6.5.8)
@n
The first term of the expression (6.5.8) for JN (r) vanishes for symmetry reasons. The
second one can be calculated with the aid of the relation:
@f (0) f (0)
= · (6.5.9)
@n n(r)
This gives: Z
1
JN (r) = ⌧ v(v.rn)f (0) dp. (6.5.10)
n(r)
Equation (6.5.10) may be identified with Fick’s law JN = D.rn. The tensor D
here is proportional to the unit matrix: D↵ = D ↵ . The di↵usion coefficient D is
given by:
D = ⌧ hv↵2 i, (6.5.11)
13
No confusion being possible, the spatial density gradient is simply denoted here by rn (and not
by rr n).
146 Transport coefficients
that is:
kT ⌧
D= , (6.5.12)
m
1
D= v`. (6.5.13)
3
D kT
= · (6.5.15)
µD q
Formulas (6.4.3) and (6.5.12) both involve a microscopic quantity, the collision time ⌧ ,
which has been taken as an estimation of the relaxation time towards a local equilib-
rium. This parameter does not appear in the Einstein relation (which, as a matter of
fact, can be obtained in the general macroscopic framework of the relations between
kinetic coefficients and transport coefficients).
Bibliography 147
Bibliography
N.W. Ashcroft and N.D. Mermin, Solid state physics, W.B. Saunders Company,
Philadelphia, 1976.
R. Balian, From microphysics to macrophysics, Vol. 2, Springer-Verlag, Berlin, 1992.
K. Huang, Statistical mechanics, Wiley, New York, second edition, 1987.
H.J. Kreuzer, Nonequilibrium thermodynamics and its statistical foundations,
Clarendon Press, Oxford, 1981.
F. Reif, Fundamentals of statistical and thermal physics, McGraw-Hill, New York,
1965.
H. Smith and H.H. Jensen, Transport phenomena, Oxford Science Publications,
Oxford, 1989.
References
P. Drude, Zur Elektronentheorie. I, Annalen der Physik 1, 566 (1900); Zur Elektro-
nentheorie. II, Annalen der Physik 3, 369 (1900).
Supplement 6A
Landau damping
e2 n1/3
⇢p = · (6A.1.1)
kT
Equations (6A.2.2) are invariant under time-reversal, that is, under the change t !
t (with H ! H). They cannot describe irreversible processes. Accordingly, they
cannot, in the strict sense, be qualified as kinetic equations.
3
These two fields are determined in a self-consistent way starting from the distribution functions
themselves (see Subsection 2.3).
150 Landau damping
4⇡
"(q, !) = 1 + i (q, !). (6A.2.7)
!
4
Formula (6A.2.7) can be derived by comparing the expressions for r ⇥ H in both descriptions
(formulas (6A.2.3) and (6A.2.4)).
Conductivity and electrical permittivity of a collisionless plasma 151
Assuming that the electric field is that of a plane wave of angular frequency ! and
wave vector q (formula (6A.2.6)), we look for a solution of equation (6A.3.5) varying
in the same manner. For reasons which will appear more clearly later,5 it is convenient
to attribute to the angular frequency ! a finite imaginary part ✏ > 0 and to take the
limit ✏ ! 0+ at the end of the calculations. The equation verified by f (1) is, in this
regime:
e
i(! + i✏ q.v)f (1) = E.rv f0 . (6A.3.6)
m
We get from equation (6A.3.6):
e E.rv f0
f (1) = i · (6A.3.7)
m ! + i✏ q.v
and: Z
J =e f (1) v dv. (6A.3.9)
and: Z
e2 (E.rv f0 )v
J= i lim dv. (6A.3.11)
m ✏!0+ ! + i✏ qvz
Formula (6A.3.11) displays the fact that the conductivity tensor is diagonal. It has
two transverse identical components xx = yy = T and a longitudinal component
zz = L . In other words, when E is perpendicular (resp. parallel) to q, the relation
between J and E amounts to J = T E (resp. J = L E). The dependence with
respect to q and ! of T and L reflects the non-local and retarded character of the
plasma response.
5
See Chapter 12.
6
We have set q = |q|.
Conductivity and electrical permittivity of a collisionless plasma 153
• Transverse conductivity
This is given by:
Z @f0
e2 vx @v
T (q, !) = lim (6A.3.12)
x
i dv.
m ✏!0+ ! + i✏ qvz
• Longitudinal conductivity
This is given by:
Z @f0
e2 vz @v
L (q, !) = lim (6A.3.13)
z
i dv.
m ✏!0+ ! + i✏ qvz
4⇡
"T (q, !) = 1 + i T (q, !) (6A.3.14)
!
and:
4⇡
"L (q, !) = 1 + i L (q, !). (6A.3.15)
!
4⇡ @E 1 @2E
r ⇥ (r ⇥ E) = + , (6A.3.16)
c2 @t c2 @t2
that is:
4⇡ !2
q ⇥ (q ⇥ E) = i!E E, (6A.3.17)
c2 c2
or:
!2
q ⇥ (q ⇥ E) = "(q, !)E (6A.3.18)
c2
(a formula which can also be directly obtained by eliminating H between the Maxwell
equations (6A.2.4)).
When projected onto the axes Ox, Oy, and Oz, equation (6A.3.18) leads to the
equations:
h !2 i h !2 i
q2 2
"T (q, !) Ex = 0, q2 2
"T (q, !) Ey = 0, (6A.3.19)
c c
and:
!2
"L (q, !)Ez = 0. (6A.3.20)
c2
154 Landau damping
Equations (6A.3.19) and (6A.3.20) admit two families of entirely decoupled solu-
tions. The first one corresponds to transverse waves (Ex 6= 0, Ey 6= 0, Ez = 0), defined
by:
q 2 c2
"T (q, !) = · (6A.3.21)
!2
These ‘new’ waves (that is, absent in vacuum), purely longitudinal, are called electro-
static waves on the grounds that they have no associated magnetic field (H = 0).
1/2
In formula (6A.4.4), the quantity !p = (4⇡ne2 /m) is a characteristic angular fre-
1/2
quency called the plasma angular frequency. Also, we have set v = (kT /m) .
When q ! 0, the longitudinal permittivity tends towards a real limit value:
!p2
"(q = 0, !) = 1 , (6A.4.5)
!2
which is also the limit when q ! 0 of the transverse permittivity8 "T (q, !).
Formula (6A.4.9) yields the dispersion relation of the longitudinal wave, that is, the
q-dependence of !. It has complex roots ! = ! 0 + i! 00 , which can be obtained through
successive approximations.
If we first neglect the q-depending terms, we get, at lowest order,
4⇡ne2
1 = 0, (6A.4.10)
m! 2
that is:
! = !p . (6A.4.11)
8
Indeed, in the limit q ! 0, there is no preferred direction in the system.
156 Landau damping
dW 1
= E02 <e L (q, !), (6A.4.12)
dt 2
that is, according to formula (6A.3.15):
dW ! 00
= " (q, !)E02 . (6A.4.13)
dt 8⇡ L
Using expression (6A.4.4) for "00L (q, !), we get:
dW 1 1 ! 2 /2q 2 v 2
= ! 2 !p2 e E02 . (6A.4.14)
dt 8(2⇡)
1/2 (qv)3
Formula (6A.4.14) shows that, in the case considered here (Maxwellian plasma), we
have dW/dt > 0. This phenomenon, known as Landau damping, was predicted by
L.D. Landau in 1946, and experimentally verified later.9
Landau damping is not linked to the collisions. It is in fact a resonance phe-
nomenon, entirely due to those electrons whose velocity in the propagation direction
of the wave is equal to the phase velocity !/q. These electrons are resonating particles
displacing themselves in phase with the wave. The field of the wave, stationary with
respect to these electrons, produces on them a work which does not vanish on average.
An electron with a velocity slightly higher than the phase velocity of the wave gives
up energy to it, whereas an electron with a velocity slightly smaller receives energy
from it. In the case of a Maxwellian plasma, the equilibrium distribution function of
the electrons decreases with the modulus of their velocity (formula (6A.4.1)). The bal-
ance of these energy exchanges thus corresponds to an energy transfer from the wave
towards the individual electrons, that is, to a damping of the wave.
9
This mechanism is used in laboratory plasmas to bring them to the very high temperatures
necessary for controlled thermonuclear fusion.
Bibliography 157
Bibliography
where n is the gas density and tot ⇠ r02 the total collision cross-section (r0 denotes the
range of the intermolecular forces). We deduce from it an estimation of the collision
1
However, when the mean free path is much larger than the dimensions of the vessel in which the
gas is enclosed, the gas, rarefied, cannot reach a local equilibrium state. The corresponding transport
regime is called the ballistic regime or the Knudsen regime. The collisions of the molecules with
the walls of the vessel then have an essential role. The ballistic regime plays an important role in
electronic transport in very small devices.
2
In place of the local density of molecules, we rather use in this context the local mass density
⇢(r, t) = mn(r, t) (m is the mass of one of the gas molecules). Besides, we often choose as variable,
instead of the local temperature T (r, t), the local density of internal energy per unit mass eint (r, t),
related to the local temperature by an equipartition formula:
3
eint (r, t) = kT (r, t).
2m
Local balance equations 161
time:
1
⌧⇠ · (7.1.3)
n tot v
The mean free path and the collision time vary inversely proportionally to the gas
density.3
Z ✓ ◆
@f
(r, p, t) dp = 0, (7.2.1)
@t coll
where (@f /@t)coll is the collision integral of the Boltzmann equation. The relation
(7.2.1) allows us to deduce from the Boltzmann equation a general balance theorem
applicable to any one of the collisional invariants.
(7.2.4)
Since the distribution function vanishes when |p| ! 1, the fifth term on the left-hand
side of equation (7.2.4) vanishes. That gives the following general balance theorem:6
@⌦ ↵ ⌦@ ↵ ⌦ ↵ ⌦ ↵ ⌦ ↵
n n + rr . n v n v.rr n F .rp = 0. (7.2.5)
@t @t
They must be computed with the aid of the solution f of the Boltzmann equation.
From the general balance theorem we deduce local balance equations for the mass,
the kinetic momentum, and the internal energy.7
@⇢
+ r.(⇢u) = 0. (7.2.8)
@t
The local balance equation for the mass has the form of a continuity equation with no
source term. The mass flux J = ⇢u is a convective flux.9
Equation (7.2.8) may be rewritten in the equivalent form:
⇣@ ⌘
+ u.r ⇢ + ⇢r.u = 0, (7.2.9)
@t
@(⇢u) ⇢
+ r.(⇢uu + P) = F . (7.2.12)
@t m
7
In the rest of this chapter (except when the context could lead to misinterpretation), rr will
simply be denoted by r.
8
For short, the local mass density ⇢(r, t) is denoted by ⇢, and the local mean velocity u(r, t) by u.
The same will be done, in the following, for some other quantities defined locally.
9
Since we consider here a gas with only one constituent, there is no di↵usive contribution to the
mass flux.
164 From the Boltzmann equation to the hydrodynamic equations
In equation (7.2.12), ⇢uu denotes the tensor of components ⇢ui uj and P the pressure
tensor, of components: ⌦ ↵
Pij = ⇢ (vi ui )(vj uj ) . (7.2.13)
In the presence of an external force F , the kinetic momentum is not a conserved
quantity. The evolution equation of its local density ⇢u involves, besides the flux term
⇢uu + P, the source term (⇢/m)F .
Taking into account equation (7.2.8), we can rewrite equation (7.2.12) in the
equivalent form:
⇣@ ⌘ 1 1
+ u.r u = F r.P. (7.2.14)
@t m ⇢
1 @⌦ 2↵ 1 ⌦ 2↵ 1 ⌦ 2↵
⇢|v u| + r. ⇢v|v u| ⇢ v.r|v u| = 0. (7.2.15)
2 @t 2 2
@(⇢eint )
+ r.(JQ + ⇢eint u) = P : ru. (7.2.20)
@t
The internal energy is not a conserved quantity. The evolution equation of its density
contains the source term P : ru. The internal energy flux is the sum of the heat
flux JQ (conductive flux) and the convective flux ⇢eint u.
The Chapman–Enskog expansion 165
Taking into account equation (7.2.8), we can rewrite equation (7.2.20) in the
equivalent form:
⇣@ ⌘ 1 1
+ u.r eint + r.JQ = P : ru. (7.2.21)
@t ⇢ ⇢
The averages of any functions of p computed with the aid of a distribution of the type
(7.3.1) are determined by the local thermodynamic variables. This is in particular
the case of the fluxes. The relations between the fluxes and the gradients of the local
thermodynamic variables are the phenomenological laws of transport, also called the
constitutive equations.
10
The normal solutions constitute a particular class of solutions of the Boltzmann equation. We
thus cannot force them to obey arbitrary conditions, either initial or boundary. For instance, near
a physical boundary such as the wall of a vessel, the distribution function is not in general of the
normal form. Similarly, immediately after a given initial condition, the distribution function is not of
the normal form either.
166 From the Boltzmann equation to the hydrodynamic equations
themselves averages over the kinetic momentum computed with the aid of f :
8 Z
>
> n(r, t) = f (r, p, t) dp
>
>
>
>
>
> Z
< 1
u(r, t) = f (r, p, t)v dp (7.3.2)
>
> n(r, t)
>
>
>
> Z
>
> 1 m 2
: kT (r, t) = f (r, p, t) |v u(r, t)| dp.
n(r, t) 3
In the method of Chapman and Enskog, we look for the normal solutions in the form
of an expansion of the type:
1 (0)
f= (f + ⇠f (1) + ⇠ 2 f (2) + · · ·), f (1) ⌧ f (0) , f (2) ⌧ f (1) . . . (7.3.3)
⇠
The conditions (7.3.4) allow us to define unambiguously the local equilibrium distri-
bution associated with f :
2
⇥ ⇤ 3/2 |p mu(r, t)|
f (0)
(r, p, t) = n(r, t) 2⇡mkT (r, t) exp . (7.3.5)
2mkT (r, t)
We then import the expansion (7.3.3) of the normal solution into the Boltzmann
equation (7.2.2). For the sake of simplicity, we write the collision term in a condensed
The Chapman–Enskog expansion 167
form displaying its quadratic character with respect to the distribution function:11
✓ ◆
@f
= I(f |f ). (7.3.6)
@t coll
This gives:
✓ ◆
1 @ (0) 1
+ · · · (f (0) + ⇠f (1) + · · ·) + (v.rr + F .rp ) (f (0) + ⇠f (1) + · · ·)
⇠ @t ⇠
1 1 h i
= 2 I(f (0) |f (0) ) + I(f (0) |f (1) ) + I(f (1) |f (0) ) + · · ·
⇠ ⇠
(7.3.7)
In equation (7.3.7), @ (0) /@t represents the zeroth-order approximation of the derivation
operator @/@t. Since f depends on t only via n(r, t), u(r, t), and T (r, t), the operator
@/@t reads:12
@ @ @n @ @ui @ @T
= + + · (7.3.8)
@t @n @t @ui @t @T @t
We compute the zeroth-order approximation @ (0) /@t of the operator @/@t by making
use of the conservation equations written at zeroth order, which amounts to taking
into account the conditions (7.3.4) imposed on f (1) . Equation (7.3.7) yields, at lowest
order (that is, by identifying the terms in 1/⇠ 2 ),
@ (0) (0)
f + v.rr f (0) + F .rp f (0) = I(f (0) |f (1) ) + I(f (1) |f (0) ). (7.3.10)
@t
Equation (7.3.9) simply confirms the fact that the distribution f (0) , which makes the
collision integral vanish, is a local equilibrium distribution. Equation (7.3.10) allows
us to determine f (1) . The procedure may be continued at higher orders, but only the
two first approximation orders are used in practice.
Even at the two first orders, the Chapman–Enskog method is fairly heavy to use
when we retain the exact bilinear structure of the collision integral. It is easier to
use when the collision term is written in the relaxation time approximation. We will
present here this simpler form of the method.
11
By definition, we set, for any functions f (r, p, t) and g(r, p, t):
Z Z
I(f |g) = dp1 d⌦ (⌦)|v v1 |[f (r, p0 , t)g(r, p01 , t) f (r, p, t)g(r, p1 , t)].
12
We use the standard convention of summation over repeated indices.
168 From the Boltzmann equation to the hydrodynamic equations
• Pressure tensor
The components of the pressure tensor defined by the general formula (7.2.13) read,
at zeroth order, Z
(0) ⇢ 2
Pij = C (vi ui )(vj uj )e A|v u| dp, (7.4.2)
n
that is, introducing the velocity U (r, t) = v u(r, t) in the reference frame moving
with the fluid: Z
(0) AU 2
Pij = m4 C Ui Uj e dU . (7.4.3)
The o↵-diagonal elements of the tensor P (0) vanish. Its diagonal elements are equal to
the local hydrostatic pressure P(r, t) = n(r, t)kT (r, t):
(0)
Pij (r, t) = ij P(r, t). (7.4.4)
• Heat flux
The heat flux defined by the general formula (7.2.17) is given, at zeroth order, by:
Z
(0) 1 4 AU 2
JQ = m C U U 2e dU . (7.4.5)
2
⇣@ ⌘ 1 1
+ u.r u = F rP. (7.4.6)
@t m ⇢
⇣@ ⌘ 2
+ u.r eint + eint r.u = 0. (7.4.7)
@t 3
⇣@ ⌘ 2
+ u.r T + T r.u = 0. (7.4.8)
@t 3
The continuity equation (7.2.8), the Euler equation (7.4.6), and the local balance
equation for the internal energy (7.4.7) (or the equation for the local temperature
(7.4.8)) form the hydrodynamic equations of a perfect fluid. Since dissipative phe-
nomena are not taken into account, the solutions of these equations correspond to
indefinitely persisting flows.
Although they have been derived here from the Boltzmann equation (thus for a
dilute gas), these equations are valid more generally. They can in fact be established
with the aid of phenomenological arguments, also valid in a denser gas or in a liquid.
We can deduce from these equations some properties of the fluids (for instance the
equation of an adiabatic transformation, the propagation equation of an acoustic wave,
the determination of the sound velocity, and so on).
⇢ ⇣ mU 2 ⌘ p
f (0) = (2⇡mkT )
3/2
exp , U= u(r, t). (7.5.3)
m 2kT m
Since f (0) depends on r and t only via the functions ⇢, T , and U , it is necessary to
compute the partial derivatives of f (0) with respect to these latter quantities:
8 (0)
> @f f (0)
>
>
> =
>
> @⇢ ⇢
>
>
< (0)
@f 1 ⇣ m 2 3 ⌘ (0) (7.5.4)
> = U f
> @T
> T 2kT 2
>
>
>
> (0)
: @f
>
=
m
Ui f (0) .
@Ui kT
We also have:
@f (0) m
= Ui f (0) . (7.5.5)
@vi kT
We deduce from these results:
h1 1 ⇣ m 2 3 ⌘ (0) m 1 i
f (1) = ⌧ (v)f (0) D(0) (⇢)+ U D (T )+ Uj D(0) (uj ) F .U ,
⇢ T 2kT 2 kT kT
(7.5.6)
with D(0) (X) = (@ (0) /@t + v.r)X. The quantities D(0) (⇢), D(0) (uj ), and D(0) (T ) are
evaluated with the aid of the zeroth-order hydrodynamic equations (formulas (7.2.8),
(7.4.6), and (7.4.8)):
8 (0)
>
> D (⇢) = ⇢r.u + U .r⇢
>
>
>
>
< (0) 1 @P 1
D (uj ) = + Fj + U .ruj (7.5.7)
> ⇢ @xj m
>
>
>
>
: D(0) (T ) = 2 T r.u + U .rT.
>
3
These expressions once imported into formula (7.5.6), give:
h1 ⇣ m 5⌘ m @uj 1 m 2 i
f (1) = ⌧ (v)f (0) (U .rT ) U2 + Ui Uj U (r.u) . (7.5.8)
T 2kT 2 kT @xi 3 kT
Introducing the symmetric tensor ⇤, of components:
m ⇣ @uj @ui ⌘
⇤ij = + , (7.5.9)
2 @xi @xj
14
For the sake of simplicity, we use the same notation f (0) for the distribution (7.3.5) and for the
function of ⇢, T , and U defined by formula (7.5.3).
The first-order approximation 171
(1)
Only the second term of the expression (7.5.10) for f (1) contributes to Pij :
Z ⇣ ⌘
(1) ⌧ m4 1
Pij = ⇤kl Ui Uj Uk Ul kl U
2
f (0) dU . (7.5.12)
kT 3
(1) P3 (1)
The tensor of components Pij is a symmetric tensor of trace i=1 Pii = 0, which
depends linearly on the symmetric tensor ⇤. It thus must be of the form:
(1) 2⌘ ⇣ m ⌘
Pij = ⇤ij ij r.u . (7.5.13)
m 3
In the constitutive equation (7.5.13), mr.u is the trace of the tensor ⇤, and ⌘ is the
viscosity coefficient.15 Identifying formulas (7.5.12) and (7.5.13), we show that the
viscosity coefficient is expressed in terms of Gaussian integrals.16 We finally obtain:
⌘ = nkT ⌧. (7.5.14)
that is: ⇣ @u
j @ui 2 @ul ⌘
Pij = ij P ⌘ + ij . (7.5.16)
@xi @xj 3 @xl
• Heat flux
(1)
The first-order heat flux, denoted JQ , only involves the first term of the expression
(7.5.10) for f (1) :
Z ⇣ m
(1) ⌧ m4 5 ⌘ 1 @T (0)
JQ = UU2 U2 Ui f dU . (7.5.17)
2 2kT 2 T @xi
(1)
JQ = .rT. (7.5.18)
5 nk 2 T ⌧
= · (7.5.19)
2 m
⇣@ ⌘ 1 1 ⌘ ⌘
+ u.r u = F rP + r2 u + r(r.u). (7.5.20)
@t m ⇢ ⇢ 3⇢
17 R1 mv 2 /2kT
The expression for involves the integrals In = 0 dv v n e with n = 6 and n = 8. We
have:
1.3.5 1/2 ⇣ 2kT ⌘1/2 1.3.5.7 1/2 ⇣ 2kT ⌘3/2
I6 = ⇡ , I8 = ⇡ .
24 m 25 m
The first-order approximation 173
⇣ ⌘
@ 2 ⌘ ⇣ 2⇤ij 2 ⌘2
⇢cv + u.r T + T (r.u) = r.(rT ) + ij r.u , (7.5.21)
@t 3 2 m 3
The first-order approximation allows us to take into account the dissipative phe-
nomena within the gas, and to obtain microscopic expressions for the viscosity coeffi-
cient and the thermal conductivity. The validity of this approximation is linked to the
smallness of f (1) as compared to f (0) , and thus in particular to the smallness of the
mean free path as compared to the typical evolution distances of the local density, the
mean local velocity, and the local temperature.
The hydrodynamic equations (7.5.20) and (7.5.21) may thus, in the case of a dilute
gas, be deduced from the Boltzmann equation. Their validity domain is actually much
larger than that of the Boltzmann equation. Accordingly, they may be established
phenomenologically in denser gases and in liquids.19
@T
⇢cp r2 T = 0. (7.5.23)
@t
18
The notation cv does not stand here for the specific heat at constant volume per particle, but
for the specific heat at constant volume per unit mass.
19
In an incompressible fluid (r.u = 0), the local balance equation for the kinetic momentum is
called the Navier–Stokes equation:
✓ ◆
@ 1 1 ⌘
+ u.r u = F rP + r2 u.
@t m ⇢ ⇢
20
The fact that it is cp , and not cv , which is involved in equation (7.5.22), comes from the fact that
the equations for the local density and the local temperature are not decoupled (See Supplement 16B).
174 From the Boltzmann equation to the hydrodynamic equations
The thermal di↵usion coefficient, also called the thermal di↵usivity, is:
Dth = · (7.5.24)
⇢cp
Equation (7.5.23) is experimentally verified in liquids (if the local mean velocity is
small compared to the sound velocity), as well as in solids.
Appendices 175
Appendices
7A. A property of the collision integral
We aim here to demonstrate the property:
Z ✓ ◆
@f
(r, p, t) dp = 0, (7A.1)
@t coll
where (@f /@t)coll is the collision integral of the Boltzmann equation and (r, p, t) a
collisional invariant.
Coming back to the definition of (@f /@t)coll for molecules undergoing binary col-
lisions, namely:
✓ ◆ Z Z
@f
= dp1 d⌦ (⌦)|v v1 | f 0 f10 f f1 , (7A.2)
@t coll
To demonstrate formula (7A.1), we first make use of the fact that the integral on the
right-hand side of formula (7A.3) is invariant under the permutation of the kinetic
momenta p = mv and p1 = mv1 , which allows us to write:
Z ✓ ◆ Z Z Z
@f 1
(r, v, t) dp = dp dp1 d⌦ (⌦)|v v1 |( + 1 )(f 0 f10 f f1 ). (7A.4)
@t coll 2
We then carry out the change of variables (p, p1 ) ! (p0 , p01 ), which amounts to consid-
ering the collision {p0 , p01 } ! {p, p1 }, inverse of the collision {p, p1 } ! {p0 , p01 }. The
cross-sections relative to both collisions are identical. Besides, we have the equalities:
|v v1 | = |v 0 v10 | (7A.5)
as well as:
dpdp1 = dp0 dp01 . (7A.6)
This gives:
Z ✓ ◆ Z Z Z
@f 1
(r, v, t) dp = dp dp1 d⌦ (⌦)|v v1 |( 0
+ 1 )(f f1
0
f 0 f10 ).
@t coll 2
(7A.7)
176 From the Boltzmann equation to the hydrodynamic equations
R
Taking the half-sum of the expressions (7A.4) and (7A.7) for r, v, t)(@f @t coll
dp,
finally gives:
Z ✓ ◆ Z Z Z
@f 1
(r, v, t) dp = dp dp1 d⌦ (⌦)|v v1 |( + 1
0
1 )(f f1
0 0 0
f f1 ).
@t coll 4
(7A.8)
As is a collisional invariant, the right-hand side of equation (7A.8) vanishes, which
demonstrates the property (7A.1).
This property of the collision integral is valid for any distribution function f ,
whether or not it is a solution of the Boltzmann equation. It is a consequence of
the fact that the collisions, which modify the kinetic momenta of the molecules, are
considered as local and instantaneous.
ux = A + By, uy = 0, uz = 0, (7B.1)
where A and B are constants. We can think of this gas as made up of di↵erent layers
sliding one on top of another, as schematically depicted in Fig. 7.1.
y
u
O x
We are trying to find the friction force per unit surface sustained by the gas
situated above a given plane (represented by the dashed line in Fig. 7.1). When the
local mean velocity gradient is small, the friction force depends linearly on it. This is
Newton’s law:
@ux
= ⌘ , (7B.2)
@y
Appendices 177
where ⌘ is the viscosity coefficient of the gas. The gas above the plane loses some
x-component of its kinetic momentum in favor of the gas below the plane. The resulting
tangential friction force per unit surface area is given by:
⌦ ↵
= mn (vx ux )(vy uy ) . (7B.3)
This force identifies with the o↵-diagonal element Pxy of the pressure tensor. To
compute the average value involved in formula (7B.3), it is convenient to make use of
the distribution function F (r, v, t). This gives:
Z
= m F (r, v, t)(vx ux )(vy uy ) dv. (7B.4)
Let us now compute the viscosity coefficient with the aid of the Boltzmann equa-
tion in the relaxation time approximation.
⇣ m ⌘3/2 2
m|v u(r)|
F (0) (r, v) = n exp · (7B.5)
2⇡kT 2kT
We denote by U (r) the velocity v u(r) of the particles in the reference frame moving
with the fluid: Ux = vx ux , Uy = vy , Uz = vz . The distribution (7B.5) is actually a
function21 of U (r):
⇣ m ⌘3/2 ⇣ mU 2 ⌘
F (0) (U ) = n exp · (7B.6)
2⇡kT 2kT
The distribution F (0) depends on y through Ux . The Boltzmann equation in the re-
laxation time approximation reads:22
@F F F (0)
+ v.rr F = · (7B.7)
@t ⌧
and we look for F (1) by expanding equation (7B.7) at first order. To begin with, we
obtain the lowest-order equation:
@F (0)
= 0. (7B.9)
@t
Equation (7B.9) is actually verified by the chosen distribution F (0) (formulas (7B.5)
or (7B.6)), since the local mean velocity u(r) is time-independent. At next order, we
get a partial-di↵erential equation for F (1) ,
@F (1) F (1)
+ v.rr F (0) = , (7B.10)
@t ⌧
@F (0)
F (1) = ⌧ vy · (7B.11)
@y
The first term of the expression (7B.12) for vanishes for symmetry reasons. To
compute the second term of this expression, we make use of the equality:
which allows us to rewrite the expression (7B.11) for F (1) in the form:
m⌧ @ux (0)
F (1) = Ux Uy F . (7B.14)
kT @y
@ux
(1)
= ⌘ · (7B.16)
@y
Appendices 179
⌘ = nkT ⌧. (7B.17)
23 R1 mv 2 /2kT
The expression for ⌘ involves the integrals In = 0 dv v n e with n = 0 and n = 2. We
have:
1 1/2 ⇣ 2kT ⌘1/2 1 1/2 ⇣ 2kT ⌘3/2
I0 = ⇡ , I2 = ⇡ ·
2 m 22 m
180 From the Boltzmann equation to the hydrodynamic equations
Bibliography
dhvi hvi
m +m = eE (8.1.1)
dt ⌧
(E is an applied electric field). The sole parameter of the model is the relaxation time,
which, in Drude’s original idea, corresponded to the collision time of the electron with
the fixed ions.
Despite notable successes such as the Drude–Lorentz formula for the electrical
conductivity, the Drude model su↵ers from important drawbacks. In particular, it
does not allow us to explain the temperature dependence of the conductivity of a
normal metal1 (that is, non-superconducting). The shortcomings of the Drude model
are essentially related to the fact that the electrons are treated like a classical gas.
Between two successive collisions, the electrons are considered as free, except for their
confinement within the sample. The periodicity of the ionic array, which gives rise
to the band structure of the metal, is thus disregarded. Moreover, the physical na-
ture of the collision mechanisms accounted for by the relaxation time is not correctly
elucidated.
To remedy these deficiencies, it is necessary to take into account both the quantum
character of the electron gas and the metal band structure, and to properly describe
the collision processes. This is the aim of the semiclassical theory of transport relying
on the Bloch theorem and the Boltzmann equation.
Since the electrons are fermions, the global equilibrium distribution at tempera-
ture T is the Fermi–Dirac distribution,
1
f0 ("k ) = , (8.1.5)
e("k µ)/kT +1
perturbation potential between states |ki and |k0 i. In such a transition, the kinetic
momentum and the energy are globally conserved.
We shall now provide some indications of the structure of the collision integral,
first for impurity scattering, then for lattice vibration scattering. Note that, if two or
more independent scattering processes come into play, the collision integral is the sum
of the integrals associated with each process.
2⇡ 0 2
Wk0 ,k = |hk |Vi |ki| ("k "k0 ). (8.2.1)
h̄
To obtain the probability of a transition from state |ki to state |k0 i during the
time interval dt, given that initially the electron is in the state |ki and that the
7
We limit ourselves here to the case of a collision with a non-magnetic impurity. Accordingly, the
electron spin is not modified by the collision.
8
This symmetry property can directly be checked in the expression (8.2.1) for the transition rates
in the Born approximation, since the perturbation potential Vi is Hermitean. However, the equality
of the rates Wk0 ,k and Wk,k0 , which stems from the time-reversal invariance of the microscopic
equations of motion, is valid independently of the calculation method (thus, possibly, beyond the
Born approximation).
186 The Bloch–Boltzmann theory of electronic transport
state |k0 i is empty, we have to multiply Wk0 ,k by fk (1 fk0 ). The collision integral for
the electron–impurity collisions thus reads:
✓ ◆ Xh i
@f
= Wk,k0 fk0 (1 fk ) Wk0 ,k fk (1 fk0 ) , (8.2.3)
@t coll 0 k
✓ ◆ Z
@f V
= Wk0 ,k (fk0 fk ) dk0 . (8.2.5)
@t coll (2⇡)3
The exclusion principle therefore has no e↵ect on the structure of the collision integral
corresponding to elastic collisions (the expression for @f /@t coll would have been the
same in the absence of restrictions on the occupation of the final state9 ).
process {|ki, ⌫} ! {|k0 i, ⌫ 0 } is of the form Wk0 ,k;⌫ 0 ,⌫ (with ⌫ 0 = ⌫ ± 1 in the most
simple case in which we neglect the processes involving more than one quantum). It is
proportional to ("k0 "k + E⌫ 0 E⌫ ), which expresses the conservation of the total
energy. The collision integral thus reads:
✓ ◆ X ⇥
@f ⇤
= Wk,k0 ;⌫,⌫ 0 p⌫ 0 fk0 (1 fk ) Wk0 ,k;⌫ 0 ,⌫ p⌫ fk (1 fk0 ) , (8.2.6)
@t coll 0 0
k ,⌫,⌫
The collision integral (8.2.7) is formally analogous to the collision integral (8.2.4)
for the electron–impurity collisions. However, whereas in formula (8.2.4), Wk0 ,k is
proportional to ("k "k0 ), this is not the case for the quantity ⇤k0 ,k involved in
formula (8.2.7). This is due to the inelasticity of the electron–phonon collisions. The
symmetry of the transition rates, that is, the relation:
actually does not imply the symmetry of the quantities ⇤k0 ,k and ⇤k,k0 . Taking into
account the relation p⌫ 0 /p⌫ = e(E⌫ E⌫ 0 )/kT and the conservation of the total energy
"k0 "k + E⌫0 E⌫ = 0, gives the relation between ⇤k0 ,k and ⇤k,k0 :
This formula does not allow further simplification of the collision integral, which re-
mains a non-linear functional of the electronic distribution function.
3. Detailed balance
The global equilibrium distribution f0 ("k ) is the time-independent solution of the
Boltzmann equation (8.1.4). Besides, the distribution f0 is also a solution of the equa-
tion (@f /@t)coll = 0. Let us examine the consequences of this property, first in the
case of impurity scattering, then in the case of lattice vibration scattering.
The equality (8.3.1) is not only a global property of the considered sums, but it
is realized term by term. Indeed, we have the stronger property, called the detailed
balance relation:12
Wk,k0 f0 ("k0 ) = Wk0 ,k f0 ("k ). (8.3.2)
Formula (8.3.2) expresses the fact that the transitions equilibrate separately for each
pair of states {|ki, |k0 i}. This property is a consequence, on the one hand, of the
microreversibility property (8.2.2), and, on the other hand, of the fact that the equi-
librium distribution f0 depends solely on the electron energy, which is conserved in
the collision with an impurity.
We can check that the equality (8.3.3) is actually realized term by term:
⇤k,k0 f0 ("k0 )[1 f0 ("k )] = ⇤k0 ,k f0 ("k )[1 f0 ("k0 )]. (8.3.4)
The detailed balance relation (8.3.4) is a consequence of the relation (8.2.10) between
⇤k,k0 and ⇤k0 ,k , and of the form (8.1.5) of the equilibrium Fermi–Dirac distribution.
where ⌧ (k) is a relaxation time towards the local equilibrium described by f (0) ("k ).
This time generally depends on the wave vector k (or merely on the energy "k in an
12
The detailed balance relation is a very general property of systems at thermodynamic equilibrium
(see Chapter 12).
Electrical conductivity 189
isotropic system). The k (or "k ) dependence of the relaxation time is related to the
details of the collision mechanisms.13
The electron gas may be submitted to electric and magnetic fields, as well as to
a chemical potential gradient and a temperature gradient. Note that the electric field
and the magnetic field act very di↵erently on the electrons; indeed, the electric field
releases energy to the electrons, while the magnetic field does not. If the departures
from local equilibrium remain small, we look for a solution of the Boltzmann equation
of the form:
(1) (1)
fk ' f (0) ("k ) + fk , fk ⌧ f (0) ("k ). (8.4.3)
(1)
The term fk is of first order with respect to the perturbation created by the electric
field, the chemical potential gradient, and the temperature gradient, but it contains
terms of any order with respect to the magnetic field. In the term of equation (8.1.4)
explicitly involving the magnetic field, the distribution function f (0) does not play any
role. Indeed, f (0) depending on k via the energy "k , gives:
@f (0)
rk f (0) = h̄vk · (8.4.4)
@"k
5. Electrical conductivity
The Bloch–Boltzmann theory relies on the knowledge of the band structure. We will
consider here the simple case of electrons belonging to one band not completely filled,
the conduction band of a metal or a semiconductor. This band is described in the e↵ec-
tive mass approximation, the e↵ective mass tensor being assumed to be proportional
to the unit matrix. The dispersion law of the electronic energy reads "k = h̄2 k 2 /2m⇤ ,
the bottom of the conduction band being taken as the origin of energies. The mean
velocity of an electron in state |ki is vk = h̄k/m⇤ .
13
The microscopic calculation of the relaxation times will be carried out in Supplement 8A.
14
The calculation for the thermal conductivity and the thermoelectric coefficients is carried out
in Supplement 8B.
190 The Bloch–Boltzmann theory of electronic transport
We assume that there is an applied electric field E, spatially uniform and time-
independent, and that there is neither temperature nor chemical potential gradient,
nor applied magnetic field. In stationary regime, the linearized Boltzmann equation
(8.4.5) reads in this case:
(1)
@f (0) fk
(vk .eE) = (8.5.1)
@"k ⌧ ("k )
(we assume here that the relaxation time is a function of the electronic energy).
The relevant local equilibrium distribution is the Fermi–Dirac distribution f0 ("k ).
We thus obtain15 from equation (8.5.1):
✓ ◆
(1) @f0
fk = e⌧ (")(v.E) · (8.5.2)
@"
The electric current density J can be obtained from fk via the integral:
Z
2e
J= fk v dk. (8.5.3)
(2⇡)3
The equilibrium distribution does not contribute to the current, which reads:
Z ✓ ◆
e2 @f0
J= ⌧ (")v(v.E) dk. (8.5.4)
4⇡ 3 @"
15
From now on, the energy and the mean velocity of the electron in state |ki are simply denoted
by " and v (instead of "k and vk ).
Electrical conductivity 191
In the present model, the density of states and the integrated density of states are
respectively of the form n(") = C"1/2 and N (") = (2C/3)"3/2 . We usually rewrite
the expression (8.5.7) for the conductivity in a form analogous to the Drude–Lorentz
formula:
ne2 h⌧ i
= · (8.5.8)
m⇤
eh⌧ i
µD = · (8.5.9)
m⇤
In formulas (8.5.8) and (8.5.9), the notation h⌧ i represents an average relaxation time
defined by:
R1 @f0
⌧ (")"3/2 d"
h⌧ i = 0
R1 @f0
@"
· (8.5.10)
0
"3/2 @" d"
Let us now make precise the expression for h⌧ i in both cases of a highly de-
generate electron gas (metal) and of a non-degenerate electron gas (non-degenerate
semiconductor).
5.2. Metal
On account of the analytic form of the Fermi–Dirac function f0 ("), we can, for a
function of energy (") slowly varying over an energy interval |" µ| ' kT and
not increasing more rapidly than a power of " when " ! 1, write the Sommerfeld
expansion:
Z 1 ✓ ◆
@f0 ⇡2 2@
2
(") d" = (µ) + (kT ) + ···, kT ⌧ µ. (8.5.11)
0 @" 6 @"2 "=µ
At low temperature, the opposite of the derivative of the Fermi–Dirac function approxi-
mately acts like a delta function centered at the chemical potential at zero temperature,
that is, at the Fermi level "F :
@f0
' (" "F ), kT ⌧ "F . (8.5.12)
@"
In the case of a metal, the integrals involved in formula (8.5.10) may thus be
2
evaluated at T = 0 up to terms of order (kT /"F ) . We get approximately:
It follows that the expression for the electrical conductivity of a metal only involves
the value of the relaxation time at the Fermi level:
ne2 ⌧ ("F )
= · (8.5.14)
m⇤
We have:
@f0
=A e "
. (8.5.16)
@"
The relaxation time generally depends on the energy according to a power law
of the type ⌧ (") / "s , the value of the exponent s being related to the nature of the
collision process.16 Formula (8.5.10) reads accordingly:
R 1 s 3/2 "
" " e d"
h⌧ i / R0 1 3/2 "
· (8.5.17)
0
" e d"
We thus get:
( 52 + s)
h⌧ i / (kT )
s
, (8.5.18)
( 52 )
E = ⇢.J , (8.6.1)
where ⇢ is the resistivity tensor. The components of the tensors and ⇢ depend on
the magnetic field. Introducing the collision time18 ⌧ , we can, while remaining within
the semiclassical regime defined by the condition h̄!c ⌧ kT , distinguish two di↵erent
magnetic field regimes. If !c ⌧ 1, the electron can move along several pitches of the
cyclotron helix between two successive collisions. The magnetic field is then considered
as being strong.19 In the opposite case !c ⌧ ⌧ 1, the magnetic field is said to be weak.
dhvi hvi h 1 i
m⇤ + m⇤ = e E + hvi ⇥ H , (8.6.2)
dt ⌧ c
from which we deduce the current J = nehvi. In stationary regime, we thus obtain
the expressions for the conductivity tensor,
0 1 !c ⌧ 1
2
ne ⌧ B 1 + ! c
2 ⌧ 2 1 + ! c
2⌧ 2
C
= @ A, (8.6.3)
m⇤ !c ⌧ 1
1 + !c2 ⌧ 2 1 + !c2 ⌧ 2
and for the resistivity tensor:
0 1
⇤ 1 !c ⌧
m
⇢= 2 @ A. (8.6.4)
ne ⌧
!c ⌧ 1
17
The foundations of the quantum theory of electronic transport will be expounded in Chapter 15.
18
The notation ⌧ stands here for a typical value of the relaxation time ⌧ (").
19
Even though this regime of magnetic field is said to be strong, it allows us to remain within the
semiclassical transport domain. The two conditions h̄!c ⌧ kT and !c ⌧ 1 must be simultaneously
fulfilled, which implies that we must have h̄/⌧ ⌧ kT (see the discussion in Section 7).
194 The Bloch–Boltzmann theory of electronic transport
xy (H) = yx ( H) (8.6.5)
is verified.
In the absence of collisions, which, in the Drude model, corresponds to ⌧ infinite,
the conductivity tensor simply reads:20
0 1
2 0 !c 1
ne
= ⇤ @ A. (8.6.6)
m
!c 1 0
By analogy with the solution (8.5.2) in the presence of the electric field alone, we are
(1)
looking for an expression for fk of the form:
✓ ◆
@f0
(1)
fk = v.C(") , (8.6.8)
@"
(1)
where C(") is a vector to be determined. Introducing the expression (8.6.8) for fk in
equation (8.6.7), we obtain for the vector C the equation:
e C.v
C.(v ⇥ H) + = eE.v. (8.6.9)
⇤
m c ⌧ (")
20
The situation described here is very di↵erent from the situation without magnetic field, in which
case the conductivity = ne2 ⌧ /m⇤ becomes infinite in the absence of collisions.
Semiclassical transport in the presence of a magnetic field 195
This gives: ✓ ◆
(1) @f0 e⌧ 2 (") E.v
fk = + (!c ⇥ E).v . (8.6.12)
@" 1 + !c2 ⌧ 2 (") ⌧ (")
0⌦ ⌧ ↵ ⌦ !c ⌧ 2 ↵ 1
ne2 B 1 + !c ⌧
2 2 1 + !c2 ⌧ 2
C
= @ A. (8.6.13)
m⇤ ⌦ !c ⌧ 2 ↵ ⌦ ⌧ ↵
1 + !c2 ⌧ 2 1 + !c2 ⌧ 2
In formula (8.6.13), the symbol h. . .i has the same meaning as in formula (8.5.10). The
result (8.6.13) has the same structure as the Drude model’s result (formula (8.6.3)).
However, in contrast to the latter, it accounts for the energy dependence of the relax-
ation time.
Let us now analyze, in the semiclassical regime, the specific e↵ects due to the
presence of the magnetic field, namely, the Hall e↵ect and the transverse magneto-
resistance.
y
H O x
v
(e/c) v × H
+ + + + + + + + + + + + + +
Ey
- - - - - - - - - - - - - - -
Ex Jx
of this electric field constitutes the Hall e↵ect, discovered by E.H. Hall in 1879. The
charge carriers (electrons or holes21 ), deflected by the magnetic field in the direction
of Oy, tend to accumulate on the sides of the sample. This results in the setting-up
of an electric field along Oy which opposes the motion of the charge carriers and their
further accumulation on the sides of the sample.
We assume here that the magnetic field is weak (!c ⌧ ⌧ 1). The tensor (8.6.13)
then reads: 0 1
h⌧ i !c h⌧ 2 i
ne2 @ A.
= ⇤ (8.6.14)
m
!c h⌧ 2 i h⌧ i
The value taken in these conditions by the Hall field Ey can be obtained by writing
that the current Jy = yx Ex + yy Ey vanishes. We usually set:
Ey = Ex tan ✓H , (8.6.15)
h⌧ 2 i
✓H = ! c · (8.6.16)
h⌧ i
The measure of the Hall e↵ect is actually a measure of the component ⇢yx of the
resistivity tensor. Indeed, as Jy = 0 we have:
Ey
⇢yx = · (8.6.17)
Jx
⇢yx = RH H. (8.6.18)
1 h⌧ 2 i
RH = · (8.6.19)
nqc h⌧ i2
Let us now provide the explicit expressions for RH in a metal and in a non-
degenerate semiconductor.
• Metal
For a metal, we have:
h⌧ i = ⌧ ("F ), h⌧ 2 i = ⌧ 2 ("F ). (8.6.20)
21
We assume here that there is only one type of charge carriers, which may be either electrons
with charge q = e or holes with charge q = e.
Semiclassical transport in the presence of a magnetic field 197
1
RH = · (8.6.21)
nec
The Hall coefficient of a metal thus does not depend on any parameter other than the
electron density.
• Non-degenerate semiconductor
The sign of the Hall coefficient allows us to determine the sign of the charge carriers.
Taking for the energy dependence of the relaxation time a power law of the type
⌧ (") / "s , gives:
1 ( 52 + 2s) ( 25 )
RH = ⇥ 5 ⇤2 · (8.6.22)
nqc ( + s) 2
2
ne2 h⌧ 2 i
Jx = ⇤ h⌧ i !c2 h⌧ 3 i + !c2 Ex . (8.6.24)
m h⌧ i
The relative modification, due to the presence of the magnetic field, of the longitudinal
conductivity is of order H 2 :
2
h⌧ ih⌧ 3 i h⌧ 2 i
= !c2 2
, !c ⌧ ⌧ 1. (8.6.25)
h⌧ i
198 The Bloch–Boltzmann theory of electronic transport
1 1
h⌧ i h⌧ i
= , !c ⌧ 1. (8.6.26)
h⌧ i
The Drude model, in which the relaxation time is considered as a constant, does
not allow us to describe this e↵ect (formulas (8.6.25) and (8.6.26) with constant ⌧
both yield = 0).
h̄
⌧ "F , (8.7.2)
⌧ ("F )
a much less restrictive condition than would be inequality (8.7.1). Condition (8.7.2)
may be rewritten equivalently in the form of the Io↵e–Regel criterion:
kF ` 1. (8.7.3)
Validity limits of the Bloch–Boltzmann theory 199
In formula (8.7.3), kF is the Fermi wave vector, and ` ⇠ h̄kF ⌧ ("F )/m the elastic mean
free path22 of the electrons. When criterion (8.7.3) is verified, the Bloch–Boltzmann
theory is applicable. When disorder increases in such a way that this condition is
no longer fulfilled, quantum interference e↵ects are likely to appear and to profoundly
modify the transport properties.23 These e↵ects have been observed in systems of small
dimensions, called mesoscopic systems, in which the spectrum of electronic states is
discrete and the electronic motion coherent (which signifies that, if an electron can
propagate in the system without undergoing inelastic collisions, its wave function keeps
a definite phase).
22
The elastic mean free path concerns the collisions which do not modify the electron energy, such
as the electron–impurity collisions.
23
See Supplement 15A.
24
See Chapter 15.
25
The random phase hypothesis is the quantum analog of the molecular chaos hypothesis of the
kinetic theory of dilute classical gases (see Chapter 9).
200 The Bloch–Boltzmann theory of electronic transport
Bibliography
N.W. Ashcroft and N.D. Mermin, Solid state physics, W.B. Saunders Company,
Philadelphia, 1976.
C. Cohen-Tannoudji, B. Diu, and F. Laloë, Quantum mechanics, Vol. 2, Hermann
and Wiley, Paris, second edition, 1977.
D.K. Ferry, Semiconductor transport, Taylor & Francis, London, 2000.
Y. Imry, Introduction to mesoscopic physics, Oxford University Press, Oxford, second
edition, 2002.
W. Jones and N.H. March, Theoretical solid-state physics: non-equilibrium and
disorder , Vol. 2, Wiley, New York, 1973. Reprinted, Dover Publications, New York,
1985.
P.S. Kireev, Semiconductor physics, Mir Publishers, second edition, Moscow, 1978.
R. Peierls, Some simple remarks on the basis of transport theory, Lecture Notes in
Physics 31 (G. Kirczenow and J. Marro editors), Springer-Verlag, Berlin, 1974.
H. Smith and H.H. Jensen, Transport phenomena, Oxford Science Publications,
Oxford, 1989.
P.Y. Yu and M. Cardona, Fundamentals of semiconductors, Springer-Verlag, Berlin,
third edition, 2003.
J.M. Ziman, Electrons and phonons: the theory of transport phenomena in solids,
Clarendon Press, Oxford, 1960. Reissued, Oxford Classic Texts in the Physical Sci-
ences, Oxford, 2001.
J.M. Ziman, Principles of the theory of solids, Cambridge University Press, Cam-
bridge, second edition, 1979.
References
P. Drude, Zur Elektronentheorie. I, Annalen der Physik 1, 566 (1900); Zur Elektro-
nentheorie. II, Annalen der Physik 3, 369 (1900).
D.A. Greenwood, The Boltzmann equation in the theory of electrical conduction in
metals, Proc. Phys. Soc. London 71, 585 (1958).
W. Pauli, in Probleme der Modernen Physik , S. Hirzel, Leipzig, 1928. Reprinted in
Collected scientific papers by W. Pauli (R. Kronig and V.E. Weisskopf editors),
Interscience, New York, 1964.
Supplement 8A
Collision processes
1. Introduction
One of the aims of the study of collision processes is to elucidate the microscopic nature
of the mechanisms limiting the electronic transport and to derive, when possible, an
expression for the relaxation time associated with the collision integral.
For elastic collisions in an isotropic system, such as electron–impurity or electron–
long-wavelength-acoustic-phonon collisions, this time can be defined and microscop-
ically computed, provided that the transition rate Wk0 ,k from state |ki to state |k0 i
only depends on the modulus of k and on the angle between the vectors k and k0 .
The relaxation time is then a function of the electron energy "k , a quantity conserved
in the collision. Once the law ⌧ ("k ) is determined, the transport coefficients can be
explicitly computed.
2. Electron–impurity scattering
As previously, we consider here electrons belonging to one band not completely filled,
described in the e↵ective mass approximation. The e↵ective mass tensor is assumed
to be proportional to the unit matrix.
If the deviations from local equilibrium remain small, we look for a solution of the
(1) (1)
Boltzmann equation of the form fk ' f (0) ("k ) + fk , where fk is a correction of first
order with respect to the perturbations which make the system depart from the local
equilibrium described by f (0) ("k ). The elasticity of the collisions implying the equality
f (0) ("k ) = f (0) ("k0 ), leaves us with:
✓ ◆ Z
@f V (1) (1)
= Wk0 ,k (fk0 fk ) dk0 . (8A.2.2)
@t coll (2⇡)3
202 Collision processes
Comparing formulas (8A.2.2) and (8A.2.3), gives a formal expression for the inverse
relaxation time: ✓
Z (1) ◆
1 V fk0
= W 0
k ,k 1 dk0 . (8A.2.4)
⌧ (k) (2⇡)3 fk
(1)
However, the expression (8A.2.4) for 1/⌧ (k) involves the solution, which remains to
be determined, of the linearized Boltzmann equation.
We can get rid of this drawback by studying the situation in which the only
perturbation present is due to a uniform and time-independent applied electric field E.
The linearized Boltzmann equation reads, in this case:
(1)
@f (0) fk
(vk .eE) = · (8A.2.5)
@"k ⌧ (k)
The relevant local equilibrium distribution f (0) being the Fermi–Dirac function f0 ("k ),
we have: ✓ ◆
(1) @f0
fk = e⌧ (k)(vk .E) · (8A.2.6)
@"k
To determine ⌧ (k), we make the assumption that, the system being isotropic, the
relaxation time is a function of "k (or, which amounts to the same thing, a function of
k = |k|). This assumption will be verified at the end of the calculation. The collisions
being elastic, gives us ⌧ (k) = ⌧ (k 0 ). Since vk = h̄k/m⇤ , we deduce from formula
(1) (1)
(8A.2.6) for fk , together with its analog for fk0 , the identity:
(1)
fk0 k0 .E
(1)
= · (8A.2.7)
fk k.E
k
E k'
θ
α
O y
φ
and, therefore:
k0 .E
= cos ✓ + tan ↵ sin ✓ cos . (8A.2.10)
k.E
At the Born approximation of scattering, the transition rate Wk0 ,k reads:
2⇡ 0 2
Wk0 ,k = |hk |Vi |ki| ("k "k0 ), (8A.2.11)
h̄
where Vi denotes the electron–impurity interaction potential. The corresponding dif-
ferential scattering cross-section is:2
m⇤ 2 V 2 0 2
(✓) = 4 |hk |Vi |ki| . (8A.2.12)
4⇡ h̄
2
This gives:
(2⇡)3 h̄
Wk0 ,k = (k k 0 ) (✓). (8A.2.13)
V 2 m⇤ k
Since the rate Wk0 ,k depends only, besides k, on the scattering angle ✓ (and not on the
absolute orientations of the vectors k and k0 ), formula (8A.2.8) involves an integration
over the surface of constant energy "k . This integration once carried out, leaves us with:
Z
1 v h̄k
= (✓)(1 cos ✓ tan ↵ sin ✓ cos ) d⌦, v = ⇤· (8A.2.14)
⌧ (k) V m
In the factor (1 cos ✓ tan ↵ sin ✓ cos ), only the term (1 cos ✓) provides a non-
vanishing contribution to the above integral. We thus have:
Z
1 v
= (✓)(1 cos ✓) d⌦. (8A.2.15)
⌧ (k) V
2
With the hypotheses made for Wk0 ,k , the di↵erential cross-section depends only on ✓ (and not
on both angles ✓ and ). It is thus denoted by (✓).
204 Collision processes
In fact, the sample does not contain a unique impurity, but a concentration ni of
impurities. If these impurities are randomly distributed in space,3 the inverse relax-
ation time is obtained by adding the contributions of the various impurities:
Z
1
= ni v (✓)(1 cos ✓) d⌦. (8A.2.16)
⌧ (k)
Formula (8A.2.16) can be compared to the estimation ⌧ 1 ⇠ n totRv of the inverse re-
laxation time in a dilute classical gas (n is the gas density, tot = (⌦) d⌦ the total
scattering cross-section, and v a typical molecular velocity). In the formula (8A.2.16)
concerning electronic collisions, the factor (1 cos ✓) appearing in the integrand ac-
counts for the greater relative importance of collisions with a large scattering angle.
The quantity:
1
`tr = v⌧ (k) = R (8A.2.17)
ni (✓)(1 cos ✓) d⌦
is called the transport mean free path.4
Let us now provide the expressions for (✓) and ⌧ (k) for di↵erent types of electron–
impurity collisions.
Ze2
Vi (r) = ± , (8A.2.18)
"d r
diverges in the limit ✓ ! 0 which corresponds to a very large impact parameter. Now,
1/3
in a solid, the impact parameter has a natural upper limit bm = ni /2, equal to
half the typical distance between two impurities. A lower limit ✓min of the deviation
angle, given by the equality tan(✓min /2) = R/bm , corresponds to the upper limit bm
of the impact parameter. Taking into account this lower limit of ✓, we rewrite formula
(8A.2.16) as: Z
1 R2 ⇡ 1 cos ✓
= ni v2⇡ sin ✓ d✓. (8A.2.21)
⌧ (k) 4 ✓min sin4 (✓/2)
This gives:
Z ⇡ ⇣
1 cos ✓ ✓min ✓min ⌘
sin ✓ d✓ = 8 log sin = 4 log 1 + cot2 · (8A.2.22)
✓min sin4 (✓/2) 2 2
• Neutral impurities
For the sake of simplicity, we consider that, as far as the scattering of an electron is
concerned, a neutral impurity may be treated like a very strongly screened ionized
impurity. We therefore write:
Ze2 e k0 r
Vi (r) = ± , (8A.2.25)
"d r
where k0 1 denotes the screening length.7
The corresponding di↵erential scattering cross-section is:
✓ 2 ⇤ ◆2
Ze m 1
(✓) = 4 2 2
, (8A.2.26)
"d h̄ (k0 + K 2 )
2
where K = k0 k represents the wave vector change in the course of the scattering
process. The collision being elastic, we have:
✓
k = k0 , K = 2k sin · (8A.2.27)
2
Formula (8A.2.26) can thus be rewritten in the form:
✓ 2 ⇤ ◆2 ✓ ◆ 2
Ze m 2 ✓
(✓) = 4 k 2
+ 4k 2
sin . (8A.2.28)
"d h̄2
0
2
The divergence of the integral (8A.2.16) giving 1/⌧ (k) is automatically avoided by the
introduction of the screening length.
If the velocity is not too large and the screening strong enough, we can neglect
k 2 as compared to k02 in formula (8A.2.28). In this case, the di↵erential cross-section
does not depend on ✓ (the scattering is isotropic) or on the velocity. We have:
1
= ni v tot . (8A.2.29)
⌧ (k)
The relaxation time thus varies inversely proportionally to the velocity, that is, in
terms of energy:
⌧ (") / " 1/2 . (8A.2.30)
The above calculation relies on the hypothesis that the scattering of an electron
by a neutral impurity may be described using the screened potential (8A.2.25) in the
limit of very strong screening. However, the actual problem of scattering by a neutral
impurity is notably more complicated. A more refined calculation shows that actually
the scattering cross-section is inversely proportional to the velocity. The corresponding
relaxation time is thus energy-independent. As a result, the mobility limited by the
scattering on neutral impurities is temperature-independent.
7
The screening of a charged impurity in a semiconductor is due to the gas of charge carriers. For
a non-degenerate electron gas of density n, the screening length is given by the Debye expression:
1/2
k0 1 = (kT /4⇡ne2 ) .
Electron–phonon scattering 207
3. Electron–phonon scattering
The inelastic scattering of electrons by lattice vibrations induces a temperature de-
pendence of the electrical conductivity. To give a correct account of this e↵ect, it is
necessary to deal with the correlation of ionic motions, which can only be described
in terms of phonons.
8 1/2
The screening length is here the Thomas–Fermi screening length: k0 1 = (kTF /4⇡ne2 ) (TF
denotes the Fermi temperature).
9
In the presence of non-magnetic scattering impurities, the resistivity of a metal decreases mono-
tonically with the temperature towards a temperature-independent term, which constitutes, by def-
inition, the residual resistivity. It is the only remaining contribution to the resistivity at very low
temperatures, when scattering by lattice vibrations has become negligible.
10
This can be achieved with the aid of a phase shift method due to J. Friedel (1956).
11
We neglect here the Umklapp processes for which k0 k = q + K, where K is a reciprocal lattice
vector. These processes, whose e↵ect is much more difficult to calculate, play an important role in
some transport properties, in particular in thermal conduction (however, they do not qualitatively
modify the temperature dependence of the resistivity of normal metals). It can be shown that, when
these processes are discarded, the only phonon modes to consider are the longitudinal modes for
which the polarization vector is parallel to the wave vector.
208 Collision processes
In formula (8A.3.1), ak and a†k0 are respectively the annihilation operator of an electron
in state |ki and the creation operator of an electron in state |k0 i, whereas bq and b† q
denote respectively the annihilation operator of a phonon of wave vector q and the
creation operator of a phonon of wave vector q.
The Hamiltonian (8A.3.1) induces electronic transitions between the di↵erent
electronic states |ki. The corresponding collision integral is of the form:
✓ ◆
@f 2⇡ X 2
= |g(k, k0 )|
@t coll h̄ 0
k
n o
⇥ fk0 (1 fk ) n q ("k "k0 h̄! q ) + (1 + nq ) ("k "k0 + h̄!q )
n o
fk (1 fk0 ) nq ("k0 "k h̄!q ) + (1 + n q) ("k "k0 + h̄! q) ,
(8A.3.2)
where nq and !q denote the mean occupation number and the angular frequency of
mode q. In formula (8A.3.2), the two contributions to the entering term in fk0 (1 fk )
correspond respectively to the absorption of a phonon of wave vector q and the
emission of a phonon of wave vector q, whereas the two contributions to the leaving
term in fk (1 fk0 ) correspond respectively to the absorption of a phonon of wave
vector q and the emission of a phonon of wave vector q.
The collisions of electrons with long wavelength acoustic phonons are quasielastic,
as pictured by the inequality h̄|!q | ⌧ "k . The corresponding collision integral has the
same structure as the electron–impurity collision integral (8A.2.1). It reads:
✓ ◆
@f 2⇡ X 2
= |g(k, k0 )| (fk0 fk )(2n0q + 1) ("k0 "k ), k0 = k + q, (8A.3.3)
@t coll h̄ 0
k
with: ✓ ◆1/2
1 N h̄
g(k, k ) =
0
i q, q = qU (q). (8A.3.4)
V 2M !q
In formula (8A.3.4), N is the number of atoms in the sample, M the mass of one of
these atoms, and U (q) the Fourier transform of the electron–ion interaction potential.
We obtain for the inverse relaxation time:
1 2⇡ X 2
= (1 cos ✓)| q | (2n0q + 1) ("k+q "k ), (8A.3.5)
⌧ (k) h̄ q
12
In order to avoid any confusion with the electron wave vector modulus k, the Boltzmann constant
is denoted here by kB .
Electron–phonon scattering 209
1 4⇡ X 2 0
= (1 cos ✓)| q | nq ("k+q "k ). (8A.3.6)
⌧ (k) h̄ q
we obtain:
Z Z ⇣
1 4⇡ V 1
2⇡q 2 m⇤ ⇡
2 0 q ⌘
= dq 2 sin ↵ cos2 ↵| q | nq cos ↵ + d↵. (8A.3.9)
⌧ (k) h̄ 8⇡ 3 0 kqh̄2 0 2k
The angular integral yields a non-vanishing contribution only if q < 2k. We thus have:
Z 2k
1 m⇤ V 2
= q 3 | q | n0q dq. (8A.3.10)
⌧ (k) 2⇡h̄3 k 3 0
k+q q
θ α
k
The function U (q) tends towards a constant as q ! 0. Therefore, for the acoustic
phonon modes for which !q ⇠ cs q, where cs denotes the sound velocity, the parameter
q introduced in formula (8A.3.4) behaves proportionally to q as q ! 0. Taking as
1/2
2
integration variable x = h̄!q /kB T = h̄cs q/kB T , and setting, as q ! 0, | q | = A2 q,
we finally get the following approximate expression,
✓ ◆5 Z
1 m⇤ V kB T 2h̄ck/kB T
' A2 x4 n(x) dx, (8A.3.13)
⌧ (k) 2⇡h̄3 k 3 h̄cs 0
1
where n(x) = (ex 1) .
15
The order of magnitude of ⇥ is the same as that of the Debye temperature ⇥D = h̄cs qD /kB
(qD is the Debye wave vector). Indeed, we have kF 3 = 3⇡ 2 n, where n is the density of conduction
3 = 6⇡ 2 N/V , where N/V is the ionic density. For a metal of valence Z, we thus have
electrons, and qD
qD = (2/Z)1/3 kF .
Bibliography 211
Bibliography
N.W. Ashcroft and N.D. Mermin, Solid state physics, W.B. Saunders Company,
Philadelphia, 1976.
C. Cohen-Tannoudji, B. Diu, and F. Laloë, Quantum mechanics, Vol. 2, Hermann
and Wiley, Paris, second edition, 1977.
W. Jones and N.H. March, Theoretical solid-state physics: non-equilibrium and
disorder , Vol. 2, Wiley, New York, 1973. Reprinted, Dover Publications, New York,
1985.
P.S. Kireev, Semiconductor physics, Mir Publishers, second edition, Moscow, 1978.
L.D. Landau and E.M. Lifshitz, Mechanics, Butterworth-Heinemann, Oxford, third
edition, 1976.
H. Smith and H.H. Jensen, Transport phenomena, Oxford Science Publications,
Oxford, 1989.
P.Y. Yu and M. Cardona, Fundamentals of semiconductors, Springer-Verlag, Berlin,
third edition, 2003.
J.M. Ziman, Electrons and phonons: the theory of transport phenomena in solids,
Clarendon Press, Oxford, 1960. Reissued, Oxford Classic Texts in the Physical Sci-
ences, Oxford, 2001.
J.M. Ziman, Principles of the theory of solids, Cambridge University Press, Cam-
bridge, second edition, 1979.
References
where n(") is the density of states in energy of the electrons. More precisely, the fluxes
JN and JQ⇤
are given by the formulas:
8
> 1
< JN = K0 (eE
> rµ) K1
T
rT
(8B.2.2)
>
>
: J ⇤ = K1 (eE 1
Q rµ) K2 rT.
T
with: 8
> L11 = K0 T
>
<
L12 = L21 = K1 T (8B.2.4)
>
>
:
L22 = K2 T.
The Onsager symmetry relation L12 = L21 is therefore verified, as expected.
3. Thermal conductivity
The isothermal electrical conductivity and the open-circuit thermal conductivity are
expressed respectively as = e2 K0 and = (K0 K2 K12 )/K0 T . According to formula
(8B.2.1) for p = 0, we have:
n
K0 = ⇤ h⌧ i, (8B.3.1)
m
where: R1 @f (0)
⌧ (")"3/2 d"
h⌧ i = 0
R1
@"
@f (0)
(8B.3.2)
0
"3/2 @" d"
(we have taken as usual n(") = C"1/2 ). Similarly we have, according to formula
(8B.2.1) for p = 1 and p = 2,
n ⌦ ↵
K1 = (" µ)⌧ , (8B.3.3)
m⇤
214 Thermoelectric coefficients
and:
n ⌦ 2 ↵
K2 =⇤
(" µ) ⌧ . (8B.3.4)
m
In formulas (8B.3.3) and (8B.3.4), the symbol h. . .i has the same meaning as in formula
(8B.3.2). We deduce from these results the expression for the open-circuit thermal
conductivity:
n h h"⌧ i i
2
= ⇤ h"2 ⌧ i · (8B.3.5)
m T h⌧ i
Let us now make precise the form of the integrals K0 , K1 , K2 , and deduce from
them the expression for in a metal and in a non-degenerate n-type semiconductor.
3.1. Metal
The Sommerfeld expansion at lowest order of the integrals K0 , K1 , and K2 yields the
approximate expressions:
⇡2 2 dk0 (") ⇡2 2
K0 ' k0 ("F ), K1 ' (kT ) , K2 ' (kT ) k0 ("F ), (8B.3.6)
3 d" "="F 3
where:
2
k0 (") = "n(")⌧ ("). (8B.3.7)
3m⇤
We have K12 ⌧ K0 K2 and thus ' K2 /T . Since = e2 K0 , there is a propor-
tionality relation between and T , known as the Wiedemann–Franz law:
⇡2 k2
' · (8B.3.8)
T 3 e2
The ratio L = / T , called the Lorenz number, is thus in this model a temperature-
independent constant:2
⇡2 k2
L= · (8B.3.9)
3 e2
Relation (8B.3.8) was established empirically in 1853 for numerous metals. The value
of the Lorenz number, obtained here in an extremely simple band structure model,
does not in fact depend on it. In the general case in which the electrical conductivity
and the thermal conductivity are tensors, we can show that the components of the
same indices of these tensors are in the ratio ⇡ 2 k 2 T /3e2 .
For a relaxation time depending on the energy according to a power law of the
type ⌧ (") / "s , we get for the Lorenz number the value:3
⇣5 ⌘ k2
L= +s 2· (8B.3.10)
2 e
The ratio L = / T is not universal as it is the case in metals, but it reflects the
energy dependence of the relaxation time.
The foregoing calculation relies on the hypothesis of the existence of a unique
relaxation time governing both electrical conduction and thermal conduction. Such
a relaxation time does exist when the collision processes are elastic (which means
in practice that the energy change of an electron in the course of a collision is small
compared to kT ). The scattering of electrons by lattice vibrations obeys this condition
at high temperatures. At low temperatures, the mechanism limiting the conductivity is
the elastic scattering of electrons by impurities. The Wiedemann–Franz law is indeed
well verified at high as well as at low temperatures.
⇡2 k d log k0 (")
⌘= kT . (8B.4.3)
3 e d" "="F
With the chosen band structure, the thermoelectric power is negative. As a general
rule, its sign depends on the band structure of the material.
3
In a semiconductor, the contribution of the electrons to the thermal conductivity is small com-
pared to the contribution of the phonons. Moreover, the above calculation is oversimplified, since it
takes only the electrons into account, and not the holes. If we take into account the contribution
of the holes, a supplementary term adds to the expression (8B.3.10) for the Lorenz number in a
semiconductor.
216 Thermoelectric coefficients
1 K1
⇡= · (8B.4.5)
e K0
The microscopic expressions (8B.4.3) and (8B.4.5) for ⌘ and ⇡ verify as expected
the second Kelvin relation:
⇡ = ⌘T. (8B.4.6)
Bibliography 217
Bibliography
N.W. Ashcroft and N.D. Mermin, Solid state physics, W.B. Saunders Company,
Philadelphia, 1976.
W. Jones and N.H. March, Theoretical solid-state physics: non-equilibrium and
disorder , Vol. 2, Wiley, New York, 1973. Reprinted, Dover Publications, New York,
1985.
P.S. Kireev, Semiconductor physics, Mir Publishers, second edition, Moscow, 1978.
H. Smith and H.H. Jensen, Transport phenomena, Oxford Science Publications,
Oxford, 1989.
J.M. Ziman, Electrons and phonons: the theory of transport phenomena in solids,
Clarendon Press, Oxford, 1960. Reissued, Oxford Classic Texts in the Physical Sci-
ences, Oxford, 2001.
J.M. Ziman, Principles of the theory of solids, Cambridge University Press, Cam-
bridge, second edition, 1979.
This page intentionally left blank
Chapter 9
Master equations
This chapter is devoted to the master equations, which play a very important role in
the description of the evolution of out-of-equilibrium physical systems.
To begin with, we consider those physical phenomena likely to be modelized
by random processes. The processes which are most commonly used for this type of
modelization are the Markov processes, often called picturesquely ‘processes without
memory’. A Markov process is governed by its transition probabilities, which allow
us, step by step, to determine its evolution from its initial distribution. The transition
probabilities of a Markov process obey a non-linear functional equation, the Chapman–
Kolmogorov equation. If we have independent information about the behavior of the
transition probabilities over short time intervals, we can deduce from the Chapman–
Kolmogorov equation a linear equation, called the master equation, describing the
evolution of the transition probabilities over much longer time intervals.
Then, adopting a more physical point of view, we investigate how such an evolu-
tion equation can be established starting from the Liouville–von Neumann equation
for the density matrix. In particular, we study the evolution in the presence of a per-
turbation of a system with a very large number of degrees of freedom. If the density
matrix of the system is diagonal at the initial time, its diagonal elements obey after-
ward –provided that some hypotheses are verified– an irreversible evolution equation,
similar to the master equation for a Markov process, called the Pauli master equation.
The Pauli master equation, which is valid in the weak-coupling long-time limit,
allows us to describe the irreversible behavior of a physical system, in particular the
approach of equilibrium starting from an initial out-of-equilibrium distribution.
220 Master equations
Inversely, the joint probability pn reads, in terms of p1 and of the conditional proba-
bilities p1|1 (x2 , t2 |x1 , t1 ), . . . , p1|n 1 (xn , tn |x1 , t1 ; . . . ; xn 1 , tn 1 ):
Equation (9.1.4) allows us to obtain the distribution of X(t) at a time t2 > t1 given
the distribution of X(t) at time t1 , provided that the elementary conditional probabil-
ity p1|1 (x2 , t2 |x1 , t1 ) is known. However, the elementary conditional probability may
depend on the past history of the process. In such a case, equation (9.1.4) is of no
practical use.
1
For convenience, the probability densities (be they conditional or joint densities) are simply here
called probabilities.
Markov processes: the Chapman–Kolmogorov equation 221
otherwise stated, if the conditional probability p1|n 1 only depends on the two most recent past values
of the variable, the process X(t) is a second-order Markov process. Higher-order Marlov processes may
be defined in an analogous way. The essential characteristic of a Markov process (whatever its order)
is that the memory of its past history does not persist indefinitely, but eventually disappears. The
first-order Markov processes being the only ones considered here, they will be simply called Markov
processes (with no explicit mention of their order).
222 Master equations
All the joint probabilities are thus determined if we know the probability p1 and the
elementary conditional probability p1|1 , called transition probability. The transition
probability of a Markov process is independent from the past history of the process.
Equation (9.1.4) actually allows us, in this case, to determine p1 (x2 , t2 ) given p1 (x1 , t1 )
(t2 > t1 ), and thus to establish the evolution equation of p1 (x, t).
Amongst the Markov processes, the stationary processes play a particularly im-
portant role. For a Markov process to be stationary, it is necessary and sufficient that
p1 be time-independent and that p1|1 only depend on the time interval involved. Most
often, for a stationary process, the probability p1 represents the distribution which is
attained after a sufficiently long time ⌧ , whatever the initial state x0 . In this case, we
have:
p1 (x) = lim p1|1 (x, ⌧ |x0 ). (9.1.9)
⌧ !1
The process is then entirely defined by the knowledge of the transition probability.
(9.1.10)
In the case of a Markov process, the definition (9.1.7) being taken into account, identity
(9.1.10) involves only p1|1 . It reads:
Z
p1|1 (x3 , t3 |x1 , t1 ) = p1|1 (x2 , t2 |x1 , t1 )p1|1 (x3 , t3 |x2 , t2 ) dx2 , (9.1.11)
with t1 < t2 < t3 . Equation (9.1.11) expresses a necessary condition for a random
process to be Markovian. This functional equation, which expresses a constraint that
the transition probability of a Markov process must obey, is known under the name of
Chapman–Kolmogorov equation, or of Smoluchowski equation.4 It was established by
M. Smoluchowski in 1906, then by S. Chapman in 1916 and A. Kolmogorov in 1931.
A specific Markov process is fully determined by p1 and p1|1 . These functions must
obey the Chapman–Kolmogorov equation (9.1.11), as well as the consistency condition
(9.1.4). Two non-negative functions p1 and p1|1 which verify these two conditions
uniquely define a Markov process.
4
For instance it is under this latter name that it is referred to in the context of Brownian motion
(see Chapter 11).
Master equation for a Markovian random process 223
We assume that, for t ⌧ ⌧ (the time scale ⌧ remaining to be defined from a physical
point of view), the behavior of p1|1 (x2 , t|x1 ) is of the form:6
In formula (9.2.2), the quantity W (x2 |x1 ) 0 is the transition probability per unit
time
R from x1 to x 2 =
6 x1 . The parameter is determined by the normalization condition
p1|1 (x2 , t|x1 ) dx2 = 1: Z
= W (x2 |x1 ) dx2 . (9.2.3)
The important point here is the existence of a transition rate W (x2 |x1 ). As displayed
by equation (9.2.2), W (x2 |x1 ) is linked to p1|1 (x2 , t|x1 ) (more precisely, to the be-
havior of p1|1 (x2 , t|x1 ) for t ⌧ ⌧ ).
On account of the stationarity property of p1|1 , we can rewrite equation (9.1.11)
in the form:
Z
p1|1 (x3 , t + t|x1 ) = p1|1 (x2 , t|x1 )p1|1 (x3 , t|x2 ) dx2 , (9.2.4)
5
No hypothesis being made about the distribution p1 (which may be time-dependent), such a
process is not necessarily stationary (only its transition probability is stationary).
6
The symbol o( t) denotes an unspecified term such that o( t)/ t ! 0 as t ! 0.
7
This implies t ⌧ ⌧ . However, no upper bound is imposed to the time t, which may be
arbitrarily large.
224 Master equations
which can also be rewritten, using the expression (9.2.3) for (with modified nota-
tions):
Z
@p1|1 (x, t|x1 ) ⇥ ⇤
= W (x|x0 )p1|1 (x0 , t|x1 ) W (x0 |x)p1|1 (x, t|x1 ) dx0 . (9.2.7)
@t
In the transition probabilities involved in equation (9.2.7), the initial value x1 is given.
We can write equivalently, simplifying the notations,9
Z
@p1|1 (x, t) ⇥ ⇤
= W (x|x0 )p1|1 (x0 , t) W (x0 |x)p1|1 (x, t) dx0 , (9.2.8)
@t
and re-expressing @p1|1 (x, t|x1 , 0)/@t by making use of the master equation (9.2.7), we
deduce from equation (9.2.8) the equation:
ZZ h i
@p1 (x, t)
= W (x|x0 )p1|1 (x0 , t|x1 ) W (x0 |x)p1|1 (x, t|x1 ) p1 (x1 , t1 ) dx0 dx1 .
@t
(9.2.12)
Carrying out the integration over x1 , we deduce from equation (9.2.12) the evolution
equation of p1 :
Z
@p1 (x, t) ⇥ ⇤
= W (x|x0 )p1 (x0 , t) W (x0 |x)p1 (x, t) dx0 . (9.2.13)
@t
Equation (9.2.13) for p1 is formally identical to the master equation (9.2.8) for p1|1 .
It is a linear first-order di↵erential equation, from which we can deduce p1 (x, t) from
the initial distribution p1 (x, 0) if the transition rates11 are known.
11
The equation (9.2.13) for p1 is often also called, by extension, a master equation. This denom-
ination may however be somewhat confusing. The fact that the distribution p1 (x, t) obeys equation
(9.2.13) does not guarantee that the same is true for the transition probability p1|1 (x, t) (and thus,
that the process under consideration is Markovian).
226 Master equations
where U (t) denotes the elementary evolution operator e i H1 t/h̄ . Equation (9.3.3)
does not reduce to a balance equation for the diagonal elements of the density matrix.
12
The Hamiltonian H1 is assumed to have no diagonal elements on the eigenbase of H0 , a hy-
pothesis which can always be made possibly through an appropriate redefinition of H0 .
The Pauli master equation 227
The terms with p 6= m give oscillating contributions which cannot be discarded with-
out further assumptions. If it is possible to make, at any time, the random phase
assumption, according to which the phases of the quantum coherences are distributed
at random (which amounts to assuming that the density matrix is diagonal at any
time), equation (9.3.3) takes the form of a balance equation involving solely the diag-
onal elements of ⇢(t):
d⇢nn (t) 1 X⇥ 2 2 ⇤
= lim |Unm | ⇢mm (t) |Umn | ⇢nn (t) . (9.3.4)
dt t!0 t m
If the energy spectrum is continuous or densely packed, we can introduce the transition
rates Wn0 ,n as defined by:
2
|Un0 n |
Wn0 ,n = lim , (9.3.5)
t!0 t
and which may be expressed with the aid of the Fermi golden rule:
2⇡ 2 2
Wn0 ,n = |h n0 |H1 | n i| ("n "n0 ). (9.3.6)
h̄
We thus get for the diagonal elements of the density matrix an evolution equation of
the form of the Pauli master equation:13
d⇢nn (t) X⇥ ⇤
= Wn,n0 ⇢n0 n0 (t) Wn0 ,n ⇢nn (t) . (9.3.7)
dt 0 n
3.2. Irreversibility
In contrast to the microscopic equations of motion, the Pauli master equation is not
time-reversal invariant.14 Thus it may conveniently describe the irreversible behavior
of a macroscopic system. However, as shown by the general evolution equation (9.3.3),
a quantum system obeys the Pauli master equation only if its density matrix may
be considered as diagonal on the base {| n i}. However, even if the density matrix is
diagonal at a given time, it is no longer diagonal later on, due to the interaction H1 .
The main difficulty we encounter when we try to establish the Pauli master equa-
tion is precisely the way of passing from the microscopic description, reversible, to a
macroscopic one, irreversible. In the following, we will present a method in which the
approximations allowing us to arrive to an irreversible description appear explicitly.
In a first step, we write, via a second-order perturbation calculation, an evolution
equation for the diagonal elements of the density matrix. This equation, reversible,
called the generalized master equation, involves only the diagonal elements of ⇢(t).
We then show how, owing to proper hypotheses (thermodynamic limit, short memory
approximation), we can deduce from the generalized master equation (reversible), the
Pauli master equation (irreversible).
13
The diagonal elements ⇢nn (t) of the density matrix have to be identified with the average
occupation probabilities pn (t).
14
Indeed, all its terms are real, and time is linearly involved in the first-order derivative.
228 Master equations
d⇢nn (t) XZ t ⇥ ⇤
= 2
dt0 ⌦n,n0 (t t0 ) ⇢n0 n0 (t0 ) ⇢nn (t0 ) + O( 3 ⇢), (9.4.1)
dt 0 0
n 6=n
Equation (9.4.1), which involves only the diagonal elements of the density matrix, is
a retarded or generalized 16 master equation. It contains a memory kernel defined by
the functions ⌦n,n0 (t).
d⇢nn ( t) XZ t ⇥ ⇤
= 2
dt0 ⌦n,n0 ( t t0 ) ⇢n0 n0 (t0 ) ⇢nn (t0 ) . (9.4.3)
dt 0
n0 6=n
15
The derivation of equation (9.4.1) is purely mathematical and does not rely on physical argu-
ments. It will not be detailed here.
16
The generalized master equation (9.4.1) is sometimes qualified as ‘non-Markovian’, owing to
the fact that it involves the diagonal elements of the density matrix at times t0 prior to t. Such a
terminology is however ambiguous, since the generalized master equation does not contain enough
information to allow us to determine whether the process it describes is, or is not, Markovian.
From the generalized master equation to the Pauli master equation 229
As displayed by formula (9.4.2), ⌦nn0 (t) is an even function. We can therefore write:
d⇢nn ( t) XZ t ⇥ ⇤
= 2
dt0 ⌦n,n0 (t + t0 ) ⇢n0 n0 (t0 ) ⇢nn (t0 ) , (9.4.4)
dt 0
n0 6=n
d⇢nn ( t) XZ t ⇥ ⇤
= 2
dt00 ⌦n,n0 (t t00 ) ⇢n0 n0 ( t00 ) ⇢nn ( t00 ) . (9.4.5)
dt 0
n0 6=n
d⇢nn (t) X Z t ⇥ ⇤
= 2
dt0 ⌦nn0 (t0 ) ⇢n0 n0 (t) ⇢nn (t) . (9.5.1)
dt 0
n0 6=n
We have:
Z ✓ ◆
t
2 2 h̄ "n " n0
dt ⌦nn0 (t ) = 2 |h
0 0
n0 |H1 | n i| sin t · (9.5.2)
0 h̄ "n " n0 h̄
In equation (9.5.1), the coefficients of the diagonal elements of the density matrix are
time-dependent. Besides, this equation remains invariant under time-reversal. Both
features di↵erentiate it from the Pauli master equation (9.3.6). However, in the case
of a system with a very large number of degrees of freedom, we can, by means of an
approximation about the form of the memory kernel, obtain, starting from (9.5.1), the
Pauli master equation.
the density matrix will be denoted respectively by H1 (", ↵; "0 , ↵0 ) and ⇢(", ↵; t). With
these notations, and on account of formula (9.5.2), equation (9.5.1) may be rewritten
in the form:
XX 2 ✓ ◆
d⇢(", ↵; t) 0 2 h̄ " "0 ⇥ 0 0 ⇤
= 2 2 |H(", ↵; " 0
, ↵ )| 0
sin t ⇢(" , ↵ ; t) ⇢(", ↵; t) .
dt 0 0 h̄
" " h̄
↵ "
(9.5.3)
In the thermodynamic limit in which the size of the system tends towards infinity, we
can introduce the density of states in energy n(") and replace the discrete sum over "0
involved on the right-hand side of equation (9.5.3) by an integral over the energy. We
can then write for any time t (bounded however by ⌧ ):
@⇢(", ↵; t) XZ 2 2
= 2 d"0 n("0 ) 2 |H(", ↵; "0 , ↵0 )|
@t h̄
↵0
✓ ◆
h̄ " "0 ⇥ 0 0 ⇤
⇥ 0
sin t ⇢(" , ↵ ; t) ⇢(", ↵; t) .
" " h̄
(9.5.4)
Note that, at this stage, the evolution equation of the diagonal elements of the density
matrix is still reversible.
d⇢(", ↵; t) XZ 2⇡ 2⇥ ⇤
= 2 d"0 n("0 ) (" "0 )|H(", ↵; "0 , ↵0 )| ⇢("0 , ↵0 ; t) ⇢(", ↵; t) .
dt h̄
↵0
(9.5.5)
The integration over "0 gives:
@⇢(", ↵; t) X⇥ ⇤
= W (↵, ↵0 )⇢(", ↵0 ; t) W (↵0 , ↵)⇢(", ↵; t) . (9.5.6)
@t 0 ↵
Equation (9.5.6), with the transition rates (9.5.7), is the Pauli master equation.
In contrast to the generalized master equation, the Pauli master equation is not
time-reversal invariant. It describes an irreversible evolution of the system. An irre-
versible evolution equation is thus obtained as soon as we make the hypothesis t ⌧c ,
called the short memory approximation. It is only then that the evolution of the di-
agonal elements of the density matrix becomes analogous to that of the probabilities
pn (t) of a Markov process.
6. Discussion
6.1. Validity domain
Let us come back on the hypotheses needed to establish the Pauli master equation.
Amongst these, the hypotheses made to obtain the generalized master equation (9.4.1)
(perturbation calculation, initial random phase hypothesis) are made first. Then, we
have the passage to the thermodynamic limit, followed by the passage to the limit
t ⌧c (short memory approximation), the order in which these two limits are taken
being crucial. When the coupling parameter is finite, the Pauli master equation is
only valid for times t ⌧ ⌧ , where ⌧ = O( 2 ). For the Pauli master equation to be
valid at any time t, we have to take the limit ( ! 0, t ! 1), the product 2 t
remaining finite.
As in the case of the master equations governing the transition probabilities of
Markovian processes, we again encounter, for the derivation of the Pauli master equa-
tion, the need for the existence of two well-separated time scales. The short time
scale ⌧c characterizes the duration of a microscopic interaction inducing a transition
from one state to another. The large time scale ⌧ corresponds to the typical time
separating two microscopic interactions, and depends on the coupling strength. The
description of the evolution in terms of a master equation is only possible if ⌧c ⌧ ⌧ .
It is then valid for times t such that ⌧c ⌧ t ⌧ ⌧ .
time resolution limited by t. We can, in this way, make use of the Pauli master
equation for times t much larger than ⌧ , and thus in particular to study the approach
of equilibrium of a macroscopic system.
Bibliography
References
In 1827, the botanist R. Brown discovered under the microscope the incessant and
irregular motion of small pollen particles suspended in water. He also remarked that
small mineral particles behave exactly in the same way (such an observation is impor-
tant, since it precludes to attributing this phenomenon to some ‘vital force’ specific
to biological objects). In a general way, a particle in suspension in a fluid executes
a Brownian motion when its mass is much larger than the mass of one of the fluid’s
molecules.
The idea according to which the motion of a Brownian particle is a consequence of
the motion of the lighter molecules of the surrounding fluid became widespread during
the second half of the nineteenth century. The first theoretical explanation of this
phenomenon was given by A. Einstein in 1905. The direct experimental checking of
the Einstein’s theory led to the foundation of the atomic theory of matter (in particular
the measurement of the Avogadro’s number by J. Perrin in 1908). A more achieved
theory of Brownian motion was proposed by P. Langevin in 1908.
However, slightly before A. Einstein, and in a completely di↵erent context,
L. Bachelier had already obtained the law of Brownian motion in his thesis entitled
“La théorie de la spéculation” (1900). Models having recourse to Brownian motion or
to its generalizations are widely used nowadays in financial mathematics. In a more
general setting, Brownian motion played an important role in mathematics: histor-
ically, it was to represent the displacement of a Brownian particle that a stochastic
process was constructed for the first time (N. Wiener, 1923).
The outstanding importance of Brownian motion in out-of-equilibrium statistical
physics stems from the fact that the concepts and methods used in its study are not
restricted to the description of the motion of a particle immersed in a fluid of lighter
molecules, but are general and may be applied to a wide class of physical phenomena.
236 Brownian motion: the Langevin model
d2 x dx
m = m + F (t), (10.1.1)
dt2 dt
or:
dv dx
m = m v + F (t), v= · (10.1.2)
dt dt
• Average value
We assume that the average value of the Langevin force vanishes:
⌦ ↵
F (t) = 0. (10.1.3)
This hypothesis is necessary to have the average value of the Brownian particle’s
velocity vanishing at equilibrium (as it should, since there is no applied external force).
• Autocorrelation function
The autocorrelation function of the random force,
⌦ ↵
g(⌧ ) = F (t)F (t + ⌧ ) , (10.1.4)
(the signification of the parameter D will be made precise later). The correlation time
is of the order of the mean time interval separating two successive collisions of the
fluid’s molecules on the Brownian particle. If this time is much shorter than the other
characteristic times, such as for instance the relaxation time of the average velocity
from a well-defined initial value,4 we can assimilate g(⌧ ) to a delta function of weight
2Dm2 :
g(⌧ ) = 2Dm2 (⌧ ). (10.1.6)
3
The averages taking place here are defined as ensemble averages computed with the aid of the
distribution function of the bath (see Supplement 10B).
4
See Subsection 2.2.
5
This hypothesis may be justified on account of the central limit theorem: indeed, due to the
numerous collisions undergone by the Brownian particle, the force F (t) may be considered as resulting
from the superposition of a very large number of identically distributed random functions.
238 Brownian motion: the Langevin model
The average velocity of the Brownian particle responds linearly to the external
applied force. We can associate with this response a transport coefficient. The Brown-
ian particle, if it carries a charge q, acquires under the e↵ect of a static electric field E
the limit velocity hvi = qE/m . Its mobility µ = hvi/E is thus:7
q
µ= = qA(! = 0). (10.2.7)
m
6
For the sake of simplicity, we use the same notation Fext (.) for the force Fext (t) and its Fourier
transform Fext (!), as well as the same notation hv(.)i for the average velocity hv(t)i and its Fourier
transform hv(!)i.
7
No confusion being possible here with a chemical potential, the drift mobility of the Brownian
particle is simply denoted by µ (and not by µD ).
Response and relaxation 239
v(0) = v0 . (10.2.8)
The solution of equation (10.1.2) corresponding to the initial condition (10.2.8) reads:
Z
1 t
(t t0 )
v(t) = v0 e t
+ F (t0 )e dt0 , t > 0. (10.2.9)
m 0
The velocity v(t) of the Brownian particle is a random process. In the above defined
conditions, this process is not stationary. We will compute the average value and the
variance of v(t) at any time t > 0.
• Average velocity
Since the fluctuating force vanishes on average, we obtain, from formula (10.2.9):
⌦ ↵
v(t) = v0 e t , t > 0. (10.2.10)
The average velocity relaxes exponentially towards zero with a relaxation time
⌧r = 1 .
• Velocity variance
The variance of the velocity is defined for instance by the formula:
⌦⇥ ⌦ ↵⇤2 ↵
v (t) = (10.2.11)
2
v(t) v(t) .
When the autocorrelation function of the Langevin force is given by the simplified
formula (10.1.6), we obtain:
Z t
2 (t t0 )
v (t) = 2D dt0 , (10.2.13)
2
e
0
that is:
D
v (t) = (1 ), t > 0. (10.2.14)
2 2 t
e
At time t = 0, the variance of the velocity vanishes (the initial velocity is a non-random
variable). Under the e↵ect of the Langevin force, velocity fluctuations arise, and the
variance v2 (t) increases with time. At first, this increase is linear:
For t ⌧r , the average velocity tends towards zero (formula (10.2.10)). Equations
(10.2.16) and (10.2.17) show that hv 2 (t)i then tends towards a limit value D/ in-
dependent of v0 . The average energy hE(t)i = mhv 2 (t)i/2 tends towards the corre-
sponding limit hEi = mD/2 . Then, the Brownian particle is in equilibrium with the
bath.
If the bath is itself in thermodynamic equilibrium at temperature T , the average
energy of the particle in equilibrium with it takes its equipartition value hEi = kT /2.
Comparing both expressions for hEi, we get a relation between the di↵usion coefficient
D in the velocity space, associated with the velocity fluctuations, and the friction
coefficient , which characterizes the dissipation:
m
= D. (10.2.18)
kT
Z 1
1 ⌦ ↵
= F (t)F (t + ⌧ ) d⌧. (10.2.19)
2mkT 1
Equation (10.2.19) relates the friction coefficient to the autocorrelation function of the
Langevin force. It is known as the second fluctuation-dissipation theorem.9 This the-
orem expresses here the fact that the friction force and the fluctuating force represent
8
It will be shown in Subsection 4.3 that this relation can be extended to the case in which the
autocorrelation function of the Langevin force is not a delta function but a function of finite width
characterized by the correlation time ⌧c , provided that we have ⌧c ⌧ ⌧r . See also on this question
Supplement 10A.
9
Generally speaking, the fluctuation-dissipation theorem, which can be formulated in di↵erent
ways, constitutes the heart of the linear response theory (see Chapter 14). In the case of Brownian
motion as described by the Langevin model, the terminology of ‘second’ fluctuation-dissipation theo-
rem, associated with formula (10.2.19) for the integral of the random force autocorrelation function,
is due to R. Kubo.
Response and relaxation 241
two aspects of the same physical phenomenon, namely, the collisions of the Brownian
particle with the molecules of the fluid which surrounds it.
x(0) = x0 . (10.2.20)
Integrating the expression (10.2.9) for the velocity between times 0 and t, we get, given
the initial condition (10.2.20):
Z (t t0 )
v0 1 t
1 e
x(t) = x0 + (1 e t
)+ F (t0 ) dt0 , t > 0. (10.2.21)
m 0
The displacement x(t) x0 of the Brownian particle is also a random process. This
process is not stationary. We will calculate the average and the variance of the dis-
2
placement as functions of time, as well as the second moment h[x(t) x0 ] i.
• Average displacement
We have: ⌦ ↵ v0
x(t) = x0 + (1 e t
), t > 0. (10.2.22)
For t ⌧r , the average displacement hx(t)i x0 tends towards the finite limit v0 / .
• Variance of the displacement
The variance of the displacement x(t) x0 is also the variance of x(t), defined for
instance by the formula:
⌦⇥ ⌦ ↵⇤2 ↵
x (t) = (10.2.23)
2
x(t) x(t) .
From formulas (10.2.21) and (10.2.22), we get:
Z Z
1 t t ⌦ ↵ (t t0 ) (t t00 )
x (t) = dt0 dt00 F (t0 )F (t00 ) [1 ][1 ], (10.2.24)
2
e e
m2 2 0 0
that is, taking for the autocorrelation function of the Langevin force the simplified
expression (10.1.6):
Z
2D t 0 2
2
x (t) = 2
(1 e t ) dt0 . (10.2.25)
0
Starting from its vanishing initial value, the variance of the displacement increases,
first as 2Dt3 /3 for t ⌧ ⌧r , then as 2Dt/ 2 for t ⌧r .
242 Brownian motion: the Langevin model
Formulas (10.2.26) and (10.2.28) show that the Brownian particle di↵uses at large
times. The di↵usion coefficient D is related to the di↵usion coefficient in the velocity
space D by the formula:
D
D= 2
· (10.2.29)
dx
⌘ = F (t). (10.2.30)
dt
Using the expression (10.2.31) for the random force autocorrelation function, we get,
for any t: ⌦ 2↵
[x(t) x0 ] = 2Dt. (10.2.33)
In this description, the motion of the Brownian particle is di↵usive at any time.
D kT
= · (10.2.35)
µ q
Formula (10.2.35) is the Einstein relation between the di↵usion coefficient D, associ-
ated with the displacement fluctuations, and the mobility µ, related to the dissipation.
The Einstein relation is a formulation of the first fluctuation-dissipation theorem.12 It
may also be written in the form of a relation between D and ⌘:
kT
D= · (10.2.36)
⌘
In these conditions, at any finite time t, the particle is in equilibrium with the bath.
Its velocity v(t) is a stationary random process.14 Since the average value of the ve-
locity vanishes at equilibrium, the autocorrelation function of v(t), which we will now
compute, represents the dynamics of the equilibrium velocity fluctuations.
3.1. Correlation function between the Langevin force and the velocity
To begin with, starting from formula (10.3.2), it is possible to compute the correlation
function hv(t)F (t0 )i:
Z
⌦ ↵ 1 t ⌦ ↵ 00
v(t)F (t ) =
0
F (t00 )F (t0 ) e (t t ) dt00 . (10.3.3)
m 1
When the autocorrelation function of the Langevin force is of the form (10.1.6),
equation (10.3.3) reads:
Z t
⌦ ↵ 00
v(t)F (t0 ) = 2Dm (t0 t00 )e (t t ) dt00 . (10.3.4)
1
<v(t)F(t')>
γ−1
t t'
τc
Fig. 10.1 Correlation function between the Langevin force and the velocity at
finite ⌧c .
14
When the force F (t) is modelized by a stationary Gaussian random process of autocorrelation
function g(⌧ ) = 2Dm2 (⌧ ), the stationary process v(t) defined by formula (10.3.2) is called the
Ornstein–Uhlenbeck process (see Chapter 11 and Supplement 11B).
Equilibrium velocity fluctuations 245
Actually, the correlation time ⌧c of the Langevin force does not vanish, and the
result (10.3.5) is correct only for |t t0 | ⌧c . Taking the finite correlation time into
account results in the smoothing out of the discontinuity exhibited by formula (10.3.5),
the correlation function hv(t)F (t0 )i passing in fact continuously from its maximum
value to zero over a time interval of order ⌧c . The shape at finite ⌧c of the curve
representing hv(t)F (t0 )i as a function of t0 , the time t being fixed, is shown in Fig. 10.1.
⌦ ↵ D
v(t)v = e |t|
. (10.3.8)
<v(t)v>
τc
γ−1
0 t
Fig. 10.2 Equilibrium velocity correlation function, at vanishing ⌧c (full line), and
at finite ⌧c (dotted line).
15
This property will be established later (see formula (10.4.13)).
246 Brownian motion: the Langevin model
d⌦ ↵ ⌦ ↵
v(t)v(t0 ) = v(t)v(t0 ) , t t0 . (10.3.9)
dt
Equation (10.3.9) is of the same form as the di↵erential equation dhv(t)i/dt = hv(t)i
(t 0) describing the relaxation of hv(t)i from a well-defined initial value v(t = 0).
This property, according to which the velocity fluctuations regress (that is, disappear)
according to the same law as the average velocity, is called the regression theorem.
⌦ ↵ kT
v(t)v = e |t|
. (10.3.10)
m
Coming back to the definition (10.2.5) of the complex admittance, we get the identity:
Z 1⌦
1 ↵
A(!) = v(t)v ei!t dt. (10.3.12)
kT 0
16
The Fourier–Laplace transformation, also called the unilateral Fourier transformation, is defined
over the integration interval (0, 1) (in contrast to the ordinary Fourier transformation, defined over
the interval ( 1, 1)). The usual Laplace transformation, also defined over the interval (0, 1), uses,
in place of i!, a complex parameter z.
17
The terminology of ‘first’ fluctuation-dissipation theorem, associated with formula (10.3.12) for
the Fourier–Laplace transform of the equilibrium velocity autocorrelation function, is due to R. Kubo.
We can also get this result by applying the Kubo formulas of the general theory of linear response to
the isolated system made up of the Brownian particle coupled with the bath (see Chapter 14).
Harmonic analysis of the Langevin model 247
4.1. Relation between the spectral densities of the random force and of the
velocity
The Fourier transforms F (!) of the random force, on the one hand, and v(!) of the
Brownian particle’s velocity, on the other hand, are defined by the formulas:18
Z 1
F (!) = F (t)ei!t dt (10.4.1)
1
and: Z 1
v(!) = v(t)ei!t dt. (10.4.2)
1
Given that F (t) and v(t) are stationary random processes, F (!) and v(!) are in fact
obtained by integrating over a large interval of finite width T of the time axis starting
from any origin, the limit T ! 1 being taken at the end of the calculations. In the
framework of the Langevin model, F (!) and v(!) are related by the formula:
1 1
v(!) = F (!). (10.4.3)
m i!
The spectral densities SF (!) and Sv (!) are defined by the formulas:
1⌦ 2↵ 1⌦ 2↵
SF (!) = lim |F (!)| , Sv (!) = lim |v(!)| . (10.4.4)
T !1 T T !1 T
1 1
Sv (!) = SF (!). (10.4.5)
m2 2 + !2
The spectral density of the Brownian particle’s velocity is thus the product of the
spectral density of the random force by a Lorentzian of width ⇠ .
18
For the sake of simplicity, we use the same notation F (.) for the random force F (t) and its Fourier
transform F (!), as well as the same notation v(.) for the velocity v(t) of the Brownian particle and
its Fourier transform v(!).
248 Brownian motion: the Langevin model
According to the Wiener–Khintchine theorem, the spectral density and the au-
tocorrelation function of a stationary random process form a Fourier transform pair.
The autocorrelation function g(⌧ ) of the random force being a very ‘peaked’ function
(of width ⇠ ⌧c ) around ⌧ = 0, the spectral density SF (!) is a very ‘broad’ function (of
width ⇠ ⌧c 1 ). The bath being at thermodynamic equilibrium, SF (!) is referred to as
the thermal noise.19
!c2
SF (!) = SF , SF = 2Dm2 . (10.4.8)
!c2 + !2
g(⌧ ) = Dm2 !c e !c |⌧ |
. (10.4.9)
19
We will come back in more detail to the study of the thermal noise (in an electric conductor at
equilibrium) in Supplement 10C.
20
Such expressions for SF (!) and g(⌧ ) may be justified starting from certain microscopic models
for the interaction of the Brownian particle with the bath (see Supplement 10B).
Time scales 249
⌦ ↵ D !c2
v(t)v = e |t|
e !c |t|
. (10.4.12)
!c2 2 !c
⌦ 2↵ D !c
v = , (10.4.13)
!c +
5. Time scales
Thus, as shown for instance by formula (10.4.12), two time scales come into play in
the dynamics of the equilibrium velocity fluctuations of a Brownian particle. The first
one, very short, is the correlation time ⌧c of the random force, whereas the other one,
much longer, is the relaxation time ⌧r = 1 of the average velocity.22
For the velocity fluctuations to regress (that is, to decrease noticeably), a time
at least of the order of the longest time scale ⌧r is needed. The Brownian particle’s
velocity is thus essentially a slow variable. The random force, whose autocorrelation
21
The arguments presented here are approximate. Indeed, it is not fully consistent to take into ac-
count the finite correlation time of the random force while retaining the instantaneous character of the
friction term as it appears in the Langevin equation (10.1.2). In the case of colored noise, we must in
fact write down a generalized Langevin equation with a retarded friction term (see Supplement 10A).
22
Despite the fact that formula (10.4.12) has been established in the particular case of an expo-
nential autocorrelation function g(⌧ ), the result concerning the characteristic times of the dynamics
of the equilibrium velocity fluctuations has a more general scope.
250 Brownian motion: the Langevin model
function decreases over a much shorter time of order ⌧c , is a rapid variable. This
separation of time scales, as pictured by the inequality:
⌧c ⌧ ⌧r , (10.5.1)
is crucial in the Langevin model. It can be shown, using microscopic models, that
inequality (10.5.1) is actually verified when the particle under study is much heavier
than the molecules of the fluid which surrounds it. It is only in this case that a
particle moving within a fluid may be qualified as ‘Brownian’ and its evolution properly
described by the Langevin equation (10.1.2).
Moreover, in the term in dv/dt involved in the Langevin equation (10.1.2), dt
does not stand for an infinitesimal time interval but for a finite time interval t
during which a finite change v of the particle velocity takes place. The interval t is
necessarily much longer than the collision time ⌧c , since the evolution of the Brownian
particle’s velocity results from the many collisions that the particle undergoes with the
fluid’s molecules. Besides, equation (10.1.2) (averaged) describes the relaxation of the
average velocity fluctuations, and accounts for a significant evolution only when t
remains small as compared to the relaxation time ⌧r . These considerations show that
the Langevin equation describes the evolution of the velocity of a Brownian particle
over a time interval t between ⌧c and ⌧r :
⌧c ⌧ t ⌧ ⌧r . (10.5.2)
In the viscous limit, in which interest is focused on the evolution of the Brow-
nian particle’s displacement (Einstein–Smoluchowski description), the evolution time
interval of interest verifies the inequality:
t ⌧r . (10.5.3)
Bibliography 251
Bibliography
P.M. Chaikin and T.C. Lubensky, Principles of condensed matter physics, Cam-
bridge University Press, Cambridge, 1995.
S. Chandrasekhar, Stochastic problems in physics and astronomy, Rev. Mod. Phys.
15, 1 (1943). Reprinted in Selected papers on noise and stochastic processes (N. Wax
editor), Dover Publications, New York, 2003.
C.W. Gardiner, Handbook of stochastic methods, Springer-Verlag, Berlin, third edi-
tion, 2004.
N.G. van Kampen, Stochastic processes in physics and chemistry, North-Holland,
Amsterdam, third edition, 2007.
R. Kubo, M. Toda, and N. Hashitsume, Statistical physics II: nonequilibrium
statistical mechanics, Springer-Verlag, Berlin, second edition, 1991.
R.M. Mazo, Brownian motion: fluctuations, dynamics, and applications, Oxford
University Press, Oxford, 2002.
F. Reif, Fundamentals of statistical and thermal physics, McGraw-Hill, New York,
1965.
R. Zwanzig, Nonequilibrium statistical mechanics, Oxford University Press, Oxford,
2001.
References
1
If we neglect ⌧c , and if we assume that F (t) is a Gaussian process, the evolution of the velocity
from time t, as described by equation (10A.1.1), depends only on its value at this time, and not
on the values it took at prior times. The velocity v(t) is thus in this case a Markov process. This
property remains approximately true if we take into account the finite character of ⌧c , provided that
the inequality ⌧c ⌧ ⌧r is verified (see Chapter 11).
2
The reason for this designation will be made clear later.
3
For the sake of simplicity, we use the same notation A(.) for the response function A(t) and its
Fourier transform A(!).
254 The generalized Langevin model
is thus:
1
A(t) = ⇥(t)
e t. (10A.1.4)
m
In formula (10A.1.4), ⇥(t) denotes the Heaviside function:
⇢
0, t < 0
⇥(t) = (10A.1.5)
1, t > 0.
The Fourier transform of A(t) is the complex admittance of the Langevin model:
1 1
A(!) = · (10A.1.6)
m i!
In some respects, this description of the motion of a particle immersed within a
fluid is too schematic. In particular, the friction cannot be established instantaneously.
Its setting up requires a time at least equal to the collision time ⌧c of the particle with
the fluid’s molecules. Retardation e↵ects are thus necessarily present in the friction
term. Their consideration leads to a modification of the form of the equation of motion
and, accordingly, of the velocity response function, as well as to a modification of the
random force autocorrelation function.
Z t
dv dx
m = m (t t0 )v(t0 ) dt0 + F (t), v= , (10A.1.7)
dt 1 dt
in which the friction force at time t is determined by the values of the velocity at times
t0 < t. Equation (10A.1.7) is called the generalized or retarded Langevin equation.
Rt
The friction force m 1 (t t0 )v(t0 ) dt0 involves a memory kernel defined by
the function (t) for t > 0. It is in fact convenient to consider that the memory kernel
R (t)
1
is defined for any t as a decreasing function of |t|, of width ⇠ ⌧c , and such that
1
(t) dt = 2 . Introducing the causal or retarded memory kernel ˜ (t) = ⇥(t) (t),
we can rewrite equation (10A.1.7) in the following equivalent form:
Z 1
dv dx
m = m ˜ (t t0 )v(t0 ) dt0 + F (t), v= · (10A.1.8)
dt 1 dt
The Langevin force F (t) is modelized here, as in the non-retarded Langevin model,
by a random stationary process of zero mean. This process is, most often, assumed to
be Gaussian. Since we now take into account the retarded character of the friction,
it is necessary, for the consistency of the model, to take equally into account the
non-vanishing correlation time of the random force. Accordingly, we assume that the
autocorrelation function g(⌧ ) = hF (t)F (t + ⌧ )i decreases over a finite time ⇠ ⌧c . In
contrast to the non-retarded case, we do not make the hypothesis ⌧c ⌧ 1.
Harmonic analysis of the generalized Langevin model 255
2. Complex admittance
In the presence of an applied external force Fext (t), the retarded equation of motion
of the particle reads:
Z 1
dv dx
m = m ˜ (t t0 )v(t0 ) dt0 + F (t) + Fext (t), v= · (10A.2.1)
dt 1 dt
On average, we have:
⌦ ↵ Z 1
⌦ ↵
d v ⌦ ↵ ⌦ ↵ d x
m = m ˜ (t t ) v(t0 ) dt0 + Fext (t),
0
v = · (10A.2.2)
dt 1 dt
with: ⌦ ↵
v = A(!)F. (10A.2.4)
The quantity:
1 1
A(!) = (10A.2.5)
m (!) i!
is the complex admittance of the generalized Langevin model. In formula (10A.2.5),
the generalized friction coefficient (!), defined by the formula:
Z 1
(!) = (t)ei!t dt, (10A.2.6)
0
denotes the Fourier–Laplace transform of the memory kernel (t) (or the Fourier trans-
form of the retarded memory kernel ˜ (t)). Note that (! = 0) = .
More generally, for an external force Fext (t) of Fourier transform Fext (!), the
Fourier transform hv(!)i of the stationary solution hv(t)i of equation (10A.2.2) is:
⌦ ↵
v(!) = A(!)Fext (!). (10A.2.7)
Z 1⌦
1 ↵
A(!) = v(t)v ei!t dt. (10A.3.2)
kT 0
Using the Wiener–Khintchine theorem, we deduce from equation (10A.3.4) the spectral
density of the velocity,
2kT <e (!)
Sv (!) = , (10A.3.5)
m | (!) i!|2
then, using formula (10A.3.1), the spectral density of the random force:
1
(⌧ ) = g(⌧ ). (10A.3.8)
mkT
4
This result can be demonstrated by applying the Kubo formulas of the linear response theory
to the isolated system made up of the particle coupled with the bath (see also Chapter 14).
An analytical model 257
The decrease of the memory kernel is thus characterized by the correlation time5 ⌧c .
The generalized friction coefficient is proportional to the Fourier–Laplace trans-
form of the random force autocorrelation function:
Z 1⌦
1 ↵
(!) = F (t)F (t + ⌧ ) ei!⌧ d⌧. (10A.3.9)
mkT 0
4. An analytical model
If we have explicit analytical expressions for the memory kernel and the random force
autocorrelation function, we can derive analytical expressions for the complex admit-
tance, and, possibly, for the velocity response function in the generalized Langevin
model.
g(⌧ ) = Dm2 !c e !c |⌧ |
, (10A.4.1)
(t) = !c e !c |t|
. (10A.4.2)
!c
(!) = · (10A.4.3)
!c i!
5
By contrast, formula (10A.3.8) shows that it is inconsistent to take into account the finite
correlation time of the random force while retaining the instantaneous character of the friction term
as it is displayed in the non-retarded Langevin equation (10A.1.1).
6
Expressions of this type can be obtained in the framework of certain microscopic models of the
interaction of the particle with the bath. This in particular the case in the Caldeira–Leggett model
(see Supplement 10B). It is then seen that !c can be given the significance of an angular frequency
characteristic of the bath.
7
The generalized friction coefficient (!) is the Fourier transform of the causal function ˜ (t) =
⇥(t) (t).
258 The generalized Langevin model
1 1
A(!) = !c · (10A.4.4)
m i!
!c i!
The poles of A(!) give access to the characteristic relaxation times of the average
velocity from a well-defined initial value. In the weak-coupling case !c / > 4, these
poles are of the form i!± , with:
⇣ ⌘1/2
!c
!± = 1± 1 4 !c 1
. (10A.4.5)
2
The expression (10A.4.6) for A(t) has to be compared with the corresponding ex-
pression in the non-retarded Langevin model (formula (10A.1.4)). In the generalized
Langevin model, we do not make the assumption ⌧c ⌧ 1. Accordingly, there is no
clear-cut separation of time scales between the random force and the velocity of the
particle.
Bibliography 259
Bibliography
References
The equation of motion of the particle coupled with the bath, which, according
to equations (10B.1.2), is of the form:
X cn
mẍ(t) = cn xn (t) x(t) , (10B.1.6)
n
mn !n2
Z t
mẍ(t) + m (t t0 )ẋ(t0 ) dt0 = mx(t0 ) (t t0 ) + F (t). (10B.1.7)
t0
The functions (t) and F (t) involved in equation (10B.1.7) are defined in terms of the
microscopic parameters of the model by the formulas:
1 X c2n
(t) = cos !n t (10B.1.8)
m n mn !n2
262 Brownian motion in a bath of oscillators
and:
X
pn (t0 )
F (t) = cn xn (t0 ) cos !n (t t0 ) + sin !n (t t0 ) . (10B.1.9)
n
mn !n
In equation (10B.1.7), (t) acts as a memory kernel and F (t) as a random force. Using,
instead of (t), the retarded memory kernel ˜ (t) = ⇥(t) (t), we can rewrite equation
(10B.1.7) in the following equivalent form:
Z 1
mẍ(t) + m ˜ (t t0 )ẋ(t0 ) dt0 = mx(t0 )˜ (t t0 ) + F (t). (10B.1.10)
t0
3
This is the classical distribution function of the bath in the phase space. The quantum general-
ization will be outlined in Section 3.
4
The averages involved here are computed with the aid of the bath’s distribution function
(10B.1.11).
The Caldeira–Leggett model 263
1 X c2n
˜ (t) = ⇥(t) cos !n t, (10B.1.15)
m n mn !n2
⇡ X c2n ⇥ ⇤
<e (!) = (! !n ) + (! + !n ) . (10B.1.17)
2m n mn !n2
At this stage, we generally introduce the spectral density of the coupling with the
environment, defined for ! > 0 by the formula:
⇡ X c2n
J(!) = (! !n ), ! > 0. (10B.1.18)
2 n mn ! n
J(!)
<e (!) = · (10B.1.19)
m!
In the continuum limit, the quantities <e (!) and J(!) may be considered as contin-
uous functions of !.
5
We thus define the Fourier transform of ˜ (t) in the distribution sense. This procedure is com-
monly used when computing the generalized susceptibilities as Fourier transforms of the response
functions (see Chapter 12).
6
We make use of the relation:
1 1
lim = vp i⇡ (x),
✏!0+ x + i✏ x
The expressions (10B.1.20) for J(!) and <e (!) are only valid at low angular fre-
quencies. Indeed, the spectral density does not in fact increase without bounds, but it
decreases towards zero as ! ! 1. To account for this behavior, we write, in place of
the relations (10B.1.20), the formulas:
✓ ◆ ✓ ◆
! !
J(!) = m !fc (! > 0), <e (!) = fc , (10B.1.21)
!c !c
in which fc (!/!c ) is a cut-o↵ function tending towards zero more rapidly than ! 1
as10 ! ! 1. We often choose for convenience a Lorentzian cut-o↵ function:
✓ ◆
! !2
fc = 2 c 2· (10B.1.22)
!c !c + !
We then have:
!c2 !c2
J(!) = m ! (! > 0), <e (!) = · (10B.1.23)
!c2 + !2 !c2 + !2
In this case, the memory kernel is modelized by an exponential function of time con-
stant !c 1 :
(t) = !c e !c |t| . (10B.1.24)
The corresponding modelization has to be done on the Langevin force autocorrelation
function, which yields:
⌦ ↵
F (t)F (t + ⌧ ) = mkT !c e !c |⌧ | . (10B.1.25)
7
The exponent may possibly take non-integer values, for instance in the case of an interaction
with a disordered or fractal environment.
8
This limit is the infinitely short memory limit (see Subsection 1.6)
9
The reason for this designation will be made clear in Subsection 1.6.
10
The angular frequency !c characterizes the width of the angular frequency band of the bath
oscillators e↵ectively coupled with the particle.
Dynamics of the Ohmic free particle 265
Z +1
dv dx
m +m ˜ (t t0 )v(t0 ) dt0 = F (t), v= · (10B.1.26)
dt 1 dt
Consider the case !c / > 4, which corresponds to a weak coupling between the particle
and the bath. We can then write:
⌦ ↵ kT ⇣ ⌘ 1/2 Z 1 ✓ 1 1
◆
v(t)v = 1 4 !c 1
e i!t d!, (10B.2.4)
1 ! +! ! 2 + !+
m⇡ 2 2 2
with:
!c ⇣ ⌘1/2
!± = 1± 1 4 !c 1
. (10B.2.5)
2
After the integration, we have:
✓ ◆
⌦ ↵ kT ⇣ ⌘ 1/2 !+ !
v(t)v = 1 4 !c 1
e ! |t|
e !+ |t|
. (10B.2.6)
m !c !c
In the infinitely short memory limit !c ! 1, we retrieve the result of the non-
retarded Langevin model:
⌦ ↵ kT
v(t)v = e |t| . (10B.2.7)
m
1 d⌦ 2↵
D(t) = [x(t) x] , t > 0. (10B.2.8)
2 dt
The limit value at large times of D(t) is, at any non-vanishing temperature, the Ein-
stein value D = kT /⌘ of the di↵usion coefficient. This result holds regardless of the
value, finite or not, of !c .
In the non-retarded model, we have:
kT
D(t) = 1 e t
, t > 0. (10B.2.10)
⌘
h̄!
SF (!) = h̄! coth m <e (!). (10B.3.1)
2
The memory kernel being independent of the temperature, its modelization by an expo-
nential (formula (10B.1.24)) may be maintained in the quantum case, which amounts
to keep the expression (10B.1.23) for <e (!). However, the relation (10B.1.12) be-
tween the memory kernel and the random force autocorrelation function has to be
modified. Equation (10B.1.26), which is still formally valid, is then referred to as a
quantum Langevin equation. It allows us to describe at any temperature, including
T = 0, the dynamics of the Ohmic free particle.
The main characteristics of the dynamics are the following:13 below some crossover
temperature linked to the bath, the description of the dynamics in terms of Brownian
motion, that is, with well-separated time scales for the random force, on the one hand,
and for the particle’s velocity, on the other hand, becomes inadequate. Indeed, at large
times, both the random force autocorrelation function and the velocity autocorrelation
function exhibit a long negative time tail / t 2 .
⌦ 2↵ h̄
[x(t) x] ⇠2 log t, t 1. (10B.3.3)
⇡m
The curves representing D(t) (and the corresponding classical di↵usion coefficient) as
a function of t for di↵erent temperatures are plotted in Fig. 10B.1.
2.5
T = 2Tc
2
πmD(t)/ħ
1.5
T = Tc
1
T = Tc/2
0.5
T=0
0
0 2 4 6 8 10
γt
Fig. 10B.1 The coefficient D(t) (in dashed lines, its classical counterpart).
Bibliography 269
Bibliography
References
1
In a general way, the fluctuation-dissipation theorem expresses a relation between the admittance
of a linear dissipative system and the equilibrium fluctuations of relevant generalized forces (see
Chapter 14).
The Nyquist theorem 271
By analogy with the hypotheses made about the Langevin force, we assume that
the fluctuating potential di↵erence V (t) is a centered stationary random process, with
fluctuations characterized by a correlation time ⌧c . We denote by g(⌧ ) = hV (t)V (t+⌧ )i
the autocorrelation function of V (t) and set:
Z 1
g(⌧ ) d⌧ = 2DL2 . (10C.2.4)
1
If ⌧c is much shorter than the other characteristic times, such as for instance the
relaxation time ⌧r = L/R, we assimilate g(⌧ ) to a delta function of weight 2DL2 :
Z 1⌦
1 ↵
R= V (t)V (t + ⌧ ) d⌧. (10C.2.10)
kT 0
Formula (10C.2.10) relates the resistance R of the circuit to the autocorrelation func-
tion of the fluctuating potential di↵erence. It constitutes the expression in the present
problem for the second fluctuation-dissipation theorem.
JV (! = 0) = 4RkT. (10C.2.12)
JV (!) = 4RkT, ! ⌧ ⌧c 1
. (10C.2.13)
JV (!) = 4R(!)kT, ! ⌧ ⌧c 1
. (10C.2.15)
2
The values of R(!) and of |A(!)| are determined experimentally. The measurement
of hI 2 (t)i thus allows us in principle to obtain the value of the Boltzmann constant.
In practice, we are more interested in the variance of V (t). We measure the contri-
bution hV 2 (t)i! to the variance of the angular frequencies belonging to a given range
(!, ! + !) of positive angular frequencies:
⌦ ↵ 1
V 2 (t) ! = JV (!) !. (10C.2.18)
2⇡
⌦ ↵ 2
V 2 (t) ! = R(!)kT !. (10C.2.19)
⇡
The measurement of the ratio hV 2 (t)i! /R(!) allows us experimental access to the
value of the Boltzmann constant.
Bibliography 275
Bibliography
References
Like the preceding one, this chapter deals with the Brownian motion of a free particle,
but from a di↵erent point of view. We aim here to determine the temporal evolution
of the distribution function of the Brownian particle’s velocity. This ‘à la Schrödinger’
approach is complementary to the previously adopted ‘à la Heisenberg’ one, which
consists in determining the solution of the Langevin equation.
The ‘à la Schrödinger’ approach relies crucially on the notion of Markov process.
If the Langevin force is assumed to be Gaussian and delta-correlated (in other words,
if the noise is a Gaussian white noise), the velocity of the Brownian particle can
indeed be considered as a Markov process. This property remains approximately true
even if the random force is not Gaussian and/or has a finite correlation time (colored
noise). The Markovian character of the velocity allows us to express its distribution
function at time t + t in terms of its distribution function at time t, provided that
t is sufficiently large to be compatible with the Markovian hypothesis ( t ⌧c ,
where ⌧c is the correlation time of the Langevin force). Under certain conditions,
we can deduce from this relation a second-order partial di↵erential equation for the
distribution function, known as the Fokker–Planck equation. Its characteristics are
obtained by studying the evolution of the velocity over the time interval t via the
Langevin equation.
Over much longer evolution time intervals ( t ⌧r , where ⌧r is the relaxation
time of the average velocity), the displacement of the Brownian particle may in turn be
represented by a Markov process. The distribution function of the displacement then
obeys an evolution equation similar to the Fokker–Planck equation, known as the
Einstein–Smoluchowski equation, which identifies with the usual di↵usion equation.
278 Brownian motion: the Fokker–Planck equation
1
See Supplement 11B.
2
If the Gaussian hypothesis is not verified, the Markovian character of v(t) is only approximate.
The same is true when we take into account the finite correlation time ⌧c of the Langevin force (that
is, when the noise is colored). In this case, the Brownian particle’s velocity may be considered as a
Markov process provided that we study its evolution over a time interval t ⌧c .
The Kramers–Moyal expansion 279
Our aim is to directly relate f (v, t + t) and f (v, t), without passing through
the initial distribution. Since v(t) is a Markov process, the transition probability p1|1
verifies equation (11.1.3), which we can rewrite, changing the notations, in the form:
Z
p1|1 (v, t + t|v0 , t0 ) = p1|1 (v 0 , t|v0 , t0 )p1|1 (v, t + t|v 0 , t) dv 0 . (11.1.6)
Importing the expression (11.1.6) for p1|1 (v, t + t|v0 , t0 ) into equation (11.1.5), we
get:
ZZ
f (v, t + t) = p1|1 (v 0 , t|v0 , t0 )p1|1 (v, t + t|v 0 , t) f (v0 , t0 ) dv 0 dv0 . (11.1.7)
Making use of the integral expression (11.1.4) for the distribution function (written
for f (v 0 , t)), we deduce from equation (11.1.7) a direct relation between f (v, t + t)
and f (v, t):
Z
f (v, t + t) = f (v 0 , t)p1|1 (v, t + t|v 0 , t) dv 0 , t ⌧c . (11.1.8)
In equation (11.1.8), the transition probability p1|1 (v, t + t|v 0 , t) does not depend
separately on times t and t + t, but only on their di↵erence3 t.
It is possible, under certain conditions, to deduce from the integral equation
(11.1.8) a partial di↵erential equation for f (v, t). For this purpose, we carry out a
systematic expansion of equation (11.1.8), called the Kramers–Moyal expansion. This
procedure relies on the physical idea that the Brownian particle’s velocity variations
produced by the collisions with the lighter fluid’s molecules are weak in relative value.
We thus get from equation (11.1.8) a partial di↵erential equation involving, besides
the time-derivative @f (v, t)/@t, partial derivatives of f (v, t) of any order with respect
to v. Owing to certain hypotheses, it is possible to keep in this equation only the two
first derivatives of f (v, t) with respect to v. The evolution equation thus obtained is
the Fokker–Planck equation. It represents a particular form of master equation.
modify the velocity of the particle. However, the Brownian particle is much heavier
than the fluid’s molecules, so that the relative variations of its velocity produced by
the collisions are small, at least as we consider the evolution of the velocity over a
time interval t ⌧ ⌧r .
Thus, considering equation (11.1.8) over such a time interval, we have to take
into account the fact that the velocity variation w = v v 0 is much smaller than the
velocity v.
For t ⌧c , we can use the simplified expression g(⌧ ) = 2Dm2 (⌧ ) for the Langevin
force autocorrelation function. We then have:
( 0 00
⌦ 0 ↵ 2Dme (t t ) , t00 < t0
v(t )F (t ) =
00
(11.2.4)
0, t00 > t0 .
From equations (11.2.2) and (11.2.3), the variance of w is:
2 ⇥ 2⇤
hw2 i hwi = 2D t + O ( t) . (11.2.5)
The Langevin force being Gaussian, the velocity variation w computed from the
integrated Langevin equation (11.2.1) is itself a Gaussian random variable.4 The tran-
sition probability p1|1 , considered as a function of w at fixed v, is thus a Gaussian law.
It is characterized by the mean and the variance of w, whose expressions for t ⌧ ⌧r
are given by formulas (11.2.2) and (11.2.5). These quantities do not depend separately
on t and t + t, but only on the time interval t. The same is true of the transition
probability, which can thus be written as a function5 p1|1 (w, v, t):
" #
2
1/2 (w + v t)
p1|1 (w, v, t) = 4⇡D t exp , ⌧c ⌧ t ⌧ ⌧r .
4D t
(11.2.8)
When p1|1 has the Gaussian form (11.2.8), only the coefficients M1 and M2 di↵er from
zero, whereas the coefficients Mn with n > 2 vanish.
Importing the series expansion (11.2.10) into the right-hand side of equation (11.2.9),
we obtain the relation:
Z
@n h i
1
X ( 1)n
f (v, t + t) = wn n f (v, t)p1|1 (w, v, t) dw, (11.2.11)
n=0
n! @v
4
See Supplement 11B.
5
Even if we do not assume that F (t) is a Gaussian process, the central limit theorem indicates
that p1|1 (w, v, t), considered as a function of w at fixed v, is a Gaussian law provided that we have
t ⌧c . The velocity variation during the time interval t results in fact from a very large number
of statistically independent velocity variations due to the many collisions that the Brownian particle
undergoes during the time interval t.
6
Note that the dependence of p1|1 with respect to its first argument w must be kept as it is.
Actually, it is not possible to make an expansion with respect to this argument since p1|1 (w, v, t)
varies rapidly with w.
282 Brownian motion: the Fokker–Planck equation
which also reads, with the aid of the moments hwn i of p1|1 (w, v, t):
( 1)n @ n h n i
X1
f (v, t + t) = hw if (v, t) , t ⌧c . (11.2.12)
n=0
n! @v n
Formula (11.2.12) is the Kramers–Moyal expansion of f (v, t + t). This expansion was
established by H.A. Kramers in 1940 and J.E. Moyal in 1949.
For t ⌧ ⌧r , the moments hwn i are proportional to t (formula (11.2.6)). Keep-
ing only the terms of order t, we get from equation (11.2.12):
X1
( 1)n @ n
f (v, t + t) f (v, t) = Mn f t. (11.2.13)
n=1
n! @v n
Taking formally the limit t ! 0, we deduce from equation (11.2.13) the partial
di↵ferential equation obeyed by f (v, t):
X1
@f ( 1)n @ n
= Mn f . (11.2.14)
@t n=1
n! @v n
@f @ 1 @2
= M1 f + M2 f , (11.3.1)
@t @v 2 @v 2
which has the form of a generalized7 di↵usion equation in the velocity space.
It may also be written in the form of a continuity equation,
@f @J
+ = 0, (11.3.3)
@t @v
in which the probability current J is defined by:
@f
J= vf D · (11.3.4)
@v
The evolution of the solution of the Fokker–Planck equation is thus described by the
hydrodynamic picture of a continuous flow in the velocity space. As displayed by
formula (11.3.4), the corresponding probability current is the sum of a ‘convective’
current vf and a ‘di↵usive’ current D@f /@v.
f (v, 0) = (v v0 ). (11.3.6)
We introduce the Fourier transform9 with respect to v of f (v, t), defined by:
Z 1
f (⇠, t) = f (v, t)ei⇠v dv. (11.3.7)
1
7
Equation (11.3.2) indeed contains, as well as the di↵usion term @ 2 [Df (v, t)]/@v 2 , the drift term
@[ vf (v, t)]/@v.
8
This property disappears in higher dimensions, where there are stationary states with non-
vanishing current.
9
For the sake of simplicity, we use the same notation f (. , t) for the distribution function f (v, t)
and its Fourier transform f (⇠, t).
284 Brownian motion: the Fokker–Planck equation
@f (⇠, t) @f (⇠, t)
+ ⇠ = D⇠ 2 f (⇠, t). (11.3.9)
@t @⇠
@f @f
dt + d⇠ = df, (11.3.10)
@t @⇠
setting:
dt d⇠ df
= = · (11.3.11)
1 ⇠ D⇠ 2 f
From equations (11.3.11), we get:
⇣ D ⇠2 ⌘
⇠ = ⇠1 e t , f = f1 exp · (11.3.12)
2
The quantity ⇠e t is thus a first integral (that is, a quantity conserved in the course
of the evolution). The general solution of equation (11.3.9) is of the form:
⇣ D ⇠2 ⌘
f (⇠, t) = ⇠e t
exp · (11.3.13)
2
In equation (11.3.13), stands for a function which must be chosen so that f (⇠, t)
obeys the initial condition (11.3.8):
⇣ D ⇠2 ⌘
(⇠) = ei⇠v0 exp · (11.3.14)
2
This gives:
D ⇠2
f (⇠, t) = exp i⇠e t
v0 exp 1 e 2 t
. (11.3.15)
2
The fundamental solution f (v, t) is obtained from expression (11.3.15) for f (⇠, t)
by inverse Fourier transformation:
⇣ ⌘1/2 2
1/2 (v v0 e t )
f (v, t) = (1 e 2 t
) exp , t > 0.
2⇡D 2D 1 e 2 t
(11.3.16)
Brownian motion and Markov processes 285
and of variance:
D
v (t) = (1 ). (11.3.18)
2 2 t
e
These results for the mean value and the variance of the velocity coincide with the
formulas deduced directly from the Langevin equation. At large times ( t 1), the
fundamental solution approaches the stationary distribution as given by:
1/2
f (v) = ( /2⇡D) exp( v 2 /2D). (11.3.19)
As shown by formula (11.1.4), the fundamental solution f (v, t) identifies with the
transition probability p1|1 (v, t|v0 , t0 = 0). In the case t ⌧ 1, we actually recover10
from formula (11.3.16) the expression (11.2.8) for p1|1 .
process, even if the random force corresponds to Gaussian white noise. However, in the
viscous or overdamped limit, the inertia term of the Langevin equation is discarded,
and the equation of motion takes the form of a first-order di↵erential equation for the
displacement:
dx
⌘ = F (t). (11.4.1)
dt
Equation (11.4.1) is valid over an evolution time interval t ⌧r . When the random
force corresponds to white noise, we have the decorrelation property:
⌦ ↵
x(t)F (t0 ) = 0, t0 > t, (11.4.2)
@p(x, t) @ 2 p(x, t)
=D · (11.4.4)
@t @x2
Equation (11.4.4) for p(x, t), analogous to the Fokker–Planck equation for f (v, t),
is nothing but the usual di↵usion equation.11 Its fundamental solution for the initial
condition p(x, 0) = (x x0 ) is a Gaussian law of mean x0 and of variance x2 (t) = 2Dt:
2
(x x0 )
p(x, t) = (4⇡Dt)
1/2
exp , t > 0. (11.4.5)
4Dt
d2 x dx
m +m + m!02 x = F (t), (11.4.6)
dt2 dt
where it is assumed that F (t) corresponds to Gaussian white noise. Equation (11.4.6)
being a second-order di↵erential equation, neither the displacement x(t) x0 of the
oscillator nor its velocity v(t) = dx(t)/dt are Markov processes. On the other hand,
the two-dimensional process {x(t), v(t)} is Markovian. It can indeed be described by
a set of two first-order di↵erential equations,
8
> dx(t)
>
< = v(t)
dt
(11.4.7)
>
>
: m dv(t) + m v(t) + m! 2 x(t) = F (t),
0
dt
and studied with the aid of arguments analogous to those developed for the velocity
in the case of a free Brownian particle. We thus can show that the joint distribution
function f (x, v, t) obeys a Fokker–Planck equation.
Generally speaking, when a process is not Markovian, it may be considered as
a ‘projection’ of a more complex Markov process (that is, a Markov process with a
higher number of dimensions), obtained by introducing additional variables in the
description. These additional variables serve to describe explicitly information which
otherwise would be implicitly contained in the past values of the relevant variables.
Obviously, this procedure is of practical interest only if the additional variables are
in limited number. In the case of the Brownian motion of a harmonic oscillator, it
is enough to consider both variables x(t) and v(t) to reduce the study to that of a
two-dimensional Markov process.
288 Brownian motion: the Fokker–Planck equation
Bibliography
P.M. Chaikin and T.C. Lubensky, Principles of condensed matter physics, Cam-
bridge University Press, Cambridge, 1995.
C.W. Gardiner, Handbook of stochastic methods, Springer-Verlag, Berlin, third edi-
tion, 2004.
N.G. van Kampen, Stochastic processes in physics and chemistry, North-Holland,
Amsterdam, third edition, 2007.
R. Kubo, M. Toda, and N. Hashitsume, Statistical physics II: nonequilibrium
statistical mechanics, Springer-Verlag, Berlin, second edition, 1991.
R.M. Mazo, Brownian motion: fluctuations, dynamics, and applications, Oxford
University Press, Oxford, 2002.
F. Reif, Fundamentals of statistical and thermal physics, McGraw-Hill, New York,
1965.
R. Zwanzig, Nonequilibrium statistical mechanics, Oxford University Press, Oxford,
2001.
References
A. Einstein, Über die von der molekularkinetischen Theorie der Wärme geforderte
Bewegung von in ruhenden Flüssigkeiten suspendierten Teilchen, Annalen der Physik
17, 549 (1905).
M. Smoluchowski, Zur kinetischen Theorie der Brownschen Molekularbewegung
und der Suspensionen, Annalen der Physik , 21, 756 (1906).
A.D. Fokker, Die mittlere Energie rotierender elektrischer Dipole im Strahlungsfeld,
Annalen der Physik 43, 810 (1914).
M. Planck, An essay on statistical dynamics and its amplification in the quantum
theory (translated title), Sitzungsberichte der preuss. Akademie der Wissenschaften
324 (1917).
G.E. Uhlenbeck and L.S. Ornstein, On the theory of the Brownian motion, Phys.
Rev. 36, 823 (1930). Reprinted in Selected papers on noise and stochastic processes
(N. Wax editor), Dover Publications, New York, 2003.
H.A. Kramers, Brownian motion in a field of force and the di↵usion model of chem-
ical reactions, Physica 7, 284 (1940).
M.C. Wang and G.E. Uhlenbeck, On the theory of the Brownian motion II, Rev.
Mod. Phys. 17, 323 (1945). Reprinted in Selected papers on noise and stochastic pro-
cesses (N. Wax editor), Dover Publications, New York, 2003.
Bibliography 289
J.E. Moyal, Stochastic processes and statistical physics, J. Royal Statist. Soc. Ser. B
11, 150 (1949).
R. Kubo, The fluctuation-dissipation theorem and Brownian motion, 1965, Tokyo
Summer Lectures in Theoretical Physics (R. Kubo editor), Syokabo, Tokyo and Ben-
jamin, New York, 1966.
R. Kubo, The fluctuation-dissipation theorem, Rep. Prog. Phys. 29, 255 (1966).
Supplement 11A
Random walk
dpn (t) ⇥ ⇤
= w pn+1 (t) + pn 1 (t) 2pn (t) . (11A.1.3)
dt
1
In equation (11A.1.3), the quantity w = (2 t) represents the transition probability
per unit time between a given site and one or the other of its two nearest neighbors.
From the very definition of the model, it results that the displacement of the
drunken walker is a Markov process. Equation (11A.1.3) is the master equation gov-
erning the evolution of the probabilities pn (t).
The solution pn (t) of the master equation (11A.1.3) for the initial condition (11A.2.1)
can be expressed with the aid of the nth-order modified Bessel function of the first
kind:
pn (t) = e 2wt I|n| (2wt). (11A.2.2)
On account of the properties of I|n| , we can deduce from formula (11A.2.2) the asymp-
totic behavior of pn (t) (t ! 1, n ! 1 with n2 /t fixed):
⇣ n2 ⌘
1/2
pn (t) ' (4⇡wt) exp · (11A.2.3)
4wt
We will be concerned with the di↵usion properties of the particle. The two lowest-
order moments of its displacement are defined by the formulas:
1
X 1
X
⌦ ↵ ⌦ ↵
x(t) = a npn (t), x2 (t) = a2 n2 pn (t), (11A.2.4)
n= 1 n= 1
In the symmetric random walk, the initial condition as well as the master equation
being invariant under the transformation n 1 ! n + 1, we have, for any value of t:
⌦ ↵
x(t) = 0. (11A.2.5)
The second moment of the particle’s displacement obeys the evolution equation:
⌦ ↵ 1
d x2 (t) X ⇥ ⇤
= wa2 n2 pn+1 (t) + pn 1 (t) 2pn (t) , (11A.2.6)
dt n= 1
with the initial condition: ⌦ ↵
x2 (0) = 0. (11A.2.7)
The Pbracketed quantity on the right-hand side of equation (11A.2.6) is simply equal
1
to 2 n= 1 pn (t), that is, equal to 2 since the sum of probabilities is conserved. To
determine hx2 (t)i, we must therefore solve the di↵erential equation:
⌦ ↵
d x2 (t)
= 2wa2 , (11A.2.8)
dt
with the initial condition (11A.2.7). This gives, at any time t:
hx2 (t)i = 2wa2 t. (11A.2.9)
The motion of the particle (random walker) is thus di↵usive. The di↵usion coefficient is:
D = wa2 . (11A.2.10)
⇣ x2 ⌘
p(x, t) = (4⇡Dt)
1/2
exp , t > 0. (11A.3.4)
4Dt
Bibliography
References
A. Einstein, Über die von der molekularkinetischen Theorie der Wärme geforderte
Bewegung von in ruhenden Flüssigkeiten suspendierten Teilchen, Annalen der Physik
17, 549 (1905).
M. Kac, Random walk and the theory of Brownian motion, American Mathematical
Monthly 54, 369 (1947). Reprinted in Selected papers on noise and stochastic processes
(N. Wax editor), Dover Publications, New York, 2003.
Supplement 11B
Brownian motion:
Gaussian processes
The limit T ! 1 is taken at the end of the calculations. Each Fourier coefficient an
is a linear combination of the {y(t)}’s with 0 t T :
Z
1 T
an = y(t)ei!n t dt. (11B.1.2)
T 0
We also write:
1
X
Y (t) = An e i!n t
, 0 t T, (11B.1.3)
n= 1
with: Z
1 T
An = Y (t)ei!n t dt. (11B.1.4)
T 0
The coefficients An , which are linear superpositions of Gaussian random variables, are
themselves Gaussian random variables.
Gaussian Markov stationary processes 295
The stationarity of the process Y (t) implies that its Fourier coefficients of unequal
angular frequencies are uncorrelated:
⌦ ↵ ⌦ 2↵
An A⇤n0 = |An | n,n0 . (11B.1.5)
Since they are Gaussian, they are statistically independent. The Fourier coefficients of a
stationary Gaussian random process are thus independent Gaussian random variables.
1/2 y12 /2
p1 (y1 , t1 ) = (2⇡) e . (11B.2.1)
We have ⇢ii = 1 and ⇢ij = ⇢ji . The covariance matrix is thus of the form:
!
1 ⇢12
M2 = , (11B.2.3)
⇢12 1
gives:
1 1/2 1
p2 (y1 , t1 ; y2 , t2 ) = (2⇡) Det A2 exp (a11 y12 + a22 y22 + 2a12 y1 y2 ) . (11B.2.6)
2
Since: !
1 1 ⇢12
A2 = (11B.2.7)
1 ⇢212 ⇢12 1
and:
1
Det A2 = , (11B.2.8)
1 ⇢212
formula (11B.2.6) also reads:
2
1 1/2 y1 + y22 2⇢12 y1 y2
p2 (y1 , t1 ; y2 , t2 ) = (2⇡) (1 ⇢212 ) exp · (11B.2.9)
2(1 ⇢212 )
For t1 < t2 , the conditional probability p1|1 (y2 , t2 |y1 , t1 ) is obtained by dividing
p2 (y1 , t1 ; y2 , t2 ) (formula (11B.2.9)) by p1 (y1 , t1 ) (formula (11B.2.1)):
2
1/2 (y2 y1 ⇢12 )
p1|1 (y2 , t2 |y1 , t1 ) = [2⇡(1 ⇢12 )]
2
exp , t1 < t2 . (11B.2.10)
2(1 ⇢212 )
⇢13 ⇢23 1
If we set: 0 1
A11 A12 A13
B C
A3 = B
@ A12 A22 A23 C
A, (11B.2.13)
A13 A23 A33
formula (11B.2.11) can be rewritten as:
3/2 1/2
p3 (y1 , t1 ; y2 , t2 ; y3 , t3 ) = (2⇡) Det A3
1
⇥ exp (A11 y12 + A22 y22 + A33 y32 + 2A12 y1 y2 + 2A13 y1 y3 + 2A23 y2 y3 ) .
2
(11B.2.14)
Application to Brownian motion 297
For t1 < t2 < t3 , the conditional probability p1|2 (y3 , t3 |y2 , t2 ; y1 , t1 ) may be obtained
by dividing p3 (y1 , t1 ; y2 , t2 ; y3 , t3 ) by p2 (y1 , t1 ; y2 , t2 ).
For the process Y (t) to be Markovian, it is necessary for the probability
p1|2 (y3 , t3 |y2 , t2 ; y1 , t1 ) = p3 (y1 , t1 ; y2 , t2 ; y3 , t3 )/p2 (y1 , t1 ; y2 , t2 ) to be independent of
y1 . From the comparison between formulas (11B.2.6) and (11B.2.14) for the two-time
and three-time probabilities, it follows that the absence of a term in y1 y3 in p2 implies
that we must have A13 = 0. As for the terms in y12 and in y1 y2 , they will cancel after
the division of p3 by p2 provided that we have:
1 ⇢12
A11 = a11 = , A12 = a12 = · (11B.2.15)
1 ⇢212 1 ⇢212
Since A13 is proportional to the minor of ⇢13 in M3 , the condition A13 = 0 yields:
We can show that, if condition (11B.2.16) is satisfied, it is also the case of conditions
(11B.2.15).
g(⌧ ) = e |⌧ |
, > 0, (11B.2.18)
dv
m + m v = F (t), (11B.3.1)
dt
where the Langevin force F (t) is a random stationary process. We assume that the
spectral density of F (t) is independent of the angular frequency (white noise), which
amounts to consider that the corresponding autocorrelation function is a delta func-
tion. The process F (t) is then completely random. We make the supplementary hy-
pothesis that it is a Gaussian process.
298 Brownian motion: Gaussian processes
We assume that the Brownian particle has been in contact with the bath for a
sufficiently long time to find itself in equilibrium at any finite time t. Its velocity v(t)
is then a random stationary process.
Since F (t) is a Gaussian random process, F (!) is a Gaussian random variable. The
same is true of v(!). Therefore, v(t) is a Gaussian random process, termed the
Ornstein–Uhlenbeck process, entirely characterized by its autocorrelation function
hv(t)v(t + ⌧ )i.
We deduce from the expressions (11B.3.5) and (11B.3.6) for p1 (v1 ) and p2 (v1 , t1 ; v2 , t2 )
the conditional probability of the Ornstein–Uhlenbeck process:
⇣ m ⌘1/2 ⇥ ✓ (t2 t1 ) 2
◆
⇤ 1/2 m [v2 v1 e ]
p1|1 (v2 , t2 |v1 , t1 ) = 1 e 2 (t2 t1 )
exp ,
2⇡kT 2kT 1 e 2 (t2 t1 )
(11B.3.7)
where t1 < t2 .
The velocity distribution function at time t, obtained starting from an initial
distribution (v v0 ), is thus:
⇣ m ⌘1/2 2
1/2 m (v v0 e t )
f (v, t) = 1 e 2 t
exp · (11B.3.8)
2⇡kT 2kT 1 e 2 t
Bibliography
References
J.L. Doob, The Brownian movement and stochastic equations, Ann. Math. 43, 351
(1942). Reprinted in Selected papers on noise and stochastic processes (N. Wax edi-
tor), Dover Publications, New York, 2003.
M.C. Wang and G.E. Uhlenbeck, On the theory of the Brownian motion II, Rev.
Mod. Phys. 17, 323 (1945). Reprinted in Selected papers on noise and stochastic pro-
cesses (N. Wax editor), Dover Publications, New York, 2003.
Chapter 12
Linear responses
and equilibrium correlations
The field a(t) is thus coupled to a conjugate physical quantity A. For instance, in
the case of a perturbation induced by an electric field, the quantity conjugate to the
field is the electric polarization,2 whereas, in the case of a perturbation induced by a
magnetic field, the quantity conjugate to the field is the magnetization.3
We are interested in a physical quantity B, of equilibrium average hBi. We want
to determine the response of B to the field a(t), in other words, to compute at any
time the modification hB(t)ia = hB(t)ia hBi of the average of B due to the applied
field. For the sake of simplicity, we assume that B is centered (hBi = 0), which allows
us to identify hB(t)ia and hB(t)ia . In the linear range, the out-of-equilibrium average
of B is written in the form:
Z 1
⌦ ↵
B(t) a = ˜BA (t, t0 )a(t0 ) dt0 , (12.1.2)
1
where the real quantity ˜BA (t, t0 ) is a linear response function depending only on the
properties of the unperturbed system.
The response function ˜BA (t) thus represents the impulsional response.5
• Causal character
The causality principle is a commonly admitted physical principle which states that
any physical e↵ect must follow in time the cause which produces it. A modification of
the applied field taking place at a time t0 may thus lead to a modification of hB(t)ia
only at times t > t0 . In other words, the response function ˜BA (t, t0 ) is causal, which
means that it may be non-vanishing only for t > t0 . The actual upper bound in the
integral on the right-hand side of formulas (12.1.2) and (12.1.3) is thus t, and not +1.
2. Generalized susceptibilities
2.1. Linear response to a harmonic perturbation
The linear response of A to the harmonic field a(t) = <e(ae i!t ) reads:
Z 1
⌦ ↵ 0
A(t) a = ˜(t t0 ) <e(ae i!t ) dt0 , (12.2.1)
1
Since the argument of ˜(t t0 ) must be positive, the upper bound of the integral
on the right-hand side of equations (12.2.1) and (12.2.2) is in fact t. Setting t t0 = ⌧ ,
gives: Z 1
⌦ ↵
A(t) a = <e ae i!t ˜(⌧ )ei!⌧ d⌧ . (12.2.3)
0
5
It is also called the retarded Green’s function.
6
The general theory of linear response will be expounded in Chapter 13, where in particular the
Kubo formulas enabling us to compute explicitly the linear response functions in terms of equilibrium
correlation functions of the relevant dynamical variables will be established.
304 Linear responses and equilibrium correlations
With the aid of the generalized susceptibility (!), defined as the Fourier transform
of the causal function ˜(t), that is, by the formula:
Z 1
(!) = ˜(t)ei!t dt, (12.2.4)
0
⌦ ↵
A(!) a = a(!) (!). (12.2.7)
and by letting ✏ tend towards zero at the end of the calculation. We then set:
The generalized susceptibility is thus defined as the Fourier transform in the distribu-
tion sense of the response function.
7
For the sake of simplicity, we use the same notation a(.) for the field a(t) and its Fourier transform
a(!), as well as the same notation hA(.)ia for the response hA(t)ia and its Fourier transform hA(!)ia .
Generalized susceptibilities 305
It is useful to introduce the function (z) defined for a complex argument z with
=m z > 0 as the Fourier–Laplace transform of ˜(t):
Z 1
(z) = ˜(t)eizt dt, =m z > 0. (12.2.10)
0
The function (z) is analytic in the upper complex half-plane. We assume in the rest
of this chapter8 that (z) ! 0 as z ! 1. As displayed by formula (12.2.9), the
generalized susceptibility (!) is the limit of (z) as the point of affix z = ! + i✏ in
the upper complex half-plane tends towards the point of abscissa ! on the real axis.
The function 0 (!) is the Fourier transform of the even part ˜p (t) = 12 [ ˜(t) + ˜( t)]
of ˜(t) and (up to a factor 1/i) the function 00 (!) is the Fourier transform of its
odd partR ˜i (t) = 12 [ ˜(t) ˜( t)]. Setting
R 1 ˜00p (t) = ˜0 (t) and ˜i (t) = i ˜00 (t), we have
1
0
(!) = 1 ˜ (t)e dt and (!) = 1 ˜ (t)e dt.
0 i!t 00 i!t
where ⇥(t) denotes the Heaviside function. Then, coming back to formula (12.2.10),
valid for =m z > 0, we can write:
Z 1 Z 1
d! 00
(z) = 2i dt eizt
(!)e i!t , =m z > 0. (12.2.13)
0 1 2⇡
The integration over time leads to the spectral representation of (z) in terms of
00
(!):
Z
1 1 00 (!)
(z) = d!. (12.2.14)
⇡ 1 ! z
8
Supplementary information on the behavior of (z) when z ! 1 is given in the appendix at
the end of this chapter.
306 Linear responses and equilibrium correlations
Formula (12.2.14) displays the fact that (z) exhibits singularities only on the real
axis. It is thus an analytic function of z in the upper complex half-plane.9
• Spectral representation of (z) in terms of 0
(!)
We can alternatively represent ˜(t) with the aid of its even part:
and: Z 1
1 0
(! 0 )
(! + i✏) = d! 0 . (12.3.2)
i⇡ 1 !0 (! + i✏)
9
The spectral representation (12.2.14) may be used to define (z) at any point of affix z situated
outside the real axis. It defines an analytic function of z either in the upper complex half-plane or in
the lower complex half-plane (see Chapter 13), in contrast to the integral definition (12.2.10) valid
only for =m z > 0.
10
See Section 4.
11
An alternative way, very commonly used, to establish the Kramers–Kronig relations, will be
presented in appendix at the end of this chapter.
Dissipation 307
8 Z 1
> 1 (! 0 ) 0
00
>
>
0
(!) = vp d!
< ⇡ 1 !0 !
Z (12.3.3)
>
> 1 1
(! 0 )
0
>
: 00
(!) = vp d! 0 .
⇡ ! 0 !
1
4. Dissipation
The instantaneous power absorbed by the system submitted to the field a(t) is:
⌦ ↵
dW d A(t) a
= a(t) · (12.4.1)
dt dt
The linear response hA(t)ia to the harmonic field a(t) = a cos !t is given by formula
(12.2.5). The corresponding instantaneous absorbed power is thus:
dW ⇥ ⇤
= a2 ! cos !t 0
(!) sin !t + 00
(!) cos !t . (12.4.2)
dt
The average power absorbed over one period (or over an integer number of periods) is
given by the formula:
dW 1 (12.4.3)
= a2 ! 00
(!).
dt 2
12
We make use of the relation:
1 1
lim = vp i⇡ (x),
✏!0+ x + i✏ x
The average energy absorption rate in the system is thus related to the imaginary part
of the generalized susceptibility. The energy supplied by the external field is eventually
dissipated irreversibly within the system, that is, transformed into heat.14 For this
reason, 00 (!) is qualified as the dissipative part of the generalized susceptibility.
At thermodynamic equilibrium, the average dissipated power is positive.15 As a
consequence, the quantity ! 00 (!) is positive.
5. Non-uniform phenomena
5.1. Linear response functions
We are still considering here a physical system at thermodynamic equilibrium. We
submit it to an inhomogeneous external field a(r, t). The perturbation Hamiltonian
may in principle be written in a form generalizing the Hamiltonian (12.1.1) of the
homogeneous case: Z
H1 (t) = dr a(r, t)A(r). (12.5.1)
In formula (12.5.1), A(r) is the physical quantity conjugate to the field a(r, t) and
the spatial integration is carried over the whole volume of the system. In some cases
however, the perturbation Hamiltonian involves more naturally, instead of the applied
field, the potential from which this latter derives. Let us take the example of a sys-
tem of electrons perturbed by a non-uniformR electric field E(r, t) = rP(r, t). The
perturbation Hamiltonian reads H1 (t) = dr (r, t)⇢(r), where ⇢(r) = e i (r ri )
is the charge density at point r (e denotes the electron charge and the {ri }’s are the
positions of the di↵erent electrons). The charge density ⇢(r) is thus coupled to the
potential (r, t).
In an inhomogeneous situation, to determine the response of the physical quantity
B(r) to the field a(r, t) amounts to computing the modification hB(r, t)ia of the
average value of B(r) due to the applied field. We assume that B(r) is centered. In
the linear range, its out-of-equilibrium average reads:
Z Z 1
⌦ ↵
B(r, t) a = dr 0
˜BA (r, t ; r 0 , t0 )a(r 0 , t0 ) dt0 . (12.5.2)
1
The real quantity ˜BA (r, t ; r 0 , t0 ) is a non-local linear response function. The spatial
integration on the right-hand side of formula (12.5.2) is carried out over the whole
volume of the system.
As in the homogeneous case, the linear response function is invariant under time-
translation and is also causal. Moreover, if the unperturbed system is invariant under
14
A simple example of a dissipative system is a harmonic oscillator damped by viscous friction
(see Supplement 12A). In this case, the energy supplied by the field is transferred to the incoherent
degrees of freedom of the damping fluid.
15
This property will be demonstrated in Chapter 14 with the aid of the microscopic expression
for the average dissipated power.
Non-uniform phenomena 309
space translation, which we assume to be the case in the following of this section, the
response function ˜BA (r, t ; r 0 , t0 ) does not depend separately on the two arguments r
and r 0 , but only on the di↵erence r r 0 .
As in the uniform case, we limit ourselves here to the study of the linear response
of a physical quantity A(r) to its own conjugate field a(r, t), denoting simply by ˜(r, t)
the corresponding response function.
More generally, the response hA(r, t)ia to an applied field of Fourier transform
a(q, !) has the spatial and temporal Fourier transform:
⌦ ↵
A(q, !) a = a(q, !) (q, !). (12.5.8)
The study of the generalized susceptibility (q, !) at given wave vector q can be
carried out in the same way as the study of (!). Using notations extending those of
the homogeneous case, we can in particular write the spectral representation:17
Z 1
1 (q, !)
00
(q, z) = d!, (12.5.9)
⇡ 1 ! z
16
Note that, because of the causality principle, the spatial and temporal variables do not play
equivalent roles as far as the response function is concerned. The spatial Fourier transform of the
response function is a true Fourier transform, whereas the so-called temporal ‘Fourier’ transform is
in fact a Fourier–Laplace transform (which must be taken in the distribution sense).
17
As in the uniform case, this spectral representation can be extended to any point of affix z
situated outside the real axis.
310 Linear responses and equilibrium correlations
6.1. Definition
The system at equilibrium is described by a time-independent Hamiltonian H0 . A
dynamical variable or observable A(r) possibly depending on the space point r is
associated with each physical quantity
P of the system. An example of such an observable
is the particle density n(r) = i (r ri ).
In classical mechanics, a dynamical variable A(r, t) is a function of the general-
ized coordinates and momenta, which evolve according to the Hamilton’s equations
governed by H0 . In quantum mechanics, we associate with an observable a Hermitean
operator A(r, t) = eiH0 t/h̄ A(r)e iH0 t/h̄ , which evolves according to the Heisenberg
equation:
dA(r, t)
ih̄ = [A(r, t), H0 ]. (12.6.1)
dt
Consider two classical operators18 (dynamical variables) Ai (r, t) and Aj (r, t),
assumed to be centered for the sake of simplicity. Their equilibrium correlation function
18
In the quantum case, because of the non-commutativity of the operators, there are several
non-equivalent definitions of the correlation function of two operators Ai (r, t) and Aj (r, t). It is
even the case for the autocorrelation function of an operator A(r, t), since A(r, t) does not in gen-
eral commute with A(r 0 , t0 ). We use in particular the symmetric correlation function S̃Ai Aj (r, t) =
1
2
hAi (r, t)Aj +Aj Ai (r, t)i, as well as, for a system in canonical equilibrium, the Kubo canonical corre-
R
lation function K̃Ai Aj (r, t) = 1
0 he
H0 A e
j
H0 A (r, t)i d , where
i = (kT ) 1 . In the classical
limit, S̃Ai Aj (r, t) and K̃Ai Aj (r, t) both reduce to C̃Ai Aj (r, t) (see Chapter 14). In the present dis-
cussion, we use the notation C̃ for the correlation function (in the quantum case, C̃ stands in fact for
either S̃ or K̃).
Equilibrium correlation functions 311
⌦ ↵
C̃Ai Aj (r, t ; r 0 , t0 ) = Ai (r, t)Aj (r 0 , t0 ) , (12.6.2)
Inversely, we write: Z 1
1
A(r, t) = A(r, !)e i!t
d!. (12.6.5)
2⇡ 1
Using the fact that the autocorrelation function C̃AA (r, t ; r 0 , t0 ) does not depend
separately on t and on t0 , but only on the di↵erence t t0 , we can show that:
⌦ ↵
A(r, !)A(r 0 , ! 0 ) = 2⇡ (! + ! 0 )CAA (r, ! ; r 0 ). (12.6.6)
In formula (12.6.6), CAA (r, ! ; r 0 ) denotes the temporal Fourier transform of the au-
tocorrelation function C̃AA (r, t; r 0 , 0), defined by:20
Z 1
CAA (r, ! ; r ) =
0
C̃AA (r, t ; r 0 , 0)ei!t dt. (12.6.7)
1
The function CAA (r, ! ; r) is called the power spectrum or the spectral density 21
of the fluctuations of A(r, t). Formula (12.6.7) constitutes the generalization of the
19
For the sake of simplicity, we use the same notation A(r, .) for the operator A(r, t) and its
temporal Fourier transform A(r, !).
20
In contrast to the response functions, the correlation functions are not causal. The temporal
Fourier transform involved in definition (12.6.7) is thus a true Fourier transform, and not a Fourier–
Laplace transform.
21
It remains to be proved that this quantity is actually positive. This property, which is a conse-
quence of the fluctuation-dissipation theorem, will be established in Chapter 14.
312 Linear responses and equilibrium correlations
We then
R compute its Fourier transform by introducing the supplementary integration
V 1 dr 0 = 1 (V denotes the volume of the system), which gives:
Z
1⌦ ↵
C̃Ai Aj (r, t)e iq.r dr = Ai (q, t)Aj ( q, 0) . (12.6.9)
V
The spatial and temporal Fourier transform CAi Aj (q, !) of the correlation func-
tion C̃Ai Aj (r, t) is defined by:
Z Z 1
CAi Aj (q, !) = dr e iq.r
dt ei!t C̃Ai Aj (r, t). (12.6.11)
1
Using formula (12.6.9), gives the relation (valid in a translationally invariant medium):
Z 1
1 ⌦ ↵
CAi Aj (q, !) = Ai (q, t)Aj ( q, 0) ei!t dt. (12.6.12)
V 1
interaction a centered operator A(r). For instance, in the case of scattering of light
by a fluid in equilibrium,23 the radiation is scattered by the density fluctuations of
the fluid, and therefore the operator A(r) is proportional to the density fluctuation
n(r) = n(r) hni.
An incident plane-wave initial state |ki is scattered to a final state which, in the
framework of the Born approximation for scattering, is considered as being a plane-
wave state24 |k0 i. The matrix
R element of the interaction operator between states |ki
0
and |k0 i is hk0 |A(r)|ki = e ik .r A(r)eik.r dr: it is thus identical to the Fourier
component A( q) of the operator A(r) (having set q = k k0 ). At lowest perturbation
order, the probability per unit time of the process (|ki, ) ! (|k0 i, 0 ) is given by the
Fermi golden rule:
2⇡ 0 2
W(k0 , 0 ),(k, ) = |h |A( q)| i| (h̄! + " " 0 ). (12.6.13)
h̄
The probability per unit time Wk0 ,k of the process |ki ! |k0 i is obtained by weighting
W(k0 , 0 ),(k, ) by the occupation probability p of the initial state of the system at
equilibrium, and by summing over all initial states and all final states:
2⇡ X 2
Wk0 ,k = p |h 0 |A( q)| i| (h̄! + " " 0 ). (12.6.14)
h̄ 0 ,
When A(r) is the density fluctuation n(r), the scattering function S(q, !) is also
designated as the dynamical structure factor. This latter denomination will be adopted
generically in the rest of this chapter.
Introducing in formula (12.6.15) the Fourier representation of the delta function,
1R1
(!) = (2⇡) 1
ei!t dt, we can express S(q, !) with the aid of an autocorrelation
function. Since the operator A(r) is Hermitean, we have:
⇤
h 0 |A( q)| i = h |A(q)| 0 i. (12.6.16)
This gives:
Z 1 X
S(q, !) = p h |ei" t/h̄
A(q)e i" 0 t/h̄
| 0 ih 0 |A( q)| iei!t dt, (12.6.17)
1 , 0
23
See Supplement 16B.
24
This implies that the interaction of the radiation with the target has be weak enough in order
to allow us to neglect multiple scattering. If the Born approximation is not appropriate, we replace
the interaction potential A(r) by a convenient pseudopotential.
314 Linear responses and equilibrium correlations
that is:
Z 1 ⌦ ↵
S(q, !) = A(q, t)A( q, 0) ei!t dt. (12.6.18)
1
Equation (12.6.18) shows that the dynamical structure factor25 is the Fourier trans-
form of the autocorrelation function hA(q, t)A( q, 0)i. This property was demon-
strated by L. van Hove in 1954.
In the particular case of a translationally invariant medium, comparing formula
(12.6.18) with the formula (12.6.12) written for Ai = Aj = A gives:
It is then possible to deduce from the dynamical structure factor the autocorrelation
function C̃AA (r, t).
7.1. Stationarity
Any equilibrium correlation function depends only on the di↵erence of its two temporal
arguments. We therefore have the equality:
⌦ ↵ ⌦ ↵
A(q, t)A( q, 0) = A(q, t + t0 )A( q, t0 ) , (12.7.1)
⌦ ↵ ⌦ ↵
A(q, t)A( q, 0) = A(q, 0)A( q, t) . (12.7.2)
7.2. Time-derivation
We have the relation:
⌦ ↵ ⌦ ↵
Ȧ(q, 0)A( q, t) = A(q, 0)Ȧ( q, t) . (12.7.3)
25
Simple examples of dynamical structure factors are presented in Supplement 12C. The dynamical
structure factor in a fluid will be studied in Supplement 16B.
Properties of the equilibrium autocorrelation functions 315
To derive formula (12.7.3), we can for instance26 take the derivative of formula (12.7.1)
with respect to t0 :
⌧ ⌧
@A(q, t + t0 ) @A( q, t0 )
A( q, t0 ) + A(q, t + t0 ) = 0. (12.7.4)
@t0 @t0
Making t0 = 0 in equation (12.7.4), we first verify the property:
⌦ ↵ ⌦ ↵
Ȧ(q, t)A( q, 0) = A(q, t)Ȧ( q, 0) , (12.7.5)
then, translating all times by t and changing then t into t, formula (12.7.3).
⌦ ↵⇤ ⌦ ↵
A(q, t)A( q, 0) = A(q, 0)A( q, t) . (12.7.6)
or, changing t into t in the integral on the right-hand side of equation (12.7.8):
Z 1
⌦ ↵
S (q, !) =
⇤
A(q, 0)A( q, t) ei!t dt. (12.7.9)
1
Using the stationarity property (12.7.2), we verify that S ⇤ (q, !) = S(q, !).
26
We can alternatively demonstrate formula (12.7.3) by writing the Heisenberg equations rel-
ative to the operators A(q, t) and A( q, t), that is, ih̄Ȧ(q, t) = [A(q, t), H0 ] and ih̄Ȧ( q, t) =
[A( q, t), H0 ], and then using the invariance of the trace under a circular permutation of the op-
erators (together with the fact that the equilibrium density operator is a function of H0 ).
27
Note that the real character of S(q, !) is already apparent in formula (12.6.15).
316 Linear responses and equilibrium correlations
⌦ ↵ ⌦ ↵
A(q, 0)A( q, t) = A( q, t ih̄ )A(q, 0) . (12.7.10)
1 ⇥ iH0 t/h̄ ⇤
Tr e A( q, 0)e iH0 (t ih̄ )/h̄
A(q, 0) , (12.7.12)
Z
or:
1 ⇥ ⇤
Tr e H0 i(t ih̄ )H0 /h̄
e A( q, 0)e i(t ih̄ )H0 /h̄
A(q, 0) . (12.7.13)
Z
This last expression is simply hA( q, t ih̄ )A(q, 0)i, which demonstrates formula
(12.7.10).
This expression for S(q, !) must be compared with that for S( q, !):
Z 1 ⌦ ↵
S( q, !) = A( q, t)A(q, 0) e i!t
dt. (12.7.18)
1
where I is the rectangular contour of large sides equal to 2R represented in Fig. 12.1
(R will be made to tend towards infinity at the end of the calculation). The function
to be integrated is analytic in the domain interior to the contour.28 The integral I
thus vanishes according to Cauchy’s theorem.
y = Im τ
-R O R x = Re τ
-R-iħβ -R-iħβ
28
We have:
1 X
hA( q, ⌧ )A(q, 0)i = |A(q, 0) 0 |2 e " +i⌧ (" " 0)
.
Z 0 ,
(12.7.20)
S( q, !) = e h̄!
S(q, !), (12.7.22)
29
This hypothesis is physically meaningful in a system with an infinite number of degrees of
freedom. However, in systems with a finite number of degrees of freedom, the correlation functions
are oscillating functions, and this limit in principle does not exist. We can nevertheless define it by
treating correlation functions as distributions, which allows us to extend the detailed balance relation
to these systems.
Appendix 319
Appendix
I
(z)
J= dz, (12A.1)
J
z !
where ! denotes a real angular frequency. The function (z), analytic in the upper
complex half-plane, is defined by the Fourier–Laplace integral (12.2.10). The integra-
tion contour J shown in Fig. 12.2 avoids, by using a semicircle of radius ✏, the pole
of abscissa ! of the function (z)/(z !). The contour J is closed by a semicircle of
large radius R.
The function (z) being analytic in the upper complex half-plane, the integral J of
formula (12A.1) vanishes according to Cauchy’s theorem. The function (z) generally
vanishes as z tends towards infinity.30 The contribution of the semicircle of radius R to
the integral J thus vanishes in the limit R ! 1. By detailing the other contributions
to J, we get:
Z Z 1 Z
! "
(! 0 ) (! 0 ) 0
d! 0 + d! 0 + i (! + ✏ei✓ ) d✓ = 0. (12A.2)
1 !0 ! !+✏ !0 ! ⇡
In the limit ✏ ! 0+ , the third integral on the left-hand side of equation (12A.2) tends
towards i⇡ (!). This gives the formula:
Z 1
i (! 0 )
(!) = vp d! 0 . (12A.3)
⇡ 1 !0 !
Identifying separately the real and imaginary parts of both sides of equation (12A.3),
we get the Kramers–Kronig relations (formulas (12.3.3)).
30
As a simple example, we can quote the generalized susceptibility xx (z) of a harmonic oscillator,
damped or not, which decreases proportionally to z 2 as z tends towards infinity, except in the viscous
limit in which it decreases proportionally to z 1 (see Supplement 12A).
320 Linear responses and equilibrium correlations
Im z
•
-R O ω-ε ω ω+ε R Re z
Fig. 12.2 Contour J used for the derivation of the Kramers–Kronig relations.
The Kramers–Kronig relations need to be modified if (z) does not tend towards
zero but towards a finite constant 1 as z tends towards infinity.31 In such a case, the
out-of-equilibrium
R1 average of the physical quantity A involves, besides the retarded
term 1 ˜(t t0 )a(t0 )dt0 , an instantaneous contribution 1 a(t) (due to a term 1 (t)
in the response function). In this case,we can still write the Kramers–Kronig relations
by working on the di↵erence (z) 1:
8 Z 1
> 1 00
(! 0 ) 0
>
>
0
(!) 1 = vp d!
< ⇡ 1 !0 !
Z (12A.4)
>
> 1 1 0
(! 0 )
>
: 00
(!) = vp
1 0
d! .
⇡ 1 !0 !
31
An example of this type of behavior appears in dielectric relaxation (see Supplement 13A).
Bibliography 321
Bibliography
P.M. Chaikin and T.C. Lubensky, Principles of condensed matter physics, Cam-
bridge University Press, Cambridge, 1995.
S. Dattagupta, Relaxation phenomena in condensed matter physics, Academic Press,
Orlando, 1987.
D. Forster, Hydrodynamic fluctuations, broken symmetries, and correlation func-
tions, Westview Press, Boulder, 1995.
C. Kittel, Introduction to solid state physics, Wiley, New York, eighth edition, 2005.
R. Kubo, M. Toda, and N. Hashitsume, Statistical physics II: nonequilibrium
statistical mechanics, Springer-Verlag, Berlin, second edition, 1991.
L.D. Landau and E.M. Lifshitz, Electrodynamics of continuous media, Butterworth-
Heinemann, Oxford, second edition, 1984.
S.W. Lovesey, Condensed matter physics: dynamic correlations, The Benjamin/
Cummings Publishing Company, Reading, second edition, 1986.
P.C. Martin, Measurements and correlation functions, Les Houches Lecture Notes
1967 (C. De Witt and R. Balian editors), Gordon and Breach, New York, 1968.
G. Parisi, Statistical field theory, Westview Press, Boulder, 1998.
M. Plischke and B. Bergersen, Equilibrium statistical physics, World Scientific,
Singapore, third edition, 2006.
References
R. de L. Kronig, On the theory of dispersion of X-rays, J. Opt. Soc. Am. 12, 547
(1926).
H.A. Kramers, La di↵usion de la lumière par les atomes, Atti del Congresso Inter-
nazionale dei Fisici (Como), 2, 545, Zanichelli, Bologna, 1927.
Supplement 12A
Linear response
of a damped oscillator
p
ẋ = , ṗ = kx. (12A.2.2)
m
To determine the modes, we assume that both x(t) and p(t) are proportional to
e i!t
. This gives the characteristic equation:
k
!2 + = 0, (12A.2.3)
m
1
See the general theory of linear response expounded in Chapter 13.
Oscillator damped by viscous friction 323
d2 x
m + m!02 x = 0. (12A.2.5)
dt2
The Hamiltonian H0 (formula (12A.2.1)), Hamilton’s equations (12A.2.2), and the
equation of motion (12A.2.5) of the undamped oscillator are invariant under time-
reversal.
d2 x dx
m +m + m!02 x = 0. (12A.3.1)
dt2 dt
In contrast to equation (12A.2.5), equation (12A.3.1) is not time-reversal invariant.
This equation, which has been introduced phenomenologically, does not follow directly
(that is, without approximations) from a microscopic Hamiltonian.2 The viscous force
represents the mean e↵ect produced on the oscillator by its interaction with the many
incoherent degrees of freedom of the fluid which surrounds it. The energy of the damped
oscillator evolving according to equation (12A.3.1) tends to flow irreversibly towards
the modes of the fluid. This flow corresponds to a dissipation of the oscillator’s energy.
3.1. Modes
The characteristic equation determining the angular frequencies of the modes of the
damped oscillator reads:
! 2 i ! + !02 = 0. (12A.3.2)
Its solutions are: ✓ ◆1/2
2
! = ±!1 i , !1 = !02 · (12A.3.3)
2 4
2
The equation of motion (12A.3.1) can be derived in the framework of the Caldeira–Leggett
dissipation model. The Caldeira–Leggett Hamiltonian describes the global system made up of the
oscillator and the modes of the bath with which it is coupled. If these modes form a continuum, we
can obtain for the oscillator an irreversible equation of motion. In the Ohmic dissipation case, this
equation takes the form (12A.3.1) (see also about this question Supplement 14A).
324 Linear response of a damped oscillator
Two di↵erent types of damped motions may take place, depending on whether < 2!0
or > 2!0 .
4. Generalized susceptibility
We apply to the system an external force Fext (t). The equation of motion now reads:
d2 x dx
m 2
+m + m!02 x = Fext (t). (12A.4.1)
dt dt
1 1
xx (!) = · (12A.4.2)
m ! 2 + !02 i !
Generalized susceptibility 325
8
>
> 1 !02 ! 2
xx (!) =
0
>
>
< m (! 2 !02 )2 + 2 !2
(12A.4.3)
>
> 1 !
>
xx (!) =
00
>
: 2 ·
m (! 2 !02 ) + 2 !2
1 1
xx (z) = , =m z > 0. (12A.4.4)
m z 2 + !02 i z
✓ ◆
1 1 1
xx (!) = + · (12A.4.5)
2m!1 ! !1 + i
2 ! + !1 + i
2
The decomposition (12A.4.5) is valid whatever the damping (that is, weak or not).
However, this decomposition presents a practical interest only when !1 is real, that is
for < 2!0 . In this case, it follows that 00xx (!) is the algebraic sum of two Lorentzians
centered at ! = ±!1 and of width (Fig. 12A.1):
1
xx (!) = (12A.4.6)
00 2 2
·
2m!1 (! !1 ) +
2 2
(! + !1 ) +
2 2
4 4
3
See Subsection 4.2.
326 Linear response of a damped oscillator
χ"xx(ω)
0 ω/ω0
-1 1
Fig. 12A.1 Imaginary part of the generalized susceptibility in the case < 2!0 .
⇡ ⇥ ⇤
xx (!) = (! !0 ) (! + !0 ) . (12A.4.7)
00
2m!0
The function:
1 1
xx (z) = , =m z > 0, (12A.4.9)
m !02 i z
decreases proportionally to z 1
as z tends towards infinity. We have:
1 !
xx (!) = (12A.4.10)
00
·
m !04 + 2 ! 2
χ"xx(ω)/ω
|
0 ω
sin !0 t
˜xx (t) = ⇥(t) · (12A.5.3)
m!0
Bibliography
P.M. Chaikin and T.C. Lubensky, Principles of condensed matter physics, Cam-
bridge University Press, Cambridge, 1995.
S.W. Lovesey, Condensed matter physics: dynamic correlations, The Benjamin/
Cummings Publishing Company, second edition, Reading, 1986.
P.C. Martin, Measurements and correlation functions, Les Houches Lecture Notes
1967 (C. De Witt and R. Balian editors), Gordon and Breach, New York, 1968.
Supplement 12B
Electronic polarization
1. Semiclassical model
In the presence of a non-resonant electromagnetic wave, an atomic system acquires an
electric polarization. The response function allowing us to determine this polarization
can be computed in the framework of a semiclassical model, in which the atom is
quantified whereas the electromagnetic field of the wave is treated classically. This
model is simple enough to enable us to directly obtain the response function and
the corresponding generalized susceptibility. This semiclassical calculation allows us
in particular to introduce the notion of oscillator strength associated with a transition,
and to justify the Lorentz model (fully classical) of the elastically bound electron.
Consider an atomic system with a fundamental level of energy "0 and excited
levels of energies "n , assumed non-degenerate for the sake of simplicity, to which
correspond eigenstates | 0 i and | n i. We assume that the atom, initially in its funda-
mental state | 0 i, is excited by a non-resonant1 plane wave of angular frequency !.
Under the e↵ect of this excitation, there appears an induced dipolar electric moment,
oscillating at the angular frequency ! and proportional to the electric field of the wave
if this field is weak.
The electric field E(t) of the wave, parallel to the Ox axis, is assumed to be
spatially uniform. The perturbation is described by the dipolar electric Hamiltonian:
H1 (t) = eE(t)x, (12B.1.1)
in which e denotes the charge of the electron and x the component of its displacement
along Ox. Our aim is to compute the out-of-equilibrium average of the induced dipolar
electric moment P = ex. This average hP (t)ia is defined by the formula:
⌦ ↵
P (t) a = eh (t)|x| (t)i, (12B.1.2)
where | (t)i represents the state of the system at time t. In the linear range, we write:
Z 1
⌦ ↵
P (t) a = ˜(t t0 )E(t0 ) dt0 . (12B.1.3)
1
In equation (12B.1.3), ˜(t) = e ˜xx (t) denotes the linear response function of the
2
polarization.
1
The angular frequency ! thus coincides with none of the Bohr angular frequencies !n0 =
("n "0 )/h̄ associated with the transitions taking place from | 0 i.
330 Electronic polarization
The atomic state just after the application of the field is:2
e X
| (0+ )i = | 0 i E | n ih n |x| 0 i. (12B.2.6)
ih̄ n
The average induced dipolar electric moment hP (t)ia = eh (t)|x| (t)i can be
computed with the aid of the state vector | (t)i given by formula (12B.2.7). The
matrix element h 0 |x| 0 i vanishing for symmetry reasons, we get at first perturbation
order:
⌦ ↵ 2e2 E X 2
P (t) a = |h n |x| 0 i| sin !n0 t, t > 0. (12B.2.8)
h̄ n
The linear response function of the atomic system is thus:
2e2 X 2
˜(t) = ⇥(t) |h n |x| 0 i| sin !n0 t. (12B.2.9)
h̄ n
2 P
We have introduced the closure relation n | n ih n | = 1.
Comparison with the Lorentz model 331
3. Generalized susceptibility
For a harmonic applied field E(t) = <e(Ee i!t ), the linear response of the polarization
reads: ⌦ ↵ ⇥ ⇤
P (t) a = <e Ee i!t (!) . (12B.3.1)
The generalized susceptibility is defined by the formula (!) = lim✏!0+ (! + i✏),
with: Z 1
2e2 X 2
(! + i✏) = |h n |x| 0 i| sin !n0 t ei!t e ✏t dt. (12B.3.2)
h̄ n 0
After the integration over time has been carried out, we get:
X ✓ ◆
e2 2 1 1
(!) = lim |h n |x| 0 i| + · (12B.3.3)
h̄ ✏!0+ n ! !n0 + i✏ ! + !n0 + i✏
We write:
ex = cl (!)E, (12B.4.3)
3
This model has historically played a very important role in the study of the optical properties
of material media.
332 Electronic polarization
where cl (!) denotes the generalized susceptibility of the Lorentz model (or classical
susceptibility). To determine it, we first compute cl (! + i✏), and we then take the
limit ✏ ! 0+ . At finite ✏, retaining in the denominator only terms of order lower than
two in ✏, we get:
e2 1
cl (! + i✏) ' · (12B.4.4)
m ! 2 + !02 2i✏!
In the limit ✏ ! 0+ , we get the generalized susceptibility of the Lorentz model, whose
real and imaginary parts are:
8 ✓ ◆
> e2 1 1
cl (!) = vp + vp
> 0
>
< 2m!0 ! !0 ! + !0
(12B.4.5)
>
> ⇡e2 ⇥ ⇤
>
: cl (!) = (! !0 ) (! + !0 ) .
00
2m!0
2m!n0 2
fn0 = |h n |x| 0 i| . (12B.4.8)
h̄
For an unperturbed Hamiltonian of the form H0 = (p2 /2m) + (r), where r and p are
the electron position and momentum operators, it is possible to prove the following
property, termed the oscillator strength sum rule (or the Thomas–Reiche–Kuhn sum
rule):4 X
fn0 = 1. (12B.4.9)
n
4
See Chapter 14 for a general discussion of sum rules.
Comparison with the Lorentz model 333
Using formulas (12B.3.4), on the one hand, and formulas (12B.4.5), on the other hand,
we then check the identities:
8 X
>
>
>
0
(!) = fn0 0cl (!)[!0 !!n0 ]
< n
X (12B.4.10)
>
>
> 00 (!) =
: fn0 00cl (!)[!0 !!n0 ] ,
n
Thus, the semiclassical calculation allows us, for a non-resonant wave, to justify
the classical Lorentz model of the elastically bound electron. The semiclassical gener-
alized susceptibility appears as a linear combination of generalized susceptibilities of
the type of the Lorentz model’s susceptibility. The angular frequencies of the di↵erent
oscillators identify with the atomic Bohr angular frequencies. The proportion of os-
cillators with a given angular frequency is equal to the oscillator strength associated
with the corresponding transition.
334 Electronic polarization
Bibliography
1. The examples
We present here two elementary examples of dynamical structure factors. The systems
under consideration, made up of a unique atom, either free or submitted to a harmonic
potential, are described by extremely simple equations of motion.1
2. Free atom
Consider a unique atom of mass m, whose position and momentum operators are
respectively denoted by r0 and p0 . This atom, which is free, is in thermal equilibrium
at temperature T . For the sake of simplicity, we assume that the equilibrium statistics
is the Maxwell–Boltzmann one. The quantum character of the equations of motion
will however be taken into account.
is: ⇥ ⇤
n(q, t) = exp iq.r0 (t) . (12C.2.2)
For a free atom, we simply have r0 (t) = r0 + (p0 /m)t, which gives:
h p0 i
n(q, t) = exp iq. r0 + t . (12C.2.3)
m
1
In more complex systems such as fluids, the determination of the dynamical structure factor
involves solving much more complicated equations of motion, but follows analogous lines. This subject
is treated in Supplement 16B.
2
We use the same notation n(. , t) for the density n(r, t) and its spatial Fourier transform n(q, t).
336 Some examples of dynamical structure factors
The expression (12C.2.3) for n(q, t) may be factorized into a product of expo-
nentials. We must take into account the fact that the operators r0 and p0 do not
commute. Using the Glauber identity,3 we end up in one or the other of the two
following equivalent factorized expressions:
8 ⇣ ih̄q 2 t ⌘ ⇣ p .q ⌘
>
> n(q, t) = exp exp i
0
t exp( iq.r0 )
< 2m m
(12C.2.4)
>
> ⇣ 2 ⌘ ⇣ ⌘
: n(q, t) = exp ih̄q t exp( iq.r ) exp i p0 .q t .
0
2m m
⌦ ↵ h q 2 t(t + ih̄ ) i
1
n(q, t)n( q, 0) = exp , = (kT ) . (12C.2.6)
2m
In the same way, with the aid of the second of formulas (12C.2.4) applied to
n( q, t), we can write:
⇣ ih̄q 2 t ⌘ ⇣ p .q ⌘
0
n(q, 0)n( q, t) = exp exp i t . (12C.2.7)
2m m
On average, we have:
⌦ ↵ h q 2 t(t ih̄ ) i
n(q, 0)n( q, t) = exp · (12C.2.8)
2m
Both hn(q, t)n( q, 0)i and hn(q, 0)n( q, t)i are invariant under the change
q! q (as a consequence of the fact that the system possesses the inversion symmetry
r! r). We then verify the Kubo–Martin–Schwinger condition:
⌦ ↵ ⌦ ↵
n(q, 0)n( q, t) = n( q, t ih̄ )n(q, 0) . (12C.2.9)
3
For two operators A and B which both commute with their commutator [A, B], we have the
Glauber identity:
1
eA+B = eA eB e 2 [A,B] .
Atom in a harmonic potential 337
Expressing hn(q, t)n( q, 0)i with the aid of formula (12C.2.6) and carrying out the
integration over time, gives:
✓ ◆1/2 " ✓ ◆2 #
2⇡m m q2
S(q, !) = exp ! h̄ . (12C.2.11)
q2 2q 2 2m
We can group together the two factors on the right-hand side of equation (12C.3.1),
taking into account the fact that x0 (t) and x0 do not commute. Using again the
Glauber identity, gives:
⇣1 ⌘
exp[ iqx0 (t)] exp(iqx0 ) = exp q 2 [x0 (t), x0 ] exp[ iqx0 (t) + iqx0 ]. (12C.3.2)
2
For an oscillator of mass m and of angular frequency !0 , we have x0 (t) = x0 cos !0 t +
(p0 /m!0 ) sin !0 t. The commutator [x0 (t), x0 ] is a scalar which needs no averaging. We
thus have:
⌦ ↵ ⇣1 ⌘⌦ ↵
n(q, t)n( q, 0) = exp q 2 [x0 (t), x0 ] exp[ iqx0 (t) + iqx0 ] . (12C.3.3)
2
To determine the density autocorrelation function, we are thus led to compute the
equilibrium average of an exponentiated operator eA , in which A = iq[x0 (t) x0 ] is
338 Some examples of dynamical structure factors
a linear combination of the position and momentum operators of the oscillator. Using
the identity:
⌦ A↵ D A2 E
e = exp , (12C.3.4)
2
applicable to this type of operator,4 we get:
⌦ ↵ ⇣ 1 ⌦ ⌘
2↵
exp[ iqx0 (t) + iqx0 ] = exp q 2 [x0 (t) x0 ] . (12C.3.5)
2
This gives:
⌦ ↵ ⇣ ⇥⌦ ↵ ⌦ ↵⇤⌘
n(q, t)n( q, 0) = exp q 2 x20 x0 (t)x0 . (12C.3.6)
Formula (12C.3.6) involves the autocorrelation function hx0 (t)x0 i, as well as the
average hx20 i. We have:
8
> ⌦ ↵ h̄
>
< x0 (t)x0 = 2m!0 [(1 + n0 )e + n0 ei!0 t ],
i!0 t
>
(12C.3.7)
>
> ⌦ 2↵ h̄ h̄!0
>
: x0 = coth ,
2m!0 2
1
where n0 = (e h̄!0 1) denotes the Bose–Einstein distribution function at temper-
1
ature T = (k ) .
is called the Debye–Waller factor 5 and is denoted by e 2W (q) . The second factor,
which depends on time,
✓ ◆
⌦ ↵ h̄q 2 ⇥ ⇤
exp q x0 (t)x0 = exp
2
(1 + n0 )e i!0 t
+ n0 ei!0 t
, (12C.3.9)
2m!0
may be written equivalently as::
✓ ◆
⌦ ↵ h̄q 2 ⇥ ( h̄!0 h̄!0 ⇤
exp q 2 x0 (t)x0 = exp h̄!0
e i!0 t+ 2 )
+ e(i!0 t 2 )
.
4m!0 sinh 2
(12C.3.10)
4
Note that identity (12C.3.4) is analogous to the formula of probability theory yielding the
average of the exponential of a Gaussian centered random variable.
5
The same type of factor, originating from the thermal motion of the atoms around their equi-
librium positions, is involved in the scattering of neutrons or X-rays by a crystal.
Atom in a harmonic potential 339
Setting:
h̄q 2
y= , (12C.3.11)
h̄!0
2m!0 sinh 2
we can rewrite the expression (12C.3.10) for exp q 2 hx0 (t)x0 i with the aid of the
following series expansion:
h1 1 i X1
exp y x+ = xn In (y), n = 0, ±1, ±2, . . . , (12C.3.12)
2 x n= 1
where the In (y)’s are the modified Bessel functions of the first kind. We finally obtain
the formula:
1
X
⌦ ↵ 1
n(q, t)n( q, 0) = e 2W (q)
In (y)e 2 n h̄!0 in!0 t
. (12C.3.13)
n= 1
1
X
2W (q)+ 12 h̄!
S(q, !) = 2⇡e In (y) (! n!0 ). (12C.3.14)
n= 1
The dynamical structure factor given by formula (12C.3.14) verifies the detailed bal-
ance relation6 S(q, !) = S(q, !) e h̄! (the system possesses inversion symmetry).
In the sum on the right-hand side of equation (12C.3.14), the term n = 0 cor-
responds to an elastic scattering process with no energy exchange between the radia-
tion and the target (here, the atom). The terms n = ±1 represent the one-quantum
contributions to the dynamical structure factor, whereas the terms with higher |n|
correspond to contributions involving a larger number of quanta.
6
To check this relation, we use the following property of the modified Bessel functions:
I n (y) = In (y).
340 Some examples of dynamical structure factors
Bibliography
Linear response theory is a general formalism applicable to any physical system slightly
departing from equilibrium. It allows us to express the linear response functions and
the generalized susceptibilities in terms of the equilibrium correlation functions of the
relevant dynamical variables. For instance, in the case of the response to an applied
external field, one of the two relevant dynamical variables represents the physical quan-
tity conjugate to the field whereas the other represents the physical quantity whose
out-of-equilibrium average is measured. The computation of the linear response func-
tions relies on a first-order perturbation expansion of the density operator (or of the
phase space distribution function) of the system coupled to the field. The expressions
thus obtained constitute the Kubo formulas of linear response theory.
In this chapter, after having introduced this general formalism and established the
Kubo formulas, we examine the symmetry properties of the linear response functions
and of the associated equilibrium correlation functions. Some of these properties de-
pend on the way that the relevant dynamical variables behave under time-reversal. The
Kubo formulas of linear response theory thus enable us in particular to demonstrate
the Onsager reciprocity relations.
342 General linear response theory
1 1
⇢0 = e H0
, = (kT ) . (13.2.1)
Z
1
This is the most frequent situation. This is why it is quite common practice to make this
hypothesis. However, in principle, in the Kubo theory, the equilibrium distribution is not necessarily
the canonical one.
First-order evolution of the density operator 343
From an initial time t0 (the limit t0 ! 1 will be taken at the end of the
calculation), we isolate the system from the thermostat and we submit it to an external
field a(t), assumed to be spatially uniform.2 The perturbation is described by the
Hamiltonian:3
H1 (t) = a(t)A, (13.2.2)
where A is the Hermitean operator4 associated with the physical quantity conjugate
to the field. The Hamiltonian of the perturbed system is:
H = H0 + H1 (t). (13.2.3)
For t > t0 , the system is isolated. Its density operator ⇢(t) obeys the Liouville–von
Neumann equation:
d⇢(t)
= iL⇢(t), (13.2.4)
dt
where L denotes the Liouville operator associated with the total Hamiltonian.5 We
aim to determine, at first perturbation order, the solution of equation (13.2.4) for the
initial condition ⇢(t0 ) = ⇢0 .
L = L0 + L1 . (13.2.5)
Making use of formulas (13.2.5) and (13.2.6) in equation (13.2.4), we get, since iL0 ⇢0
vanishes:
d ⇢(t)
= iL1 ⇢0 iL0 ⇢(t) iL1 ⇢(t). (13.2.7)
dt
2
The case of a non-uniform field will be treated in Section 9.
3
The forces whose e↵ect may be described by a Hamiltonian of the type (13.2.2) are termed
mechanical forces. There are other types of force, whose e↵ect cannot be expresssed in this way. For
instance, temperature or chemical potential inhomogeneities controlled from outside produce within
a system generalized forces which give rise to a heat flux or to a particle flux. Such forces are termed
thermal forces. The response to this latter type of force will be studied in Chapter 16.
4 P
The perturbation Hamiltonian H1 (t) may in general appear as a sum of the type i ai (t)Ai .
The e↵ect of each term of the sum can be studied separately in the linear regime. In such a situation,
each operator Ai is not necessarily Hermitean (the overall hermiticity of H1 (t) only requires that, if
Ai = A†j , we have ai = a⇤j ).
5
In this first stage, we make use of the Liouville operator formalism which allows us to treat
formally in a unified manner classical and quantum problems as well. We will use in all cases the
denomination of density operator and the notation ⇢(t) (for a classical system, ⇢(t) will stand for the
phase space distribution function).
344 General linear response theory
The third term on the right-hand-side of equation (13.2.7) is of higher order. To get
the first-order correction to ⇢0 , it is enough to solve the evolution equation:
d ⇢(t)
= iL1 ⇢0 iL0 ⇢(t), (13.2.8)
dt
for the initial condition ⇢(t0 ) = 0.
To this end, let us set ⇢(t) = e iL0 t
F (t). The evolution equation of F (t) reads:
dF (t)
= ieiL0 t L1 ⇢0 . (13.2.9)
dt
The solution of equation (13.2.9) for the initial condition F (t0 ) = 0 is:
Z t
0
F (t) = i eiL0 t L1 ⇢0 dt0 . (13.2.10)
t0
We deduce from equation (13.2.10) the expression for ⇢(t):
Z t
0
⇢(t) = i e iL0 (t t ) L1 ⇢0 dt0 . (13.2.11)
t0
The major simplification arising from the linearity hypothesis is the presence of the
unperturbed Liouville operator in the exponent of the evolution operator.
At this stage, it is convenient to make precise the meaning6 of L1 ⇢0 :
1
L1 ⇢0 = [H1 , ⇢0 ]. (13.2.12)
h̄
Formula (13.2.11) thus reads:
Z
i t iL0 (t t0 )
⇢(t) = e [H1 , ⇢0 ] dt0 , (13.2.13)
h̄ t0
that is, taking into account the expression (13.2.2) for the perturbation Hamiltonian:
Z
i t 0 0
⇢(t) = a(t )e iL0 (t t ) [A, ⇢0 ] dt0 . (13.2.14)
h̄ t0
In formula (13.2.14), the operator A, written in the Schrödinger picture, does not
depend on time. The Liouville operator L0 acts only on A and not on ⇢0 , since ⇢0 and
H0 commute. This gives:
Z
i t 0 I 0
⇢(t) = a(t )[A (t t), ⇢0 ] dt0 , (13.2.15)
h̄ t0
where AI (t) = eiL0 t A = eiH0 t/h̄ Ae iH0 t/h̄ denotes the operator A in the interaction
picture (that is, in the Heisenberg picture with respect to H0 ).
The limit t0 ! 1 once taken,we get:
Z t
i
⇢(t) = a(t0 )[AI (t0 t), ⇢0 ] dt0 . (13.2.16)
h̄ 1
6
The notations of the quantum case being more familiar, it is this latter case that we choose to
develop here, the classical case being treated in an appendix at the end of this chapter.
The linear response function 345
Using the invariance of the trace under a circular permutation of the operators gives:
Z t
⌦ ↵ i
B(t) a = a(t0 )Tr [B, AI (t0 t)]⇢0 dt0 . (13.3.4)
h̄ 1
In equation (13.3.4), the quantity Tr [B, AI (t0 t)]⇢0 is simply the equilibrium average
h[B, AI (t0 t)]i. It is thus possible to carry out a translation of its temporal arguments
without modifying it and to write:7
Z t
⌦ ↵ i ⌦ ↵
B(t) a = a(t0 ) [B I (t t0 ), A] dt0 . (13.3.5)
h̄ 1
Formula (13.3.5) shows that the linear response function ˜BA (t) is expressed in terms
of an equilibrium average of a commutator of unperturbed dynamical variables.
7
Whereas hB(t)ia denotes the out-of-equilibrium average of B, a quantity such as h[B I (t t0 ), A]i
denotes an equilibrium average of a commutator of dynamical variables in the interaction picture. No
misinterpretation being possible, we will suppress from now on the superscript label I for the operators
in the interaction picture appearing in the response functions and the associated correlation functions.
346 General linear response theory
i ⌦ ↵
˜BA (t) = ⇥(t) [B(t), A] , (13.3.7)
h̄
where ⇥(t) denotes the Heaviside function, is the Kubo formula8 for ˜BA (t). The
average involved in formula (13.3.7) is an equilibrium average computed with the aid
of the density operator ⇢0 . Interestingly, if A or B commutes with H0 , ˜BA (t) vanishes.
3.2. Expression for ˜ BA (t) with the aid of the eigenstates and eigenvalues
of H0
We denote by {| n i} a base of eigenstates of H0 , of energies "n (the levels are assumed
non-degenerate). It is also a base of eigenstates of ⇢0 , since ⇢0 and H0 commute. We
can rewrite formula (13.3.7) in the form:
i X
˜BA (t) = ⇥(t) h n |[B(t), A]⇢0 | n i, (13.3.8)
h̄ n
i X
˜BA (t) = ⇥(t) ⇧n Bnq Aqn ei!nq t Anq Bqn ei!qn t . (13.3.9)
h̄ n,q
In formula (13.3.9), the quantities !nq = ("n "q )/h̄ are the Bohr angular frequencies
of the unperturbed system. We can also write, inverting the indices n and q in the
second sum on the right-hand side of equation (13.3.9):
i X
˜BA (t) = ⇥(t) (⇧n ⇧q )Bnq Aqn ei!nq t . (13.3.10)
h̄ n,q
The response function ˜BA (t) thus appears as a linear superposition of imaginary
exponentials oscillating at the unperturbed Bohr angular frequencies. For a system
with a finite number of degrees of freedom, the spectrum of H0 is discrete, and the same
is true of the Fourier spectrum of ˜BA (t). The response function is then a countable
sum of periodic functions. Such a function does not vanish as t ! 1: a finite system
thus possesses in this sense an infinitely long ‘memory’.9 We will now illustrate this
feature of the response function with two simple examples.
8
The classical analog of formula (13.3.7) is demonstrated in an appendix at the end of this
chapter.
9
The system considered here is a finite system described by a Hamiltonian. This excludes a system
such as the oscillator damped by viscous friction, whose displacement response function vanishes as
t ! 1.
Relation with the canonical correlation function 347
• Harmonic oscillator
The response function ˜xx (t) of the displacement of a harmonic oscillator with mass m
and angular frequency !0 is a sinusoidal function of angular frequency !0 :
sin !0 t
˜xx (t) = ⇥(t) · (13.3.11)
m!0
i X
˜xx (t) = ⇥(t) x0n xn0 ei!0n t x0n xn0 ei!n0 t . (13.3.13)
h̄ n
2X 2
˜xx (t) = ⇥(t) |h 0 |x| n i| sin !n0 t. (13.3.14)
h̄ n
1
⇧n = e "n
. (13.4.1)
Z
We will show that, in these conditions, the average value h[B(t), A]i involved in the
Kubo formula (13.3.7) can be rewritten in a more simple way (that is, without involving
a commutator).
348 General linear response theory
Inserting in formula (13.4.2) the evolution equation for A (ih̄Ȧ = [A, H0 ]), we get:
Z
[A, ⇢0 ] = ih̄⇢0 e H0 Ȧe H0 d . (13.4.3)
0
Z
1 ⌦ ↵
K̃BA (t) = A( ih̄ )B(t) d , (13.4.5)
0
where A( ih̄ ) = e H0 Ae H0
denotes the operator A in the interaction picture at
the imaginary time ih̄ .
Formula (13.4.4) shows that the response function ˜BA (t) of a system initially
at equilibrium with a thermostat at temperature T can be expressed in terms of the
canonical correlation function11 of B with Ȧ:
5. Generalized susceptibility
The linear response hB(t)ia to an applied field a(t) of Fourier transform a(!), conjugate
to a physical quantity A, has the Fourier transform:
⌦ ↵
B(!) a = BA (!)a(!). (13.5.1)
The generalized susceptibility BA (!) is the Fourier transform in the distribution sense
of the response function ˜BA (t). To get it, we first compute:
Z 1
BA (! + i✏) = ˜BA (t)ei!t e ✏t dt, ✏ > 0, (13.5.2)
0
10
We can check the identity (13.4.2) by computing the matrix elements of both sides on the base
{| n i}.
11
The classical analog of formula (13.4.6) is demonstrated in an appendix at the end of this
chapter.
Generalized susceptibility 349
If ˜BA (t) is expressed with the aid of formula (13.3.10), the above procedure yields
the expression:
1X 1
BA (!) = (⇧n ⇧q )Bnq Aqn lim+ · (13.5.4)
h̄ n,q ✏!0 !qn ! i✏
1X 1
BA (z) = (⇧n ⇧q )Bnq Aqn · (13.5.6)
h̄ n,q !qn z
In general, as the system’s size tends towards infinity, the poles of BA (z) push
closer and closer together. In the limit of an infinite system, they form a continuum.
The discrete set of poles situated on the real axis then becomes a branch cut and
formula (13.5.6) is only valid for =m z 6= 0. It is then convenient to introduce a
function characterizing the density of poles on the real axis as a function of !: this is
the spectral function ⇠BA (!).
350 General linear response theory
6. Spectral function
6.1. Definition
The spectral function ⇠BA (!) is defined by:
⇡X
⇠BA (!) = (⇧n ⇧q )Bnq Aqn (!qn !). (13.6.1)
h̄ n,q
Only the terms n 6= q of the double sum on the right-hand side of formula (13.6.1)
yield a non-vanishing contribution to ⇠BA (!). Like the response function, the spectral
function vanishes if A or B commutes with H0 .
Since the operators A and B have been assumed Hermitean, we have the property:
⇤
⇠BA (!) = ⇠AB (!). (13.6.2)
Importing the expression (13.6.1) for ⇠BA (!) into formula (13.6.3), we demonstrate
the formula:
1 ⌦ ↵
⇠˜BA (t) = [B(t), A] . (13.6.4)
2h̄
Comparing formula (13.6.4) with the Kubo formula (13.3.7) for the response function,
we verify that the relation:
holds between the response function and the inverse Fourier transform of the spectral
function.
Formulas (13.5.6) and (13.6.1) show that we can write a spectral representation of
BA (z) in terms of ⇠BA (!):
Z 1
1 ⇠BA (!)
BA (z) = d!. (13.6.6)
⇡ 1 ! z
Spectral function 351
In contrast to the integral definition (13.5.5), which is only valid for =m z > 0, the
spectral representation (13.6.6) allows us to define BA (z) at any point of affix z
outside the real axis.
We have the property (valid for any z outside the real axis):
BA (z) = AB (z ). (13.6.7)
⇤ ⇤
On the real axis, according to formula (13.5.2) and to formula (13.6.6) (written for
z = ! + i✏), we have:
Z
1 ⇠BA (! 0 )
BA (!) = lim 0
d! 0 . (13.6.8)
⇡ ✏!0+ ! ! i✏
The function BA (z) defined by the spectral representation (13.6.6) reaches dif-
ferent values on opposite sides of the cut:
lim BA (! + i✏) 6= lim BA (! i✏). (13.6.9)
✏!0+ ✏!0+
We have:
lim BA (! + i✏) lim BA (! i✏) = 2i ⇠BA (!). (13.6.10)
✏!0+ ✏!0+
1
The spectral function ⇠BA (!) thus represents, up to the factor (2i) , the di↵erence
between the values taken by BA (z) at the upper edge and at the lower edge of the
cut at the point of abscissa ! = <e z. On the left-hand side of equation (13.6.10), the
first term is simply BA (!), whereas the second term, which involves BA (! i✏) =
AB (! + i✏), is equal to lim✏!0+ BA (! i✏) = ⇤AB (!). Formula (13.6.10) thus reads:
⇤
1⇥ ⇤
⇠BA (!) = BA (!) AB (!) .
⇤
(13.6.11)
2i
In general, the susceptibilities BA (!) and AB (!) are unequal, and the spectral
function ⇠BA (!) does not identify with the imaginary part of BA (!). However, in the
particular case B = A, we have:
In the case of a finite system, ⇠BA (!) is a countable sum of Dirac distributions.
Thus, it is not a function, but a distribution. We provide below its expression in the
two simple examples already considered.
• Harmonic oscillator
We have:
⇡ ⇥ ⇤
⇠xx (!) = xx (!)
00
= (! !0 ) (! + !0 ) . (13.6.13)
2m!0
• Atom perturbed by an electric field
The spectral function ⇠xx (!) for an atom, initially in its fundamental state | 0 i and
perturbed by an electric field, reads:
⇡X 2⇥ ⇤
⇠xx (!) = 00xx (!) = |h 0 |x| n i| (! !n0 ) (! + !n0 ) . (13.6.14)
h̄ n
352 General linear response theory
7. Relaxation
In the course of a relaxation measurement, the system is first submitted to a field
a(t) for a sufficiently long time. This field is then suddenly removed. Our aim is
to study the ensuing relaxation of a physical quantity B, that is, the way that the
out-of-equilibrium average hB(t)ia tends towards the equilibrium average12 hBi. The
deviation hB(t)ia hBi will be denoted for short by hB(t)ia .
a(t)
a
η-1
0 t
At time t = 0, the field is suddenly suppressed. The system then evolves freely
under the e↵ect of its unperturbed Hamiltonian H0 alone. The behavior of hB(t)ia
for t > 0 describes the relaxation of B from the out-of-equilibrium state attained at
t = 0. The applied field is assumed to be weak enough for the relaxation to be properly
described by a linear theory.
12
Here, the physical quantity B is not assumed to be centered.
Relaxation 353
that is: Z 1
⌦ ↵ ⌘t0
B(t) a = ae⌘t ˜BA (t0 )e dt0 . (13.7.3)
t
R1 R1 0
where the integral t ˜BA (t0 )dt0 is defined as lim✏!0+ t ˜BA (t0 )e ✏t dt0 . The general
formula (13.7.4) is valid for t 0 as well as for t 0. We will now analyze these two
cases successively.
• Case t 0
For t 0, the e↵ective lower bound of the integral on the right-hand side of equation
(13.7.4) is 0. This gives:
Z 1
⌦ ↵
B(t 0) a = a lim ˜BA (t)e ✏t
dt. (13.7.5)
✏!0+ 0
Thus, the deviation from equilibrium h B(t)ia for t 0 does not depend on time.
It is expressed as the product of theR amplitude a of the applied field by the static
1
susceptibility BA (! = 0) = lim✏!0+ 0 ˜BA (t)e ✏t dt:
⌦ ↵
B(t 0) a = a BA (! = 0). (13.7.6)
• Case t 0
For t 0, we start from formula (13.7.4), rewritten in the form:
Z t
⌦ ↵
B(t 0) a = a BA (! = 0) a ˜BA (t0 ) dt0 . (13.7.7)
0
We compute the second term on the right-hand side of equation (13.7.7) using the
Kubo formula (13.4.6), rewritten as:
Z
⌦ ↵
˜BA (t) = ⇥(t) A( ih̄ )Ḃ(t) d . (13.7.8)
0
The integration over time once carried out, we deduce from formula (13.7.7) the ex-
pression for hB(t)ia for t 0:
Z Z
⌦ ↵ ⌦ ↵ ⌦ ↵
B(t 0) 0 = a BA (! = 0) + a A( ih̄ )B(t) d a A( ih̄ )B d .
0 0
(13.7.9)
354 General linear response theory
To deduce from formula (13.7.10) a Kubo formula for BA (! = 0), we carry out
an integration by parts at finite ✏ > 0 of the integral over time on the right-hand side.
This gives:
Z Z 1 Z
⌦ ↵ ⌦ ↵
BA (! = 0) = A( ih̄ )B d lim ✏ A( ih̄ )B(t) d .
✏t
dt e
0 ✏!0+ 0 0
(13.7.11)
The matrix elements of B(t) oscillate at the Bohr angular frequencies of the unper-
turbed system. Therefore we have:13
Z 1 Z
⌦ ↵ ⌦ ↵
lim ✏ dt e ✏t
A( ih̄ )B(t) d = A0 B 0 . (13.7.12)
✏!0+ 0 0
In formula (13.7.12), the operators A0 and B 0 are defined14 by their matrix elements
on the eigenbase of H0 :
(
h n |A| q i, "n = "q
h n |A0 | q i = (13.7.13)
0, "n 6= "q .
We deduce from these results an expression for the static susceptibility:15
Z
⌦ ↵ ⌦ 0 0↵
BA (! = 0) = A( ih̄ )B d A B , (13.7.14)
0
which involves the equal time canonical correlation function K̃B B 0 ,A A0 (t = 0):
13
We use the formula:
Z 1 ⇢
✏t+i⌫t 0, ⌫=6 0
lim ✏ e dt =
✏!0+ 0 1, ⌫ = 0.
14
If the energy levels of H0 are non-degenerate, the operators A0 and B 0 possess only diagonal
elements on the eigenbase of H0 . Then they may be called the diagonal parts of A and B with respect
to H0 .
15
A more thorough discussion of the static susceptibility is presented in an appendix at the end
of this chapter. In particular, the static susceptibility BA (! = 0) is compared with the isothermal
susceptibility TBA .
Relaxation 355
For t 0, the quantity B relaxes. Introducing the Kubo formula (13.7.15) for
BA (! = 0) into formula (13.7.9) gives:
Z
⌦ ↵ ⌦ ↵ ⌦ ↵
B(t 0) a = a A( ih̄ )B(t) d a A0 B 0 , (13.7.17)
0
that is:
⌦ ↵
B(t 0) a = a K̃B B 0 ,A A0 (t). (13.7.18)
we have: ⌦ ↵
B(t 0) a = a BA (t). (13.7.20)
According to formula (13.7.4), the response and relaxation functions are related by:
Z 1
BA (t) = ˜BA (t0 ) dt0 , t 0. (13.7.21)
t
We have, inversely:
d
˜BA (t) = BA (t), t 0. (13.7.22)
dt
This is the Debye relaxation law.16 The relaxation function (13.7.23) is fully charac-
terized by the relaxation time ⌧ and the initial value (0).
In many systems however,17 the relaxation dynamics depart from exponential
laws (and are generally much slower). The relaxation processes may then be described
by means of various other functions, including the Kohlrausch–Williams–Watts or
stretched exponential law:
In linear regime, hB(z)ia is related to BA (z). From formulas (13.7.7) and (13.7.8),
we get the relation:18
⌦ ↵
⌦ ↵ B(t = 0) a BA (z)
B(z) a = 1 . (13.7.27)
iz BA (z = 0)
Introducing into formula (13.7.27) the expression for hB(z)ia directly deduced from
the equations of motion, we can obtain the expression for BA (z) and deduce from it
the equilibrium correlation function involved in the appropriate Kubo formula. This
method can be used for the computation of transport coefficients.19
16
The Debye model of dielectric relaxation is described in Supplement 13A.
17
This is, for instance, the case in spin glasses and structural glasses, as well as in other complex
systems.
18
Formula (13.7.27) contains the same information as formula (13.7.7). It may be viewed as more
aesthetic than this latter formula. However it gets complicated when there
P are several input operators,
otherwise stated, when the perturbation Hamiltonian is of the form i ai (t)Ai .
19
The Green–Kubo formulas allowing us to compute the transport coefficients from the equilibrium
correlation functions will be established in Chapters 15 and 16.
Symmetries of the response and correlation functions 357
8.1. Stationarity
The unperturbed system being at equilibrium, the function ⇠˜BA (t) is stationary, that
is, invariant under time translation. We therefore have:
1 ⌦ ↵ 1 ⌦ ↵
[B(t), A] = [A( t), B] , (13.8.1)
2h̄ 2h̄
that is:
⇠˜BA
⇤
(t) = ⇠˜BA (t), (13.8.5)
and:
⇤
⇠BA (!) = ⇠BA ( !) = ⇠AB (!). (13.8.6)
Therefore, if the spectral function ⇠AB (!) is invariant when we exchange A and B,
it is a real and odd function of !. If ⇠AB (!) changes sign in this permutation, it is a
purely imaginary and even function of !.
where ✏A = ±1. The operator A is then said to have a well-defined signature ✏A under
time-reversal. For instance, time-reversal does not change the position operator of a
particle, but it reverses its velocity operator:
The signature of the position is thus ✏x = +1, while that of the velocity is ✏v = 1.
• E↵ect of a magnetic field
In the absence of a magnetic field, H0 and ⇢0 are time-reversal invariant. However, H0
and ⇢0 may depend on an external magnetic field H breaking this invariance. Such a
field is coupled to physical quantities of signature 1. We have:
From formula (13.8.6), it follows that, if A and B have identical signatures, the spectral
function ⇠AB (!) = ⇠BA (!) is a real and odd function of !, whereas, if A and B have
opposite signatures, the spectral function ⇠AB (!) = ⇠BA (!) is a purely imaginary
and even function of !.
As for the response function and the generalized susceptibility, the reciprocity
relations read:
˜BA (t) = ✏A ✏B ˜AB (t), (13.8.14)
and:
BA (!) = ✏A ✏B AB (!). (13.8.15)
and:
BA (!, H) = ✏A ✏B AB (!, H). (13.8.17)
Formula (13.8.17) is verified by the electrical conductivity tensor in the presence of an
external magnetic field, as well as by the magnetic susceptibility tensor.20
The Kubo formulas of linear response theory thus provide a microscopic way to
compute the kinetic coefficients. These coefficients are shown to obey the Onsager or
Onsager–Casimir reciprocity relations, which are thus demonstrated.
9. Non-uniform phenomena
The Kubo formula for the response function generalizes to the case in which the system
is perturbed by a non-homogeneous external field a(r, t). Then, the perturbation is
described by a Hamiltonian of the form:
Z
H1 (t) = dr a(r, t)A(r). (13.9.1)
In the linear range, the out-of-equilibrium average of a centered physical quantity B(r)
reads: Z Z 1
⌦ ↵
B(r, t) a = dr 0 ˜BA (r, t ; r 0 , t0 )a(r 0 , t0 ) dt0 . (13.9.2)
1
The Kubo formula generalizing formula (13.3.7) for the homogeneous case reads:
i ⌦ ↵
˜BA (r, t ; r 0 , t0 ) = ⇥(t t0 ) [B(r, t), A(r 0 , t0 )] . (13.9.3)
h̄
Since the unperturbed system is at equilibrium, the response function ˜BA (r, t ; r 0 , t0 )
only depends on t t0 . In addition, when the unperturbed system is space-translational
invariant, ˜BA (r, t ; r 0 , t0 ) does not depend separately on r and r 0 , but on the di↵erence
r r 0 alone.
Thus, in linear response theory, the response of the system to an external per-
turbation, describing an out-of-equilibrium situation, is expressed in terms of a two-
time average in the equilibrium state, that is, of an equilibrium correlation function.
The out-of-equilibrium average hB(r, t)ia of a centered physical quantity B(r), linear
for weak excitations, is computed with the aid of a response function ˜BA (r, t ; r 0 , t0 )
causal and time-translational invariant (as well as, if the case arises, space-translational
invariant).
Appendices 361
Appendices
i
{, } ! [, ] (13A.1)
h̄
between commutators and Poisson brackets.
However, without recourse to this correspondence, we will present here a direct
derivation of the classical formulas for the first-order correction to the distribution
function and the linear response function. We will limit ourselves to the homogeneous
case.
that is, taking into account the expression (13.2.2) for the perturbation Hamiltonian:
Z t
iL0 (t t0 )
⇢(t) = a(t0 )e {A, ⇢0 } dt0 . (13A.4)
t0
13A.2. The Kubo formula for the classical linear response function
The physical quantity B is assumed to be centered. Its average in the presence of the
external field is: Z
⌦ ↵ 1
B(t) a = ⇢(t)B dqdp. (13A.7)
N !h3N
Importing into formula (13A.7) the expression (13A.6) for ⇢(t), we obtain:
Z t Z
⌦ ↵ 1
B(t) a = a(t ) dt {A(t0 t), ⇢0 }B dqdp.
0 0
(13A.8)
N !h3N 1
we get:
Z Z
⌦ ↵ 1 t
B(t) a = a(t0 ) dt0 {B, A(t0 t)}⇢0 dqdp, (13A.10)
N !h3N 1
that is, after a translation of the temporal arguments of the equilibrium correlation
1R
function h{B, A(t0 t)}i = (N !h3N ) {B, A(t0 t)}⇢0 dqdp:
Z t
⌦ ↵ ⌦ ↵
B(t) a = a(t0 ) {B(t t0 ), A} dt0 . (13A.11)
1
Formula (13A.11) shows that the classical linear response function is expressed in
terms of an equilibrium average of a Poisson bracket of dynamical variables (Kubo
formula):
⌦ ↵
˜BA (t) = ⇥(t) {B(t), A} . (13A.12)
⌦ ↵
˜BA (t) = ⇥(t) B(t)Ȧ . (13A.13)
where the index i denotes the di↵erent degrees of freedom of the system of Hamilto-
nian H0 . Since ⇢0 is the canonical distribution function, we have:
@⇢0 @H0 @⇢0 @H0
= ⇢0 , = ⇢0 , (13A.16)
@qi @qi @pi @pi
that is:
{A, ⇢0 } = ⇢0 {A, H0 }. (13A.17)
On account of the evolution equation Ȧ = {H0 , A}, formula (13A.17) reads:
To obtain BA (! = 0), we use the Kubo formula which allows us to express ˜BA (t)
in terms of the canonical correlation function23 K̃B Ȧ (t):
Z
⌦ ↵
˜BA (t) = ⇥(t) A( ih̄ )Ḃ(t) d . (13B.2)
0
Importing the expression (13B.2) for ˜BA (t) into formula (13B.1), the integration over
time having been carried out and the limit ✏ ! 0+ taken, gives:
Z Z
⌦ ↵ ⌦ ↵
BA (! = 0) = A( ih̄ )B d lim A( ih̄ )B(t) d . (13B.3)
0 t!1 0
22
This formula can also be deduced very directly from its quantum analog (formula (13.4.6)):
indeed, in the classical case, the di↵erent operators commute and the Kubo canonical correlation
function K̃B Ȧ (t) identifies with hB(t)Ȧi.
23
We assume that the system is initially at equilibrium with a thermostat at temperature T .
364 General linear response theory
Formula (13B.3) will be especially useful when comparing BA (! = 0) with the isother-
mal susceptibility.24
Now, to recover the result previously obtained in the main text, we can show25
that the matrix elements between eigenstates of same energy of the operators A and B
are the only ones which are left in the second term on the right-hand side of equation
(13B.3) in the limit t ! 1. We therefore have:
Z
⌦ ↵ ⌦ ↵
BA (! = 0) = A( ih̄ )B d A0 B 0 , (13B.4)
0
that is, in terms of the equal time canonical correlation function of the operators
A A0 and B B 0 :
hBia hBi
T
BA = lim · (13B.7)
a!0 a
24
See Subsection 13B.3.
25
We use Abel’s theorem:
Z Z 1 Z
✏t
lim hA( ih̄ )B(t)i d = lim ✏ e hA( ih̄ )B(t)i d dt,
t!1 0 ✏!0+ 0 0
In formula (13B.7), hBia denotes the equilibrium average of B in the presence of the
applied field: ⇥ ⇤
Tr Be (H0 aA)
hBia = ⇥ ⇤ , (13B.8)
Tr e (H0 aA)
whereas hBi is the equilibrium average of B in the absence of a field:
⌦ ↵ Tr Be H0
B = · (13B.9)
Tr e H0
To compute hBia at first order, we write the first-order expansion for the exponential
e (H0 aA) :
h Z i
e (H0 aA)
=e H0
1+a A( ih̄ ) d + O(a2 ) , (13B.10)
0
and we deduce from it the first-order expansion for the density operator at equilibrium
in the presence of the field:
✓ Z ◆
e (H0 aA)
e H0 ⇥ ⇤
⇥ ⇤' ⇥ ⇤ 1+a A( ih̄ ) hAi d . (13B.11)
Tr e (H0 aA) Tr e H0
0
In formula (13B.11), hAi denotes the equilibrium average of A in the absence of a field.
We then deduce from formula (13B.7) the expression for the isothermal susceptibility:
Z
⌦ ↵ ⌦ ↵⌦ ↵
T
BA = A( ih̄ )B d A B . (13B.12)
0
Formula (13B.12) involves the equal time canonical correlation function of the opera-
tors A hAi and B hBi:
T
BA = K̃B hBi,A hAi (t = 0). (13B.13)
In the particular case B = A, the isothermal susceptibility is related to the equal time
canonical correlation function of the fluctuations of A:
T
AA = K̃A hAi,A hAi (t = 0). (13B.15)
13B.3. Discussion
The comparison between the static susceptibility of an isolated system computed via
the Kubo formula, on the one hand, and the isothermal susceptibility, on the other
hand, displays the fact that these two susceptibilities are in principle di↵erent.
As shown by formulas (13B.4) and (13B.12), BA (! = 0) and T
BA are equal only
if we have the identity: ⌦ 0 0 ↵ ⌦ ↵⌦ ↵
A B = A B . (13B.16)
Condition (13B.16) also reads:
Z
⌦ ↵ ⌦ ↵⌦ ↵
lim A( ih̄ )B(t) d = A B . (13B.17)
t!1 0
Bibliography
P.M. Chaikin and T.C. Lubensky, Principles of condensed matter physics, Cam-
bridge University Press, Cambridge, 1995.
S. Dattagupta, Relaxation phenomena in condensed matter physics, Academic Press,
Orlando, 1987.
D. Forster, Hydrodynamic fluctuations, broken symmetries, and correlation func-
tions, Westview Press, Boulder, 1995.
R. Kubo, M. Toda, and N. Hashitsume, Statistical physics II: nonequilibrium
statistical mechanics, Springer-Verlag, Berlin, second edition, 1991.
S.W. Lovesey, Condensed matter physics: dynamic correlations, The Benjamin/
Cummings Publishing Company, Reading, second edition, 1986.
P.C. Martin, Measurements and correlation functions, Les Houches Lecture Notes
1967 (C. De Witt and R. Balian editors), Gordon and Breach, New York, 1968.
M. Plischke and B. Bergersen, Equilibrium statistical physics, World Scientific,
Singapore, third edition, 2006.
D. Zubarev, V. Morozov, and G. Röpke, Statistical mechanics of nonequilibrium
processes, Vol. 2: Relaxation and hydrodynamic processes, Akademie Verlag, Berlin,
1997.
R. Zwanzig, Nonequilibrium statistical mechanics, Oxford University Press, Oxford,
2001.
References
E2 , termed the Lorentz cavity field, whereas the dipoles inside this cavity create a
field E3 . On account of these various contributions, the local field reads:
Eloc = Eext + E1 + E2 + E3 . (13A.1.2)
The Maxwell field is the sum of the external field and the depolarizing field:
EMax = Eext + E1 . (13A.1.3)
Between the Maxwell field and the local field, we have the relation:
Eloc = EMax + E2 + E3 . (13A.1.4)
• Depolarizing field
We generally write the depolarizing field as:
E1x = N x Px , E1y = N y Py , E1z = N z Pz , (13A.1.5)
where Nx , Ny , and Nz are the depolarizing factors.2
• Lorentz field
The Lorentz field E2 due to the polarization charges on the surface of the fictitious
spherical cavity is:
4⇡
E2 = P. (13A.1.6)
3
• Field of the dipoles inside the cavity
Among the fields contributing to the local field, the field E3 is the only one which
depends on the structure of the material. In a medium that is either isotropic or of
cubic symmetry this field vanishes:
E3 = 0. (13A.1.7)
4⇡
Eloc = EMax + P, (13A.1.8)
3
independently of the form (spherical or not) of the sample. Formula (13A.1.8), called
the Lorentz relation, is always valid (provided that the considered medium is either
isotropic or of cubic symmetry).
In the case of a spherical sample for which Nx = Ny = Nz = 4⇡/3, the relation
(13A.1.2) simply becomes:
Eloc = Eext . (13A.1.9)
2
The values of the depolarizing factors depend on the ratios between the principal axes of the
ellipsoid. In the case of a spherical sample, we have N = 4⇡/3 for any axis.
370 Dielectric relaxation
" 1 4⇡
= N ↵. (13A.1.14)
"+2 3
1
(ε'-1)/4πNα0
0.8
0.6
ε"/4πNα0
0.4
0.2
0 1 2 ωτ 3
Fig. 13A.1 Debye model: ["0 (!) 1]/4⇡N ↵0 and "00 (!)/4⇡N ↵0 as functions of !⌧ .
If the angular frequency ! of the applied field is much larger than ⌧ 1 , the
molecule becomes unable to ‘follow’ the field. Debye proposed accordingly to write
the polarizability in the form:
↵0
↵(!) = , (13A.3.1)
1 i!⌧
where ↵0 denotes the static orientational polarizability. In a low density liquid, we
have "(!) ' 1. We can therefore assimilate the local field and the Maxwell field, and
simplify the Clausius–Mossotti formula (13A.1.14) into:
4⇡N ↵0
"(!) = 1 + · (13A.3.3)
1 i!⌧
The real and imaginary parts of "(!) are given by the formulas:
8
> 1
< " (!) 1 = ("s 1)
> 0
1 + !2 ⌧ 2
(13A.3.4)
>
> !⌧
: " (!) = ("s
00
1) ,
1 + !2 ⌧ 2
with "s 1 = 4⇡N ↵0 . The corresponding curves are represented in Fig. 13A.1.
5
We have here "1 = 1, since the other polarization mechanisms are not taken into account.
The Debye theory of dielectric relaxation 373
ε"
ωτ = 1
↗
ωτ >> 1 ωτ << 1
0 1 εs ε'
Thus, the Cole–Cole diagram associated with the Debye model is a semicircle ("00 > 0).
The relaxation time may be directly read on this diagram, since "00 (!) takes its max-
imum value at ! = ⌧ 1 (Fig. 13A.2).
!"00
K' · (13A.3.6)
c
From the Debye model, we get:
"s 1 !2 ⌧
K= · (13A.3.7)
c 1 + !2 ⌧ 2
The absorption coefficient as given by formula (13A.3.7) first increases with !, then
it saturates for !⌧ 1: this is the Debye plateau.
3.4. Discussion
The Debye theory is actually valid only for !⌧ ⌧ 1, as displayed by the fact that the
part !⌧ 1 of the experimental Cole–Cole diagram exhibits a kind of protuberance,
374 Dielectric relaxation
that is, a zone in which "0 < 1, absent from the semicircular Cole–Cole diagram
corresponding to the Debye model.
Similarly, for !⌧ 1, the measured absorption coefficient does not exhibit a
plateau, but instead decreases. The Debye model of orientational polarization thus
proves to be insufficient to explain the shape of the dielectric relaxation curves actually
observed in polar liquids. We shall now see what improvements can be brought by a
microscopic description based on linear response theory.
The liquid being isotropic on average, the induced average dipolar moment is parallel
to Ox. In the linear response regime, it is of the form:
Z 1
⌦ ↵
M (t) a = ˜M M (t t0 )Eext (t0 ) dt0 . (13A.4.2)
1
We assume that the relation between the local field and the external field is
simply given by formula (13A.1.9). We can then identify N ↵(!) and V 1 M M (!). On
6
We exclude molecules with free or nearly free internal degrees of freedom.
7
For the sake of simplicity, we simply denote by M the component Mx of the induced dipolar
moment.
A microscopic model of orientational polarization 375
account of the Clausius–Mossotti relation (13A.1.14), M M (!) and "(!) are related
by the formula:
"(!) 1 4⇡ 1
= M M (!), (13A.4.6)
"(!) + 2 3 V
that is: Z 1
"(!) 1 4⇡ 1
= ˜M M (t)ei!t dt. (13A.4.7)
"(!) + 2 3 V 0
Importing the expression (13A.4.5) for ˜M M (t) into equation (13A.4.7), we get the
formula: Z 1 ⌦ ↵
"(!) 1 4⇡ 1 d M (t)M i!t
= e dt, (13A.4.8)
"(!) + 2 3 V 0 dt
which, due to the average isotropy of the liquid, may be rewritten as:
Z 1 ⌦ ↵
"(!) 1 4⇡ d M (t).M i!t
= e dt. (13A.4.9)
"(!) + 2 9kT V 0 dt
Formula (13A.4.9) is valid whatever the origin of the dipolar moment of the sample
(ionic, orientational . . .). We will now apply it to the determination of the orientational
polarization in a polar fluid of low density.
When the molecules are of low polarizability, the electric moment of the system
originates essentially from the orientation of the molecular dipoles. We therefore write:
Z 1 ⌦ ↵
4⇡ d M0 (t).M0 i!t
"(!) 1 = e dt, (13A.4.11)
3kT V 0 dt
that is:
M = M0 + M1 ,
where M0 is the contribution to the dipolar moment due to the orientation of the molecules (consid-
ered as rigid rods) and M1 the contribution of the vibrations. The vibrations being assumed to be of
weak amplitude, the most significant contribution to M is M0 . It is the only contribution which we
will take into account here, which amounts to setting "1 = 1.
376 Dielectric relaxation
The fluid being of low density, the orientational correlations between di↵erent molecules
may be neglected. Accordingly, we write:
⌦ ↵ ⌦ ↵
M0 (t).M0 ' N V µi (t).µi (13A.4.13)
If ui denotes the unit vector parallel to µi , we can rewrite equation (13A.4.16) in the
form:
Z 1
"(!) 1 dS(t) i!t
= e dt, (13A.4.17)
"s 1 0 dt
where the function S(t) = hui (t).ui i is proportional to the autocorrelation function
of an individual dipole. At initial time, we have S(0) = u2i = 1. Due to the colli-
sions undergone by the molecule, S(t) decreases as time increases (and it eventually
vanishes).
S(t) = e t/⌧
, t > 0, (13A.4.18)
the time ⌧ being a measure of the duration of the correlation of ui (t) with ui (0), we
recover the Debye expression for the dielectric permittivity:
"(!) 1 1
= · (13A.4.19)
"s 1 1 i!⌧
A microscopic model of orientational polarization 377
"(!) 1 S̈(0)
' 2 , (13A.4.22)
"s 1 !
in which S̈(0) < 0. Equating the real and imaginary parts of both sides of the approx-
imate equation (13A.4.22) gives:
8
>
< "0 (!) S̈(0)
1 ' ("s 1)
!2 (13A.4.23)
>
:
" (!) ' 0.
00
Formulas (13A.4.23) allow us to understand the origin9 of the protuberance "0 < 1
observed on the experimental Cole–Cole diagram. They show in addition that the
absorption coefficient K = !"00 /nc vanishes as ! ! 1, which explains the lowering of
the curve K(!) observed for !⌧ 1.
9
The precise microscopic mechanism accounting for the analyticity of S(t) at t = 0 is not described
here. By analogy with the description of the velocity autocorrelation function of a Brownian particle,
we can think of this mechanism as involving some microscopic time much shorter than the relaxation
time.
378 Dielectric relaxation
Bibliography
N.W. Ashcroft and N.D. Mermin, Solid state physics, W.B. Saunders Company,
Philadelphia, 1976.
W. Jones and N.H. March, Theoretical solid-state physics: non-equilibrium and
disorder , Vol. 2, Wiley, New York, 1973. Reprinted, Dover Publications, New York,
1985.
C. Kittel, Introduction to solid state physics, Wiley, New York, eighth edition, 2005.
R. Kubo, M. Toda, and N. Hashitsume, Statistical physics II: nonequilibrium
statistical mechanics, Springer-Verlag, Berlin, second edition, 1991.
D.A. McQuarrie, Statistical mechanics, University Science Books, Sausalito, second
edition, 2000.
R. Zwanzig, Nonequilibrium statistical mechanics, Oxford University Press, Oxford,
2001.
References
In the absence of any other interactions, the average magnetic moment hM (t)ia
induced by the applied field evolves according to the equation:
⌦ ↵
d M (t) a ⌦ ↵
= M (t) a ⇥ H(t). (13B.1.2)
dt
In the presence of interactions inducing a relaxation of the average magnetic moment,
we write, instead of equation (13B.1.2), an evolution equation of the form:
⌦ ↵ ⌦ ↵
d M (t) a ⌦ ↵ d M (t) a
= M (t) a ⇥ H(t) + , (13B.1.3)
dt dt relax
in which the relaxation term dhM (t)ia /dt|relax accounts for the di↵erent phenomena
producing the relaxation of hM (t)ia towards its equilibrium value.
We will first present the phenomenological theory of magnetic resonance relying
on the Bloch equations. Then we will solve the same problem microscopically via the
quantum theory of linear response.3
1
There are several types of magnetic resonance phenomena, such as nuclear magnetic resonance
(NMR), electronic paramagnetic resonance (EPR) . . ., thus denominated depending on whether the
spins are carried by nuclei, electrons . . . Here, we only provide a general description of the principle
of the magnetic resonance phenomenon.
2
The gyromagnetic ratio is negative for electronic spins, but most often positive for nuclear spins.
3
This calculation will be restricted to the simple case in which damping is neglected.
380 Magnetic resonance
2. Phenomenological theory
2.1. The Bloch equations
At equilibrium in the presence of the static field H0 assumed to be applied along the
Oz axis, the only non-vanishing component of the average magnetic moment is hMz ia ,
as expressed by the formulas:
We introduce for convenience the static magnetic susceptibility, defined by the relation:
M0 = 0 H0 . (13B.2.2)
⌦ ↵ ⌦ ↵
d Mx,y (t) a Mx,y (t) a
= · (13B.2.4)
dt relax T2
The relaxation of hMx (t)ia and hMy (t)ia is mainly due to spin–spin interactions.
Clearly, the microscopic determination of T1 and T2 requires a knowledge of the
interaction mechanisms of a spin with its environment. Depending on the particular
conditions of the system under study, we can have either T1 ' T2 or T1 T2 .
The evolution equations deduced from equation (13B.1.3) by projection on the
three axes, with the relaxation terms as given by formulas (13B.2.3) and (13B.2.4),
constitute the Bloch equations.
Phenomenological theory 381
(as well as an analogous formula for hMy (t)ia ). The average magnetic moment de-
scribed by hMx (t)ia and hMy (t)ia is a response to the rotating field H1 (t). We can
write hMx (t)ia in the form:
⌦ ↵ ⇥ ⇤
Mx (t) a = H1 0T (!) cos !t + 00T (!) sin !t , (13B.2.14)
where: 8
>
> !0 0 (!0 !)
T (!) =
0
>
> 2
>
< (! !0 ) + T2 2
1 (13B.2.15)
>
> !0 0 T2
T (!) =
>
>
00
>
: (!
2
!0 ) + T2 2
are the real and imaginary parts of the transverse susceptibility T (!).
For ! ' !0 , the leading contribution to 00xx (!) is that of the first term of formula
(13B.2.21). The average absorbed power takes approximately the form of a Lorentzian
centered at !0 and of width 2T2 1 :
dW 1 1 1
' H12 !02 0 T2 · (13B.2.22)
dt 4 (!
2
!0 ) + T2 2
The width of the resonance line is thus determined by T2 . Inversely, the measurement
of the linewidth provides information about the microscopic processes governing T2 .
3. A microscopic model
Consider a spin S placed in a static magnetic field H0 . Submit it in addition to a
transverse magnetic field H1 (t) of small amplitude. We are looking for the average
magnetic moment hM (t)ia at first perturbation order.
The unperturbed Hamiltonian is:6
h0 = M .H0 . (13B.3.1)
6
In order to avoid any confusion with the magnetic field, the unperturbed Hamiltonian is denoted
here by a lower case letter (the same is done below for the perturbed Hamiltonian). The expression
(13B.3.1) for h0 corresponds to the spin part of the unperturbed Hamiltonian (this latter may also
have an orbital part). The spin–lattice and spin–spin interactions are not taken into account here.
384 Magnetic resonance
The eigenvalues of h0 are "m = m h̄H0 = mh̄!0 , the magnetic quantum number m
being an integer or a half-integer. The corresponding eigenstates are denoted by |mi.
The perturbation Hamiltonian is:
⇥ ⇤
h1 (t) = M .H1 (t) = Mx H1x (t) + My H1y (t) . (13B.3.2)
Depending on the choices made for the functions H1x (t) and H1y (t), the form (13B.3.2)
of h1 (t) may allow us in particular to describe either a rotating field or a linearly
polarized field.
According to linear response theory, the average induced magnetic moment
hMx (t)ia is given by:
Z 1
⌦ ↵ ⇥ ⇤
Mx (t) a = ˜xx (t t0 )H1x (t0 ) + ˜xy (t t0 )H1y (t0 ) dt0 . (13B.3.3)
1
The response functions ˜Mx Mx (t) and ˜Mx My (t), respectively denoted here by ˜xx (t)
and ˜xy (t) for the sake of simplicity, may be determined by using the Kubo formulas,
which read in this case:
i ⌦ ↵
˜xx (t) = ⇥(t) 2
[Sx (t), Sx (0)] , (13B.3.4)
h̄
and:
i ⌦ ↵
˜xy (t) = ⇥(t) 2
[Sx (t), Sy (0)] . (13B.3.5)
h̄
We can also deduce ˜xx (t) and ˜xy (t) from the generalized susceptibilities xx (!) and
xy (!) by inverse Fourier transformation. We will adopt here this latter method.
7
and:
X ✓ h s|Sx |mihm|Sy | si hm|Sx | sih s|Sy |mi
◆
1
xy (!) = lim
2
h̄ ·
✏!0+ !m, s ! i✏ ! s,m ! i✏
m6= s
(13B.3.7)
7
We compute the susceptibilities by using the general formula:
1X 1
BA (!) = (⇧n ⇧q )Bnq Aqn lim ·
h̄ n,q ✏!0+ !qn ! i✏
.
A microscopic model 385
To make explicit the expressions (13B.3.6) and (13B.3.7) for xx (!) and xy (!), we
compute the matrix elements h s|Sx |mi and h s|Sy |mi. For a spin s (m then varying
from s to +s), we have:
and: ✓ ◆
2
sh̄ 1 1
xy (!) = i lim+ + · (13B.3.10)
2 ✏!0 ! !0 + i✏ ! + !0 + i✏
These results may be generalized at finite temperature. In this case, we have:
✓ ◆
2
1 1
xx (!) = hSz i lim+ (13B.3.11)
2 ✏!0 ! !0 + i✏ ! + !0 + i✏
and: ✓ ◆
i 2
1 1
xy (!) = hSz i lim+ + · (13B.3.12)
2 ✏!0 ! !0 + i✏ ! + !0 + i✏
In formulas (13B.3.11) and (13B.3.12), hSz i denotes the equilibrium average of the
component of the spin along Oz. At T = 0, we have hSz i = sh̄. Formulas (13B.3.11)
and (13B.3.12) then respectively coincide with formulas (13B.3.9) and (13B.3.10).
The response functions ˜xx (t) and ˜xy (t) are obtained by inverse Fourier trans-
formation. The limit ✏ ! 0+ having been taken, we obtain the formulas:
˜xx (t) = 2
hSz ih̄⇥(t) sin !0 t (13B.3.13)
and:
˜xy (t) = 2
hSz ih̄⇥(t) cos !0 t. (13B.3.14)
8
We assume here < 0, that is, !0 > 0.
386 Magnetic resonance
From the expression (13B.3.11) for xx (!) = Mx Mx (!), we deduce, the limit ✏ ! 0+
once taken: ⇥ ⇤
⇡
Sx Sx (!) = hSz i (! + !0 ) (! !0 ) . (13B.3.22)
00
2
This verifies the relation:10
h̄!
Sxx (!) = h̄ coth Sx Sx (!).
00
(13B.3.23)
2
1 h̄!0
S̃xx (t) = hSz ih̄ coth cos !0 t. (13B.3.24)
2 2
10
The relation (13B.3.23) between Sxx (!) and 00xx (!) constitutes the expression for this problem
of the fluctuation-dissipation theorem (see Chapter 14).
11
We have the identity:
2
hSz i = !0 0,
Bibliography
This chapter deals with the fluctuation-dissipation theorem, which constitutes one of
the fundamental statements of the statistical physics of linear irreversible processes.
The fluctuation-dissipation theorem expresses a relation between the dissipative
part of a generalized susceptibility and the associated equilibrium correlation function.
In practice, this relation is used either to describe the intrinsic fluctuations of a dy-
namical variable using the characteristics of the susceptibility or, inversely, to deduce
the susceptibility from the relevant equilibrium fluctuations.
The fluctuation-dissipation theorem is derived from the Kubo formulas of linear
response theory. As a general rule, the energy absorbed on average by a dissipative
system coupled to an external field a(t) is larger than the energy restored to the field
by the system. The energy supplied by the field is eventually dissipated irreversibly
within the system. In the case of a linear coupling described by the perturbation
Hamiltonian a(t)A, the average dissipated power is related to the imaginary part
of the generalized susceptibility AA (!). This susceptibility is expressed in terms of
the equilibrium autocorrelation function of the relevant dynamical variable (namely,
the physical quantity A). The autocorrelation function of an operator characterizing
the fluctuations of the associated physical quantity, the Kubo formula for 00AA (!)
establishes a relation between the dissipation and the equilibrium fluctuations of A.
This relation constitutes the fluctuation-dissipation theorem.
390 The fluctuation-dissipation theorem
1. Dissipation
The energy received by a system of infinite size coupled to an external field is even-
tually transformed into heat. It is thus dissipated irreversibly within the system. In
a harmonic linear regime, the average dissipated power is related to the generalized
susceptibility. Consequently, one of the means of determining experimentally the sus-
ceptibility consists in measuring the average dissipated power.
The relation between the average dissipated power and the susceptibility can be
established in several ways, either in the general framework of the linear response
theory, or by comparing the expression for the average dissipated power deduced from
the Fermi golden rule with the Kubo formula for the susceptibility.
1.1. Computation of the average dissipated power via the linear response
theory
Consider a system of unperturbed Hamiltonian H0 , submitted to a spatially uniform
applied field a(t). The corresponding perturbation is described by the Hamiltonian:1
H1 (t) = a(t)A, (14.1.1)
in which the Hermitean operator A represents the physical quantity coupled to the
external field. We assume that the latter is harmonic (a(t) = <e(ae i!t )). Our aim is to
calculate the average dissipated power dW/dt. This quantity may be obtained, either
by calculating the average power received by the system and eventually dissipated, or
from the average evolution rate of the total energy of the system coupled to the field.
• Average power received by the system
The instantaneous power received by the system submitted to the field a(t) is:
⌦ ↵
dW d A(t) a
= a(t) · (14.1.2)
dt dt
In a stationary harmonic regime of angular frequency !, we have:
⌦ ↵ ⇥ ⇤
A(t) a = <e ae i!t AA (!) . (14.1.3)
For a real amplitude a, the instantaneous power received by the system is thus:
dW ⇥ 0 ⇤
= a2 ! cos !t AA (!) sin !t + AA (!) cos !t .
00
(14.1.4)
dt
On average, the power dissipated within the system is equal to the power it receives.
The average dissipated power is thus proportional to ! 00AA (!):
dW 1 (14.1.5)
= a2 ! AA (!).
00
dt 2
1 P
The study may be generalized to the case in which H1 (t) is a sum of the type j aj (t)Aj . To
compute the dissipation, it is then necessary to take into account the linear response of the di↵erent
observables Ai , that is, the set of generalized susceptibilities Ai Aj (!).
Dissipation 391
• Average evolution rate of the energy of the system coupled to the field
We can also obtain the average dissipated power by considering as the ‘system’ (in
the thermodynamic sense) the system of unperturbed Hamiltonian H0 placed in the
field a(t), and by determining the average evolution rate of the energy Etot of this
global ensemble. This latter energy is given by:
⇥ ⇤
Etot = Tr ⇢(t)H , (14.1.6)
where H = H0 a(t)A is the Hamiltonian of the system coupled to the field and ⇢(t)
the density operator of the system. We have:
dEtot d⇣ ⇥ ⇤⌘
= Tr ⇢(t)H , (14.1.7)
dt dt
that is: h d⇢(t) i da(t) ⇥
dEtot ⇤
= Tr H Tr ⇢(t)A . (14.1.8)
dt dt dt
The first term on the right-hand side of equation (14.1.8) vanishes on account of the
evolution equation of ⇢(t). We are left with:
dEtot da(t) ⌦ ↵
= A(t) a . (14.1.9)
dt dt
In the considered harmonic regime, we therefore have, at any time,
dEtot ⇥ ⇤
= a2 ! sin !t AA (!) cos !t
0
+ AA (!) sin !t
00
, (14.1.10)
dt
and, on average:
dEtot 1 (14.1.11)
= a2 ! AA (!).
00
dt 2
1.2. Computation of the average dissipated power via the Fermi golden rule
The Hamiltonian H0 has a base of eigenstates {| n i}, of energies "n . The system
is submitted to the perturbation aA cos !t. We associate with this perturbation a
transition rate between an initial state | n i and final states | q i given, according to
the Fermi golden rule, by:
⇡a2 X 2⇥ ⇤
|h q |A| n i| (!qn !) + (!nq !) . (14.1.12)
2h̄2 q
392 The fluctuation-dissipation theorem
dW ⇡a2 X 2
= ⇧n h̄!|Aqn | (!qn !). (14.1.13)
dt abs 2h̄2 n,q
Similarly, we denote by dW/dt|em the energy emitted per unit time by the system. We
compute it in an analogous way from the second term of expression (14.1.12). This
gives:
dW ⇡a2 X 2
= ⇧n h̄!|Aqn | (!nq !), (14.1.14)
dt em 2h̄2 n,q
dW ⇡a2 X 2
= 2 ⇧q h̄!|Aqn | (!qn !). (14.1.15)
dt em 2h̄ n,q
The energy actually received per unit time by the system is equal to the di↵erence
dW/dt|abs dW/dt|em :
dW dW ⇡a2 X 2
= (⇧n ⇧q )!|Aqn | (!qn !), (14.1.16)
dt abs dt em 2h̄ n,q
dW 1 (14.1.17)
= a2 !⇠AA (!).
dt 2
The spectral function ⇠AA (!) being just the imaginary part 00AA (!) of the generalized
susceptibility, the expressions (14.1.5) and (14.1.17) for dW/dt are e↵ectively identical.
2
Indeed, for ! > 0, the non-vanishing terms of the sum on the right-hand side of equation
(14.1.16) correspond to "q > "n , that is, to ⇧n > ⇧q .
3
The spectral function ⇠AA (!) is defined by the formula:
⇡ X
⇠AA (!) = (⇧n ⇧q )|Aqn |2 (!qn !).
h̄ n,q
Equilibrium fluctuations 393
The close relation between the calculations carried out via the linear response theory,
on the one hand, and via the Fermi golden rule, on the other hand, comes out of the
fact that both calculations are first perturbation order treatments.
2. Equilibrium fluctuations
At equilibrium, the correlation between the fluctuations of two physical quantities is
characterized by the correlation function of the associated operators. In the classical
case, the correlation function C̃BA (t) of two operators A and B is defined unambigu-
ously (for centered operators, we have C̃BA (t) = hB(t)Ai). However, in the quantum
case, because of the non-commutativity of operators, we generally use one or the other
of two non-equivalent definitions for the correlation function of two operators A and B.
It is even so for the autocorrelation function of an operator A, since A(t) does not in
general commute with A(t0 ).
1⌦ ↵
S̃BA (t) = {A, B(t)}+ , (14.2.2)
2
where {A, B(t)}+ = AB(t) + B(t)A denotes the anticommutator of A and B(t). The
quantity:
1⌦ ↵
S̃AA (t) = {A, A(t)}+ (14.2.3)
2
is the symmetric autocorrelation function of the operator A.
394 The fluctuation-dissipation theorem
Z
1 ⌦ ↵
K̃BA (t) = e H0
Ae H0
B(t) d , = (kT )
1
. (14.2.4)
0
2.2. Expression for the Fourier transforms S BA (!) and K BA (!) over an
eigenbase of H0
Let SBA (!) and KBA (!) be the respective Fourier transforms of S̃BA (t) and K̃BA (t):
Z 1 Z 1
SBA (!) = S̃BA (t)ei!t dt, KBA (!) = K̃BA (t)ei!t dt. (14.2.5)
1 1
and: ⌦ ↵ X
AB(t) = ⇧q Aqn Bnq e i!qn t
, (14.2.7)
n,q
we deduce from formulas (14.2.8) and (14.2.10) a direct relation (not involving the
eigenstates of H0 ) between SBA (!) and KBA (!):
h̄! h̄!
SBA (!) = coth KBA (!). (14.2.12)
2 2
Both symmetric and canonical correlation functions are thus identical in this limit.
Accordingly, we have, in the classical limit h̄! ⌧ 1:
4
Violations of this theorem are only observed in out-of-equilibrium systems such as spin glasses
or structural glasses. These systems relax very slowly and display aging properties: the time scale of
the response to an external perturbation or of a correlation function between two physical quantities
increases with the age of the system, that is, with the time elapsed since its preparation. In other
words, these functions decrease all the more slowly when the system is older. The aging manifests itself
in particular by the separate dependance of response and/or correlation functions of some dynamical
variables with respect to their two temporal arguments. In such a case, the fluctuation-dissipation
theorem does not hold.
396 The fluctuation-dissipation theorem
3.1. Relation between S BA (!) or K BA (!) and the spectral function ⇠BA (!)
Using the expressions (14.2.8) and (14.2.10) for SBA (!) and KBA (!), together with
the definition of the spectral function5 ⇠BA (!), we verify the relations:
h̄!
SBA (!) = h̄ coth ⇠BA (!) (14.3.1)
2
and:
2
KBA (!) = ⇠BA (!). (14.3.2)
!
and: Z 1
!
AA (!) = K̃AA (t)ei!t dt, (14.3.4)
00
2 1
which show that the knowledge of the energy dissipation in the system perturbed by
the field a(t) coupled to the physical quantity A is equivalent to that of the dynamics
of the equilibrium fluctuations of A.
In the classical limit, S̃AA (t) and K̃AA (t) identify with C̃AA (t) (formula (14.2.14)),
and formulas (14.3.3) and (14.3.4) simply read:
Z
! 1⌦ ↵
00
(!) = A(t)A ei!t dt. (14.3.5)
AA
2 1
6
See Subsection 4.2.
7
The examples presented here are classical. A quantum example, concerning the displacement of
a harmonic oscillator coupled with a phonon bath, will be treated in Supplement 14A.
The fluctuation-dissipation theorem 397
10
See Supplement 16B.
398 The fluctuation-dissipation theorem
4. Positivity of ! AA (!)
00
AA (!) 0. (14.4.1)
00
!
and:
!2
AA (!) = KAA (!). (14.4.3)
00
!
2
Formulas (14.4.2) and (14.4.3) show that the positivity of ! AA (!)
00
implies the posi-
tivity of SAA (!) as well as the positivity of KAA (!):
5. Static susceptibility
5.1. Expression for AA (! = 0)
The imaginary part 00AA (!) of the generalized susceptibility is an odd function of !.
Taking advantage of this property, we can deduce from the fluctuation-dissipation
theorem the expression for AA (! = 0) in terms of an equilibrium correlation function.
The function 00AA (!) being odd, 0AA (! = 0) identifies with the static suscepti-
bility. The Kramers–Kronig relation for 0AA (! = 0),
Z 1
1 AA (!)
00
AA (! = 0) = (14.5.1)
0
d!,
⇡ 1 !
Static susceptibility 399
We thus retrieve the proportionality relation between the static susceptibility and the
equal time canonical correlation function:11
⌦ ↵
AA (! = 0) = A2 . (14.5.5)
@A
AA (! = 0) = = (14.5.6)
T
AA ·
@a T
@A
= K̃AA (t = 0). (14.5.7)
@a T
11
The expression for BA (! = 0) involves in principle the equal time correlation function
K̃B = 0), which reduces to K̃BA (t = 0) since we have assumed A0 = 0, B 0 = 0.
B 0 ,A A0 (t
400 The fluctuation-dissipation theorem
6. Sum rules
Generally speaking, the sum rules of the linear response theory are exact integral rela-
tions that any generalized susceptibility must verify. They stipulate that each moment
of BA (!) must be given by an equal time equilibrium correlation function of certain
derivatives of the operators A(t) and B(t). The thermodynamic sum rule and the sum
rule of the oscillator strengths are well-known examples of sum rules. In practice, the
sum rules impose constraints on the phenomenological models likely to be proposed
for the generalized susceptibilities (in particular concerning their behavior at large
angular frequencies).
We will establish the sum rules, limiting ourselves to the simple case of the re-
sponse of a physical quantity A to its own conjugate field.
6.1. Derivation
To study the behavior of the susceptibility AA (!) at large angular frequencies, we in-
troduce the function AA (z), defined by its spectral representation in terms of 00AA (!):
Z 1
1 AA (!)
00
AA (z) = d!, =m z 6= 0. (14.6.1)
⇡ 1 ! z
The coefficients of the expansion (14.6.2) are proportional to the moments of order
n 1 of 00AA (!)/!.
According to the fluctuation-dissipation theorem (14.3.4), we have the equality:
Z 1 Z 1
1 AA (!)
00
!n d! = ! n KAA (!) d!, n 0. (14.6.3)
⇡ 1 ! 2⇡ 1
The expansion (14.6.2) thus involve the moments of order n 1 of KAA (!). According
to the properties of the Fourier transformation, each moment of KAA (!) (including the
zeroth-order moment, which does not appear in the expansion (14.6.2)) is proportional
to the derivative of corresponding order,12 computed at t = 0, of the autocorrelation
12
We have, by definition:
Z 1
1 i!t
K̃AA (t) = KAA (!)e d!.
2⇡ 1
Hence the expression for the moment of order n of KAA (!) is:
Z 1 ✓ ◆n
1 d
! n KAA (!) d! = i K̃AA (t) , n 0.
2⇡ 1 dt t=0
Sum rules 401
Z 1 ⇣ d ⌘n
1 AA (!)
00
!n d! = i K̃AA (t) , n 0. (14.6.4)
⇡ 1 ! dt t=0
The expression appearing on the right-hand side of equation (14.6.4) is a finite quantity.
The relations obtained in this way for the various values of n 0 constitute the sum
rules that 00AA (!) must satisfy.
Formulas (14.6.4) are of interest for even n only. Indeed, since 00AA (!) is an odd
function of !, the moments of odd order of 00AA (!)/! vanish. Accordingly, K̃AA (t)
is an even function of t analytic at t = 0. Its derivatives of odd order at the origin
vanish.
In the classical case, we deduce from equation (14.6.4), written for the moment
of order 2p of 00AA (!)/!, the relation:
Z 1
1 ⌦ 2↵
AA (!) d! = [A(p) ] , 0.
1 00
! 2p p (14.6.5)
⇡ 1
This property is known as the f -sum rule. We will check it for the example of the
polarization of an atom perturbed by an electric field.
The generalized susceptibility associated with the polarization of an atom, initially
in its fundamental state | 0 i and perturbed by an electric field parallel to the Ox axis,
is:
X ✓ ◆
e2 2 1 1
(!) = lim |h n |x| 0 i| + · (14.6.9)
h̄ ✏!0+ n ! !n0 + i✏ ! + !n0 + i✏
402 The fluctuation-dissipation theorem
In formula (14.6.9), the states {| n i} (n > 0) are the excited unperturbed eigenstates
and the !n0 ’s are the Bohr angular frequencies. In particular, we have:
⇡e2 X 2⇥ ⇤
00
(!) = |h n |x| 0 i| (! !n0 ) (! + !n0 ) . (14.6.10)
h̄ n
The f -sum rule (14.6.8) reads, in this case:
2m X 2
|h n |x| 0 i| !n0 = 1. (14.6.11)
h̄ n
Formula (14.6.11) is simply the Thomas–Reiche–Kuhn (oscillator strength) sum rule.
The odd order moments of 00xx (!)/! actually vanish. The first two even order moments
of 00xx (!)/! are finite. The thermodynamic sum rule and the f -sum rule are verified
for any . We actually have the equalities:
Z 1
1 1
2 d! = (14.6.14)
m⇡ 1 (! 2 !02 ) + 2 ! 2 m!02
(thermodynamic sum rule) and:
Z 1
1 !2 1
2 d! = (14.6.15)
m⇡ 1 (! 2 !02 ) + 2 !2 m
(f -sum rule). Formulas (14.6.14) and (14.6.15) express the fact that the oscillator in
a viscous fluid reaches thermal equilibrium as a consequence of collisions with the
molecules of the fluid: m!02 hx2 i/2 = kT /2, mhẋ2 i = kT /2.
However, the moments of even order n 4 of 00xx (!)/! diverge. This divergence
displays the fact that the viscous damping model with an !-independent friction coef-
ficient is not satisfactory at large angular frequencies. This model can be improved by
the introduction of a generalized friction coefficient (!) (the equation of motion of the
oscillator then involves a retarded friction force). For instance, with the generalized
friction coefficient:
!c
(!) = , (14.6.16)
!c i!
in which !c denotes an angular frequency characteristic of the damping fluid, the
moment of order 4 of 00xx (!)/! becomes finite, the moments of even order n > 4 of
this quantity being, however, still divergent.
Bibliography 403
Bibliography
P.M. Chaikin and T.C. Lubensky, Principles of condensed matter physics, Cam-
bridge University Press, Cambridge, 1995.
C. Cohen-Tannoudji, B. Diu, and F. Laloë, Quantum mechanics, Vol. 2, Hermann
and Wiley, Paris, second edition, 1977.
S. Dattagupta, Relaxation phenomena in condensed matter physics, Academic Press,
Orlando, 1987.
D. Forster, Hydrodynamic fluctuations, broken symmetries, and correlation func-
tions, Westview Press, Boulder, 1995.
R. Kubo, M. Toda, and N. Hashitsume, Statistical physics II: nonequilibrium
statistical mechanics, Springer-Verlag, Berlin, second edition, 1991.
P.C. Martin, Measurements and correlation functions, Les Houches Lecture Notes
1967 (C. De Witt and R. Balian editors), Gordon and Breach, New York, 1968.
A. Messiah, Quantum mechanics, North-Holland, Amsterdam, 1970.
D. Zubarev, V. Morozov, and G. Röpke, Statistical mechanics of nonequilibrium
processes, Vol. 2: Relaxation and hydrodynamic processes, Akademie Verlag, Berlin,
1997.
R. Zwanzig, Nonequilibrium statistical mechanics, Oxford University Press, Oxford,
2001.
References
H. Nyquist, Thermal agitation of electric charge in conductors, Phys. Rev. 32, 110
(1928).
L. Onsager, Reciprocal relations in irreversible processes. I., Phys. Rev. 37, 405
(1931); II., Phys. Rev. 38, 2265 (1931).
H.B. Callen and T.A. Welton, Irreversibility and generalized noise, Phys. Rev.
83, 34 (1951).
H.B. Callen and R.F. Greene, On a theorem of irreversible thermodynamics, Phys.
Rev. 86, 702 (1952).
R. Kubo, The fluctuation-dissipation theorem and Brownian motion, 1965, Tokyo
Summer Lectures in Theoretical Physics (R. Kubo editor), Syokabo, Tokyo and Ben-
jamin, New York, 1966.
R. Kubo, The fluctuation-dissipation theorem, Rep. Prog. Phys. 29, 255 (1966).
Supplement 14A
Dissipative dynamics
of a harmonic oscillator
p2 1
H0 = + m!02 x2 , (14A.2.1)
2m 2
sin !0 t
˜xx (t) = ⇥(t) · (14A.2.8)
m!0
✓ ◆
1 1 1
xx (!) = lim + · (14A.2.10)
2m!0 ✏!0+ ! !0 + i✏ ! + !0 + i✏
h̄ h̄!0
S̃xx (t) = coth cos !0 t, (14A.2.16)
2m!0 2
⇡h̄ h̄!0 ⇥ ⇤
Sxx (!) = coth (! !0 ) + (! + !0 ) . (14A.2.17)
2m!0 2
h̄!
Sxx (!) = h̄ coth xx (!).
00 (14A.2.19)
2
In fact, the uncoupled oscillator, which is not damped, does not constitute a
genuine dissipative system. According to expression (14A.2.11) for 00xx (!), the oscil-
lator absorbs energy only at its own angular frequency !0 . As shown by expression
(14A.2.17) for Sxx (!), the displacement fluctuations show up exclusively at this angu-
lar frequency. The fluctuation-dissipation theorem treats this limiting case consistently.
Response functions and susceptibilities of the coupled oscillator 407
where the index n = 1, . . . , N denotes each of the bath’s oscillators (we can consider
that these oscillators represent phonon modes). The bilinear coupling term between
the oscillator under study and the bath is:
N
X N
X
cn xn x = h̄ gn (bn + b†n )(a + a† ), (14A.3.2)
n=1 n=1
where bn and b†n are the annihilation and creation operators relative to the mode n.
1/2
The quantities cn or gn = cn (4mmn !0 !n ) are coupling constants. We generally
neglect the terms bn a + bn a , which correspond to processes in which two quanta are
† †
The coupling between the system and the bath is performed through processes in
which the oscillator gains a quantum to the detriment of mode n, and conversely.
d ˜BA (t) i ⌦ ↵ i ⌦ ↵
= (t) [B, A] + ⇥(t) [Ḃ(t), A] . (14A.3.5)
dt h̄ h̄
2
These terms correspond to rapidly oscillating contributions whose e↵ect may be neglected. Such
an approximation is commonly used in optics, where it takes the name of rotating wave approximation.
3
We use here the notations of the general linear response theory.
408 Dissipative dynamics of a harmonic oscillator
The second term on the right-hand side of equation (14A.3.5) identifies with ˜ḂA (t).
We thus have:
d ˜BA (t) i ⌦ ↵
= (t) [B, A] + ˜ḂA (t). (14A.3.6)
dt h̄
Hence, the susceptibilities bn a† (! + i✏) once eliminated and the limit ✏ ! 0+ once
taken, we have:
1 1
aa† (!) = lim P 1· (14A.3.10)
h̄ ✏!0+ ! + i✏ !0 n gn (!
2 + i✏ !n )
Analysis of xx (!) 409
Since aa (!) and a† a† vanish even in the presence of the coupling, the general-
ized susceptibility xx (!) of the coupled oscillator can be deduced from aa† (!) and
a† a (!):
h̄ ⇥ ⇤
xx (!) = aa† (!) + a† a (!) . (14A.3.12)
2m!0
To study xx (!), we thus have to analyze formulas (14A.3.10) and (14A.3.11).
4. Analysis of xx (!)
P
It is proportional to the density of phonon modes Z(!) = N 1
n (! !n ). Setting
2
ge↵ = N 1 g 2 , gives:
(!) = ⇡Z(!). (14A.4.6)
(
3! 2 /!D
3
, 0 < ! < !D
Z(!) = (14A.4.7)
0, ! > !D .
Z !D
3! 02 1
(!) = vp 3 ! d! 0 . (14A.4.8)
0 !D !0
In particular, we have:
1
(!) ' , |!| !D , (14A.4.9)
!
and:4
3
(! = 0) = · (14A.4.10)
2!D
The integration on the right-hand side of equation (14A.4.8) once carried out, we get:
3 h1 ! ⇣ ! ⌘2 !D ! i
(!) = + + ln · (14A.4.11)
!D 2 !D !D !
4
As displayed by the general formula (14A.4.4), the behavior (!) ' ! 1 at large angular
frequencies of the function (!) is not specific to the Debye model, but it is realized in all cases (this
behavior is solely linked to the normalization of the density of modes Z(!)). Besides, the property
(! = 0) < 0 also holds
R independently of the modelization chosen for the density of modes, since we
have (! = 0) = vp 11 [Z(!)/!] d! with Z(!) a positive quantity.
Analysis of xx (!) 411
Δ(ω)
ωD ω
-3/2ωD
As (! = 0) < 0, the stability condition (14A.4.16) implies that the coupling constant
g cannot exceed some value gmax . In the particular case of the Debye model, in which
1/2
(! = 0) is given by formula (14A.4.10), we have gmax = (2!0 !D /3) .
ω0 > ωD Δ(ω)
◦
◦
ωD ω0 ω
◦
◦
ω0 < ωD Δ(ω)
◦
◦
ω0 ωD ω
◦
◦
When the stability condition (14A.4.16) is fulfilled, the ordinate at the origin of the
straight line of equation f (!) = g 2 (! !0 ) is smaller than (! = 0) = 3/2!D . All
Analysis of xx (!) 413
then correctly represented, for the whole set of values of !, by a Lorentzian centered
at !m ' !0 + g 2 (!0 ), and of width 2g 2 (!0 ):
1
aa† (!)
00
=
m
2
, m = g 2 (!0 ). (14A.4.21)
h̄ (! !m ) + 2
m
6
There is a unique solution in weak coupling, whatever the value of !0 . The logarithmic divergence
of the function (!) at ! = !D is an artefact of the Debye model (in a more realistic description of
the density of modes, the function (!) would be negative for small values of ! and positive for large
ones, the curve (!) having the shape of a dispersion curve with a maximum close to the maximal
value of the phonon angular frequency).
414 Dissipative dynamics of a harmonic oscillator
• Resonance
When !0 > !D , and whatever the coupling strength, one of the solutions of equation
(14A.4.17) is larger than !0 , that is, larger than !D . For this solution, denoted by !m1 ,
we have Z(!m1 ) = 0 and (!m1 ) = 0. The corresponding response is not damped. In
the vicinity of !m1 , we have:
1 1
aa† (!) ⇡ (! !m1 ). (14A.4.23)
00
'
h̄ 1 g2 0 (!
m1 )
h̄ ⇥ ⇤
xx (!) ' aa† (!) + a† a (!) . (14A.4.25)
2m!m
h̄ ⇥ ⇤
xx (!) aa† (!) aa† ( (14A.4.26)
00 00 00
' !) .
2m!m
h̄ ⇥ ⇤
xx (!) aa† (!) + aa† ( (14A.4.28)
0 0 0
' !) .
2m!m
Dynamics of the weakly coupled oscillator 415
Using then the weak-coupling expression (14A.4.22) for 0aa† (!) gives:
1 ! !m ! + !m
0
(!) ' + · (14A.4.29)
xx
2m!m 2
(! !m ) + m 2 2
(! + !m ) + m 2
sin !m t
˜xx (t) ' ⇥(t) e mt
. (14A.5.1)
m!m
h̄ h̄! m m
Sxx (!) = coth · (14A.5.2)
2m!m 2 (! !m )
2
+ 2
m (! + !m )
2
+ 2
m
We can compute S̃xx (t) by inverse Fourier transformation. The function S̃xx (t) has
in general no simple analytic expression.7 However, in the classical limit h̄! ⌧ 1 in
which S̃xx (t) is given by the formula:
Z
2kT m 1 1
S̃xx (t) = ⇥ 2 ⇤⇥ 2 ⇤ e i!t d!, (14A.5.3)
⇡m 1 (! !m ) + m (! + !m ) + m
2 2
7
The poles of the function coth h̄z/2 indeed play a role in the computation of the Fourier integral
giving S̃xx (t).
416 Dissipative dynamics of a harmonic oscillator
kT ⇣ m
⌘
S̃xx (t) ' 2
cos ! m t + sin ! m |t| e m |t|
. (14A.5.4)
m!m !m
Formulas (14A.5.1) and (14A.5.4) show that, in the presence of a weak coupling
with the phonon bath, the motion of the oscillator in classical regime consists of a
damped oscillation whose angular frequency !m is shifted with respect to !0 by a
value proportional to the squared coupling constant. The correlation time ⌧c = m1
varies as the inverse of the squared coupling constant. It diverges as the latter vanishes,
the oscillation then persisting indefinitely without decreasing.
Bibliography 417
Bibliography
P.M. Chaikin and T.C. Lubensky, Principles of condensed matter physics, Cam-
bridge University Press, Cambridge, 1995.
C. Cohen-Tannoudji, J. Dupont-Roc, and G. Grynberg, Atom–photon interac-
tions: basic processes and interactions, Wiley, New York, 1992.
H. Haken, Synergetics, Springer-Verlag, Berlin, third edition, 1983.
W.H. Louisell, Quantum statistical properties of radiation, Wiley, New York, 1973.
S.W. Lovesey, Condensed matter physics: dynamic correlations, The Benjamin/
Cummings Publishing Company, Reading, second edition, 1986.
P.C. Martin, Measurements and correlation functions, Les Houches Lecture Notes
1967 (C. De Witt and R. Balian editors), Gordon and Breach, New York, 1968.
H.L. Pécseli, Fluctuations in physical systems, Cambridge University Press, Cam-
bridge, 2000.
U. Weiss, Quantum dissipative systems, World Scientific, Singapore, third edition,
2008.
References
In formula (15.1.1), the {ri }’s are the position operators of the di↵erent electrons of
the considered sample.
The electric current density operator at point r is defined by:
e X⇥ ⇤
J (r) = vi (r ri ) + (r ri )vi , (15.1.2)
2 i
where vi = ṙi is the velocity of the electron i. Since the field is uniform, the current
density is also uniform, and it can therefore be written as:
Z
1
J= J (r) dr (15.1.3)
V
(V denotes the volume of the sample). Using formula (15.1.2), this gives:
e X
J= vi . (15.1.4)
V i
In the linear response regime, the average value hJ↵ (t)i of the component of J
parallel to the direction ↵ reads:
Z 1
⌦ ↵
J↵ (t) = ˜BA (t t0 )E (t0 ) dt0 , (15.1.5)
1
The Kubo–Nakano formula 421
P
with A = e i ri, and B = J↵ . According to the general linear response theory, the
response function ˜BA (t) is given by the formula:
i ⌦ X ↵
˜BA (t) = ⇥(t) [J↵ (t), e ri, ] . (15.1.6)
h̄ i
Z 1 Z
⌦ ↵
↵ (!) = V lim+ dt e (i! ✏)t
J ( ih̄ )J↵ (t) d . (15.1.8)
✏!0 0 0
1
Using the fact that the response function ˜BA (t) is real, we can also write the components of
the conductivity tensor in uniform electric field in the equivalent form:
Z 1 Z
↵ (!) = V lim dt e(i! ✏)t hJ↵ J ( t + ih̄ )i d ,
✏!0+ 0 0
that is: Z Z
⇥ ⇤
Ḣ1 = r. (r, t)J (r) dr + J (r).r (r, t) dr. (15.1.12)
The first term on the right-hand side of equation (15.1.12) can be transformed into a
surface integral of the flux (r, t)J (r), and then cancelled owing to a proper choice of
boundary conditions. Formula (15.1.12) then reduces to:
Z
Ḣ1 = E (r, t)J (r) dr. (15.1.13)
and: Z Z
⌦ ↵ ⌦ ↵
J↵ (q, !) = dr J↵ (r, t) ei(!t q.r)
dt. (15.1.18)
Z 1 Z
✏)(t t0 )
⌦ ↵
⇥ lim+ d(t t0 ) e(i! J (r 0 , ih̄ )J↵ (r, t t0 ) d .
✏!0 0 0
(15.1.20)
R
Introducing the supplementary integration V1 dr = 1, and carrying out the change of
variables (r, r r 0 ) ! (r, r 0 ) in the double integral over space variables thus obtained,
we can recast formula (15.1.20) in a form involving a correlation function of the spatial
Fourier transforms of the current densities:2
Z 1 Z
1 ⌦ ↵
↵ (q, !) = lim dt e(i! ✏)t
J ( q, ih̄ )J↵ (q, t) d , (15.1.21)
V ✏!0+ 0 0
If the one-particle eigenstates are plane waves, the field operators (annihilation and
creation of particles at point r) are defined by the formulas:
8 X
>
> (r, t) = V 1/2 aki (t)eiki .r
>
< i
X (15.2.2)
>
> †
: (r, t) = V aki (t)e
† 1/2 iki .r
> .
i
More generally, if the one-particle eigenstates are functions ki (r), we define the field
operators as follows: 8 X
>
> (r, t) = aki (t) ki (r)
>
< i
X † (15.2.3)
>
>
: (r, t) = aki (t) ⇤ki (r).
†
>
i
5
In formulas (15.2.5) and (15.2.6), the operators a†k1 and ak2 are taken at time t (for the sake of
simplicity, this time-dependence is not explicitly displayed).
6
The eigenstates | ki i are denoted for short by |ki i in the remainder of this section even if they
are not plane waves.
The Kubo–Greenwood formula 425
where nki = ha†ki aki i is the average number of electrons in state |ki i at thermodynamic
equilibrium. Since there is no current at equilibrium, we have:
X
hk|p|ki nk = 0. (15.2.11)
k
Taking into account equations (15.2.13), (15.2.14), and (15.2.15), we finally obtain the
Kubo–Greenwood formula for the real part of the conductivity tensor at vanishing
wave vector:7
⇡e2 X
<e ↵ (!) = hk1 |p↵ |k2 ihk2 |p |k1 i(nk1 nk2 ) [h̄! ("k2 "k1 )].
m2 !V
k1 ,k2
(15.2.16)
Formula (15.2.16) allows us to relate the real part of the conductivity at vanishing wave
vector of a non-interacting electron gas to transitions between one-particle stationary
states.
Therefore, for the dissipative part of this susceptibility, that is, for the spectral function
⇠BA (!), we get:
Formula (15.2.18) displays the fact that, as stated above, the real part of the conduc-
tivity tensor corresponds to the dissipative part of this susceptibility.
In a conductor, the real part of the conductivity tensor and the imaginary part
of the dielectric permittivity tensor are related to one another by the formula:8
4⇡
=m "(!) = <e (!). (15.2.19)
!
In the framework of an independent electron model, the Kubo–Greenwood formula
for <e (!) thus also gives access to the electromagnetic absorption properties as
characterized by =m "(!).
(H0 + Vi )| ki i = E ki | ki i, (15.3.1)
The matrix elements which have to be taken into account in the Kubo–Greenwood
formula (15.2.16) are these matrix elements between perturbed states. The system,
disordered, can be considered as isotropic: ↵ = ↵ . We obtain:
⇡e2 X 2 1
2
2 |hk1 |Vi |k2 i| ⇥ ⇤
<e (!) = h̄ |k1 k2 | (nk1 nk2 ) h̄! (✏k2 ✏k1 ) .
m2 !V 3 2
(h̄!)
k1 ,k2
(15.3.5)
This perturbative calculation is only valid for Vi ⌧ h̄!, where Vi represents a typical
value of the modulus of the matrix elements hk1 |Vi |k2 i. Therefore, it cannot be applied
at vanishing angular frequency.
In a metal, both |k1 | and |k2 | are close to the modulus kF of the Fermi wave
vector. We therefore have, ✓ denoting the angle between k1 and k2 :
2
|k1 k2 | ' 2kF2 (1 cos ✓). (15.3.6)
Also, we have:
n k1 nk2 = f0 ("k1 ) f0 ("k2 ), (15.3.7)
where f0 denotes the Fermi–Dirac function. Since, in the corresponding term of the
Kubo–Greenwood formula, "k2 and "k1 are related by "k2 "k1 = h̄!, we can write,
for h̄! ⌧ "F , where "F is the Fermi energy,
@f0
f0 ("k1 ) f0 ("k2 ) ' h̄! , (15.3.8)
@"k1
e2 X h̄2 k12 1
<e (!) = ("k1 "F ). (15.3.10)
V m2 ! 2 3 ⌧ (k1 )
k1
1
In formula (15.3.10), the inverse relaxation time [⌧ (k1 )] is defined by the usual
formula for electron–impurity collisions:
1 2⇡ X 2
= |hk1 |Vi |k2 i| (1 cos ✓) ("k2 "k1 ). (15.3.11)
⌧ (k1 ) h̄
k2
3.2. Conductivity
The real part of the conductivity may be easily computed due to the presence of the
delta function in the sum on the right-hand side of equation (15.3.10). Assuming that
Conductivity of an electron gas in the presence of impurities 429
the relaxation time is a function of the electron energy, this latter equation can be
rewritten as:
where n = kF3 /3⇡ 2 is the density of the electron gas. This gives:
ne2
<e (!) = · (15.3.14)
m! 2 ⌧ ("F )
ne2 ⌧ ("F )
<e (!) = , (15.3.15)
m! 2 ⌧ 2 ("F )
ne2 ⌧ ("F ) 1
<e (!) = · (15.3.16)
m 1 + ! ⌧ 2 ("F )
2
dhvi hvi
m +m = eE(t).
dt ⌧
ne2 ⌧ 1
(!) = ·
m 1 i!⌧
10
An alternative derivation of <e (!) from the Kubo–Greenwood formula, in which the condition
!⌧ ("F ) 1 is not assumed to be fulfilled, is presented in Supplement 15A.
430 Quantum theory of electronic transport
"F ⌧ ("F ) h̄ is not fulfilled that specific quantum e↵ects are likely to appear. This
latter condition also reads, more simply, in the form of the Io↵e–Regel criterion,
kF ` 1, (15.3.17)
where ` ⇠ h̄kF ⌧ ("F )/m is the elastic mean free path of the electrons.
When disorder increases in such a way that the Io↵e–Regel criterion is violated,
corrections to Drude’s theory must be accounted for. These corrections, which can
be shown to involve terms in inverse powers of kF `, are due to quantum interference
e↵ects involving length scales much larger than `. The elastic scattering by impurities
does not destroy the phase coherence of the scattered waves. This coherence disappears
over a coherence length ` (T ) `, which generally depends on dynamical mechanisms
(impurity motion, phonon scattering . . .).
In mesoscopic systems of size L ` (T ), quantum interference e↵ects, likely
to lead to electronic localization (that is, to the absence of conduction) may modify
the transport properties.11 For their study, we have to devise a convenient method
of calculation of the electrical conductivity. Such a method was first proposed by
R. Landauer in 1957.12
11
The coherence length ` (T ) is a decreasing function of temperature. Any system at sufficiently
low temperature is thus mesoscopic.
12
See Supplement 15A.
Bibliography 431
Bibliography
N.W. Ashcroft and N.D. Mermin, Solid state physics, Holt-Saunders, Philadel-
phia, 1976.
A.L. Fetter and J.D. Walecka, Quantum theory of many-particle systems,
McGraw-Hill, New York, 1971. Reprinted, Dover Publications, New York, 2003.
D. Forster, Hydrodynamic fluctuations, broken symmetries, and correlation func-
tions, Westview Press, Boulder, 1995.
Y. Imry, Introduction to mesoscopic physics, Oxford University Press, Oxford, second
edition, 2002.
W. Jones and N.H. March, Theoretical solid-state physics: non-equilibrium and
disorder , Vol. 2, Wiley, New York, 1973. Reprinted, Dover Publications, New York,
1985.
C. Kittel, Quantum theory of solids, Wiley, New York, second edition, 1967.
R. Kubo, M. Toda, and N. Hashitsume, Statistical physics II: nonequilibrium
statistical mechanics, Springer-Verlag, Berlin, second edition, 1991.
M. Plischke and B. Bergersen, Equilibrium statistical physics, World Scientific,
Singapore, third edition, 2006.
H. Smith and H.H. Jensen, Transport phenomena, Oxford Science Publications,
Oxford, 1989.
R. Zwanzig, Nonequilibrium statistical mechanics, Oxford University Press, Oxford,
2001.
References
M.S. Green, Marko↵ random processes and the statistical mechanics of time-
dependent phenomena, J. Chem. Phys. 20, 1281 (1952).
M.S. Green, Marko↵ random processes and the statistical mechanics of time-
dependent phenomena. II. Irreversible processes in fluids, J. Chem. Phys. 22, 398
(1954).
R. Kubo, Statistical-mechanical theory of irreversible processes. I. General theory
and simple applications to magnetic and conduction problems, J. Phys. Soc. Japan
12, 570 (1957).
H. Nakano, A method of calculation of electrical conductivity, Prog. Theor. Phys.
17, 145 (1957).
432 Quantum theory of electronic transport
1. Introduction
This supplement deals with the electrical conductivity in uniform electric field of a
weakly disordered conductor at zero temperature. In the case of a macroscopic sample,
the conductivity can be computed with the aid of the Kubo–Greenwood formula. This
way to determine the conductivity, which makes use of the relation between its real
part and the electromagnetic absorption properties, corresponds to an experimental
procedure in which, after having placed the conducting sample devoid of contacts in an
electromagnetic cavity, we measure the supplementary absorption due to its presence.
There is another approach, due to R. Landauer, allowing us to compute the con-
ductivity of a disordered conductor. Landauer’s approach is better suited to the case
of systems of small dimensions (mesoscopic) and to the discussion of the localization
phenomenon (absence of conduction) in one-dimensional disordered systems.
In equations (15A.2.1), J denotes the current density and ⇢ the density of free charges.
Defining as usual the spatial and temporal Fourier transforms of the various fields
involved, we write in linear regime the relations D(q, !) = " (q, !).E(q, !) and
B(q, !) = µ(q, !).H(q, !), as well as Ohm’s law J (q, !) = (q, !).E(q, !). The
tensors " and µ are respectively the dielectric permittivity tensor and the magnetic
permeability tensor of the medium,1 whereas is the electrical conductivity tensor.
1
Remember that we are using Gauss units, in which the dielectric permittivity of vacuum equals
unity.
434 Conductivity of a weakly disordered metal
4⇡
"(!) = " (!) + i (!). (15A.2.2)
!
The global permittivity "(!) includes the contribution of electrons of all bands, even
of those which are only partially filled. This can be shown by analyzing in two di↵erent
ways the Maxwell equation (15A.2.1) for r ⇥ H.
If all the e↵ects of the response of the electrons are described through a global
dielectric permittivity "(!), we consider that there is no current in the medium and
we write D = "E. This gives:
1⇥ ⇤
r⇥H = i!"(!) E. (15A.2.3)
c
If we adopt the point of view according to which the response of the electrons is
described both by a dielectric permittivity " (!) and an electrical conductivity (!),
the medium is crossed through by a current J = E and we have:
1⇥ ⇤
r⇥H = 4⇡ (!) i!" (!) E. (15A.2.4)
c
Identifying the expressions (15A.2.3) and (15A.2.4) for r⇥H, we get formula (15A.2.2).
As a general rule, the Kubo formula for the electrical susceptibility tensor reads:
e2 X h n |x↵ | q ih q |x | ni
↵ (!) = lim+ (⇧n ⇧q ) , (15A.2.6)
V h̄ ✏!0 n,q !qn ! i✏
where, at fixed temperature and chemical potential, the average occupation probability
at equilibrium ⇧n of the state | n i is the Fermi–Dirac function f0 ("n ). Like the tensors
" (!) and (!), the tensor (!) is assumed here diagonal: ↵ (!) = (!) ↵ . We
have:
e2 X⇥ ⇤ |h q |x| n i|2
(!) = lim+ f0 ("n ) f0 ("q ) · (15A.2.7)
V h̄ ✏!0 n,q !qn ! i✏
We can then deduce from (!) the dielectric permittivity "(!) = 1 + 4⇡ (!).
The polarization of the electrons of the completely filled bands generally leads to
" (!) real. According to formula (15A.2.2), this gives:
4⇡
=m "(!) = <e (!), (15A.2.8)
!
that is:
<e (!) = ! =m (!). (15A.2.9)
Making use of formula (15A.2.7) for (!), we obtain an expression for the real part of
the conductivity at vanishing wave vector:
⇡e2 ! X⇥ ⇤ 2
<e (!) = f0 ("n ) f0 ("q ) |h q |x| n i| (!qn !). (15A.2.10)
V h̄ n,q
⇡e2 X X 2⇥ ⇤
<e (!) = |h q |vx | n i| f0 ("n ) f0 ("q ) ("q "n h̄!).
!V n
q6=n
(15A.2.11)
2
We assume here that the polarization of the medium is uniquely of electronic origin.
3
The fields varying little over a distance of the order of the electronic mean free path, they may
be considered as spatially uniform.
436 Conductivity of a weakly disordered metal
We thus retrieve the Kubo–Greenwood formula, previously deduced from the general
Kubo–Nakano formula giving the conductivity in terms of a correlation function of
currents.
The average power absorbed per unit volume by the conductor placed in an elec-
tromagnetic wave of electric field E(t) = <e[E0 e i!t ] is:
dW 1
= E02 ! 00
(!), (15A.2.12)
dt 2
that is:
dW 1
= E02 <e (!). (15A.2.13)
dt 2
At zero temperature, the excitations of the system are electron–hole pairs: owing to
the absorption of a quantum of energy h̄!, an electron in a state of energy "n < "F is
put in a state of energy "q (= "n + h̄!) > "F , leaving behind a hole ("F is the Fermi
energy). Owing to the presence of the delta function on the right-hand side of equation
(15A.3.1), we get:
Z "F
n("n )n("n + h̄!) 2
<e (!) = 2⇡e V h̄
2
|h q |vx | n i| d"n . (15A.3.2)
"F h̄! h̄!
Averaging over the positions of the impurities, assumed randomly distributed, gives:
Z "F
n("n )n("n + h̄!) ⌦ 2↵
<e (!) = 2⇡e2 V h̄ |(vx )qn | d"n . (15A.3.3)
"F h̄! h̄!
2 2
In formula (15A.3.3), h|(vx )qn | i denotes the average of |h q |vx | n i| over the di↵erent
configurations of the disordered sample.
If the condition kF ` 1 is fulfilled, where kF denotes the Fermi wave vector and `
the elastic mean free path of the electrons, the eigenstates of H0 may be decomposed
over a base of plane waves |ki, the various plane-wave states forming a wave-packet
of width ⇠ ` 1 : X X q
| ni = ank |ki, | qi = ak |ki. (15A.3.4)
k k
Conductivity of a macroscopic system 437
We set "n = h̄2 kn2 /2m and "q = h̄2 kq2 /2m (with kn ` 1, kq ` 1). We assume that,
as an e↵ect of disorder, the amplitudes ank may be modelized by independent Gaussian
random variables of zero mean and of variance approximately given by a Lorentzian4
of width ` 1 :
⌦ n⇤ q ↵ ⇡ 1
ak ak0 ' nq kk0 2 · (15A.3.5)
`kn V (k kn )2 + (4`2 ) 1
The normalization of | n i is thus guaranteed on average.5 On account of the Gaussian
character of the amplitudes, we can deduce from formulas (15A.3.4) and (15A.3.5) the
2
expression for the quantity h|(vx )qn | i for q 6= n:
⌦ 2↵
X X⌦ h̄kx q q⇤ h̄kx0 n ↵
|(vx )qn | = an⇤
k a a 0 a 0 , q 6= n. (15A.3.6)
m k k m k
k k0
According to formula (15A.3.5), only the terms with k = k0 contribute to the above
average. This gives:
⌦ 2↵ ⇡2 X h̄2 k 2 /3m2
|(vx )qn | = ⇥ 2 1 ⇤⇥ 2 1⇤
, (15A.3.7)
2 2 2 2
` kn kq V
k
(k kn ) + (4`2 ) (k kq ) + (4`2 )
that is:6
⌦ 2↵ 2⇡h̄2 ` 1
|(vx )qn | ' · (15A.3.8)
3V m2 1 + (kq kn )2 `2
Then, coming back to formula (15A.3.3), we obtain:
Z
4⇡ 2 e2 h̄3 ` "F
n("n )n("n + h̄!) 1
<e (!) = d"n . (15A.3.9)
3m2 "F h̄! h̄! 2
1 + (kq kn ) `2
m2 ! 2
1 + (kq
2
kn ) `2 ' 1 + `2 , (15A.3.10)
h̄2 kF2
that is:
2
1 + (kq kn ) `2 ' 1 + ! 2 ⌧ 2 ("F ), (15A.3.11)
4
We have set k = |k|.
5
We have: X
2
h n| ni = |an
k| ,
k
where ⌧ ("F ) = `m/h̄kF is the relaxation time at the Fermi level. The real part of the
conductivity thus reads approximately:
4⇡ 2 e2 h̄3 ` 2 1
<e (!) ' n ("F ) · (15A.3.12)
3m2 1 + ! 2 ⌧ 2 ("F )
To continue the calculation, we need an expression for the density of states in the
disordered metal. We will assume that the density of states is not very di↵erent from
3/2
that of a free electron gas n(") = ⇡ 2 2 1/2 (m/h̄2 ) "1/2 . We then get, on account of
the relation n = kF3 /3⇡ 2 , where n denotes the number of electrons per unit volume:
ne2 ⌧ ("F ) 1
<e (!) ' · (15A.3.13)
m 1 + ! 2 ⌧ 2 ("F )
This is the result, for a harmonic field at the angular frequency !, of the classical
Drude model.
The above method can also be applied to a one-dimensional sample, provided
that it is macroscopic. To carry out the calculation, we have to take into account the
2
appropriate modifications in the expressions for h|(vx )qn | i and n("F ), as well as in
the relation between n and kF .
R
1 T
The gradient of relative density through the barrier is equal to 2R/L. At zero
temperature, assuming for the sake of simplicity that all the electrons have the same
velocity, equal to the Fermi velocity vF , we write the current associated with this
gradient as J = vF (1 R). The current J is a di↵usion current. The coefficient D of
Fick’s law is given by:
vF L 1 R
D= · (15A.4.1)
2 R
The conductance of the sample follows via the fluctuation-dissipation theorem, written
in the form of the relation:
D n @µ
= (15A.4.2)
µD e @n T
between the di↵usion coefficient and the drift mobility (n denotes the number of elec-
trons per unit length). At T = 0, we have:
@µ 1
= , (15A.4.3)
@n T =0 2n("F )
where n("F ) is here the density of states per unit length and per spin direction at the
1
Fermi level. Taking the value n("F ) = (⇡h̄vF ) corresponding to a free electron gas,
we obtain for the conductivity = neµD the expression:
2De2
= · (15A.4.4)
⇡h̄vF
The resistivity ⇢ = 1
is:
⇡h̄vF
⇢= · (15A.4.5)
2De2
440 Conductivity of a weakly disordered metal
e2 T
G= · (15A.4.7)
⇡h̄ R
1 B Beiφ D
A C Ce-iφ
t1 t2 ei
D= · (15A.5.2)
1 e2i r2 r10
T1 T 2
T (2) = 1/2
· (15A.5.4)
1 + R1 R2 2(R1 R2 ) cos ✓
1/2
R 1 + R2 2(R1 R2 ) cos ✓
R(2) = 1/2
· (15A.5.6)
1 + R1 R2 2(R1 R2 ) cos ✓
According to the Landauer formula, the conductance of this obstacle made of two
barriers in series is:
e2 T (2)
G(2) = , (15A.5.7)
⇡h̄ R(2)
442 Conductivity of a weakly disordered metal
The dephasing depends on the distance separating the barriers. We now con-
sider a statistical ensemble of samples made up of the two above described barriers,
separated by distances varying from one sample to another. We denote by h. . .i the
average over this ensemble. We can assume that is uniformly distributed over the
interval (0, 2⇡). Formula (15A.5.5) yields in this case hcos ✓i = 0. We then deduce from
formula (15A.5.8) the average resistance:
D 1 E ⇡h̄ R1 + R2
= 2 · (15A.5.9)
G(2) e (1 R1 )(1 R2 )
Interestingly enough, formula (15A.5.9) shows that Ohm’s law of addition of resis-
tances in series is not valid in general7 (remember, however, that we are carrying out
a zero-temperature calculation).
In view of the calculation of the average resistance of a system of N barriers in
series, it is convenient to rewrite formula (15A.5.9) in the form:
D 1 E ⇡h̄ ✓ R R2 R1 R2
◆
1
= 2 + +2 · (15A.5.10)
G(2) e 1 R1 1 R2 1 R1 1 R2
The addition law (15A.5.10) allows us to compute the average resistance h⇢(2) i of
a set of two systems of resistances ⇢1 and ⇢2 in series:
h⇢(2) i = ⇢1 + ⇢2 + 2⇢1 ⇢2 . (15A.5.11)
Averaging over the parameters related to each system, assumed statistically indepen-
dent of one another, gives:9
h⇢(2) i = h⇢1 i + h⇢2 i + 2h⇢1 ih⇢2 i. (15A.5.12)
7
Ohm’s law would indeed give, for the resistance of the two-barrier set (independently of the
value of ): ✓ ◆
1 ⇡h̄ R1 R2
= + ·
G(2) e2 1 R1 1 R2
In the limit of small resistances, that is, for small reflection coefficients (Ri ⌧ 1), Ohm’s law identifies
with the exact result (15A.5.9).
8
Note that ⇢ does not represent here the resistivity, but rather the inverse of the dimensionless
conductance g.
9
To work with simple notations, we keep the same designation for the average resistance either
before or after the supplementary averaging over the parameters related to each system.
Addition of quantum resistances in series: localization 443
Formula (15A.5.16) allows us to show that log(1 + ⇢) is a relevant scaling variable for
this problem. We indeed get, for the conductor of N barriers considered precedently,
after averaging over the parameters related to each barrier:
⌦ ↵ ⌦ ↵
log(1 + ⇢(N ) ) = N log(1 + ⇢) . (15A.5.17)
10
Indeed, the ratio of the intensity transmitted after passage through a sample of length L and of
absorption coefficient K to the incident intensity is T = e KL .
11
We make use of the formula:
Z 2⇡
1h 1/2
i
log(a + b cos ✓) d✓ = ⇡ log a + (a2 b2 ) .
0 2
444 Conductivity of a weakly disordered metal
Formula (15A.5.17) implies that the typical resistance of a sample of length L follows
the scaling law log(1 + ⇢) = ↵L, from which we deduce:
⇢ = e↵L 1. (15A.5.18)
The inverse localization length ↵ acts as the scaling variable. Formula (15A.5.18) shows
that, for ↵L 1, the typical resistance of a sample of length L increases exponentially
with L (⇢ ' e↵L ), whereas, for ↵L ⌧ 1, we recover a ‘classical’ additive resistance
(⇢ ' ↵L).
Bibliography 445
Bibliography
References
and:
e Xn ⇥ ⇤ ⇥ ⇤ o
J (r, t) = vi (t) r ri (t) + r ri (t) vi (t) . (16.1.2)
2 i
In formulas (16.1.1) and (16.1.2), the {ri }’s and the {vi = ṙi }’s are the position and
velocity operators of the di↵erent electrons. The charge and electric current densities
are related by the continuity equation:
˙ t) + r.J (r, t) = 0.
⇢(r, (16.1.3)
We are interested in the local mean values h⇢(r, t)i and hJ (r, t)i. In the linear
response regime, the average electric current density is related to the electric field
and to the density gradient through the non-local and retarded phenomenological
constitutive equation:
Z Z
⌦ ↵ X
J↵ (r, t) = dr 0 dt0 ˜↵ (r r 0 , t t0 )E (r 0 , t0 )
XZ Z
⌦ ↵
dr 0 dt0 D̃↵ (r r0 , t t0 )r ⇢(r 0 , t0 ) ,
(16.1.4)
in which ˜ (r, t) and D̃(r, t) are respectively the electrical conductivity and di↵usion
tensors.
As usual, we define the spatial and temporal Fourier transforms of the various
quantities involved by formulas of the type:1
Z Z
A(q, !) = dr A(r, t)ei(!t q.r) dt. (16.1.5)
1
We use the same notation A(. , .) for a quantity A(r, t) and its spatial Fourier transform A(q, t),
as well as for its spatial and temporal Fourier transform A(q, !).
The indirect Kubo method 449
By Fourier transforming equations (16.1.3) and (16.1.4), we see that h⇢(q, !)i and
(q, !) are related by:
P
↵ (q, !)q↵ q
⌦ ↵ ↵,
⇢(q, !) = P (q, !). (16.1.6)
i! + D↵ (q, !)q↵ q
↵,
This gives us the following linear relation between the Fourier transforms of the average
electric current density and the electric field:
P
" ↵0 0 (q, !)q↵0 q 0 #
⌦ ↵ X ↵0 , 0
J↵ (q, !) = ↵ (q, !) D↵ (q, !) P E (q, !).
i! + D↵0 0 (q, !)q↵0 q 0
↵0 , 0
(16.1.7)
The relation (16.1.13) may only be satisfied for arbitrary q if there is a propor-
tionality relation between ↵ (q, ! = 0) and D↵ (q, ! = 0) of the form:
After the proportionality constant a has been determined, formula (16.1.15) allows us
to deduce D↵ from ↵ .
Formulas (16.1.16) and (16.1.17) show that the proportionality constant a between
↵ and D↵ represents a static susceptibility of the type BA (! = 0), in which
The indirect Kubo method 451
the operators A and B are both identical to the charge density ⇢(r) = en(r). More
precisely, we have:
a = e2 lim nn (q). (16.1.18)
q!0
⌦ 2↵
nn (q = 0) = V ( n) . (16.1.19)
⌦ 2↵ kT @n
( n) = , (16.1.20)
V @µ T
where µ denotes the chemical potential of the charge carriers. We deduce from formulas
(16.1.19) and (16.1.20) the expression for nn (q = 0):
@n
nn (q = 0) = · (16.1.21)
@µ T
✓ ◆ 1
@µ
a = e2 · (16.1.22)
@n T
We deduce from equations (16.1.15) and (16.1.22) the following relation between
the components of the same indices of the tensors and D in uniform and static
regime:
D↵ 1 @µ
= · (16.1.23)
↵ e2 @n T
In the case of a non-degenerate electron gas, we have (@µ/@n)T = kT /n, and formula
(16.1.23) is the usual Einstein relation.
As a general rule, formula (16.1.23) allows us to determine the di↵usion tensor
given the conductivity tensor, which itself is computed using the Kubo–Nakano formula
(16.1.11). The whole procedure is, for this reason, qualified as the ‘indirect Kubo
method’.2
2
The indirect Kubo method may be used to derive the Green–Kubo formulas associated with
other thermal transport coefficients, such as the thermal conductivity and the viscosity coefficient of
a gas or a liquid.
452 Thermal transport coefficients
where ⇢(r)
˙ corresponds to the unperturbed evolution of ⇢(r), we verify (using inte-
gration by parts) the relation:
Z
Ḣ1 = E(r, t).J (r) dr. (16.2.2)
The entropy source, related to the Joule e↵ect, reads in this case:
1
S = E(r, t).J (r). (16.2.3)
T
The comparison of formulas (16.2.2) and (16.2.3) shows that Ḣ1 /T identifies with the
opposite of the entropy production within the system:
Z
Ḣ1 = T S dr. (16.2.4)
We now intend to generalize this relation in order to be able to define Ḣ1 in situ-
ations for which we cannot write a perturbation Hamiltonian, such as heat conduction
in a fluid. Such an approach of the calculation of the thermal transport coefficients
can, in this sense, be qualified as mechanical.3
but remains however in local equilibrium. The temperature, the chemical potential,
and the local mean velocity become respectively:
8
> T (r, t) = T0 + T (r, t)
>
<
µ(r, t) = µ0 + µ(r, t) (16.2.5)
>
>
:
u(r, t) = u(r, t).
The irreversible processes generated by this departure from equilibrium give rise to a
production of entropy within the fluid. In a pure fluid, there are no di↵usive fluxes. Be-
sides, for the sake of simplicity, we do not take into account here the viscous dissipative
e↵ects. The source of entropy is thus solely related to heat transport. It reads:
1
S = JQ .r( ), (16.2.6)
T
with:
T (r, t) h "+P i
h1 (r, t) = "(r, t) n(r, t) . (16.2.8)
T n
In formula (16.2.8), n(r, t) and "(r, t) denote respectively the local densities of par-
ticles and energy, whereas n, ", P are the global thermodynamic equilibrium values
of the particle density,⇥ the energy
⇤ density, and the pressure. It can be shown that
4
the quantity "(r, t) (" + P)/n n(r, t) represents a local density of thermal energy.
Indeed, at fixed number of particles, we have the thermodynamic relation:
T dS = dE + PdV. (16.2.9)
T "+P
dS = d" dn, (16.2.10)
V n
which allows us to interpret ["(r, t) (" + P)/n]n(r, t) as a local density of thermal
energy.
4
For short, the temperature T0 is from now on denoted by T .
454 Thermal transport coefficients
In order to apply the linear response theory to the system perturbed by H1 (t), we first
note that this operator is of the general form:
Z
H1 = a(r, t)A(r) dr, (16.2.11)
where a(r, t) denotes an applied field coupled to a physical quantity A(r). In the case
of a thermal perturbation, formula (16.2.8) shows that the quantity coupled to the
applied field a(r, t) = T 1 T (r, t) is the local density of thermal energy.
The derivative Ḣ1 is given by:
Z
Ḣ1 = a(r, t)Ȧ(r) dr. (16.2.12)
The evolution of the local density of thermal energy can be deduced from the evolution
of n(r, t) and "(r, t), as governed by the hydrodynamic equations:
8
> @n(r, t) 1
>
< @t + r.g(r, t) = 0
m
(16.2.13)
>
>
: @"(r, t) + r.J (r, t) = 0,
E
@t
in which g(r, t) denotes the local density of kinetic momentum and JE (r, t) the en-
ergy flux. In its linearized form, valid for fluctuations of small amplitude, the kinetic
momentum density reads:
g(r, t) = mnu(r, t). (16.2.14)
As for the linearized energy flux, it reads:5
To examine the relevance of the expression (16.2.8) proposed for h1 (r, t), we have
to check that the derivative Ḣ1 deduced from it e↵ectively verifies the relation (16.2.4)
(the entropy source S being given by formula (16.2.6)). Coming back to the definition
(16.2.12) of Ḣ1 , and making use of the evolution equations (16.2.13), we obtain:
Z
T (r, t) ⇣ "+P ⌘
Ḣ1 = r. JE g dr, (16.2.16)
T mn
5
The use of the hydrodynamic equations assumes slow variations in space and in time of the
various quantities involved. The hydrodynamic regime is defined by the inequalities:
!⌧ ⌧ 1, q` ⌧ 1,
in which ! and q denote an angular frequency and a wave vector typical of the perturbations imposed
on the medium, ⌧ the collision time, and ` the mean free path.
The source of entropy and the equivalent ‘Hamiltonian’ 455
that is: Z
T (r, t)
Ḣ1 = r.JQ dr. (16.2.17)
T
The integral on the right-hand side of equation (16.2.17) can be recast in the following
form:
Z Z Z
⇥ ⇤ ⇥ ⇤
T (r, t) r.JQ (r) dr = r. T (r, t)JQ (r) dr JQ (r).r T (r, t) dr.
(16.2.18)
The first term on the right-hand side of equation (16.2.18) may be transformed into
a surface integral of the flux T (r, t)JQ , and then cancelled owing to a convenient
choice of boundary conditions. This gives:
Z
1
Ḣ1 = T JQ .r( ) dr. (16.2.19)
T
Comparing formulas (16.2.19) and (16.2.6), we actually verify the relation (16.2.4)
between Ḣ1 and S . This justifies a posteriori the introduction of the operator H1 (t)
defined by formulas (16.2.7) and (16.2.8) as an equivalent perturbation ‘Hamiltonian’.
Expressing the response function ˜BA (r r 0 , t t0 ) with the aid of the canonical Kubo
correlation function6 gives:
Z Z Z
⌦ ↵ 1 t ⌦ ↵
JQ↵ (r, t) = dr 0
dt 0
JQ (r 0 , ih̄ )JQ↵ (r, t t0 ) r T (r 0 , t0 ) d .
T 1 0
(16.2.21)
By Fourier transformation of equation (16.2.21), we get:
⌦ ↵
JQ↵ (q, !) = ↵ (q, !)[r T ](q, !). (16.2.22)
6
The derivation of the thermal conductivity tensor may be carried out in a classical framework
if we are interested in the case of a gas or a liquid in the ordinary sense. However, the quantum
formulation of the Green–Kubo formula leaves open the possibility of applying it in other situations,
for instance to get the thermal conductivity tensor of a degenerate electron gas (metal).
456 Thermal transport coefficients
The components ↵ (q, !) of the thermal conductivity tensor (q, !) are given by:7
Z 1 Z
1 ⌦ ↵
↵ (q, !) = lim dt e(i! ✏)t
JQ ( q, ih̄ )JQ↵ (q, t) d .
V T ✏!0+ 0 0
(16.2.23)
The Green–Kubo formula (16.2.23) is an expression for the thermal conductivity in
terms of an equilibrium correlation function of the spatial Fourier transforms of the
heat current densities.
In an isotropic fluid, the thermal conductivity tensor is diagonal: ↵ (q, !) =
(q, !) ↵ . We simply have (when the classical description suffices):
Z 1
1 ⌦ ↵
(q, !) = lim dt e(i! ✏)t
JQ ( q, 0)JQ (q, t) . (16.2.24)
kT 2 V ✏!0+ 0
7
The calculations are similar to those carried out to derive the Kubo–Nakano formula for
↵ (q, !).
Bibliography 457
Bibliography
J.-P. Hansen and I.R. McDonald, Theory of simple liquids, Academic Press, Lon-
don, third edition, 2006.
R. Kubo, M. Toda, and N. Hashitsume, Statistical physics II: nonequilibrium
statistical mechanics, Springer-Verlag, Berlin, second edition, 1991.
L.D. Landau and E.M. Lifshitz, Fluid mechanics, Butterworth-Heinemann, Ox-
ford, second edition, 1987.
References
M.S. Green, Marko↵ random processes and the statistical mechanics of time-depen-
dent phenomena, J. Chem. Phys. 20, 1281 (1952).
M.S. Green, Marko↵ random processes and the statistical mechanics of time-depen-
dent phenomena. II. Irreversible processes in fluids, J. Chem. Phys. 22, 398 (1954).
R. Kubo, Statistical-mechanical theory of irreversible processes. II. Response to ther-
mal disturbance, J. Phys. Soc. Japan 12, 1203 (1957).
L.P. Kadanoff and P.C. Martin, Hydrodynamic equations and correlation func-
tions, Ann. Phys. 24, 419 (1963).
J.M. Luttinger, Theory of thermal transport coefficients, Phys. Rev. 135, A1505
(1964).
R. Zwanzig, Time-correlation functions and transport coefficients in statistical me-
chanics, Ann. Rev. Phys. Chem. 16, 67 (1965).
Supplement 16A
Di↵usive light waves
@I(r, t)
= Dr2 I(r, t) D2 I(r, t) + S(r, t), (16A.1.1)
@t
where D denotes the di↵usion coefficient of light intensity, and S(r, t) a source term
possibly present in the medium. The second term on the right-hand side of equation
(16A.1.1) accounts for absorption, the absorption length being Labs = 1 .
We now assume that there is no source term (S(r, t) = 0), and we take the
initial condition I(r, t = 0) = I0 (r). Considering for the sake of simplicity an infinite
medium, we introduce the spatial Fourier transform1 of I(r, t),
Z
I(q, t) = I(r, t)e iq.r
dr, (16A.1.2)
and then the Fourier–Laplace transform I(q, z) of I(q, t), defined for a complex argu-
ment z of positive imaginary part:
Z 1
I(q, z) = I(q, t)eizt dt, =m z > 0. (16A.1.3)
0
1
We use the same notation I(. , .) for the intensity I(r, t), its spatial Fourier transform I(q, t),
and the Fourier–Laplace transform of this latter quantity, I(q, z).
Di↵usion coefficient of light intensity 459
(16A.2.1)
2
We can also directly solve the di↵erential equation obeyed by I(q, t), which reads:
@I(q, t)
= D(q 2 + 2 )I(q, t), I(q, t) = I0 .
@t
This ends in the result (16A.1.7) without recourse to the Fourier–Laplace transformation. However,
the use of this transformation is highly recommended in the more complex situations in which coupled
variables are involved (an example of such a case is provided in Supplement 16B).
3
The study of the di↵usion of light waves began in astrophysics, with a view to understanding
how the radiation emitted at the center of stars is a↵ected by passing through an interstellar cloud.
The radiative transfer equation was first introduced and used in this context.
460 Di↵usive light waves
dΩ
dσ
In equation (16A.2.1), p(n, n0 )(d⌦0 /4⇡)(d⌦/4⇡) is the fraction of the radiation inten-
sity entering a cone of solid angle d⌦0 around the incident direction n0 , and leaving
a cone of solid angle d⌦ around the scattered direction n. For a medium containing
spherically symmetric scatterers, p(n, n0 ) is a function of n0 .n = cos ⇥, where ⇥ is the
scattering angle (that is, the angle between the incident and scattered directions). The
time ⌧ is the average time separating two successive collision events, and the length `
is the corresponding mean free path. Both parameters are assumed to be constant.
The albedo is a positive quantity, less than 1 or equal to 1, depending on whether the
scattering takes place with or without absorption of light energy. Equation (16A.2.4)
is a conservation equation for I(r, t) only if a = 1 (which is indeed the case when there
is no absorption).
Similarly, multiplying both sides of equation (16A.2.1) by n and integrating
over ⌦, we obtain the evolution equation of J (r, t):
Z
⌧ 2 @J (r, t) ⌧
+ 1 hcos ⇥i J (r, t) = ` n(n.rr )I(r, n, t) d⌦. (16A.2.6)
` @t `
In equation (16A.2.6), we have set:
Z
1
p(n, n0 )n d⌦ = n0 hcos ⇥i, (16A.2.7)
4⇡
where the quantity hcos ⇥i is defined by:
Z
1
hcos ⇥i = hn.n0 i = p(n, n0 )n.n0 d⌦. (16A.2.8)
4⇡
Equations (16.2.4) and (16.2.6) doR not constitute a closed system of equations for
I(r, t) and J (r, t), since the term n(n.rr )I(r, n, t) d⌦ involved on the right-hand
side of equation (16.2.6) cannot be expressed in terms of I(r, t) and J (r, t) only.
However, when the distribution of the light intensity is almost
R isotropic, we can re-
place the specific intensity I(r, n, t) by I(r, t)/4⇡ in the term n(n.rr )I(r, n, t) d⌦.
This approximation allows us to obtain a closed system of equations for I(r, t) and
J (r, t), equation (16.2.6) being replaced by the following approximate equation:5
⌧ 2 @J (r, t) ⌧ `
+ 1 hcos ⇥i J (r, t) = rI(r, t). (16A.2.9)
` @t ` 3
`2
J (q, z) ' iqI(q, z), (16A.2.11)
3⌧ 1 hcos ⇥i
and, consequently:
I0
I(q, z) ' · (16A.2.12)
`2 1 a
iz + q 2
+
3⌧ (1 hcos ⇥i) ⌧
Comparing formula (16.2.12) with formula (16.1.4) shows that the intensity I(r, t)
obeys, in the hydrodynamic limit, a di↵usion equation with absorption. We identify
the di↵usion coefficient of light intensity,
`2
D= , (16A.2.13)
3⌧ 1 hcos ⇥i
1
D= v`tr , (16A.2.15)
3
where:
`
`tr = (16A.2.16)
1 hcos ⇥i
and:
`
v= (16A.2.17)
⌧
are respectively the transport mean free path and the transport velocity.
where the symbol h. . .i denotes the configuration average over the positions of the
scatterers. The function g2 (t) is the autocorrelation function of light intensity. As the
concentration of scatterers is high, we can write approximately:6
⌦ ↵ ⌦ ↵⌦ ↵ ⌦ ↵⌦ ⇤ ↵
I(t)I(0) = (t) ⇤ (t) (0) ⇤ (0) + (t) ⇤ (0) (t) (0) . (16A.3.2)
Using formula (16A.3.2), we obtain for the autocorrelation of light intensity the Siegert
formula: ⌦ ↵
2 (t) ⇤ (0)
g2 (t) = |g1 (t)| , g1 (t) = ⌦ 2↵ · (16A.3.3)
| |
We set:
1
X
(t) = (n)
(t), (16A.3.4)
n=1
where (n) (t) denotes the amplitude of the field of the wave associated with the tra-
jectories in which the light undergoes n scattering events, assumed elastic, between
the source and the measurement point. We can write:
n
Y
(n)
(t) = (n)
(0) exp{ iqi .[ri (t) ri (0)]}, (16A.3.5)
i=1
where the {ri (t)}’s represent the positions of the scatterers at time t whereas the
{qi }’s are the associated scattering wave vectors. We have:
⇥i
|qi | = 2q sin , (16A.3.6)
2
6
To justify formula (16A.3.2), we assume that, as a consequence of the high concentration of
scatterers, the amplitude (t) may be considered as a Gaussian random process.
464 Di↵usive light waves
where ⇥i is the scattering angle for the ith scattering and q the modulus of the
light wave vector. From the decomposition (16A.3.4) of (t) we get the following
decomposition of g1 (t),
1
X (n)
g1 (t) = g1 (t), (16A.3.7)
n=1
The quantity:
⌦ 2↵
| (n)
(0)|
P (n) = ⌦ 2↵ (16A.3.9)
| |
is the probability for a trajectory terminating at the measurement point to undergo n
scattering events. Besides P (n), which concerns the di↵usive properties of light waves,
(n)
the expression (16A.3.8) for g1 (t) involves a configuration average relative to the
suspended particles. This latter average brings into play, on the one hand, an average
over the positions ri of the scatterers, and, on the other hand, an average over the
possible directions of the wave vectors qi . Both sets of random variables will be, in
first approximation, treated as independent.
⇥ ⇤ ⇣ |r (t) r (0)|2 ⌘
3/2 i i
p ri (t) ri (0) = (4⇡Dp t) exp · (16A.3.10)
4Dp t
The average over the positions of the various scatterers once carried out, we get, on
account of formula (16A.3.6):
n
(n) ⌦Y ⇥i ↵
g1 (t) = P (n) exp 4Dp q 2 t sin2 . (16A.3.11)
i=1
2
In formula (16A.3.11), it remains to average over the scattering angles. We can set7
1
t0 = (Dp q 2 ) . At lowest order in t/t0 , the average over the scattering angles may be
approximated by:
n
⌦Y ⇥i ↵ h t i
exp 4Dp q 2 t sin2 ' exp 2n 1 hcos ⇥i i . (16A.3.12)
i=1
2 t0
7
Physically, t0 represents the time necessary for a suspended particle to di↵use over a distance
of the order of the light wavelength.
Di↵usive wave spectroscopy 465
As a consequence we get, in terms of the mean free path ` and the transport mean
free path `tr defined by formula (16A.2.16):
⇣ t ` ⌘
(n)
g1 (t) ' P (n) exp 2n (16A.3.13)
t0 `tr
and:
1
X ⇣ t ` ⌘
g1 (t) ' P (n) exp 2n · (16A.3.14)
n=1
t0 `tr
The exact detailed calculation of p(s) is intricate. We can show that this quantity
is dominated by an exponential factor exp( r2 /4D⌧ph ), resulting from the di↵usive
character of light waves (⌧ph = s/v is the photon transit time and D = v`tr /3 the
di↵usion coefficient of the waves). More precisely, we can show that the density p(s)
is of the form:
⇣ 3r2 ⌘
p(s) = (s) exp , (16A.3.17)
4s`tr
where (s) is a function decreasing like a power law.
We can deduce from formula (16A.3.18) the leading behavior of g1 (t). The argu-
ment of the exponential involved in the integrand has indeed a maximum, situated at
the point s0 where the derivative of the argument vanishes:
3r2 1 t 1
2 = 0. (16A.3.19)
4`tr s20 t0 `tr
466 Di↵usive light waves
1/2
We have s0 = r (3t0 /8t) . Since this maximum takes place in an exponent, its posi-
tion determines the behavior of g1 (t). This gives, approximately:
h r ⇣ 6t ⌘1/2 i
g1 (t) ' exp , (16A.3.20)
`tr t0
which yields for the light intensity autocorrelation function the behavior:
h r ⇣ 6t ⌘1/2 i
g2 (t) ' exp 2 . (16A.3.21)
`tr t0
Bibliography
References
1. Introduction
When a beam of monochromatic light passes through a transparent, dense fluid, some
part of the light is scattered, since the density of the medium is not uniform. Now,
‘frozen’ (that is, time-independent) density fluctuations cannot exist in a fluid. The
light scattered by the density fluctuations of a fluid thus exhibits a spectrum of angular
frequencies characteristic of the time dependence of the density fluctuations.
We aim here to show how we can determine this spectrum in the case of a normal
fluid, that is, a fluid isotropic, made of spherically symmetric molecules, not charged,
and not superfluid. We are interested in the hydrodynamic regime which prevails
when the system, after many collisions, has reached a state of local thermodynamic
equilibrium. If ! and q denote respectively an angular frequency and a wave vector
typical of the perturbations imposed to the medium,1 the hydrodynamic regime is that
of the excitations of low angular frequency and large wavelength, as pictured by the
inequalities:
!⌧ ⌧ 1, q` ⌧ 1, (16B.1.1)
where ⌧ represents the collision time and ` the mean free path.2
In this regime, the fluid is fully described by the local values of the thermodynamic
quantities, whose evolution is determined by the hydrodynamic equations. When the
amplitude of the fluctuations is not too high, these latter equations may be linearized.
To simplify the study of hydrodynamic fluctuations, we first get rid of the bound-
ary conditions by considering an infinite medium, which allows us to carry out a spatial
Fourier transformation. Then the remaining initial conditions problem can be solved
by using the Fourier–Laplace transformation.
3. Transverse fluctuations
We write the kinetic momentum density as the sum of a longitudinal component and
a transverse one,
g(r, t) = gL (r, t) + gT (r, t), (16B.3.1)
with:
r ⇥ gL = 0, r.gT = 0. (16B.3.2)
We introduce the spatial Fourier transform g(q, t) of g(r, t), defined by:
5
Z
g(q, t) = g(r, t)e iq.r dr, (16B.3.3)
as well as the Fourier transforms gL (q, t) and gT (q, t), defined in a similar way from
gL (r, t) and gT (r, t). Note that gL (q, t) and gT (q, t) are respectively parallel and
perpendicular to q.
@gT (r, t) ⌘
r2 gT (r, t) = 0. (16B.3.4)
@t mn
Although gT (r, t) is not, in general, the quantity in which we are primarily interested,6
its evolution equation, decoupled from those of the other hydrodynamic variables, is
simpler to analyze. Equation (16B.3.4) is a di↵usion equation involving a transverse
di↵usion coefficient, related to the viscosity of the fluid and also called the kinematic
viscosity:
⌘
DT = · (16B.3.5)
mn
gT (q, t = 0)
gT (q, z) = · (16B.3.8)
iz + DT q 2
Owing to the di↵usion process, gT (q, z) exhibits a di↵usion pole of affix z0 = iDT q 2 .
We compute gT (q, t) by inverting the Fourier–Laplace transformation:
Z
gT (q, t = 0) e izt
gT (q, t) = dz. (16B.3.9)
2⇡ C iz + DT q 2
we have:
gT (q, t = 0) = gT , (16B.3.12)
⇣ r2 ⌘
gT (r, t) = gT (4⇡DT t)
3/2
exp , t > 0. (16B.3.14)
4DT t
4. Longitudinal fluctuations
The local density of particles, the longitudinal component of the local density of ki-
netic momentum, and the local density of energy obey the following coupled evolution
equations:
8
> @n(r, t) 1
>
> + r.gL (r, t) = 0
>
> @t m
>
>
<
@gL (r, t) ⌘ ⇥ ⇤ ⌘ 2
+ rP(r, t) r r.gL (r, t) r gL (r, t) = 0 (16B.4.1)
>
>
> @t 3mn mn
>
>
>
> @"(r, t) " + P
: + r.gL (r, t) r2 T (r, t) = 0.
@t mn
We choose here as thermodynamic variables the local density of particles and the
local temperature. Once gL (r, t) is eliminated, the problem reduces to the resolution
of two coupled equations for n(r, t) and T (r, t).
4.1. The coupled equations for the density of particles and the temperature
First, taking the divergence of the evolution equation of gL (r, t) (the second of equa-
tions (16B.4.1)), and taking into account the evolution equation of n(r, t) (the first of
equations (16B.4.1)), we get the equation:
⇣ @2 @ 2⌘ 1 2
DL r n(r, t) r P(r, t) = 0, (16B.4.2)
@t2 @t m
where:
4⌘
DL = (16B.4.3)
3mn
Longitudinal fluctuations 473
In equation (16B.4.5), we have introduced the squared adiabatic sound velocity in the
fluid c2s = ( /m)(@P/@n)T ( is the ratio of the specific heats at constant pressure
and constant volume). A priori, the initial conditions term on the right-hand side of
equation (16B.4.5) involves both n(q, t = 0) and ṅ(q, t = 0). It is always possible to
choose ṅ(q, t = 0) = 0 (for instance by assuming that g(q, t = 0) is purely transverse).
Equation (16B.4.5) then reads:9
⇣ c2s q 2 ⌘ q 2 @P
z2 izDL q 2 + n(q, z) + T (q, z) = ( iz + DL q 2 )n(q, 0).
m @T n
(16B.4.6)
We now have to take into account the evolution equation of "(r, t). Eliminat-
ing gL (r, t) between the first and third of equations (16B.4.1), we get the following
equation:
@h "+P i
"(r, t) n(r, t) r2 T (r, t) = 0. (16B.4.7)
@t n
In terms of the chosen thermodynamic
⇥ ⇤ variables n(r, t) and T (r, t), the local density
of thermal energy "(r, t) (" + P)/n n(r, t) reads:
"+P T @S
"(r, t) n(r, t) = n(r, t) + ncv T (r, t), (16B.4.8)
n V @n T
where cv denotes the specific heat at constant volume per particle. Importing expres-
sion (16B.4.8) for the local density of thermal energy into equation (16B.4.7), gives:
T @S @n(r, t) @T (r, t)
+ ncv r2 T (r, t) = 0, (16B.4.9)
V @n T @t @t
9
In the absence of ambiguity, the quantity n(q, t = 0) will simply be written n(q, 0).
474 Light scattering by a fluid
Setting:
a= , (16B.4.11)
ncv
we obtain, after some transformations, the equation:
⇣@ ⌘ T @s @n(r, t)
ar2 T (r, t) + = 0, (16B.4.12)
@t cv @n T @t
T @s T @s
( iz + aq 2 )T (q, z) iz n(q, z) = n(q, t = 0) + T (q, t = 0). (16B.4.13)
cv @n T cv @n T
The initial conditions term on the right-hand side of equation (16B.4.13) involves
both10 n(q, t = 0) and T (q, t = 0). There are no instantaneous correlations between
the density fluctuations and the temperature. To calculate the spectrum of density
fluctuations, it is thus not necessary to include the term in T (q, 0) in the expression
for n(q, z). Otherwise stated, we can make T (q, 0) = 0. Equation (16B.4.13) then
reduces to:
T @s T @s
( iz + aq 2 )T (q, z) iz n(q, z) = n(q, 0). (16B.4.14)
cv @n T cv @n T
10
When no misinterpretation is possible, these quantities will be respectively denoted by n(q, 0)
and T (q, 0).
Longitudinal fluctuations 475
T @s @P 1
= c2s , (16B.4.17)
mcv @n T @T n
established in footnote below,11 we deduce the following expression for the determinant
(16B.4.16):
⇣ c2s q 2 ⌘ 1
Det = i (z + iaq 2 ) z 2 + izDL q 2 zc2s q 2 · (16B.4.18)
The solution of the linear system (16B.4.15) is, as for the particle density, represented
here by n(q, z):
1
(z + iaq 2 )(z + iDL q 2 ) c2s q 2
n(q, z) = in(q, t = 0) ⇣ · (16B.4.19)
c2s q 2 ⌘ 1
(z + iaq 2 ) z 2 + izDL q 2 zc2s q 2
11
Coming back to the expressions = cp /cv for the ratio of specific heats and c2s = ( /m)(@P/@n)T
for the squared sound velocity, we write the thermodynamic relation (16B.4.17) in the form:
(@s/@n)T (@P/@T )n
cp cv = T ,
(@P/@n)T
(@s/@V )T (@P/@T )V
cp cv = T ·
(@P/@V )T
Given the Maxwell relation (@S/@V )T = (@P/@T )V , we have, in order to establish identity (16B.4.17),
to demonstrate the formula:
T [(@P/@T )V ]2
cp c v = ·
N (@P/@V )T
This may be rewritten as:
T V ↵2
N (cp cv ) = ,
T
where ↵ = (1/V )(@V /@T )P is the dilatation coefficient at constant pressure of the fluid and T =
(1/V )(@V /@P )T its isothermal compressibility. We recognize in the above formula a standard ther-
modynamic identity.
476 Light scattering by a fluid
1
(z + iaq 2 )(z 2 + izDL q 2 c2s q 2 ) = ( iaq 2 )c2s q 2 · (16B.4.21)
(0) 1
z0 = z0 · (16B.4.26)
1 ia
z0 = iaq 2 + iaq 2 = q2 = iDth q 2 , (16B.4.27)
where we have introduced the thermal di↵usion coefficient Dth = a/ = /ncp (cp is
the specific heat at constant pressure per particle). A similar calculation can be carried
12
This is realized in a liquid, in which case ' 1.
Longitudinal fluctuations 477
(0)
out for the sound poles z1,2 , which we look for in the form z1,2 + z1,2 . In the limit
of small wave vectors, z1,2 is of order q 2 :
i 2 1
z1,2 = aq · (16B.4.28)
2
i i 2 1
z1,2 = ±cs q DL q 2 aq = ±cs q i q2 , (16B.4.29)
2 2
where:
DL a 1
= + (16B.4.30)
2 2
is the sound attenuation coefficient. All three poles z0 and z1,2 lie in the lower complex
half-plane.
Let us now come back to formula (16B.4.19). At the order at which the poles have
been calculated, we can write the denominator in the form of the following product:
Decomposing n(q, z)/n(q, t = 0) into partial fractions and expressing the residues at
lowest order in q gives:
✓ ◆
n(q, z) 1 i i 1 1
= + + · (16B.4.32)
n(q, t = 0) z + iDth q 2 2 z cs q + i q 2 z + cs q + i q 2
n(q, t) 1 Dth q 2 t 1 q2 t
= e + e cos cs qt, t > 0. (16B.4.33)
n(q, 0)
This gives:
1
IR = 2⇡S(q) · (16B.5.9)
1
IB = 2⇡S(q) · (16B.5.10)
2
We therefore have:
IR + 2IB = 2⇡S(q). (16B.5.11)
The ratio:
IR
= 1 (16B.5.12)
2IB
Bibliography
B.J. Berne and R. Pecora, Dynamic light scattering. With applications to chem-
istry, biology, and physics, Wiley, New York, 1976. Reprinted, Dover Publications,
New York, 2000.
P.M. Chaikin and T.C. Lubensky, Principles of condensed matter physics, Cam-
bridge University Press, Cambridge, 1995.
D. Forster, Hydrodynamic fluctuations, broken symmetries, and correlation func-
tions, Westview Press, Boulder, 1995.
J.-P. Hansen and I.R. McDonald, Theory of simple liquids, Academic Press, Lon-
don, third edition, 2006.
L.D. Landau and E.M. Lifshitz, Statistical physics, Butterworth-Heinemann, Ox-
ford, third edition, 1980.
References
L.P. Kadanoff and P.C. Martin, Hydrodynamic equations and correlation func-
tions, Ann. Phys. 24, 419 (1963).
R.D. Mountain, Spectral distribution of scattered light in a simple fluid, Rev. Mod.
Phys. 38, 205 (1966).
Index