An Introduction To Geometrical Physics (R. Aldrovandi, J. G. Pereira)
An Introduction To Geometrical Physics (R. Aldrovandi, J. G. Pereira)
An Introduction To Geometrical Physics (R. Aldrovandi, J. G. Pereira)
GEOMETRICAL PHYSICS
R. Aldrovandi & J.G. Pereira
Instituto de Fsica Teorica
State University of Sao Paulo UNESP
Sao Paulo Brazil
To our parents
Nice, Dina, Jose and Tito
ii
The simplest geometrical setting used consciously or not by physicists in their everyday work is the 3-dimensional euclidean space E3 . It consists of the set R3 of ordered triples of real numbers such as p = (p1 , p2 , p3 ), q
= (q 1 , q 2 , q 3 ), etc, and is endowed with a very special characteristic, a metric
defined by the distance function
" 3
#1/2
X
d(p, q) =
(pi q i )2
.
i=1
It is the space of ordinary human experience and the starting point of our
geometric intuition. Studied for two-and-a-half millenia, it has been the
object of celebrated controversies, the most famous concerning the minimum
number of properties necessary to define it completely.
From Aristotle to Newton, through Galileo and Descartes, the very word
space has been reserved to E3 . Only in the 19-th century has it become clear
that other, different spaces could be thought of, and mathematicians have
since greatly amused themselves by inventing all kinds of them. For physicists, the age-long debate shifted to another question: how can we recognize,
amongst such innumerable possible spaces, that real space chosen by Nature
as the stage-set of its processes? For example, suppose the space of our everyday experience consists of the same set R3 of triples above, but with a
different distance function, such as
d(p, q) =
3
X
|pi q i |.
i=1
iv
any other space given by any distance function with R3 as set point. It so
happens, however, that Nature has chosen the former and not the latter space
for us to live in. To know which one is the real space is not a simple question
of principle something else is needed. What else? The answer may seem
rather trivial in the case of our home space, though less so in other spaces
singled out by Nature in the many different situations which are objects of
physical study. It was given by Riemann in his famous Inaugural Address1 :
... those properties which distinguish Space from other conceivable triply extended quantities can only be deduced from experience.
Thus, from experience! It is experiment which tells us in which space we
actually live in. When we measure distances we find them to be independent
of the direction of the straight lines joining the points. And this isotropy
property rules out the second proposed distance function, while admitting
the metric of the euclidean space.
In reality, Riemanns statement implies an epistemological limitation: it
will never be possible to ascertain exactly which space is the real one. Other
isotropic distance functions are, in principle, admissible and more experiments are necessary to decide between them. In Riemanns time already
other geometries were known (those found by Lobachevsky and Boliyai) that
could be as similar to the euclidean geometry as we might wish in the restricted regions experience is confined to. In honesty, all we can say is that
E3 , as a model for our ambient space, is strongly favored by present day
experimental evidence in scales ranging from (say) human dimensions down
to about 1015 cm. Our knowledge on smaller scales is limited by our capacity to probe them. For larger scales, according to General Relativity, the
validity of this model depends on the presence and strength of gravitational
fields: E3 is good only as long as gravitational fields are very weak.
These data are like all data not logically necessary,
but only of empirical certainty . . . one can therefore investigate
their likelihood, which is certainly very great within the bounds of
observation, and afterwards decide upon the legitimacy of extending them beyond the bounds of observation, both in the direction of
the immeasurably large and in the direction of the immeasurably
small.
1
A translation of Riemanns Address can be found in Spivak 1970, vol. II. Cliffords
translation (Nature, 8 (1873), 14-17, 36-37), as well as the original transcribed by David
R. Wilkins, can be found in the site http://www.emis.de/classics/Riemann/.
v
The only remark we could add to these words, pronounced in 1854, is
that the bounds of observation have greatly receded with respect to the
values of Riemann times.
. . . geometry presupposes the concept of space, as well as
assuming the basic principles for constructions in space .
In our ambient space, we use in reality a lot more of structure than
the simple metric model: we take for granted a vector space structure, or
an affine structure; we transport vectors in such a way that they remain
parallel to themselves, thereby assuming a connection. Which one is the
minimum structure, the irreducible set of assumptions really necessary to
the introduction of each concept? Physics should endeavour to establish on
empirical data not only the basic space to be chosen but also the structures
to be added to it. At present, we know for example that an electron moving
in E3 under the influence of a magnetic field feels an extra connection (the
electromagnetic potential), to which neutral particles may be insensitive.
Experimental science keeps a very special relationship with Mathematics. Experience counts and measures. But Science requires that the results
be inserted in some logically ordered picture. Mathematics is expected to
provide the notion of number, so as to make countings and measurements
meaningful. But Mathematics is also expected to provide notions of a more
qualitative character, to allow for the modeling of Nature. Thus, concerning
numbers, there seems to be no result comforting the widespread prejudice
by which we measure real numbers. We work with integers, or with rational
numbers, which is fundamentally the same. No direct measurement will sort
out a Dedekind cut. We must suppose, however, that real numbers exist:
even from the strict experimental point of view, it does not matter whether
objects like or e are simple names or are endowed with some kind of an
sich reality: we cannot afford to do science without them. This is to say that
even pure experience needs more than its direct results, presupposes a wider
background for the insertion of such results. Real numbers are a minimum
background. Experience, and logical necessity, will say whether they are
sufficient.
From the most ancient extant treatise going under the name of Physics2 :
When the objects of investigation, in any subject, have first
principles, foundational conditions, or basic constituents, it is
through acquaintance with these that knowledge, scientific knowledge, is attained. For we cannot say that we know an object before
2
vi
we are acquainted with its conditions or principles, and have carried our analysis as far as its most elementary constituents.
The natural way of attaining such a knowledge is to start
from the things which are more knowable and obvious to us and
proceed towards those which are clearer and more knowable by
themselves . . .
Euclidean spaces have been the starting spaces from which the basic geometrical and analytical concepts have been isolated by successive, tentative,
progressive abstractions. It has been a long and hard process to remove the
unessential from each notion. Most of all, as will be repeatedly emphasized,
it was a hard thing to put the idea of metric in its due position.
Structure is thus to be added step by step, under the control of experiment. Only once experiment has established the basic ground will internal
coherence, or logical necessity, impose its own conditions.
Contents
I
MANIFOLDS
1 GENERAL TOPOLOGY
1.0 INTRODUCTORY COMMENTS
1.1 TOPOLOGICAL SPACES . . .
1.2 KINDS OF TEXTURE . . . . .
1.3 FUNCTIONS . . . . . . . . . . .
1.4 QUOTIENTS AND GROUPS . .
1.4.1 Quotient spaces . . . . .
1.4.2 Topological groups . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
2 HOMOLOGY
2.1 GRAPHS . . . . . . . . . . . . . . . . . . .
2.1.1 Graphs, first way . . . . . . . . . .
2.1.2 Graphs, second way . . . . . . . . .
2.2 THE FIRST TOPOLOGICAL INVARIANTS
2.2.1 Simplexes, complexes & all that . .
2.2.2 Topological numbers . . . . . . . .
3 HOMOTOPY
3.0 GENERAL HOMOTOPY . . . . .
3.1 PATH HOMOTOPY . . . . . . . .
3.1.1 Homotopy of curves . . . .
3.1.2 The Fundamental group .
3.1.3 Some Calculations . . . .
3.2 COVERING SPACES . . . . . .
3.2.1 Multiply-connected Spaces
3.2.2 Covering Spaces . . . . . .
3.3 HIGHER HOMOTOPY . . . . .
vii
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
3
3
5
15
27
36
36
41
.
.
.
.
.
.
49
50
50
52
57
57
64
.
.
.
.
.
.
.
.
.
73
73
78
78
85
92
98
98
105
115
viii
CONTENTS
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
121
. 121
. 121
. 123
. 125
5 DIFFERENTIABLE MANIFOLDS
133
5.1 DEFINITION AND OVERLOOK . . . . . . . . . . . . . . . . . 133
5.2 SMOOTH FUNCTIONS . . . . . . . . . . . . . . . . . . . . . . 135
5.3 DIFFERENTIABLE SUBMANIFOLDS . . . . . . . . . . . . . . 137
II
DIFFERENTIABLE STRUCTURE
6 TANGENT STRUCTURE
6.1 INTRODUCTION . . . . . . . . . . . . .
6.2 TANGENT SPACES . . . . . . . . . . . .
6.3 TENSORS ON MANIFOLDS . . . . . . .
6.4 FIELDS & TRANSFORMATIONS . . . .
6.4.1 Fields . . . . . . . . . . . . . . .
6.4.2 Transformations . . . . . . . . . .
6.5 FRAMES . . . . . . . . . . . . . . . . . .
6.6 METRIC & RIEMANNIAN MANIFOLDS
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
141
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
7 DIFFERENTIAL FORMS
7.1 INTRODUCTION . . . . . . . . . . . . . . . . . . .
7.2 EXTERIOR DERIVATIVE . . . . . . . . . . . . . .
7.3 VECTOR-VALUED FORMS . . . . . . . . . . . . .
7.4 DUALITY AND CODERIVATION . . . . . . . . . .
7.5 INTEGRATION AND HOMOLOGY . . . . . . . . .
7.5.1 Integration . . . . . . . . . . . . . . . . . . .
7.5.2 Cohomology of differential forms . . . . . . .
7.6 ALGEBRAS, ENDOMORPHISMS AND DERIVATIVES
8 SYMMETRIES
8.1 LIE GROUPS . . . . . . . . . . . . . . .
8.2 TRANSFORMATIONS ON MANIFOLDS .
8.3 LIE ALGEBRA OF A LIE GROUP . . .
8.4 THE ADJOINT REPRESENTATION .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
143
. 143
. 145
. 154
. 161
. 161
. 167
. 175
. 180
.
.
.
.
.
.
.
.
189
. 189
. 197
. 210
. 217
. 225
. 225
. 232
. 239
.
.
.
.
247
. 247
. 252
. 259
. 265
CONTENTS
ix
9 FIBER BUNDLES
9.1 INTRODUCTION . . . . . . . . . .
9.2 VECTOR BUNDLES . . . . . . . . .
9.3 THE BUNDLE OF LINEAR FRAMES
9.4 LINEAR CONNECTIONS . . . . . .
9.5 PRINCIPAL BUNDLES . . . . . . .
9.6 GENERAL CONNECTIONS . . . .
9.7 BUNDLE CLASSIFICATION . . . .
III
273
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
FINAL TOUCH
.
.
.
.
.
.
.
273
275
277
284
297
303
316
321
10 NONCOMMUTATIVE GEOMETRY
323
10.1 QUANTUM GROUPS A PEDESTRIAN OUTLINE . . . . . . 323
10.2 QUANTUM GEOMETRY . . . . . . . . . . . . . . . . . . . . 326
IV
MATHEMATICAL TOPICS
331
333
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
334
338
341
344
348
379
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
379
380
380
380
382
382
CONTENTS
4.7
5 BANACH ALGEBRAS
5.1 Quantization . . . . . . .
5.2 Banach algebras . . . . . .
5.3 *-algebras and C*-algebras
5.4 From Geometry to Algebra
5.5 Von Neumann algebras . .
5.6 The Jones polynomials . .
385
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
385
387
389
390
393
397
6 REPRESENTATIONS
403
6.1 A Linear representations . . . . . . . . . . . . . . . . . . . . . 404
6.2 B Regular representation . . . . . . . . . . . . . . . . . . . . . 408
6.3 C Fourier expansions . . . . . . . . . . . . . . . . . . . . . . . 409
7 VARIATIONS & FUNCTIONALS
7.1 A Curves . . . . . . . . . . . . . . . . . .
7.1.1 Variation of a curve . . . . . . . . .
7.1.2 Variation fields . . . . . . . . . . .
7.1.3 Path functionals . . . . . . . . . . .
7.1.4 Functional differentials . . . . . . .
7.1.5 Second-variation . . . . . . . . . .
7.2 B General functionals . . . . . . . . . . . .
7.2.1 Functionals . . . . . . . . . . . . .
7.2.2 Linear functionals . . . . . . . . .
7.2.3 Operators . . . . . . . . . . . . . .
7.2.4 Derivatives Frechet and Gateaux
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
8 FUNCTIONAL FORMS
8.1 A Exterior variational calculus . . . . . . . . . .
8.1.1 Lagrangian density . . . . . . . . . . .
8.1.2 Variations and differentials . . . . . . .
8.1.3 The action functional . . . . . . . . .
8.1.4 Variational derivative . . . . . . . . . .
8.1.5 Euler Forms . . . . . . . . . . . . . . .
8.1.6 Higher order Forms . . . . . . . . . .
8.1.7 Relation to operators . . . . . . . . .
8.2 B Existence of a lagrangian . . . . . . . . . . .
8.2.1 Inverse problem of variational calculus
8.2.2 Helmholtz-Vainberg theorem . . . . . .
8.2.3 Equations with no lagrangian . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
415
. 415
. 415
. 416
. 417
. 418
. 420
. 421
. 421
. 422
. 423
. 423
425
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
426
426
427
428
428
429
429
429
430
430
430
431
CONTENTS
8.3
xi
. . . . .
The homotopy formula .
Examples . . . . . . . .
Symmetries of equations
C Building lagrangians
8.3.1
8.3.2
8.3.3
.
.
.
.
.
.
.
.
9 SINGULAR POINTS
9.1 Index of a curve . . . . . . . . . . .
9.2 Index of a singular point . . . . . .
9.3 Relation to topology . . . . . . . .
9.4 Basic two-dimensional singularities
9.5 Critical points . . . . . . . . . . . .
9.6 Morse lemma . . . . . . . . . . . .
9.7 Morse indices and topology . . . .
9.8 Catastrophes . . . . . . . . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
432
432
434
436
439
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
439
442
443
443
444
445
446
447
449
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
450
450
450
450
451
451
452
452
452
453
453
455
455
457
457
458
459
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
459
460
461
464
466
467
xii
CONTENTS
12 GEODESICS
12.1 Selfparallel curves . . . . . . . . . . .
12.1.1 In General Relativity . . . . . .
12.1.2 The absolute derivative . . . . .
12.1.3 Selfparallelism . . . . . . . . .
12.1.4 Complete spaces . . . . . . . .
12.1.5 Fermi transport . . . . . . . . .
12.1.6 In Optics . . . . . . . . . . . .
12.2 Congruences . . . . . . . . . . . . . .
12.2.1 Jacobi equation . . . . . . . . .
12.2.2 Vorticity, shear and expansion .
12.2.3 LandauRaychaudhury equation
471
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
472
472
473
474
475
475
476
476
476
480
483
PHYSICAL TOPICS
485
1 HAMILTONIAN MECHANICS
1.1 Introduction . . . . . . . . . . .
1.2 Symplectic structure . . . . . .
1.3 Time evolution . . . . . . . . .
1.4 Canonical transformations . . .
1.5 Phase spaces as bundles . . . .
1.6 The algebraic structure . . . . .
1.7 Relations between Lie algebras .
1.8 Liouville integrability . . . . . .
487
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
MORE MECHANICS
2.1
HamiltonJacobi . . . . . . . . . . . . . . . . . . . . . .
2.1.1 Hamiltonian structure . . . . . . . . . . . . . . .
2.1.2 Hamilton-Jacobi equation . . . . . . . . . . . . .
2.2 The Lagrange derivative . . . . . . . . . . . . . . . . . .
2.2.1 The Lagrange derivative as a covariant derivative
2.3 The rigid body . . . . . . . . . . . . . . . . . . . . . . .
2.3.1 Frames . . . . . . . . . . . . . . . . . . . . . . . .
2.3.2 The configuration space . . . . . . . . . . . . . .
2.3.3 The phase space . . . . . . . . . . . . . . . . . . .
2.3.4 Dynamics . . . . . . . . . . . . . . . . . . . . . .
2.3.5 The space and the body derivatives . . . . .
2.3.6 The reduced phase space . . . . . . . . . . . . . .
2.3.7 Moving frames . . . . . . . . . . . . . . . . . . .
2.3.8 The rotation group . . . . . . . . . . . . . . . . .
487
488
490
491
494
496
498
501
503
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
503
503
505
507
507
510
510
511
511
512
513
513
514
515
CONTENTS
xiii
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
521
. 521
. 521
. 522
. 526
. 526
. 529
. 531
. 535
. 536
. 537
. 537
. 542
. 547
. 550
.
.
.
.
553
. 553
. 554
. 558
. 561
.
.
.
.
.
.
.
565
. 565
. 566
. 567
. 568
. 570
. 571
. 572
.
.
.
.
575
. 575
. 576
. 579
. 579
xiv
CONTENTS
7 GAUGE FIELDS
7.1 A The gauge tenets . . . . . . . . . . .
7.1.1 Electromagnetism . . . . . . . .
7.1.2 Nonabelian theories . . . . . . .
7.1.3 The gauge prescription . . . . .
7.1.4 Hamiltonian approach . . . . .
7.1.5 Exterior differential formulation
7.2 B Functional differential approach . . .
7.2.1 Functional Forms . . . . . . . .
7.2.2 The space of gauge potentials .
7.2.3 Gauge conditions . . . . . . . .
7.2.4 Gauge anomalies . . . . . . . .
7.2.5 BRST symmetry . . . . . . . .
7.3 C Chiral fields . . . . . . . . . . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
589
. 590
. 590
. 591
. 593
. 594
. 595
. 596
. 596
. 598
. 601
. 602
. 603
. 603
8 GENERAL RELATIVITY
605
8.1 Einsteins equation . . . . . . . . . . . . . . . . . . . . . . . . 605
8.2 The equivalence principle . . . . . . . . . . . . . . . . . . . . . 608
8.3 Spinors and torsion . . . . . . . . . . . . . . . . . . . . . . . . 612
9 DE
9.1
9.2
9.3
9.4
9.5
SITTER SPACES
General characteristics . . . . .
Curvature . . . . . . . . . . . .
Geodesics and Jacobi equations
Some qualitative aspects . . . .
Wigner-Inon
u contraction . . .
.
.
.
.
.
.
.
.
.
.
VI
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
615
. 615
. 619
. 620
. 621
. 621
625
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
625
628
629
630
631
635
Part I
MANIFOLDS
Chapter 1
GENERAL TOPOLOGY
Or, the purely qualitative properties of spaces.
1.0
INTRODUCTORY COMMENTS
1.1
TOPOLOGICAL SPACES
context provides a general and abstract concept of open sets and gives a clear
meaning to the above rather elusive word neighbourhood, while providing
the general background against which the fundamental notions of continuity
and convergence acquire well defined contours.
1.1.2 A space will be, to begin with, a set endowed with some decomposition allowing us to talk about its parts. Although the elements belonging
to a space may be vectors, matrices, functions, other sets, etc, they will be
called, to simplify the language, points. Thus, a space will be a set S of
points plus a structure leading to some kind of organization, such that we
may speak of its relative parts and introduce spatial relationships. This
structure is introduced as a well-performed division of S, as a convenient family of subsets. There are various ways of dividing a set, each one designed to
accomplish a definite objective.
We shall be interested in getting appropriate notions of neighbourhood,
distinguishability of points, continuity and, later, differentiability. How is a
fitting decomposition obtained? A first possibility might be to consider S
with all its subsets. This conception, though acceptable in principle, is too
particular: it leads to a quite disconnected space, every two points belonging
to too many unshared neighbourhoods. It turns out (see section 1.3) that
any function would be continuous on such a pulverized space and in consequence the notion of continuity would be void. The family of subsets is
too large, the decomposition would be too fine-grained. In the extreme
opposite, if we consider only the improper subsets, that is, the whole point
set S and the empty set , there would be no real decomposition and again
no useful definition of continuity (subsets distinct from and S are called
proper subsets). Between the two extreme choices of taking a family with
all the subsets or a family with no subsets at all, a compromise has been
found: good families are defined as those respecting a few well chosen, suitable conditions. Each one of such well-bred families of subsets is called a
topology.
Given a point set S, a topology is a family of subsets of S (which are
called, by definition, its open sets) respecting the 3 following conditions:
(a) the whole set S and the empty set belong to the family;
(b) given a finite number
T of members of the family, say U1 , U2 , U3 , . . . , Un ,
their intersection ni=1 Ui is also a member;
(c) given any number (finite or infinite) of open sets, their union belongs to
the family.
The topology is formed by the set of the open balls. It is a standard practice
to designate a topological space by its point set when there is no doubt as
to which topology is meant. That is why the euclidean space is frequently
denoted simply by Rn . We shall, however, insist on the notational difference: En will be Rn plus the ball topology. En is the basic, starting space,
as even differential manifolds will be presently defined so as to generalize it.
We shall see later that the introduction of coordinates on a general space S
requires that S resemble some En around each one of its points. It is important to notice, however, that many of the most remarkable properties of the
euclidean space come from its being, besides a topological space, something
else. Indeed, one must be careful to distinguish properties of purely topological nature from those coming from additional structures usually attributed
to En , the main one being that of a vector space.
1.1.7 In metric spaces, any point p has a countable set of open neighbourhoods {Ni } such that for any set U containing p there exists at least one Nj
included in U . Thus, any set U containing p is a neighbourhood. This is not
a general property of topological spaces. Those for which this happens are
said to be first-countable spaces (Figure 1.1).
3
Some authors (Kolmogorov & Fomin 1977, for example) do define a neighbourhood
of p as an open set of T to which p belongs. In our language, a neighbourhood which is
also an open set of T will be an open neighbourhood.
Figure 1.1: In first-countable spaces, every point p has a countable set of open neighbourhoods {Nk }, of which at least one is included in a given U 3 p. We say that all
points have a local countable basis. All metric spaces are of this kind.
10
B and B 0 are basis defining the same topology iff, for every U B and
every p U , there exists some U0 B 0 such that p B0 U and
vice-versa.
Again, it is instructive to give some thought to disks and rectangles in E2 . A
basis for the real euclidean line E1 is provided by all the open intervals of the
type (r 1/n, r + 1/n), where r runs over the set of rational numbers and n
over the set of the integer numbers. This is an example of countable basis.
When a topology has at least one countable basis, it is said to be secondcountable. Second countable topologies are always first-countable ( 7) but
the inverse is not true. We have said above that all metric spaces are firstcountable. There are, however, metric spaces which are not second countable
(Figure 1.2).
Figure 1.2: A partial hierarchy: not all metric spaces are second-countable, but all of
them are first-countable.
11
Figure 1.3: The sphere S 2 with some of its open sets, which are defined as the intersections of S 2 with the open balls of the euclidean 3-dimensional space.
(1.1)
12
all the open balls entirely included in Rn+ ; (ii) for each ball tangent to the
hyperplane pn = 0, the union of that ball with (the set containing only) the
tangency point.
1.1.11 Notice that, for the 2-dimensional case (the upper-half plane, Figure
1.4) for example, sets of type ,
including intersections with the horizontal line,
2
are not open in E but are open in E2+ . One speaks of the above topology as
the swimmers topology: suppose a fluid flows upwardly from the horizontal
borderline into the space with a monotonously decreasing velocity which is unit at
the bottom. A swimmer with a constant unit velocity may start swimming in any
direction at any point of the fluid. In a unit interval of time the set of all possible
swimmers will span a basis.
Figure 1.4: The upper-half plane E2+ , whose open sets are the intersections of the point
1.1.12 A cautionary remark: the definitions given above (and below) may
sometimes appear rather queer and irksome, as if invented by some skewminded daemon decided to hide simple things under tangled clothes. They
have evolved, however, by a series of careful abstractions, starting from the
properties of metric spaces and painstakingly checked to see whether they
lead to useful, meaningful concepts. Fundamental definitions are, in this
sense, the products of Experimental Mathematics. If a simpler, more direct
definition seems possible, the reader may be sure that it has been invalidated
by some counter-example (see as an example the definition of a continuous
function in section 1.3.4).
13
1.1.13 Consider two topologies T1 and T2 defined on the same point set
S. We say that T1 is weaker than T2 if every member of T1 belongs also to
T2 . The topology T1 is also said to be coarser than T2 , and T2 is finer than
T1 (or T2 is a refinement of T1 , or still T2 is stronger than T1 ). The topology
T for the finite space of 1.1.3 is clearly weaker than the discrete topology
for the same point set.
1.1.14 We have said that the topology for Minkowski space time cannot
be obtained from the Lorentz metric, which is unable to define balls. The
specification of a topology is of fundamental importance because (as will
be seen later) it is presupposed every time we talk of a continuous (say,
wave) function. We could think of using an E4 topology, but this would
be wrong because (besides other reasons) no separation would then exist
between spacelike and timelike vectors. The fact is that we do not know the
real topology of spacetime. We would like to retain euclidean properties both
in the space sector and on the time axis. Zeeman4 has proposed an appealing
topology: it is defined as the finest topology defined on R4 which induces an
E3 topology on the space sector and an E1 topology on the time axis. It
is not first-countable and, consequently, cannot come from any metric. In
their everyday practice, physicists adopt an ambiguous behaviour and use
the balls of E4 whenever they talk of continuity and/or convergence.
1.1.15 Given the subset C of S, its complement is the set C 0 = {p S
such that p
/ C}. The subset C is a closed set in the topological space
(S, T ) if C 0 is an open set of T . Thus, the complement of an open set is (by
definition) closed. It follows that and S are closed (and open!) sets in all
topological spaces.
1.1.16 Closedness is a relative concept: a subset C of a topological subspace Y of S can be closed in the induced topology even if open in S; for
instance, Y itself will be closed (and open) in the induced topology, even if
Y is an open set of S.
Retain that closed, just as open, depends on the chosen topology. A set
which is open in a topology may be closed in another.
1.1.17 A connected space is a topological space in which no proper
subset is simultaneously open and closed.
In this case S cannot be decomposed into the union of two disjoint open sets.
One should not confuse this concept with path-connectedness, to be defined
4
Zeeman 1964, 1967; later references may be traced from Fullwood 1992.
14
later ( 1.3.15) and which, intuitively, means that one can walk continuously
between any two points of the space on a path entirely contained in it. Pathconnectedness implies connectedness, but not vice-versa. Clearly the line E1
is connected, but the line-minus-zero space E1 {0} (another notation:
E1 \{0}) is not. The finite space of 1.1.3 is connected.
1.1.18 The discrete topology : set S and all its subsets are taken as
open sets. The set of all subsets of a set S is called its power set, denoted
P (S), so that we are taking the topological space (S, P (S)). This does yield
a topological space. For each point p, {p} is open. All open sets are also
closed and so we have extreme unconnectedness. Lest the reader think this
example to be physically irrelevant, we remark that the topology induced on
the light cone by Zeemans topology for spacetime ( 1.1.14) is precisely of
this type. Time is usually supposed to be a parameter running in E1 and
a trajectory on some space S is a mapping attributing a point of S to each
instant in E1 . It will be seen later (section 1.3) that no function from
E1 to a discrete space may be continuous. A denizen of the light cone, like
the photon, would not travel continuously through spacetime but bound
from point to point. The discrete topology is, of course, the finest possible
topology on any space. Curiously enough, it can be obtained from a metric,
the so-called discrete metric: d(p, q) = 1 if p 6= q, and d(p, q) = 0 if p = q.
The indiscrete (or trivial ) topology is T = {, S}. It is the weakest possible
topology on any space andbeing not first-countablethe simplest example
of topology which cannot be given by a metric. By the way, this is an
illustration of the complete independence of topology from metrics: a nonmetric topology may have finer topologies which are metric, and a metric
topology can have finer non-metric topologies. And a non-metric topology
may have weaker topologies which are metric, and a metric topology can
have weaker non-metric topologies.
1.1.19 Topological product Given two topological spaces A and B, their
topological product (or cartesian product) AB is the set of pairs (p, q) with
p A and q B, and a topology for which a basis is given by all the pairs of
type U V , U being a member of a basis in A and V a member of a basis in
B. Thus, the cartesian product of two topological spaces is their cartesian set
product ( Math.1.11) endowed with a product topology. The usual torus
imbedded in E3 , denoted T2 , is the cartesian product of two 1-dimensional
spheres (or circles) S 1 . The n-torus Tn is the product of S 1 by itself n times.
1.1.20 We have clearly separated topology from metric and found examples of non-metric topologies, but it remains true that a metric does define
15
a topology. A reverse point of view comes from asking the following question: are all the conditions imposed in the definition of a distance function
necessary to lead to a topology? The answer is no. Much less is needed. A
prametric suffices. On a set S, a prametric is a mapping
: S S R+ such that (p, p) = 0 for all p S.
1.1.21 The consideration of spatial relationships requires a particular way
of dividing a space into parts. We have chosen, amongst all the subsets of S,
particular families satisfying well chosen conditions to define topologies. A
family of subsets of S is a topology if it includes S itself, the empty set , all
unions of subsets and all intersections of a finite number of them. A topology
is that simplest, minimal structure allowing for precise non-trivial notions of
convergence and continuity. Other kinds of families of subsets are necessary
for other purposes. For instance, the detailed study of convergence in a nonmetric topological space S requires cuttings of S not including the empty
set, called filters. And, in order to introduce measures and define integration
on S, still another kind of decomposition is essential: a -algebra. In other
to make topology and integration compatible, a particular -algebra must be
defined on S, the Borel -algebra. A sketchy presentation of these questions
is given in Mathematical Topic 3.
1.2
KINDS OF TEXTURE
We have seen that, once a topology is provided, the set point acquires a kind
of elementary texture, which can be very tight (as in the indiscrete topology),
very loose (as in the discrete topology), or intermediate. We shall see now
that there are actually optimum decompositions of spaces. The best behaved
spaces have not too many open sets: they are compact. Nor too few: they
are of Hausdorff type.
There are many ways of probing the topological makeup of a space. We
shall later examine two external approaches: one of them (homology) tries
to decompose the space into its building bricks by relating it to the decomposition of euclidean space into triangles, tetrahedra and the like. The other
(homotopy) examines loops (of 1 or more dimensions) in the space and their
continuous deformations. Both methods use relationships with other spaces
and have the advantage of providing numbers (topological numbers) to
characterize the space topology.
For the time being, we shall study two internal ways of probing a space
(S, T ). One considers subsets of S, the other subsets of T. The first considers
16
samples of isolated points, or sequences, and gives a complete characterization of the topology. The second consists of testing the texture by rarefying
the family of subsets and trying to cover the space with a smaller number of
them. It reveals important qualitative traits. We shall start by introducing
some concepts which will presently come in handy.
1.2.1 Consider a space (S, T ). Given an arbitrary set U S, not necessarily belonging to T , in general there will be some closed sets C containing
U . The intersection C of all closed sets containing U is the closure of
U , denoted U . An equivalent, more intuitive definition is U = {p such that
every neighbourhood of p has nonvanishing intersection with U }. The bestknown example is that of an open interval (a, b) in E1 , whose closure is the
closed interval [a, b].
The closure of a closed set V is V itself, and V being the closure of itself implies
that V is closed.
Given an arbitrary set W S, not necessarily belonging to T , its interior ,
denoted int W or W 0 , is the largest open subset of W . Given all the open sets
O contained in W , then
W 0 = O .
W 0 is the set of points of S for which W is an open neighbourhood. The boundary
b(U ) of a set U is the complement of its interior in its closure,
U0 = U
\U 0 .
b(U ) = U
\b(U ). If U is an open set of T , then U 0 = U and
It is also true that U 0 = U
\U . If U is a closed set, then U
= U and b(U ) = U \U 0 . These definitions
b(U ) = U
correspond to the intuitive view of the boundary as the skin of the set. From
this point of view, a closed set includes its own skin. The sphere S 2 , imbedded in
E3 , is its own interior and closure and consequently has no boundary. A set has
empty boundary when it is both open and closed. This allows a rephrasing of the
definition of connectedness: a space S is connected if, except for and S itself, it
has no subset whose boundary is empty.
Let again S be a topological space and U a subset. A point p U is an isolated point
of U if it has a neighbourhood that contains no other point of U . A point p of S is a
limit point of U if each neighbourhood of p contains at least one point of U distinct of
p. The set of all the limit points of U is called the derived set of U , written D(U ). A
= U D(U ): we may obtain the closure of a set by adding to it all
theorem says that U
its limiting points. U is closed iff it already contains them all, U D(U ). When every
neighbourhood of p contains infinite points of U , p is an accumulation point of U (when
such infinite points are not countable, p is a condensation point ). Though we shall not be
using all these notions in what follows, they appear frequently in the literature and give a
17
taste of the wealth and complexity of the theory coming from the three simple axioms of
1.1.2.
En .
n=0
5
18
Recall that we can define real numbers as the limit points of sequences of
rational numbers. This is possible because the subset of rational numbers Q is
everywhere dense in the set R of the real numbers with the euclidean topology
(which turns R into E1 ). The set Q has derivative D(Q) = R and interior Q0 = .
Its closure is the same as that of its complement, the set J = R\Q of irrational
numbers: it is R itself. As said in 1.1.9, both Q and J are topological subspaces
of R.
19
points, clearly defined on metric spaces, becomes rather loose. All we can
say is that the points of a convergent sequence get progressively closer to its
limit, when this point is unique.
1.2.6 Roughly speaking, linear spaces, or vector spaces, are spaces allowing for addition and rescaling of their members. We leave the definitions
and the more algebraic aspects to Math.1, the details to Math.4, and concentrate in some of their topological possibilities. What imports here is that
a linear space over the set of complex numbers C may have a norm, which
is a distance function and defines consequently a certain topology called the
norm topology. Once endowed with a norm, a vector space V is a metric
topological space. For instance, a norm may come from an inner product, a
mapping from the cartesian set product V V into C,
V V C,
(v, u) < v, u >
(1.2)
< v, v >
will be the norm of v induced by the inner product. This is a special norm,
as norms may be defined independently of inner products. Actually, one
must impose certain compatibility conditions between the topological and
the linear structures (see Math.4).
1.2.7 Hilbert space6 Everybody knows Hilbert spaces from (at least)
Quantum Mechanics courses. They are introduced there as spaces of wavefunctions, on which it is defined a scalar product and a consequent norm.
There are basic wavefunctions, in terms of which any other may be expanded.
This means that the set of functions belonging to the basis is dense in the
whole space. The scalar product is an inner product and defines a topology. In Physics textbooks two kinds of such spaces appear, according to
whether the wavefunctions represent bound states, with a discrete spectrum,
or scattering states. In the first case the basis is formed by a discrete set
of functions, normalized to the Kronecker delta. In the second, the basis is
formed by a continuum set of functions, normalized to the Dirac delta. The
latter are sometimes called Dirac spaces.
Formally, a Hilbert space is an inner product space which is complete
under the inner product norm topology. Again we leave the details to Math.4,
6
Halmos 1957.
20
and only retain here some special characteristics. It was originally introduced
as an infinite space H endowed with a infinite but discrete basis {vi }iN ,
formed by a countably infinite orthogonal family of vectors. This family is
dense in H and makes of H a separable space.
member of the space
P Each
i
can be written in terms of the basis: X = i=1 X vi . The space L2 of all
absolutely square integrable functions on the interval (a, b) R,
Z b
2
L = {f on [a, b] with
|f (x)|2 dx < },
a
21
in our ambient space E3 , such as the sphere, the ellipsoid and the torus:
they are contained in finite portions of E3 , while the plane, the hyperboloid
and the paraboloid are not. It is possible, however, to give an intrinsic
characterization of finite extension, dependent only on the internal properties
of the space itself and not on any knowledge of larger spaces containing it.
We may guess from the above examples that spaces whose extensions are
limited have a lesser number of open sets than those which are not. In
fact, in order to get an intrinsic definition of finite extension, it is necessary
to restrict the number of open sets in a certain way, imposing a limit to the
divisibility of space. And, to arrive at that restriction, the preliminary notion
of covering is necessary.
1.2.10 Suppose a topological space S and a collection C = {U } of open
sets such that S is their union, S = U . The collection C is called an
open covering of S. The interval I has a well known property, which is the
Heine-Borel lemma: with the topology induced by E1 , every covering of I has
a finite subcovering. An analogous property holds in any euclidean space: a
subset is bounded and closed iff any covering has a finite subcovering. The
general definition of compactness is thereby inspired.
1.2.11 Compactness A topological space S is a compact space if each
covering of S contains a finite subcollection of open sets which is also a
covering.
Cases in point are the historical forerunners, the closed balls in euclidean
spaces, the spheres S n and, as expected, all the bounded surfaces in E3 .
Spaces with a finite number of points (as that in 1.1.3) are automatically
compact. In Physics, compactness is usually introduced through coordinates
with ranges in suitably closed or half-closed intervals. It is, nevertheless, a
purely topological concept, quite independent of the very existence of coordinates. As we shall see presently, not every kind of space accepts coordinates.
And most of those which do accept require, in order to be completely described, the use of many distinct coordinate systems. It would not be possible
to characterize the finiteness of a general space by this method.
On a compact space, every sequence contains a convergent subsequence,
a property which is equivalent to the given definition and is sometimes used
instead: in terms of sequences,
a space is compact if, from any sequence of its points,
one may extract a convergent subsequence.
1.2.12 Compact spaces are mathematically simpler to handle than noncompact spaces. Many of the topological characteristics physicists became
22
23
24
E1 , X and Z (Figure 1.5), of which we identify all (and only!) the points
which are strictly negative: pX pZ iff p < 0. The points pX = 0 and pZ = 0
are distinct, pX lying in the region of X not identified with Z and pZ lying
in Z. But they have no disjoint neighbourhoods.
The space has a Y aspect in this case, but not all non-Hausdorff spaces
exhibit such a bifurcation. All Hausdorff spaces are of course necessarily firstseparable, but they go much further, allowing to discern points in an way ideal
for physicists, after the discussion of 1.1.2. Actually, each point is a closed
set. The Hausdorff property is also highly praised by analysts: it ensures
the uniqueness of a converging sequence limit point. And it is fortunate that
most spaces appearing in Physics are Hausdorff, as only on such spaces are the
solutions of differential equations (with fixed initial conditions) assured to be
unique the Hausdorff character is included in the hypotheses necessary to
prove the unicity of solution theorem. On non-Hausdorff spaces, solutions are
only locally unique.9 It would seem that physicists should not worry about
possible violations of so desirable condition, but non-Hausdorff spaces turn
up in some regions of spacetime for certain solutions of Einsteins equations,10
giving rise to causal anomalies.11 Although the Hausdorff character is also
necessary to the distinction of events in spacetime,12 Penrose has speculated
on its possible violation.13
An open set can be the union of disjoint point sets. Take the interval
I = [0, 1]. Choose a basis containing I, and all the sets obtained by omitting from I at most a countable number of points. A perfect though rather
9
Arnold 1973.
Hajicek 1971.
11
Hawking & Ellis 1973.
12
Geroch & Horowitz in Hawking & Israel 1979. An interesting article on the topology
of the Universe.
13
Penrose in Hawking & Israel 1979, mainly in those pages (591-596) dedicated to
psychological time.
10
25
For instance, the book of Kolmogorov & Fomin, 1977, chap.II. A general resume with
many (counter) examples is Steen & Seebach 1970.
15
For an example of Hausdorff but not normal space, see Kolmogorov & Fomin 1970,
p. 86.
26
Figure 1.6: First-separable, second-separable and normal spaces: (left) first separable
every two points have exclusive neighbourhoods; (center) Hausdorff every two points
have disjoint neighbourhoods; (right) normal disjoint closed sets are included in disjoint
open sets.
given a point p, amongst the intervals of type [p, r) with r rational, there
will be some one included in any U containing p. The Sorgenfrey topology
is finer than the usual euclidean line topology, though it remains separable
(by denseness). This favorite pet of topologists is non-metrizable.
1.2.18 Consider a covering {U } of S. It is a locally finite covering if each
point p has a neighbourhood U 3 p such that U intersects only a finite number of the U . A covering {Vi } is a refinement of {U } if, for every Vi , there
is a U such that U Vi . A space is paracompact if it is Hausdorff and all
its coverings have local finite refinements. Notice: finite subcoverings lead to
compactness and finite refinements (plus Hausdorff) to paracompactness. A
connected Hausdorff space is paracompact if it is second-countable. Figure
1.7 is a scheme of the separability hierarchy. Every metric space is paracompact. Every paracompact space is normal. Paracompactness is a condition
of importance for integration, as it is sufficient for attributing a partition of
unity (see Math.3.5) to any locally finite covering. It is also necessary to
the existence of linear connections on the space.16 Paracompact spaces are
consequently essential to General Relativity, in which these connections are
16
1.3. FUNCTIONS
27
1.3
FUNCTIONS
Geroch 1968.
28
1.3. FUNCTIONS
29
Hawking, King & McCarthy 1976; Gobel 1976 a,b; Malament 1977.
30
1.3. FUNCTIONS
31
could in reality define a topology as an equivalence class under homeomorphisms. And the ultimate (perhaps too pretentious?) goal of Topology
as a discipline is the classification of all topological spaces, of course up to
such equivalences. Intuitively, A is homeomorphic to B means that A can
be deformed, without being torn or cut, to look just like B. Under a homeomorphism, images and pre-images of open sets are open, and images and
pre-images of closed sets are closed. A sphere S 2 can be stretched to become
an ellipsoid or an egg-shaped surface or even a tetrahedron. Such surfaces
are indistinguishable from a purely topological point of view. They are the
same topological space. The concept of homeomorphism gives in this way
a precise meaning to those rather loose notions of suave deformations we
have been talking about in 1.0.4. Of course, there is no homeomorphism
between either the sphere, the ellipsoid, or the tetrahedron and (say) a torus
T 2 , which has quite different topological properties.
From the point of view of sequences: a homeomorphism is a mapping
h : A B such that, if {pn } is a sequence in A converging to a point
p, then the sequence {h(pn )} in B converges to h(p); and vice-versa: if
the sequence {qn } in B converges to a point q, then the sequence given by
{h1 (qn )} in A converges to {h1 (q)}.
A condensation is a mapping f : A B which is one-to-one [x 6= y
f (x) 6= f (y)] and onto [f (A) = B]; a homeomorphism is a condensation
whose inverse is also a condensation.
The study of homeomorphisms between spaces can lead to some surprises.
It was found that a complete metric space [ 1.2.5] can be homeomorphic to
an incomplete space. Consequently, the completeness of metric spaces is
not really a topological characteristic. Another result: the space of rational
numbers with the usual topology is homeomorphic to any metric countable
space without isolated points.
The main objective of Zeemans cited papers on spacetime was to obtain a
topology whose automorphic homeomorphisms preserve the causal structure
and constitute a group including the Poincare and/or the conformal group.
1.3.11 Take again the euclidean vector space En . Any isometry will be
a homeomorphism, in particular any translation. Also homothecies with
reason 6= 0 are homeomorphisms. From these two properties it follows
that any two open balls are homeomorphic to each other, and any open ball
is homeomorphic to the whole space. As a hint of the fundamental role which
euclidean spaces will come to play, suppose a space S has some open set U
which is by itself homeomorphic to an open set (a ball) in some En : there is
a homeomorphic mapping f : U ball, f (p U ) = x = (x1 , x2 , . . . , xn ).
Such a homeomorphism is a local homeomorphism. Because the image space
32
The name Frechet space is also used with another, more fundamental and quite
distinct meaning. See, for instance, Sierpi
nski 1956.
1.3. FUNCTIONS
33
34
that subset of Y X whose members are the functions f such that o f (c). Call
(C, O) the collection of such (c, o). A topology on Y X is then generated from
a basis consisting of all the finite intersections of the sets (c, o) (C, O). This
is the compact-open topology, which coincides with the previous one when X is
compact and Y is metric. Other topologies may be defined on function spaces
and their choice is a matter of convenience. A point worth emphasizing is that,
besides other requirements, a topology is presupposed in functional differentiation
and integration. Function spaces are more complicated because their dimension
is (highly) infinite and many of the usual properties of finite spaces do not carry
over to them.20 As stated in 1.2.13, there is one remarkable difference: when the
topology is given by a norm, an infinite-dimensional space is never locally compact.
Differences between finite- and infinite-dimensional spaces are summarized in DeWittMorette, Masheshwari & Nelson 1979, appendix A.
21
Ezawa 1978; idem, 1979.
1.3. FUNCTIONS
35
36
1.4
1.4.1
Quotient spaces
1.4.1 Consider a spherically symmetric physical system in E3 (say, a central potential problem). We usually (for example, when solving the potential
problem by separation of variables) perform some manipulations to reduce
the space to sub-spaces and arrive finally to an equation in the sole variable
r. All the points on a sphere of fixed radius r will be equivalent, so that
each sphere will correspond to an equivalence class. We actually merge all
the equivalent points into one of them, labelled by the value of r, which is
taken as the representative of the class. On the other hand, points on spheres
37
into the region between the two spheres of radii (r and (r + . This
region is an open set in E3 so that J is an open set in the quotient space
if is supposed continuous. As distinctions between equivalent points are
irrelevant, the physical configuration space reduces to E1+ . The symmetry
transformations constitute the rotation group in E3 , the special orthogonal
group SO(3). In such cases, when the equivalence is given by symmetry
under a transformation group G, the quotient space is denoted S/G. Here,
<1>
E1+ = E3 /SO(3).
1.4.2 Inspired by the example above, we now formalize the general case
of spaces of equivalence classes. Suppose a topological space S is given on
which an equivalence relation R is defined: two points p and q are equivalent,
p q, if linked by the given relation. We can think of S as the configuration
space of some physical system, of which R is a symmetry: two points are
equivalent when attainable from each other by a symmetry transformation.
All the points obtainable from a given point p by such a transformation
constitute the equivalence class of p, indicated in general by [p] when no
simpler label is at hand. This class may be labelled by any of its points
instead of p, of course, and points of distinct classes cannot be related by
transformations. The set of equivalence classes, which we call {[p]} = S/R,
can be made into a topological space, whose open sets are defined as follows:
let : S S/R be the projection : p [p] (class to which p belongs).
Then a set U contained in S/R is defined to be open iff <1> (U ) is an
38
39
Figure 1.10: Kinds of open sets in the quotient topology: (left) first-kind: <1> (u) =
nZ ( + n, + n); (center) second-kind: <1> (u v) = nZ ( + n, + n 1); (right)
with the quotient topology, f (x) = exp[i2x] becomes a homeomorphism.
40
The resulting quotient space is the Klein bottle. Experiments with a paper
leaf will still be instructive, but frustating the real construction of the
bottle will show itself impossible (because we live in E3 ; it would be possible
in E4 !). Higher dimensional toruses T n = S 1 S 1 . . . S 1 (n times) are obtained
equally as quotients, T n = En /Zn .
1.4.6 The above examples and the previous chapter considerations on
compactification illustrate the basic fact: for spaces appearing in physical
problems, the topological characteristics are as a rule fixed by boundary
conditions and/or symmetry properties.
1.4.7 As previously said, it is customary to write G instead of R in the
quotient, S/G, when the relation is given by a group G as with the toruses,
the spherical case E1+ = E3 /SO(3), and S 1 = E1 /Z in 1.4.3. Consider
now the configuration space of a system of n particles. From the classical
point of view, it is simply E3n , the product of the configuration spaces of
the n particles. If we now take into account particle indistinguishability,
the particle positions become equivalent. The configuration space will be
the same, but with all the particle positions identified, or glued together.
In an ideal gas, for example, the points occupied by the n particles are to
be identified. The space is insensitive to the exchange of particles. These
exchanges constitute the symmetric group Sn (the group of all permutations
41
Figure 1.12: Scheme of E1 (1, +1) showing the equivalent points to form a Mobius
band.
1.4.2
Topological groups
42
43
The circle S 1 is thus a topological group and the set of positive real numbers
R+ is a topological subgroup of R\{0}. As suggested by the example of En ,
normed vector spaces in general (reviewed in Math.4), and consequently their
vector subspaces, are topological groups with the addition operation
m(a, b) = a + b.
1.4.14 The set Q of quaternions is a vector space on R with basis {1, i, j, k}.
It constitutes topological groups with respect to the addition and multiplication operations, and is isomorphic to R4 as a vector space. An element of
Q can be written as
q = a 1 + bi + cj + dk ,
with a, b, c, d R. The basis member 1 acts as the identity element. The
multiplications of the other members are given in Table 1.1.
i
j
k
i
-1
-k
+j
j
+k
-1
-i
k
-j
+i
-1
a2 + b2 + c2 + d2 = 1}.
44
euclidean spaces. These linear groups are neither abelian nor compact. Generalizing a bit: let V be a vector space; the sets of automorphisms and
endomorphisms of V are respectively
Aut V = {f: V V , such that f is linear and invertible}
End V = {f: V V , such that f is linear} .
Then, Aut V End V is a topological group with the composition of
linear mappings as operation. If we represent the linear mappings by their
matrices, this composition is nothing more than the matrix product, and we
have precisely the case GL(m, K).
The groups O(n) GL(m, R) of orthogonal n n matrices are other
examples. The special orthogonal groups SO(n) of orthogonal matrices with
det = +1 are ubiquitous in Physics and we shall come to them later.
1.4.17 The set of all matrices of the form ( L0 1t ), where L GL(n, R) and
t Rn , with the matrix product operation, is a topological group called the
affine group of dimension n. It is denoted A(n, R).
1.4.18 Linear projective groups Take the set Mn of all n n matrices
with entries in field K. It constitutes more than a vector space it is
an algebra. Each matrix A GL(n, K) defines an automorphism of the
form AM A1 , for all M Mn . This automorphism will be the same if A is
replaced by kA = kIn A, for any k K, and where In is the unit nn matrix.
The subgroup formed by the matrices of the type kIn is indeed ineffective on
Mn . The group of actual automorphisms is the quotient GL(n, K)/{kIn }.
Consider the subalgebra of Mn formed by the n projectors Pk = matrix
whose only nonvanishing element is the diagonal k-th entry, which is 1:
(Pk )rs = rs ks .
The Pk s are projectors into 1-dimensional subspaces. They satisfy the relations Pi Pj = ij Pi . Each one of the above automorphisms transforms this set
of projectors into another set of projectors with the same characteristics. For
this reason the automorphisms are called projective transformations, and the
group of automorphisms GL(n, K)/{kIn } is called the projective group, denoted P L(n, K). The n-dimensional space of projectors, which is taken into
an isomorphic space by the transformations, is the projective space, indicated
by KP n . There are, however, other approaches to this type of space.
1.4.19 Projective spaces Every time we have a problem involving only
the directions (not the senses) of vectors and in which their lengths are
irrelevant, we are involved with a projective space. Given a n-dimensional
vector space V over the field K, its corresponding projective space KP n is
45
1977.
46
quotient space formed by the equivalence classes of the relation. Any function
f on Cn+1 such that f (z) = f (cz) (function which is homogeneous of degree
zero) is actually a function on CP n .
1.4.20 Real Grassmann spaces Real projective spaces are generalized
in the following way: given the euclidean space EN with its vector space structure, its d-th Grassmann space GN d (E) [another usual notation: Gd (EN )] is
the set of all d-dimensional vector subspaces of EN . Remark that, as vectors
subspaces, they all include the origin (the zero vector) of EN . All Grassmannians are compact spaces (see 8.1.14). Projective spaces are particular
cases:
RP 1 = G1 (E2 ), RP 2 = G1 (E3 ), . . . , RP n = G1 (En+1 ).
Projectors are in a one-to-one correspondence with the subspaces of a vector
space. Those previously used were projectors of rank one. In the present
case the euclidean structure is to be preserved, so that the projectors must
be skew-symmetric endomorphisms, or matrices p satisfying p = pT and
p2 = p. If they project into a d-dimensional subspace, they must furthermore
be matrices of rank d. Consequently, GN d may be seen also as the space of
such projectors:
GN d = {p End(EN ) such that p2 = p = pT , rank p = d}.
Notice that GN d = GN,N d and the dimension is dim GN d = d(N d).
The set of orthogonal frames in the d-dimensional subspaces form another
space, the Stiefel space (see 8.1.15), which is instrumental in the general
classification of fiber bundles (section 9.7).
1.4.21 Complex Grassmann spaces Projective spaces are generalized
to the complex case in an immediate way. One starts from a euclideanized
(as above) complex vector space CN = (z1 , z2 , . . . , zn , zn+1 )}. Now, projectors
must be hermitian endomorphisms and the space is
N
GC
N d = Gd (C ) =
GN d (C) = {p End(CN ) such that p2 = p = p , rank p = d }.
47
are devoted to some notions coming under the name of algebraic topology,
which lead to the computation of some numbers of topological significance.
Roughly speaking, one looks for defects (such as holes and forbidden subspaces) in topological spaces, while remaining inside them. One way to find
such faults comes from the observation that defects are frequently related to
closed subspaces which do not bound other subspaces, as it happens with
some of the closed lines on the torus. Such subspaces give origin to discrete
groups and are studied in homology theory (chapter 2). Another way is by
trying to lasso them, drawing closed curves (and their higher dimensional
analogues) to see whether or not they can be continuously reduced to points
(or to other closed curves) in the space. Such loops are also conveniently
classified by discrete groups. This is the subject of homotopy theory (chapter 3). Both kinds of groups lead to some integer numbers, invariant under
homeomorphisms, which are examples of topological numbers. We shall later
introduce more structure on topological spaces and sometimes additional
structure can be used to signal topological aspects. Within differentiable
structure, vector and tensor fields arise which are able to reveal topological
defects (singular points, see Math.9) through their behaviour around them,
in a way analogous to the velocity field of a fluid inside an irregular container.
Later, in section 7.5, a little will be said about differential topology.
48
Chapter 2
HOMOLOGY
Dissection of a space into cellular building bricks provides information about
its topology.
49
50
2.1
CHAPTER 2. HOMOLOGY
GRAPHS
Spaces usually appearing in Physics have much more structure than a mere
topology. Exceptions are the graphs (or diagrams) used in perturbation techniques of Field Theory, cluster expansions in Statistical Mechanics, circuit
analysis, etc, whose interest comes in part from their mere topological properties. Graph Theory is a branch of Mathematics by itself, with important
applications in traffic problems, management planning, electronics, epidemic
propagation, computer design and programming, and everywhere else. Only
a very sketchy outline of the subject will be given, although hopefully pedantic enough to prepare for ensuing developments. Graphs provide a gate into
homological language. The first part below introduces them through a mere
formalization of intuitive notions. The second rephrases (and extends) the
first: its approach allows the introduction of the ideas of chain and boundary,
and sets the stage for the basic notions of simplex and complex.
2.1.1
2.1.1 We shall call (closed) edge any subspace in E3 which is homeomorphic to the interval I = [0, 1] with the induced topology. Indicating the
edge itself by e, and by h : I e the homeomorphism, the points h(0) and
h(1) will be the vertices of the edge. A graph G E3 is a topological space
defined in the following way:
(i) the point set consists of the points of a union of edges satisfying the
condition that the intersection of any two edges is either or one common
vertex;
(ii) the open sets are subsets X G whose intersection with each edge
is open in that edge.
2.1.2 Because edges are homeomorphic to I with the induced topology
(and not, for example, with the quotient topology of 1.4.3), no isolated
one-vertex bubbles like
are admitted. Notice that knots (Mathematical
Topic 2.3) are defined as subspaces in E3 homeomorphic to the circle S 1 . So,
absence of bubbles means that no knots are parts of graphs. Graphs can be
defined in a less restrictive way with no harm to the results involving only
themselves, but the above restrictions are essential to the generalization to
be done later on (to simplexes and the like). From (ii), a subset is closed if its
intersection with each edge is closed in that edge. The open edges, obtained
by stripping the closed edges of their vertices, are open sets in G. The set of
vertices is discrete and closed in G. A graph is compact iff it is finite, that is,
if it has a finite number of edges. It is connected iff any two of its vertices can
2.1. GRAPHS
51
52
CHAPTER 2. HOMOLOGY
vertices up the number of peaks (V) minus the number of passes (E) plus
the number of valleys (L) equals one! To compare with the non-planar case,
consider the tetrahedron of Figure 2.2. A simple counting shows that (G)
= V E + L = 2. Its plane, flattened version beside has (G) = 1 because
the lower face is no more counted. Notice that only elementary loops,
those not containing sub-loops, are counted (see 2.2.3 below).
Figure 2.1: Scheme of an island, with peaks (vertices), passes (edges) and valleys (loops).
2.1.2
2.1.4 Consider now E3 as a vector space, and choose any three of its
vectors v0 , v1 , v2 , imposing however that (v1 - v0 ) and (v2 - v0 ) be linearly
independent:
k 1 (v1 v0 ) + k 2 (v2 v0 ) = 0 implies k 1 = k 2 = 0.
Defining a1 = k 1 , a2 = k 2 , a0 = (k 1 + k 2 ), this is equivalent to saying that
the two conditions
a0 v 0 + a1 v 1 + a2 v 2 = 0
a0 + a1 + a2 = 0
2.1. GRAPHS
53
2
X
i=0
b vi ;
2
X
bi = 1.
(2.1)
i=0
2.1.5 Suppose we consider only two of the vectors, say v0 and v1 . They
are of course linearly independent. We can in this case define, just as above,
their dependent vectors by
e = b0 v0 + b1 v1 , b0 + b1 = 1.
Now, the coordinates b0 and b1 determine points on a straight line. In the
same way, we can take only one of the vectors, say v0 , and its dependent
vector as v = v0 itself: this will determine a point.
Add now an extra condition on the dependent vectors: that each coordinate be strictly positive, bi > 0. Then, the vector b above will span the
54
CHAPTER 2. HOMOLOGY
interior of a triangle; the vector e will span an open edge; and v0 again will
span an isolated point, or a vertex. Notice that the coordinates related to
the vector e give actually a homeomorphism between the interval (0, 1) and a
line segment in E3 , justifying the name open edge. If instead we allow bi 0,
a segment homeomorphic to the closed interval [0, 1] results, a closed edge.
With these edges and vertices, graphs may now be defined as previously. An
edge can be indicated by the pair (vi , vk ) of its vertices, and a graph G by a
set of vertices plus a set of pairs. An oriented graph is obtained when all the
pairs are taken to be ordered pairs which is a formal way of putting arrows
on the edges. A path from vertex v1 to vertex vn+1 is a sequence of edges
e1 e2 e3 . . . en with ei = (vi , vi+1 ) or ei = (vi+1 , vi ). We have said that a graph is
connected when, given two vertices, there exists at least one path connecting
them. It is multiply-connected when there are at least two independent, nonintersecting paths connecting any two vertices. It is simply-connected when
it is connected but not multiply-connected. In Physics, multiply-connected
graphs are frequently called irreducible graphs.
2.1.6 A path is not supposed to accord itself to the senses of the arrows.
A path is called simple if all its edges are distinct (one does not go twice
through the same edge) and all its vertices are distinct except possibly v1 =
vn+1 . In this last case, it is a loop. An Euler path on G is a path with all edges
distinct and going through all the vertices in G. The number nk of edges
starting or ending at a vertex vk is its degree (coordination number would
be more to the physicists taste; chemists would probablyP
prefer valence).
Clearly the sum of all degrees on a graph is even, as V1 ni = 2E. The
number of odd vertices (that is, those with odd degrees) is consequently even.
2.1.7 The Bridges of K
onigsberg Graph theory started up when Euler
faced this problem. There were two islands in the river traversing Kants
town, connected between each other and to the banks by bridges as in the
scheme of Figure 2.4. People wanted to know whether it was possible to do
a tour traversing all the bridges only once and finishing back at the starting
point. Euler found that it was impossible: he reasoned that people should
have a departure for each arrival at every point, so that all degrees should
be even which was not the case.
2.1.8 Graph Theory has scored a beautiful victory for Mathematics with
the recent developments on the celebrated (see Phys.3.2.5) four-color problem. The old conjecture, recently demonstrated, was that four colors were
sufficient for the coloring of a map. This is a problem of graph theory. As said
above, graphs are also of large use in many branches of Physics. Through
2.1. GRAPHS
55
56
CHAPTER 2. HOMOLOGY
57
2.2
2.2.1
2.2.1 Let us now proceed to generalize the previous considerations to objects which are more than graphs. Consider En with its structure of vector
space. A set of vertices is said to be linearly independent if, for the set of
vectors (v0 , v1 , v2 , . . . , vp ) fixing them, the two conditions
a0 v 0 + a1 v 1 + . . . + ap v p = 0
a0 + a1 + . . . + ap = 0
imply a0 = a1 = . . . = ap = 0. This means that the vectors (vi v0 )
are linearly independent. We define a vector b dependent on the vectors
v0 , v1 , . . . , vp by
P
P
b = pi=0 bi vi ; pi=0 bi = 1.
58
CHAPTER 2. HOMOLOGY
The points determined by the barycentric coordinates bi describe a p-dimensional subspace of En , more precisely, an euclidean subspace of En .
2.2.2 A (closed) simplex of dimension p (or a p-simplex ) with vertices
v0 , v1 , v2 , . . . , vp is the set of points determined by the barycentric coordinates satisfying the conditions bi 0 for i = 0, 1, 2, . . . , p. Special cases are
points (0-simplexes), closed intervals (1-simplexes), triangles (2-simplexes)
and tetrahedra (3-simplexes). A p-simplex is indicated by sp and is said to
be generated by its vertices. The points with all nonvanishing barycentric
coordinates are interior to the simplex, their set constituting the open simplex. The boundary sp of sp is the set of points with at least one vanishing
coordinate bi = 0. Given sp generated by vo , v1 , v2 , . . . , vp , any subset of
vertices will generate another simplex: if sq is such a subsimplex, we use the
notation sq h sp . It is convenient to take the empty set as a subsimplex of
any simplex. Dimension theory gives the empty set the dimension (-1), so
that the empty simplex is designated by s1 . The edge in the left branch of
Figure 2.6 is a 1-simplex in E3 . Its boundary is formed by the vertices v0
and v1 , which are also subsimplexes:
v 0 h s1 ; v 1 h s1 .
s1 h vo
s1 h v1 .
59
60
CHAPTER 2. HOMOLOGY
61
n
X
()i Ni .
i=0
62
CHAPTER 2. HOMOLOGY
like to have for the boundary an orientation coherent with that of the 2simplex. The figure suggests that the faces coherently oriented with respect
to the triangle are v0 v1 , v1 v2 and v2 v0 , so that the oriented boundary is
(v0 v1 v2 ) = v0 v1 + v1 v2 + v2 v0 .
(2.2)
2.2.14 Notice that the opposite orientation would be coherent with the
opposite orientation of the edges. All the possible orientations are easily
found to be equivalent to one of these two. As a general rule, for a p-simplex
(v0 v1 v2 . . . vp ) = (vi0 vi1 vi2 . . . vip ) , the sign being + or - according to
whether the permutation
0 1 2 ... p
i0 i1 i2 ... ip
is even or odd. An equivalent device is to think of (v0 v1 v2 . . . vp ) as an antisymmetric product of the vertices, or to consider that vertices anticommute.
This is consistent with the boundary of an edge,
(v0 v1 ) = v1 v0 .
As a mnemonic rule,
(v0 v1 v2 . . . vp ) = v1 v2 . . . vp v0 v2 . . . vp + v0 v1 v3 . . . vp . . .
The successive terms are obtained by skipping each time one vertex in the
established order and alternating the sign. For a 0-simplex, the boundary is
defined to be 0 (which is the zero chain). Each term in the sum above, with
the corresponding sign, is a (oriented) face of sp .
2.2.15 Let us go back to eq.[2.2], giving the boundary of the simplex
(v0 v1 v2 ). What is the boundary of this boundary? An immediate calculation
shows that it is 0. The same calculation, performed on the above general
definition of boundary, gives the same result because, in (v0 v1 v2 . . . vp ),
63
each (p-2)-simplex appears twice and with opposite signs. This is a result of
fundamental significance:
sp 0.
(2.3)
cp =
m1 s1p
m2 s2p
+ ... +
p
m Np sN
p
Np
X
mj sjp ,
j=1
cp =
Np
X
mj sjp .
j=1
Each term in the right hand side is a face of cp . Equation [2.3] implies that,
also for chains,
cp 0.
This is one of the many avatars of one of the most important results of all
Mathematics, called by historical reasons Poincare lemma. We shall meet it
again, under other guises. In the present context, it is not far from intuitive.
Think of the sphere S2 , the boundary of a 3-dimensional ball, or of the torus
T2 which bounds a full-torus, or of any other usual boundary in E3 : they
have themselves no boundary.
64
2.2.2
CHAPTER 2. HOMOLOGY
Topological numbers
There are topological invariants of various types: some are general qualities of
the space, like connectedness and compactness; other are algebraic structures
related to it, as the homotopic groups to be seen in chapter 3; still other are
numbers, like the Euler number and dimension. Whatever they may be,
their common point is that they cannot be changed by homeomorphisms.
Homology provides invariants of two kinds: groups and numbers.
2.2.18 The p-chains satisfying
cp = 0
are called closed p-chains, or p-cycles. The zero p-chain is a trivial p-cycle.
The p-cycles constitute a subgroup of Cp (K), denoted Zp (K).
2.2.19 We have been using integer numbers for the chain coefficients:
mj Z. In fact, any abelian group can be used instead of Z. Besides
the integer homology we have been considering, other cases are of great importance, in particular the real homology with coefficients in R. As we can
also multiply chains by numbers in Z (or R), chains constitute actually a
vector space. To every vector space V corresponds its dual space V formed
by all the linear mappings taking V into Z (or R). V is isomorphic to V
(as long as V has finite dimension) and we can introduce on V constructs
analogous to chains, called cochains. A whole structure dual to homology,
cohomology, can then be defined in a purely algebraic way. This would take
us a bit too far. We shall see later that chains, in the case of differentiable
manifolds, are fundamentally integration domains. They are dual to differential forms and to every property of chains correspond an analogous property
of differential forms. Taking the boundary of a chain will correspond to taking the differential of a form, and Poincare lemma will correspond to the
vanishing of the differential of a differential. Differential forms will have the
role of cochains and we shall leave the study of cohomology to that stage,
restricting the treatment here to a minimum.
p is the sum of all (p+1)-chains
2.2.20 Given a chain cp , its coboundary c
of which cp is an oriented face. Although this is more difficult to see, the
65
A
C
D
Figure 2.9: A tetrahedron (left) and a triangulation of the sphere S 2 (right).
2.2.23 Consider the complex formed by the surface of a tetrahedron (Figure 2.9). The vertices A, B, C and D can be used to define the edges AB,
AC, BD, etc. Let us begin by listing and commenting on some results:
(i) A = 0 ; (AB) = B A.
Notice: B is a face of AB, while A is not. Only the oriented vertex (A)
is a face.
= BA + CA + DA (notice the correct signs!);
(ii) A
(iii) 2 (AB) = 0 ; 2 (AB) = 0;
(iv) (AB) = 4AB.
66
CHAPTER 2. HOMOLOGY
67
Figure 2.10: Only the lower two simplexes are real triangulations of the circle.
is a juxtaposition of 1-simplexes joined in such a way that the resulting complex is homeomorphic to it. Now, an edge must have two distinct vertices,
so that simplex (1) in Figure 2.10 is not suitable. The two edges in (2) are
supposed to be distinct but they have two identical vertices. The complexes
(3) and (4) are good triangulations of S 1 . Take the case (3). It is easily seen
that:
= CA + BA; B
= AB + CB;
A
68
CHAPTER 2. HOMOLOGY
ker p = {cp Cp (K) such that p cp = 0}
is, of course, the set of p-cycles. We have already said that it constitutes a
group, which we shall denote by Zp (K). Consider p-cycles p , p , p Zp (K),
and (p + 1)-chains p+1 , p+1 Cp+1 (K). If
p = p + p+1 and p = p + p+1 ,
then
p = p + (p+1 + p+1 ) .
Consequently, the relation between p-cycles which differ by a boundary is an
equivalence, and divides Zp (K) into equivalence classes. Each class consists
of a p-cycle and all other p-cycles which differ from it by a boundary. The
equivalence classes can be characterized by those p , p such that no p+1
exists for which p p = p+1 . When such a p+1 does exist, p and p
are said to be homologous to each other. The relation between them is a
homology and the corresponding classes, homology classes. Let us be formal
once more: consider the image of p+1 , the operator acting on (p+1)-chains:
Im p+1 = {those cp which are boundaries of some cp+1 } .
The set of these p-boundaries form still another group, denoted Bp (K). From
the Poincare lemma, Bp (K) Zp (K). Every boundary is a cycle although
not vice-versa. Bp (K) is a subgroup of Zp (K) and there is a quotient group
Hp (K) = Zp (K)/Bp (K) .
2.2.27 This quotient group is precisely the set of homology classes referred
to above. Roughly speaking, it counts how many independent p-cycles exist
which are not boundaries. We have been talking of general complexes, which
can in principle be homeomorphic to some topological spaces. If we restrict
ourselves to finite complexes, which can only be homeomorphic to compact
spaces, then it can be proved that the ranks (number of generators) of all the
groups above are finite numbers. More important still, for finite complexes,
the groups H p (K) are isomorphic to the groups
of harmonic p-cycles .
Consequently,
bp (K) = rank Hp (K) .
69
2.2.28 These Hp (K) are the homology groups of K and, of course, of any
space homeomorphic to K. Let us further state a few important results:
(i) the meaning of b0 : b0 (S) is the number of connected components of S;
(ii) the Poincare duality theorem: in an n-dimensional space S,
bnp (S) = bp (S);
(iii) the Euler-Poincare characteristic is related to the Betti numbers by:
(S) =
Pn
j=0 bj (S).
70
CHAPTER 2. HOMOLOGY
the calculations to (about) one half. The Euler characteristic may be found
either by counting in the triangulation
X
()i Ni = 9 27 + 18 = 0,
i
or from
X
()i bi = 1 2 + 1 = 0.
71
fk : Hk (P ) Hk (P 0 )
between the corresponding homology groups. Such homomorphisms become
isomorphisms when f is a homeomorphism.
The homology group H1 is the abelianized subgroup of the more informative fundamental homotopy group 1 , to be examined in section 3.1.2.
2.2.32 Once the Betti numbers for a space S are found, they may be put
together as coefficients in the Poincare polynomial
PS (t) =
Pn
j
j=0 bj (S) t
72
CHAPTER 2. HOMOLOGY
Feynman diagrams1 or in the incredibly beautiful version of General Relativity found by Regge.2 Much more appears when differential forms are
at work and we shall see some of it in chapter 7. Let us finish with a few
words on cubic homology. In many applications it is easier to use squares
and cubes instead of the triangles and tetrahedra we have being employing.
Mathematicians are used to calculate homotopic properties (next section is
devoted to applying complexes to obtain the fundamental group) in cubic
complexes. For physicists they are still more important, as cubic homology
is largely used in the lattice approach to gauge theories.3 Now: cubes cannot,
actually, be used as simplexes in the simple way presented above. The resulting topological numbers are different: for instance, a space with a single
point would have all of the bj = 1. It is possible, however, to proceed to
a transformation on the complexes (called normalization), in such a way
that the theory becomes equivalent to the simplex homology.4 Triangulations
are more fundamental because (i) any compact space (think on the square)
can be triangulated, but not every compact space can be squared (think on
the triangle), and (ii) the topological numbers coming up from triangulations
coincide with those obtained from differential forms. In a rather permissive
language, it is the triangles and tetrahedra which have the correct continuum
limits.
Chapter 3
HOMOTOPY
3.0
GENERAL HOMOTOPY
We have said that, intuitively, two topological spaces are equivalent if one can
be continuously deformed into the other. Instead of the complete equivalence
given by homeomorphisms in general difficult to uncover we can more
modestly look for some special deformations preserving only a part of the
topological characteristics. We shall in this section examine one-parameter
continuous deformations. Roughly speaking, homotopies are function deformations regulated by a continuous parameter, which may eventually be
translated into space deformations. The topological characterization thus
obtained, though far from complete, is highly significant.
In this section we shall state many results without any proof. This is
not to be argued as evidence for a neurotic dogmatic trend in our psychism.
In reality, some of them are intuitive and the proofs are analogous to those
given later in the particular case of homotopy between paths.
3.0.1 Homotopy between functions Let f and g be two continuous
functions between the topological spaces X and Y , f, g : X Y . Let again
I designate the interval [0, 1] included in E1 . Then
f is homotopic to g (f g) if there exists a continuous function
F : X I Y such that F (p, 0) = f (p) and F (p, 1) = g(p) for every
p X. The function F is a homotopy between f and g.
Two functions are homotopic when they can be continuously deformed into
each other, in such a way that the intermediate deformations constitute a
73
74
CHAPTER 3. HOMOTOPY
family of continuous functions between the same spaces: F (p, t) is a oneparameter family of continuous functions interpolating between f and g. For
fixed p, it gives a curve linking their values.
3.0.2 Two constant functions f, g : X Y with f (p) = a and g(q) = b for
all p, q X are homotopic if Y is path-connected. In this case, a and b can
be linked by a path : I Y , with (0) = a and (1) = b. Consequently,
F (p, t) = (t) for all (p, t) X I is a homotopy between f and g (Figure
3.1).
Figure 3.1: On the left, the cartesian product X I; on the right, the space
Y where the images of f and g lie and, for each value of t, of the mediating
function F (p, t).
75
76
CHAPTER 3. HOMOTOPY
77
Figure 3.3: (left) The plane is homotopically equivalent to a point (here, the
origin), that is, it is contractible.
Figure 3.4: (right) The punctured plane E2 /{0} is homotopically equivalent
to the circle.
78
CHAPTER 3. HOMOTOPY
3.0.11 We have repeatedly said that homotopies preserve part of the topological aspects. Contractibility, when extant, is one of them. We shall in the
following examine some other, of special relevance because they appear as
algebraic structures, the homotopy groups. The homology groups presented
in chapter 2 are also invariant under homotopic transformations.
3.0.12 General references Two excellent and readable books on homotopy are: Hilton 1953 and Hu 1959. A book emphasizing the geometrical
approach, also easily readable, is Croom 1978.
3.1
PATH HOMOTOPY
3.1.1
Homotopy of curves
79
Figure 3.5: For each value of t in I, F (s, t) gives a path, intermediate between
f and g.
Figure 3.6: For each fixed value of s, F (s, t) gives a path going from f (s) to
g(s).
80
CHAPTER 3. HOMOTOPY
I I, it is continuous on I I. It is a homotopy f h.
3.1.4 By definition, every pair of points inside each path-component of a
given space X can be linked by some path. Let {C } be the set of pathcomponents of X, which is denoted by 0 (X):
0 (X) = {C }.
This notation is a matter of convenience: 0 will be included later in the
family of homotopy groups and each one of them is indicated by n for some
integer n. The relation between the path homotopy classes on a space X and
the path components of the space X I of all paths on X is not difficult to
understand (Figure 3.8). A homotopy F between two paths on X is a path
on X I and two homotopic paths f and g can be thought of as two points of
X I linked by the homotopy F .
3.1.5 The intermediate curves representing the successive stages of the
continuous deformation must, of course, lie entirely on the space X. Suppose
X to have a hole as in Figure 3.9. Paths f and g are homotopic to each other
and so are j and h. But f is homotopic neither to j nor to h. The space X I
is so divided in path-components, which are distinct for f and j.
There are actually infinite components for X I when X has a hole: a
curve which turns once around the hole can only be homotopic to curves
turning once around the hole; a curve which turns twice around the hole
will be continuously deformable only into curves turning twice around the
hole; and the same will be true for curves turning n times around the hole,
suggesting a relation between the homotopy classes and the counting of the
number of turns the curves perform around the hole. Another point is that
81
Figure 3.8: Relation between path homotopy classes on a space X and the
path components of the space X I of all paths on X.
82
CHAPTER 3. HOMOTOPY
83
84
CHAPTER 3. HOMOTOPY
defines f g, and
H(s, 1) =
f 0 (s)
for s [0, 21 ]
g 0 (2s 1) for s [ 12 , 1].
(f (g h))(s) =
f (2s)
for s [0, 12 ]
h(2s 1)
for s [ 12 , 1].
85
f (4s)
for s [0, 14 ]
g(4s 1) for s [ 14 , 12 ]
h(2s 1) for s [ 12 , 1].
4s
f ( t+1
)
for s [0, t+1
]
, t+2
]
g(4s t 1) for s [ t+1
F (s, t) =
4
4
h( 4st2
)
for s [ t+2
, 1].
2t
4
The arguments of f, g and h are all in [0, 1]. It is easily checked that
F (s, 1) is the same as (f (g h))(s), and that F (s, 0) is the same as ((f
g) h)(s).
3.1.2
Classes of closed curves constitute a discrete group, which is one of the main
topological characteristics of a space. Here we shall examine the general
properties of this group. Important examples of discrete groups, in particular
the braid and knot groups, are described in Math.2. Next section will be
devoted to a method to obtain the fundamental group for some simple spaces
as word groups.
3.1.10 We have seen in the previous section that the path product
does not define a group structure on the whole set of homotopy classes. We
shall now restrict that set so as to obtain a group the fundamental group
which turns out to be a topological characteristic of the underlying space.
3.1.11 Let X be a topological space and x0 a point in X. A path whose
both end-points are x0 is called a loop with base point x0 . It is a continuous
function f : I X, with f (0) = f (1) = x0 .
3.1.12 The set of homotopy classes of loops with a fixed base point constitutes a group with the operation . More precisely:
86
CHAPTER 3. HOMOTOPY
(i) given two loops f and g with the same base point x0 , their composition
f g is well defined and is a loop with base point x0 (Figure 3.13);
(ii) the properties of associativity, existence of an identity element [ex0 ],
and existence of an inverse [f <1> ] for each [f ] hold evidently .
3.1.13 The group formed by the homotopy classes of loops with base point
x0 is called the fundamental group of space X relative to the base point x0
and is denoted 1 (X, x0 ). It is also known as the Poincare group (a name
not used by physicists, to avoid confusion) and first homotopy group of X at
x0 , because of the whole series of groups n (X, x0 ) which will be introduced
in section 3.3.
3.1.14 A question arising naturally from the above definition is the following: how much does the group depend on the base point? Before we answer
87
Fraleigh 1974.
88
CHAPTER 3. HOMOTOPY
if X is path-connected and both x0 and x1 are in X, then
1 (X, x0 ) is isomorphic to 1 (X, x1 ) .
3.1.16 Suppose now that X is a topological space and C is a path-component of X to which x0 belongs. As all the loops at x0 belong to C and
all the groups are isomorphic, we may write 1 (C) = 1 (X, x0 ). The group
depends only on the path-component and gives no information at all about
the remaining of X. In reality, we should be a bit more careful, as the
isomorphism # above is not natural (or canonical ): it depends on the path
. This means that different paths between x0 and x1 take one same element
of 1 (X, x0 ) into different elements of 1 (X, x1 ), although preserving the
overall abstract group structure. The isomorphism is canonical only when
the group is abelian. We shall be using this terminology without much ado
in the following, though only in the abelian case is writing 1 (C) entirely
justified. These considerations extend to the whole space X if it is pathconnected, when we then talk of the group 1 (X).
Figure 3.15:
3.1.17 A class [f ] contains its representative f and every other loop continuously deformable into f . Every space will contain a class [c(t)] of trivial,
or constant, closed curves, c(t) = x0 for all values of t(Figure 3.15). If this
class is the only homotopy class on X, 1 (X, x0 ) = [c(t)], we say that X is
simply-connected. More precisely,
X is simply-connected when it is path-connected
and 1 (X, x0 ) is trivial for some x0 X.
89
Figure 3.16:
3.1.18 When X is simply-connected, every two loops on X can be continuously deformed into each other, and all loops to a point. When this is
not the case, X is multiply-connected. This is a first evidence of how 1 (X)
reflects, to a certain degree, the qualitative structure of X. The fact that
closed laces in a multiply-connected space cannot be caused to shrink to a
point while remaining inside the space signals the presence of some defect,
of a hole of some kind. The plane E2 is clearly simply-connected, but this
changes completely if one of its points is excluded, as in E2 {0}: loops not
encircling the zero are trivial but those turning around it are not (Figure
3.16). There is a simple countable infinity of loops not deformable into each
other. We shall see later (on 3.1.33) that 1 (E2 \{0}) = Z, the group of
integer numbers. We shall also see ( 3.1.34) that the delection of two points,
say as in E2 \{ 1, +1}, leads to a non-abelian group 1 . The space E3 with
an infinite line deleted (say, x = 0, y = 0) is also a clear non-trivial case.
Some thought dedicated to the torus T 2 = S 1 S 1 will, however, convince
the reader that such intuitive insights are yet too rough.
3.1.19 The trivial homotopy class [c(t)] acts as the identity element of
1 and is sometimes denoted [ex0 ]. More frequently, faithful to the common
usage of the algebrists, it is denoted by 0 when 1 is abelian, and 1 when
it is not (or when we do not know or care).
3.1.20 All the euclidean spaces En are simply-connected: 1 (En ) = 1. The
n-dimensional spheres S n for n 2 are simply-connected: 1 (S n ) = 1 for
90
CHAPTER 3. HOMOTOPY
Figure 3.17: A continuous mapping between spaces X and Y induces a homomorphism between their fundamental groups.
91
92
CHAPTER 3. HOMOTOPY
3.1.3
Some Calculations
3.1.27 Triangulations provide a simple means of obtaining the fundamental group. The method is based on a theorem4 whose statement requires
some preliminaries.
(i) Take a path-connected complex with vertices v1 , v2 , v3 , . . . , and edges
(v1 v2 ), (v1 v3 ), (v2 v3 ), etc. To each oriented edge we attribute a symbol,
(vi vk ) gik .
1
We then make the set of such symbols into a group, defining gik
= gki and
imposing, for each triangle (vi vj vk ), the rule
Schulman 1968.
DeWitt-Morette 1969; Laidlaw & DeWitt-Morette 1971; DeWitt-Morette 1972.
Hu 1959; see also Nash & Sen 1983.
93
3.1.29 Take again the (surface) tetrahedron as in Figure 3.18 and, for P ,
the connected path 1 2 3 4. For each triangle there is a condition:
g12 g24 g41 = 1; g41 g13 g34 = 1 ; g23 g34 g42 = 1; g12 g23 g31 = 1.
We now impose also g12 = g23 = g34 = 1, because the corresponding edges
belong to the chosen path. As a consequence, all the group elements reduce
to the identity. The fundamental group is then 1 (tetrahedron) = {1}. The
tetrahedron is simply-connected, and so are the spaces homeomorphic to it:
the sphere S 2 , the ellipsoid, etc.
3.1.30 The disc in E2 , that is, the circle S 1 and the region it circumscribes
(Figure 3.19); it is perhaps the simplest of all cases. The triangulation in the
figure makes it evident that
94
CHAPTER 3. HOMOTOPY
Figure 3.19:
Figure 3.20:
95
g51 = g 0 g 00 = g 00 g 0 ,
so that g 0 and g 00 commute. The group has two independent generators
commuting with each other. Consequently, 1 (T 2 ) = Z Z.
3.1.33 The disc with one hollow. A triangulating complex for the oncepunctured disk is given in Figure 3.22, as well as a chosen path. When two
sides of a triangle are on the path, the third is necessarily in correspondence
with the identity element. From this, it comes out immediately that g12 =
96
CHAPTER 3. HOMOTOPY
g23 = g45 = g78 = g48 = 1. It follows also that g68 = 1. We remain with two
conditions, g51 g16 = 1 and g16 g67 = 1, from which g51 = g67 = g61 =: g. One
independent generator: 1 = Z. In fact, the hollowed disk is homotopically
equivalent to S 1 . Notice that the dark, absent triangle was not used as a
simplex: it is just the hollow. If it were used, one more condition would be
at our disposal, which would enforce g = 1. Of course, this would be the disk
with no hollow at all, for which 1 = {1}.
3.1.34 The twice-punctured disk. Figure 3.23 shows, in its left part,
a triangulation for the disk with two hollows and a chosen path, the upper
part being a repetition of the previous case. We find again g51 = g67 =
g61 =: g. The same technique, applied to the lower part gives another,
independent generator: g4,13 = g3,13 = g3,11 =: g 0 . The novelty here is that
the group is non-commutative. Of course, we do not know a priori the
relative behaviour of g and g 0 . To see what happens, let us take (right part
of Figure 3.23) the three loops , and starting at point (10), and examine
their representatives in the triangulation:
97
As and are not homotopic, their classes are different: [] 6= []. But
[] = [ 1 ][][] ,
so that [][] 6= [][], or gg 0 6= g 0 g. The group 1 is non-abelian with two
generators and no specific name. It is an unnamed group given by its presentation (Math 2.2). It might be interesting to examine the Bohm-Aharonov
effect corresponding to this case, with particles described by wavefunctions
with two or more components to (possibly) avoid the loss of information on
the group in 1-dimensional representations.
3.1.35 The projective line RP1 (see 1.4.19): as the circle is homotopically equivalent to the hollowed disk, we adapt the triangulation of
3.1.33 for S 1 (see Figure 3.24, compared with Figure 3.22) with identified
antipodes: 1 = 1, 2 = 2, etc. Notice that this corresponds exactly
to a cone with extracted vertex. The path is 1 2 3 4 5 6 and,
of course, its antipode. An independent generator remains,
g = g14 = g26 = g16 = g36 = g56 = g15 ,
so that 1 (RP1 ) = Z. This is to be expected, as RP1 S 1 .
Figure 3.24:
3.1.36 The projective plane RP2 : the sphere S 2 with identified antipodes may be described as in Figure 3.25 (see 1.4.19). From the equations
g13 g34 = 1; g34 g42 = 1; g24 g41 = 1; g31 g14 = 1 ,
98
CHAPTER 3. HOMOTOPY
Figure 3.25:
3.2
3.2.1
COVERING SPACES
Multiply-connected Spaces
99
Figure 3.26:
Unless we had carefully chosen the boundary values at the start, will
become multivalued on L1 = L2 . Of course, this operation is a violence
against the space: it changes radically its topology from a topology analogous
to that of 1.3.12 to another, akin to that of 1.4.3. The initial box is simplyconnected, the final annulus is multiply-connected. But the point we wish to
stress is that, unless the boundary conditions were previously prepared by
hand so as to match, will become multivalued. Had we started with the
annulus as configuration space from the beginning, no boundaries as L1 and
L2 would be present.
3.2.2 If we want to be single-valued, we have to impose it by hand (say,
through periodic conditions). This happens in Quantum Mechanics when
100
CHAPTER 3. HOMOTOPY
Figure 3.27:
where A is some vector. The integration is to be performed along some curve,
such as 1 or 2 . The question of interest is: consider 1 and 2 to be curves
linking points a and b. Are the two integrals
R
R
1 (b) = 1 A dl and 2 (b) = 2 A dl
equal to each other? It is clear that
5
6
7
101
1 (b) 2 (b) =
R
1 2
A dl =
A dl,
the last integration being around the closed loop starting at a, going through
1 up to b, then coming back to a through the inverse of 2 , which is given
by (2 )1 = 2 . We see that (b) will be single-valued iff
H
A dl = 0 .
If the region S circumvented by 1 2 = 1 + (2 )1 is simply-connected,
then 1 2 is just the boundary S of S, and Greens theorem implies
R
R
A dl = S rotA d.
1 2
For general x, the single-valuedness condition for (x) is consequently given
by rot rotA = 0. In a contractible domain this means that some (x) exists
such that A = grad. In this case,
Rx
(x) = a grad dl = (x) (a) ,
so that we can choose (a) = 0 and (x) = (x). When this is the case, A
is said to be integrable or of potential type ( is its integral, or potential ), a
nomenclature commonly extended to (x) itself.
Figure 3.28:
102
CHAPTER 3. HOMOTOPY
The monodromy condition, B = rot A = 0, is the absence of field. Nonvanishing electromagnetic fields are, for this reason, non-integrable phase
factors.8
3.2.5 Consider now a multiply-connected configuration space, as the annulus in Figure 3.28. Because the central region is not part of the space,
1 2 is no more the boundary of a domain. We can still impose A =
rot [v], for some v(x), but then v(x) is no longer single-valued.9 We can
still write B =rotA, but neither is A single-valued! In Quantum Mechanics
there is no reason for the phases to be single-valued: only the states must be
unique for the description to be physically acceptable. A state corresponds
to a ray, that is, to a set of wavefunctions differing from each other only
by phases. Starting at a and going through 1 2 , the phase changes in
proportion to the integral of A around the hole which, allowing multivalued
animals, can be written
I
Z
Z
2e
2e
2e
2e
rotA d =
B d =
A dl =
,
(phase) =
~c
hc S
hc S
hc
being the flux of B through the surface S circumvented by 1 2 (which
is not its boundary now!). In order to have
e
(a) = ei ~c (a)
single-valued, we must have (2e/hc) = 2n, that is, the flux is quantized:
=n
hc
.
e
It may happen that this condition does not hold, as in the Bohm-Aharonov
effect (see 4.2.18).10
3.2.6 All these considerations (quite schematic, of course) have been made
only to emphasize that multi-connectedness can have very important physical
consequences.
3.2.7 We shall in the next section examine more in detail the relation between monodromy and multiple-connectedness. Summing it up, the following
will happen: let X be a multiply-connected space and a function on X.
The function will be multivalued in general. A covering space E will be
8
9
10
Yang 1974.
Budak & Fomin 1973.
Aharonov & Bohm 1959; 1961.
103
an unfolding of X, another space on which becomes single-valued. Different functions will require different covering spaces to become single-valued,
but X has a certain special covering, the universal covering U (X), which
is simply-connected and on which all functions become single-valued. This
space is such that X = U (X)/1 (X). The universal covering U (X) may be
roughly seen as that unfolding of X with one copy of X for each element of
1 (X).
3.2.8 Recall the considerations on the configuration space of a system of
n identical particles ( 1.4.7; see also 3.2.29 below, and Math.2.9), which is
E3n /Sn . In that case, E3n is the universal covering and 1 Sn . Consider,
to fix the ideas, the case n = 2. Call x1 and x2 the positions of the first
and the second particles. The covering space E6 is the set {(x1 , x2 )}. The
physical configuration space X would be the same, but with the points (x1 ,
x2 ) and (x2 , x1 ) identified. Point (x2 , x1 ) is obtained from (x1 , x2 ) by the
action of a permutation P12 , an element of the group
S2 : (x2 , x1 ) = P12 (x1 , x2 ).
A complex function (x1 , x2 ) (say, the wavefunction of the 2-particle system)
will be single-valued on the covering space, but 2-valued on the configuration
space. To make a drawing possible, consider instead of E3 , the two particles
on the plane E2 . The scheme in Figure 3.29 shows how (x1 , x2 ) is singlevalued on E, where (x1 , x2 ) 6= (x2 , x1 ), and double-valued on X, where the
two values correspond to the same point (x1 , x2 ) (x2 , x1 ). There are two
sheets because P12 applied twice is the identity.
Wavefunctions commonly used are taken on the covering space, where they
are single-valued. The function [P12 (x1 , x2 )] is obtained from (x1 , x2 ) by
the action of an operator U (P12 ) representing P12 on the Hilbert space of
wavefunctions:
(x2 , x1 ) = [P12 (x1 , x2 )] = U (P12 )(x1 , x2 ).
This is a general fact: the different values of a multivalued function are
obtained by the action of a representation of a group, a distinct subgroup
of 1 (X) for each function. Above, the group S2 is isomorphic to the cyclic
group Z2 (notice
the analogy of this situation with the covering related to
the function z in next section). The whole fundamental group will give all
the values for any function. There are as many covering spaces of X as there
are subgroups of 1 (X).
3.2.9 Percolation Phase transitions are more frequently signalled by singularities in physical quantities, as the specific heat. However, sometimes
104
CHAPTER 3. HOMOTOPY
Figure 3.29:
Broadbent & Hammersley 1957; Essam 1972. Old references, but containing the
qualitative aspects here referred to.
12
An intuitive introduction is given in Efros 1986.
105
Figure 3.30: The infinite unfolding of the 2-particle configuration space for
braid statistics.
3.2.2
Covering Spaces
3.2.12 The Riemann surface of a multivalued analytic function14 is a covering space of its analyticity domain, a space on which the function becomes
single-valued. It is the most usual example of a covering space. The considerations of the previous section hint at the interest of covering spaces to
Quantum Mechanics. As said above, if X is the configuration space and
13
14
106
CHAPTER 3. HOMOTOPY
that z simply takes one into another two values which are taken back to a
same value by the inverse. Only by going twice around the loop in C {0}
can we obtain a closed curve in the image space. On the other hand, a loop
not circumventing the zero is taken into two loops in the image space. The
trouble, of course, comes from
Figure 3.31:
the two-valued character of z and the solution is well known: the function
becomes monodromous if, instead of C {0}, we take as definition domain
a space formed by two Riemann sheets with a half-infinite line
(say, E1+
) in
common. This new surface is a covering of C\{0}. The function z is + | z|
15
16
107
[
1
1
1
p (U ) =
(n , n + ).
4
4
n=
108
CHAPTER 3. HOMOTOPY
Figure 3.32:
The open intervals Vn = (n1/4, n+1/4) are (see Figure 3.33) homeomorphically mapped onto U by p. As E1 is simply-connected, it is by definition
the universal covering space of S 1 . Other covering spaces are, for instance,
S 1 itself given as
S 1 = {z C such that |z| = 1}
with the mappings pn : S 1 S 1 , pn (z) = z n , n Z+ .
3.2.18 Consider the torus T 2 = S 1 S 1 . It can be shown that the product
of two covering maps is a covering map. Then, the product
p p : E1 E1 S 1 S 1 , (p p)(x, y) = (p(x), p(y)),
with p the mapping of 3.2.17, is a covering map and E2 , which is simplyconnected, is the universal covering of T 2 .
3.2.19 There are several techniques to calculate the fundamental groups
of topological spaces, all of them rather elaborate. One has been given in
109
Figure 3.33:
Figure 3.34:
section 3.1.3. We shall here describe another, which exploits the universal
covering space. As our aim is only to show the ideas, we shall concentrate in
obtaining the fundamental group of the circle, 1 (S 1 ).
3.2.20 Let p : E X be a covering map. If f is a continuous function of
Y to X, the mapping f : Y E such that p f = f is the lift, or covering,
of f . Pictorially, we say that the diagram 3.34 is commutative.
This is a very important definition. We shall be interested in lifts of two
kinds of mappings: paths and homotopies between paths. In the following,
some necessary results will simply be stated and, when possible, illustrated.
They will be useful in our search for the fundamental group of S 1 .
3.2.21 Let (E, p) be a covering of X and f : I X a path. The lift f is
the path-covering of f. If F : I I X is a homotopy, then the homotopy
F : I I E such that p F = F is the covering homotopy of F .
3.2.22 Take again the covering mapping of 3.2.17, p(x) = (cos 2x,
sin 2x) = ei2x . Then,
110
CHAPTER 3. HOMOTOPY
Figure 3.35:
i) the path f : I S 1 given by f (t) = (cos t, sin t), with initial endpoint
(1, 0), has the lift f(t) = t/2, with initial endpoint 0 and final endpoint 1/2
(Figure 3.35, left);
Figure 3.36:
ii) the path f : I S 1 , given by f (t) = (cos t, sin t), has the lift
f(t) = t/2 (Figure 3.35, right);
iii) the path h(t) = (cos 4t, sin 4t) traverses twice the circle S 1 ; it has the
= 2t (Figure 3.36).
lift h(t)
3.2.23 Let us enunciate some theorems concerning the uniqueness of pathand homotopy-coverings.
Theorem 1: Let (E, p) be the universal covering of X, and f : I X
be a path with initial endpoint x0 . If e0 E is such that p(e0 ) = x0 , then
there is a unique covering path of f beginning at e0 (Figure 3.37).
Theorem 2: Let (E, p) be the universal covering of X, and F : II X
be a homotopy with F (0, 0) = e0 . If e0 E is such that p(e0 ) = x0 , then
there is a unique homotopy-covering F : I I E such that F (0, 0) = e0
(Figure 3.38).
111
Figure 3.37:
Fraleigh 1974.
112
CHAPTER 3. HOMOTOPY
Figure 3.38:
f(2t)
for t [0, 21 ]
n + g(2t 1) for t [ 12 , 1].
begins at 0. It
and suppose that f(1) = n and g(1) = m. By construction, h
is the lift of f g, as the functions sine and cosine have
is easy to see that h
= f g. On the other hand,
periods 2[p(n + t) = p(t)]. Consequently p h
113
114
CHAPTER 3. HOMOTOPY
115
By the way, this manifold is the configuration space for the spherical top,
whose quantization18 is consequently rather involved. In fact, that is why
there are two kinds of wavefunctions: those for which U (2) = U (0) = 1, and
those for which U (2) = - 1. The first type describes systems with integer
angular momentum. Wavefunctions of the second type, called spinors, describe systems with half-integer angular momentum. The universal covering
of SO(3) is the group SU (2) of the unitary complex matrices of determinant
+ 1, whose manifold is the 3-sphere S 3 . Consequently, SO(3) = SU (2)/Z2 .
The group SU (2) stands with respect to SO(3) in a way analogous as the
square-root covering of 3.2.13 stands to C {0}: in order to close a loop
in SU (2), we need to turn twice on SO(3), so that only when = 4 is the
identity recovered. As SU (2) = S 3 is simply-connected, it is the universal
covering of SO(3).
3.2.31 The simplest case of Quantum Mechanics on a multiply-connected
space comes out in the well known Young double-slit interference experiment.
We shall postpone its discussion to 4.2.17 and only state the result. On a
multiply-connected space, the wavefunction behaves as the superposition of
its values on all the leaves of the covering space.
3.3
HIGHER HOMOTOPY
Besides lassoing them with loops, which are 1-dimensional objects, we can
try to capture holes in space with higher-dimensional closed objects. New
groups revealing space properties emerge.
3.3.1 The attentive (and suspicious) reader will have frowned upon the
notation 0 (A) used for the set of path-connected components of space A
in 3.1.4. There, and in 3.1.13, when the fundamental group 1 was
also introduced as the first homotopy group, a whole series of groups n
was announced. Indeed, both 0 and 1 are members of a family of groups
involving classes of loops of dimensions 0, 1, 2, . . . We shall now say a few
words on the higher groups n (X, x0 ), for n = 2, 3, 4, etc.
3.3.2 With loops, we try to detect space defects by lassoing them, by
throwing loops around them. We have seen how it works in the case of
a point extracted from the plane, but the method is clearly inefficient to
apprehend, say, a missing point in E3 . In order to grasp it, we should tend
18
116
CHAPTER 3. HOMOTOPY
Applications of homotopy to defects in a medium are examined in Nash & Sen 1983.
117
F (In ; 0) = f ; F (In ; 1) = g;
F ( In , I) = x0 .
That this homotopy is an equivalence relation can be shown in a way analogous to the 1-dimensional case. The set n (X, x0 ) is consequently decomposed into disjoint subsets, the homotopy classes. The class to which f
belongs will be once again indicated by [f ].
3.3.7 Notice that, in the process of closing the curve to obtain a loop, the
interval I itself becomes equivalent to a loop from the homotopic point
of view, we could take S 1 instead of I with identified endpoints. Actually,
each loop could have been defined as a continuous mapping f : S 1 X. In
the same way, we might consider the n-loops as mappings f : S n X. This
alternative definition will be formalized towards the end of this section.
3.3.8 Let us introduce a certain algebraic structure by defining an operation analogous to that of 3.1.6. Given f and g n (X, x0 ),
h(t1 , t2 , . . . , tn ) =
(f g)(t1 , t2 , . . . , tn ) =
f (2t1 , t2 , . . . , tn )
for t1 [0, 21 ]
g(2t1 1, t2 , . . . , tn ) for t1 [ 12 , 1].
118
CHAPTER 3. HOMOTOPY
119
n 2.
120
CHAPTER 3. HOMOTOPY
3.3.18 A last word on 0 : 0-loops would be mappings from {0} I into X; such
loops can be deformed into each other when their images lay in the same path-component
of X. It is natural to put their classes, which correspond to the components, in the family
{n (X)}.
3.3.19 Hard-sphere gas A gas of impenetrable particles in E3 has, of course, nontrivial 2 . The classical problem of the hard-sphere gas is a good example of the difficulties
appearing in such spaces. After some manipulation, the question reduces to the problem
of calculating the excluded volume21 left by penetrable particles, which is as yet unsolved.
21
Chapter 4
MANIFOLDS & CHARTS
4.1
4.1.1
MANIFOLDS
Topological manifolds
122
ii) the space S has the same dimension n at all points. The number n is then
the dimension of S, n = dim S. The union of a surface and a line in
E3 is not a manifold. This condition can be shown to be a consequence
of the following one, frequently used in its stead:
ii) S is connected. When necessary, a non-connected S can be decomposed
into its connected components. For space-time, for instance, connectedness is supposed to hold because we would have no knowledge of
any disconnected component not our own. Nowadays, with the belief
in the existence of confined quarks and shielded gluons based on an
ever increasing experimental evidence, one should perhaps qualify this
statement.
iii) S has a countable basis (i.e., it is second-countable).
This is a pathology-exorcizing requirement the Sorgenfrey line of
1.2.17, for example, violates it.
iv) S is a Hausdorff space.
Again to avoid pathological behaviours of the types we have talked
about in 1.2.15.
4.1.2 Not all the above conditions are really essential: some authors call
topological manifold any locally-euclidean or locally-half-euclidean connected topological space. In this case, the four conditions above define
a countable Hausdorff topological manifold. We have already said that
Einsteins equations in General Relativity have solutions exhibiting nonHausdorff behaviour in some regions of spacetime, and that some topologies
proposed for Minkowski space are not second-countable. The fundamental
property for all that follows is the local-euclidean character, which will allow
the definition of coordinates and will have the role of a complementarity
principle: in the local limit, the differentiable manifolds whose study is our
main objective will be fairly euclidean. That is why we shall suppose from
now on the knowledge of the usual results of Analysis on En , which will be
progressively adapted to manifolds in what follows.
4.1. MANIFOLDS
123
4.1.2
4.1.3 The reader will have noticed our circumspection concerning the concept of dimension. This seems intuitively a very topological idea, because
it is so fundamental. We have indeed used it as a kind of primitive concept.
Just above, we have taken it for granted in euclidean spaces and defined dimensions of more general spaces in consequence. But only locally-euclidean
spaces have been contemplated. The trouble is that a well established theory for dimension only exists for metric second-countable spaces, of which
euclidean spaces are a particular case. The necessity of a theory to provide a well defined meaning to the concept became evident at the end of
last century, when intuition clearly showed itself a bad guide. Peano found
a continuous surjective mapping from the interval I = [0, 1] into its square
I I, so denying that dimension could be the least number of continuous
real parameters required to describe the space. Cantor exhibited a one-toone correspondence between E1 and E2 , so showing that the plane is not
richer in points than the line, dismissing the idea of dimension as a measure
of the point content of space and even casting some doubt on its topological nature. Mathematicians have since then tried to obtain a consistent and
general definition.
4.1.4 A simplified rendering1 of the topological dimension of a space X is
given by the following series of statements:
i) the empty set , and only , has dimension equal to 1: dim = 1;
ii) dim X n if there is a basis of X whose members have boundaries of
dimension n 1;
iii) dim X = n if dim X n is true, and dim X n 1 is false;
iv) dim X = if dim X n is false for each n.
This definition has at least two good properties: it does give n for the euclidean En and it is monotonous (X Y implies dim X dim Y ). But a
close examination shows that it is rigorous only for separable metric (equivalently, metric second-countable) spaces. If we try to apply it to more general
1
A classic on the subject is Hurewicz & Wallman 1941; it contains a historical introduction and a very commendable study of alternative definitions in the appendix.
124
spaces, we get into trouble: for instance, one should expect that the dimension of a countable space be zero, but this does not happen with the above
definition. Furthermore, there are many distinct definitions which coincide
with that above for separable metric spaces but give different results for less
structured spaces, and none of them is satisfactory in general.2
4.1.5 In another direction, explicitly metric, we may try to define dimension by another procedure: given the space as a subset of some En , we count
the number N () of n-cubes of side necessary to cover it. We then make
smaller and smaller, and calculate the Kolmogorov capacity (or capacity
dimension)
ln N ()
0 ln(1/)
dc = lim
Suppose a piece of line: divide it in k pieces and take = 1/k. Then, the
number of pieces is N () = k, and dc = 1. A region of the plane E2 may
be covered by k 2 = N () squares of side = 1/k, so that dc =2. The
capacity dimension gives the expected results for simple, usual spaces. It is
the simplest case of a whole series of dimension concepts based on ideas of
measure which, unlike the topological dimension, are not necessarily integer.
Consider a most enthralling example: take the Cantor set of 1.2.4. After
the j-th step of its construction, 2j intervals remain, each of length ( 31j ).
Thus, N () = 2j intervals of length = 31j are needed to cover it, and
dc =
ln 2
.6309...!
ln 3
Spaces with fractional dimension seem to peep out everywhere in Nature and
have been christened fractals by their champion, B. B. Mandelbrot.3 Notice
that the fractal character depends on the chosen concept of dimension: the
topological dimension of the Cantor set is zero. Fractals are of great (and
ever growing) importance in dynamical systems.4
4.1.6 Spaces with non-integer dimensions have been introduced in the
thirties (von Neumann algebras, see Math.5.5).
2
4.2
125
4.2.1 Let us go back to item (i) in the definition of a topological manifold: every point p of the manifold has an euclidean open neighbourhood U ,
homeomorphic to an open set in some En , and so to En itself. The homeomorphism
: U open set in En
will give local coordinates around p. The neighbourhood U is called a coordinate neighbourhood of p. The pair (U, ) is a chart, or local system of
coordinates (LSC) around p. To be more specific: consider the manifold En
itself; an open neighbourhood V of a point q En is homeomorphic to another open set of En . Each homeomorphism of this kind will define a system
of coordinate functions, as u in Figure 4.1. For E2 , for instance, we can use
126
127
4.2.6 The fact that a manifold may require more than one chart has a
remarkable consequence. Transformations are frequently treated in two supposedly equivalent ways, the so called active (in which points are moved)
and passive (in which their coordinates are changed) points of view. This
can be done in euclidean spaces, and its generality in Physics comes from the
euclidean supremacy among usual spaces. On general manifolds, only the
active point of view remains satisfactory.
4.2.7 Given any two charts (U, x) and (V, y) with U V 6= , to a given
point p U V will correspond coordinates x = x(p) and y = y(p) (see
Figure 4.2). These coordinates will be related by a homeomorphism between
open sets of En ,
y x<1> : En En ,
which is a coordinate transformation and can be written as
y i = y i (x1 , x2 , . . . , xn ).
(4.1)
xj = xj (y 1 , y 2 , . . . , y n ).
(4.2)
4.2.8 Given two charts (U , ) and (U , ) around a point, the coordinate transformation between them is commonly indicated by a transition
function
g :(U , ) (U , )
1
and its inverse g
.
7
128
4.2.9 Consider now the euclidean coordinate spaces as linear spaces, that
is, considered with their vector structure. Coordinate transformations are
relationships between points in linear spaces. If both x y <1> and y x<1>
are C (that is, differentiable to any order) as functions in En , the two local
systems of coordinates (LSC) are said to be differentially related .
4.2.10 An atlas on the manifold S is a collection of charts {(U , )} such
that
S
U = S.
4.2.11 The following theorem can be proven:
any compact manifold can be covered by a finite atlas,
that is, an atlas with a finite number of charts.
4.2.12 If all the charts are related by linear transformations in their intersections, it will be a linear atlas.
4.2.13 If all the charts are differentially related in their intersections, it
will be a differentiable atlas. This requirement of infinite differentiability can
be reduced to k-differentiability. In this case, the atlas will be a C k -atlas.
Differentiating [4.1] and [4.2] and using the chain rule,
ki =
y i xj
.
xj y k
(4.3)
129
4.2.14 This means that both jacobians are 6= 0. If some atlas exists on S
whose jacobians are all positive, S is orientable . Roughly speaking, it has
two faces. Most commonly found manifolds are orientable. The Mobius strip
and the Klein bottle are examples of non-orientable manifolds.
4.2.15 Suppose a linear atlas is given on S, as well as an extra chart not belonging
to it. Take the intersections of the coordinate-neighbourhood of this chart with all the
coordinate-neighbourhoods of the atlas. If in these intersections all the coordinate transformations from the atlas LSCs to the extra chart are linear, the chart is admissible to
the atlas. If we add to a linear atlas all its admissible charts, we get a (linear) complete
atlas, or (linear) maximal atlas. A topological manifold with a complete linear atlas is
called a piecewise-linear manifold (usually, a PL manifold).
4.2.16 A topological manifold endowed with a certain differentiable atlas
is a differentiable manifold (see Chapter 5 , section 5.1 ). These are the most
important manifolds for Physics and will deserve a lot of attention in the
forthcoming chapters.
4.2.17 Electron diffraction experiment The simplest case of Quantum Mechanics on a multiply connected space appears in the well known
Young double-slit interference experiment, or in the 1927 Davisson & Germer electron diffraction experiment. We suppose the wave function to be
represented by plane waves (corresponding to free particles of momentum
p = mv) incident from the left (say, from an electron source S situated far
enough to the left). A more complete scheme is shown in Figure 4.5), but let
us begin by considering only the central part B of the future doubly-slitted
obstacle (Figure 4.3). We are supposing the scene to be E3 , so that B extends
to infinity in both directions perpendicular to the drawing. Of course, the
simple exclusion represented by B makes the space multiply-connected, actually homeomorphic to E3 \E1 . A manifold being locally euclidean, around
each point there is an open set homeomorphic to an euclidean space. When
the space is not euclidean, it must be somehow divided into intersecting regions, each one euclidean and endowed with a system of coordinates. Here,
it is already impossible to use a unique chart, at least two as in Figure 4.4
being necessary. We now add the parts A and C of the barrier (Figure 4.5).
Diffraction sets up at the slits 1 and 2, which distort the wave. The slits
act as new sources, from which waves arrive at a point P on the screen after
propagating along straight paths 1 and 2 perpendicular to the wavefronts.
If the path lengths are respectively l1 and l2 , the wavefunction along k will
k
. The two waves will have at P a phase difference
get the phase 2l
2|l1 l2 |
2mv|l1 l2 |
=
.
130
Let us recall the usual treatment of the problem8 and learn something from
Quantum Mechanics. At the two slit-sources, we have the same wavefunction
0 . The waves superpose and interfere all over the region at the right, in
particular at P . The waves arrive at P as
1
2l2
1 = 0 exp i 2l
and
exp
i .
2
0
131
132
1 = 0 e
i2
h
R
1
pdq
= 0 e
i2
h
e
1 [mv+ c A]dl
= 0 e
i2
h
R
1
mvdl+ i2
h
e
c
e
c
Adl
Adl
i2
h
R
2
pdq
= 0 e
i2
h
e
2 [mv+ c A]dl
= 0 e
i2
h
R
2
mvdl+ i2
h
Taking a simplified view, analogous to the double slit case, of the kinematic
part, the total phase difference at P will be given by
h
i
H
2|l1 l2 |
i
+ 2e
Adl
hc
2 = 1 e
The closed integral, which is actually a line integral along the curve 2 1 , is
the magnetic flux. The effect, once the kinematical contribution is accounted
for, shows that the vector potential, though not directly observable, has an
observable circulation. The wavefunctions 1 and 2 are values of on
distinct leaves of the covering space, and should be related by a representation
of the fundamental group, which is here 1 = Z. A representation of the
group Z, acting on any (one-dimensional) wavefunction, will be given by any
phase factor like exp[i2], for each real value of , :n exp[i2 n]. The
value of is obtained from above,
H
e
2|
= |l1 l
+ hc
A dl.
Once this is fixed, one may compute the contribution of other leaves, such as
those corresponding to paths going twice around the forbidden region. For
each turn, the first factor will receive the extra contribution of the length of
a circle around B, and the contribution of many-turns paths are negligible
in usual conditions.
Chapter 5
DIFFERENTIABLE
MANIFOLDS
5.1
134
well: it has been found recently, to general surprise, that E4 has infinite
distinct differentiable structures! Another point illustrating the pitfalls of
intuition: not every topological manifold admits of a differentiable structure.
In 1960, Kervaire had already found a 10-dimensional topological manifold
which accepts no complete differentiable atlas at all. In the eighties, a whole
family of non-smoothable compact simply-connected 4-dimensional manifolds was found. And, on the other hand, it has been found that every
non-compact manifold accepts at least one smooth structure.1
5.1.2 The above definitions concerning C -atlases and manifolds can be
extended to C k -atlases and C k -manifolds in an obvious way. It is also possible to relax the Hausdorff conditions, in which case, as we have already
said, the unicity of the solutions of differential equations holds only locally.
Broadly speaking, one could define differentiable manifolds without imposing
the Hausdorff condition, second-countability and the existence of a maximal
atlas. These properties are nevertheless necessary to obtain some very powerful results, in particular the Whitney theorems concerning the imbedding
of a manifold in other manifolds of higher dimension. We shall speak on these
theorems later on, after the notion of imbedding has been made precise.
5.1.3 A very important theorem by Whitney says that a complete C k -atlas
contains a C k+1 -sub-atlas for k 1. Thus, a C 1 -structure contains a C structure. But there is much more: it really contains an analytic sub-atlas.
The meaning here is the following: in the definition of a differentiable atlas,
replace the C -condition by the requirement that the coordinate transformations be analytic. This will define an analytic atlas, and a manifold with an
analytic complete atlas is an analytic manifold . The most important examples of such are the Lie groups. Of course not all C functions are analytic,
as the formal series formed with their derivatives as coefficients may diverge.
5.1.4 General references An excellent introduction is Boothby l975. A
short reference, full of illuminating comments, is the 5-th chapter of Arnold
l973. Nomizu 1956 is a very good introduction to the very special geometrical
properties of Lie groups. The existence of many distinct differentiable structures on E4 was found in l983. It is an intricate subject, the proof requiring
the whole volume of Freed & Uhlenbeck l984. Before proceeding to the main
onslaught, Donaldson & Kronheimer 1991 summarize the main results in a
nearly readable way. According to recent rumors (as of November 1994), a
suggestion of Witten has led people to obtain all the main results in a much
simpler way.
1
Quinn 1982.
5.2
135
SMOOTH FUNCTIONS
5.2.1 In order to avoid constant repetitions when talking about spaces and
their dimensions, we shall as a rule use capital letters for manifolds and the
corresponding small letters for their dimensions: dim N = n, dim M = m,
etc.
5.2.2 A function f : N M is differentiable (or C k , or still smooth) if,
for any two charts (U, x) of N and (V, y) of M , the function
y f x<1> : x(U ) y(V )
is differentiable ( C k ) as a function between euclidean spaces.
5.2.3 Recall that all the analytic notions in euclidean spaces are presupposed. This function y f x<1> , taking an open set of En into an open set
of Em , is the expression of f in local coordinates. We usually write simply
y = f (x), a very concise way of packing together a lot of things. We should
keep in mind the complete meaning of this expression (see Figure 5.1): the
point of N whose coordinates are x = (x1 , x2 , . . . , xn ) in chart (U, x) is taken
by f into the point of M whose coordinates are y = (y 1 , y 2 , . . . , y m ) in chart
(V, y).
5.2.4 The composition of differentiable functions between euclidean spaces
is differentiable. From this, it is not difficult to see that the same is true for
functions between differentiable manifolds, because
136
137
5.2.7 We have seen that two spaces are equivalent from a purely topological point of view when related by a homeomorphism, a topology-preserving
transformation. A similar role is played, for spaces with a differentiable
structure, by a diffeomorphism:
A diffeomorphism is a differentiable homeomorphism
whose inverse is also smooth.
5.2.8 Two smooth manifolds are diffeomorphic when some diffeomorphism
exists between them. In this case, besides being topologically the same, they
have equivalent differentiable structures. The famous result by Milnor cited
in the previous section can be put in the following terms: on the sphere S 7 one
can define 28 distinct smooth structures, building in this way 28 differentiable
manifolds. They are all distinct from each other because no diffeomorphism
exists between them. The same holds for the infinite differentiable manifolds
which can be defined on E4 .
5.2.9 The equivalence relation defined by diffeomorphisms was the starting point of an ambitious program: to find all the equivalence classes of
smooth manifolds. For instance, it is possible to show that the only classes
of 1-dimensional manifolds are two, represented by E1 and S 1 . The complete
classification has also been obtained for two-dimensional manifolds, but not
for 3-dimensional ones, although many partial results have been found. The
program as a whole was shown not to be realizable by Markov, who found
4-dimensional manifolds whose class could not be told by finite calculations.
5.3
DIFFERENTIABLE SUBMANIFOLDS
138
The above f (M ) is a differentiable imbedded submanifold of N . It corresponds precisely to our intuitive idea of submanifold, as it preserves globally
all the differentiable structure.
5.3.4 A weaker kind of inclusion is the following. A smooth function
f : M N is an immersion if, given any point p M , it has a neighbourhood U , with p U M , such that f restricted to U is an imbedding. An
immersion is thus a local imbedding and every imbedding is automatically an
immersion. The set f (M ), when f is an immersion, is an immersed submanifold . Immersions are consequently much less stringent than imbeddings. We
shall later ( 6.4.33 below) give the notion of ntegral submanifold.
5.3.5 These things can be put down in another (equivalent) way. Let us go back to
the local expression of the function f : M N (supposing n m). It is a mapping
between the euclidean spaces Em and En , of the type y f x<1> , to which corresponds
a matrix (y i /xj ). The rank of this matrix is the maximum order of non-vanishing
determinants, or the number of linearly independent rows. It is also (by definition) the
rank of y f x<1> and (once more by definition) the rank of f . Then, f is an immersion
iff its rank is m at each point of M . It is an imbedding if it is an immersion and else an
homeomorphism into f (M ). It can be shown that these definitions are quite equivalent to
those given above.
139
140
fixing the minimum dimension of the imbedding euclidean space of a differentiable manifold. It is a theorem that 2-dimensional orientable surfaces
are imbeddable in E3 : spheres, hyperboloids, toruses are perfect imbedded
submanifolds of our ambient space. On the other hand, it can be shown that
non-orientable surfaces without boundary are not, which accounts for our inability to visualize a Klein bottle. Non-orientable surfaces are, nevertheless,
imbeddable in E4 .
Part II
DIFFERENTIABLE
STRUCTURE
141
Chapter 6
TANGENT STRUCTURE
6.1
INTRODUCTION
144
da(t)
dt
can then be defined at the point A, since this derivative is a constant, and
will be a vector tangent to the curve at A. This high-school view of a
vector is purely euclidean: it suggests that point B lies in the same space
as the curve. Actually, there is no mathematical reason to view the tangent
space as attached to the point on the manifold (as in Figure 6.1(b)). This
is done for purely intuitive, picturesque reasons.
145
6.2
TANGENT SPACES
146
Figure 6.2: A curve maps E1 into N , and a real function proceeds in the
converse way. The definition of vector uses the notion of derivative on E1 .
147
Vp = Vpi i .
(6.2)
x
Notice that the coordinates are particular functions belonging to the algebra R(N ). The basis {/xi } is the natural basis, or coordinate basis
associated to the given LSC, alternatively defined through the conditions
xj
= ij .
i
x
(6.3)
.
xi
(6.4)
6.2.6 The tangent vectors are commonly called simply vectors, or still
contravariant vectors. As it happens to any vector space, the linear mappings
p : Tp N E1 constitute another vector space, denoted here Tp N , the
dual space of Tp N . It is the cotangent space to N at p. Its members are
covectors, or covariant vectors, or still 1-forms. Given an arbitrary basis
148
(6.5)
p = p (ei ) i .
(6.6)
For any p Tp N ,
6.2.7 The dual space is again an n-dimensional vector space. Nevertheless,
its isomorphism to the tangent space is not canonical (i.e, basis independent),
and no internal product is defined on Tp N . An internal product will be
present if a canonical isomorphism exists between a vector space and its
dual, which is the case when a metric is present.
6.2.8 Let f : M N be a C function between the differentiable manifolds M and N . Such a function induces a mapping
f : Tp M Tf (p) N
between tangent spaces (see Figure 6.3). If g is an arbitrary real function on
N , g R(N ), this mapping is defined by
[f (Xp )](g) = Xp (g f )
(6.7)
for every Xp Tp M and all g R(N ). This mapping is a homomorphism of vector spaces. It is called the differential of f , by extension of
the euclidean case: when M = Em and N = En , f is precisely the jacobian matrix. In effect, we can in this case use the identity mappings as
coordinate-homeomorphisms and write p = (x1 , x2 , . . . , xm ) and f (p) = y =
(y 1 , y 2 , . . . , y n ). Take the natural basis,
Xp (g f ) = Xpi
[g(y)]
xi
= Xpi
g y j
y j xi
Thus, the vector f [ (Xp )] is obtained from Xp by the (left-)product with the
jacobian matrix:
y j i
[f (Xp )]j =
X.
(6.8)
xi p
The differential f is also frequently written df . Let us take in the above
definition N = E1 , so that f is a real function on M (see Figure 6.4). As g
in [6.7] is arbitrary, we can take for it the identity mapping. Then,
f (Xp ) = df (Xp ) = Xp (f ).
(6.9)
149
(6.11)
) = ij .
(6.12)
xi
Consequently, the mappings {dxj } form a basis for the cotangent space,
dual to the natural basis {/xi }. Using [6.4], [6.10] and [6.11], one finds
dxj (
df (Xp ) =
f
xi
dxi (Xp ) .
As this is true for any vector Xp , we can write down the operator equation
f
dxi ,
(6.13)
xi
which is the usual expression for the differential of a real function. This is
the reason why the members of the cotangent space are also called 1-forms:
the above equation is the expression of a differential form in the natural
basis {dxj }, so that df Tp M . One should however keep in mind that, on a
general differentiable manifold, the dxj are not simple differentials, but linear
operators: as said by [6.11], they take a member of Tp M into a number.
df =
150
)dxi .
i
x
(6.14)
Figure 6.5: c
151
d
dt t0
Notice that (f x<1> ) is just the local expression of f , with the identity
coordinate mapping in E1 . In the usual notation of differential
calculus, it is
simply written f . In that notation, the tangent vector c dtd t0 is fixed by
[ dtd f c(t)]t0 =
f dcj
|
cj dt t0
f
= cj (t0 ) c
j.
Figure 6.6: Using a chart to show the connection with usual notation.
The tangent vector is in this case the velocity vector of the curve at the
point c(t0 ). Notice the practical way of taking the derivatives: one goes first,
by inserting identity maps like x x<1> , to a local coordinate representation
of the function, and then derive in the usual way.
152
= x(1 x2 )1/2 ,
whose critical points are (0, 1) and (0, 1). All the remaining points are
Figure 6.7: Two regular points corresponding to one regular value by a horizontal projection of the circle.
regular points of f . The points f (x, y) which are images of regular points are
the regular values of f . To the regular value P in Figure 6.7 correspond two
regular points: p and p0 . At p, the jacobian is positive; at p0 , it is negative. If
we go along S 1 starting (say) from (1, 0) and follow the concomitant motion
of the image by f , we see that the different signs of the jacobian indicate the
distinct senses in which P is attained in each point of the inverse image. The
sign of the jacobian at a point p is called the degree of f at p: degp f = +1;
153
154
6.3
TENSORS ON MANIFOLDS
6.3.1 Tensors at a point p on a differentiable manifold are defined as tensors on the tangent space to the manifold at p. Let us begin by recalling
the invariant (that is, basis-independent) definition of a tensor in linear algebra. Take a number r of vector spaces V1 , V2 , . . . , Vr and an extra vector
space W . Take the cartesian product V1 V2 . . . Vr . Then, the set
Lr (V1 , V2 , . . . , Vr ; W ) of all the multilinear mappings of the cartesian product of r-th order into W is itself another vector space.2 The special case
L1 (V ; W ), with W the field over which V is defined, is the dual space V of
V ; for a real vector space, V = L1 (V ; R).
6.3.2 Consider now the cartesian product of a vector space V by itself
s times, and suppose for simplicity V to be a real space. A real covariant
tensor of order s on V is a member of the space L(V V . . . V ; R), that
is, a multilinear mapping
Ts0 : V
{z. . . V} R.
| V
s times
By multilinear we mean, of course, that the mapping is linear on each copy
of V . On the other hand, given two such tensors, say T and S, the linear
(vector) structure of their own space is manifested by
(a T + b S)(v1 , v2 , . . . , vs ) = a T (v1 , v2 , . . . , vs ) + b S(v1 , v2 , . . . , vs ).
6.3.3 The tensor product T S is defined by
T S(v1 , v2 , . . . , vs , vs+1 , vs+2 , . . . vs+q )
= T (v1 , v2 , . . . , vs ) S(vs+1 , vs+2 , . . . vs+q ),
(6.15)
dim V .
155
(6.16)
T0r : V
. . . V } R.
| V {z
r times
The space dual to the dual of a (finite dimensional) vector space is the space
itself3 : (V ) = V . Given a basis {ei } for V , a basis for the contravariant
r-tensors is
{ei1 ei2 ei3 . . . eir }, 1 i1 , i2 , . . . ir
dim V .
156
the T i1 i2 i3 ir being its components in the given basis. Of course, the same
considerations made for covariant tensors and their components hold here.
6.3.6 A mixed tensor , covariant of order s and contravariant of order r, is
a multilinear mapping
Tsr : V
. . . V } R.
{z. . . V} |V V {z
| V
s times
r times
Given a basis {ei } for V and its dual basis {i } for V , a general mixed
tensor will be written
...ir
e ei2 ei3 . . . eir j1 j2 j3 . . . js . (6.17)
T = Tji11ji22ji33 ...j
s i1
It is easily verified that, just like in the cases for vectors and covectors, the
components are the results of applying the tensor on the basis elements:
...ir
= T (i1 , i2 , i3 , . . . , ir , ej1 , ej2 , ej3 , . . . , ejs ).
Tji11ji22ji33 ...j
s
(6.18)
The components in the above basis will be (notice the contracted index j)
n n n ...n
jn
...n
3
i1
i+1
r
Tm11 m22 m
.
3 ...mk1 j mk+1 ...ms
6.3.7 Important special cases of covariant tensors are the symmetric ones,
those satisfying
T (v1 , v2 , . . . , vk , . . . , vj , . . .) = T (v1 , v2 , . . . , vj , . . . , vk , . . .)
157
the summation taking place over all the permutations P of the indices. Notice
that T W Sk+j (V ). With this symmetrizing operation, the linear space
S(V ) becomes a commutative algebra. The same can be made, of course, for
contravariant tensors.
6.3.8 An algebra like those above defined (see 6.3.3 and 6.3.7), which
is a sum of vector spaces, V ==
k=0 Vk , with the binary operation taking
Vi Vj Vi+j , is a graded algebra.
6.3.9 Let {i } be a basis for the dual space V . A mapping p : V E1
defined by first introducing
P V
j1
j2
j3
jk
P = dim
ji =1 Pj1 j2 j3 jk . . . ,
with Pj1 j2 j3 ...jk symmetric in the indices, and then putting
P
j1 j2 j3
jk
p(v) = P (v, v, . . . , v) = dimV
ji =1 Pj1 j2 j3 ...jk v v v . . . v
gives a polynomial in the components of the vector v. The definition is
actually basis-independent, and p is called a polynomial function of degree
k. The space of such functions constitutes a linear space Pk (V ). The sum of
these spaces, P (V ) =
k=0 Pk (V ), is an algebra which is isomorphic to the
algebra S(V) of 6.3.7.
6.3.10 Of special interest are the antisymmetric tensors, which satisfy
T (v1 , v2 , . . . , vk , . . . , vj , . . .) = T (v1 , v2 , . . . , vj , . . . , vk , . . .)
for every pair j, k of indices. Let us examine the case of the antisymmetric
covariant tensors. At a fixed order, they constitute a vector space by themselves. The tensor product of two antisymmetric tensors of order p and q is
158
where the summation takes place on all the permutations P of the numbers (1, 2, . . . , s). Symbol sign P represents the parity of P . The tensor
Alt(T ) is antisymmetric by construction. If n is the number of elementary
transpositions (Mathematical Topic 2) necessary to take (1, 2, . . . , s) into
(p1 , p2 , . . . , ps ), the parity sign P = ()n .
6.3.11 Given two antisymmetric tensors, of order p and of order q,
their exterior product is the (p + q)-antisymmetric tensor given by
=
(p+q)!
p!q!
Alt( ).
This operation does make the set of antisymmetric tensors into an associative
graded algebra, the exterior algebra, or Grassmann algebra. Notice that only
tensors of the same order can be added, so that this algebra includes in reality
all the vector spaces of antisymmetric tensors. We shall here only list some
properties of real tensors which follow from the definition above:
(i)
(ii)
(iii)
(iv)
(v)
( + ) = +
( + ) = +
a ( ) = (a ) = (a ), for any a R;
( ) = )
(6.20)
= ()
In the last property, concerning commutation, and are the orders respectively of and .
If {i } is a basis for the covectors, the space of s-order antisymmetric
tensors has a basis
{i1 i2 i3 . . . is }, 1 i1 , i2 , . . . , is dim V .
An antisymmetric s-tensor can then be written
=
1
s!
i1 i2 i3 ...is i1 i2 i3 . . . is ,
(6.21)
159
the i1 i2 i3 ...is s being the components of in this basis. The space of antisymmetric s-tensors reduces automatically to zero for s > dim V .
Notice further that the dimension of the vector space formed by the antisymmetric covariant s-tensors is dims V . The dimension of the whole Grassmann algebra is 2dim V .
6.3.12 The exterior product is preserved by mappings between manifolds.
Let f : M N be such a mapping and consider the antisymmetric s-tensor
f (p) on the vector space Tf (p) N . The function f determines then a tensor
on Tp M through
(f )p (v1 , v2 , . . . , vs ) = f (p) (f v1 , f v2 , . . . , f vs ).
(6.22)
f
is linear;
f ( ) = f f ;
(f g) = g f .
(6.23)
(6.24)
(6.25)
160
1
s!
The well defined behaviour when mapped between different manifolds renders
the differential forms the most interesting of all tensors. But of course we
shall come to them later on.
6.3.14 Let us now go back to differentiable manifolds. A tensor at a point
p M is a tensor defined on the tangent space Tp M . One
can
choose a jchart
around p and use for Tp M and Tp M the natural bases xi and {dx }. A
general tensor can be written
Tsr =
...ir
Tji11ji22ji33 ...j
s
In another chart, the natural basis will be {/xi } and {dxj }, the same
tensor being written
0
0
0
j10
dxj2 . . . dxj3 dxjs
0
0
0
0 . . . dx
i
i
i
i
x r
x 1
x 2
x 3
0
0
0
i2
i3
i1
0
0
0
0
x
x
xir
xj1 xj2
xjs
i1 i2 i3 ...ir x
= Tj 0 j 0 j 0 ...js0
. . . j1 j2 . . . js
0
0
0
1 2 3
xi0r
x x
x
xi1
xi2
xi3
which gives the transformation of the components under changes of coordinates in the charts intersection. Changes of basis unrelated to coordinate
changes will be examined later on. We find frequently tensors defined by
eq.[6.27]: they are those entities whose components transform in that way.
6.3.15 It should be understood that a tensor is always a tensor with respect to a given group. In [6.27], the group of coordinate transformations
is involved. General basis transformations (section 6.5 below) constitute
another group, and the general tensors above defined are related to that
group. Usual tensors in E3 are actually tensors with respect to the group of
rotations, SO(3). Some confusion may arise because rotations may be represented by coordinate transformations in E3 . But not every transformation
is representable through coordinates, and it is better to keep this in mind.
6.4
6.4.1
161
Fields
6.4.1 Let us begin with an intuitive view of vector fields. In the preceding
sections, vectors and tensors have been defined at a fixed point p of a differentiable manifold M . Although we have been lazily negligent
about this
aspect, the natural bases we have used are actually { xi p }. Suppose now
that we extend these vectors throughout the whole charts coordinate neighbourhood, and that the components are differentiable functions f i : M R,
f i (p) = Xpi . New vectors are then obtained, tangent to M at other points
of the coordinate neighbourhood. Through changes of charts, vectors can
eventually be got all over the manifold. Now, consider a fixed vector at p,
tangent to some smooth curve: it can be continued in the above way along
the curve. This set of vectors, continuously and differentiably related along
a differentiable curve, is a vector field. At p, Xp : R(M ) R. At different
points, X will map R(M ) into different points of R, that is, a vector field is
a mapping X : R(M ) R(M ). In this way, generalizing that of a vector,
one gets the formal definition of a vector field:
a vector field X on a smooth manifold M is a linear mapping
X : R(M ) R(M ) obeying the Leibniz rule:
X(f g) = f X(g) + g X(f ), f, g R(M ).
6.4.2 The tangent bundle A vector field is so a differentiable choice of
a member of Tp M for each p of M . It can also be seen as a mapping from M
into the set of all the vectors on M , the union T M = pM Tp M , with the
proviso that p is taken into Tp M :
X : M T M,
X : p Xp Tp M.
(6.28)
162
= 1 (V )
x(V ) Em ,
x : Xp (x1 (p), x2 (p), . . . , xm (p), Xp1 , Xp2 , . . . , Xpm ),
(6.29)
(6.30)
(6.31)
Xpi
xi
.
p
, y) on M , with W
V 6= , the mapping y x<1> ,
Given another chart (W
defined by
y x<1> (x1 (p), x2 (p), . . . , xm (p), Xp1 , Xp2 , . . . , Xpm )
= (y 1 x<1> (x1 , x2 , . . . , xm ), y 2 x<1> (x1 , x2 , . . . , xm ), . . .
. . . y m x<1> (x1 , x2 , . . . , xm ), Yp1 , Yp2 , . . . , Ypm ),
where (see 6.2.8) YpI = (jacobian of y x<1> )ij Xpj is differentiable. The
two charts are, in this way, differentiably related. A complete atlas can in
this way be defined on T M , making it into a differentiable manifold. This
differentiable manifold T M is the tangent bundle, the simplest example of
a differentiable fiber bundle, or bundle space. The tangent space to a point
p, Tp M , is called, in the bundle language, the fiber on p. The field X itself,
defined by eq.[6.28], is a section of the bundle. Notice that the bundle space
in reality depends on the projection for the definition of its topology.
6.4.4 The commutator Take the field X, given in some coordinate neighbourhood as X = X i x i . As X(f ) R(M ), one could consider the action of
another field Y = Y j x j on X(f ):
2
i f
j i f
(X
)
+
Y
X
.
xj
xi
xj xi
This expression tells us that the operator Y X, defined by
Y Xf = Y j
(Y X)f = Y (Xf ),
does not belong to the tangent space, due to the presence of the last term.
This annoying term is symmetric in XY , and would disappear under antisymmetrization. Indeed, as easily verified, the commutator of two fields
j
j
i Y
i X
[X, Y ] := (XY Y X) = X
Y
(6.32)
i
i
x
x
xj
does belong to the tangent space and is another vector field.
163
6.4.5 ... and its algebra The operation of commutation defines on the
space T M a structure of linear algebra. It is also easy to check that
[X, X] = 0,
[[X, Y ], Z] + [[Z, X], Y ] + [[Y, Z], X] = 0,
the latter being the Jacobi identity. An algebra satisfying these two conditions is a Lie algebra. Thus, the vector fields on a manifold constitute, with
the commutation, a Lie algebra.
6.4.6 Notice that a diffeomorphism f preserves the commutator:
f [X, Y ] = [f X, f Y ].
Furthermore, given two diffeomorphisms f and g, (f g) X = f g X.
6.4.7 The cotangent bundle Analogous definitions lead to general tensor bundles. In particular, consider the union
T M = pM Tp M.
A covariant vector field, cofield or 1-form is a mapping
: M T M
such that (p) = p T M, p M .
6.4.8 This corresponds, in just the same way as has been seen for the
vectors, to a differentiable choice of a covector on each p M . In general,
the action of a form on a vector field X is denoted
(X) = < , X >,
so that
: T M R(M ).
(6.33)
(6.34)
(6.35)
(6.36)
164
6.4.10 The cofield bundle above defined is the cotangent bundle, or the
bundle of forms.
We shall see later (chapter 7) that not every 1-form is the differential of
a function. Those who are differentials of functions are the exact forms.
6.4.11 We have obtained vector and covector fields. An analogous procedure leads to tensor fields. We first consider the tensor algebra over a point
p M , consider their union for all p, topologize and smoothen the resultant
set, then define a general tensor field as a section of this tensor bundle.
6.4.12 At each p M , we have two m-dimensional vector spaces, Tp M
and Tp M , of course isomorphic. Nevertheless, their isomorphism is not natural (or canonical ). It depends on the chosen basis. Different basis fix
isomorphisms taking the same vector into different covectors. Only the presence of an internal product on Tp M (due for instance to the presence of a
metric, a case which will be seen later) can turn the isomorphism into a
natural one. By now, it is important to keep in mind the total distinction
between vectors and covectors.
6.4.13 Think of En as a vector space: its tangent vector bundle is a Cartesian product. A tangent vector to En at a point p can be completely specified
by a pair (p, V ), where also V is a vector in En . This comes from the isomorphism between each Tp En and En itself. Forcing a bit upon the trivial,
we say that En is parallelizable, the same vector V being defined at all the
different points of the manifold En . Given a general manifold M , it is said
to be parallelizable if its tangent bundle is trivial, that is, a mere Cartesian
product T M = M Em . In this case, a vector field V can be globally (that
is, everywhere on M ) given by (p, V ). Recalling the definition of a vector
field as a section on the tangent bundle, this means that there exists a global
section. Actually, the existence of a global section implies the triviality of the
bundle. This holds for any bundle: if some global section exists, the bundle
is a Cartesian product.
All toruses are parallelizable. Of all the spheres S n , only S 1 , S 3 and
7
S are parallelizable. The sphere S 2 is not a result sometimes called the
hedgehog theorem: you cannot comb a hairy hedgehog so that all its prickles
stay flat. There will be always at least one point like the crown of the head.
The simplest way to find out whether M is parallelizable or not is based on
the simple idea that follows: consider a vector V 6= 0. Then, the vector field
(, V ) will not vanish at any point of M . Suppose that we are able to show
that no vector field on M is everywhere nonvanishing. This would imply that
T M is not trivial. A necessary condition for parallelizability is the vanishing
165
A detailed treatment of the subject, with plenty of examples, is given in the little
masterpiece Arnold 1973.
166
6.4.15 Dynamical systems: maps Dynamical systems are also described, mainly in model building, by iterating maps like
xn+1 = f (xn ) ,
where x is a vector describing the state of some system. To help visualization,
we may consider n as a discrete time. The state at the n-th stage is given
by a function of the (n 1)-th stage, and so by its n-th iterate applied on
the initial seed state x0 . The set of points {xn } by which the system evolves
is the orbit of the system. An important concept in both the flow and the
map pictures is the following: suppose there is a compact set A to which the
sequence xn (or, in the flow case, the state when t becomes larger and larger)
converges for a given subset of the set of initial conditions. It may consist of
one, many or infinite points and is called an attractor . It may also happen
that A is a fractal, in which case it is a strange (or chaotic) attractor.5 This
is the case of the simple mapping
f : I I, I = [0, 1], f (x) = 4x(1 x),
popularly known as the logistic map, which for certain values of I
tends to a strange attractor akin to a Cantor set. Strange attractors are
fundamental in the recent developments in the study of chaotic behaviour in
non-linear dynamics.6
6.4.16 Let us go back to the beginning of this chapter, where a vector at
p M was defined as the tangent to a curve a(t) onM , with a(0) = p. It is
interesting to associate a vector to each point of the curve by liberating the
variation of the parameter t in eq.[6.1]:
Xa(t) (f ) =
d
(f a)(t).
dt
(6.37)
Then, Xa(t) is the tangent field to a(t), and a(t) is the integral curve of
X through p. In general, this is possible only locally, in a neighbourhood
of p. When X is tangent to a curve globally, the above definition being
extendable to the whole M , X is said to be a complete field . Let us for the
sake of simplicity take a neighbourhood U of p and suppose a(t) U , with
coordinates (a1 (t), a2 (t), . . . , am (t)). Then, from [6.37],
Xa(t) =
5
6
dai
.
dt ai
(6.38)
See Farmer, Ott & Yorke 1983, where a good discussion of dimensions is also given.
For a short review, see Grebogi, Ott & Yorke 1987.
167
Thus,
dai
(6.39)
dt
i
. In this sense, the field whose integral curve is a(t)
is the component Xa(t)
da
is given by dt . In particular, Xp = da
. Conversely, if a field is given by
dt t=0
k 1
2
m
its components X (x (t), x (t), . . . , x (t)) in some natural basis, its integral
curve x(t) is obtained by solving the system of differential equations X k =
dxk
. The existence and unicity of solutions for such systems is in general
dt
valid only locally.
Xa(t) (ai ) =
6.4.2
Transformations
168
(q0 , p0 ) (qt , pt ) .
Given a domain U M , the Liouville theorem says that the above flow preserves
its volume: vol [(t)U ] = vol [(0)U ]. Suppose now that M itself has a finite volume.
Then, after a large enough time interval, forcibly ((t)U ) U 6= . In words: given any
neighbourhood U of a state point (q, p), it contains at least one point which comes back
to U for t > some tr . For large enough periods of time, a system comes back as near as
one may wish to its initial state. This is Poincares theor`eme du retour.
t0
1
[f (p (t)) f (p)] ,
t
(6.40)
169
d
dt
!
t0
d
f=
dt
(f p )
t0
1
= lim
[f p (t0 + t) f p (t0 )]
t0
t
1
= lim
{f [(q + t)] f (q)}
t0 t
1
{f [q (t)] f (q)} = Xp (t0 ) f.
lim
t0 t
As f is any element of R(M ), we have indeed
p (t0 ) = Xp (t0 ) .
(6.41)
The above definition generalizes to manifolds, though only locally, the well
known case of matrix transformations engendered by an invertible matrix
g(t) = etX , of which the matrix X is the generator,
X = dg
= XetX |t=0 .
dt t=0
A matrix Y will transform according to
Y 0 = g(t)Y g 1 (t) = etX Y etX (1 + tX)Y (1 tX) Y + t[X, Y ]
to first order in t, and we find that the first derivative is
[X, Y ] = limt0
1
t
{g(t) Y g 1 (t) Y } .
170
more general objects. We face, to begin with, the problem of finding the
variation rate of a vector field Y at p M with respect to Xp . This can be
done by using the fact that X generates locally a 1-parameter group, which
induces an isomorphism t : Tp M Tt (p) M , as well as its inverse t .
It becomes then possible to compare values of vector fields. We shall just
state three different definitions, which can be shown to be equivalent. The
Lie derivative of a vector field Y on M with respect to the vector field X on
M , at a point p M , is given by any of the three expressions:
(LX Y )p = limt0 1t t (Y(t,p) ) Yp
1
Yp t (Y(t,p) )
t0 t
= dtd {t (Y )} t=0 .
= lim
(6.42)
Each expression is more convenient for a different purpose. Notice that the
vector character of Y is preserved by the Lie derivative: LX Y is a vector
field.
Let us examine the definition given in the first equality of eqs.[6.42] (see
Figure 6.10). The action p (t) induces an isomorphism between the tangent
spaces Tp M and Tq M , with q = p (t). By this isomorphism, Yp is taken into
p (t) (Yp ), which is in general different from Yq , the value of Y at q. By
using the inverse isomorphism p (t) we bring Yq back to Tp M . In this last
vector space we compare and take the limit. As it might be expected, the
same definition can also be shown to reduce to
LX Y = [X, Y ].
One should observe that the concept of Lie derivative does not require any
extra structure on the differentiable manifold M . Given the differentiable
structure, Lie derivatives are automatically present.
6.4.26 Let us consider, as shown in Figure 6.11, a little planar model for
the water flow in a river of breadth 2a: take the velocity field X = (a2 y 2 )e1 ,
and e2 = y
. It generates the 1-parameter group
with e1 = x
X (t, p) = (a2 y 2 )te1 + p,
or
X (t, p) = [x + (a2 y 2 )t]e1 + ye2 ,
with p = (x, y). The flow leaves the border points invariant. Consider now a
constant transversal field, Y = e2 . It generates the group
171
= 2yt x
+
f = t [X, Y ]f + Y f,
so that
1t {X (t, p)(Yp ) Yp } (f ) = [X, Y ]f ,
which is the expression of the third definition in [6.42]. Thus, the Lie derivative turns up when we try to find out how a field Y experiences a small
transformation generated by another field X.
172
6.4.27 Consider on the plane (Figure 6.12) the two fields X = e1 (con
(6.43)
(6.44)
173
},
xi
(6.45)
they satisfy
Lj (Y ) =
Y i
;
xj xi
(6.46)
(6.47)
(6.48)
Thus,
(LX )Y = X((Y )) ([X, Y ]).
6.4.30 This comes out as a consequence of the general definition for the
Lie derivative of a tensor field of any kind, which is (cf. eq.[6.42])
(LX T )p == lim
t0
1
Tp t (T(t,p) ) .
t
(6.49)
174
Once the action of the Lie derivative is known on fields and cofields, the
general definition is given as
(LX T )(Y1 , Y2 , . . . Ys , 1 , 2 , . . . r )
= LX [T (Y1 , Y2 , . . . Ys , 1 , 2 , . . . r )]
T (LX Y1 , Y2 , . . . , Ys , 1 , 2 , . . . r )
T (Y1 , LX Y2 , . . . Ys , 1 , 2 , . . . r ) . . .
T (Y1 , Y2 , . . . , LX Ys , 1 , 2 , . . . r )
T (Y1 , Y2 , . . . , Ys , LX 1 , 2 , . . . r )
T (Y1 , Y2 , . . . Ys , 1 , LX 2 , . . . r ) . . .
T (Y1 , Y2 , . . . , Ys , 1 , 2 , . . . , LX r ). (6.50)
6.4.31 Notice that LX preserves the tensor character: it takes an rs
tensor into another tensor of the same type. In terms of the components: in
a natural basis {i } the components of LX T are
ab...r
a
ib...r
b
ai...r
(LX T )ab...r
ef ...s = X(Tef ...s ) (i X )Tef ...s (i X )Tef ...s . . .
ab...i
i
ab...r
i
ab...r
i
ab...r
(i X r )Tef
...s + (e X )Tif ...s ) + (f X )Tei...s ) . . . + (s X )Tei...i ). (6.51)
6.4.32 The Lie derivative LX provides the infinitesimal changes of tensorial objects under the 1-parameter group of transformations of which X is the
generator. For this reason, Lie derivatives are basic instruments in the study
of transformations imposed on differentiable manifolds (automorphisms) and,
a fortiori, in the study of symmetries (see section 8.2).
6.5. FRAMES
175
6.5
FRAMES
176
dr
;
dt
V =
d
;
dt
V =
d
.
dt
dr
;
dt
V = r d
;
dt
V = r sin d
,
dt
;
r
X =
1
;
r
X =
.
r sin
(6.52)
1
r sin
6.5. FRAMES
177
and their permutations in the lower indices (in which the coefficients are
clearly antisymmetric). Notice: the coefficients are not necessarily constant
and depend on the chosen basis. Clearly, a necessary condition for the basis
to be holonomic is that Cijk = 0 for all commutators of the basis members.
This condition, Cijk = 0 for all basis members, may be shown to be also
sufficient for holonomy. The Jacobi identity, required by the Lie algebra,
implies
C n kl C i jn + C n jk C i ln + C n lj C i kn = 0.
(6.54)
6.5.5 Let us re-examine the question of frame transformations. Given two
natural basis on the intersection of two charts, a field X will be written
0
X = X i x i = X i
.
xi0
xj i
X =
X.
xi
j0
(6.55)
This expression gives the way in which field components in natural bases
change when these bases are themselves changed. Here, basis transformations
are intimately related to coordinate transformations. However, other basis
transformations are possible: for example, going from the holonomic basis
(r , , ) to the basis (Xr , X , X ) of eq.[6.52] in the spherical case above
is a basis transformation unrelated to a change of coordinates.
6.5.6 Given an anholonomous basis {Xi }, it will always be possible to
write locally each one of its members in some coordinate basis as
Xi = Xij
.
xj
xj
,
y i
where {y i } is
xj
(6.57)
178
(6.60)
so that the group of transformations acts on the left on the 1-form basis. Dual
basis transform inversely to each other, so that, under the action, the value
< , X > is invariant. That is to say that < , X > is basis-independent.
6.5.8 The bundle of linear frames Let Bp M be the set of all linear basis
for Tp M . As we have said, it is a vector space and a group, just GL(m, R).
In a way similar to that used to build up T M as a manifold, the set
6.5. FRAMES
179
BM = pM Bp M
180
of frame,
the set of coordinates will transform according to x0 = Ax, or xj
j0 i
= Ai x , as any contravariant vector. This leads to
0
dAji
dt
u i + Aj
dui
.
dt
A velocity, for example, as seen from two frames, will have its components
related by
0
dAji i
j0 i
j0
v.
v = Ai v +
dt
Of course, we are here supposing that also the frames are in relative motion.
We shall come back to such moving frames later ( 7.2.17 , 7.3.12 and
9.3.6).
6.6
181
is some compressible fluid, with temperature gradients and all which is necessary to render point-dependent the derivative of the pressure with respect
to the
density at fixed entropy. The sound velocity will be given by
fluid
p
and the sound propagation will be governed by geodesics of still
c2s =
S
another metric, gij00 = c12 ij . Nevertheless, in both cases we use also the euS
clidean metric to make measurements, and much of geometrical optics and
acoustics comes from comparing the results in both metrics involved. This is
only to call attention to the fact that there is no such a thing like the metric
of a space. It happens frequently that more than one is important in a given
situation (for an example in elasticity, see Physical Topic 3, section 3.3.2).
Let us approach the subject a little more formally.
6.6.1 In the space of differential forms, a basis dual to the basis {Xi } for
fields in T M is given by those j such that
j (Xi ) = < j , Xi > = j i,
(6.61)
(6.62)
(6.64)
(6.65)
182
(ii) it is symmetric:
X Y = Y X;
(iii) it is non-singular :
if X Y = 0 for every field Y , then X = 0.
6.6.4 In the basis introduced in 6.6.2, we have
g(Xi , Xj ) = Xi Xj = gmn m (Xi ) n (Xj ),
so that
gij = g(Xi , Xj ) = Xi Xj .
(6.66)
The relationship between metrics and general frames (in particular, tetrads)
will be seen in 9.3.6. As gij = gji and we commonly write simply i j for
the symmetric part of the bilinear basis, then
i j = (i j) =
1
2
( i j + j i ) ,
we have
g = gij i j
(6.67)
(6.68)
6.6.5 This is the usual notation for a metric. Notice also the useful symmetrizing notation (ij) for indices. All indices (ijk . . .) inside the parenthesis
are to be symmetrized. For antisymmetrization the usual notation is [ijk . . .],
meaning that all the indices inside the brackets are to be antisymmetrized.
Knowledge of the diagonal terms is enough: the off-diagonal may be obtained
by polarization, that is, by using the identity
g(X, Y ) = 12 [g(X + Y, X + Y ) g(X, X) g(Y, Y )].
6.6.6 A metric establishes a relation between vector and covector fields:
Y is said to be the contravariant image of a form z if, for every X,
g(X, Y ) = z(X).
183
(6.69)
(6.70)
x2
a2
y2
.
b2
a2
x
b2
y
y .
184
2x
a2
dx +
2y
b2
dy
some local solution f = T df (see 7.2.12 for a systematic method to get it);
this solution can be extended to the whole space, giving the whole ellipse.
This is a special case, as of course not every field or cofield is integrable. In
most cases they are only locally integrable, or nonintegrable at all.
Suppose now that a metric gij is present, which relates fields and cofields.
In the case above gij = diag(1/a2 , 1/b2 ) is of evident interest, as f (v) =
g(v, v), with v the position vector (x, y). To the vector v of components (xj )
will correspond the covector of components (pk = gkj xj ) and the action of
this covector on v will give simply p(v) = pk xk = gij xi xj . As we are also in a
euclidean space, the euclidean metric mij = ij may be used to help intuition.
We may consider p and v as two euclidean vectors of components (pk ) and
(xk ). Comparison of the two metrics is made by using g(v, v) = m(p, v).
Consider the curve p(v) = g(v, v) = m(p, v) = C, which is an ellipse. The
vector v gives a point on the ellipse and the covector p, now assimilated to an
euclidean vector, is orthogonal to the curve at each point, or to its tangent
at the point. This construction, allowing one to relate a 1-form to a field
in the presence of a non-trivial metric, is very much used in Physics. For
rigid bodies, the metric m is the inertia tensor,, the vector v is the angular
velocity and its covector is the angular momentum. The ellipsoid is the
inertia ellipsoid, the whole construction going under the name of Poinsot
(more details can be found in Physical Topic 2, section 2.3.10). In crystal
optics, the Fresnel ellipsoid ij xi xj = C regulates the relationship between
the electric field E and the electric displacement D = E, where the metric is
the electric permeability (or dielectric) tensor. In this case, another ellipsoid
is important, given by the inverse metric 1 : it is the index, or Fletchers
ellipsoid (Physical Topic 5.6). In all the cases, the ellipsoid is defined by
equating some hamiltonian to a constant.
6.6.8 An important property of a space V endowed with an internal product is the following: given any linear function f R(V ), there is a unique
vf V such that, for every u V , f (u) = (u, vf ). So, the forms include all
the real linear functions on Tp M (which is expected, they constituting its dual
space), and the vectors include all the real linear functions on Tp M (equally
not unexpected, the dual of the dual being the space itself). The presence of
a metric establishes a natural (or canonical) isomorphism between a vector
space (here, Tp M ) and its dual.
6.6.9 The above definition has used fixed bases. As in general no base covers the whole manifold, convenient transformations are to be performed in the
intersections of the definition domains of every pair of bases. If some of the
185
(6.72)
186
(6.74)
whose meaning will become clear later ( 9.4.23), then the Lie derivative
acquires the form
(LX g) = X X + X X .
Introducing the covariant derivative
X; = X X ,
(6.75)
(LX g) = X; + X; .
(6.76)
it can be written as
The condition for isometry, LX g = 0, then becomes
X; + X; = 0,
(6.77)
187
For applications in gravitation and cosmology, see Weinberg 1972, chap. 13.
188
12
Chapter 7
DIFFERENTIAL FORMS
7.1
INTRODUCTION
T dxdydz.
189
190
to eliminate redundancy, but there is a deeper reason for that: integrals are
invariant under basis transformations and the corresponding jacobian determinants are already included in the integration measures, which are henceby
antisymmetric. We could almost say that, as soon as one thinks of integration, only antisymmetric objects are of interest. This is a bit too strong as,
for instance, a metric may be involved in the integration measure. However,
differential calculus at least is basically concerned with antisymmetric objects
with a well defined behaviour under transformations, that is, antisymmetric
tensors.
7.1.2 In the case of 1-forms (frequently called Pfaffian forms), of course,
antisymmetry is of no interest. We have seen, however, that they provide
basis for higher-order forms, obtained by exterior product ( 6.3.11). Recall
that the exterior product of two 1-forms (say, two members of a basis { i })
is an antisymmetric mapping
: T10 (M ) T10 (M ) T20 (M ),
where Tsr (M ) is the space of (r, s)-tensors on M . In the basis formed in this
way, a 2-form F , for instance, will be written
F = 12 Fij i j .
7.1.3 We shall denote k (M ) the space of the antisymmetric covariant
tensors of order k on the space M , henceforth simply called k-forms. Recall
that they are tensor fields, so that in reality the space of the k-forms on the
manifold M is the union
S
k (M ) = pM k (Tp M ).
In a way quite similar to the previously defined bundles, the above space can
be topologized and made into another fiber bundle, the bundle of kforms
on M . A particular k-form is then a section
: M k (M )
: p p k (Tp M ).
It is a universally accepted abuse of language to call k-forms differential
forms of order k.
We say abuse, of course, because not every differential form is the
differential of something else.
7.1.4 The exterior product also called wedge product is the generalization of the vector product in E3 to spaces of any dimension and thus,
through their tangent spaces, to general manifolds. It is a mapping
7.1. INTRODUCTION
191
: p (M ) q (M ) p+q (M ),
which makes the whole space of forms into a graded associative algebra.
Recall that dim p (M ) = mp , and the spaces of order p > m reduce to
zero. Thus, if p is a p-form and q is a q-form, p q = 0 whenever
p + q > m. The space of 0-forms has as elements the real functions on M
whose compositions, by the way, exhibit trivially the pull-back property.
7.1.5 A basis for the maximal-order space m (M ) is a single m-form
1 2 3 . . . m.
In other words, m (M ) is a 1-dimensional space. The nonvanishing elements
of m (M ) are called volume elements, or volume forms. Two volume elements v1 and v2 are said to be equivalent if a number c > 0 exists such that
v1 = cv2 . This equivalence divides the volume forms into two classes, each
one called an orientation. We shall come back to volume forms later.
This definition of orientation can be shown to be equivalent to that given
in 4.2.14.
Some nave considerations in euclidean spaces provide a more pictorial
view of Pfaffian forms. Let us proceed to them.
7.1.6 Perhaps the most elementary and best known 1-form in Physics is
the mechanical work, a Pfaffian form in E3 . In a natural basis, it is written
W = Fk dxk ,
with the components Fk representing the force. The total work realized in
taking a particle from a point a to point b along a line is
R
Wab [] = Fk dxk ,
and in general depends on the chosen line. It will be path-independent only
when the force comes from a potential as a gradient Fk = (gradU )k . In
this case W is W = dU , truly the differential of a function, and
Wab = U (a) U (b).
A much used criterion for this integrability is to see whether W [] = 0
when is any closed curve. However, the work related to displacements in
a non-potential force field is a typical non-differential 1-form: its integral
around a closed curve does not vanish, and its integral between two points
will depend on the path. Thus, the simplest example of a form which is not
a differential is the mechanical work of a non-potential force. We shall later
find another simple example, the heat exchange ( 7.2.10).
192
7.1. INTRODUCTION
193
194
Figure 7.2: (a) The field dr; (b) The field dr: the moduli change from
point to point, but the directions remain radial.
k k k k
j11j22j33jpp
7.1. INTRODUCTION
195
Figure 7.3: (a) The line field dy; (b) The line field dy; (c) The line field
dx.
k k k k
j11j22j33jpp
jk11
jk21
...
jkp1
jk12
jk22
...
jkp2
...
...
...
...
j1p
k
j2p
...
k
jpp
(7.1)
q n1 n2 n3 np
2 k3 kq n1 n2 n3 np
= q! km11km
.
j11j22j33jq q jm1 1j2mj32j
m3 mq+p
2 m3 mq+p
(7.2)
(7.3)
j1 j2 j3 jm j1 j2 j3 jm = m!
(7.4)
When p = m = dim M ,
(7.5)
(7.6)
j1 jn i1 ...in A
i1 j1
i2 j2
...A
in jn
= n!detA.
(7.7)
Notice that we are using here upper indices only for notational convenience.
We shall later meet Kronecker symbols of type [7.3] with indices raised by
the action of a metric.
196
Figure 7.4: (a) The line field w = bdy adx; (b) The line field w = 2xdx +
dy.
7.1.11 Kronecker symbols are instrumental in calculations involving components of forms. Given a p-form
=
1
p!
j1 j2 jp j1 j2 . . . jp ,
(7.8)
1
p!
k k k k
j11j22j33jpp k1 k2 k3 kp .
(7.9)
(7.10)
7.1.12 Given the p-form and the q-form , the components of the wedge
product are, in terms of the components of and ,
( )i1 i2 ip+q =
1
p!
k k k k j j j jp
q 1 2 3
3
i11i22i3
iq+p
k1 k2 k3 kq j1 j2 j3 jp .
(7.11)
(7.12)
without the factor 1/p! but respecting j1 < j2 < j3 < . . . < jp in the
summation. We shall use one or another of the conventions, according to
convenience.
197
7.1.14 The main properties of the exterior product have been outlined in
eqs.[6.20]. Let us only restate the rule concerning commutation:
p q = ()pq q p
(7.13)
(7.14)
f = f = f.
(7.15)
Of course,
Given any p-form , we define the operation of exterior product by a 1-form
through
() : p (M ) p+1 (M )
() = ,
p < m.
(7.16)
7.1.15 It is easy to check that the vanishing of the wedge product of two
Pfaffian forms is a necessary and sufficient condition for their being linearly
dependent.
7.2
EXTERIOR DERIVATIVE
f
f
dxi =
dxi ,
i
i
x
x
(7.17)
198
i
xi
dxi .
dx
dx
=
199
(7.18)
(7.19)
The computation is done by repeated use of properties (iii) and (ii) of 7.2.1.
One sees immediately the great advantage of using the natural basis i = dxi .
In this case, from property (iv), only one term remains:
dF = 21 dFij dxi dxj .
The component is a function, so its differential is just as given by eq.[7.17]:
dF =
1
2
Fij
dxk dxi dxj .
xk
(7.20)
We would like to have this 3-form put into the canonical form [7.8], with the
components fully symmetrized. If we antisymmetrize now in pairs of indices
(k, i) and (k, j), we in reality get 3 equal terms,
dF =
1
3!
1
q!
200
1
q!
1
(q+1)!
1
q!
1
q!
j j j ...j
i00i11i22...iqq
j1 j2 ...jq
.
x0
(7.23)
=
x0
1
q!
j1 j2 ...jq
dxj1 dxj2 . . . dxjq
xj0
(7.24)
so that
.
(7.25)
x0
The expression for the exterior derivative in an arbitrary basis will be found
below (see eq.[7.74]). We shall see later (eq.[7.161) the real meaning of
eq.[7.24], and give, in consequence, still another closed expression for d.
d = dxj0
i=0
k
X
i+j
()
b
b
[Xi , Xj ], X0 , X1 , . . . , Xi , . . . , Xj , . . . , Xk , (7.26)
i<j
=
;
e
=
=
;
e
=
=
.
2
3
x1
x
x2
y
x3
z
201
Its dual is {dx, dy, dz}. The euclidean metric is, in the related cartesian
coordinates,
g = ij dxi dxj .
Given a vector V = V i x i , its covariant image Z is a form such that, for any
vector U ,
Z(U ) = Zi U i = g(U, V ) = ij V j U i ,
so that Zi = ij Vj . One uses the same names for a vector and for its covariant
image, writing Zi = Vi . So, to the vector V corresponds the form V = Vi dxi .
Its differential is
dV =
Vi k
dx dxi = 12 (k Vi i Vk ) dxk dxi ,
xk
or
dV = 21 (rotV )ki dxk dxi .
Think of electromagnetism in E3 : the vector potential is the 1-form A =
Ai dxi , and the magnetic field is H = dA = rotA.
The derivative of a 0-form f is
df = (gradf )i dxi .
Suppose the form V above to be just this gradient form. Then,
d2 f = rot gradf
and the Poincare lemma is here the well known property rot grad f 0.
When the vector potential is the gradient of a function, A = df , the magnetic
field vanishes:
H = d2 f 0.
Consider now the second order tensor
T = 21 Tij dxi dxj .
In E3 , to this tensor will correspond a unique vector (or 1-form) U , fixed by
Tij = ijk Uk . The differential is
dT = 21 k Tij dxk dxi dxj
= 21 ijk k Uk dxk dxi dxj = (div U ) dx1 dx2 dx3 .
Taking Ui = 21 ijk (rotV )jk , the Poincare lemma assumes still another well
known avatar, namely, div rot V 0. The expression for the laplacian of a
0-form f , div gradf = i i f , is easily obtained.
202
A criterion to see the difference between a true vector and a second order tensor is the
behaviour under parity (xi xi ) transformation. A true vector changes sign, while a
second order tensor does not. The magnetic field is such a tensor, and Maxwells equation
divH = 0 is dH = 0, actually the identity d2 A 0 if H = dA.
0
H3 H2 E1
H3
0
H1 E2
[F ] =
(7.27)
H2 H1
0
E3
E1 E2 E3 0
The fourth row and column in the matrix correspond to the zeroth, or time
components. The field strength can be written as a 2-form
F = 21 F dx dx ,
(7.28)
with , = 1, 2, 3, 0. In detail,
F = H1 dx2 dx3 + H2 dx3 dx1 + H3 dx1 dx2
+ E1 dx1 dx0 + E2 dx2 dx0 + E3 dx3 dx0 ,
or
F = 21 ijk Hi dxj dxk + Ej dxj dx0 .
(7.29)
From [7.21],
dF =
1
3!
{ F + F + F } dx dx dx .
(7.30)
From [7.29],
~ Hdx
~ 1 dx2 dx3 +
dF =
+
H3
x0
E2
x1
E1
x2
H
x0
E2
x3
H2
x0
E3
x2
E1
x3
E3
x1
dx0 dx3 x1 .
(7.31)
~ H
~ = 0 and 0 H
~ = rot E.
~
(7.32)
is the same as
203
This is the first pair of Maxwells equations. Of course, this could have been
seen already in eq.[7.30], which gives them directly in the usual covariant
expression
F + F + F = 0.
(7.33)
Equation [7.31] says that the electromagnetic form F is closed. In Minkowski pseudo-euclidean space (supposedly contractible; recall that we do
not know much about its real topology, 1.1.14, 1.1.18 and 1.3.6), there
exists then a 1-form
A = A dx
such that
F = dA =
1
2
[ A A ] dx dx ,
(7.34)
or, in components,
F = A A .
The potential form A is not unique: given any 0-form f , we can also write
F = d(A + df ). The potentials A and A0 = A + df,
A0 = A + f,
(7.35)
give both the same field F . This is a gauge transformation. The gauge
invariance of F is thus related to its closedness and to the arbitrariness born
from the Poincare lemma. We could formally define F as dA. In that case, the
first pair of Maxwells are not really equations, but constitute an identity.
This point of view is justified in the general framework of gauge theories.
From the quantum point of view, the fundamental field is the potential A,
and not the field strength F . Although itself not measurable, its integral
along a closed line is measurable (Aharonov-Bohm effect, seen in 4.2.18).
Furthermore, it is the field whose quanta are the photons. Even classically,
there is a hint of its more fundamental character, coming from the lagrangian
formalism: interactions with a current j are given by A j . There is also a
further suggestion of the special character of dF = 0: unlike the second pair
of Maxwells equations (see below, 7.4.17), the first pair does not follow
from variations of the electromagnetic lagrangian L = 41 F F .
7.2.10 Thermodynamics of very simple systems We call very simple systems those whose states are described by points on a two-dimensional
manifold with boundary, usually taken as diffeomorphic to the upper right
quadrant of the plane E2 . Thermodynamical coordinates are conveniently
chosen so as to represent measurable physical variables. We shall use the
entropy S and the volume V . The remaining physical quantities are then
204
T
S V
dS +
T
V
dP =
P
S V
dS +
P
V
S
S
dV,
dV.
Thus,
T
V
S
dV dS =
P
S V
dS dV.
Consequently,
T
V
S
P
S V
which is one of Maxwells reciprocal relations. The other relations are obtained in the same way, however using different independent variables from
the start. All of them are integrability conditions, here embodied in the
Poincare lemma. A mathematically well founded formulation of Thermodynamics was initiated by Caratheodory3 and is nowadays advantageously spelt
in terms of differential forms, but we shall not proceed to it here.4
7.2.11 We have introduced 1-forms, to start with, as differentials of functions (or 0-forms). We have afterwards said that not every 1-form is the
differential of some function, and have found some examples of scu nondifferential forms: mechanical work ( 7.1.6) and thermodynamical heat and
work exchanges ( 7.2.10). This happens also for forms of higher order: not
every p-form is the differential of some (p 1)-form. This is obviously related
to integrability: given an exact form
3
4
Very nice resumes are found in Chandrasekhar 1939 and Born 1964.
See for instance Mrugala 1978 and references therein.
205
= d ,
Pp
j1
j=1 ()
R1
0
(7.36)
(7.37)
The proof of this fundamental formula is rather involved and will not be
given here.5 It can nevertheless be directly verified from eq.[7.36], by using
the identity
d
d
[t(tx)] t [(tx)].
(tx) =
dt
dt
5
A constructive proof for general manifolds is found in Nash & Sen 1983; on En , proofs
are given in every textbook: Goldberg l962, Burke l985, etc.
206
(7.38)
so that is indeed exact and the looked for above is just = T (up to s
such that d = 0). Of course, the formulae above hold globally on euclidean
spaces, which are contractible.
~ = constant. It is closed by
7.2.14 Take a constant magnetic field, B
~ B
~ = 0. It is the rotational of A
~ = TB
~ =B
~ ~r, as
Maxwells equation
comes directly from eq.[7.36].
7.2.15 A simple test: take in E2 the form
p
xdx+ydy
v = dr = d x2 + y 2 =
.
2
2
x +y
Then, as expected,
Tv =
R1
0
2 +ty 2
dt tx2
t x2 +t2 y 2
x2 + y 2 = r.
(7.39)
r
r
g
1 2
n
i g y
i f
g[y
,
y
,
.
.
.
,
y
]
=
X
=
X
.
i
r
i
i
x
y x
x y r
(f X) = X i f
xi
f r
= X(f r ) f r .
207
Figure 7.5:
(f )(X) = (f X) = ( f )(X).
In a local basis, we have = j dy j and
(f X) = j dy j [X i
In the case case = dg =
g
y r
f r
]
xi y r
= j
f r
xi
Xi .
dy r :
(f dg)(X) = dg(f X) =
g f r
y r xi
X i.
1
p!
[f i1 i2 i3 ...ip dy i1 dy i2 y i3 . . . y ip1 ] f dy ip .
Therefore
d[f ] =
1
p!
208
1
p!
d[f ] =
1
p!
Consequently,
d[f ] = f [d] .
7.2.17 General Basis As far as derivations are involved, calculations
are simpler in natural bases, but other bases may be more convenient when
some symmetry is present. Let us go back to general basis and reexamine the
question of derivation. Suppose {e } is a general basis for the vector fields
in an open coordinate neighbourhood U of M , and { } its dual basis. They
can be related to the natural basis of the chart (U, x):
e = e
;
x
(7.40)
= dx .
(7.41)
= e e ,
x
(7.42)
dx = ,
(7.43)
Conversely,
(7.44)
(7.45)
209
= 12 [e e (e ) e e (e )] e e
= 12 [e e (e ) e e (e )] .
Using the relations like e e = and the derivatives
e e (e ) = e e (e ),
etc, we finally get
d = 21 C .
(7.46)
d (e , e ) = C
.
(7.47)
We shall see later that, on Lie groups, there is always a basis in which
the structure coefficients are constant (the structure constants). In that
case, eq.[7.46] bears the name of Maurer-Cartan equation.
Suppose now the 1-form A = A . By using [7.45], one easily finds
dA = 12 [e (A ) e A ) C ] .
(7.49)
C = 0.
7.2.18 We can go back to the anholonomic spherical basis of 6.5.3 for
E3 , and find the dual forms to the fields Xr , X and X . They are given
respectively by: r = dr, = rd and = r sin . We may then write
the gradient in this basis,
df = ek (f ) k = Xr (f ) r + X (f ) + X (f )
and check that this is the same as
df = (r f )dr + ( f )d + ( f )d,
which would be the expression in the natural basis related to the coordinates
(r, , ). The invariance of the exterior derivative is clear.
210
7.3
VECTOR-VALUED FORMS
7.3.1 Up to now, we have been considering ordinary q-forms, antisymmetric linear mappings taking q vector fields into the real line. A vectorvalued q-form will take the same fields into a vector space. If V is the vector
space and q (M ) is the space of ordinary q-forms on the manifold M , a qform with values in V is an element of the direct product V q (M ); if
{Va } is a basis for V , is written
= Va a ,
(7.51)
(7.53)
remains invariant. Usual forms a have already been introduced as basisindependent objects in q (M ), so that the whole object s basis-independent
in both spaces.
7.3.3 Algebra-valued forms Of special interest is the case in which the
vector space V has an additional structure of Lie algebra. The generators
will satisfy commutation relations [Ja , Jb ] = f c ab Jc and may be used as a
basis for the linear space V . It is precisely what happens in the examples
quoted above: gauge fields and potentials are forms with values in the Lie
algebra of the gauge group. There are two possible operations on such forms,
the exterior product and the Lie algebra operation. Due to the possible
anticommutation properties of exterior products, one must be careful when
handling operations with the complete vector-valued forms. A bracket can be
211
(7.54)
(7.55)
C being the order of the algebra-valued form C. This is a graded commutator . When at least one of the involved forms is of even degree, no signs
will come from the exterior product and the bracket reduces to a simple
commutator. Otherwise, an anticommutator comes out.
7.3.5 For example, a 1-form will be
A = Ja Aa dx .
(7.56)
Start by calculating
A A = Ja Jb Aa Ab dx x = 21 [Ja , Jb ]Aa Ab dx x
= 12 Jc f c ab Aa Ab dx x ,
(7.57)
the fc ab s being the Lie algebra structure constants. If we compare with [7.54],
it comes out that
A A = ( 12 [A, A].
(7.58)
As announced, this is actually a graded commutator, though we use for it
the usual commutator symbol.
7.3.6 A particular example is given by the algebra su(2) of the Lie group
SU (2) of special (that is, with determinant = +1) unitary complex matrices.
The lowest-dimensional representation has generators Ji = 21 i , where for
the i s we may take the Pauli matrices in the forms
1 = [ 01 10 ] ; 2 = [ 0i 0 i ] ; ; 3 = [ 10 01 ] ; .
The 1-form [7.56] is then the matrix
h
A3 dx
A = 21 A1 dx +iA
2 dx
A1 dx iA2 dx
A3 dx
Lichnerowicz 1955.
212
(7.59)
(7.60)
(7.61)
The Ja s are the generators of the gauge group Lie algebra. To obtain the
relations between the components, use [[7.61], [[7.59] and [[7.57] to write
F = 12 Ja ( Aa Aa + f a bc Ab Ac )dx dx .
(7.62)
(7.63)
F a = Aa Aa + f a bc Ab Ac .
(7.64)
(7.65)
213
DA () = d() + A ()
is the covariant derivative of a 1-form. This will be a bit more formalized
below, in 7.3.11. As we stand, such names are mere analogies to the Riemannian case. Within this interpretation, the field is the covariant derivative
of the connection proper, and the Bianchi identity establishes the vanishing
of the covariant derivative of the field. By the same analogy, the field F is
the curvature of the connection A. In components, eq.[7.65] reads
0 = dF + [A, F ] =
1
3!
Ja {[ F a ] + f a bc Ab [ F c ] } dx dx dx ,
214
dW 0 may be written as
dW= g dW g1 + g{ - (-)W W }g1 = g{dW + |[,]|}g1 ,
in terms of the graded commutator [7.55]. We look now for a compensating
form, a 1-form A transforming according to
A0 = gAg 1 + gdg 1 = g{A }g 1 ,
and we verify that
dW 0 + |[A0 , W 0 ]| = g{dW + |[A, W ]|}g 1 .
In geometrical language, an A transforming as above is a connection on the
manifold. It follows from the last expression that
DW = dW + |[A, W ]|
is the covariant derivative for any form transforming according to the expression W 0 = gW g 1 . This reduces to the previous expressions in the case
of gauge fields. On the other hand, if W is a column vector of forms, and
the matrices act as usual on column vectors, W 0 = gJa W a = gW , then the
covariant derivative is
DW = dW + AW,
with A the same connection as above. Of this type are the covariant derivatives of the source fields in gauge theories. In general, the gs generate a
group G. Quantities transforming as gW g 1 are said to belong to the adjoint representation of G, and quantities transforming as gW belong to linear
representations. In gauge theories, the potentials A and their curvatures
(field strengths) F belong to the adjoint representation. Source fields usually belong to linear representations. Let us retain that the expression of
the covariant derivative depends both on the order of the form and on the
representation to which it belongs in the transformation group.
It is good to keep in mind that some of the so-called covariant derivatives found in
many texts are actually covariant coderivatives, to be seen later ( 7.4.20).
215
(7.67)
(7.68)
(7.69)
(7.70)
(7.71)
From < ei , ej > = ij , one obtains < dei , ej > + < ei , dej > = 0, with the
consequence
ij = ji .
(7.72)
We can define matrices inverse to those appearing in [7.69] and [7.70] so that
K = ej ej and (K T ) = i i . Basis duality enforces ej = j and i = ei .
Differentiating j ei = ij , one finds that
dj = j dei ei = ij ei .
Consequently,
d j = dj (K T ) = ij ei (K T )
and
d j = ij i .
ij
(7.73)
The forms are the Cartan connection forms of space En . Basis like {ei },
which can be defined everywhere on En , are called moving frames (rep`eres
mobiles). Equations [7.68] and [7.73] are the (Cartan) structure equations
of En . We have said that, according to a theorem by Whitney, every differentiable manifold can be locally imbedded (immersed) in some En , for n
large enough. Cartan has used moving frames to analyze the geometry of
general smooth manifolds immersed in large enough euclidean spaces. More
about this subject can be found in Mathematical Topic 10. For an example
of moving frames in elasticity, see Physical Topic 3.3.2.
216
7.3.13 Let us come back to the expressions of the exterior derivative. They
have been given either in the basis-independent form [7.26] or in the natural
basis [7.25]. When a form is given in a general basis,
=
1
p!
j1 j2 jp j1 j2 . . . jp ,
things get more involved as the derivatives of each basis element jk must be
taken into account also. Suppose the basis { i } and its dual {ej }. Writing
the Cartan 1-form in the basis { i } as ij = j ik k , where the j ik s are
the connection components, eqs.[7.73] and [7.45] tell us that the structure
coefficients are the antisymmetric parts of these components, C j ik = j [ik] ,
and we can choose
j ik = ejm ek (em
i ).
In terms of this connection, a covariant derivative j is defined whose
components just appear in the expression for the exterior derivative in a
general basis:
1
d = ()p (p+1)!
j1 j2 . . . jp jp+1
1
= ()p (p+1)!
[p+1 j1 j2 jp ] j1 j2 . . . jp jp+1 ,
(7.74)
(7.75)
This is a bit more general than the well known formulae giving the differential
in general coordinate systems, and reduce to them for natural basis.
7.3.14 Frobenius theorem, alternative version We have said ( 6.4.33)
that a set of linearly independent tangent vector fields X1 , X2 , . . . , Xn on a
manifold M are locally tangent to a submanifold N (of dimension n < m)
around a point p M if they are in involution, [Xj , Xk ] = ci jk Xi . This
means that, if we take such fields as members of a local basis {Xa } on M ,
with a = 1, 2, . . . , n, n + 1, . . . , m, the structure coefficients ca jk vanish whenever a n + 1. The dual version is the following: consider a set of Pfaffian
forms 1 , . . . , n . Linear independence means that their exterior product is
nonvanishing (Cartan lemma, Mathematical Topic 10.1.2). If such forms are
to be cotangent to a submanifold, they must close the dual algebra to the
involution condition, and we must have eq.[7.46] for the ci jk s restricted to
217
7.4
7.4.1 In some previous examples, we have been using relationships between forms like Uk = 2!1 kij Tij (in E3 ) and Fe = 2!1 F (in Minkowski
space). These are particular cases of a general relation between p-forms and
(n p)-forms on a manifold N . Recall that the dimension
of the space p of
n
n
p-forms on an n-dimensional manifold is p = np . Of course, the space of
(n p)-forms is a vector space of the same dimension and so both spaces are
isomorphic. The presence of a metric makes of this isomorphism a canonical
one.
7.4.2 Given the Kronecker symbol i1 i2 ...in and a metric g, we can define
mixed-index symbols by raising some of the indices with the help of the
contravariant metric,
j j ...j
p
1 2
j1 k1 j2 k2
ip+1
g
. . . g jp kp k1 k2 ...kp ip+1 ...in .
...in = g
(np)!
g
j j ...j
i11...i2 p p ,
218
|g|
jp
j2
j1
[ . . . ] = (np)! j1 j2 ...jp jp+1 ...jn jp +1 . . . jn .
(7.76)
Here |g| is the modulus of g = det(gij ). A p-form will be taken into its
dual, the (n p)-form
h
i
j1
j2
jp
1
= p! j1 j2 jp . . .
|g|
= (np)! p! j1 j2 ...jp jp+1 ...jn j1 j2 ...jp jp+1 . . . jn ,
or
|g|
(np)! p!
(7.77)
|g|
( )jp+1 ...jn = p! j1 j2 ...jp jp+1 ...jn j1 j2 ...jp .
(7.78)
|g|
v = 1 = n! j1 j2 ...jn j1 j2 . . . jn
(7.79)
p
1
2
n
(7.80)
= |g| . . . ,
which is an especial volume form ( 7.1.5) called the canonical volume form
corresponding to the metric g. Given the basis {ej }, dual to { j },
v(e1 , e2 , . . . , en ) =
|g|
n!
j
1 (e1 ) j1 (e2 )
j
2 (e1 ) j2 (e2 )
j1 j2 ...jn
...
j. . .
n (e1 ) j2 (e1 )
. . . j1 (en )
. . . j2 (en )
...
...
jn
. . . (en )
219
or
v(e1 , e2 , . . . , en ) =
|g|
n!
j1 j2 ...jn j1 j2 ...jn =
p
|g|
(7.81)
by [7.1] and [7.4]. This could of course have been obtained directly from
[7.80]. Given an arbitrary set of n fields X1 , X2 , . . . , Xn ,
v(X1 , X2 , . . . , Xn ) =
|g|
n!
X1j1 X2j1
X1j2 X2j2
... ...
X1jn X2jn
j1 j2 ...jn
. . . Xnj1
. . . Xnj2
. . . . . .
. . . Xnjn
or
v(X1 , X2 , . . . , Xn ) =
(7.82)
(7.83)
for all sets of p fields {X1 , X2 , . . . , Xp }. This is the invariant, basis independent, definition of operator . It coincides with [7.77], and this tells us that
what we have done there (and might be not evident), supposing that the
operator ignores the components, is correct.
7.4.6 Let us go back to eq.[7.76] and check what comes out when we apply
twice the star operator:
[ j1 j2 . . . jp ]
=
=
|g|
(np)! p!
|g|
(np)! p!
Now,
k1 k2 ...kp jp+1 jp+2 ...jn g jp+1 ip+1 g jp+2 ip+2 . . . g jn in g j1 k1 g j2 k2 . . . g jp kp
= j1 j2 ...jp ip+1 ip+2 ...in det (g ij ) = j1 j2 ...jp ip+1 ip+2 ...in g 1 ,
220
so that
[ j1 j2 . . . jp ]
=
|g|
1
g (np)! p!
j j ...j
|g|
1
()p(np) i11i22...ipp (n p)! i1 . . . ip
g (np)! p!
1
= |g|
()p(np) (n p)!p! i1 . . . ip
g (np)! p!
= |g|
()p(np) i1
g
. . . ip .
Thus, taking twice the dual of a p-form yields back the original form up to
a sign and the factor |g|/g. The components of a metric constitute always a
symmetric matrix, which can always be put in diagonal form in a convenient
basis. The number of positive diagonal terms (P ) minus the number of
negative diagonal terms (N ), s = P N = (n N ) N = n 2N is
an invariant property of the metric (a theorem due to Sylvester), called its
signature.7 Minkowski metric, for instance, has signature s = 2. The factor
|g|/g is simply a sign ()N = ()(ns)/2 . We find thus that, for any p-form,
p = ()p(np)+(ns)/2 p ,
(7.84)
(7.85)
when applied to a p-form and the metric has signature s. Of course, |g|/g = 1
for a strictly Riemannian metric and for this reason the signature dependence
of 1 is frequently ignored in textbooks.
7.4.7 Let {ej } be a basis in which the metric components are gij . The
metric volume element introduced in 7.4.4 could have been alternatively
defined as the n-form v such that
v(e1 , e2 , . . . , en ) v(e1 , e2 , . . . , en ) = det (gij ) = g.
(7.86)
In reality, this only fixes v up to a sign. The manifold N has been supposed
to be orientable and the choice of the sign in this case corresponds to a choice
of orientation.
7.4.8 An inner product (, ) between two forms of the same order can
then be introduced: it is such that
( ) = (, ) v.
(7.87)
It generalizes the inner product generated by the metric on the space of the
1-forms.
7
We could have defined s = N P instead, without any change for our purposes.
221
(7.88)
7.4.11 We have already used the star operator in 7.2.8. The trivial character of the euclidean metric has hidden it somewhat, but the correspondence
between vectors and second order antisymmetric tensors in E3 is given precisely by the star operator. It is essential to the definition of the laplacian of
a function f, f = divgradf . It is a simple exercise to check that
d df = (f )dx dy dz.
7.4.12 Take Minkowski space, with g00 = 1 and gii = 1, in the cartesian basis {dxi }, and with the convention 0123 = +1:
dx1 =
1
3!
p
|g| 1 dx x dx = 1 230 dx2 dx3 dx0
= dx2 dx3 dx0 = dx0 dx2 dx3 ;
1
2!
1
3!
|g| A x x x .
F dx dx ,
1 1
[
2! 2!
F ]dx dx =
1
2!
Fe dx dx .
222
(7.89)
Perhaps codifferential would be a more appropriate name, but coderivative is more usual. A quick counting will tell that this additional exterior
differentiation takes a p-form into a (p 1)-form. There is more:
dede = ()p1 ()p 1 d 1 d = 1 d d 0.
(7.90)
(7.91)
Still another expression will be given in 7.6.13. Only after that an expression
will be found for the coderivative de(p q ) of the wedge product of two
forms.
7.4.16 par:7laplacian Now, a laplacian operator can be defined which acts
on forms of any order:
e 2 = d de + ded.
:= (d + d)
(7.92)
223
On 0-forms, reduces (up to a sign!) to the usual Laplace-Beltrami operator acting on functions. Notice that the laplacian of a p-form is a p-form.
Harmonic analysis can be extended to antisymmetric tensors of any order.
A p-form such that = 0 is said to be harmonic. The harmonic p-forms
constitute a vector space by themselves. From the very definition of , a
form simultaneously closed and coclosed is harmonic.
The laplacian has the commutation properties
d = d;
= ;
e
de = d.
(7.94)
Comparison with eq.[7.29] shows that the Hodge operator takes H E and
E H. It corresponds so to the usual dual transformation in electromagnetism, a symmetry of the theory in the sourceless case. If we calculate the
coderivative of the electromagnetic form F , we find
~ Edx
~ 0 + (0 E
~ rotH)
~ d~x.
deF =
(7.95)
We have seen in 7.2.9 that the first pair of Maxwells equations is summarized in dF = 0. Now we see that, in the absence of sources, the second
pair
~ E
~ = 0 and 0 E
~ = rotH
~
(7.96)
is equivalent to deF = 0. The first pair is metric-independent. The coderivative is strongly metric-dependent, and so is the second pair of Maxwells
equations. Equations [7.31] and [7.96] tell us that, in the absence of sources,
F is a harmonic form. The first pair does not change when charges and
currents are present, but the the second does: the first equation in [7.96]
acquires a term 4 in the right-hand side, and the second a term 4J. If
we define the current 1-form
h
i
j := 4 dx0 J~ d~x ,
(7.97)
224
(7.98)
(7.99)
~ J~ = 0,
0 +
(7.100)
or, in components,
which is the continuity equation: charge conservation is a consequence of
the coexactness of the current form. Notice that this is metric dependent.
In the presence of charges, the electromagnetic form is no more harmonic,
but remains closed. Consequently, in every contractible region of Minkowski
space there exists a 1-form A = A dx such that F = dA. From eq.[7.98] we
see that this potential form obeys the wave equation
dedA = j,
(7.101)
A + A = j .
(7.102)
or, in components,
The Lorenz gauge is the choice deA = A = 0.
7.4.18 Here is a list of relations valid in Minkowski space, for forms of
degrees 0 to 4 (we use the compact notations dx = dx dx , dx =
dx dx dx , dx dx dx dx dx ; the bracket [, , . . .] means a
complete antisymmetrization in the included indices).
form
f
A dx
1
F dx
2!
1
W dx
3!
1
V
dx
4!
f dx1230
1
A dx
3!
1 1
[ F ]dx
2! 2!
W [ dx]
V1230
d
( f )dx
1
A dx
2! [ ]
1
F dx
3! [ ]
1
W dx
4! [ ]
0
de
0
A
F dx
2!1 W dx
1
V1230 dx
3!
7.4.19 With the notation of ( 7.3.8), the field equations for gauge theories
are the Yang-Mills equations
F a + f a bc Ab F c = J a ,
(7.103)
(7.104)
225
(7.105)
Thus, in the sourceless case, the field equations are just the Bianchi identities
[7.65], [7.66] written for the dual of F . A point of interest of self-dual fields
( 7.4.13) is that for them the sourceless Yang-Mills equations coincide with
the Bianchi identities. Any F of the form F = dA + A A will solve the field
equations. Recall the discussion of 7.3.11 on types of covariant derivatives,
depending on the form degrees. The expression at the left-hand side of the
equations above is the covariant coderivative of the 2-form F according to
the connection A, as will be seen below.
7.4.20 Covariant coderivative To adapt the discussion ( 7.3.11) on
covariant derivatives to the case of coderivatives, we simply notice that the
transformations in the value (vector) space ignore the tensor content of the
form, so that ( W )0 = W 0 = g W g 1 . We can therefore apply the same
reasoning to obtain
d W 0 + |[A0 , W 0 ]| = g{d W + |[A, W ]|}g 1 .
e = ()W 1 d W , we apply ()W 1 to this expression to find the
As dW
covariant coderivative
e + ()W 1 |[A, W ]|.
e = dW
DW
(7.106)
7.5
(7.107)
We present here a quick, simplified view of an involved subject. The excuse for quickness
is precisely simplicity. The aim is to introduce the crucial ideas and those simple notions
which are of established physical relevance.
7.5.1
Integration
7.5.1 Let us go back to our small talk of Section 7.1, where it was said that
exterior differential forms are entities living under the shadow of the integral
sign. We shall not have space for any serious presentation of the subject
only the main ideas and results will find their place here. The fundamental
226
as a limit of a summation.
(7.109)
227
f (P )M
f .
:=
f (Sp )
(7.111)
Sp
j
X
i=1
Z
mi
=
pi
j
X
i=1
Z
mi
f .
(7.112)
Spi
The boundary of a singular chain is defined in a natural way: first, the boundary of a curvilinear simplex is the image of the restriction of the mapping
f to the boundary of the corresponding euclidean simplex, pj = f (Spj ).
Thus,
j
X
p =
mi pi .
(7.113)
i=1
As the mapping f is a homeomorphism, it will take the empty set into the
empty set. Thus, pj = f (Spj ) = f () = 0, and
p 0.
(7.114)
228
7.5.8 The above definition of singular chains carries over to them all the
algebraic properties of euclidean chains. It is then possible to define closed
singular chains, or cycles, as well as exact singular chains, and transpose the
same homology groups to chains on differentiable manifolds.
7.5.9 We are now ripe to state one of the most important theorems of all
Mathematics: the integral of the exterior derivative of a form, taken on a
singular chain, equals the integral of the form taken on the boundary of that
chain:
Z
Z
d =
.
(7.115)
It is called by a great mathematician of today (i.e., Arnold) the Newton Leibniz - Gauss - Ostrogradski - Green - Stokes - Poincare theorem, a name
tracing its historical evolution. Its fantastic generality was recognized by the
last member of the illustrious team, and includes as particular cases all the
vector analysis theorems associated to those eminent names. The first two
patriarchs belong because, if f is a real function and is the real oriented
interval (a, b), then we have that = b a, and
Z
Z
df =
Z
f = f (b) f (a).
f=
a
(7.116)
Most authors call the theorem after the last-but-one member of the list.
7.5.10 Although we shall not pretend to demonstrate it, let us give only
one indication: it is enough to suppose it valid for the euclidean case:
Z
Z
Z
d =
f d =
d(f );
=f (S)
=
f =
=
.
S
f (S)
(7.117)
229
=S
S2
B 3
B3
Z
(rot A) d =
dA =
S2
B3
S2
A = 0.
S 2 =0
=S
As already said, H is a 2-tensor; by using the Hodge operation, it can be confused with
e = rot H.
an (axial) vector. Then dH = div H and dH
S3
S3
S 3
230
7.5.14 Consider a force field on E3 : suppose that the work necessary for a
displacement between two points a and b is independent of the path (Figure
7.6). This is the same as saying that the integral along any closed path
through a and b is independent of that closed path, and thus vanishes.
For a
H
i
conservative system, the work W = Fi dx is a closed form. From Fi dxi = 0
231
assumed to exist for any domain D. We have later assumed the existence of
a differentiable homeomorphism from euclidean spaces to the differentiable
manifolds to the extent of taking chains from one to the other, etc, etc. It
should however be said, in favour of the above exposition, that it allows a
fair view of the main relationships of integration to the global characteristics of the space. The interested reader is urged to consult more complete
treatments.9
7.5.17 On hypersurfaces10 A hypersurface is an (n 1)-dimensional
space immersed in En . It may be an imbedded manifold, or be only locally differentiable. We shall here suppose it an imbedding for the sake of
simplicity. Suppose in En a hypersurface given by the equation (x) =
(x1 , x2 , . . . , xn ) = 0. Then to will correspond a special (n 1)-form, its
volume form , in the following way. A requirement will be that the surface
be nonsingular, grad = d 6= 0, at least locally. Let v be the En volume
form, v = dx1 dx2 dxn . Then is defined by v = d . Around
a point p one is in general able to define new coordinates {uk } with positive
x
|| and such that one of them, say uj , is (x). Then,
jacobian || u
x
v = du1 du2 duj1 d duj+1 dun .
(7.118)
u
If around a point p it so happens that j 6= 0, we may simply choose
ui6=j = xi and uj = (x), in which case
= ()j1
(7.119)
232
D, a function which is 1 inside D and zero outside it. If D is the set of points
x such that (x) 0, we define () = 1 for (x) 0 and () = 0 for
(x) > 0, so that () is the characteristic function of D. In this case we
have that 0 () = (), and
grad = () grad .
This corresponds to the usual practice of using, on a surface, the normal
directed outward, and leads to general expressions for well known relations
of vector analysis.
Such distributions are of use, for example, in getting Maxwells equations
in integral form (Physical Topic 4.3).
7.5.2
We begin by recalling and in the meantime rephrasing some of the comments about
chains and their homologies, made in chapter 2. Then we describe the dual structures
which are found among the forms, cohomologies. Although cohomology is an algebraic
structure with a very wide range of applications, differential forms are a subject of choice
to introduce the main ideas involved.11 Here we shall suppose all the chains already
displayed on a general smooth manifold M .
233
which has all the properties of a scalar product, just as in the case of a vector
space and its dual. This product is in reality an action between homology
and cohomology classes, and not between individual forms and chains.
This is a consequence of the two following properties:
(i) The integral of a closed form over a cycle depends only on the homology class
of the cycle : if = , then
Z
Z
Z
Z
Z
Z
Z
=
= +
= + d = .
(ii) The integral of a closed form over a cycle depends only on the cohomology
class of the form : if = d, then
Z
Z
Z
Z
Z
Z
Z
= ( + d) =
+ d =
+
d =
.
234
7.5.21 On a compact metric manifold, the star operator allows the definition of a global inner product between p-forms through an integration over
the whole manifold M :
Z
(, ) :=
(7.121)
M
7.5.22 A few comments: first, the compactness requirement is made to ensure the
existence of the integrals. It can be softened to the exigency that at least one of the
forms involved has a compact carrier: that is, it is different from zero only on a compact
domain. Second, the complete duality between forms and chains really requires something
else: chains on general smooth manifolds have been introduced through mappings which
are only locally homeomorphisms, and nothing else has been required concerning the
differentiability of their inverses. That is why they are called singular chains. On the
other hand forms, as we have introduced them, are fairly differentiable objects. The above
relation between forms and chains only exists when forms are enlarged so as to admit
components which are distributions. Then, a general theory can be built with forms and
chains as the same objects this has been done by Schwartz and De Rham, who used for
the new general objects, including both forms and chains, the physically rather misleading
name (proposed by Schwartz) current.
(7.123)
Here, another caveat: we have mainly talked about a particular kind of homology, the so-called integer homology, for chains with integer multiplicities.
The parallelism between forms and chains is valid for real homology: chains
with real multiplicities are to be used. The vector spaces are then related to
the real numbers, and the line R takes the place of the previously used set of
integer numbers Z. In the above example, the space H 0 (B n ) is isomorphic
to the real line R. This means that 0-forms (that is, functions) on the balls
235
can be closed but not exact, although in this case they will be constants:
their set is isomorphic to R. This trivial result is no more valid for p 6= 0: in
these cases, every closed form is exact. This is simply a pedantic rephrasing
of what has been said before, since B n is contractible.
7.5.24 The next simplest compact manifolds are the n-dimensional spheres S n imbedded in En+1 . For them,
H 0 (S 0 ) n Rn2
Hn0)(Sn )0 when
H (S ) p6=R0, n.
H p (S
(7.124)
On S 4 , for instance, closed 1-, 2-, and 3-forms are exact. All 4-forms (which
are of course necessarily closed) and closed functions are constants. On the
sphere S 2 , every closed 1-form (irrotational covariant vector field) is exact
(that is, a gradient). All 2-forms and closed functions are constant.
7.5.25 Spheres are the simplest examples of compact manifolds. Life is
much simpler on such manifolds, on which the internal product (7.121) has
many important properties. Take for instance a p-form and a (p 1)-form
. Then,
Z
(d, ) =
d
M
Z
Z
Z
p
p1
d
d( ) + ()
=
d( ) () d =
M
M
M
Z
Z
Z
p 1
=
+
() d =
de.
M
Consequently,
e
(d, ) = (, d).
(7.125)
The operators d and de are, thus, adjoint to each other in this internal product.
It is also easily found that 1 is adjoint to and that the laplacian
e
= dde + dd
is self-adjoint:
(, ) = (, ).
(7.126)
236
7.5.26 Kodaira-Hodge-De Rham decomposition theorem This theorem, in its present-day form, is the grown-up form of a well known result,
of which primitive particular versions were known to Stokes and Helmholtz.
Called a fundamental theorem of vector analysis by Sommerfeld,12 it says
that a differentiable enough vector field V in E3 , with a good enough behaviour at infinity, may be written in the form
V = grad f + rot T + c,
where c is a constant vector. In its modern form, it is perhaps the deepest
result of the above general harmonic analysis. It says that the inner product
divides the space of p-forms into three orthogonal sub-spaces. In a rather
weak version,
on a compact-without-boundary manifold, every form can be decomposed in
a unique way into the sum of one exact, one co-exact and one harmonic
form:
= d + de + h,
(7.127)
with h = 0. The authors the theorem is named after have shown that in
e and =
reality, with the above notation, a form exists such that = d
d, which puts as the sum of a laplacian plus a harmonic form:
= + h.
(7.128)
(7.129)
237
The boundary conditions must be stated in an invariant way. For that, let
us introduce some notation. Let i: M M be the inclusion mapping
238
(7.132)
(7.133)
(7.135)
(7.136)
7.5.31 A last remark: the inner product is used to obtain invariants. The
action for electromagnetism, for example, is the functional of the basic fields
A given by
Z
I[A] = (F, F ) =
F F,
(7.137)
7.6
239
While physicists have been striving to learn some geometry, geometers were progressively
converting their subject into algebra. We have been leaning heavily on analytical terminology and way-of-thinking in the last chapters. In this paragraph we shall rephrase some
of the previous results and proceed in a more algebraic tune.
7.6.1 Let us start by stating (or restating) briefly some known facts. We
refer to Mathematical Topic .1 for more detailed descriptions of algebraic
concepts. An algebra is a vector space V with a binary operation V V V ,
submitted to certain general conditions. It may be associative or not, and
commutative or not.
7.6.2 A given algebra is a Lie algebra if its operation is anticommutative
and satisfies the Jacobi identity. When the algebra is associative but not a
Lie algebra, it can be made into one: starting from any binary operation a
Lie bracket can be defined as the commutator
[, ] = ,
and makes of any associative algebra a Lie algebra.
7.6.3 An endomorphism (or linear operator) on a vector space V is a
mapping V V preserving its linear structure. If V is any algebra (not
necessarily associative, it may be merely a vector space), End V = {set of
endomorphisms on V} is an associative algebra. Then the set [End V ] of its
commutators is a Lie algebra. The generic name derivation is given to any
endomorphism D : V V satisfying Leibniz law:
D() = (D) + (D).
The Lie algebra [End V ] contains D(V ), the vector subspace of all the derivations of V , and the Lie bracket makes of D(V ) a Lie subalgebra of [End V ].
This means that the commutator of two derivations is a derivation. Given an
element V , it defines an endomorphism ad(a) = ada , called the adjoint
action of a, by
ada (b) = [a, b].
This is a derivative because
ada (bc) = [a, bc] = b[a, c] + [a, b]c = ada (b)c + bada (c).
The set ad(A) = {ada for all a A} contains all the internal derivations of
a Lie algebra A, and is itself a Lie algebra homomorphic to A.
240
(7.139)
(7.140)
If is a 1-form,
iX = i(X) = < , X > = (X).
(7.141)
241
p
X
j=1
(7.143)
X = ()
p1 [1 X 0 p ]
Z
p1+(np+1)(p1)+(ns)/2
= ()
p1 [X 0 p ]
Z
n(p1)+(ns)/2
= ()
p1 [(X 0 ) p ]
Z
= p1 [i(X) p ] = (p1 , i(X) p ).
7.6.8 Some properties of interest are (f being a real function and g a
mapping between manifolds):
iX f = 0;
(7.144)
iX iX 0;
(7.145)
if X = f iX = iX (f );
g (ig X ) = iX (g );
iX ( ) = (iX ) + () iX .
(7.146)
(7.147)
(7.148)
242
(7.149)
(7.150)
7.6.10 Consequently, LX is a derivation, and d and i(X) are antiderivations. In reality, these three endomorphisms are not independent. Let us
recall the expression for the Lie derivative of a p-form: if
= j1 j2 j3 ...jp xj1 dxj2 . . .dxjp ,
then
(LX ) =
+
+
... +
(7.151)
Take p = 1, = j dxj :
LX = X i (i j ) dxj + (j X i ) i dxj .
On the other hand,
i(X)d = i(X)[(j i ) dxj dxi ]
= < X, j i dxj > dxi (j i ) dxj < X, dxi >
= X j (j i ) dxi X i (j i ) dxj
.
(7.152)
(7.153)
(7.154)
Therefore,
d[iX ] = d[X i i ] = (i X j ) j dxi + X i (j i )dxj .
We see that
LX = iX d + d iX = {iX , d}.
(7.155)
This is actually a general result, valid for s of any order and extremely
useful in calculations. It also illustrates another general property: the anticommutator of two antiderivations is a derivation. There is more in this
243
line. Consider derivatives generically indicated by D, D0 , etc, and antiderivatives A, A0 , etc. Using the definitions, one finds easily that the square of an
antiderivative is a derivative, AA = D. In the same token, one finds for
the anticommutator {A, A0 } = D0 , as well as the following relations for the
commutator: [D, D0 ] = D00 and [D, A] = A.
7.6.11 Consequences of eqs.[7.155] and (7.145) are the commutation properties
L X iX = iX L X
(7.156)
d(LX ) = LX (d).
(7.157)
The Lie derivative commutes both with the interior poduct and the exterior
derivative. Other interesting properties are:
Lf X = f LX + df iX
(7.158)
[LX , iY ] = i[X,Y ]
(7.159)
L[X,Y ] = [LX , LY ] .
(7.160)
xj0
in eq.[7.25],
(7.161)
(7.162)
f
This equation, which generalizes the formula df = dxi x
i valid for 0forms, is called the Koszul formula and shows how Lie derivatives generalize
partial derivatives. We may use it to check the coherence between the Lie
derivative and the exterior derivative. As the Lie derivative is a derivation,
d( ) = dxj L j ( ) = dxj (L j ) + (L j )] =
= (dxj L j ) + () (dxj L j ) = (d) + () d).
7.6.13 It is also possible to establish a new expression for the codifferential. From [7.89] and [7.25],
244
Therefore,
1
n(np)+(ns)/2
de = i(j )
= ()
i(j ) j . (7.163)
xj
x
In an euclidean space with cartesian basis,
j = ()p(np) j ,
x
x
and we have
de = i(j ) j = i(j )L j
x
In a general basis, the latter is written as center
de = i(ej )Lej .
(7.164)
(7.165)
This is not unexpected if we look at [7.161], and remember the given adjoint
relation between (dxj ) and i(j ). With [7.155], it leads directly to
de = i(j ) d i(j ).
(7.166)
7.6.14 Using [7.164], the derivation character of the Lie derivative, and
[7.144]-[7.148] we find that
de( ) = de + () de
() (L j ) [i(j )] [i(j )] (L j ). (7.167)
7.6.15 From eq.[7.142] it turns out that, if {Xj } is the basis dual to {k },
i(Xj )[i ] = ji i i(Xj ),
so that
i
i(Xj ) + i(Xj ) i
= (i ) i(Xj ) + i(Xj ) (i )
= {(i ), i(Xj )} = ji (7.168)
for any . Consequently, we find the anticommutator
{(i ), i(Xj )} = ji .
(7.169)
245
(j )i(Xj ) p = p p .
(7.170)
j=1
(7.171)
(7.172)
(7.173)
as they should. These formulas are the starting point of what some people call
supersymmetric quantum mechanics and have been beautifully applied14
to Morse theory and in the study of instantons.
7.6.16 The Lie algebra (M ) of fields on a smooth manifold M acts on
the space C of differentiable functions on M . We have an algebra, and
another space on which it acts as endomorphisms. The second space is a
module and we say that C is a (M )-module. With the action of the Lie
derivatives, which due to eq.[7.160] represent the Lie algebra, also the space
of p-forms is a module.
14
Witten 1982a, b.
246
Chapter 8
SYMMETRIES
8.1
LIE GROUPS
The study of a topological group is much easier when the group operation
is analytic. Even the algebraic structure becomes more accessible. This,
however, requires that the topological space have an analytic structure, more
precisely: that it be an analytic manifold. We have already said ( 5.1.3)
that every C structure has an analytic substructure. That is why most
authors prefer to define a Lie group as a C manifold, ready however to
make good use of the analytic subatlas when necessary. A topological group
has been defined in section 1.4.2 as a topological space G endowed with a
group structure and such that both the mappings m : G G G given by
(g, h) g h, and inv: G G given by g g 1 are continuous.
8.1.1 A Lie group is a C manifold G on which a structure of algebraic
group is defined, in such a way that the mappings
G G G , (g, h) g h
and
G G , g g 1
are all C .
8.1.2 It follows from this definition that the mappings
Lg : G G , h g h
and
Rg : G G , h h g
are diffeomorphisms for every g G. These mappings are called respectively
lefttranslation and righttranslation induced by the element g.
247
248
CHAPTER 8. SYMMETRIES
L t
,
0 1
8.1.5 Other examples may be obtained as direct products : let G1 and G2 be two Lie
groups. The product G = G1 G2 can be endowed with the C structure of the cartesian
product, which makes of G the direct product group G1 G2 if the following operation is
defined:
GGG
((g1 , g2 ), (g10 , g20 )) (g1 g10 , g2 g20 ) .
It is important to notice that the affine group is not to be considered as the Lie group
GL(n, R) Rn , because the product for AL(n, R) is
GGG
((L, t), (L0 , t0 )) (LL0 , Lt0 + t) .
This is an example of semi-direct product, denoted usually by GL(n, R) Rn .
8.1.6 As S 1 is a Lie group, the product of n copies of S 1 is a Lie group, the toral
group or ntorus T n = S 1 S 1 . . . S 1 .
249
250
CHAPTER 8. SYMMETRIES
p
X
i i
xy
i=1
n
X
xj y j .
j=p+1
with p + q = n and Ip,q the diagonal matrix with the first p elements equal
to (+1), and the q remaining ones equal to ( 1). A case of particular
importance is
SO(p, q) = {X O(p, q) such that det X = 1}.
These are non-compact groups, the noblest example of which is the Lorentz
group SO(3, 1). More about these groups will be said in 9.3.4.
8.1.10 One more definition: let be an algebraic homomorphism between the Lie
groups G1 and G2 . Then, will be a homomorphism of Lie groups iff it is C .
251
252
CHAPTER 8. SYMMETRIES
8.2
TRANSFORMATIONS ON MANIFOLDS
253
254
CHAPTER 8. SYMMETRIES
homogeneous manifold are isomorphic. The Lie algebra has then a canonical
decomposition of the form G0 = H 0 + T , for some T .
A homogeneous space G/H has always a Riemann metric which is invariant under
the action of G, and is said to be a homogeneous Riemannian space. We might in larger
generality define a homogeneous Riemannian space as a Riemannian space M whose group
of motions acts transitively on M .
The fact that a space may be obtained as a quotient of two Lie groups as G/H has
very deep consequences, particularly in the case of homogeneous symmetric spaces. This
happens when G has an involutive automorphism, : G G, 2 = 1. The Lie algebra
canonical decomposition G0 = H 0 + T , in the presence of such an involution, has the form
[H 0 , H 0 ] H 0 ; [H 0 , T ] T ; [T, T ] H 0 .
The bundle G = (G/H, H) admits an invariant connection, called its canonical connection,
which is determined by the space T . It has vanishing torsion and a very special form for
the curvature. This special canonical connection is a restriction, to the quotient space, of
the MaurerCartan form of the group G.
8.2.8 The action of G is said to be free when no element but the identity
e of G preserves any point of M , that is, Rg p = p for some p M implies g
= e. Thus a free action admits no fixed point.
8.2.9 The set of points of M which is obtained from a given point p by
the action of G is the orbit of p:
orbit(p) = {q = pg, g G} .
Thus, one obtains the orbit of p by acting on p with all the elements of G.
An orbit may be seen as the invariant subset generated by a single point:
orbit(p) = p G .
Every orbit is a transitive Gspace by itself. A transitive space is clearly an
orbit by one of its points.
8.2.10 Everything which was said above about right action can be repeated for left action, with the notation L, and Lg (p) = gp. No confusion
should arise by the use of the same notation for the (left or right) action on
a manifold M and the (left or right) translation on the group G itself. For a
physical example illustrating the difference between right and leftactions,
see Physical Topic 2.3.9.
8.2.11 As R is a Lie group, a oneparameter group ( 6.4.17) on a manifold
M is an example of action, as it satisfies conditions (i) and (ii) of 8.2.1,
and is C .
255
8.2.13 The best known case is the usual action of matrices on column
vectors. We can rephrase it in a pedantic and precise way and call it the
natural action (on the left) of GL(n, R) on Rn :
L : GL(n, R) Rn Rn
(A, x) Ax.
In an analogous way, we have the action on the right
R : Rn GL(n, R)) Rn
(x, A) xT A.
8.2.14 The left-action of the affine group ( 8.1.4, 8.1.5) on Rn is given
by
x1
x1
2
2
x
x
L
t
L
t
... ,
, ...
0 1
n
n
0 1
x
x
1
1
where L GL(n, R) and t Rn . If L O(n), then we have action of a
subgroup of A(n, R) called the group of rigid motions, or euclidean group on
Rn .
8.2.15 The Poincar
e group P O(4, R) (or inhomogeneous Lorentz group,
see Physical Topic 6) is a subgroup of the affine group for n = 4: A(n, R)
P O(4, R), where
L t
P O(4, R) =
such that L SO(3, 1) and t Minkowski space .
0 1
8.2.16 Linear and Affine basis
256
CHAPTER 8. SYMMETRIES
X
X
X
(f, L) f L :=
fj Lj1 ,
fj Lj2 , . . . ,
fj Ljn .
Given two basis f and f in B(Rn ), there exists a unique L GL(n, R) such that
f = f L .
Thus, the action is simply transitive. This means that there is a one-to-one correspondence
between GL(n, R) and B(Rn ) given by
L e L ,
where e = {e1 , e2 , . . . , en } is the canonical basis for Rn :
e1 = (1, 0, 0, . . . , 0),
Once this correspondence is established, it is possible to endow the set B(Rn ) with a
topology and a C structure so as to make it diffeomorphic to GL(n, R). B(Rn ) is called
the basis space of Rn .
8.2.17 What was done above can be repeated with a general vector space V instead
of Rn . Let us examine now the set of the affine basis on a vector space V :
A(V ) = {[f1 , f2 , . . . , fn ; x] =: [f, x] such that f B(V )
and x V } .
L t
L t
[f L, f t + x] := [f, x]
0 1
0 1
Pn
where f t := i=1 fi ti . Given any two basis f and f A(V ), there is a unique element of
A(n, R) such that
L t
f = f
.
0 1
Thus, there is a one-to-one correspondence between A(n, R) and A(V ), given by
L t
L t
a
,
0 1
0 1
where a = {e1 , e2 , . . . , en , 0} is the canonical basis for A(V ). With this correspondence,
we can endow the set A(V ) with a topology and a C structure making it diffeomorphic
to A(n, R). As a manifold, A(V ) is the affine basis space on V . If, instead of a, another
basis a0 were used, the same structure would result, up to a diffeomorphism.
257
8.2.18 We have learned (section 1.4.1) how to obtain new manifolds from
a given one, as quotient spaces by an equivalence relation. Suppose the
relation between two points p and q M defined by
p q there exists some g G such that q = Rg p.
It is an equivalence. The set [p] = {q M such that q p}, which is the
orbit(p), is the equivalence class with representative p. The set of all these
classes is denoted by M/G: it is said to be the quotient space of M by G.
The canonical projection is defined by : M M/G, p [p]. A quotient
topology can then be introduced on M/G, as well as a C structure.
An important particular case appears when M is itself a Lie group and
G a subgroup of M . Then the quotient space M/G is an analytic manifold. The action G M/G M/G is transitive and M/G is consequently
a homogeneous manifold under the action of G. Manifolds of this type
(group/subgroup) have many interesting special characteristics, as also the
action is an analytic mapping, as well as the projection M M/G.
8.2.19 Suppose the group G acts simultaneously on two manifolds M and N , with
actions denoted mg and ng (see Figure 8.1). A mapping : M N is equivariant (or
an intertwining map) when mg = ng . The diagram at the right of Figure 8.1 is
commutative.
Figure 8.1:
258
CHAPTER 8. SYMMETRIES
the set resulting from whatever action of G on the set A. In particular, the
measure is leftinvariant if (LgA) = (A) and rightinvariant if (Rg A)
= (A). The Lebesgue measure on an euclidean space is invariant under
translations and rotations; the Lebesgue measure on the sphere is invariant
under the action of the rotation group on the sphere. Such Haar measures
are sure to exist only on locally compact groups ( 1.2.13, 8.1.12). On
such groups, they are unique up to positive factors. For the groups T n and
(R, +), the Haar measures are simply the (normalized) Lebesgue measures.
Haar measures provide a characterization of compactness: the Haar measure
on G is finite iff G is compact. This property stays behind the well known
fact that Fourier expansions on compact spaces (groups!) are series while
Fourier expansions on non-compact spaces are integrals (see Mathematical
Topic 6.3).
8.2.21 Invariant integration Given
an invariant Rmeasure on M , the corresponding
R
integral is invariant in the sense that M f (gx)d(x) = M f (x)d(x). An integral may be
left-invariant, right-invariant, or both.
8.2.22 Function spaces, revisited Let us go back to 8.2.7 and consider the space
C(M ) of complex functions on a homogeneous space M . The space C(M ) may be made
into a Gspace by introducing the action (Rg f )(m) = f (mg 1 ). ). Now, C(M ) is a vector
space, which is fair, but it is not necessarily homogeneous. It has, in general, invariant
subsets which are themselves linear spaces. Take for example a simple group G and M
= G, the action being given by the group multiplication, say Rg : x xg. Take C h (G),
the set of all homomorphisms of G into the non-vanishing complex numbers. Then if
h C h (G), we have that h C(G) and the set of all constant multiples of h constitutes
an invariant (one-dimensional) subspace. Such subspaces are independent for distinct hs.
. When G is finite and commutative, each member of C(G) is a unique sum of members,
one from each invariant subspace. When G is infinite, such sums become infinite and
some extra requirements are necessary. In general, one restricts C(G) to a subspace of
measurable (square integrable) functions with some (say, Lebesgue) Haar measure. Take
for instance G as the circle S 1 . The only measurable homomorphisms are hn (x) = einx ,
with n = 0, 1, 2, . . .. Then, any square integrable function f on G may be written in a
unique way as
X
f (x) =
fn einx ,
n
Let us say a few words on the compact non-commutative case. Take G as the rotation
group in E3 , and M = S 2 , the spherical surface in E3 . The space of all squareintegrable
259
functions on M divides itself into invariant subspaces Mj , one for each odd number (2j +1)
and such that dim Mj = (2j+1). They are formed by 3variable homogeneous polynomials
of degree j which are harmonic, that is, satisfy the Laplace equation on M . Each Mj
provides an irreducible representation, the fj Mj being
P the surface harmonics, and
any function f C(S 2 ) is uniquely expanded as f =
n fj , with fj Mj (see also
Mathematical Topic 6.3).
8.3
We have said in section 6.4.5 that the vector fields on a smooth manifold
constitute a Lie algebra. Lie groups are differentiable manifolds of a very
special type. We describe now (very superficially) the general properties of
fields and forms on Lie groups.
8.3.1 Consider the left action of a Lie group G on itself:
Lg : G G
h Lg (h) = gh .
(8.1)
A first thing which is peculiar to the present case is that this action is a
diffeomorphism. It induces of course the differential mapping between the
respective tangent spaces,
Lg = dLg : Th G Tgh G .
(8.2)
(8.3)
(8.4)
(8.5)
260
CHAPTER 8. SYMMETRIES
Figure 8.2:
(8.6)
(8.7)
for any leftinvariant fields X, Y . This is to say that the Lie algebra of
left-invariant fields at any point on a Lie group G is determined by the Lie
algebra of such fields at the identity point of G.
8.3.3 This algebra of invariant fields, a subalgebra of the general Lie algebra of all the fields on G, is the Lie algebra of the Lie group G. It is usually
denoted by L(G), or simply G0 . The vector space of G0 is T eG. A basis for
G0 will be given by d (= dim G) linearly independent leftinvariant fields X ,
which will satisfy
[X , X ] = C X .
(8.8)
8.3.4 According to (8.7), this relation must hold (with the same C s) at
any point of G, so that the structure coefficients are now point-independent.
261
They are, for this reason, called the structure constants of G. The Lie algebra
of G is thus a small (as compared with the infinite algebra of all fields) Lie
algebra of d fields fixed by their values at one point of G.
8.3.5 Right-invariant fields can be defined in an analogous way. They
constitute a Lie algebra isomorphic to that of the left-invariant fields.
8.3.6 A p-form w on G is leftinvariant if
Lg w = w .
(8.9)
Let us see how things work for 1-forms: given a form wgh at gh, its pull-back
is defined by
hLg1 w, Xih = hw, Lg1 Xigh .
(8.10)
If w is invariant,
hw, Xih = hw, Lg1 Xigh .
(8.11)
(8.12)
If also X is invariant,
Therefore, an invariant form, when applied to an invariant field, gives a
constant.
8.3.7 Invariant Pfaffian forms on Lie groups are commonly called Maurer
Cartan forms. They constitute a basis {w } for L (G), dual to that of
invariant fields satisfying eq.(8.8). As a consequence, eq.[7.46] tells us that
they obey
(8.13)
dw = 12 C w w .
This is the MaurerCartan equation, which can be put in a basis-independent
form: define the vector-valued canonical form
w = X w .
(8.14)
(8.15)
(8.16)
(8.17)
are matrices in the Lie algebra satisfying eq.(8.16) (not forgetting that dg 1
= - g 1 dgg 1 ).
262
CHAPTER 8. SYMMETRIES
8.3.9 Given any n n matrix A, its exponential is defined by the (convergent) series
P
1 j
eA = I + A + 21 A2 + . . . =
(8.18)
j=0 j! A .
The set of matrices of type exp(tA), with t R, constitutes an abelian group:
etA esA = e(t+s)A ; etA etA = I ; etc.
The mapping a: R GL(n, R), a(t) = exp(tA), is a curve on GL(n, R)
whose tangent at t = 0 is
d tA
e = AetA t=0 = A .
(8.19)
dt
t=0
So, A TI GL(n, R), or A G0 L(n, R). The set of matrices exp(tA) is the
group generated by A. As A is arbitrary, we have shown that any n n
matrix belongs to G0 L(n, R). Thus, G0 L(n, R) is formed by all the n n
matrices, while GL(n, R) is formed by those which are invertible.
8.3.10 A very important result is Ados theorem:
every Lie algebra of a Lie group is a subalgebra of G0 L(n, R), for some
value of n.
For Lie groups, an analogous statement holds, but only locally: every Lie
group is locally isomorphic to a subgroup of some GL(n, R).
Concerning matrix notation and the use of a basis: a general matrix in
GL(n, R) will be written as g = exp(X p ), where the X s constitute a
basis in G0 L(n, R). The components p are the group parameters of g.
The vector-valued form
w = X w = g 1 dg = g 1 (X dp )g = g 1 X g dp
will be a matrix of forms, with entries
wi k = (X )i k w = [g 1 X g]i k dp .
8.3.11 Exponential mapping We have seen in 6.4.17 how the group
R acts on a manifold. Let us apply what was said there to the case in which
the manifold is itself a Lie group:
:RGG
: (t, h) (t, h) .
(8.20)
263
(t, e) = e (t) .
The theory of ordinary differential equations tells us that, in this case, there
is an open U G around e in which the solution of eq.[6.41],
d
e (t) = Xe (t)
dt
(8.21)
is unique, for any X. Then, a(t) = e (t) is the integral curve of X through
the identity and Xe G0 . Now, when the manifold is a Lie group, this is
a global result: the field X is complete, that is, a(t) is defined for every
t R. Still more, the set {aX (t)}, for all t R is a oneparameter subgroup of G generated by X. We can then introduce the exponential mapping,
generalizing the case of GL(n, R), as
exp : G0 G
exp(X) = aX (0) ,
(8.22)
(8.23)
264
CHAPTER 8. SYMMETRIES
0 p1 p2 . . . pn s
0 0 0 . . . 0 q 1
0 0 0 . . . 0 q 2
h(p, q, s) = . .
..
..
..
.. .
.. ..
.
.
.
.
0 0 0 . . . 0 q n
0 0 0 ... 0
0
It is immediate that the products of two matrices of this kind are given by
h(p, q, s) h(p0 , q 0 , s0 ) = h(0, 0, pq 0 ) ;
[h(p, q, s)]2 = h(0, 0, pq) and [h(p, q, s)]m = 0 for m > 2 .
The commutator
[h(p, q, s), h(p0 , q 0 , s0 )] = h(0, 0, pq 0 p0 q)
will define the Lie algebra. The group Hn is arrived at by the exponential map
h(p, q, s) H(p, q, s) = exp[h(p, q, s)] ,
which is
1 p1
0 1
0 0
H(p, q, s) = . .
.. ..
0 0
0 0
p2
0
1
..
.
...
...
...
..
.
pn
0
0
..
.
0
0
...
...
1
0
s + 21 pq
q1
q2
.
..
q
1
8.4
265
G0
> G0
exp
exp
>
is commutative.
8.4.2 When working with matrices, we are used to seeing a matrix h be
transformed by another matrix g as ghg 1 . This is a special case of a certain very special representation which is rooted in the very nature of a Lie
group. The automorphisms j above belong to the group Aut(G) of the
linear transformations of G (seen as a vector space). The differential operation j j = dj takes Aut(G) into Aut(G). It is a homomorphism, since
d(j k) = dj dk. Consequently, it is a representation of Aut(G) on G. An
important subgroup of Aut(G) is formed by the inner automorphisms of G,
which are combinations of left-and right-translations ( 8.1.2) induced by an
element g and its inverse g 1 :
jg = Lg Rg1 = Rg1 Lg
266
CHAPTER 8. SYMMETRIES
jg (h) = ghg 1 .
(8.24)
(8.25)
(8.26)
(8.27)
(8.28)
Thus: take the curve exp(tX); transform it by jg ; then, Adg X is the tangent
to the transformed curve at the identity (Figure 8.3). This representation is
of fundamental importance in modern field theory, as both gauge potentials
and fields belong to it (see sections 7.3 and 7.4; see also Physical Topic 7).
8.4.5 The mapping
ad := d(Ad)
ad : G0 (AutG)0
X adX
(8.29)
(8.30)
267
Figure 8.3:
In a basis {Xi },
adXi Xj = [Xi , Xj ] = C k ij Xk .
(8.31)
(8.32)
8.4.6 Notice that if g has its actions given by the matrices U (g), and X
is also a matrix (case of GL(n, R)), eq.[8.26] gives simply the usual rule for
matrix transformation U XU 1 . From a purely algebraic point of view, the
adjoint representation is defined by [8.32]: it is that representation by matrices whose entries are the structure constants. It is sometimes called regular
representation,1 but we shall use this name for another kind of representation
(see Mathematical Topic 6).
We may consider also the representations given by the action on the
forms, through the pull-back L(g1 ) and R(g1 ) . Such representations on
the covectors are called coadjoint representations. In the matrix notation of
8.3.10, the adjoint representation will be given by a matrix A, such that
X0 = g 1 (X )g = A X
1
268
CHAPTER 8. SYMMETRIES
Figure 8.4:
Let us change a bit the point of view (see Figure 8.4): take p fixed and
change g G. The action Rg becomes a mapping of G into M , which we
shall denote pe:
pe : G M
pe(g) = Rg (p) = pg.
The set of points pe(g) is, as said in 8.2.9, the orbit of p by G.
8.4.8 Then, the differential mapping
(8.33)
269
pe : Te G Tp M
of Tp M ,
will take a field X of G into some field X
= pe (X).
X
(8.34)
(8.36)
270
CHAPTER 8. SYMMETRIES
Notice the change g g 1 between the two cases. The process is rather
knotty. When we want to know how a fundamental field X changes in some
point p of M , we begin by going back to the field X in G which X represents;
transform X by the adjoint representation and then bring the result back by
the mapping pe (Figure 8.5).
Figure 8.5:
This process is pictorially represented in the commutative diagram
p
> Tp M
G0
Adg1
Rg
G0
> Tp M
8.4.11 All this discussion is strictly local. It shows where the importance
of the adjoint representation comes from and will be instrumental in the
study of fiber bundles.
8.4.12 The Killing form is a bilinear form on G defined by
ij = (Xi , Xj ) = tr [ad (Xi ) ad(Xj )],
(8.37)
271
or
ij = C k im C m jk .
(8.38)
Goldberg l962.
272
CHAPTER 8. SYMMETRIES
Chapter 9
FIBER BUNDLES
9.1
INTRODUCTION
We have already met a few examples of fiber bundles: the tangent bundle, the
cotangent bundle, the bundle of linear frames. Also fibered spaces of another
kind have been seen: the covering spaces, whose fibers are discrete spaces
acted upon by the fundamental group. We shall in the following consider only
bundles with differentiable fibers differential fiber bundles. Of this type,
we have glimpsed tensorial bundles in general, which include the tangent
and the cotangent bundles as particular cases. Locally, bundles are directproduct manifolds, but globally they are nothing of sort. In reality, their
importance comes out mainly when global effects, or effects in the large,
are at stake. As Physics is largely based on (local) differential equations,
such effects are usually taken into account in the boundary conditions.
We start with an intuitive presentation of the bundle idea, then proceed
to examine the simplest cases, vector bundles. In vector bundles, whose prototypes are the tangent bundles, the fibers are vector spaces. We shall then
proceed to the more involved bundle of frames, on which linear connections
play their game. The frame bundle is the prototype of principal bundles,
whose fibers are groups and which are the natural setting summing up all
differential geometry. The frame bundle provides the natural background for
General Relativity, and abstract principal bundles do the same for Gauge
Theory.
9.1.1 Intuitively, a fiber bundle is a manifold (the base) to every point
of which one glues another manifold (the fiber). For example, the sphere
S 2 and all the planes ( E2 ) tangent to it (the sphere tangent bundle); or the
same sphere S 2 and all the straight half-lines ( E1+ normal to it (the normal
bundle). The classical phase space of a free particle is a combination of the
configuration space (base) and its momentum space (fiber), but it is a simple
cartesian product and, as such, a trivial bundle (Physical Topic 1). Notice
however that the base and the fiber do not by themselves determine the
273
274
9.2
275
VECTOR BUNDLES
(9.1)
(9.2)
Steenrod l970.
276
(9.3)
Trautman l970.
Yang 1974.
277
depends, fundamentally, only on the topologies of the base and the group.
It does not depend on the fiber. Consequently, the classification can be
realized by taking into account only the principal bundles, in which the fiber
is replaced by the group itself (see section 9.7).
9.2.9 In gauge theories, the source fields belong usually to (associated)
vector bundles.
9.2.10 Fibration All the fibers are isomorphic to the typical fiber in a
fiber bundle. The notion may be generalized to that of a fibration, in which
the fibers are not necessarily the same on each point of the base space. Let us
briefly describe the idea. Given two spaces E and M , a fibration is a mapping
: E M possessing a property called homotopy lifting, which consists
in the following. Consider another space K, and a map f : K E. This
leads to the composition g = f : K M . The map f is the lift of g.
Consider now the homotopy class of g, a set of homotopic maps gt with g0 = g.
If this homotopy lifts to a homotopy ft of f , with f0 = f , then is said to
have the homotopy property (a fiber bundle is a particular case, the bundle
projection being a locally trivial fibration). Thus, the only requirement now
is that all the fibers 1 (x), for x M , be of the same homotopy type. An
example of recent interest is the loop space: start by taking as total space the
set M0I of curves x(t) with initial endpoint x0 = x(0). The final endpoint of
each curve will be x1 = x(1). Take as fibration the mapping : x(t) x1
taking each path into its final endpoint. The mapping 1 (x) is the set of
all curves c0 (x) from x0 to x. Now, choose some fixed path c0 (x) from x0 to
x. Any other path of 1 (x) will differ from c0 (x) by a loop through x0 ,
so that each new c0 (x) determines an element of the loop space LM. The
mapping 1 (x) is homotopically equivalent to LM. Thus, in the fiber above,
the initial endpoint is LM, and satisfies the fibration requirements.
9.3
There are many reasons to dedicate a few special pages to the bundle of linear
frames BM. On one hand, it epitomizes the notion of principal bundle and
is fundamental for geometry in general. This makes it of basic importance
for General Relativity, for example (Physical Topic 8). On the other hand,
precisely because of this particular link to the geometry of the base space, it
has some particular properties not shared by other principal bundles. The
intent here is to make clear its general aspects, which have provided the
historical prototype for the general bundle scheme, as well as stressing those
aspects well known to gravitation physicists which find no counterpart in the
bundles which underlie gauge theories.
278
(9.5)
(9.6)
Being a linear mapping between two vector spaces of the same dimension, b
is an isomorphism. It appoints the base member bj as the representative
of the canonical vector Kj belonging to the typical fiber Em . Consequently,
two vectors X and Y of T(b) M are images of two vectors r and s on Em :
X = b(r) = b(ri Ki ) = ri b(Ki ) = ri bi
and
Y = b(s) = b(si Ki ) = si b(Ki ) = si bi .
9.3.2 Structure group To see more of the action of the structure group
on BM, notice that GL(m, R) acts on the space Em on the left (actually,
we might build up an associated vector bundle with fiber Em ): if V Em ,
V = V k Kk , and a = (aj i ) GL(m, R),
(aV )j = aj i V i or aV = Kj aj i V i .
The mapping b will give
279
(9.7)
a Em
ba&
b
T(b) M
(9.8)
when applied to the group identity, gives just the frame b: b(e) = be = b.
The derivative mapping will be
b : Ta GL(m, R) Tba BM
b (Ja ) = Xba .
(9.9)
(9.10)
280
(9.12)
(9.13)
(9.15)
(9.16)
(9.18)
The name is not gratuitous. The presence of the solder form signals a coupling between the tangent spaces, to the bundle and to the base manifold,
which is much stronger for BM than for other principal bundles. A consequence is that a connection on BM will have, besides the curvature it shares
with all connections, another related form, torsion. Soldering is absent in
bundles with internal spaces, which consequently exhibit no torsion.
281
Figure 9.1:
9.3.4 Orthogonal groups Let us say a few more words on the relation
between the orthogonal groups and bilinear forms, introduced in 8.1.9. A
group of continuous transformations preserving a symmetric bilinear form
is an orthogonal group or, if the form is not positive-definite, a pseudoorthogonal group. If an element of the group is written , this defining
property takes the matrix form
T = ,
(9.19)
(9.20)
and will have trA = 0. Let us go back to the real linear group GL(m, R, and
the basis [9.10]. Given the bilinear form , both basis and entry indices can
be lowered and raised, as in (a b )i j = ai bj . One finds for instance
[ab , cd ] = bc ad da cb .
(9.21)
In this basis a member K = K ab ab of the algebra will have as components its own matrix elements: (K)ij = K ij . The use of double-indexed basis
for the algebra generators is the origin of the double-indexed notation (peculiar to the linear and orthogonal algebras) for the algebra-valued forms, as
282
(9.22)
All this leads to the usual commutation relations for the generators of
orthogonal and pseudo-orthogonal groups,
[J ab , J cd ] = bc J ad + ad J bc bd J ac ac J bd .
(9.23)
The usual group of rotations in 3-dimensional euclidean space is the special orthogonal group, indicated SO(3). Being special means connected
to the identity, that is, represented by 3 3 matrices of determinant = +1.
Orthogonal and pseudo-orthogonal groups are usually indicated by SO()
= SO(p, q), with (p, q) fixed by the signs in the diagonalized form of .
The group of rotations in n-dimensional euclidean space will be SO(n), the
Lorentz group will be SO(3, 1), etc.
9.3.5 Reduction We may in many cases replace GL(m, R) by some subgroup in such a way as to obtain a sub-bundle. The procedure is called
bundle reduction. For example, GL(m, R) can be reduced to the orthogonal subgroup O(m) (or to its pseudo-orthogonal groups). The bundle BM
of the linear frames reduces to the sub-bundle OM = (M, Op M, O(m), ),
where Op M is the set of orthogonal frames on Tp M . Now, Tp M is isomorphic
to Em , the typical fiber of the associated tangent bundle and on which there
exists an internal product which is just invariant under the action of O(m).
Let us consider a consequence of the reduction to SO().
9.3.6 Tetrads The most interesting point of reduction is that, in the
process, each basis of BM defines on M a Riemannian metric. Suppose on
Em the invariant internal product is given by the euclidean (or a pseudoeuclidean) metric , with
(r, s) = r s .
(9.24)
(9.25)
283
subset in BM formed by the b = (b1 , b2 , . . . , bm ) which are orthogonal according to g. The resulting bundle, OM , is the bundle of orthogonal frames on
M . In the case of interest for General Relativity, it is the pseudo-orthogonal
Lorentz group SO(3, 1) which is at work in the tangent space, and we must
take for the Lorentz metric of Minkowski E3,1 space. Of course, SO(3, 1)
is also a subgroup of GL(4, R). Given a natural basis { }, a general basis
{h } of BM has elements
h = h ,
(9.26)
and its dual basis is {h },
h = h dx
(9.27)
h (h ) = h h = .
(9.28)
with
To reduce to the bundle OM , we impose orthogonality by some g:
g(h , h ) = g h h =
(9.29)
(9.30)
284
9.4
LINEAR CONNECTIONS
Connections materialize in the notion of parallel transport. Consider a (piecewise differentiable) curve on a manifold. A vector X undergoes parallel
transport if it is displaced along in such a way that its angle with the curve
(i.e. with the tangent to ) remains constant (Figure 9.2). This intuitive view
supposes a metric (see 9.4.23), but actually a connection suffices. A connection determines a covariant derivative of the type described in 7.3.11, and
the vector is parallel-transported when its covariant derivative vanishes. The
covariant derivative can be extended to any tensor, and a tensor is paralleltransported when its covariant derivative vanishes. The notion of parallel
transport for general principal bundles will be introduced (in 9.6.21) in an
analogous way.
9.4.1 Ask a friend to collaborate in the following experiment.5 He must
stand before you, with his arm straight against his side but with his thumb
pointing at you. He must keep rigidly the relative position of the thumb with
respect to the arm: no rotation of thumb around arm axis allowed. He will
then (i) lift the arm sideways up to the horizontal position; (ii) rotate the
arm horizontally so that it (the arm) points at you at the end; you will be
seeing his fist, with the thumb towards your right; (iii) finally he will drop
the arm back to his side. The thumb will still be pointing to your right. The
net result will be a 90 rotation of the thumb in the horizontal plane. Notice
that his hand will have been moving along three great arcs of a sphere S 2 of
radius L (the arms length). Its is just as if you looked at the behaviour of a
vector on earth: (a) initially at the south pole S and pointing at you (see
Figure 9.3); (b) transported
Figure 9.2:
285
Figure 9.3:
along the rim from S to H, all the time pointing at you; (c) taken along
the equator to the meridian just facing you (the vector becomes progressively visible and, at F , it will be entirely toward your right; (d) transported
southwards back to S. The net rotation is a measure of earths curvature.
For a small parallel displacement dxk , the variation of a vector X will be
given by
X i = i jk X j dxk ,
where i jk represents precisely the connection. Along a curve, one must
integrate. It so happens that the curvature is the rotational of , and that,
in the case above, the curve bounds one-eigth of the whole earths surface,
4L2 (see details in Mathematical Topic 10.3.1). Stokes theorem transforms
the line integral into the surface integral of the curvature over the octant.
The curvature is constant, and equal to 1/L2 , so that what remains is just
an octants area divided by L2 , which gives the rotation angle, /2.
The same procedure, if followed on the plane, which is a flat (zero curvature) space, would take the vector back to its initial position quite unmodified.
9.4.2 There is no such a thing as curvature of space. This is perhaps
still more evident when general connections are introduced (section 9.6).
Curvature is a property of a connection, and a great many connections may
be defined on the same space. Different particles feel different connections
and different curvatures. There might be a point for taking the Levi-Civita
connection as part of the very definition of spacetime, as is frequently done. It
seems far wiser, however, to take space simply as a manifold, and connection
(with its curvature) as an additional structure (see Physical Topic 8.2). We
shall here be concerned with the general notion of linear connection and
286
(9.31)
(9.32)
(9.33)
Notice that the fields Ei are exactly dual to the solder forms. Also
(Ei ) = b (Ei ) = bi .
Thus, on the base manifold and in a given basis {bi }, the soldering forms are
represented by that base of forms which is dual to {bi }.
287
(9.34)
(9.35)
(9.36)
(9.37)
The notation is not without a purpose. The detailed calculations give for
the coefficients Fm n ij and T k ij the values of the curvature and the torsion,
whose meaning we shall examine in the following. Let us retain here that the
curvature appears as the vertical part of the commutator of the basic fields,
and the torsion as its part along themselves.
The projection is a linear mapping from the start, which becomes
an isomorphism of vector spaces (between the horizontal subspace Hb and
T(b) M ) when a connection is added. The decomposition of Tb BM into horizontal and vertical is, therefore, complete in what concerns the vector space,
but not in what concerns the algebra.
288
(9.38)
It is consequently given by
T (X, Y ) = d(HX, HY ) = Ki di (X j Ej , X k Ek ).
We find that
T = 21 Kk T k ij i j .
(9.39)
(9.40)
(9.41)
y
,
x
which furthermore can be chosen to be locally the same for the manifold and
for the tangent space, h = , then
T = .
(9.42)
(9.43)
(9.44)
R = + .
(9.45)
In a natural basis,
289
(9.46)
(9.47)
(9.48)
(9.49)
dX k = k ij X i d j .
(9.50)
or
A geodesic curve is a self-parallel curve, that is, a curve along which its
i
is parallel-transported. It obeys conseown tangent vector (velocity) d
ds
quently the geodesic equation
d2 k
ds
+ k ij d
ds
d j
ds
= 0.
9.4.13 More about geodesics is said in Mathematical Topic 12. Here, only
a few general aspects of them will interest us. Geodesics provide an easy
view of the meaning of curvature. Consider two bits of geodesics, d and
ds starting at O as in Figure 9.4. Displace d along d to obtain d0 , and d
along d to obtain d 0 , thereby constituting an infinitesimal parallelogram.
Take then a field X. If we parallel-transport X along d, it will change
according to [9.50]; if we propagate the resulting field along d 0 , it will again
be changed analogously. We shall find a vector X 0 at point c. Now start
again from O, going however first along d and then along d0 . We find X 00
at point c. The difference will then be fixed by the curvature:
X k = Rk ipq X i dp d q .
(9.51)
290
Figure 9.4:
9.4.14 Geodesics allows also an intuitive view of the effect of torsion, and
this in a rather abrupt way: torsion disrupts the above infinitesimal geodesic
parallelograms. In effect, if we transport as above the very geodesic bits in
the presence of torsion (as in Figure 9.5), we find that a gap between the
extremities shows up, and such that
k = [k ij k ji ] d i dj = T k ij d i dj .
(9.52)
291
Figure 9.5:
(9.54)
(9.55)
and
(Z T )(X, Y ) R(X, Y )Z + T (T (X, Y ), Z) + (cyclic permutations) = 0.
(9.56)
Notice that, even when T = 0, neither is trivial. The covariant derivative
has the properties
f X Y = f X Y,
(9.57)
X+Z Y = X Y + Z Y.
(9.58)
(9.59)
Let us profit to give some explicit expressions. Take a vector field basis
{ea }, with [ea , eb ] = f c ab ec and find first that
X Y = X a [ea Y c + Y b c ba ] ec ,
(9.60)
(9.61)
292
Then, we calculate
X Y = X a ea (Y b eb ) = X a [ea Y c + Y b c ba ]ec .
(9.62)
We find next:
[X, Y ] = [X a ea , Y b eb ] = [X a ea (Y c ) Y a ea (X c ) + X a Y b f c ab ]ec ;
(9.63)
(9.64)
(9.65)
and
Finally, using eq.[9.59], we obtain
Rf cab = ea f cb eb f ca + d cb f da d ca f db f g ab f cg .
(9.66)
(9.68)
(9.69)
(9.70)
293
294
Dt = lim
(9.71)
Figure 9.6:
The covariant derivative in the direction of a field X at a point p is the
covariant derivative along a curve which is tangent to X at p. From the
above definition of covariant derivative, a section (in general, a tensor) is
said to be parallel-displaced along a curve iff the corresponding covariant
derivative vanishes. When its covariant derivative is zero along any field,
it will be parallel-transported along any curve. In this case, we say that
the connection preserves the tensor. The definition is actually an usual
295
derivative, only taking into account the whole, invariant version of a tensor.
This means that, besides derivating the components, it derivates also the
basis members involved. Take a tensor like T = T e w . The covariant
derivative will be
DT = dT e w + T de w + T e dw .
Using eqs.[7.67] and [7.73] for the adapted frames of Mathematical Topic
10.1.3, it becomes
DT = [e (T ) + T T ] w e w .
The covariant derivative along a curve will be the contraction of this
derivative with the vector tangent to the curve at each point (that is, if u is
the tangent field, its index will be contracted with the derivative index). In
the above example, it will be
Du T = u [e (T ) + T T ] e w .
9.4.22 In this way, the above notion of covariant derivative applies to
general tensors, sections of associated bundles of the frame bundle, and gives
the usual expressions in terms of components (say) in a natural basis, duly
projected along the curve, that is, contracted with its tangent vector. Say,
for a covariant vector, we find the expressions used in 6.6.14,
D X = X; = X X .
(9.72)
This semicolon notation for the covariant derivative, usual among physicists, does not include the antisymmetrization. In invariant language,
DX = 21 D[ X] dx dx .
For a contravariant field,
D X = X; = X + X .
(9.73)
(9.74)
This will vanish when the connection preserves the metric. The compo
nents of the torsion tensor in a natural basis are T
= . In
principle, there exists an infinity of connections preserving a given metric,
296
but only one of them has vanishing torsion. In this case, the connection is
symmetric in the lower indices and we can solve the above expression to find
= = 12 g [ g + g g ],
(9.75)
just the Christoffel symbol of eq.[6.74]. Summing up: given a metric, there
exists a unique torsionless connection which preserves it, whose components
in a natural basis are the usual Christoffel symbols and is called the LeviCivita connection of the metric. Usual Riemannian curvature is the curvature of this connection, which is the connection currently used in General
Relativity (see Phys.Topic 8). The hypothesis of gravitation universality
gives priority to this connection, as it says that all particles respond to its
presence in the same way.
It is with respect to this connection that parallel transport acquires the
simple, intuitive meaning of the heuristic introduction: a vector is paralleldisplaced along a curve if its modulus and its angle with the tangent to the
curve remain constant. Of course, measuring modulus and angle presupposes
the metric. And it is the curvature of this connection which is meant when
one speaks of the curvature of a (metric) space. The discovery of curved
spaces, or non-euclidean geometries (Math.Topic 11), has been historically
the germ of modern geometry.
9.4.24 Consider a manifold M and a point p M . We define the symmetry sp at p as a diffeomorphism of a neighbourhood U of p into itself which
sends exp(X) into exp(X) for all X Tp M . This means in particular that
normal coordinates change signs. When such a symmetry exists, the space
is said to be locally symmetric.
Suppose then that a linear connection is defined on M . We denote M
with this fixed connection by (M, ). A differentiable mapping f of M into
itself will be an affine transformation if the induced mapping f : T M
T M maps horizontal curves into horizontal curves. This means that f maps
each parallel vector field along each curve into a parallel vector field along
the curve f (). The affine transformations on M constitute a Lie group. If
the symmetry sp above is an affine transformation, (M, ) is an affine locally
symmetric manifold. This only happens when T = 0 and R = 0.
On the other hand, (M, ) is said to be an affine symmetric manifold if,
for each p M , the symmetry sp can be extended into a global affine transformation (compare with section 8.2.7). On every affine symmetric manifold
M the group of affine transformations acts transitively. Thus, M may be
seen as a homogeneous space, M = G/H. The connection on G will be the
torsion-free connection above referred to.
9.5
297
PRINCIPAL BUNDLES
(9.76)
(9.77)
The space F may be, for example, a space of column vectors on which G is
represented by n n matrices (g):
gf := (g)f.
(9.78)
298
A bundle (M, F, G, ) is got in this way, which is associated to (M, G, ). Notice that different representations lead to different associated bundles. There
are therefore infinite such bundles for each group.
9.5.3 Locally, a point in (M, F, G, ) will be represented by coordinates
(p, f ) = (x1 , x2 , . . . , xm , f 1 , . . . , f n ). A LSC transformation will lead to some
(p0 , f 0 ). Both f and f 0 belong to F . If the action of G on F is (simply)
transitive, there will be a (unique) matrix (g) such that f 0 = (g)f . Thus,
the group action accounts for the LSC transformations in the fiber.
9.5.4 A point on the principal bundle will be found by a section :
(p, f ) = (p),
(p, f ) = (p) = p.
(9.79)
9.5.5 Let us now proceed to the formal definition. It requires a lot of things
in order to ensure that the bundle as a whole is a differentiable manifold.
A C principal fiber bundle is a triplet
P = (M, G, )
such that P (the complete space) and M (the base space) are C differentiable manifolds, and G (the structure group) is a Lie group satisfying the
following conditions:
(1) G acts freely and effectively on P on the right, Rg : P G P ;
this means that no point of P is fixed under the action, and no subgroup of
G is the stability group of some point of P ;
(2) M is the quotient space of P under the equivalence defined by G,
M = P/G; the projection : P M is C ; for each p of M , G is simply
transitive on the fiber 1 (p); so, the fiber is homogeneous, and we say that
the group preserves the fiber;
(3) P is locally trivial: for every p of M , there exists a neighbourhood
U 3 p and a C mapping
FU : 1 (U ) G
such that FU commutes with Rg for every g in G; the combined mapping
fU : 1 (U ) U G
fU (b) = ((b), FU (b))
(9.80)
299
With the first condition, we may generalize here the mapping b of eq.[9.6].
When the action is free and effective, there is a Lie algebra isomorphism
between the group Lie algebra and the tangent to the fiber ( 8.4.8). Take
an associated bundle AM , with a typical fiber F which is usually a vector
space and whose copies in the bundle we shall call realized fibers. Choose
a starting basis on F . A point b of the principal bundle may be seen as a
mapping from F into the realized fiber on (b), with image zb = Lb z0 , and
with z0 indicating the zero-section, which will be defined in 9.5.8. The
group identity will deputize ze = Le z0 = z0 as the set of representatives of
the starting basis members. Each member of F will be thus translated into a
point of the realized fiber, and each point z of the realized fiber will come
from a member b1 (z) of F , its paradigm. The typical fiber F is in this
way installed on (b).
Figure 9.7:
9.5.6 We have seen in sections 8.2 and 8.4 that, when a group G acts on
a manifold M , each point p of M defines a mapping
pe : G M, pe(g) = pg = Rg (p).
If the action is free and we restrict ourselves to orbits of p, this mapping
is a diffeomorphism. Well, given a point b in a fiber, every other point in
the fiber is on its orbit since G acts in a simply transitive way. Thus, eb is a
diffeomorphism between the fiber and G,
eb : G 1 ((b)) P
eb(g) = Rg (b) = bg.
(9.81)
300
9.5.7 To say that M = P/G is to say that (bg) = (b). To say that FU
commutes with Rg means that FU (bg) = FU (b)g, or
FU Rg (b) = Rg FU (b).
As a consequence (see Figure 9.8),
Rg (Xb ) = (Xb )
FU Rg = Rg FU .
(9.82)
(9.83)
Figure 9.8:
(9.84)
301
(9.85)
(9.86)
Notice that something like [FU (b)]1 will always take b into the point
(p, e), in the respective trivialization, but the point corresponding to the
identity e may be different in each trivialization. And, as each chart (U, x)
around p will lead to a different trivialization, the point on the fiber corresponding to e will be different for each chart. It is usual to write FU (x),
where x Em is the coordinate of p = (b) in the chart (U, x).
The bundle commonly used in gauge theories has an atlas (Ui , x(i) ) with
all the Ui identical. Changes of LSC reduce then to changes in the coordinate
mappings, and the transition functions gU V represent exactly the local gauge
transformations,6 which correspond here to changes of zero sections. From
b = U (p)FU (b) and V (p) = bFV1 (b)
it follows that
V (p) = U (p)FU (b)FV1 (b) = U (p)gU V (p),
(9.87)
which shows precisely how sections change under LSC transformations. Equation [9.85] says else that, given in the fiber,
U = gU V V .
(9.88)
j .
(9.89)
(9.90)
302
This means that LSC transformations are at work only in the internal
spaces, the fibers. The fields carry a representation of the gauge group
in this way. In the fiber, a change of LSC can be looked at in two ways:
either as a coordinate change as above or as a point transformation with
fixed coordinates. The non-triviality appears only in the point dependence
of the transition function.
9.5.10 Giving the coordinate neighbourhoods and transition functions
completely characterizes the bundle. It is possible to show that:
(i) if either G or M is contractible, the bundle is trivial;
(ii) if a principal bundle is trivial, every bundle associated to it is trivial.
9.5.11 Sub-bundles Fiber bundles are differentiable manifolds. Can we
introduce the notion of immersed submanifolds, while preserving the bundle
characteristics? In other words, are there sub-bundles? The answer is yes,
but under rather strict conditions. We have seen a particular case in 9.3.5.
As there is a lot of structure to preserve, including group actions, we must
begin by defining homomorphisms between bundles. Given two bundles P
and P, with structure groups G and G, a bundle homomorphism between
them includes a mapping f : P P 0 and a group homomorphism h : G
G0 , with
f (bg) = f (b)h(g).
(9.91)
If f is an immersion (an injection) and h is a monomorphism (an injective
homomorphism), then we have an immersion of P in P . In this case, P is a
sub-bundle of P . If furthermore P and P have the same base space, which
remains untouched by the homomorphism, then we have a group reduction,
and P is a reduced bundle of P .
9.5.12 Induced bundles We may require less than that. Suppose now
a bundle P with base space B. If both another manifold B 0 and a continuous mapping f : B 0 B are given, then by a simple use of function
compositions and pull-backs one may define a projection, as well as charts
and transition functions defining a bundle P 0 over B 0 . It is usual to say that
the map f between base spaces induces a mapping f : P 0 P between
complete spaces and call P 0 = f P the induced bundle, or the pull-back
bundle. Suppose there is another base-space-to-be B 00 and another analogous
map f 0 : B 00 B leading to a bundle P 00 over B 00 in just the same way. If
B 0 = B 00 and the maps are homotopic, then P 0 and P 00 are equivalent. Such
maps are used to show the above quoted results on the triviality of bundles involving contractible bases and/or fibers. They are also used to obtain
general bundles as induced bundles of Stiefel manifolds, which allows their
classification ( 9.7.2)
303
9.5.13 It might seem that the above use of a general abstract group is farfetched and that Physics is concerned only with transformation groups acting
on physical spaces, such as spacetime and phase spaces. But the above
scheme is just what appears in gauge theories (Physical Topic 7). Gauge
groups (usually of the type SU (N )) are actually abstract, acting on some
internal spaces of wavefunctions defined on Minkowski base space. The
first statement in 9.5.10 would say that, if Minkowski space is contractible,
the bundles involved in gauge theories are trivial if no additional constraints
are imposed via boundary conditions.
9.6
GENERAL CONNECTIONS
(9.92)
304
(9.93)
Each Ja will be a vertical field and the algebra generated by {Ja }, which
is tangent to the fiber, will be isomorphic to G0 and will represent it on P .
Given a vertical field X, there is a unique X in G0 such that X = eb (X).
The field X is said to be engendered by X, which it represents on P .
The set {Ja } may be used as a basis for Vb .
9.6.3 3.A fundamental field X, under a group transformation, will change
according to eq.[8.35]:
Rg (X) = Adg1 (X) = eb [Adg1 (X)].
(9.94)
(9.95)
305
(9.97)
(9.98)
A field Xb on the complete space will thus have two contributions: this one
from Tp M , and another, purely vertical, coming from G as in eq.[9.95]:
bb + X b .
Xb = Rg U (Yp ) + eb (Xe ) = X
(9.99)
bb is
Now comes the trouble: X b is purely vertical by construction, but X
not necessarily purely horizontal! It may have a vertical part (Figure 9.9).
The total vertical component will be, consequently, X b plus some contribubb .
tion from X
Figure 9.9:
306
or Z = Xb .
(9.100)
(9.101)
(J a ) = Ja .
(9.102)
(9.103)
a (J b ) = ba .
(9.104)
It is, in a restricted sense, the inverse of eb . It is equivariant ( 8.2.19)
because it transforms under the action of G in the following way: for any
fundamental field X,
(Rg )(X) = (Rg X) = (Adg1 X) =
eb (Adg1 X) = Adg1 X = Adg1 X(X).
307
Therefore,
Rg = Rg = Adg1 ,
(9.105)
(9.106)
or
Figure 9.10:
308
9.6.11 Well, all this may be very beautiful, but is a form on P , it inhabits (the cotangent bundle of) the bundle space. We need a formulation
in terms of forms on the base space M . In order to bring down to M ,
we will be forced to resort to the use of sections the only means to pull
back. Let us consider two zero sections related to two charts on M , with
open sets U and V . They will be related by eq.[9.87],
V (p) = U (p)gU V (p),
(9.107)
where the transition function gU V , given by [9.85], mediates the transformation of section U to section V . It is a mapping
gU V : U V G.
Its differential will take a field X on M into a field on G,
gU V : Tp M TgU V G
gU V : X gU V (X).
(9.108)
(9.109)
(9.110)
V = V
(9.111)
Notice that U and V are forms (only locally defined) on M with values
in G.
9.6.13 Now, let us take the differential of [9.107] by using the Leibniz
formula, and apply the result to a field X on M :
V (X) = U (X)gU V (p) + U (p)gU V (X),
(9.112)
309
(9.113)
When we apply to this field on P , the first term on the right-hand side
will be
{RgU V [U (X)]} = (RgU V )(U X) = (Adg1 )U X
UV
UV
The second term, on the other hand, will be {gU V (X)} by eq.[9.110].
This will be a certain field of G, in reality the field gU V (X) on G brought to
G0 , of which gU V (X) = U gU V (X) is the fundamental representative field
on P . A field on G is brought to G0 by the Maurer-Cartan form w, so that
{gU V (X)} = w[gU V (X)] = (gU V w)(X) = wU V (X).
Consequently, (V (X)) = (V )(X) will be
V (X) = Adg1 U (X) + wU V (X),
(9.114)
V = Adg1 U + wU V .
(9.115)
UV
or
UV
(9.117)
Ja 0a = g 1 Ja ga + g 1 g.
(9.118)
and
310
9.6.14 In these expressions the reader will have recognized the behaviour
of a gauge potential under the action of a gauge transformation, or the change
in the Christoffeln due to a change of basis. Here, the small correction we
promised in 9.6.10: gauge potentials are pulled-back connections on principal fiber bundles with the gauge group as structure group and space-time
as base space.7 They are defined on the base space and so they are sectiondependent. A section is what is commonly called a gauge. Changes of
sections are gauge transformations. The geometrical interpretation of the
underlying structure of gauge theories, pioneered by Trautman and a bit resented at first, is nowadays accepted by everybody.8 It helps clarifying many
important points. Let us only call attention to one of such. The vacuum
term g 1 dg in [9.116] is a base-space representative of the Maurer-Cartan
form, eq.[9.109]. This form is a most important geometrical characteristic of
the group, in reality connected to many of its topological properties. The
vacuum of gauge theories is thereby strongly related to the basic properties
of the gauge group.
9.6.15 A 1-form satisfying condition [9.101] is said to be a vertical form.
On the other hand, a form on P which vanishes when applied to any vertical
field,
(Vp X) = 0, X,
(9.119)
is a horizontal 1-form. Clearly, the canonical form [9.103] on the frame
bundle is horizontal. Vertical (and horizontal) forms of higher degrees are
those which vanish when at least one horizontal (respectively, vertical) vector
appears as their arguments.
9.6.16 Given a connection , horizontal spaces are defined at each point
of P . Given a p-form on P with values in some vector space V , its absolute
derivative (or covariant derivative) according to the connection is the (p +
1)-form
D = Hd = d H,
(9.120)
where H is the projection to the horizontal space:
D(X1 , X2 , . . . , Xp+1 ) = d(HX1 , HX2 , . . . , HXp+1 ).
(9.121)
The covariant derivative is clearly a horizontal form. An important property of D is that it preserves the representation: if belongs to a representation, so does D. For example, if
Rg = Adg1 ,
7
8
Trautman l970.
Daniel & Viallet l980; Popov l975; Cho l975.
311
then also
Rg D = Adg1 D.
(9.122)
(9.124)
(9.125)
(9.126)
F 0 = g 1 F g.
(9.127)
or
(9.128)
A careful case study for the vertical and horizontal components of X and Y
leads to9
F (X, Y ) = d(X, Y ) + 12 [(X), (Y )],
(9.129)
which is the compact version of Cartans structure equations. This is to
be compared with the results of 7.3.12: there, we had a zero-curvature
connection (also called a flat connection) on En . A better analogy is found
in submanifolds (Mathematical Topic 10).
9
312
(9.130)
(9.131)
(9.132)
F a = a a + C a bc b c .
(9.133)
where
313
Figure 9.11:
(9.134)
(9.135)
that is, when D0 at x. What is the meaning of it? For linear connections
and tangent vectors, this generalizes the usual notion of parallelism in euclidean spaces. In gauge theories, in which the associated fibers contain the
source fields, it represents a sort of internal state preservation. Suppose
a gauge model for the group SU (2), as the original Yang-Mills case. The
nucleon wavefields will have two components, being isotopic-spin Pauli
spinors. Suppose that at point x the field is a pure up spinor, representing a proton. If D = 0 at x, then (x + dx) given by eq.[9.135] will also be
a pure proton at (x + dx). All this is summed up again in the following: all
over a parallel transport of an object, the corresponding element in typical
fiber stays the same (Figure 9.12).
314
Figure 9.12:
315
Berry 1984.
Simon 1983.
316
the role of an effective vector potential. This is of course a very crude example, which is far from doing justice to a beautiful and vast subject.12 Even
in a fixed problem, what is a parameter and what is a variable frequently
depends on the conditions. And it was found later13 that this geometrical
phase (sometimes also called holonomy phase) can appear even in the absence of parameters, in the time evolution of some systems. There is now a
large amount of experimental confirmation of its existence in many different
physical situations.
9.7
BUNDLE CLASSIFICATION
Let us finish the chapter with a few words on the classification of fiber bundles. Steenrods theorem ( 9.2.3) is a qualitative result, which says that
there is always at least one bundle with a base space M and a group G. We
have seen ( 9.1.1) that with the line and the circle at least two bundles are
possible: the trivial cylinder and the twisted Mobius band. Two questions14
come immediately to the mind, and can in principle be answered:
(1) In how many ways can M and G be assembled to constitute a complete
space P ? The answer comes out from the universal bundle approach.
(2) Given P , is there a criterion to measure how far it stands from the
trivial bundle? The answer is given by the theory of characteristic classes.
Notice to begin with that each associated bundle is trivial when the principal bundle P is trivial. Consequently, an eventual classification of vector
bundles is induced by that of the corresponding principal and we can concentrate on the latter.
9.7.1 Back to homogeneous spaces Let us recall something of what
was said about homogeneous spaces in 8.2.7. If M is homogeneous under
the action of a group G, then we can go from any point of M to any other
point of M by some transformation belonging to G. A homogeneous space
has no invariant subspace but itself. Simple Lie groups are homogeneous
by the actions (right and left) defined by the group multiplication. Other
homogeneous spaces can be obtained as quotient spaces. The group action
establishes an equivalence relation, p q, if there exists some g G such
that q = Rg p. The set
[p] = { q M such that q p } = orbitG (p)
12
13
14
317
is the equivalence class with representative p. The set of all these classes is the
quotient space of M by the group G, denoted by M/G and with dimension
given by: dim M/G = dim M - dim G. The canonical projection is
: M M/G,
p [p].
All the spheres are homogeneous spaces: S n = SO(n + 1)/SO(n). Consider in particular the hypersphere S 4 = SO(5)/SO(4). It so happens that
SO(5) is isomorphic to the bundle of orthogonal frames on S 4 . We have thus
a curious fact: the group SO(5) can be seen as the principal bundle of the
SO(4)-orthogonal frames on S 4 . This property can be transferred to the de
Sitter spacetimes (Physical Topic 9): they are homogeneous spaces with the
Lorentz group as stability subgroup, respectively
DS(4, 1) = SO(4, 1)/SO(3, 1) and DS(3, 2) = SO(3, 2)/SO(3, 1).
And the de Sitter groups are the respective bundles of (pseudo-) orthogonal
frames. Can we generalize these results? Are principal bundles always related
to homogeneous spaces in this way? The answer is that it is almost so. We
shall be more interested in special orthogonal groups SO(n). In this case, we
recall that the Stiefel manifolds ( 8.1.15) are homogeneous spaces,
Snk = SO(n)/SO(k).
They do provide a general classification of principal fiber bundles.
9.7.2 Universal bundles Given a smooth manifold M and a Lie group
G, it is in principle possible to build up a certain number of principal fiber
bundles with M as base space and G as structure group. One of them is
the direct product, P = M G. But recall that the general definition of
a principal bundle includes the requirement that M be the quotient space
of P under the equivalence defined by G, that is, M = P/G. Let us take
a bit seriously the following easy joke. Say, to start with, that principal
bundles are basically product objects obeying the relation M = P/G, which
are trivial when in this relation we can multiply both sides by G in the nave
arithmetic way to obtain P = M G. Thus, nontrivial bundles are objects
P = M G, where is a twisted product generalizing the cartesian
product as the Mobius band generalizes the cylinder. how many of such
twisted products are possible? Answering this question would correspond
to somehow classifying the principal fiber bundles. It is actually possible to
obtain a homotopic classification of the latter, and this is done through a
construct which is itself a very special fiber bundle. This special fiber bundle
is the universal bundle for G.
318
Figure 9.13:
319
320
or how far it stands from the direct product. To obtain them, one first
finds them for the universal bundle, thereby estimating how far it is from
triviality; then one pull the forms back to the induced bundles; as the pullback preserves the cohomological properties, the same twisting measure
holds for the bundle of interest. Thus, the first task is to determine the
cohomology groups H n [O(m)/GxO(m d)]. For the most usual groups,
such classes are named after some illustrious mathematicians. For G = O(n)
and SO(n), they are the Pontryagin classes, though the double covering of
special groups produce an extra series of classes, the Euler classes. For U (n),
they are the Chern classes. These are the main ones. Other may be arrived
at by considering non-real cohomologies.
Part III
FINAL TOUCH
321
Chapter 10
NONCOMMUTATIVE GEOMETRY
Only a glimpse into promising new developments.
10.1
10.1 Think of a group of usual matrices. Matrices in general do not commute. Their entries consist of real or complex numbers ai j . We are used to
multiplying matrices (that is, to perform the group operation) and, in doing
so, the entries themselves get multiplied according to the well-known rules.
Complex numbers commute with each other and we do not trouble with their
order when multiplying them: ai j am n = am n ai j .
10.2 Matrix groups are Lie groups, that is, smooth manifolds whose points
are the group elements. To each group element, a point on the group manifold, corresponds a matrix. Thus, each matrix will have its coordinates.
What are the coordinates of a matrix? Just the entries am n if they are real
and, if they are complex, their real and imaginary parts ( 6.4.3 and 6.5.8).
Thus, although themselves noncommuting, matrices are represented by sets
of commuting real numbers. Their non-commutativity is embodied in the
rules to obtain the entries of the product matrix from those of the matrices
being multiplied.
Suppose now that, in some access of fantasy, we take the very entries
i
a j as noncommutative. This happens to be not pure folly, provided some
rules are fixed to ensure a minimum of respectability to the new structure
so obtained. With convenient restrictions, such as associativity, the new
structure is just a Hopf algebra (Math.1.25).
1
Connes 1990.
323
324
Woronowicz uses the term pseudo-group. Concerning the name quantum groups,
the physicists are not alone in their guilt: people looking for material on Hopf algebras
in the older mathematical literature will have to look for annular groups ... The best
name is probably bialgebras.
3
McGuire 1964.
4
See different contributions in Yang & Ge 1989.
325
10.6 A first idea would be to take the above noncommuting matrix elements also as matrices. Matrices of matrices then turn up. This is where
the large use of direct-product matrices ( Math.2.10) comes out. In diagonal block matrices, the blocks work independently from each other and
will correspond to abelian bialgebras. Drop the diagonal character and you
will have general block matrices. By the way, finite von Neumann algebras
are precisely algebras of block matrices (well, actually almost everything a
physicist bothers about is a von Neumann algebra. The novelty is that up
to recent times most of these were commutative).
10.7 Groups are usually introduced in Physics as sets of transformations
on some carrier space. Would quantum groups also preside over transformations? If so, the above matrices with non-commuting entries would be
expected to act upon column vectors, these with non-commuting components. This is actually so,5 and the study of the carrier spaces (called quantum spaces, or Manin spaces) seems simpler than that of the Hopf algebras
themselves.6 The general case is, however, very involved also.7
10.8 There are other gates into the realm of bialgebras. Hopf introduced
them originally in homology theory, but other simpler cases have been found
since then. One may, for instance, proceed in a way similar to that used to
introduce the classical groups as transformation groups ( 8.1.9). These are
sets of transformations preserving given sesquilinear forms.8 This is probably
the most apealling approach for physicists. In knot theory, they appear in
presentations of groups and algebras. In physics, the original approach was
related to Lie algebras deformations.
Classical Sine-Gordon equation is related to the Lie algebra sl(2, R). Once
quantized, instead of this algebra, this integrable equation exhibitted another
structure, which was recognized by Kulish and Reshetikin as a deformation
of sl(2, R), and called it its quantum version. Drinfeld9 has then given
the new structure a general descripton, through the consideration of phase
spaces defined on Lie groups (see Phys.10).
10.9 A most interesting approach is that pioneered by Woronowicz,10
which we could call the Fourier gate. It relates to harmonic analysis on
groups. It goes from the Pontryagin duality for abelian groups, through
5
6
7
8
9
10
Manin 1989.
See for instance Fairlie, Fletcher & Zachos 1989, and references therein.
See Ocneanus postface to Enoch & Schwartz 1992.
Dubois-Violette & Launer 1990.
Drinfeld 1983.
Woronowicz 1987.
326
the Tanaka-Krein duality for non-abelian groups, to still more general theories (see Math.6.14 and those following it). The whole subject is very
involved in its formal aspects, and still a research subject for physicists and
mathematicians.11 It involves deep points of the theory of Banach algebras
(Math.5) and Hopf-Banach algebras. We retain here only the point that coordinates can become non-commutative and proceed to a heuristic discussion
of a formalism well known to physicists.
10.2
QUANTUM GEOMETRY
An idea of the present state of affairs can be got in Gerstenhaber & Stasheff 1992.
Dirac 1926.
327
Dubois-Violette 1991.
Coquereaux 1989.
15
In this chapter, the symbol is taken for the so-called star-product of quantum
Wigner functions. It has nothing to do with its previous use for the composition of
mapppings. It is the Fourier transform of the convolution, for which the symbol * is
used of old.
14
328
(10.3)
AW = F [Fb1 [A]].
(10.4)
and
The Wigner functions are, despite their c-number appearence, totally quantum objects. They are representatives of quantum quantities (they will include powers of ~, for example) and only become classical in the limit ~ 0,
when they give the corresponding classical quantities. They embody and
materialize the correspondence principle. The densities
A = F 1 [AW ] = Fb1 [A]
16
17
(10.5)
329
include usually Dirac deltas and their derivatives. All this may seem rather
strange, that c-number functions can describe Quantum Mechanics. Actually, this is a lie: the Wigner functions are c-number functions indeed, but
they do not multiply each other by the usual pointwise product. In order to
keep faith to the above correspondence rule with quantum operators, a new
product has to be introduced, as announced in 10.12. This means that
we can describe quantum phenomena through functions, provided we change
the operation in their algebra. The product related to quantization is
called (rather unfortunately) the star-product.18 The simplest way to introduce it is by changing the usual multiplication rule of the Fourier basic
functions
(a,b) (q, p) = ei2(aq+ibp)/h .
(10.6)
We impose
(a,b) (q, p) (c,d) (q, p) = ei(adbc)h/4 (a+c,b+d) (q, p)
(10.7)
(10.8)
The last expression says that the functions provide a basis for a representation of the group R2 , which is self-dual under Fourier transformations
( Math.6.13). Imposing equation [10.7], with the extra phase, means to
change R2 into the Heisenberg group ( 8.3.13). This is enough to establish
a complete correspondence between the functions and the quantum operators. The commutator of two functions, once the new product is defined, is
no longer trivial:
{AW , BW }M oyal (q, p) = [AW , BW ] (q, p)
Z
h
i 4
2 2
dxdydzdwA(x,y) B (z,w)
sin[ (yz xw)](x+z,y+w) (q, p).
=( )
h
h
4
(10.9)
This quantum bracket, called Moyal bracket after its discoverer, allows in
principle to look at quantum problems in a way analogous to hamiltonian
mechanics. Derivatives are introduced through the bracket: for instance,
{q, F }M oyal is the derivative of F with respect to the function q. A symplectic structure appears naturally,19 whose differences with respect to that
of classical mechanics (shown in Physical Topic 1) correspond exactly to the
18
19
330
quantum effects. Calculations are of course more involved than those with
the Poisson bracket. Due to the star product, for instance, the derivative of
p with respect to q is no more zero, but just the expected {q, p} = i~. If
we want to attribute coordinates to the quantum phase space, the only way
to keep sense is to use no longer the usual euclidean space with commutative numbers, but to consider coordinate functions with a product. In
this picture, we repeat, quantization preserves the dynamical functions, but
changes the product of their algebra.
10.14 And here comes the crux: in the limit ~ 0, the Moyal bracket
gives just the Poisson bracket. The strangeness of the Poisson bracket is
thus explained: though not of purely algebraic content, it is the limit of an
algebraic and nontrivial bracket coming from a non-commutative geometry.
The central algebraic operation of Classical Mechanics is thus inherited from
Quantum Mechanics, but this only is seen after the latter has been given
its full geometrical content. There is much to be done as yet, mainly when
we consider more general phase spaces, and a huge new field of research on
noncommutative geometrical structures is open.
All this is to say that Physics has still much to receive from Geometry.
It is an inspiring fact that an age-old structure of classical mechanics finds
its explanation in the new geometry.
We shall not cease from exploration
And the end of all our exploring
Will be to arrive where we started
And know the place for the first time.20
20
Part IV
MATHEMATICAL TOPICS
331
Math.Topic 1
THE BASIC ALGEBRAIC STRUCTURES
GROUPS AND LESSER STRUCTURES
1
2
3
4
5
Definitions
Transformation groups
Representations
Groupoids, monoids, semigroups
Subgroups
Algebras
Kinds of algebras
Lie algebra
Enveloping algebra
Algebra of a group
Dual algebra
Derivation
COALGEBRAS
24 Coalgebras
25 Bialgebras, or Hopf algebras
26 R-matrices
Current mathematical literature takes for granted the notions given here, as
a minimum minimorum, the threshold of literacy. As a rule, we only retain
below those ideas which are necessary to give continuity to the presentation.
Some concepts are left to the Glossary.
333
334
1.1
Groups are the most important of algebraic structures. Men enjoy their
presence in some special way, through the bias of symmetry, and learning
processes seem to make use of them. Since their unheeded introduction by
an ill-fated young man, its fortune has been unsurpassed in Physics.
1.1 Definitions Asking for the readers tolerance, let us only recall, to
keep the language at hand, that the cartesian set product U V of two sets
U and V is the set of all pairs (u, v) with u U and v V . A group is a set
point G on which is defined a binary operation
: G G G,
taking the cartesian set product of G by itself into G, with the four following
properties:
(a) for all g, g 0 G, the result g g 0 belongs to G;
(b) there exists in G an element e (the identity, or neutral element) such
that, for all g G,
e g = g e = g;
(c) to every g G corresponds an inverse element g 1 which is such that
g 1 g = g g 1 = e;
(d) the operation is associative: for all g, g 0 , g 00 in G,
(g g 0 ) g 00 = g (g 0 g 00 ).
The group (G, ) is commutative (or abelian) when g g 0 = g 0 g holds for all
g, g 0 G. The operation symbol is usually omited for multiplicative groups.
The center of G is the set of the gG which commute with all the elements
of G. The order of a group G, written |G|, is the number of elements of G.
Given a group G, if there is a setQof ai G such that every g G can be
written in the monomial form g = i (ai )ni for some set of exponents {ni },
we say that the set {ai } generates G, and call the ai s generators of G. In
the monomials, as the ai s are not necessarily commutative, there may be
repetitions of each ai in different positions. When the set {ai } is finite, G is
finitely generated . The number of generators of G is, in this case, called the
rank of G.
335
There are some fluctuations concerning the definitions of monoid and semigroup.
Some authors (such as Hilton & Wylie l967) include the existence of neutral element for
semigroups, others (Fraleigh 1974) only associativity. In consequence, the schemes here
shown, concerned with these structures, are author-dependent. We follow, for simplicity,
the glossary Maurer 1981.
336
more widely known, it may be simpler to get the other structures from them,
by eliminating some conditions, as in the diagram
drop
associativity
GROUPOID
MONOID
there is no
neutral element
SEMIGROUP
&
GROUP
A general scheme is
Set
G = {g,g 0 ,...}
&
operation
is associative
SEMIGROUP
T HEN
IF
there exists
neutral element
MONOID
GROUPOID
ABELIAN
GROUP
.
GROUP
every element
has an inverse
operation
is commutative
337
338
Comment 1.1.5 Solvable and nilpotent groups Groups can be classified according to their degree of non-commutativity. Given two subsets A and B of a group G, we
define their commutator as [A, B] = { all elements in G of the form aba1 b1 , with a A
and b B}. We start with the abelian quotient G/[G, G]. We may then form two sequences of subgroups in the following way. Define G1 = [G, G], called the commutator
group of G; proceed to G2 = [G1 , G1 ] and so on to the n-th term of the series, Gn =
[Gn1 , Gn1 ]. Putting G = G0 , then
Gn1 Gn for all n. If it so happens that there exists an integer k such that Gn =
identity for all n k, the group G is solvable of class k. A second sequence is obtained by
defining G0 = G and Gn = [G, Gn1 ]. If Gn = identity for n k, the group G is nilpotent
of class k. Of course, both sequences are trivial for abelian groups. And the less trivial
they are, the farther G is from an abelian group.
1.2
339
add another
operation x
Nonassociative
Ring
every element, up to
Unit Ring
impose associativity
Division
Ring
there is neutral
element for x
0, has an inverse
by x00
by x
RING
FIELD
Commutative
Ring
if x is
commutative
The sets of real numbers (R) and complex numbers (C) constitute fields.
Familiarity with real and complex numbers makes fields the best known of
such structures. It is consequently more pedagogical to arrive at the concept
of ring by lifting progressively their properties:
FIELD
drop product
commutativity
Division ring
drop existence of
product inverse
RING
drop existence of
product neutral
Unit Ring
340
natural way:
a+b=
a(g) g +
gG
b(g) g =
gG
ab =
[a(g) + b(g)] g;
gG
a(g) b(h) g h.
g,hG
The summations constitute a new ring, the group ring R(G) of G over R.
If R is a unital ring, each g G can be identified with that element a of
R(G) whose single coefficient is a(g) = 1. The group is thereby extended
to a ring. If G is non-abelian, so will be the ring. Now, to each a R(G)
corresponds anPR-valued function on G, fa : G R, such that fa (g) =
a(g) and a =
R will
gG fa (g)g.PConversely, to any function f : GP
correspond
a
ring
member
f
(g)g.
Notice
that,
given
a
=
gG
gG a(g)g
P
P
and b = hG b(h)h, the product ab = gH fab (g)g will be given by
X X
ab =
[
fa (h) fb (gh1 )]g.
g inH hH
We arrive thus at another definition: given a group G, its group ring R(G)
over R is the set of mappings f : G R, with addition defined as
(f1 + f2 )(g) = f1 (g) + f2 (g),
and multiplication in the ring given by the convolution product
X
(f1 f2 )(g) =
f1 (h) f2 (h1 g).
hG
The condition concerning the finite number of coefficients in the first definition is necessary for the summations to be well-defined. Also the convolution
P
requires a good definition of the sum over all members of the group, h .
That is to say that it presupposes a measure on G. We shall briefly discuss
measures in Mathematical Topic 3. If G is a multiplicative group, R(G) is indicated by < R, +, > and is the ring of G over R. The restriction < R, >
contains G. If G is noncommutative, R(G) is a noncommutative ring. R(G)
is sometimes called the convolution ring of G, because another ring would
come out if the multiplication were given by the pointwise product,
(f1 f2 )(g) = f1 (g)f2 (g).
341
The analogy with Fourier analysis is not fortuitous (see Mathematical Topic
6). In order to track groups immersed in a ring R, we can use idempotents.
An idempotent (or projector ), as the name indicates, is an element p of the
ring such that p p = p. A group has exactly one idempotent element.
1.3
342
to the more current one of vector space, but as this pious statement seems
useless, we use both of them indifferently.
1.11 The notion of action We repeat that the cartesian set product
U V of two sets U and V is the set of all pairs (u, v) with u U and v V .
A well-defined mapping U V V goes under the general name of action
of U on V . The above axioms for vector spaces define an action of the field
F on V .
Comment 1.3.1 Linear representations of algebras are actions on modules and the
classification of modules is intimately related to that of representations.
343
a subfamily {wk } with dim W members. The set of vectors of V with zero
coefficients ak along the wk s will be the linear complement of W .
1.13 Dual space Vector spaces are essentially duplicated. This is so
because the space V * formed by linear mappings on a vector space is another
vector space, its dual. The image of v V by the linear mapping k V *
is indicated by < k, v >. When V is finite-dimensional, the dual of V is a
twin space, isomorphic to V . In the infinite-dimensional case, V is in general
only isomorphic to a subspace of V **. A mapping of V into its dual is an
involution. The isomorphism V V *, even in the finite-dimensional case, is
in general not canonical. This means that it depends of the basis chosen for
V . The presence of a norm induces a canonical isomorphism between V and
(at least part of) V *, and involutions are frequently defined as a mapping
V V (see below, 1.17). The action of the dual on V may define an inner
product V V F .
1.14 Inner product An inner product is defined as a mapping from the
cartesian set product V V into C, V V C, (v, u) < v, u > with the
following properties:
(a) < v1 + v2 , u > = < v1 , u > + < v2 , u >;
(b) < a v, u > = a < v, u >;
(c) < v, u > = < u, v > (so that < v, v > R );
(d) < v, v > 0;
(e) < v, v > = 0 v = 0.
The action of the dual V * on V defines an inner product if there exists
an isomorphism between V and V *, which is the case when V is of finite
dimension. An inner product defines a topology on V . Vector spaces endowed
with a topology will be examined in Mathematical Topics 4 and 5.
Comment 1.3.2 Infinite dimensional vector spaces differ deeply from finite dimensional ones. For example, closed bounded subsets are not compact neither in the norm
nor in the weak topology.
1.15 Endomorphisms and projectors An endomorphism (or linear
operator) on a vector space V is a mapping V V preserving its linear
structure. Projectors, or idempotents, are here particular endomorphisms p
satisfying p2 = p. Given a vector space E and a subspace P of E, it will be
possible to write any vector v of E as v = vp + vq , with vp P . The set Q of
vectors vq will be another subspace, the supplement of P in E. Projectors p
on P (which are such as p(vp ) = vp , Im p = P ) are in canonical (that is, basis
344
1.4
Algebras
The word algebra denotes of course one of the great chapters of Mathematics. But, just as the word topology, it also denotes a very specific
algebraic structure.
1.17 Algebras An algebra is a vector space A over a field F (for us,
R or C), on which is defined a binary operation (called multiplication) m :
A A A, (, ) m(, ) = such that, for all a F and , , A,
the following conditions hold:
(i) (a) = a() = (a);
(ii) ( + ) = + ;
1.4. ALGEBRAS
345
(iii) ( + ) = + .
This defines an action of the vector space on itself. Once so defined, A
is an algebra on F . When F = C, a mapping A A is an
involution, or adjoint operation, if it satisfies the postulates
(i) = ;
(ii) () = ;
(iii) (a + b) = a + b .
In that case is the adjoint of .
&m
A.
%m
AAA
id m &
AA
is commutative. A module can be assimilated to a commutative algebra.
A is a unit algebra (or unital algebra) if it has a unit, that is, an element
e such that e = e = , A. Notice that each A defines a
mapping h :F A by h (a) = a . Of course, = h (1). The existence
of a unit can be stated as the existence of an element e such that he (a) =
a e = a, for all a F . This is the same as the commutativity of the diagram
AA
h id
F A
&
&id h
m
AF .
.
A
346
scheme:
VECTOR SPACE
add associative
multiplication
ALGEBRA
add unit
INNER PRODUCT
SPACE
INVOLUTIVE
UNIT ALGEBRA
add involution
INVOLUTIVE
ALGEBRA
(1.1)
An element of an unit-algebra A is an invertible element if there exists in A
an element 1 such that 1 = 1 = e. The set G(A) of the invertible
elements of A is a group with the multiplication, the group of the algebra
A.
A graded algebra is a direct sum of vector spaces, A = k Ak , with the
binary operation taking Ai Aj Ai+j . If Ak , we say that k is the
degree (or order) of , and write = k.
1.19 Lie algebra An algebra is a Lie algebra if its multiplication (called
the Lie bracket) is anticommutative and satisfies the Jacobi identity
m(, m(, )) + m(, m(, )) + m(, m(, )) = 0.
Starting from any binary operation, the Lie bracket can always be defined
as the commutator [, ] = , and builds, from any associative
algebra A, a Lie algebra AL . If A is any algebra (not necessarily associative,
even merely a vector space), End A = {set of endomorphisms on A} is an
associative algebra. Then, the set [End A] of its commutators is a Lie algebra.
A vector basis 1 , 2 , . . . , n for the underlying vector space will be a basis
for the Lie algebra.
Comment 1.4.1 Lie algebras have a classification analogous to groups. They may be
solvable, nilpotent, simple, semisimple, etc, with definitions analogous to those given for
groups.
Comment 1.4.2 Drop in the algebra A the external multiplication by scalars of its
underlying vector space. What remains is a ring. As long as this external multiplication is
irrelevant, we can talk of a ring. The usual language is rather loose in this respect, though
most people seem to prefer talking about algebras.
1.4. ALGEBRAS
347
1.21 Algebra of a group The algebra of a group G comes out when the
group ring of G is a field, usually R or C. It is frequently called the group
convolution algebra, to distinguish it from the other algebra which would
come if the pointwise product were used.
Comment 1.4.5 This other algebra is the set of real or complex functions on the
group G, with addition and multiplication given by the pointwise addition and product:
(f + g)(x) = f (x) + g(x) and (f g)(x) = f (x) g(x). It is sometimes also called the group
algebra.
1.22 Dual algebra On algebras, the involution is required to submit to
some conditions, by which it transfers the algebraic properties into the dual
space, which thereby becomes a dual algebra.
A norm, for example that coming from an inner product, can define a
topology on a linear space. Addition of a topology, through an inner product
or by other means, turns linear spaces into much more complex objects.
These will be left to the Mathematical Topic 5.
1.23 Derivation The generic name derivation is given to any endomorphism D:A A for which Leibnizs rule holds: D() = (D) + (D).
The Lie algebra [End A] contains D(A), the vector subspace of all the derivations of A, and the Lie bracket makes of D(A) a Lie subalgebra of [End
A], called its derived algebra. This means that the commutator of two
derivations is a derivation. Each member A defines an endomorphism
ad(a) = ada , called the adjoint action of a, by
ada (b) = [a, b] b.
348
1.5
Coalgebras
1.24 General case Suppose now that spaces A and B are algebras over
a certain field, say C. Then A B is also an algebra (the tensor product of
algebras A and B) with the product defined by (a b)(a0 b0 ) = (a a0 ) (b b0 ).
When A and B are associative unit algebras, so will be their tensor product.
The product in A is in reality a mapping m : A A A, a a0 a a0 .
Its dual mapping : A A A is called the coproduct, or comultiplication
(or still diagonal mapping). It is supposed to be associative. The analogue
to associativity (the coassociativity) of the comultiplication would be the
property
(id ) (x) = ( id) (x)
as homomorphisms of A in A A A:
AA
&id
AAA.
%id
A
&
AA
Once endowed with this additional structure, A is a coalgebra. The coalgebra is commutative if (A) is included in the symmetric part (see Comment
1.3.4 above) of A A. Let us put it in other words: define a permutation
map : A A A A,
(x y = y x;
then the coalgebra is commutative if = .
1.25 Bialgebras, or Hopf algebras An associative unit algebra A is a
Hopf algebra (or bialgebra, or an old name annular group) if it is a
coalgebra satisfying
(i) the product is a homomorphism of unit coalgebras;
(ii) the coproduct is a homomorphism of unit algebras,
(xy) = (x)(y).
1.5. COALGEBRAS
349
xj yj ,
&id
xA.
%id
xA
&
AA
It is an algebra homomorphism:
(xy) = (x)(y).
2. The antipodal map : A A, an antihomomorphism (xy) =
(y)(x) such that
m(id ) (x) = m( id)(x) = (x)I,
which is described in the diagram
id
A A
AA
&m
xA
(x) I A
&
%m
A A
id
AA
The map (x) is called the antipode (or co-inverse) of x. Given the
permutation map (x y) = y x, then 0 = is another coproduct
on A, whose antipode is 0 = 1 .
Comment 1.5.1 Some people call bialgebras structures as the above up to the existence of the counit, and reserve the name Hopf algebras to those having further the
antipode.
350
1
(2 x) x
and 2 x =
1
(1 x) x.
Math.Topic 2
DISCRETE GROUPS. BRAIDS AND KNOTS
A DISCRETE GROUPS
1
2
3
4
B BRAIDS
5 Geometrical braids
6 Braid groups
7 Braids in everyday life
8 Braids presented
9 Braid statistics
10 Direct product representations
11 The Yang-Baxter equation
C KNOTS AND LINKS
12
13
14
15
16
2.1
Knots
Links
Knot groups
Links and braids
Invariant polynomials
A Discrete groups
These are the original groups, called into being by Galois. Physicists became
so infatuated with Lie groups that it is necessary to say what we mean by
discrete groups: those which are not continuous, on which the topology is
either undefined or the discrete topology. They can be of finite order (like
351
352
Practically all the cases of discrete groups we shall meet here are fundamental groups of some topological spaces. These are always found in terms
of some generators and relations between them. Let us say a few words on
this way of treating discrete groups, taking for simplicity a finite rank.
2.1 Words and free groups
Consider a set A of n elements, A = {a1 , a2 , . . . , an }. We shall use the
names letters for the elements aj and alphabet for A itself. An animal with
p times the letter aj will be written apj and will be called a syllable. A finite
string of syllables, with eventual repetitions, is (of course) a word . Notice
that there is no reason to commute letters: changing their orders lead to
different words. The empty word 1 has no syllables.
There are two types of transformations acting on words, called elementary
contractions. They correspond to the usual manipulations of exponents: by
a contraction of first type, a symbol like api aqi becomes ap+q
; by a second type
i
contraction, a symbol like a0j is replaced by the empty word 1, or simply
dropped from the word. With these contractions, each word can be reduced
to its simplest expression, the reduced word. The set F [A] of all the reduced
words of the alphabet A can be made into a group: the product u v of two
words u and v is just the reduced form of the juxtaposition uv. It is possible
to show that this operation is associative and ascribes an inverse to every
reduced word. The resulting group F [A] is the word group generated by the
alphabet A. Each letter ak is a generator .
Words may look at first as too abstract objets. They are actually extremely useful. Besides obvious applications in Linguistics and decoding,1
the word groups are largely used in Mathematics, and have found at least
1
Schreider 1975.
353
one surprising application in Geometry: they classify the 2-dimensional manifolds.2 In Physics, they are used without explicit mention in elementary
Quantum Mechanics. Recall the Weyl prescription3 (the correspondence
rule) to obtain the quantum operator Weyl(pm q n ) = W(pm q n ) corresponding to a classical dynamical quantity like pm q n :
m n
W(p q ) =
1
2n
n
X
n
k=0
m nk
q p q
1
2m
m
X
m
k=0
pk qn pmk
(2.1)
where bold-faced letters represent operators. For the first few cases, W(pq)
= 21 (pq + qp), W(pq 2 ) = 13 (pq2 + qpq + q2 p), etc. The quantum operator
corresponding to a polynomial pm qn in the classical degree of freedom q
and its conjugate momentum p is the (normalized) sum of all the words
one can obtain with m times the letter p and n times the letter q.
Now, given a general group G, it will be a free group if it has a set
A = {a1 , a2 , . . . , an } of generators such that G is isomorphic to the word
group F [A]. In this case, the aj are the free generators of G. The number of
letters, which is the rank of G, may eventually be infinite.
The importance of free groups comes from the following theorem:
Every group G is a homomorphic image of some free group F [A].
This means that a mapping f : F [A] G exists, preserving the group operation. In a homomorphism, in general, something is lost: many elements
in F [A] may be taken into a same element of G. F [A] is in general too rich.
Something else must be done in order to obtain an isomorphism. As a rule, a
large F [A] is taken and the freedom of its generators is narrowed by some
relationships between them.
2.2 Presentations
Consider a subset {rj } of F [A]. We build the minimum normal subgroup
R with the rj as generators. The quotient F [A]/R will be a subgroup, corresponding to putting all the rj = 1. An isomorphism of G onto F [A]/R
will be a presentation of G. The set A is the set of generators and each rj
is a relator . Each r R is a consequence of {rj }. Each equation rj = 1 is a
relation. Now, another version of the theorem of the previous section is:
Every group G is isomorphic to some quotient group of a free group.
2
3
354
Comment 2.1.2 In this way, groups are introduced by giving generators and relations
between them. Free groups have for discrete groups a role analogous to that of coordinate
systems for surfaces: these are given, in a larger space, by the coordinates and relations
between them. Of course, such coordinates being non-commutative, things are much
more complicated than with usual coordinates and equations seldom lead to the elimination
of variables. Related to this point, there is a difficulty with presentations: the same group
can have many of them, and it is difficult to know whether or not two presentations refer
to the same group.
2.3 Cyclic groups The simplest discrete groups are the cyclic groups,
which are one-letter groups. A group G is a cyclic group if there is an element
a such that any other element (including the identity) may be obtained as
ak for some k. It is of order n if the identity is an . Good examples are the
n-th roots of 1 in the complex plane. They form a group isomorphic to the
set {0, 1, 2, 3, . . . , n 1} of integers with the operation of addition modulo n.
This is the cyclic group Zn . There is consequently one such group Zn for each
integer n = |Zn |. The simplest case is Z2 , which can be alternatively seen as
a multiplicative group of generator a = 1: Z2 = {1, - 1; the operation is
the usual multiplication}. Every cyclic group is abelian. Every subgroup of
a cyclic group is a cyclic group. Any two cyclic groups of the the same finite
order are isomorphic. Thus, the groups Zn classify all cyclic groups and for
this reason Zn is frequently identified as the cyclic group of order n. Any
infinite cyclic group is isomorphic to the group Z of integers under addition.
Given a group G and an element a G, then the cyclic subgroup of G
generated by a, hai = {an : n Z} is the smallest subgroup of G which
contains a. If hai = G, then a generates G entirely, and G is itself a cyclic
group.
Comment 2.1.3 Consider an element g G. If an integer n exists such that gn = e,
then n is the order of the element g, and g belongs to a cyclic subgroup. When no such
integer exists, g is said to be of infinite order. If every element of G is of finite order, G is
a torsion group. G is torsion-free if only its identity is of finite order. In an abelian group
G, the set T of all elements of finite order is a subgroup of G, the torsion subgroup of G.
355
Comment 2.1.4 The expression a permutation group is used for any (proper or
improper) subgroup of a symmetric group. This is very important because every finite
group is isomorphic to some permutation group (Cayleys theorem).
1 2
i
i + 1 n 1 n
1 2 i + 1
i
n 1 n
(2.3)
Each permutation will be a word with the alphabet {sj }. The si s obey
the relations
sj sj+1 sj = sj+1 sj sj+1
si sj = sj sj for |i j| > 1
(si )2 = 1,
(2.4)
(2.5)
(2.6)
which determine completely the symmetric group Sn : any group with generators satisfying these relations is isomorphic to Sn .
Comment 2.1.5 Many groups are only attained through presentations. This is frequently the case with fundamental groups of spaces. A good question is the following:
given two presentations, can we know whether or not they present the same group?
This is a version of the so-called word problem. It was shown by P. S. Novikov that
there can exist no general procedure to answer this question.
356
n!
tn1 tn2 tn3 . . . tnr r
nj 1 2 3
n
!j
j
j=1
Qn
{nj }
(2.7)
The summation
takes place over the sets {ni } of non-negative integers for
P
which ni=1 i ni = n. Of course, such a summation of invariant objects is
itself invariant. This is an example of a very important way of characterizing
discrete groups: by invariant polynomials. Though not the case here, it is
sometimes easier to find the polynomials than to explicit the group itself.
This happens for example for the knot groups (see Math.2.16 below). The
above invariant polynomial for the symmetric group appears as partition
functions in Statistical Mechanics of systems of identical particles,5 which
are invariant under particle permutations (Phys.3).
2.2
B Braids
j = 1, 2, . . . , n,
j = 1, 2, . . . , n,
Comtet 1974.
For a description of the symmetric and braid groups leading to braid statistics, see
Aldrovandi 1992.
6
Doubrovine, Novikov & Fomenko 1979, vol. II.
5
2.2. B BRAIDS
357
Figure 2.1:
(2.9)
Then 1 [Bn M ] is the full braid group Bn , or simply braid group. Artins
braid group is the full braid group for M = E2 , but the above formal definition allows generalization to braid groups on any manifold M . Of course,
Fn M is just the configuration space for a gas of n impenetrable particles,
358
Leinaas & Myrheim 1977: a very good discussion of the configuration spaces of
identical particles systems. Wavefunctions for bosons and fermions are found without
resource to the summations of wavefunctions of distinguishable particles usually found in
textbooks. Braid groups, although not given that name, are clearly at play.
8
Schulman l968; Laidlaw & DeWitt-Morette 1971; DeWitt-Morette 1972; DeWittMorette, Masheshwari & Nelson 1979.
9
Birman 1975: the standard mathematical reference.
2.2. B BRAIDS
359
Figure 2.2:
Figure 2.3:
A basic point is the following: consider the Figure 2.3. Each point on
it is, ultimately, sent into itself. It would appear that it corresponds to the
identity, but that is not the case! Identity is 2(a) and Figure 2.3 cannot be
reduced to it by any continuous change of point positions on E2 . It cannot be
unwoven! A short experiment shows that it would be possible to disentangle
it if the space were E3 . As every braid is a composition of the elementary
braids, that would mean that any braid on E3 may be unbraided ... as
360
Figure 2.4:
(2.10)
(2.11)
They are the same as the two relations [2.4] and [2.5] for Sn . The absence of
condition [2.6], however, makes of Bn a quite different group. It suffices to
say that, while Sn is of finite order, Bn is infinite.
2.9 Braid statistics
2.2. B BRAIDS
361
The absence of the relation [2.6] has, as we have said, deep consequences.
Unlike the elementary exchanges of the symmetric group, the square of an
elementary braid is not the identity. In many important applications, however, it is found that j2 differs from the identity in a well-defined way. In
the simplest case, j2 can be expressed in terms of the identity and j , which
means that it satisfies a second order equation like (j x)(j y) = 0,
where x and y are numbers. In this case, the j s belong to a subalgebra
of the braid group algebra, called Hecke algebra. This is the origin of the
so-called skein relations, which are helpful in the calculation of the invariant
polynomials of knot theory.
In Quantum Mechanics, a basis for a representation of a braid group will
be given by operators U (j ) acting on wavefunctions according to 0 (x) =
U (j )(x) = ei (x). But now there is no constraint enforcing U (j2 ) = 1, so
that U 2 (j )(x) = U (j2 )(x) = ei2 (x), U (j3 )(x) = ei3 (x), etc. The
representation is now, like the group, infinite. It is from the condition U (j2 )
= 1 that the possibilities of phase values for the usual n-particle wavefunctions are reduced to two: as twice the same permutation leads to the same
state, U (j2 )(x) = (x) so that ei = 1. The two signs correspond to
wave-functions which are symmetric and antisymmetric under exchange of
particles, that is, to bosons and fermions. When statistics is governed by the
braid groups, as is the case for two-dimensional configuration spaces of impenetrable particles, the phase ei remains arbitrary and there is a different
statistics for each value of . Such statistics are called braid statistics.
2.10 Direct product representations Representations of the braid
groups can be obtained with the use of direct products of matrix algebras.
Suppose the direct product of two matrices A and B. By definition, the
matrix elements of their direct product A B are
< ij|A B|mn > = < i|A|m >< j|B|n > .
(2.12)
(2.13)
(2.14)
(2.15)
(2.16)
362
(2.17)
b12 R
b23 R
b12 = R
b23 R
b12 R
b23 ,
R
(2.18)
is equivalent to
which is the braid equation, name usually given to [2.10]. To show it, look
at s1 , s2 as s1 = S12 and s2 = S23 , S being some direct product as above.
Then find < ijk|s1 s2 s1 |mnr > = S ij pq S qk vr S pv mn and < kji|s2 s1 s2 |mnr >
= S ji qs S kq mv S vs nr , so that the braid equation is
S kj ab S bi cr S ac mn = S ji ca S kc mb S ba nr .
We have found above conditions for representations of B3 . Higher order
direct products of projectors will produce representations for higher order
b Aut(V V ) satisfying
braid groups. In the general case, given a matrix R
relations as above and the identity E Aut(V ), a representation of BN on
V N is obtained with generators
bi,i+1 . . . E E = (E)i1 R
bi,i+1 (E)N i . (2.19)
i = E E E . . . R
2.11 The Yang-Baxter equation With the notation above, we may
easily establish a direct connection of the braid relations to the Yang-Baxter
equation, usually written
R12 R13 R23 = R23 R13 R12 ,
(2.20)
(2.21)
b = P R, R
bij mn =
Define now another product matrix by the permutation R
ji
R mn . The above expressions are then equivalent to
b12 R
b23 R
b12 = R
b23 R
b12 R
b23 ,
R
(2.22)
just the braid equation. The permutation relation is thus a very interesting
tool to obtain representations of the braid groups from Yang-Baxter solutions
and vice-versa. Notice that, due to this equivalence, many people give the
name Yang-Baxter equation to the braid equation. An important point is
that Yang-Baxter equations come out naturally from the representations of
363
the Lie algebra of any Lie group. Thus, each such Lie algebra representation
will provide a solution for the braid relations.10
The relation between this matrix formulation and our first informal representation of braids by their plane drawings leads to an instructive matrixdiagrammatic formulation. It is enough to notice the relationship
a /b
c/ d
ab cd
R
and proceed to algebrize diagrams by replacing concatenation by matrix multiplication, paying due attention to the contracted indices. Looking at Figure
2.5, we see that the braid equation [2.10], becomes exactly the Yang-Baxter
equation in its form [2.17].
2.3
The classification of knots has deserved a lot of attention from physicists like
Tait and Kelvin at the end of the last century, when it was thought that the
disposition of the elements in the periodic table might be related to some
kind of knotting in the ether. Motivated by the belief in the possibility of a
fundamental role to be played by weaving patterns in the background of physical reality, they have been the pioneers in the (rather empirical) elaboration
of tables11 of distinct knots. Nowadays the most practical classification of
knots and links is obtained via invariant polynomials. Braids constitute
10
11
364
Figure 2.6:
one from the other without somehow tying or untying, that is, passing one of
the ends through some loop. The mathematical formalization of this intuitive
notion of different knots is the so called knot problem and leads to an
involved theory. The characterization is completely given by the notion of
knot- (or link-) type, which is as sound as unpractical. Actually, there is no
practical way to establish the distinction of every two given knots. There are
however two methods allowing an imperfect solution of the problem. One
of them attributes to every given knot (or link) a certain group, the other
attributes a polynomial. They are imperfect because two different knots may
have the same polynomials or group. The characterization by the knot groups
is stronger: two knots with the same group have the same polynomials, but
not vice-versa. On the other hand, polynomials are easier to find out.
We must somehow ensure the stability of the knot, and we do it by
eliminating the possibility of untying. We can either extend the ends to
infinity or simply connect them. We shall choose the latter, obtaining the
closed versions drawn more symmetrically as in the Figure 2.7. The example
to the right is equivalent to the circle, which is the trivial knot.
12
For an introductory, intuitive view, see Neuwirth 1979. For a recent qualitative
appraisal, see Birman 1991; an involved treatment, but with a very readable first
chapter is Atiyah 1991.
365
Figure 2.7:
2.13 Links
Links are intertwined knots, consequently defined as spaces homeomorphic to the disjoint union of circles. The left example of Figure 2.8 shows
a false link, whose component knots are actually independent. Such links
are also called unknots. The center and right examples of Figure 2.8 show
two of the simplest collectors favorites. Of course, knots are particular onecomponent links, so that we may use links to denote the general case.
We have above talked loosely of continuous deformation of the host
space taking one knot into another. Let us make the idea more precise. The
first step is the following: two knots A and B are equivalent when there exists
a homeomorphism of E3 into itself which takes A into B. This is fine, but
13
Crowell & Fox 1963; Doubrovine, Novikov & Fomenko 1979, vol. II.
366
Figure 2.8:
there is better. We would like to put arrows along the lines, to give knots
an orientation. The equivalence just introduced would not take into account
different orientations of the knots. A more involved notion will allow A and
B to be equal only if also their orientations coincide once A is deformed
into B. An isotopy (or isotopic deformation) of a topological space M is a
family of homeomorphisms ht of M into itself, parametrized by t [0, 1],
and such that (i) h0 (p) = p for all p M , and (ii) the function ht (p) is
continuous in both p and t. Isotopies provide the finest notion of continuous orientation-preserving deformations of M into itself. They constitute a
special kind of homotopy, in which each member of the one-parameter family
of deformations is invertible. When M is the host space of links, this definition establishes an equivalence relation, whose classes are called link-types.
Link-types provide the complete characterization of links, but it has a serious drawback: given a link, it is a very difficult task to perform isotopies
to verify whether or not it may be taken into another given link. That is
why the experts content themselves with incomplete characterizations, such
as the link group and the invariant polynomials.
2.14 Knot groups
Knots, as defined above, are all homeomorphic to the circle and consequently topologically equivalent as 1-dimensional spaces. We have seen that
knot theory in reality is not concerned with such 1-dimensional spaces themselves, but with how E3 englobes these deformed circles. Given the knot
K, consider the complement E3 \K. The knot group of K is the fundamental
group of this complement, 1 (E3 \K). It is almost evident that the group of
the trivial knot is Z. Simple experiments with a rope will convince the reader
that such groups may be very complicated. The trefoil group is the second
braid group B2 . As already said, knot groups do not completely characterize
367
knot types: two inequivalent knots may have the same group.
2.15 Links and braids
The relation between links and braids is given by Alexanders theorem,
which requires a preliminary notion. Given a braid, we can obtain its closure
simply by identifying corresponding initial and end points. Experiments with
pieces of rope will once again be helpful. For instance, with the sole generator
1 of the two-strand-group B2 we can build the Hopf link and the trefoil: they
are respectively the braids 12 and 13 when their corresponding ends meet.
Alexanders theorem says that
to every link-type corresponds a closed braid,
provided both braid and link are tame .
corresponds to a link K, we
Given the braid whose closure (denoted )
Commendable collections of papers on the subject are Yang & Ge 1989 and Kohno
1990.
15
Goldberg 1962.
368
t1
(2.23)
< )( >A = 0,
the index A indicating Alexander. This relation says that the j s are
in a Hecke algebra: it is a graphic version of
t1
j1 j + I = 0.
t
(2.24)
S
T
S
T
>,
>,
(2.25)
(2.26)
369
we can go down and down to simpler and simpler links, and at the end only
the identity and simple blobs O remain.
The animal represents a projector. Kauffmans decomposition is justified by Jones discovery of representations of the braid group in some special
von Neumann algebras, which are generated by projectors. Jones has thereby
found other polynomials, and also clarified the meaning of the skein relations.
The cycle indicator polynomial appears as the partition function of a system
of identical particles (Phys.3.1.2). Jones polynomials appear as the partition
function of a lattice model (Math.5.6).
Adams 1994
Birman 1991
Crowell & Fox 1963
Fraleigh 1974
Kauffman 1991
Yang & Ge 1989
Neuwirth 1965
370
Math.Topic 3
SETS AND MEASURES
MEASURE SPACES
1
2
3
4
5
6
7
ERGODISM
8 Types of flow
9 The ergodic problem
3.1
MEASURE SPACES
372
373
374
generated by the open intervals (a, b) with b a and the measure function
is m[(a, b)] = b a. The Lebesgue measure extends easily to En .
3.5 Partition of identity
Consider a closed subset U of a differentiable manifold M . Then, there is a
theorem which says that there exists a smooth function fU (the characteristic
function of U ) such that fU (p) = 1 for all p U , and fU (p) = 0 for all p
/ U.
Suppose further that M is paracompact. This means that M is Hausdorff
and each covering has a locally finite sub-covering. Given a smooth atlas,
there will be a locally finite coordinate covering {Uk }. Then, another theorem
says that a family {fk } of smooth functions exists such that
(i) the support of fk Uk ;
(ii) 0 fk (p) 1 for all p M ;
P
(iii) k fk (p) = 1 for all p M .
The family {fk } is a partition of the identity . The existence of a
partition of the identity can be used to extend a general local property to
the whole space, as in the important examples below.
3.6 Riemannian metric
Once assured that a partition of the identity exists, we may show that a
differentiable manifold has always a Riemannian metric (6.6.10). As each
coordinate neighbourhood is euclidean, we may define on each Uk the eu(k)
clidean metric g = . A Riemannian metric on M will then be given
by
X
(k)
g (p) =
fk (p)g
(p).
k
Uk
3.2. ERGODISM
3.2
375
ERGODISM
376
(3.3)
<A>=a
.
(3.4)
then
The interval T is supposed to be large not only with respect to the times
involved in the detailed microscopic processes (like scattering times), but also
as compared with those times relevant for the establishment of equilibrium
(relaxation time, free flight between the walls, etc).
The ergodic problem is summarized in the question: is the ergodic theorem valid? Or, which is the same, can we replace one average by the other?
If the answer is positive, we can replace statistics by a dynamical average.
3.2. ERGODISM
377
Roughly speaking, the answer is that the theorem is true provided the measure on phase space has a certain property, so that statistics is actually never
eliminated. To give an idea of the kind of questions involved, we shall briefly
describe the basic results.
A first point refers to the very existence of the limit in [3.3]. A second
point is concerned with the independence of (
a on the particular flow (the
particular hamiltonian).
Concerning the limit, there are two main points of view, depending on
the type of convergence assumed. One is that of Birkhoff, the other that of
von Neumann. Suppose a finite-volume subset S of phase space. Then, we
have two different theorems:
(a) Birkhoff s theorem: if the dynamical function f (q, p) on S is such
that
R
f [q(0), p(0)]d < ,
S
then
1
limT { 2T
RT
T
exists for all points (q, p) of S and is independent of the chosen origin of
time. This limit will be identified with f, but notice that this is a particular
definition, assuming a particular type of convergence.
(b) von Neumanns theorem: consider the Hilbert space of squareintegrable dynamical functions on S. The inner product
R
(f, g) = S f (q, p)g(q, p)dS
defines a norm ||f ||. Then there exists a function f such that
limT ||f f|| = 0.
If f is simultaneously integrable and square-integrable, Birkhoffs and von
Neumanns limits coincide over S, except for functions defined on sets of zero
measures. Of course, everything here holds only up to such sets.
This seems to settle the question of the limit, though it should be noticed
that other topologies on the function spaces could be considered. Equation
[3.4] is valid for both cases above, provided an additional hypothesis concerning the measure d is assumed. In simple words, the measure d should not
divide the phase space into non-communicating sub-domains. Phase space
must not be decomposed into flow-invariant sub-regions.
More precisely: a space is metric-indecomposable (or metrically transitive) if it cannot be separated into two (or more) regions whose measures
378
are invariant under the dynamical flow and different from 0 or 1. The condition for the ergodic theorem to be true is that the phase space be metrically
transitive. This means that there is no tendency for a point to abide in a
sub-domain of phase space, or that no trajectory remains confined to some
sub-region. In particular, there must be no hidden symmetries. It is in general very difficult to know whether or not this is the case for a given physical
system, or even for realistic models.
Balescu 1975
Arnold 1976
Jancel 1969
Mackey 1978
Math.Topic 4
TOPOLOGICAL LINEAR SPACES
4.1
380
uv
4.2
Norm
4.3
4.4
Hilbert space
Helmerg 1969.
Halmos 1957.
381
i=1
vi ui .
Dieudonne 1960.
382
are nevertheless
Pnew problems in this continuum-label case: the convergent
summations i=1 used in the very definition of Hilbert space become integrals. In order to define integrals over a set, one needs additional structures:
those of a -algebra of subsets, and that of a measure (see Math.3). Such
Hilbert spaces will depend also on the choice of these structures.
4.5
Banach space
We have seen that Hilbert space is an inner product space which is complete
under the inner product norm topology. More general, a Banach space is a
normed vector space which is complete under the norm topology. Thus, each
one of its Cauchy sequences in the norm topology is convergent.
4.6
A Banach space is a topological vector space when both the addition operation and the scalar multiplication are continuous in the norm topology.
Although these rather abstract concepts hold in finite-dimensional spaces,
they are actually fundamental in the study of infinite-dimensional spaces,
which have quite distinct characteristics.
A general scheme is shown in Figure 4.1. Normed spaces have metric
topologies. If they are also complete, they are Banach spaces. On the other
hand, the norm may come from an inner product, or not. When it does
and furthermore the space is complete, it is a Hilbert space. If the linear
operations (addition and scalar multiplication) are continuous in the norm
topology (inner product or not), we have topological vector spaces.
Comment 4.6.1 The word metrizable, when applied to such spaces, means that its
topology is given by a translation-invariant metric.
Let us recall that the dual space to a given linear space V is that linear
space (usually denoted V ) formed by all the linear mappings from V into
its own field. When V is finite-dimensional, V is related to V by an isomorphism which in general is not canonical, but V is canonically isomorphic
to V . In the infinite-dimensional case, V is in general only isomorphic to a
subspace of V . The image of v V by k V is indicated by < k, v >.
On a topological vector space V , another topology is defined through the
action of the V . It is called the weak topology and may be defined through
convergence: a sequence {vn } converges weakly to v V if, for every k V ,
< k, vn > < k, v > as n . As the names indicate, the norm topology is finer than the weak topology: a sequence may converge weakly and
383
Figure 4.1:
not converge in the norm topology. There are many other possible topologies, in effect. Figure 4.2 shows a scheme of linear spaces and some of their
topologies.
A very useful notion is the following: a subset U of a topological vector
space is a bounded set if it obeys the following condition of archimedean
type: for any neighbourhood V of the origin there exists a number n > 0
such that nV U .
4.7
Function spaces
Consider the space C (M, C) of differentiable complex functions on a manifold M . It is a vector space to start with. Define on this space an internal
operation of multiplication C (M, C) C (M, C) C (M, C). To help
the mind, we may take the simplest multiplication, the pointwise product,
defined by (f g)(x) = f (x)g(x). Then C (M, C) becomes an algebra. Actually, it is as associative, commutative *-algebra (described in Math.5). We
may in principle introduce other kinds of multiplication and obtain other
algebras within the same function space.
As hinted in the discussion on the Hilbert space, there are very important
cases in which the norm involves a measure. They combine in this way the
above ideas with those given in Math.3, and the resulting structure is far more
384
Figure 4.2:
complicated. For this reason we leave Banach and *-algebras to Math.5. Differentiability on infinite-dimensional manifolds, which supposes topological
vector spaces to provide a tangent structure, is discussed in Math.7.
Kolmogorov & Fomin 1977
Bratelli & Robinson 1979
Math.Topic 5
BANACH ALGEBRAS
1 Quantization
2 Banach algebras
3 *-algebras and C*-algebras
4 From Geometry to Algebra
5 von Neumann algebras
6 The Jones polynomials
The wealth of vector spaces endowed with a topology has been seen in
Math.4. We give now a sketchy account of what happens when the vector spaces are, furthermore, algebras. This includes spaces of operators, of
particular interest to quantum physics. That is why we start by recalling
some basic points behind the usual idea of quantization, and use some well
known aspects of quantum theory to announce some notions to be developed
afterwards.
5.1
Quantization
386
A =
<|A|>
.
<|>
5.2
387
Banach algebras
388
This homomorphism includes the above polynomial case, and the good
subset o(f ), as the notation suggests, depends on the function f . The integration is along , which is any closed curve circumscribing the entire set
SpA.
Comment 5.2.2 We learn thus, by the way, how to get a function of a given quantum
operator.
Consider now the case of commutative Banach algebras, which are the
natural setting for standard harmonic analysis1 (Mat.6.11). To each such
algebra one associates I(B), the set of its maximal proper (that is, 6= B)
ideals. I(B) is compact. An important point is that to each maximal ideal
corresponds a character of B, a homomorphic mapping : B C. Given ,
then 1 (0) is a maximal ideal of B. This interpretation of the characters as
maximal ideals leads to the Gelfand transformation: it relates a function on
b
I(B) to each element of B, given by A()
= (A). The set of values of the
b coincides with SpA.
function A
Let B = R(X) be the algebra of real continuous functions on the compact
X with the pointwise product (f g)(x) = f (x)g(x) as operation. The set of
functions vanishing at a point x is an ideal: f (x) = 0 implies (f g)(x) = 0
for any g. Each closed ideal is formed by those functions which vanish on
some closed subset Y X. There is a correspondence between the maximal
ideals and the points of X: we can identify x X to the maximal ideal I(x)
of functions f such that f (x) = 0 and the space X itself to the quotient
R(X)/I, where I is the union of all I(x).
1
Katznelson 1976.
5.3
389
Let us recall that, as an algebra, A will have a dual space, formed by all
the linear mappings of A into the real line. We write dual, with quotation
marks, because infinite dimensional vector spaces are deeply different from
the finite dimensional vector spaces and one of the main differences concerns
precisely the dual. For finite dimensional spaces, the dual of the dual is the
2
Dixmier 1982.
390
5.4
We have said in section 5.2 that a compact space X can be seen as a subspace
of the algebra C(X) of continuous functions on X with the pointwise product
as multiplication. Each point of X is an ideal formed by those functions
which vanish at the point. Maximal ideals are identifiable to characters of
the algebra and in section 5.3 we have indeed said that if X is a compact then
391
it can be identified to the characters of C(X). This has been the starting
point of a process by which Geometry has been recast as a chapter of Algebra.
We shall only say a few words on the subject, which provides one of the gates
into non-commutative geometry and may become important to Physics in the
near future.
Consider again the space C (M, C) of differentiable complex functions
on a manifold M . It is clearly a linear space. Define on this space an internal operation of multiplication C (M, C) C (M, C) C (M, C). The
function space becomes an algebra. When there is no unit in this algebra,
we can always add it. What results is a unital algebra. To help the mind, we
can take the simplest, usual commutative pointwise product of complex functions: (f g)(x) = f (x)g(x). Suppose we are able to introduce also some norm.
With the pointwise product, C (M, C) becomes an associative commutative
*-algebra.
We can in principle introduce other kinds of multiplication and obtain
other algebras with the same starting space. Inspired by phase spaces, for
instance, we might think of introducing a Poisson bracket. From a purely
algebraic point of view, a Poisson algebra is a commutative algebra A as
above endowed with a map {, } : A A A such that:
(i) A is Lie algebra with the operation {, };
(ii) the bracket is a derivative in A: {a, bc} = b{a, c} + {a, b}c.
Once endowed with such a function algebra and bracket, a space M is
said to have a Poisson structure. Notice that M is not necessarily a phase
space: a Poisson structure can in principle be introduced on any differentiable manifold. But there is still more. The differentiable structure of M is
encoded3 in the *-algebra C (M, C). Each property of M has a translation
into a property of C (M, C). If we restrict ourselves to the space C 0 (M, C)
of functions which are only continuous, only the topological properties of M
remain encoded, the information on the differentiable structure of M being
forgotten. For instance, the mentioned theorem by Gelfand and Naimark
throws a further bridge between the two structures. It states, roughly, that
any unit abelian *-algebra is isomorphic to some algebra like C 0 (M, C), with
M some compact manifold. The difference between C (M, C) and C 0 (M, C)
is that the former has a great number of derivations, the vector fields on M .
In the algebraic picture, such derivations4 are replaced by the derivations
in the *-algebra, that is, endomorphisms of C (M, C) satisfying Leibniz
rule,5 like the above example with the Poisson bracket. Such derivations
3
4
5
Dubois-Violette 1991.
Dubois-Violette 1988.
Connes 1980.
392
Connes1986.
5.5
393
Dixmier 1981.
Takesaki 1978.
394
a Hilbert space H, which is closed under the weak operator topology. This
is the locally convex topology in L(H) induced by the family of semi-norms
x L(H)
395
Projectors are ordered as follows. Let p and q be two projectors in the von
Neumann algebra M . We say that p and q are equivalent, and write p q,
if there exists u M such that p = uu and u u = q. We say that p q
(q dominates p) if there exists u M such that p = uu and u uq = u u.
This means that u u projects into a subspace of qH, that is, that u u q.
If p q and q p, then p q. We say further that pq if pq = 0.
Now, if M is a factor, then it is true that, given two projections p and
q, either p q or q p. The terminology is not without recalling that of
transfinite numbers. The projector q is finite if the two conditions p q
and p q together imply p = q. The projector q is infinite if the conditions
p q, p q and p 6= q can hold simultaneously. The projector p is minimal
if p 6= 0 and q p implies q = 0 (p only dominates 0). The factors are then
classified in types, denoted I, II1 , II , and III:
I: if there exists a minimal projector;
II1 : if there exists no minimal projector and all projectors are finite;
II : if there exists no minimal projector and there are finite and infinite
projectors;
III: if the only finite projection is 0.
On a factor there exists a dimension function d: {projections on M }
[0, ] with the suitable properties:
(i) d(0) = 0;
(ii) d(k pk ) = k d(pk ) if pi pj for i 6= j;
(iii) d(p) = d(q) if p q.
It is possible then to show that, conversely, d(p) = d(q) implies p q.
This Murray-von Neumann dimension d can then be normalized so that
its values have the following ranges:
type I: d(p) { 0, 1, 2, . . . , n, with possibly n = }; if n is finite, type
In ; if not, type I ;
type II1 : d(p) [0, 1];
type II : d(p) [0, ];
type III: d(p) {0, }.
We have been talking about a *-algebra A as a set of operators acting on
some Hilbert space H. For many purposes, it is interesting to make abstraction of the supposed carrier Hilbert space and consider the algebra by itself,
taking into account as far as possible only its own properties, independent
396
Comment 5.5.2 The name W*-algebra is frequently reserved to the abstract von
Neumann algebras.
5.6
397
Let us now examine some particular finite-dimensional von Neumann algebras, of special interest because of their relationship with braids and knots.
A finite-dimensional von Neumann algebra is just a product of matrix algebras, and can be represented in the direct-product notation. In his work10
dedicated to the classification of factors, Jones11 was led to examine12 certain
complex von Neumann algebras13 An+1 , generated by the identity I plus n
projectors p1 , p2 , . . . , pn satisfying
p2i = pi = pi
pi pi1 pi = pi
pi pj = pj pi for |i j|
2.
(5.1)
(5.2)
(5.3)
The complex number , the inverse of which is called the Jones index, is
usually written by Jones as
=
t
,
(1+t)2
where t is another complex number, more convenient for later purposes. For
more involved von Neumann algebras, the Jones index is a kind of dimension
(notice: in general a complex number) of subalgebras in terms of which
the whole algebra can be decomposed in some sense. In lattice models of
Statistical Mechanics, with a spin variable at each vertex, the Jones index is
the dimension of the spin-space (Phys.3). Conditions [5.2] and [5.3] involve
clearly a nearest neighbor prescription, and are reminiscent of the braid
relations. We shall see below that some linear combinations of the projectors
and the identity do provide braid group generators.
Consider the sequence of algebras An . We can add the algebra A0 = C
to the club. If we impose that each algebra An embeds naturally in An+1 , it
turns out that this is possible for arbitrary n only if t takes on some special
values: either
(a) t is real positive, or
(b) t is of the form t = e2i/k , with k = 3, 4, 5, . . . in which case
=
10
11
12
13
1
4 cos2 (/k)
Jones 1983.
Jones 1987.
Jones 1985, and his contribution in Kohno 1990.
Wenzl 1988.
398
For these values of t there exists a trace defined on the union of the An s,
defined as a function into the complex numbers, tr: n An C, entirely
determined by the conditions
tr(ab) = tr(ba);
tr(w pn+1 ) = tr(w) for w An
(5.4)
(5.5)
tr(a a) > 0 if a 6= 0;
tr(I) = 1.
(5.6)
(5.7)
(5.8)
(5.9)
(5.10)
(5.11)
r(i ) = Gi = t [tpi (I pi )].
Actually, these operators Gi are just the invertible elements of the algebra,
so that Bn appears here as the group of the algebra An . The inverse to
the generators are
1 1
G1
(5.12)
i = [t pi (I pi )].
t
Each generator G will satisfy a condition of the type (G a)(G b) = 0.
This means that the squared generators are linear functions of the generators,
14
15
399
(normalized so that
so that we have a Hecke algebra. One has P = Ga
ba
2
P = P ) as the projector on the eigenspace of b, in terms of which G =
(b a)P + aI = a(I P ) + bP . The projectors are
(Gi t t)(Gi + t) = 0,
or still
t1
1
t G1
I = 0.
(5.13)
i t Gi +
t
Let us introduce an inspiring notation. Indicate each braid generator by
/
Gi = /\ , where it is implicit that the first line is the i-th, and all the unspecified lines are identity lines. The identity itself is indicated by )( and G1
i
\
suitably by /\ . Relation [5.13] can then be drawn as
\
t /\ - t1 /\ +
t1
)( = 0 .
(5.14)
S
t3/2 +t1/2 T
d
- t1/2
)( .
(5.15)
The first projectors are shown in Figure 5.2 (for the case n = 4).
Figure 5.2:
And a simple blob gives just d as a number multiplier: The bubble is
normalized by
400
= d.
Condition [5.10] is immediate. Conditions [5.9] and [5.8] are shown respectively in Figure 5.3 and Figure 5.4.
Figure 5.3:
Figure 5.4:
The reason for the name monoid diagrams is simple to understand here.
Projectors, with the sole exception of the identity, are not invertible. Once
we add projectors to the braid group generators, what we have is no more
a group, but a monoid (Math.1). The addition of projectors to the braid
generators, or the passing into the group algebra, turns the matrix-diagram
relationship into a very powerful technique.
The Jones polynomials are obtained as follows. Given a knot, obtain it
401
(5.16)
Given a knot, draw it on the plane, with all the crossings well-defined.
Choose a crossing and decompose it according to [5.14] and [5.15]. Two new
knots come out, which are simpler than the first one. The polynomial of the
starting knot is equal to the sum of the polynomials of these two new knots.
Do it again for each new knot. In this way, the polynomial is related to the
polynomials of progressively simpler knots. At the end, only the identity and
the blob remain.
Jones has shown that his polynomials are isotopic invariants. The Jones
polynomial is able to distinguish links which have the same Alexander polynomial. Perhaps the simplest example is the trefoil knot: there are actually
two such knots, obtained from each other by inverting the three crossings.
This inversion corresponds, in the Jones polynomial, to a transformation
t t1 , which leads to a different Laurent polynomial. This means that the
two trefoils are not isotopic.
Kirillov 1974
Haag 1993
Bratelli & Robinson 1979
Dixmier 1981
Dixmier 1982
Kaufman 1991
Jones 1991
402
Math.Topic 6
REPRESENTATIONS
0 Introduction
A LINEAR REPRESENTATIONS
1 Generalities
2 Dimension
3 Unitary representations
4 Equivalent representations
5 Characters
6 Irreducible representations
7 Tensor products
B REGULAR REPRESENTATION
8 Invariant spaces
9 Invariant measures
10 Generality
11 Relation to von Neumann algebras
C FOURIER EXPANSIONS
12 The standard cases
13 Pontryagin duality
14 Noncommutative harmonic analysis
15 The Peter-Weyl theorem
16 Tanaka-Krein duality
17 Quantum groups
Introduction
In general, a representation1 of a group G is a homomorphism of G into
some other group H . A representation of a Lie algebra G0 of a Lie group
G will be a homomorphism of G0 into some other Lie algebra H 0 . All this is
1
403
404
MATH.TOPIC 6. REPRESENTATIONS
6.1
A Linear representations
6.1 Generalities
Consider a vector space V and the set of transformations defined on
it. Of all such transformations, those which are continuous and invertible
constitute the linear group on V , indicated by L(V ). A linear representation
of the group G is a continuous function T defined on G , taking values on
L(V ) and satisfying the homomorphism requirement,
T (gh) = T (g)T (h).
(6.1)
Mackey 1955.
Gelfand, Graev & Vilenkin 1966.
Mackey 1978.
(6.2)
405
Ados theorem (see 8.3.10) may be rephrased as follows: every finitedimensional Lie algebra has a faithful linear representation. Or still: every
finite-dimensional Lie algebra can be obtained as a matrix algebra. If V
is a finite space, such matrices are finite and dT (the dimension of V ) is
the dimension of representation T (g). The homomorphism condition is then
simply
dT
X
Tij (gh) =
Tik (g) Tkj (h).
(6.3)
k=1
When V is infinite, we must worry about the convergence of the involved series. V can be, for instance, a Hilbert space. In the general case of a carrier
space endowed with an inner product (u, v), the adjoint A of a matrix A
satisfies (Au, v) = (u, A v) for all u, v V . It comes out that, if T (g) is a
representation, then T (g 1 ) is another representation, called the representation adjoint to the representation T (g) (this is not to be confounded with
the adjoint representation of a Lie group).
6.3 Unitary representations
A representation T (g) of a group G in the inner product space V is unitary
if, for all u, v in V ,
(T (g)u, T (g)v) = (u, v).
(6.4)
In this case, T (g)T (g) = E and T (g) = T 1 (g) = T (g 1 ), so that a unitary
representation coincides with its adjoint.
Comment 6.1.1 The name comes from the matrix case: in a matrix unitary representation, all the representative matrices are unitary.
6.4 4 Equivalent representations
Let T (g) be a representation on some carrier space V , and let us suppose
there exists a linear invertible mapping f into some other vector space U ,
f : V U . Then, T 0 (g) = f T (g) f 1 is a representation with representation space U . The relation between such representations, obtained
from each other by a linear invertible mapping, is an equivalence. Two such
representations are thus equivalent representations and may be seen as different realizations of one same representation. A well known example is
the rotation group in euclidean 3-dimensional space, SO(3): the vector representation, corresponding to angular momentum 1, is that generated by the
fields
ijk (xj k xk j )
if seen as transformations on functions and fields on E3 ; it is that of the real
orthogonal 33 matrices if seen as acting on column vectors.
406
MATH.TOPIC 6. REPRESENTATIONS
6.5 Characters
Consider, as a simple and basic example, the case in which the carrier
space V is a one-dimensional space. Matrices reduce to functions, so that
T (g) = (g)I, with I the identity operator and a non-vanishing complex
function on G. As T is a homomorphism, (gh) = (g)(h) for all g, h
G. A function like , taking G homomorphically into C\{0}, is called a
character . If G is a finite group, there exists some integer n such that (g n ) =
n (g) = 1, so that |(g)| = 1 and *(g) = (g)1 .
Every finite abelian group is the direct product of cyclic groups ZN . Such
a cyclic group has a single generator g such that g N = I. Hence, N (g) = 1.
For ZN , consequently, the characters are the N -th roots of the identity, one
for each group element. ZN is, therefore, isomorphic to its own group of
characters.
The characters of the direct product of two groups are the respective
product of characters, and it follows that every finite abelian group is isomorphic to its group of characters. The characters of finite abelian groups
have the following further properties:
(i) gG (g) = 0;
(ii) gG 1 (g)2 (g) = 0, whenever 1 6= 2 .
The characters are thus orthogonal to each other. They are in consequence linearly independent, they span the space of functions on the group,
and form a basis for the vector space of complex functions on G . If the
group is unitary, so is the character representation: *(g) = 1 (g).
6.6 Irreducible representations
Suppose that the carrier space V has a subspace V 0 such that T (g)v 0 V 0
for all v 0 V 0 and g G. The space V 0 is an invariant subspace, and
T (g) is reducible. When no invariant subspace exists, the representation is
irreducible. In the reducible case there are two new representations, a T 0 (g)
on V 0 and another, T 00 (g) on the quotient space V 00 = V /V 0 . There exists
always a basis in which the matrices of a reducible representation acquires
the bloc-triangular form
0
T (g) K(g)
.
0
T 00 (g)
There is in general no basis in which K(g) vanishes. This happens, however,
when V 0 admits a linear complement V 00 which is also an invariant subspace,
so that V is the direct sum V = V 0 V 00 of both. Recall that this means
that any v V can be written as v = v 0 + v 00 , with v 0 V 0 and v 00 V 00 . In
this case
T (g)v = T (g)v 0 + T (g)v 00 = T 0 (g)v 0 + T 00 (g)v 00 .
407
dT
X
Tkk (g),
(6.5)
k=1
(6.6)
(6.7)
408
6.2
MATH.TOPIC 6. REPRESENTATIONS
B Regular representation
(6.8)
Recall that on locally-compact groups, the existence of a left-invariant measure is guaranteed, as is the existence of a right-invariant measure. Such Haar
measures, by the way, do not necessarily coincide (groups for which they coincide are called unimodular ). Given the actions of a group G on itself, the
functions defined on G will carry the left and right representations:
L(h)f (g) = (Lh f )(g) = f (h1 g)
R(h)f (g) = (Rh f )(g) = f (gh).
(6.11)
(6.12)
6.10 Generalities
If G has a left-invariant measure, we may take for function space the set
of square-integrable functions. L(g) is called the left-regular representation.
In an analogous way, R(g) is the right-regular representation. Such representations, given by operators acting on functions defined on the group itself,
are of the utmost importance. First, because it happens that
5
Notice to avoid confusion that, as already mentioned, some authors use the name
regular representation in other senses: see for instance Hamermesh 1962, Gilmore 1974.
409
6.3
C Fourier expansions
We discuss some small bits of Fourier analysis, starting with the main qualitative aspects of the elementary case.
6.12 The standard cases
6
7
Mackey 1968.
Mackey 1978, chap. 8.
410
MATH.TOPIC 6. REPRESENTATIONS
einx f(n),
(6.13)
dxeinx f (x)
(6.14)
n=
with both x and y E1 . Analogous statements hold, with the difference that
now both groups are (E1 , +), which is non-compact.
Suppose finally we started with the cyclic group ZN , which is a kind of
lattice circle. Then,
f (m) =
N
1 X i(2/N )mn
e
f (m).
ei(2/N )mn f(n) ; f(n) =
2
m=1
n=1
N
X
(6.16)
411
(6.17)
(6.18)
412
MATH.TOPIC 6. REPRESENTATIONS
d p
XX
d [T (g)]ij f ij ,
(6.19)
i,j=1
where
f
Z
ij
= d
(6.20)
Vilenkin 1969.
413
Vilenkin 1969
Mackey 1978
Kirillov 1974
Katznelson 1976
Enoch & Schwartz 1992
414
MATH.TOPIC 6. REPRESENTATIONS
Math.Topic 7
A CURVES
[1] Variation of a curve
[2] Variation fields
[3] Path functionals
[4] Functional differentials
[5] Second-variation
B GENERAL FUNCTIONALS
[6] Functionals
[7] Linear functionals
[8] Operators
[9] Derivatives Frechet and Gateaux
A geometrical approach to variations is given that paves the way to the introduction of functionals and creates the opportunity for a glimpse at infinitedimensional manifolds.
7.1
7.1.1
A Curves
Variation of a curve
416
We shall consider only paths with fixed ends, from the initial end-point a
= (1) to the final end-point b = (1). Given paths , , going from a
to b, they are homotopic (we write ) if there exists a continuous
mapping F : I I M such that, for every t, s in I,
F (t, 0) = (t); F (t, 1) = (t); F (t, 1) = (t); F (1, s) = a; F (1, s) = b .
The family of curves F is a homotopy including , and , or, if we wish,
a continuous deformation of the curve into the curves and .
For each fixed s, s (t) = F (t, s) is a curve from a to b, intermediate
between and , with (t) = 0 (t) somewhere in the middle. But also, for
each fixed t, t (s) = F (t, s) is a curve going from the point (t) to the point
(t) while s treads I and meets (t) somewhere in between (see Figure 7.1).
Let us fix this notation: s (t) is a curve with parameter t at fixed s; t (s)
is a curve with parameter s at fixed t. The one-parameter family s of curves
is called a variation of .
Figure 7.1:
7.1.2
Variation fields
We now add a smooth structure: M is supposed to be a differentiable manifold and all curves are differentiable paths. Each s (t) will have tangent
vectors
7.1. A CURVES
417
Figure 7.2:
Vs (t) =
d
(t)
dt s
= s ,
which are its velocities at the points s (t). But also the transversal curves
d
t (s) will have their tangent fields, ds
t (s). Consider now the curve (t):
besides its velocity
V (t) = V0 (t) = (t)
d
(t)
dt 0
it will display along itself another family of vectors, a vector field induced by
the variation, (see Figure 7.2)
X(t) =
d
(s)|s=0
ds t
d
(0).
ds t
X(t) will be a vector field on M , defined on each point of (t). This vector
field X is called an infinitesimal variation of . It is sometimes called a
Jacobi field, though this designation is more usually reserved to the case in
which is a geodesic.
7.1.3
Path functionals
Non-exact Pfaffian forms, whose integrals depend on the curve along which
it is taken, open theR way to path functionals. A trivial example is the length
itself: the integral ds depends naturally on the path. Another example is
418
the work done by a non-potential force, which depends on the path along
which the force is transported from point a to point b: the integral
R
Wab [] = Fk dxk
is actually a function of the path, it will depend on the functional form of
(t). Paths belong to a space of functions, and mappings from function spaces
into R or C are functionals. Thus, work is a functional of the trajectory. But
perhaps the most important example in mechanics is the action functional
related to the motion between two points a and b along the path . Let the
points on be given by coordinates i and the corresponding velocity be
given by i . We can use time as the parameter t, so that the velocity
above is the true speed. Let the lagrangian density at each point be given
by L[ i (t), i (t), t]. The action functional of the motion along will be
Z
(7.1)
A[] = L[ i (t), i (t), t] dt.
7.1.4
Functional differentials
7.1. A CURVES
419
the space Hab with a chart. The high cardinality involved is hidden in the
apparently innocuous , which contains an arbitrary functional freedom.
This assumption makes life much simpler than a detailed examination of Banach spaces characteristics, and we shall admit it. Let us only remark that,
Banach spaces being vector spaces, we are justified in adding the coordinates
i as we shall presently do. There is, however, a clear problem in multiplying
them by scalars, as then the end-points could change. This means that paths
do not constitute a topological vector space. It is fortunate that we shall only
need addition in the following.
A differential of the action A (which does not always exist) can be introduced as follows. Take the path and one of its neighbours in the variation
family, given by the coordinates i + i . We represent collectively the coordinates of the path points by and + . Suppose then that we may
separate the difference between the values of their actions into two terms,
A[ + ] A[] = F [] + R,
(7.2)
has many names. In functional calculus its is called Frechet (or strong)
derivative. In mechanics, due to the peculiar form it exhibits, it is called
Lagrange derivative. For the action functional [7.1], we have
A[ + ] A[]
Z
Z
i
i
i
i
= L[ (t) + (t), (t) + (t)]dt [ i (t), i (t)]dt
Z
L i L i
=
+ i dt + O( 2 ),
i
so that
Z
F [] =
L i L i
+ i dt.
i
(7.3)
420
Z
Z
d L i
d L
L d i
dt =
i dt
dt
i
i
i dt
dt
dt
Z
L i
L i
d L
=
i dt.
i
i
i
dt
t=1
t=0
As the variations are null at the end-points, only the last term remains and
we get
Z
L
d L
i dt.
(7.4)
F [] =
i
i
dt
The integrand is now the Lagrange derivative of L. The Euler-Lagrange
equations
L
d L
=0
(7.5)
i
dt i
are extremum conditions on the action: they fix those curves which extremize A[].
7.1.5
Second-variation
d
(t),
dt
DV
Dt
= V V.
7.2
421
B General functionals
A heuristic introduction to functionals which generalize the path functionals will be given
here. Attention will be called on the topologies necessarily involved. The approach is voluntarily repetitive. Functionals generalize to operators. Functionals are real- or complexvalued functions on some spaces of functions, while operators take such spaces into other
spaces of the same kind. Only that minimum necessary to the presentation of the very
useful notion of functional derivative will be treated.
7.2.1
Functionals
422
Functionals appear in this way as functions depending on many (actually, infinite) variables, or on variables whose indices belong to a continuum.
For instance, t may be time, x(t) a trajectory between two fixed points in
E3 , and f will be a functional on trajectories (say, the classical mechanical
action [7.1]). We may of course take the converse point of view and consider
usual few-variable functions as functionals on spaces of functions whose arguments belong to a discrete finite set. Or we may go further in the first
direction, and consider many-dimensional sets of continuous indices. For instance, in relativistic field theory (see Phys.6) such indices are the coordinates
in Minkowski space, the role of the degrees of freedom being then played by
fields (x, y, z, ct). A splendid functional will be the action functional for the
fields .
7.2.2
Linear functionals
The reasoning above makes clear the interest of functionals: they are the
functions defined on infinite dimensional spaces, specially on spaces whose
infinity has the power of the continuum or more. As function spaces can
be always converted into linear spaces, a functional can be defined as any
complex function defined on a linear space. Once we have established that
a functional is a function from a function space {f } to C, we may think of
applying to it the usual procedures of geometry and analysis. But then, to
start with, the function space should be a good (topological) space. Better
still, it should be a manifold. Banach spaces will play here the role played
by the euclidean spaces in the finite case. Some authors2 give another name
to topological vector spaces: they call euclidean spaces any inner-product
linear space, be it finite-dimensional or not. It is necessary to adapt euclidean
properties to the infinite-dimensional case and then consider linear objects.3
We shall below list the main results in words as near as possible to those
describing the finite dimensional case, while stressing the notions, here more
delicate, of continuity and boundedness.
In general a functional is a mapping from a topological linear space into
C. The mapping W is a linear functional when
W [f + g] = W [f ] + W [g] and W [kf ] = kW [f ].
2
7.2.3
423
Operators
7.2.4
Derivatives Fr
echet and Gateaux
(7.6)
when ||g|| 0. When this happens, Mf0 [g] Y is the strong differential
(or Frechet differential ) of M at f ; the linear operator Mf0 is the strong
derivative, (or Frechet derivative), of M at f . The theorem for the derivative
of the function of a function holds for the strong derivative. We may define
another differential: the weak differential (or Gateaux differential ) is
d
M [f + tg] M [f ]
Df M [g] =
M [f + tg]
= lim
,
(7.7)
t0
dt
t
t=0
the convergence being understood in Y s norm topology. In general, Df M [g]
is not linear in g. When Df M [g] happens to be linear in g, then the linear
operator Df M is called the weak derivative, (or Gateaux derivative) at f .
The theorem for the derivative of the function of a function does not hold in
general for the weak derivative.
424
Math.Topic 8
FUNCTIONAL FORMS
0 Introduction
A EXTERIOR VARIATIONAL CALCULUS
1
2
3
4
5
6
7
8
Lagrangian density
Variations and differentials
The action functional
Variational derivative
Euler Forms
Higher order Forms
Relation to operators
Continuum Einstein convention
B EXISTENCE OF A LAGRANGIAN
426
8.1 Introduction
Exterior differential calculus is a very efficient means to compactify notation and reduce expressions of tensor analysis to their essentials. It has
been for long the privileged language of the geometry of finite dimensional
spaces, though it has become a matter of necessity for physicists only in recent times. On the other hand, functional techniques and the closely related
variational methods have long belonged to the common lore of Theoretical
Physics. What follows is a mathematically nave introduction to exterior
variational calculus,1 which involves differential forms on spaces of infinite
dimension in close analogy with the differential calculus on finite dimensional
manifolds. We insist on the word calculus because the approach will be
purely descriptive, operational and practical. It is of particular relevance in
field theory and specially helpful in clarifying the geometry of field spaces.
Only local aspects will be under consideration, meaning by that properties valid in some open set in the functional space of fields. As it has been
the case with differential forms on finite-dimensional manifolds, functional
forms may come to be of help also in the search of topological functional
characteristics.
Special cases of it have been diffusely applied to the study of some specific problems,
such as the BRST symmetry,2 and anomalies,3 but its scope is far more general.
8.1
8.1.1
Lagrangian density
427
(8.1)
f
L
a
is the
where a L is a shorthand for
a (in analogy to f = x ) , and
a
purely functional variation of . We are thus using a natural basis for these
1-forms on the bundle. Vector fields (in the geometrical sense of the word)
X = X a
a . In the spirit of field theory, all information is contained in the
fields, which are the degrees of freedom. Consequently, L will be supposed
to have no explicit dependence on x , so that
L = (a L) a .
Of course, for the part concerned with variations in spacetime (i.e., in the
arguments of the fields), we have usual forms and
d2 L =
8.1.2
1
2
[ L L] dx dx = 0.
(8.2)
8.3 This is precisely one of the results of exterior calculus which we wish
to extend to the -space. This extension is natural for the operator:
2L =
1
2
[a b L b a L] a b = 0.
(8.3)
(8.5)
428
Comment 8.1.1 This behaviour is no real novelty, as it appears normally in differential calculus when we separate a manifold into two subspaces.
8.4 The total variation T does not commute with spacetime variations,
as
[ , T ]f = ( x ) f,
(8.6)
but the purely functional variation does.
8.1.3
8.5 We shall from now on consider only purely functional variations. Furthermore, instead of densities as the L above, we shall consider only objects
integrated on spacetime, such as the action functional
Z
(8.7)
S[] = d4 xL[(x)].
For simplicity of language, we shall sometimes interchange the terms lagrangian and action. Differentiating the expression above,
Z
Z
4
S[] = d xL[(x)] = d4 x{a L[(x)]}a (x).
(8.8)
8.1.4
Variational derivative
8.6 We shall suppose that the conditions allowing to identify the strong
and the weak derivatives are satisfied (see Math.7). Thus, the differential5
of a functional F [] at a point of the function space along a direction (x)
in that space will be defined by
dF [ + ]
F [ + ] F []
0
=
F [] = lim
.
(8.9)
0
d
=0
It is a linear operator on (x), and F 0 [] = 0 is a linearized version of the
equation F [] = 0. The integrand in (8.8) is the Frechet derivative of L[]
along = /. The presence of the integration, allied to property (8.5),
justifies the usual procedures of nave variational calculus, such as taking
variations (in reality, functional derivatives and not differentials) inside the
common derivatives which, allied to an indiscriminate use of integrations by
parts (that is, assuming convenient boundary conditions), lends to it a great
simplicity.
5
The expression given is actually that of the weak (Gateaux) differential. When the
strong derivative exists, so does the weak and both coincide. We suppose it to be the case,
and use the most practical expression.
8.1.5
429
Euler Forms
8.7 The vanishing of the expression inside the curly bracket {} in (8.8)
gives the field equations, a L[(x)] = 0. Given a set of field equations
Ea [(x)] = 0, we shall call its Euler Form the expression
Z
E[] = d4 xEa [(x)]a (x).
(8.10)
The exterior functional (or variational) differential of such an expression will
be defined as
Z
E[] = d4 xEa [(x)] a (x)
R
= 12 d4 x{b Ea [(x)] a Eb [(x)]}b (x) a (x). (8.11)
The differential of (8.8) is immediately found to be zero: 2 S[] = 0.
8.1.6
8.1.7
Relation to operators
(8.14)
430
(8.15)
8.2
B Existence of a lagrangian
8.2.1
8.2.2
Helmholtz-Vainberg theorem
8.11 Consider the expression (8.11). The Ea [(x)] are densities just as
L[(x)], and the differentials appearing are Frechet differentials,
Ea = {b Ea []} b = Ea0 [].
6
(8.16)
8.2.
B EXISTENCE OF A LAGRANGIAN
431
(8.17)
8.2.3
(8.19)
(8.20)
(8.21)
with the relative sign conveniently chosen. After putting E under the form
E = (vj j vi )v i + [ 21 (j vi j v i ) p(j v j )],
7
8
Vainberg 1964.
Aldrovandi & Pereira 1986, 1988.
432
(8.22)
(8.23)
(8.24)
However, when the choice of the fundamental physical field is given by some
other reason, as in quantum field theory, it is of no great help that a lagrangian may be found by some smart change of variable.
There is an obvious ambiguity in writing the Euler Form for a set of two
or more field equations, as multiplying each equation by some factor leads
to an equivalent set. Such a freedom may be used to choose an exact Euler
Form and to give the lagrangian a correct sign, for example leading to a
positive hamiltonian.
8.3
C Building lagrangians
8.3.1
Finlayson 1972.
8.3.
C BUILDING LAGRANGIANS
433
(8.25)
The fields a appearing in the argument of Za1 a2 ...ap are multiplied by the
variable t before the integration is performed. As t runs from 0 to 1, the
field values are continuously deformed from 0 to a . This is a homotopy
operation10 in -space. A more general homotopy t = t + (1 t) 0
with 0 6= 0 can be used, but without real gain of generality. The important
point is that the -space is supposed to be a starshaped domain around some
zero field (each point may be linked to zero by straight lines). Spaces of
this kind are called affine spaces by some authors. Some important field
spaces, however, are not affine. For example, the space of metrics used in
General Relativity includes no zero, nor does the space of chiral fields with
values on a Lie group. For such cases, the use of (8.26) is far from immediate
(see Phys.7).
8.15 The Poincare inverse lemma says that, on affine functional spaces, Z
can always be written locally as
Z[] = (T Z) + T (Z).
(8.27)
434
(8.29a)
(8.29b)
8.3.2
Examples
(8.31)
F 2 /(2k)
A .
(8.32)
In this case,
TE = A
R1
0
dt
t F
1t2 F 2 /(2k)
(8.33)
8.3.
435
C BUILDING LAGRANGIANS
(8.34)
E = g R 12 g (R + ) g ,
E = [ g(R + )] g g R .
Of these two terms, the latter is known to be a divergence12 and the first
436
L = [ + ieA ] [ ieA ] + m + 4 41 F F .
8.21 Second order fermion equation
Applying twice the Dirac operator, we obtain a second order equation
for the fermion, which includes a spin-field coupling introduced by Fermi to
account for the anomalous magnetic moment of the neutron. In this case,
i
e
E = u
t ie( A ) 2ieA e A A F
2
i
h
e
2
+ u
t + ie( A ) + 2ieA e A A F
2
+ F + e + e ( ) A .
h
8.3.3
LF
A
LF
A
Symmetries of equations
(8.35)
8.3.
C BUILDING LAGRANGIANS
437
symmetries of the lagrangian. This is a well known fact, but here we find a
necessary condition for that: LX = 0. Still other notions of differential
calculus translate easily to Forms, keeping quite analogous properties. Such
is the case, for example, of the interior product iX W of a field X by a Form
W , which has the usual relation to the Lie derivative,
LX W = iX (W ) + (iX W ).
8.23 We have shown some examples of the power of exterior variational
calculus in treating in a very economic way some involved aspects of field
theories. All examined cases were local, valid in some open set of the field
space. Recent years have witnessed an ever growing interest in the global,
topological properties of such spaces. Anomalies, BRST symmetry and other
peculiarities (see Phys.7) are now firmly believed to be related to the cohomology of the field functional spaces, this belief coming precisely from results
obtained through the use of some special variational differential techniques.
Many global properties of finite dimensional manifolds are fairly understood
and transparently presented in the language of exterior differential forms.
The complete analogy of the infinite dimensional calculus suggests that, besides being of local interest, it is a natural language to examine also global
properties of field spaces.
438
Math.Topic 9
SINGULAR POINTS
1 Index of a curve
2 Index of a singular point
3 Relation to topology
4 Examples
5 Critical points
6 Morse lemma
7 Morse index and topology
8 Catastrophes
9.1
Index of a curve
A very complete treatment of the subject is given in Doubrovine, Novikov & Fomenko
1979, Vol. II, 13-15.
439
440
X will turn around itself a certain number of times, both in the clockwise
sense (taken by convention as negative) and in the counterclockwise sense
(positive by convention). The algebraic sum of this number of turns is the
index of the curve with respect to the field X. For instance, the index equals
+1 in the case pictured in Figure 9.1. If it so happens that p is not a singular
point of the field X, we can always find, thanks to the continuity of X, a
small enough neighbourhood of p inside which all curves have null index.
In Figure 9.2 we show an example: in the complex plane version of E2 , the
behaviour of the field
X(z) = z 2 = (x + iy)2 = x2 y 2 + i2xy = (x2 y 2 ) x + 2xy y ,
when it traverses the circle |z| = 1 around its singular point z = 0. The index
is + 2. It is easy to see that, had we taken a point outside the curve, the
index would be zero. In general, the circle will have index (+n) with respect
to the field X(z) = z n and (n) with relation to the field X(z) = z n (which
illustrates the role of the orientation). A practical way to find the index is to
draw the vectors directly at the origin and follow the angle it makes with
R+ .
Figure 9.1:
Let us formalize the above picture: suppose a chart (U, c) and a field
X = X 1 e1 + X 2 e2 .
Let U 0 = U { singular points of X}, and p U 0 . Define the mapping
f : U 0 S 1 , from U 0 into the unit circle given by
f (p) =
Xp
|Xp |
441
Figure 9.2:
x2 y 2
x2 +y 2
+ i x22xy
= X 1 + iX 2 .
+y 2
Figure 9.3:
d = d arctan xx1 =
x2 dx1 x1 dx2
|x|2
442
is a differential form well defined on the whole plane outside the singular
points. The index of a closed curve : S 1 U 0 is then defined as
I
1
ind =
d
(9.1)
2
In the example X(z) = z 2 ,
f = arctan x22xy
y 2
It is easier to define z = rei , X(z) = r2 e2i , so that f = 2 . While goes
from 0 to 2, f goes from 0 to 4. Consequently,
H
1
ind = 2
d(2) = 2.
Notice an important thing: the mapping f (z), defined through the field
X, takes two points of U 0 in one same point of S 1 .
The index is just the number of points from the domain of f taken into
one point of its image. This is a general property, which will allow the
generalization to higher dimensional manifolds. Let us list some results which
can be proven about the indices:
(i) they do not change under continuous deformations of the curve ,
provided never touches any singular point;
(ii) they do not change under continuous deformations of the field, provided X never has a singular point on the curve;
(iii) every disk whose contour has a non vanishing index related to a field
contains some singular point of that field;
(iv) the index of a curve contained in a small enough neighbourhood of
a singular point is independent of the curve, that is, it is the same for any
curve; the index so obtained is the index of the singular point; this index is
chart-independent;
(v) if a curve encloses many singular points, its index is the sum of the
indices of each point. A good example is given by the field V (z) = z 2 (z/2);
it has two singular points in the unit disk. The index of z = 0 is 2, that of z
=1/2 is zero, and the index of the unit curve |z|2 = 1 is 2;
(vi) the index of a singular point is invariant under homeomorphism; this
allows the passage from the plane to any bidimensional manifold, as it is a
purely local property.
9.2
In order to generalize all this to singular points of fields on general differentiable manifolds, let us start by recalling what was said in section 5.3: any
443
Xp
f [Xp ]
=: r
.
|f [Xp ]|
|Xp |
(9.2)
The index of the singular point is the Brouwer degree (6.2.15) of this mapping.
9.3
Relation to topology
9.4
In the two dimensional case, drawing the local integral lines is of great help
to get intuition on the corresponding fields. Figure 9.4 shows some simple
singular points, with their names and corresponding indices. Notice that the
index does not change if we invert the field orientation. Figure 9.5 shows
444
two other singularities of great importance: the saddle point and the dipole.
Sources, crosspoints and sinks may be taken as elementary singular points.
For 2-dimensional manifolds, we may cast a bridge towards the homological version of the Euler number by taking a triangulation and putting (i) a
source at each vertex, (ii) a crosspoint replacing each edge, and (iii) a sink
at the center of each loop.
Figure 9.4:
9.5
Critical points
Notice that, from the drawings above, the index is not enough to characterize completely the kind of singular point. Such a characterization requires
further analysis, involving a field linearization: near the singular point, it
is approximated so as to acquire a form x = Ax, x being a set of coordinates
and A a matrix. The eigenvalues of A provide a complete classification.2
On n-dimensional metric spaces, as said, connection may be made with the
critical points of functions, which are singular points of their gradient fields.3
Let us go back to the differentiable function f : M N . We have
defined its rank as the rank of the jacobian matrix of its local expression in
2
445
Figure 9.5:
9.6
Morse lemma
Milnor 1973.
446
the critical point p such that y(p) = 0 and, for y y(W ), the function
f y 1 (y1 , y2 , . . . , yn ) can be written as a quadratic form:
2
2
f y 1 (y1 , y2 , . . . , yn ) = f (p) y12 y22 . . . yk2 + yk+1
+ yk+2
+ . . . + yn2 (9.4)
9.7
Another enthralling result links the critical points of any smooth function to
the topology of the space: if N is compact, and nk denotes the number of
critical points with index k, then
n
X
i=0
()k nk = (N ),
(9.5)
9.8. CATASTROPHES
447
9.8
Catastrophes
Thom 1972. The main stimulus for Thom came from optics and the idea of form
evolution in biology. A more readable text is Thom 1974. Zeeman has applied the theory
to biology, medicine, social sciences, etc. His works are collected in Zeeman 1977, where
much material and references can be found. A general appraisal of the theory, as well as
of the controversy its applications have raised, is found in Woodcock & Davis 1980.
448
Figure 9.6:
Berry 1976.
Math.Topic 10
EUCLIDEAN SPACES AND SUBSPACES
0 Introduction
A STRUCTURE EQUATIONS
1 Moving frames
2 The Cartan lemma
3 Adapted frames
4 Second quadratic form
5 First quadratic form
B RIEMANNIAN STRUCTURE
6 Curvature
7 Connection
8 Gauss, Ricci and Codazzi equations
9 Riemann tensor
C GEOMETRY OF SURFACES
10 Gauss Theorem
D RELATION TO TOPOLOGY
11 The Gauss-Bonnet theorem
12 The Chern theorem
0 Introduction
Euclidean spaces are, as we have seen, the fundamental spaces to which
manifolds are locally homeomorphic. In addition, differentiable manifolds
can always be imbedded in some euclidean space of high enough dimension
(section 5.3). For a n-dimensional manifold N , the Whitney theorem only
guarantees that this is always possible in a (2n + 1)-dimensional space, but
sometimes a smaller dimension is enough: for instance, S 2 may be imbedded
449
450
in E3 . On the other hand, one more dimension is not always sufficient: the 4dimensional Schwarzschild space, solution of Einsteins equations, cannot be
imbedded in E5 . No general rule giving the minimal dimension for euclidean
imbedding is known. We shall here consider the manifold N immersed in
some En+d , with d large enough. This will allow us to touch on some results
of what is nowadays called classical differential geometry.
10.1
A Structure equations
10.1.1
Moving frames
Recall (section 7.3) that, given a moving frame {ei } on En+d , the structure
equations are
d j = j i i ;
j
d i = i k .
10.1.2
(10.1)
(10.2)
10.1.3
Adapted frames
451
Comment 10.1.1 We are used, in the simple cases we usually meet in current Physics,
to take a space (preferably euclidean), there fix a frame once for all, and refer everything
to it. We even lose the notion that some frame is always involved. This simple-minded
procedure fails, of course, whenever the space is somehow non-euclidean. The amazingly
simple idea of Cartan was to consider instead a bunch of moving frames, valid actually
in a neighbourhood of each point, and whose set will finally constitute the bundle of frames
(section 9.3).
The imbedding i induces forms i ( j ) and i ( j i ) on U . The pull-back
commutes with the operations of exterior product and exterior differentiation.
The basic point of the method of moving frames comes thereof: the structure
equations valid on V hold also on some open of N . It will be better to use
the indices i, j, k, . . . from 1 to n + d, as above; , , , . . . in the range 1 to n;
and indices a, b, c, . . . from n + 1 to n + d. Separating the structure equations
in an obvious way, they become
d = + a a ;
a
d =
d = + c c ;
(10.6)
c;
(10.7)
d b a = a b + c a b c .
(10.8)
=
=
(10.5)
(10.4)
(10.3)
c;
10.1.4
(10.9)
(10.10)
10.1.5
The first quadratic form is the metric on U induced by the imbedding: given
u and v Tp U , this metric is defined by
< u, v >p := < ip (u), ip (v) > .
(10.11)
The metric and the fields {e } determine the and . We say then that
all these objects belong to the intrinsic geometry of U .
452
10.2
B Riemannian structure
10.2.1
Curvature
Let us compare [10.5] with [10.2]. The latter is valid for the euclidean space.
The 2-forms = c c measure how much N departs from an euclidean
space, they characterize its curvature. In terms of intrinsic animals, that is,
in terms of objects on N itself, they are, from [[10.5]],
= d .
(10.12)
They are the curvature forms on N . With the forms acting on the space
tangent to N , the structure equation [10.3] reduces to
d = .
10.2.2
(10.13)
Connection
(10.14)
(10.15)
(10.16)
(10.17)
Matrix behaves, under base changes, in the usual way matrices do under
linear transformations.
10.2.3
453
(10.18)
They are the normal curvature forms. This expression may be combined with
[10.9] and [10.13] to give the Gauss equation
X
(ha ha ha ha )
(10.19)
= 21
a
(10.20)
The imbedding divides the geometry into two parts, which are related
by the second quadratic forms and by eq.[10.6], which is called the Codazzi
equation:
d a = a + c a c .
(10.21)
The above equations constitute the basis of classical differential geometry.
The important point is that the geometrical objects on N (fields, forms,
tensors, etc) may be given a treatment independent of the exterior objects.
All the relations involving the indices , , , etc may be written without
making appeal to objects with indices a, b, c, etc. This fact, pointed out
by Gauss, means that the manifold N has its own geometry, its intrinsic
geometry, independently of the particular imbedding. This may be a matter
of course from the point of view we have been following, but was far from
evident in the middle of the nineteenth century, when every manifold was
considered as a submanifold of an euclidean space. The modern approach has
grown exactly from the discovery of such intrinsic character: the properties of
a manifold ought to be described independently of references from without.
10.2.4
Riemann tensor
(10.22)
(10.23)
454
(10.24)
These components constitute the Riemann curvature tensor. The components of the connection form [10.23] are the Christoffel symbols, which may
be written in terms of derivatives of the components of the metric tensor.
These metric components are, when restricted to the intrinsic sector,
g = e eg .
Of course, they are now point-dependent since the adapted basis vectors
change from point to point.
The Ricci tensor R = R is symmetric on a Riemannian manifold.
A manifold whose Ricci tensor satisfies R = g , with a constant, is
called an Einstein space. There are very interesting theorems concerning the
immersion of Einstein spaces. One of them is the following:
if an Einstein space as above has dimension m and is immersed
in Em+1 , then necessarily 0.
Another curious result is the following:
suppose that, on a connected manifold of dimension m, R = f g ,
with f a function; then, if m 3, f is necessarily a constant.
To get the Christoffel symbols (see Phys.8), we start by differentiating
the function g ,
dg = dx g = de e + e de
= e e + e e = g + g
= g + g dx .
Defining := g , we see that
g = + .
Calculating g + g g , we arrive at
= g = 12 g [ g + g g ] .
(10.25)
10.3
C Geometry of surfaces
10.3.1
Gauss Theorem
455
To get some more insight, as well as to make contact with the kernel of
classical geometrical lore, let us examine surfaces imbedded in E3 . Consider
then some surface S, dim S = 2, and an imbedding i : S E3 . Two vectors
u, v Tp S will have an internal product given by [10.11], the metric on
S induced through i by the euclidean metric of E3 . To examine the local
geometry around a point p S, take an open U, S U 3 p, and an open V
such that E3 V i(U ). Choose on V a moving frame (e1 , e2 , e3 ) adapted
to i in such a way that e1 and e2 are tangent to i(U ), and e3 is normal. The
orientation may be such that (e1 , e2 , e3 ) is positive in E3 . Here, to simplify
the notation, we take the imbedding as a simple inclusion, all the geometrical
objects on S being considered as restrictions to S of objects on E3 . Given
any vector v = v 1 e1 + v 2 e2 on S, it follows that 3 (v) = 0. Equation [10.4]
becomes
0 = d 3 = 3 1 1 + 3 2 2 .
It follows from the Cartan lemma that
3 1 = h11 1 + h12 2
3 2 = h21 1 + h22 2 ,
with
h12 = h21 ,
(10.26)
(10.27)
456
(10.28)
(10.29)
(10.30a)
(10.30b)
1 = cos 1 + sin 2 ,
02
= sin 1 + cos 2 .
(10.31a)
(10.31b)
d 1 = 2 ( 1 2 + d),
d 2 = 1 ( 2 1 + d).
(10.32)
As the forms satisfying such equations are unique, it follows that the connection form of the new basis is
0
2 1 = 2 1 + d.
0
(10.33)
457
were very helpful in finding fundamental properties and making them more
easily understood, all of them can be arrived at in an intrinsic way, the
only difficulty being the necessity of a more involved formalism. In higher
dimensional spaces, the curvature is characterized by all the components
R of the Riemann tensor which, for a n-dimensional space, amount to
n2 (n2 1)/12. When n = 2, only one component is enough to characterize
the curvature, as above. In this case, using equations [10.19], [10.21] and
[10.28], we find that
K = 12 R1212 .
10.4
D Relation to topology
10.4.1
21
Ui
21 =
Ui
21
Ui
R
Ui
d 2 1 .
0
Keep in mind that 2R1 is defined on the whole S, while 2 1 is only defined
on S 0 . The integral Ui d 2 1 can be made to vanish by taking Ui smaller
0
and smaller, so that Ui {pi }. The form 2 1 keeps itself out from the pi s.
Recalling that the sum of all the indices is the Euler characteristic, we arrive,
in the limit, at
P R
P R
0
0
2 = i Ui 2 1 = i SUi d 2 1
because, in a compact, the union Ui is also the boundary of S = S - Ui
with reversed orientation. From [10.29],
458
2 =
0
R
SUi
K 1 2 .
As is independent of the induced metric and of the chosen field, this relation
depends only on S. It holds clearly also for any manifold diffeomorphic to
S. It is a relation between a differentiable characteristic of the manifold,
the curvature, and the topological substratum. It is a famous result, the
Gauss-Bonnet theorem. For a sphere S 2 of radius r, the Gaussian curvature
is K = 1/r2 , = r2 sin dd, and we obtain = 2 again. Notice that
1
= 2
K g dx1 dx2 .
(10.35)
10.4.2
The theorem above has been generalized to manifolds of dimension 2n. First,
imbeddings were used. Allendoerfer and Weil found its first intrinsic proof
in 1943. Two years later, a simpler (for mathematicians) proof, using fiber
bundles, was given by Chern. For these 2n-manifolds, the Euler characteristic
is
Z
2
= a2n
KT g dx1 dx2 . . . dx2n .
(10.36)
S
Math.Topic 11
NON-EUCLIDEAN GEOMETRIES
11.1
Euclids postulate of the parallels (his 5-th postulate) stated that, given a
straight line L and a point P not belonging to it, there was only one straight
line going through P that never met L. The eon-long debate on this postulate
was concerned with the question of its independence: is it an independent
postulate, or can it be deduced from the other postulates? Euclid himself
was aware of the problem and presented separately the propositions coming
exclusively from the first four postulates. These propositions came to constitute absolute geometry. Those dependent on the 5-th postulate, as for
example the statement that the sum of the internal angles of a triangle is
equal to , he set apart. The debate was given a happy ending around the
middle of the 19th century through the construction of spaces which kept
in validity the first four postulates, but violated precisely the 5-th. On one
side, the independence was thereby proved. On the other, the very way by
which the solution had been found pointed to the existence of new, hitherto
unsuspected kinds of space. Such new non-euclidean spaces are at present
time called Riemannian spaces and their character is deeply rooted in their
metric properties. Though the word space has since then acquired a much
more general, metric-independent meaning, we shall in this chapter follow
the widespread usage of using it to denote a metric space. The main fact
about non-euclidean spaces is spelled out by saying that they are curved.
On such spaces, the role of straight lines is played by geodesics.
459
460
11.2
461
Figure 11.1:
will be to consider both S 2 and a hyperbolic space imbedded in E3 , and
project them into a plane, thereby obtaining curved spaces on R2 . The
projection to be used, the stereographic projection, has very nice properties,
in special that of preserving circles. Geodesics are sent into geodesics, and
angles are also preserved. It turns out that, if we want to preserve metric
properties in the hyperbolic case, the imbedding space must be not E3 , but
the pseudo-euclidean space E2,1 instead.
The treatment is, up to the dimension, identical to that of the de Sitter
spaces (Phys.9). We shall, consequently, concentrate here on some basic aspects and refer to that chapter for others, as for example, for the justification
of the above statements on the values of the scalar curvature.
11.3
462
4L2
x2 +y 2 +4L2
LZ
.
2L
If we call
r2 = (x2 + y 2 )/4L2 ,
the proportionality coefficient will be
n=
1
.
1 + r2
(11.2)
x
y
1 r2
;
Y
=
;
Z
=
L
.
1 + r2
1 + r2
1 + r2
(11.3)
2LX
2LY
; y=
.
LZ
LZ
(11.4)
dx2 + dy 2
.
(1 + r2 )2
(11.5)
463
L+Z
,
LZ
so that lines at constant Z (horizontal circles) will be led into points satisfying
r2 = constant. The equator (Z = 0), in particular, is taken into r2 = 1, or
x2 + y 2 = 4L2 . Each great circle meeting the equator at (X, Y ) will meet it
again at (X, Y ), and this will correspond to (x, y) and (x, y). In the
general case, a great circle is the intersection of S 2 with a plane through the
origin, with equation uX + vY + wZ = 0. This plane is orthogonal to the
Figure 11.2:
vector (u, v, w), whose modulus is irrelevant, so that actually the constants
are not independent (if u, v, w are direction cosines, u2 + v 2 + w2 = 1).
464
Figure 11.3:
Examine first the planes going through the axis OZ: w = 0 and then Y =
(u/v)X =: X. The intersection projection will satisfy the equation y =
x, representing straight lines through the plane origin. Now, w 6= 0 when
the circle is in general position, which leads to
Z = X + Y = n(x + y)
with = u/w and = v/w. We then obtain from n = (L Z)/2L the
equation
(x 2L)2 + (y 2L)2 = 4L2 (1 + 2 + 2 ),
representing circles centered at the point (2L, 2L).
Geodesics are great circles. Consequently, all the geodesics starting at a
given point of S 2 will intersect again at its antipode. And we see in this way
how Euclids postulate of the parallels is violated: there are no parallels in
such a space, as any two geodesics will meet at some point.
11.4
X 2 + Y 2 Z 2 = L2 or Z = X 2 + Y 2 + L2 .
465
Again we choose the point (0, 0, L) as projection center (it is now the lowest
point of the upper branch) and project each point of the hyperboloid on
the plane tangent at the point (0, 0, L) (which is now the highest point on
the lower branch), as indicated in Figure 11.3. The same reasoning applied
above to the spherical case will lead again to the relations X/x = Y /y =
(L Z)/2L = n, but the form X 2 + Y 2 Z 2 = L2 leads to another
expression for the function n. Instead of [11.2], we have now
n(x, y) =
1
,
1 r2
(11.6)
x
y
1 + r2
;
Y
=
;
Z
=
L
.
1 r2
1 r2
1 r2
(11.7)
Now we have a problem. We would like to have the equations defining the
surfaces to represent relations between measured distances, and the interval
to be ds2 = dX 2 +dY 2 dZ 2 . This is obviously impossible with the euclidean
metric. In order to preserve that idea, we must change the ambient space
and consider pseudo-euclidean spaces. The above non-compact surfaces will
be (pseudo-) spheres in such spaces.
The pseudo-euclidean space E2,1 is the set point R3 with, instead of [11.1],
the pseudo-distance function
p
(11.8)
d(p, q) = + (p1 q1 )2 + (p2 q2 )2 (p3 q3 )2 .
Due to the negative sign before the last term, this is not a real distance
function and does not define a topology. Taking the origin q = (0, 0, 0) as
the center, there are spheres of three types, according to the values of their
radius L: precisely spheres of real radius, satisfying X 2 + Y 2 Z 2 = L2 ;
those of vanishing radius, X 2 +Y 2 Z 2 = 0; and those with purely imaginary
radius, satisfying X 2 +Y 2 Z 2 = L2 . Thus, the above surfaces in E3 appear
as the possible spheres in E2,1 , with the extra bonus that now the equations
have a metric sense.
Returning to the specific case of the two-sheeted hyperboloid, we have
that the interval in the plane coordinates turns out to be
ds2 = dX 2 + dY 2 dZ 2 = n2 (dx2 + dy 2 ) =
dx2 +dy 2
.
(1r2 )2
1
ij .
(1 r2 )2
(11.9)
466
Metrics like this one and that defined in [11.5], which are proportional to a
flat metric, are called conformally flat metrics. Though they give measures
of distance different from those of the euclidean metric, they give the same
measure for angles.
The plane R2 with the above metric is called the Lobachevsky plane.
Notice that the metric actually brings the hyperboloid down and up to the
plane. The lower sheet has Z/L < 1 and is mapped into the disc bounded
by the circle r2 =1. The disc r2 < 1 with the metric [11.9] is called the
Poincare space. We shall come to it later. The upper sheet has Z/L > 1
and is mapped into the complement of the disc in the plane. As we go to
infinity in the upper and lower sheets we approach the circle (respectively)
from outside and from inside.
We may analyze the projections of intersections of the hyperboloids with
planes (the now eventually open great circles) in the same way used for
the spherical case. Putting together the two metrics by writing
n = (1 r2 )1 ,
we find on the plane the curves
(x 2L)2 + (y 2L)2 = 4L2 (2 + 2 1).
11.5
(11.10)
=
+ 2v k
+ 2 k n2 = 0.
(11.12)
ij
2
ds
ds ds
ds
ds
It is a happy fact that, whenever we may define a momentum by
pi = gij v j ,
(11.13)
the geodesics equation simplify because the first two contributions in the
Christoffeln [11.11] just cancel the term coming from the derivative of the
metric. We remain with
dpk 1
+ 2 [k g ij ]pi pj = 0.
(11.14)
ds
SPACE
11.6. THE POINCARE
467
[1r2 ]2
.
4L2
(11.16)
or
D2 X
1
+
[X g(X, V )V ] = 0.
(11.17)
Ds2
L2
Now, X = [X g(X, V )V ] is the component of X transversal to the curve.
D2
||
As the tangential part X || has Ds
= 0, one arrives at
2X
1
D2 X
+ 2 X = 0.
2
Ds
L
11.6
(11.18)
The Poincar
e space
Consider now the Poincare space. We may consider the plane E2 as the
complex plane. It is known that an open disc on the complex plane can be
taken into the upper-half-plane by a homographic transformation, which is
furthermore a conformal transformation. Actually, there is one transformation for each point of the half-plane. For each arbitrarily chosen point a
on the half-plane, there will be one homographic half-plane disc transformation taking a into the circle center.1 Introducing the complex variables
z = x + iy and w = u + iv, the transformation
wa
z=K
(11.19)
w a
1
468
takes the open upper-half-plane onto a disc of radius K whose center is the
transformed of a. We choose for convenience a = iaL. The relations between
the coordinates are:
x = 2L
u2 + v 2 4L2
4Lu
; y = 2L 2
,
2
2
u + (v + 2L)
u + (v + 2L)2
(11.20)
gij =
1
v2
ij ,
1
[ij k2 + ik j2 jk i2 ].
v
(11.22)
(11.23)
1 2
[v u 2 ] = 0.
v
(11.24)
u
v
Putting
z = ln v; z = v1 v = v1 v v1 v 2 ; z = cu
will lead to z = cu + d. The two remaining equations, v1 v = d cu and
u = cv 2 are put together in
SPACE
11.6. THE POINCARE
469
dv
du
dcu
,
cv
Figure 11.4:
Figure 11.5:
The Poincare disc has a very curious zoology of curves. Some are indicated in Figure 11.5. There are equidistant curves (as a and b), circles (c),
470
and circles which are tangent to the infinity circle r2 = 1. The latter goes
under the honest name of limiting circles, but also under those of oricycles, or
still horocycles (curve h in Figure 11.5). They, and their higher-dimensional
analogues (horispheres), are important in the study of representations of the
groups of motions on hyperboloids.2 Comparison with the corresponding
curves in the half-plane is an amusing exercise. It serves at least to illustrate
the vanity of a curve aspect, which depends heavily on the coordinate system.
Whittaker 1958
Rosenfeld & Sergeeva 1986
Lavrentiev & Chabat 1977
Vilenkin 1969.
Math.Topic 12
GEODESICS
A wellknown aphorism says that there is only one experiment in Physics: Rutherfords.
In order to acquire information on a blackbox system (an atom, a molecule, a nucleus),
we examine the resulting paths of a wellknown probe (a particle, a light beam) sent
against it. Comparison is made with the paths of the free case, in which the black
box system is not there. That is to say that Physics dwells frequently with the study of
trajectories. In this chapter we describe how to obtain information on a geometry through
the study of its curves. It turns out that what is best probed by trajectories is always a
connection.
Introduction
The word geodesic is used in more than one sense. In the common lore, it
means the shortestlength curve between two points. In a more sophisticated
sense, it represents a selfparallel curve, whose tangent vectors keep parallel
to themselves along the curve. The first concept assumes a metric to measure
the lengths, whereas the second supposes a welldefined meaning for the idea
of parallelism. These are quite different concepts. The ambiguity comes
from the fact that geodesics have been conceived in order to generalize to
curved spaces the straight lines of euclidean spaces. They are as straight as
possible on noneuclidean spaces. The trouble is that the starting notion
of a straight line is itself ambiguous.
A straight line can be seen either as the shortestlength curve between
two points or as a curve keeping a fixed direction all along. The two ideas
coincide for euclidean spaces, but not for more general spaces. The shortest
length point of view is meaningful only on strictly Riemannian manifolds
(endowed with a positivedefinite metric). The selfparallel idea has a
meaning for pseudoRiemannian spaces, and even in nonmetric spaces. As
it happens, parallelism is an idea related to a connection, which may be or
not related to a metric. As the selfparallel concept is more general and
reduces to the shortestlength point of view in the metric case, it is to be
471
472
12.1
Selfparallel curves
12.1.1
In General Relativity
d2 y k
.
dt2
Time is absolute, so that t has the same value for all events in space. Space
time is thus a direct product E1 E3 . Now pass into another, arbitrary
coordinate system {xk }, in terms of which the distance between two infinitesimally close points in spacetime is written ds2 = gij dxi dxj . Then the
expression for the acceleration becomes
A0k =
i
j
d2 xk
k dx dx
+
,
ij
dt2
dt dt
(12.1)
(12.2)
There is nothing new in this change of appearance, only a change of coordinates. The equation stating the absence of forces,
i
j
d2 xk
k dx dx
+ ij
=0,
dt2
dt dt
(12.3)
gives simply a straight line written in the new coordinates. The set {gij }
contains the components of the euclidean metric of E3 , which are {ij } in
cartesian coordinates {y k }, written in the new coordinates {xk }. It only
happens that, in euclidean cartesian coordinates, the k ij s vanish, and so
do their derivatives.
Suppose now that this is no more the case, that gij is another, non
euclidean metric. The k ij s are then the Christoffel symbols, representing
473
12.1.2
(12.5)
d
d
( ) ,
(12.6)
DV
= V V .
D
(12.8)
This expression holds in any basis and differs from (12.1) only because
the connection involved is quite general. In the basis {ea = a }, the velocity
field V is dd and the acceleration will be written
d b d c
d2 a
a
A=
+ bc
ea .
d 2
d d
(12.9)
474
DS
D
dS
d
if
12.1.3
Selfparallelism
The curve will be a selfparallel curve when the tangent velocity keeps
parallel to itself along , that is, if
DV
= V V = 0 .
(12.10)
A=
D
In the basis {ea = a },
dV a
d2 a
d b d c
a
+ bc
=
+ a bc V b V c = 0 .
(12.11)
2
d
d d
d
A selfparallel curve is more suitably called a geodesic when the connection
is a metric connection (as said, there are fluctuations in this nomenclature: many people call geodesic any selfparallel curve). The equation above,
spelling the vanishing of acceleration, is the geodesic equation.
Comment 12.1.2 The parametrization t (t) = 0 (t)), 1 (t), 2 (t), 3 (t) is, when
(t) satisfies the geodesic equation, defined up to an affine transformation t s = at + b,
where a 6= 0 and b are real constants. For this reason, such a t is frequently called an
affine parameter.
(12.12)
475
12.1.4
Complete spaces
12.1.5
Fermi transport
(12.13)
FX
A vector X such that DD
= 0 is said to be Fermipropagated, or Fermi
transported, along the curve. This derivative has the following properties:
476
1. when is a geodesic,
2.
DF V
D
DF X
D
DX
D
= 0 on any curve;
FX
FY
3. if both DD
= 0 and DD
= 0 hold, then g(X, Y ) is constant along the
curve, meaning in particular that orthogonal vectors remain so along
the curve;
12.1.6
In Optics
In geometrical optics, geodesics turn up as light rays, which obey just the
geodesic equation. There, actually, geodesics justify their primitive geodesical role. Not as the shortest length line between the two endpoints, but as
that line corresponding to the shortest optical length. The refraction index
is essentially a 3dimensional space metric, and the light ray follows that
line which has the shortest length as measured in that metric. This is seen
in some detail in Topic Phys.5. Light rays will follow geodesics also in the
4dimensional spacetime of General Relativity. We only notice here that,
as the interval ds2 = 0 for massless particles like the photon, fourvelocities
cannot be defined as they can [see eq.(12.6)] for massive particles following
a timelike curve. Some other curve parameter must be used.
12.2
Congruences
Our objective in the following will be to give a general, qualitative description of the relative behaviour of neighbouring curves on a manifold. We
shall concentrate on geodesics, and by neighbouring we shall mean those
geodesics which are near to each other in some limited region of the manifold.
A family of neighbouring curves is called a pencil or congruence (or still a
ray bundle, or bunch).
12.2.1
Jacobi equation
12.2. CONGRUENCES
477
(12.14)
Comment 12.2.1 The field X = 0 is a trivial solution. Notice that T (V, V ) = 0 and
R(V, V ) = 0. On a geodesic, as A = 0, both A and V are Jacobi fields. On a nongeodesic
curve, to impose that A is a Jacobi field is to say that the second acceleration vanishes:
V V V = V A =
D2 V
=0.
D 2
As
DW
= V W = [V a Da W c ]ec ,
D
eq. (12.14) is the same as
D2 X
D
+
T (X, V ) + R(X, V )V = 0 .
2
D
D
(12.15)
(12.16)
(12.17)
In components,
D2 X
D 2
or
D2 X
D 2
d
d
On a geodesic, as
D2 X
D 2
d
or
D2 X
D 2
+ V b V c Dc (X a T d ab ) + X a Rd cab = 0 ,
d
+V
D
Vb
D
D
a d
a d
c
(X T ab ) + X R cab V = 0 .
D
(12.18)
(12.19)
= 0,
+ V c Dc (X a T d ab V b ) + X a Rd cab V b V c = 0 ,
D
a d
b
a d
b c
+
(X T ab V ) + X R cab V V = 0 .
D
(12.20)
(12.21)
478
the curve, colinear with its velocity vector. The remaining directions are
transversal. The metric can be projected into a metric on that subspace:
hmn = gmn Vm Vn (see eq. (12.33) below). Thus, to each point p on the
curve will correspond a plane Pp orthogonal to the curve at p. The next step
is to examine congruences of curves, together with their variations.
Consider a congruence of curves (s) and a variation of it, (s, t). At fixed
s, these curves will cross (s) and constitute another congruence, parametrid
and U = dtd tangent to the respective
zed by t. Consider the fields V = ds
congruences of curves and normalized to unity. Each one of these fields will
be taken into itself by the others congruence. That is, their Lie derivatives
will vanish: [V, U ] = 0. From this commutativity it follows that V a a U b
= U a a V b . But then it follows also that V a Da U b = U a Da V b , or V a U b ;a =
D
U a V b ;a , or still Ds
Ub = D
V b . This may be written as
Dt
D b
U = U a Da V b = U a V b ;a ,
Ds
or
V U = U V .
(12.22)
(12.23)
12.2. CONGRUENCES
479
(12.24)
which is the simplest case of the Jacobi equation. Other forms are
D2 U
+ R(U, V )V = 0
D 2
(12.25)
and
D2 U
D2 U
a b
+
U
V
R(e
,
e
)V
=
+ U a V b Rd cab ed = 0 .
a
b
D 2
D 2
In a basis {ea =
(12.26)
},
a
D2 U d
D2 U d
d b d c
a d
+
U
R
=
+ U a Rd cab V b V c = 0 .
cab
D 2
d d
D 2
(12.27)
and s(s).
(12.28)
480
12.2.2
It is clear that the study of Jacobi fields is the main source of information
on the relative behaviour of neighbouring geodesics. They reflect all the
qualitative behaviour: whether geodesics tend to crowd or to separate, to
cross or to wind around each other. But there are some other tensor an
scalar quantities which can be of great help. In particular, when dealing with
the effect of curvature on families of curves in spacetime, a hydrodynamic
terminology is very convenient, as such curves (if timelike or null) represent
possible flow lines (of a fluid constituted by test particles), or are histories
of massless particles. We can introduce the notions of vorticity and shear,
expansion tensor and volume expansion, all of them duly projected into the
transversal space. The intuitive meaning of such tensor quantities is the
following. Suppose given around the curve a congruence of curves, whose
orthogonal cross section draw a circle on Pp . As we proceed along the curve
from point p to another point q at a distance ds, the congruence will take
the points on the circle at p into points on a line on Pq . Volume expansion
will measure the enlargement of the circle, tensor expansion will measure its
deformations, which may be larger or smaller in each direction (Figure 12.2).
At vanishing tensor and volume expansion, the circle will be taken into an
equalsized circle, as shown in Figure 12.3. The same Figure illustrates the
vorticity of the central curve , which measures how much the neighbouring
geodesics turn around it.
The procedure consists of looking at expansion, tensor and volume, and
12.2. CONGRUENCES
481
see how such tensorial quantities behave along the curve. This means to
D
calculate D
acting on them.
Given a metric g, the general invariant definitions of vorticity and expansion are:
vorticity:
(X, Y ) = g(X, Y V ) g(X V, Y )
expansion: (X,
Y) =
1
2
[g(X, Y V ) + g(X V, Y )] .
They are related to a curve of tangent field V in the following way: vorticity is its covariant curl
(X, Y ) = X a Y b V[a;b]
(12.29)
ab =
(ea , eb ) = Va;b Vb;a .
(12.30)
(12.31)
482
Figure 12.3: In the absence of shear and volume expansion, a circle around
the curve is taken into another circle of the same size. Vorticity will, however,
cause a rotation.
a , eb ). If the tangent V to
Comment 12.2.3 Consequently, Va;b = 12 (ea , eb ) + (e
ab = 0.
the curve is a Killing vector,
ab = 2 Va;b and
In spacetime, the fourvelocity has been defined in such a way that g(V, V )
= 1. This allows one to define a transversal metric by
h(X, Y ) = g(X, Y ) g(X, V ) g(Y, V ) ,
(12.32)
(12.33)
(12.34)
(12.35)
12.2. CONGRUENCES
483
At each point of a (timelike) curve, they give the vorticity and the expansion
in space.
When n = 4, a vorticity vector is defined by
a = 21 abcd Vb cd ; ab = ha c hb d V[c;d] .
The volume expansion is = V a ;a . The shear tensor is the traceless part of
the expansion tensor
(12.36)
ab = ab 13 hab
12.2.3
LandauRaychaudhury equation
We quote for completeness this equation, which relates expansion to curvature, vorticity and shear:
DV a
d
a b
2
2
2
1
= Rab V V + 2( ) 3 +
.
(12.37)
ds
Ds ;a
As 2 2 = ab ab 0 and 2 2 = ab ab 0, we see that vorticity induces
expansion, and shear induces contraction. Another important equation is
(with everything transversal)
DV[
d
ab = 2 [ ] +
.
(12.38)
ds
Ds ;]
The use of the above concepts and techniques allowed Penrose and Hawking to show that Einsteins equations lead to a singularity in the primaeval
universe. In General Relativity light rays follow null geodesics, and material objects follow timelike geodesics. In the Standard Cosmological Model
[Weinberg 72], the motion of the always receding galaxies are represented by
an expanding timelike congruence of geodesics in spacetime. If we start
from the presentday remarkably isotropic state of the universe and look
backward in time, the above equations lead, under certain physically reasonable conditions, to a starting point of any geodesics. Cosmic spacetime is
consequently not a complete space.
Synge & Schild 1978
Kobayashi & Nomizu 1963
Synge 1960
Hawking & Ellis 1973
484
Part V
PHYSICAL TOPICS
485
Phys. Topic 1
HAMILTONIAN MECHANICS
1.
2.
3.
4.
5.
6.
7.
8.
Introduction
Symplectic structure
Time evolution
Canonical transformations
Phase spaces as bundles
The algebraic structure
Relations between Lie algebras
Liouville integrability
This summary of hamiltonian mechanics has three main objectives: (i) to provide
a good exercise on differential forms, (ii) to show how forms give a far more precise
formulation of a well-known subject and (iii) to present an introductory resume of this
classical theory in what is nowadays the mathematicians colloquial language.1
1.1
Introduction
dq i
dpi
+
.
dt q i
dt pi
(1.1)
and
dpi
H
=
dt
q i
(1.2)
The emphasis here is neither that usually found in physicists texts, standardized
for instance in Goldstein 1980, nor that highly mathematical of the Bible Abraham &
Marsden 1978. It is rather in the line of Arnold 1976.
487
488
H
H
.
i
pi q
q i pi
(1.3)
H i H
dpi .
dq +
q i
pi
(1.4)
Applied to (1.3), this form gives dH(XH ) = dH/dt = 0, which says that the
value of H is conserved along the integral curve of XH .
The hamiltonian formalism provides deep insights into the details of
any mechanical system, even when the number of degrees of freedom is
continuum-infinite, as they are in Field Theory. We are here confining ourselves to particle mechanics, but those willing to see an example of that kind
can take a look at Phys.7.1.3, where the formalism is applied to gauge fields.
1.2
Symplectic structure
(1.5)
(1.6)
489
(1.8)
490
1.3
Time evolution
+ Xqi i ,
pi
q
(1.9)
(1.10)
X = Xpi
will correspond the covariant field
F H
dF
F H
F dq i F dpi
+
=
.
=
i
i
i
q pi
pi q
q dt
pi dt
dt
(1.11)
F H
F H
,
i
q pi
pi q i
(1.12)
(1.13)
the Liouville equation. The field XH flows the function along time: this
is precisely the role of a generator of infinitesimal transformations (section
6.4.2).
The operator XH is known in Statistical Mechanics as the Liouville operator, or liouvillian. Functions like F (q, p) are the classical observables, or
dynamical functions. The hamiltonian function presides over the time evolution of the physical system under consideration: we shall say that H(q, p) is
the generating function of the field XH . The time evolution of a dynamical
quantity F (q, p) is given by the solution of equation [1.13],
F (t) = F [q(t), p(t)] = etXH F (0) = F (0) + tXH F (0) +
t2
XH XH F (0) + . . .
2!
t2
{{F (0), H}, H} + . . .
2!
= F [etXH q(0), etXH p(0)] .
(1.14)
491
(1.15)
(1.16)
of the Lie derivative establishes with [1.15] the invariance of the whole series
of Poincare invariants . . . , including that one with the number
n of s, which is proportional to the volume form of M :
n = ()n dq 1 dq 2 . . . dq n dp1 dp2 . . . dpn .
(1.17)
The preservation of n by the hamiltonian flow is the general Liouville theorem. Dissipative systems will violate it.
Comment 1.3.1 Another structure is usually introduced on the manifold M : an
euclidean metric (, ) which, applied to two fields X and Y written in the manner of
[1.9], is written (X, Y ) = Xqi Yqi + Xpi Ypi . Amongst all the transformations on the 2n
dimensional space M preserving this structure (that is, amongst all the isometries), those
which are linear constitute a group, the orthogonal group O(2n).
1.4
Canonical transformations
1.4 The properties of the hamiltonian function and its related evolution
field are generalized as follows. A field X is a hamiltonian field if the form
is preserved by the local transformations generated by X. This is to say
that
LX = 0 .
(1.18)
Such transformations leaving invariant are the canonical transformations.
When n = 1, as [1.5] is the phase space area form, canonical transformations
appear as area-preserving diffeomorphisms.4 The 1form corresponding to
such a field will be closed because, being closed,
d(iX ) = LX = 0 .
4
(1.19)
Arnold 1966. Areapreserving diffeomorphisms are more general, as they may act on
non-symplectic manifolds.
492
(1.20)
for some F (q, p), then X is said to be a strictly hamiltonian field5 (or globally
hamiltonian field) and F is its generating function. In a more usual language,
F is the generating function of the corresponding canonical transformation.
In reality, most hamiltonian fields do not correspond to a generating function
at all, as iX is not exact in most cases a generating function exists only
locally. On the other hand, as a closed form is always locally exact, around
any point of M there is a neighbourhood where some F (q, p) does satisfy
iX = dF . Notice that any field XF related to some dynamical function F
by (1.20) automatically fulfills
LXF = 0 .
(1.21)
F
F
pi q i q i pi
(1.22)
Given two functions F and G, their Poisson bracket will have several different
though equivalent expressions:
F G F G
q i pi pi q i
= (XF , XG )
= XF (G) = XG (F ) = iXF iXG
= dF (XG ) = dG(XF ) .
{F, G} =
(1.23)
(1.24)
(1.25)
(1.26)
493
(1.28)
(1.29)
r being the parameter along the integral curve of XG . The formal solution
of this equation is alike to [1.14],
F [q(r), p(r)] = erXG F (0) = F (0) + rXG F (0) +
= F (0) + r{F (0), G} +
r2
XG XG F (0) + . . .
2!
r2
{{F (0), G}, G} + . . .
2!
= F [erXG q(0), erXG p(0)] . (1.30)
+ Xqi i ,
pi
q
(1.31)
494
Suppose another field is given, Y = Ypi p i + Yqi q i . The action of the 2-form
on X and Y will give
(X, Y ) = Xqi Ypi Xpi Yqi .
(1.32)
(0)
1.5
1.5 As said above, phase spaces are very particular cases of symplectic
manifolds.6 Not all symplectic manifolds are phase spaces. Think for instance
of the sphere S 2 . The area form endows S 2 with a symplectic structure
which does not correspond to a phase space. Phase spaces are the cotangent
bundles (6.4.7) of the configuration spaces. As such, they have, for each
point in configuration space, a non-compact subspace, the cotangent space,
dual and isomorphic to the tangent space. Consequently, phase spaces are
always non-compact spaces, which excludes, for instance, all the spheres.
Actually, the cotangent bundle T N of any differentiable manifold N has
a natural symplectic structure. Forms defined on T N are members of the
space T (T N ). Now, it so happens that T (T N ) has a canonical element
which is such that, given any section s : N T N , , the relation s
= s holds. Recall how we have introduced topology and charts (see for
example 6.4.3) on a tensor bundle. Take a chart (U, q) on N with coordinates
q 1 , q 2 , . . . , q n ; let be the natural projection : T N N , : Tq N q
N . Then, on 1 (U ) in T N a chart is given by (q 1 , q 2 , . . . , q n , p1 , p2 , . . . , pn ).
). These coordinates represent on T N the Pfaffian form which in the natural
basis on U is written
U = pi dq i .
(1.33)
We recognize here the [1.8]. The differentiable structure allows the extension of U to the whole T N , giving a 1form (the Liouville form of the
cotangent bundle) which reduces to U in 1 (U ), to some V in 1 (V ),
etc. Define then the 2form = - d. It will be closed and nondegenerate
6
495
and, in the chart (U, q) it will be given by expression [1.5]. Such a 2form
establishes a bijection between Tq N and Tq N at each point q N . To each
X T N corresponds iX T N .
A particularly simple case appears when the infinitesimal transformations
preserve the Liouville form itself, that is, when
LX = 0 .
(1.34)
q2
X = q1
2
q
q 1
is related to an integral of motion if X(H) = 0. The integral of motion will
then be (X) = q 1 p2 q 2 p1 . The same holds for the linear momenta p1 and
p2 , generating functions for the translation generators /q 1 and /q 2 . If
we calculate directly the Poisson brackets, we find
{q 1 p2 q 2 p1 , p1 } = p2 ; {q 1 p2 q 2 p1 , p2 } = p1 ;
{q 1 p2 q 2 p1 , q 1 } = q 2 ; {q 1 p2 q 2 p1 , q 2 } = q 1 .
We see that they reproduce the algebra of the plane euclidean group. The
field algebra, nevertheless, is
=0.
X, 1 = 2 ; X, 2 = 1 ; X,
q
q
q
q
pj
In reality, (X, /p1 ) = q 2 and (X, /p2 ) = - q 1 as it should be,
but
(X, 1 ) = (X, 2 ) = 0,
q
q
so that
(X, 1 ) 6= {q 1 p2 q 2 p1 , p1 } .
q
Comment 1.5.1 This is related to lagrangian manifolds. An n-dimensional subspace
of the 2n-dimensional phase space M is a Lagrange manifold if (X, Y ) = 0 for any
two vectors X, Y tangent to it. That is, the restriction of to is zero. Examples
are the configuration space itself, or the momentum space. The angular momentum X
above is a field on configuration space. One must be careful when Lagrange manifolds
are present, since may be degenerate. Of course, canonical transformations preserve
Lagrange manifolds, that is, they take a Lagrange manifold into another one.
496
1.6
1.6 We have been using above the holonomic base {/q i , /pj } for the
vector fields on phase space. In principle, any set of 2n linearly independent
fields may be taken as a basis. Such a general basis {ei } will have its dual,
the base { j } with j (ei ) = ij , and its members will have commutators
[ei , ej ] = ck ij ek ,
where the structure coefficients ck ij give a measure of the basis anholonomicity. A general field will be written X = X i ei = i (X)ei ; a general 1-form,
= i i = (ei ) i ; the differential of a function F will be dF = ei (F ) i ; and
so on. The symplectic 2form will be
= 21 ij i j = 12 (ei , ej ) i i .
(1.35)
(1.36)
(1.37)
(1.38)
(1.39)
This gives the Poisson bracket in terms of the inverse to the symplectic
matrix. An interesting case occurs when the Liouville form is preserved by
all the basis elements, which are consequently all strictly hamiltonian: from
Lek = 0 it follows that iek = dfk , with fk = (ek ). As = - d, then
ij = d(ei , ej ) = 12 ej (fi ) ei (fj ) + ck ij fk .
497
(1.41)
Modern approaches frequently define the hamiltonian formalism via the Poisson bracket, introduced as
{F, G} = hij ei (F )ej (G)
with hij = ij .
(1.42)
Novikov 1982.
498
(1.44)
Comment 1.6.1 A short additional comment on the lagrangian manifolds of Comment 1.5.1. Consider a simple phase space M = E2n . Given an nplane En in M , it is
called a lagrangian plane if, for any two vectors X and Y in En , (X, Y ) = 0. An alternative definition of lagrangian manifold is just as follows: any ndimensional submanifold
of M whose tangent spaces are all lagrangian planes. An interesting point is that the set
of all such planes is itself a manifold, called the lagrangian grassmannian (n) of M . This
manifold has important topological characteristics:
H1 ((n), Z) H 1 ((n), Z) 1 ((n), Z) Z .
Actually, it happens that (n) = U (n)/O(n). The group U (n) acts transitively on (n)
and makes here a rare intrusion in Classical Mechanics.
1.7
1.7 The Poisson bracket is antisymmetric and satisfies the Jacobi identity.
It is an operation defined on the space C (M, R) of real differentiable functions on M . Consequently, C (M, R) is an infinite-dimensional Lie algebra
499
(1.45)
[XF , XG ] = X{F,G}
(1.46)
(1.47)
dF[X,Y ] = dFX , FY
(1.48)
Therefore,
and
This means that
(which can be alternatively obtained by using the identity i[X,Y ] = LX iY iX LY ), from which it follows that
F[X,Y ] = {FX , FY } + (X, Y ) ,
(1.49)
500
(1.51)
501
1.8
Liouville integrability
(1.52)
(1.53)
Arnold 1976
Goldstein 1980
Abraham & Marsden 1978
Arnold, Kozlov & Neishtadt 1988
502
Phys. Topic 2
MORE MECHANICS
HamiltonJacobi
1 Hamiltonian structure
2 HamiltonJacobi equation
The Lagrange derivative
4
5
6
7
8
9
10
11
12
13
Frames
The configuration space
The phase space
Dynamics
The space and the body derivatives
The reduced phase space
Moving frames
The rotation group
Left- and right-invariant fields
The Poinsot construction
2.1
2.1.1
HamiltonJacobi
Hamiltonian structure
(2.1)
504
PHYS. TOPIC 2.
MORE MECHANICS
where the functions hij (x) are such that the bracket is antisymmetric and
satisfies the Jacobi identity. The space M is then called a Poisson manifold, and the bracket is said to endow M with a hamiltonian structure. Of
course, hij (x) is just the inverse symplectic matrix ij (x) when it is invertible (compare Eq.(2.1) with Eq.(1.39) of Phys.1). The preference given to
this definition of hamiltonian structure, based on the bracket, comes from
the study of systems for which the matrix [hij (x)] is well defined but not
invertible at every point x.
Consider two Poisson manifolds, M and N . A mapping f : M N is a
Poisson mapping if
{f F, f G}M = {F, G}N
(2.2)
where f is the pullback, f F = F f . Take two symplectic manifolds
(M1 , 1 ) and (M2 , 2 ); call 1 and 2 respectively the projections of M1 M2
into M1 and M2 , 1 : M1 M2 M1 and 2 : M1 M2 M2 . Consider also
a mapping f : M1 M2 , with graph f . Let further if : f M1 M2 be
the inclusion. Then,1
= 1 1 2 2
is a symplectic form on the product M1 M2 . If if = 0, the mapping f is
said to be a symplectic mapping.
Comment 2.1.1 We shall not in the following make any distinction between Poisson
and symplectic manifolds and/or mappings.
2.1. HAMILTONJACOBI
2.1.2
505
Hamilton-Jacobi equation
(2.3)
S
S
enforces pk = q
k and Pk = - Qk . Suppose we find a symplectic mapping f
such that S is independent of the Qj s. Then, S becomes simply a function
on the configuration space. In this case all the Pj = 0 and the hamiltonian
is a constant, H = E or, in the remaining variables,
S
k
=E .
(2.4)
H q ,
q k
X
d
i
F (q, p, t) =
XH
(x) i F (x, t) ,
dt
dx
i=1
(2.5)
(2.6)
506
PHYS. TOPIC 2.
MORE MECHANICS
If Ft (x) is a flow, the solution will be F (x, t) = fo (Ft (x)). The orbits of
the vector field XH are the characteristics of the above differential equation.
Notice that they will fix the evolution of any function. The curve solving
Hamilton equations (which are ordinary differential equations) in configuration space is the characteristic curve of the solutions of the HamiltonJacobi
equations (which are partial differential equations). More will be said on
characteristics in Phys.4.
A famous application comes out in Quantum Mechanics, where S appears
as the phase of the wavefunction: = exp[iS/~]. Then, with H(q, p) =
p2 /2m + V (q), the Schrodinger equation
~2
+ V = E
2m
(2.7)
x1
2
+
x2
2
+ ... +
xn
2
=1
(2.8)
is, of course, an example (see Phys.4 and Phys.5). The function is the
optical length, and the level surfaces of are the wave fronts. The characteristics of H(x, x ) = 0 obey differential equations which are just the
Hamilton equations
H
H
, p =
.
x =
p
x
The projections of the trajectories on the xspace are the rays.
Given a hypersurface in En , define (x) as the distance of the point x
to . Then satisfies the eikonal equation. Actually, any solution of this
equation is, locally and up to an additive constant, the distance of x to some
hypersurface.
5
Arnold 1980.
2.2
507
We shall not really examine the lagrangian formalism, which is summarized in Math.8
and Phys.6 and pervades many other chapters. The intention here is only to present the
Lagrange derivative of Classical Mechanics as an example of covariant derivative.
2.2.1
The configuration space M is the space spanned by the values of the degrees
of freedom. Its points are described by a coordinate set x = xk , one xk for
each degree. The velocity space is described accordingly by x = {x k }, where
k
x k = dx
. The lagrangian function L(x, x,
t) is defined on the combined
dt
configurationvelocity space, which is actually the tangent bundle T (M ) of
M . The extremals of L(x, x,
t) are curves (t) satisfying the EulerLagrange
equations, or equations of motion (see Math.7).
Comment 2.2.1 That the combined space is T (M ) is a simplicity assumption. No
system has been exhibited where this hypothesis has been found wanting. There is no
problem in identifying the point set, but the supposition means also that there is a projection, with the charts of the configuration space being related to those of the combined
space, etc.
Comment 2.2.2 When the configuration space is non-trivial, that is, when the degrees of freedom take values in a non-euclidean space, many local systems of coordinates
may be necessary to cover it, but the lagrangian function should be independent of the
number and choice of the charts.
Consider a time-independent change of coordinates in configuration space,
x y k = y k (xj ), which:
(i) is invertible, that is, we can find xj = xj (y k );
(ii) takes time as absolute, so that the velocities simply follow the configk
uration transformation, y k = dydt ;
(iii) leaves invariant the lagrangian function, L(x, x,
t) = L(y, y,
t).
In the lagrangian formalism, the coordinates of the configuration and
velocity spaces are independent in each chart but, once a coordinate transformation is performed, the new velocities may depend on (say) the old
coordinates. A first important thing is that, despite the highly arbitrary
character of the transformation, the velocities (tangent vectors) are linearly
k
k
transformed: y k = dydt = y
x j . This simply says that they are indeed vectors
xj
under coordinate transformations on the configuration space. Notice that
j
x n
xn
=
.
y j
y j
(2.9)
508
PHYS. TOPIC 2.
From
x k =
follows
MORE MECHANICS
xk n
y
y n
x k
2 xk
=
y n .
j
j
n
y
y y
(2.10)
(2.11)
.
y j
dt y j
y j xk dt x k
(2.13)
(2.14)
509
:=
.
(2.15)
k
k
x
x
dt x k
The reasoning remains true for higher-order lagrangians, dependent on the
acceleration, second acceleration, etc. The Lagrange derivative must be accordingly adapted, the compensating terms being then an alternate sum of
higher-order contributions:
d
d2
d3
:=
... + . . .
xk
xk dt x k dt2 xk dt3 x k
(2.16)
Comment 2.2.3 The Lagrange derivative is not invariant under a general transformation in velocity space. Notice that, in the simple L = L(x, x)
case, the non-covariant
d L
d k
compensating term dt
p
,
the
newtonian
expression
for the force. That force
is
just
k
dt
x
has, consequently, no coordinate-independent meaning.
The complete Lagrange derivative is essential to obtain consistent expressions6 for the force Qi coming from velocity dependent potentials V (q, q),
V
d V
V
= k +
.
k
q
q
dt qk
(2.17)
510
PHYS. TOPIC 2.
MORE MECHANICS
Comment 2.2.4 The degrees of freedom xk are indexed here by the discrete labels
k. When there is a continuum of degrees of freedom, each degree becomes a function
of the labels, which are then indicated (say) by x. Each degree is then a field
(x). Systems with a continuous infinity of degrees of freedom are discussed in Math.8
and Phys.6.
Comment 2.2.5 In a system with several degrees of freedom, the individual Euler
Lagrange equations are not necessarily invariant under coordinate transformations. It is
their set which is invariant. Each equation is said to be covariant, not invariant.
This reminds us of the components of vector fields and forms: components are covariant,
though vector and covector fields are invariant. In effect, the Euler-Lagrange equations
can be assembled in a certain functional differential form, which is invariant and can be
defined also for sets of equations which do not come as extremal conditions on a Lagrange
function (Math.8).
Comment 2.2.6 We have been cheating a little. Recall that t is not necessarily
time: it is actually a curve parameter, and all the above derivatives are concerned
with points on a curve. What really happens is that the Lagrange derivative is the good
derivative on a functional space, the space of functionals on the space of trajectories. A
wellknown example of such a functional is the action functional (see Math.7).
2.3
The study of rigid body motion has many points of interest: (i) it gives an example of
nontrivial configuration space, which is furthermore a Lie group whose Maurer-Cartan
form has a clear physical interpretation; (ii) it gives a simple example of a metricinduced
canonical isomorphism between a vector space and its dual; (iii) it illustrates the use of
moving frames; (iv) it shows the difference between left and rightaction of a group.
2.3.1
Frames
A rigid body is defined as a set of material points in the metric space E3 such
that the distance between any pair of points is fixed. Any three non-colinear
points on the so defined rigid body define a frame, in general nonorthogonal,
in E3 . This frame attached to the body will be indicated by F 0 . It is called
the body frame, in oposition to a fixed frame F given a priori in the vector
space E3 , called the space frame, or laboratory frame. We are taking
advantage of the double character of E3 , which is both a manifold and a
vector space. Given the position of the three points, the position of any
other point of the rigid body will be completely determined. We say that the
configuration is given. Thus, in order to specify the position of any body
point, we need only to give the positions of the three points. These would
511
2.3.2
We shall consider the case in which the body has a fixed point, and take that
point as the origin O for both frames F and F 0 . There are no translations
anymore: the configuration space of a rigid body moving around a fixed
point reduces to SO(3). We are supposing a fixed orientation for the body,
otherwise the configuration space would be O(3). As a manifold, SO(3) is a
half-sphere: SO(3) S 3 /Z2 . When we use angles as coordinates on SO(3)
(say, the Euler angles), the members of the tangent space will be angular
velocities. We may go from such starting coordinates to other generalized
coordinates, of course. It is of practical interest to identify the frame origin
O also to the group identity element. The space tangent to the configuration
space at O will then be identifiable to the Lie algebra so(3) of the group,
whose generators {Li } obey [Li , Lj ] = ijk Lk . As the structure constants are
just given by the Kronecker symbol ijk , the Lie operation coincides with the
usual vector product Li Lj = ijk Lk in E3 .
2.3.3
The phase space will be the cotangent bundle T SO(3). The members of
the cotangent space will be the dual to the angular velocities, that is, the
angular momenta. In problems involving rotational symmetry, one frequently
512
PHYS. TOPIC 2.
MORE MECHANICS
starts with phase space coordinates (q i , pk ) and find the angular momenta
M1 = q 2 p3 q 3 p2 , etc, as conserved quantities. They satisfy the Poisson
bracket algebra {Mi , Mj } = ijk Mk . This means that the Mk s provide a
representation of so(3) (the Lie algebra of the group SO(3)) in the Poisson
bracket Lie algebra of all functions on the phase space. In some cases, it
is convenient to use the Mk s themselves as generalized coordinates. The
Poisson bracket
{F, G} = hij (x) i F j G
in this case has hij (x) = ijk M k for i, j = 1, 2, 3; for i, j = 4, 5, 6 it has hij (x)
= 1 when j = i 3; and hij (x) = 0 for all the remaining cases. Furthermore,
the invariant Cartan-Killing metric of SO(3) is constant and euclidean, so
that in this metric we are tempted to write Mi = M i . Nevertheless, another
metric is present, the moment of inertia Iij . A comparison with the case of
optics is helpful here. The Cartan-Killing metric plays the same role played
by the euclidean E3 metric in optics, while the moment of inertia has some
analogy to the refractive metric (see Phys.5).
2.3.4
Dynamics
1
2
ij
I ij Mi Mj ,
(2.19)
Mi i H.
Ms
dt
2.3.5
513
The space and the body derivatives of the components of a vector quantity G are related by
dG
dt
=
space
dG
dt
+ b G .
(2.21)
body
2.3.6
514
2.3.7
PHYS. TOPIC 2.
MORE MECHANICS
Moving frames
Let us consider again the two frames F and F 0 with the same origin O. Take
a cartesian system of coordinates in each one. A point will have coordinates
x = (x1 , x2 , . . . , xn ) in F and x0 = (x01 , x02 , . . . , x0n ) in F 0 . Let us first simply
consider the motion of an arbitrary particle with respect to both frames.
The coordinates will be related by transformations x0i = Aij xj and xi =
(A1 )ij x0j .
Compare now the velocities in the two frames. To begin with, the point
k
will have velocity vF = x of components x k = xdt in the space frame F and
0k
vF0 0 = x 0 of components x 0k = xdt in the space frame F 0 . Here, of course, the
absolute character of time (t is the same in both frames) is of fundamental
importance. But there is more: we may want to consider the velocity with
respect to F as seen from F 0 , and vice-versa. Let us call vF0 the first and vF 0
the velocity with respect to F 0 as seen from F . Let us list the velocities in
the convenient notation
(seen from)
v(with respect to)
vF = x = velocity with respect to, and seen from, the space frame F ;
vF0 = velocity with respect to F as seen from F 0 ;
vF 0 = velocity with respect to F 0 as seen from F ;
vF0 0 = x 0 = velocity with respect to, and seen from, the rotating frame F 0 .
As velocities are vectors with respect to coordinates transformations, vF0 =
(2.23)
ki = (A1 )kj A ji
(2.24)
Let us call
the angular velocity tensor. Then we have
vFk = ki xi ,
(2.25)
515
which is the equation v = r when n = 3. In this case, the usual relationship of antisymmetric tensors to vectors allows one to define the vector
angular velocity j from the tensor angular velocity by ki = kij j , or
j = 12 jki (A1 )kn A ni .
(2.26)
2.3.8
Each matrix A taking a vector given in F into the same vector in F 0 represents
a rotation. It is a member of the rotation group SO(3), and the form =
A1 dA is the SO(3) canonical (or Maurer-Cartan) form. The angular velocity
tensor is the result of applying this form to the field dtd , tangent to the particle
trajectory,
d
dA
1
.
= A1
A dA
dt
dt
Thus, the angular velocity is the canonical form along the trajectory. But
the role of the canonical form is to take any vector field on the group into
a vector field at the identity. That is, into the Lie algebra of the group. If
i
A = e Ji , then
= e
so that
d
= e
dt
i Ji
i Ji
dk Jk e
j Jj
Jk e
jJ
d
dt
= e
i Ji
Jk e
jJ
dk ,
dk
dk
= (AdA1 )k j Jj
= Jk0
dt
dt
belongs to the Lie algebra. We see in this way how the angular velocities turn
up in the Lie algebra so(3) of SO(3). By the way, the above considerations
show also that = (AdA1 J)k dk = AdA1 (d).
2.3.9
We can transport a tangent vector into the group identity by two other
means: left-translation and right-translation. To each position of the body
516
PHYS. TOPIC 2.
MORE MECHANICS
d J
e = Jg
d
(2.28)
dL + L L = 0 ;
(2.29)
dR R R = 0 .
(2.30)
i
i and that the time variation of the anholonomic form is given by
d
j
= hi j i = hi j Ri .
dt
In matrix notation, g = (gij ), and g 1 = (g ij ) implies R = dgg 1 , the
ijk k
relation with vector notation being ij
= ()
R = R . We check that
d
k
ijk
1
and then that = g ir (g )rj , as it should. The formula
= d dt
517
d
= i
= k Jk =
i
dt
d
dt
= .
d
sX
X F (g) =
F (g e )
= dR F (g)(X) .
ds
s=0
R
One sees that dR F (g)(X) = dL F (g)(Adg X). In particular, for the field X =
d
tangent to a curve,
dt
"
#
R
L
L
d
d
d
d
1
1
F (g) = g
g
g F (g) +
F (g) ,
F (g) = dg
dt
dt
dt
dt
or
dL F
dR F
=
R F .
dt
dt
dL V
dR V
=
+ R V .
dt
dt
Comparison with (2.21) shows that:
(i) R dtd is the usual space velocity;
(ii) L dtd is the usual body velocity;
R
(iii) ddt is the usual space derivative;
(iv)
dL
dt
(2.31)
518
2.3.10
PHYS. TOPIC 2.
MORE MECHANICS
519
520
PHYS. TOPIC 2.
MORE MECHANICS
Phys. Topic 3
A STATISTICAL MECHANICS
Introduction
General overview
B LATTICE MODELS
The Ising model
Spontaneous breakdown of symmetry
The Potts model
Cayley tree and Bethe lattice
The four-color problem
C ELASTICITY
Regularity and defects
Classical elasticity
Nematic systems
The Franck index
3.1
3.1.1
A Statistical Mechanics
Introduction
The objective of Statistical Mechanics is to describe the behaviour of macroscopic systems, composed by a large number of elements, assuming the knowledge of the underlying dynamics of the individual constituents. In the effort
to describe real systems, usually very involved objects, it is forced to resort
to simplified models. Some models are actually reference models, supposed
to give a first approximation to a whole class of systems and playing the role
of guiding standards. They are fundamental to test calculation methods and
as starting points for more realistic improvements. For low-density gases, for
instance, the main reference models are the ideal gases, classical and quantal,
and the hard-sphere gas. For solids, lattices with oscillators in the vertices
521
522
are standard when the involved atoms or molecules have no structure. The
next step involves attributing some simple internal structure to the atoms.
It is of course very tempting to look at the lattice as a space of which the
cells are building blocks. And then consider the case of negligible spacing
between the atoms as a model for the continuous media.
As implied in section 2.2, the structure of a space is at least in part revealed by its building blocks. However, it is in general difficult to find out
which ones the necessary blocks are. We have there used irregular tetrahedra
to cover E3 . It would have been impossible to do it with regular tetrahedra.
Because historically this problem was at first studied in the 2-dimensional
case, we refer to it as the problem of the tilings (or pavings) of a space.
Thus, we say that we cannot pave E3 with regular tetrahedra, though we
can do it with irregular ones; and we cannot pave the plane E2 with regular
pentagons, though we can pave a sphere with them, and then project into E2 ,
as we do in order to endow the plane with a spherical metric (see Math.11).
This procedure is helpful in modeling some distortions in crystals, caused by
defects which, in the continuum limit, induce a curvature in the medium.
Indeed, once the lattices become very tightly packed, some continuity and
differentiability can be assumed. Vectors become vector fields, and tensors
alike. Whether or not the system is a continuum is a question of the physical scales involved. When we can only see the macroscopic features, we may
look at the limiting procedure as either an approximation (the continuum
approximation) or as a real description of the medium. Elasticity theory
treats the continuum case, but the lattice picture is too suggestive to be
discarded even in the continuum approximation.
We are thus led to examine continuum media, elastic or not. Introducing
defects into regular lattices can account for many properties of amorphous
media. The addition of defects to regular model crystals, for example, provide
good insights into the qualitative structure of glasses.
Modern theory of glasses sets up a bridge between lattice models and
Elasticity Theory. Adding defects changes the basic euclidean character of
regular lattices. It turns out that some at least of the amorphous aspects of
glasses can be seen as purely geometrical and that adding defects amounts
to attributing torsion and/or curvature to the medium.
3.1.2
General overview
523
case. But we are none the wiser after assuming microscopic hamiltonian dynamics, which supposes the knowledge of the boundary values. It is clearly
impossible to have detailed information on the boundary conditions for all
the particles in any realistic situation, such as a normal gas with around
1023 particles. It is essential to take averages on these boundary conditions,
and that is where the statistical comes forth. Different assumptions, each
one related to a different physical situation, lead to different ways of taking
the average. This is the subject of the ensemble theory of Statistical Mechanics. The microcanonical ensemble is used when all the energy values
are equally probable; the canonical ensemble describes systems plunged
in a thermal bath, for which only the average energy of the system is conserved; the grand-canonical ensemble describes systems for which only the
average number of particles is preserved; and so on. From a more mathematical point of view, Statistical Mechanics is a privileged province of Measure
Theory. Each ensemble defines a measure on phase space, making of it a
probability space (Math.3).
The volume of a 2n-dimensional phase space M is given by the Lebesgue
measure dqdp dq 1 dq 2 . . . dq n dp1 dp2 . . . dpn . The measure (the volume) of a
domain D will be
R
m(D) = hn D dqdp.
There are, however, two converging aspects leading to the choice of another
measure. First, M is a non-compact space and m(M ) is infinite. Second, in
physical situations there are preferred regions on phase space. An n-particle
gas with total hamiltonian h(q, p), for example, will have a distribution proportional to the Boltzmann factor exp[h(q, p)/kT ]. In general, the adopted
measure on phase space includes a certain non-negative distribution function
F (q, p) 0 which, among other qualities, cuts down contributions from high
qs and ps. Such a measure will give to D the value
R
m(D) = D F (q, p)dqdp.
If we want to have a probability space, we have to normalize F (q, p) so that
m(M ) = 1. The canonical ensemble, for example, adopts the measure
F (q, p) = R
eh(q,p)/kT
.
eh(q,p)/kT dqdp
M
(3.1)
524
H(x, t) =
R
M
This is quite general: given any microscopic mechanical quantity r(q, p; x, t),
its macroscopic correspondent R(x, t) will be given by the average
Z
R(x, t) =
F (q, p)r(q, p; x, t)dqdp.
(3.2)
M
(3.3)
eH/kT
.
treH/kT
(3.4)
(3.5)
525
The state of the system is now given by the density matrix, and the
partition function is the denominator in [3.4]. The canonical ensemble is
particularly convenient, as the fixed number of particles makes it easier to
define the hamiltonian H. Consider for a moment a gas with N particles
of the same species. All average values can be obtained from the partition
function. The semi-classical case is obtained as the limit of Plancks constant going to zero, but it is wise to preserve one quantum characteristic
incorporated in the Gibbs rule: the particles are indistinguishable. For the
state of the system, fixed by the expectation values of the observables, it is
irrelevant which individual particle is in this or that position in phase space.
This means that the partition function, in terms of which the average values
can be obtained, must be invariant under the action of the symmetric group
SN , which presides over the exchange of particles. The canonical partition
function QN (, V ) of a real non-relativistic gas of N particles contained in
a d-dimensional volume V at temperature T = 1/k is an invariant polynomial of SN , just the SN cycle indicator polynomial CN (Math.2). If is the
mean thermal wavelength and bj is the j-th cluster integral which takes into
account the interactions of j particles at a time,
QN (, V ) =
1
N!
CN
V
V
V
b1 d , 2b2 d , 3b3 d , . . .
(3.6)
526
3.2
3.2.1
B Lattice models
The Ising model
Figure 3.1:
The factor Jij represents the coupling between the molecules situated in the
i-th and the j-th sites. To solve such a model means to obtain the explicit
form of the partition function. A reasonably realistic case is the Heisenberg
model, for which d = q = 3. Though no analytic solution has been obtained,
many results can be arrived at by numerical methods. Some other cases have
been solved, none of them realistic enough. The problem is less difficult in
the nearest-neighbour approximation, which supposes that Jij 6= 0 only when
i and j are immediate neighbours. For 1-dimensional systems with no
external magnetic field (H = 0), exact solutions are known for all values of
the spin dimension q. For higher dimensions, the best known model is the
527
The symbol < ij > recalls that the summation takes place only on nearest
neighbours. The partition function is
X
X
X
QN (, H) =
exp{K
si sj + h
sk },
(3.9)
sk =1
<ij>
(3.10)
To illustrate the general method of solution and to introduce the important concept of transfer matrix, let us see how to arrive to this result. The
model consists of a simple line of spins 1/2, disposed in N sites. The left
segment of Figure 3.2 shows it twice, once to exhibit the site numeration and
the other to give an example of possible spin configuration. Identification of
the sites 1 and N + 1 (N +1 = 1 ) corresponds to a periodic boundary
condition. We have in this case an Ising chain (Figure 3.2, right segment),
which becomes a torus in the 2-dimensional case.
Figure 3.2:
N
X
i=1
i i+1 H
N
X
i=1
i ,
(3.11)
528
eE() =
eK
i i+1 +h
(3.12)
The partition function is the trace of the N -power of the transfer matrix. In
other cases, more than one transfer matrix can be necessary, as in the Potts
model (see Phys.3.2.3 below), in which each transfer matrix is related to a
generator of the braid group. Here, V alone suffices. It can be diagonalized
with two eingenvalues 1 and 2 . Then, [3.15] will say that
N
QN (H, T ) = N
1 + 2 .
e
= 2 + (2eK cosh h) + 2 sinh(2K) = 0.
Kh
eK
e
One finds
= eK cosh h
(3.16)
(3.17)
e2 K cosh2 h 2 sinh(2 K) ,
so that
QN (H, T ) =
N
q
2
2
K
cosh h + e cosh h 2 sinh(2 K)
+ e
cosh h
e2 K
N
cosh h 2 sinh(2 K) . (3.18)
2
529
3.2.2
The main interest in these models lies in the study of phase transitions. The
1-dimensional solution [3.10] shows no transition at all, which is an example of
a general result, known as the van Hove theorem: 1-dimensional systems with
short-range interactions between constituents exhibit no phase transition.
The 2-dimensional Onsager solution, however, shows a beautiful transition,
signaled by the behaviour of the specific heat C, whose derivative exhibits
a singularity near the critical (Curie) temperature kTc 2, 269J. The
specific heat itself behaves, near this temperature, as C constant ln(|T
Tc |1 ). A logarithmic singularity is considered to be a weak singularity. The
magnetization M , however, has a more abrupt behaviour, of the form
M
1.2224
N
1
2
Tc T
Tc
, with = 1/8.
(3.21)
530
Numerical studies show that in the 3-dimensional case the phase transition
is more accentuated. For a cubic lattice the specific heat near the critical
temperature behaves as
C |T Tc | , with 0, 125.
(3.22)
3.2.3
531
The Potts model3 may be defined on any graph (see section 2.1), that is, any
set of vertices (sites) with only (at most) one edge between each pair. This
set of sites and edges constitutes the basic lattice, which in principle models
some crystalline structure. A variable si , taking on N values, is defined on
each site labelled i. For simplicity of language, we call this variable spin.
Dynamics is introduced by supposing that only adjacent spins interact, and
that with interaction energy eij = J si sj , where is a Kronecker delta. The
total energy will be E = J (ij) si sj , the summation being on all the edges
(i, j). Then, with K = J/kT, the partition function for an M -site lattice will
be
QM = s exp[(ij) si sj ] ,
(E2i1 )
s,s0
(E2i )s,s0 =
1
N
n
Y
sj s0j ,
j6=i=1
n
Y
N si si+1
sj s0j .
j=1
(3.23)
(3.24)
532
Let us give some examples, with the notation |s > = |s1 , s2 , s3 , . . . , sn > :
< s|E2 |s0 > = s1 s2 (s1 s01 s2 s02 . . . sn s0n );
< s|E4 |s0 > = s2 s3 (s1 s01 s2 s02 . . . sn s0n );
...
0
1
N
s1 s01 N1
s2 s02 . . . sn s0n ;
s3 s03 s4 s04 . . . sn s0n ;
E2i =
N si si+1 E E E E E . . . E
N si si+1 (E n ). (3.25)
Matrix E2i is, thus, a diagonal matrix, with entries N si si+1 . The oddindexed matrices are
h
i
1
E2i1 = E E E E E . . . N . . . E E
i
h
= E (i1) 1N E (ni) , (3.26)
where
1
N
v
N
E2j , Wj =
v
N
I + E2j1 ,
(3.27)
\ = )( +
v
N
(3.28)
533
v
N
/\ =
)( +
.
(3.29)
(3.30)
There will be two global transfer matrices, which can be put into the
forms
V = exp{K(E2 + E4 + . . . + E2n2 )} = exp{K
n1
X
sj sj+1 } E n =
j=1
n1
Y
j=1
n1
Y 2j 2j+1
v
)(
+
I + E2j =
N
j=1
2j
v
N
2j+1
=
n1
Y 2j
2j1
2j
(3.31)
j=1
and
W =
n h
Y
vI +
N E2j1 = N
n/2
j=1
n
Y
2j1
v
N
2j
)( +
j=1
=N
n/2
n 2j1
Y
\
/\
(3.32)
j=1
534
V = (I +
v E2 )
N
, W = (vI +
N E1 )(vI +
N E3 ).
Figure 3.3:
theory. Of course, there will be two extra factors from the bubbles, which
must be extracted. This solution is general: for M = 2n vertices, we add 2n
projectors and then close the result, obtaining n extra bubbles which must
then be extracted. Thus, the partition function is
Q = N n/2 < K > .[4.11]
(3.33)
The general relationship is thus the following: given a lattice, draw its
medium alternate link K, which weaves itself around the vertices going
alternatively up and down the edges. Figure 3.5 shows the case m = n =2,
which corresponds simply to T closed by pairs of cups. To each edge of the
lattice will correspond a crossing, a generator of Bn (or its inverse). Vertices
will correspond to regions circumvented by loops. With the convenient choice
535
Figure 3.4:
Figure 3.5:
3.2.4
536
each vertex in the terminology of section 2.1. The first q vertices constitute
the first shell, the added q(q 1) ones form the second shell. Proceed
iteratively in this way, adding (q 1) edges to each point of the r-th shell to
obtain the (r + 1)-th shell. There are q(q 1)r1 vertices in the r-th shell.
Suppose we stop in the n-th shell. The result is a tree with
V = q[(q 1)n 1]/(q 2).
This graph is called a Cayley tree. It is used in Statistical Mechanics, each
vertex being taken as a particle endowed with spin. The partition function
will be the sum over all possible spin configurations. There is a problem,
though. The number of vertices in the n-th shell is not negligible with respect
to V , so that one of the usual assumptions of the thermodynamical limit
that border effects are negligible is jeopardized. One solution is to take
n , consider averages over large regions not including last-shell vertices,
and take them as representative of the whole system. The tree so obtained
is called the Bethe lattice and the model is the Ising model on the Bethe
lattice. Its interest is twofold: (i) it is exactly solvable and (ii) it is a first
approximation to models with more realistic lattices (square, cubic, etc).
3.2.5
The intuitive notion of a map on the plane may be given a precise definition
in the following way. A map M is a connected planar graph G and an
embedding (a drawing) of G in the plane E2 . The map divides the plane
into components, the regions or countries. G is the underlying graph of M ,
each edge corresponding to a piece of the boundary between two countries.
Actually, one same graph corresponds to different maps, it can be drawn in
different ways.
However, there is another graph related to a given map M : place a vertex
in each country of M and join vertices in such a way that to each common
border correspond an edge (as we have done in drawing the graph for the
Konigsberg bridges in 2.1.7). This graph is the dual graph of M , denoted
D(M ). When we talk of coloring a map we always suppose that no two
regions with a common border have the same color. The celebrated four-color
conjecture says that 4 colors are sufficient. That 4 colors are the minimum
necessary number is easily seen from some counter-examples. That they are
also sufficient is believed to have been demonstrated in the 70s by Appel
and Haken. The proof, involving very lengthy computer checkings and
some heuristic considerations, originated a warm debate.4 What matters
4
3.3. C ELASTICITY
537
here is that the question is a problem in graph theory and related to lattice
models. In effect, the problem is equivalent to that of coloring the vertices
of the dual graph with different colors whenever the vertices are joined by
a common edge. Or, if we like, to consider the dual graph as a lattice,
and colors as values of a spin variable, with the proviso that neighbouring
spins be different. Given a graph G, the number P (G, t) of colorings of G
using t or fewer colors may be extended to any value of the variable t. It
is then called a chromatic polynomial . It comes not as a great surprise that
such polynomials are related to partition functions of some lattice models in
Statistical Mechanics.
Pathria 1972
Baxter 1982
3.3
3.3.1
C Elasticity
Regularity and defects
538
Comment 3.3.3 Beauty is sometimes related to slightly broken symmetry. And some
masterpieces are what they are because they are slightly uncomfortable to the eye. Some
of Eschers woodprints, such as the celebrated Waterfall, are good illustrations of torsion
5
3.3. C ELASTICITY
539
Figure 3.6:
(Figure 3.8). On the other hand, we can pave a hyperboloid with heptagons,
which signals to positive curvature when we remove an edge, and negative
curvature if we add one. The crystal would be globally transformed if all
the edges were changed, but the presence of a localized defect would only
change the curvature locally. The local curvature will thus be either positive
or negative around a defect of this kind. It is a general rule that, when a
localized defect is inserted in a lattice, it deforms the region around but its
effect dies out progressively with distance. Figure 8 shows a typical case of
(two-dimensional) dislocation through the insertion of a limited extra line.
Experiments with a paper sheet can be of help here.
Well, in real media we have to do with atoms placed on the vertices. It
happens in some amorphous semi-conductors6 that, due to the presence of
6
540
Figure 3.7:
Figure 3.8:
3.3. C ELASTICITY
541
Figure 3.9:
542
Figure 3.10:
3.3.2
Classical elasticity
(3.34)
Second, we can say whether or not two vectors at distinct points are parallel.
We do it by transporting one of them along the lines defining the lattice into
the others position, in such a way that the angles it forms with the lines are
kept the same. If their direction and sense coincide when they are superposed,
they are parallel. This corresponds to parallel-transporting according to the
trivial euclidean connection (the Levi-Civita connection for the metric ab ),
whose covariant derivatives coincide with usual derivatives in a Cartesian
natural basis, and for which the lattice lines are the geodesics. The lattice
itself play the role of a geodesic grid. We can place at each vertex a set of
lattice vectors {ea }, oriented along the lines and parallel-transported all
over the lattice. Given the set at one vertex, we know it at any other vertex.
The euclidean metric will fix ab = (ea , eb ) for this initial dreibein.
The first clear visible signal of a deformation is that the measure of distance between the points changes. In the general case, the change is different
in different regions. Two neighbouring points initially separated by an eu-
3.3. C ELASTICITY
543
Figure 3.11:
(3.35)
as if the distance were given by some other, point-dependent metric gij . The
{dxi } are the same as the previous {dxa }: we are simply concentrating the
deformation in the metric. The euclidean and the new metric tensors are
related by some point-dependent transformation ha i (x),
gij = ab ha i hb j = ha i hb j (ea , eb ) = (ei , ej ).
(3.36)
(3.37)
544
for some fields ba i , which represent the departure from the trivial dreibein
related to the unstrained state. The new metric has the form gij = ij + 2uij ,
where
uij = 21 (bij + bji + ab ba i bb j )
(3.38)
is the strain tensor .
If the new basis is holonomic, ba i = i ua for some field ua (x), the deformation field . In this case the field of deformations is the variation in the
coordinates, given by x0k = xk + uk (x). This gives
dx0k = dxk + duk (x) = dxk + j uk (x)dxj ,
so that the length element changes by
dl2 dl02 = dxk dxk +j uk (x)dxk dxj +j uk (x)dxj dxk +i uk (x)j uk (x)dxi dxj .
The derivative wij = i uj is the distortion tensor . To first order in u and
its derivatives, dl02 = [jk + 2ujk ]dxj dxk , where
ujk = w(jk) = 21 [j uk + k uj ]
(3.39)
is the strain tensor for this holonomic case. Notice from gij = ij + 2uij =
ab ha i hb j that the Cartan frames (dreibeine) are
ha i = ia + wa i .
(3.40)
The fields bai of [3.37], consequently, generalize the distortion tensor to the
anholonomic case. The holonomic case comes up when they are the derivatives of some deformation field. Some authors define defects as the loci of
singular points, and dislocations as lines of singularity of the deformation
field. This only has a meaning, of course, when this field exists.
Summarizing, a deformation creates a new metric and new dreibeine. It
changes consequently also the connection. We might think at first the new
connection to be the Levi-Civita connection of the new metric, but here
comes a novelty. The connection is, in principle, a metric-independent object and can acquire proper characteristics. For example, it can develop a
non-vanishing torsion. Impurities, besides changing the distances between
the atoms, can also disrupt some of the bonds, in such a way that the original (say) hexagon is no more closed. They may become open rings in the
plane, but they may also acquire a helicoidal aspect in 3 dimensional media.
The euclidean geodesic grid collapses. These deformations are called dislocations and are of different kinds. Figures 3.12, 3.13 show what happens to
a loop in the case of the dislocation of Figures 3.9, 3.10. If we keep using the
original geodesic grid, we find a breach: the grid is destroyed and there are
3.3. C ELASTICITY
545
Figure 3.12:
B :=
du =
0
Burgers 1940.
uk i
dx =
xi
Z
0
(3.41)
546
Figure 3.13:
(3.44)
Thus, the Nye index is precisely the dual to the torsion field. The torsion is
T a ij = i ha j j ha i + a bi hb j a bj hb i ,
so that, by a change of basis, T k ij = ha k T a ij = k ji k ij . Thus,
rk = rij k [ij] ,
(3.45)
3.3. C ELASTICITY
547
(3.46)
The metric can only fix a piece of the symmetric part of ijr in the two last
indices (see Phys.8.1). To determine ijr completely we need to know the
torsion through experimental measurements of the Nye index.
3.3.3
Nematic systems
On such spin glasses and gauge field theories, see Toulouse & Vannimenus 1980.
548
Figure 3.14:
3.3. C ELASTICITY
549
Figure 3.15:
Comment 3.3.4 Systems with the same general characteristics, but without parity
invariance, are thermodynamically unstable they decay into stable systems of another
kind, cholesterics.
Consider a finite nematic system. By that we mean that the system covers a compact domain V in E3 with boundary V . The distribution of the
direction field will be fixed up when the values of n are given on V . Two
standard examples10 come to the scene when we consider an infinite cylindrical system with axis (say) along the axis Oz. In cylindrical coordinates
(, z, ), the field n(r) will not depend on z because it is infinite in that
direction, and will not depend on because there is not in the system any
parameter with the dimension of length, in terms of which we could write a
non-dimensional variable like n(r). Thus, we have only to consider a plane
transverse section of the cylinder and the only significant variable is the angle
10
Landau & Lifchitz 1990; the last editions have chapters concerning dislocations and
nematic systems, written respectively in collaboration with A.M. Kossevitch and L.P.
Pitayevski.
550
: n(r) = n(). The two cases correspond to two distinct kinds of boundary
conditions: the values of n at the boundary, n|V , are either orthogonal to
V or parallel to V . Continuity will then fix the field all over the section.
It is clear then, by symmetry, that n(r) is in both cases ill-defined on the Oz
axis, which is supposed perpendicular to the plane at the origin in Figure
3.16. This line of singularity Oz is a disclination line.
Figure 3.16:
3.3.4
Franck 1951.
3.3. C ELASTICITY
551
Figure 3.17:
552
Phys. Topic 4
PROPAGATION OF DISCONTINUITIES
1
2
3
4
4.1
Characteristics
Partial differential equations
Maxwells equations in a medium
The eikonal equation
Characteristics
Given a system of first-order partial differential equations, its solution can always be obtained from the solution of a certain system of ordinary differential
equations. The latter are called the characteristic equations of the original
system. This is also true for some important higher-order equations. Such a
conversion from partial-differential into ordinary equations can be seen as a
mere method for finding solutions. For some physical problems, however, it
is much more than that. The physical trajectory, solution of Hamilton equations (which are ordinary differential equations) in configuration space, is the
characteristic curve of the solutions of the Hamilton-Jacobi equations (which
are partial differential equations). This points to their main interest for us:
the solutions of the characteristic equations (frequently called the characteristics) have frequently a clear physical meaning: particle trajectories, light
rays, etc. The classical example is Hamiltons approach to geometrical optics
(Phys.5.1).
There are two main views on characteristics:
(i) they are lines (or surfaces, or still hypersurfaces, depending on the
dimensionality of the problem) along which disturbances or discontinuities
propagate, in the limit of short wavelength (geometric acoustics and/or geometric optics); thus, they appear as lines of propagation of the quickest
perturbations. The surface (or line, or still hypersurface) bordering the
region attained by the disturbance originated at some point, called the char553
554
d
i
F (q, p, t) =
XH
(x)
F (x, t)
i
dt
x
i=1
(4.1)
(4.2)
If Ft (x) is a flow, the solution will be F (x, t) = f0 (Ft (x)). The orbits of the
vector field XH are the characteristics of equation (4.1).
The other example is given by the eikonal equation
2
2
2
+
+ ... +
= 1,
(4.3)
x1
x2
xn
in which is the optical length and the level surfaces of are the wavefronts
(Phys.5).
4.2
(4.4)
555
f
f
2f
2f
2f
; q=
; r=
;
t
=
;
s
=
,
x
y
x2
xy
y 2
(4.5)
dp = rdx + sdy
(4.6)
dq = sdx + tdy.
(4.7)
f
f
f
and ( ) we can obtain ( x ) and ( y ), f and its first derivatives, p and q,
are known on . In order to find the solution in a neighborhood of , we
should start by (in principle at least) determining the second derivatives r, s
and t on . In order to obtain them we must solve eqs.(4.6) and (4.7). The
condition for that is that the determinant
be different from zero. There are then two directions (dy/dx) at each point
(x, y) for which there are no solutions. The two families of curves on which
= 0 are the characteristic curves. Along them one cannot find r, s and
t from the knowledge of f , p and q. Thus, in this line of attack, one must
require that be nowhere tangent to the characteristics. We shall see later
the opposite case, in which just coincides with a characteristic. Once the
non-tangency condition is fulfilled, there must be a solution in a neighborhood of . The miracle of the story is that, when looking for the higher order
derivatives in terms of the preceding ones, one finds, step by step, always the
same condition, with the same determinant. Thus, if is different from zero,
f can be obtained as a Taylor series.
The characteristic equation
Ady 2 2Bdxdy + Cdx2 = 0,
whose solutions correspond to
B
y
=
x
B 2 AC
C
(4.8)
556
determines, in principle, two families of curves on the plane (xy), which are
the characteristic curves.
There are three quite distinct cases, according to the values of the discriminant B 2 AC, and this leads to the classification of the equations and
of the corresponding differential operators appearing in (4.4):
B 2 AC < 0: the equation is of elliptic type;
B 2 AC > 0: the equation is of hyperbolic type;
B 2 AC = 0: the equation is of parabolic type.
Notice that A, B and C depend at least on x and y, so that the character
may be different in different points of the plane. Thus, the above conditions
are to be thought of in the following way: if the discriminant is negative in
all the points of a region D, then the equation is elliptic on D. And in an
analogous way for the other two cases.
Only for the hyperbolic type, for which there are two real roots 1 and
2 , is the above process actually applied. If the coefficients A, B, C are
functions of x and y only, then these curves are independent of the specific
solution of the differential equation (see the Klein-Gordon example below).
The families of curves are then given by (x, y) = c1 , which is the integral of
y 0 + 1 (x, y) = 0, and (x, y) = c2 , which is the integral of y 0 + 2 (x, y) = 0.
Suppose the differential equation has a fixed solution f = f0 (x, y). In
order to pass into geometric acoustics and/or optics, we
(i) add to it a perturbation f1 . Usually certain conditions are imposed on
such perturbations, conditions related to geometric acoustics or optics:
f1 is small, their first derivatives are small, but their second derivatives are relatively large. This means that f1 varies strongly at small
distances. It obeys the linearized equation,
A
2 f1
2 f1
2 f1
+
2B
+
C
= 0,
x2
xy
y 2
A
+ 2B
+C
= 0.
(4.9)
x
x
y
y
Finally, we
557
, =
and dx
= d
the
(iii) find the ray propagation by putting k =
x
y
dy
dk
latter being the group velocity. The eikonal equation turns into the
dispersion relation Ak 2 2Bk + C 2 = 0, and then
dx
B Ak
=
,
dy
C Bk
which is an alternative form of
dx
B
=
dy
B 2 AC
.
C
Comment 4.2.1 We have been treating scalar optics: f is a scalar. In the real case
of optics, the procedure above must be followed for each component of the electric and
magnetic fields, as well as the four-potential.
As a very simple though illustrative case, consider the Klein-Gordon equation,
1 2f
2f
+ m2 c2 f = 0.
2
2
2
x
c t
As the equation is already linear, the perturbations will obey the same equation. The characteristics are (dx/dt) = c, that is, the light cone. The
dispersion relation for the eikonal equation will be = + kc; the source
term F of [4.4] is here the term ( m2 c2 f ), but it does not influence the
characteristics. This is general: whenever the coefficients A, B and C dependent only on the independent variables x and t, the characteristics are
independent of the special solution of the starting equation.
Only a few words on geometric acoustics.2 In a medium with sound
velocity c, the differential equations of the two families of characteristic curves
C+ and C are (dx/dt) = v + c and (dx/dt) = v c. The disturbances
propagate with the sound velocity with respect to a local frame moving with
the fluid. The velocities v + c and v c are the velocities with respect to the
fixed reference frame. Along each characteristic, the fluid velocity v remains
constant. Some perturbations are simply transported with the fluid, that is,
propagate along a third characteristic C0 , given by (dx/dt) = v. In general,
a disturbation propagates along the three characteristics passing through a
certain point on the plane (x, t). It can nevertheless be decomposed into
components, each one going along one of them.
Comment 4.2.2 What we gave here is a telegraphic sketch of a large, wonderful
theory. Systems of first order partial differential equations are equivalent to systems of
Pfaffian forms, from which a systematic theory of characteristics can be more directly
formulated.3
2
3
558
4.3
Geometrical optics was traditionally regarded as an asymptotic approximation, for large wave numbers, of the wave solutions of Maxwells equations.
In two series of lectures delivered in the forties, Luneburg changed the tune.
He noticed the identity of the eikonal equation and the equation of characteristics of Maxwells equations. Think of a light signal emitted at t = 0,
which attains at an instant t the points of a surface defined by some function (x, y, z) = ct. The surface is a border, separating the region already
attained by the waves from the region not yet reached by any field. In the
inner side of the surface, the field has some nonvanishing value; at the other
side, the field is zero. The wavefront (x, y, z) represents a discontinuity of
the field, propagating at speed c, which may be point-dependent. Though a
little more involved at the start, this point of view has the great advantage
of treating light propagation no more as an approximation, but as a particular class of exact solutions of Maxwells equations, light rays appearing in
this view as lines along which discontinuities propagate. The equations, of
course, coincide with those obtained in the short wavelength treatment.
Consider an isotropic but non-homogeneous medium which is otherwise
electromagnetically inert (neither macroscopic magnetization nor electric polarization). This means that the electric and the magnetic permeabilities
depend on the positions but not on the directions. In anisotropic media
and become symmetric tensors, but we shall not consider this case here.
Then, with D = E and B = H, the sourceless Maxwell equations are
D
= 0;
t
B
= 0;
c rot E +
t
div D = 0;
div B = 0.
c rot H
(4.10a)
(4.10b)
(4.10c)
(4.10d)
The second and the fourth may be obtained respectively from the first and
the third by the duality symmetry: , E H, H E. The energy
is
W =
1
8
(E D + H B),
c
4
(E H).
559
+ div S = 0.
(i)
k f =
()f
()k f = k
E3
E3
k ()f
3
ZE
Z
(k )()f =
f (k );
E3
(ii)
Z
()i Vj =
i Vj =
E3
E3
i ()Vj
Z
Z
=
Vj (i )() = Vj (i );
E3
(iii)
(rot V )k = kij
Z
()i Vj = kij
i Vj = kij
3
ZE
i ()Vj
E3 Z
= kij
Vj (i )()f = kij
E3
(i )Vj ,
or
R
D
rot V =
(grad ) V.
= Pk
i (i )
These results are in general of local validity, the surfaces being supposed to
be piecewise differentiable.
Take the Maxwell equation div D = 0. Its integral form will be obtained
by integrating it on a domain D and using the formula (ii) above:
Z
Z
Z
Z
div D = Di (i ) = D grad = (D grad ) .
D
4
Gelfand & Shilov 1964. More details are given below, in 7.5.17.
560
Z
(c rot H t D) =
(cgrad H t D)
and
Z
Z
(cgrad E + t B)
(c rot E + t B) =
D
561
respective brackets:
[H] c grad + [D] t = 0;
[E] c grad [B] t = 0;
[D] grad = 0;
[B] grad = 0.
(4.11a)
(4.11b)
(4.11c)
(4.11d)
It might seem that these are conditions on the field discontinuities. But
usually the discontinuities are given, or supposed, so that actually these are
conditions on the surface, on the function .
The formulae contain two main possibilities. Either the fields are discontinuous, or the permeabilities are. Using D = E and B = H, the
equations become:
[H] c grad + [E] t = 0;
[E] c grad [H] t = 0;
[E] grad = 0;
[H] grad = 0.
(4.12a)
(4.12b)
(4.12c)
(4.12d)
The case in which the permeabilities are not continuous is the usual one in
optical instruments. In such instruments, the hypersurface is fixed in time,
(x1 , x2 , x3 , x4 ) = (x1 , x2 , x3 ) = 0, so that t = 0 and the conditions
become:
[H] c grad = 0;
[E] c grad = 0;
[E] grad = 0;
[H] grad = 0.
(4.13a)
(4.13b)
(4.13c)
(4.13d)
4.4
562
(t )2 }[E]
c2
= 0.
(t )2 }[H]
c2
= 0.
(t )2 = 0.
c2
(4.15)
, it becomes
(grad )2
n2
(t )2 = 0.
c2
(4.16)
Because one starts by integrating all over D, the terms without derivatives just compensate and disappear. This will always be the case with
discontinuities: only the derivative terms contribute to the conditions on the
surface. Comparing with the extra terms and looking back to those terms
really giving some contribution to the ultimate result, one sees that only
those constituting a wave equation contribute (which, by the way, answers
for its ubiquity in Physics). What happens to the large-frequency approach
is then clear: the hypotheses made in the asymptotic approach are such that
only the higher derivative terms are left, so that the results coincide.
Of course, once the results are obtained, we may consider the surface to
be, at the beginning, the source of a disturbance, taking the field to be zero
at one side. The disturbance which is propagated is then the field itself.
Notice that the equation
[H] c grad + [E] t = 0
comes from c rot H E
= 0, and that
t
[E] c grad [H] t = 0
comes from c rot E + H
= 0. Recall how we get the wave equation: we
t
take the curl of
c rot H E
=0
t
563
to get
E
= 0,
c2 rot rot H c rot
t
2 H
c2 t2
= H +
2 H
c2 t2
= 0.
We see thus the parallelism of the two procedures, the operation (grad )
playing a role dual to taking the curl.
Luneburg 1966
Abraham & Marsden 1978
Guillemin & Sternberg 1977
Choquet-Bruhat, DeWitt-Morette & Dillard-Bleick 1977
564
Phys. Topic 5
GEOMETRICAL OPTICS
1
2
3
4
5
6
7
Introduction
The light ray equation
Hamiltons point of view
Relation to geodesics
The Fermat principle
Maxwells fish-eye
Fresnels ellipsoid
5.0
Introduction
The central equation of Geometrical Optics is the light ray equation. The
usual approach to it is to start by looking for asymptotic solutions of Maxwells
equations, fall upon the eikonal equation and then examine the ray curvature. The result is an equation in euclidean 3-dimensional space, which may
be interpreted as the geodesic equation in a metric defined by the refractive
index. We shall here proceed from the eikonal equation as obtained in Phys.4,
therefore parting with this traditional, asymptotic approach. Consequently,
we look at rays as characteristics, curves along which certain electromagnetic
discontinuities, the wave fronts, propagate.
The characteristic equation for Maxwells equations in an isotropic (but
not necessarily homogeneous) medium of dielectric function (r), magnetic
2 2 2
2
2
2
2 n
+
+
2
= (grad ) 2
= 0. (5.1)
y
z
c
t
c
t
Looking for a solution in the form (x, y, z, t) = (x, y, z) ct, we fall upon
565
566
the eikonal equation, or equation for the wave fronts, under the form
5.1
2
+
2
+
2
= (grad )2 = n2 (r).
(5.2)
1
2n
grad [(grad )2 ] =
1
2n
grad [n2 ];
thus,
d
ds
dr
n
ds
= grad n.
(5.4)
(5.5)
which is the differential equation for the light rays. The Poynting vector, and
thus the energy flux, is oriented along the direction of u. The equations for
the light rays are the characteristic equations for the eikonal equations, which
are themselves the characteristic equations of Maxwells equations. For this
reason they are called the bicharacteristic equations of Maxwells equations.
The curvature of a curve at one of its points is 1/R, with R the radius of
/R, where n
is the unit vector along
the osculating circle. As a vector, it is n
the radius. When u is the unit tangent vector and s is the length parameter,
567
/R = (du/ds). The ray curvature (or curvature vector of a ray) is thus the
n
vector field K = (du/ds). Equation [5.5] is equivalent to
K=
1
(u grad n) u,
n
(5.6)
5.2
1
2n2
[j p2j ],
1 X 2 n
X = 2 k p k k + (
p)
.
(5.7)
n
x
n j j xk pk
This field is symplectically dual to the form
dH = iX = iX (dxk pk ) .
The bicharacteristic curve is an integral curve of X lying on the characteristic manifold H(p) = 12 . Notice that we are here talking about bicharacteristic or characteristic objects (fields, equations and curves) on the phase space,
of which the characteristic objects on the configuration space are projections.
As k pk x k = 2H, the lagrangian corresponding to H above is
L[] =
1 2 2
n x + y 2 + z 2 .
2
(5.8)
568
This gives the kinetic energy which is related to the Riemann metric
p
n dx2 + dy 2 + dz 2 = nds.
(5.9)
(5.10)
5.3
Relation to geodesics
Let us show that light rays are simply the geodesics of the refractive metric
[5.10]. The corresponding Christoffel symbols are
k ij = n1 ik j n + jk i n ij kr r n = (ik j) ij kr r (ln n), (5.11)
569
1
n
k j
2v v j n i (v i v i ) k n = 0.
(5.12)
Changing variables from and v to s and u, we find just the light ray
equation. In the inverse way, the equation for the light rays is
d
d
d
2 dr
n dr
=
n
= ds
(n2 v)
grad n = ds
ds
ds
d
d
d
d
d 2
n + n2 ds
v = 2vn ds
n + n2 ds
v,
= v ds
or
d
ds
Using
dv k
d
1 dv k
,
n ds
v+
2
n
d
v ds
n
1
n2
grad n = 0.
we find
dv k
d
2
n
v k dd n
1
n3
k n = 0,
the geodesic equation above. By the way, this equation becomes particularly
simple and significant when written in terms of pk : it reads
dpk
= k (ln n).
d
(5.13)
The logarithm of the refractive index acts as (minus) the potential in the
mechanical picture.
Thus, the equation for the ray curvature just states that the light ray is
a geodesic curve in the refractive metric gij = n2 ij . The procedure above is
general. If we write pi = gij vj = i , then the calculation of (dpk /d ) leads
automatically to
dv k
+ k ij v i v j = 0,
(5.14)
d
being the Levi-Civitta connection of the refractive metric. The inverse
procedure works if p is an exact 1-form, that is, a gradient of some , because
in a certain moment we are forced to use i pj = j pi . Anyhow, one always
finds
dpk
= 12 g ij k (pi pj ) = 12 k (g ij ) pi pj .
(5.15)
d
Comment 5.3.1 The condition rot (nu) = rot (n2 v) = 0, known in Optics as the
condition for the existence of the eikonal, is an obvious consequence of the Poincare
lemma, as p = d.
570
Comment 5.3.2 As seen, the relationship between optical media and metrics is deep
indeed. Mathematicians go as far as identifying the expressions optical instrument and
Riemannian manifold.1
n
,
(5.16)
xk
D
L
so that x
In this way the
k = 0 is equivalent to the geodesic equation.
equivalence is established between the mechanical and the optical points
of view, at least in what concerns the equations.
5.4
We have seen that the differential equations given by the hamiltonian field
[5.7], in the form of Hamilton equations, are equivalent to the geodesic equation for the refractive metric, or still to the Lagrange equations
R written in
hamiltonian form. The geodesics extremize the arc length nds for this
metric. Now,
R
R
nds = d
on a path is the optical length of , or its time of flight. Thus, the light
rays are those paths between two given points which extremize the optical
length. This is Fermats principle.
The surfaces = constant may be seen as surfaces of discontinuity
(Phys.4), or as wavefronts. The higher the value of |grad |, the closer
are these surfaces packed together. The eikonal equation would say that the
1
571
refractive index is a measure of the density of such surfaces. Thus, the discontinuities propagate more slowly in regions of higher index. If we interpret
grad as a vector, it will be tangent to some curve, it will be a velocity
which is larger in higher-index regions.
5.5
Maxwells fish-eye
Consider the unit sphere S 2 and its stereographic projection into the plane.
A point on S 2 will be fixed by the values X, Y, Z of its coordinates, with X 2
+ Y 2 + Z 2 = 1 (for more details, see Math.11). The relation to spherical
coordinates are X = sin cos ; Y = sin sin ; Z = cos . We choose the
north pole (0, 0, 1) as projection center and project each point of the
sphere on the plane Z = 0. The corresponding plane coordinates (x, y) will
be given by
Y
X
; y=
.
(5.17)
x=
1Z
1Z
Call
X2 + Y 2
2
2
2
r =x +y =
.
(5.18)
(1 Z)2
The line element will then be
ds2 = dX 2 + dY 2 + dZ 2 = 4
dx2 + dy 2
.
(1 + r2 )2
(5.19)
2
.
1 + r2
(5.20)
(x2 R2 1)2 + y 2 = C 2 ,
with some constants R and C. Thus, they are all the circles through the
points (0, 1). All the light rays starting at a given point will intersect
again at another point, corresponding to its antipode on S 2 . This is an
example of perfect focusing.
Comment 5.5.1 A manifold such that all points have this property is called a wiedersehen manifold. The sphere S 2 is a proven example, but it is speculated that others exist,
and also conjectured that only (higher dimensional) spheres may be wiedersehen manifolds.
But things become far more exciting in anisotropic media. Let us say a
few words on crystal optics.
572
5.6
Fresnels ellipsoid
(5.21)
The Fresnel ellipsoid is given by ij ij xi xj =constant= 2We . The 2tensor = (ij ) is non-degenerate and symmetric, the latter property being a
consequence of the requirement that the work done in building up the field,
dWe = 12 E d D,
be an exact differential form. Consequently, is a metric, E is a field and
D its covariant version according to this metric. Now, the metric can be
diagonalized, in which case the field and cofield are colinear in the 3-space
E3 , or parallel: Di = i Ei . The ellipsoid becomes
i i (xi )2 = 2We .
(5.22)
The metric eigenvalues i are the principal dielectric constants. The construction to get D from E using the ellipsoid is analogous to the Poinsot
construction for obtaining the angular momentum of a rigid body from its
angular velocity (Phys.2.3.10). It is also the same given above to obtain p
from v. Diagonalizing the metric corresponds to taking the three cartesian
axes along the three main axes of the ellipsoid.
Usual crystals are magnetically isotropic, or insensitive, so that the magnetic permeability may be taken as constant: = 0 . The magnetic induction B and the magnetic field H are simply related by B = H = 0 H.
Thus, the (squared) refraction index is given by the tensor (n2 )ij = 0 ij . By
diagonalization as above, Fresnels ellipsoid becomes
i n2i Ei2 = constant = 2We .
The ni s are the principal refraction indices. Along the principal axes of the
ellipsoid, of size (1/ni ), light will travel with the so called principal light
velocities,
c
1
ui =
=
.
(5.23)
ni
i 0
In the above procedure, D is seen as a form, a covector, while E is a vector.2
2
That is why Sommerfeld 1954 (p. 139, footnote) talks of the E components as point
coordinates and of those of D as plane coordinates.
573
Synge 1937
Sommerfeld 1954
Born & Wolf 1975
Gelfand & Shilov 1964
Luneburg 1966
574
Phys. Topic 6
6.1
6.1 Introduction
Elementary particles must have a well-defined behaviour under changes of
inertial frames in Minkowski spacetime. Such changes constitute the Poincare
group (inhomogeneous Lorentz group). In order to have a well-defined behaviour under the transformations of a group, an object must belong to a
representation, to a multiplet. If the representation is reducible, the object
is composite, in the sense that it can be decomposed into more elementary
objects belonging to irreducible representations. Thus, truly elementary particles must belong to irreducible representations and be classified in multiplets of the Poincare group. Each multiplet will have well-defined mass and
helicity.
575
576
6.2
B Spacetime transformations
(6.1)
Let us first consider the Lorentz group. If is the Lorentz metric matrix,
the matrices = ( ) satisfy T = , with T standing for the transpose
matrix. This is the defining condition for the Lorentz group. The matrices for
a general Lorentz transformation will have the form = exp[ 21 J ], where
= are the transformation parameters, and J are the generators
obeying the commutation relations
[J , J ] = J + J J J
(6.2)
577
Each representation will have as generators some matrices J . Expression [6.2] holds for anti-adjoint generators. There is no special reason to use
self-adjoint operators, as anyhow the group SO(3, 1), being non-compact, will
have no unitary finite-dimensional representations. Furthermore, SO(3, 1)
cannot accommodate half-integer spin particles. The true Lorentz group of
Nature is SL(2, C), which has also spinor representations and is the covering
group of SO(3, 1). We put, consequently, L = SL(2, C). Thus, the classifying group of elementary particles is the covering group of the Poincare
group. It is more practical to rechristen the Poincare group and write
P = SL(2, C) T . The translation group has generators T , and the remaining commutation relations are the following:
[J , T ] = T T ,
(6.3)
[T , T ] = 0.
(6.4)
Besides mass and helicity, particles (and their fields) carry other quantum
numbers (charges in general: electric charge, flavor, isotopic spin, color, etc),
related to other symmetries, not concerned with spacetime. For simplicity, we
shall call them internal symmetries. If G is their group, the total symmetry
is the direct product P G. A particle (a field) is thus characterized by being
put into multiplets of P and G. A particle with a zero charge is invariant
under the respective transformation and is accommodated in a singlet (zerodimensional) representation of the group.
There are two kinds of internal symmetries. They may be global (as that
related to the conservation of baryon number), independent of the point in
spacetime; or they may be local (those involved in gauge invariance). In the
last case, the above direct product is purely local, the fields being either in
a principal (if a gauge potential) or in an associated fiber bundle (if a source
field).
6.3 The basic cases
Relativistic fields are defined according to their behaviour under Lorentz
transformations, that is, according to the representation they belong to. Fortunately, Nature seems to use only the lowest representations: the scalar, the
vector and the spinor representations.
Scalar fields (belonging to a singlet) are those which remain unchanged:
0 (x0 ) = (x).
(6.5)
(6.6)
578
(6.7)
(6.8)
(6.9)
F = J
(6.10)
i
}](x),
4
(6.11)
(6.12)
(6.13)
6.3
579
C Internal transformations
(6.14)
(6.15)
(6.16)
(6.18)
6.4
D Lagrangian formalism
580
where represents collectively all the involved fields (Math.8). The variation of a field i may be decomposed as
i (x) + x i (x).
i (x) =
(6.20)
The second term in the right-hand side is the variation due to changes
x = x0 x in the coordinate argument. As to the first,
i
is the change in the functional form of i at a fixed value of the argument.
Notice that
= 0.
[ , ]
(6.21)
The general variation of S[] is given by:
Z
4
i (x) +
i (x) + [(x)]x . (6.22)
+ dx
i (x)
i (x)
with [d4 x] a symbolic notation to signify the variation of the integration
measure, which is ( x )d4 x in Cartesian coordinates. Collecting terms,
Z
[(x)]
[(x)]
4
i (x) +
i (x) + [(x)]x
S[] = d x
.
i (x)
i (x)
(6.23)
where the Lagrange derivative (Math.7) is simply
[(x)]
[(x)]
[(x)]
=
i (x)
i (x)
i (x)
(6.24)
581
Ja =
+ [(x)] a
i (x) a
(6.26)
is conserved. Internal symmetries are concerned with the first term, while
spacetime symmetries are concerned with the last one. Let us then examine
some examples.
(i) Translations are given by:
x0 = x + x = x +
x
a .
a
(6.27)
[]
i .
i
(6.28)
(6.29)
and consequently
=
i
2
[ ] .
(6.30)
= ( x x ) ,
and the Noether current will be the total angular momentum current density:
M
=
i
i
[]
x
. (6.32)
i
i
i
,
i
(6.33)
582
which appears only when the field is not a Lorentz singlet. From the conservation laws M = 0 and = 0, it follows that
S = .
(6.34)
1
2
m2 ,
(6.35)
(6.36)
= i m
(6.37)
1
4
{ + } .
(6.38)
0 = eiq ; = eiq .
(6.39)
i (x)
[(x)]
[(x)]
[(x)]
=
[Ta ]ji j =
Ta . (6.40)
a
i (x)
i (x)
i (x)
(6.41)
(6.42)
(6.43)
583
,
a
(6.44)
=
a
a
a
and not simply
.
a
(6.45)
As a consequence,
D = + Aa
+ Aa
.
a
(6.46)
0
G
= a ; ja = a .
A
A
(6.47)
(6.48)
G
,
Aa
(6.49)
G
= F a + f a bc Ab F c = J a
Aa
(6.50)
with equation [6.42]. This is an example of the well known result of Mechanics
(Phys.2.2), which says that simple functional derivatives are not covariant,
584
(6.53)
585
(6.54)
(x)
= (x) 4 (x y)
a (y)
(6.55)
We see that the deltas concentrate the variations at the interior point y. The
last, derivative term of [6.23] is an integration on the surface of the system,
which for the generic field is
i R
h
i
h
R 4
[(x)]
[(x)]
3
d x i (x) i (x) = S d i (x) i (x) .
The integration variable x will be at the surface of the system, whereas
y represents a point inside the system. Due to the deltas, this term will
vanish. The remaining terms will give, after integration,
S
(y)
G (y) b
G (y)
c
=
(T
)
(y)
+
f
A
(y)
+
. (6.56)
a
ij
j
ac
a (y)
i (y)
Ab (y)
Ab (y)
S
(y)
=
(Ta )ij j (y) = 0
a
(y)
i (y)
(6.57)
(6.58)
for the gauge field. Relations like [6.57] and [6.58], coming solely from the
invariance requirement and independent of the field equations, are said to be
strong relations. Consider first the latter. The last expression is
[ab + f b ac Ac ]
G (y)
= 0.
Ab (y)
(6.59)
586
(6.60)
for the covariant derivative, and write the strong relation for the gauge field
as
D D Fa = 0.
(6.61)
For the source field, [6.57] gives
(y)
(y)
(y)
(Ta )ij j (y) =
(Ta )ij j (y) = 0.
i (y)
i (y)
i (y)
(6.62)
Take for example the case of a fermion field, whose lagrangian is given by
(see Phys.7)
(6.63)
= 2i [ ] m
The source current is
Ja = i Ta Ta ,
which is also
Ja =
,
Aa
(6.64)
(6.65)
(6.66)
The strong relations are not real conservation laws, but mere manifestations
of the local invariance. As said in Phys.6.7, only the sum J + j of the
source current and the gauge field self-current has zero divergence. And
this fact only leads to a meaningful (that is, covariant) conserved charge
under the additional proviso that the local transformations become constant
transformations outside the system.
6.10 Using general frames1
We have used above a Cartesian coordinated system. An alternative2 is to
introduce general coordinate systems or, still better, general frames. Notice
to begin with that, even on a curved spacetime of metric g, the tetrads
h = (h ) satisfy, under Lorentz transformations, h0T h0 = hT h; thus, the
1
2
587
g = g x .
g = h =
(6.67)
d x h(i , i , h ) + h h
h
, (6.68)
i + (
i ) + ( )x
+h
i
i
4
(6.69)
Comment 6.4.1 We take the opportunity to comment upon the notion of covariant
derivative and the behavior of the connection. Let us examine the vector field case: to the
first order, a change in the coordinates leads to
0 (x0 ) = 0 (x + dx) = 0 (x) + dx
= (x)
i
2
(M ) = (x) dx
Thus,
(x) = 0 (x0 ) (x) = [ + ]dx =: D dx = D ,
and we have 0 (x) = (x) when D = 0, that is, when is parallel-transported. To
see how should change, it is enough to compare with
0 (x0 ) = 0 (x + dx) = (x)
i
2
(M ) .
Comment 6.4.2 The variation along a curve of tangent vector (velocity) u will be
(u) = [ + ]dx (u) = u D =
D
Ds
D
When = u itself, we have the acceleration Ds
u . The condition of no variation of the
velocity field along the curve will lead to the geodesic equation
D
Ds u
= u [ u + u ] =
d
ds u
+ u u = 0.
588
6.11 If we were to consider gravitation, h would be taken as an independent field. This is not our interest here (see Phys.8 for that). We shall here
remain on flat Minkowski space, though using tetrads, point-dependent general frames. Lorentz metric, as in general coordinates, becomes coordinate
dependent. The tetrads will here only represent coordinate transformations
on Minkowski space, through which Poincare transformations can be simulated. Thus, h will only relate to the change x and
h = h h x = h x .
The connection , on its side, is a mere transform of a cartesian = 0, so
that, in the basis {h } = {h dx }, it is
= h h .
6.12 For coordinate transformations x0 = x + x , we have h0 = h +
h , with
h h =
(x0 x )
y
x
.
y
x
y
y x
x
=
h
.
x y
x
(6.70)
We have consequently
= ,
h
the energy-momentum tensor previously obtained.
(6.71)
Phys. Topic 7
GAUGE FIELDS
0 Introduction
A THE GAUGE TENETS
1 Electromagnetism
2 Nonabelian theories
3 The gauge prescription
4 Hamiltonian approach
5 Exterior differential formulation
B FUNCTIONAL DIFFERENTIAL APPROACH
6 Functional Forms
7 The space of gauge potentials
8 Gauge conditions
9 Gauge anomalies
10 BRST symmetry
C CHIRAL FIELDS
11 Some comments on chiral fields
0 Introduction
General Relativity has been for a long time the prototype of a physical theory with a geometric flavor. The badge of this geometrical character is given
by the fact that, under the action of a gravitational field, a test particle
moves freely in a curved space, the curvature being that of the Levi-Civita
connection of the gravitational field, which is a metric. The gravitational
interaction is thereby geometrized. Gauge theories are also of geometrical
character, as the gauge potentials are connections on general principal bundles. In both cases, a significant geometric stage set is supposed as a kind
of kinematical background to which dynamics is added. Gauge theories are,
in a sense, still more geometric than General Relativity, because there is a
duality symmetry between their dynamics and the geometric background.
589
590
Comment 7.0.3 Trautman and Yang have greatly emphasized the geometric approach to gauge theories. This has been roughly presented in the main text. We shall give
here a resume of the physicists approach.
7.1
7.1.1
(7.1)
(7.2)
(7.5)
591
7.1.2
Nonabelian theories
But gauge theories2 in their modern sense were really inaugurated3 when
Yang and Mills proceeded to consider the isospin group SU (2), whose nonabelian character made a lot of difference. A fermionic source will now be a
multiplet, a field transforming in a well-defined way under the action of the
group. Utiyama4 generalized their procedure to other Lie groups and we shall
prefer to recall once for all the general case of a group G. The phase factors,
elements of the gauge group G in that representation, are now operators of
a
the form ei (x)Ta , and the source multiplets will transform according to
(x) 0 (x) = U (x)(x) = ei
a (x)T
a
(x).
(7.6)
(7.7)
The subject is treated in every modern text on Field Theory, from the classical
treatise by Bogoliubov & Shirkov 1980, to the more recent Itzykson & Zuber 1980; there
are many other books, as Faddeev & Slavnov 1978; for an excellent short introduction
covering practically all the main points, see Jackiw 1980.
3
Weyls pioneering version, in which gauge invariance appears as an indifferent scaling in ambient space, is summarized in Weyl 1932 and in Synge & Schild 1978. The original
pioneering literature is of difficult access: it includes Weyl 1919 and Weyl 1929; London
1927; Fock 1926; and O. Klein, contribution in New Theories in Physics, Conference of
the International Union of Physics, Warsaw, 1938. Excerpts can be found as addenda in
Okun 1984.
4
Utiyama 1955.
592
field. If the group generators are Ta0 in the representation of , then A (x)
must be written A (x) = Ta0 Aa (x). An examination of the sourceless case
shows that A (x) must actually be in the adjoint representation of G. By this
representation, the group acts on its own Lie algebra. If the generators of G
are Ja , they will satisfy the general commutation relations [Ja , Jb ] = f c ab Jc ,
with f c ab the structure constants, and will transform by Ja U 1 Ja U . The
vector potential will then be
A (x) = Ja Aa (x).
(7.8)
(7.9)
a matrix F = Ja F a in the adjoint representation which will, as a consequence of [7.7], transform according to
0
a
F (x) F
(x) = U (x)1 F (x)U (x) = U 1 Ja U F
.
(7.10)
1
2
m 14 F a Fa . (7.11)
The gauge field (strength) F is a 2-form,
F = 12 Ja F a dx dx
(7.12)
(7.13)
F = dA + 12 [A, A] = dA + A A.
(7.14)
as
In components, this is
F = 21 Ja Aa Aa + f a bc Ab Ac dx dx ,
(7.15)
593
(7.16)
(7.17)
(7.18)
Notice that the dual presupposes a metric, so that the former version is in
principle to be preferred.
The field equations for gauge theories are the Yang-Mills equations
F a + f a bc Ab F c = J a
(7.19)
(7.20)
We observe that, in the sourceless case, the field equations are just the
Bianchi identities written for the dual of F . This is the duality symmetry. If
we know the geometrical background, we know the dynamics. For this reason we have said that gauge theories are still more geometric than General
Relativity. While in the latter dynamics is introduced independently, the
Yang-Mills equations are related by duality to the Bianchi identities, which
are purely geometric.
Comment 7.1.1 Notice that, when G is an N -dimensional abelian group, the theory
reduces formally to N electromagnetisms.
7.1.3
(7.21)
594
(7.22)
Because the covariant derivative is different when acting on fields in different representations, it is important to pay careful attention to its form on
(y)
each object. It will be [7.21] when acting on , but (y) is a (co-)vector
object alike to A , so that D will have the rotational form given in eq.
[6.48] of Phys.6. Once this is said and retained, it is a common practice to
use always the same symbol, and write in all cases.
7.1.4
Hamiltonian approach
(7.23)
G
= E ai = F ai0 .
0 Aai
(7.24)
(7.25)
(7.26)
5
6
595
where each Ga (x) in tr[Aa 0 Ga (x)] is the expression appearing in the nonabelian Gauss law, which reads
Ga (x, t) = Dk E ak = k E ak + f a bc Ab k E ck = 0.
(7.27)
We are here, as announced, ignoring factors involving the coupling constants. We see in [7.26] that A0 is the Lagrange multiplier enforcing Gauss
law, which appears as a constraint. Another constraint is the magnetic field
expression
(7.28)
Ba i = 12 ijk F a jk = ijk j Aa k + 21 abc Ab j Ac k .
The ensuing dynamical equations are Hamiltons equations: Amp`eres
law
1 i
E = ( B)i + ijk [Aj , Bk ] + [A0 , Ei ],
c t
and the time variation of the vector potential,
1
A = E A0 + [A0 , A].
c t
(7.29)
(7.30)
(7.31)
[A],
Aa k
(7.32)
7.1.5
We have given the basic formulae in the main text. Let us only repeat a few
of them for sake of completeness. The field strength [7.14] is the curvature
of the gauge potential A, which is a connection. Taking the differential of F
leads directly to the Bianchi identity
dF + [A, F ] = 0.
(7.33)
596
(7.34)
Thus, in the sourceless case, we see clearly the invariant version of the duality
symmetry. A self-dual (or antiself-dual) 2-form in a 4-dimensional space,
solution of
F = F,
(7.35)
will respect
F = F = [F ] = F = ()(4s)/2 F = ()s/2 F.
(7.36)
7.2
Functional forms (Math.8, to which we refer for the calculations) enlarge the
geometrical meaning of gauge fields.
7.2.1
Functional Forms
(7.37)
The coefficient, whose vanishing gives the Yang-Mills equations, is the covariant coderivative of the curvature F of the connection A according to that
same connection. Each component Aa is a variable labelled by the double index (a, ), and f a bc are the gauge group structure constants. Let us examine
the condition for the existence of a lagrangian. Taking the differential,
E = F a + f a bc Ab F c + f a bc Ab F c Aa ,
(7.38)
the last term vanishes if we use the complete antisymmetry (or cyclic symmetry) of f a bc : the coefficients become symmetric under the change (a, )
597
(b, ). Integrating by parts the first term, using again the cyclic symmetry and conveniently antisymmetrizing in (, ), we arrive at the necessary
condition for the existence of a lagrangian:
E = 12 dF a dFa = 0.
(7.39)
The cyclic symmetry used above holds for semisimple groups, for which
the Cartan-Killing form is an invariant metric well defined on the group.
Actually, we have been using this metric to raise and lower indices all along.
No lagrangian exists in the nonsemisimple case.7 In the semisimple case, we
obtain
a
= 14 F a Fa .
(7.40)
G = 12 Aa D F
R
The action is just that of eq.[7.137], which is F F . We might consider
the action
Z
CG = F F.
(7.41)
It is not difficult to find that there exists a 3-form K such that F F = dK.
This means that CG is a surface term, that would not lead to local equations
by variation (though a na
R ve variationRwould lead to the Bianchi identities).
In the euclidean case, E4 F F = S 3 K. We now consider the field F
concentrated in a limited region, so that only the vacuum exists far enough,
that is, on a sphere S 3 of large enough radius. The potential will be given
by the last term of equation [7.7] (or [7.43] just below). Examination of the
detailed form of K shows that, with a convenient normalization, CG can be
put in the form
Z
1
d3 xtr ijk g 1 (x) i g(x) g 1 (x) j g(x) g 1 (x) k g(x) .
n = 242
S3
(7.42)
When the gauge group is SU (2), whose manifold is also S , the function
g(x) takes S 3 into S 3 . Once more, we can show8 that the integrand is actually
a volume form on S 3 , so that in the process of integration we are counting
the number of times the values g(x) cover SU (2) = S 3 while the variable
x covers S 3 one time. The normalization above is chosen so that just such
integer number n comes out. This is the winding number (see 3.3.13
and 6.2.15) of the function g. The values of the number n can be used to
classify the vacua, which are of the form g 1 dg. This topological number is a
generalization of the Chern number (section Math.10.4.2), introduced in the
bundle of frames, to general bundles on E4 with structure group SU (2).
3
7
8
598
7.2.2
The functional approach gives an important role to the space of the gauge
potentials, on which the state functional [A] is defined. This space is usually
called the A-space and will be denoted by . As a function, the potential A
depends (i) on some fixed starting value a, and (ii) on the group element
g by whose transformation A is obtained from a. Thus, we rewrite [7.7]
as
A(a, g) = g 1 ag + g 1 dg.
(7.43)
This decomposition corresponds to specially convenient coordinates on
A-space. The vacuum term, v = g 1 dg, corresponds to the Maurer-Cartan
form of the group: one checks easily that dv + v v = 0.
In the functional case, A itself becomes a functional of the functions g(x)
and a(x). Notice that, in this case, also g(x) is to be seen not as an element
of the gauge group G, but as a member g of the space of G-valued
Figure 7.1: Local decomposition of into components along and perpendicular to the large group.
functions (g(x) is actually a chiral field, see below). This space, an infinite
group formed by all the gauge transformations on spacetime, is called the
large group, and will be denoted by .
The space is starshaped,9 so that the use of the homotopy formula
to get the Lagrangian G of [7.40] is straightforward, and G will be valid
on the whole as far as no subsidiary gauge condition is imposed. Of
course this is not the real physical space, which requires a choice of gauge
9
Singer 1981.
599
and is far more complicated.10 Given the large group , the physical space
is formed by the gauge-inequivalent points of , the quotient space /
constituted by the gauge orbits, or the space of points a(x). This leads
to a (local!) decomposition of into components along and perpendicular
to the functional large group (Figure 7.1). Variations on may be locally
decomposed into a part along and a part orthogonal to ,
Aa = || Aa + Aa .
(7.44)
(7.45)
(7.46)
(7.47)
A = g 1 ag,
(7.48)
(7.49)
(7.50)
600
= g 1 g = g 1 ()g = g 1 (a Ja )g
= (a )g 1 Ja g = a Ka b Jb = b Jb ,
or
b = a Ka b .
(7.51)
a
|| a
A +
A
a
A
Aa
a
=
D a +
A
a
A
Aa
a
a
A . (7.52)
+
= D
Aa
Aa
= Xa () a +
Aa
Aa .
We find in this way that the group generators acting on the functionals are
Xa = D
.
Aa
(7.53)
(7.54)
601
(7.56)
(7.57)
|| a = 12 f a bc b c .
(7.58)
or
The components a , or the matrix i j = a (Ja )i j = a f i aj , are alternatively
used when convenient, the same holding for A , F , A , etc.
7.2.3
Gauge conditions
Gauge subsidiary conditions correspond to 1-Forms along . Take, for example, the one dimensional abelian case of electromagnetism, for which the
Maxwell Euler Form is
E = ( F )A
(7.59)
As A = || A + A and || A = for some parameter field , an integration by parts shows that the contribution along vanishes. The Lorenz
gauge condition is specified by the 1-Form
H = ( A ) = A A || A .
(7.60)
602
in [7.37], the contribution along the group vanishes again. The Lorenz Form
is now
H = ( Aa )a = Aa (D a [A , ]a )
= Aa || Aa = 2 || (Aa Aa ). (7.62)
Supposing a convenient normalization for the Cartan-Killing metric which
we have been using implicitly, the total Euler Form can be written
E = tr {D F A
|| (A A )}.
(7.63)
We have used, for the sector along , the holonomic (or coordinate)
basis {a }, composed of exact Forms. We could likewise have used a nonholonomic basis. With differential forms, the choice of basis is in general
dictated by the symmetry of the problem.
7.2.4
Gauge anomalies
The expressions for the gauge anomaly are components of 1-Forms along
in the anholonomic basis {a }:
U = Ua a .
(7.64)
|| U = 12 [Ta Ub Tb Ua Uc f c ab ]a b .
(7.65)
The vanishing of the expression inside the brackets is the usual Wess-Zumino
consistency condition, which in this language becomes simply
|| U = 0.
(7.66)
Again, U must be locally an exact Form, but only along , so that TU is not
a lagrangian.
Notice that, unlike the case of the action, the last equation does not
express the invariance of U under gauge transformations. Only when acting
on 0-Forms does || represent gauge transformations. As seen in Math.8, the
situation is again analogous to differential geometry, where transformations
are represented by Lie derivatives. Let us consider vector fields on , say
entities such as = a Ta or X = X a / a . Transformations on Forms will
be given by the Lie derivatives
LX = iX + iX .
603
For 0-Forms, only the last term remains, but for U the first will also contribute. The invariance of a Form W under a transformation whose generator
is represented by a Killing Field X is expressed by LX W = 0. In the case
of an Euler Form coming from a lagrangian, E = S. The commutativity between the Lie derivative and the differential operator leads to LX E = LX S,
a well known result: the invariance of S (LX S = 0) implies the invariance of
E (LX E = 0), but not vice-versa (see Math.8.3.3). The invariance of E only
implies the closeness of LX S, and the equations may have symmetries which
are not in the lagrangian.11
7.2.5
BRST symmetry
(7.67)
(7.68)
(7.69)
7.3
C Chiral fields
We make now a few comments on pure chiral fields, here understood simply
as the group-valued fields g(x) met above.
11
12
13
14
Okubo 1980.
Stora 1984; Baulieu 1984.
Stora 1984; Leinaas & Olaussen 1982.
Ma
nes 1985.
604
(i) The functional space reduces to , and will coincide with the previous
. Neither G nor are starshaped spaces, so that we must work with
tensor fields on the Lie algebra (which, being a vector space, is starshaped)
and their functional counterparts. The variation of the Maurer-Cartan form
= g 1 g is the covariant derivative of its corresponding Form:
||
= g 1 (g)g 1 g + g 1 (gg 1 g)
= + [ , ] = D . (7.70)
(ii) To obtain the Euler Form corresponding to the 2-derivative contribution to the chiral field dynamics, we start from the usual action
S = 21 tr ( ),
(7.71)
from which
E = S = tr( ) = tr { ( + [ , ])}
= tr { } = tr {( )}
= tr { (g 1 g)g 1 g}. (7.72)
(iii) The existence of a lagrangian here is a consequence of the functional
Maurer-Cartan equation. In effect,
E = ( a )a ] = a a + ( a )a
= ( a + f a bc b c ) a + a )a
= ( a )[a + 21 f a bc b c ] = 0.
The presence of in the trace argument in [7.71] would not be evident
from the field equation
(g 1 g) = 0.
(7.73)
The variation was entirely made in terms of and , which belong to
starshaped spaces, and not in terms of g(x). We can consequently follow the
inverse way: put [7.72] in the form E = tr ( ) and get [7.71] back.
(iv) The above considerations are examples of the power of exterior variational calculus, on which more is said in Math.8.
Phys. Topic 8
GENERAL RELATIVITY
1 Einsteins equation
2 The equivalence principle
3 Spinors and torsion
Little more than a topical formulary, with emphasis on some formal points.
8.1
Einsteins equation
The geometrical stage set is provided by the bundle of linear frames. This
means that the cast of characters will include linear connections
= a b a b dx ,
their curvatures
F = a b Ra b = 12 a b Ra b dx x
= 21 a b [ a b a b + a c c b a c c b ] dx dx ,
and tensors in general. We shall see, however, that the main actors will be
metrics. It is good to keep in mind the different character of the indices
in a b as in Ra b : the first two are algebraic, as they indicate algebra
components, while the remaining indices are those of tensorial components.
In the curvature form F ,
Ra b = 12 Ra b dx dx
is the component of R along a b .
Consider a manifold with a Riemannian metric. Given an arbitrary linear
connection , the covariant derivative of a metric tensor will have components
D g = g; = g g g
605
(8.1)
606
This will vanish when the connection preserves the metric, that is, when
the metric is parallel transported by :
g = ()
(8.2)
Recall that we use the notations () and [] for symmetrized and antisymmetrized indices. From this we find that
g + g g = () + [] + []
(8.3)
so that T
= 0 implies that the connection is symmetric in the lower indices.
In order to see it, recall that one changes algebra and tensor indices with the
tetrads and that, for the algebra indices in a connection,
a b = ha hb + ha hb
Replace this in the torsion component
T a = ha ha + a b hb a b hb
(8.4)
to obtain
T a = ha ha
or
T = ha T a = .
The frequently used symmetry = is thus a consequence of a vanishing torsion.
1
8.1.
EINSTEINS EQUATION
607
Comment 8.1.2 As a comparison of the bundle of frames (section 9.3) with general
bundles involving internal symmetry groups (section 9.5) shows, having a vanishing
torsion is quite distinct from having no torsion at all.
For a symmetric connection, the last two terms in [8.3] vanish and we
find that the connection is given by the Christoffel symbol
} = 21 g [ g + g g ] .
= {
(8.5)
This is the connection at work in gravitation as described by General Relativity. In Gauge Theories, the connection is the basic field. In General
Relativity, it is a tributary field: it is completely fixed by the metric, which
is thus the true fundamental field.
Comment 8.1.3 On the other hand, the two theories have much in common. The
field is in both cases the curvature. The absence of field is given by the vanishing of
the curvature. There is another point. For fields in general, one has a more physical
criterion to know where the field is: the energy-momentum density being a positive
characteristic, the field is present where the energy-momentum density is different from
zero. The gravitational field has no well-defined (that is, covariant) energy momentum and
thus this characterization fails. But also here there is something in common, because the
current density of a gauge field is not well defined (covariant) either. Energy-momentum
is the source of gravitation but there is no way of defining a (covariant) energy-momentum
for the gravitational field proper. Color (say) is the source of a gauge field, but there is
no covariant characterization of the color density of the gauge field proper.
The vanishing of torsion implies some extra symmetries in the curvature
components. Besides the antisymmetry in the second pair of indices in the
Riemann tensor
Ra b = a b a b + a b a b ,
(8.6)
(8.7)
(8.8)
Chandrasekhar 1972.
(8.9)
608
is the only symmetric 2-tensor with vanishing covariant derivative. Concerning the source fields, their symmetrized energy-momentum tensor ,
modified by the presence of g , is the only symmetric 2-tensor with vanishing covariant derivative. The source, in Newtonian gravitation, is the mass,
whose concept is broadened into energy by Special Relativity. Energy, which
in field theory is represented by the energy-momentum tensor, is to be the
source of gravitation. It is thus natural to write G = k , where k is
some constant. Comparing with Newtons law in the static weak-field limit,
one determines k = (8G/c4 ) and the field equation, Einsteins equation,
comes as
.
(8.10)
R 12 R g = 8G
c4
In the absence of sources, contraction of R 21 R g = 0 with g leads to
R = 0, and consequently to
R = 0.
(8.11)
This equation can be obtained from the Einstein-Hilbert action
Z
S[g] = d4 x g R,
(8.12)
whose variation is
Z
Z
S[g] = d x g g
R + d4 x g g R
Z
Z
4
1
d x g R 2 g R g + d4 x [total divergence]. (8.13)
Comment 8.1.4 A difference with respect to gauge theories turns up here: unlike
that of gauge fields, action [8.12] is linear in the curvature. Dynamics is in consequence
quite different.
8.2
Let us now concern ourselves with the coupling of gravitation to other fields.
Despite well known qualms,3 we shall choose a nave course, whose main
advantage is that of being short.
The manifold metric g relates to the flat Lorentz tangent metric through
the tetrad fields,
g = ab ha hb .
(8.14)
Suppose for a moment that the ha s are trivial four-legs, mere coordinate
choices. We can calculate the corresponding Christoffel and curvature. We
3
609
a
find then that Rb
= 0. This is a matter of course, as trivial tetrads will
only lead to other representations, in terms of non-cartesian coordinates,
of the flat Lorentz metric. We pass from the tangent metric to another,
Riemannian metric only through a non-trivial tetrad field. This leads to
a rule, which we shall call equivalence principle. To obtain the effect of
gravitation on sources in general (particles or fields), (i) write all the usual
equations they obey in Minkowski space in general coordinates, represented
by trivial tetrads, and (ii) keep the same formulae, but with the trivial tetrads
replaced by general tetrads, related to the metric by [8.14]. The presence of
tetrads enforces also the passage of simple derivatives to covariant derivatives
with the frames Cartan connection, so that usual derivatives are replaced
by covariant ones. The resulting equation holds in General Relativity. This
is reminiscent of the gauge prescription (Phys.7).
Comment 8.2.1 Of course, the equivalence principle takes its roots in the inverse
reasoning: when gravitation becomes progressively weaker, the equations for the source
fields approach those valid in Special Relativity (Phys.6). There is a particular system of
(normal) coordinates in which = 0.
Comment 8.2.2 Another difference with respect to gauge theories lies in the attribution of particles to the group multiplets. In gauge theories, source particles are placed
in convenient multiplets, the attribution being based on phenomenological grounds. A
particle which is insensitive to a certain gauge field is supposed to be in a singlet representation of the gauge group. The linear group (and its subgroups, Lorentz and Poincare)
acting on spacetime can always act upon fields (x) via the regular representation, which
changes the very arguments (points of spacetime) of the field (Math.6). Thus, every field
feels gravitation through this representation, a property that goes under the name of
universality . Fields endowed with spin will transform according to a direct product of
this representation and that (vector, spinor, etc) representation related to spin.
R
The total action will be [8.12] plus d4 x g S , the action of the remaining source fields, modified as indicated. The resulting Euler-Lagrange
equations will be (i) for the gravitational field, Einsteins equations, to which
the other fields provide the source in the form of their modified energymomentum density tensor, and (ii) for the source fields, simply their modified
free equations. As an example, a free scalar field, liege to the Klein-Gordon
equation,
t (x) + m2 (x) = 0,
u
(8.15)
will obey that same equation formally, but with the dAlembertian replaced
by the four-dimensional Laplace-Beltrami operator
t =
u
1
g
g g .
(8.16)
610
The laplacian is a second order operator, and here all derivatives should
be covariant. It should be
t = D D .
u
(8.17)
And indeed it is. But we should notice that the first covariant derivative,
hitting on the scalar , coincides with the simple derivative. And the second
one, hitting on the vector , will have the Cristoffel coupled with the vector
representation (Phys.6), leading to the above expression.
Non-trivial geometry affects the scalar field only through the presence of
the metric in the dAlembertian operator. We shall see below that spinor
fields probe deeper into the geometry. They are sensitive to the tetrads and,
through their spin, to a part of the connection.
The source energy-momentum current density is
=
S
.
g
(8.18)
Fock 1964.
= g R
1
2
g R
611
and the total current density vanishes. Einsteins equations [8.10] may, however, be put
into a form5 analogous to eq.[6.50] of Phys.6, that is,
+ g t = g ,
(8.19)
where is antisymmetric in and , and t represents the energy-momentum of the
gravitational field. The total current equals , so that
(8.20)
g( + t ) = 0.
The quantity t is a pseudo-tensor, not covariant under general frame transformations,
just as the self-current j a nu of the gauge fields (see Phys.6.7) is not gauge covariant. Nevertheless, t is asymptotically a tensor: at large distances, it becomes a tensor under linear
coordinate transformations which mimic isometries of Minkowski space, transformations
of the Poincare group.6 Again, this is in complete analogy with the gauge self-current,
which becomes covariant only at large distances, when the gauge transformations must
become global.
Up to this point, we have seen the standard case, which works when no
fields with spin higher than zero is present.
In the light of the geometric vision acquired by all our previous discussion,
we can indulge in some instructive reflection. As repeatedly stated, there is
no curvature of space. Curvature is a property of a connection, and a great
many connections may be defined on the same space. Take an electron on a
Riemannian spacetime. It responds to the action of the Levi-Civita connection given by the Riemannian metric. Now add an electromagnetic field. The
electron will now answer to the appeal of two connections, the previous one
and the electromagnetic potential. Add further a neutrino: it will feel (probably) the Levi-Civita connection, but not the electromagnetic potential. As
long as it stays far from the electron, there will be no manifestation of the
weak-force connection. Thus, different particles feel different connections,
different curvatures, and will consequently show distinctly curved trajectories to our euclidean eyes. If we included connections in the very definition of
space, the electron and the neutrino would live in different spaces. Now, there
is a point for taking the Levi-Civita connection of spacetime as part of its
definition, as universality of gravitation would imply that all particles would
feel it the same. There are two reasons for not doing this. First, universality
is as yet a pious postulate, based more on simplicity requirements than on
experimental evidence. Second, there is theoretical evidence that spinning
particles feel torsion, that is, that they deviate from the purely metric behaviour.7 As it is, it seems far wiser to take space simply as a manifold, and
connection (with its curvature) as an additional structure.
5
6
7
612
8.3
= 12 () + 12 [] = 21 () 12 T
(8.21)
Comment 8.3.1 There is another, frequently used, decomposition. Indicate the Levi
Civita connection (that is, the Christoffel) of a given metric by . Then, one writes
= + K ,
where K is called the contorsion tensor (the difference between two connections is
always a tensor). The two decompositions do not coincide, as the contorsion can have a
symmetric part. Actually, if also preserves the metric, one can find that
K =
1
2
(T + T T ) ,
Comment 8.3.2 Given a linear connection , then each expression of the form
(t) = t + (1 t) ,
for t [0, 1], defines a linear connection. The particular case t = (1/2) has vanishing
torsion. The torsion of is the difference between and (1/2) . As only the symmetric part
appears in the geodesic equation (Math.12), torsion does not contribute. As (t) () =
() , all these connections have the same geodesics.
For a recent discussion on deviations from the geodesic behaviour of particles, see
Yee & Bander 1993.
613
Spinors (see Phys.6) have interesting properties, which make of them ideal
detectors of torsion. Their treatment requires the explicit use of the four-legs
and their spin couples to torsion. Let us briefly examine what happens to a
Dirac spinor in a background space endowed with curvature and torsion.9
In the presence of a (external) curvature, spinors respond no more to
the usual derivative, but to the Fock-Ivanenko derivative, which takes into
account the spin coupling:
(8.22)
D = iha a 4i ab ab
where the 4 4 matrices Jab = 21 ab generate the bispinor representation.10
The lagrangian is
= 2i ha a + bc a bc ( ) a bc bc a m
(8.23)
and leads to the Dirac equation in the presence of a connection,
ha a 4i bc bc m = 0.
(8.24)
If we recall the expressions for the energy-momentum density (eq. Phys.6.30)
= 2i nu [ ] .
(8.25)
and for the spin density current
S ab = 14 [ ab + ab ] ,
(8.26)
(8.27)
Dirac 1958.
Bjorken & Drell 1964, to be consulted also on the little bit of gammalogy used
below.
10
614
Comment 8.3.5 Recall (9.4.14) that torsion does not affect geodesics (though the
symmetric part of contorsion does), but breaks infinitesimal parallelograms.
Phys. Topic 9
DE SITTER SPACES
1
2
3
4
5
General characteristics
Curvature
Geodesics and Jacobi equations
Some qualitative aspects
Wigner-In
on
u contraction
For once, though in the specially simple case of constant curvature, we
present some detailed calculations on Riemannian spaces. As de Sitter spaces
and their groups of isometries (the de Sitter groups) are related by WignerInon
u contraction to the Minkowski space and and its group of isometries
(the Poincare group), we profit to give also an example of the contraction
procedure.
9.1
General characteristics
A Riemannian space is said to be of constant curvature when its scalar curvature R = g R is a constant. Four-dimensional constant-curvature spaces
have Riemann tensor components always given locally by
R =
1
12
(g g g g ) .
(9.1)
We adopt here the terminology of Hawking & Ellis 1973, though not their sign
conventions. Notice in particular that the sign of R depends on the conventions used.
615
616
Figure 9.1: Stereographic coordinates for the 4-sphere: consider the northpole N = (0, 0, 0, 0, +L) E5 and the euclidean space E4 tangent to the
sphere at the south pole S = (0, 0, 0, 0, L). Given a point S 4 , draw a
straight line from N through . The coordinate mapping will take into the
point x E4 at which the straight line intersects E4 . The coordinates of
are then the cartesian coordinates of x.
Comment 9.1.1 The invariance of Minkowski spacetime M under the transformations of the Poincare group P reflects its uniformity. P is the group of motions of M
(6.6.14), with the maximal possible number of Killing vectors, which is ten for a 4dimensional space. Notice that the Lorentz subgroup provides an isotropy around a given
point of M , and the invariance under translations enforces this isotropy around any other
point. This is the meaning of uniformity: all the points of spacetime are ultimately
equivalent. Amongst curved spacetimes, only those of constant curvature can lodge the
highest number of Killing vectors. Given the metric signature and the value of R, the
maximally-symmetric torsionless space is unique.2 General Relativity does not consider
torsionned spaces and, in its picture, the de Sitter spaces are the only uniform curved
spacetimes.
A de Sitter space [which we call DS(4, 1) for reasons given below] may
be seen as an inclusion in E4,1 of the hypersurface whose points ( 1 , 2 , 3 ,
4 , 5 ) satisfy
( 1 )2 + ( 2 )2 + ( 3 )2 ( 4 )2 + ( 5 )2 = L2 ,
(9.2)
with the induced topology and the metric induced through the inclusion by
the pseudo-euclidean metric of E4,1 . This space is homeomorphic to S 3 E1
2
Weinberg 1972.
617
Figure 9.2: The anti-de Sitter space. Points of the upper branch correspond
to points outside the circle. Points of the lower branch, to points inside.
and its group of motions is the pseudo-orthogonal group SO(4, 1).3 An antide Sitter space DS(3, 2) may be seen as an inclusion in E3,2 , the manifold
whose points satisfy
( 1 )2 + ( 2 )2 + ( 3 )2 ( 4 )2 ( 5 )2 = 1,
(9.3)
again with the induced topology and metric. The space is now homeomorphic
to S 1 E3 and its group of motions is the pseudo-orthogonal group SO(3, 2).
Both SO(4, 1) and SO(3, 2) are called de Sitter groups and each contains the
Lorentz group SO(3, 1) as a subgroup.
There are closed timelike geodesics in the anti-de Sitter space DS(3, 2).
Actually, DS(3, 2) has a fascinating property, the wiedersehen faculty
(Phys.5.5): all timelike geodesics passing through a point will converge
unanimously to another point, and then to a third one, und so weiter. Its
3
618
universal covering is obtained by simply unwrapping the circle. This covering, which no more supports causal bizarreries, is then simply the topological
space E4 . Some authors reserve the name anti-de Sitter space to this covering space.
The metric is, in both cases, simply the Lorentz metric multiplied by a
point function. Metrics differing by the product by a function are conformally
equivalent, meaning that all angle measures are the same in both cases. A
space which is equivalent in this way to a flat space is called conformally flat.
The de Sitter spaces are conformally flat.
Consider in E5 the hypersphere S 4 given in Cartesian coordinates { a }, a =
1, 2, . . . , 5 by
( 1 )2 + ( 2 )2 + ( 3 )2 + ( 4 )2 + ( 5 )2 = L2 .
We can project it stereographically from the point 5 = +L (north
pole) into the hyperplane E4 tangent at the point 5 = L (south pole).
This will provide every point of the hypersphere (except the north pole) with
coordinates (, = 1, . . . , 4)
x =
2
1 5 /L
(9.4)
(9.5)
n = 21 (1 5 /L)
(9.6)
9.2. CURVATURE
619
(9.8)
1
,
2
1 4L
2
(9.9)
9.2
Curvature
We can easily obtain the Riemann curvature for the spherical and the hyperbolic cases. To treat all of them simultaneously, we write
n=
1
,
1 + s 2 /4L2
(9.10)
the sign s = 55 = (+1) referring to the spherical and DS(4,1) cases, the sign
s = 55 = (1) to the DS(3,2) case. The notation will be used for both
the euclidean and the Lorentz metric. Noticing that
n =
s 2
n x ,
2
2L
(9.11)
n
[ x x x ].
2L2
(9.12)
We then calculate
n
[ ]
L2
n2
4 x x ( ) + x x
4L
= s
620
and
=
n2
{x x ( ) + 2 ( ) + x x ( )}.
4
4L
Using again (9.7) and (9.9), the Riemann tensor components
R = +
are found to be
R = s
1
[ g g ].
L2
(9.13)
3
g
L2
(9.14)
and the scalar curvature is, as expected if we compare (9.13) and (9.1), the
constant
12
R = s 2.
(9.15)
L
These results are, up to the numerical factors, the same for spacetimes of
any dimension.
9.3
+
= 0.
x 2x
ds2
2nL2
ds ds
(9.16)
(9.17)
D2 X
s
+ 2 [X g(X, V )V ] = 0.
(9.18)
2
Ds
L
Now, the expression X = [X g(X, V )V ] represents the component of X
D2
||
transverse to the curve. As the tangential part X || satisfies Ds
= 0, we
2X
arrive at
D2 X
s
+ 2 X = 0.
(9.19)
2
Ds
L
621
Formally, this equation is of harmonic oscillator type, which hints at periodic solutions and to the above mentioned focusing property. The transverse
field simply oscillates around the curve. This phenomenon of perfect focusing is not quite surprising if we recall how similar this space is to a higher
dimensional replica of that favorite chimaera of optics, the perfectly focusing
Maxwells fish-eye (Phys.5.5).
9.4
9.5
Wigner-In
on
u contraction
622
(9.21)
= P 55 (4L2 )1 K
(9.22)
with
x
(9.23)
[ , J ] = ;
(9.24)
[ , ] = 55 L J .
(9.25)
lim = P
(9.26)
In the limit L ,
L
and the de Sitter algebra contracts to the usual Poincare algebra of the
generators J and P . Of course, if we look at (9.9), (9.12) and (9.13), we
see that this limit leads exactly to the Minkowsky geometry. This is the
usual Wigner-Inon
u contraction.
Let us now consider another possibility by taking the opposite limit, that
is, L 0. In this case,
(9.27)
lim = 41 55 K
L0
U
CONTRACTION
9.5. WIGNER-INON
623
and the de Sitter algebra contracts to the algebra given by eq.(9.23) and
[K , J ] = K K
(9.28)
[K , K ] = 0.
(9.29)
We see in this way that the Lie group formed by the Lorentz (J ) and the
special conformal generators (K ) has the same Lie algebra as the Poincare
group.
Summing up: by the process of Wigner-Inon
u group contraction with
L , both de Sitter groups reduce to the Poincare group, both de Sitter
spacetimes reduce to Minkowski spacetime, and the de Sitter translations
reduce to the ordinary translations in space and time. In a similar procedure, but considering the limit L 0, the de Sitter translations reduce
to the special conformal transformations, and the resulting group, despite
presenting the same algebra, is deeply different from the Poincare group. We
can conjecture in this way that, if somehow the de Sitter group appears as
the symmetry group of a physical theory, the Poincare group generated by
J and P would be related to the weak field limit (L , R 0) of
this theory, while the Poincare-like group generated by J and K would be
related to the strong field limit (L 0, R ) of the theory.
Eisenhart 1949
G
ursey 1962
Hawking & Ellis 1973
Kobayashi & Nomizu l963
624
Phys. Topic 10
SYMMETRIES ON PHASE SPACE
1
2
3
4
5
10.1
Suppose that a group G acts transitively on the phase space M (so that M is
homogeneous under G), in such a way that its transformations are canonical.
This means that G will act through a representation in a subgroup of the
huge group of canonical transformations. In this case, M is said to be a
symplectic homogeneous manifold. The generators Ja of the Lie algebra G0
of G will have commutation relations
[Ja , Jb ] = f c ab Jc .
(10.1)
(10.2)
In this case, the algebra representation is said to be linear. But this is not
what happens in general. Usually, the actions of groups on manifolds are
625
626
(10.4)
(10.5)
(10.6)
627
Always in the subspace of the Xa s, this expression says that the 2-form K
of components Kab is the (exterior) differential of the 1-form of components
c . The condition reduces to K = d, a coboundary. The cohomology class
of K is trivial for representations like [10.3]. From [10.2], [10.5] and [10.6],
the linear case is seen to require
Xa [b ] + Xb [a ] f c ab c = 0.
(10.7)
(10.8)
Fa is the generating function of the canonical transformation whose infinitesimal generator is Xa . But this means that there is still another representation at work here, that of the algebra of hamiltonian fields in the
algebra of differentiable functions on M , with the Poisson bracket as algebra
operation. With a simplified notation, it is the homomorphism : (G0 )
C (M, R), : Xa Fa , with the corresponding operations [, ] {}. From
what is seen in (Phys.1.7),
dF[Xa,Xb] (Y ) = i[Xa,Xb] (Y )
for any field Y . The last expression is the same as
([Xa , Xb ], Y ) = f c ab (Xc , Y ),
1
1984.
There is more than a mere analogy here. See for instance Faddeev & Shatashvilli
628
(10.9)
The presence of the constant (Xa , Xb ), which comes out from applying
a 2-form to the two fields Xa and Xb , says that, in principle, also is a
projective representation. This is related to the fact that generating functions
are defined up to constants. We may proceed in a way quite analogous to the
previous case. The Jacobi identity applied to [10.9] will say that the 2-form
is a cocycle. Let us add a constant to each of the above functions, and
consider the modified functions: Fa0 = Fa + a , Fb0 = Fb + b , etc. The relation
becomes
{Fa0 , Fb0 } = f c ab dFc0 + 0 (Xa , Xb ),
with 0 (Xa , Xb ) = (Xa , Xb ) c f c ab . The constants a may be seen as the
result of applying an invariant 1-form to the respective fields, a = (Xa ),
etc. As (d)ab = c f c ab , we see that 0 = + d. If some exists
whose choice leads to 0 = 0, showing as the exact form = d, then
the functions may be displaced by arbitrary constants so that the algebra
reduces to {Fa , Fb } = f c ab Fc . The projective representation reduces to a
linear representation when the cohomology class of is trivial. In this case,
we have {Fa , Fb } = F[Xa,Xb] , and the symplectic manifold M is said to be
strictly homogeneous (or Poisson) under the action of G.
10.2
When the above Fa s exist, we can also consider directly the composite mapping F : G0 C (M, R given by , such that F (Ja ) = Fa (x). We
are supposing that G is a symmetry group of the system. The transformations generated by its representative generators will preserve the hamiltonian
function H. As
{Fa (x), H} = Xa H = LXa H,
the invariance of H under the transformations whose generating function is
Fa , LXa H = 0 gives just the usual {Fa (x), H} = 0. Each Fa is a constant
of motion. This is the hamiltonian version of Noethers theorem (Phys.6.6).
Each symmetry yields a conserved quantity.
Let us place ourselves in the particular case in which the Liouville form
is also preserved: LXa = 0. Then,
629
(10.10)
(10.11)
(10.12)
10.3
(10.13)
630
10.4
Integrability revisited
Consider2 on a phase space two functions L(q, p) and M (q, p) with values in
some Lie algebra G0 of a Lie group G:
L = Ja La (q, p) ; M = Ja M a (q, p).
(10.14)
They are said to constitute a Lax pair if the evolution equation for L is
d
L = [L, M ] = Ja f a bc Lb M c .
dt
(10.15)
If now we take for the hamiltonian M = g 1 dtd g, the solution of this equation is
L(t) = g 1 (t)L(0)g(t).
(10.16)
Recognizing the action of the adjoint representation, we can write this also
as
L(t) = Adg1 (t) L(0).
(10.17)
The evolution is governed by the adjoint action. Consider now any polynomial I(L) of L which is invariant under the adjoint representation. It will
not change its form under the group action, and therefore
d
I(L) = 0.
dt
(10.18)
631
10.5
The classical Yang-Baxter equation is the Jacobi identity for the Poisson
bracket for phase spaces defined on Lie groups, written in terms of the inverse
to the symplectic matrix. Let us see how it comes out.3 The Poisson bracket
of two functions F and G is related to the symplectic cocycle by
{F, G} = (XF , XG ) = ek (G)kj ej (F ),
(10.19)
Drinfeld 1983.
(10.20)
632
where iea is the interior product. With the cocycle condition d = 0, this
implies the closedness of the form iea . It follows that there exists locally
a function Fa such that iea = dFa . The function Fa will be the generating
function of the canonical transformation generated by ea . Consequently,
{Fa , Fb } = ea (Fb ) = ab . The Jacobi identity is then
{{Fa , Fb }, Fc } + {{Fc , Fa }, Fb } + {{Fb , Fc }, Fa }
= {ab , Fc } + {ca , Fb } + {bc , Fa }
= ec (ab ) + eb (ca ) + ea (bc ) = 0. (10.21)
Using ec (ab ) = 12 [ec eb (Fa ) ec ea (Fb )] for each term of [10.21], we find
f d cb ad + f d ac bd + f d ba cd = 0.
Contracting with the product ka jb ic , we arrive finally at
f i ab ka jb + f j ab ia kb + f k ab ja ib = 0.
(10.22)
(10.23)
This is the classical Yang-Baxter equation, though not in its most usual form,
which is given in direct product notation. The contravariant tensors (rkj )
may be seen as a map
G T G T G , g r(g) = rkj ek ej .
This represents on the group manifold a general member of the direct product
G0 G0 which will have the form r = rab Ja Jb . In this notation (Math.2.10),
the algebra is included in higher product spaces by adjoining the identity
algebra. For an element of G0 , we write, for example,
X1 = X 1 = X a (Ja 1); X2 = 1 X = X a (1 Ja ),
or
X1 = X 1 1; X2 = 1 X 1; X3 = 1 1 X,
and so on. Elements of G0 G0 may then be written
r12 = rab Ja Jb 1; r13 = rab Ja 1 Jb ; r23 = rab 1 Ja Jb .
We can make use of the multiple index notation:
633
(10.24)
The name classical comes from the fact that, when conveniently parametrized,
this equation is the limit h 0 of the Yang-Baxter equation [2.22] of Math.2.
Arnold 1976
Kirillov 1974
Babelon & Viallet 1989
634
Part VI
Glossary and Bibliography
635
GLOSSARY
Abelianizer: a canonical homomorphism : G G/[G,G] taking a group
G into its abelianized subgroup.
Affine space: a subspace of a linear space V whose elements may be
written in the form a = k + v0 , with k in a linear subspace of V and v0 a
fixed point of V.
Baire space: a space which is not a countable union of nowhere dense
subsets. A complete metric space is a Baire space.
Bijective: a mapping which covers all the target space (surjective or
onto) and is one-to-one (see function). Also called a condensation.
Canonical: in general, a basis independent object, or mapping. For an
isomorphism between linear spaces, a basis-independent isomorphism. For
groups, see homomorphism.
C*-algebra: see *-algebra. An involutive Banach algebra satisfying the
further condition ||u*u|| = ||u||2 . Only in such algebras can we talk about
self-adjointness: u is self-adjoint if u = u*.
Center of a group G: the set of elements of G commuting with all the
elements of G. The center of an algebra A is the subalgebra formed by those
elements commuting with all elements of A.
Centralizer of a subset X of a group G: the set of G elements commuting with every member of X; the centralizer of G itself is the center of G;
centralizer of a subset X of an algebra A: the set of elements of A commuting
with every member of X; the centralizer of A itself is the center of algebra A.
Characteristic of a ring: for a in ring R, call (na) the expression a +
a + . . . + a, with n summands. Positive integers ni may exist for which
(ni a) = 0 for all a R. Then n = minimum {ni } is the characteristic of R.
When no such ni s exist, R is of characteristic zero. The rings Z, R and C
are of this kind. Zn is of characteristic n.
Commutant S of a subset S of elements of an algebra A: S = {a A
such that a s = sa for all s S}. The commutant of A itself is its center.
Condensation: a bijective mapping.
Congruency: for a fixed positive integer n, the number r (0 r <n)
637
638
such that h = nq + r for some q is congruent to h modulo n. Notation:
h = r(mod n). The number r such that h + k = nq + r = r(mod n) is the
sum modulo n of h and k. This can be adapted to multiplication: the
multiplication modulo n of two integers p and q is the remainder of their
usual product when divided by n, the number m such that pq = m(mod n).
Conjugate class: if a is an element of a group (G, .), its conjugate
class [a] is the set of all elements of G which can be put under the form x a
x1 for some x G. Two conjugate classes [a] and [b] are either identical or
disjoint. The element a belongs to the center of G if and only if [a] = {a}.
Degree of an element of a graded algebra: see graded algebra.
Deformation of a topological space X: a family {hs } of mappings hs :
X X, with parameter s I [0, 1] such that h0 is the identity mapping
and the function H: I X I X defined by H(s,p) = hs (p) is continuous.
When the mappings hs are homeomorphisms, the deformation is an isotopy,
or isotopic deformation. A deformation into or onto a subspace Y is a
deformation of X such that the image h1 X is contained in (or is equal to) Y.
A subspace Y of a topological space X is a deformation retract of X if there
is a retraction r: X Y which is a deformation. Deformation retractions
preserve homotopy type.
Differential graded algebra: a graded algebra on which is defined a
graded derivative , a derivation D such that D() = (D) + (-) D.
The standard example is the exterior derivative. Of special interest because
especially prone to cohomology.
Distance function: a function d taking any pair (p, q) of points of a
set X into the real line R and satisfying the following four conditions : (i)
d(p, q) 0 for all pairs (p, q); (ii) d(p, q) = 0 if and only if p = q; (iii)
d(p, q) = d(q, p) for all pairs (p, q); (iv) d(p, r) + d(r, q) d(p, q) for any
three points p, q, r. It is thus a mapping d : X X R+ . A space on
which a distance function is defined is a metric space. For vector spaces
this expression is usually reserved to translation-invariant distance functions.
A distance function is sometimes called a metric, but it is better to separate
the two concepts, though in effect a definite-positive metric tensor defines a
distance function.
Divisors of zero: two elements l and r of a ring < R, +, . > are (respectively left and right) divisors of zero if they are nonzero and such that
l.r = 0. In a commutative ring, left (right) divisors of zero are right (left)
divisors of zero. In the ring < Zn , +, . >, all numbers not relatively prime
to n are divisors of zero (and only them). So, 2 and 3 are divisors of zero in
Z6 . In particular, Zn has no divisors of zero when n is a prime number. See
integral domain.
Endomorphism: a homomorphism of a set endowed with some algebraic
639
structure (such as a group, a ring, a linear space, . . . ) into itself.
Epimorphism: a surjective homomorphism, that is, a map whose image
covers entirely the target space and which preserves the algebraic structure.
Equivalence relation: see relation.
Filter: a filter on the set S is a family F of subsets of S such that (i) the
empty set does not belong to F, (ii) the intersection of two members of F
is also a member, and (iii) any subset of S containing a member of F is also
a member. The simplest example is the set of all open neighbourhoods of
a fixed point p S. An ultrafilter is a filter which is identical to any filter
finer to it. Filters and ultrafilters are essential to the study of continuity and
convergence in non-metric topological spaces. The notion of filter is dual to
that of set ideal. See ideal.
Fr
echet space: a metrizable complete topological vector space.
Function: a mapping with a unique value in the target space for each
point of its domain. As a point set function, f: A B will be (i) surjective
(or onto) if f(A) = B; that is, the values of f for all points of A cover the whole
of B; (ii) injective (or one-to-one) if, for all a and a A, the statement
f(a) = f(a) implies a = a; that is, it takes distinct points of A into distinct
points of B; (iii) bijective (also called a condensation) if it is both onto
and one-to-one. If B A, an inclusion is an injective map i : B A with
i(p) = p if p B.
Graded algebra: a sum of vector spaces, V = k Vk , with the binary
operation taking Vi Vj Vi+j . If Vk , we say that k is the degree
(or order) of , and write = k. The standard example is the space of
differential forms of every order on a manifold M.
Graded derivative: see differential graded algebra.
Graph of a function F: M N : the set f(M) of points of N which are the
image by f of some point of M.
Homomorphism: in general, a mapping preserving algebraic structure.
A mapping f: (G, ) (H, # ) between two groups is a homomorphism
when for all a,b M G, f(a b) = f(a) # f(b). When such a mapping exists,
G and H are homomorphic. A homomorphism is canonical when it takes
a given member of G always into the same member of H (see trivializer and
abelianizer). Straightforward generalization for rings. See isomorphism.
Ideal: the sub-ring R of a ring R is a left-ideal if a b M R for all a
M R and b M R; it is a right-ideal if a b M R when a M R and b M
R; and a bilateral ideal if a b M R when either a M R or b M R. When
nothing else is said, the word ideal is used for bilateral ideals. When R is
such a bilateral ideal, there is a natural multiplication defined on the group
R/R. The resulting ring is a quotient ring. The ring Zn , +, . is the
quotient of Z by the ideal formed by all the multiples of n: Zn = Z/nZ. The
640
ring R is a maximal ideal of R if, for any other ideal R, R c R implies
R = R. In the ring of complex functions on a topological space S, those
functions vanishing at a certain point p M S form an ideal. Analogous to,
and sometimes confused with, a normal subgroup. If R is a ring with unity,
and N is an ideal of R containing an element with a multiplicative inverse,
then N = R. On a set S, an ideal is a family I of subsets of S such that (i)
the whole set S does not belong to I, (ii) the union of two members of I is
also a member, and (iii) any subset of S contained in a member of I is also a
member.
Injective mapping: a one-to-one mapping, taking distinct points into
distinct points (see function).
Integral domain: a commutative ring with unity and with no divisors
of zero. Every field is an integral domain. Every finite integral domain is a
field. Zn is a field when n is prime.
Isomorphism: in general, a one-to-one onto (or bijective) mapping preserving algebraic structure. An isomorphism between two groups (G, ) and
(H, # ) is a bijective mapping f: G H such that for all a,b M G, f(a b) =
f(a) # f(b). When such a mapping exists, G and H are isomorphic. The
generalization for rings is straightforward. An isomorphism is a bijective
homomorphism.
Isotopy: see deformation.
Kernel of a mapping f: X Y: the set ker f = {x M X such that f(x)
= 0}. For a homomorphism f: G H, the kernel (ker f) is the set of all the
elements of G which are mapped into the identity element of H.
Metric: a second order nonsingular symmetric tensor. See distance
function.
Metric space: see distance function.
Metrizable space: a space whose topology may be generated by the
balls defined by a distance function. It is always first-countable. For topological vector spaces, this metric must be translation-invariant.
Monomorphism: an injective homomorphism, that is, a one-to-one
mapping preserving algebraic structure.
Multiplication modulo n: see congruency.
One-to-one: same as injective; see function.
Onto: see surjective and function.
Operation: a binary operation o on a set S is a rule assigning to each
ordered pair of elements (a, b) of S another element a o b of S. It establishes
structure on S, and a good notation would be S, o but a structured set is
frequently denoted simply by the symbol of the set point, S. When a o b =
b o a for all a, b M S, the operation is commutative. When a o (b o c) =
(a o b) o c for all a, b, c M S, the operation is associative.
641
Power set P(S) of a given set S: the set of all subsets of S. Sometimes
indicated by the notation Exp S. If S is finite with n elements, P(S) has
2n elements. Also for S infinite, P(S) is larger then S (Cantor theorem).
Prametric: on a set S, a prametric is a mapping : S S R+ such
that (p, p) = 0 for all p S. Once endowed with a prametric, the space S is a
prametric space, and (p, q) is the prametric distance between p and q. It
is possible to have (p, q) = 0 even if p 6= q. If (p, q) 6= 0 implies p 6= q, then
the prametric is separating. If for all pairs of arguments (p, q) = (q, p), the
prametric is symmetric. A symmetric is a separating symmetric prametric.
Metrics are very special cases of symmetrics (cf. distance function). A
prametric is enough to define a topology on S.
Relation on a set S: a subset of S S. Given a relation R between
the points of a set, we write pRq to mean that p is in that relation with
q. The relation R is an equivalence relation when it is reflexive (each
point p is in that relation with itself, pRp, p), symmetric (pRq implies
qRp) and transitive (pRq and qRr together imply pRr). Equality is the
obvious example. Loosely speaking, near is an equivalence but far is
only symmetric. On the plane with coordinates (x, y), the relation has the
same coordinate x = x0 establishes an equivalence between all the points
on the same vertical line. An equivalence relation (for which the notation
instead of R is usual) divides a point set into equivalence classes, subsets
of points having that relation between each other. In the example of the
plane, each vertical line is an equivalence class and can be labelled by the
value of x0 .
Retraction of a topological space X onto a subspace Y : a continuous
mapping r : X Y such that, for any p Y , r(p) = p. When such a
retraction exists, Y is a retract of X.
Rn : the set of ordered n-uples of real numbers. Each n-uple is in general
represented as x = (x1 , x2 , ..., xn ).
Separating: a prametric m is separating if m(p, q) 6= 0 implies p 6= q. A
metric is always separating.
Sum modulo n: see congruency.
Surjective mapping: a map whose image covers entirely the target space
(see function); same as onto.
Topological invariant: a property (quality or number) of a topological
space which is shared by all spaces homeomorphic to it (invariant under
homeomorphisms).
Ultrafilter: see filter.
642
REFERENCES
Abraham R & Marsden J 1978: Foundations of Mechanics (2nd ed.), BenjaminCummings, Reading, Mass.
Adams C C 1994: The Knot Book, W.H. Freeman and Co., New York.
Agarwal G S & Wolf E 1970: Phys. Rev. D2 2161.
Aharonov Y & Anandan J 1987: Phys. Rev. Lett. 58 1593.
Aharonov Y & Anandan J 1988: Phys. Rev. D38 1863.
Aharonov Y & Bohm D 1959: Phys. Rev. 115 485.
Aharonov Y & Bohm D 1961: Phys. Rev. 123 1511.
Aldrovandi R 1992: Fort. Physik 40 631.
Aldrovandi R & Galetti D 1990: J. Math. Phys. 31 2987.
Aldrovandi R & Kraenkel R A 1988: J. Phys. A21 1329.
Aldrovandi R & Pereira J G 1986: Phys. Rev. D33 2788.
Aldrovandi R & Pereira J G 1988: Rep.Math.Phys. 26 237.
Aldrovandi R & Pereira J G 1991: in MacDowell S, Nussenzveig H M
& Salmeron R A (eds.), Frontier Physics: Essays in Honour of J. Tiomno,
World Scientific, Singapore.
Alexandrov P 1977: Introduction `
a la Theorie Homologique de la Dimension et la Topologie Combinatoire, MIR, Moscow.
Anderson I M and Duchamp T 1980: Am. J. Math. 102 781.
Arafune J , Freund P G O & Gobel C J 1975: J. Math. Phys. 16 433.
Arkhangelskii A V & Fedorchuk V V 1990: in Arkhangelskii A V &
Pontryagin L S (eds.), General Topology I, Encyclopaedia of Mathematical
Sciences, vol.17, Springer, Berlin.
Arnold V I 1966: Ann. Inst. Fourier 16 319.
Arnold V I 1973: Ordinary Differential Equations, MIT Press, Cambridge.
Arnold V I 1976: Les Methodes Mathematiques de la Mechanique Classique, MIR, Moscow.
644
Arnold V I, Kozlov V V & Neishtadt A I 1988: in Arnold V I (ed.),
Dynamical Systems III, Encyclopaedia of Mathematical Sciences, Springer,
Berlin.
Assis A K T 1989: Found. Phys. Lett. 2 301.
Askar A 1985: Lattice Dynamical Foundations of Continuum Theories,
World Scientific, Singapore.
Atherton R W and Homsy G M 1975: Stud. Appl. Math. 54 31.
Atiyah M F, Drinfeld V G, Hitchin N J & Manin Yu. I 1978: Phys. Lett.
A65 185.
Atiyah M F 1979: Geometry of Yang-Mills Fields, Acad. Naz. Lincei,
Pisa.
Atiyah M F 1991: The Geometry and Physics of Knots (Lezioni Lincee,
Accademia Nazionale dei Lincei, Pisa), Cambridge University Press, Cambridge.
Avis S J & Isham C J 1978: Proc. Roy. Soc. 362 581.
Avis S J & Isham C J 1979: in Levy M & Deser S, Recent Developments
in Gravitation, Proceedings of the 1978 Carg`ese School, Gordon and Breach,
New York.
Babelon O & Viallet C -M 1989: Integrable Models, Yang-Baxter Equation, and Quantum Groups, SISSA-ISAS preprint 54 EP.
Baker G A 1958: Phys. Rev. 109 2198.
Balescu R 1975: Equilibrium and Nonequilibrium Statistical Mechanics,
J. Wiley, New York.
Baulieu L 1984: Nucl. Phys. B241 557.
Baxter R J 1982: Exactly Solved Models in Statistical Mechanics,
Academic Press, London.
Bayen F, Flato M, Fronsdal C, Lichnerowicz A & Sternheimer D 1978:
Ann . Phys.
111 61 and 111.
Becher P & Joos H 1982: Z. Phys. C15 343.
Berry M V 1976: Advances in Physics 25 1.
Berry M V 1984: Proc. Roy. Soc. A392 45.
Biggs N L, Lloyd E K & Wilson R J 1977: Graph Theory 1736-1936,
Oxford University Press, Oxford.
Birman J S 1975: Braids, Links, and Mapping Class Groups, Princeton
University Press, Princeton.
Birman J S 1991: Math. Intell. 13 52.
Bjorken J D & Drell S D 1964: Relativistic Quantum Mechanics, McGrawHill,
New York.
645
Bjorken J D & Drell S D 1965: Relativistic Quantum Fields, McGraw-Hill,
New York.
Bogoliubov N N & Shirkov D V 1980: Introduction to the Theory of
Quantized Fields
(3rd ed.), J. Wiley, New York.
Bonora L & Cotta-Ramusino P 1983: Comm. Math. Phys. 87 589.
Boothby W M l975 : An Introduction to Differentiable Manifolds and
Riemannian Geometry, Academic Press, New York.
Born M & Wolf E 1975: Principles of Optics, Pergamon, Oxford.
Born M 1964: Natural Philosophy of Cause and Chance, Dover, New
York.
Born M & Infeld L 1934: Proc. Roy. Soc. A144 425.
Boya L J, Cari
nena J F & Mateos J 1978: Fort. Physik 26 175.
Bratelli O & Robinson D W 1979: Operator Algebras and Quantum Statistical Mechanics I, Springer, New York.
Broadbent S R & Hammersley J M 1957: Proc. Camb. Phil. Soc. 53
629.
Budak B M & Fomin S V 1973: Multiple Integrals, Field Theory and
Series,
MIR, Moscow.
Burgers J M 1940: Proc. Phys. Soc. (London) 52 23.
Burke W L l985: Applied Differential Geometry, Cambridge University
Press, Cambridge.
Chandrasekhar S 1939: An Introduction to the Study of Stellar Structure,
U. Chicago Press, Chicago.
Chandrasekhar S 1972: Am. J. Phys. 40 224.
Chillingworth D 1974: in Global Analysis and its Applications, IAEA,
Vienna, vol.I.
Chinn W G & Steenrod N E 1966: First Concepts of Topology, Random
House,
New York
Cho Y M l975: J. Math. Phys. 16 2029.
Choquet-Bruhat Y, DeWitt-Morette C & Dillard-Bleick M 1977: Analysis, Manifolds and Physics, North-Holland, Amsterdam.
Christie D E 1976: Basic Topology, MacMillan, New York.
Coleman S 1977 : Classical Lumps and their Quantum Descendants, in
A. Zichichi (ed.), New Phenomena in Subnuclear Physics, Proceedings of
the 1975 International School of Subnuclear Physics (Erice, Sicily), Plenum
Press, New York.
Coleman S 1979 : The Uses of Instantons, in A. Zichichi (ed.), The
Whys of Subnuclear Physics , Proceedings of the 1977 International School
646
of Subnuclear Physics (Erice, Sicily), Plenum Press, New York
Connes A 1980: C. R. Acad. Sci. Paris 290A 599.
Connes A 1986: Noncommutative Differential Geometry, Pub. IHES
(Paris) 62 257.
Connes A 1990: Geometrie Non-Commutative, InterEditions, Paris.
Comtet L 1974: Advanced Combinatorics, Reidel, Dordrecht.
Coquereaux R 1989: J. Geom. Phys. 6 425.
Croom F H 1978: Basic Concepts of Algebraic Topology, Springer, Berlin.
Crowell R H & Fox R H 1963: Introduction to Knot Theory, Springer,
Berlin.
Daniel M & Viallet C M l980: Rev. Mod. Phys. 52 175.
Davis W R & Katzins G H 1962: Am. J. Phys. 30 750.
de Gennes P G & Prost J 1993: The Physics of Liquid Crystals (2nd ed.),
Clarendon Press, Oxford.
de Rham G 1960: Varietes Differentiables, Hermann, Paris.
DeWitt-Morette C 1969: Ann. Inst. Henri Poincare XI 153.
DeWitt-Morette C 1972: Comm. Math. Phys. 28 47.
DeWitt-Morette C, Masheshwari A & Nelson B 1979: Phys. Rep. 50
255.
Dirac P A M 1926: Proc. Roy. Soc. A109 642.
Dirac P A M 1958: in W. Frank (ed.), Planck Festschrift, VEB Deutscher
Verlag der Wissenschaften, Berlin, p. 339.
Dirac P A M 1984: The Principles of Quantum Mechanics, Clarendon
Press, Oxford.
Dittrich J 1979: Czech. J. Phys. B29 1342.
Dixmier J 1981: von Neumann Algebras, North-Holland, Amsterdam.
Dixmier J 1982: C*-Algebras, North-Holland, Amsterdam.
Donaldson S K & Kronheimer P B 1991: The Geometry of Four-Manifolds,
Clarendon Press, Oxford.
Doubrovine B, Novikov S & Fomenko A 1979: Geometrie Contemporaine,
MIR, Moscow.
Dowker J S 1979: Selected Topics in Topology and Quantum Field Theory,
lectures delivered at the Center for Relativity, Austin, Texas.
Drinfeld V G 1983: Sov. Math. Dokl. 27 222.
Dubois-Violette M 1991: in Bartocci C, Bruzzo U & Cianci R (eds.),
Differential Geometric Methods in Theoretical Physics, Proceedings of the
1990 Rapallo Conference, Springer Lecture Notes in Physics 375, Springer,
Berlin.
Dubois-Violette M 1988: C. R. Acad. Sci. Paris 307I 403.
Dubois-Violette M, Kerner R & Madore J 1990: J. Math. Phys. 31 316.
Dubois-Violette M & Launer G 1990: Phys. Lett. B245 175.
647
Efros A L 1986: Physics and Geometry of Disorder, MIR, Moscow.
Eisenhart L P 1949: Riemannian Geometry, Princeton University Press,
Princeton.
Enoch M & Schwartz J.-M 1992: Kac Algebras and Duality of Locally
Compact Groups, Springer, Berlin.
Essam J W 1972: in Domb C & Green M S (eds.), Phase Transitions and
Critical Phenomena, vol.2, Academic Press, London.
Ezawa Z F 1978: Phys. Rev. D18 2091.
Ezawa Z F 1979: Phys. Lett. B81 325.
Faddeev L D 1982: Sov. Phys. Usp. 25 130.
Faddeev L D & Shatashvilli S L 1984: Theor. Math. Phys. 60 770.
Faddeev L D & Slavnov A A 1978: Gauge Fields. Introduction to Quantum Theory, Benjamin/Cummings, Reading.
Fairlie D B, Fletcher P & Zachos C Z 1989: Phys. Lett. B218 203.
Farmer J D, Ott E & Yorke J E 1983: Physica D7 153.
Feynman R P, Leighton R B & Sands M 1965: The Feynman Lectures in
Physics, Addison-Wesley, Reading.
Finkelstein D 1966: J. Math. Phys. 7 1218.
Finlayson B A l972: Phys. Fluids 13 963.
Flanders H l963: Differential Forms, Academic Press, New York.
Fleming H 1987: Rev. Bras. Fs. 17 236.
Fock V 1926: Z. Physik 39 226.
Fock V A 1964:The Theory of Space, Time and Gravitation (2nd ed.),
Pergamon, New York.
Forsyth A R 1965: The Theory of Functions, Dover, New York.
Fraleigh J B 1974: A First Course in Abstract Algebra, Addison-Wesley,
Reading.
Franck F C 1951: Phil. Mag. 42 809.
Freed D S & Uhlenbeck K K l984: Instantons and Four-Manifolds, Springer,
Berlin.
Fullwood D T 1992: J. Math. Phys. 33 2232.
Furry W H 1963: in Brittin W E, Downs B W & Downs J (eds.), Lectures in Theoretical Physics, vol. V, 1962 Summer Institute for Theoretical
Physics, (Boulder) Interscience, New York.
Galetti D & Toledo Piza A F 1988: Physica A149 267.
Gelfand I M & Shilov G E 1964: Generalized Functions, vol.1,
Academic Press, New York.
Gelfand I M, Graev I M & Vilenkin N Ya 1966: Generalized Functions,
Vol. 5,
Academic Press, New York.
Geroch R P 1968 : J. Math. Phys. 9 1739.
648
Gerstenhaber M & Stasheff J (eds.) 1992: Deformation Theory and
Quantum Groups with Applications to Mathematical Physics, AMS, Providence.
Gibling P J 1977: Graphs, Surfaces and Homology, Chapman & Hall,
London.
Gilmore R 1974: Lie Groups, Lie Algebras, and Some of Their Applications,
J.Wiley, New York.
Godbillon G 1971: Elements de Topologie Algebrique, Hermann, Paris.
Goldberg S I l962: Curvature and Homology, Dover, New York.
Goldstein H 1980: Classical Mechanics (2nd ed.), Addison-Wesley, Reading.
Gobel R 1976a: J. Math. Phys. 17 845.
Gobel R 1976b: Comm. Math. Phys. 46 289.
Graver J E & Watkins M E 1977: Combinatorics with Emphasis on the
Theory of Graphs, Springer, Berlin.
Greenberg M J 1967: Lectures on Algebraic Topology, Benjamin, Reading.
Grebogi C, Ott E & Yorke J A 1987: Science 238 585.
Guckenheimer J & Holmes P 1986: Nonlinear Oscillations, Dynamical
Systems, and Bifurcations of Vector Fields, Springer, Berlin.
Guillemin V & Sternberg S 1977: Geometric Asymptotics, AMS, Providence.
Gurarie D 1992: Symmetries and Laplacians, North Holland, Amsterdam.
G
ursey F 1962: Introduction to the de Sitter Groups, in G
ursey F (ed.),
Group Theoretical Concepts and Methods in Elementary Particle Physics,
Istanbul Summer School of Theoretical Physics, Gordon and Breach, New
York.
Haag R 1993: Local Quantum Physics, Springer, Berlin.
Hajicek P 1971: Comm. Math. Phys. 21 75.
Halmos P R 1957: Introduction to Hilbert Spaces, Chelsea Pub.Co., New
York.
Hamermesh M 1962: Group Theory and its Applications to Physical Problems,
Addison-Wesley, Reading.
Harris W F 1975: Phil. Mag. B32 37.
Hawking S W & Ellis G F R 1973: The Large Scale Structure of SpaceTime,
Cambridge University Press, Cambridge.
Hawking S W & Israel W (eds.) 1979: General Relativity: An Einstein
Centenary Survey, Cambridge University Press, Cambridge.
Hawking S W, King A R & McCarthy P J 1976: J. Math. Phys. 17 174.
649
Hehl F W, von der Heyde P, Kerlick G D & Nester J M 1976: Rev. Mod.
Phys. 48 393.
Hilton P J & Wylie S l967: Homology Theory, Cambridge University
Press, Cambridge.
Hilton P J 1953: An Introduction to Homotopy Theory, Cambridge University Press, Cambridge.
Hocking J C & Young G S 1961: Topology, Addison-Wesley, Reading.
Hu S T 1959: Homotopy Theory, Academic Press, New York.
Huang K 1987: Statistical Mechanics (2nd ed.), J. Wiley, New York.
Hurewicz W & Wallman H 1941: Dimension Theory, Princeton University Press, Princeton.
Hwa R W & Teplitz V L 1966: Homology and Feynman Integrals, Benjamin, Reading.
Isham C J, Penrose R & Sciama D W (eds.) 1975: Quantum Gravity,
Oxford University Press, Oxford.
Isham C J 1978: Proc. Roy. Soc. 362 383.
Isham C J 1984: in DeWitt B S & Stora R (eds.), Relativity, Groups and
Topology II, Les Houches Summer School, North-Holland, Amsterdam.
Itzykson C & Zuber J.-B. 1980: Quantum Field Theory, McGraw-Hill,
New York.
Jackiw R 1980: Rev. Mod. Phys. 52 661.
Jackson J D 1975: Classical Electrodynamics (2nd ed.) John Wiley, New
York.
Jacobson N 1962: Lie Algebras, Dover, New York.
Jancel R 1969: Foundations of Classical and Quantum Statistical Mechanics,
Pergamon, Oxford.
Jones V F R 1983: Invent. Math. 72 (1983) 1.
Jones V F R 1985: Bull. Amer. Math. Soc. 12 103.
Jones V F R 1987: Ann. Math. 126 335.
Jones V F R 1991: Subfactors and Knots, CBMS, Amer. Math. Soc.,
Providence, Rhode Island.
Katznelson Y 1976: An Introduction to Harmonic Analysis, Dover, New
York.
Kauffman L H 1991: Knots and Physics, World Scientific, Singapore.
Khavin V P 1991: in Khavin V P & Nikolskij N K (eds.), Commutative Harmonic Analysis I, Encyclopaedia of Mathematical Sciences, vol.15,
Springer, Berlin.
650
Kohno T 1990: New Developments in the Theory of Knots, World Scientific, Singapore.
Kobayashi S & Nomizu K l963: Foundations of Differential Geometry,
Interscience, New York.
Kolmogorov A N & Fomin S V 1970: Introductory Real Analysis, PrenticeHall, Englewood Cliffs.
ements de la Theorie des FoncKolmogorov A N & Fomin S V 1977: El
tions et de
lAnalyse Fonctionelle, MIR, Moscow.
Konopleva N P & Popov V N 1981: Gauge Fields, Harwood, Chur.
Laidlaw M G G & DeWitt-Morette C 1971: Phys. Rev. D3 1375.
Lanczos C 1986:The Variational Principles of Mechanics (4th ed.) Dover,
New York.
Landau L D & Lifshitz E M l969: Statistical Physics (2nd ed.), Pergamon,
Oxford.
Landau L D & Lifshitz E M l975: The Classical Theory of Fields, Pergamon, Oxford.
Landau L D & Lifshitz E M l989: Mechanique des Fluides, MIR, Moscow.
Landau L D & Lifshitz E M l990:Theory of Elasticity, MIR, Moscow.
Lavrentiev M & Chabat B 1977: Methodes de la Theorie des Fonctions
dUne Variable Complexe, MIR, Moscow.
Leinaas J M & Myrheim J 1977: Nuovo Cimento B37 1.
Levi B G 1993: Physics Today, 46 (#3) 17.
Levy M & Deser S (eds.) 1979: Recent Developments in Gravitation,
Carg`ese Lectures 1978, Plenum Press, New York.
Lichnerowicz A 1955: Theorie Globale des Connexions et des Groupes
dHolonomie, Dunod, Paris.
London F 1927: Z. Physik 42 375.
Loos H G 1967: J. Math. Phys. 8 2114.
Love A E H 1944: A Treatise on the Mathematical Theory of Elasticity,
Dover, New York.
Lovelock P & Rund J l975: Tensors, Differential Forms and Variational
Principles,
J. Wiley, New York.
Lu Qi-keng 1975: Chin. J. Phys. 23 153.
Luneburg R K 1966: Mathematical Theory of Optics, University of California Press, Berkeley.
Mackey G W 1955: The Theory of Group Representations, Lecture Notes,
Department of Mathematics, University of Chicago.
Mackey G W 1968: Induced Representations of Groups and Quantum
Mechanics, Benjamin, New York.
651
Mackey G W 1978: Unitary Group Representations in Physics, Probability and Number Theory, Benjamin/Cummings, Reading.
Majid S 1990: Int. J. Mod. Phys. A5 1.
Malament D B 1977: J. Math. Phys. 18 1399.
Mandelbrot B B 1977: The Fractal Geometry of Nature, W. H. Freeman,
New York.
Ma
nes J, Stora R & Zumino B 1985: Comm. Math. Phys. 108 157.
Manin Yu I 1989: Comm. Math. Phys. 123 163.
Marsden J 1974: Applications of Global Analysis in Mathematical Physics,
Publish or Perish, Inc., Boston.
Maslov V P 1972: Theorie des Perturbations et Methodes Asymptotiques,
Dunod-Gauthier-Villars, Paris.
Maurer J 1981: Mathemecum, Vieweg & Sohn, Braunschweig, Wiesbaden.
McGuire J B 1964: J. Math. Phys. 5 622.
Michel L 1964: in G
ursey F (ed.), Group Theoretical Concepts and Methods in Elementary Particle Physics, Istanbul Summer School of Theoretical
Physics (1963), Gordon and Breach, New York.
Milnor J 1973: Morse Theory, Princeton University Press, Princeton.
Misner C W & Wheeler J A 1957: Ann. Phys. 2 525.
Misner C W, Thorne K & Wheeler J A 1973: Gravitation, W. H. Freeman,
San Francisco.
Mrugala R 1978: Rep. Math. Phys. 14 419.
Munkres J R l975: Topology: a First Course, Prentice Hall, New York.
Nabarro F R N 1987: Theory of Crystal Dislocations, Dover, New York.
Nash C & Sen S l983: Topology and Geometry for Physicists, Academic
Press, London.
Neuwirth L P 1965:Knot Groups, Princeton University Press, Princeton.
Neuwirth L P 1979: Scient. Am. 240 84.
Nomizu K 1956: Lie Groups and Differential Geometry, Publ. Math.Soc.
Japan.
Novikov S P 1982: Sov. Math. Rev. 3 3.
Okubo S 1980: Phys. Rev. D22 919.
Okun L B 1984: Introduction to Gauge Theories, publications of the
Institute of Theoretical and Experimental Physics ITEP - 43, Moscow.
Olver P J 1986: Applications of Lie Groups to Differential Equations,
Springer, New York.
Ore O 1963: Graphs and their Uses, L. W. Singer, New York.
Papapetrou A 1951: Proc. Roy. Soc. A209 248.
Pathria R K 1972: Statistical Mechanics, Pergamon, Oxford.
Penrose R & MacCallum M A H 1977: Phys. Rep. 6 241.
Penrose R 1972: Rep. Math. Phys. 12 65.
652
Pontryagin L S 1939: Topological Groups, Princeton University Press,
Princeton.
Popov D A l975: Theor. Math. Phys. 24 347.
Porteous I R 1969: Topological Geometry, van Nostrand, London.
Quinn F 1982: J. Diff. Geom. 17 503.
Rasband S N 1990: Chaotic Dynamics of Nonlinear Systems, J. Wiley,
New York.
Regge T 1961: Nuovo Cimento 19 558.
Rolfsen D 1976: Knots and Links, Publish or Perish, Berkeley.
Rosenfeld B A & Sergeeva N D 1986: Stereoghaphic Projection, MIR,
Moscow.
Rourke C & Stewart I 1986: New Scientist, 4th September, pg. 41
Saaty T L & Kainen P C 1986: The Four-Color Problem, Dover, New
York.
Sadoc J F & Mosseri R 1982: Phil. Mag. B45 467.
Schmidt H.-J. 1993: Fort. Physik 41 179.
Schreider Ju. A 1975: Equality, Resemblance and Order, MIR, Moscow.
Schulman L 1968: Phys. Rev. 176 1558.
Schutz B 1985: Geometrical Methods of Mathematical Physics, Cambridge University Press, Cambridge.
Sierpi
nski W 1956: General Topology, University of Toronto Press; Dover
edition, New York, 2000.
Simmons G F 1963: Introduction to Topology and Modern Analysis, McGrawHill- Kogakusha, N.Y.-Tokyo.
Simon B 1983: Phys. Rev. Lett. 51 2167.
Singer I M & Thorpe J A 1967: Lecture Notes on Elementary Topology
and Geometry, Scott Foresman & Co., Glenview.
Singer I M 1981: Physica Scripta 24 817.
Skyrme T H R 1962: Nucl. Phys. 31 556.
Slebodzinski W. 1970: Exterior Forms and their Applications, PWN,
Warszawa.
Sommerfeld A 1954: Optics, Academic Press, New York.
Sommerfeld A 1964a: Mechanics of Deformable Bodies, Academic Press,
New York.
Sommerfeld A 1964b: Partial Differential Equations in Physics, Academic Press,
New York.
Spivak M 1970: Comprehensive Introduction to Differential Geometry,
Publish or Perish, Brandeis.
Steen L A & Seebach, Jr J A 1970: Counterexamples in Topology, Holt,
Rinehart and Winston, New York.
653
Steenrod N l970: The Topology of Fibre Bundles, Princeton University
Press, Princeton.
Stora R 1984: in tHooft G et al. (eds.), Progress in Gauge Field Theory,
Plenum,
New York.
Streater R F & Wightman A S 1964: PCT, Spin and Statistics, and all
That, Benjamin, New York.
Synge J L 1937: Geometrical Optics, Cambridge University Press, Cambridge.
Synge J L 1960: Relativity: The General Theory, J. Wiley, New York.
Synge J L & Schild A 1978: Tensor Calculus, Dover, New York.
Takesaki M 1978: in DellAntonio G et al (eds.), Springer Lecture Notes
in Physics 80, Springer, Berlin.
Temperley H N V & Lieb E H 1971: Proc. Roy. Soc. 322 251.
Thom R 1972: Stabilite Structurelle et Morphegen`ese , Benjamin, New
York.
Thom R 1974: Mod`eles Mathematiques de la Morphegen`ese, Union Generale
d Editions, Paris.
Timoshenko S & Goodier J N 1970: Theory of Elasticity (3rd ed.) McGrawHill,
New York.
Toulouse G & Vannimenus J 1980: Phys. Rep. 67 47.
Trautman A l970: Rep. Math. Phys. 1 29.
Trautman A l979: Czech. J. Phys. B29 107.
Utiyama R 1955: Phys. Rev. 101 1597.
Vainberg M M l964: Variational Methods for the Study of Nonlinear Operators,
Holden-Day, San Francisco.
Vilenkin N Ya 1969: Fonctions Speciales et Theorie de la Representation
des Groupes, Dunod, Paris.
Warner F W l983: Foundations of Differential Manifolds and Lie Groups,
Scott-Foreman, Glenview.
Warner G 1972: Harmonic Analysis on Semi-Simple Lie Groups I, Springer,
Berlin.
Weinberg S 1972: Gravitation and Cosmology, J. Wiley, New York.
Wenzl H 1988: Invent. Math. 92 349.
Westenholz C l978: Differential Forms in Mathematical Physics, NorthHolland, Amsterdam.
Weyl H 1919: Annalen der Physik 59 101.
Weyl H 1929: Z. Physik 56 330.
654
Weyl H 1932: Theory of Groups and Quantum Mechanics, E. P. Dutton,
New York.
Wheeler J A & Zurek W H (eds.) 1983: Quantum Theory and Measurement, Princeton University Press, Princeton.
Whittaker E T 1944: Analytical Dynamics, Dover, New York.
Whittaker E T 1953: A History of the Theories of Aether and Electricity,
Thomas Nelson, London.
Woodcock A & Davis M 1980: Catastrophe Theory, Penguin Books, Harmondsworth.
Wootters W & Zurek W H 1979: Phys.Rev. D19 473.
Woronowicz S L 1987: Commun. Math. Phys. 111 613.
Wu T T & Yang C N 1975: Phys. Rev. D12 3843.
Wu Y S & Zee A 1985: Nucl. Phys. B258 157.
Yang C N & Mills R L 1954: Phys. Rev. 96 191.
Yang C N 1974: Phys. Rev. Lett. 33 445.
Yang C N & Ge M L 1989: Braid Group, Knot Theory and Statistical
Mechanics,
World Scientific, Singapore.
Yee K & Bander M 1993: Phys. Rev. D48 2797.
Zeeman E C 1964: J. Math. Phys. 5 490.
Zeeman E C 1967: Topology 6 161.
Zeeman E C 1977: Catastrophe Theory, Addison-Wesley, Reading.
Zumino B, Wu Y S & Zee A 1984: Nucl. Phys. B241 557.
Zwanziger J W, Koenig M & Pines A 1990: Ann. Rev. Phys. Chem. 41
601.
Index
E 3 , 189
RP n , 45
S 1 , 32, 38, 42
S 3 , 43
S 4 , 22, 235, 237
S n , 21, 89, 133, 235, 317
Sn , 354
S n , 11, 45
Rn , 641
0 , 120
1 , 86, 92
n , 116
1 , 113
-algebra, 35
E1 , 20
bad, 25
1
E+ , 37
E2 , 9
E3 , 3, 200, 218, 221
E4 , 22, 29
En , 8, 20, 22, 30, 31, 89, 214
R+ , 42
*-algebra, 391
Cn , 45
E4 , 134
En , 45
abelian, 334
abelianized, 337
abelianizer, 337, 637
accumulation point, 16
acoustics, 556
acoustics., 557
action, 626
defined, 342
of a group, 167
on phase space, 625
of a set, 342
on a basis, 178
action functional, 579
adjoint
action, 239
in algebra, 345
representation, 265
Ados theorem, 405
affine
group, 44
space, 637
transformation, 296
affine: locally symmetric manifold, 296
Aharonov-Bohm effect, 77, 97, 102,
131
Alexander polynomial, 368
Alexandrovs compactification, 35
algebra, 344, 347
Banach, 386
C*, 389
defined, 344
enveloping, 347
exterior, 158
factor, 394
general definition, 239
graded, 157, 240, 346
Grassmann, 158
Hopf, 348
Lie, 163, 239
noncommutative, 154
655
656
INDEX
of a group, 347
Bn , 359, 398
of functions, 145, 392
bad E1 , 25
of sets, 15
Baire space, 637
semisimple, 271
Banach
von Neumann, 393
algebra, 326, 386, 387
W*, 393
space, 20, 382, 386, 387, 409
algebra-valued form, 210
barycentric coordinates, 53
algebraic topology, 46, 67
basic field, 287
Allendoerfer, 458
basis, 175
alphabet, 352
anholonomic, 176, 209
alternating group, 355
canonical, 178
alternation, 158
coordinate, 176
Ambrose-Singer holonomy theorem, 315
covector, 178
analytic
dual, 148
atlas, 134
for a topology, 9
manifold, 134
general, 208
angular
global, 175
momentum
holonomic, 176, 209
current, 581
local, 175
anholonomic basis, 176, 209
non-coordinate, 176
annular group, 348
transformations, 177
anomaly, 602, 625, 627
vector
anti-de Sitter spacetime., 615
coordinate, or natural, 147
antiderivation, 242
Berrys phase, 315
antipode, 349
Bethe lattice, 55, 536
antisymmetric tensors, 157
Betti number, 65, 67, 69, 234, 271,
arcwise-connected, 34
367, 447
Aristotle, iii
bialgebra, 348
Arnold, 165
Bianchi identity, 212, 225, 291, 312,
Artins braid groups, 357
593
associated bundle, 179, 276, 293, 297
first, 289
asymptotic flatness, 537
second, 289
atlas
bijective, 637, 639
analytic, 134
billiard theorem, 376
differentiable, 128
bimodule, 341
fibered, 275
Birkhoffs theorem, 377
in general, 128
Bohm-Aharonov effect, 77, 97, 102,
linear, 128
131
attractor, 166
Boliyai, iv
automorphism, 30, 43, 265
Boltzmann, 524
inner, 265
Borel
INDEX
-algebra, 373
measure, 373
set, 373
Born-Infeld electrodynamics, 434
boundary, 122
homomorphism, 56
conditions, 22, 40, 99
and bundles, 274
homomorphism, 67
of a boundary, 63
of a set, 16
boundary conditions, 39
bounded
operator, 423
set, 423
Braid
statistics, 360
braid
equation, 362
geometrical, 356
group, 114, 357, 360, 525, 533
statistics, 104, 114, 525
tame, 357
branch (in a graph), 51
bridges of Konigsberg, 54
Brouwer degree, 153, 443
BRST transformations, 603
bundle
associated, 179, 276, 293, 297
existence of, 275
fiber, 162
heuristic introduction, 273
induced, 302
normal, 273
of 1-forms, 164
of frames, 179, 284, 605
of k-forms, 190
principal, 179
reduced, 302
reduction, 282
space, 162
657
structure group, 179
tangent, 162
tensor, 163
trivial, 164, 274
vector, 275
Burgers
circuit, 545
vector, 290, 545
C -structure, 133
C*-algebra, 389, 637
calculating theorem, 93
canonical, 637
homomorphism, 639
isomorphism, 164
canonical form, 280, 310
Cantor set, 17, 35, 124, 166
capacity dimension, 124
carrier space, 404
Cartan, 189
connection form, 215
lemma, 450
metric, 271
cartesian product, 14, 35
cartesian set product, 334, 342, 372
cartography, 126
catastrophe, 448
Cauchy sequence, 18
Cauchy-Schwarz inequality, 379
causal
anomalies, 24
structure, 31
Cayley
theorem on finite groups, 355
tree, 55, 536
Cech, 116
center, 637
of a group, 334
centralizer, 637
chain, 226
boundary, 63
658
closed, 64
coboundary, 64
defined, 63
face, 63
harmonic, 65
homologous, 68
of a path, 55
singular, 226, 234
chaos
chaotic behaviour, 166
deterministic, 376
character of a representation, 406, 407
characteristic
class, 319
equation, 565
function, 231, 374
of a ring, 637
characteristic equations, 553
characteristics, 553
charge, 584
chart, 125, 177
Chern
theorem, 458
class, 320
chiral field, 603
cholesterics, 549
chord (in a graph), 51
Christoffel symbol, 27, 186, 296, 454,
568, 607, 619
chromatic polynomial, 537
circle, 32, 42, 75
Classical Mechanics, 330
classification
of differentiable manifolds, 137
of fiber bundles, 319
closed
chain, 64
curve, 32
form, 198
set, 13
closure
INDEX
of a braid, 367
of a set, 16
coadjoint representation, 267, 630
coalgebra, 348
coarser topology, 13
coassociativity, 348
coboundary, 64
operator, 64
cochain, 64
coclosed
form, 222
cocycle, 64
Codazzi equation, 453
coderivative, 222, 243, 596
codifferential, 222
coexact
form, 222
cofield, 163
cohomology, 64
group, 69
of Lie algebra representations, 627
color, 607
commutant, 637
commutator, 162
subgroup, 337
compact, 25
group, 42
manifold, 232, 443
without boundaries, 235
space, 446
surface, 457
compact space, 20
compact-open topology, 33, 74, 120
compactification, 22, 35
compactness, 20
local, 22
compensating form, 214
complement, 13
linear, 343
complete
atlas, 129
INDEX
bundle space, 275, 298, 303
differentiable atlas, 134
field, 166
space, 18, 31
completion, 18
complex, 59, 68
components, 156, 160
composition, 135
of functions, 75, 82
comultiplication, 348
condensation, 31, 637, 639
condensation point, 16
conformal
group, 31
spaces, 618
transformation, 187
conformally flat metric, 466, 570
congruency, 637
conjugate class, 638
connected
graph, 54
space, 13
connectedness, 25, 34, 122
multiple, 89
path, 87
simple, 88
connection, 26, 621
as a distribution, 306
flat, 311
form, 306, 452
Cartan, 215
general, 605
general treatment, 303
Levi-Civitta, 611
linear, 286
continuity, 28, 373
continuity equation, 224
continuous operator, 423
contorsion
tensor, 612
contractible, 75, 91, 101, 198, 206
659
contraction, 156
contravariant
image, 182
tensor, 155
vector, 147
convergence, 15, 18, 373
convex, 59
convolution, 340
Conway, 368
coordinate, 32, 125, 136
function, 149
basis, 147, 176
function, 32
neighbourhood, 161
transformation, 127, 177, 588
coproduct, 348
cotangent
bundle, 164, 567
space, 147
counit, 349
covariant
coderivative, 214, 225
derivative, 186, 212, 213, 225, 284,
287, 583, 586
of a section, 294
tensor, 154
vector, 147
covector, 150
covering, 21
homotopy, 109
map, 107
space, 103, 105, 119
covering space, 105
CPn , 113
critical phenomena, 530
critical point, 445, 446
crosspoint, 444
crunode, 121
crystal optics, 184
Curie, 529
current form, 223
660
curvature, 288, 290, 452, 605, 611
constant, 460
form, 311, 452
of a curve, 566
of a light ray, 567
principal radii, 456
tensor, 454
total, Gaussian, 456
curve
closed, or loop, 32, 136
defined, 32
differentiable, 136
index of a, 440
integral, 166
cusp, 121
cycle, 64, 355
indicator polynomial, 356
structure of a permutation, 356
cyclic group, 354
cylinder, 39, 114
De Rham
current, 234
decomposition theorem, 236
theorem on harmonic forms, 236
theorem on periods, 230
de Sitter spacetime, 317
de Sitter spacetime., 615
decomposition theorem, 236
defect, 538
deformation, 538, 638
field, 544
retract, 638
degree, 152, 638
Brower, 153
dense subset, 17, 375
derivation, 242, 391
algebraic, 239, 347
derivative
Frechet, 423
functional, 423
INDEX
Gateaux, or weak, 423
Lie, 169
strong, 423
derived
algebra, 347
set, 16
Descartes, iii, 187
determinants, 195
dielectric constant, 572
diffeomorphic manifolds, 137
diffeomorphism, 137, 163
differentiable
atlas, 128, 177
curve, 136, 145
distribution, 175
manifold, 160, 185
operator, 423
structure, 391
differential, 148
Frechet, 423
Gateaux, 423
strong, 423
weak, 423
differential equation, 183
differential form, 64, 160, 181, 189
differential operator
classification, 556
differential topology, 47, 443
differentially related, 128
dimension, 122, 123
of vector space, 342
defined, 342
dimension theory, 17
Dirac, 326
equation, 578
equation, 613
monopole, 77
space, 381
spinor, 613
Dirac space, 19
direct product, 361
INDEX
661
disc
2
in E , 93
once punctured, 95
twice punctured, 96
disclination, 538
discontinuity
propagation of, 558
discrete
group, 351
metric, 14
topology, 7, 14, 29
dislocation, 538
dispersion relation, 557
distance, 185
distance function, iii, 3, 33, 185, 638
distortion tensor, 544
distribution, 175, 231, 306
divergence, 222
division ring, 338
divisors of zero, 638
dual
basis, 148
of a form
definition, 218
invariant definition, 219
space, 154, 343, 382
tensor, 593
transformation, 223
dual
tensor, 213
dual space
defined, 343
duality
of chains and forms, 233
operation, 218
Pontryagin, 411
symmetry, 558, 593
Tanaka-Krein, 412
dynamical system: flows, 165
dynamical system:maps, 166
edge
closed, 50
open, 50
eikonal equation, 554
eikonal equation, 556, 562, 566
Einstein, 187
equation, 435, 608
space, 454
tensor, 607
Einstein-Hilbert action, 608
electrodynamics
Born-Infeld, 434
electromagnetic form, 203
electromagnetism, 202, 223, 229, 238
electron diffraction, 129
elementary particle, 575
ellipsoid
of inertia, or Poinsot, 184
Fletchers, 184, 573
Fresnel, 184
index, 184, 573
of wave normal, 573
reciprocal, 573
elliptic type operator, 556
endomorphism, 43, 239, 343, 638
defined, 343
energy-momentum, 588
ensemble
statistical, 523
enveloping
algebra, 347
epimorphism, 639
equation
Einsteins, 435, 608
Gauss, 453
Killing, 186
Maurer-Cartan, 209
Maxwells, 202
Maxwells second pair, 223
Ricci, 453
equivalence
662
class, 37, 74
classes, 641
relation, 30, 37, 639, 641
equivalence principle, 608
equivalent representations, 405
Ergodic
flow, 375
ergodic
problem, 376
theorem, 376, 524
theory, 375
Euclid, 459
euclidean plane, 9
euclidean space, iii, 3, 7, 21, 147, 214,
449, 460
pseudo-, 465
euclideanization, 29
Euler
characteristic, 447, 457
class, 320
number, 51, 444
Euler-Lagrange equation, 580
Euler-Poincare characteristic, 60, 69,
165, 443
event
in probability theory, 373
exact form, 164, 192, 198
exchange of particles, 41
exterior
derivative
introduced, 197
algebra, 158, 159
coderivative, 217
derivative, 210
in a general basis, 216
invariant definition, 200
product, 158, 159, 190, 197
as operation, 197
variational calculus, 437, 604
Faraday law, 229
INDEX
Fermats principle, 570
fermion field, 582
fermion number, 245
Feynmans picture, 92
fiber, 107, 162, 179
bundle, 162
heuristic introduction, 273
fibered chart, 275
fibration, 277
field, 161, 457, 576
basic, 287
complete, 166
defined, 339
fundamental, 269, 280
hamiltonian, 627
horizontal, 286
Jacobi, 417
relativistic, 577
scalar, 577
spinor, 578
strictly hamiltonian, 627
theory, 576
vector, 578
vertical, 280
fields of planes, 192
filter, 15, 639
finer topology, 13
finite space, 7, 13, 21
finitely generated group, 334
first homotopy group, 86
first quadratic form, 451
first-countable, 8, 10, 13
first-separability, 23, 25
flat
connection, 311
space, 460
Fletchers ellipsoid, 184, 573
flow, 165, 168
ergodic, 375
hamiltonian, 168
mixing, 375
INDEX
non-ergodic, 375
flow diagram, 165
Fock-Ivanenko derivative, 613
focusing
perfect, 571
force, 191
form, 163
algebra-valued, 210
canonical, 310
frame bundle, 280
closed, 198
coclosed, 222
coexact, 222
curvature, 311
differential, 147, 189
exact, 164, 198
harmonic, 223
horizontal, 310
Kirillov, 629
vector-valued, 150
vertical, 310
four-color problem, 54, 536
Fourier
analysis, 22, 409, 410
coefficient, 381
duality, 413
Fox, 116
Frechet
derivative, 419
derivative, 423
space, 32, 639
fractals, 124
frame, 175, 586
adapted, 450
transformations, 177
Franck index, 119, 550
free group, 353
Fresnel ellipsoid, 184, 572
Friedmann model, 229
Frobenius theorem, 175, 192, 199
in terms of forms, 216
663
full braid group, 357
function, 639
between differentiable manifolds,
135
continuous, 28
distance, 185
homotopic, 73
in coordinates, 135
monodromous, 27
polynomial, 157
function space, 33
functional, 418, 421
linear, 422
functorial properties, 90
fundamental field, 269, 280, 304
fundamental group, 83, 86, 112, 116,
352
fundamental sequence, 18
Gd (EN )], 46
Galileo, iii
Galois, 351
Gateaux derivative, or weak derivative, 423
gauge, 203, 210, 212, 301, 303, 310,
312
anomaly, 602
field, 592, 596
group, 591
potential, 307, 592
prescription, 594
theory, 276, 589, 609
transformation, 579, 598
gauge theory, 212, 238
Gauss
theorem, 455
curvature, 456
equation, 453
normal mapping, 455
theorem, 228, 456
Gelfand-Mazur theorem, 387
664
Gelfand-Naimark theorem, 390
Gelfand-Naimark-Segal construction,
396
general basis, 208
General Relativity, iv, 26, 187, 607
generator, 168, 334, 352
genus, 70
geodesic, 587, 612, 620
equation, 569
equation, 289, 612
geometrical phase, 316
Geometry, 187
geometry
intrinsic, 451
of surfaces, 455
GL(m, R), 178
GL(m, C), 43
GL(m, K), 43
GL(m, R), 43, 282
GNS construction, 396
Graded
algebra, 639
graded
algebra, 157, 191, 638
derivative, 638
ring, 338
graded algebra, 240, 346
graded commutator, 211
gradient, 192
Graph
of a function, 639
graph
defined, 50
theory, 54
Grassmann
algebra, 158
space
complex, 46
real, 46
Grassmann space, 318
gravitation, 607
INDEX
Greens theorem, 101, 228
group, 41
1-parameter, 167
abelian, 334
abstract, 335
action, 167
affine, 44
algebra of a, 347
alternating, 355
annular, 348
compact, 42
defined, 334
discrete, 351
finitely generated, 334
free, 353
Heisenberg, 329
nilpotent, 338
of an algebra, 346
of quaternions, 43
orthogonal, 281
presentation, 353
representation, 22, 335, 403
ring, 340
semisimple, 271, 337
simple, 337
solvable, 338
symmetric, 354
topological, 41
torsion-free, 354
transformation, 335
type of, 409
group reduction, 302
groupoid, 335
Haar measure, 22, 408, 411
Hahn extension theorem, 373
Hamilton equation, 554
Hamilton-Jacobi equation, 554
hamiltonian
field, 627
flow, 168
INDEX
formalism, 594
mechanics, 329
harmonic analysis, 409
non-commutative, 411
harmonic chain, 65
harmonic form, 223
on a Lie group, 271
Hausdorff, 25, 122
space, 134, 373
Hausdorff space, 23
Hecke algebra, 361, 399
hedgehog theorem, 164
Heine-Borel lemma, 21
Heisenberg group, 329
Helmholtz-Korteweg lagrangian, 434
hessian, 445
higher homotopy groups, 118
Hilbert space, 19, 32, 103, 380, 385,
409
Hilbert-Einstein lagrangian, 435
Hodge star operator, 218, 219
Hodge theorem
manifolds with boundary, 237
manifolds without boundary, 236
holonomic
basis, 176, 209
holonomy group, 314
homeomorphism, 31, 35, 75, 137
homogeneous
space, 316, 621
homology, 68, 78
class, 68
cubic, 72
group, 68, 69, 71, 228, 233
integer, 64, 234
real, 64, 234
singular, 60
theory, 47
homomorphism, 70, 335, 639
induced, 118
homothecies, 30
665
homotopic
path, 78
functions, 73, 116
homotopy, 73, 80
class, 74, 85, 89, 116, 153
classification of bundles, 319
group, 71, 91
first, 86
n-th, 117
path, 78
theory, 47
type, 75
homotopy formula, 205
Hopf, 153
Hopf algebra, 323, 413
defined, 348
Hopf-Banach algebra, 326
Hopf-von Neumann algebra, 413
horizontal
curve, 292
field, 286
form, 310
lift
of a curve, 292
of a vector field, 292
horocycle, 470
Hurewicz, 116, 120
hyperbolic space, 618
hyperbolic type operator, 556
hyperfinite algebra, 396
hypersurface, 231, 559
ideal, 639
idempotent, 341, 343
image, 335
imbedded submanifold, 138
imbedding, 138, 140, 450
immersed submanifold, 138
immersion, 138
improper subset, 6
incidence number, 55
666
inclusion, 639
indefinite metric, 185
index
Franck, 550
Morse, 446
of a curve, 440
of a singular point, 442, 457
of a subgroup, 337
index ellipsoid, 573
indiscrete topology, 7, 14, 29
indistinguishability, 41
induced bundle, 302, 319
induced topology, 11, 75
inertia
ellipsoid, 184
tensor, 184
infinite-dimensional spaces, 22
infinitesimal operator, 168
infinitesimal variation, 417
injective, 639, 640
inner product, 19, 220, 234, 343
defined, 343
space, 19, 379, 380
instanton, 22, 222, 596
integrability, 199, 630
integrable, 101
forms, 192
integral, 226
integral curve, 166
integral domain, 640
integrating denominator, 194
integration, 374
interference experiment, 129
interior, 16
interior product, 240
interval I = [0,1], 20
invariance
local, 585
invariant
measure, 408
form on a Lie group, 271
INDEX
polynomial, 71, 367, 398
space, 408
subgroup, 337
inverse image, 28
inverse Poincare lemma, 205
invertible element, 346
involution, 631
of algebra, 345, 347
of vector space, 343
involutive
algebra, 387
irrational numbers, 11
Ising model, 527, 531, 536
isolated point, 16
isometry, 30, 31, 185
isomorphism, 640
canonical, 164, 184
natural, 164
isotopy, 638
(or isotopic deformation), 366
Jacobi
equation, 467, 620
field, 417
identity, 163, 177, 346, 631
Jones, 369, 394
index, 397, 532
polynomial, 398
Konigsberg bridges, 54
Kac algebras, 413
Kauffman, 368, 369
kernel, 67, 335, 640
Kervaire, 134
Killing
equation, 186
field, 186
form, 270
vector, 186, 616
kink, 34
Kirillov form, 629
INDEX
667
Klein bottle, 40
algebra, 163, 210, 239, 266, 346,
350
Klein-Gordon equation, 557, 577, 609
bracket, 239, 346
knot, 531
derivative, 185, 242
group, 366
functional, 436
theory, 534
properties, 172
Kolmogorov capacity, 124
group, 134, 165, 276
Koszul formula, 243
Lie
derivative,
169
Kronecker symbol, 194, 217
lift, 109, 277
Lagrange
light cone, 29
derivative, 419, 580, 583, 584
light ray, 566
theorem, 337
limit point, 16
lagrangian, 580
limiting circle, 470
existence of, 596
line, 11
Laplace-Beltrami operator, 223, 274, line field, 193
609
Linear
laplacian, 221, 222, 245
independence, 216
in homology, 65
linear
large group, 599
atlas, 128
lattice
complement, 343
Bethe, 55, 535
connection, 26, 286, 605
models, 526
dependence, 342
parameter, 526
frame, 278
Lax pair, 630
functional, 422
Lebesgue measure, 35, 373
group, 43
left-regular representation, 408
independence, 342
Leibniz
operator, 239, 343, 423
rule, 145
representation, 214, 626
lemma
space, 380
Cartan, 450
linear space
inverse Poincare, 205
defined, 341
Morse, 445
link, 365, 367
Poincare, 198
Liouville
length
equation, 554
of curve, 185
theorem, 168, 375
of vector, 185
liquid crystal, 547
Lenz, 527
Lobachevsky, iv
letter, 352
plane, 466
Levi-Civita connection, 295, 611
local
Lichnerowicz bracket, 211
compactness, 22, 34, 373
Lie
coordinate, 125
668
homeomorphism, 31
invariance, 584
system of coordinates, 125
transformation, 583
locally
euclidean space, 121
locally finite covering, 26
logistic map, 166
loop (curve), 32, 85
higher-dimensional, 116
of loops, 120
loop (in a graph), 51
loop space, 277
loop(curve)
differentiable, 136
Lorentz
group, 576
group, 577, 616
metric, 4, 13, 27, 181, 185, 618
singlet, 582
transformation, 187, 581
Lorenz gauge, 224, 601
LSC, 125, 147
Luneburg, 554, 558
Lyapunov exponent, 376
Mobius band, 38, 94, 274
magnetic
field, 202, 206
monopole, 126, 153
magnetization, 529
manifold
analytic, 134
imbedded, 138
immersed, 138
Poisson, 628
strictly homogeneous, 628
symplectic
homogeneous, 625
manifolds
diffeomorphic, 137
INDEX
manifolds-with-boundary, 11, 122
Manin space, 325
map, 166
Markov, 137
Maurer-Cartan
equation, 209, 601
form, 604
maximal
atlas, 133
ideal, 640
maximum, 446
Maxwell
equations, 558
equations, first pair, 202
equations, second pair, 223
reciprocal relations, 204
measurable
space, 372
subsets, 373
measure, 20, 35, 340, 382, 523
finite, 373
positive, 373
space, 373
measurement, 185
mechanical
work, 191
metric, iii, 4, 148, 164, 605, 640
conformally equivalent, 618
conformally flat, 570
defined, 181
indefinite, 185
Lorentz, 185
refractive, 568
Riemannian, 185
semi-Riemannian, 185
metric space, iii, 8, 10, 25, 240, 638
metric topologies, 8
metric-indecomposable, 377
metrically transitive, 377
metrizable space, 25, 26, 59, 382, 640
Milnor, 133
INDEX
minimum, 446
Minkowski space, 4, 13, 181, 186, 221,
615
forms in, 224
mixed tensor, 156
Mixing flow, 375
module
defined, 341
momentum
Souriau, 628, 629
monodromy, 102
theorem, 111
monoid, 335
monoid diagrams, 368, 399
monomorphism, 640
monopole, 119
Morse
index, 446
inequality, 447
lemma, 445
theory, 445
motion, 30, 185, 576
moving frame, 214, 450
Moyal bracket, 329, 392
multiply-connected
space, 98
graph, 54
space
defined, 89
wavefunctions, 129
wavefunctions on, 102
natural
basis, 147, 149
isomorphism, 164
neighbourhood, 8
nematic system, 548
network, 10
Newton, iii
nilpotent group, 338
Noethers theorem, 35, 628
669
first, 580
second, 584
non-coordinate basis, 176
non-degenerate critical point, 445
non-Hausdorff space, 24
non-integrable phase factors, 102
non-linear representation, 269, 626
non-metric topological space, 15
non-metric topological spaces, 373
non-orientable manifolds, 129
non-potential force, 191
noncommutative algebra, 154
noncommutative geometry, 327, 392
noncommutative geometry,, 323
norm, 19, 34, 380, 386
norm topology, 19, 380
normal
bundle, 273
coordinates, 612
operator, 389
space, 25
subgroup, 337
normed
ring, 387
vector space, 22, 42, 380
nowhere dense subset, 17
O(n), 44
one-parameter
group, 167
one-to-one, 640
Onsager, 529
onto, 640
open r-balls, 3
open ball, 4, 11
open set, 5, 6, 24
operation, internal, 640
operator, 386
coboundary, 64
continuous, 423
defined, 423
670
laplacian, 65
types of, 556
optical indicatrix, 573
optical length, 570
optics, 554, 556, 565
orbit, 167, 268, 630
order
of a group, 334
of an element, 354
oricycle, 470
orientation, 129, 191
of a simplex, 61
orthogonal
group, 281
sequence, 381
system, 381
parabolic type operator, 556
paracompact, 26, 314, 374
parallel
lines, 459
transport, 284, 289, 293, 294
parallelism, 312
parallelizable
manifold, 164, 276, 287
parallelogram rule, 379
parameterization, 126
parity, 202
partial differential equation, 554
partition
function, 524, 525
of a space, 107
of the identity, 374
path, 32, 54, 74, 78
closed, or loop, 54
Euler, 54
simple, 54
path homotopy, 78
path-component, 34, 74, 80
path-connectedness, 13, 34, 74, 87
path-covering, 109
INDEX
Pauli matrices, 211
paving, 522
percolation, 103
permutation, 354, 362
Peter-Weyl theorem, 412
Pfaffian
equation, 193
form, 190, 197
phase space, 168
enlarged, 168
symmetries, 625
phase transformation, 582
phase transition, 529
piecewise-linear manifold, 129
planar graphs, 51
plastic crystals, 548
Poincare, 116, 367
group, 576
duality, 69
group, 31, 187, 616
inverse lemma, 205
lemma, 63, 64, 198, 569
polynomial, 71
space, 466
theorem du retour, 168
theorem on integration, 228
Poincare
conjecture, 104
Poinsot
construction, 572
ellipsoid, 184
point
critical, 445
critical non-degenerate, 445
maximum, 446
minimum, 446
regular, 445
saddle, 446
singular, 439
pointwise product, 340, 391
Poisson
INDEX
bracket, 330
bracket, 326, 392
equation, 236
manifold, 628
structure, 391
polarization, 182
polyhedron, 226
of a complex, 59
polynomial algebra, 388
polynomial function, 157
Pontryagin
class, 320
duality, 411
positive measure, 373
potential, 101
potential form, 203, 224
Potts model, 55, 531
power set, 14, 27, 372, 641
Poynting vector, 558, 566
prametric, 15, 641
predual, 390
presentation, 353
principal curvature radii, 456
principal fiber bundle, 179, 317
probability
space, 373
theory, 373
product topology, 14
projection, 107, 161
projective
group, 44
line, 45, 97
plane, 45, 97, 113
representation, 626
space, 44
transformation, 44
projector, 44, 341, 343
proper subset, 6
pull-back, 159, 206, 451
pull-back bundle, 319
pure braid group, 357
671
pure state, 45
push-foward, 159
Pythagoras theorem, 379
quadratic form
first, 451
second, 451
quantization, 39, 100, 330, 385
quantum group, 324, 350, 392, 413
Quantum Mechanics, 92, 129
quantum mechanics, 326
quantum space, 325
quaternions, 43
quotient
space, 37, 621
topology, 37, 75, 107
rank
of a finitely generated group, 334
of a function, 138
rational numbers, 11
reduction
of bundle, 282, 302
refinement, 13, 29
of a covering, 26
refractive index, 571
regular point, 445
regular values, 152
relation
defined, 641
equivalence, 641
relative
topology, 11
compactness, 22
relator, 353
rep`ere mobile, 215
representation, 335, 626
adjoint, 265
of a group, 266
of an algebra, 266
coadjoint, 267, 630
672
completely reducible, 407
defined, 403
dimension of, 405
equivalent, 405
faithful, 297, 404
generalities, 265, 403
irreducible, 406
linear, 404
non-linear, 269
projective, 626
reducible, 406
regular, 267, 408
right-regular, 408
ring, 411
singlet, 407
space, 404
theory, 404
trivial, 404
unitary, 405
resolvent, 387
restricted holonomy group, 314
retraction, 641
Ricci, 620
equation, 453
tensor, 454, 607
theorem, 606
Riemann, iv, 619
inaugural address, iv
integral, 226
sheet, 106
surface, 105
tensor, 453, 607
Riemannian, 570
manifold, 185
metric, 185, 220, 282, 374
Riesz theorem, 22
right-regular representation, 408
ring
defined, 338
of a group, 339
of subsets, 372
INDEX
RP1 , 97
RP2 , 97, 113
RP3 , 114
RPn , 113
S1 , 32, 75
S2 , 36, 127, 446, 458
S4 , 237
S7 , 137
saddle-point, 446
scalar
curvature, 607
invariant, 607
scalar field, 577
Schwartz, 234
Schwarzschild
radius, 126
space, 450
second quadratic form, 451
second-countable, 11, 25, 122
topology, 10
second-separable, 23
section, 162, 276
self-adjoint, 389, 637
algebra, 387
self-dual forms, 222, 596
semigroup, 335
seminorm, 380
semisimple algebra, 271
semisimple group, 337
separability, 23
first, 23
second, 23
separable space, 17, 23
separated space, 23
sequence, 18, 21, 31
Cauchy, 18
set function, 373
countably additive, 373
finitely additive, 373
positive, 373
INDEX
set product, 372
signature, 220
simple
group, 337
simplex, 58, 226
simply-connected, 88
graph, 54
Sina theorem, 376
singular chain, 226
singular point, 439, 442, 457
sink, 444
skein relations, 361, 368
smectic
phase, 547
smooth function, 446
SO(3), 114
SO(n), 44
soldering, 280
form, 280
solvable group, 338
Sorgenfrey line, 25
source, 444
Souriau momentum, 628, 629
space, 6
compact, 20
concept of, iii, 4, 129
conformally flat, 618
contractible, 75
cotangent, 147
Dirac, 19
Hilbert, 19
hyperbolic, 618
inner product, 19
separable, 23
separated, 23
tangent, 147
vector, 19
spaces of paths, 35
spacetime, 29, 616
spacetime topology, 13
specific heat, 529
673
spectral radius, 387
spectrum, 387
sphere, 235, 276, 458
sphere S 7 , 133
sphere S n , 133
sphere S n , 11
sphere S7 , 137
spherical top, 115
spin current, 581, 582
spinor, 613
field, 578
spontaneous breakdown of symmetry,
530
standard field, 287
star-operation, 218
star-product, 329
starshaped sets, 380
state
of a physical system, 385, 524
stereographic projection, 36, 618
Stiefel manifold, 46, 317, 319
Stokes theorem, 228
straight line, 380
strain tensor, 544
strange attractor, 166
strong
derivative, 419, 423
differential, 423
topology, 380
stronger topology, 13
structure
coefficients, 176
structure equations, 311, 450
En , 215
structure group, 179, 275, 278
su(2) algebra, 211
SU(2) group, 115, 211
sub-bundle, 302
subgroup, 337
topological, 42
submanifold, 137
674
subring, 338
subsets, 15
improper, 6
proper, 6
superconductor, 100
superfluid, 100
superselection rules, 34
supersymmetry, 245
surfaces, 4, 21, 455, 456
surgery, 104
surjective, 639, 641
syllable, 352
symmetric, 641
algebra, 387
symmetric group, 354
symmetric tensor, 156
symmetry, 577
properties, 40
and anomaly, 625
of a lagrangian, 436
on phase space, 625
transformation, 37
system
of coordinate functions, 125
Tn , 40
tail, 18
tame braid, 357
Tanaka-Krein duality, 412
tangent
bundle, 162, 273
field, 166
space, 147, 154
vector, 146
Temperley and Lieb, 398
tensor, 154, 190
antisymmetric, 157
bundle, 163
components, 156
contraction, 156
contravariant, 155
INDEX
covariant, 154
field, 164
mixed, 156
product, 154, 344
of representations, 407
symmetric, 156
tensor product
defined, 344
tetrad, 282
trivial, 588
tetrad field, 283
tetrahedron, 52, 65, 66, 93
theorem
calculating, 93
billiard, 376
bundle classification, 319
du retour, 168
ergodic, 376
hedgehog, 164
thermodynamics, 204
tiling, 522
topological
number, 22
conservation laws, 34
defect, 47
dimension, 123
group, 22, 41
invariant, 35, 64, 90, 641
manifold, 121
number, 15, 47, 61, 64
product, 14, 39
space, 4
defined, 7
vector space, 382
vector space., 35
topology, 4, 31
basis, 9
coarser, 13
compact-open, 33
defined, 6
discrete, 7, 14
INDEX
finer, 13
general, 3
indiscrete, 7, 14
induced, or relative, 11
norm, 19, 380
quotient, 36
strong, 380
stronger, 13
trivial, 14
uniform, 380
weaker, 13
torsion, 280, 288, 290, 606, 612
-free group, 354
group, 354
subgroup, 354
torus, 40, 69, 94, 108
trace, 398
transfer matrix, 528, 531, 533
local, 532
transformation
conformal, 187
Lorentz, 187
transformation group, 335, 404
transgression, 205
transition function, 127
transition functions, 301
translation, 30, 581
tree
Cayley, 55, 535
tree graph, 51
triangular inequality, or sub-additivity,
379
triangulation, 59, 69, 72, 444
trivial
bundle, 164
representation, 335, 404
topology, 14
trivialization, 298, 304
trivializer, 335
twisted field, 274
type
675
of a factor algebra, 395
of a group, 409
typical fiber, 274
ultrafilter, 639
uniform topology, 380
unimodular group, 408
uniqueness of solutions, 24
unit
algebra, 345, 387
ring, 338, 372
universal
bundle, 317
covering space, 107
enveloping algebra, 347, 350
universality of gravitation, 609
universe
Friedmann model, 229
upper-half
plane, 12, 25
space, 11
Urysohns theorem, 25
vacuum, 119, 530
van Hove theorem, 529
variation, 78, 416
vector, 145, 146
bases, or frames, 175
contravariant, 147
covariant, 147
field, 161, 276, 391, 578, 582
space, 8, 19, 76, 77
normed, 379
vector fibered atlas, 275
vector space
defined, 341
vector-valued form, 150, 210
vertex, 50
vertical
field, 280, 286
form, 310
676
space, 294, 303, 307
volume form, 191, 220
canonical, 218
of a hypersurface, 231
von Neumann
algebra, 124, 369, 393, 409, 413
bicommutant theorem, 394
decomposition theorem, 394
ergodic theorem, 377
W*-algebra, 393
Wang
theorem, 621
wave equation, 224
wave front, 566
wavefunction, 100, 101, 106, 114
weak
derivative, (or Gateaux derivative),
423
differential, 423
topology, 382
weaker topology, 13
wedge product, 190
Weil, 458
Wess-Zumino
condition, 602, 627
Weyl
prescription, 328, 353
Weyl-Wigner picture of Quantum Mechanics, 328, 392
Whitney theorem, 134, 139, 215, 449
on metric, 185
wiedersehen manifold, 571
Wigner
function, 328
Wigner-Inon
u
group contraction, 621
winding number, 119, 153, 550, 597
word, 352
group, 352
work, 230
INDEX
as a differential form, 191
Yang-Baxter equation, 324, 350, 362
classical, 631
Yang-Mills
equation, 224, 237, 583, 593, 596
Young double-slit experiment, 129
Zeeman, 31
topology for spacetime, 13, 14, 29
zero section, 300