Lnotes ds2007f
Lnotes ds2007f
Lnotes ds2007f
net/publication/268444274
CITATIONS READS
4 21,565
1 author:
Rainer Klages
Queen Mary, University of London
96 PUBLICATIONS 2,225 CITATIONS
SEE PROFILE
Some of the authors of this publication are also working on these related projects:
All content following this page was uploaded by Rainer Klages on 18 November 2014.
Rainer Klages
School of Mathematical Sciences
Queen Mary, University of London
Notes for version 1.1 compiled by Phil Howard
c
Rainer Klages, Phil Howard, QMUL
Contents
1 Preliminaries . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
Bibliography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 75
4 Contents
1 Preliminaries
There exist two essentially different approaches to the study of dynamical systems, based on
the following distinction:
time-continuous nonlinear differential equations ⇋ time-discrete maps
One approach starts from time-continuous differential equations and leads to time-discrete
maps, which are obtained from them by a suitable discretization of time. This path is
pursued, e.g., in the book by Strogatz [Str94].1 The other approach starts from the study of
time-discrete maps and then gradually builds up to time-continuous differential equations,
see, e.g., [Ott93, All97, Dev89, Has03, Rob95]. After a short motivation in terms of nonlinear
differential equations, for the rest of this course we shall follow the latter route to dynamical
systems theory. This allows a generally more simple way of introducing the important
concepts, which can usually be carried over to a more complex and physically realistic
context.
As far as the style of these lectures is concerned, it is important to say that this course, and
thus these notes, are presented not in the spirit of a pure but of an applied mathematician
(actually, of a mathematically minded theoretical physicists). That is, we will keep technical
mathematical difficulties to an absolute minimum, and if we present any proofs they will
be very short. For more complicated proofs or elaborate mathematical subtleties we will
usually refer to the literature. In other words, our goal is to give a rough outline of crucial
concepts and central objects of this theory as we see it, as well as to establish crosslinks
between dynamical systems theory and other areas of the sciences, rather than dwelling
upon fine mathematical details. If you wish, you may consider this course as an applied
follow-up of the 3rd year course MAS308 Chaos and Fractals.
That said, it is also not intended to present an introduction to the context and history
of the subject. However, this is well worth studying, the field now encompassing over
a hundred years of activity. The book by Gleick [Gle96] provides an excellent starting
point for exploring the historical development of this field. The very recent book by Smith
[Smi07] nicely embeds the modern theory of nonlinear dynamical systems into the general
socio-cultural context. It also provides a very nice popular science introduction to basic
concepts of dynamical systems theory, which to some extent relates to the path we will
follow in this course.
This course consists of three main parts: The introductory Part I starts by exploring some
examples of dynamical systems exhibiting both simple and complicated dynamics. We
then discuss the interplay between time-discrete and time-continuous dynamical systems
in terms of Poincaré surfaces of section. We also provide a first rough classification of
different types of dynamics by using the Poincaré-Bendixson theorem. Part II introduces
1
see also books by Arrowsmith, Percival and Richards, Guckenheimer and Holmes
6 1 Preliminaries
2
see http://www.maths.qmul.ac.uk/˜mathres/dynsys for further information
Part I
Pivot Torque
θ
Rod
Figure 2.1: Driven pendulum of mass M with a torque applied at the pivot and subject to
gravity.
Without worrying too much about how one can use physics to obtain an equation of motion
for this system starting from Newton’s equation of motion, see [Ott93, Str94] for such
derivations, we move straight to the equation itself and merely indicate whence each term
arises:
8 2 Examples of realistic dynamical systems
where for sake of simplicity we have set the mass M equal to one. Here we write θ̇ := dθ dt
to
denote the derivative of θ with respect to time, which is also sometimes called the angular
velocity. In (2.1) θ is an example of a dynamical variable describing the state of the system,
whereas ν, A, f are called control parameters. Here ν denotes the friction coefficient, A the
amplitude of the periodic driving and f the frequency of the driving force. In contrast to
dynamical variables, which depend on time, the values for the control parameters are chosen
once for the system and then kept fixed, that is, they do not vary in time. Equation (2.1)
presents an example of a driven (driving force), nonlinear (because of the sine function,
sin x ≃ x − x3 /3!), dissipative (because of driving and damping) dynamical system.
It is generally impossible to analytically solve complicated nonlinear equations of motion
such as (2.1). However, they can still be integrated by numerical methods (such as Runge-
Kutta integration schemes), which allows the production of simulations such as the ones
that can be explored in “The Pendulum Lab”, a very nice interactive webpage [Elm98].
Playing around with varying the values of control parameters there, one finds the following
four different types of characteristic behaviour: This systems has already been studied in
many experiments, even by high school pupils!
ω ω
(1) (2)
θ θ
ω (4) ω
(3)
θ θ
Figure 2.2: Plots of four different types of motion of the driven nonlinear pendulum (2.1)
in the (θ, ω)-space under variation of control parameters.
θ̈ + θ = 0 . (2.2)
2.2 Bouncing ball 9
This system is called the harmonic oscillator, and the resulting dynamics is known as
libration. If we represent the dynamics of this system in a coordinate system where we
plot θ and the angular velocity ω := θ̇ as functions of time, we obtain pictures which
look like the one shown in Fig. 2.2 (a).
2. same as 1. except that θ can be arbitrarily large, −π ≤ θ < π: Fixed points exist at
θ = 0 (stable) and θ = π (unstable). If the initial condition θ(0) = π is taken with
non-zero initial velocity ω 6= 0, continual rotation is obtained; see Fig. 2.2 (b).
3. same as 2. except that the damping is not zero anymore, ν 6= 0: The pendulum comes
to rest at the stable fixed point θ = 0 for t → ∞. This is represented by some spiraling
motion in Fig. 2.2 (c).
4. same as 3. except that now we have a periodic driving force, that is, A, f 6= 0: In this
case we explore the dynamics of the full driven nonlinear pendulum equation (2.1).
We observe a ‘wild’, rather unpredictable, chaotic-like dynamics in Fig. 2.2 (d).
We conclude this discussion by mentioning that the driven nonlinear pendulum is a paradig-
matic example of a non-trivial dynamical system, which also displays chaotic behavior. It
is found in many physical and other systems such as in Josephson junctions (which are
microscopic semi-conductor devices), in the motion of atoms and planets and in biological
clocks. It is also encountered in many engineering problems.
G
M
Vibrating plate ν A, f
Figure 2.3: A ball of mass M subject to gravity, that elastically bounces off a vibrating
plate.
look at the dynamics of a point particle with constant speed that collides elastically with
the disk and the walls of the box. That is, we have specular reflection at the walls where
in- and outgoing angles are the same, θ = θ′ .2 This system strongly reminds one of the
idealistic case of a billiard table (without friction but with a circular obstacle) and is indeed
referred to as a particle billiard in the literature. The particular example shown in Fig. 2.4
is known as the Sinai billiard.
111111
000000
000000
111111
000000
111111
000000
111111
000000
111111
000000
111111
θ
θ’
Normal
One may now ask the question about sensitivity of initial conditions, that is: What happens
2
see http://www.dynamical-systems.org/sinai/index.html for a simulation (with optional gravity)
2.3 Particle billiards 11
to two nearby particles in this billiard (which do not interact with each other) with slightly
different directions of their initial velocities? Since at the moment no computer simulation
of this case is available to us, we switch to a slightly more complicated dynamical system
as displayed in Fig. 2.5, where we can numerically explore this situation [Miy03].
M G
Figure 2.5: A billiard where a point particle bounces off semicircles on a plate.
Stretch Fold
Figure 2.6: Schematic representation of the stretch and fold mechanism of an ensemble of
particles in a chaotic dynamical system.
As we shall see later in this course, this is one of the fundamental mechanisms of what is
called “chaotic behavior” in nonlinear dynamical systems. The purpose of these lectures
is to put the handwaving assessment of the behavior illustrated above for some “realistic”
dynamical systems onto a more rigorous mathematical basis such that eventually we can
answer the question what it means to say that a system exhibits “chaotic” dynamics. To
the end of this course we will also consider the dynamics of statistical ensembles of particles
such as in the simulation.
Definition 1 [Las94, Kat95]1 A dynamical system consists of a phase (or state) space P
and a family of transformations φt : P → P , where the time t may be either discrete, t ∈ Z,
or continuous, t ∈ R. For arbitrary states x ∈ P the following must hold:
1. φ0 (x) = x identity and
2. φt (φs (x)) = φt+s (x) ∀t, s ∈ R additivity
In other words, a dynamical system may be understood as a mathematical prescription for
evolving the state of a system in time [Ott93, All97]. Property 2 above ensures that the
transformations φt form an Abelian group. As an exercise, you may wish to look up different
definitions of dynamical systems on the internet.
Here φt is the transformation that we have already encountered in Definition 1 above. Note
that this does not answer the question of how to construct the flow for arbitrary initial
conditions x(0).
Definition 3 A single path in phase space followed by x(t) in time is called the trajectory
or orbit of the dynamical system.2
See Fig. 3.1 for an example.
x3
x(t)
x(0)
x2
x1
Figure 3.1: Example of a trajectory in 3-dimensional phase space (note that the trajectory
is not supposed to cross itself).
In the following we will assume the existence and uniqueness of solutions of the vector field
F . A proof of this exists if F is smooth (continuously differentiable, C 1 ) [Str94].
Corollary 1 For smooth F two distinct trajectories cannot intersect, nor can a single tra-
jectory cross itself for t < ∞.
Proof: Assume a single trajectory crosses itself, see Fig. 3.2. Starting with the initial
condition at the point of intersection there are two choices of direction to proceed in, which
contradicts uniqueness. The same argument applies to two distinct trajectories. q.e.d.
If ∃ uniqueness, x(0) determines the outcome of a flow after time t: we have what is called
determinism.
2
In this course we use both words synonymously.
14 3 Definition of dynamical systems and the Poincaré-Bendixson theorem
One may now ask about the general ‘character’ of a flow φt (x), intuitively speaking: Is it
‘simple’ or ‘complicated’ ? In the following we give a first, rather qualitative answer.
Definition 4 We call a flow simple if for t → ∞ all x(t) are either fixed points x =
constant, respectively ẋ = 0, or periodic orbits, i.e., closed loops x(t + τ ) = x(t), τ ∈ R.
Two examples of fixed points are shown in Figs. 3.3: In both cases the fixed point is located
at the origin of the coordinate system in the two-dimensional plane. However, if we look
at initial conditions in an environment around these fixed points and how the trajectory
determined by it evolves in time, we may observe different types of behavior: For example, in
one case the trajectory ‘spirals in’ to the fixed point by approaching it time-asymptotically,
whereas in the other case the trajectory ‘spirals out’.
We remark that these are by far not the only cases of possible dynamics [Str94]. So we see
that a fixed point, apart from its mere existence, can have a very different impact onto its
environment: In the first case we speak of a stable fixed point, whereas in the second case
the fixed point is said to be unstable.
x2 x2
x1 x1
The same reasoning applies to the simple example of a periodic orbit shown in Fig. 3.4: The
left hand side depicts a circular periodic orbit. However, if we choose initial conditions that
are not on this circle we may observe, for example, the behavior illustrated on the right
hand side of this figure: trajectories ‘spiral in’ onto the periodic orbit both from the interior
of the circle and from the exterior. An isolated closed trajectory such as this periodic orbit
is called a limit cycle, which in this case is stable.
Of course, as for the fixed point the opposite case is also possible, that is, the limit cycle is
unstable if all nearby trajectories spiral out (you may wish to draw a figure of this). The
detailed determination and classification of fixed points and periodic orbits is in the focus
of what is called linear stability analysis in the literature, see [Ott93, Str94] in case of flows.
For maps we will learn right this later in the course.
x2
x2
x1
x1
Limit cycle
Figure 3.4: left: circular periodic orbit; right: approach to this orbit as a limit cycle
One may now ask the further question of how large the dimensionality N of the phase space
has to be for complicated behaviour to be possible. The answer is given in terms of the
following important theorem:[Str94, Rob95]4
2d 3d
Figure 3.5: Trajectories in 2-d (left) cannot cross whereas in 3-d (right), they can.
Example 1 Let us revisit the driven nonlinear pendulum that we have seen before,
θ̈ + ν θ̇ + sin θ = A sin(2πf t) .
x1 := θ̇, x2 := θ, x3 := 2πf t .
The ‘trick’ of incorporating the time dependence of the differential equation as a third
variable allows us to make the vector field autonomous leading to
is called a time-discrete map. xn+1 = M (xn ) are sometimes called the equations of motion
of the dynamical system.
Choosing the initial condition x0 determines the outcome after n discrete time steps (hence
determinism) in the following way:
x1 = M (x0 ) = M 1 (x0 ),
x2 = M (x1 ) = M (M(x0 )) = M 2 (x0 )(6= M (x0 )M (x0 )!).
Example 2 [Ott93] For N = 1 let zn ∈ N0 be the number of insects hatching out of eggs
in year n ∈ N0 . Let r > 0 be the average number of eggs laid per insect. In an (for the
insects) ‘ideal’ case we have
where we assume that insects live no longer than for one year. This straightforwardly implies
for r > 1 an exponentially increasing and for r < 1 an exponentially decreasing population.
3.2 Time-discrete dynamical systems 17
Example 3 Cobweb plot for the logistic map restricted to the parameter range 1 < r < 2
and for 0 ≤ x ≤ 1, see Fig. 3.6.
y
y
y=x
y=x 1
M
y =M(x 0) unstable
fixed point
x0 x1 x2 xs x
x0 x1 x=1/2 x
Figure 3.6: left: first step of a cobweb plot for the logistic map with 1 < r < 2; right:
magnification of the left hand side with the result for the iterated procedure.
Constructing a cobweb plot for a map proceeds in five steps according to the following
algorithm:
1. Choose an initial point x0 and draw a vertical line until it intersects the graph of the
chosen map M(x). The point of intersection defines the value y = M(x0 ) of the map
after the first iteration.
2. By using Eq. (3.5) we now identify y with the next position in the domain of the map
at time step n = 1 which leads to x1 = y.
18 3 Definition of dynamical systems and the Poincaré-Bendixson theorem
3. This is represented in the figure by drawing a horizontal line from the previous point
of intersection until it intersects with the diagonal y = x.
4. Drawing a vertical line from the new point of intersection down to the x-axis makes
clear that we have indeed graphically obtained the new position x1 on the abscissa.
5. However, in practice this last vetical line is neglected for making a cobweb plot. In-
stead, we immediately construct the next iterate by repeating step 1 to 3 discussed
above for the new starting point x1 instead of x0 , and so on.
In summary, for a cobweb plot we continuously alternate between vertical and horizontal
lines that intersect with the graph of the given map and the diagonal; see also p.5 of [All97]
and [Has03] for details.
Example 4
1. The logistic map is not invertible, since one image has two preimages, as you may
convince yourself by a simple drawing.
xn = yn+1
1
yn = (f (xn ) − xn+1 )
k
1
= (f (yn+1) − xn+1 ) , (3.11)
k
so we can conclude that the map is invertible.
3.3 Poincaré surface of section 19
We may now be wondering whether there is any condition on N for the existence of ‘com-
plicated solutions’ in time-discrete maps, in analogy to Corollary 2 for time-continuous
dynamical systems. The answer is given as follows:
Example 5
1. Consider the logistic map Eq. (3.9) for 0 < r ≤ 4. It is smooth, defined on a compact
set but not invertible, hence complicated solutions are possible.
2. Consider Hénon’s map Eq. (3.10) for smooth f (x). This map is smooth, however,
whether the time asymptotic dynamics is defined on a closed, bounded set R is not
clear. On the other hand, the map is invertible and N = 2, hence complicated solutions
are in any case possible.
from the nth to the (n + 1)st piercing under the condition that x3 = K is kept fixed.
This definition appears to be fairly simple, however, there are at least two subtleties: Firstly,
it is quite easy to produce a Poincaré surface of section for a given differential equation nu-
merically, but only in exceptional cases can the corresponding Poincaré map be obtained
analytically. Secondly, if the times τ between two piercings are constant, then some triv-
ial information about the dynamical system is conveniently separated out by producing a
20 3 Definition of dynamical systems and the Poincaré-Bendixson theorem
x3
C(n+2)
B(n+1)
A(n)
K
x2
x1
Figure 3.7: A continuous trajectory in space pierces the plane x3 = K at several points at
discrete time n.
Poincaré surface of section. However, most often there is no reason why τ should be con-
stant. If this is not the case, the complete dynamics is defined by what is called a suspended
flow, or suspension flow,
Eq. (3.14) is called the first-return time map.5 This allows some idea of how the Poincaré-
Bendixson theorem 1 applies to reduced-dimension discrete systems: The above suspended
flow still provides an exact representation of the underlying time-continuous dynamical
system. However, if we refer only to the first Eq. (3.13) by neglecting the second Eq. (3.14),
as we did before, we reduce the dimensionality of the whole dynamical system by one.
Of course we could also consider the whole suspended flow Eqs. (3.13),(3.14) as an N-
dimensional map, however, the set on which Eq. (3.14) is defined is always unbounded and
hence we could not relate to the Poincaré-Bendixson theorem.
Still, by construction our Poincaré map Eq. (3.13) is invertible. However, if we neglect any
single component of the vector in Eq. (3.13), implying that we further reduce the phase space
dimensionality by one, the resulting map will become noninvertible, because we have lost
some information. Hence, in view of the Poincaré-Bendixson theorem 1 it is not surprising
that despite their minimal dimensionality, one-dimensional noninvertible maps can exhibit
complicated time-asymptotic behavior.
5
Obviously, this equation then replaces the dynamics for the eliminated phase space variable, however,
the detailed relation between these two different quantities is typically non-trivial [Gas98, Kat95].
3.3 Poincaré surface of section 21
A second version of a Poincaré map is obtained by a Poincaré surface of section in time. For
this purpose, we sample a time-continous dynamical system not with respect to a constraint
in phase space but at discrete times tn = t0 + nτ, n ∈ N. In this case, the map determining
Eq. (3.14) of the suspended flow boils thus down to T (xn ) = nτ : We have what is called a
stroboscopic sampling of the phase space.
This variant is very convenient for dynamical systems driven by periodic forces such as,
for example, our nonlinear pendulum Eq. (2.1) (why?). However, we wish to illustrate this
technique for a simpler system, where this is easier to see. This system provides a rare
example for which the Poincaré map can be calculated analytically. In order to do so, we
first need to learn about a mathematical object called the (Dirac) ‘δ-function’, which we
may introduce as follows.
We are looking for a ‘function’ having the following properties:
0 , x 6= 0
δ(x) := (3.15)
∞, x=0
For example, think of the sequence of functions defined by
1 x2
δγ (x) := √ exp(− 2 ) (3.16)
γ 2π 2γ
as shown in Fig. 3.8. It is not hard to see that in the limit of γ → 0 this sequence has the
desired properties, that is, δγ (x) → δ(x) (γ → 0).
Y
γ 0
γ3
γ2
γ1
x
Figure 3.8: A sequence of functions approaching the δ-function.
We remark that many other representations of the δ-function exist. Strictly speaking the
‘δ-function’ is not a function but rather a functional, respectively a distribution defined on
a specific (Schwartz) function space.6
6
A nice short summary of properties of the δ-function, written from a physicist’s point of view, is given,
for example, in the German version of the book by F.Reif, Statistische Physik und Theorie der Wärme, see
Appendix A.7.
22 3 Definition of dynamical systems and the Poincaré-Bendixson theorem
Two important properties of the δ-function that we will use in the following are its normal-
ization, Z ǫ
∀ǫ > 0 dx δ(x) = 1 , (3.17)
−ǫ
length l, ’inertia’ I
frictionless pivot
_
kicks of strengthK
l
no gravity
The equation of motion for this dynamical system is straightforwardly derived from physical
arguments [Ott93], however, we just state it here in form of
∞
X
θ̈ = k sin θ δ(t − mτ ) , (3.19)
m=0
where the dynamical variable θ describes the turning angle and k := K/I is a control
parameter. As before we can rewrite this differential equation as a vector field,
θ̇ = ω (3.20)
∞
X
ω̇ = k sin θ δ(t − mτ ) . (3.21)
m=0
According to Eq. (3.21) ω is constant during times t 6= mτ between the kicks but changes
discontinuously at the kicks, which take place at times t = mτ . Eq. (3.20) then implies that
θ ∼ t between the kicks by changing continuously at the kicks, reflecting the discontinuous
changes in the slope ω.
Let us now construct a suitable Poincaré surface of section in time for this dynamical system.
Let us define θn := θ(t) and ωn := ω(t) at t = nτ + 0+ , where 0+ is a positive infinitesimal.
That is, we look at both dynamical variables right after each kick. We then integrate
Eq. (3.20) through the δ-function at t = (n + 1)τ leading to
Z (n+1)τ +0+
dt θ̇ = θn+1 − θn = ωn τ (3.22)
nτ +0+
θn+1 = θn + ωn (3.24)
ωn+1 = ωn + k sin θn+1 . (3.25)
This two-dimensional mapping is called the standard map, or sometimes also the Chirikov-
Taylor map. It exhibits a dynamics that is typical for time-discrete Hamiltonian dynamical
systems thus serving as a standard example for this class of systems [Ott93].7 For conve-
nience, the dynamics of (θn , ωn ) is often considered mod 2π. We remark in passing that
Eqs. (3.24),(3.25) can also be derived from a toy model for cyclotron dynamics, where
a charged particle moves under a constant magnetic field and is accelerated by a time-
dependent voltage drop.8
We may now further reduce the kicked rotor dynamics by feeding Eq. (3.24) into Eq. (3.25)
leading to
ωn+1 = ωn + k sin(θn + ωn ) (3.26)
By assuming ad hoc that θn ≪ ωn , which of course would have to be justified in detail if
one wanted to claim the resulting equation to be a realistic model for a kicked rotor,9 one
arrives at
ωn+1 = ωn + k sin ωn . (3.27)
This one-dimensional climbing sine map is depicted in Fig. 3.10. It provides another example
of a seemingly simple mapping exhibiting very non-trivial dynamics that changes in a very
complicated way under parameter variation.
Part II
7
see, e.g., the book by Lichtenberg and Lieberman for further studies of this specific type of dynamical
systems.
8
see, e.g., a review by J.D.Meiss on symplectic twist maps for some details (Rev. Mod. Phys. 64, 795
(1992).
9
This justification is not trivial and depends very much on the choice of the control parameter k; see
research papers by Bak et al.
24 4 Some basic ingredients of nonlinear dynamics
Ln+1
π 2π Ln
Topological properties of
one-dimensional maps
One-dimensional maps are the simplest systems capable of chaotic motion. They are thus
very convenient for learning about some fundamental properties of dynamical systems. They
also have the advantage that they are quite amenable to rigorous mathematical analysis.
On the other hand, it is not straightforward to relate them to realistic dynamical systems.
However, as we have tried to illustrate in the first part of these lectures, one may argue for
such connections by carefully using discretizations of time-continuous dynamics.1
Let us start the second part of our lectures with another very simple but prominent example
of a one-dimensional map.
1
The ‘physicality’ of one-dimensional maps is a delicate issue on which there exist different points of
view in the literature [Tél06].
25
x
x0 1
Figure 4.2: Approximation of the piecewise linear nondifferentiable tent map by a sequence
of nonlinear differentiable unimodal maps.
A second question may arise about the necessity, or the meaning, of the noninvertibility
of the tent map compared with realistic (Hamiltonian) dynamical systems, which are most
26 4 Some basic ingredients of nonlinear dynamics
2. Then it folds the resulting line segment due to the presence of the cusp at x = 1/2,
which leads to motion bounded on the unit interval.
2 y=2x
stretch
2x 1
0 1/2 1 0 1 2
fold
2 0
2−2x 0 1/2 1 x
0 1
fold
The tent map thus yields a simple example for an essentially nonlinear stretch-and-fold
mechanism, as it typically generates chaos. This mechanism is encountered not only in
the bouncing ball billiard but also in many other realistic dynamical systems. We may
remark that ‘stretch and cut’ or ‘stretch, twist and fold’ provide alternative mechanisms for
generating complicated dynamics. You may wish to play around with these ideas in thought
experiments, where you replace the sets of points by kneading dough.
From now on we essentially consider one-dimensional time-discrete maps only. However,
most of the concepts that we are going to introduce carry over, suitably amended, to n-
dimensional and time-continuous dynamical systems as well.
−π/2 π/2
x
Example 9
1. F (x) = tan x is a C ∞ -diffeomorphism, as one can see by playing around with the
functional forms for arctan x and tan x under differentiation.
Example 10
1. Let F (x) = x = Id(x). Then the set of fixed points is determined by Fix(F ) = R; see
Fig. 4.5.
2. Let F (x) = −x. The fixed points must fulfill F (x) = x = −x ⇒ 2x = 0 ⇒ x = 0, so
Fix(F ) = 0.
28 4 Some basic ingredients of nonlinear dynamics
F(x)
The period two points must fulfill F 2 (x) = F (F (x)) = x. However, F (x) = −x and
F (−x) = x, so Per2 (F ) = R. But note that Prime Per2 (F ) = R \ {0}!
Both results could also have been inferred directly from Fig. 4.6.
F(x)
Figure 4.6: Set of fixed points and points of period two for F (x) = −x.
Remark 2 In typical dynamical systems the fixed points and periodic orbits are isolated
with ‘more complicated’ orbits in between, as will be discussed in detail later on.
There exists also a nice fixed point theorem: A continuous function F mapping a compact
interval onto itself has at least one fixed point; see [Dev89] for a proof and for related
theorems. The detailed discussion of such theorems is one of the topics of the module
‘Chaos and Fractals’.
3x−x3
Example 11 Let F (x) = 2
.
3x−x3
1. The fixed points of this mapping are calculated to F (x) = x = 2
⇒ x − x3 =
x(1 − x2 ) = 0 ⇒ Fix(F ) = {0, ±1}.
2. Let us illustrate the dynamics of this map in a cobweb plot. For this purpose, note
that the extrema are F ′ (x) = 21 (3 − 3x2 ) = 0 ⇒ x = ±1 with F (1) = 1, F (−1) = −1.
4.3 Dense sets, Bernoulli shift and topological transitivity 29
3 √
The roots of the map are F (x) = 0 = 3x−x
2
⇒ x(3 − x2
) = 0 ⇒ x ∈ {0, ± 3 ≃ 1.73}.
We can now draw the graph of the map, see Fig. 4.7. The stability of the fixed points
can be assessed by cobweb plots of nearby orbits as we have discussed before.
y
stable
1
−1 sqrt(3)
−sqrt(3) unstable 1 x
−1
stable
3x−x3
Figure 4.7: Cobweb plot for the map defined by the function F (x) = 2
.
3. The points of period 2 are determined by F 2 (x) = x, so one has to calculate the roots of
this polynomial equation. In order to do so it helps to separate the subset of solutions
Fix(F ) ⊆ Per2 (F ) from the polynomial equation, which defines the remaining period
2 solutions by division.
Example 12
2. One can easily construct eventually periodic orbits via backward iteration as illustrated
in Fig. 4.9.
−1 1 x
fixed point
i=1
i=2 i=2 x
Figure 4.9: Construction of eventually periodic orbits for an example via backward iteration.
4.3 Dense sets, Bernoulli shift and topological transitivity 31
ε ε
p
Figure 4.10: Illustration of an ǫ neighbourhood Nǫ (p).
ε 0
∋
x B
y Α
∋
Certainly, much more could be said on the denseness and related properties of rational and
irrational numbers in R. However, this is not what we need for the following. We will
continue by introducing another famous map:
Example 13 The Bernoulli shift (also shift map, doubling map, dyadic transformation)
The Bernoulli shift shown in Fig. 4.12 is defined by
2x , 0 ≤ x < 1/2
B : [0, 1) → [0, 1) , B(x) := 2x mod 1 = . (4.2)
2x − 1 , 1/2 ≤ x < 1
Proposition 2 The cardinality |Pern (B)|, i.e., the number of elements of Pern (B), is equal
to 2n − 1 and the periodic points are dense on [0, 1).
32 4 Some basic ingredients of nonlinear dynamics
0 1/2 1
Proof: Let us prove this proposition by using a more convenient representation of the
Bernoulli shift dynamics, which is defined on the circle.2 Let
denote the unit circle in the complex plane [Rob95, Has03], see Fig. 4.13.
Im(z)
|z|=1
Re(z)
Figure 4.13: Representation of a complex number z = cos φ + i sin φ on the unit circle.
Then define
We have thus lifted the Bernoulli shift dynamics onto R in form of φ → 2φ. Note that
the map on the circle is C 0 , whereas B in Eq. (4.2) is discontinuous at x = 1/2. This is
one of the reasons why mathematically it is sometimes more convenient to use the circle
representation; see also below.
2
Strictly speaking one first has to show that Eq. (4.2) is topologically conjugate to Eq. (4.4), which implies
that many dynamical properties such as the ones we are going to prove are the same. Topological conjugacy
is an important concept in dynamical systems theory which is discussed in detail in the module ‘Chaos and
Fractals’, hence we do not introduce it here; see also [Dev89, Ott93, Has03, All97].
4.3 Dense sets, Bernoulli shift and topological transitivity 33
Let us now calculate the periodic orbits for B(z). Let n ∈ N and
n
B n (z) = ((z 2 )2 . . .)2 = z 2 = z . (4.5)
| {z }
n times
where for the last equality we have introduced a phase k ∈ Z expressing the possibility that
we have k windings around the circle. Matching the left with the right hand side yields
2n φ = φ + k and, if we solve for φ,
k
φ= . (4.7)
2n −1
Now let us restrict onto 0 ≤ φ < 1 in order to reproduce the domain defined in Eq. (4.2).
This implies the constraint
0 ≤ k < 2n − 1, k ∈ N0 . (4.8)
Let us discuss this solution for two examples before we state the general case: For n = 1
we have according to Eq. (4.8) 0 ≤ k < 1, which implies φ = 0 = Fix(B). Analogously, for
n = 2 we get 0 ≤ k < 3, which implies φ ∈ {0, 13 , 23 } = Per2 (B).
n
In the general case, we thus find that the roots of z 2 −1 = 1 are periodic points of B of
period n. There exist exactly 2n − 1 of them, i.e. |Pern (B)| = 2n − 1, and they are uniformly
spread over the circle, see n = 2 (if you wish, please convince yourself for higher n′ s), with
equal intervals of size δn → 0 (n → ∞). Hence, the periodic points are dense on [0, 1).
q.e.d.
There are at least two different ways of how to prove this. The idea of the first version is
to use the above theorem. For convenience (that is, to avoid the point of discontinuity) we
may consider again the map on the circle. Then for whatever subset U we choose ∃N ∈ N
such that B N (U) ⊇ S 1 as sketched in Fig. 4.14, which is due to the linear expansion of the
map. This just needs to be formalized [Has03]. It is no coincidence that this mechanism
reminds us of what we have already encountered as “mixing”.
B(U)
U
2φ
B2(U)
A second way is to use what is called a symbolic dynamics, or coding of an orbit. Special
cases are, for example, the binary or decimal representations of a real number. However,
since symbolic dynamics is covered by the module Chaos and Fractals we do not introduce
this concept in our course; see p.42 of [Dev89], p.212 of [Has03] or p.39 of [Kat95] for a
proof and for further details.
We have thus learned that the Bernoulli shift has both a dense set of periodic orbits and
is topologically transitive, which earlier we have associated with ‘simple’ and ‘complicated’
dynamics, respectively. So surprisingly this simple map simultaneously exhibits both types
of dynamics, depending on initial conditions.
Example 15
1. The Bernoulli shift is not minimal (why?).
2. The rotation on the circle Rλ (θ) as defined on the exercise sheet is minimal if λ is
irrational. The proof is a slight extension of what we have already shown in the
coursework, that is, for irrational λ there is no periodic orbit [Has03, Kat95, Dev89].
Remark 3
1. Obviously, minimality implies topological transitivity.
2. Minimality is rather rare in dynamical systems, because it allows no existence of
periodic orbits.
3. Later on we will encounter ideas that are very similar to topological transitivity by
additionally employing what is called a measure, leading us to concepts of ergodic
theory.
4.4 Stability 35
4.4 Stability
In this chapter we make our qualitative assessment of the stability of periodic points quan-
titative, which we have explored in terms of cobweb plots. We start with the following
definition:
Remark 4
1. This means that for a sink all points “sufficiently close” to p are attracted to p.
Accordingly, for a source all such points are repelled from p. For a marginal fixed
point both is possible, and one can further classify this point as being marginally
stable, unstable or both, depending on direction.
2. Above we followed Devaney [Dev89] by defining stability for differentiable fixed points
F ′ (p) only. One can dig a little bit deeper without requiring differentiability for F (p)
by using ǫ-environments and then proving the above inequalities as a theorem [All97].
Definition 19 [All97] The set W s (p) of initial conditions x whose orbits converge to a sink
p, limn→∞ F n (x) = p, is called the stable set or basin of p. W u (p) denotes the unstable set
of initial conditions, whose orbits are repelled from a source.
For invertible maps the definition of W u (p) can be made precise via backward iteration for
given x. For non-invertible maps W u (p) can be defined as the set of all x for which there
exists a set of preimages F (x−i ) = x−i+1 with limn→∞ F −n (x) = p [Rob95].
3
Example 16 Consider the map defined in Example 11 by the function F (x) = 3x−x 2
. Let
′ 1 2 ′ 3
us recall that Fix(F ) = {0, ±1}. Since F (x) = 2 (3 − 3x ) we have F (0) = 2 > 1, so as we
already know from cobwep plots in Fig 4.7 it is a source, whereas F ′ (±1) = 0 < 1 are sinks.
Fig. 4.15 illustrates that the basin of attraction for p = 1 has actually a very complicated,
intertwined structure [All97]. This is revealed by iterating backwards the set of initial
conditions B1 leading to the two preimages B2 and B2′ , and so on. Because of symmetry,
an analogous topology is obtained for p = −1. This points towards the fact that basins of
attraction can exhibit fractal structures,3 which indeed is often encountered in dynamical
systems.
We now check for the stability of periodic points.
U
x
B2’ B1 B2
S B3’
Figure 4.15: An example of a map exhibiting a complicated basin structure, which here is
constructed by generating preimages of a given subset of the basin.
Proof: This trivially follows from Definition 18. Define G(x) := F k (x) and look at
G(p) = p. Then p is an attracting/repelling/marginal fixed point of G ⇐⇒ p is an
attracting/repelling/marginal fixed point of F k ⇐⇒ p is an attracting/repelling/marginal
periodic orbit of F . q.e.d.
That is, for hyperbolic maps we wish to rule out any marginal behavior in the nth iterates.
Remark 5
Example 17
1. The tent map has a point of non-differentiability at x = 1/2, hence it is piecewise
C 1 and piecewise expanding, which according to the remark above implies piecewise
hyperbolicity.
2. Let us consider again the logistic map L(x) = rx(1 − x) , x ∈ [0, 1] , r > 0, see also
Eq. (3.9). It is L′ (x) = r(1 − 2x) with L′ (1/2) = 0 ∀r. Consequently L cannot be
expanding.
We are finally able to define mathematically what previously we already referred to, in a
very loose way, as “chaos”.
Remark 6 This definition highlights two basic ingredients of a chaotic dynamical system:
1. A dense set of periodic orbits related to ‘simple’ dynamics, which provide an element
of regularity.
2. Topological transitivity reflecting ‘complicated’ dynamics and thus an element of irreg-
ularity.
Note that according to the Birkhoff Transitivity Theorem 2 topological transitivity also
ensures indecomposability of the map, which in turn implies that these two sets must be
intertwined in a non-trivial way.
38 5 Definitions of deterministic chaos
Example 18 In Proposition 2 we have seen that the Bernoulli shift has a dense set of
periodic orbits. However, Proposition 3 stated that it is also topologically transitive. Con-
sequently it is D-chaotic.
ε δ
x y Fn(y)
Fn(x)
Nε (x)
Let us now choose δ > 0. Then the above solution stipulates that ∀x and ∀Nǫ (x) ∃y ∈ Nǫ (x)
such that after n > m iterations
q.e.d.
2. Topological transitivity which means that there exists a dense orbit that, despite
stretching, eventually returns close to its initial starting point. This implies that
the dynamical systems must exhibit as well some folding mechanism.
This interplay between “stretch and fold” is precisely what we have seen in the earlier
computer simulations of Chapter 2.
One may now wonder why we have two different definitions of chaos and whether there
exists any relation between them. A first answer to this question is given by the following
theorem:
We thus see that there is an exponential separation between two nearby points as we follow
their trajectories. The rate of separation λ(x0 ) := ln 2 is called the (local) Ljapunov exponent
of B(x).
This simple example can be generalized as follows, leading to a general definition of the
Ljapunov exponent for one-dimensional maps F . Consider
δxn = |x′n − xn | = |F n (x′0 ) − F n (x0 )| =: δx0 enλ(x0 ) (δx0 → 0) (5.5)
for which we presuppose that an exponential separation of trajectories exists.3 By further-
more assuming that F is differentiable we can rewrite this equation to
1 δxn
λ(x0 ) = lim lim ln
n→∞ δx0 →0 n δx0
1 |F n (x0 + δx0 ) − F n (x0 )|
= lim lim ln
n→∞ δx0 →0 n δx0
n
1 dF (x)
= lim ln . (5.6)
n→∞ n dx x=x0
Using the chain rule we obtain
dF n (x)
= F ′ (xn−1 )F ′ (xn−2 ) . . . F ′ (x0 ) , (5.7)
dx x=x0
which leads to our final result
n−1
1 Y ′
λ(x0 ) = lim ln F (xi )
n→∞ n
i=0
n−1
1X
= lim ln |F ′ (xi )| . (5.8)
n→∞ n
i=0
The last expression defines a time average (in the mathematics literature this is sometimes
called a Birkhoff average), where n terms along the trajectory with initial condition x0 are
summed up by averaging over n. These considerations motivate the following important
definition:
Definition 25 [All97] Let F ∈ C 1 be a map of the real line. The local Ljapunov number
L(x0 ) of the orbit {x0 , x1 , . . . , xn−1 } is defined as
n−1 n1
Y
′
L(x0 ) := lim F (xi ) (5.9)
n→∞
i=0
Remark 9
1. If F is not C 1 but piecewise C 1 , the definition can still be applied by excluding single
points of non-differentiability.
3. If F ′ (xi ) = 0 ⇒6 ∃λ(x). However, typically this concerns only a finite set of points.
4. For an expanding map it follows from Definition 21 that λ(x) > 0. Likewise, for a
contracting map λ(x) < 0.
The concept of Ljapunov exponents enables us to come up with a third definition of deter-
ministic chaos:
1. F is topologically transitive.
Remark 10
1. “Typical” means, loosely speaking, that this statement applies to any point that we
randomly pick on J with non-zero probability.5
2. The reason why we require typicality for initial conditions is illustrated in Fig. 5.2. It
shows an example of a map with an unstable fixed point at x = 1/2, where λ(1/2) > 0.
However, this value of λ is atypical for the map, since all the other points are attracted
to the stable fixed points at {0, 1}, where the Ljapunov exponents are negative. Thus
it would be misleading to judge for L-chaos based only on “some” initial condition.
4. This is the reason why in the applied sciences “chaos in the sense of Ljapunov” became
such a popular concept. However, it seems that mathematicians rather prefer to speak
of chaos in the sense of Devaney or Wiggins.
5. λ > 0 implies that a map is sensitive, however, the other direction is not true. That
is, W-chaos is weaker than L-chaos, since trajectories can separate more weakly than
exponentially.
6. Note also that W-chaos requires sensitivity ∀x, whereas for L-chaos λ > 0 for one
(typical) x is enough. This is so because, in contrast to sensitivity, the Ljapunov
exponent defines an average (statistical) quantity.
5
We will give a rigorous definition of this statement later on, after we know what a measure is, see
Def. 48.
42 5 Definitions of deterministic chaos
M’>1
0 1
M’<1
Figure 5.2: Example of a map that has a positive local Ljapunov exponent, whereas for
typical initial conditions the Ljapunov exponent is negative.
is the Ljapunov exponent along the periodic orbit {p0 , p1 , . . . , pn−1 }. For the fixed point case
n = 1 we have λ(p) = ln |F ′ (p)|.
Example 22
1. For the Bernoulli shift B(x) = 2x mod 1 we have B ′ (x) = 2 ∀x ∈ [0, 1) , x 6= 21 , hence
λ(x) = ln 2 at these points. So here the Ljapunov exponent is the same for almost all
x, because the map is uniformly expanding.
2. For the map F (x) = (3x − x3 )/2 discussed previously, see Figs. 4.7 and 4.15, we
already know that Fix(F ) = {0, ±1}. Let us calculate the map’s Ljapunov exponents
at these fixed points: With F ′ (x) = 23 − 23 x2 we have F ′ (0) = 32 ⇒ λ(0) = ln 23 >
0 corresponding to a repelling fixed point. For the other two fixed points we get
F ′ (±1) = 0, so unfortunately in these (rare) cases the Ljapunov exponent is not
defined. The calculation of Ljapunov exponents for period two orbits will be subject
of the coursework sheet.
5.4 Summary
We conclude this second part with what may be called a tentative summary: Let us assume
that we have a map F ∈ C 1 defined on a compact set and that it is topologically transitive.
Figure 5.3 then outlines relationships between different dynamical systems properties, all
topologically characterizing “chaotic behavior” by assessing the stability of a dynamical
system in slightly different ways.
However, there is one definition in the figure that we have not discussed so far:
Definition 28 F is topologically mixing if for any two open sets U, V ⊂ J ∃N ∈ N such
that ∀n > N, n ∈ N , F n (U) ∩ V 6= ∅.
You may wish to explore yourself in which respect this property relates to what we have
previously encountered as topological transitivity, see Theorem 2.
expanding
HK HK cw
lm lm
D-chaos topologically
L-chaos
mixing
Banks HK cw
W-chaos
Figure 5.3: Tentative summary of relations between different topological chaos properties.
Figure 5.3 may be understood as follows: Open arrows represent logical implications that
hold basically under the conditions stated above,6 crossed-out bold arrows denote situations
where a counterexample can be found ruling out the existence of a respective implication.
The label ‘HK refers to proofs contained in the book by Hasselblatt and Katok [Has03], ‘cw’
means that you were supposed to show this in a coursework problem, ‘lm’ stands for the
logistic map, which provides counterexamples for certain values of the control parameter,
‘Banks’ points to Theorem 3. The arrows without any labels hold for counterexamples that
can be found elsewhere in the literature.7 See also the masters thesis by A.Fotiou [Fot05]
for details, which summarizes recent literature on all these relations.
Obviously there are still gaps in this diagram: Of particular importance would be to learn
whether W-chaos can imply topological mixing. One may furthermore wonder whether D-
chaos and L-chaos can imply topological mixing, whereas it appears unlikely that topological
mixing implies D-chaos or that it implies that a map is expanding. For the author of these
lecture notes all these questions remain to be answered, and respective hints on solutions
would be most welcome.
We may finally remark that apart from the topological chaos properties discussed here, one
can come up with further definitions of chaos which, however, are based on more sophisti-
cated concepts of dynamical systems theory that go beyond the scope of these lectures.8
6
In some cases the detailed technical requirements for the validity of the proofs can be more involved.
7
For W-chaos not implying D-chaos see counterexamples in [Ban92] and in a paper by Assaf, for W-chaos
not implying L-chaos see the Pomeau-Manneville map, for topologically mixing not implying L-chaos see
some research papers by Prozen and Campbell on irrational triangular billiards.
8
This relates to notions of dynamical randomness in terms of, e.g., topological and measure-theoretic
entropies or symbolic dynamics as they are encountered in what is called the ergodic hierarchy in the
literature [Arn68, Ott93, Guc90].
Part III
Probabilistic description of
one-dimensional maps
For this part of our lectures we recommend Refs. [Las94, Dor99, Bec93] as supplemetary
literature. The first book introduces the topic from a rigorous mathematical point of view,
whereas the other two books represent the physicist’s approach.
We first derive a fundamental equation that describes how a whole collection of points in
phase space evolves in time under iteration of a mapping.
6.1 The Frobenius-Perron equation 45
x x
0 1/6 2/6 3/6 4/6 5/6 1 0 1
Figure 6.1: Left: The logistic map for control parameter r = 4 with M = 6 bins; right:
Histogram for an ensemble of points iterated n times in the logistic map, computed by using
these bins.
Example 23 [Las94]
Let us consider again the logistic map Eq. (3.9), say, at the parameter value r = 4 for which
it reads L(x) = 4x(1 − x) , x ∈ [0, 1], see Fig 6.1 left for a picture. Let us now sample the
nonlinear dynamics of this map by doing statistics in the following way:
1. Take what is called a statistical ensemble in phase space, which here are N points with
initial conditions xi0 , i = 1, . . . , N that are, e.g., uniformly distributed on [0, 1].1
2. Iterate all xi0 according to the usual equation of motion xin+1 = L(xin ).
4. Construct histograms by counting the fraction Nn,j of the N points that are in the jth
bin at time step n. This yields a result like the one depicted in Fig. 6.1 right.
This very intuitive approach, which is easily implemented on a computer, motivates the
following definition:
Definition 29
number of points Nn,j in bin at position xj at time step n
ρn (xj ) := (6.1)
total number of points Ntimes bin width δx = M1
where xj := 21 ( j−1
M
+ j
M
) = 2j−1
2M
is the centre of the bin, defines a probability density.
1
The idea of a statistical ensemble is that one looks at a “swarm” of representative points in the phase
space of a dynamical system for sampling its statistical properties; see [Bec93] or Wikipedia in the internet
for further information.
46 6 Dynamics of statistical ensembles
Remark 11
1. ρn (xj ) is normalised,
M P
X Nn,j δx
j N
ρn (xj )δx = = =1 , (6.2)
j=1
Nδx N
The question is now whether there exists an equation that describes the dynamics of ρn (x)
in the limits of infinitesimally small bins, δx → 0, and an infinite number of points, N → ∞.
That is, we want to understand how ρn (x) changes with discrete time n by iterating the
map. The answer is provided by the Frobenius-Perron equation. Here we derive it in an
intuitive way, based on the idea of the conservation of the number of points for an ensemble
under iteration [Ott93]. Conservation of probability implies that2
with y = F (x) if the map under consideration is one-to-one. However, a chaotic map such
as, for example, the Bernoulli shift, is not injective. According to Fig. 6.2 we thus have to
modify Eq. (6.4) to X
ρ(y)δy = ρ(xi )δxi . (6.5)
y=F (xi )
δy
0 δ x0 δ x1 1 x
Figure 6.2: Idea of deriving the Frobenius-Perron equation, demonstrated for the example
of the Bernoulli shift.
2
In probability theory this is called transformation of variables.
6.1 The Frobenius-Perron equation 47
This leads to
X δxi
ρ(y) = ρ(xi )
i
δy
X 1
= ρ(xi )
i
δy/δxi
X 1
= ρ(xi ) (δxi → 0) (6.6)
i
|F ′(xi )|
for which we have assumed that the map is at least piecewise differentiable and F ′ (xi ) 6= 0
except at single points. We now identify ρn+1 (x) = ρ(y) and ρn (x) = ρ(x) yielding our final
result:
Remark 12
1. In physical terms this is an example of a continuity equation expressing the conserva-
tion of the number of points in the phase space of a dynamical system.
2. This equation has the form of a recursive functional equation and can thus be solved
by (numerical ) iteration.
Eq. (6.8) makes clear that there exists a formal analogy between the FPE and the equation
of motion xn+1 = F (xn ) of a map. Note, however, that the latter equation determines the
dynamics of single points, whereas the FPE describes the dynamics of a statistical ensemble
of them.
It is easy to show (see coursework) that P exhibits the following two properties:
1. Linearity:
P (λ1 ρ1 + λ2 ρ2 ) = λ1 P ρ1 + λ2 P ρ2 , λ1 , λ2 ∈ R (6.10)
48 6 Dynamics of statistical ensembles
2. Positivity:
P ρ ≥ 0 if ρ ≥ 0 (6.11)
So we are dealing with a linear positive operator. We could explore further mathematical
properties of it along these lines. For this purpose we first would have to define the function
space on which P is acting.3 We could then characterize this operator, for example, in
terms of Markov and semigroup properties. However, these are advanced topics of functional
analysis on which we do not want to elaborate in these lectures. Ref. [Las94] gives a nice
introduction if you are interested in this direction of research.
Focusing onto more applied topics, an important question can be posed on the basis of the
following definition.
is called a fixed point of P . ρ∗ (x) is also called an invariant density, which expresses the
fact that it does not change with time n.
A crucial problem is now to calculate ρ∗ (x) for a given map, which requires to solve the
FPE Eq. (6.7). This is of fundamental interest in the theory of dynamical systems, because
ρ∗ (x) provides a statistical characteristisation of a map in terms of what eventually happens
to an ensemble of points. However, to obtain analytical solutions is only possible in rather
special cases.
In the next section we will learn about a basic method that works for certain classes of
piecewise linear maps. However, let us first illustrate the above concept of the Frobenius-
Perron operator and invariant densities for the simplest example we can think of:
Example 24 [Bec93]
For the Bernoulli shift B(x) = 2x mod 1 the FPE Eq. (6.7) reads
1 1
P ρ(x) = ρ(x1 ) + ρ(x2 ) (6.13)
|B ′ (x1 )| |B ′ (x2 )|
with
x = B(xi ) , i = 1, 2
1
2x , 0 ≤ x1 < 21
= , (6.14)
2x2 − 1, 21 ≤ x2 < 1
see Fig. 6.3. However, we can do better by explicitly constructing the Frobenius-Perron
operator P for this map, as follows: We first observe that B ′ (x) = 2 (except at x = 1/2).
We can then calculate the two points x1 and x2 as functions of x by piecewise inverting
Eq. (6.14) to xi = B −1 (x). This leads to
x 1
x1 = B −1 (x) = , 0 ≤ x1 < and 0 ≤ x < 1
2 2
x+1 1
x2 = B −1 (x) = , ≤ x2 < 1 and 0 ≤ x < 1 . (6.15)
2 2
3
We wish to work on a Lebesgue space of measurable functions, which is at least L1 .
6.1 The Frobenius-Perron equation 49
0 x1 x2 1 x
Figure 6.3: Idea of deriving the Frobenius-Perron equation, demonstrated for the example
of the Bernoulli shift.
which represents the explicit form of the Frobenius-Perron operator for B(x). Let us now
look for a fixed point of the FPE of this map, which for this operator must satisfy
∗ ∗ 1 x 1 x+1
ρ (x) = P ρ (x) = ρ + ρ . (6.17)
2 2 2 2
A naive guess might be that the invariant density for this map is simply uniform on the
whole interval, ρ∗ (x) = c, where c is a constant (why?). We can easily check this hypothesis
by plugging our ansatz into the above equation yielding ρ∗ (x) = c = 12 c + 12 c, which verifies
that our guess is correct. Since ρ∗ (x) is a probability density it must be normalized, which
for a distribution that is uniform on the unit interval implies ρ∗ (x) = 1. So we can conclude
that B(x) has an invariant density of ρ∗ (x) = 1.
Remark 13
1. We have obtained one fixed point of the FPE, however, is this the only one? That
poses the question about uniqueness of the solution. Or in other words, if we choose
different initial densities ρ0 (x) and iteratively solve the FPE for this map, is there any
dependence on initial densities in that these densities converge to different invariant
densities?
That such a situation can easily occur is illustrated by the “double Bernoulli shift”
depicted in Fig. 6.4: Start, for example, with a density that is concentrated on 0 ≤
x ≤ 1/2 only but is zero otherwise. Then it will converge to a uniform density on the
same interval while it is zero otherwise. On the other hand, if we choose a density
that is concentrated on 1/2 ≤ x ≤ 1 only but is zero otherwise, the invariant density
will be uniform on the latter interval while it is zero otherwise. Consequently, we have
here a map that does not exhibit a unique invariant density – what we get depends
on our choice of the initial density.
50 6 Dynamics of statistical ensembles
0 1 x
Figure 6.4: The “double Bernoulli shift” as an example of a map that has no unique invariant
density.
In general, the question about existence and uniqueness of an invariant density for a
given map is difficult to answer [Las94, Bec93]. The result also depends on what kind
of density we are looking for, as characterized by further mathematical properties like
continuity, differentiability etc. that we may require for our result. We will get back
to these questions in the following sections, which will give us some further insight.
2. In case of the Bernoulli shift we have solved the FPE for ρ∗ (x) by just guessing a
correct solution. In the following section we will present a method by which, for a
certain class of maps, one can systematically solve the FPE for obtaining invariant
densities.
2. intIi ∩ intIj = ∅ (i 6= j), that is, there is only overlap at the boundaries.
Example 25 Consider the unit interval I = [0, 1]. Then {Ii }i=1,2 = {I1 , I2 } = {[0, 21 ], [ 12 , 1]}
is a partition; see Fig. 6.5. Note that one can find infinitely many different partitions here.
6.2 Markov partitions and transition matrices 51
I1 I2
0 1/2 1
An interesting question is now what happens to partition parts if we let a map act onto
them. Let us study this by means of the example sketched in Fig. 6.6, where we apply the
map D(x) defined in the figure on the previous sample partition. The map then yields
I2 D(x)
1/2
I1
0 1 x
I1 1/2 I2
Figure 6.6: The map D(x) applied onto a partition consisting of the two parts {I1 , I2 }.
Note that here partition parts or unions of them are recovered exactly under the action of
the map. This is the essence of what is called a Markov property of a map:
If you like formal definitions, you may prefer the following one:
Example 26 Let us check this formal definition for the map D(x) of Fig. 6.6. We just have
52 6 Dynamics of statistical ensembles
to verify that the logical implication holds for the four cases
I2 E(x)
1/2
E(I 2)
I1
0 1 x
I1 1/2 I2
Figure 6.7: An example of a map E(x) with a partition that is not Markov.
The first branch of the map does not cause any problem, since it is identical with the one
of our previous map D(x). But let us look at the action of the second branch:
Remark 14
1. Maps for which there exists a Markov partition are often called Markov maps. How-
ever, note that sometimes further conditions such as hyperbolicity, smoothness, etc.,
are required in the literature.
2. Markov partitions appear to have first been constructed by Sinai in 1968 for the Sinai
billiard. For such higher dimensional systems the principle of Markov partitions is the
same as the one described above, however, the precise definition is technically a bit
more involved than for one-dimensional maps [Kat95, Guc90].
3. Markov partitions are not unique for a given map, see the coursework.
6.2 Markov partitions and transition matrices 53
4. Markov partitions are important because they define, loosely speaking, an ‘invariant
frame’ according to which the action of a map on statistical ensembles can be described.
They enable to map the statistical dynamics of a system onto a Markov process – hence
the name –, which is precisely what we will explore in the following.
Example 28 The map F shown in Fig. 6.8 is defined on a Markov partition (please convince
yourself that this is the case). Let Nni be the number of points in partition P part i at the
i
nth iteration of this map. Let us start with a uniform distribution of N = i Nn points on
[0, 1] at time step n = 0. As one can infer from the left part of Fig. 6.8, the dynamics of Nni
is then given as follows:
y y
1 1
N12 I2 F(x) t 21 t 22
1/2
N 11 I1 t 11 t 12
0 1 x 0 1 x
I1 1/2 I2
N 01 N0 2
Figure 6.8: Left: A map that enables a matrix representation for the dynamics of a statistical
ensemble of points; right: Illustration of where the transition matrix elements come from
for this map.
1 1 1 1
N11 = N + N02 = [N + 2N02 ]
2 0 2 0
1 1 1
N12 = N + 0 · N02 = N01 . (6.21)
2 0 2
This can equivalently be written as
1 1
N1 1 1 2 N0
2 = (6.22)
N1 2 1 0 N02
or
1
N n+1 = T N , (6.23)
2 F n
where N n ∈ R2 is a column vector and T F ∈ M2,2 a topological transition matrix (or transfer
matrix) acting on N n .
Let us focus for a moment on the (topological) structure of the matrix T F . There is a simple
recipe of how to obtain this matrix, as follows:
54 6 Dynamics of statistical ensembles
Applied to the above example of map F , this prescription leads directly to the transition
matrix T F of Eq. (6.22). An alternative representation of a transition matrix for a map is
obtained by using the following definition:
Example 29 The Markov graph for the map F of Fig. 6.8 acting on the given Markov
partition is sketched in Fig. 6.9.
t 21
t 11 1 2
t 12
Figure 6.9: Markov graph for the map F on the given Markov partition.
Remark 15
1. What we encounter here is a representation of the action of F in terms of a symbolic
dynamics of a map. The above Markov graph defines it with respect to the given
Markov partition.4
2. Note that our definition is such that tij ∈ N0 . In other literature you will usually find
that tij ∈ {0, 1}.5 Why our definition is a bit more convenient for practical purposes
you will see later on.
4
Other partitions could be used as well. Symbolic dynamics is a fundamental concept in dynamical
systems theory, which would deserve much more detailed explanations. However, since this is again a topic
of the module on Chaos and Fractals we do not further discuss it in our course and refer to the literature
instead [Bec93, All97].
5
see, e.g., work by Ruelle and Bowen
6.2 Markov partitions and transition matrices 55
3. Markov graphs can be constructed without having any transition matrix T F available.
It suffices to just look at how a map F acts onto a Markov partition. If you find it
simpler, you may thus first sketch the Markov graph and then obtain T F from it.
The following definition now makes mathematically precise what we actually mean with a
topological transition matrix:
Example 30 Let us apply this definition to the map F already studied above. According
to Fig. 6.10 F is invertible on I1 , and we obtain
y
I2 F(x)
1/2
I1
0 I21 I22 1 x
I1 1/2 I2
Figure 6.10: Construction of a transition matrix for the map F according to Definition 37.
The proof of this proposition is a bit too lengthy for our lectures, but we outline the basic
idea: First of all, note that for our example of map F considered above the corresponding
Frobenius-Perron operator 1/2 T F actually defines a stochastic matrix, where the sum of all
P
entries of a column is equal to one, M i=1 tij /2 = 1.
6
However, this case is rather special
– consider, for example, map G depicted in Fig. 6.11. Based on the information given in
the figure one can show that the constant slope |G′ (x)| 6∈ N. It is then easily seen that
the corresponding matrix T G /|G′ (x)| cannot be a stochastic one, which refers to the typical
situation.
Nevertheless, this connection with stochastic matrices points in the right direction: If the
conditions of the above proposition are fulfilled there exists a systematic procedure of how to
6
We remark that there is no unique definition of a stochastic matrix in the literature: some authors
define it as above [Kat95], others require that the sum of all row entries is one, instead of the column entries
[Bec93]. Some authors identify a stochastic matrix with a Markov matrix, others distinguish between them
depending on whether the row or the column entries add up to one.
6.2 Markov partitions and transition matrices 57
construct the Frobenius-Perron operator in terms of a transition matrix that is generally not
a topological but a stochastic one, which acts on probabilities defined on Markov partition
parts, see Section 17.5 of Ref. [Bec93] for details. For uniformly constant slope this stochastic
matrix can be simplified to a topological transition matrix divided by the absolute value of
the slope, which acts onto probability density vectors as given in the proposition.
y
1
G(x)
I2
1/2
I1
x
0 1
I1 I2 I3
Figure 6.11: Example of a map G yielding a topological transition matrix that is not directly
related to a stochastic one.
where c = const. Note that all matrix elements tij defining the transition from states j at
time step n to state i at time step n + 1 are constants. That is, they do not depend on
the history of the process at all previous time steps but are defined with respect to the
states i and j at time steps n + 1 and n only, which is the hallmark of a Markov process.
Proposition 5 thus transforms the dynamics generated by the FPE onto a Markov chain
[Bec93]. Note, however, that any transition matrix exhibits a specific topological structure,
which to some extent characterizes a given map. This structure is encoded in the values of
the tij ’s, and summing over all these matrix elements can make the corresponding Markov
chain arbitrarily complicated.7
If one can transform a Frobenius-Perron operator onto a topological transition matrix,
the resulting Frobenius-Perron matrix equation Eq. (6.30) can be used for calculating the
invariant density ρ∗ in form of the following algorithm:
7
Interestingly, the topological structure of the transition matrix for a given map can change drastically
under tiny variation of control parameters. Such a topological instability of a dynamical system produces a
sensitivity of the whole dynamics with respect to control parameters [Kla07].
58 6 Dynamics of statistical ensembles
T F ρτ = τ ρτ , (6.32)
where τ is an eigenvalue and ρτ the corresponding eigenvector. If you did this calcu-
lation correctly, you will find that the largest eigenvalue is always equal to |F ′ (x)|,8
with eigenvector ρτ .
max
3. It follows: Choose
ρn = c ρτ , (6.34)
max
c ρτ = ρ∗ , (6.36)
max
This recipe of how to calculate invariant densities has a rigorous mathematical underpinning:
Example 31 We now apply the above method for calculating invariant measures of piece-
wise linear Markov maps to our previous map F introduced in Example 28:
8
This fact is not trivial but a consequence of the following Proposition 4.
9
A nonnegative matrix that fulfills this property is called eventually positive.
6.2 Markov partitions and transition matrices 59
1. The first step in this algorithm we have already performed before, see Examples 28
and 30. They yielded the Frobenius-Perron matrix equation
1
ρn+1 = T ρ , (6.38)
2 F n
see Eq. (6.29), where
1 2
TF = (6.39)
1 0
and the probability density vectors are defined by
1
ρn
ρn = ∈ R2 , (6.40)
ρ2n
by calculating
det(T F − τ I) = 0 , (6.42)
where
1 0
I= (6.43)
0 1
is the identity matrix. We find
1−τ 2
det = 0, (1 − τ )(−τ ) − 2 = 0
1 −τ
r
2 1 1 1 3
−τ +τ −2= 0⇒τ = ± + 2 = ± = 2 or − 1 . (6.44)
2 4 2 2
We see that indeed τmax = |F ′ (x)|, as stated to the end of step 2 of our algorithm.
3. We now have to calculate the largest eigenvector by solving the equation
1 1
1 2 ρτmax ρτmax
=2 , (6.45)
1 0 ρ2τmax ρ2τmax
which leads to
ρ∗ = cρτ (6.49)
max
with
2 2
X
∗i
X 3
1= ρ · diam(Ii ) = c ρiτmax · diam(Ii ) = c (6.50)
i=1 i=1
4
It is not a bad idea to cross-check whether our solution indeed fulfills the Frobenius-Perron
matrix equation Eq. (6.38),
4 4 2 2
3
1 1 2 3 3
+3
2 = 2 = 2 . X (6.52)
3 2 1 0 3 3
7 Measure-theoretic description of
dynamics
So far we have analyzed the dynamics of statistical ensembles by using probability densities,
which represents rather a physicist’s approach towards a probabilistic theory of dynamical
systems. Mathematicians prefer to deal with measures, which are often more well-behaved
than probability densities.
In this chapter we introduce these quantities, learn about basic properties of them and
eventually briefly elaborate on some ergodic properties of dynamical systems. The first
section particularly draws on Ref. [Las94], the second one on Refs. [Dor99, Arn68]. If you
have problems to associate any meaning with measures you may wish to take a look into
Ref. [All97].
is the probability of finding a point in A. µ(A) is called a measure of A, in the sense that we
‘assign a positive number to the set A’. Let us demonstrate this idea for a simple example:
7.1 Probability measures 61
Example 32 Let I = [0, 1] and A = [0, 12 ] with ρ(x) = 1 – think, for example, of the
Bernoulli shift. Then Z 1
2 1
µ(A) = dx 1 = (7.2)
0 2
is the probability of finding a point in A.
We make this basic idea mathematically precise in terms of two definitions:
1. A ∈ A ⇒ I \ A ∈ A complementarity
S
2. A1 , A2 , . . . , AN ∈ A ⇒ k Ak ∈ A additivity
Example 33 Consider the three intervals I = [0, 1], A1 = [0, 1/2), A2 = [1/2, 1]. Is
A = {A1 , A2 } a σ-algebra? Let us check the two parts of the above definition:
Even for an amended set {A1 , A2 , I} the first condition yields I \ I = ∅, which again is not
an element of this new set. We thus find that A˜ = {A1 , A2 , I, ∅} must be a σ-algebra of
I. It is a general feature of σ-algebras that the whole set I as well as the empty set must
always be elements of the algebra.
Definition 39 [Las94]
A function µ : A → R+ ∪ ∞ defined on a σ-algebra A is called a measure if
1. µ(A) ≥ 0 ∀A ∈ A .
2. µ(∅) = 0.
3. If A1 , A2 , . . . , An ∈ A and Ai ∩ Aj = ∅ (i 6= j) then
[ X
µ( Ak ) = µ(Ak ) σ − additivity . (7.3)
k k
Note the analogy between this definition and Kolmogorov’s axioms of probability theory.
Example 34 Let A˜ = {A1 , A2 , I, ∅} as before. Let µ(A) be the length of the interval A;
for example, if A = [a, b) then µ(A) := b − a. Is this function a measure? Let us check the
above definition step by step:
3. For the third condition we have to check all the different cases:
1 1
A1 ∩ A2 = ∅ ⇒ µ(A1 ∪ A2 ) = µ(I) = 1 = µ(A1 ) + µ(A2 ) = + X
2 2
A1 ∩ I =
6 ∅ X
A2 ∩ I =6 ∅ X
1
A1 ∩ ∅ = ∅ ⇒ µ(A1 ∪ ∅) = µ(A1 ) = = µ(A1 ) + µ(∅) X , (7.4)
2
since µ(∅) = 0. The same holds for A2 ∩ ∅ and I ∩ ∅.
We can conclude that µ(A) is a measure. This specific measure, which is defined by the
length of the interval, is called the Lebesgue measure. From now on we will denote it by µL .
These considerations motivate the following definition:
Remark 17 We emphasize that µL is only one example of a measure. There exist plenty of
other types of measure such as, for example, the Dirac measure discussed in the coursework
[Las94].
Let us get back to our motivation Eq. (7.1). Let ρ(x) be a Rprobability density on I ⊆ R and
let A be a σ-algebra on the same interval. Then µ(A) := A dxρ(x) , A ∈ A , see Eq. (7.1),
defines a probability measure on I. That this holds true can be verified straightforwardly
by applying Definition 39 and using properties of Riemann integration.
Let us now assume that ρ(x) is generated by a map such as, e.g., the Bernoulli shift, see
Fig. 7.1. We have the Frobenius-Perron equation (FPE)
with Frobenius-Perron operator PF for a map F governing the time evolution of the prob-
ability densities ρn (x). The above choice of µ(A) raises the question whether there exists
an equation for the evolution of this measure associated with a density that is analogous to
the FPE.
In order to answer this question we employ again the principle of conservation of probability
as for our derivation of the FPE in Section 6.1. With A = F ◦ F −1 (A) for a one-dimensional
map F , where F −1 (A) defines the preimages of the set A, the number of points xn in F −1 (A)
7.1 Probability measures 63
y
B(x)
1
0 1 x
B−1(A)
Figure 7.1: Construction of the analogue of the FPE for the measure obtained by integrating
a probability density.
must be Requal to the number of points xn+1 in A. This implies that the probability measures
µ(A) = A dxρ(x) on image and preimages must be the same under iteration,
µn+1 (A) = µn (F −1 (A)) . (7.6)
See the Bernoulli shift B(X) = 2x mod 1 in Fig. 7.1 for an illustration. We thus arrive
at the measure-theoretic version of the FPE Eq. (6.7) [Bec93]. Alternatively, this equation
can be obtained by just integrating the FPE.
Recall that the fixed point of the FPE, ρ∗ (x) = PF ρ∗ (x), yielded the invariant density
of the map F , see Definition 31. This leads to the following analogous defnition for the
corresponding probability measure:
Definition 42 [Bec93]
µn+1 (A) = µn (A) =: µ∗ (A) is called an invariant measure, i.e., the measure does not change
with time.
Hence, for our µ-version of the FPE it must hold that µ∗ (A) = µ∗ (F −1 (A)) [All97]. A map
F that fulfills this property is also sometimes called measure preserving in the mathematics
literature [Las94].
We may now highlight another important property of the probability measure µ(A) obtained
from a density in form of the following definition:
Remark 18
1. Note that for the Lebesgue measure in the integral one may also write dx = dµL =
µL(dx). This more general notation is often used in the mathematics literature.
2. Loosely speaking one may say that a measure is absolutely continuous if it is given
by integrating a density [Kat95, Has03]. However, this only implies continuity for the
measure µ if we specify the space of probability density functions on which we are
working. The function space should then exclude densities ρ that are singular such
as, e.g., the δ−function that we have encountered in Section 3.3, see Eq. (3.15). This
is also illustrated in the following examples.
3. Rigorously mathematically speaking, the function space on which we should be work-
ing here is the space of Lebesgue-integrable functions [Las94]: Let us replace the
Lebesgue measure µL (dx) by some other measure ν(dx). Then Lp (I, A , ν) , 1 ≤
Rp < ∞ is thep family of all possible real-valued measurable functions ρ satisfying
ν(dx) |ρ(x)| < ∞, where this integral should be understood as a Lebesgue in-
tegral.2 Note that if ρ ∈ Lp is not Riemann-integrable anymore, it may still be
Lebesgue-integrable, that is, the above integral may still exist.
4. Some authors define a measure µ to be absolutely continuous with respect to another
measure ν if for all measurable sets A ∈ I µ(A) = 0 whenever ν(A) = 0 [dM93, Las94].
If ρ ∈ L1 and ν is σ-finite,3 then one can prove that this definition is equivalent to
Definition 43. This is the contents of the Radon-Nikodym Theorem, and ρ = dµ/dν is
called the Radon-Nikodym derivative of µ with respect to ν [dM93, Las94].
5. If ρ ∈ C 0 and bounded it follows from the fundamental theorem of calculus that µ is
C 1 -smooth and dµ/dx = ρ(x).
6. If ρ(x) is continuous except at finitely many points and bounded, then there still exists
a measure µ according to Eq. (7.7), however, it is not everywhere C 1 anymore.
Let us illustrate the last two points of this remark by two examples:
Example 36
1. Consider the Bernoulli shift B(x) = 2x mod 1 on [0, 1) with ρ∗ (x) = 1. It is easy to
∗
see
R x that∗ there exists a differentiable absolutely continuous invariant measure µ (x) =
0
dx̃ ρ (x̃) = x.
2. Consider the step function ρ∗ (x) sketched
Rx in the left part of Fig. 7.2. Integrating over
this function according to µ∗ (x) = 0 dx̃ρ∗ (x̃) yields the graph depicted in the right
part of this figure, which ist still C 0 but not differentiable at the points of discontinuity
of ρ∗ (x).
Definition 44 [Lev94, Las94]
If the measure µ defined by Eq. (7.7) is not absolutely continuous, it is called singular.4
2
I found the entry on Lebesgue integration in the online Wikipedia quite helpful to get an intuition of
what Lebesgue integration is about compared with ordinary Riemann integration.
3
A measure is σ-finite if it can be written as the countable union of sets with finite measure.
4
One may also say that the measure µ defined by Eq. (7.7) decomposes into an absolutely continuous
one µa and a singular one µs , which is the contents of Lebesgue’s decomposition theorem [Las94].
7.1 Probability measures 65
ρ µ
a3 1
a1
a2
a4
x x
0 x1 x2 x3 1 0 x1 x2 x3 1
Figure 7.2: A step function density and the corresponding absolutely continuous probability
measure obtained by integrating the density.
Example 37 For B(x) let ρ̃∗ (x) := 21 [δ(x − 13 ) + δ(x − 32 )] be the invariant density
R x defined
1 2 ∗
on the period 2 orbit { 3 , 3 } only, see the left part of Fig. 7.3. Then µ̃ (x) = 0 dx̃ ρ∗ (x̃),
see the right hand side of this figure, is singular.
~*
ρ ~*
µ
8
1/2
x x
0 1/3 2/3 1 0 1/3 2/3 1
Figure 7.3: A delta function density yields a singular measure via integration.
Definition 45 [All97]
A measure µ is called atomic5 if all of the measure is contained in a finite or countably
infinite set of points.
Example 38 The measure µ̃∗ of the above example is atomic,6 which in fact is better seen
if we look at the measure µ̃∗ (A) on sets A ⊆ [0, 1] rather than µ̃∗ (x).
5
In measure theory, an atom is a measurable set which has positive measure and contains no ”smaller”
set of positive measure.
6
This measure actually represents a linear combination of the Dirac measure as defined on a coursework
sheet.
66 7 Measure-theoretic description of dynamics
Example 39 [Dor99]
y
1
111111
000000
000000
111111
000000
111111
000000
111111
000000
111111 11111111111
00000000000
000000
111111
000000
111111
1/2
00000000000
11111111111
000000
111111 00000000000
11111111111
00000000000
11111111111
000000
111111 00000000000
11111111111
0 111111
000000
000000
111111 00000000000
11111111111
1 x 00000000000
11111111111
0 1 2
n=1 11111111111
00000000000
00000000000
11111111111
area conservation
00000000000
11111111111
00000000000
11111111111
00000000000
11111111111
00000000000
11111111111
0 11111111111
00000000000
1
11111111111
00000000000
11111111111
00000000000 00000000000
11111111111
00000000000
11111111111 n
00000000000
11111111111 n times
00000000000
11111111111
8
00000000000
11111111111
n=2
00000000000
11111111111 00000000000
11111111111
00000000000
11111111111 00000000000
11111111111
00000000000
11111111111
2n layers 00000000000
11111111111
00000000000
11111111111 00000000000
11111111111
00000000000
11111111111
00000000000
11111111111
Figure 7.4: The two-dimensional area-preserving baker map and its iterative action onto a
set of points initially concentrated on the left half of the unit square.
In Fig. 7.4 we geometrically define the famous baker map, a paradigmatic example
of a two-dimensional area-preserving chaotic dynamical system. This map stretches
the unit square by a factor of two along the x-axis and squeezes it by a factor of 1/2
in the y-direction. The resulting rectangle is cut in the middle, and both halfes are
put on top of each other recovering the unit square. If we colour the left half of the
unit square (or fill this half uniformly with points while leaving the right half empty,
respectively) and iterate this map several times we obtain the layered structure shown
in the figure.
Note that the corresponding density ρn (x, y) is uniform in the x direction, whereas it
is a step function along y. However, in the limit of n → ∞ the strips become thinner
7.1 Probability measures 67
and thinner such that all points eventually distribute uniformly over the whole square,
so there are no “steps” anymore along y. In other words, the invariant density of the
baker map is simply constant everywhere. But if we modify the map a little bit the
situation changes dramatically leading us to our second example.
2. The dissipative baker map:
y
11111111
00000000
1
00000000
11111111
00000000
11111111
00000000
11111111
00000000
11111111
00000000
11111111 n=1 n=2
11111111111
00000000000
00000000
11111111 00000000000
11111111111
00000000
11111111 11111111111
00000000000 00000000000
11111111111
00000000
11111111
00000000
11111111 00000000000
11111111111
0 11111111
00000000 00000000000
11111111111 11111111111
00000000000
00000000
11111111 1 x
00000000000
11111111111
n=3
11111111111
00000000000
00000000000
11111111111 n times
00000000000
11111111111
00000000000
11111111111
Complicated
00000000000
11111111111 ‘Fractal’ structure
11111111111
00000000000
00000000000
11111111111
Figure 7.5: The dissipative baker map eventually generates an invariant measure exhibiting
a fractal structure in the y direction.
In Fig. 7.5 we introduce a little variant of the previous baker transformation, where we
still stretch and squeeze the whole unit square parallel to the x- and to the y-direction,
respectively. However, we do this now such that the big left (coloured) rectangle is
compressed into the little rectangle at the bottom of the second unit square, whereas
the little right (empty) rectangle is expanded into the big empty one on top. If we
repeat this action, again a layered structure evolves which is similar to the one of the
previous figure. One observes that, as before, the density ρn (x, y) is uniform in the
x-direction and that in the limit of n → ∞ the strips piling up along y become thinner
and thinner. However, in contrast to the previous example it can be shown that in
the limit of n → ∞ the invariant density is not uniform anymore on the whole unit
square. Strictly speaking, in this case an invariant density does not exist anymore,
because a fractal structure was generated in the y direction which makes the density
ρn (x, y) non-integrable in the limit of n → ∞. However, there still exists an invariant
measure µ∗ (x, y). This measure turns out to be absolutely continuous along x with
µ∗ (x) ∈ C 1 but it is singular along y with µ∗ (y) ∈ C 0 . See [Dor99] for further details.
Definition 46 SRB measure [You02, Eck85, Bal00]
An invariant probability measure µ∗ is called an SRB measure7 if it is absolutely continuous
7
The acronym SRB holds for the initials of the mathematicians Ya.Sinai, D.Ruelle and R.Bowen who
first introduced this measure.
68 7 Measure-theoretic description of dynamics
Recall that the unstable manifold W u (p) of a source p of a map can be defined as the set of
initial conditions whose orbits are repelled from p, see Definition 19. The same applies to p
being an unstable periodic point.8
Remark 20
1. Some authors define an SRB measure by asking for slightly stronger C 1 -smoothness
along unstable manifolds [Eck85].
Example 40
1. The Bernoulli shift B(x) = 2x mod 1 is unstable ∀x ∈ [0, 1), since with B ′ (x) = 2
it is everywhere expanding (except
R x at x∗ = 1/2).
R x Hence the whole system lives on an
∗
unstable manifold, and µ (x) = 0 dx̃ρ (x̃) = 0 dx̃1 = x is an SRB measure.
2. The baker map of Fig. 7.4 has an SRB measure that is smooth along the x axis, which
iidentifies the direction of all unstable manifolds. It is also smooth in the y direction,
although this defines the stable manifolds.
3. The dissipative baker map of Fig. 7.5 has an SRB measure that is smooth along
the unstable manifolds in the x direction. However, it is singular along the stable
manifolds in the y direction.
Remark 21
1. SRB measures are important, because the property of being smooth along at least
one direction in phase space enables one to calculate ensemble averages of (physical)
observables, see also the following section.
2. An important question is whether an SRB measure exists for a given dynamical system.
These measures are supposed to be quite common, however, their existence is usually
hard to prove. For so-called Anosov systems an important theorem (sometimes called
the SRB theorem [Dor99]) ensures the existence of SRB measures [You02]. The above
two baker maps are examples of systems which have ‘Anosov-like’ properties [Dor99].
8
For further details about the definition of stable and unstable manifolds see [Bec93, Ott93, Eck85].
7.2 Basics of ergodic theory 69
where, if it exists, dµ∗ = ρ∗ (x) dx, is called the ensemble (or space) average of g with respect
to F . Note that g(x) may depend on x, whereas hgi does not.
so calculating the local Ljapunov exponent actually means taking a time average along a
trajectory. If we choose the same observable for the ensemble average Eq. (7.9) we obtain
Z
λe := hln |F (x)|i := dxρ∗ (x) ln |F ′ (x)| .
′
(7.11)
I
Let us check these two definitions for the Bernoulli shift B(x) = 2x mod 1, where B ′ (x) = 2
for almost every x: We have that
n−1
1X
λt (x) = ln 2 = ln 2 (7.12)
n k=0
In other words, time and ensemble average are the same for almost every x,
λt (x) = λe = ln 2 . (7.14)
70 7 Measure-theoretic description of dynamics
1111111111111111
0000000000000000
gi
I
Figure 7.6: Uniform partition of the interval I ⊆ R and a local average gi defined on the
ith partition part of an observable g(x).
In the following we will show that this result is not a coincidence. We do so in two steps:
Step 1: Let us define a uniform partition on the interval I, see Fig. 7.6, with a total
number of N partition parts, where the ith partition part is defined by the subinterval
[xi , xi+1 ]. Take a long trajectory {x0 , x1 , . . . , xn−1 } generated by the map F and let ni
Ii := P
with N i=1 ni = n be the number of occurrences when the trajectory is in partition part i.
Let gi be the local average of g(x) with respect to the ith partition part,
1 X
gi := g(xk ) ∀xk ∈ Ii . (7.15)
ni k
Then
n−1 N N
1X 1X X ni
g(x) = g(xk ) = ni gi = gi . (7.16)
n n i=1 i=1
n
k=0
There is now a famous hypothesis relating the number of occurrences ni in the bin i to the
invariant probability measure µ∗i of the region i. Before we can state it we need the following
definition:
Step 1 summarizes Boltzmann’s intuitive approach to relate time and ensemble averages of
observables. His hypothesis forms the rigorous foundation of the whole theory of statistical
7.2 Basics of ergodic theory 71
mechanics [Tod92]. Step 2 now makes his idea mathematically precise by starting with an
important theorem:
µ∗ -almost everywhere.
This implies that g(x) does not depend on x anymore. That the time average is constant is
sometimes also taken as a definition of ergodicity in the literature [Dor99, Bec93]. In fact,
there is even a third definition, which is based on the following definition:
Example 42
1. The “double tent map” shown in Fig. 7.7 is certainly decomposable. According to
Proposition 6 it is therefore not ergodic.
2. The Bernoulli shift B(x) is indecomposable and hence ergodic. However, a proof of
this property is a bit cumbersome and goes along the lines as the one of ergodicity for
the baker map [Dor99].
I2
1/2
I1
0 1 x
I1 1/2 I2
Figure 7.7: An example of a map that is decomposable and hence not ergodic.
Remark 22 You may have noticed that the properties of a map being topologically tran-
sitive and being ergodic are of a very similar nature. However, note that ergodicity requires
us to use a measure, whereas topological transitivity works without a measure. In fact, for
a continuous map it can be shown under fairly general conditions that ergodicity implies
topological transitivity, whereas the reverse does not hold [Kat95, Rob95].10
Let us now get back to the Ljapunov exponent studied in Example 41. For time average
λt (x) and ensemble average λe we have found that
for the Bernoulli shift. From Definition 49 we have learned that this result must hold
whenever a map F is ergodic. However, this means that in case of ergodicity the Ljapunov
exponent λ becomes a global quantity characterising a given map F for a typical point
x, irrespective of what value we choose for the initial condition, λt (x) = λe = λ. This
observation very much facilitates the calculation of λ, as is demonstrated by the following
example:
Example 43 Let us reconsider our specific map F which has already been introduced
previously, see Example 28. For convenience we display it again in Fig. 7.8.
In Example 31 the invariant density of this map was calculated to
4
∗ 3
, 0 ≤ x < 21
ρ (x) = 2 1 , (7.21)
3
, 2
≤x<1
see Eq. (6.51). This motivates us to calculate the Ljapunov exponent of this map by using
10
For example, the reverse implication does not hold for the Pomeau-Manneville map.
7.2 Basics of ergodic theory 73
F(x)
1/2
0 1 x
1/2
Figure 7.8: A simple map that we have seen before, here used in order to demonstrate the
calculation of Ljapunov exponents via ensemble averages.
Remark 23
1. Note that by restricting ourselves to typical points and considering measures that are
defined with respect to integrating probability densities we sort out periodic orbits,
which are all of Lebesgue measure zero.
2. It is important to observe that a measure being SRB neither implies that it is physical,
nor is the reverse true. For both directions one can construct counterexamples [You02,
Bal00, Eck85]. For example, our double tent map considered above has an SRB
74 7 Measure-theoretic description of dynamics
measure but not an ergodic one. On the other hand, the map C(x) = x/2 contracting
onto a fixed point has trivially an ergodic measure on the attracting fixed point which,
however, is not SRB.
3. Some authors define an SRB measure via ergodicity, so the definitions of SRB and
physical measure are not unique in the literature.
5. The SRB theorem mentioned before proves that for Anosov systems the SRB measure
is equal to the physical one [You02], which identifies an important class of dynamical
systems where both measures are identical.
6. You may wonder why physical measures are important at all. This is for the following
reason: Pick your favourite (physical) system such as, for example, the driven pen-
dulum that we have encountered at the very beginning of these lectures. You may
wish to characterize such a system in a statistical sense by calculating your favourite
observable in terms of time or ensemble averages. You would hope that your averages
exist (i.e., that your computations converge to specific values in the limit of long times
or of large ensembles of points), and also that your results do not depend on the ini-
tial conditions you choose. In mathematical terms, you would thus expect that there
exists a “nice” measure which guarantees that your averages exist and are the same.
If your system exhibits a physical measure, it is the one you will typically “see” by
measuring your observable. And being SRB guarantees that you can calculate your
observables by projection along the unstable manifolds.
If you are further interested, I would highly recommend that you make yourself familiar
with what is called mixing in ergodic theory, see, e.g., [Arn68, Dor99].
Bibliography
[All97] K.T. Alligood, T.S. Sauer and J.A. Yorke, Chaos - An introduction to dynamical
systems. (Springer, New York, 1997).
[Arn68] V.I. Arnold and A. Avez, Ergodic problems of classical mechanics. (W.A. Benjamin,
New York, 1968).
[Eck85] J.-P. Eckmann and D. Ruelle, Ergodic theory of chaos and strange attractors. Rev.
Mod. Phys. 57, 617–656 (1985).
[Elm98] F.-J. Elmer. The Pendulum Lab. University of Basel, 1998. Online simulation at
http://monet.unibas.ch/∼elmer/pendulum.
[Kat95] A. Katok and B. Hasselblatt, Introduction to the modern theory of dynamical sys-
tems, volume 54 of Encyclopedia of Mathematics and its Applications. (Cambridge
University Press, Cambridge, 1995).
[Las94] A. Lasota and M.C. Mackay, Chaos, Fractals, and Noise. (Springer-Verlag, 1994).
[Lev94] R.W. Leven, B.-P. Koch and B. Pompe, Chaos in dissipativen Systemen. (Akademie
Verlag, Berlin, second edition, 1994).
[Smi07] L. Smith, Chaos - A very short introduction. (Oxford University Press, Oxford,
2007).
[Tél06] T. Tél and M. Gruiz, Chaotic Dynamics. (Cambridge University Press, Cambridge,
2006).
[Tod92] M. Toda, R. Kubo and N. Saitô, Statistical Physics, volume 1 of Solid State Sci-
ences. (Springer, Berlin, 2nd edition, 1992).
[You02] L.-S. Young, What are SRB measures, and which dynamical systems have them?
J. Stat. Phys. 108, 733–754 (2002).