Mira

Download as pdf or txt
Download as pdf or txt
You are on page 1of 426

Measure, Integration

& Real Analysis

preliminary edition
1 November 2019

Sheldon Axler

xiv Contents

Z Z 1/p Z 1/p0
p p0
| f g| dµ ≤ | f | dµ | g| dµ

©Sheldon Axler
Dedicated to

Paul Halmos, Don Sarason, and Allen Shields,

the three mathematicians who most


helped me become a mathematician.
About the Author

Sheldon Axler was valedictorian of his high school in Miami, Florida. He received his
AB from Princeton University with highest honors, followed by a PhD in Mathematics
from the University of California at Berkeley.
As a postdoctoral Moore Instructor at MIT, Axler received a university-wide
teaching award. He was then an assistant professor, associate professor, and professor
at Michigan State University, where he received the first J. Sutherland Frame Teaching
Award and the Distinguished Faculty Award.
Axler received the Lester R. Ford Award for expository writing from the Mathe-
matical Association of America in 1996. In addition to publishing numerous research
papers, he is the author of six mathematics textbooks, ranging from freshman to
graduate level. His book Linear Algebra Done Right has been adopted as a textbook
at over 300 universities and colleges.
Axler has served as Editor-in-Chief of the Mathematical Intelligencer and As-
sociate Editor of the American Mathematical Monthly. He has been a member of
the Council of the American Mathematical Society and a member of the Board of
Trustees of the Mathematical Sciences Research Institute. He has also served on the
editorial board of Springer’s series Undergraduate Texts in Mathematics, Graduate
Texts in Mathematics, Universitext, and Springer Monographs in Mathematics.
He has been honored by appointments as a Fellow of the American Mathematical
Society and as a Senior Fellow of the California Council on Science and Technology.
Axler joined San Francisco State University as Chair of the Mathematics Depart-
ment in 1997. In 2002, he became Dean of the College of Science & Engineering at
San Francisco State University. After serving as Dean for thirteen years, he returned
to a regular faculty appointment as a professor in the Mathematics Department.

Cover figure: Hölder’s Inequality, which is proved in Section 7A.

iv
Contents

About the Author iv

Preface for Students xi

Preface for Instructors xii

Acknowledgments xvi

1 Riemann Integration 1
1A Review: Riemann Integral 2
Exercises 1A 7

1B Riemann Integral Is Not Good Enough 9


Exercises 1B 12

2 Measures 13
2A Outer Measure on R 14
Motivation and Definition of Outer Measure 14
Good Properties of Outer Measure 15
Outer Measure of Closed Bounded Interval 18
Outer Measure is Not Additive 21
Exercises 2A 23

2B Measurable Spaces and Functions 25


σ-Algebras 26
Borel Subsets of R 28
Inverse Images 29
Measurable Functions 31
Exercises 2B 38

2C Measures and Their Properties 41


Definition and Examples of Measures 41
Properties of Measures 42
Exercises 2C 45
v
vi Contents

2D Lebesgue Measure 47
Additivity of Outer Measure on Borel Sets 47
Lebesgue Measurable Sets 52
Cantor Set and Cantor Function 55
Exercises 2D 60

2E Convergence of Measurable Functions 62


Pointwise and Uniform Convergence 62
Egorov’s Theorem 63
Approximation by Simple Functions 65
Luzin’s Theorem 66
Lebesgue Measurable Functions 69
Exercises 2E 71

3 Integration 73
3A Integration with Respect to a Measure 74
Integration of Nonnegative Functions 74
Monotone Convergence Theorem 77
Integration of Real-Valued Functions 81
Exercises 3A 84

3B Limits of Integrals & Integrals of Limits 88


Bounded Convergence Theorem 88
Sets of Measure 0 in Integration Theorems 89
Dominated Convergence Theorem 90
Riemann Integrals and Lebesgue Integrals 93
Approximation by Nice Functions 95
Exercises 3B 99

4 Differentiation 101
4A Hardy–Littlewood Maximal Function 102
Markov’s Inequality 102
Vitali Covering Lemma 103
Hardy–Littlewood Maximal Inequality 104
Exercises 4A 106

4B Derivatives of Integrals 108


Lebesgue Differentiation Theorem 108
Derivatives 110
Density 112
Exercises 4B 115
Contents vii

5 Product Measures 116


5A Products of Measure Spaces 117
Products of σ-Algebras 117
Monotone Class Theorem 120
Products of Measures 123
Exercises 5A 128

5B Iterated Integrals 129


Tonelli’s Theorem 129
Fubini’s Theorem 131
Area Under Graph 133
Exercises 5B 135

5C Lebesgue Integration on Rn 136


Borel Subsets of Rn 136
Lebesgue Measure on Rn 139
Volume of Unit Ball in Rn 140
Equality of Mixed Partial Derivatives Via Fubini’s Theorem 142
Exercises 5C 144

6 Banach Spaces 146


6A Metric Spaces 147
Open Sets, Closed Sets, and Continuity 147
Cauchy Sequences and Completeness 151
Exercises 6A 153

6B Vector Spaces 155


Integration of Complex-Valued Functions 155
Vector Spaces and Subspaces 159
Exercises 6B 162

6C Normed Vector Spaces 163


Norms and Complete Norms 163
Bounded Linear Maps 167
Exercises 6C 170

6D Linear Functionals 172


Bounded Linear Functionals 172
Discontinuous Linear Functionals 174
Hahn–Banach Theorem 177
Exercises 6D 181
viii Contents

6E Consequences of Baire’s Theorem 184


Baire’s Theorem 184
Open Mapping Theorem and Inverse Mapping Theorem 186
Closed Graph Theorem 188
Principle of Uniform Boundedness 189
Exercises 6E 190

7 L p Spaces 193
7A L p (µ) 194
Hölder’s Inequality 194
Minkowski’s Inequality 198
Exercises 7A 199

7B L p (µ) 202
Definition of L p (µ) 202
L p (µ) Is a Banach Space 204
Duality 206
Exercises 7B 208

8 Hilbert Spaces 211


8A Inner Product Spaces 212
Inner Products 212
Cauchy–Schwarz Inequality and Triangle Inequality 214
Exercises 8A 221

8B Orthogonality 224
Orthogonal Projections 224
Orthogonal Complements 229
Riesz Representation Theorem 233
Exercises 8B 234

8C Orthonormal Bases 237


Bessel’s Inequality 237
Parseval’s Identity 243
Gram–Schmidt Process and Existence of Orthonormal Bases 245
Riesz Representation Theorem, Revisited 250
Exercises 8C 251
Contents ix

9 Real and Complex Measures 255


9A Total Variation 256
Properties of Real and Complex Measures 256
Total Variation Measure 259
The Banach Space of Measures 262
Exercises 9A 265

9B Decomposition Theorems 267


Hahn Decomposition Theorem 267
Jordan Decomposition Theorem 268
Lebesgue Decomposition Theorem 270
Radon–Nikodym Theorem 272
Dual Space of L p (µ) 275
Exercises 9B 278

10 Linear Maps on Hilbert Spaces 280


10A Adjoints and Invertibility 281
Adjoints of Linear Maps on Hilbert Spaces 281
Null Spaces and Ranges in Terms of Adjoints 285
Invertibility of Operators 286
Exercises 10A 292

10B Spectrum 294


Spectrum of an Operator 294
Self-adjoint Operators 299
Normal Operators 302
Isometries and Unitary Operators 305
Exercises 10B 309

10C Compact Operators 312


The Ideal of Compact Operators 312
Spectrum of Compact Operator and Fredholm Alternative 316
Exercises 10C 323

10D Spectral Theorem for Compact Operators 326


Orthonormal Bases Consisting of Eigenvectors 326
Singular Value Decomposition 332
Exercises 10D 336
x Contents

11 Fourier Analysis 339


11A Fourier Series and Poisson Integral 340
Fourier Coefficients and Riemann–Lebesgue Lemma 340
Poisson Kernel 344
Solution to Dirichlet Problem on Disk 348
Fourier Series of Smooth Functions 350
Exercises 11A 352

11B Fourier Series and L p of Unit Circle 355


Orthonormal Basis forL2 of Unit Circle 355
Convolution on Unit Circle 357
Exercises 11B 361

11C Fourier Transform 363


Fourier Transform on L1 (R) 363
Convolution on R 368
Poisson Kernel on Upper Half-Plane 370
Fourier Inversion Formula 374
Extending Fourier Transform to L2 (R) 375
Exercises 11C 377

12 Probability Measures 380


Probability Spaces 381
Independent Events and Independent Random Variables 383
Variance and Standard Deviation 388
Conditional Probability and Bayes’ Theorem 390
Distribution and Density Functions of Random Variables 392
Weak Law of Large Numbers 396
Exercises 12 398

Photo Credits 400

Bibliography 402

Notation Index 403

Index 406

Colophon: Notes on Typesetting 411


Preface for Students

You are about to immerse yourself in serious mathematics, with an emphasis on


attaining a deep understanding of the definitions, theorems, and proofs related to
measure, integration, and real analysis. This book aims to guide you to the wonders
of this subject.
You cannot read mathematics the way you read a novel. If you zip through a page
in less than an hour, you are probably going too fast. When you encounter the phrase
as you should verify, you should indeed do the verification, which will usually require
some writing on your part. When steps are left out, you need to supply the missing
pieces. You should ponder and internalize each definition. For each theorem, you
should seek examples to show why each hypothesis is necessary.
Working on the exercises should be your main mode of learning after you have
read a section. Discussions and joint work with other students may be especially
effective. Active learning promotes long-term understanding much better than passive
learning. Thus you will benefit considerably from struggling with an exercise and
eventually coming up with a solution, perhaps working with other students. Finding
and reading a solution on the internet will likely lead to little learning.
As a visual aid, throughout this book definitions are in yellow boxes and theorems
are in blue boxes, in both print and electronic versions. Each theorem has an informal
descriptive name. The electronic version of this manuscript has links in blue.
Please check the website below (or the Springer website) for additional information
about the book. These websites link to the electronic version of this book, which is
free to the world because this book has been published under Springer’s Open Access
program. Your suggestions for improvements and corrections for a future edition are
most welcome (send to the email address below).
The prerequisite for using this book includes a good understanding of elementary
undergraduate real analysis. You can download from the website below or from the
Springer website the document titled Supplement for Measure, Integration & Real
Analysis. That supplement can serve as a review of the elementary undergraduate real
analysis used in this book.
Best wishes for success and enjoyment in learning measure, integration, and real
analysis!

Sheldon Axler
Mathematics Department
San Francisco State University
San Francisco, CA 94132, USA
website: http://measure.axler.net
e-mail: [email protected]
Twitter: @AxlerLinear

xi
Preface for Instructors

You are about to teach a course, or possibly a two-semester sequence of courses, on


measure, integration, and real analysis. In this textbook, I have tried to use a gentle
approach to serious mathematics, with an emphasis on students attaining a deep
understanding. Thus new material often appears in a comfortable context instead
of the most general setting. For example, the Fourier transform in Chapter 11 is
introduced in the setting of R rather than Rn so that students can focus on the main
ideas without the clutter of the extra bookkeeping needed for working in Rn .
The basic prerequisite for your students to use this textbook is a good understand-
ing of elementary undergraduate real analysis. Your students can download from the
book’s website (http://measure.axler.net) or from the Springer website the document
titled Supplement for Measure, Integration & Real Analysis. That supplement can
serve as a review of the elementary undergraduate real analysis used in this book.
As a visual aid, throughout this book definitions are in yellow boxes and theorems
are in blue boxes, in both print and electronic versions. Each theorem has an informal
descriptive name. The electronic version of this manuscript has links in blue.
Mathematics can be learned only by doing. Fortunately, real analysis has many
good homework exercises. When teaching this course, during each class I usually
assign as homework several of the exercises, due the next class. I grade only one
exercise per homework set, but the students do not know ahead of time which one. I
encourage my students to work together on the homework or to come to me for help.
However, I tell them that getting solutions from the internet is not allowed and would
be counterproductive for their learning goals.
If you go at a leisurely pace, then covering Chapters 1–5 in the first semester may
be a good goal. If you go a bit faster, then covering Chapters 1–6 in the first semester
may be more appropriate. For a second-semester course, covering some subset of
Chapters 6 through 12 should produce a good course. Most instructors will not have
time to cover all those chapters in a second semester; thus some choices need to
be made. The following chapter-by-chapter summary of the highlights of the book
should help you decide what to cover and in what order:

• Chapter 1: This short chapter begins with a brief review of Riemann integration.
Then a discussion of the deficiencies of the Riemann integral helps motivate the
need for a better theory of integration.

• Chapter 2: This chapter begins by defining outer measure on R as a natural


extension of the length function on intervals. After verifying some nice properties
of outer measure, we see that it is not additive. This observation leads to restricting
our attention to the σ-algebra of Borel sets, defined as the smallest σ-algebra on R
containing all the open sets. This path leads us to measures.

xii
Preface for Instructors xiii

After dealing with the properties of general measures, we come back to the setting
of R, showing that outer measure restricted to the σ-algebra of Borel sets is
countably additive and thus is a measure. Then a subset of R is defined to be
Lebesgue measurable if it differs from a Borel set by a set of outer measure 0. This
definition makes Lebesgue measurable sets seem more natural to students than the
other competing equivalent definitions. The Cantor set and the Cantor function
then stretch students’ intuition.
Egorov’s Theorem, which states that pointwise convergence of a sequence of
measurable functions is close to uniform convergence, has multiple applications in
later chapters. Luzin’s Theorem, back in the context of R, sounds spectacular but
has no other uses in this book and thus can be skipped if you are pressed for time.
• Chapter 3: Integration with respect to a measure is defined in this chapter in a
natural fashion first for nonnegative measurable functions, and then for real-valued
measurable functions. The Monotone Convergence Theorem and the Dominated
Convergence Theorem are the big results in this chapter that allow us to interchange
integrals and limits under appropriate conditions.
• Chapter 4: The highlight of this chapter is the Lebesgue Differentiation Theorem,
which allows us to differentiate an integral. The main tool used to prove this
result cleanly is the Hardy–Littlewood maximal inequality, which is interesting
and important in its own right. This chapter also includes the Lebesgue Density
Theorem, showing that a Lebesgue measurable subset of R has density 1 at almost
every number in the set and density 0 at almost every number not in the set.
• Chapter 5: This chapter deals with product measures. The most important results
here are Tonelli’s Theorem and Fubini’s Theorem, which allow us to evaluate
integrals with respect to product measures as iterated integrals and allow us to
change the order of integration under appropriate conditions. As an application of
product measures, we get Lebesgue measure on Rn from Lebesgue measure on R.
To give students practice with using these concepts, this chapter finds a formula for
the volume of the unit ball in Rn . The chapter closes by using Fubini’s Theorem to
give a simple proof that a mixed partial derivative with sufficient continuity does
not depend upon the order of differentiation.
• Chapter 6: After a quick review of metric spaces and vector spaces, this chapter
defines normed vector spaces. The big result here is the Hahn–Banach Theorem
about extending bounded linear functionals from a subspace to the whole space.
Then this chapter introduces Banach spaces. We see that completeness plays
a major role in the key theorems: Open Mapping Theorem, Inverse Mapping
Theorem, Closed Graph Theorem, and Principle of Uniform Boundedness.
• Chapter 7: This chapter introduces the important class of Banach spaces L p (µ),
where 1 ≤ p ≤ ∞ and µ is a measure, giving students additional opportunities to
use results from earlier chapters about measure and integration theory. The crucial
results called Hölder’s inequality and Minkowski’s inequality are key tools here.
0
This chapter also shows that the dual of ` p is ` p for 1 ≤ p < ∞.
Chapters 1 through 7 should be covered in order, before any of the later chapters.
After Chapter 7, you can cover Chapter 8 or Chapter 12.
xiv Preface for Instructors

• Chapter 8: This chapter focuses on Hilbert spaces, which play a central role in
modern mathematics. After proving the Cauchy–Schwarz inequality and the Riesz
Representation Theorem that describes the bounded linear functionals on a Hilbert
space, this chapter deals with orthonormal bases. Key results here include Bessel’s
inequality, Parseval’s identity, and the Gram–Schmidt process.

• Chapter 9: Only positive measures have been discussed in the book up until this
chapter. In this chapter, real and complex measures get consideration. These con-
cepts lead to the Banach space of measures, with total variation as the norm. Key
results that help describe real and complex measures are the Hahn Decomposition
Theorem, the Jordan Decomposition Theorem, and the Lebesgue Decomposition
Theorem. The Radon–Nikodym Theorem is proved using von Neumann’s slick
Hilbert space trick. Then the Radon–Nikodym Theorem is used to prove that the
0
dual of L p (µ) can be identified with L p (µ) for 1 < p < ∞ and µ a (positive)
measure, completing a project that started in Chapter 7.
The material in Chapter 9 is not used later in the book. Thus this chapter can be
skipped or covered after one of the later chapters.
• Chapter 10: This chapter begins by discussing the adjoint of a bounded linear
map between Hilbert spaces. Then the rest of the chapter presents key results
about bounded linear operators from a Hilbert space to itself. The proof that each
bounded operator on a complex nonzero Hilbert space has a nonempty spectrum
requires a tiny bit of knowledge about analytic functions. Properties of special
classes of operators (self-adjoint operators, normal operators, isometries, and
unitary operators) are described.
Then this chapter delves deeper into compact operators, proving the Fredholm
Alternative. The chapter concludes with two major results: the Spectral Theorem
for compact operators and the popular Singular Value Decomposition for compact
operators. Throughout this chapter, the Volterra operator is used as an example to
illustrate the main results.
Some instructors may prefer to cover Chapter 10 immediately after Chapter 8,
because both chapters live in the context of Hilbert space. I chose the current order
to give students a breather between the two Hilbert space chapters, thinking that
being away from Hilbert space for a little while and then coming back to it might
strengthen students’ understanding and provide some variety. However, covering
the two Hilbert space chapters consecutively would also work fine.
• Chapter 11: Fourier analysis is a huge subject with a two-hundred year history.
This chapter gives a gentle but modern introduction to Fourier series and the
Fourier transform.
This chapter first develops results in the context of Fourier series, but then comes
back later and develops parallel concepts in the context of the Fourier transform.
For example, the Fourier coefficient version of the Riemann–Lebesgue Lemma is
proved early in the chapter, with the Fourier transform version proved later in the
chapter. Other examples include the Poisson kernel, convolution, and the Dirichlet
problem, all of which are first covered in the context of the unit disk and unit circle;
then these topics are revisited later in the context of the half-plane and real line.
Preface for Instructors xv

Convergence of Fourier series is proved in the L2 norm and also (for sufficiently
smooth functions) pointwise. The book emphasizes getting students to work with
the main ideas rather than on proving all possible results (for example, pointwise
convergence of Fourier series is proved only for twice continuously differentiable
functions rather than using a weaker hypothesis).
The proof of the Fourier Inversion Formula is the highlight of the material on the
Fourier transform. The Fourier Inversion Formula is then used to show that the
Fourier transform extends to a unitary operator on L2 (R).
This chapter uses some basic results about Hilbert spaces, so it should not be
covered before Chapter 8. However, if you are willing to skip or hand-wave
through one result that helps describe the Fourier transform as an operator on
L2 (R) (see 11.87), then you could cover this chapter without doing Chapter 10.
• Chapter 12: A thorough coverage of probability theory would require a whole
book instead of a single chapter. This chapter takes advantage of the book’s earlier
development of measure theory to present the basic language and emphasis of
probability theory. For students not pursuing further studies in probability theory,
this chapter gives them a good taste of the subject. Students who go on to learn
more probability theory should benefit from the head start provided by this chapter
and the background of measure theory.
Features that distinguish probability theory from measure theory include the
notions of independent events and independent random variables. In addition to
those concepts, this chapter discusses standard deviation, conditional probabilities,
Bayes’ Theorem, and distribution functions. The chapter concludes with a proof of
the Weak Law of Large Numbers for independent identically distributed random
variables.
You could cover this chapter anytime after Chapter 7.

Please check the website below (or the Springer website) for additional information
about the book. These websites link to the electronic version of this book, which is
free to the world because this book has been published under Springer’s Open Access
program. Your suggestions for improvements and corrections for a future edition are
most welcome (send to the email address below).
I enjoy keeping track of where my books are used as textbooks. If you use this
book as the textbook for a course, please let me know.
Best wishes for teaching a successful class on measure, integration, and real
analysis!

Sheldon Axler Contact the author, or Springer if the


Mathematics Department author is not available, for permission
San Francisco State University for translations or other commercial
San Francisco, CA 94132, USA re-use of the contents of this book.
website: http://measure.axler.net
e-mail: [email protected]
Twitter: @AxlerLinear
Acknowledgments

I owe a huge intellectual debt to the many mathematicians who created real analysis
over the past several centuries. The results in this book belong to the common heritage
of mathematics. A special case of a theorem may first have been proved by one
mathematician and then sharpened and improved by many other mathematicians.
Bestowing accurate credit for all the contributions would be a difficult task that I have
not undertaken. In no case should the reader assume that any theorem presented here
represents my original contribution. However, in writing this book I tried to think
about the best way to present real analysis and to prove its theorems, without regard
to the standard methods and proofs used in most textbooks.
The manuscript for this book received an unusually large amount of class testing
at several universities before publication. Thus I received many valuable suggestions
for improvements and corrections. I am deeply grateful to all the faculty and students
who helped class test the manuscript. I implemented suggestions or corrections from
the following faculty and students, all of whom helped make this a better book:

Sunayan Acharya, Ali Al Setri, Nick Anderson, Kevin Bui, Tony Cairatt,
Eric Carmody, Timmy Chan, Logan Clark, Sam Coskey, Yerbolat
Dauletyarov, Evelyn Easdale, Ben French, Loukas Grafakos, Michael
Hanson, Michah Hawkins, Eric Hayashi, Nelson Huang, Jasim Ismaeel,
Brody Johnson, Hannah Knight, Oliver Knitter, Chun-Kit Lai, Lee Larson,
Vens Lee, Hua Lin, David Livnat, Shi Hao Looi, Dante Luber, Stephanie
Magallanes, Jan Mandel, Juan Manfredi, Zack Mayfield, Calib Nastasi,
Lamson Nguyen, Kiyoshi Okada, Célio Passos, Isabel Perez, Ricardo Pires,
Hal Prince, Noah Rhee, Ken Ribet, Spenser Rook, Arnab Dey Sarkar,
Wayne Small, Emily Smith, Keith Taylor, Ignacio Uriarte-Tuero, Alexander
Wittmond, Run Yan, Edward Zeng.

Loretta Bartolini, Mathematics Editor at Springer, is owed huge thanks for her
multiple major contributions to this project. Paula Francis, the copy editor for this
book, provided numerous useful suggestions and corrections that improved the book.
Thanks also to the people who allowed photographs they produced to enhance this
book. For a complete list, see the Photo Credits pages near the end of the book.
Special thanks to my wonderful partner Carrie Heeter, whose understanding and
encouragement enabled me to work intensely on this book. Our cat Moon, whose
picture is on page 44, helped provide relaxing breaks.

Sheldon Axler

xvi
Chapter 1
Riemann Integration

This brief chapter reviews Riemann integration. Riemann integration uses rectangles
to approximate areas under graphs. This chapter begins by carefully presenting
the definitions leading to the Riemann integral. The big result in the first section
states that a continuous real-valued function on a closed bounded interval is Riemann
integrable. The proof depends upon the theorem that continuous functions on closed
bounded intervals are uniformly continuous.
The second section of this chapter focuses on several deficiencies of Riemann
integration. As we will see, Riemann integration does not do everything we would
like an integral to do. These deficiencies provide motivation in future chapters for the
development of measures and integration with respect to measures.

Digital sculpture of Bernhard Riemann (1826–1866),


whose method of integration is taught in calculus courses.
©Doris Fiebig

1
2 Chapter 1 Riemann Integration

1A Review: Riemann Integral


We begin with a few definitions needed before we can define the Riemann integral.
Let R denote the complete ordered field of real numbers.

1.1 Definition partition

Suppose a, b ∈ R with a < b. A partition of [ a, b] is a finite list of the form


x0 , x1 , . . . , xn , where

a = x0 < x1 < · · · < xn = b.

We use a partition x0 , x1 , . . . , xn of [ a, b] to think of [ a, b] as a union of closed


subintervals, as follows:

[ a, b] = [ x0 , x1 ] ∪ [ x1 , x2 ] ∪ · · · ∪ [ xn−1 , xn ].

The next definition introduces clean notation for the infimum and supremum of
the values of a function on some subset of its domain.

1.2 Definition notation for infimum and supremum of a function

If f is a real-valued function and A is a subset of the domain of f , then

inf f = inf{ f ( x ) : x ∈ A} and sup f = sup{ f ( x ) : x ∈ A}.


A A

The lower and upper Riemann sums, which we now define, approximate the
area under the graph of a nonnegative function (or, more generally, the signed area
corresponding to a real-valued function).

1.3 Definition lower and upper Riemann sums

Suppose f : [ a, b] → R is a bounded function and P is a partition x0 , . . . , xn


of [ a, b]. The lower Riemann sum L( f , P, [ a, b]) and the upper Riemann sum
U ( f , P, [ a, b]) are defined by
n
L( f , P, [ a, b]) = ∑ (x j − x j−1 ) [x inf, x ] f
j =1 j −1 j

and
n
U ( f , P, [ a, b]) = ∑ ( x j − x j −1 ) sup f .
j =1 [ x j −1 , x j ]

Our intuition suggests that for a partition with only a small gap between consecu-
tive points, the lower Riemann sum should be a bit less than the area under the graph,
and the upper Riemann sum should be a bit more than the area under the graph.
Section 1A Review: Riemann Integral 3

The pictures in the next example help convey the idea of these approximations.
The base of the jth rectangle has length x j − x j−1 and has height inf f for the
[ x j −1 , x j ]
lower Riemann sum and height sup f for the upper Riemann sum.
[ x j −1 , x j ]

1.4 Example lower and upper Riemann sums


Define f : [0, 1] → R by f ( x ) = x2 . Let Pn denote the partition 0, n1 , n2 , . . . , 1
of [0, 1].

The two figures here show


the graph of f in red. The
infimum of this function f
is attained at the left end-
point of each subinterval
[ j−n 1 , nj ]; the supremum is
attained at the right end-
point.

L( x2 , P16 , [0, 1]) is the U ( x2 , P16 , [0, 1]) is the


sum of the areas of these sum of the areas of these
rectangles. rectangles.
1
For the partition Pn , we have x j − x j−1 = n for each j = 1, . . . , n. Thus
n
1 ( j − 1)2 2n2 − 3n + 1
L( x2 , Pn , [0, 1]) = ∑ =
n j =1
n2 6n2

and
n
1 j2 2n2 + 3n + 1
U ( x2 , Pn , [0, 1]) = ∑ n2 = ,
n j =1
6n2

n(2n2 +3n+1)
as you should verify [use the formula 1 + 4 + 9 + · · · + n2 = 6 ].

The next result states that adjoining more points to a partition increases the lower
Riemann sum and decreases the upper Riemann sum.

1.5 inequalities with Riemann sums

Suppose f : [ a, b] → R is a bounded function and P, P0 are partitions of [ a, b]


such that the list defining P is a sublist of the list defining P0 . Then

L( f , P, [ a, b]) ≤ L( f , P0 , [ a, b]) ≤ U ( f , P0 , [ a, b]) ≤ U ( f , P, [ a, b]).

Proof To prove the first inequality, suppose P is the partition x0 , . . . , xn and P0 is the
partition x00 , . . . , x 0N of [ a, b]. For each j = 1, . . . , n, there exist k ∈ {0, . . . , N − 1}
and a positive integer m such that x j−1 = xk0 < xk0 +1 < · · · < xk0 +m = x j . We have
4 Chapter 1 Riemann Integration

m
( x j − x j−1 ) inf
[ x j −1 , x j ]
f = ∑ (xk0 +i − xk0 +i−1 ) [x inf, x ] f
i =1 j −1 j

m
≤ ∑ (xk0 +i − xk0 +i−1 ) [x0 inf
0
f.
i =1 k + i −1 , x k + i ]

The inequality above implies that L( f , P, [ a, b]) ≤ L( f , P0 , [ a, b]).


The middle inequality in this result follows from the observation that the infimum
of each set of real numbers is less than or equal to the supremum of that set.
The proof of the last inequality in this result is similar to the proof of the first
inequality and is left to the reader.

The following result states that if the function is fixed, then each lower Riemann
sum is less than or equal to each upper Riemann sum.

1.6 lower Riemann sums ≤ upper Riemann sums

Suppose f : [ a, b] → R is a bounded function and P, P0 are partitions of [ a, b].


Then
L( f , P, [ a, b]) ≤ U ( f , P0 , [ a, b]).

Proof Let P00 be the partition of [ a, b] obtained by merging the lists that define P
and P0 . Then

L( f , P, [ a, b]) ≤ L( f , P00 , [ a, b])


≤ U ( f , P00 , [ a, b])
≤ U ( f , P0 , [ a, b]),

where all three inequalities above come from 1.5.

We have been working with lower and upper Riemann sums. Now we define the
lower and upper Riemann integrals.

1.7 Definition lower and upper Riemann integrals

Suppose f : [ a, b] → R is a bounded function. The lower Riemann integral


L( f , [ a, b]) and the upper Riemann integral U ( f , [ a, b]) of f are defined by

L( f , [ a, b]) = sup L( f , P, [ a, b])


P

and
U ( f , [ a, b]) = inf U ( f , P, [ a, b]),
P

where the supremum and infimum above are taken over all partitions P of [ a, b].
Section 1A Review: Riemann Integral 5

In the definition above, we take the supremum (over all partitions) of the lower
Riemann sums because adjoining more points to a partition increases the lower
Riemann sum (by 1.5) and should provide a more accurate estimate of the area under
the graph. Similarly, in the definition above, we take the infimum (over all partitions)
of the upper Riemann sums because adjoining more points to a partition decreases
the upper Riemann sum (by 1.5) and should provide a more accurate estimate of the
area under the graph.
Our first result about the lower and upper Riemann integrals is an easy inequality.

1.8 lower Riemann integral ≤ upper Riemann integral

Suppose f : [ a, b] → R is a bounded function. Then

L( f , [ a, b]) ≤ U ( f , [ a, b]).

Proof The desired inequality follows from the definitions and 1.6.

The lower Riemann integral and the upper Riemann integral can both be reasonably
considered to be the area under the graph of a function. Which one should we use?
The pictures in Example 1.4 suggest that these two quantities are the same for the
function in that example; we will soon verify this suspicion. However, as we will see
in the next section, there are functions for which the lower Riemann integral does not
equal the upper Riemann integral.
Instead of choosing between the lower Riemann integral and the upper Riemann
integral, the standard procedure in Riemann integration is to consider only functions
for which those two quantities are equal. This decision has the huge advantage of
making the Riemann integral behave as we wish with respect to the sum of two
functions (see Exercise 4 in this section).

1.9 Definition Riemann integrable; Riemann integral

• A bounded function on a closed bounded interval is called Riemann


integrable if its lower Riemann integral equals its upper Riemann integral.
Z b
• If f : [ a, b] → R is Riemann integrable, then the Riemann integral f is
a
defined by
Z b
f = L( f , [ a, b]) = U ( f , [ a, b]).
a

Let Z denote the set of integers and Z+ denote the set of positive integers.

1.10 Example computing a Riemann integral


Define f : [0, 1] → R by f ( x ) = x2 . Then

2n2 + 3n + 1 1 2n2 − 3n + 1
U ( f , [0, 1]) ≤ inf = = sup ≤ L( f , [0, 1]),
n∈Z+ 6n2 3 n∈Z+ 6n2
6 Chapter 1 Riemann Integration

where the two inequalities above come from Example 1.4 and the two equalities
easily follow from dividing the numerators and denominators of both fractions above
by n2 .
The paragraph above shows that
Our definition of Riemann
U ( f , [0, 1]) ≤ 13 ≤ L( f , [0, 1]). When
integration is actually a small
combined with 1.8, this shows that
1 modification of Riemann’s definition
L( f , [0, 1]) = U ( f , [0, 1]) = 3 . Thus
that was proposed by Gaston
f is Riemann integrable and
Darboux (1842–1917).
Z 1
1
f = .
0 3

Now we come to a key result regarding Riemann integration. Uniform continuity


provides the major tool that makes the proof work.

1.11 continuous functions are Riemann integrable

Every continuous real-valued function on each closed bounded interval is


Riemann integrable.

Proof Suppose a, b ∈ R with a < b and f : [ a, b] → R is a continuous function


(thus by a standard theorem from undergraduate real analysis, f is bounded and is
uniformly continuous). Let ε > 0. Because f is uniformly continuous, there exists
δ > 0 such that

1.12 | f (s) − f (t)| < ε for all s, t ∈ [ a, b] with |s − t| < δ.

Let n ∈ Z+ be such that b− a


n < δ.
Let P be the equally spaced partition a = x0 , x1 , . . . , xn = b of [ a, b] with
b−a
x j − x j −1 =
n
for each j = 1, . . . , n. Then

U ( f , [ a, b]) − L( f , [ a, b]) ≤ U ( f , P, [ a, b]) − L( f , P, [ a, b])


b−a n  
= ∑ sup f − inf f
n j =1 [ x , x ] [ x j −1 , x j ]
j −1 j

≤ (b − a)ε,
where the first equality follows from the definitions of U ( f , [ a, b]) and L( f , [ a, b])
and the last inequality follows from 1.12.
We have shown that U ( f , [ a, b]) − L( f , [ a, b]) ≤ (b − a)ε for all ε > 0. Thus
1.8 implies that L( f , [ a, b]) = U ( f , [ a, b]). Hence f is Riemann integrable.
Rb Rb
An alternative notation for a f is a f ( x ) dx. Here x is a dummy variable, so
Rb
we could also write a f (t) dt or use another variable. This notation becomes useful
R1
when we want to write something like 0 x2 dx instead of using function notation.
Section 1A Review: Riemann Integral 7

The next result gives a frequently used estimate for a Riemann integral.

1.13 bounds on Riemann integral

Suppose f : [ a, b] → R is Riemann integrable. Then


Z b
(b − a) inf f ≤ f ≤ (b − a) sup f
[ a, b] a [ a, b]

Proof Let P be the trivial partition a = x0 , x1 = b. Then


Z b
(b − a) inf f = L( f , P, [ a, b]) ≤ L( f , [ a, b]) = f,
[ a, b] a

proving the first inequality in the result.


The second inequality in the result is proved similarly and is left to the reader.

EXERCISES 1A
1 Suppose f : [ a, b] → R is a bounded function such that

L( f , P, [ a, b]) = U ( f , P, [ a, b])

for some partition P of [ a, b]. Prove that f is a constant function on [ a, b].


2 Suppose a ≤ s < t ≤ b. Define f : [ a, b] → R by
(
1 if s < x < t,
f (x) =
0 otherwise.
Rb
Prove that f is Riemann integrable on [ a, b] and that a f = t − s.
3 Suppose f : [ a, b] → R is a bounded function. Prove that f is Riemann inte-
grable if and only if for each ε > 0, there exists a partition P of [ a, b] such
that
U ( f , P, [ a, b]) − L( f , P, [ a, b]) < ε.
4 Suppose f , g : [ a, b] → R are Riemann integrable. Prove that f + g is Riemann
integrable on [ a, b] and
Z b Z b Z b
( f + g) = f+ g.
a a a

5 Suppose f : [ a, b] → R is Riemann integrable. Prove that the function − f is


Riemann integrable on [ a, b] and
Z b Z b
(− f ) = − f.
a a
8 Chapter 1 Riemann Integration

6 Suppose f : [ a, b] → R is Riemann integrable. Suppose g : [ a, b] → R is a


function such that g( x ) = f ( x ) for all except finitely many x ∈ [ a, b]. Prove
that g is Riemann integrable on [ a, b] and
Z b Z b
g= f.
a a

7 Suppose f : [ a, b] → R is a bounded function. For n ∈ Z+, let Pn denote the


partition that divides [ a, b] into 2n intervals of equal size. Prove that

L( f , [ a, b]) = lim L( f , Pn , [ a, b]) and U ( f , [ a, b]) = lim U ( f , Pn , [ a, b]).


n→∞ n→∞

8 Suppose f : [ a, b] → R is Riemann integrable. Prove that


Z b n
b−a j(b− a) 
a
f = lim
n→∞ n ∑f a+ n .
j =1

9 Suppose f : [ a, b] → R is Riemann integrable. Prove that if c, d ∈ R and


a ≤ c < d ≤ b, then f is Riemann integrable on [c, d].
[To say that f is Riemann integrable on [c, d] means that f with its domain
restricted to [c, d] is Riemann integrable.]
10 Suppose f : [ a, b] → R is a bounded function and c ∈ ( a, b). Prove that f is
Riemann integrable on [ a, b] if and only if f is Riemann integrable on [ a, c] and
f is Riemann integrable on [c, b]. Furthermore, prove that if these conditions
hold, then
Z b Z c Z b
f = f+ f.
a a c
11 Suppose f : [ a, b] → R is Riemann integrable. Define F : [ a, b] → R by

0 if t = a,
F (t) = Z t
 f if t ∈ ( a, b].
a

Prove that F is continuous on [ a, b].


12 Suppose f : [ a, b] → R is Riemann integrable. Prove that | f | is Riemann
integrable and that
Z b Z b

f ≤ | f |.
a a
13 Suppose f : [ a, b] → R is an increasing function, meaning that c, d ∈ [ a, b] with
c < d implies f (c) ≤ f (d). Prove that f is Riemann integrable on [ a, b].
14 Suppose f 1 , f 2 , . . . is a sequence of Riemann integrable functions on [ a, b] such
that f 1 , f 2 , . . . converges uniformly on [ a, b] to a function f : [ a, b] → R. Prove
that f is Riemann integrable and
Z b Z b
f = lim fn.
a n→∞ a
Section 1B Riemann Integral Is Not Good Enough 9

1B Riemann Integral Is Not Good Enough


The Riemann integral works well enough to be taught to millions of calculus students
around the world each year. However, the Riemann integral has several deficiencies.
In this section, we discuss the following three issues:
• Riemann integration does not handle functions with many discontinuities;
• Riemann integration does not handle unbounded functions;
• Riemann integration does not work well with limits.
In Chapter 2, we will start to construct a theory to remedy these problems.
We begin with the following example of a function that is not Riemann integrable.

1.14 Example a function that is not Riemann integrable


Define f : [0, 1] → R by
(
1 if x is rational,
f (x) =
0 if x is irrational.
If [ a, b] ⊂ [0, 1] with a < b, then
inf f = 0 and sup f = 1
[ a, b] [ a, b]

because [ a, b] contains an irrational number and contains a rational number. Thus


L( f , P, [0, 1]) = 0 and U ( f , P, [0, 1]) = 1 for every partition P of [0, 1]. Hence
L( f , [0, 1]) = 0 and U ( f , [0, 1]) = 1. Because L( f , [0, 1]) 6= U ( f , [0, 1]), we
conclude that f is not Riemann integrable.
This example is disturbing because (as we will see later), there are far fewer
rational numbers than irrational numbers. Thus f should, in some sense, have
integral 0. However, the Riemann integral of f is not defined.

Trying to apply the definition of the Riemann integral to unbounded functions


would lead to undesirable results, as shown in the next example.

1.15 Example Riemann integration does not work with unbounded functions
Define f : [0, 1] → R by
(
√1 if 0 < x ≤ 1,
x
f (x) =
0 if x = 0.
If x0 , x1 , . . . , xn is a partition of [0, 1], then sup f = ∞. Thus if we tried to apply
[ x0 , x1 ]
the definition of the upper Riemann sum to f , we would have U ( f , P, [0, 1]) = ∞
for every partition P of [0, 1].
However, we should consider the area under the graph of f to be 2, not ∞, because
Z 1 √
lim f = lim(2 − 2 a) = 2.
a ↓0 a a ↓0
10 Chapter 1 Riemann Integration

R1
Calculus courses deal with the previous example by defining √1 dx to be
0 x
R1
lima↓0 a √1x dx. If using this approach and

1 1
f (x) = √ + √ ,
x 1−x
R1
then we would define 0 f to be
Z 1/2 Z b
lim f + lim f.
a ↓0 a b ↑1 1/2

However, the idea of taking Riemann integrals over subdomains and then taking
limits can fail with more complicated functions, as shown in the next example.

1.16 Example area seems to make sense, but Riemann integral is not defined
Let r1 , r2 , . . . be a sequence that includes each rational number in (0, 1) exactly
once and that includes no other numbers. For k ∈ Z+, define f k : [0, 1] → R by

 √ x1−r if x > rk ,
k
f k (x) =

0 if x ≤ rk .

Define f : [0, 1] → [0, ∞] by



f k (x)
f (x) = ∑ 2k
.
k =1

Because every nonempty open subinterval of [0, 1] contains a rational number, the
function f is unbounded on every such subinterval. Thus the Riemann integral of f
is undefined on every subinterval of [0, 1] with more than one element.
However, the area under the graph of each f k is less than 2. The formula defining
f then shows that we should expect the area under the graph of f to be less than 2
rather than undefined.

The next example shows that the pointwise limit of a sequence of Riemann
integrable functions bounded by 1 need not be Riemann integrable.

1.17 Example Riemann integration does not work well with pointwise limits
Let r1 , r2 , . . . be a sequence that includes each rational number in [0, 1] exactly
once and that includes no other numbers. For k ∈ Z+, define f k : [0, 1] → R by

1 if x ∈ {r1 , . . . , rk },
f k (x) =
0 otherwise.

R1
Then each f k is Riemann integrable and 0 f k = 0, as you should verify.
Section 1B Riemann Integral Is Not Good Enough 11

Define f : [0, 1] → R by
(
1 if x is rational,
f (x) =
0 if x is irrational.

Clearly
lim f k ( x ) = f ( x ) for each x ∈ [0, 1].
k→∞

However, f is not Riemann integrable (see Example 1.14) even though f is the
pointwise limit of a sequence of integrable functions bounded by 1.

Because analysis relies heavily upon limits, a good theory of integration should
allow for interchange of limits and integrals, at least when the functions are appropri-
ately bounded. Thus the previous example points out a serious deficiency in Riemann
integration.
Now we come to a positive result, but as we will see, even this result indicates that
Riemann integration has some problems.

1.18 interchanging Riemann integral and limit

Suppose a, b, M ∈ R with a < b. Suppose f 1 , f 2 , . . . is a sequence of Riemann


integrable functions on [ a, b] such that

| f k ( x )| ≤ M

for all k ∈ Z+ and all x ∈ [ a, b]. Suppose limk→∞ f k ( x ) exists for each
x ∈ [ a, b]. Define f : [ a, b] → R by

f ( x ) = lim f k ( x ).
k→∞

If f is Riemann integrable on [ a, b], then


Z b Z b
f = lim fk .
a k→∞ a

The result above suffers from two problems. The first problem is the undesirable
hypothesis that the limit function f is Riemann integrable. Ideally, that property
would follow from the other hypotheses, but Example 1.17 shows that this need not
be true.
The second problem with the result
The difficulty in finding a simple
above is that its proof seems to be more
Riemann-integration-based proof of
intricate than the proofs of other results
the result above suggests that
involving Riemann integration. We do not
Riemann integration is not the ideal
give a proof here of the result above. A
theory of integration.
clean proof of a stronger result is given in
Chapter 3, using the tools of measure theory that we develop starting with the next
chapter.
12 Chapter 1 Riemann Integration

EXERCISES 1B
1 Define f : [0, 1] → R as follows:

0 if a is irrational,

f ( a) = 1 if a is rational and n is the smallest positive integer

n
such that a = mn for some integer m.
Z 1
Show that f is Riemann integrable and compute f.
0

2 Suppose f : [ a, b] → R is a bounded function. Prove that f is Riemann inte-


grable if and only if

L(− f , [ a, b]) = − L( f , [ a, b]).

3 Suppose f , g : [ a, b] → R are bounded functions. Prove that

L( f , [ a, b]) + L( g, [ a, b]) ≤ L( f + g, [ a, b])

and
U ( f + g, [ a, b]) ≤ U ( f , [ a, b]) + U ( g, [ a, b]).
4 Give an example of bounded functions f , g : [0, 1] → R such that

L( f , [0, 1]) + L( g, [0, 1]) < L( f + g, [0, 1])

and
U ( f + g, [0, 1]) < U ( f , [0, 1]) + U ( g, [0, 1]).
5 Give an example of a sequence of continuous real-valued functions f 1 , f 2 , . . .
on [0, 1] and a continuous real-valued function f on [0, 1] such that

f ( x ) = lim f k ( x )
k→∞

for each x ∈ [0, 1] but


Z 1 Z 1
f 6= lim fk .
0 k→∞ 0
Chapter 2
Measures

The last section of the previous chapter discusses several deficiencies of Riemann
integration. To remedy those deficiencies, in this chapter we extend the notion of the
length of an interval to a larger collection of subsets of R. This leads us to measures
and then in the next chapter to integration with respect to measures.
We begin this chapter by investigating outer measure, which looks promising but
fails to have a crucial property. That failure leads us to σ-algebras and measurable
spaces. Then we define measures in an abstract context that can be applied to settings
more general than R. Next, we construct Lebesgue measure on R as our desired
extension of the notion of the length of an interval.

Fifth-century AD Roman ceiling mosaic in what is now a UNESCO World Heritage


site in Ravenna, Italy. Giuseppe Vitali, who in 1905 proved result 2.18 in this chapter,
was born and grew up in Ravenna, where perhaps he saw this mosaic. Could the
memory of the translation-invariant feature of this mosaic have suggested to Vitali
the translation invariance that is the heart of his proof of 2.18?
CC-BY-SA Petar Milošević

13
14 Chapter 2 Measures

2A Outer Measure on R
Motivation and Definition of Outer Measure
The Riemann integral arises from approximating the area under the graph of a function
by sums of the areas of approximating rectangles. These rectangles have heights that
approximate the values of the function on subintervals of the function’s domain. The
width of each approximating rectangle is the length of the corresponding subinterval.
This length is the term x j − x j−1 in the definitions of the lower and upper Riemann
sums (see 1.3).
To extend integration to a larger class of functions than the Riemann integrable
functions, we will write the domain of a function as the union of subsets more
complicated than the subintervals used in Riemann integration. We will need to
assign a size to each of those subsets, where the size is an extension of the length of
intervals.
For example, we expect the size of the set (1, 3) ∪ (7, 10) to be 5 (because the
first interval has length 2, the second interval has length 3, and 2 + 3 = 5).
Assigning a size to subsets of R that are more complicated than unions of open
intervals becomes a nontrivial task. This chapter focuses on that task and its extension
to other contexts. In the next chapter, we will see how to use the ideas developed in
this chapter to create a rich theory of integration.
We begin by giving the expected definition of the length of an open interval, along
with a notation for that length.

2.1 Definition length of open interval; `( I )

The length `( I ) of an open interval I is defined by




 b − a if I = ( a, b) for some a, b ∈ R with a < b,



0 if I = ∅,
`( I ) =

 ∞ if I = (−∞, a) or I = ( a, ∞) for some a ∈ R,



∞ if I = (−∞, ∞).

Suppose A ⊂ R. The size of A should be at most the sum of the lengths of a


sequence of open intervals whose union contains A. Taking the infimum of all such
sums gives a reasonable definition of the size of A, denoted | A| and called the outer
measure of A.

2.2 Definition outer measure; | A|

The outer measure | A| of a set A ⊂ R is defined by


n ∞ ∞
[ o
| A| = inf ∑ `( Ik ) : I1 , I2 , . . . are open intervals such that A ⊂ Ik .
k =1 k =1
Section 2A Outer Measure on R 15

The definition of outer measure involves an infinite sum. The infinite sum ∑∞ k =1 t k
of a sequence t1 , t2 , . . . of elements of [0, ∞] is defined to be ∞ if some tk = ∞.
Otherwise, ∑∞ k=1 tk is defined to be the limit (possibly ∞) of the increasing sequence
t1 , t1 + t2 , t1 + t2 + t3 , . . . of partial sums; thus
∞ n
∑ tk = nlim
→∞
∑ tk .
k =1 k =1

2.3 Example finite sets have outer measure 0


Suppose A = { a1 , . . . , an } is a finite set of real numbers. Suppose ε > 0. Define
a sequence I1 , I2 , . . . of open intervals by
(
( ak − ε, ak + ε) if k ≤ n,
Ik =
∅ if k > n.

Then I1 , I2 , . . . is a sequence of open intervals whose union contains A. Clearly


∑∞k=1 `( Ik ) = 2εn. Hence | A | ≤ 2εn. Because ε is an arbitrary positive number, this
implies that | A| = 0.

Good Properties of Outer Measure


Outer measure has several nice properties that are discussed in this subsection. We
begin with a result that improves upon the example above.

2.4 countable sets have outer measure 0

Every countable subset of R has outer measure 0.

Proof Suppose A = { a1 , a2 , . . .} is a countable subset of R. Let ε > 0. For k ∈ Z+,


let  ε ε
Ik = ak − k , ak + k .
2 2
Then I1 , I2 , . . . is a sequence of open intervals whose union contains A. Because

∑ `( Ik ) = 2ε,
k =1

we have | A| ≤ 2ε. Because ε is an arbitrary positive number, this implies that


| A| = 0.

The result above, along with the result that the set Q of rational numbers is
countable, implies that Q has outer measure 0. We will soon show that there are far
fewer rational numbers than real numbers (see 2.17). Thus the equation |Q| = 0
indicates that outer measure has a good property that we want any reasonable notion
of size to possess.
16 Chapter 2 Measures

The next result shows that outer measure does the right thing with respect to set
inclusion.

2.5 outer measure preserves order

Suppose A and B are subsets of R with A ⊂ B. Then | A| ≤ | B|.

Proof Suppose I1 , I2 , . . . is a sequence of open intervals whose union contains B.


Then the union of this sequence of open intervals also contains A. Hence

| A| ≤ ∑ `( Ik ).
k =1

Taking the infimum over all sequences of open intervals whose union contains B, we
have | A| ≤ | B|.

We expect that the size of a subset of R should not change if the set is shifted to
the right or to the left. The next definition allows us to be more precise.

2.6 Definition translation; t + A

If t ∈ R and A ⊂ R, then the translation t + A is defined by

t + A = { t + a : a ∈ A }.

If t > 0, then t + A is obtained by moving the set A to the right t units on the real
line; if t < 0, then t + A is obtained by moving the set A to the left |t| units.
Translation does not change the length of an open interval. Specifically, if t ∈ R
and a, b ∈ [−∞, ∞], then t + ( a, b) = (t + a, t + b) and thus ` t + ( a, b) =

` ( a, b) . Here we are using the standard convention that t + (−∞) = −∞ and
t + ∞ = ∞.
The next result states that translation invariance carries over to outer measure.

2.7 outer measure is translation invariant

Suppose t ∈ R and A ⊂ R. Then |t + A| = | A|.

Proof Suppose I1 , I2 , . . . is a sequence of open intervals whose union contains A.


Then t + I1 , t + I2 , . . . is a sequence of open intervals whose union contains t + A.
Thus
∞ ∞
|t + A| ≤ ∑ `(t + Ik ) = ∑ `( Ik ).
k =1 k =1
Taking the infimum of the last term over all sequences I1 , I2 , . . . of open intervals
whose union contains A, we have |t + A| ≤ | A|.
To get the inequality in the other direction, note that A = −t + (t + A). Thus
applying the inequality from the previous paragraph, with A replaced by t + A and t
replaced by −t, we have | A| = |−t + (t + A)| ≤ |t + A|. Hence |t + A| = | A|.
Section 2A Outer Measure on R 17

The union of the intervals (1, 4) and (3, 5) is the interval (1, 5). Thus
  
` (1, 4) ∪ (3, 5) < ` (1, 4) + ` (3, 5)
because the left side of the inequality above equals 4 and the right side equals 5. The
direction of the inequality above is explained by noting that the interval (3, 4), which
is the intersection of (1, 4) and (3, 5), has its length counted twice on the right side
of the inequality above.
The example of the paragraph above should provide intuition for the direction of
the inequality in the next result. The property of satisfying the inequality in the result
below is called countable subadditivity because it applies to sequences of subsets.

2.8 countable subadditivity of outer measure

Suppose A1 , A2 , . . . is a sequence of subsets of R. Then


[∞ ∞

Ak ≤ ∑ | A k |.
k =1 k =1

Proof If | Ak | = ∞ for some k ∈ Z+, then the inequality above clearly holds. Thus
assume | Ak | < ∞ for all k ∈ Z+.
Let ε > 0. For each k ∈ Z+, let I1,k , I2,k , . . . be a sequence of open intervals
whose union contains Ak such that

ε
∑ `( Ij,k ) ≤ 2k + | Ak |.
j =1

Thus
∞ ∞ ∞
2.9 ∑ ∑ `( Ij,k ) ≤ ε + ∑ | Ak |.
k =1 j =1 k =1

The doubly indexed collection of open intervals { Ij,k : j, S k ∈ Z+} can be rearranged
into a sequence of open intervals whose union contains ∞ k=1 Ak as follows, where
in step k (start with k = 2, then k = 3, 4, 5, . . . ) we adjoin the k − 1 intervals whose
indices add up to k:
I1,1 , I1,2 , I2,1 , I1,3 , I2,2 , I3,1 , I1,4 , I2,3 , I3,2 , I4,1 , I1,5 , I2,4 , I3,3 , I4,2 , I5,1 , . . . .
|{z} | {z } | {z } | {z } | {z }
2 3 4 5 sum of the two indices is 6

Inequality 2.9 shows that the sum of the


S∞lengths of the intervals listed above is less
than or equal to ε + ∑∞
k=1 | Ak |. Thus

k= 1SAk ≤ ε + ∑k=1 | Ak |. Because ε is an
arbitrary positive number, this implies that ∞ ∞
k =1 A k ≤ ∑ k =1 | A k | .

Countable subadditivity implies finite subadditivity, meaning that


| A1 ∪ · · · ∪ A n | ≤ | A1 | + · · · + | A n |
for all A1 , . . . , An ⊂ R, because we can take Ak = ∅ for k > n in 2.8.
The countable subadditivity of outer measure, as proved above, adds to our list of
nice properties enjoyed by outer measure.
18 Chapter 2 Measures

Outer Measure of Closed Bounded Interval


One more good property of outer measure that we should prove is that if a < b,
then the outer measure of the closed interval [ a, b] is b − a. Indeed, if ε > 0, then
( a − ε, b + ε), ∅, ∅, . . . is a sequence of open intervals whose union contains [ a, b].
Thus |[ a, b]| ≤ b − a + 2ε. Because this inequality holds for all ε > 0, we conclude
that
|[ a, b]| ≤ b − a.
Is the inequality in the other direction obviously true to you? If so, think again,
because a proof of the inequality in the other direction requires that the completeness
of R is used in some form. For example, suppose R was a countable set (which is not
true, as we will soon see, but the uncountability of R is not obvious). Then we would
have |[ a, b]| = 0 (by 2.4). Thus something deeper than you might suspect is going
on with the ingredients needed to prove that |[ a, b]| ≥ b − a.
The following definition will be useful when we prove that |[ a, b]| ≥ b − a.

2.10 Definition open cover

Suppose A ⊂ R.
• A collection C of open subsets of R is called an open cover of A if A is
contained in the union of all the sets in C .
• An open cover C of A is said to have a finite subcover if A is contained in
the union of some finite list of sets in C .

2.11 Example open covers and finite subcovers

• The collection
S∞
{(k, k + 2) : k ∈ Z+} is an open cover of [2, 5] because
[2, 5] ⊂ k=1 (k, k + 2). This open cover has a finite subcover because [2, 5] ⊂
(1, 3) ∪ (2, 4) ∪ (3, 5) ∪ (4, 6).
• The collection
S∞
{(k, k + 2) : k ∈ Z+} is an open cover of [2, ∞) because
[2, ∞] ⊂ k=1 (k, k + 2). This open cover does not have a finite subcover
because there do not exist finitely many sets of the form (k, k + 2) whose union
contains [2, ∞).
• The collection {(0, 2 − 1k ) : k ∈ Z+} is an open cover of (1, 2) because
S
(1, 2) ⊂ ∞ 1
k=1 (0, 2 − k ). This open cover does not have a finite subcover
because there do not exist finitely many sets of the form (0, 2 − 1k ) whose union
contains (1, 2).

The next result will be our major tool in the proof that |[ a, b]| ≥ b − a. Although
we need only the result as stated, be sure to see Exercise 4 in this section, which
when combined with the next result gives a characterization of the closed bounded
subsets of R. Note that the following proof uses the completeness property of the real
numbers (by asserting that the supremum of a certain nonempty bounded set exists).
Section 2A Outer Measure on R 19

2.12 Heine–Borel Theorem

Every open cover of a closed bounded subset of R has a finite subcover.

Proof Suppose F is a closed bounded


To provide visual clues, we usually
subset of R and C is an open cover of F.
denote closed sets by F and open
First consider the case where F =
sets by G.
[ a, b] for some a, b ∈ R with a < b. Thus
C is an open cover of [ a, b]. Let
D = {d ∈ [ a, b] : [ a, d] has a finite subcover from C}.
Note that a ∈ D (because a ∈ G for some G ∈ C ). Thus D is not the empty set. Let
s = sup D.
Thus s ∈ [ a, b]. Hence there exists an open set G ∈ C such that s ∈ G. Let δ > 0
be such that (s − δ, s + δ) ⊂ G. Because s = sup D, there exist d ∈ (s − δ, s] and
n ∈ Z+ and G1 , . . . , Gn ∈ C such that
[ a, d] ⊂ G1 ∪ · · · ∪ Gn .
Now
2.13 [ a, d0 ] ⊂ G ∪ G1 ∪ · · · ∪ Gn
for all d0 ∈ [s, s + δ).Thus d0 ∈ D for all d0 ∈ [s, s + δ) ∩ [ a, b]. This implies that
s = b. Furthermore, 2.13 with d0 = b shows that [ a, b] has a finite subcover from C ,
completing the proof in the case where F = [ a, b].
Now suppose F is an arbitrary closed bounded subset of R and that C is an open
cover of F. Let a, b ∈ R be such that F ⊂ [ a, b]. Now C ∪ {R \ F } is an open cover
of R and hence is an open cover of [ a, b] (here R \ F denotes the set complement of
F in R). By our first case, there exist G1 , . . . , Gn ∈ C such that
[ a, b] ⊂ G1 ∪ · · · ∪ Gn ∪ (R \ F ).
Thus
F ⊂ G1 ∪ · · · ∪ Gn ,
completing the proof.

Saint-Affrique, the small


town in southern France
where Émile Borel
(1871–1956) was born.
Borel first stated and
proved what we call the
Heine–Borel Theorem in
1895. Earlier, Eduard
Heine (1821–1881) and
others had used similar
results.
CC-BY-SA Fagairolles 34
20 Chapter 2 Measures

Now we can prove that closed intervals have the expected outer measure.

2.14 outer measure of a closed interval

Suppose a, b ∈ R, with a < b. Then |[ a, b]| = b − a.

Proof See the first paragraph of this subsection for the proof that |[ a, b]| ≤ b − a.
To prove the inequality in theSother direction, suppose I1 , I2 , . . . is a sequence of
open intervals such that [ a, b] ⊂ ∞k=1 Ik . By the Heine–Borel Theorem (2.12), there
exists n ∈ Z+ such that
2.15 [ a, b] ⊂ I1 ∪ · · · ∪ In .
We will now prove by induction on n that the inclusion above implies that
n
2.16 ∑ `( Ik ) ≥ b − a.
k =1
This will then imply that ∑∞ n
k=1 `( Ik ) ≥ ∑k=1 `( Ik ) ≥ b − a, completing the proof
that |[ a, b]| ≥ b − a.
To get started with our induction, note that 2.15 clearly implies 2.16 if n = 1.
Now for the induction step: Suppose n > 1 and 2.15 implies 2.16 for all choices of
a, b ∈ R with a < b. Suppose I1 , . . . , In , In+1 are open intervals such that
[ a, b] ⊂ I1 ∪ · · · ∪ In ∪ In+1 .
Thus b is in at least one of the intervals I1 , . . . , In , In+1 . By relabeling, we can
assume that b ∈ In+1 . Suppose In+1 = (c, d). If c ≤ a, then `( In+1 ) ≥ b − a and
there is nothing further to prove; thus we can assume that a < c < b < d, as shown
in the figure below.

Hence
Alice was beginning to get very tired
[ a, c] ⊂ I1 ∪ · · · ∪ In . of sitting by her sister on the bank,
By our induction hypothesis, we have and of having nothing to do: once or
twice she had peeped into the book
∑nk=1 `( Ik ) ≥ c − a. Thus
her sister was reading, but it had no
n +1 pictures or conversations in it, “and
∑ `( Ik ) ≥ (c − a) + `( In+1 ) what is the use of a book,” thought
k =1
Alice, “without pictures or
= (c − a) + (d − c)
conversation?”
= d−a – opening paragraph of Alice’s
Adventures in Wonderland, by Lewis
≥ b − a, Carroll
completing the proof.

The result above easily implies that the outer measure of each open interval equals
its length (see Exercise 6).
Section 2A Outer Measure on R 21

The previous result has the following important corollary. You may be familiar
with Georg Cantor’s (1845–1918) original proof of the next result. The proof using
outer measure that is presented here gives an interesting alternative to Cantor’s proof.

2.17 nontrivial intervals are uncountable

Every interval in R that contains at least two distinct elements is uncountable.

Proof Suppose I is an interval that contains a, b ∈ R with a < b. Then

| I | ≥ |[ a, b]| = b − a > 0,
where the first inequality above holds because outer measure preserves order (see 2.5)
and the equality above comes from 2.14. Because every countable subset of R has
outer measure 0 (see 2.4), we can conclude that I is uncountable.

Outer Measure is Not Additive


We have had several results giving nice
Outer measure led to the proof
properties of outer measure. Now we
above that R is uncountable. This
come to an unpleasant property of outer
application of outer measure to
measure.
prove a result that seems
If outer measure were a perfect way to
unconnected with outer measure is
assign a size as an extension of the lengths
an indication that outer measure has
of intervals, then the outer measure of the
serious mathematical value.
union of two disjoint sets would equal the
sum of the outer measures of the two sets. Sadly, the next result states that outer
measure does not have this property.
In the next section, we begin the process of getting around the next result, which
will lead us to measure theory.

2.18 nonadditivity of outer measure

There exist disjoint subsets A and B of R such that

| A ∪ B | 6 = | A | + | B |.

Proof For a ∈ [−1, 1], let ã be the set of numbers in [−1, 1] that differ from a by a
rational number. In other words,

ã = {c ∈ [−1, 1] : a − c ∈ Q}.

If a, b ∈ [−1, 1] and ã ∩ b̃ 6= ∅, then


Think of ã as the equivalence class
ã = b̃. (Proof: Suppose there exists d ∈
of a under the equivalence relation
ã ∩ b̃. Then a − d and b − d are rational
that declares a, c ∈ [−1, 1] to be
numbers; subtracting, we conclude that
equivalent if a − c ∈ Q.
a − b is a rational number. The equation
a − c = ( a − b) + (b − c) now implies that if c ∈ [−1, 1], then a − c is a rational
number if and only if b − c is a rational number. In other words, ã = b̃.)
≈ 7π Chapter 2 Measures

[
Clearly a ∈ ã for each a ∈ [−1, 1]. Thus [−1, 1] = ã.
a∈[−1,1]
Let V be a set that contains exactly one
This step involves the Axiom of
element in each of the distinct sets in
Choice, as discussed after this proof.
{ ã : a ∈ [−1, 1]}. The set V arises by choosing one
element from each equivalence
In other words, for every a ∈ [−1, 1], the class.
set V ∩ ã has exactly one element.
Let r1 , r2 , . . . be a sequence of distinct rational numbers such that

[−2, 2] ∩ Q = {r1 , r2 , . . .}.


Then

[
[−1, 1] ⊂ (r k + V ),
k =1
where the set inclusion above holds because if a ∈ [−1, 1], then letting v be the
unique element of V ∩ ã, we have a − v ∈ Q, which implies that a = rk + v ∈
rk + V for some k ∈ Z+.
The set inclusion above, the order-preserving property of outer measure (2.5), and
the countable subadditivity of outer measure (2.8) imply

|[−1, 1]| ≤ ∑ |r k + V |.
k =1

We know that |[−1, 1]| = 2 (from 2.14). The translation invariance of outer measure
(2.7) thus allows us to rewrite the inequality above as

2≤ ∑ |V |.
k =1

Thus |V | > 0.
Note that the sets r1 + V, r2 + V, . . . are disjoint. (Proof: Suppose there exists
t ∈ (r j + V ) ∩ (rk + V ). Then t = r j + v1 = rk + v2 for some v1 , v2 ∈ V, which
implies that v1 − v2 = rk − r j ∈ Q. Our construction of V now implies that v1 = v2 ,
which implies that r j = rk , which implies that j = k.)
Let n ∈ Z+. Clearly
n
[
(rk + V ) ⊂ [−3, 3]
k =1
because V ⊂ [−1, 1] and each rk ∈ [−2, 2]. The set inclusion above implies that
[n

2.19 (rk + V ) ≤ 6.
k =1

However
n n
2.20 ∑ |r k + V | = ∑ |V | = n |V |.
k =1 k =1
Section 2A Outer Measure on R 23

Now 2.19 and 2.20 suggest that we choose n ∈ Z+ such that n|V | > 6. Thus
[n n

2.21 (r k + V ) < ∑ |r k + V |.
k =1 k =1

If we had | A ∪ B| = | A| + | B| for all disjoint subsets A, B of R, then by induction


[n n

on n we would have Ak = ∑ | Ak | for all disjoint subsets A1 , . . . , An of R.
k =1 k =1
However, 2.21 tells us that no such result holds. Thus there exist disjoint subsets
A, B of R such that | A ∪ B| 6= | A| + | B|.

The Axiom of Choice, which belongs to set theory, states that if E is a set whose
elements are disjoint nonempty sets, then there exists a set D that contains exactly
one element in each set that is an element of E . We used the Axiom of Choice to
construct the set D that was used in the last proof.
A small minority of mathematicians objects to the use of the Axiom of Choice.
Thus we will keep track of where we need to use it. Even if you do not like to use the
Axiom of Choice, the previous result warns us away from trying to prove that outer
measure is additive (any such proof would need to contradict the Axiom of Choice,
which is consistent with the standard axioms of set theory).

EXERCISES 2A
1 Prove that if A and B are subsets of R and | B| = 0, then | A ∪ B| = | A|.
2 Suppose A ⊂ R and t ∈ R. Let tA = {ta : a ∈ A}. Prove that |tA| = |t| | A|.
[Assume that 0 · ∞ is defined to be 0.]
3 Prove that if A, B ⊂ R and | A| < ∞, then | B \ A| ≥ | B| − | A|.
4 Suppose F is a subset of R with the property that every open cover of F has a
finite subcover. Prove that F is closed and bounded.
T
5 Suppose A is a set of closed subsets of R such that F∈A F = ∅. Prove that if A
contains at least one bounded set, then there exist n ∈ Z+ and F1 , . . . , Fn ∈ A
such that F1 ∩ · · · ∩ Fn = ∅.
6 Prove that if a, b ∈ R and a < b, then

|( a, b)| = |[ a, b)| = |( a, b]| = b − a.

7 Suppose a, b, c, d are real numbers with a < b and c < d. Prove that

|( a, b) ∪ (c, d)| = (b − a) + (d − c) if and only if ( a, b) ∩ (c, d) = ∅.



8 Prove that if A ⊂ R and t > 0, then | A| = | A ∩ (−t, t)| + A ∩ R \ (−t, t) .
9 Prove that | A| = lim | A ∩ (−t, t)| for all A ⊂ R.
t→∞
24 Chapter 2 Measures

10 Prove that |[0, 1] \ Q| = 1.


11 Prove that if I1 , I2 , . . . is a disjoint sequence of open intervals, then
[∞ ∞

Ik = ∑ `( Ik ).
k =1 k =1

12 Suppose r1 , r2 , . . . is a sequence that contains every rational number. Let


∞ 
[ 1 1
F = R\ rk − k , rk + k .
k =1
2 2
(a) Show that F is a closed subset of R.
(b) Prove that if I is an interval contained in F, then I contains at most one
element.
(c) Prove that | F | = ∞.

13 Suppose ε > 0. Prove that there exists a subset F of [0, 1] such that F is closed,
every element of F is an irrational number, and | F | > 1 − ε.
14 Consider the following figure, which is drawn accurately to scale.

(a) Show that the right triangle whose vertices are (0, 0), (20, 0), and (20, 9)
has area 90.
[We have not defined area yet, but just use the elementary formulas for the
areas of triangles and rectangles that you learned long ago.]
(b) Show that the yellow (lower) right triangle has area 27.5.
(c) Show that the red rectangle has area 45.
(d) Show that the blue (upper) right triangle has area 18.
(e) Add the results of parts (b), (c), and (d), showing that the area of the colored
region is 90.5.
(f) Seeing the figure above, most people expect parts (a) and (e) to have the
same result. Yet in part (a) we found area 90, and in part (e) we found area
90.5. Explain why these results differ.
[You may be tempted to think that what we have here is a two-dimensional
example similar to the result about the nonadditivity of outer measure
(2.18). However, genuine examples of nonadditivity require much more
complicated sets than in this example.]
Section 2B Measurable Spaces and Functions 25

2B Measurable Spaces and Functions


The last result in the previous section showed that outer measure is not additive.
Could this disappointing result perhaps be fixed by using some notion other than
outer measure for the size of a subset of R? The next result answers this question by
showing that there does not exist a notion of size, called the Greek letter mu (µ) in
the result below, that has all the desirable properties.
Property (c) in the result below is called countable additivity. Countable additivity
is a highly desirable property because we want to be able to prove theorems about
limits (the heart of analysis!), which requires countable additivity.

2.22 nonexistence of extension of length to all subsets of R

There does not exist a function µ with all the following properties:

(a) µ is a function from the set of subsets of R to [0, ∞].


(b) µ( I ) = `( I ) for every open interval I of R.

[  ∞
(c) µ Ak = ∑ µ( Ak ) for every disjoint sequence A1 , A2 , . . . of subsets
k =1 k =1
of R.
(d) µ(t + A) = µ( A) for every A ⊂ R and every t ∈ R.

Proof Suppose there exists a function µ


We will show that µ has all the
with all the properties listed in the state-
properties of outer measure that
ment of this result.
were used in the proof of 2.18.
Observe that µ(∅) = 0, as follows
from (b) because the empty set is an open interval with length 0.
If A ⊂ B ⊂ R, then µ( A) ≤ µ( B), as follows from (c) because we can write B
as the union of the disjoint sequence A, B \ A, ∅, ∅, . . . ; thus

µ ( B ) = µ ( A ) + µ ( B \ A ) + 0 + 0 + · · · = µ ( A ) + µ ( B \ A ) ≥ µ ( A ).

If a, b ∈ R with a < b, then ( a, b) ⊂ [ a, b] ⊂ ( a − ε, b + ε) for every ε > 0.


Thus b − a ≤ µ([ a, b]) ≤ b − a + 2ε for every ε > 0. Hence µ([ a, b]) = b − a.
If A1 , A2 , . . . is a sequence of subsets of R, then A1 , A2 \ A1 , A3 \ ( A1 ∪ A2 ), . . .
S
is a disjoint sequence of subsets of R whose union is ∞ k=1 Ak . Thus

[    
µ A k = µ A1 ∪ ( A2 \ A1 ) ∪ A3 \ ( A1 ∪ A2 ) ∪ · · ·
k =1

= µ ( A1 ) + µ ( A2 \ A1 ) + µ A3 \ ( A1 ∪ A2 ) + · · ·

≤ ∑ µ ( A k ),
k =1

where the second equality follows from the countable additivity of µ.


26 Chapter 2 Measures

We have shown that µ has all the properties of outer measure that were used
in the proof of 2.18. Repeating the proof of 2.18, we see that there exist disjoint
subsets A, B of R such that µ( A ∪ B) 6= µ( A) + µ( B). Thus the disjoint sequence
A, B, ∅, ∅, . . . does not satisfy the countable additivity property required by (c). This
contradiction completes the proof.

σ-Algebras
The last result shows that we need to give up one of the desirable properties in our
goal of extending the notion of size from intervals to more general subsets of R. We
cannot give up 2.22(b) because the size of an interval needs to be its length. We
cannot give up 2.22(c) because countable additivity is needed to prove theorems
about limits. We cannot give up 2.22(d) because a size that is not translation invariant
does not satisfy our intuitive notion of size as a generalization of length.
Thus we are forced to relax the requirement in 2.22(a) that the size is defined for
all subsets of R. Experience shows that to have a viable theory that allows for taking
limits, the collection of subsets for which the size is defined should be closed under
complementation and closed under countable unions. Thus we make the following
definition.

2.23 Definition σ-algebra

Suppose X is a set and S is a set of subsets of X. Then S is called a σ-algebra


on X if the following three conditions are satisfied:
• ∅ ∈ S;
• if E ∈ S , then X \ E ∈ S ;

[
• if E1 , E2 , . . . is a sequence of elements of S , then Ek ∈ S .
k =1

Make sure you verify that the examples in all three bullet points below are indeed
σ-algebras. The verification is obvious for the first two bullet points. For the third
bullet point, you need to use the result that the countable union of countable sets
is countable (see the proof of 2.8 for an example of how a doubly indexed list can
be converted to a singly indexed sequence). The exercises contain some additional
examples of σ-algebras.

2.24 Example σ-algebras

• Suppose X is a set. Then clearly {∅, X } is a σ-algebra on X.


• Suppose X is a set. Then clearly the set of all subsets of X is a σ-algebra on X.
• Suppose X is a set. Then the set of all subsets E of X such that E is countable or
X \ E is countable is a σ-algebra on X.
Section 2B Measurable Spaces and Functions 27

Now we come to some easy but important properties of σ-algebras.

2.25 σ-algebras are closed under countable intersection

Suppose S is a σ-algebra on a set X. Then

(a) X ∈ S ;
(b) if D, E ∈ S , then D ∪ E ∈ S and D ∩ E ∈ S and D \ E ∈ S ;

\
(c) if E1 , E2 , . . . is a sequence of elements of S , then Ek ∈ S .
k =1

Proof Because ∅ ∈ S and X = X \ ∅, the first two bullet points in the definition
of σ-algebra (2.23) imply that X ∈ S , proving (a).
Suppose D, E ∈ S . Then D ∪ E is the union of the sequence D, E, ∅, ∅, . . . of
elements of S . Thus the third bullet point in the definition of σ-algebra (2.23) implies
that D ∪ E ∈ S .
De Morgan’s Laws tell us that

X \ ( D ∩ E ) = ( X \ D ) ∪ ( X \ E ).

If D, E ∈ S , then the right side of the equation above is in S ; hence X \ ( D ∩ E) ∈ S ;


thus the complement in X of X \ ( D ∩ E) is in S ; in other words, D ∩ E ∈ S .
Because D \ E = D ∩ ( X \ E), we see that if D, E ∈ S , then D \ E ∈ S ,
completing the proof of (b).
Finally, suppose E1 , E2 , . . . is a sequence of elements of S . De Morgan’s Laws
tell us that

\ ∞
[
X\ Ek = ( X \ Ek ).
k =1 k =1
The right side of
T
the equation above is in S . Hence
T∞
the left side is in S , which implies
that X \ ( X \ ∞ k =1 Ek ) ∈ S . In other words, k=1 Ek ∈ S , proving (c).

The word measurable is used in the terminology below because in the next section
we introduce a size function, called a measure, defined on measurable sets.

2.26 Definition measurable space; measurable set

• A measurable space is an ordered pair ( X, S), where X is a set and S is a


σ-algebra on X.

• An element of S is called an S -measurable set, or just a measurable set if S


is clear from the context.

For example, if X = R and S is the set of all subsets of R that are countable or
have a countable complement, then the set of rational numbers is S -measurable but
the set of positive real numbers is not S -measurable.
28 Chapter 2 Measures

Borel Subsets of R
The next result guarantees that there is a smallest σ-algebra on a set X containing a
given set A of subsets of X.

2.27 smallest σ-algebra containing a collection of subsets

Suppose X is a set and A is a set of subsets of X. Then the intersection of all


σ-algebras on X that contain A is a σ-algebra on X.

Proof There is at least one σ-algebra on X that contains A because the σ-algebra
consisting of all subsets of X contains A.
Let S be the intersection of all σ-algebras on X that contain A. Then ∅ ∈ S
because ∅ is an element of each σ-algebra on X that contains A.
Suppose E ∈ S . Thus E is in every σ-algebra on X that contains A. Thus X \ E
is in every σ-algebra on X that contains A. Hence X \ E ∈ S .
Suppose E1 , E2 , . . . is a sequence of elements of S . Thus each Ek is in every σ-
S
algebra on SX that contains A. Thus ∞ k=1 Ek is in every σ-algebra on X that contains

A. Hence k=1 Ek ∈ S , which completes the proof that S is a σ-algebra on X.

Using the terminology smallest for the intersection of all σ-algebras that contain
a set A of subsets of X makes sense because the intersection of those σ-algebras is
contained in every σ-algebra that contains A.

2.28 Example smallest σ-algebra

• Suppose X is a set and A is the set of subsets of X that consist of exactly one
element: 
A = {x} : x ∈ X .
Then the smallest σ-algebra on X containing A is the set of all subsets E of X
such that E is countable or X \ E is countable, as you should verify.
• Suppose A = {(0, 1), (0, ∞)}. Then the smallest σ-algebra on R containing
A is {∅, (0, 1), (0, ∞), (−∞, 0] ∪ [1, ∞), (−∞, 0], [1, ∞), (−∞, 1), R}, as you
should verify.

Now we come to a crucial definition.

2.29 Definition Borel set

The smallest σ-algebra on R containing all open subsets of R is called the


collection of Borel subsets of R. An element of this σ-algebra is called a Borel
set.

We have defined the collection of Borel subsets of R to be the smallest σ-algebra


on R containing all the open subsets of R. We could have defined the collection of
Borel subsets of R to be the smallest σ-algebra on R containing all the open intervals
(because every open subset of R is the union of a sequence of open intervals).
Section 2B Measurable Spaces and Functions 29

2.30 Example Borel sets

• Every closed subset of R is a Borel set because every closed subset of R is the
complement of an open subset of R.
• EveryScountable subset of R is a Borel set because if B = { x1 , x2 , . . .}, then
B= ∞ k=1 { xk }, which is a Borel set because each { xk } is a closed set.

• Every
T∞
half-open interval [ a, b) (where a, b ∈ R) is a Borel set because [ a, b) =
1
k =1 a − k , b ).
(
• If f : R → R is a function, then the set of points at which f is continuous is the
intersection of a sequence of open sets (see Exercise 12 in this section) and thus
is a Borel set.

The intersection of every sequence of open subsets of R is a Borel set. However,


the set of all such intersections is not the set of Borel sets (because it is not closed
under countable unions). The set of all countable unions of countable intersections
of open subsets of R is also not the set of Borel sets (because it is not closed under
countable intersections). And so on ad infinitum—there is no finite procedure involv-
ing countable unions, countable intersections, and complements for constructing the
collection of Borel sets.
We will see later that there exist subsets of R that are not Borel sets. However, any
subset of R that you can write down in a concrete fashion is a Borel set.

Inverse Images
The next definition is used frequently in the rest of this chapter.

2.31 Definition inverse image; f −1 ( A)

If f : X → Y is a function and A ⊂ Y, then the set f −1 ( A) is defined by

f −1 ( A ) = { x ∈ X : f ( x ) ∈ A } .

2.32 Example inverse images


Suppose f : [0, 4π ] → R is defined by f ( x ) = sin x. Then

f −1 (0, ∞) = (0, π ) ∪ (2π, 3π ),
f −1 ([0, 1]) = [0, π ] ∪ [2π, 3π ] ∪ {4π },
f −1 ({−1}) = { 3π 7π
2 , 2 },

f −1 (2, 3) = ∅,

as you should verify.


30 Chapter 2 Measures

Inverse images have good algebraic properties, as is shown in the next two results.

2.33 algebra of inverse images

Suppose f : X → Y is a function. Then

(a) f −1 (Y \ A) = X \ f −1 ( A) for every A ⊂ Y;


S S
(b) f −1 ( A∈A A) = A∈A f −1 ( A) for every set A of subsets of Y;
T T
(c) f −1 ( A∈A A) = A∈A f −1 ( A) for every set A of subsets of Y.

Proof Suppose A ⊂ Y. For x ∈ X we have


x ∈ f −1 (Y \ A) ⇐⇒ f ( x ) ∈ Y \ A
/A
⇐⇒ f ( x ) ∈
/ f −1 ( A )
⇐⇒ x ∈
⇐⇒ x ∈ X \ f −1 ( A).
Thus f −1 (Y \ A) = X \ f −1 ( A), which proves (a).
To prove (b), suppose A is a set of subsets of Y. Then
[ [
x ∈ f −1 ( A) ⇐⇒ f ( x ) ∈ A
A∈A A∈A
⇐⇒ f ( x ) ∈ A for some A ∈ A
⇐⇒ x ∈ f −1 ( A) for some A ∈ A
[
⇐⇒ x ∈ f −1 ( A ).
A∈A
S S
Thus f −1 (A∈A A ) = A∈A f −1 ( A ),
which proves (b).
Part (c) is proved in the same fashion as (b), with unions replaced by intersections
and for some replaced by for every.

2.34 inverse image of a composition

Suppose f : X → Y and g : Y → W are functions. Then



( g ◦ f ) −1 ( A ) = f −1 g −1 ( A )

for every A ⊂ W.

Proof Suppose A ⊂ W. For x ∈ X we have



x ∈ ( g ◦ f )−1 ( A) ⇐⇒ ( g ◦ f )( x ) ∈ A ⇐⇒ g f ( x ) ∈ A
⇐⇒ f ( x ) ∈ g−1 ( A)

⇐⇒ x ∈ f −1 g−1 ( A) .

Thus ( g ◦ f )−1 ( A) = f −1 g−1 ( A) .
Section 2B Measurable Spaces and Functions 31

Measurable Functions
The next definition tells us which real-valued functions behave reasonably with
respect to a σ-algebra on their domain.

2.35 Definition measurable function

Suppose ( X, S) is a measurable space. A function f : X → R is called


S -measurable (or just measurable if S is clear from the context) if

f −1 ( B ) ∈ S

for every Borel set B ⊂ R.

2.36 Example measurable functions


• If S = {∅, X }, then the only S -measurable functions from X to R are the
constant functions.
• If S is the set of all subsets of X, then every function from X to R is S -
measurable.
• If S = {∅, (−∞, 0), [0, ∞), R} (which is a σ-algebra on R), then a function
f : R → R is S -measurable if and only if f is constant on (−∞, 0) and f is
constant on [0, ∞).
Another class of examples comes from characteristic functions, which are defined
below. The Greek letter chi (χ) is traditionally used to denote a characteristic function.

2.37 Definition characteristic function; χ E

Suppose E is a subset of a set X. The characteristic function of E is the function


χ E : X → R defined by (
1 if x ∈ E,
χ E( x ) =
0 if x ∈
/ E.

The set X that contains E is not explicitly included in the notation χ E because X will
always be clear from the context.

2.38 Example inverse image with respect to a characteristic function


Suppose ( X, S) is a measurable space, E ⊂ X, and B ⊂ R. Then


 E if 0 ∈
/ B and 1 ∈ B,

 X \ E if 0 ∈ B and 1 ∈
/ B,
χ E −1 ( B ) =


 X if 0 ∈ B and 1 ∈ B,

∅ if 0 ∈
/ B and 1 ∈
/ B.
Thus we see that χ E is an S -measurable function if and only if E ∈ S .
32 Chapter 2 Measures

The definition of an S -measurable


Note that if f : X → R is a function and
function requires the inverse image
a ∈ R, then
of every Borel subset of R to be in
S . The next result shows that to ver- 
f −1 ( a, ∞) = { x ∈ X : f ( x ) > a}.
ify that a function is S -measurable,
we can check the inverse images of
a much smaller collection of subsets
of R.

2.39 condition for measurable function

Suppose ( X, S) is a measurable space and f : X → R is a function such that



f −1 ( a, ∞) ∈ S

for all a ∈ R. Then f is an S -measurable function.

Proof Let
T = { A ⊂ R : f −1 ( A) ∈ S}.
We want to show that every Borel subset of R is in T . To do this, we will first show
that T is a σ-algebra on R.
Certainly ∅ ∈ T , because f −1 (∅) = ∅ ∈ S .
If A ∈ T , then f −1 ( A) ∈ S ; hence

f −1 ( R \ A ) = X \ f −1 ( A ) ∈ S

by 2.33(a), and thus R \ A ∈ T . In other words, T is closed under complementation.


If A1 , A2 , . . . ∈ T , then f −1 ( A1 ), f −1 ( A2 ), . . . ∈ S ; hence

[  ∞
[
f −1 Ak = f −1 ( A k ) ∈ S
k =1 k =1
S∞
by 2.33(b), and thus k=1 Ak ∈ T . In other words, T is closed under countable
unions. Thus T is a σ-algebra on R.
By hypothesis, T contains {( a, ∞) : a ∈ R}. Because T is closed under
complementation, T also contains {(−∞, b] : b ∈ R}. Because the σ-algebra T is
closed under finiteSintersections (by 2.25), we see that
S∞
T contains {( a, b] : a, b ∈ R}.
Because ( a, b) = ∞ k =1 ( a, b − 1
k ] and (− ∞, b ) = 1
k=1 (− k, b − k ] and T is closed
under countable unions, we can conclude that T contains every open subset of R.
Thus the σ-algebra T contains the smallest σ-algebra on R that contains all open
subsets of R. In other words, T contains every Borel subset of R. Thus f is an
S -measurable function.

In the result above, we could replace the collection of sets {( a, ∞) : a ∈ R}


by any collection of subsets of R such that the smallest σ-algebra containing that
collection contains the Borel subsets of R. For specific examples of such collections
of subsets of R, see Exercises 3–6.
Section 2B Measurable Spaces and Functions 33

We have been dealing with S -measurable functions from X to R in the context


of an arbitrary set X and a σ-algebra S on X. An important special case of this
setup is when X is a Borel subset of R and S is the set of Borel subsets of R that are
contained in X (see Exercise 11 for another way of thinking about this σ-algebra). In
this special case, the S -measurable functions are called Borel measurable.

2.40 Definition Borel measurable function

Suppose X ⊂ R. A function f : X → R is called Borel measurable if f −1 ( B) is


a Borel set for every Borel set B ⊂ R.

If X ⊂ R and there exists a Borel measurable function f : X → R, then X must


be a Borel set [because X = f −1 (R)].
If X ⊂ R and f : X → R is a function, then f is a Borel measurable function if
and only if f −1 ( a, ∞) is a Borel set for every a ∈ R (use 2.39).
Suppose X is a set and f : X → R is a function. The measurability of f depends
upon the choice of a σ-algebra on X. If the σ-algebra is called S , then we can discuss
whether f is an S -measurable function. If X is a Borel subset of R, then S might
be the set of Borel sets contained in X, in which case the phrase Borel measurable
means the same as S -measurable. However, whether or not S is a collection of Borel
sets, we consider inverse images of Borel subsets of R when determining whether a
function is S -measurable.
The next result states that continuity interacts well with the notion of Borel
measurability.

2.41 every continuous function is Borel measurable

Every continuous real-valued function defined on a Borel subset of R is a Borel


measurable function.

Proof Suppose X ⊂ R is a Borel set and f : X → R is continuous. To prove that f


is Borel measurable, fix a ∈ R.
If x ∈ X and f ( x ) > a, then (by the continuity of f ) there exists δx > 0 such that
f (y) > a for all y ∈ ( x − δx , x + δx ) ∩ X. Thus
  [ 
f −1 ( a, ∞) = ( x − δx , x + δx ) ∩ X.

x ∈ f −1 ( a, ∞)

The union inside the large parentheses above is an open subset of R; hence its
intersection with X is a Borel set. Thus we can conclude that f −1 ( a, ∞) is a Borel
set.
Now 2.39 implies that f is a Borel measurable function.

Next we come to another class of Borel measurable functions. A similar definition


could be made for decreasing functions, with a corresponding similar result.
34 Chapter 2 Measures

2.42 Definition increasing function

Suppose X ⊂ R and f : X → R is a function.

• f is called increasing if f ( x ) ≤ f (y) for all x, y ∈ X with x < y.


• f is called strictly increasing if f ( x ) < f (y) for all x, y ∈ X with x < y.

2.43 every increasing function is Borel measurable

Every increasing function defined on a Borel subset of R is a Borel measurable


function.

Proof Suppose X ⊂ R is a Borel set and f : X → R is increasing. To prove that f


is Borel measurable, fix a ∈ R.
Let b = inf f −1 ( a, ∞) . Then it is easy to see that
 
f −1 ( a, ∞) = (b, ∞) ∩ X or f −1 ( a, ∞) = [b, ∞) ∩ X.

Either way, we can conclude that f −1 ( a, ∞) is a Borel set.
Now 2.39 implies that f is a Borel measurable function.

The next result shows that measurability interacts well with composition.

2.44 composition of measurable functions

Suppose ( X, S) is a measurable space and f : X → R is an S -measurable


function. Suppose g is a real-valued Borel measurable function defined on a
subset of R that includes the range of f . Then g ◦ f : X → R is an S -measurable
function.

Proof Suppose B ⊂ R is a Borel set. Then (see 2.34)



( g ◦ f ) −1 ( B ) = f −1 g −1 ( B ) .

Because g is a Borel measurable function, g−1 ( B) is a Borel subset of R. Because f


is an S -measurable function, f −1 g−1 ( B) ∈ S . Thus the equation above implies
that ( g ◦ f )−1 ( B) ∈ S . Thus g ◦ f is an S -measurable function.

2.45 Example if f is measurable, then so are − f , 12 f , | f |, f 2


Suppose ( X, S) is a measurable space and f : X → R is S -measurable. Then 2.44
implies that the functions − f , 12 f , | f |, f 2 are all S -measurable functions because
each of these functions can be written as the composition of f with a continuous (and
thus Borel measurable) function g.
Specifically, take g( x ) = − x, then g( x ) = 21 x, then g( x ) = | x |, and then
g( x ) = x2 .
Section 2B Measurable Spaces and Functions 35

Measurability also interacts well with algebraic operations, as shown in the next
result.

2.46 algebraic operations with measurable functions

Suppose ( X, S) is a measurable space and f , g : X → R are S -measurable. Then

(a) f + g, f − g, and f g are S -measurable functions;


f
(b) if g( x ) 6= 0 for all x ∈ X, then g is an S -measurable function.

Proof Suppose a ∈ R. We will show that


 [  
2.47 ( f + g)−1 ( a, ∞) = f −1 (r, ∞) ∩ g−1 ( a − r, ∞) ,
r ∈Q

which implies that ( f + g)−1 ( a, ∞) ∈ S .
To prove 2.47, first suppose

x ∈ ( f + g)−1 ( a, ∞) .

Thus a < f ( x ) + g( x ). Hence the open interval a − g( x ), f ( x ) is nonempty, and
thus it contains some rational number r. This implies that r < f ( x ), which means 
that x ∈ f −1 (r, ∞) , and a − g( x ) < r, which implies that x ∈ g−1 ( a − r, ∞) .
Thus x is an element of the right side of 2.47, completing the proof that the left side
of 2.47 is contained in the right side.
The proof of the inclusion in theother direction is easier. Specifically, suppose
x ∈ f −1 (r, ∞) ∩ g−1 ( a − r, ∞) for some r ∈ Q. Thus
r < f (x) and a − r < g ( x ).
Adding these two inequalities, we see that a < f ( x ) + g( x ). Thus x is an element of
the left side of 2.47, completing the proof of 2.47. Hence f + g is an S -measurable
function.
Example 2.45 tells us that − g is an S -measurable function. Thus f − g, which
equals f + (− g) is an S -measurable function.
The easiest way to prove that f g is an S -measurable function uses the equation
( f + g )2 − f 2 − g2
fg = .
2
The operation of squaring an S -measurable function produces an S -measurable
function (see Example 2.45), as does the operation of multiplication by 12 (again, see
Example 2.45). Thus the equation above implies that f g is an S -measurable function,
completing the proof of (a).
Suppose g( x ) 6= 0 for all x ∈ X. The function defined on R \ {0} (a Borel subset
of R) that takes x to 1x is continuous and thus is a Borel measurable function (by
2.41). Now 2.44 implies that 1g is an S -measurable function. Combining this result
with what we have already proved about the product of S -measurable functions, we
f
conclude that g is an S -measurable function, proving (b).
36 Chapter 2 Measures

The next result shows that the pointwise limit of a sequence of S -measurable
functions is S -measurable. This is a highly desirable property (recall that the set of
Riemann integrable functions on some interval is not closed under taking pointwise
limits; see Example 1.17).

2.48 limit of S -measurable functions

Suppose ( X, S) is a measurable space and f 1 , f 2 , . . . is a sequence of


S -measurable functions from X to R. Suppose limk→∞ f k ( x ) exists for each
x ∈ X. Define f : X → R by

f ( x ) = lim f k ( x ).
k→∞

Then f is an S -measurable function.

Proof Suppose a ∈ R. We will show that


 ∞ [
[ ∞ \
∞ 
2.49 f −1 ( a, ∞) = f k −1 ( a + 1j , ∞) ,
j =1 m =1 k = m

which implies that f −1 ( a, ∞) ∈ S .

To prove 2.49, first suppose x ∈ f −1 ( a, ∞) . Thus there exists j ∈ Z+ such that
f ( x ) > a + 1j . The definition of limit now implies that there exists m ∈ Z+ such
that f k ( x ) > a + 1j for all k ≥ m. Thus x is in the right side of 2.49, proving that the
left side of 2.49 is contained in the right side.
To prove the inclusion in the other direction, suppose x is in the right side of 2.49.
Thus there exist j, m ∈ Z+ such that f k ( x ) > a + 1j for all k ≥ m. Taking the
limit as k → ∞, we see that f ( x ) ≥ a + 1j > a. Thus x is in the left side of 2.49,
completing the proof of 2.49. Thus f is an S -measurable function.

Occasionally we need to consider functions that take values in [−∞, ∞]. For
example, even if we start with a sequence of real-valued functions in 2.53, we might
end up with functions with values in [−∞, ∞]. Thus we extend the notion of Borel
sets to subsets of [−∞, ∞], as follows.

2.50 Definition Borel subsets of [−∞, ∞]

A subset of [−∞, ∞] is called a Borel set if its intersection with R is a Borel set.

In other words, a set C ⊂ [−∞, ∞] is a Borel set if and only if there exists
a Borel set B ⊂ R such that C = B or C = B ∪ {∞} or C = B ∪ {−∞} or
C = B ∪ {∞, −∞}.
You should verify that with the definition above, the set of Borel subsets of
[−∞, ∞] is a σ-algebra on [−∞, ∞].
Next, we extend the definition of S -measurable functions to functions taking
values in [−∞, ∞].
Section 2B Measurable Spaces and Functions 37

2.51 Definition measurable function

Suppose ( X, S) is a measurable space. A function f : X → [−∞, ∞] is called


S -measurable if
f −1 ( B ) ∈ S
for every Borel set B ⊂ [−∞, ∞].

The next result, which is analogous to 2.39, states that we need not consider all
Borel subsets of [−∞, ∞] when taking inverse images to determine whether or not a
function with values in [−∞, ∞] is S -measurable.

2.52 condition for measurable function

Suppose ( X, S) is a measurable space and f : X → [−∞, ∞] is a function such


that 
f −1 ( a, ∞] ∈ S
for all a ∈ R. Then f is an S -measurable function.

The proof of the result above is left to the reader (also see Exercise 27 in this
section).
We end this section by showing that the pointwise infimum and pointwise supre-
mum of a sequence of S -measurable functions is S -measurable.

2.53 infimum and supremum of a sequence of S -measurable functions

Suppose ( X, S) is a measurable space and f 1 , f 2 , . . . is a sequence of


S -measurable functions from X to [−∞, ∞]. Define g, h : X → [−∞, ∞] by

g( x ) = inf{ f k ( x ) : k ∈ Z+} and h( x ) = sup{ f k ( x ) : k ∈ Z+}.

Then g and h are S -measurable functions.

Proof Let a ∈ R. The definition of the supremum implies that

 ∞
[ 
h−1 ( a, ∞] = f k −1 ( a, ∞] ,
k =1

as you should verify. The equation above, along with 2.52, implies that h is an
S -measurable function.
Note that
g( x ) = − sup{− f k ( x ) : k ∈ Z+}
for all x ∈ X. Thus the result about the supremum implies that g is an S -measurable
function.
38 Chapter 2 Measures

EXERCISES 2B
S
1 Show that S = { n∈K ( n, n + 1] : K ⊂ Z} is a σ-algebra on R.
2 Verify both bullet points in Example 2.28.
3 Suppose S is the smallest σ-algebra on R containing {(r, s] : r, s ∈ Q}. Prove
that S is the collection of Borel subsets of R.
4 Suppose S is the smallest σ-algebra on R containing {(r, n] : r ∈ Q, n ∈ Z}.
Prove that S is the collection of Borel subsets of R.
5 Suppose S is the smallest σ-algebra on R containing {(r, r + 1) : r ∈ Q}.
Prove that S is the collection of Borel subsets of R.
6 Suppose S is the smallest σ-algebra on R containing {[r, ∞) : r ∈ Q}. Prove
that S is the collection of Borel subsets of R.
7 Prove that the collection of Borel subsets of R is translation invariant. More
precisely, prove that if B ⊂ R is a Borel set and t ∈ R, then t + B is a Borel set.
8 Prove that the collection of Borel subsets of R is dilation invariant. More
precisely, prove that if B ⊂ R is a Borel set and t ∈ R, then tB (which is
defined to be {tb : b ∈ B}) is a Borel set.
9 Give an example of a measurable space ( X, S) and a function f : X → R such
that | f | is S -measurable but f is not S -measurable.
10 Show that the set of real numbers that have a decimal expansion with the digit 5
appearing infinitely often is a Borel set.
11 Suppose T is a σ-algebra on a set Y and X ∈ T . Let S = { E ∈ T : E ⊂ X }.
(a) Show that S = { F ∩ X : F ∈ T }.
(b) Show that S is a σ-algebra on X.

12 Suppose f : R → R is a function.
(a) For k ∈ Z+, let
1
Gk = { a ∈ R : there exists δ > 0 such that | f (b) − f (c)| < k
for all b, c ∈ ( a − δ, a + δ)}.

Prove that Gk is an open subset of R for each k ∈ Z+.


T∞
(b) Prove that the set of points at which f is continuous equals k =1 Gk .
(c) Conclude that the set of points at which f is continuous is a Borel set.

13 Suppose ( X, S) is a measurable space, E1 , . . . , En are disjoint subsets of X, and


c1 , . . . , cn are distinct nonzero real numbers. Prove that c1 χ E + · · · + cn χ E is
1 n
an S -measurable function if and only if E1 , . . . , En ∈ S .
Section 2B Measurable Spaces and Functions 39

14 (a) Suppose f 1 , f 2 , . . . is a sequence of functions from a set X to R. Explain


why

{ x ∈ X : the sequence f 1 ( x ), f 2 ( x ), . . . has a limit in R}


∞ [
\ ∞ \
∞ 
= ( f j − f k )−1 (− n1 , n1 ) .
n =1 j =1 k = j

(b) Suppose ( X, S) is a measurable space and f 1 , f 2 , . . . is a sequence of S -


measurable functions from X to R. Prove that

{ x ∈ X : the sequence f 1 ( x ), f 2 ( x ), . . . has a limit in R}

is an S -measurable subset of X.
15 Suppose
S
X is a set and E1 , E2 ,S. . . is a disjoint sequence of subsets of X such
that ∞k=1 Ek = X. Let S = { k∈K Ek : K ⊂ Z }.
+

(a) Show that S is a σ-algebra on X.


(b) Prove that a function from X to R is S -measurable if and only if the function
is constant on Ek for every k ∈ Z+.

16 Suppose S is a σ-algebra on a set X and A ⊂ X. Let

S A = { E ∈ S : A ⊂ E or A ∩ E = ∅}.

(a) Prove that S A is a σ-algebra on X.


(b) Suppose f : X → R is a function. Prove that f is measurable with respect
to S A if and only if f is measurable with respect to S and f is constant
on A.

17 Suppose X is a Borel subset of R and f : X → R is a function such that


{ x ∈ X : f is not continuous at x } is a countable set. Prove f is a Borel
measurable function.
18 Suppose f : R → R is differentiable at every element of R. Prove that f 0 is a
Borel measurable function from R to R.
19 Suppose X is a nonempty set and S is the σ-algebra on X consisting of all
subsets of X that are either countable or have a countable complement in X.
Give a characterization of the S -measurable real-valued functions on X.
20 Suppose ( X, S) is a measurable space and f , g : X → R are S -measurable
functions. Prove that if f ( x ) > 0 for all x ∈ X, then f g (which is the function
whose value at x ∈ X equals f ( x ) g( x) ) is an S -measurable function.
21 Prove 2.52.
22 Suppose B ⊂ R and f : B → R is an increasing function. Prove that f is
continuous at every element of B except for a countable subset of B.
40 Chapter 2 Measures

23 Suppose f : R → R is a strictly increasing function. Prove that the inverse


function f −1 : f (R) → R is a continuous function.
[Note that this exercise does not have as a hypothesis that f is continuous.]
24 Suppose B ⊂ R is a Borel set and f : B → R is an increasing function. Prove
that f ( B) is a Borel set.
25 Suppose B ⊂ R and f : B → R is an increasing function. Prove that there exists
a sequence f 1 , f 2 , . . . of strictly increasing functions from B to R such that

f ( x ) = lim f k ( x )
k→∞

for every x ∈ B.
26 Suppose B ⊂ R and f : B → R is a bounded increasing function. Prove that
there exists an increasing function g : R → R such that g( x ) = f ( x ) for all
x ∈ B.
27 Prove or give a counterexample: If ( X, S) is a measurable space and

f : X → [−∞, ∞]

is a function such that f −1 ( a, ∞) ∈ S for every a ∈ R, then f is an
S -measurable function.
28 Suppose f : B → R is a Borel measurable function. Define g : R → R by
(
f ( x ) if x ∈ B,
g( x ) =
0 if x ∈ R \ B.

Prove that g is a Borel measurable function.


29 Give an example of a measurable space ( X, S) and a family { f t }t∈R such
that each f t is an S -measurable function from X to [0, 1], but the function
f : X → [0, 1] defined by

f ( x ) = sup{ f t ( x ) : t ∈ R}

is not S -measurable.
[Compare this exercise to 2.53, where the index set is Z+ rather than R.]
30 Show that
(
 2k  1 if x is rational,
lim lim cos( j!πx ) =
j→∞ k→∞ 0 if x is irrational

for every x ∈ R.
[This example is due to Henri Lebesgue.]
Section 2C Measures and Their Properties 41

2C Measures and Their Properties


Definition and Examples of Measures
The original motivation for the next definition came from trying to extend the notion
of the length of an interval. However, the definition below allows us to discuss size
in many more contexts. For example, we will see later that the area of a set in the
plane or the volume of a set in higher dimensions fits into this structure. The word
measure allows us to use a single word instead of repeating theorems for length, area,
and volume.

2.54 Definition measure

Suppose X is a set and S is a σ-algebra on X. A measure on ( X, S) is a function


µ : S → [0, ∞] such that µ(∅) = 0 and

[  ∞
µ Ek = ∑ µ(Ek )
k =1 k =1

for every disjoint sequence E1 , E2 , . . . of sets in S .

The countable additivity that forms the key


In the mathematical literature,
part of the definition above allows us to prove
sometimes a measure on ( X, S)
good limit theorems. Note that countable ad-
is just called a measure on X if
ditivity implies finite additivity: if µ is a mea-
the σ-algebra S is clear from
sure on ( X, S) and E1 , . . . , En are disjoint
the context.
sets in S , then
The concept of a measure, as
µ( E1 ∪ · · · ∪ En ) = µ( E1 ) + · · · + µ( En ), defined here, is sometimes called
a positive measure (although the
as follows from applying the equation
phrase nonnegative measure
µ(∅) = 0 and countable additivity to the dis-
would be more accurate).
joint sequence E1 , . . . , En , ∅, ∅, . . . of sets
in S .

2.55 Example measures

• If X is a set, then counting measure is the measure µ defined on the σ-algebra


of all subsets of X by setting µ( E) = n if E is a finite set containing exactly n
elements and µ( E) = ∞ if E is not a finite set.
• Suppose X is a set, S is a σ-algebra on X, and c ∈ X. Define the Dirac measure
δc on ( X, S) by (
1 if c ∈ E,
δc ( E) =
0 if c ∈
/ E.
This measure is named in honor of mathematician and physicist Paul Dirac (1902–
1984), who won the Nobel Prize for Physics in 1933 for his work combining
relativity and quantum mechanics at the atomic level.
42 Chapter 2 Measures

• Suppose X is a set, S is a σ-algebra on X, and w : X → [0, ∞] is a function.


Define a measure µ on ( X, S) by

µ( E) = ∑ w( x )
x∈E

for E ∈ S . [Here the sum is defined as the supremum of all finite subsums
∑ x∈ D w( x ) as D ranges over all finite subsets of E.]
• Suppose X is a set and S is the σ-algebra on X consisting of all subsets of X
that are either countable or have a countable complement in X. Define a measure
µ on ( X, S) by (
0 if E is countable,
µ( E) =
3 if E is uncountable.

• Suppose S is the σ-algebra on R consisting of all subsets of R. Then the function


that takes a set E ⊂ R to | E| (the outer measure of E) is not a measure because
it is not finitely additive (see 2.18).
• Suppose B is the σ-algebra on R consisting of all Borel subsets of R. We will
see in the next section that outer measure is a measure on (R, B).

The following terminology is frequently useful.

2.56 Definition measure space

A measure space is an ordered triple ( X, S , µ), where X is a set, S is a σ-algebra


on X, and µ is a measure on ( X, S).

Properties of Measures
The hypothesis that µ( D ) < ∞ is needed in part (b) of the next result to avoid
undefined expressions of the form ∞ − ∞.

2.57 measure preserves order; measure of a set difference

Suppose ( X, S , µ) is a measure space and D, E ∈ S are such that D ⊂ E. Then

(a) µ( D ) ≤ µ( E);
(b) µ( E \ D ) = µ( E) − µ( D ) provided that µ( D ) < ∞.

Proof Because E = D ∪ ( E \ D ) and this is a disjoint union, we have

µ ( E ) = µ ( D ) + µ ( E \ D ) ≥ µ ( D ),

which proves (a). If µ( D ) < ∞, then subtracting µ( D ) from both sides of the
equation above proves (b).
Section 2C Measures and Their Properties 43

The countable additivity property of measures applies to disjoint countable unions.


The following countable subadditivity property applies to countable unions that may
not be disjoint unions.

2.58 countable subadditivity

Suppose ( X, S , µ) is a measure space and E1 , E2 , . . . ∈ S . Then



[  ∞
µ Ek ≤ ∑ µ(Ek ).
k =1 k =1

Proof Let D1 = ∅ and Dk = E1 ∪ · · · ∪ Ek−1 for k ≥ 2. Then


E1 \ D1 , E2 \ D2 , E3 \ D3 , . . .
S∞
is a disjoint sequence of subsets of X whose union equals k =1 Ek . Thus

[  ∞
[ 
µ Ek = µ ( Ek \ Dk )
k =1 k =1

= ∑ µ(Ek \ Dk )
k =1

≤ ∑ µ(Ek ),
k =1

where the second line above follows from the countable additivity of µ and the last
line above follows from 2.57(a).

Note that countable subadditivity implies finite subadditivity: if µ is a measure on


( X, S) and E1 , . . . , En are sets in S , then
µ( E1 ∪ · · · ∪ En ) ≤ µ( E1 ) + · · · + µ( En ),
as follows from applying the equation µ(∅) = 0 and countable subadditivity to the
sequence E1 , . . . , En , ∅, ∅, . . . of sets in S .
The next result shows that measures behave well with increasing unions.

2.59 measure of an increasing union

Suppose ( X, S , µ) is a measure space and E1 ⊂ E2 ⊂ · · · is an increasing


sequence of sets in S . Then

[ 
µ Ek = lim µ( Ek ).
k→∞
k =1

Proof If µ( Ek ) = ∞ for some k ∈ Z+, then the equation above holds because
both sides equal ∞. Hence we can consider only the case where µ( Ek ) < ∞ for all
k ∈ Z+.
44 Chapter 2 Measures

For convenience, let E0 = ∅. Then



[ ∞
[
Ek = ( E j \ E j −1 ),
k =1 j =1

where the union on the right side is a disjoint union. Thus



[  ∞
µ Ek = ∑ µ ( E j \ E j −1 )
k =1 j =1
k
= lim
k→∞
∑ µ ( E j \ E j −1 )
j =1

k 
= lim
k→∞
∑ µ ( E j ) − µ ( E j −1 )
j =1

= lim µ( Ek ),
k→∞
Another mew.
as desired.

Measures also behave well with respect to decreasing intersections (but see Exer-
cise 10, which shows that the hypothesis µ( E1 ) < ∞ below cannot be deleted).

2.60 measure of a decreasing intersection

Suppose ( X, S , µ) is a measure space and E1 ⊃ E2 ⊃ · · · is a decreasing


sequence of sets in S , with µ( E1 ) < ∞. Then

\ 
µ Ek = lim µ( Ek ).
k→∞
k =1

Proof One of De Morgan’s Laws tells us that



\ ∞
[
E1 \ Ek = ( E1 \ Ek ).
k =1 k =1

Now E1 \ E1 ⊂ E1 \ E2 ⊂ E1 \ E3 ⊂ · · · is an increasing sequence of sets in S .


Thus 2.59, applied to the equation above, implies that
 \∞ 
µ E1 \ Ek = lim µ( E1 \ Ek ).
k→∞
k =1
Use 2.57(b) to rewrite the equation above as
\∞ 
µ( E1 ) − µ Ek = µ( E1 ) − lim µ( Ek ),
k→∞
k =1
which implies our desired result.
Section 2C Measures and Their Properties 45

The next result is intuitively plausible—we expect that the measure of the union of
two sets equals the measure of the first set plus the measure of the second set minus
the measure of the set that has been counted twice.

2.61 measure of a union

Suppose ( X, S , µ) is a measure space and D, E ∈ S , with µ( D ∩ E) < ∞. Then

µ ( D ∪ E ) = µ ( D ) + µ ( E ) − µ ( D ∩ E ).

Proof We have
  
D ∪ E = D \ ( D ∩ E) ∪ E \ ( D ∩ E) ∪ D ∩ E .

The right side of the equation above is a disjoint union. Thus


  
µ( D ∪ E) = µ D \ ( D ∩ E) + µ E \ ( D ∩ E) + µ D ∩ E
 
= µ( D ) − µ( D ∩ E) + µ( E) − µ( D ∩ E) + µ( D ∩ E)
= µ ( D ) + µ ( E ) − µ ( D ∩ E ),

as desired.

EXERCISES 2C
1 Explain why there does not exist a measure space ( X, S , µ) with the property
that {µ( E) : E ∈ S} = [0, 1).

+
Let 2Z denote the σ-algebra on Z+ consisting of all subsets of Z+.
+
2 Suppose µ is a measure on (Z+ , 2Z ). Prove that there is a sequence w1 , w2 , . . .
in [0, ∞] such that
µ ( E ) = ∑ wk
k∈ E

for every set E ⊂ Z+.


+
3 Give an example of a measure µ on (Z+ , 2Z ) such that

{µ( E) : E ⊂ Z+} = [0, 1].

4 Give an example of a measure space ( X, S , µ) such that



[
{µ( E) : E ∈ S} = {∞} ∪ [3k, 3k + 1].
k =0
46 Chapter 2 Measures

5 Suppose ( X, S , µ) is a measure space such that µ( X ) < ∞. Prove that if A is


a set of disjoint sets in S such that µ( A) > 0 for every A ∈ A, then A is a
countable set.
6 Find all c ∈ [3, ∞) such that there exists a measure space ( X, S , µ) with

{µ( E) : E ∈ S} = [0, 1] ∪ [3, c].

7 Give an example of a measure space ( X, S , µ) such that

{µ( E) : E ∈ S} = [0, 1] ∪ [3, ∞].

8 Give an example of a set X, a σ-algebra S of subsets of X, a set A of subsets of X


such that the smallest σ-algebra on X containing A is S , and two measures µ and
ν on ( X, S) such that µ( A) = ν( A) for all A ∈ A and µ( X ) = ν( X ) < ∞,
but µ 6= ν.
9 Suppose µ and ν are measures on a measurable space ( X, S). Prove that µ + ν
is a measure on ( X, S). [Here µ + ν is the usual sum of two functions: if E ∈ S ,
then (µ + ν)( E) = µ( E) + ν( E).]
10 Give an example of a measure space ( X, S , µ) and a decreasing sequence
E1 ⊃ E2 ⊃ · · · of sets in S such that

\ 
µ Ek 6= lim µ( Ek ).
k→∞
k =1

11 Suppose ( X, S , µ) is a measure space and C, D, E ∈ S are such that

µ(C ∩ D ) < ∞, µ(C ∩ E) < ∞, and µ( D ∩ E) < ∞.

Find and prove a formula for µ(C ∪ D ∪ E) in terms of µ(C ), µ( D ), µ( E),


µ(C ∩ D ), µ(C ∩ E), µ( D ∩ E), and µ(C ∩ D ∩ E).
12 Suppose X is a set and S is the σ-algebra of all subsets E of X such that E is
countable or X \ E is countable. Give a complete description of the set of all
measures on ( X, S).
Section 2D Lebesgue Measure 47

2D Lebesgue Measure
Additivity of Outer Measure on Borel Sets
Recall that there exist disjoint sets A, B ∈ R such that | A ∪ B| 6= | A| + | B| (see
2.18). Thus outer measure, despite its name, is not a measure on the σ-algebra of all
subsets of R.
Our main goal in this section is to prove that outer measure, when restricted to the
Borel subsets of R, is a measure. Throughout this section, be careful about trying to
simplify proofs by applying properties of measures to outer measure, even if those
properties seem intuitively plausible. For example, there are subsets A ⊂ B ⊂ R
with | A| < ∞ but | B \ A| 6= | B| − | A| [compare to 2.57(b)].
The next result is our first step toward the goal of proving that outer measure
restricted to the Borel sets is a measure.

2.62 additivity of outer measure if one of the sets is open

Suppose A and G are disjoint subsets of R and G is open. Then

| A ∪ G | = | A | + | G |.

Proof We can assume that | G | < ∞ because otherwise both | A ∪ G | and | A| + | G |


equal ∞.
Subadditivity (see 2.8) implies that | A ∪ G | ≤ | A| + | G |. Thus we need to prove
the inequality only in the other direction.
First consider the case where G = ( a, b) for some a, b ∈ R with a < b. We
can assume that a, b ∈/ A (because changing a set by at most two points does not
change its outer measure). Let I1 , I2 , . . . be a sequence of open intervals whose union
contains A ∪ G. For each n ∈ Z+, let

Jn = In ∩ (−∞, a), Kn = In ∩ ( a, b), Ln = In ∩ (b, ∞).

Then
`( In ) = `(Jn ) + `(Kn ) + `( Ln ).
Now J1 , L1 , J2 , L2 , . . . is a sequence of open intervals whose union contains A and
K1 , K2 , . . . is a sequence of open intervals whose union contains G. Thus
∞ ∞  ∞
∑ `( In ) = ∑ `(Jn ) + `( Ln ) + ∑ `(Kn )
n =1 n =1 n =1
≥ | A | + | G |.
The inequality above implies that | A ∪ G | ≥ | A| + | G |, completing the proof that
| A ∪ G | = | A| + | G | in this special case.
Using induction on m, we can now conclude that if m ∈ Z+ and G is a union of
m disjoint open intervals that are all disjoint from A, then | A ∪ G | = | A| + | G |.
Now
S
suppose G is an arbitrary open subset of R that is disjoint from A. Then
G= ∞ n=1 In for some sequence of disjoint open intervals I1 , I2 , . . ., each of which
is disjoint from A. Now for each m ∈ Z+ we have
48 Chapter 2 Measures

m
[ 
| A ∪ G| ≥ A ∪ In
n =1
m
= | A| + ∑ `( In ).
n =1

Thus

| A ∪ G | ≥ | A| + ∑ `( In )
n =1
≥ | A | + | G |,
completing the proof that | A ∪ G | = | A| + | G |.

The next result shows that the outer measure of the disjoint union of two sets is
what we expect if at least one of the two sets is closed.

2.63 additivity of outer measure if one of the sets is closed

Suppose A and F are disjoint subsets of R and F is closed. Then

| A ∪ F | = | A | + | F |.

Proof Suppose I1 , I2 , . . . is a sequence of open intervals whose union contains A ∪ F.


S
Let G = ∞ k=1 Ik . Thus G is an open set with A ∪ F ⊂ G. Hence A ⊂ G \ F, which
implies that
2.64 | A | ≤ | G \ F |.
Because G \ F = G ∩ (R \ F ), we know that G \ F is an open set. Hence we can
apply 2.62 to the disjoint union G = F ∪ ( G \ F ), getting
| G | = | F | + | G \ F |.
Adding | F | to both sides of 2.64 and then using the equation above gives
| A| + | F | ≤ | G |

≤ ∑ `( Ik ).
k =1

Thus | A| + | F | ≤ | A ∪ F |, which implies that | A| + | F | = | A ∪ F |.

Recall that the collection of Borel sets is the smallest σ-algebra on R that con-
tains all open subsets of R. The next result provides an extremely useful tool for
approximating a Borel set by a closed set.

2.65 approximation of Borel sets from below by closed sets

Suppose B ⊂ R is a Borel set. Then for every ε > 0, there exists a closed set
F ⊂ B such that | B \ F | < ε.
Section 2D Lebesgue Measure 49

Proof Let

L = { D ⊂ R : for every ε > 0, there exists a closed set


F ⊂ D such that | D \ F | < ε}.

The strategy of the proof is to show that L is a σ-algebra. Then because L contains
every closed subset of R (if D ⊂ R is closed, take F = D in the definition of L), by
taking complements we can conclude that L contains every open subset of R and
thus every Borel subset of R.
To get started with proving that L is a σ-algebra, we want to prove that L is closed
under countable intersections. Thus suppose D1 , D2 , . . . is a sequence in L. Let
ε > 0. For each k ∈ Z+, there exists a closed set Fk such that
ε
Fk ⊂ Dk and | Dk \ Fk | < .
2k
T∞
Thus k=1 Fk is a closed set and

\ ∞
\ ∞
\  ∞
\  ∞
[
Fk ⊂ Dk and Dk \ Fk ⊂ ( Dk \ Fk ).
k =1 k =1 k =1 k =1 k =1

The last set inclusion and the countable subadditivity of outer measure (see 2.8) imply
that
\∞  \∞ 

Dk \ Fk < ε.
k =1 k =1
T
Thus ∞ k=1 Dk ∈ L, proving that L is closed under countable intersections.
Now we want to prove that L is closed under complementation. Suppose D ∈ L
and ε > 0. We want to show that there is a closed subset of R \ D whose set
difference with R \ D has outer measure less than ε, which will allow us to conclude
that R \ D ∈ L.
First we consider the case where | D | < ∞. Let F ⊂ D be a closed set such that
| D \ F | < 2ε . The definition of outer measure implies that there exists an open set G
such that D ⊂ G and | G | < | D | + 2ε . Now R \ G is a closed set and R \ G ⊂ R \ D.
Also, we have

(R \ D ) \ (R \ G ) = G \ D
⊂ G \ F.

Thus

|(R \ D ) \ (R \ G )| ≤ | G \ F |
= |G| − | F|
= (| G | − | D |) + (| D | − | F |)
ε
< + |D \ F|
2
< ε,
50 Chapter 2 Measures

where the equality in the second line above comes from applying 2.63 to the disjoint
union G = ( G \ F ) ∪ F, and the fourth line above uses subadditivity applied to
the union D = ( D \ F ) ∪ F. The last inequality above shows that R \ D ∈ L, as
desired.
Now, still assuming that D ∈ L and ε > 0, we consider the case where | D | = ∞.
For k ∈ Z+, let Dk = D ∩ [−k, k]. Because Dk ∈ L and | Dk | < ∞, the previous
S
case implies that R \ Dk ∈ L. Clearly D = ∞ k=1 Dk . Thus

\
R\D = ( R \ Dk ) .
k =1

Because L is closed under countable intersections, the equation above implies that
R \ D ∈ L, which completes the proof that L is a σ-algebra.

Now we can prove that the outer measure of the disjoint union of two sets is what
we expect if at least one of the two sets is a Borel set.

2.66 additivity of outer measure if one of the sets is a Borel set

Suppose A and B are disjoint subsets of R and B is a Borel set. Then

| A ∪ B | = | A | + | B |.

Proof Let ε > 0. Let F be a closed set such that F ⊂ B and | B \ F | < ε (see 2.65).
Thus

| A ∪ B| ≥ | A ∪ F |
= | A| + | F |
= | A| + | B| − | B \ F |
≥ | A| + | B| − ε,

where the second and third lines above follow from 2.63 [use B = ( B \ F ) ∪ F for
the third line].
Because the inequality above holds for all ε > 0, we have | A ∪ B| ≥ | A| + | B|,
which implies that | A ∪ B| = | A| + | B|.

You have probably long suspected that not every subset of R is a Borel set. Now
we can prove this suspicion.

2.67 existence of a subset of R that is not a Borel set

There exists a set B ⊂ R such that | B| < ∞ and B is not a Borel set.

Proof In the proof of 2.18, we showed that there exist disjoint sets A, B ⊂ R such
that | A ∪ B| 6= | A| + | B|. For any such sets, we must have | B| < ∞ because
otherwise both | A ∪ B| and | A| + | B| equal ∞ (as follows from the inequality
| B| ≤ | A ∪ B|). Now 2.66 implies that B is not a Borel set.
Section 2D Lebesgue Measure 51

The tools we have constructed now allow us to prove that outer measure, when
restricted to the Borel sets, is a measure.

2.68 outer measure is a measure on Borel sets

Outer measure is a measure on (R, B), where B is the σ-algebra of Borel subsets
of R.

Proof Suppose B1 , B2 , . . . is a disjoint sequence of Borel subsets of R. Then for


each n ∈ Z+ we have [∞ [ n

Bk ≥ Bk
k =1 k =1
n
= ∑ | Bk |,
k =1
where the first line above follows from 2.5 and the last line S
follows from 2.66 (and
induction on n). Taking the limit as n → ∞, we have ∞ ∞
k=1 Bk ≥ ∑k=1 | Bk |.
The inequality in the other directions follows from countable subadditivity of outer
measure (2.8). Hence
[∞ ∞

Bk = ∑ | Bk |.
k =1 k =1
Thus outer measure is a measure on the σ-algebra of Borel subsets of R.

The result above implies that the next definition makes sense.

2.69 Definition Lebesgue measure

Lebesgue measure is the measure on (R, B), where B is the σ-algebra of Borel
subsets of R, that assigns to each Borel set its outer measure.

In other words, the Lebesgue measure of a set is the same as its outer measure,
except that the term Lebesgue measure should not be applied to arbitrary sets but
only to Borel sets (and also to what are called Lebesgue measurable sets, as we will
soon see). Unlike outer measure, Lebesgue measure is actually a measure, as shown
in 2.68. Lebesgue measure is named in honor of its inventor, Henri Lebesgue.

The cathedral in Beauvais, the


French city where Henri
Lebesgue (1875–1941) was
born. Much of what we call
Lebesgue measure and
Lebesgue integration was
developed by Lebesgue in his
1902 PhD thesis. Émile Borel
was Lebesgue’s PhD thesis
advisor. CC-BY-SA James Mitchell
52 Chapter 2 Measures

Lebesgue Measurable Sets


We have accomplished the major goal of this section, which was to show that outer
measure restricted to Borel sets is a measure. As we will see in this subsection, outer
measure is actually a measure on a somewhat larger class of sets called the Lebesgue
measurable sets.
The mathematics literature contains many different definitions of a Lebesgue
measurable set. These definitions are all equivalent—the definition of a Lebesgue
measurable set in one approach becomes a theorem in another approach. The ap-
proach chosen here has the advantage of emphasizing that a Lebesgue measurable set
differs from a Borel set by a set with outer measure 0. The attitude here is that sets
with outer measure 0 should be considered small sets that do not matter much.

2.70 Definition Lebesgue measurable set

A set A ⊂ R is called Lebesgue measurable if there exists a Borel set B ⊂ A


such that | A \ B| = 0.

Every Borel set is Lebesgue measurable because if A ⊂ R is a Borel set, then we


can take B = A in the definition above.
The result below gives several equivalent conditions for being Lebesgue measur-
able. The equivalence of (a) and (d) is just our definition and thus is not discussed in
the proof.
Although there exist Lebesgue measurable sets that are not Borel sets, you are
unlikely to encounter one. The most important application of the result below is that
if A ⊂ R is a Borel set, then A satisfies conditions (b), (c), (e), and (f). Condition (c)
implies that every Borel set is almost a countable union of closed sets, and condition
(f) implies that every Borel set is almost a countable intersection of open sets.

2.71 equivalences for being a Lebesgue measurable set

Suppose A ⊂ R. Then the following are equivalent:

(a) A is Lebesgue measurable.


(b) For each ε > 0, there exists a closed set F ⊂ A with | A \ F | < ε.

[

(c) There exist closed sets F1 , F2 , . . . contained in A such that A \ Fk = 0.
k =1

(d) There exists a Borel set B ⊂ A such that | A \ B| = 0.

(e) For each ε > 0, there exists an open set G ⊃ A such that | G \ A| < ε.
 \
∞ 

(f) There exist open sets G1 , G2 , . . . containing A such that Gk \ A = 0.
k =1

(g) There exists a Borel set B ⊃ A such that | B \ A| = 0.


Section 2D Lebesgue Measure 53

Proof Let L denote the collection of sets A ⊂ R that satisfy (b). We have already
proved that every Borel set is in L (see 2.65). As a key part of that proof, which we
will freely use in this proof, we showed that L is a σ-algebra on R (see the proof
of 2.65). In addition to containing the Borel sets, L contains every set with outer
measure 0 [because if | A| = 0, we can take F = ∅ in (b)].
(b) =⇒ (c): Suppose (b) holds. Thus for each n ∈ Z+, there exists a closed set
Fn ⊂ A such that | A \ Fn | < n1 . Now

[
A\ Fk ⊂ A \ Fn
k =1
S
for each
S∞
n ∈ Z+. Thus | A \ ∞ 1
k=1 Fk | ≤ | A \ Fn | ≤ n for each n ∈ Z . Hence
+
| A \ k=1 Fk | = 0, completing the proof that (b) implies (c).
(c) =⇒ (d): Because every countable union of closed sets is a Borel set, we see
that (c) implies (d).
(d) =⇒ (b): Suppose (d) holds. Thus there exists a Borel set B ⊂ A such that
| A \ B| = 0. Now
A = B ∪ ( A \ B ).
We know that B ∈ L (because B is a Borel set) and A \ B ∈ L (because A \ B has
outer measure 0). Because L is a σ-algebra, the displayed equation above implies
that A ∈ L. In other words, (b) holds, completing the proof that (d) implies (b).
At this stage of the proof, we now know that (b) ⇐⇒ (c) ⇐⇒ (d).
(b) =⇒ (e): Suppose (b) holds. Thus A ∈ L. Let ε > 0. Then because
R \ A ∈ L (which holds because L is closed under complementation), there exists a
closed set F ⊂ R \ A such that
|(R \ A) \ F | < ε.
Now R \ F is an open set with R \ F ⊃ A. Because (R \ F ) \ A = (R \ A) \ F,
the inequality above implies that |(R \ F ) \ A| < ε. Thus (e) holds, completing the
proof that (b) implies (e).
(e) =⇒ (f): Suppose (e) holds. Thus for each n ∈ Z+, there exists an open set
Gn ⊃ A such that | Gn \ A| < n1 . Now
\∞ 
Gk \ A ⊂ Gn \ A
k =1
T∞ 
for each n ∈ Z+. Thus | k=1 Gk \ A| ≤ | Gn \ A| ≤ n1 for each n ∈ Z+. Hence
T∞ 
| k=1 Gk \ A| = 0, completing the proof that (e) implies (f).
(f) =⇒ (g): Because every countable intersection of open sets is a Borel set, we
see that (f) implies (g).
(g) =⇒ (b): Suppose (g) holds. Thus there exists a Borel set B ⊃ A such that
| B \ A| = 0. Now 
A = B ∩ R \ ( B \ A) .
We know that B ∈ L (because B is a Borel set) and R \ ( B \ A) ∈ L (because this
set is the complement of a set with outer measure 0). Because L is a σ-algebra, the
displayed equation above implies that A ∈ L. In other words, (b) holds, completing
the proof that (g) implies (b).
Our chain of implications now shows that (b) through (g) are all equivalent.
54 Chapter 2 Measures

In addition to the equivalences in the


In practice, the most useful part of
previous result, see Exercise 13 in this
Exercise 6 is the result that every
section for another condition that is equiv-
Borel set with finite measure is
alent to being Lebesgue measurable. Also
almost a finite disjoint union of
see Exercise 6, which shows that a set
bounded open intervals.
with finite outer measure is Lebesgue mea-
surable if and only if it is almost a finite disjoint union of bounded open intervals.
Now we can show that outer measure is a measure on the Lebesgue measurable
sets.

2.72 outer measure is a measure on Lebesgue measurable sets

(a) The set L of Lebesgue measurable subsets of R is a σ-algebra on R.


(b) Outer measure is a measure on (R, L).

Proof Because (a) and (b) are equivalent in 2.71, the set L of Lebesgue measurable
subsets of R is the collection of sets satisfying (b) in 2.71. As noted in the first
paragraph of the proof of 2.71, this set is a σ-algebra on R, proving (a).
To prove the second bullet point, suppose A1 , A2 , . . . is a disjoint sequence of
Lebesgue measurable sets. By the definition of Lebesgue measurable set (2.70), for
each k ∈ Z+ there exists a Borel set Bk ⊂ Ak such that | Ak \ Bk | = 0. Now
[∞ [ ∞

Ak ≥ Bk
k =1 k =1


= ∑ | Bk |
k =1


= ∑ | A k |,
k =1

where the second line above holds because B1 , B2 , . . . is a disjoint sequence of Borel
sets and outer measure is a measure on the Borel sets (see 2.68); the last line above
holds because Bk ⊂ Ak and by subadditivity of outer measure (see 2.8) we have
| Ak | = | Bk ∪ ( Ak \ Bk )| ≤ | Bk | + | Ak \ Bk | = | Bk |.
The inequality above, S combined with countable subadditivity of outer measure
(see 2.8), implies that ∞ A
k =1 k
= ∑ ∞
k=1 | Ak |, completing the proof of (b).

If A is a set with outer measure 0, then A is Lebesgue measurable (because we


can take B = ∅ in the definition 2.70). Our definition of the Lebesgue measurable
sets thus implies that the set of Lebesgue measurable sets is the smallest σ-algebra
on R containing the Borel sets and the sets with outer measure 0. Thus the set of
Lebesgue measurable sets is also the smallest σ-algebra on R containing the open
sets and the sets with outer measure 0.
Because outer measure is not even finitely additive (see 2.18), 2.72(b) implies that
there exist subsets of R that are not Lebesgue measurable.
Section 2D Lebesgue Measure 55

We previously defined Lebesgue measure as outer measure restricted to the Borel


sets (see 2.69). The term Lebesgue measure is sometimes used in mathematical
literature with the meaning as we previously defined it and is sometimes used with
the following meaning.

2.73 Definition Lebesgue measure

Lebesgue measure is the measure on (R, L), where L is the σ-algebra of Lebesgue
measurable subsets of R, that assigns to each Lebesgue measurable set its outer
measure.

The two definitions of Lebesgue measure disagree only on the domain of the
measure—is the σ-algebra the Borel sets or the Lebesgue measurable sets? You
may be able to tell which is intended from the context. In this book, the domain is
specified unless it is irrelevant.
If you are reading a mathematics paper and the domain for Lebesgue measure
is not specified, then it probably does not matter whether you use the Borel sets
or the Lebesgue measurable sets (because every Lebesgue measurable set differs
from a Borel set by a set with outer measure 0, and when dealing with measures,
what happens on a set with measure 0 usually does not matter). Because all sets that
arise from the usual operations of analysis are Borel sets, you may want to assume
that Lebesgue measure means outer measure on the Borel sets, unless what you are
reading explicitly states otherwise.
A mathematics paper may also refer to
The emphasis in some textbooks on
a measurable subset of R, without further
Lebesgue measurable sets instead of
explanation. Unless some other σ-algebra
Borel sets probably stems from the
is clear from the context, the author prob-
historical development of the subject,
ably means the Borel sets or the Lebesgue
rather than from any common use of
measurable sets. Again, the choice prob-
Lebesgue measurable sets that are
ably does not matter, but using the Borel
not Borel sets.
sets can be cleaner and simpler.
Lebesgue measure on the Lebesgue measurable sets does have one small advantage
over Lebesgue measure on the Borel sets: every subset of a set with (outer) measure
0 is Lebesgue measurable but is not necessarily a Borel set. However, any natural
process that produces a subset of R will produce a Borel set. Thus this small
advantage does not often come up in practice.

Cantor Set and Cantor Function


Every countable set has outer measure 0 (see 2.4). A reasonable question arises
about whether the converse holds. In other words, is every set with outer measure
0 countable? The Cantor set, which is introduced in this subsection, provides the
answer to this question.
The Cantor set also gives counterexamples to other reasonable conjectures. For
example, Exercise 17 in this section shows that the sum of two sets with Lebesgue
measure 0 can have positive Lebesgue measure.
56 Chapter 2 Measures

2.74 Definition Cantor set


S
The Cantor set C is [0, 1] \ ( ∞ 1 2
n=1 Gn ), where G1 = ( 3 , 3 ) and Gn for n > 1 is
S
the union of the middle-third open intervals in the intervals of [0, 1] \ ( nj=−11 Gj ).

One way to envision the Cantor set C is to start with the interval [0, 1] and then
consider the process that removes at each step the middle-third open intervals of all
intervals left from the previous step. At the first step, we remove G1 = ( 13 , 23 ).

G1 is shown in red.

After that first step, we have [0, 1] \ G1 = [0, 31 ] ∪ [ 32 , 1]. Thus we take the
middle-third open intervals of [0, 13 ] and [ 32 , 1]. In other words, we have

G2 = ( 19 , 92 ) ∪ ( 79 , 98 ).

G1 ∪ G2 is shown in red.

Now [0, 1] \ ( G1 ∪ G2 ) = [0, 19 ] ∪ [ 29 , 31 ] ∪ [ 23 , 79 ] ∪ [ 98 , 1]. Thus


1 2 7 8
G3 = ( 27 , 27 ) ∪ ( 27 , 27 ) ∪ ( 19 20 25 26
27 , 27 ) ∪ ( 27 , 27 ).

G1 ∪ G2 ∪ G3 is shown in red.

Base 3 representations provide a useful way to think about the Cantor set. Just
1
as 10 = 0.1 = 0.09999 . . . in the decimal representation, base 3 representations
are not unique for fractions whose denominator is a power of 3. For example,
1
3 = 0.13 = 0.02222 . . .3 , where the subscript 3 denotes a base 3 representations.
Notice that G1 is the set of numbers in [0, 1] whose base 3 representations have
1 in the first digit after the decimal point (for those numbers that have two base 3
representations, this means both such representations must have 1 in the first digit).
Also, G1 ∪ G2 is the set of numbers in [0, 1] whose base 3 representations S have 1 in
the first digit or the second digit after the decimal point. And so on. Hence ∞ n=1 Gn
is the set of numbers in [0, 1] whose base 3 representations have a 1 somewhere.
Thus we have the following description of the Cantor set. In the following
result, the phrase a base 3 representation indicates that if a number has two base 3
representations, then it is in the Cantor set if and only if at least one of them contains
no 1s. For example, both 13 (which equals 0.02222 . . .3 and equals 0.13 ) and 23 (which
equals 0.23 and equals 0.12222 . . .3 ) are in the Cantor set.
Section 2D Lebesgue Measure 20103

2.75 base 3 description of the Cantor set

The Cantor set C is the set of numbers in [0, 1] that have a base 3 representation
containing only 0s and 2s.

The two endpoints of each interval in


It is unknown whether or not every
each Gn are in the Cantor set. However,
number in the Cantor set is either
many elements of the Cantor set are not
rational or transcendental (meaning
endpoints of any interval in any Gn . For
not the root of a polynomial with
example, Exercise 14 asks you to show
integer coefficients).
that 14 and 13
9
are in the Cantor set; neither
of those numbers is an endpoint of any interval in any Gn . An example of an irrational

2
number in the Cantor set is ∑ n! .
n =1 3
The next result gives some elementary properties of the Cantor set.

2.76 C is closed, has measure 0, and contains no nontrivial intervals

(a) The Cantor set is a closed subset of R.


(b) The Cantor set has Lebesgue measure 0.
(c) The Cantor set contains no interval with more than one element.

Proof Each set Gn used in the definition of the Cantor set is a union of open intervals.
S
Thus each Gn is open. Thus ∞ Gn is open,
n =1 S  and hence its complement is closed.
The Cantor set equals [0, 1] ∩ R \ ∞ G
n=1 n , which is the intersection of two closed
sets. Thus the Cantor set is closed, completing the proof of (a).
By induction on n, each Gn is the union of 2n−1 disjoint open intervals, each of
n −1
which has length 31n . Thus | Gn | = 23n . The sets G1 , G2 , . . . are disjoint. Hence
[ ∞ 1 2 4

Gn = + + +···
n =1
3 9 27

1 2 4 
= 1+ + +···
3 3 9
1 1
= · 2
3 1− 3

= 1.
S
Thus the Cantor set, which equals [0, 1] \ ∞
n=1 Gn , has Lebesgue measure 1 − 1 [by
2.57(b)]. In other words, the Cantor set has Lebesgue measure 0, completing the
proof of (b).
A set with Lebesgue measure 0 cannot contain an interval that has more than one
element. Thus (b) implies (c).
58 Chapter 2 Measures

Now we can define an amazing function.

2.77 Definition Cantor function

The Cantor function Λ : [0, 1] → [0, 1] is defined by converting base 3 represen-


tations into base 2 representations as follows:
• If x ∈ C, then Λ( x ) is computed from the unique base 3 representation of
x containing only 0s and 2s by replacing each 2 by 1 and interpreting the
resulting string as a base 2 number.
• If x ∈ [0, 1] \ C, then Λ( x ) is computed from a base 3 representation of x
by truncating after the first 1, replacing each 2 before the first 1 by 1, and
interpreting the resulting string as a base 2 number.

2.78 Example values of the Cantor function

20
 5
• Λ(0.02023 ) = 0.01012 ; in other words, Λ 81 = 16 .

664
• Λ(0.2201213 ) = 0.11012 ; in other words Λ 729 = 13 16 .

• Suppose x ∈ 13 , 23 . Then x ∈ / C because x was removed in the first step of
the definition of the Cantor set. Each base 3 representation of x begins with 0.1.
Thus we truncate and interpret 0.1 as a base 2 number, getting 12 . Hence the

Cantor function Λ has the constant value 12 on the interval 31 , 32 , as shown on
the graph below.

• Suppose x ∈ 79 , 98 . Then x ∈ / C because x was removed in the second step
of the definition of the Cantor set. Each base 3 representation of x begins with
0.21. Thus we truncate, replace the 2 by 1, and interpret 0.11 as a base 2 number,
getting 34 . Hence the Cantor function Λ has the constant value 34 on the interval
7 8

9 , 9 , as shown on the graph below.

Graph of the Cantor function on the intervals from first three steps.
Section 2D Lebesgue Measure 59

As shown in the next result, in some mysterious fashion the Cantor function
manages to map [0, 1] onto [0, 1] even though the Cantor function is constant on each
open interval in the complement of the Cantor set—see the graph in Example 2.78.

2.79 Cantor function

The Cantor function Λ is a continuous, increasing function from [0, 1] onto [0, 1].
Furthermore, Λ(C ) = [0, 1].

Proof We begin by showing that Λ(C ) = [0, 1]. To do this, suppose y ∈ [0, 1]. In
the base 2 representation of y, replace each 1 by 2 and interpret the resulting string in
base 3, getting a number x ∈ [0, 1]. Because x has a base 3 representation consisting
only of 0s and 2s, the number x is in the Cantor set C. The definition of the Cantor
function shows that Λ( x ) = y. Thus y ∈ Λ(C ). Hence Λ(C ) = [0, 1], as desired.
Some careful thinking about the meaning of base 3 and base 2 representations and
the definition of the Cantor function shows that Λ is an increasing function. This step
is left to the reader.
If x ∈ [0, 1] \ C, then the Cantor function Λ is constant on an open interval
containing x and thus Λ is continuous at x. If x ∈ C, then again some careful
thinking about base 3 and base 2 representations shows that Λ is continuous at x.
Alternatively, you can skip the paragraph above and note that an increasing
function on [0, 1] whose range equals [0, 1] is automatically continuous (although
you should think about why that holds).

Now we can use the Cantor function to show that the Cantor set is uncountable
even though it is a closed set with outer measure 0.

2.80 C is uncountable

The Cantor set is uncountable.

Proof If C were countable, then Λ(C ) would be countable. However, 2.79 shows
that Λ(C ) is uncountable.

As we see in the next result, the Cantor function shows that even a continuous
function can map a set with Lebesgue measure 0 to nonmeasurable sets.

2.81 continuous image of a Lebesgue measurable set can be nonmeasurable

There exists a Lebesgue measurable set A ⊂ [0, 1] such that | A| = 0 and Λ( A)


is not a Lebesgue measurable set.

Proof Let E be a subset of [0, 1] that is not Lebesgue measurable (the existence
of such a set follows from the discussion after 2.72). Let A = C ∩ Λ−1 ( E). Then
| A| = 0 because A ⊂ C and |C | = 0 (by 2.76). Thus A is Lebesgue measurable
because every subset of R with Lebesgue measure 0 is Lebesgue measurable.
Because Λ maps C onto [0, 1] (see 2.79), we have Λ( A) = E.
60 Chapter 2 Measures

EXERCISES 2D
1 (a) Show that the set consisting of those numbers in (0, 1) that have a decimal
expansion containing one hundred consecutive 4s is a Borel subset of R.
(b) What is the Lebesgue measure of the set in part (a)?
2 Prove that there exists a bounded set A ⊂ R such that | F | ≤ | A| − 1 for every
closed set F ⊂ A.
3 Prove that there exists a set A ⊂ R such that | G \ A| = ∞ for every open set G
that contains A.
4 The phrase nontrivial interval is used to denote an interval of R that contains
more than one element. Recall that an interval might be open, closed, or neither.
(a) Prove that the union of each collection of nontrivial intervals of R is the
union of a countable subset of that collection.
(b) Prove that the union of each collection of nontrivial intervals of R is a Borel
set.
(c) Prove that there exists a collection of closed intervals of R whose union is
not a Borel set.

5 Prove that if A ⊂ R is Lebesgue measurable, then there exists an increasing


sequence F1 ⊂ F2 ⊂ · · · of closed sets contained in A such that

[

A \ Fk = 0.
k =1

6 Suppose A ⊂ R and | A| < ∞. Prove that A is Lebesgue measurable if and


only if for every ε > 0 there exists a set G that is the union of finitely many
disjoint bounded open intervals such that | A \ G | + | G \ A| < ε.
7 Prove that if A ⊂ R is Lebesgue measurable, then there exists a decreasing
sequence G1 ⊃ G2 ⊃ · · · of open sets containing A such that
 \
∞ 

Gk \ A = 0.
k =1

8 Prove that the collection of Lebesgue measurable subsets of R is translation


invariant. More precisely, prove that if A ⊂ R is Lebesgue measurable and
t ∈ R, then t + A is Lebesgue measurable.
9 Prove that the collection of Lebesgue measurable subsets of R is dilation invari-
ant. More precisely, prove that if A ⊂ R is Lebesgue measurable and t ∈ R,
then tA (which is defined to be {ta : a ∈ A}) is Lebesgue measurable.
10 Prove that if A and B are disjoint subsets of R and B is Lebesgue measurable,
then | A ∪ B| = | A| + | B|.
Section 2D Lebesgue Measure 61

11 Prove that if A ⊂ R and | A| > 0, then there exists a subset of A that is not
Lebesgue measurable.
12 Suppose b < c and A ⊂ (b, c). Prove that A is Lebesgue measurable if and
only if | A| + |(b, c) \ A| = c − b.
13 Suppose A ⊂ R. Prove that A is Lebesgue measurable if and only if
|(−n, n) ∩ A| + |(−n, n) \ A| = 2n
for every n ∈ Z+.
1 9
14 Show that 4 and 13 are both in the Cantor set.
13
15 Show that 17 is not in the Cantor set.
16 List the eight open intervals whose union is G4 in the definition of the Cantor
set (2.74).
17 Let C denote the Cantor set. Prove that 12 C + 21 C = [0, 1].
18 Prove that every open interval of R contains either infinitely many or no elements
in the Cantor set.
Z 1
19 Evaluate Λ, where Λ is the Cantor function.
0
20 Evaluate each of the following:
9

(a) Λ 13 ;
(b) Λ(0.93).

21 Find each of the following sets:



(a) Λ−1 { 13 } ;
5

(b) Λ−1 { 16 } .

22 (a) Suppose x is a rational number in [0, 1]. Explain why Λ( x ) is rational.


(b) Suppose x ∈ C is such that Λ( x ) is rational. Explain why x is rational.
23 Show that there exists a function f : R → R such that the image under f of
every nonempty open interval is R.
24 For A ⊂ R, the quantity
sup{| F | : F is a closed bounded subset of R and F ⊂ A}
is called the inner measure of A.
(a) Show that if A is a Lebesgue measurable subset of R, then the inner measure
of A equals the outer measure of A.
(b) Show that inner measure is not a measure on the σ-algebra of all subsets
of R.
62 Chapter 2 Measures

2E Convergence of Measurable Functions


Recall that a measurable space is a pair ( X, S), where X is a set and S is a σ-algebra
on X. We defined a function f : X → R to be S -measurable if f −1 ( B) ∈ S for
every Borel set B ⊂ R. In Section 2B we proved some results about S -measurable
functions; this was before we had introduced the notion of a measure.
In this section, we return to study measurable functions, but now with an emphasis
on results that depend upon measures. The highlights of this section are the proofs of
Egorov’s Theorem and Luzin’s Theorem.

Pointwise and Uniform Convergence


We begin this section with some definitions that you probably saw in an earlier course.

2.82 Definition pointwise convergence; uniform convergence

Suppose X is a set, f 1 , f 2 , . . . is a sequence of functions from X to R, and f is a


function from X to R.
• The sequence f 1 , f 2 , . . . converges pointwise on X to f if

lim f k ( x ) = f ( x )
k→∞

for each x ∈ X.
In other words, f 1 , f 2 , . . . converges pointwise on X to f if for each x ∈ X
and every ε > 0, there exists n ∈ Z+ such that | f k ( x ) − f ( x )| < ε for all
integers k ≥ n.
• The sequence f 1 , f 2 , . . . converges uniformly on X to f if for every ε > 0,
there exists n ∈ Z+ such that | f k ( x ) − f ( x )| < ε for all integers k ≥ n and
all x ∈ X.

2.83 Example a sequence converging pointwise but not uniformly


Suppose f k : [−1, 1] → R is the
function whose graph is shown here
and f : [−1, 1] → R is the function
defined by
(
1 if x 6= 0,
f (x) =
2 if x = 0.

Then f 1 , f 2 , . . . converges pointwise


on [−1, 1] to f but f 1 , f 2 , . . . does
not converge uniformly on [−1, 1] to The graph of f k .
f , as you should verify.
Section 2E Convergence of Measurable Functions 63

Like the difference between continuity and uniform continuity, the difference
between pointwise convergence and uniform convergence lies in the order of the
quantifiers. Take a moment to examine the definitions carefully. If a sequence of
functions converges uniformly on some set, then it also converges pointwise on the
same set; however, the converse is not true, as shown by Example 2.83.
Example 2.83 also shows that the pointwise limit of continuous functions need not
be continuous. However, the next result tells us that the uniform limit of continuous
functions is continuous.

2.84 uniform limit of continuous functions is continuous

Suppose B ⊂ R and f 1 , f 2 , . . . is a sequence of functions from B to R that


converges uniformly on B to a function f : B → R. Suppose b ∈ B and f k is
continuous at b for each k ∈ Z+. Then f is continuous at b.

Proof Suppose ε > 0. Let n ∈ Z+ be such that | f n ( x ) − f ( x )| < 3ε for all x ∈ B.


Because f n is continuous at b, there exists δ > 0 such that | f n ( x ) − f n (b)| < 3ε for
all x ∈ (b − δ, b + δ) ∩ B.
Now suppose x ∈ (b − δ, b + δ) ∩ B. Then

| f ( x ) − f (b)| ≤ | f ( x ) − f n ( x )| + | f n ( x ) − f n (b)| + | f n (b) − f (b)|


< ε.

Thus f is continuous at b.

Egorov’s Theorem
A sequence of functions that converges
Dmitri Egorov (1869–1931) proved
pointwise need not converge uniformly.
the theorem below in 1911. You may
However, the next result says that a point-
encounter some books that spell his
wise convergent sequence of functions on
last name as Egoroff.
a measure space with finite total measure
almost converges uniformly, in the sense that it converges uniformly except on a set
that can have arbitrarily small measure.
As an example of the next result, consider Lebesgue measure λ on the inter-
val [−1, 1] and the sequence of functions f 1 , f 2 , . . . in Example 2.83 that con-
verges pointwise but not uniformly on [−1, 1]. Suppose ε > 0. Then taking
E = [−1, − 4ε ] ∪ [ 4ε , 1], we have λ([−1, 1] \ E) < ε and f 1 , f 2 , . . . converges uni-
formly on E, as in the conclusion of the next result.

2.85 Egorov’s Theorem

Suppose ( X, S , µ) is a measure space with µ( X ) < ∞. Suppose f 1 , f 2 , . . . is a


sequence of S -measurable functions from X to R that converges pointwise on
X to a function f : X → R. Then for every ε > 0, there exists a set E ∈ S such
that µ( X \ E) < ε and f 1 , f 2 , . . . converges uniformly to f on E.
64 Chapter 2 Measures

Proof Suppose ε > 0. Temporarily fix n ∈ Z+. The definition of pointwise


convergence implies that
∞ \
[ ∞
2.86 { x ∈ X : | f k ( x ) − f ( x )| < n1 } = X.
m =1 k = m

For m ∈ Z+, let



\
Am,n = { x ∈ X : | f k ( x ) − f ( x )| < n1 }.
k=m

Then clearly A1,n ⊂ A2,n ⊂ · · · is an increasing sequence of sets and 2.86 can be
rewritten as

[
Am,n = X.
m =1

The equation above implies (by 2.59) that limm→∞ µ( Am,n ) = µ( X ). Thus there
exists mn ∈ Z+ such that
ε
2.87 µ ( X ) − µ ( Amn , n ) < .
2n
Now let

\
E= Amn , n .
n =1
Then
 ∞
\ 
µ( X \ E) = µ X \ Amn , n
n =1


[ 
=µ ( X \ Amn , n )
n =1


≤ ∑ µ ( X \ Amn , n )
n =1

< ε,

where the last inequality follows from 2.87.


To complete the proof, we must verify that f 1 , f 2 , . . . converges uniformly to f
on E. To do this, suppose ε0 > 0. Let n ∈ Z+ be such that n1 < ε0 . Then E ⊂ Amn , n ,
which implies that
| f k ( x ) − f ( x )| < n1 < ε0
for all k ≥ mn and all x ∈ E. Hence f 1 , f 2 , . . . does indeed converge uniformly to f
on E.
Section 2E Convergence of Measurable Functions 65

Approximation by Simple Functions

2.88 Definition simple function

A function is called simple if it takes on only finitely many values.

Suppose ( X, S) is a measurable space, f : X → R is a simple function, and


c1 , . . . , cn are the distinct nonzero values of f . Then
f = c1 χ E + · · · + c n χ E ,
1 n

where Ek = f −1 ({c k }). Thus this function f is an S -measurable function if and


only if E1 , . . . , En ∈ S (as you should verify).

2.89 approximation by simple functions

Suppose ( X, S) is a measure space and f : X → [−∞, ∞] is S -measurable.


Then there exists a sequence f 1 , f 2 , . . . of functions from X to R such that

(a) each f k is a simple S -measurable function;


(b) | f k ( x )| ≤ | f k+1 ( x )| ≤ | f ( x )| for all k ∈ Z+ and all x ∈ X;
(c) lim f k ( x ) = f ( x ) for every x ∈ X;
k→∞

(d) f 1 , f 2 , . . . converges uniformly on X to f if f is bounded.

Proof The idea of the proof is that for each k ∈ Z+ and n ∈ Z, the interval
[n, n + 1) is divided into 2k equally sized half-open subintervals. If f ( x ) ∈ [0, k),
we define f k ( x ) to be the left endpoint of the subinterval into which f ( x ) falls; if
f ( x ) ∈ (−k, 0), we define f k ( x ) to be the right endpoint of the subinterval into
which f ( x ) falls; and if | f ( x )| ≥ k, we define f k ( x ) to be ±k. Specifically, let
  

 2
m
k if 0 ≤ f ( x ) < k and m ∈ Z is such that f ( x ) ∈ 2mk , m2+k 1 ,



  

 m+1 if − k < f ( x ) < 0 and m ∈ Z is such that f ( x ) ∈ m , m+1 ,
2k 2k 2k
f k (x) =



 k if f ( x ) ≥ k,




−k if f ( x ) ≤ −k.

Each f −1 [ 2mk , m2+k 1 ) ∈ S because f is an S -measurable function. Thus each f k
is an S -measurable simple function; in other words, (a) holds.
Also, (b) holds because of how we have defined f k .
The definition of f k implies that
1
2.90 | f k ( x ) − f ( x )| ≤ 2k
for all x ∈ X such that f ( x ) ∈ [−k, k].
Thus we see that (c) holds.
Finally, 2.90 shows that (d) holds.
66 Chapter 2 Measures

Luzin’s Theorem
Our next result is surprising. It says that
Nikolai Luzin (1883–1950) proved
an arbitrary Borel measurable function is
the theorem below in 1912. Most
almost continuous, in the sense that its
mathematics literature in English
restriction to a large closed set is contin-
refers to the result below as Lusin’s
uous. Here, the phrase large closed set
Theorem. However, Luzin is the
means that we can take the complement
correct transliteration from Russian
of the closed set to have arbitrarily small
into English; Lusin is the
measure.
transliteration into German.
Be careful about the interpretation of
the conclusion of Luzin’s Theorem that f | B is a continuous function on B. This is
not the same as saying that f (on its original domain) is continuous at each point
of B. For example, χQ is discontinuous at every point of R. However, χQ|R\Q is a
continuous function on R \ Q (because this function is identically 0 on its domain).

2.91 Luzin’s Theorem

Suppose g : R → R is a Borel measurable function. Then for every ε > 0, there


exists a closed set F ⊂ R such that |R \ F | < ε and g| F is a continuous function
on F.

Proof First consider the special case where g = d1 χ D + · · · + dn χ D for some


1 n
distinct nonzero d1 , . . . , dn ∈ R and some disjoint Borel sets D1 , . . . , Dn ⊂ R.
Suppose ε > 0. For each k ∈ {1, . . . , n}, there exist (by 2.71) a closed set Fk ⊂ Dk
and an open set Gk ⊃ Dk such that
ε ε
| Gk \ Dk | < and | Dk \ Fk | < .
2n 2n
ε
Because Gk \ Fk = ( Gk \ Dk ) ∪ ( Dk \ Fk ), we have | Gk \ Fk | < n for each k ∈
{1, . . . , n}.
Let
[n  \ n
F= Fk ∪ (R \ Gk ).
k =1 k =1
S
Then F is a closed subset of R and R \ F = nk=1 ( Gk \ Fk ). Thus |R \ F | < ε.
Because Fk ⊂ Dk , we see that g is identically dk on Fk . Thus g| Fk is continuous
for each k ∈ {1, . . . , n}. Because
n
\ n
\
(R \ Gk ) ⊂ ( R \ Dk ) ,
k =1 k =1
T
we see that g is identically 0 on nk=1 (R \ Gk ). Thus g|Tn (R\Gk ) is continuous.
k =1
Putting all this together, we conclude that g| F is continuous (use Exercise 9 in this
section), completing the proof in this special case.
Now consider an arbitrary Borel measurable function g : R → R. By 2.89, there
exists a sequence g1 , g2 , . . . of functions from R to R that converges pointwise on R
to g, where each gk is a simple Borel measurable function.
Section 2E Convergence of Measurable Functions 67

Suppose ε > 0. By the special case already proved, for each k ∈ Z+, there exists
a closed set Ck ⊂ R such that |R \ Ck | < 2kε+1 and gk |Ck is continuous. Let

\
C= Ck .
k =1

Thus C is a closed set and gk |C is continuous for every k ∈ Z+. Note that

[
R\C = (R \ Ck );
k =1

thus |R \ C | < 2ε .
For each m ∈ Z, the sequence g1 |(m,m+1) , g2 |(m,m+1) , . . . converges pointwise
on (m, m + 1) to g|(m,m+1) . Thus by Egorov’s Theorem (2.85), for each m ∈ Z,
there is a Borel set Em ⊂ (m, m + 1) such that g1 , g2 , . . . converges uniformly to g
on Em and
ε
|(m, m + 1) \ Em | < |m|+3 .
2
Thus g1 , g2 , . . . converges uniformly to g on C ∩ Em for each m ∈ Z. Because each
gk |C is continuous, we conclude (using 2.84) that g|C∩Em is continuous for each
m ∈ Z. Thus g| D is continuous, where
[
D= (C ∩ Em ).
m∈Z

Because [ 
R\D ⊂ Z∪ (m, m + 1) \ Em ∪ ( R \ C ),
m∈Z
we have |R \ D | < ε.
There exists a closed set F ⊂ D such that | D \ F | < ε − |R \ D | (by 2.65). Now

|R \ F | = |(R \ D ) ∪ ( D \ F )| ≤ |R \ D | + | D \ F | < ε.
Because the restriction of a continuous function to a smaller domain is also continuous,
g| F is continuous, completing the proof.

We need the following result to get another version of Luzin’s Theorem.

2.92 continuous extensions of continuous functions

• Every continuous function on a closed subset of R can be extended to a


continuous function on all of R.
• More precisely, if F ⊂ R is closed and g : F → R is continuous, then there
exists a continuous function h : R → R such that h| F = g.

Proof Suppose F ⊂ R is closed and g : F → R is continuous. Thus R \ F is the


union of a collection of disjoint open intervals { Ik }. For each such interval of the
form ( a, ∞) or of the form (−∞, a), define h( x ) = g( a) for all x in the interval.
68 Chapter 2 Measures

For each interval Ik of the form (b, c) with b < c and b, c ∈ R, define h on [b, c]
to be the linear function such that h(b) = g(b) and h(c) = g(c).
Define h( x ) = g( x ) for all x ∈ R for which h( x ) has not been defined by the
previous two paragraphs. Then h : R → R is continuous and h| F = g.

The next result gives a slightly modified way to state Luzin’s Theorem. You can
think of this version as saying that the value of a Borel measurable function can be
changed on a set with small Lebesgue measure to produce a continuous function.

2.93 Luzin’s Theorem, second version

Suppose E ⊂ R and g : E → R is a Borel measurable function. Then for every


ε > 0, there exists a closed set F ⊂ E and a continuous function h : R → R such
that | E \ F | < ε and h| F = g| F .

Proof Suppose ε > 0. Extend g to a function g̃ : R → R by defining


(
g( x ) if x ∈ E,
g̃( x ) =
0 if x ∈ R \ E.

By the first version of Luzin’s Theorem (2.91), there is a closed set C ⊂ R such
that |R \ C | < ε and g̃|C is a continuous function on C. There exists a closed set
F ⊂ C ∩ E such that |(C ∩ E) \ F | < ε − |R \ C | (by 2.65). Thus

| E \ F | ≤ (C ∩ E) \ F ∪ (R \ C ) ≤ |(C ∩ E) \ F | + |R \ C | < ε.

Now g̃| F is a continuous function on F. Also, g̃| F = g| F (because F ⊂ E) . Use


2.92 to extend g̃| F to a continuous function h : R → R.

The building at Moscow State University where the mathematics seminar organized
by Egorov and Luzin met. Both Egorov and Luzin had been students at Moscow State
University and then later became faculty members at the same institution. Luzin’s
PhD thesis advisor was Egorov.
CC-BY-SA A. Savin
Section 2E Convergence of Measurable Functions ≈ 100 ln 2

Lebesgue Measurable Functions

2.94 Definition Lebesgue measurable function

A function f : A → R, where A ⊂ R, is called Lebesgue measurable if f −1 ( B)


is a Lebesgue measurable set for every Borel set B ⊂ R.

If f : A → R is a Lebesgue measurable function, then A is a Lebesgue measurable


subset of R [because A = f −1 (R)]. If A is a Lebesgue measurable subset of R, then
the definition above is the standard definition of an S -measurable function, where S
is a σ-algebra of all Lebesgue measurable subsets of A.
The following list summarizes and reviews some crucial definitions and results:
• A Borel set is an element of the smallest σ-algebra on R that contains all the
open subsets of R.
• A Lebesgue measurable set is an element of the smallest σ-algebra on R that
contains all the open subsets of R and all the subsets of R with outer measure 0.
• The terminology Lebesgue set would make good sense in parallel to the termi-
nology Borel set. However, Lebesgue set has another meaning, so we need to
use Lebesgue measurable set.
• Every Lebesgue measurable set differs from a Borel set by a set with outer
measure 0. The Borel set can be taken either to be contained in the Lebesgue
measurable set or to contain the Lebesgue measurable set.
• Outer measure restricted to the σ-algebra of Borel sets is called Lebesgue
measure.
• Outer measure restricted to the σ-algebra of Lebesgue measurable sets is also
called Lebesgue measure.
• Outer measure is not a measure on the σ-algebra of all subsets of R.
• A function f : A → R, where A ⊂ R, is called Borel measurable if f −1 ( B) is a
Borel set for every Borel set B ⊂ R.
• A function f : A → R, where A ⊂ R, is called Lebesgue measurable if f −1 ( B)
is a Lebesgue measurable set for every Borel set B ⊂ R.
Although there exist Lebesgue measur-
“Passing from Borel to Lebesgue
able sets that are not Borel sets, you are
measurable functions is the work of
unlikely to encounter one. Similarly, a
the devil. Don’t even consider it!”
Lebesgue measurable function that is not
–Barry Simon (winner of the
Borel measurable is unlikely to arise in
American Mathematical Society
anything you do. A great way to simplify
Steele Prize for Lifetime
the potential confusion about Lebesgue
Achievement), in his five-volume
measurable functions being defined by in-
series A Comprehensive Course in
verse images of Borel sets is to consider
Analysis
only Borel measurable functions.
70 Chapter 2 Measures

The next result states that if we adopt


“He professes to have received no
the philosophy that what happens on a
sinister measure.”
set of outer measure 0 does not matter
– Measure for Measure,
much, then we might as well restrict our
by William Shakespeare
attention to Borel measurable functions.

2.95 every Lebesgue measurable function is almost Borel measurable

Suppose f : R → R is a Lebesgue measurable function. Then there exists a Borel


measurable function g : R → R such that

|{ x ∈ R : g( x ) 6= f ( x )}| = 0.

Proof There exists a sequence f 1 , f 2 , . . . of Lebesgue measurable simple functions


from R to R converging pointwise on R to f (by 2.89). Suppose k ∈ Z+. Then there
exist c1 , . . . , cn ∈ R and disjoint Lebesgue measurable sets A1 , . . . , An ⊂ R such
that
f k = c1 χ A + · · · + c n χ A .
1 n

For each j ∈ {1, . . . , n}, there exists a Borel set Bj ⊂ A j such that | A j \ Bj | = 0
[by the equivalence of (a) and (d) in 2.71]. Let

gk = c1 χ B + · · · + c n χ B .
1 n

Then gk is a Borel measurable function and |{ x ∈ R : gk ( x ) 6= f k ( x )}| = 0.


S
/ ∞
If x ∈ k=1 { x ∈ R : gk ( x ) 6 = f k ( x )}, then gk ( x ) = f k ( x ) for all k ∈ Z and
+
hence limk→∞ gk ( x ) = f ( x ). Let

E = { x ∈ R : lim gk ( x ) exists in R}.


k→∞

Then E is a Borel subset of R [by Exercise 14(b) in Section 2B]. Also,



[
R\E ⊂ { x ∈ R : gk ( x ) 6= f k ( x )}
k =1

and thus |R \ E| = 0. For x ∈ R, let

2.96 g( x ) = lim (χ E gk )( x ).
k→∞

If x ∈ E, then the limit above exists by the definition of E; if x ∈ R \ E, then the


limit above exists because (χ E gk )( x ) = 0 for all k ∈ Z+.
For each k ∈ Z+, the function χ E gk is Borel measurable. Thus 2.96 implies that g
is a Borel measurable function (by 2.48). Because

[
{ x ∈ R : g( x ) 6= f ( x )} ⊂ { x ∈ R : gk ( x ) 6= f k ( x )},
k =1

we have |{ x ∈ R : g( x ) 6= f ( x )}| = 0, completing the proof.


Section 2E Convergence of Measurable Functions 71

EXERCISES 2E
1 Suppose X is a finite set. Explain why a sequence of functions from X to R that
converges pointwise on X also converges uniformly on X.
2 Give an example of a sequence of functions from Z+ to R that converges
pointwise on Z+ but does not converge uniformly on Z+.
3 Give an example of a sequence of continuous functions f 1 , f 2 , . . . from [0, 1]
to R that converge pointwise to a function f : [0, 1] → R that is not a bounded
function.
4 Prove or give a counterexample: If A ⊂ R and f 1 , f 2 , . . . is a sequence of
uniformly continuous functions from A to R that converge uniformly to a
function f : A → R, then f is uniformly continuous on A.
5 Give an example to show that Egorov’s Theorem can fail without the hypothesis
that µ( X ) < ∞.
6 Suppose ( X, S , µ) is a measure space with µ( X ) < ∞. Suppose f 1 , f 2 , . . . is a
sequence of S -measurable functions from X to R such that limk→∞ f k ( x ) = ∞
for each x ∈ X. Prove that for every ε > 0, there exists a set E ∈ S such that
µ( X \ E) < ε and f 1 , f 2 , . . . converges uniformly to ∞ on E (meaning that for
every t > 0, there exists n ∈ Z+ such that f k ( x ) > t for all integers k ≥ n and
all x ∈ E).
[The exercise above is an Egorov-type theorem for sequences of functions that
converge pointwise to ∞.]
7 Suppose F is a closed bounded subset of R and g1 , g2 , . . . is an increasing
sequence of continuous real-valued functions on F (thus g1 ( x ) ≤ g2 ( x ) ≤ · · ·
for all x ∈ F) such that sup{ g1 ( x ), g2 ( x ), . . .} < ∞ for each x ∈ F. Define a
real-valued function g on F by

g( x ) = lim gk ( x ).
k→∞

Prove that g is continuous on F if and only if g1 , g2 , . . . converges uniformly on


F to g.
[The result above is called Dini’s Theorem.]
+
8 Suppose µ is the measure on (Z+ , 2Z ) defined by
1
µ( E) = ∑ 2n
.
n∈ E

Prove that for every ε > 0, there exists a set E ⊂ Z+ with µ(Z+ \ E) < ε
such that f 1 , f 2 , . . . converges uniformly on E for every sequence of functions
f 1 , f 2 , . . . from Z+ to R that converges pointwise on Z+.
[This result does not follow from Egorov’s Theorem because here we are asking
for E to depend only on ε. In Egorov’s Theorem, E depends on ε and on the
sequence f 1 , f 2 , . . . .]
72 Chapter 2 Measures

9 Suppose F1 , . . . , Fn are disjoint closed subsets of R. Prove that if

g : F1 ∪ · · · ∪ Fn → R

is a function such that g| Fk is a continuous function for each k ∈ {1, . . . , n},


then g is a continuous function.
10 Suppose F ⊂ R is such that every continuous function from F to R can be
extended to a continuous function from R to R. Prove that F is a closed subset
of R.
11 Prove or give a counterexample: If F ⊂ R is such that every bounded continuous
function from F to R can be extended to a continuous function from R to R,
then F is a closed subset of R.
12 Give an example of a Borel measurable function f from R to R such that there
does not exist a set B ⊂ R such that |R \ B| = 0 and f | B is a continuous
function on B.
13 Prove or give a counterexample: If f t : R → R is a Borel measurable function
for each t ∈ R and f : R → (−∞, ∞] is defined by

f ( x ) = sup{ f t ( x ) : t ∈ R},

then f is a Borel measurable function.


14 Suppose b1 , b2 , . . . is a sequence of real numbers. Define f : R → [0, ∞] by
∞
 1
∑ / {b1 , b2 , . . .},
if x ∈
4 k |x − b |
f (x) = k =1 k


∞ if x ∈ {b1 , b2 , . . .}.

Prove that |{ x ∈ R : f ( x ) < 1}| = ∞.


[This exercise is a variation of a problem originally considered by Borel. If
b1 , b2 , . . . contains all the rational numbers, then it is not even obvious that
{ x ∈ R : f ( x ) < ∞} 6= ∅.]
15 Suppose B is a Borel set and f : B → R is a Lebesgue measurable function.
Show that there exists a Borel measurable function g : B → R such that

|{ x ∈ B : g( x ) 6= f ( x )}| = 0.
Chapter 3
Integration

To remedy deficiencies of Riemann integration that were discussed in Section 1B,


in the last chapter we developed measure theory as an extension of the notion of the
length of an interval. Having proved the fundamental results about measures, we are
now ready to use measures to develop integration with respect to a measure.
As we will see, this new method of integration fixes many of the problems with
Riemann integration. In particular, we will develop good theorems for interchanging
limits and integrals.

Statue in Milan of Maria Gaetana Agnesi,


who in 1748 published one of the first calculus textbooks.
A translation of her book into English was published in 1801.
In this chapter, we develop a method of integration more powerful
than methods contemplated by the pioneers of calculus.
©Giovanni Dall’Orto

73
74 Chapter 3 Integration

3A Integration with Respect to a Measure


Integration of Nonnegative Functions
We will first define the integral of a nonnegative function with respect to a measure.
Then by writing a real-valued function as the difference of two nonnegative functions,
we will define the integral of a real-valued function with respect to a measure. We
begin this process with the following definition.

3.1 Definition S -partition

Suppose S is a σ-algebra on a set X. An S -partition of X is a finite collection


A1 , . . . , Am of disjoint sets in S such that A1 ∪ · · · ∪ Am = X.

The next definition should remind you


We adopt the convention that 0 · ∞
of the definition of the lower Riemann
and ∞ · 0 should both be interpreted
sum (see 1.3). However, now we are
to be 0.
working with an arbitrary measure and
thus X need not be a subset of R. More importantly, even in the case when X is a
closed interval [ a, b] in R and µ is Lebesgue measure on the Borel subsets of [ a, b],
the sets A1 , . . . , Am in the definition below do not need to be subintervals of [ a, b] as
they do for the lower Riemann sum—they need only be Borel sets.

3.2 Definition lower Lebesgue sum

Suppose ( X, S , µ) is a measure space, f : X → [0, ∞] is an S -measurable


function, and P is an S -partition A1 , . . . , Am of X. The lower Lebesgue sum
L( f , P) is defined by
m
L( f , P) = ∑ µ( A j ) inf
Aj
f.
j =1

Suppose ( X, S , µ) is a measure space. We R will denote the integral of an S -


measurable function f with respect
R to µ by f dµ. Our basic requirements for
an integral are that we want χ E dµ to equal µ( E) for all E ∈ S , and we want
R R R
( f + g) dµ = f dµ + g dµ. As we will see, the following definition satisfies
both of those requirements (although this is not obvious). Think about why the
following definition is reasonable in terms of the integral equaling the area under the
graph of the function (in the special case of Lebesgue measure on an interval of R).

3.3 Definition integral of a nonnegative function

Suppose ( X, S , µ) is a measure space and f : X → [0,R ∞] is an S -measurable


function. The integral of f with respect to µ, denoted f dµ, is defined by
Z
f dµ = sup{L( f , P) : P is an S -partition of X }.
Section 3A Integration with Respect to a Measure 75

Suppose ( X, S , µ) is a measure space and f : X → [0, ∞] is an S -measurable


function. Each S -partition A1 , . . . , Am of X leads to an approximation of f from
below by the S -measurable simple function ∑m j=1 inf f χ A . This suggests that
Aj j

m
∑ µ( A j ) inf
Aj
f
j =1
R
should be an approximation from below of our intuitive notionRof f dµ. Taking the
supremum of these approximations leads to our definition of f dµ.
The following result gives our first example of evaluating an integral.

3.4 integral of a characteristic function

Suppose ( X, S , µ) is a measure space and E ∈ S . Then


Z
χ E dµ = µ( E).

Proof If P is the S -partition of X con-


sisting of E and its complement X \ E, R symbol d in the expression
The
f dµ has no independent meaning,
then clearly L(χ E , P) = µ( E). Thus
R but it often usefully separates
R f from
χ E dµ ≥ µ( E). µ. Because the d in f dµ does not
To prove the inequality in the other represent another object, some
direction, suppose P is an S -partition mathematicians prefer typesetting
A1 , . . . , Am of X. Then µ( A j ) inf χ E an upright Rd in this situation,
Aj
equals µ( A j ) if A j ⊂ E and equals 0 producing f dµ. However, the
otherwise. Thus upright d looks jarring to some
readers who are accustomed to
L(χ E , P) = ∑ µ( A j ) italicized symbols. This book takes
{ j : A j ⊂ E} the compromise position of using
 [  slanted d instead of math-mode
=µ Aj italicized d in integrals.
{ j : A j ⊂ E}

≤ µ ( E ).
R
Thus χ E dµ ≤ µ( E), completing the proof.

3.5 Example integrals of χQ and χ[0, 1] \ Q


Suppose
R λ is Lebesgue measure on R. As a special case of the result above, we
have χQ dλ = 0 (because |Q| = 0). Recall that χQ is not Riemann integrable on
[0, 1]. Thus even at this early stage in our development of integration with respect to
a measure, we have fixed one of theRdeficiencies of Riemann integration.
Note also that 3.4 implies that χ[0, 1] \ Q dλ = 1 (because |[0, 1] \ Q| = 1),
which is what we want. In contrast, the lower Riemann integral of χ[0, 1] \ Q on [0, 1]
equals 0, which is not what we want.
76 Chapter 3 Integration

3.6 Example integration with respect to counting measure is summation


Suppose µ is counting measure on Z+ and b1 , b2 , . . . is a sequence of nonnegative
numbers. Think of b as the function from Z+ to [0, ∞) defined by b(k) = bk . Then
Z ∞
b dµ = ∑ bk ,
k =1
as you should verify.
Integration with respect to a measure can be called Lebesgue integration. The
next result shows that Lebesgue integration behaves as expected on simple functions
represented as linear combinations of characteristic functions of disjoint sets.

3.7 integral of a simple function

Suppose ( X, S , µ) is a measure space, E1 , . . . , En are disjoint sets in S , and


c1 , . . . , cn ∈ [0, ∞]. Then
Z  n  n
∑ ck χ E
k
dµ = ∑ ck µ(Ek ).
k =1 k =1

Proof Without loss of generality, we can assume that E1 , . . . , En is an S -partition of


X [by replacing n by n + 1 and setting En+1 = X \ ( E1 ∪ . . . ∪ E  n ) and cn+1 = 0].
If P is the S -partition E1 , . . . , En of X, then L ∑nk=1 ck χ E , P = ∑nk=1 ck µ( Ek ).
k
Thus Z n  n
∑ ck χE dµ ≥ ∑ ck µ(Ek ).
k
k =1 k =1
To prove the inequality in the other direction, suppose that P is an S -partition
A1 , . . . , Am of X. Then
 n  m
L ∑ ck χ E , P = ∑ µ( A j ) min ci
k =1
k
j =1 {i : A j ∩ Ei 6=∅}

m n
= ∑ ∑ µ( A j ∩ Ek ) {i : Amin
∩ E 6=∅}
ci
j =1 k =1 j i

m n
≤ ∑ ∑ µ( A j ∩ Ek )ck
j =1 k =1
n m
= ∑ ck ∑ µ( A j ∩ Ek )
k =1 j =1
n
= ∑ ck µ(Ek ).
k =1
R 
The inequality above implies that ∑nk=1 ck χ Ek dµ ≤ ∑nk=1 ck µ( Ek ), completing
the proof.
Section 3A Integration with Respect to a Measure 77

The next easy result gives an unsurprising property of integrals.

3.8 integration is order preserving

Suppose ( X, S , µ) is a measure space and f , g : X →R[0, ∞] areR S -measurable


functions such that f ( x ) ≤ g( x ) for all x ∈ X. Then f dµ ≤ g dµ.

Proof Suppose P is an S -partition A1 , . . . , Am of X. Then


inf f ≤ inf g
Aj Aj
R R
for each j = 1, . . . , m. Thus L( f , P) ≤ L( g, P). Hence f dµ ≤ g dµ.

Monotone Convergence Theorem


For the proof of the Monotone Convergence Theorem (and several other results), we
will need to use the following mild restatement of the definition of the integral of a
nonnegative function.

3.9 integrals via simple functions

Suppose ( X, S , µ) is a measure space and f : X → [0, ∞] is S -measurable. Then


Z n m
3.10 f dµ = sup ∑ c j µ( A j ) : A1 , . . . , Am are disjoint sets in S ,
j =1
c1 , . . . , cm ∈ [0, ∞), and
m o
f (x) ≥ ∑ c j χ A (x) for every x ∈ X
j
.
j =1

Proof First note that the left side of 3.10 is bigger than or equal to the right side by
3.7 and 3.8.
To prove that the right side of 3.10 is bigger than or equal to the left side, first
assume that inf f < ∞ for every A ∈ S with µ( A) > 0. Then for P an S -partition
A
A1 , . . . , Am of nonempty subsets of X, take c j = inf f , which shows that L( f , P) is
Aj
R
in the set on the right side of 3.10. Thus the definition of f dµ shows that the right
side of 3.10 is bigger than or equal to the left side.
The only remaining case to consider is when there exists a set A ∈ S such that
µ( A) > 0 and inf f = ∞ [which implies that f ( x ) = ∞ for all x ∈ A]. In this case,
A
for arbitrary t ∈ (0, ∞) we can take m = 1, A1 = A, and c1 = t. These choices
show that the right side of 3.10 is at least tµ( A). Because t is an arbitrary positive
number, this shows that the right side of 3.10 equals ∞, which of course is greater
than or equal to the left side, completing the proof.

The next result allows us to interchange limits and integrals in certain circum-
stances. We will see more theorems of this nature in the next section.
78 Chapter 3 Integration

3.11 Monotone Convergence Theorem

Suppose ( X, S , µ) is a measure space and 0 ≤ f 1 ≤ f 2 ≤ · · · is an increasing


sequence of S -measurable functions. Define f : X → [0, ∞] by

f ( x ) = lim f k ( x ).
k→∞

Then Z Z
lim f k dµ = f dµ.
k→∞

Proof The function f is S -measurable by 2.53. R R


Because f k ( x ) ≤ f ( x ) for every
R x ∈ X,
R we have f k dµ ≤ f dµ for each
k ∈ Z+ (by 3.8). Thus limk→∞ f k dµ ≤ f dµ.
To prove the inequality in the other direction, suppose A1 , . . . , Am are disjoint
sets in S and c1 , . . . , cm ∈ [0, ∞) are such that
m
3.12 f (x) ≥ ∑ c j χ A (x)j
for every x ∈ X.
j =1

Let t ∈ (0, 1). For k ∈ Z+, let


n m o
Ek = x ∈ X : f k (x) ≥ t ∑ c j χ A (x) .
j
j =1

Then E1 ⊂ E2 ⊂ · · · is an increasing sequence of sets in S whose union equals X.


Thus limk→∞ µ( A j ∩ Ek ) = µ( A j ) for each j ∈ {1, . . . , m} (by 2.59).
If k ∈ Z+, then
m
f k (x) ≥ ∑ tc j χ A ∩ E (x) j k
j =1

for every x ∈ X. Thus (by 3.9)


Z m
f k dµ ≥ t ∑ c j µ( A j ∩ Ek ).
j =1

Taking the limit as k → ∞ of both sides of the inequality above gives


Z m
lim f k dµ ≥ t ∑ c j µ( A j ).
k→∞ j =1

Now taking the limit as t increases to 1 shows that


Z m

k→∞
lim f k dµ ≥ ∑ c j µ ( A j ).
j =1

Taking the supremum of the inequality above over all S -partitions A1 , . . . , Am


of X andR all c1 , . . . R, cm ∈ [0, ∞) satisfying 3.12 shows (using 3.9) that we have
limk→∞ f k dµ ≥ f dµ, completing the proof.
Section 3A Integration with Respect to a Measure 79

The proof that the integral is additive will use the Monotone Convergence Theorem
and our next result. The representation of a simple function h : X → [0, ∞] in the
form ∑nk=1 ck χ E is not unique. Requiring the numbers c1 , . . . , cn to be distinct and
k
E1 , . . . , En to be nonempty and disjoint with E1 ∪ · · · ∪ En = X produces what
is called the standard representation of a simple function [take Ek = h−1 ({ck }),
where c1 , . . . , cn are the distinct values of h]. The following lemma shows that all
representations (including representations with sets that are not disjoint) of a simple
measurable function give the same sum that we expect from integration.

3.13 integral-type sums for simple functions

Suppose ( X, S , µ) is a measure space. Suppose a1 , . . . , am , b1 , . . . , bn ∈ [0, ∞]


and A1 , . . . , Am , B1 , . . . , Bn ∈ S are such that ∑m n
j=1 a j χ A = ∑k=1 bk χ B . Then
j k

m n
∑ a j µ( A j ) = ∑ bk µ( Bk ).
j =1 k =1

Proof We assume A1 ∪ · · · ∪ Am = X (otherwise add the term 0χ X \ ( A ∪ · · · ∪ A )).


1 m
Suppose A1 and A2 are not disjoint. Then we can write
3.14 a1 χ A + a2 χ A = a1 χ A \ A2
+ a 2 χ A2 \ A1 + ( a 1 + a 2 ) χ A1 ∩ A2 ,
1 2 1

where the three sets appearing on the right side of the equation above are disjoint.
Now A1 = ( A1 \ A2 ) ∪ ( A1 ∩ A2 ) and A2 = ( A2 \ A1 ) ∪ ( A1 ∩ A2 ); each
of these unions is a disjoint union. Thus µ( A1 ) = µ( A1 \ A2 ) + µ( A1 ∩ A2 ) and
µ( A2 ) = µ( A2 \ A1 ) + µ( A1 ∩ A2 ). Hence
a1 µ ( A1 ) + a2 µ ( A2 ) = a1 µ ( A1 \ A2 ) + a2 µ ( A2 \ A1 ) + ( a1 + a2 ) µ ( A1 ∩ A2 ).
The equation above, in conjunction with 3.14, shows that if we replace the two
sets A1 , A2 by the three disjoint sets A1 \ A2 , A2 \ A1 , A1 ∩ A2 and make the
appropriate adjustments to the coefficients a1 , . . . , am , then the value of the sum
∑m j=1 a j µ ( A j ) is unchanged (although m has increased by 1).
Repeating this process with all pairs of subsets among A1 , . . . , Am that are
not disjoint after each step, in a finite number of steps we can convert the ini-
tial list A1 , . . . , Am into a disjoint list of subsets without changing the value of
∑m j =1 a j µ ( A j ).
The next step is to make the numbers a1 , . . . , am distinct. This is done by replacing
the sets corresponding to each a j by the union of those sets, and using finite additivity
of the measure µ to show that the value of the sum ∑m j=1 a j µ ( A j ) does not change.
Finally, drop any terms for which A j = ∅, getting the standard representation
for a simple function. We have now shown that the original value of ∑m j =1 a j µ ( A j )
is equal to the value if we use the standard representation of the simple function
∑m n
j=1 a j χ A j . The same procedure can be used with the representation ∑k=1 bk χ Bk to
show that ∑nk=1 bk µ(χ B ) equals what we would get with the standard representation.
k
Thus the equality of the functions ∑m n
j=1 a j χ A and ∑k=1 bk χ B implies the equality
j k

∑m n
j=1 a j µ ( A j ) = ∑k=1 bk µ ( Bk ).
80 Chapter 3 Integration

Now we can show that our definition


If we had already proved that
of integration does the right thing with
integration is linear, then we could
simple measurable functions that might
quickly get the conclusion of the
not be expressed in the standard represen-
previous result by integrating both
tation. The result below differs from 3.7
sides of the equation
mainly because the sets E1 , . . . , En in the
∑m n
j=1 a j χ A j = ∑k=1 bk χ Bk with
result below are not required to be dis-
joint. Like the previous result, the next respect to µ. However, we need the
result would follow immediately from the previous result to prove the next
linearity of integration if that property had result, which is used in our proof
already been proved. that integration is linear.

3.15 integral of a linear combination of characteristic functions

Suppose ( X, S , µ) is a measure space, E1 , . . . , En ∈ S , and c1 , . . . , cn ∈ [0, ∞].


Then Z  n  n
∑ ck χE dµ = ∑ ck µ(Ek ).
k
k =1 k =1

Proof The desired result follows from writing the simple function ∑nk=1 ck χ E in
k
the standard representation for a simple function and then using 3.7 and 3.13.

Now we can prove that integration is additive on nonnegative functions.

3.16 additivity of integration

Suppose ( X, S , µ) is a measure space and f , g : X → [0, ∞] are S -measurable


functions. Then Z Z Z
( f + g) dµ = f dµ + g dµ.

Proof The desired result holds for simple nonnegative S -measurable functions (by
3.15). Thus we approximate by such functions.
Specifically, let f 1 , f 2 , . . . and g1 , g2 , . . . be increasing sequences of simple non-
negative S -measurable functions such that
lim f k ( x ) = f ( x ) and lim gk ( x ) = g( x )
k→∞ k→∞

for all x ∈ X (see 2.89 for the existence of such increasing sequences). Then
Z Z
( f + g) dµ = lim ( f k + gk ) dµ
k→∞
Z Z
= lim f k dµ + lim gk dµ
k→∞ k→∞
Z Z
= f dµ + g dµ,

where the first and third equalities follow from the Monotone Convergence Theorem
and the second equality holds by 3.15.
Section 3A Integration with Respect to a Measure 81

The lower Riemann integral is not additive, even for bounded nonnegative measur-
able functions. For example, if f = χQ ∩ [0, 1] and g = χ[0, 1] \ Q, then

L( f , [0, 1]) = 0 and L( g, [0, 1]) = 0 but L( f + g, [0, 1]) = 1.


In contrast, if λ is Lebesgue measure on the Borel subsets of [0, 1], then
Z Z Z
f dλ = 0 and g dλ = 1 and ( f + g) dλ = 1.
R R R
More generally, we have just proved that ( f + g) dµ = f dµ + g dµ for
every measure µ and for all nonnegative measurable functions f and g. Recall that
integration with respect to a measure is defined via lower Lebesgue sums in a similar
fashion to the definition of the lower Riemann integral via lower Riemann sums
(with the big exception of allowing measurable sets instead of just intervals in the
partitions). However, we have just seen that the integral with respect to a measure
(which could have been called the lower Lebesgue integral) has considerably nicer
behavior (additivity!) than the lower Riemann integral.

Integration of Real-Valued Functions


The following definition gives us a standard way to write an arbitrary real-valued
function as the difference of two nonnegative functions.

3.17 Definition f +; f −

Suppose f : X → [−∞, ∞] is a function. Define functions f + and f − from X to


[0, ∞] by
( (
f ( x ) if f ( x ) ≥ 0, 0 if f ( x ) ≥ 0,
f + (x) = and f − ( x ) =
0 if f ( x ) < 0 − f ( x ) if f ( x ) < 0.

Note that if f : X → [−∞, ∞] is a function, then


f = f+ − f− and | f | = f + + f −.
The decomposition above allows us to extend our definition of integration to functions
that take on negative as well as positive values.

3.18 Definition integral of a real-valued function

R and f : XR→ [−∞, ∞] is an S -measurable


Suppose ( X, S , µ) is a measure space
function such that at least oneR of f + dµ and f − dµ is finite. The integral of
f with respect to µ, denoted f dµ, is defined by
Z Z Z
f dµ = f + dµ − f − dµ.

If f ≥ 0, then f + = f and f − = 0; thus this definition is consistent with the


previous definition of the integral of a nonnegative function.
82 Chapter 3 Integration

R R
R The condition | f | dµ < ∞ is equivalent to the condition f + dµ < ∞ and
f − dµ < ∞ (because | f | = f + + f − ).

3.19 Example a function whose integral is not defined


Suppose λ is Lebesgue measure on R and f : R → R is the function defined by
(
1 if x ≥ 0,
f (x) =
−1 if x < 0.
R R R
Then f dλ is not defined because f + dλ = ∞ and f − dλ = ∞.

The next result says that the integral of a number times a function is exactly what
we expect.

3.20 integration is homogeneous

R ( X, S , µ) is a measure space and f : X → [−∞, ∞] is a function such


Suppose
that f dµ is defined. If c ∈ R, then
Z Z
c f dµ = c f dµ.

Proof R and c ≥ 0.
First consider the case where f is a nonnegative function R If P is
an S -partition of X, then clearly L(c f , P) = cL( f , P). Thus c f dµ = c f dµ.
Now consider the general case where f takes values in [−∞, ∞]. Suppose c ≥ 0.
Then
Z Z Z
c f dµ = (c f )+ dµ − (c f )− dµ
Z Z
= c f + dµ − c f − dµ
Z Z 
=c +
f dµ − f − dµ
Z
=c f dµ,

where the third line follows from the first paragraph of this proof.
Finally, now suppose c < 0 (still assuming that f takes values in [−∞, ∞]). Then
−c > 0 and
Z Z Z
c f dµ = (c f )+ dµ − (c f )− dµ
Z Z

= (−c) f dµ − (−c) f + dµ
Z Z 
= (−c) f − dµ − f + dµ
Z
=c f dµ,

completing the proof.


Section 3A Integration with Respect to a Measure 83

Now we prove that integration with respect to a measure has the additive property
required for a good theory of integration.

3.21 additivity of integration

Suppose ( X, S , µ) Ris a measure space


R and f , g : X → R are S -measurable
functions such that | f | dµ < ∞ and | g| dµ < ∞. Then
Z Z Z
( f + g) dµ = f dµ + g dµ.

Proof Clearly
( f + g)+ − ( f + g)− = f + g
= f + − f − + g+ − g− .
Thus
( f + g)+ + f − + g− = ( f + g)− + f + + g+ .
Both sides of the equation above are sums of nonnegative functions. Thus integrating
both sides with respect to µ and using 3.16 gives
Z Z Z Z Z Z
( f + g)+ dµ + f − dµ + g− dµ = ( f + g)− dµ + f + dµ + g+ dµ.

Rearranging the equation above gives


Z Z Z Z Z Z
( f + g)+ dµ − ( f + g)− dµ = f + dµ − f − dµ + g+ dµ − g− dµ,

where the left side is not of the form ∞ − ∞ because ( f + g)+ ≤ f + + g+ and
( f + g)− ≤ f − + g− . The equation above can be rewritten as
Z Z Z Gottfried Leibniz (1646–1716)
( f + g) dµ = f dµ + g dµ, R
invented the symbol to denote
completing the proof. integration in 1675.

The next result resembles 3.8, but now the functions are allowed to be real valued.

3.22 integration is order preserving

Suppose ( X, S , µ)R is a measure


R space and f , g : X → R are S -measurable
functions such that Rf dµ and R g dµ are defined. Suppose also that f ( x ) ≤ g( x )
for all x ∈ X. Then f dµ ≤ g dµ.
R R
Proof The casesR where f dµ = R±∞ or g dµ = ±∞ are left to the reader. Thus
we assume that | f | dµ < ∞ and | g| dµ < ∞.
The additivity (3.21) and homogeneity (3.20 with c = −1) of integration imply
that Z Z Z
g dµ − f dµ = ( g − f ) dµ.
The last integral is nonnegative because g( x ) − f ( x ) ≥ 0 for all x ∈ X.
84 Chapter 3 Integration

The inequality in the next result receives frequent use.

3.23 absolute value of integral ≤ integral of absolute value

R ( X, S , µ) is a measure space and f : X → [−∞, ∞] is a function such


Suppose
that f dµ is defined. Then
Z Z

f dµ ≤ | f | dµ.

R
Proof
R Because
R f dµ is defined, f is an S -measurable function and at least one of
f + dµ and f − dµ is finite. Thus
Z Z Z

f dµ = f + dµ − f − dµ
Z Z
≤ f + dµ + f − dµ
Z
= ( f + + f − ) dµ
Z
= | f | dµ,

as desired.

EXERCISES 3A
1 Suppose ( X, S , µ)Ris a measure space and f : X → [0, ∞] is an S -measurable
function such that f dµ < ∞. Explain why

inf f = 0
E

for each set E ∈ S with µ( E) = ∞.


2 Suppose X is a set, S is a σ-algebra on X, and c ∈ X. Define the Dirac measure
δc on ( X, S) by (
1 if c ∈ E,
δc ( E) =
0 if c ∈
/ E.
R
Prove that if f : X → [0, ∞] is S -measurable, then f dδc = f (c).
[Careful: {c} may not be in S .]
3 Suppose ( X, S , µ) is a measure space and f : X → [0, ∞] is an S -measurable
function. Prove that
Z
f dµ > 0 if and only if µ({ x ∈ X : f ( x ) > 0}) > 0.
Section 3A Integration with Respect to a Measure 85

4 Give an example of a Borel measurable function f : [0, 1] → (0, ∞) such that


L( f , [0, 1]) = 0.
[Recall that L( f , [0, 1]) denotes the lower Riemann integral, which was defined
in Section R1A. If λ is Lebesgue measure on [0, 1], then the previous exercise
states that f dλ > 0 for this function f , which isRwhat we expect of a positive
function. Thus even though both L( f , [0, 1]) and f dλ are defined by taking
the supremum of approximations from below, Lebesgue measure captures the
right behavior for this function f and the lower Riemann integral does not.]
5 Verify the assertion that integration with respect to counting measure is summa-
tion (Example 3.6).
6 Suppose ( X, S , µ) is a measure space, f : X → [0, ∞] is S -measurable, and P
and P0 are S -partitions of X such that each set in P0 is contained some set in P.
Prove that L( f , P) ≤ L( f , P0 ).
7 Suppose X is a set, S is the σ-algebra of all subsets of X, and w : X → [0, ∞]
is a function. Define a measure µ on ( X, S) by

µ( E) = ∑ w( x )
x∈E

for E ⊂ X. Prove that if f : X → [0, ∞] is a function, then


Z
f dµ = ∑ w ( x ) f ( x ),
x∈X

where the infinite sums above are defined as the supremum of all sums over
finite subsets of X.
8 Suppose λ denotes Lebesgue measure on R. Given an example of a sequence
f 1 , f 2 , . . . of simple Borel measurable functionsR from R to [0, ∞) such that
limk→∞ f k ( x ) = 0 for every x ∈ R but limk→∞ f k dλ = 1.
9 Suppose µ is a measure on a measurable space ( X, S) and f : X → [0, ∞] is an
S -measurable function. Define ν : S → [0, ∞] by
Z
ν( A) = χ A f dµ

for A ∈ S . Prove that ν is a measure on ( X, S).


10 Suppose ( X, S , µ) is a measure space and f 1 , f 2 , . . . is a sequence of nonnegative
S -measurable functions. Define f : X → [0, ∞] by f ( x ) = ∑∞ k=1 f k ( x ). Prove
that Z Z∞
f dµ = ∑ f k dµ.
k =1

11 Suppose ( X, S , µ) is a measure
R space and f 1 , f 2 , . . . are S -measurable functions
from X to R such that ∑∞ k=1 | f k | dµ < ∞. Prove that there exists E ∈ S such
that µ( X \ E) = 0 and limk→∞ f k ( x ) = 0 for every x ∈ E.
86 Chapter 3 Integration

12 Show
R that there exists a Borel measurable function f : R → (0, ∞) such that
χ I f dλ = ∞ for every nonempty open interval I ⊂ R, where λ denotes
Lebesgue measure on R.
13 Give an example to show that the Monotone Convergence Theorem (3.11) can
fail if the hypothesis that f 1 , f 2 , . . . are nonnegative functions is dropped.
14 Give an example to show that the Monotone Convergence Theorem can fail if
the hypothesis of an increasing sequence of functions is replaced by a hypothesis
of a decreasing sequence of functions.
[This exercise shows that the Monotone Convergence Theorem should be called
the Increasing Convergence Theorem. However, see Exercise 20.]
15 Suppose λ is Lebesgue Rmeasure on R and f : R → [−∞, ∞] is a Borel measur-
able function such that f dλ is defined.

R t ∈ R,R define f t : R → [−∞, ∞] by f t ( x ) = f ( x − t). Prove that


(a) For
f t dλ = f dλ for all t ∈ R

R t ∈ R \1{R0}, define f t : R → [−∞, ∞] by f t ( x ) = f (tx ). Prove that


(b) For
f t dλ = |t| f dλ for all t > 0.

16 Suppose S and T are σ-algebras on a set X and S ⊂ T . Suppose µ1 is a


measure on ( X, S), µ2 is a measure on ( X, T ), and µ1 ( ER) = µ2 ( E)R for all
E ∈ S . Prove that if f : X → [0, ∞] is S -measurable, then f dµ1 = f dµ2 .

For x1 , x2 , . . . a sequence in [−∞, ∞], define lim inf xk by


k→∞

lim inf xk = lim inf{ xk , xk+1 , . . .}.


k→∞ k→∞

Note that inf{ xk , xk+1 , . . .} is an increasing function of k; thus the limit above
on the right exists in [−∞, ∞].
17 Suppose that ( X, S , µ) is a measure space and f 1 , f 2 , . . . is a sequence of non-
negative S -measurable functions on X. Define a function f : X → [0, ∞] by
f ( x ) = lim inf f k ( x ).
k→∞

(a) Show that f is an S -measurable function.


(b) Prove that Z Z
f dµ ≤ lim inf f k dµ.
k→∞

(c) Give an example showing that the inequality in (b) can be a strict inequality
even when µ( X ) < ∞ and the family of functions { f k }k∈Z+ is uniformly
bounded.
[The result in (b) is called Fatou’s Lemma. Some textbooks prove Fatou’s Lemma
and then use it to prove the Monotone Convergence Theorem. Here we are taking
the reverse approach—you should be able to use the Monotone Convergence
Theorem to give a clean proof of Fatou’s Lemma.]
Section 3A Integration with Respect to a Measure 87

18 Give an example of a sequence x1 , x2 , . . . of real numbers such that


n
lim
n→∞
∑ xk exists in R,
k =1
R
but x dµ is not defined, where µ is counting measure on Z+ and x is the
function from Z+ to R defined by x (k) = xk .
19 Show that if ( X, S , µ) is a measure space and f : X → [0, ∞) is S -measurable,
then Z
µ( X ) inf f ≤ f dµ ≤ µ( X ) sup f .
X X

20 Suppose ( X, S , µ) is a measure space and f 1 , f 2 , . . . is a monotone (meaning


either increasing or decreasing) sequence of S -measurable functions. Define
f : X → [−∞, ∞] by
f ( x ) = lim f k ( x ).
k→∞
R
Prove that if | f 1 | dµ < ∞, then
Z Z
lim f k dµ = f dµ.
k→∞

21 Henri Lebesgue wrote the following about his method of integration:


I have to pay a certain sum, which I have collected in my pocket. I
take the bills and coins out of my pocket and give them to the creditor
in the order I find them until I have reached the total sum. This is the
Riemann integral. But I can proceed differently. After I have taken
all the money out of my pocket I order the bills and coins according
to identical values and then I pay the several heaps one after the other
to the creditor. This is my integral.
Use 3.15 to explain what Lebesgue meant and to explain why integration of
a function with respect to a measure can be thought of as partitioning the
range of the function, in contrast to Riemann integration, which depends upon
partitioning the domain of the function.
[The quote above is taken from page 796 of The Princeton Companion to
Mathematics, edited by Timothy Gowers.]
88 Chapter 3 Integration

3B Limits of Integrals & Integrals of Limits


This section focuses on interchanging limits and integrals. Those tools allow us to
characterize the Riemann integrable functions in terms of Lebesgue measure. We
also develop some good approximation tools that will be useful in later chapters.

Bounded Convergence Theorem


We begin this section by introducing some useful notation.

3.24 Definition integration on a subset

R space and E ∈ S . If f : X → [−∞, ∞] is an


Suppose ( X, S , µ) is a measure
S -measurable function, then E f dµ is defined by
Z Z
f dµ = χ E f dµ
E
R
if the right side of the equation above is defined; otherwise E f dµ is undefined.
R R
Alternatively, you can think of E f dµ as f | E dµ E , where µ E is the measure
R in E.
obtained by restricting µ to the elements of S that are contained
RNotice that according to the definition above, the notation X f dµ means the same
as f dµ. The following easy result illustrates the use of this new notation.

3.25 bounding an integral

Suppose ( X, S , µ)R is a measure space, E ∈ S , and f : X → [−∞, ∞] is a


function such that E f dµ is defined. Then
Z

f dµ ≤ µ( E) sup| f |.
E E

Proof Let c = sup| f |. We have


E Z Z

f dµ = χ E f dµ
E
Z
≤ χ E| f | dµ
Z
≤ cχ E dµ

= cµ( E),

where the second line comes from 3.23, the third line comes from 3.8, and the fourth
line comes from 3.15.
Section 3B Limits of Integrals & Integrals of Limits 89

The next result could be proved as a special case of the Dominated Convergence
Theorem (3.31), which we prove later in this section. Thus you could skip the proof
here. However, sometimes you get more insight by seeing an easier proof of an
important special case. Thus you may want to read the easy proof of the Bounded
Convergence Theorem that is presented next.

3.26 Bounded Convergence Theorem

Suppose ( X, S , µ) is a measure space with µ( X ) < ∞. Suppose f 1 , f 2 , . . . is a


sequence of S -measurable functions from X to R that converges pointwise on X
to a function f : X → R. If there exists c ∈ (0, ∞) such that

| f k ( x )| ≤ c

for all k ∈ Z+ and all x ∈ X, then


Z Z
lim f k dµ = f dµ.
k→∞

Proof The function f is S -measurable


Note the key role of Egorov’s
by 2.48.
Theorem, which states that pointwise
Suppose c satisfies the hypothesis of
convergence is close to uniform
this theorem. Let ε > 0. By Egorov’s
convergence, in proofs involving
Theorem (2.85), there exists E ∈ S such
interchanging limits and integrals.
that µ( X \ E) < 4cε and f 1 , f 2 , . . . con-
verges uniformly to f on E. Now

Z Z Z Z Z

f k dµ − f dµ = f k dµ − f dµ + ( f k − f ) dµ
X\E X\E E
Z Z Z
≤ | f k | dµ + | f | dµ + | f k − f | dµ
X\E X\E E
ε
< + µ( E) sup| f k − f |,
2 E

where the last inequality follows from 3.25. Because f 1 , f 2 , . . . converges uniformly
to f on E and µ( E) < ∞, the right side of the inequality above is less than ε for k
sufficiently large, which completes the proof.

Sets of Measure 0 in Integration Theorems


Suppose ( X, S , µ) is a measure space. If f , g : X → [−∞, ∞] are S -measurable
functions and
µ({ x ∈ X : f ( x ) 6= g( x )}) = 0,
R R
then the definition of an integral implies that f dµ = g dµ (or both integrals are
undefined). Because what happens on a set of measure 0 often does not matter, the
following definition is useful.
90 Chapter 3 Integration

3.27 Definition almost every

Suppose ( X, S , µ) is a measure space. A set E ∈ S is said to contain µ-almost


every element of X if µ( X \ E) = 0. If the measure µ is clear from the context,
then the phrase almost every can be used (abbreviated by some authors to a. e.).

For example, almost every real number is irrational (with respect to the usual
Lebesgue measure on R) because |Q| = 0.
Theorems about integrals can almost always be relaxed so that the hypotheses
apply only almost everywhere instead of everywhere. For example, consider the
Bounded Convergence Theorem (3.26), one of whose hypotheses is that
lim f k ( x ) = f ( x )
k→∞
for all x ∈ X. Suppose that the hypotheses of the Bounded Convergence Theorem
hold except that the equation above holds only almost everywhere, meaning there
is a set E ∈ S such that µ( X \ E) = 0 and the equation above holds for all x ∈ E.
Define new functions g1 , g2 , . . . and g by
( (
f k ( x ) if x ∈ E, f ( x ) if x ∈ E,
gk ( x ) = and g( x ) =
0 if x ∈ X \ E 0 if x ∈ X \ E.
Then
lim gk ( x ) = g( x )
k→∞
for all x ∈ X. Hence the Bounded Convergence Theorem implies that
Z Z
lim gk dµ = g dµ,
k→∞
which immediately implies that
Z Z
lim f k dµ = f dµ
k→∞
R R R R
because gk dµ = f k dµ and g dµ = f dµ.

Dominated Convergence Theorem


The next result tells us that if a nonnegative function has a finite integral, then its
integral over all small sets (in the sense of measure) is small.

3.28 integrals on small sets are small

Suppose
R ( X, S , µ) is a measure space, g : X → [0, ∞] is S -measurable, and
g dµ < ∞. Then for every ε > 0, there exists δ > 0 such that
Z
g dµ < ε
B

for every set B ∈ S such that µ( B) < δ.


Section 3B Limits of Integrals & Integrals of Limits 91

Proof Suppose ε > 0. Let h : X → [0, ∞) be a simple S -measurable function such


that 0 ≤ h ≤ g and Z Z
ε
g dµ − h dµ < ;
2
the existence of a function h with these properties follows from 3.9. Let
H = max{ h( x ) : x ∈ X }
and let δ > 0 be such that Hδ < 2ε .
Suppose B ∈ S and µ( B) < δ. Then
Z Z Z
g dµ = ( g − h) dµ + h dµ
B B B
Z
≤ ( g − h) dµ + Hµ( B)
ε
< + Hδ
2
< ε,
as desired.

Some theorems, such as Egorov’s Theorem (2.85) have as a hypothesis that the
measure of the entire space is finite. The next result sometimes allows us to get
around this hypothesis by restricting attention to a key set of finite measure.

3.29 integrable functions live mostly on sets of finite measure

Suppose
R ( X, S , µ) is a measure space, g : X → [0, ∞] is S -measurable, and
g dµ < ∞. Then for every ε > 0, there exists E ∈ S such that µ( E) < ∞ and
Z
g dµ < ε.
X\E

Proof Suppose ε > 0. Let P be an S -partition A1 , . . . , Am of X such that


Z
3.30 g dµ < ε + L( g, P).

Let E be the union of those A j such that inf f > 0. Then µ( E) < ∞ (because
Aj
otherwise
R we would have L( g, P) = ∞, which contradicts the hypothesis that
g dµ < ∞). Now Z Z Z
g dµ = g dµ − χ E g dµ
X\E

< ε + L( g, P) − L(χ E g, P)
= ε,
where the second line follows from 3.30 and the definition of the integral of a
nonnegative function, and the last line holds because inf f = 0 for each A j not
Aj
contained in E.
92 Chapter 3 Integration

Suppose ( X, S , µ) is a measure space and f 1 , f 2 , . . . is a sequence of S -measurable


functions on X such that limk→∞ f k (Rx ) = f ( x )Rfor every (or almost every) x ∈ X.
In general, it is not true that limk→∞ f k dµ = f dµ (see Exercises 1 and 2).
We already have two good theorems about interchanging limits and integrals.
However, both of these theorems have restrictive hypotheses. Specifically, the Mono-
tone Convergence Theorem (3.11) requires all the functions to be nonnegative and
it requires the sequence of functions to be increasing. The Bounded Convergence
Theorem (3.26) requires the measure of the whole space to be finite and it requires
the sequence of functions to be uniformly bounded by a constant.
The next theorem is the grand result in this area. It does not require the sequence
of functions to be nonnegative, it does not require the sequence of functions to
be increasing, it does not require the measure of the whole space to be finite, and
it does not require the sequence of functions to be uniformly bounded. All these
hypotheses are replaced only by a requirement that the sequence of functions is
pointwise bounded by a function with a finite integral.
Notice that the Bounded Convergence Theorem follows immediately from the
result below (take g to be an appropriate constant function and use the hypothesis in
the Bounded Convergence Theorem that µ( X ) < ∞).

3.31 Dominated Convergence Theorem

Suppose ( X, S , µ) is a measure space, f : X → [−∞, ∞] is S -measurable, and


f 1 , f 2 , . . . are S -measurable functions from X to [−∞, ∞] such that

lim f k ( x ) = f ( x )
k→∞

for almost every x ∈ X. If there exists an S -measurable function g : X → [0, ∞]


such that Z
g dµ < ∞ and | f k ( x )| ≤ g( x )

for every k ∈ Z+ and almost every x ∈ X, then


Z Z
lim f k dµ = f dµ.
k→∞

Proof Suppose g : X → [0, ∞] satisfies the hypotheses of this theorem. If E ∈ S ,


then
Z Z Z Z Z Z

f k dµ − f dµ = f k dµ − f dµ + f k dµ − f dµ
X\E X\E E E
Z Z Z Z

≤ f k dµ + f dµ + f k dµ − f dµ
X\E X\E E E
Z Z Z

3.32 ≤2 g dµ + f k dµ − f dµ .
X\E E E
Section 3B Limits of Integrals & Integrals of Limits 93

Case 1: Suppose µ( X ) < ∞.


Let ε > 0. By 3.28, there exists δ > 0 such that
Z
ε
3.33 g dµ <
B 4
for every set B ∈ S such that µ( B) < δ. By Egorov’s Theorem (2.85), there exists
a set E ∈ S such that µ( X \ E) < δ and f 1 , f 2 , . . . converges uniformly to f on E.
Now 3.32 and 3.33 imply that
Z Z ε
Z

f k dµ − f dµ < + ( f k − f ) dµ .
2 E

Because f 1 , f 2 , . . . converges uniformly to f on E and µ( E) < ∞,


R the last term
R on
the right is less than 2ε for all sufficiently large k. Thus limk→∞ f k dµ = f dµ,
completing the proof of case 1.
Case 2: Suppose µ( X ) = ∞.
Let ε > 0. By 3.29, there exists E ∈ S such that µ( E) < ∞ and
Z
ε
g dµ < .
X\E 4

The inequality above and 3.32 imply that


Z Z ε
Z Z

f k dµ − f dµ < + f k dµ − f dµ .
2 E E

By case 1 as applied to the sequence f 1 | E , f 2 | E , . R. ., the last term


R on the right is less
than 2ε for all sufficiently large k. Thus limk→∞ f k dµ = f dµ, completing the
proof of case 2.

Riemann Integrals and Lebesgue Integrals


We can now use the tools we have developed to characterize the Riemann integrable
functions. In the theorem below, the left side of the last equation denotes the Riemann
integral.

3.34 Riemann integrable ⇐⇒ continuous almost everywhere

Suppose a < b and f : [ a, b] → R is a bounded function. Then f is Riemann


integrable if and only if

|{ x ∈ [ a, b] : f is not continuous at x }| = 0.

Furthermore, if f is Riemann integrable and λ denotes Lebesgue measure on R,


then f is Lebesgue measurable and
Z b Z
f = f dλ.
a [ a, b]
94 Chapter 3 Integration

Proof Suppose n ∈ Z+. Consider the partition Pn that divides [ a, b] into 2n sub-
intervals of equal size. Let I1 , . . . , I2n be the corresponding closed subintervals, each
of length (b − a)/2n . Let

2n  2n 
3.35 gn = ∑ inf f χ I
Ij j
and hn = ∑ sup f χ I .
j
j =1 j =1 Ij

The lower and upper Riemann sums of f for the partition Pn are given by integrals.
Specifically,
Z Z
3.36 L( f , Pn , [ a, b]) = gn dλ and U ( f , Pn , [ a, b]) = hn dλ,
[ a, b] [ a, b]

where λ is Lebesgue measure on R.


The definitions of gn and hn given in 3.35 are actually just a first draft of the
definitions. A slight problem arises at each point that is in two of the intervals
I1 , . . . , I2n (in other words, at endpoints of these intervals other than a and b). At
each of these points, change the value of gn to be the infimum of f over the union
of the two intervals that contain the point, and change the value of hn to be the
supremum of f over the union of the two intervals that contain the point. This change
modifies gn and hn on only a finite number of points. Thus the integrals in 3.36 are
not affected. This change is needed in order to make 3.38 true (otherwise the two
sets in 3.38 might differ by at most countably many points, which would not really
change the proof but which would not be as aesthetically pleasing).
Clearly g1 ≤ g2 ≤ · · · is an increasing sequence of functions and h1 ≥ h2 ≥ · · ·
is a decreasing sequence of functions on [ a, b]. Define functions f L : [ a, b] → R and
f U : [ a, b] → R by

f L ( x ) = lim gn ( x ) and f U ( x ) = lim hn ( x ).


n→∞ n→∞

Taking the limit as n → ∞ of both equations in 3.36 and using the Bounded Conver-
gence Theorem (3.26) along with Exercise 7 in Section 1A, we see that f L and f U
are Lebesgue measurable functions and
Z Z
3.37 L( f , [ a, b]) = f dλ
L
and U ( f , [ a, b]) = f U dλ.
[ a, b] [ a, b]

Now 3.37 implies that f is Riemann integrable if and only if


Z
( f U − f L ) dλ = 0.
[ a, b]

Because f L ( x ) ≤ f ( x ) ≤ f U ( x ) for all x ∈ [ a, b], the equation above holds if and


only if
|{ x ∈ [ a, b] : f U ( x ) 6= f L ( x )}| = 0.
The remaining details of the proof can be completed by noting that

3.38 { x ∈ [ a, b] : f U ( x ) 6= f L ( x )} = { x ∈ [ a, b] : f is not continuous at x }.


Section 3B Limits of Integrals & Integrals of Limits 95

Rb
We previously defined the notation a f to mean the Riemann integral of f .
Because the Riemann integral and Lebesgue integral agree for Riemann integrable
Rb
functions (see 3.34), we now redefine a f to denote the Lebesgue integral.

Rb
3.39 Definition a f

Suppose −∞ ≤ a < b ≤ ∞ and f : ( a, b) → R is Lebesgue measurable. Then


Rb Rb R
• a f and a f ( x ) dx mean (a,b) f dλ, where λ is Lebesgue measure on R;
Ra Rb
• b f is defined to be − a f.

The definition in the second bullet point above is made so that equations such as
Z b Z c Z b
f = f+ f
a a c
remain valid even if, for example, a < b < c.

Approximation by Nice Functions


In the next definition, the notation k f k1 should be k f k1,µ because it depends upon
the measure µ as well as upon f . However, µ is usually clear from the context. In
some books, you may see the notation L1 ( X, S , µ) instead of L1 (µ).

3.40 Definition k f k1 ; L1 ( µ )

Suppose ( X, S , µ) is a measure space. If f : X → [−∞, ∞] is S -measurable,


then the L1 -norm of f is denoted by k f k1 and is defined by
Z
k f k1 = | f | dµ.

The Lebesgue space L1 (µ) is defined by

L1 (µ) = { f : f is an S -measurable function from X to R and k f k1 < ∞}.

The terminology and notation used above are convenient even though k·k1 might
not be a genuine norm (to be defined in Chapter 6).

3.41 Example L1 (µ) functions that take on only finitely many values
Suppose ( X, S , µ) is a measure space and E1 , . . . , En are disjoint subsets of X.
Suppose a1 , . . . , an are distinct nonzero real numbers. Then
a1 χ E + · · · + a n χ E ∈ L1 ( µ )
1 n

if and only if Ek ∈ S and µ( Ek ) < ∞ for all k ∈ {1, . . . , n}. Furthermore,


k a1 χ E1 + · · · + an χ En k1 = | a1 |µ( E1 ) + · · · + | an |µ( En ).
96 Chapter 3 Integration

3.42 Example `1
If µ is counting measure on Z+ and x = x1 , x2 , . . . is a sequence of real numbers
(thought of as a function on Z+ ), then k x k1 = ∑∞ 1
k=1 | xk |. In this case, L ( µ ) is
1 1
often denoted by ` (pronounced little-el-one). In other words, ` is the set of all
sequences x1 , x2 , . . . of real numbers such that ∑∞
k=1 | xk | < ∞.

The easy proof of the following result is left to the reader.

3.43 properties of the L1 -norm

Suppose ( X, S , µ) is a measure space and f , g ∈ L1 (µ). Then


• k f k1 ≥ 0;
• k f k1 = 0 if and only if f ( x ) = 0 for almost every x ∈ X;
• kc f k1 = |c|k f k1 for all c ∈ R;
• k f + g k1 ≤ k f k1 + k g k1 .

The next result states that every function in L1 (µ) can be approximated in L1 -
norm by measurable functions that take on only finitely many values.

3.44 approximation by simple functions

Suppose µ is a measure and f ∈ L1 (µ). Then for every ε > 0, there exists a
simple function g ∈ L1 (µ) such that

k f − gk1 < ε.

Proof Suppose ε > 0. Then there exist simple functions g1 , g2 ∈ L1 (µ) such that
0 ≤ g1 ≤ f + and 0 ≤ g2 ≤ f − and
Z Z
ε ε
( f + − g1 ) dµ < and ( f − − g2 ) dµ < ,
2 2
where we have used 3.9 to provide the existence of g1 , g2 with these properties.
Let g = g1 − g2 . Then g is a simple function in L1 (µ) and

k f − gk1 = k( f + − g1 ) − ( f − − g2 )k1
Z Z
= ( f + − g1 ) dµ + ( f − − g2 ) dµ

< ε,

as desired.
Section 3B Limits of Integrals & Integrals of Limits 97

3.45 Definition L1 ( R ); k f k1

• The notation L1 (R) denotes L1 (λ), where λ is Lebesgue measure on either


the Borel subsets of R or the Lebesgue measurable subsets of R.
• When working with L1 (R), the notation k f k1 denotes the integral of the
absolute value of f with respect to Lebesgue measure on R.

3.46 Definition step function

A step function is a function g : R → R of the form

g = a1 χ I + · · · + a n χ I ,
1 n

where I1 , . . . , In are intervals of R and a1 , . . . , an are nonzero real numbers.

Suppose g is a step function of the form above and the intervals I1 , . . . , In are
disjoint. Then
k gk1 = | a1 | | I1 | + · · · + | an | | In |.
In particular, g ∈ L1 (R) if and only if all the intervals I1 , . . . , In are bounded.
The intervals in the definition of a step
Even though the coefficients
function can be open intervals, closed in-
a1 , . . . , an in the definition of a step
tervals, or half-open intervals. We will be
function are required to be nonzero,
using step functions in integrals, where
the function 0 that is identically 0 on
the inclusion or exclusion of the endpoints
R is a step function. To see this, take
of the intervals does not matter.
n = 1, a1 = 1, and I1 = ∅.

3.47 approximation by step functions

Suppose f ∈ L1 (R). Then for every ε > 0, there exists a step function
g ∈ L1 (R) such that
k f − gk1 < ε.

Proof Suppose ε > 0. By 3.44, there exist Borel (or Lebesgue) measurable subsets
A1 , . . . , An of R and nonzero numbers a1 , . . . , an such that | Ak | < ∞ for all k ∈
{1, . . . , n} and
n ε

f − ∑ a k χ Ak < .
k =1 1 2
For each k ∈ {1, . . . , n}, there is an open subset Gk of R that contains Ak and
whose Lebesgue measure is as close as we want to | Ak | [by part (e) of 2.71]. Each
open subset of R, including each Gk , is a countable union of disjoint open intervals.
Thus for each k, there is a set Ek that is a finite union of bounded open intervals
contained in Gk whose Lebesgue measure is as close as we want to | Gk |. Hence for
each k, there is a set Ek that is a finite union of bounded intervals such that
98 Chapter 3 Integration

| Ek \ Ak | + | Ak \ Ek | ≤ | Gk \ Ak | + | Gk \ Ek |
ε
< ;
2| a k | n
in other words,
ε
χ − χ E 1 < .
Ak k 2| a k | n
Now
n n n n

f − ∑ ak χ E ≤ f −
k 1
∑ ak χ A + ∑ ak χ A −
k 1 k
∑ ak χ E
k 1
k =1 k =1 k =1 k =1
n
ε
< + ∑ | a k | χ A − χ E 1
2 k =1 k k

< ε.
Each Ek is a finite union of bounded intervals. Thus the inequality above completes
the proof because ∑nk=1 ak χ E is a step function.
k

Luzin’s Theorem (2.91 and 2.93) gives a spectacular way to approximate a Borel
measurable function by a continuous function. However, the following approximation
theorem is usually more useful than Luzin’s Theorem. For example, the next result
plays a major role in the proof of the Lebesgue Differentiation Theorem (4.10).

3.48 approximation by continuous functions

Suppose f ∈ L1 (R). Then for every ε > 0, there exists a continuous function
g : R → R such that
k f − g k1 < ε
and { x ∈ R : g( x ) 6= 0} is a bounded set.

Proof For every a1 , . . . , an , b1 , . . . , bn , c1 , . . . , cn ∈ R and g1 , . . . , gn ∈ L1 (R),


we have
n n n

f − ∑ a k gk ≤ f − ∑ a k χ [ b , c ] + ∑ a k ( χ [ b , c ] − gk )
k =1 1 k =1
k k 1 k =1
k k 1

n n

≤ f − ∑ ak χ[b , c ] +
k k 1
∑ | a k | k χ [ b , c ] − gk k1 ,
k k
k =1 k =1

where the inequalities above follow


from 3.43. By 3.47, we can choose
a1 , . . . , an , b1 , . . . , bn , c1 , . . . , cn ∈ R to
make k f − ∑nk=1 ak χ[b , c ]k1 as small
k k
as we wish. The figure here then
shows that there exist continuous func-
tions g1 , . . . , gn ∈ L1 (R) that make
∑nk=1 | ak |kχ[bk , ck ] − gk k1 as small as we The graph of a continuous function gk
wish. Now take g = ∑nk=1 ak gk . such that kχ[b , c ] − gk k1 is small.
k k
Section 3B Limits of Integrals & Integrals of Limits 99

EXERCISES 3B
1 Give an example of a sequence f 1 , f 2 , . . . of functions from Z+ to [0, ∞) such
that
lim f k (m) = 0
k→∞
Z
for every m ∈ Z+ but lim f k dµ = 1, where µ is counting measure on Z+.
k→∞

2 Give an example of a sequence f 1 , f 2 , . . . of continuous functions from R to


[0, 1] such that
lim f k ( x ) = 0
k→∞
Z
for every x ∈ R but lim f k dλ = ∞, where λ is Lebesgue measure on R.
k→∞

3 Suppose λ is Lebesgue
R measure on R and f : R → R is a Borel measurable
function such that | f | dλ < ∞. Define g : R → R by
Z
g( x ) = f dλ.
(−∞, x )

Prove that g is uniformly continuous on R.


4 (a) Suppose ( X, S , µ) is a measure space with µ( X ) < ∞. Suppose that
f : X → [0, ∞) is a bounded S -measurable function. Prove that
Z n m o
f dµ = inf ∑ µ( A j ) sup f : A1 , . . . , Am is an S -partition of X .
j =1 Aj

R if the hypothesis that f is


(b) Show that the conclusion of part (a) can fail
bounded is replaced by the hypothesis that f dµ < ∞.
(c) Show that the conclusion of part (a) can fail if the condition that µ( X ) < ∞
is deleted.
[Part (a) of this exercise shows that if we had defined an upper Lebesgue sum,
then we could have used it to define the integral. However, parts (b) and (c) show
that the hypotheses that f is bounded and that µ( X ) < ∞ would be needed if
defining the integral via the equation above. The definition of the integral via the
lower Lebesgue sum does not require these hypotheses, showing the advantage
of using the approach via the lower Lebesgue sum.]
5 R measure on R. Suppose f : R → R is a Borel measurable
Let λ denote Lebesgue
function such that | f | dλ < ∞. Prove that
Z Z
lim f dλ = f dλ.
k→∞ [−k, k]

6 Let λ denote Lebesgue measure onRR. Give an example of a continuous


R function
f : [0, ∞) → R such that limt→∞ [0, t] f dλ exists (in R) but [0, ∞) f dλ is not
defined.
100 Chapter 3 Integration

7 Let λ denote Lebesgue measure onRR. Give an example of a continuous


R function
f : (0, 1) → R such that limn→∞ ( 1 , 1) f dλ exists (in R) but (0, 1) f dλ is not
n
defined.
8 Verify the assertion in 3.38.
9 Verify the assertion in Example 3.41.
10 (a) Suppose ( X, S , µ) is a measure space such that µ( X ) < ∞. Suppose
p, r are positive numbers with p R< r. Prove that if Rf : X → [0, ∞) is an
S -measurable function such that f r dµ < ∞, then f p dµ < ∞.
(b) Give an example to show that the result in part (a) can be false without the
hypothesis that µ( X ) < ∞.
11 Suppose ( X, S , µ) is a measure space and f ∈ L1 (µ). Prove that
{ x ∈ X : f ( x ) 6 = 0}
is the countable union of sets with finite µ-measure.
12 Suppose
(1 − x )k cos xk
f k (x) = √ .
x
Z 1
Prove that lim f k = 0.
k→∞ 0
13 Give an example of a sequence of nonnegative Borel measurable functions
f 1 , f 2 , . . . on [0, 1] such that both the following conditions hold:
Z 1
• lim f k = 0;
k→∞ 0

• sup f k ( x ) = ∞ for every m ∈ Z+ and every x ∈ [0, 1].


k≥m

14 Let λ denote Lebesgue measure on R.


√ R
(a) Let f ( x ) = 1/ x. Prove that [0, 1] f dλ = 2.
R
(b) Let f ( x ) = 1/(1 + x2 ). Prove that R f dλ = π.
R
(c) Let f ( x ) = (sin x )/x. Show that the integral (0, ∞) f dλ is not defined
R
but limt→∞ (0, t) f dλ exists in R.

15 Prove or give a counterexample: If G is an open subset of (0, 1), then χG is


Riemann integrable on [0, 1].
16 Suppose f ∈ L1 (R).
(a) For t ∈ R, define f t : R → R by f t ( x ) = f ( x − t). Prove that
limk f − f t k1 = 0.
t →0

(b) For t > 0, define f t : R → R by f t ( x ) = f (tx ). Prove that


limk f − f t k1 = 0.
t →1
Chapter 4
Differentiation

Does there exist a Lebesgue measurable set that fills up exactly half of each interval?
To get a feeling for this question, consider the set E = [0, 18 ] ∪ [ 41 , 38 ] ∪ [ 21 , 58 ] ∪ [ 34 , 78 ].
This set E has the property that

b
| E ∩ [0, b]| =
2

for b = 0, 14 , 12 , 34 , 1. Does there exist a Lebesgue measurable set E ⊂ [0, 1], perhaps
constructed in a fashion similar to the Cantor set, such that the equation above holds
for all b ∈ [0, 1]?
In this chapter we see how to answer this question by considering differentia-
tion issues. We begin by developing a powerful tool called the Hardy–Littlewood
maximal inequality. This tool is used to prove an almost everywhere version of the
Fundamental Theorem of Calculus. These results lead us to an important theorem
about the density of Lebesgue measurable sets.

Trinity College at the University of Cambridge in England. G. H. Hardy


(1877–1947) and John Littlewood (1885–1977) were students and later faculty
members here. If you have not already done so, you should read Hardy’s remarkable
book A Mathematician’s Apology (do not skip the fascinating Foreword by C. P.
Snow) and see the movie The Man Who Knew Infinity, which focuses on Hardy,
Littlewood, and Srinivasa Ramanujan (1887–1920).
CC-BY-SA Rafa Esteve

101
102 Chapter 4 Differentiation

4A Hardy–Littlewood Maximal Function


Markov’s Inequality
The following result, called Markov’s inequality, has a sweet, short proof. We will
make good use of this result later in this chapter (see the proof of 4.10). Markov’s
inequality also leads to Chebyshev’s inequality (see Exercise 2 in this section).

4.1 Markov’s inequality

Suppose ( X, S , µ) is a measure space and h ∈ L1 (µ). Then

1
µ({ x ∈ X : |h( x )| ≥ c}) ≤ k h k1
c
for every c > 0.

Proof Suppose c > 0. Then


Z
1
µ({ x ∈ X : |h( x )| ≥ c}) = c dµ
c { x ∈ X : |h( x )|≥c}
Z
1
≤ |h| dµ
c { x ∈ X : |h( x )|≥c}

1
≤ k h k1 ,
c
as desired.

St. Petersburg University along the Neva River in St. Petersburg, Russia.
Andrei Markov (1856–1922) was a student and then a faculty member here.
CC-BY-SA A. Savin
Section 4A Hardy–Littlewood Maximal Function 103

Vitali Covering Lemma

4.2 Definition 3 times a bounded nonempty open interval

Suppose I is a bounded nonempty open interval of R. Then 3 ∗ I denotes the


open interval with the same center as I and three times the length of I.

4.3 Example 3 times an interval


If I = (0, 10), then 3 ∗ I = (−10, 20).

The next result is a key tool in the proof of the Hardy–Littlewood maximal
inequality (4.8).

4.4 Vitali Covering Lemma

Suppose I1 , . . . , In is a list of bounded nonempty open intervals of R. Then there


exists a disjoint sublist Ik1 , . . . , Ikm such that

I1 ∪ · · · ∪ In ⊂ (3 ∗ Ik1 ) ∪ · · · ∪ (3 ∗ Ikm ).

4.5 Example Vitali Covering Lemma


Suppose n = 4 and
I1 = (0, 10), I2 = (9, 15), I3 = (14, 22), I4 = (21, 31).
Then
3 ∗ I1 = (−10, 20), 3 ∗ I2 = (3, 21), 3 ∗ I3 = (6, 30), 3 ∗ I4 = (11, 41).
Thus
I1 ∪ I2 ∪ I3 ∪ I4 ⊂ (3 ∗ I1 ) ∪ (3 ∗ I4 ).
In this example, I1 , I4 is the only sublist of I1 , I2 , I3 , I4 that produces the conclusion
of the Vitali Covering Lemma.

Proof of 4.4 Let k1 be such that


| Ik1 | = max{| I1 |, . . . , | In |}.
Suppose k1 , . . . , k j have been chosen.
The technique used here is called a
Let k j+1 be such that | Ik j+1 | is as large
greedy algorithm because at each
as possible subject to the condition that stage we select the largest remaining
Ik1 , . . . , Ik j+1 are disjoint. If there is no interval that is disjoint from the
choice of k j+1 such that Ik1 , . . . , Ik j+1 are previously selected intervals.
disjoint, then the procedure terminates.
Because we start with a finite list, the procedure must eventually terminate after some
number m of choices.
104 Chapter 4 Differentiation

Suppose j ∈ {1, . . . , n}. To complete the proof, we must show that

Ij ⊂ (3 ∗ Ik1 ) ∪ · · · ∪ (3 ∗ Ikm ).

If j ∈ {k1 , . . . , k m }, then the inclusion above obviously holds.


Thus assume that j ∈ / {k1 , . . . , k m }. Because the process terminated without
selecting j, the interval Ij is not disjoint from all of Ik1 , . . . , Ikm . Let Ik L be the first
interval on this list not disjoint from Ij ; thus Ij is disjoint from Ik1 , . . . , Ik L−1 . Because
j was not chosen in step L, we conclude that | Ik L | ≥ | Ij |. Because Ik L ∩ Ij 6= ∅, this
last inequality implies (easy exercise) that Ij ⊂ 3 ∗ Ik L , completing the proof.

Hardy–Littlewood Maximal Inequality


Now we come to a brilliant definition that turns out to be extraordinarily useful.

4.6 Definition Hardy–Littlewood maximal function

Suppose h : R → R is a Lebesgue measurable function. Then the Hardy–


Littlewood maximal function of h is the function h∗ : R → [0, ∞] defined by
Z b+t
1
h∗ (b) = sup | h |.
t >0 2t b−t

In other words, h∗ (b) is the supremum over all bounded intervals centered at b of
the average of |h| on those intervals.

4.7 Example Hardy–Littlewood maximal function of χ[0, 1]

As usual, let χ[0, 1] denote the characteristic function of the interval [0, 1]. Then

 1

 if b ≤ 0,
 2(1− b )

(χ[0, 1]) (b) = 1 if 0 < b < 1,


1
if b ≥ 1,
2b The graph of (χ[0, 1])∗ on [−2, 3].

as you should verify.

If h : R → R is Lebesgue measurable and c ∈ R, then {b ∈ R : h∗ (b) > c} is


an open subset of R, as you are asked to prove in Exercise 9 in this section. Thus h∗
is a Borel measurable function.
Suppose h ∈ L1 (R) and c > 0. Markov’s inequality (4.1) estimates the size of
the set on which |h| is larger than c. Our next result estimates the size of the set on
which h∗ is larger than c. The Hardy–Littlewood maximal inequality proved in the
next result is a key ingredient in the proof of the Lebesgue Differentiation Theorem
(4.10). Note that this next result is considerably deeper than Markov’s inequality.
Section 4A Hardy–Littlewood Maximal Function 105

4.8 Hardy–Littlewood maximal inequality

Suppose h ∈ L1 (R). Then

3
|{b ∈ R : h∗ (b) > c}| ≤ k h k1
c
for every c > 0.

Proof Suppose F Ris a closed bounded subset of {b ∈ R : h∗ (b) > c}. We will

show that | F | ≤ 3c −∞ |h|, which implies our desired result [see Exercise 24(a) in
Section 2D].
For each b ∈ F, there exists tb > 0 such that
Z b+t
1 b
4.9 |h| > c.
2tb b−tb

Clearly [
F⊂ ( b − t b , b + t b ).
b∈ F
The Heine–Borel Theorem (2.12) tells us that this open cover of a closed bounded set
has a finite subcover. In other words, there exist b1 , . . . , bn ∈ F such that
F ⊂ (b1 − tb1 , b1 + tb1 ) ∪ · · · ∪ (bn − tbn , bn + tbn ).
To make the notation cleaner, relabel the open intervals above as I1 , . . . , In .
Now apply the Vitali Covering Lemma (4.4) to the list I1 , . . . , In , producing a
disjoint sublist Ik1 , . . . , Ikm such that
I1 ∪ · · · ∪ In ⊂ (3 ∗ Ik1 ) ∪ · · · ∪ (3 ∗ Ikm ).
Thus
| F | ≤ | I1 ∪ · · · ∪ In |
≤ |(3 ∗ Ik1 ) ∪ · · · ∪ (3 ∗ Ikm )|
≤ |3 ∗ Ik1 | + · · · + |3 ∗ Ikm |
= 3(| Ik1 | + · · · + | Ikm |)
Z Z
3 
< |h| + · · · + |h|
c Ik1 Ik m
Z ∞
3
≤ | h |,
c −∞

where the second-to-last inequality above comes from 4.9 (note that | Ik j | = 2tb for
the choice of b corresponding to Ik j ) and the last inequality holds because Ik1 , . . . , Ikm
are disjoint.
The last inequality completes the proof.
106 Chapter 4 Differentiation

EXERCISES 4A
1 Suppose ( X, S , µ) is a measure space and h : X → R is an S -measurable
function. Prove that
Z
1
µ({ x ∈ X : |h( x )| ≥ c}) ≤ |h| p dµ
cp
for all positive numbers c and p.
2 Suppose ( X, S , µ) is a measure space with µ( X ) = 1 and h ∈ L1 (µ). Prove
that
n Z o Z Z 2 
1
µ x ∈ X : h( x ) − h dµ ≥ c ≤ 2 h2 dµ − h dµ
c

for all c > 0.


[The result above is called Chebyshev’s inequality; it plays an important role
in probability theory. Pafnuty Chebyshev (1821–1894) was Markov’s thesis
advisor.]
3 Suppose ( X, S , µ) is a measure space. Suppose h ∈ L1 (µ) and khk1 > 0.
Prove that there is at most one number c ∈ (0, ∞) such that

1
µ({ x ∈ X : |h( x )| ≥ c}) = k h k1 .
c
4 Show that the constant 3 in the Vitali Covering Lemma (4.4) cannot be replaced
by a smaller positive constant.
5 Prove the assertion left as an exercise in the last sentence of the proof of the
Vitali Covering Lemma (4.4).
6 Verify the formula in Example 4.7 for the Hardy–Littlewood maximal function
of χ[0, 1].

7 Find a formula for the Hardy–Littlewood maximal function of the characteristic


function of [0, 1] ∪ [2, 3].
8 Find a formula for the Hardy–Littlewood maximal function of the function
h : R → [0, ∞) defined by
(
x if 0 ≤ x ≤ 1,
h( x ) =
0 otherwise.

9 Suppose h : R → R is Lebesgue measurable. Prove that

{b ∈ R : h∗ (b) > c}

is an open subset of R for every c ∈ R.


Section 4A Hardy–Littlewood Maximal Function 107

10 Prove or give a counterexample: If h : R → [0, ∞) is an increasing function,


then h∗ is an increasing function.
11 Give an example of a Borel measurable function h : R → [0, ∞) such that
h∗ (b) < ∞ for all b ∈ R but sup{h∗ (b) : b ∈ R} = ∞.

12 Show that |{b ∈ R : h∗ (b) = ∞}| = 0 for every h ∈ L1 (R).


13 Show that there exists h ∈ L1 (R) such that h∗ (b) = ∞ for every b ∈ Q.
14 Suppose h ∈ L1 (R). Prove that

3
|{b ∈ R : h∗ (b) ≥ c}| ≤ k h k1
c
for every c > 0.
[This result slightly strengthens the Hardy–Littlewood maximal inequality (4.8)
because the set on the left side above includes those b ∈ R such that h∗ (b) = c.
A much deeper strengthening comes from replacing the constant 3 in the Hardy–
Littlewood maximal inequality with a smaller constant. In 2003, Antonios
Melas answered what had been an open question about the best constant. He
√ that can replace 3 in the Hardy–Littlewood
proved that the smallest constant
maximal inequality is (11 + 61)/12 ≈ 1.56752; see Annals of Mathematics
157 (2003), 647–688.]
108 Chapter 4 Differentiation

4B Derivatives of Integrals
Lebesgue Differentiation Theorem
The next result states that the average amount by which a function in L1 (R) differs
from its values is small almost everywhere on small intervals. The 2 in the denomi-
nator of the fraction in the result below could be deleted, but its presence makes the
length of the interval of integration nicely match the denominator 2t.
The next result is called the Lebesgue Differentiation Theorem, even though no
derivative is in sight. However, we will soon see how another version of this result
deals with derivatives. The hard work takes place in the proof of this first version.

4.10 Lebesgue Differentiation Theorem, first version

Suppose f ∈ L1 (R). Then


Z b+t
1
lim | f − f (b)| = 0
t ↓0 2t b−t

for almost every b ∈ R.

Before getting to the formal proof of this first version of the Lebesgue Differen-
tiation Theorem, we pause to provide some motivation for the proof. If b ∈ R and
t > 0, then 3.25 gives the easy estimate
Z b+t
1
| f − f (b)| ≤ sup{| f ( x ) − f (b)| : | x − b| ≤ t}.
2t b−t

If f is continuous at b, then the right side of this inequality has limit 0 as t ↓ 0,


proving 4.10 in the special case in which f is continuous on R.
To prove the Lebesgue Differentiation Theorem, we will approximate an arbitrary
function in L1 (R) by a continuous function (using 3.48). The previous paragraph
shows that the continuous function has the desired behavior. We will use the Hardy–
Littlewood maximal inequality (4.8) to show that the approximation produces ap-
proximately the desired behavior. Now we are ready for the formal details of the
proof.
Proof of 4.10 Let δ > 0. By 3.48, for each k ∈ Z+ there exists a continuous
function hk : R → R such that

δ
4.11 k f − h k k1 < .
k2k
Let
1
Bk = {b ∈ R : | f (b) − hk (b)| ≤ k and ( f − hk )∗ (b) ≤ 1k }.
Then

4.12 R \ Bk = {b ∈ R : | f (b) − hk (b)| > 1k } ∪ {b ∈ R : ( f − hk )∗ (b) > 1k }.


Section 4B Derivatives of Integrals 109

Markov’s inequality (4.1) as applied to the function f − hk and 4.11 imply that
δ
4.13 |{b ∈ R : | f (b) − hk (b)| > 1k }| <
.
2k
The Hardy–Littlewood maximal inequality (4.8) as applied to the function f − hk
and 4.11 imply that

4.14 |{b ∈ R : ( f − hk )∗ (b) > 1k }| < .
2k
Now 4.12, 4.13, and 4.14 imply that
δ
|R \ Bk | < .
2k −2
Let

\
B= Bk .
k =1
Then
[∞ ∞ ∞
δ

4.15 |R \ B| = (R \ Bk ) ≤ ∑ |R \ Bk | < ∑ 2 k −2
= 4δ.
k =1 k =1 k =1

Suppose b ∈ B and t > 0. Then for each k ∈ Z+ we have


Z b+t Z b+t
1 1 
| f − f (b)| ≤ | f − hk | + |hk − hk (b)| + |hk (b) − f (b)|
2t b−t 2t b−t
 1 Z b+t 
≤ ( f − hk )∗ (b) + |hk − hk (b)| + |hk (b) − f (b)|
2t b−t
Z b+t
2 1
≤ + |hk − hk (b)|.
k 2t b−t

Because hk is continuous, the last term is less than 1k for all t > 0 sufficiently close to
0 (how close is sufficiently close depends upon k). In other words, for each k ∈ Z+,
we have Z
1 b+t 3
| f − f (b)| <
2t b−t k
for all t > 0 sufficiently close to 0.
Hence we conclude that
Z
1 b+t
lim | f − f (b)| = 0
t↓0 2t b−t
for all b ∈ B.
Let A denote the set of numbers a ∈ R such that
Z
1 a+t
lim | f − f ( a)|
t↓0 2t a−t

either does not exist or is nonzero. We have shown that A ⊂ (R \ B). Thus
| A| ≤ |R \ B| < 4δ,
where the last inequality comes from 4.15. Because δ is an arbitrary positive number,
the last inequality implies that | A| = 0, completing the proof.
110 Chapter 4 Differentiation

Derivatives
You should remember the following definition from your calculus course.

4.16 Definition derivative

Suppose g : I → R is a function defined on an open interval I of R and b ∈ I.


The derivative of g at b, denoted g0 (b), is defined by

g(b + t) − g(b)
g0 (b) = lim
t →0 t
if the limit above exists, in which case g is called differentiable at b.

We now turn to the Fundamental Theorem of Calculus and a powerful extension


that avoids continuity. These results show that differentiation and integration can be
thought of as inverse operations.
You saw the next result in your calculus class, except now the function f is
only required to be Lebesgue measurable (and its absolute value must have a finite
Lebesgue integral). Of course, we also need to require f to be continuous at the
crucial point b in the next result, because changing the value of f at a single number
would not change the function g.

4.17 Fundamental Theorem of Calculus

Suppose f ∈ L1 (R). Define g : R → R by


Z x
g( x ) = f.
−∞

Suppose b ∈ R and f is continuous at b. Then g is differentiable at b and

g 0 ( b ) = f ( b ).

Proof If t 6= 0, then
g(b + t) − g(b)
R b+t Rb
−∞ f− −∞ f
− f (b) = − f (b)
t t
R b + t
f
= b − f (b)
t
R b+t f − f (b )

4.18 = b
t
≤ sup | f ( x ) − f (b)|.
{ x ∈R : | x −b|<|t|}

If ε > 0, then by the continuity of f at b, the last quantity is less than ε for t
sufficiently close to 0. Thus g is differentiable at b and g0 (b) = f (b).
Section 4B Derivatives of Integrals 111

A function in L1 (R) need not be continuous anywhere. Thus the Fundamental


Theorem of Calculus (4.17) might provide no information about differentiating the
integral of such a function. However, our next result states that all is well almost
everywhere, even in the absence of any continuity of the function being integrated.

4.19 Lebesgue Differentiation Theorem, second version

Suppose f ∈ L1 (R). Define g : R → R by


Z x
g( x ) = f.
−∞

Then g0 (b) = f (b) for almost every b ∈ R.

Proof Suppose t 6= 0. Then from 4.18 we have


g(b + t) − g(b) R b+t f − f (b )

− f (b) = b
t t
Z b+t
1
≤ | f − f (b)|
t b
Z b+t
1
≤ | f − f (b)|
t b−t
for all b ∈ R. By the first version of the Lebesgue Differentiation Theorem (4.10),
the last quantity has limit 0 as t → 0 for almost every b ∈ R. Thus g0 (b) = f (b) for
almost every b ∈ R.

Now we can answer the question raised on the opening page of this chapter.

4.20 no set constitutes exactly half of each interval

There does not exist a Lebesgue measurable set E ⊂ [0, 1] such that

b
| E ∩ [0, b]| =
2
for all b ∈ [0, 1].

Proof Suppose there does exist a Lebesgue measurable set E ⊂ [0, 1] with the
property above. Define g : R → R by
Z b
g(b) = χE .
−∞

Thus g(b) = 2b for all b ∈ [0, 1]. Hence g0 (b) = 12 for all b ∈ (0, 1).
The Lebesgue Differentiation Theorem (4.19) implies that g0 (b) = χ E(b) for
almost every b ∈ R. However, χ E never takes on the value 12 , which contradicts the
conclusion of the previous paragraph. This contradiction completes the proof.
112 Chapter 4 Differentiation

The next result says that a function in L1 (R) is equal almost everywhere to the
limit of its average over small intervals. These two-sided results generalize more
naturally to higher dimensions (take the average over balls centered at b) than the
one-sided results.

4.21 L1 (R) function equals its local average almost everywhere

Suppose f ∈ L1 (R). Then


Z b+t
1
f (b) = lim f
t ↓0 2t b−t

for almost every b ∈ R.

Proof Suppose t > 0. Then


 1 Z b+t  1 Z b+t 

f − f (b) = f − f (b)
2t b−t 2t b−t
Z b+t
1
≤ | f − f (b)|.
2t b−t
The desired result now follows from the first version of the Lebesgue Differentiation
Theorem (4.10).

Again, the conclusion of the result above holds at every number b at which f is
continuous. The remarkable part of the result above is that even if f is discontinuous
everywhere, the conclusion holds for almost every real number b.

Density
The next definition captures the notion of the proportion of a set in small intervals
centered at a number b.

4.22 Definition density

Suppose E ⊂ R. The density of E at a number b ∈ R is

| E ∩ (b − t, b + t)|
lim
t ↓0 2t

if this limit exists (otherwise the density of E at b is undefined).

4.23 Example density of an interval



1
 if b ∈ (0, 1),
The density of [0, 1] at b = 1
if b = 0 or b = 1,
2

0 otherwise.
Section 4B Derivatives of Integrals 113

The next beautiful result shows the power of the techniques developed in this
chapter.

4.24 Lebesgue Density Theorem

Suppose E ⊂ R is a Lebesgue measurable set. Then the density of E is 1 at


almost every element of E and is 0 at almost every element of R \ E.

Proof First suppose | E| < ∞. Thus χ E ∈ L1 (R). Because


Z b+t
| E ∩ (b − t, b + t)| 1
= χE
2t 2t b−t
for every t > 0 and every b ∈ R, the desired result follows immediately from 4.21.
Now consider the case where | E| = ∞ [which means that χ E ∈ / L1 (R) and hence
4.21 as stated cannot be used]. For k ∈ Z , let Ek = E ∩ (−k, k). If |b| < k, then the
+
density of E at b equals the density of Ek at b. By the previous paragraph as applied
to Ek , there are sets Fk ⊂ Ek and Gk ⊂ R \ Ek such that | Fk | = | Gk | = 0 and the
density of Ek equals 1 at every element of Ek \ Fk and the density of Ek equals 0 at
every elementS
of (R \ Ek ) \ GS
k.
Let F = ∞ k =1 Fk and G = ∞
k=1 Gk . Then | F | = | G | = 0 and the density of E is
1 at every element of E \ F and is 0 at every element of (R \ E) \ G.

The bad Borel set provided by the next


The Lebesgue Density Theorem
result leads to a bad Borel measurable
makes the example provided by the
function. Specifically, let E be the bad
next result somewhat surprising. Be
Borel set in 4.25. Then χ E is a Borel
sure to spend some time pondering
measurable function that is discontinuous
why the next result does not
everywhere. Furthermore, the function χ E
contradict the Lebesgue Density
cannot be modified on a set of measure 0
Theorem. Also, compare the next
to be continuous anywhere (in contrast to
result to 4.20.
the function χQ).
Even though the function χ E discussed in the paragraph above is continuous
nowhere and every modification of this function on a set of measure 0 is also continu-
ous nowhere, the function g defined by
Z b
g(b) = χE
0
is differentiable almost everywhere (by 4.19).
The proof of 4.25 given below is based on an idea of Walter Rudin.

4.25 bad Borel set

There exists a Borel set E ⊂ R such that

0 < |E ∩ I | < | I |

for every nonempty bounded open interval I.


114 Chapter 4 Differentiation

Proof We use the following fact in our construction:

4.26 Suppose G is a nonempty open subset of R. Then there exists a closed set
F ⊂ G \ Q such that | F | > 0.

To prove 4.26, let J be a closed interval contained in G such that 0 < |J |. Let
r1 , r2 , . . . be a list of all the rational numbers. Let
∞ 
[ |J | |J | 
F=J\ rk − , r k + .
k =1
2k +2 2k +2

1
Then F is a closed subset of R and F ⊂ J \ Q ⊂ G \ Q. Also, |J \ F | ≤ 2 |J |
S  
because J \ F ⊂ ∞
|J | |J |
k =1 r k − 2 k + 2 , r k + 2 k + 2 . Thus

| F | = |J | − |J \ F | ≥ 12 |J | > 0,

completing the proof of 4.26.


To construct the set E with the desired properties, let I1 , I2 , . . . be a sequence
consisting of all nonempty bounded open intervals of R with rational endpoints. Let
F0 = F̂0 = ∅, and inductively construct sequences F1 , F2 , . . . and F̂1 , F̂2 , . . . of closed
subsets of R as follows: Suppose n ∈ Z+ and F0 , . . . , Fn−1 and F̂0 , . . . , F̂n−1 have
been chosen as closed sets that contain no rational numbers. Thus

In \ ( F̂0 ∪ . . . ∪ F̂n−1 )

is a nonempty open set (nonempty because it contains all rational numbers in In ).


Applying 4.26 to the open set above, we see that there is a closed set Fn contained in
the set above such that Fn contains no rational numbers and | Fn | > 0. Applying 4.26
again, but this time to the open set

In \ ( F0 ∪ . . . ∪ Fn ),

which is nonempty because it contains all rational numbers in In , we see that there is
a closed set F̂n contained in the set above such that F̂n contains no rational numbers
and | F̂n | > 0.
Now let

[
E= Fk .
k =1

Our construction implies that Fk ∩ F̂n = ∅ for all k, n ∈ Z+. Thus E ∩ F̂n = ∅ for
all n ∈ Z+. Hence F̂n ⊂ In \ E for all n ∈ Z+.
Suppose I is a nonempty bounded open interval. Then In ⊂ I for some n ∈ Z+.
Thus
0 < | Fn | ≤ | E ∩ In | ≤ | E ∩ I |.
Also,
| E ∩ I | = | I | − | I \ E| ≤ | I | − | In \ E| ≤ | I | − | F̂n | < | I |,
completing the proof.
Section 4B Derivatives of Integrals 115

EXERCISES 4B

For f ∈ L1 (R) and I an interval of R with R 0 < | I | < ∞, let f I denote the
average of f on I. In other words, f I = |1I | I f .

1 Suppose f ∈ L1 (R). Prove that


Z b+t
1
lim | f − f [b−t, b+t] | = 0
t ↓0 2t b−t

for almost every b ∈ R.


2 Suppose f ∈ L1 (R). Prove that
n 1 Z o
lim sup | f − f I | : I is an interval of length t containing b = 0
t ↓0 |I| I

for almost every b ∈ R.


3 Suppose f : R → R is a Lebesgue measurable function such that f 2 ∈ L1 (R).
Prove that Z
1 b+t
lim | f − f (b)|2 = 0
t↓0 2t b−t

for almost every b ∈ R.


4 Prove thatR the Lebesgue Differentiation Theorem (4.19) stillRholds if the hypoth-
∞ x
esis that −∞ | f | < ∞ is weakened to the requirement that −∞ | f | < ∞ for all
x ∈ R.
5 Suppose f : R → R is a Lebesgue measurable function. Prove that

| f (b)| ≤ f ∗ (b)

for almost every b ∈ R.


Rs
6 Prove that if h ∈ L1 (R) and −∞ h = 0 for all s ∈ R, then h = 0.
7 Give an example of a Borel subset of R whose density at 0 is not defined.
8 Give an example of a Borel subset of R whose density at 0 is 13 .

9 Prove that if t ∈ [0, 1], then there exists a Borel set E ⊂ R such that the density
of E at 0 is t.
10 Suppose E is a Lebesgue measurable subset of R such that the density of E
equals 1 at every element of E and equals 0 at every element of R \ E. Prove
that E = ∅ or E = R.
Chapter 5
Product Measures

Lebesgue measure on R generalizes the notion of the length of an interval. In this


chapter, we see how two-dimensional Lebesgue measure on R2 generalizes the notion
of the area of a rectangle. More generally, we construct new measures that are the
products of two measures.
Once these new measures have been constructed, the question arises of how to
compute integrals with respect to these new measures. Beautiful theorems proved in
the first decade of the twentieth century allow us to compute integrals with respect to
product measures as iterated integrals involving the two measures that produced the
product. Furthermore, we will see that under reasonable conditions we can switch
the order of an iterated integral.

Main building of Scuola Normale Superiore di Pisa, the university in Pisa, Italy,
where Guido Fubini (1879–1943) received his PhD in 1900. In 1907 Fubini proved
that under reasonable conditions, an integral with respect to a product measure can
be computed as an iterated integral and that the order of integration can be switched.
Leonida Tonelli (1885–1943) also taught for many years in Pisa; he also proved a
crucial theorem about interchanging the order of integration in an iterated integral.
CC-BY-SA Lucarelli

116
Section 5A Products of Measure Spaces 117

5A Products of Measure Spaces


Products of σ-Algebras
Our first step in constructing product measures is to construct the product of two
σ-algebras. We begin with the following definition.

5.1 Definition rectangle

Suppose X and Y are sets. A rectangle in X × Y is a set of the form A × B,


where A ⊂ X and B ⊂ Y.

Keep the figure shown here in mind


when thinking of a rectangle in the sense
defined above. However, remember that
A and B need not be intervals as shown
in the figure. Indeed, the concept of an
interval makes no sense in the generality
of arbitrary sets.

Now we can define the product of two σ-algebras.

5.2 Definition product of two σ-algebras; S ⊗ T

Suppose ( X, S) and (Y, T ) are measurable spaces. Then


• the product S ⊗ T is defined to be the smallest σ-algebra on X × Y that
contains
{ A × B : A ∈ S , B ∈ T };

• a measurable rectangle in S ⊗ T is a set of the form A × B, where A ∈ S


and B ∈ T .

Using the terminology introduced in


The notation S × T is not used
the second bullet point above, we can say
because S and T are sets (of sets),
that S ⊗ T is the smallest σ-algebra con-
and thus the notation S × T
taining all the measurable rectangles in
already is defined to mean the set of
S ⊗ T . Exercise 1 in this section asks
all ordered pairs of the form ( A, B),
you to show that the measurable rectan-
where A ∈ S and B ∈ T .
gles in S ⊗ T are the only rectangles in
X × Y that are in S ⊗ T .
The notion of cross sections plays a crucial role in our development of product
measures. First, we define cross sections of sets, and then we define cross sections of
functions.
118 Chapter 5 Product Measures

5.3 Definition cross sections of sets; [ E] a and [ E]b

Suppose X and Y are sets and E ⊂ X × Y. Then for a ∈ X and b ∈ Y, the cross
sections [ E] a and [ E]b are defined by

[ E] a = {y ∈ Y : ( a, y) ∈ E} and [ E]b = { x ∈ X : ( x, b) ∈ E}.

5.4 Example cross sections of a subset of X × Y

5.5 Example cross sections of rectangles


Suppose X and Y are sets and A ⊂ X and B ⊂ Y. If a ∈ X and b ∈ Y, then
( (
B if a ∈ A, A if b ∈ B,
[ A × B] a = and [ A × B]b =
∅ if a ∈
/A ∅ if b ∈/ B,
as you should verify.

The next result shows that cross sections preserve measurability.

5.6 cross sections of measurable sets are measurable

Suppose S is a σ-algebra on X and T is a σ-algebra on Y. If E ∈ S ⊗ T , then

[ E] a ∈ T for every a ∈ X and [ E]b ∈ S for every b ∈ Y.

Proof Let E denote the collection of subsets E of X × Y for which the conclusion
of this result holds. Then A × B ∈ E for all A ∈ S and all B ∈ T (by Example 5.5).
The collection E is closed under complementation and countable unions because
[( X × Y ) \ E] a = Y \ [ E] a
and
[ E1 ∪ E2 ∪ · · · ] a = [ E1 ] a ∪ [ E2 ] a ∪ · · ·
for all subsets E, E1 , E2 , . . . of X × Y and all a ∈ X, as you should verify, with
similar statements holding for cross sections with respect to all b ∈ Y.
Because E is a σ-algebra containing all the measurable rectangles in S ⊗ T , we
conclude that E contains S ⊗ T .
Section 5A Products of Measure Spaces 119

Now we define cross sections of functions.

5.7 Definition cross sections of functions; [ f ] a and [ f ]b

Suppose X and Y are sets and f : X × Y → R is a function. Then for a ∈ X and


b ∈ Y, the cross section functions [ f ] a : Y → R and [ f ]b : X → R are defined
by

[ f ] a (y) = f ( a, y) for y ∈ Y and [ f ]b ( x ) = f ( x, b) for x ∈ X.

5.8 Example cross sections


• Suppose f : R × R → R is defined by f ( x, y) = 5x2 + y3 . Then
[ f ]2 (y) = 20 + y3 and [ f ]3 ( x ) = 5x2 + 27
for all y ∈ R and all x ∈ R, as you should verify.
• Suppose X and Y are sets and A ⊂ X and B ⊂ Y. If a ∈ X and b ∈ Y, then
[χ A × B] a = χ A( a)χ B and [χ A × B]b = χ B(b)χ A ,
as you should verify.

The next result shows that cross sections preserve measurability, this time in the
context of functions rather than sets.

5.9 cross sections of measurable functions are measurable

Suppose S is a σ-algebra on X and T is a σ-algebra on Y. Suppose


f : X × Y → R is an S ⊗ T -measurable function. Then

[ f ] a is a T -measurable function on Y for every a ∈ X

and
[ f ]b is an S -measurable function on X for every b ∈ Y.

Proof Suppose D is a Borel subset of R and a ∈ X. If y ∈ Y, then

y ∈ ([ f ] a )−1 ( D ) ⇐⇒ [ f ] a (y) ∈ D
⇐⇒ f ( a, y) ∈ D
⇐⇒ ( a, y) ∈ f −1 ( D )
⇐⇒ y ∈ [ f −1 ( D )] a .
Thus
([ f ] a )−1 ( D ) = [ f −1 ( D )] a .
Because f is an S ⊗ T -measurable function, f −1 ( D ) ∈ S ⊗ T . Thus the equation
above and 5.6 imply that ([ f ] a )−1 ( D ) ∈ T . Hence [ f ] a is a T -measurable function.
The same ideas show that [ f ]b is an S -measurable function for every b ∈ Y.
120 Chapter 5 Product Measures

Monotone Class Theorem


The following standard two-step technique often works to prove that every set in a
σ-algebra has a certain property:
1. show that every set in a collection of sets that generates the σ-algebra has the
property;
2. show that the collection of sets that has the property is a σ-algebra.
For example, the proof of 5.6 used the technique above—first we showed that every
measurable rectangle in S ⊗ T has the desired property, then we showed that the
collection of sets that has the desired property is a σ-algebra (this completed the proof
because S ⊗ T is the smallest σ-algebra containing the measurable rectangles).
The technique outlined above should be used when possible. However, in some
situations there seems to be no reasonable way to verify that the collection of sets
with the desired property is a σ-algebra. We will encounter this situation in the next
subsection. To deal with it, we need to introduce another technique that involves
what are called monotone classes.
The following definition will be used in our main theorem about monotone classes.

5.10 Definition algebra

Suppose W is a set and A is a set of subsets of W. Then A is called an algebra


on W if the following three conditions are satisfied:
• ∅ ∈ A;
• if E ∈ A, then W \ E ∈ A;
• if E and F are elements of A, then E ∪ F ∈ A.

Thus an algebra is closed under complementation and under finite unions; a


σ-algebra is closed under complementation and countable unions.

5.11 Example collection of finite unions of intervals is an algebra


Suppose A is the collection of all finite unions of intervals of R. Here we are in-
cluding all intervals—open intervals, closed intervals, bounded intervals, unbounded
intervals, sets consisting of only a single point, and intervals that are neither open nor
closed because they contain one endpoint but not the other endpoint.
Clearly A is closed under finite unions. You should also verify that A is closed
under complementation. Thus A is an algebra on R.

5.12 Example collection of countable unions of intervals is not an algebra


Suppose A is the collection of all countable unions of intervals of R.
Clearly A is closed under finite unions (and also under countable unions). You
should verify that A is not closed under complementation. Thus A is neither an
algebra nor a σ-algebra on R.
Section 5A Products of Measure Spaces 121

The following result provides an example of an algebra that we will exploit.

5.13 the set of finite unions of measurable rectangles is an algebra

Suppose ( X, S) and (Y, T ) are measurable spaces. Then

(a) the set of finite unions of measurable rectangles in S ⊗ T is an algebra


on X × Y;
(b) every finite union of measurable rectangles in S ⊗ T can be written as a
finite union of disjoint measurable rectangles in S ⊗ T .

Proof Let A denote the set of finite unions of measurable rectangles in S ⊗ T .


Obviously A is closed under finite unions.
The collection A is also closed under finite intersections. To verify this claim,
note that if A1 , . . . , An , C1 , . . . , Cm ∈ S and B1 , . . . , Bn , D1 , . . . , Dm ∈ T , then
  \ 
( A1 × B1 ) ∪ · · · ∪ ( An × Bn ) (C1 × D1 ) ∪ · · · ∪ (Cm × Dm )
n [
[ m  
= ( A j × Bj ) ∩ (Ck × Dk )
j =1 k =1
n [
[ m  
= ( A j ∩ Ck ) × ( Bj ∩ Dk ) ,
j =1 k =1
Intersection of two rectangles is a rectangle.
which implies that A is closed under finite intersections.
If A ∈ S and B ∈ T , then
   
( X × Y ) \ ( A × B ) = ( X \ A ) × Y ∪ X × (Y \ B ) .

Hence the complement of each measurable rectangle in S ⊗ T is in A. Thus the


complement of a finite union of measurable rectangles in S ⊗ T is in A (use De
Morgan’s Laws and the result in the previous paragraph that A is closed under finite
intersections). In other words, A is closed under complementation, completing the
proof of (a).
To prove (b), note that if A × B and C × D are measurable rectangles in S ⊗ T ,
then (as can be verified in the figure above)
   
5.14 ( A × B) ∪ (C × D ) = ( A × B) ∪ C × ( D \ B) ∪ (C \ A) × ( B ∩ D ) .

The equation above writes the union of two measurable rectangles in S ⊗ T as the
union of three disjoint measurable rectangles in S ⊗ T .
Now consider any finite union of measurable rectangles in S ⊗ T . If this is not
a disjoint union, then choose any nondisjoint pair of measurable rectangles in the
union and replace those two measurable rectangles with the union of three disjoint
measurable rectangles as in 5.14. Iterate this process until obtaining a disjoint union
of measurable rectangles.
122 Chapter 5 Product Measures

Now we define a monotone class as a collection of sets that is closed under


countable increasing unions and under countable decreasing intersections.

5.15 Definition monotone class

Suppose W is a set and M is a set of subsets of W. Then M is called a monotone


class on W if the following two conditions are satisfied:

[
• If E1 ⊂ E2 ⊂ · · · is an increasing sequence of sets in M, then Ek ∈ M;
k =1

\
• If E1 ⊃ E2 ⊃ · · · is a decreasing sequence of sets in M, then Ek ∈ M.
k =1

Clearly every σ-algebra is a monotone class. However, some monotone classes


are not closed under even finite unions, as shown by the next example.

5.16 Example a monotone class that is not an algebra


Suppose A is the collection of all intervals of R. Then A is closed under countable
increasing unions and countable decreasing intersections. Thus A is a monotone
class on R. However, A is not closed under finite unions, and A is not closed under
complementation. Thus A is neither an algebra nor a σ-algebra on R.

If A is a collection of subsets of some set W, then the intersection of all mono-


tone classes on W that contain A is a monotone class that contains A. Thus this
intersection is the smallest monotone class on W that contains A.
The next result provides a useful tool when the standard technique for showing
that every set in a σ-algebra has a certain property does not work.

5.17 Monotone Class Theorem

Suppose A is an algebra on a set W. Then the smallest σ-algebra containing A


is the smallest monotone class containing A.

Proof Let M denote the smallest monotone class containing A. Because every σ-
algebra is a monotone class, M is contained in the smallest σ-algebra containing A.
To prove the inclusion in the other direction, first suppose A ∈ A. Let

E = { E ∈ M : A ∪ E ∈ M}.

Then A ⊂ E (because the union of two sets in A is in A). A moment’s thought


shows that E is a monotone class. Thus the smallest monotone class that contains A
is contained in E , meaning that M ⊂ E . Hence we have proved that A ∪ E ∈ M
for every E ∈ M.
Now let
D = { D ∈ M : D ∪ E ∈ M for all E ∈ M}.
Section 5A Products of Measure Spaces 123

The previous paragraph shows that A ⊂ D . A moment’s thought again shows that D
is a monotone class. Thus, as in the previous paragraph, we conclude that M ⊂ D .
Hence we have proved that D ∪ E ∈ M for all D, E ∈ M.
The paragraph above shows that the monotone class M is closed under finite
unions. Now if E1 , E2 , . . . ∈ M, then

E1 ∪ E2 ∪ E3 ∪ · · · = E1 ∪ ( E1 ∪ E2 ) ∪ ( E1 ∪ E2 ∪ E3 ) ∪ · · · ,

which is an increasing union of a sequence of sets in M (by the previous paragraph).


We conclude that M is closed under countable unions.
Finally, let
M0 = { E ∈ M : W \ E ∈ M}.
Then A ⊂ M0 (because A is closed under complementation). Once again, you
should verify that M0 is a monotone class. Thus M ⊂ M0 . We conclude that M is
closed under complementation.
The two previous paragraphs show that M is closed under countable unions and
under complementation. Thus M is a σ-algebra that contains A. Hence M contains
the smallest σ-algebra containing A, completing the proof.

Products of Measures
The following definitions will be useful.

5.18 Definition finite measure; σ-finite measure

• A measure µ on a measurable space ( X, S) is called finite if µ( X ) < ∞.


• A measure is called σ-finite if the whole space can be written as the countable
union of sets with finite measure.
• More precisely, a measure µ on a measurable space ( X, S) is called σ-finite
if there exists a sequence X1 , X2 , . . . of sets in S such that

[
X= Xk and µ( Xk ) < ∞ for every k ∈ Z+ .
k =1

5.19 Example finite and σ-finite measures

• Lebesgue measure on the interval [0, 1] is a finite measure.


• Lebesgue measure on R is not a finite measure but is a σ-finite measure.
• Counting measure on R is not a σ-finite measure (because the countable union
of finite sets is a countable set).
124 Chapter 5 Product Measures

The next result will allow us to define the product of two σ-finite measures.

5.20 measure of cross section is a measurable function

Suppose ( X, S , µ) and (Y, T , ν) are σ-finite measure spaces. If E ∈ S ⊗ T ,


then

(a) x 7→ ν([ E] x ) is an S -measurable function on X;


(b) y 7→ µ([ E]y ) is a T -measurable function on Y.

Proof We will prove (a). If E ∈ S ⊗ T , then [ E] x ∈ T for every x ∈ X (by 5.6);


thus the function x 7→ ν([ E] x ) is well defined on X.
We first consider the case where ν is a finite measure. Let

M = { E ∈ S ⊗ T : x 7→ ν([ E] x ) is an S -measurable function on X }.

We need to prove that M = S ⊗ T .


If A ∈ S and B ∈ T , then ν([ A × B] x ) = ν( B)χ A( x ) for every x ∈ X (by
Example 5.5). Thus the function x 7→ ν([ A × B] x ) equals the function ν( B)χ A (as
a function on X), which is an S -measurable function on X. Hence M contains all
the measurable rectangles in S ⊗ T .
Let A denote the set of finite unions of measurable rectangles in S ⊗ T . Suppose
E ∈ A. Then by 5.13(b), E is a union of disjoint measurable rectangles E1 , . . . , En .
Thus

ν([ E] x ) = ν([ E1 ∪ · · · ∪ En ] x )
= ν([ E1 ] x ∪ · · · ∪ [ En ] x )
= ν([ E1 ] x ) + · · · + ν([ En ] x ),

where the last equality holds because ν is a measure and [ E1 ] x , . . . , [ En ] x are disjoint.
The equation above, when combined with the conclusion of the previous paragraph,
shows that x 7→ ν([ E] x ) is a finite sum of S -measurable functions and thus is an
S -measurable function. Hence E ∈ M. We have now shown that A ⊂ M.
Our next goal is to show that M is a monotone class on X × Y. To do this, first
suppose E1 ⊂ E2 ⊂ · · · is an increasing sequence of sets in M. Then
 [∞  ∞
[ 
ν [ Ek ] x = ν ([ Ek ] x )
k =1 k =1

= lim ν([ Ek ] x ),
k→∞

where we have used 2.59. Because the pointwise limit of S -measurable


S
functions

is S -measurable (by 2.48), the equation above shows that x 7→ ν [ ∞k = 1 Ek ] x is
S∞
an S -measurable function. Hence k=1 Ek ∈ M. We have now shown that M is
closed under countable increasing unions.
Section 5A Products of Measure Spaces 125

Now suppose E1 ⊃ E2 ⊃ · · · is a decreasing sequence of sets in M. Then


 \∞  ∞
\ 
ν [ Ek ] x = ν ([ Ek ] x )
k =1 k =1
= lim ν([ Ek ] x ),
k→∞

where we have used 2.60 (this is where we use the assumption that ν is a finite
measure). Because the pointwise limit of S -measurable T∞
functions
 is S -measurable
(by 2.48), the equation T
above shows that x 7 → ν [ k = 1 E k ] x is an S -measurable
function. Hence ∞ k=1 Ek ∈ M. We have now shown that M is closed under
countable decreasing intersections.
We have shown that M is a monotone class that contains the algebra A of all
finite unions of measurable rectangles in S ⊗ T [by 5.13(a), A is indeed an algebra].
The Monotone Class Theorem (5.17) implies that M contains the smallest σ-algebra
containing A. In other words, M contains S ⊗ T . This conclusion completes the
proof of (a) in the case where ν is a finite measure.
Now consider the case where ν is a σ-finite measure. Thus there exists a sequence
S
Y1 , Y2 , . . . of sets in T such that ∞ k=1 Yk = Y and ν (Yk ) < ∞ for each k ∈ Z .
+
Replacing each Yk by Y1 ∪ · · · ∪ Yk , we can assume that Y1 ⊂ Y2 ⊂ · · · . If
E ∈ S ⊗ T , then
ν([ E] x ) = lim ν([ E ∩ ( X × Yk )] x ).
k→∞

The function x 7→ ν([ E ∩ ( X × Yk )] x ) is an S -measurable function on X, as follows


by considering the finite measure obtained by restricting ν to the σ-algebra on Yk
consisting of sets in T that are contained in Yk . The equation above now implies that
x 7→ ν([ E] x ) is an S -measurable function on X, completing the proof of (a).
The proof of (b) is similar.

5.21 Definition integration notation

Suppose ( X, S , µ) is a measure space and g : X → [−∞, ∞] is a function. The


notation Z Z
g( x ) dµ( x ) means g dµ,

where dµ( x ) indicates that variables other than x should be treated as constants.

5.22 Example integrals


If λ is Lebesgue measure on [0, 4], then
Z Z
64
( x2 + y) dλ(y) = 4x2 + 8 and ( x2 + y) dλ( x ) = + 4y.
[0, 4] [0, 4] 3
R R
The intent in the next definition is that X Y f ( x, y) dν(y) dµ( x ) is defined only
when the inner integral and then the outer integral both make sense.
126 Chapter 5 Product Measures

5.23 Definition iterated integrals

Suppose ( X, S , µ) and (Y, T , ν) are measure spaces and f : X × Y → R is a


function. Then
Z Z Z Z 
f ( x, y) dν(y) dµ( x ) means f ( x, y) dν(y) dµ( x ).
X Y X Y
R R
R compute X Y f ( x, y) dν(y) dµ( x ), first (temporarily) fix x ∈
In other words, to
X and compute Y f ( x, y) dν(y) [if this integralRmakes sense]. Then compute the
integral with respect to µ of the function x 7→ Y f ( x, y) dν(y) [if this integral
makes sense].

5.24 Example iterated integrals


If λ is Lebesgue measure on [0, 4], then
Z Z Z
( x2 + y) dλ(y) dλ( x ) = (4x2 + 8) dλ( x )
[0, 4] [0, 4] [0, 4]

352
=
3
and
Z Z Z  64 
( x2 + y) dλ( x ) dλ(y) = + 4y dλ(y)
[0, 4] [0, 4] [0, 4] 3
352
= .
3
The two iterated integrals in this example turned out to both equal 352
3 , even though
they do not look alike in the intermediate step of the evaluation. As we will see in the
next section, this equality of integrals when changing the order of integration is not a
coincidence.

The definition of (µ × ν)( E) given below makes sense because the inner integral
below equals ν([ E] x ), which makes sense by 5.6 (or use 5.9), and then the outer
integral makes sense by 5.20(a).
The restriction in the definition below to σ-finite measures is not bothersome be-
cause the main results we seek are not valid without this hypothesis (see Example 5.30
in the next section).

5.25 Definition product of two measures; µ × ν

Suppose ( X, S , µ) and (Y, T , ν) are σ-finite measure spaces. For E ∈ S ⊗ T ,


define (µ × ν)( E) by
Z Z
(µ × ν)( E) = χ E( x, y) dν(y) dµ( x ).
X Y
Section 5A Products of Measure Spaces 127

5.26 Example measure of a rectangle


Suppose ( X, S , µ) and (Y, T , ν) are σ-finite measure spaces. If A ∈ S and
B ∈ T , then
Z Z
(µ × ν)( A × B) = χ A × B( x, y) dν(y) dµ( x )
X Y
Z
= ν( B)χ A( x ) dµ( x )
X

= µ ( A ) ν ( B ).

Thus product measure of a measurable rectangle is the product of the measures of the
corresponding sets.

For ( X, S , µ) and (Y, T , ν) σ-finite measure spaces, we defined the product µ × ν


to be a function from S ⊗ T to [0, ∞] (see 5.25). Now we show that this function is
a measure.

5.27 product of two measures is a measure

Suppose ( X, S , µ) and (Y, T , ν) are σ-finite measure spaces. Then µ × ν is a


measure on ( X × Y, S ⊗ T ).

Proof Clearly (µ × ν)(∅) = 0.


Suppose E1 , E2 , . . . is a disjoint sequence of sets in S ⊗ T . Then

[  Z  [ ∞ 
(µ × ν) Ek = ν [ Ek ] x dµ( x )
X
k =1 k =1
Z ∞
[ 
= ν ([ Ek ] x ) dµ( x )
X
k =1
Z  ∞ 
=
X k =1
∑ ν([Ek ]x ) dµ( x )

∞ Z
= ∑ ν([ Ek ] x ) dµ( x )
k =1 X

= ∑ (µ × ν)(Ek ),
k =1

where the fourth equality follows from the Monotone Convergence Theorem (3.11;
or see Exercise 10 in Section 3A). The equation above shows that µ × ν satisfies the
countable additivity condition required for a measure.
128 Chapter 5 Product Measures

EXERCISES 5A
1 Suppose ( X, S) and (Y, T ) are measurable spaces. Prove that if A is a
nonempty subset of X and B is a nonempty subset of Y such that A × B ∈
S ⊗ T , then A ∈ S and B ∈ T .
2 Suppose ( X, S) is a measurable space. Prove that if E ∈ S ⊗ S , then

{ x ∈ X : ( x, x ) ∈ E} ∈ S .

3 Let B denote the σ-algebra of Borel subsets of R. Show that there exists a set
E ⊂ R × R such that [ E] a ∈ B and [ E] a ∈ B for every a ∈ R, but E ∈
/ B ⊗ B.

4 Suppose ( X, S) and (Y, T ) are measurable spaces. Prove that if f : X → R is


S -measurable and g : Y → R is T -measurable and h : X × Y → R is defined
by h( x, y) = f ( x ) g(y), then h is (S ⊗ T )-measurable.
5 Verify the assertion in Example 5.11 that the collection of finite unions of
intervals of R is closed under complementation.
6 Verify the assertion in Example 5.12 that the collection of countable unions of
intervals of R is not closed under complementation.
7 Suppose A is a nonempty collection of subsets of a set W. Show that A is an
algebra on W if and only if A is closed under finite intersections and under
complementation.
8 Suppose µ is a measure on a measurable space ( X, S). Prove that the following
are equivalent:
(a) The measure µ is σ-finite.
(b) There exists an increasing sequence X1 ⊂ X2 ⊂ · · · of sets in S such that
S
X= ∞ k=1 Xk and µ ( Xk ) < ∞ for every k ∈ Z .
+

(c) ThereSexists a disjoint sequence X1 , X2 , X3 , . . . of sets in S such that


X= ∞ k=1 Xk and µ ( Xk ) < ∞ for every k ∈ Z .
+

9 Suppose µ and ν are σ-finite measures. Prove that µ × ν is a σ-finite measure.


10 Suppose ( X, S , µ) and (Y, T , ν) are σ-finite measure spaces. Prove that if ω is
a measure on S ⊗ T such that ω ( A × B) = µ( A)ν( B) for all A ∈ S and all
B ∈ T , then ω = µ × ν.
[The exercise above means that µ × ν is the unique measure on S ⊗ T that
behaves as we expect on measurable rectangles.]
Section 5B Iterated Integrals 129

5B Iterated Integrals
Tonelli’s Theorem
Relook at Example 5.24 in the previous section and notice that the value of the
iterated integral was unchanged when we switched the order of integration, even
though switching the order of integration led to different intermediate results. Our
next result states that the order of integration can be switched if the function being
integrated is nonnegative and the measures are σ-finite.

5.28 Tonelli’s Theorem

Suppose ( X, S , µ) and (Y, T , ν) are σ-finite measure spaces. Suppose


f : X × Y → [0, ∞] is S ⊗ T -measurable. Then
Z
(a) x 7→ f ( x, y) dν(y) is an S -measurable function on X,
Y
Z
(b) y 7→ f ( x, y) dµ( x ) is a T -measurable function on Y,
X

and
Z Z Z Z Z
f d( µ × ν ) = f ( x, y) dν(y) dµ( x ) = f ( x, y) dµ( x ) dν(y).
X ×Y X Y Y X

Proof We begin by considering the special case where f = χ E for some E ∈ S ⊗ T .


In this case, Z
χ E( x, y) dν(y) = ν([ E] x ) for every x ∈ X
Y
and Z
χ E( x, y) dµ( x ) = µ([ E]y ) for every y ∈ Y.
X
Thus (a) and (b) hold in this case by 5.20.
First assume that µ and ν are finite measures. Let
n Z Z Z Z o
M = E ∈ S ⊗T : χ E( x, y) dν(y) dµ( x ) = χ E( x, y) dµ( x ) dν(y) .
X Y Y X

If A ∈ S and B ∈ T , then A × B ∈ M because both sides of the equation defining


M equal µ( A)ν( B).
Let A denote the set of finite unions of measurable rectangles in S ⊗ T . Then
5.13(b) implies that every element of A is a disjoint union of measurable rectangles
in S ⊗ T . The previous paragraph now implies A ⊂ M.
The Monotone Convergence Theorem (3.11) implies that M is closed under
countable increasing unions. The Bounded Convergence Theorem (3.26) implies
that M is closed under countable decreasing intersections (this is where we use the
assumption that µ and ν are finite measures).
We have shown that M is a monotone class that contains the algebra A of all
finite unions of measurable rectangles in S ⊗ T [by 5.13(a), A is indeed an algebra].
130 Chapter 5 Product Measures

The Monotone Class Theorem (5.17) implies that M contains the smallest σ-algebra
containing A. In other words, M contains S ⊗ T . Thus
Z Z Z Z
5.29 χ E( x, y) dν(y) dµ( x ) = χ E( x, y) dµ( x ) dν(y)
X Y Y X

for every E ∈ S ⊗ T .
Now relax the assumption that µ and ν are finite measures. Write X as an
increasing union of sets X1 ⊂ X2 ⊂ · · · in S with finite measure, and write Y
as an increasing union of sets Y1 ⊂ Y2 ⊂ · · · in T with finite measure. Suppose
E ∈ S ⊗ T . Applying the finite-measure case to the situation where the measures
and the σ-algebras are restricted to X j and Yk , we can conclude that 5.29 holds
with E replaced by E ∩ ( X j × Yk ) for all j, k ∈ Z+. Fix k ∈ Z+ and use the
Monotone Convergence Theorem (3.11) to conclude that 5.29 holds with E replaced
by E ∩ ( X × Yk ) for all k ∈ Z+. One more use of the Monotone Convergence
Theorem then shows that
Z Z Z Z Z
χ E d( µ × ν ) = χ E( x, y) dν(y) dµ( x ) = χ E( x, y) dµ( x ) dν(y)
X ×Y X Y Y X

for all E ∈ S ⊗ T , where the first equality above comes from the definition of
(µ × ν)( E) (see 5.25).
Now we turn from characteristic functions to the general case of an S ⊗ T -
measurable function f : X × Y → [0, ∞]. Define a sequence f 1 , f 2 , . . . of simple
S ⊗ T -measurable functions from X × Y to [0, ∞) by
 h m m + 1
 m

k
if f ( x, y ) < k and m is the integer with f ( x, y ) ∈ , ,
f k ( x, y) = 2 2k 2k

k if f ( x, y) ≥ k.

Note that

0 ≤ f 1 ( x, y) ≤ f 2 ( x, y) ≤ f 3 ( x, y) ≤ · · · and lim f k ( x, y) = f ( x, y)
k→∞

for all ( x, y) ∈ X × Y.
Each f k is a finite sum of functions of the form cχ E, where c ∈ R and E ∈ S ⊗ T .
Thus the conclusions of this theorem hold for each function f k .
The Monotone Convergence Theorem implies that
Z Z
f ( x, y) dν(y) = lim f k ( x, y) dν(y)
Y k→∞ Y
R
for every x ∈ X. Thus the function x 7→ Y f ( x, y) dν(y) is the pointwise limit on
X of a sequence of S -measurable functions. Hence (a) holds, as does (b) for similar
reasons.
The last line in the statement of this theorem holds for each f k . The Monotone
Convergence Theorem now implies that the last line in the statement of this theorem
holds for f , completing the proof.
Section 5B Iterated Integrals 131

See Exercise 1 in this section for an example (with finite measures) showing that
Tonelli’s Theorem can fail without the hypothesis that the function being integrated
is nonnegative. The next example shows that the hypothesis of σ-finite measures also
cannot be eliminated.

5.30 Example Tonelli’s Theorem can fail without the hypothesis of σ-finite
Suppose B is the σ-algebra of Borel subsets of [0, 1], λ is Lebesgue measure on
([0, 1], B), and µ is counting measure on ([0, 1], B). Let D denote the diagonal of
[0, 1] × [0, 1]; in other words,
D = {( x, x ) : x ∈ [0, 1]}.
Then Z Z Z
χ D( x, y) dµ(y) dλ( x ) = 1 dλ = 1,
[0, 1] [0, 1] [0, 1]
but Z Z Z
χ D( x, y) dλ( x ) dµ(y) = 0 dµ = 0.
[0, 1] [0, 1] [0, 1]

The following useful corollary of Tonelli’s Theorem states that we can switch the
order of summation in a double-sum of nonnegative numbers. Exercise 2 asks you
to find a double-sum of real numbers in which switching the order of summation
changes the value of the double sum.

5.31 double sums of nonnegative numbers

If { x j,k : j, k ∈ Z+} is a doubly indexed collection of nonnegative numbers, then


∞ ∞ ∞ ∞
∑ ∑ x j,k = ∑ ∑ x j,k .
j =1 k =1 k =1 j =1

Proof Apply Tonelli’s Theorem (5.28) to µ × µ, where µ is counting measure


on Z+.

Fubini’s Theorem
Our next goal is Fubini’s Theorem, which
Historically, Fubini’s Theorem
has the same conclusions as Tonelli’s
(proved in 1907) came before
Theorem but has a different hypothesis.
Tonelli’s Theorem (proved in 1909).
Tonelli’s Theorem requires the function
However, presenting Tonelli’s
being integrated to be nonnegative. Fu-
Theorem first, as is done here, seems
bini’s Theorem instead requires the inte-
to lead to simpler proofs and better
gral of the absolute value of the function
understanding. The hard work here
to be finite. When using Fubini’s The-
went into proving Tonelli’s Theorem;
orem to evaluate the integral of f , you
thus our proof of Fubini’s Theorem
will usually first use Tonelli’s Theorem as
consists mainly of bookkeeping
applied to | f | to verify the hypothesis of
details.
Fubini’s Theorem.
132 Chapter 5 Product Measures

As you will see in the proof of Fubini’s Theorem, the function in 5.32(a) is defined
only for almost every x ∈ X and the function in 5.32(b) is defined only for almost
every y ∈ Y. For convenience, you can think of these functions as equaling 0 on the
sets of measure 0 on which they are otherwise undefined.

5.32 Fubini’s Theorem

Suppose ( X, S , µ) and (Y, T , ν) are σ-finite measure


R spaces. Suppose
f : X × Y → [−∞, ∞] is S ⊗ T -measurable and X ×Y | f | d(µ × ν) < ∞.
Then Z
| f ( x, y)| dν(y) < ∞ for almost every x ∈ X
Y
and Z
| f ( x, y)| dµ( x ) < ∞ for almost every y ∈ Y.
X
Furthermore,
Z
(a) x 7→ f ( x, y) dν(y) is an S -measurable function on X,
Y
Z
(b) y 7→ f ( x, y) dµ( x ) is a T -measurable function on Y,
X

and
Z Z Z Z Z
f d( µ × ν ) = f ( x, y) dν(y) dµ( x ) = f ( x, y) dµ( x ) dν(y).
X ×Y X Y Y X

Proof R Tonelli’s Theorem (5.28) applied to the nonnegative function | f | implies that
x 7→ Y | f ( x, y )| dν ( y ) is an S -measurable function on X. Hence
n Z o
x∈X: | f ( x, y)| dν(y) = ∞ ∈ S .
Y

Tonelli’s Theorem applied to | f | also tells us that


Z Z
| f ( x, y)| dν(y) dµ( x ) < ∞
X Y
R
because the iterated integral above equals X ×Y | f | d ( µ × ν ) . The inequality above
implies that n Z o
µ x∈X: | f ( x, y)| dν(y) = ∞ = 0.
Y

Recall that f + and f − are nonnegative S ⊗ T -measurable functions such that


| f | = f + + f − and f = f + − f − (see 3.17). Applying Tonelli’s Theorem to f +
and f − , we see that
Z Z
5.33 x 7→ f + ( x, y) dν(y) and x 7→ f − ( x, y) dν(y)
Y Y
Section 5B Iterated Integrals 133


R functions from X to [0, ∞]. Because f R ≤ | f | and f ≤ | f |, the
are S -measurable +
sets { x ∈ X : Y f + ( x, y) dν(y) = ∞} and { x ∈ X : Y f − ( x, y) dν(y) = ∞}
have µ-measure
R 0. Thus the intersection of these two sets, which is the set of x ∈ X
such that Y f ( x, y) dν(y) is not defined, also has µ-measure 0.
Subtracting the second function in 5.33 from the first function in 5.33, we see that
the function that we define to be 0 for those x ∈ XRwhere we encounter ∞ − ∞ (a
set of µ-measure 0, as noted above) and that equals Y f ( x, y) dν(y) elsewhere is an
S -measurable function on X.
Now
Z Z Z
f d( µ × ν ) = f + d( µ × ν ) − f − d( µ × ν )
X ×Y X ×Y X ×Y
Z Z Z Z
= f + ( x, y) dν(y) dµ( x ) − f − ( x, y) dν(y) dµ( x )
X Y X Y
Z Z 
= f + ( x, y) − f − ( x, y) dν(y) dµ( x )
X Y
Z Z
= f ( x, y) dν(y) dµ( x ),
X Y

where the first line above comes from the definition of the integral of a function that
is not nonnegative R(note that neither of the two terms on the right side of the first line
equals ∞ because X ×Y | f | d(µ × ν) < ∞) and the second line comes from applying
Tonelli’s Theorem to f + and f − .
We have now proved all aspects of Fubini’s Theorem that involve integrating first
over Y. The same procedure provides proofs for the aspects of Fubini’s theorem that
involve integrating first over X.

Area Under Graph

5.34 Definition region under the graph

Suppose X is a set and f : X → [0, ∞] is a function. Then the region under the
graph of f , denoted U f , is defined by

U f = {( x, t) ∈ X × (0, ∞) : 0 < t < f ( x )}.

The figure indicates why we call U f


the region under the graph of f , even
in cases when X is not a subset of R.
Similarly, the informal term area in the
next paragraph should remind you of the
area in the figure, even though we are
really dealing with the measure of U f in
a product space.
134 Chapter 5 Product Measures

The first equality in the result below can be thought of as recovering Riemann’s
conception of the integral as the area under the graph (although now in a much more
general context with arbitrary σ-finite measures). The second equality in the result
below can be thought of as reinforcing Lebesgue’s conception of computing the area
under a curve by integrating in the direction perpendicular to Riemann’s.

5.35 area under the graph of a function equals the integral

Suppose ( X, S , µ) is a σ-finite measure space and f : X → [0, ∞] is an


 of Borel subsets of (0, ∞),
S -measurable function. Let B denote the σ-algebra
and let λ denote Lebesgue measure on (0, ∞), B . Then U f ∈ S ⊗ B and
Z Z
(µ × λ)(U f ) = f dµ = µ({ x ∈ X : t < f ( x )}) dλ(t).
X (0, ∞)

Proof For k ∈ Z+, let


k2[
−1   
Ek = f −1 [ mk , m+ 1 m
k ) × 0, k and Fk = f −1 ([k, ∞]) × (0, k ).
m =0
Then Ek is a finite union of S ⊗ B -measurable rectangles and Fk is an S ⊗ B -
measurable rectangle. Because

[
Uf = ( Ek ∪ Fk ),
k =1
we conclude that U f ∈ S ⊗ B .
Now the definition of the product measure µ × λ implies that
Z Z
(µ × λ)(U f ) = χU ( x, t) dλ(t) dµ( x )
f
X (0, ∞)
Z
= f ( x ) dµ( x ),
X
which completes the proof of the first equality in the conclusion of this theorem.
Tonelli’s Theorem (5.28) tells us that we can interchange the order of integration
in the double integral above, getting
Z Z
(µ × λ)(U f ) = χU ( x, t) dµ( x ) dλ(t)
f
(0, ∞) X
Z
= µ({ x ∈ X : t < f ( x )}) dλ(t),
(0, ∞)
which completes the proof of the second equality in the conclusion of this theorem.

Markov’s inequality (4.1) implies that if f and µ are as in the result above, then
R
f dµ
µ({ x ∈ X : f ( x ) > t}) ≤ X
t
R
for all t > 0. Thus if X f dµ < ∞, then the result above should be considered to be
R
somewhat stronger than Markov’s inequality (because (0, ∞) 1t dλ(t) = ∞).
Section 5B Iterated Integrals 135

EXERCISES 5B
1 (a) Let λ denote Lebesgue measure on [0, 1]. Show that
Z Z
x 2 − y2 π
dλ(y) dλ( x ) =
[0, 1] [0, 1] ( x 2 + y2 )2 4

and Z Z
x 2 − y2 π
dλ( x ) dλ(y) = − .
[0, 1] [0, 1] ( x 2 + y2 )2 4
(b) Explain why (a) violates neither Tonelli’s Theorem nor Fubini’s Theorem.
2 (a) Give an example of a doubly indexed collection { xm,n : m, n ∈ Z+} of
real numbers such that
∞ ∞ ∞ ∞
∑ ∑ xm,n = 0 and ∑ ∑ xm,n = ∞.
m =1 n =1 n =1 m =1

(b) Explain why (a) violates neither Tonelli’s Theorem nor Fubini’s Theorem.
3 Suppose ( X, S) is a measurable space and f : X → [0, ∞] is a function. Let B
denote the σ-algebra of Borel subsets of (0, ∞). Prove that U f ∈ S ⊗ B if and
only if f is an S -measurable function.
4 Suppose ( X, S) is a measurable space and f : X → R is a function. Let
graph( f ) ⊂ X × R denote the graph of f :

graph( f ) = { x, f ( x ) : x ∈ X }.

Let B denote the σ-algebra of Borel subsets of R. Prove that graph( f ) ∈ S ⊗ B


if and only if f is an S -measurable function.
136 Chapter 5 Product Measures

5C Lebesgue Integration on Rn
Throughout this section, assume that m and n are positive integers. Thus, for example,
5.36 should include the hypothesis that m and n are positive integers, but theorems
and definitions become easier to state without explicitly repeating this hypothesis.

Borel Subsets of Rn
We begin with a quick review of notation and key concepts concerning Rn .
Recall that Rn is the set of all n-tuples of real numbers:

Rn = {( x1 , . . . , xn ) : x1 , . . . , xn ∈ R}.

The function k·k∞ from Rn to [0, ∞) is defined by

k( x1 , . . . , xn )k∞ = max{| x1 |, . . . , | xn |}.

For x ∈ Rn and δ > 0, the open cube B( x, δ) with side length 2δ is defined by

B( x, δ) = {y ∈ Rn : ky − x k∞ < δ}.

If n = 1, then an open cube is simply a bounded open interval. If n = 2, then an


open cube might more appropriately be called an open square. However, using the
cube terminology for all dimensions has the advantage of not requiring a different
word for different dimensions.
A subset G of Rn is called open if for every x ∈ G, there exists δ > 0 such that
B( x, δ) ⊂ G. Equivalently, a subset G of Rn is called open if every element of G is
contained in an open cube that is contained in G.
The union of every collection (finite or infinite) of open subsets of Rn is an open
subset of Rn . Also, the intersection of every finite collection of open subsets of Rn is
an open subset of Rn.
A subset of Rn is called closed if its complement in Rn is open. A set A ⊂ Rn is
called bounded if sup{k ak∞ : a ∈ A} < ∞.
We adopt the following common convention:

Rm × Rn is identified with Rm+n .

To understand the necessity of this convention, note that R2 × R 6= R3 because


R2 × R and R3 contain different kinds of objects. Specifically, an element of R2 × R
is an ordered pair, the first of which is an element of R2 and the second  of which is
an element of R; thus an element of R2 × R looks like ( x1 , x2 ), x3 . An element
of R3 is an ordered triple of real numbers that looks like ( x1 , x2 , x3 ). However, we
can identify ( x1 , x2 ), x3 with ( x1 , x2 , x3 ) in the obvious way. Thus we say that
R2 × R “equals” R3 . More generally, we make the natural identification of Rm × Rn
with Rm+n .
To check that you understand the identification discussed above, make sure that
you see why B( x, δ) × B(y, δ) = B ( x, y), δ for all x ∈ Rm , y ∈ Rn , and δ > 0.
We can now prove that the product of two open sets is an open set.
Section 5C Lebesgue Integration on Rn 137

5.36 product of open sets is open

Suppose G1 is an open subset of Rm and G2 is an open subset of Rn . Then


G1 × G2 is an open subset of Rm+n .

Proof Suppose ( x, y) ∈ G1 × G2 . Then there exists an open cube D in Rm centered


at x and an open cube E in Rn centered at y such that D ⊂ G1 and E ⊂ G2 . By
reducing the size of either D or E, we can assume that the cubes D and E have the
same side length. Thus D × E is an open cube in Rm+n centered at ( x, y) that is
contained in G1 × G2 .
We have shown that an arbitrary point in G1 × G2 is the center of an open cube
contained in G1 × G2 . Hence G1 × G2 is an open subset of Rm+n .

When n = 1, the definition below of a Borel subset of R1 agrees with our previous
definition (2.29) of a Borel subset of R.

5.37 Definition Borel set; Bn

• A Borel subset of Rn is an element of the smallest σ-algebra on Rn containing


all open subsets of Rn.
• The σ-algebra of Borel subsets of Rn is denoted by Bn .

Recall that a subset of R is open if and only if it is a countable disjoint union of


open intervals. Part (a) in the result below provides a similar result in Rn, although
we must give up the disjoint aspect.

5.38 open sets are countable unions of open cubes

(a) A subset of Rn is open in Rn if and only if it is a countable union of open


cubes in Rn.
(b) Bn is the smallest σ-algebra on Rn containing all the open cubes in Rn.

Proof We will prove (a), which clearly implies (b).


The proof that a countable union of open cubes is open is left as an exercise for
the reader (actually, arbitrary unions of open cubes are open).
To prove the other direction, suppose G is an open subset of Rn. For each x ∈ G,
there is an open cube centered at x that is contained in G. Thus there is a smaller
cube Cx such that x ∈ Cx ⊂ G and all coordinates of the center of Cx are rational
numbers and the side length of Cx is a rational number. Now
[
G= Cx .
x∈G

However, there are only countably many distinct cubes whose center has all rational
coordinates and whose side length is rational. Thus G is the countable union of open
cubes.
138 Chapter 5 Product Measures

The next result tells us that the collection of Borel sets from various dimensions
fit together nicely.

5.39 product of the Borel subsets of Rm and the Borel subsets of Rn

Bm ⊗ Bn = Bm+n .

Proof Suppose E is an open cube in Rm+n . Thus E is the product of an open cube
inRm and an open cube in Rn . Hence E ∈ Bm ⊗ Bn . Thus the smallest σ-algebra
containing all the open cubes in Rm+n is contained in Bm ⊗ Bn . Now 5.38(b) implies
that Bm+n ⊂ Bm ⊗ Bn .
To prove the set inclusion in the other direction, temporarily fix an open set G in
Rn . Let
E = { A ⊂ R m : A × G ∈ B m + n }.
Then E contains every open subset of Rm (as follows from 5.36). Also, E is closed
under countable unions because

[  ∞
[
Ak × G = ( A k × G ).
k =1 k =1

Furthermore, E is closed under complementation because


   
Rm \ A × G = (Rm × Rn ) \ ( A × G ) ∩ Rm × G .

Thus E is a σ-algebra on Rm that contains all open subsets of Rm , which implies that
Bm ⊂ E . In other words, we have proved that if A ∈ Bm and G is an open subset of
Rn , then A × G ∈ Bm+n .
Now temporarily fix a Borel subset A of Rm . Let
F = { B ⊂ R n : A × B ∈ B m + n }.
The conclusion of the previous paragraph shows that F contains every open subset of
Rn . As in the previous paragraph, we also see that F is a σ-algebra. Hence Bn ⊂ F .
In other words, we have proved that if A ∈ Bm and B ∈ Bn , then A × B ∈ Bm+n .
Thus Bm ⊗ Bn ⊂ Bm+n , completing the proof.

The previous result implies a nice associative property. Specifically, if m, n, and


p are positive integers, then two applications of 5.39 give
(Bm ⊗ Bn ) ⊗ B p = Bm+n ⊗ B p = Bm+n+ p .
Similarly, two more applications of 5.39 give
Bm ⊗ (Bn ⊗ B p ) = Bm ⊗ Bn+ p = Bm+n+ p .
Thus (Bm ⊗ Bn ) ⊗ B p = Bm ⊗ (Bn ⊗ B p ); hence we can dispense with parentheses
when taking products of more than two Borel σ-algebras. More generally, we could
have defined Bm ⊗ Bn ⊗ B p directly as the smallest σ-algebra on Rm+n+ p containing
{ A × B × C : A ∈ Bm , B ∈ Bn , C ∈ B p } and obtained the same σ-algebra (see
Exercise 3 in this section).
Section 5C Lebesgue Integration on Rn 139

Lebesgue Measure on Rn

5.40 Definition Lebesgue measure; λn

Lebesgue measure on Rn is denoted by λn and is defined inductively by

λ n = λ n −1 × λ 1 ,

where λ1 is Lebesgue measure on (R, B1 ).

Because Bn = Bn−1 ⊗ B1 (by 5.39), the measure λn is defined on the Borel


subsets of Rn. Thinking of a typical point in Rn as ( x, y), where x ∈ Rn−1 and
y ∈ R, we can use the definition of the product of two measures (5.25) to write
Z Z
λn ( E) = χ E( x, y) dλ1 (y) dλn−1 ( x )
R n −1 R

for E ∈ Bn . Of course, we could use Tonelli’s Theorem (5.28) to interchange the


order of integration in the equation above.
Because Lebesgue measure is the most commonly used measure, mathematicians
often dispense with explicitly displaying the measure and just use a variable name.
In other words, if no measure is explicitly displayed in an integral and the context
indicates no other measure, then you should assume that the measure involved
is Lebesgue measure in the appropriate dimension. For example, the result of
interchanging the order of integration in the equation above could be written as
Z Z
λn ( E) = χ E( x, y) dx dy
R R n −1

for E ∈ Bn ; here dx means dλn−1 ( x ) and dy means dλ1 (y).


In the equations above giving formulas for λn ( E), the integral over Rn−1 could be
rewritten as an iterated integral over Rn−2 and R, and that process could be repeated
until reaching iterated integrals only over R. Tonelli’s Theorem could then be used
repeatedly to swap the order of pairs of those integrated integrals, leading to iterated
integrals in any order.
Similar comments apply to integrating functions on Rn other than characteristic
functions.R For example, if f : R3 → R is a B3 -measurable function such that either
f ≥ 0 or R3 | f | dλ3 < ∞, then by either Tonelli’s Theorem or Fubini’s Theorem we
have Z Z Z Z
f dλ3 = f ( x1 , x2 , x3 ) dx j dxk dxm ,
R3 R R R
where j, k, m is any permutation of 1, 2, 3.
Although we defined λn to be λn−1 × λ1 , we could have defined λn to be λ j × λk
for any positive integers j, k with j + k = n. This potentially different definition
would have led to the same σ-algebra Bn (by 5.39) and to the same measure λn
[because both potential definitions of λn ( E) can be written as identical iterations of
n integrals with respect to λ1 ].
140 Chapter 5 Product Measures

Volume of Unit Ball in Rn


The proof of the next result provides good experience in working with the Lebesgue
measure λn . Recall that tE = {tx : x ∈ E}.

5.41 measure of a dilation

Suppose t > 0. If E ∈ Bn , then tE ∈ Bn and λn (tE) = tn λn ( E).

Proof Let
E = { E ∈ Bn : tE ∈ Bn }.
Then E contains every open subset of Rn (because if E is open in Rn then tE is open
in Rn ). Also, E is closed under complementation and countable unions because
[ ∞  ∞
[
t(Rn \ E) = Rn \ (tE) and t Ek = (tEk ).
k =1 k =1
Hence E is a σ-algebra on Rn containing the open subsets of Rn. Thus E = Bn . In
other words, tE ∈ Bn for all E ∈ Bn .
To prove λn (tE) = tn λn ( E), first consider the case n = 1. Lebesgue measure on
R is a restriction of outer measure. The outer measure of a set is determined by the
sum of the lengths of countable collections of intervals whose union contains the set.
Multiplying the set by t corresponds to multiplying each such interval by t, which
multiplies the length of each such interval by t. In other words, λ1 (tE) = tλ1 ( E).
Now assume n > 1. We will use induction on n and assume that the desired result
holds for n − 1. If A ∈ Bn−1 and B ∈ B1 , then
 
λn t( A × B) = λn (tA) × (tB)
= λn−1 (tA) · λ1 (tB)
= tn−1 λn−1 ( A) · tλ1 ( B)
5.42 = t n λ n ( A × B ),
giving the desired result for A × B.
For m ∈ Z+, let Cm be the open cube in Rn centered at the origin and with side
length m. Let
Em = { E ∈ Bn : E ⊂ Cm and λn (tE) = tn λn ( E)}.
From 5.42 and using 5.13(b), we see that finite unions of measurable rectangles
contained in Cm are in Em . You should verify that Em is closed under countable
increasing unions (use 2.59) and countable decreasing intersections (use 2.60, whose
finite measure condition holds because we are working inside Cm ). From 5.13 and
the Monotone Class Theorem (5.17), we conclude that Em is the σ-algebra on Cm
consisting of Borel subsets of Cm . Thus λn (tE) = tn λn ( E) for all E ∈ Bn such that
E ⊂ Cm .
Now suppose E ∈ Bn . Then 2.59 implies that

λn (tE) = lim λn t( E ∩ Cm ) = tn lim λn ( E ∩ Cm ) = tn λn ( E),
m→∞ m→∞
as desired.

Section 5C Lebesgue Integration on Rn ≈ 100 2

5.43 Definition open unit ball in Rn

The open unit ball in Rn is denoted by Bn and is defined by

Bn = {( x1 , . . . , xn ) ∈ Rn : x1 2 + · · · + xn 2 < 1}.

The open unit ball Bn is open in Rn (as you should verify) and thus is in the
collection Bn of Borel sets.

5.44 volume of the unit ball in Rn




 π n/2


 (n/2)! if n is even,
λn (Bn ) =

 (n+1)/2 π (n−1)/2
2

 if n is odd.
1·3·5·····n

Proof Because λ1 (B1 ) = 2 and λ2 (B2 ) = π, the claimed formula is correct when
n = 1 and when n = 2.
Now assume that n > 2. We will use induction on n, assuming that the claimed for-
mula is true for smaller values of n. Think of Rn = R2 × Rn−2 and λn = λ2 × λn−2 .
Then
Z Z
5.45 λn (Bn ) = χB ( x, y) dy dx.
R2 R n −2 n

Temporarily fix x = ( x1 , x2 ) ∈ R2 . If x1 2 + x2 2 ≥ 1, then χB ( x, y) = 0 for


n
all y ∈ Rn−2 . If x1 2 + x2 2 < 1 and y ∈ Rn−2 , then χB ( x, y) = 1 if and only if
n

y ∈ (1 − x1 2 − x2 2 )1/2 Bn−2 . Thus the inner integral in 5.45 equals


 
λn−2 (1 − x1 2 − x2 2 )1/2 Bn−2 χB ( x ),
2

which by 5.41 equals


(1 − x1 2 − x2 2 )(n−2)/2 λn−2 (Bn−2 )χB2 ( x ).
Thus 5.45 becomes the equation
Z
λ n ( B n ) = λ n −2 ( B n −2 ) (1 − x1 2 − x2 2 )(n−2)/2 dλ2 ( x1 , x2 ).
B2

To evaluate this integral, switch to the usual polar coordinates that you learned about
in calculus (dλ2 = r dr dθ), getting
Z π Z 1
λ n ( B n ) = λ n −2 ( B n −2 ) (1 − r2 )(n−2)/2 r dr dθ
−π 0

= λ n −2 ( B n −2 ).
n
The last equation and the induction hypothesis give the desired result.
142 Chapter 5 Product Measures

The table here gives the first five val-


ues of λn (Bn ), using 5.44. The last col-
n λn (Bn ) ≈ λn (Bn )
umn of this table gives a decimal approx-
1 2 2.00
imation to λn (Bn ), accurate to two dig-
its after the decimal point. From this ta- 2 π 3.14
ble, you might guess that λn (Bn ) is an 3 4π/3 4.19
increasing function of n, especially be-
4 π 2 /2 4.93
cause the smallest cube containing the
ball Bn has n-dimensional Lebesgue mea- 5 8π 2 /15 5.26
sure 2n . However, Exercise 12 in this
section shows that λn (Bn ) behaves much
differently.

Equality of Mixed Partial Derivatives Via Fubini’s Theorem

5.46 Definition partial derivatives; D1 f and D2 f

Suppose G is an open subset of R2 and f : G → R is a function. For ( x, y) ∈ G,


the partial derivatives ( D1 f )( x, y) and ( D2 f )( x, y) are defined by

f ( x + t, y) − f ( x, y)
( D1 f )( x, y) = lim
t →0 t
and
f ( x, y + t) − f ( x, y)
( D2 f )( x, y) = lim
t →0 t
if these limits exist.

Using the notation for the cross section of a function (see 5.7), we could write the
definitions of D1 and D2 in the following form:

( D1 f )( x, y) = ([ f ]y )0 ( x ) and ( D2 f )( x, y) = ([ f ] x )0 (y).

5.47 Example partial derivatives of x y


Let G = {( x, y) ∈ R2 : x > 0} and define f : G → R by f ( x, y) = x y . Then

( D1 f )( x, y) = yx y−1 and ( D2 f )( x, y) = x y ln x,
as you should verify. Taking partial derivatives of those partial derivatives, we have

D2 ( D1 f ) ( x, y) = x y−1 + yx y−1 ln x

and 
D1 ( D2 f ) ( x, y) = x y−1 + yx y−1 ln x,
as you should also verify. The last two equations show that D1 ( D2 f ) = D2 ( D1 f )
as functions on G.
Section 5C Lebesgue Integration on Rn 143

In the example above, the two mixed partial derivatives turn out to equal each
other, even though the intermediate results look quite different. The next result shows
that the behavior in the example above is typical rather than a coincidence.
Some proofs of the result below do not use Fubini’s Theorem. However, Fubini’s
Theorem leads to the clean proof below.
The integrals that appear in the proof
Although the continuity hypotheses
below make sense because continuous
in the result below can be slightly
real-valued functions on R2 are measur-
weakened, they cannot be
able (because for a continuous function,
eliminated, as shown by Exercise 14
the inverse image of each open set is open)
in this section.
and because continuous real-valued func-
2
tions on closed bounded subsets of R are
bounded.

5.48 equality of mixed partial derivatives

Suppose G is an open subset of R2 and f : G → R is a function such that D1 f ,


D2 f , D1 ( D2 f ), and D2 ( D1 f ) all exist and are continuous functions on G. Then

D1 ( D2 f ) = D2 ( D1 f )

on G.

Proof Fix ( a, b) ∈ G. For δ > 0, let Sδ = [ a, a + δ] × [b, b + δ]. If Sδ ⊂ G, then


Z Z b+δ Z a+δ 
D1 ( D2 f ) dλ2 = D1 ( D2 f ) ( x, y) dx dy
Sδ b a
Z b+δ  
= ( D2 f )( a + δ, y) − ( D2 f )( a, y) dy
b

= f ( a + δ, b + δ) − f ( a + δ, b) − f ( a, b + δ) + f ( a, b),
where the first equality comes from Fubini’s Theorem (5.32) and the second and third
equalities come from the Fundamental
R Theorem of Calculus.
A similar calculation of S D2 ( D1 f ) dλ2 yields the same result. Thus
δ
Z
[ D1 ( D2 f ) − D2 ( D1 f )] dλ2 = 0

 
for all δ such that Sδ ⊂ G. If D1 ( D2 f ) ( a, b) > D2 ( D1 f ) ( a, b), then by
the continuity of D1 ( D2 f ) and D2 ( D1 f ), the integrand in the equation above is
positive on Sδ for δ sufficiently small, which
 contradicts the integral above equaling
0. Similarly, the inequality D1 ( D2 f ) ( a, b) < D2 ( D1 f ) ( a, b) also contradicts
the equation above for small δ. Thus we conclude that
 
D1 ( D2 f ) ( a, b) = D2 ( D1 f ) ( a, b),

as desired.
144 Chapter 5 Product Measures

EXERCISES 5C
1 Show that a set G ⊂ Rn is open in Rn if and only if for each (b1 , . . . , bn ) ∈ G,
there exists r > 0 such that
n q o
( a1 , . . . , an ) ∈ Rn : ( a1 − b1 )2 + · · · + ( an − bn )2 < r ⊂ G.

2 Show that there exists a set E ⊂ R2 (thinking of R2 as equal to R × R) such


that the cross sections [ E] a and [ E] a are open subsets of R for every a ∈ R, but
E∈ / B2 .
3 Suppose ( X, S), (Y, T ), and ( Z, U ) are measurable spaces. We can define
S ⊗ T ⊗ U to be the smallest σ-algebra on X × Y × Z that contains

{ A × B × C : A ∈ S , B ∈ T , C ∈ U }.

Prove that if we make the obvious identifications of the products ( X × Y ) × Z


and X × (Y × Z ) with X × Y × Z, then

S ⊗ T ⊗ U = (S ⊗ T ) ⊗ U = S ⊗ (T ⊗ U ).

4 Show that Lebesgue measure on Rn is translation invariant. More precisely,


show that if E ∈ Bn and a ∈ Rn, then a + E ∈ Bn and λn ( a + E) = λn ( E),
where
a + E = { a + x : x ∈ E }.
5 Suppose f : Rn → R is Bn -measurable and t ∈ R \ {0}. Define f t : Rn → R
by f t ( x ) = f (tx ).
(a) Prove that f t is Bn measurable.
Z
(b) Prove that if f dλn is defined, then
Rn
Z Z
1
f t dλn = f dλn .
Rn |t|n Rn

6 Suppose λ denotes Lebesgue measure on (R, L), where L is the σ-algebra of


Lebesgue measurable subsets of R. Show that there exist subsets E and F of R2
such that
• F ∈ L ⊗ L and (λ × λ)( F ) = 0;
• / L ⊗ L.
E ⊂ F but E ∈
[The measure space (R, L, λ) has the property that every subset of a set with
measure 0 is measurable. This exercise asks you to show that the measure space
(R2 , L ⊗ L, λ × λ) does not have this property.]
7 Suppose M ∈ Z+. Verify that the collection of sets E M that appears in the proof
of 5.41 is a monotone class.
Section 5C Lebesgue Integration on Rn 145

8 Show that the open unit ball in Rn is an open subset of Rn.


9 Suppose G1 is a nonempty subset of Rm and G2 is a nonempty subset of Rn .
Prove that G1 × G2 is an open subset of Rm × Rn if and only if G1 is an open
subset of Rm and G2 is an open subset of Rn .
[One direction of this result was already proved (see 5.36); both directions are
stated here to make the result look prettier and to be comparable to the next
exercise, where neither direction has been proved.]
10 Suppose F1 is a nonempty subset of Rm and F2 is a nonempty subset of Rn .
Prove that F1 × F2 is a closed subset of Rm × Rn if and only if F1 is a closed
subset of Rm and F2 is a closed subset of Rn .
11 Suppose E is a subset of Rm × Rn and

A = { x ∈ Rm : ( x, y) ∈ E for some y ∈ Rn }.

(a) Prove that if E is an open subset of Rm × Rn , then A is an open subset


of Rm .
(b) Prove or give a counterexample: If E is a closed subset of Rm × Rn , then
A is a closed subset of Rm .

12 (a) Prove that limn→∞ λn (Bn ) = 0.


(b) Find the value of n that maximizes λn (Bn ).
13 For readers familiar with the gamma function Γ: Prove that

π n/2
λn (Bn ) =
Γ( n2 + 1)

for every positive integer n.


14 Define f : R2 → R by
 2 2
 xy( x − y ) if ( x, y) 6= (0, 0),
f ( x, y) = x 2 + y2

0 if ( x, y) = (0, 0).

(a) Prove that D1 ( D2 f ) and D2 ( D1 f ) exist everywhere on R2 .


 
(b) Show that D1 ( D2 f ) (0, 0) 6= D2 ( D1 f ) (0, 0).
(c) Explain why (b) does not violate 5.48.
Chapter 6
Banach Spaces

We begin this chapter with a quick review of the essentials of metric spaces. Then
we extend our results on measurable functions and integration to complex-valued
functions. After that, we rapidly review the framework of vector spaces, which
allows us to consider natural collections of measurable functions that are closed under
addition and scalar multiplication.
Normed vector spaces and Banach spaces, which are introduced in the third section
of this chapter, play a hugely important role in modern analysis. Most interest focuses
on linear maps on these vector spaces. Key results about linear maps that we develop
in this chapter include the Hahn–Banach Theorem, the Open Mapping Theorem, the
Closed Graph Theorem, and the Principle of Uniform Boundedness.

Market square in Lwów, a city that has been in several countries because of changing
international boundaries. Before World War I, Lwów was in Austria–Hungary.
During the period between World War I and World War II, Lwów was in Poland.
During this time, mathematicians in Lwów, particularly Stefan Banach (1892–1945)
and his colleagues, developed the basic results of modern functional analysis. After
World War II, Lwów was in the USSR. Now Lwów is in Ukraine and is called Lviv.
CC-BY-SA Petar Milošević

146
Section 6A Metric Spaces 147

6A Metric Spaces
Open Sets, Closed Sets, and Continuity
Much of analysis takes place in the context of a metric space, which is a set with a
notion of distance that satisfies certain properties. The properties we would like a
distance function to have are captured in the next definition, where you should think
of d( f , g) as measuring the distance between f and g.
Specifically, we would like the distance between two elements of our metric space
to be a nonnegative number that is 0 if and only if the two elements are the same. We
would like the distance between two elements not to depend on the order in which
we list them. Finally, we would like a triangle inequality (the last bullet point below),
which states that the distance between two elements is less than or equal to the sum
of the distances obtained when we insert an intermediate element.
Now we are ready for the formal definition.

6.1 Definition metric space

A metric on a nonempty set V is a function d : V × V → [0, ∞) such that


• d( f , f ) = 0 for all f ∈ V;
• if f , g ∈ V and d( f , g) = 0, then f = g;
• d( f , g) = d( g, f ) for all f , g ∈ V;
• d( f , h) ≤ d( f , g) + d( g, h) for all f , g, h ∈ V.
A metric space is a pair (V, d), where V is a nonempty set and d is a metric on V.

6.2 Example metric spaces

• Suppose V is a nonempty set. Define d on V × V by setting d( f , g) to be 1 if


f 6= g and to be 0 if f = g. Then d is a metric on V.
• Define d on R × R by d( x, y) = | x − y|. Then d is a metric on R.
• For n ∈ Z+, define d on Rn × Rn by

d ( x1 , . . . , xn ), (y1 , . . . , yn ) = max{| x1 − y1 |, . . . , | xn − yn |}.
Then d is a metric on Rn .
• Define d on C ([0, 1]) × C ([0, 1]) by d( f , g) = sup{| f (t) − g(t)| : t ∈ [0, 1]};
here C ([0, 1]) is the set of continuous real-valued functions on [0, 1]. Then d is
a metric on C ([0, 1]).

• Define d on `1 × `1 by d ( a1 , a2 , . . .), (b1 , b2 , . . .) = ∑∞
k=1 | ak − bk |; here `
1

is the set of sequences ( a1 , a2 , . . .) of real numbers such that ∑k=1 | ak | < ∞.
Then d is a metric on `1 .
148 Chapter 6 Banach Spaces

The material in this section is proba-


This book often uses symbols such as
bly review for most readers of this book.
f , g, h as generic elements of a
Thus more details than usual are left to the
generic metric space because many
reader to verify. Verifying those details
of the important metric spaces in
and doing the exercises is the best way
analysis are sets of functions; for
to solidify your understanding of these
example, see the fourth bullet point
concepts. You should be able to transfer
of Example 6.2.
familiar definitions and proofs from the
n
context of R or R to the context of a metric space.
We will need to use a metric space’s topological features, which we introduce
now.

6.3 Definition open ball; B( f , r )

Suppose (V, d) is a metric space, f ∈ V, and r > 0.


• The open ball centered at f with radius r is denoted B( f , r ) and is defined
by
B ( f , r ) = { g ∈ V : d ( f , g ) < r }.

• The closed ball centered at f with radius r is denoted B( f , r ) and is defined


by
B ( f , r ) = { g ∈ V : d ( f , g ) ≤ r }.

Abusing terminology, many books (including this one) include phrases such as
suppose V is a metric space without mentioning the metric d. When that happens,
you should assume that a metric d lurks nearby, even if it is not explicitly named.
Our next definition declares a subset of a metric space to be open if every element
in the subset is the center of an open ball that is contained in the set.

6.4 Definition open set

A subset G of a metric space V is called open if for every f ∈ G, there exists


r > 0 such that B( f , r ) ⊂ G.

6.5 open balls are open

Suppose V is a metric space, f ∈ V, and r > 0. Then B( f , r ) is an open subset


of V.

Proof Suppose g ∈ B( f , r ). We need to show that an open ball  centered at g is


contained in B( f , r ). To do this, note that if h ∈ B g, r − d( f , g) , then

d( f , h) ≤ d( f , g) + d( g, h) < d( f , g) + r − d( f , g) = r,

which implies that h ∈ B( f , r ). Thus B g, r − d( f , g) ⊂ B( f , r ), which implies
that B( f , r ) is open.
Section 6A Metric Spaces 149

Closed sets are defined in terms of open sets.

6.6 Definition closed subset

A subset of a metric space V is called closed if its complement in V is open.

For example, each closed ball B( f , r ) in a metric space is closed, as you are asked
to prove in Exercise 3.
Now we define the closure of a subset of a metric space.

6.7 Definition closure

Suppose V is a metric space and E ⊂ V. The closure of E, denoted E, is defined


by
E = { g ∈ V : B( g, ε) ∩ E 6= ∅ for every ε > 0}.

Limits in a metric space are defined by reducing to the context of real numbers,
where limits have already been defined.

6.8 Definition limit in V

Suppose (V, d) is a metric space, f 1 , f 2 , . . . is a sequence in V, and f ∈ V. Then

lim f k = f means lim d( f k , f ) = 0.


k→∞ k→∞

In other words, a sequence f 1 , f 2 , . . . in V converges to f ∈ V if for every ε > 0,


there exists n ∈ Z+ such that
d( f k , f ) < ε for all integers k ≥ n.
The next result states that the closure of a set is the collection of all limits of
elements of the set. Also, a set is closed if and only if it equals its closure. The proof
of the next result is left as an exercise that provides good practice in using these
concepts.

6.9 closure

Suppose V is a metric space and E ⊂ V. Then

(a) E = { g ∈ V : there exist f 1 , f 2 , . . . in E such that lim f k = g};


k→∞
(b) E is the intersection of all closed subsets of V that contain E;

(c) E is a closed subset of V;


(d) E is closed if and only if E = E;
(e) E is closed if and only if E contains the limit of every convergent sequence
of elements of E.
150 Chapter 6 Banach Spaces

The definition of continuity that follows uses the same pattern as the definition for
a function from a subset of R to R.

6.10 Definition continuity

Suppose (V, dV ) and (W, dW ) are metric spaces and T : V → W is a function.


• For f ∈ V, the function T is called continuous at f if for every ε > 0, there
exists δ > 0 such that

dW T ( f ) , T ( g ) < ε

for all g ∈ V with dV ( f , g) < δ.


• The function T is called continuous if T is continuous at f for every f ∈ V.

The next result gives equivalent conditions for continuity. Recall that T −1 ( E) is
called the inverse image of E and is defined to be { f ∈ V : T ( f ) ∈ E}. Thus the
equivalence of the (a) and (c) below could be restated as saying that a function is
continuous if and only if the inverse image of every open set is open. The equivalence
of the (a) and (d) below could be restated as saying that a function is continuous if
and only if the inverse image of every closed set is closed.

6.11 equivalent conditions for continuity

Suppose V and W are metric spaces and T : V → W is a function. Then the


following are equivalent:

(a) T is continuous.
(b) lim f k = f in V implies lim T ( f k ) = T ( f ) in W.
k→∞ k→∞
(c) T −1 ( G ) is an open subset of V for every open set G ⊂ W.
(d) T −1 ( F ) is a closed subset of V for every closed set F ⊂ W.

Proof We first prove that (b) implies (d). Suppose (b) holds. Suppose F is a closed
subset of W. We need to prove that T −1 ( F ) is closed. To do this, suppose f 1 , f 2 , . . .
is a sequence in T −1 ( F ) and limk→∞ f k = f for some f ∈ V. Because (b) holds, we
know that limk→∞ T ( f k ) = T ( f ). Because f k ∈ T −1 ( F ) for each k ∈ Z+, we know
that T ( f k ) ∈ F for each k ∈ Z+. Because F is closed, this implies that T ( f ) ∈ F.
Thus f ∈ T −1 ( F ), which implies that T −1 ( F ) is closed [by 6.9(e)], completing the
proof that (b) implies (d).
The proof that (c) and (d) are equivalent follows from the equation
T − 1 (W \ E ) = V \ T − 1 ( E )
for every E ⊂ W and the fact that a set is open if and only if its complement (in the
appropriate metric space) is closed.
The proof of the remaining parts of this result are left as an exercise that should
help strengthen your understanding of these concepts.
Section 6A Metric Spaces 151

Cauchy Sequences and Completeness


The next definition is useful for showing (in some metric spaces) that a sequence has
a limit, even when we do not have a good candidate for that limit.

6.12 Definition Cauchy sequence

A sequence f 1 , f 2 , . . . in a metric space (V, d) is called a Cauchy sequence if for


every ε > 0, there exists n ∈ Z+ such that d( f j , f k ) < ε for all integers j ≥ n
and k ≥ n.

6.13 every convergent sequence is a Cauchy sequence

Every convergent sequence in a metric space is a Cauchy sequence.

Proof Suppose limk→∞ f k = f in a metric space (V, d). Suppose ε > 0. Then
there exists n ∈ Z+ such that d( f k , f ) < 2ε for all k ≥ n. If j, k ∈ Z+ are such that
j ≥ n and k ≥ n, then
ε ε
d( f j , f k ) ≤ d( f j , f ) + d( f , f k ) < 2 + 2 = ε.
Thus f 1 , f 2 , . . . is a Cauchy sequence, completing the proof.

Metric spaces that satisfy the converse of the result above have a special name.

6.14 Definition complete metric space

A metric space V is called complete if every Cauchy sequence in V converges to


some element of V.

6.15 Example
• All five of the metric spaces in Example 6.2 are complete, as you should verify.
• The metric space Q, with metric defined by d( x, y) = | x − y|, is not complete.
To see this, for k ∈ Z+ let
1 1 1
xk = + 2! + · · · + k! .
101! 10 10
If j < k, then
1 1 2
| xk − x j | = + · · · + k! < ( j+1)! .
10( j+1)! 10 10
Thus x1 , x2 , . . . is a Cauchy sequence in Q. However, x1 , x2 , . . . does not con-
verge to an element of Q because the limit of this sequence would have a decimal
expansion 0.110001000000000000000001 . . . that is neither a terminating deci-
mal nor a repeating decimal. Thus Q is not a complete metric space.
152 Chapter 6 Banach Spaces

Entrance to the École Polytechnique (Paris), where Augustin-Louis Cauchy


(1789–1857) was a student and a faculty member. Cauchy wrote almost 800
mathematics papers and the highly influential textbook Cours d’Analyse (published
in 1821), which greatly influenced the development of analysis.
CC-BY-SA NonOmnisMoriar

Every nonempty subset of a metric space is a metric space. Specifically, suppose


(V, d) is a metric space and U is a nonempty subset of V. Then restricting d to
U × U gives a metric on U. Unless stated otherwise, you should assume that the
metric on a subset is this restricted metric that the subset inherits from the bigger set.
Combining the two bullet points in the result below shows that a subset of a
complete metric space is complete if and only if it is closed.

6.16 connection between complete and closed

(a) A complete subset of a metric space is closed.


(b) A closed subset of a complete metric space is complete.

Proof We begin with a proof of (a). Suppose U is a complete subset of a metric


space V. Suppose f 1 , f 2 , . . . is a sequence in U that converges to some g ∈ V.
Then f 1 , f 2 , . . . is a Cauchy sequence in U (by 6.13). Hence by the completeness
of U, the sequence f 1 , f 2 , . . . converges to some element of U, which must be g
(see Exercise 7). Hence g ∈ U. Now 6.9(e) implies that U is a closed subset of V,
completing the proof of (a).
To prove (b), suppose U is a closed subset of a complete metric space V. To show
that U is complete, suppose f 1 , f 2 , . . . is a Cauchy sequence in U. Then f 1 , f 2 , . . . is
also a Cauchy sequence in V. By the completeness of V, this sequence converges to
some f ∈ V. Because U is closed, this implies that f ∈ U (see 6.9). Thus the Cauchy
sequence f 1 , f 2 , . . . converges to an element of U, showing that U is complete. Hence
(b) has been proved.
Section 6A Metric Spaces 153

EXERCISES 6A
1 Verify that each of the claimed metrics in Example 6.2 is indeed a metric.
2 Prove that every finite subset of a metric space is closed.
3 Prove that every closed ball in a metric space is closed.
4 Suppose V is a metric space.
(a) Prove that the union of each collection of open subsets of V is an open
subset of V.
(b) Prove that the intersection of each finite collection of open subsets of V is
an open subset of V.

5 Suppose V is a metric space.


(a) Prove that the intersection of each collection of closed subsets of V is a
closed subset of V.
(b) Prove that the union of each finite collection of closed subsets of V is a
closed subset of V.

6 (a) Prove that if V is a metric space, f ∈ V, and r > 0, then B( f , r ) ⊂ B( f , r ).


(b) Give an example of a metric space V, f ∈ V, and r > 0 such that
B ( f , r ) 6 = B ( f , r ).
7 Show that a sequence in a metric space has at most one limit.
8 Prove 6.9.
9 Prove that each open subset of a metric space V is the union of some sequence
of closed subsets of V.
10 Prove or give a counterexample: If V is a metric space and U, W are subsets
of V, then U ∪ W = U ∪ W.
11 Prove or give a counterexample: If V is a metric space and U, W are subsets
of V, then U ∩ W = U ∩ W.
12 Suppose (U, dU ), (V, dV ), and (W, dW ) are metric spaces. Suppose also that
T : U → V and S : V → W are continuous functions.
(a) Using the definition of continuity, show that S ◦ T : U → W is continuous.
(b) Using the equivalence of 6.11(a) and 6.11(b), show that S ◦ T : U → W is
continuous.
(c) Using the equivalence of 6.11(a) and 6.11(c), show that S ◦ T : U → W is
continuous.

13 Prove the parts of 6.11 that were not proved in the text.
154 Chapter 6 Banach Spaces

14 Suppose a Cauchy sequence in a metric space has a convergent subsequence.


Prove that the Cauchy sequence converges.
15 Verify that all five of the metric spaces in Example 6.2 are complete metric
spaces.

16 Suppose (U, d) is a metric space. Let W denote the set of all Cauchy sequences
of elements of U.
(a) For ( f 1 , f 2 , . . .) and ( g1 , g2 , . . .) in W, define ( f 1 , f 2 , . . .) ≡ ( g1 , g2 , . . .)
to mean that
lim d( f k , gk ) = 0.
k→∞
Show that ≡ is an equivalence relation on W.
(b) Let V denote the set of equivalence classes of elements of W under the
equivalence relation above. For ( f 1 , f 2 , . . .) ∈ W, let ( f 1 , f 2 , . . .)ˆ denote
the equivalence class of ( f 1 , f 2 , . . .). Define dV : V × V → [0, ∞) by

dV ( f 1 , f 2 , . . .)ˆ, ( g1 , g2 , . . .)ˆ = lim d( f k , gk ).
k→∞

Show that this definition of dV makes sense and that dV is a metric on V.


(c) Show that (V, dV ) is a complete metric space.
(d) Show that the map from U to V that takes f ∈ U to ( f , f , f , . . .)ˆ preserves
distances, meaning that

d( f , g) = dV ( f , f , f , . . .)ˆ, ( g, g, g, . . .)ˆ

for all f , g ∈ U.
(e) Explain why (d) shows that every metric space is a subset of some complete
metric space.
Section 6B Vector Spaces 155

6B Vector Spaces
Integration of Complex-Valued Functions
Complex numbers were invented so that we can take square roots of negative numbers.
The idea is to assume we have a square root of −1, denoted i, that obeys the usual
rules of arithmetic. Here are the formal definitions:

6.17 Definition complex numbers; C

• A complex number is an ordered pair ( a, b), where a, b ∈ R, but we write


this as a + bi.
• The set of all complex numbers is denoted by C:

C = { a + bi : a, b ∈ R}.

• Addition and multiplication on C are defined by

( a + bi ) + (c + di ) = ( a + c) + (b + d)i,
( a + bi )(c + di ) = ( ac − bd) + ( ad + bc)i;

here a, b, c, d ∈ R.

If a ∈ R, then we identify a + 0i
with a. Thus we think of R as a subset of √ symbol i was first used to denote
The
−1 by Leonhard Euler
C. We also usually write 0 + bi as bi, and
(1707–1783) in 1777.
we usually write 0 + 1i as i. You should
verify that i2 = −1.
With the definitions as above, C satisfies the usual rules of arithmetic. Specifically,
with addition and multiplication defined as above, C is a field, as you should verify.
Thus subtraction and division of complex numbers are defined as in any field.
The field C cannot be made into an or-
Much of this section may be review
dered field. However, the useful concept
for many readers.
of an absolute value can still be defined
on C.

6.18 Definition Re z; Im z; absolute value; limits

Suppose z = a + bi, where a and b are real numbers.


• The real part of z, denoted Re z, is defined by Re z = a.
• The imaginary part of z, denoted Im z, is defined by Im z = b.

• The absolute value of z, denoted |z|, is defined by |z| = a2 + b2 .
• If z1 , z2 , . . . ∈ C and L ∈ C, then lim zk = L means lim |zk − L| = 0.
k→∞ k→∞
156 Chapter 6 Banach Spaces

For b a real number, the usual definition of |b| as a real number is consistent with
the new definition just given of |b| with b thought of as a complex number. Note that
if z1 , z2 , . . . is a sequence of complex numbers and L ∈ C, then
lim zk = L ⇐⇒ lim Re zk = Re L and lim Im zk = Im L.
k→∞ k→∞ k→∞
We will reduce questions concerning measurability and integration of a complex-
valued function to the corresponding questions about the real and imaginary parts of
the function. We begin this process with the following definition.

6.19 Definition measurable complex-valued function

Suppose ( X, S) is a measurable space. A function f : X → C is called


S -measurable if Re f and Im f are both S -measurable functions.

See Exercise 5 in this section for two natural conditions that are equivalent to
measurability for complex-valued functions.
We will make frequent use of the following result. See Exercise 6 in this section
for algebraic combinations of complex-valued measurable functions.

6.20 | f | p is measurable if f is measurable

Suppose ( X, S) is a measurable space, f : X → C is an S -measurable function,


and 0 < p < ∞. Then | f | p is an S -measurable function.

Proof The functions (Re f )2 and (Im f )2 are S -measurable because the square
of an S -measurable function is measurable (by Example 2.45). Thus the function
(Re f )2 + (Im f )2 is S -measurable (because the sum of two S -measurable functions
 p/2
is S -measurable by 2.46). Now (Re f )2 + (Im f )2 is S -measurable because it
is the composition of a continuous function on [0, ∞) and an S -measurable function
(see 2.44 and 2.41). In other words, | f | p is an S -measurable function.

Now we define integration of a complex-valued function by separating the function


into its real and imaginary parts.

6.21 Definition integral of a complex-valued function

Suppose ( X, SR , µ) is a measure space and f : X → C is an S -measurable


1
R with | f | dµ < ∞ [the collection of such functions is denoted L (µ)].
function
Then f dµ is definedZ
by Z Z
f dµ = (Re f ) dµ + i (Im f ) dµ.

The integral of a complex-valued measurable function is defined above only when


the absolute value of the function has a finite integral. In contrast, the integral of
every nonnegative measurable R function is definedR (althoughRthe value may be ∞),
andR if f is real valued
R then f dµ is defined to be f + dµ − f − dµ if at least one
of f + dµ and f − dµ is finite.
Section 6B Vector Spaces 157

R You can easily Rshow that if f , g : X → C are S -measurable functions such that
| f | dµ < ∞ and | g| dµ < ∞, then
Z Z Z
( f + g) dµ = f dµ + g dµ.

Similarly, the definition of complex multiplication leads to the conclusion that


Z Z
α f dµ = α f dµ

for all α ∈ C (see Exercise 8).


The inequality in the result below concerning integration of complex-valued
functions does not follow immediately from the corresponding result for real-valued
functions. However, the small trick used in the proof below does give a reasonably
simple proof.

6.22 bound on the absolute value of an integral

Suppose ( X, S , µ)R is a measure space and f : X → C is an S -measurable


function such that | f | dµ < ∞. Then
Z Z

f dµ ≤ | f | dµ.

R R
Proof The result clearly holds if f dµ = 0. Thus assume that f dµ 6= 0.
Let R
| f dµ|
α= R .
f dµ
Then
Z Z Z

f dµ = α f dµ = α f dµ
Z Z
= Re(α f ) dµ + i Im(α f ) dµ
Z
= Re(α f ) dµ
Z
≤ |α f | dµ
Z
= | f | dµ,

where
R the second equality holds by Exercise 8, the fourth equality holds because
| f dµ| ∈ R, the inequality on the fourth line holds because Re z ≤ |z| for every
complex number z, and the equality in the last line holds because |α| = 1.

Because of the result above, the Bounded Convergence Theorem (3.26) and the
Dominated Convergence Theorem (3.31) hold if the functions f 1 , f 2 , . . . and f in the
statements of those theorems are allowed to be complex valued.
158 Chapter 6 Banach Spaces

We now define the complex conjugate of a complex number.

6.23 Definition complex conjugate; z

Suppose z ∈ C. The complex conjugate of z ∈ C, denoted z (pronounced z-bar),


is defined by
z = Re z − (Im z)i.

For example, if z = 5 + 7i then z = 5 − 7i. Note that a complex number z is a


real number if and only if z = z.
The next result gives basic properties of the complex conjugate.

6.24 properties of complex conjugates

Suppose w, z ∈ C. Then
• product of z and z
z z = | z |2 ;
• sum and difference of z and z
z + z = 2 Re z and z − z = 2(Im z)i;
• additivity and multiplicativity of complex conjugate
w + z = w + z and wz = w z;
• complex conjugate of complex conjugate
z = z;
• absolute value of complex conjugate
| z | = | z |;
• integral of complex conjugate of a function
Z Z
f dµ = f dµ for every measure µ and every f ∈ L1 (µ).

Proof The first item holds because


zz = (Re z + i Im z)(Re z − i Im z) = (Re z)2 + (Im z)2 = |z|2 .
To prove the last item, suppose µ is a measure and f ∈ L1 (µ). Then
Z Z Z Z
f dµ = (Re f − i Im f ) dµ = Re f dµ − i Im f dµ
Z Z
= Re f dµ + i Im f dµ
Z
= f dµ,

as desired.
The straightforward proofs of the remaining items are left to the reader.
Section 6B Vector Spaces 159

Vector Spaces and Subspaces


The structure and language of vector spaces will help us focus on certain features of
collections of measurable functions. So that we can conveniently make definitions
and prove theorems that apply to both real and complex numbers, we adopt the
following notation.

6.25 Definition F

From now on, F stands for either R or C.

In the definitions that follow, we use f and g to denote elements of V because in


the crucial examples the elements of V are functions from a set X to F.

6.26 Definition addition; scalar multiplication

• An addition on a set V is a function that assigns an element f + g ∈ V to


each pair of elements f , g ∈ V.
• A scalar multiplication on a set V is a function that assigns an element
α f ∈ V to each α ∈ F and each f ∈ V.

Now we are ready to give the formal definition of a vector space.

6.27 Definition vector space

A vector space (over F) is a set V along with an addition on V and a scalar


multiplication on V such that the following properties hold:

commutativity
f + g = g + f for all f , g ∈ V;
associativity
( f + g) + h = f + ( g + h) and (αβ) f = α( β f ) for all f , g, h ∈ V and α, β ∈ F;
additive identity
there exists an element 0 ∈ V such that f + 0 = f for all f ∈ V;
additive inverse
for every f ∈ V, there exists g ∈ V such that f + g = 0;
multiplicative identity
1 f = f for all f ∈ V;
distributive properties
α( f + g) = α f + αg and (α + β) f = α f + β f for all α, β ∈ F and f , g ∈ V.

Most vector spaces that you will encounter are subsets of the vector space F X
presented in the next example.
160 Chapter 6 Banach Spaces

6.28 Example the vector space F X


Suppose X is a nonempty set. Let F X denote the set of functions from X to F.
Addition and scalar multiplication on F X are defined as expected: for f , g ∈ F X and
α ∈ F, define

( f + g)( x ) = f ( x ) + g( x ) and (α f )( x ) = α f ( x )

for x ∈ X. Then, as you should verify, F X is a vector space; the additive identity in
this vector space is the function 0 ∈ F X defined by 0( x ) = 0 for all x ∈ X.

Fn ; FZ
+
6.29 Example
Special case of the previous example: if n ∈ Z+ and X = {1, . . . , n}, then F X is
the familiar space Rn or Cn , depending upon whether F = R or F = C.
+
Another special case: FZ is the vector space of all sequences of real numbers or
complex numbers, again depending upon whether F = R or F = C.

By considering subspaces, we can greatly expand our examples of vector spaces.

6.30 Definition subspace

A subset U of V is called a subspace of V if U is also a vector space (using the


same addition and scalar multiplication as on V).

The next result gives the easiest way to check whether a subset of a vector space
is a subspace.

6.31 conditions for a subspace

A subset U of V is a subspace of V if and only if U satisfies the following three


conditions:
• additive identity
0 ∈ U;
• closed under addition
f , g ∈ U implies f + g ∈ U;
• closed under scalar multiplication
α ∈ F and f ∈ U implies α f ∈ U.

Proof If U is a subspace of V, then U satisfies the three conditions above by the


definition of vector space.
Conversely, suppose U satisfies the three conditions above. The first condition
above ensures that the additive identity of V is in U.
The second condition above ensures that addition makes sense on U. The third
condition ensures that scalar multiplication makes sense on U.
Section 6B Vector Spaces 161

If f ∈ V, then 0 f = (0 + 0) f = 0 f + 0 f . Adding the additive inverse of 0 f to


both sides of this equation shows that 0 f = 0. Now if f ∈ U, then(−1) f is also in
U by the third condition above. Because f + (−1) f = 1 + (−1) f = 0 f = 0, we
see that (−1) f is an additive inverse of f . Hence every element of U has an additive
inverse in U.
The other parts of the definition of a vector space, such as associativity and
commutativity, are automatically satisfied for U because they hold on the larger
space V. Thus U is a vector space and hence is a subspace of V.

The three conditions in 6.31 usually enable us to determine quickly whether a


given subset of V is a subspace of V, as illustrated below. All the examples below
except for the first bullet point involve concepts from measure theory.

6.32 Example subspaces of F X


• The set C ([0, 1]) of continuous real-valued functions on [0, 1] is a vector space
over R because the sum of two continuous functions is continuous and a constant
multiple of a continuous functions is continuous. In other words, C ([0, 1]) is a
subspace of R[0,1] .
• Suppose ( X, S) is a measurable space. Then the set of S -measurable functions
from X to F is a subspace of F X because the sum of two S -measurable functions
is S -measurable and a constant multiple of an S -measurable function is S -
measurable.
• Suppose ( X, S , µ) is a measure space. Then the set Z (µ) of S -measurable
functions f from X to F such that f = 0 almost everywhere [meaning that
µ { x ∈ X : f ( x ) 6= 0} = 0] is a vector space over F because the union of
two sets with µ-measure 0 is a set with µ-measure 0 [which implies that Z (µ)
is closed under addition]. Note that Z (µ) is a subspace of F X .
• Suppose ( X, S) is a measurable space. Then the set of bounded measurable
functions from X to F is a subspace of F X because the sum of two bounded
S -measurable functions is a bounded S -measurable function and a constant mul-
tiple of a bounded S -measurable function is a bounded S -measurable function.
• Suppose ( X, S , µ) is a measure
R space. Then the set of S -measurable functions
f from X to F such that f dµ = 0 is a subspace of F X because of standard
properties of integration.
• Suppose ( X, S , µ) is a measureRspace. Then the set L1 (µ) of S -measurable
functions from X to F such that | f | dµ < ∞ is a subspace of F X [we are now
redefining L1 (µ) to allow for the possibility that F = R or F R= C]. The set
1 and scalarRmultiplication because | f + g| dµ ≤
RL (µ) is closed
R under addition
R
| f | dµ + | g| dµ and |α f | dµ = |α| | f | dµ.
• The set `1 of all sequences ( a1 , a2 , . . .) of elements of F such that ∑∞
k =1 | a k | < ∞
Z + 1
is a subspace of F . Note that ` is a special case of the example in the previous
bullet point (take µ to be counting measure on Z+ ).
162 Chapter 6 Banach Spaces

EXERCISES 6B
1 Show that if a, b ∈ R with a + bi 6= 0, then
1 a b
= 2 2
− 2 i.
a + bi a +b a + b2
2 Suppose z ∈ C. Prove that

max{|Re z|, |Im z|} ≤ |z| ≤ 2 max{|Re z|, |Im z|}.
|Re z| + |Im z|
3 Suppose z ∈ C. Prove that √ ≤ |z| ≤ |Re z| + |Im z|.
2
4 Suppose w, z ∈ C. Prove that |wz| = |w| |z| and |w + z| ≤ |w| + |z|.
5 Suppose ( X, S) is a measurable space and f : X → C is a complex-valued
function. For conditions (b) and (c) below, identify C with R2 . Prove that the
following are equivalent:
(a) f is S -measurable.
(b) f −1 ( G ) ∈ S for every open set G in R2 .
(c) f −1 ( B) ∈ S for every Borel set B ∈ B2 .

6 Suppose ( X, S) is a measurable space and f , g : X → C are S -measurable.


Prove that
(a) f + g, f − g, and f g are S -measurable functions;
f
(b) if g( x ) 6= 0 for all x ∈ X, then g is an S -measurable function.

7 Suppose ( X, S) is a measurable space and f 1 , f 2 , . . . is a sequence of S -


measurable functions from X to C. Suppose lim f k ( x ) exists for each x ∈ X.
k→∞
Define f : X → C by
f ( x ) = lim f k ( x ).
k→∞
Prove that f is an S -measurable function.
8 Suppose ( X, S , µ)R is a measure space and f : X → C is an S -measurable
function such that | f | dµ < ∞. Prove that if α ∈ C, then
Z Z
α f dµ = α f dµ.

9 Suppose V is a vector space. Show that the intersection of every collection of


subspaces of V is a subspace of V.
10 Suppose V and W are vector spaces. Define V × W by
V × W = {( f , g) : f ∈ V and g ∈ W }.
Define addition and scalar multiplication on V × W by
( f 1 , g1 ) + ( f 2 , g2 ) = ( f 1 + f 2 , g1 + g2 ) and α( f , g) = (α f , αg).
Prove that V × W is a vector space with these operations.
Section 6C Normed Vector Spaces 163

6C Normed Vector Spaces


Norms and Complete Norms
This section begins with a crucial definition.

6.33 Definition norm; normed vector space

A norm on a vector space V (over F) is a function k·k : V → [0, ∞) such that


• k f k = 0 if and only if f = 0 (positive definite);
• kα f k = |α| k f k for all α ∈ F and f ∈ V (homogeneity);
• k f + gk ≤ k f k + k gk for all f , g ∈ V (triangle inequality).
A normed vector space is a pair (V, k·k), where V is a vector space and k·k is a
norm on V.

6.34 Example norms

• Suppose n ∈ Z+. Define k·k1 and k·k∞ on Fn by


k( a1 , . . . , an )k1 = | a1 | + · · · + | an |
and
k( a1 , . . . , an )k∞ = max{| a1 |, . . . , | an |}.
Then k·k1 and k·k∞ are norms on Fn , as you should verify.
• On `1 (see the last bullet point in Example 6.32 for the definition of `1 ), define
k·k1 by

k( a1 , a2 , . . .)k1 = ∑ | a k |.
k =1

Then k·k1 is a norm on `1 , as you should verify.


• Suppose X is a nonempty set and b( X ) is the subspace of F X consisting of the
bounded functions from X to F. For f a bounded function from X to F, define
k f k by
k f k = sup{| f ( x )| : x ∈ X }.
Then k·k is a norm on b( X ), as you should verify.
• Let C ([0, 1]) denote the vector space of continuous functions from the interval
[0, 1] to F. Define k·k on C ([0, 1]) by
Z 1
kfk = | f |.
0

Then k·k is a norm on C ([0, 1]), as you should verify.


164 Chapter 6 Banach Spaces

Sometimes examples that do not satisfy a definition help you gain understanding.

6.35 Example not norms


• Let L1 (R) denote theR vector space of Borel (or Lebesgue) measurable functions
f : R → F such that | f | dλ < ∞, where λ is Lebesgue measure on R. Define
k·k1 on L1 (R) by Z
k f k1 = | f | dλ.
Then k·k1 satisfies the homogeneity condition and the triangle inequality on
L1 (R), as you should verify. However, k·k1 is not a norm on L1 (R) because
the positive definite condition is not satisfied. Specifically, if E is a nonempty
Borel subset of R with Lebesgue measure 0 (for example, E might consist of a
single element of R), then kχ Ek1 = 0 but χ E 6= 0. In the next chapter, we will
discuss a modification of L1 (R) that removes this problem.
• If n ∈ Z+ and k·k is defined on Fn by
k( a1 , . . . , an )k = | a1 |1/2 + · · · + | an |1/2 ,
then k·k satisfies the positive definite condition and the triangle inequality (as
you should verify). However, k·k as defined above is not a norm because it does
not satisfy the homogeneity condition.
• If k·k1/2 is defined on Fn by
2
k( a1 , . . . , an )k1/2 = | a1 |1/2 + · · · + | an |1/2 ,
then k·k1/2 satisfies the positive definite condition and the homogeneity condi-
tion. However, if n > 1 then k·k1/2 is not a norm on Fn because the triangle
inequality is not satisfied (as you should verify).

The next result shows that every normed vector space is also a metric space in a
natural fashion.

6.36 normed vector spaces are metric spaces

Suppose (V, k·k) is a normed vector space. Define d : V × V → [0, ∞) by

d ( f , g ) = k f − g k.

Then d is a metric on V.

Proof Suppose f , g, h ∈ V. Then


d( f , h) = k f − hk = k( f − g) + ( g − h)k
≤ k f − gk + k g − hk
= d( f , g) + d( g, h).
Thus the triangle inequality requirement for a metric is satisfied. The verification of
the other required properties for a metric are left to the reader.
Section 6C Normed Vector Spaces 165

From now on, all metric space notions in the context of a normed vector space
should be interpreted with respect to the metric introduced in the previous result.
However, usually there is no need to introduce the metric d explicitly—just use the
norm of the difference of two elements. For example, suppose (V, k·k) is a normed
vector space, f 1 , f 2 , . . . is a sequence in V, and f ∈ V. Then in the context of a
normed vector space, the definition of limit (6.8) becomes the following statement:

lim f k = f means lim k f k − f k = 0.


k→∞ k→∞

As another example, in the context of a normed vector space, the definition of a


Cauchy sequence (6.12) becomes the following statement:

A sequence f 1 , f 2 , . . . in a normed vector space (V, k·k) is a Cauchy se-


quence if for every ε > 0, there exists n ∈ Z+ such that k f j − f k k < ε for
all integers j ≥ n and k ≥ n.

Every sequence in a normed vector space that has a limit is a Cauchy sequence
(see 6.13). Normed vector spaces that satisfy the converse have a special name.

6.37 Definition Banach space

A complete normed vector space is called a Banach space.

In other words, a normed vector space


In a slight abuse of terminology, we
V is a Banach space if every Cauchy se-
often refer to a normed vector space
quence in V converges to some element
V without mentioning the norm k·k.
of V.
When that happens, you should
The verifications of the assertions in
assume that a norm k·k lurks nearby,
Examples 6.38 and 6.39 below are left to
even if it is not explicitly displayed.
the reader as exercises.

6.38 Example Banach spaces

• The vector space C ([0, 1]) with the norm defined by k f k = sup| f | is a Banach
space. [0, 1]

• The vector space `1 with the norm defined by k( a1 , a2 , . . .)k1 = ∑∞


k=1 | ak | is a
Banach space.

6.39 Example not a Banach space


R1
• The vector space C ([0, 1]) with the norm defined by k f k = 0 | f | is not a
Banach space.
• The vector space `1 with the norm defined by k( a1 , a2 , . . .)k∞ = sup | ak | is
not a Banach space. k ∈Z+
166 Chapter 6 Banach Spaces

6.40 Definition infinite sum in a normed vector space

Suppose g1 , g2 , . . . is a sequence in a normed vector space V. Then ∑∞


k=1 gk is
defined by
∞ n
∑ gk = lim
n→∞
∑ gk
k =1 k =1
if this limit exists, in which case the infinite series is said to converge.

Recall from your calculus course that if a1 , a2 , . . . is a sequence of real numbers


such that ∑∞ ∞
k=1 | ak | < ∞, then ∑k=1 ak converges. The next result states that the
analogous property for normed vector spaces characterizes Banach spaces.

 
6.41 ∑∞ ∞
k=1 k gk k < ∞ =⇒ ∑k=1 gk converges ⇐⇒ Banach space

Suppose V is a normed vector space. Then V is a Banach space if and only if


∑∞ ∞
k=1 gk converges for every sequence g1 , g2 , . . . in V such that ∑k=1 k gk k < ∞.

Proof First suppose V is a Banach space. Suppose g1 , g2 , . . . is a sequence in V such


that ∑∞ ∞
k=1 k gk k < ∞. Suppose ε > 0. Let n ∈ Z be such that ∑m=n k gm k < ε.
+
For j ∈ Z , let f j denote the partial sum defined by
+

f j = g1 + · · · + g j .

If k > j ≥ n, then

k f k − f j k = k g j +1 + · · · + g k k
≤ k g j +1 k + · · · + k g k k

≤ ∑ k gm k
m=n

< ε.
Thus f 1 , f 2 , . . . is a Cauchy sequence in V. Because V is a Banach space, we conclude
that f 1 , f 2 , . . . converges to some element of V, which is precisely what it means for
∑∞k=1 gk to converge, completing one direction of the proof.
To prove the other direction, suppose ∑∞ k=1 gk converges for every sequence
g1 , g2 , . . . in V such that ∑∞ k =1 k g k k < ∞. Suppose f 1 , f 2 , . . . is a Cauchy sequence
in V. We want to prove that f 1 , f 2 , . . . converges to some element of V. It suffices to
show that some subsequence of f 1 , f 2 , . . . converges (by Exercise 14 in Section 6A).
Dropping to a subsequence (but not relabeling) and setting f 0 = 0, we can assume
that

∑ k f k − f k−1 k < ∞.
k =1
Hence ∑∞k=1 ( f k − f k−1 ) converges. The partial sum of this series after n terms is f n .
Thus limn→∞ f n exists, completing the proof.
Section 6C Normed Vector Spaces 167

Bounded Linear Maps


When dealing with two or more vector spaces, as in the definition below, assume that
the vector spaces are over the same field (either R or C, but denoted in this book as F
to give us the flexibility to consider both cases).
The notation T f , in addition to the standard functional notation T ( f ), is often
used when considering linear maps, which we now define.

6.42 Definition linear map

Suppose V and W are vector spaces. A function T : V → W is called linear if


• T ( f + g) = T f + Tg for all f , g ∈ V;
• T (α f ) = αT f for all α ∈ F and f ∈ V.
A linear function is often called a linear map.

The set of linear maps from a vector space V to a vector space W is itself a vector
space, using the usual operations of addition and scalar multiplication of functions.
Most attention in analysis focuses on the subset of bounded linear functions, defined
below, which we will see is itself a normed vector space.
In the next definition, we have two normed vector spaces, V and W, which may
have different norms. However, we use the same notation k·k for both norms (and
for the norm of a linear map from V to W) because the context makes the meaning
clear. For example, in the definition below, f is in V and thus k f k refers to the norm
in V. Similarly, T f ∈ W and thus k T f k refers to the norm in W.

6.43 Definition bounded linear map; k T k; B(V, W )

Suppose V and W are normed vector spaces and T : V → W is a linear map.


• The norm of T, denoted k T k, is defined by

k T k = sup{k T f k : f ∈ V and k f k ≤ 1}.

• T is called bounded if k T k < ∞.


• The set of bounded linear maps from V to W is denoted B(V, W ).

6.44 Example bounded linear map


Let C ([0, 3]) be the normed vector space of continuous functions from [0, 3] to F,
with k f k = sup| f |. Define T : C ([0, 3]) → C ([0, 3]) by
[0, 3]
( T f )( x ) = x2 f ( x ).

Then T is a bounded linear map and k T k = 9, as you should verify.


168 Chapter 6 Banach Spaces

6.45 Example linear map that is not bounded


Let V be the normed vector space of sequences ( a1 , a2 , . . .) of elements of F such
that ak = 0 for all but finitely many k ∈ Z+, with k( a1 , a2 , . . .)k∞ = maxk∈Z+ | ak |.
Define T : V → V by
T ( a1 , a2 , a3 , . . .) = ( a1 , 2a2 , 3a3 , . . .).
Then T is a linear map that is not bounded, as you should verify.

The next result shows that if V and W are normed vector spaces, then B(V, W ) is
a normed vector space with the norm defined above.

6.46 k·k is a norm on B(V, W )

Suppose V and W are normed vector spaces. Then kS + T k ≤ kSk + k T k


and kαT k = |α| k T k for all S, T ∈ B(V, W ) and all α ∈ F. Furthermore, the
function k·k is a norm on B(V, W ).

Proof Suppose S, T ∈ B(V, W ). then


kS + T k = sup{k(S + T ) f k : f ∈ V and k f k ≤ 1}
≤ sup{kS f k + k T f k : f ∈ V and k f k ≤ 1}
≤ sup{kS f k : f ∈ V and k f k ≤ 1}
+ sup{k T f k : f ∈ V and k f k ≤ 1}
= k S k + k T k.
The inequality above shows that k·k satisfies the triangle inequality on B(V, W ).
The verification of the other properties required for a normed vector space is left to
the reader.

Be sure that you are comfortable using all four equivalent formulas for k T k shown
in Exercise 16. For example, you should often think of k T k as the smallest number
such that k T f k ≤ k T k k f k for all f in the domain of T.
Note that in the next result, the hypothesis requires W to be a Banach space but
there is no requirement for V to be a Banach space.

6.47 B(V, W ) is a Banach space if W is a Banach space

Suppose V is a normed vector space and W is a Banach space. Then B(V, W ) is


a Banach space.

Proof Suppose T1 , T2 , . . . is a Cauchy sequence in B(V, W ). If f ∈ V, then


k Tj f − Tk f k ≤ k Tj − Tk k k f k,
which implies that T1 f , T2 f , . . . is a Cauchy sequence in W. Because W is a Banach
space, this implies that T1 f , T2 f , . . . has a limit in W, which we call T f .
Section 6C Normed Vector Spaces 169

We have now defined a function T : V → W. The reader should verify that T is a


linear map. Clearly

k T f k ≤ sup{k Tk f k : k ∈ Z+}

≤ sup{k Tk k : k ∈ Z+} k f k

for each f ∈ V. The last supremum above is finite because every Cauchy sequence is
bounded (see Exercise 4). Thus T ∈ B(V, W ).
We still need to show that limk→∞ k Tk − T k = 0. To do this, suppose ε > 0. Let
n ∈ Z+ be such that k Tj − Tk k < ε for all j ≥ n and k ≥ n. Suppose j ≥ n and
suppose f ∈ V. Then

k( Tj − T ) f k = lim k Tj f − Tk f k
k→∞

≤ ε k f k.

Thus k Tj − T k ≤ ε, completing the proof.

The next result shows that the phrase bounded linear map means the same as the
phrase continuous linear map.

6.48 continuity is equivalent to boundedness for linear maps

A linear map from one normed vector space to another normed vector space is
continuous if and only if it is bounded.

Proof Suppose V and W are normed vector spaces and T : V → W is linear.


First suppose T is not bounded. Thus there exists a sequence f 1 , f 2 , . . . in V such
that k f k k ≤ 1 for each k ∈ Z+ and k T f k k → ∞ as k → ∞. Hence

fk  fk  T fk
lim = 0 and T = 6→ 0,
k→∞ kT fk k kT fk k kT fk k

where the nonconvergence to 0 holds because T f k /k T f k k has norm 1 for every


k ∈ Z+. The displayed line above implies that T is not continuous, completing the
proof in one direction.
To prove the other direction, now suppose T is bounded. Suppose f ∈ V and
f 1 , f 2 , . . . is a sequence in V such that limk→∞ f k = f . Then

k T f k − T f k = k T ( f k − f )k
≤ k T k k f k − f k.

Thus limk→∞ T f k = T f . Hence T is continuous, completing the proof in the other


direction.

Exercise 18 gives several additional equivalent conditions for a linear map to be


continuous.
170 Chapter 6 Banach Spaces

EXERCISES 6C
1 Show that the map f 7→ k f k from a normed vector space V to F is continuous
(where the norm on F is the usual absolute value).
2 Prove that if V is a normed vector space, f ∈ V, and r > 0, then

B ( f , r ) = B ( f , r ).

3 Show that the functions defined in the last two bullet points of Example 6.35 are
not norms.
4 Prove that each Cauchy sequence in a normed vector space is bounded (meaning
that there is a real number that is greater than the norm of every element in the
Cauchy sequence).
5 Show that if n ∈ Z+, then Fn is a Banach space with both the norms used in the
first bullet point of Example 6.34.
6 Suppose X is a nonempty set and b( X ) is the vector space of bounded functions
from X to F. Prove that if k·k is defined on b( X ) by k f k = sup| f |, then b( X )
is a Banach space. X

7 Show that `1 with the norm defined by k( a1 , a2 , . . .)k∞ = sup | ak | is not a


Banach space. k ∈Z+

8 Show that `1 with the norm defined by k( a1 , a2 , . . .)k1 = ∑∞


k=1 | ak | is a Banach
space.
9 Show that the vector space C ([0, 1]) of continuous functions from [0, 1] to F
R1
with the norm defined by k f k = 0 | f | is not a Banach space.
10 Suppose U is a subspace of a normed vector space V such that some open ball
of V is contained in U. Prove that U = V.
11 Prove that the only subsets of a normed vector space V that are both open and
closed are ∅ and V.
12 Suppose V is a normed vector space. Prove that the closure of each subspace of
V is a subspace of V.
13 Suppose U is a normed vector space. Let d be the metric on U defined by
d( f , g) = k f − gk for f , g ∈ U. Let V be the complete metric space con-
structed in Exercise 16 in Section 6A.
(a) Show that the set V is a vector space under natural operations of addition
and scalar multiplication.
(b) Show that there is a natural way to make V into a normed vector space and
that with this norm, V is a Banach space.
(c) Explain why (b) shows that every normed vector space is a subspace of
some Banach space.
Section 6C Normed Vector Spaces 171

14 Suppose U is a subspace of a normed vector space V. Suppose also that W is a


Banach space and S : U → W is a bounded linear map.
(a) Prove that there exists a unique continuous function T : U → W such that
T |U = S.
(b) Prove that the function T in part (a) is a bounded linear map from U to W
and k T k = kSk.
(c) Give an example to show that part (a) can fail if the assumption that W is
a Banach space is replaced by the assumption that W is a normed vector
space.

15 For readers familiar with the quotient of a vector space and a subspace: Suppose
V is a normed vector space and U is a subspace of V. Define k·k on V/U by

k f + U k = inf{k f + gk : g ∈ U }.

(a) Prove that k·k is a norm on V/U if and only if U is a closed subspace of V.
(b) Prove that if V is a Banach space and U is a closed subspace of V, then
V/U (with the norm defined above) is a Banach space.
(c) Prove that if U is a Banach space (with the norm it inherits from V) and
V/U is a Banach space (with the norm defined above), then V is a Banach
space.

16 Suppose V and W are normed vector spaces with V 6= {0} and T : V → W is


a linear map.
(a) Show that k T k = sup{k T f k : f ∈ V and k f k < 1}.
(b) Show that k T k = sup{k T f k : f ∈ V and k f k = 1}.
(c) Show that k T k = inf{c ∈ [0, ∞) : k T f k ≤ ck f k for all f ∈ V }.
n kT f k o
(d) Show that k T k = sup : f ∈ V and f 6= 0 .
kfk

17 Suppose U, V, and W are normed vector spaces and T : U → V and S : V → W


are linear. Prove that kS ◦ T k ≤ kSk k T k.
18 Suppose V and W are normed vector spaces and T : V → W is a linear map.
Prove that the following are equivalent:
(a) T is bounded.
(b) There exists f ∈ V such that T is continuous at f .
(c) T is uniformly continuous (which means that for every ε > 0, there exists
δ > 0 such that k T f − Tgk < ε for all f , g ∈ V with k f − gk < δ).

(d) T −1 B(0, r ) is an open subset of V for some r > 0.
172 Chapter 6 Banach Spaces

6D Linear Functionals
Bounded Linear Functionals
Linear maps into the scalar field F are so important that they get a special name.

6.49 Definition linear functional

A linear functional on a vector space V is a linear map from V to F.

When we think of the scalar field F as a normed vector space, as in the next
example, the norm kzk of a number z ∈ F is always intended to be just the usual
absolute value |z|. This norm makes F a Banach space.

6.50 Example linear functional


Let V be the vector space of sequences ( a1 , a2 , . . .) of elements of F such that
ak = 0 for all but finitely many k ∈ Z+. Define ϕ : V → F by

ϕ ( a1 , a2 , . . . ) = ∑ ak .
k =1
Then ϕ is a linear functional on V.

• If we make V a normed vector space with the norm k( a1 , a2 , . . .)k1 = ∑ | ak |,
then ϕ is a bounded linear functional on V, as you should verify. k =1

• If we make V a normed vector space with the norm k( a1 , a2 , . . .)k∞ = max | ak |,


then ϕ is not a bounded linear functional on V, as you should verify. k∈Z
+

6.51 Definition null space; null T

Suppose V and W are vector spaces and T : V → W is a linear map. Then the
null space of T is denoted by null T and is defined by

null T = { f ∈ V : T f = 0}.

If T is a linear map on a vector space


The term kernel is also used in the
V, then null T is a subspace of V, as you
mathematics literature with the
should verify. If T is a continuous linear
same meaning as null space. This
map from a normed vector space V to a
book uses null space instead of
normed vector space W, then null T is a
kernel because null space better
closed subspace of V because null T =
captures the connection with 0.
T −1 ({0}) and the inverse image of the
closed set {0} is closed [by 6.11(d)].
The converse of the last sentence fails, because a linear map between normed
vector spaces can have a closed null space but not be continuous. For example, the
linear map in 6.45 has a closed null space (equal to {0}) but it is not continuous.
However, the next result states that for linear functionals, as opposed to more
general linear maps, having a closed null space is equivalent to continuity.
Section 6D Linear Functionals 173

6.52 bounded linear functionals

Suppose V is a normed vector space and ϕ : V → F is a linear functional that is


not identically 0. Then the following are equivalent:

(a) ϕ is a bounded linear functional.


(b) ϕ is a continuous linear functional.
(c) null ϕ is a closed subspace of V.

(d) null ϕ 6= V.

Proof The equivalence of (a) and (b) is just a special case of 6.48.
To prove that (b) implies (c), suppose ϕ is a continuous linear functional. Then
null ϕ, which is the inverse image of the closed set {0}, is a closed subset of V by
6.11(d). Thus (b) implies (c).
To prove that (c) implies (a), we will show that the negation of (a) implies the
negation of (c). Thus suppose ϕ is not bounded. Thus there is a sequence f 1 , f 2 , . . .
in V such that k f k k ≤ 1 and | ϕ( f k )| ≥ k for each k ∈ Z+. Now
f1 f
− k ∈ null ϕ
ϕ( f 1 ) ϕ( f k ) This proof makes major use of
dividing by expressions of the form
for each k ∈ Z+ and ϕ( f ), which would not make sense
 f f  f1 for a linear mapping into a vector
1
lim − k = .
k→∞ ϕ ( f 1 ) ϕ ( f k ) ϕ ( f1 ) space other than F.
Clearly
 f1  f1
ϕ = 1 and thus / null ϕ.

ϕ( f 1 ) ϕ( f 1 )
The last three displayed items imply that null ϕ is not closed, completing the proof
that the negation of (a) implies the negation of (c). Thus (c) implies (a).
We now know that (a), (b), and (c) are equivalent to each other.
Using the hypothesis that ϕ is not identically 0, we see that (c) implies (d). To
complete the proof, we need only show that (d) implies (c), which we will do by
showing that the negation of (c) implies the negation of (d). Thus suppose null ϕ is
not a closed subspace of V. Because null ϕ is a subspace of V, we know that null ϕ
is also a subspace of V (see Exercise 12 in Section 6C). Let f ∈ null ϕ \ null ϕ.
Suppose g ∈ V. Then
 ϕ( g)  ϕ( g)
g = g− f + f.
ϕ( f ) ϕ( f )

The term in large parentheses above is in null ϕ and hence is in null ϕ. The term
above following the plus sign is a scalar multiple of f and thus is in null ϕ. Because
the equation above writes g as the sum of two elements of null ϕ, we conclude that
g ∈ null ϕ. Hence we have shown that V = null ϕ, completing the proof that the
negation of (c) implies the negation of (d).
174 Chapter 6 Banach Spaces

Discontinuous Linear Functionals


The second bullet point in Example 6.50 shows that there exists a discontinuous linear
functional on a certain normed vector space. Our next major goal is to show that every
infinite-dimensional normed vector space has a discontinuous linear functional (see
6.62). Thus infinite-dimensional normed vector spaces behave in this respect much
differently from Fn , where all linear functionals are continuous (see Exercise 4).
We need to extend the notion of a basis of a finite-dimensional vector space to an
infinite-dimensional context. In a finite-dimensional vector space, we might consider
a basis of the form e1 , . . . , en , where n ∈ Z+ and each ek is an element of our vector
space. We can think of the list e1 , . . . , en as a function from {1, . . . , n} to our vector
space, with the value of this function at k ∈ {1, . . . , n} denoted by ek with a subscript
k instead of by the usual functional notation e(k). To generalize, in the next definition
we allow {1, . . . , n} to be replaced by an arbitrary set that might not be a finite set.

6.53 Definition family

A family {ek }k∈Γ in a set V is a function e from a set Γ to V, with the value of
the function e at k ∈ Γ denoted by ek .

Even though a family in V is a function mapping into V and thus is not a subset
of V, the set terminology and the bracket notation {ek }k∈Γ are useful, and the range
of a family in V really is a subset of V.
We now restate some basic linear algebra concepts, but in the context of vector
spaces that might be infinite-dimensional. Note that only finite sums appear in the
definition below, even though we might be working with an infinite family.

6.54 Definition linearly independent; span; basis

Suppose {ek }k∈Γ is a family in a vector space V.


• {ek }k∈Γ is called linearly independent if there does not exist a finite
nonempty subset Ω of Γ and a family {α j } j∈Ω in F \ {0} such that
∑ j∈Ω α j e j = 0.

• The span of {ek }k∈Γ is denoted by span{ek }k∈Γ and is defined to be the set
of all sums of the form
∑ αj ej ,
j∈Ω

where Ω is a finite subset of Γ and {α j } j∈Ω is a family in F.


• A vector space V is called finite-dimensional if there exists a finite set Γ and
a family {ek }k∈Γ in V such that span{ek }k∈Γ = V.
• A vector space is called infinite-dimensional if it is not finite-dimensional.
• A family in V is called a basis of V if it is linearly independent and its span
equals V.
Section 6D Linear Functionals 175

For example, { x n }n∈{0, 1, 2, ...} is a ba-


The term Hamel basis is sometimes
sis of the vector space of polynomials.
used to denote what has been called
Our definition of span does not take
a basis here. The use of the term
advantage of the possibility of summing
Hamel basis emphasizes that only
an infinite number of elements in contexts
finite sums are under consideration.
where a notion of limit exists (as is the
case in normed vector spaces). When we get to Hilbert spaces in Chapter 8, we
consider another kind of basis that does involve infinite sums. As we will soon see,
the kind of basis as defined here is just what we need to produce discontinuous linear
functionals.
Now we introduce terminology that
No one has ever produced a
will be needed in our proof that every vec-
concrete example of a basis of an
tor space has a basis.
infinite-dimensional Banach space.

6.55 Definition maximal element

Suppose A is a collection of subsets of a set V. A set Γ ∈ A is called a maximal


element of A if there does not exist Γ0 ∈ A such that Γ $ Γ0 .

6.56 Example maximal elements


For k ∈ Z, let kZ denote the set of integer multiples of k; thus kZ = {km : m ∈ Z}.
Let A be the collection of subsets of Z defined by A = {kZ : k = 2, 3, 4, . . .}.
Suppose k ∈ Z+. Then kZ is a maximal element of A if and only if k is a prime
number, as you should verify.

A subset Γ of a vector space V can be thought of as a family in V by considering


{e f } f ∈Γ , where e f = f . With this convention, the next result shows that the bases of
V are exactly the maximal elements among the collection of linearly independent
subsets of V.

6.57 bases as maximal elements

Suppose V is a vector space. Then a subset of V is a basis of V if and only if it is


a maximal element of the collection of linearly independent subsets of V.

Proof Suppose Γ is a linearly independent subset of V.


First suppose also that Γ is a basis of V. If f ∈ V but f ∈ / Γ, then f ∈ span Γ,
which implies that Γ ∪ { f } is not linearly independent. Thus Γ is a maximal element
among the collection of linearly independent subsets of V, completing one direction
of the proof.
To prove the other direction, suppose now that Γ is a maximal element of the
collection of linearly independent subsets of V. If f ∈ V but f ∈ / span Γ, then
Γ ∪ { f } is linearly independent, which would contradict the maximality of Γ among
the collection of linearly independent subsets of V. Thus span Γ = V, which means
that Γ is a basis of V, completing the proof in the other direction.
176 Chapter 6 Banach Spaces

The notion of a chain plays a key role in our next result.

6.58 Definition chain

A collection C of subsets of a set V is called a chain if Ω, Γ ∈ C implies Ω ⊂ Γ


or Γ ⊂ Ω.

6.59 Example chains

• The collection C = {4Z, 6Z} of subsets of Z is not a chain because neither of


the sets 4Z or 6Z is a subset of the other.
• The collection C = {2n Z : n ∈ Z+} of subsets of Z is a chain because if
m, n ∈ Z+, then 2m Z ⊂ 2n Z or 2n Z ⊂ 2m Z.

The next result follows from the Ax-


Zorn’s Lemma is named in honor of
iom of Choice, although it is not as intu-
Max Zorn (1906–1993), who
itively believable as the Axiom of Choice.
published a paper containing the
Because the techniques used to prove the
result in 1935, when he had a
next result are so different from tech-
postdoctoral position at Yale.
niques used elsewhere in this book, the
reader is asked either to accept this result without proof or find one of the good proofs
available via the internet or in other books. The version of Zorn’s Lemma stated here
is simpler than the standard more general version, but this version is all that we need.

6.60 Zorn’s Lemma

Suppose V is a set and A is a collection of subsets of V with the property that


the union of all the sets in C is in A for every chain C ⊂ A. Then A contains a
maximal element.

Zorn’s Lemma now allows us to prove that every vector space has a basis. The
proof does not help us find a concrete basis because Zorn’s Lemma is an existence
result rather than a constructive technique.

6.61 bases exist

Every vector space has a basis.

Proof Suppose V is a vector space. If C is a chain of linearly independent subsets


of V, then the union of all the sets in C is also a linearly independent subset of V (this
holds because linear independence is a condition that is checked by considering finite
subsets, and each finite subset of the union is contained in one of the elements of the
chain).
Thus if A denotes the collection of linearly independent subsets of V, then A
satisfies the hypothesis of Zorn’s Lemma (6.60). Hence A contains a maximal
element, which by 6.57 is a basis of V.
Section 6D Linear Functionals 177

Now we can prove the promised result about the existence of discontinuous linear
functionals on every infinite-dimensional normed vector space.

6.62 discontinuous linear functionals

Every infinite-dimensional normed vector space has a discontinuous linear


functional.

Proof Suppose V is an infinite-dimensional vector space. By 6.61, V has a basis


{ek }k∈Γ . Because V is infinite-dimensional, Γ is not a finite set. Thus we can assume
Z+ ⊂ Γ (by relabeling a countable subset of Γ).
Define a linear functional ϕ : V → F by setting ϕ(e j ) equal to jke j k for j ∈ Z+,
setting ϕ(e j ) equal to 0 for j ∈ Γ \ Z+, and extending linearly. More precisely, define
a linear functional ϕ : V → F by
 
ϕ ∑ α j e j = ∑ α j jke j k
j∈Ω j∈Ω∩Z+

for every finite subset Ω ⊂ Γ and every family {α j } j∈Ω in F.


Because ϕ(e j ) = jke j k for each j ∈ Z+, the linear functional ϕ is unbounded,
completing the proof.

Hahn–Banach Theorem
In the last subsection, we showed that there exists a discontinuous linear functional
on each infinite-dimensional normed vector space. Now we turn our attention to the
existence of continuous linear functionals.
The existence of a nonzero continuous linear functional on each Banach space is
not obvious. For example, consider the Banach space `∞ /c0 , where `∞ is the Banach
space of bounded sequences in F with
k( a1 , a2 , . . .)k∞ = sup | ak |
k ∈Z+
and c0 is the subspace of `∞ consisting of those sequences in F that have limit 0. The
quotient space `∞ /c0 is an infinite-dimensional Banach space (see Exercise 15 in
Section 6C). However, no one has ever exhibited a concrete nonzero linear functional
on the Banach space `∞ /c0 .
In this subsection, we show that infinite-dimensional normed vector spaces have
plenty of continuous linear functionals. We do this by showing that a bounded linear
functional on a subspace of a normed vector space can be extended to a bounded
linear functional on the whole space without increasing its norm—this result is called
the Hahn–Banach Theorem (6.69).
Completeness plays no role in this topic. Thus this subsection deals with normed
vector spaces instead of Banach spaces.
We do most of the work needed to prove the Hahn–Banach Theorem in the next
lemma, which shows that we can extend a linear functional to a subspace generated
by one additional element, without increasing the norm. This one-element-at-a-time
approach, when combined with a maximal object produced by Zorn’s Lemma, gives
us the desired extension to the full normed vector space.
178 Chapter 6 Banach Spaces

If V is a real vector space, U is a subspace of V, and h ∈ V, then U + Rh is the


subspace of V defined by

U + Rh = { f + αh : f ∈ U and α ∈ R}.

6.63 Extension Lemma

Suppose V is a real normed vector space, U is a subspace of V, and ψ : U → R


is a bounded linear functional. Suppose h ∈ V \ U. Then ψ can be extended to a
bounded linear functional ϕ : U + Rh → R such that k ϕk = kψk.

Proof Suppose c ∈ R. Define ϕ(h) to be c, and then extend ϕ linearly to U + Rh.


Specifically, define ϕ : U + Rh → R by

ϕ( f + αh) = ψ( f ) + αc

for f ∈ U and α ∈ R. Then ϕ is a linear functional on U + Rh.


Clearly ϕ|U = ψ. Thus k ϕk ≥ kψk. We need to show that for some choice of
c ∈ R, the linear functional ϕ defined above satisfies the equation k ϕk = kψk. In
other words, we want

6.64 |ψ( f ) + αc| ≤ kψk k f + αhk for all f ∈ U and all α ∈ R.

It would be enough to have

6.65 |ψ( f ) + c| ≤ kψk k f + hk for all f ∈ U,


f
because replacing f by α in the last inequality and then multiplying both sides by |α|
would give 6.64.
Rewriting 6.65, we want to show that there exists c ∈ R such that

−kψk k f + hk ≤ ψ( f ) + c ≤ kψk k f + hk for all f ∈ U.

Equivalently, we want to show that there exists c ∈ R such that

−kψk k f + hk − ψ( f ) ≤ c ≤ kψk k f + hk − ψ( f ) for all f ∈ U.

The existence of c ∈ R satisfying the line above follows from the inequality
 
6.66 sup −kψk k f + hk − ψ( f ) ≤ inf kψk k g + hk − ψ( g) .
f ∈U g ∈U

To prove the inequality above, suppose f , g ∈ U. Then

−kψk k f + hk − ψ( f ) ≤ kψk(k g + hk − k g − f k) − ψ( f )
= kψk(k g + hk − k g − f k) + ψ( g − f ) − ψ( g)
≤ k ψ k k g + h k − ψ ( g ).

The inequality above proves 6.66, which completes the proof.


Section 6D Linear Functionals 179

Because our simplified form of Zorn’s Lemma deals with set inclusions rather
than more general orderings, we need to use the notion of the graph of a function.

6.67 Definition graph

Suppose T : V → W is a function from a set V to a set W. Then the graph of T


is denoted graph( T ) and is the subset of V × W defined by

graph( T ) = { f , T ( f ) ∈ V × W : f ∈ V }.

Formally, a function from a set V to a set W equals its graph as defined above.
However, because we usually think of a function more intuitively as a mapping, the
separate notion of the graph of a function remains useful.
The easy proof of the next result is left to the reader. The first bullet point
below uses the vector space structure of V × W, which is a vector space with natural
operations of addition and scalar multiplication, as given in Exercise 10 in Section 6B.

6.68 function properties in terms of graphs

Suppose V and W are normed vector spaces and T : V → W is a function.

(a) T is a linear map if and only if graph( T ) is a subspace of V × W.


(b) Suppose U ⊂ V and S : U → W is a function. Then T is an extension of S
if and only if graph(S) ⊂ graph( T ).
(c) If T : V → W is a linear map and c ∈ [0, ∞), then k T k ≤ c if and only if
k gk ≤ ck f k for all ( f , g) ∈ graph( T ).

The proof of the Extension Lemma


Hans Hahn (1879–1934) was a
(6.63) used inequalities that do not make
student and later a faculty member
sense when F = C. Thus the proof of the
at the University of Vienna, where
Hahn–Banach Theorem below requires
one of his PhD students was Kurt
some extra steps when F = C.
Gödel (1906–1978).

6.69 Hahn–Banach Theorem

Suppose V is a normed vector space, U is a subspace of V, and ψ : U → F is a


bounded linear functional. Then ψ can be extended to a bounded linear functional
on V whose norm equals kψk.

Proof First we consider the case where F = R. Let A be the collection of subsets
E of V × R that satisfy all the following conditions:
• E = graph( ϕ) for some linear functional ϕ on some subspace of V;
• graph(ψ) ⊂ E;
• |α| ≤ kψk k f k for every ( f , α) ∈ E.
180 Chapter 6 Banach Spaces

Then A satisfies the hypothesis of Zorn’s Lemma (6.60). Thus A has a maximal
element. The Extension Lemma (6.63) implies that this maximal element is the graph
of a linear functional defined on all of V. This linear functional is an extension of ψ
to V and it has norm kψk, completing the proof in the case where F = R.
Now consider the case where F = C. Define ψ1 : U → R by
ψ1 ( f ) = Re ψ( f )
for f ∈ U. Then ψ1 is an R-linear map from U to R and kψ1 k ≤ kψk (actually
kψ1 k = kψk, but we need only the inequality). Also,
ψ( f ) = Re ψ( f ) + i Im ψ( f )

= ψ1 ( f ) + i Im −iψ(i f )

= ψ1 ( f ) − i Re ψ(i f )
6.70 = ψ1 ( f ) − iψ1 (i f )
for all f ∈ U.
Temporarily forget that complex scalar multiplication makes sense on V and
temporarily think of V as a real normed vector space. The case of the result that
we have already proved then implies that there exists an extension ϕ1 of ψ1 to an
R-linear functional ϕ1 : V → R with k ϕ1 k = kψ1 k ≤ kψk.
Motivated by 6.70, we define ϕ : V → C by
ϕ( f ) = ϕ1 ( f ) − iϕ1 (i f )
for f ∈ V. The equation above and 6.70 imply that ϕ is an extension of ψ to V. The
equation above also implies that ϕ( f + g) = ϕ( f ) + ϕ( g) and ϕ(α f ) = αϕ( f ) for
all f , g ∈ V and all α ∈ R. Also,

ϕ(i f ) = ϕ1 (i f ) − iϕ1 (− f ) = ϕ1 (i f ) + iϕ1 ( f ) = i ϕ1 ( f ) − iϕ1 (i f ) = iϕ( f ).
The reader should use the equation above to show that ϕ is a C-linear map.
The only part of the proof that remains is to show that k ϕk ≤ kψk. To do this,
note that
 
| ϕ( f )|2 = ϕ ϕ( f ) f = ϕ1 ϕ( f ) f ≤ kψk k ϕ( f ) f k = kψk | ϕ( f )| k f k

for all f ∈ V, where the second equality holds because ϕ ϕ( f ) f ∈ R. Dividing by
| ϕ( f )|, we see from the line above that | ϕ( f )| ≤ kψk k f k for all f ∈ V (no division
necessary if ϕ( f ) = 0). This implies that k ϕk ≤ kψk, completing the proof.

We have given the special name linear functionals to linear maps into the scalar
field F. The vector space of bounded linear functionals now also gets a special name
and a special notation.

6.71 Definition dual space

Suppose V is a normed vector space. Then the dual space of V, denoted V 0, is the
normed vector space consisting of the bounded linear functionals on V. In other
words, V 0 = B(V, F).

By 6.47, the dual space of every normed vector space is a Banach space.
Section 6D Linear Functionals 181

6.72 k f k = max{| ϕ( f )| : ϕ ∈ V 0 and k ϕk = 1}

Suppose V is a normed vector space and f ∈ V \ {0}. Then there exists ϕ ∈ V 0


such that k ϕk = 1 and k f k = ϕ( f ).

Proof Let U be the 1-dimensional subspace of V defined by


U = { α f : α ∈ F }.
Define ψ : U → F by
ψ(α f ) = αk f k
for α ∈ F. Then ψ is a linear functional on U with kψk = 1 and ψ( f ) = k f k. The
Hahn–Banach Theorem (6.69) implies that there exists an extension of ψ to a linear
functional ϕ on V with k ϕk = 1, completing the proof.

The next result gives another beautiful application of the Hahn–Banach Theorem,
with a useful necessary and sufficient condition for an element of a normed vector
space to be in the closure of a subspace.

6.73 condition to be in the closure of a subspace

Suppose U is a subspace of a normed vector space V and h ∈ V. Then h ∈ U if


and only if ϕ(h) = 0 for every ϕ ∈ V 0 such that ϕ|U = 0.

Proof First suppose h ∈ U. If ϕ ∈ V 0 and ϕ|U = 0, then ϕ(h) = 0 by the


continuity of ϕ, completing the proof in one direction.
/ U. Define ψ : U + Fh → F by
To prove the other direction, suppose now that h ∈
ψ( f + αh) = α
for f ∈ U and α ∈ F. Then ψ is a linear functional on U + Fh with null ψ = U and
ψ(h) = 1.
Because h ∈ / U, the closure of the null space of ψ does not equal U + Fh. Thus
6.52 implies that ψ is a bounded linear functional on U + Fh.
The Hahn–Banach Theorem (6.69) implies that ψ can be extended to a bounded
linear functional ϕ on V. Thus we have found ϕ ∈ V 0 such that ϕ|U = 0 but
ϕ(h) 6= 0, completing the proof in the other direction.

EXERCISES 6D
1 Suppose V is a normed vector space and ϕ is a linear functional on V. Suppose
α ∈ F \ {0}. Prove that the following are equivalent:
(a) ϕ is a bounded linear functional.
(b) ϕ−1 (α) is a closed subset of V.
(c) ϕ−1 (α) 6= V.
182 Chapter 6 Banach Spaces

2 Suppose ϕ is a linear functional on a vector space V. Prove that if U is a


subspace of V such that null ϕ ⊂ U, then U = null ϕ or U = V.
3 Suppose ϕ and ψ are linear functionals on the same vector space. Prove that
null ϕ ⊂ null ψ
if and only if there exists α ∈ F such that ψ = αϕ.

For the next two exercises, Fn should be endowed with the norm k·k∞ as defined
in Example 6.34.
4 Suppose n ∈ Z+ and V is a normed vector space. Prove that every linear map
from Fn to V is continuous.
5 Suppose n ∈ Z+, V is a normed vector space, and T : Fn → V is a linear map
that is one-to-one and onto V.
(a) Show that
inf{k Tx k : x ∈ Fn and k x k∞ = 1} > 0.
(b) Prove that T −1 : V → Fn is a bounded linear map.

6 Suppose n ∈ Z+.
(a) Prove that all norms on Fn have the same convergent sequences, the same
open sets, and the same closed sets.
(b) Prove that all norms on Fn make Fn into a Banach space.

7 Suppose V and W are normed vector spaces and V is finite-dimensional. Prove


that every linear map from V to W is continuous.
8 Prove that every finite-dimensional normed vector space is a Banach space.
9 Prove that every finite-dimensional subspace of each normed vector space is
closed.
10 Give a concrete example of an infinite-dimensional normed vector space and a
basis of that normed vector space.
11 Show that the collection A = {kZ : k = 2, 3, 4, . . .} of subsets of Z satisfies
the hypothesis of Zorn’s Lemma (6.60).
12 Prove that every linearly independent family in a vector space can be extended
to a basis of the vector space.
13 Suppose V is a normed vector space, U is a subspace of V, and ψ : U → R is a
bounded linear functional. Prove that ψ has a unique extension to a bounded
linear functional ϕ on V with k ϕk = kψk if and only if
 
sup −kψk k f + hk − ψ( f ) = inf kψk k g + hk − ψ( g)
f ∈U g ∈U

for every h ∈ V \ U.
Section 6D Linear Functionals 183

14 Show that there exists a linear functional ϕ : `∞ → F such that

| ϕ( a1 , a2 , . . .)| ≤ k( a1 , a2 , . . .)k∞

for all ( a1 , a2 , . . .) ∈ `∞ and

ϕ( a1 , a2 , . . .) = lim ak
k→∞

for all ( a1 , a2 , . . .) ∈ `∞ such that the limit above on the right exists.
15 Suppose B is an open ball in a normed vector space V such that 0 ∈
/ B. Prove
that there exists ϕ ∈ V 0 such that

Re ϕ( f ) > 0

for all f ∈ B.
16 Show that the dual space of each infinite-dimensional normed vector space is
infinite-dimensional.

A normed vector space is called separable if it has a countable subset whose clo-
sure equals the whole space.
17 Suppose V is a separable normed vector space. Explain how the Hahn–Banach
Theorem (6.69) for V can be proved without using any results (such as Zorn’s
Lemma) that depend upon the Axiom of Choice.
18 Suppose V is a normed vector space such that the dual space V 0 is a separable
Banach space. Prove that V is separable.
19 Prove that the dual of the Banach space C ([0, 1]) is not separable; here the norm
on C ([0, 1]) is defined by k f k = sup| f |.
[0, 1]

The double dual space of a normed vector space is defined to be the dual space of
the dual space. If V is a normed vector space, then the double dual space of V is
0
denoted by V 00 ; thus V 00 = (V 0 ) . The norm on V 00 is defined to be the norm it
0
receives as the dual space of V .
20 Define Φ : V → V 00 by
(Φ f )( ϕ) = ϕ( f )
for f ∈ V and ϕ ∈ V 0.
Show that kΦ f k = k f k for every f ∈ V.
[The map Φ defined above is called the canonical isometry of V into V 00.]
21 Suppose V is an infinite-dimensional normed vector space. Show that there is a
convex subset U of V such that U = V and such that the complement V \ U is
also a convex subset of V with V \ U = V.
[See 8.25 for the definition of a convex set. This exercise should stretch your
geometric intuition because this behavior cannot happen in finite dimensions.]
184 Chapter 6 Banach Spaces

6E Consequences of Baire’s Theorem


This section focuses on several important
The result here called Baire’s
results about Banach spaces that depend
Theorem is often called the Baire
upon Baire’s Theorem. This result was
Category Theorem. This book uses
first proved by René-Louis Baire (1874–
the shorter name of this result
1932) as part of his 1899 doctoral disserta-
because we do not need the
tion at École Normale Supérieure (Paris).
categories introduced by Baire.
Even though our interest lies primar-
Furthermore, the use of the word
ily in applications to Banach spaces, the
category in this context can be
proper setting for Baire’s Theorem is the
confusing because Baire’s
more general context of complete metric
categories have no connection with
spaces.
the category theory that developed
decades after Baire’s work.
Baire’s Theorem
We begin with some key topological notions.

6.74 Definition interior

Suppose U is a subset of a metric space V. The interior of U, denoted int U, is


the set of f ∈ U such that some open ball of V centered at f with positive radius
is contained in U.

You should verify the following elementary facts about the interior.

• The interior of each subset of a metric space is open.


• The interior of a subset U of a metric space V is the largest open subset of V
contained in U.

6.75 Definition dense

A subset U of a metric space V is called dense in V if U = V.

For example, Q and R \ Q are both dense in R, where R has its standard metric
d( x, y) = | x − y|.
You should verify the following elementary facts about dense subsets.

• A subset U of a metric space V is dense in V if and only if every nonempty open


subset of V contains at least one element of U.
• A subset U of a metric space V has an empty interior if and only if V \ U is
dense in V.

The proof of the next result uses the following fact, which you should first prove:
If G is an open subset of a metric space V and f ∈ G, then there exists r > 0 such
that B( f , r ) ⊂ G.
Section 6E Consequences of Baire’s Theorem 185

6.76 Baire’s Theorem

(a) A complete metric space is not the countable union of closed subsets with
empty interior.
(b) The countable intersection of dense open subsets of a complete metric space
is nonempty.

Proof We will prove (b) and then use (b) to prove (a).
To prove (b), suppose (V, d) is a complete metric space T
and G1 , G2 , . . . is a
sequence of dense open subsets of V. We need to show that ∞ k=1 Gk 6 = ∅.
Let f 1 ∈ G1 and let r1 ∈ (0, 1) be such that B( f 1 , r1 ) ⊂ G1 . Now suppose
n ∈ Z+, and f 1 , . . . , f n and r1 , . . . , rn have been chosen such that

6.77 B ( f 1 , r1 ) ⊃ B ( f 2 , r2 ) ⊃ · · · ⊃ B ( f n , r n )

and

6.78 r j ∈ 0, 1j and B( f j , r j ) ⊂ Gj for j = 1, . . . , n.

Because B( f n , rn ) is an open subset of V and Gn+1 is dense in V, there exists


1
f n+1 ∈ B( f n , rn ) ∩ Gn+1 . Let rn+1 ∈ 0, n+1 be such that

B( f n+1 , rn+1 ) ⊂ B( f n , rn ) ∩ Gn+1 .

Thus we inductively construct a sequence f 1 , f 2 , . . . that satisfies 6.77 and 6.78 for
all n ∈ Z+.
If j ∈ Z+, then 6.77 and 6.78 imply that
1
6.79 f k ∈ B( f j , r j ) and d( f j , f k ) ≤ r j < j for all k > j.

Hence f 1 , f 2 , . . . is a Cauchy sequence. Because (V, d) is a complete metric space,


there exists f ∈ V such that limk→∞ f k = f .
Now 6.79 T∞
and 6.78 imply that for each
T∞
j ∈ Z+, we have f ∈ B( f j , r j ) ⊂ Gj .
Hence f ∈ k=1 Gk , which means that k=1 Gk is not the empty set, completing the
proof of (b).
To prove (a), suppose (V, d) is a complete metric space and F1 , F2 , . . . is a se-
quence of closed subsets of V with empty interior. Then V \ F1 , V \ F2 , . . . is a
sequence of dense open subsets of V. Now (b) implies that

\
∅ 6= (V \ Fk ).
k =1

Taking complements of both sides above, we conclude that



[
V 6= Fk ,
k =1

completing the proof of (a).


186 Chapter 6 Banach Spaces

Because [
R= {x}
x ∈R

and each set { x } has empty interior in R, Baire’s Theorem implies R is uncountable.
Thus we have yet another proof that R is uncountable, different than Cantor’s original
diagonal proof and different from the proof via measure theory (see 2.17).
The next result is another nice consequence of Baire’s Theorem.

6.80 the set of irrational numbers is not a countable union of closed sets

There does not exist a countable collection of closed subsets of R whose union
equals R \ Q.

Proof This will be a proof by contradiction. Suppose F1 , F2 , . . . is a countable


collection of closed subsets of R whose union equals R \ Q. Thus each Fk contains
no rational numbers, which implies that each Fk has empty interior. Now
[ ∞
 [ 
R= {r } ∪ Fk .
r ∈Q k =1

The equation above writes the complete metric space R as a countable union of
closed sets with empty interior, which contradicts Baire’s Theorem [6.76(a)]. This
contradiction completes the proof.

Open Mapping Theorem and Inverse Mapping Theorem


The next result shows that a surjective bounded linear map from one Banach space
onto another Banach space maps open sets to open sets. As shown in Exercises 10
and 11, this result can fail if the hypothesis that both spaces are Banach spaces is
weakened to allow either of the spaces to be a normed vector space.

6.81 Open Mapping Theorem

Suppose V and W are Banach spaces and T is a bounded linear map of V onto W.
Then T ( G ) is an open subset of W for every open subset G of V.

Proof Let B denote the open unit ball B(0, 1) = { f ∈ V : k f k < 1} of V. For any
open ball B( f , a) in V, the linearity of T implies that

T B( f , a) = T f + aT ( B).

Suppose G is an open subset of V. If f ∈ G, then there exists a > 0 such that


B( f , a) ⊂ G. If we can
 show that 0 ∈ int T ( B), then the equation above shows that
T f ∈ int T B( f , a) . This would imply that T ( G ) is an open subset of W. Thus to
complete the proof we need only show that T ( B) contains some open ball centered
at 0.
Section 6E Consequences of Baire’s Theorem 187

The surjectivity and linearity of T imply that



[ ∞
[
W= T (kB) = kT ( B).
k =1 k =1
S
Thus W = ∞ k=1 kT ( B ). Baire’s Theorem [6.76(a)] now implies that kT ( B ) has a
nonempty interior for some k ∈ Z+. The linearity of T allows us to conclude that
T ( B) has a nonempty interior.
Thus there exists g ∈ B such that Tg ∈ int T ( B). Hence

0 ∈ int T ( B − g) ⊂ int T (2B) = int 2T ( B).

Thus there exists r > 0 such that B(0, 2r ) ⊂ 2T ( B) [here B(0, 2r ) is the closed ball
in W centered at 0 with radius 2r]. Hence B(0, r ) ⊂ T ( B). The definition of what it
means to be in the closure of T ( B) [see 6.7] now shows that
h ∈ W and khk ≤ r and ε > 0 =⇒ ∃ f ∈ B such that kh − T f k < ε.
r
For arbitrary h 6= 0 in W, applying the result in the line above to khk
h shows that
khk
6.82 h ∈ W and ε > 0 =⇒ ∃ f ∈ r B such that kh − T f k < ε.

Now suppose g ∈ W and k gk < 1. Applying 6.82 with h = g and ε = 12 , we see


that
there exists f 1 ∈ 1r B such that k g − T f 1 k < 12 .
Now applying 6.82 with h = g − T f 1 and ε = 41 , we see that
1
there exists f 2 ∈ 2r B such that k g − T f 1 − T f 2 k < 14 .

Applying 6.82 again, this time with h = g − T f 1 − T f 2 and ε = 18 , we see that


1
there exists f 3 ∈ 4r B such that k g − T f 1 − T f 2 − T f 3 k < 18 .
Continue in this pattern, constructing a sequence f 1 , f 2 , . . . in V. Let

f = ∑ fk ,
k =1

where the infinite sum converges in V because


∞ ∞
1 2
∑ k fk k < ∑ 2 k −1 r
= ;
r
k =1 k =1

here we are using 6.41 (this is the place in the proof where we use the hypothesis that
V is a Banach space). The inequality displayed above shows that k f k < 2r .
Because
1
kg − T f1 − T f2 − · · · − T fn k < n
2
and because T is a continuous linear map, we have g = T f .
We have now shown that B(0, 1) ⊂ 2r T ( B). Thus 2r B(0, 1) ⊂ T ( B), completing
the proof.
188 Chapter 6 Banach Spaces

The next result provides the useful in-


The Open Mapping Theorem was
formation that if a bounded linear map
first proved by Banach and his
from one Banach space to another Banach
colleague Juliusz Schauder
space has an algebraic inverse (meaning
(1899–1943) in 1929–1930.
that the linear map is injective and surjec-
tive), then the inverse mapping is automatically bounded.

6.83 Bounded Inverse Theorem

Suppose V and W are Banach spaces and T is a one-to-one bounded linear map
from V onto W. Then T −1 is a bounded linear map from W onto V.

Proof The verification that T −1 is a linear map from W to V is left to the reader.
To prove that T −1 is bounded, suppose G is an open subset of V. Then
−1
( T −1 ) ( G ) = T ( G ).
By the Open Mapping Theorem (6.81), T ( G ) is an open subset of W. Thus the
equation above shows that the inverse image under the function T −1 of every open
set is open. By the equivalence of parts (a) and (c) of 6.11, this implies that T −1 is
continuous. Thus T −1 is a bounded linear map (by 6.48).

The result above shows that completeness for normed vector spaces sometimes
plays a role analogous to compactness for metric spaces (think of the theorem stating
that a continuous one-to-one function from a compact metric space onto another
compact metric space has an inverse that is also continuous).

Closed Graph Theorem


Suppose V and W are normed vector spaces. Then V × W is a vector space with
the natural operations of addition and scalar multiplication as defined in Exercise 10
in Section 6B. There are several natural norms on V × W that make V × W into a
normed vector space; the choice used in the next result seems to be the easiest. The
proof of the next result is left to the reader as an exercise.

6.84 product of Banach spaces

Suppose V and W are Banach spaces. Then V × W is a Banach space if given


the norm defined by
k( f , g)k = max{k f k, k gk}
for f ∈ V and g ∈ W. With this norm, a sequence ( f 1 , g1 ), ( f 2 , g2 ), . . . in
V × W converges to ( f , g) if and only if lim f k = f and lim gk = g.
k→∞ k→∞

The next result gives a terrific way to show that a linear map between Banach
spaces is bounded. The proof is remarkably clean because the hard work has been
done in the proof of the Open Mapping Theorem (which was used to prove the
Bounded Inverse Theorem).
Section 6E Consequences of Baire’s Theorem 189

6.85 Closed Graph Theorem

Suppose V and W are Banach spaces and T is a function from V to W. Then T


is a bounded linear map if and only if graph( T ) is a closed subspace of V × W.

Proof First suppose T is a bounded linear map. Suppose ( f 1 , T f 1 ), ( f 2 , T f 2 ), . . . is


a sequence in graph( T ) converging to ( f , g) ∈ V × W. Thus

lim f k = f and lim T f k = g.


k→∞ k→∞

Because T is continuous, the first equation above implies that limk→∞ T f k = T f ;


when combined with the second equation above this implies that g = T f . Thus
( f , g) = ( f , T f ) ∈ graph( T ), which implies that graph( T ) is closed, completing
the proof in one direction.
To prove the other direction, now suppose graph( T ) is a closed subspace of
V × W. Thus graph( T ) is a Banach space with the norm that it inherits from V × W
[from 6.84 and 6.16(b)]. Consider the linear map S : graph( T ) → V defined by

S( f , T f ) = f .

Then
kS( f , T f )k = k f k ≤ max{k f k, k T f k} = k( f , T f )k
for all f ∈ V. Thus S is a bounded linear map from graph( T ) onto V with kSk ≤ 1.
Clearly S is injective. Thus the Bounded Inverse Theorem (6.83) implies that S−1 is
bounded. Because S−1 : V → graph( T ) satisfies the equation S−1 f = ( f , T f ), we
have

k T f k ≤ max{k f k, k T f k}
= k( f , T f )k
= k S −1 f k
≤ k S −1 k k f k

for all f ∈ V. The inequality above implies that T is a bounded linear map with
k T k ≤ kS−1 k, completing the proof.

Principle of Uniform Boundedness


The next result states that a family of
The Principle of Uniform
bounded linear maps on a Banach space
Boundedness was proved in 1927 by
that is pointwise bounded is bounded in
Banach and Hugo Steinhaus
norm (which means that it is uniformly
(1887–1972). Steinhaus recruited
bounded as a collection of maps on the
Banach to advanced mathematics
unit ball). This result is sometimes called
after overhearing him discuss
the Banach–Steinhaus Theorem. Exercise
Lebesgue integration in a park.
17 is also sometimes called the Banach–
Steinhaus Theorem.
190 Chapter 6 Banach Spaces

6.86 Principle of Uniform Boundedness

Suppose V is a Banach space, W is a normed vector space, and A is a family of


bounded linear maps from V to W such that

sup{k T f k : T ∈ A} < ∞ for every f ∈ V.

Then
sup{k T k : T ∈ A} < ∞.

Proof Our hypothesis implies that



[
V= { f ∈ V : k T f k ≤ n for all T ∈ A} ,
n =1
| {z }
Vn

where Vn is defined by the expression above. Because each T ∈ A is continuous, Vn


is a closed subset of V for each n ∈ Z+. Thus Baire’s Theorem [6.76(a)] and the
equation above imply that there exist n ∈ Z+ and h ∈ Vand r > 0 such that

6.87 B(h, r ) ⊂ Vn .

Now suppose g ∈ V and k gk < 1. Thus rg + h ∈ B(h, r ). Hence if T ∈ A, then


6.87 implies k T (rg + h)k ≤ n, which implies that
T (rg + h) Th k T (rg + h)k k Thk n + k Thk

k Tgk = − ≤ + ≤ .
r r r r r
Thus
n + sup{k Thk : T ∈ A}
sup{k T k : T ∈ A} ≤ < ∞,
r
completing the proof.

EXERCISES 6E
1 Suppose U is a subset of a metric space V. Show that U is dense in V if and
only if every nonempty open subset of V contains at least one element of U.
2 Suppose U is a subset of a metric space V. Show that U has an empty interior if
and only if V \ U is dense in V.
3 Prove or give a counterexample: If V is a metric space and U, W are subsets
of V, then (int U ) ∪ (int W ) = int(U ∪ W ).
4 Prove or give a counterexample: If V is a metric space and U, W are subsets
of V, then (int U ) ∩ (int W ) = int(U ∩ W ).
Section 6E Consequences of Baire’s Theorem 191

5 Suppose
∞ 
[
1
X = {0} ∪ k
k =1
and d( x, y) = | x − y| for x, y ∈ X.
(a) Show that ( X, d) is a complete metric space.
(b) Each set of the form { x } for x ∈ X is a closed subset of R that has an
empty interior as a subset of R. Clearly X is a countable union of such sets.
Explain why this does not violate the statement of Baire’s Theorem that
a complete metric space is not the countable union of closed subsets with
empty interior.

6 Give an example of a metric space that is the countable union of closed subsets
with empty interior.
[This exercise shows that the completeness hypothesis in Baire’s Theorem cannot
be dropped.]
7 (a) Define f : R → R as follows:

0 if a is irrational,

f ( a) = 1 if a is rational and n is the smallest positive integer

n
such that a = mn for some integer m.
At which numbers in R is f continuous?
(b) Show that there does not exist a countable collection of open subsets of R
whose intersection equals Q.
(c) Show that there does not exist a function f : R → R such that f is continu-
ous at each element of Q and discontinuous at each element of R \ Q.
8 Suppose ( X, d) is a complete metric
T
space and G1 , G2 , . . . is a sequence of
dense open subsets of X. Prove that ∞k=1 Gk is a dense subset of X.
9 Prove that there does not exist an infinite-dimensional Banach space with a
countable basis.
[This exercise implies, for example, that there is not a norm that makes the
vector space of polynomials with coefficients in F into a Banach space.]
10 Give an example of a Banach space V, a normed vector space W, a bounded
linear map T of V onto W, and an open subset G of V such that T ( G ) is not an
open subset of W.
[This exercise shows that the hypothesis in the Open Mapping Theorem that
W is a Banach space cannot be relaxed to the hypothesis that W is a normed
vector space.]
11 Show that there exists a normed vector space V, a Banach space W, a bounded
linear map T of V onto W, and an open subset G of V such that T ( G ) is not an
open subset of W.
[This exercise shows that the hypothesis in the Open Mapping Theorem that V
is a Banach space cannot be relaxed to the hypothesis that V is a normed vector
space.]
192 Chapter 6 Banach Spaces

A linear map T : V → W from a normed vector space V to a normed vector space


W is called bounded below if there exists c ∈ (0, ∞) such that k f k ≤ ck T f k
for all f ∈ V.
12 Suppose T : V → W is a bounded linear map from a Banach space V to a
Banach space W. Prove that T is bounded below if and only if T is injective and
the range of T is a closed subspace of W.
13 Give an example of a Banach space V, a normed vector space W, and a one-to-
one bounded linear map T of V onto W such that T −1 is not a bounded linear
map of W onto V.
[This exercise shows that the hypothesis in the Bounded Inverse Theorem (6.83)
that W is a Banach space cannot be relaxed to the hypothesis that W is a
normed vector space.]
14 Show that there exists a normed space V, a Banach space W, and a one-to-one
bounded linear map T of V onto W such that T −1 is not a bounded linear map
of W onto V.
[This exercise shows that the hypothesis in the Bounded Inverse Theorem (6.83)
that V is a Banach space cannot be relaxed to the hypothesis that V is a normed
vector space.]
15 Prove 6.84.
16 Suppose V is a Banach space with norm k·k and that ϕ : V → F is a linear
functional. Define another norm k·k ϕ on V by
k f k ϕ = k f k + | ϕ( f )|.
Prove that if V is a Banach space with the norm k·k ϕ , then ϕ is a continuous
linear functional on V (with the original norm).
17 Suppose V is a Banach space, W is a normed vector space, and T1 , T2 , . . . is a
sequence of bounded linear maps from V to W such that limk→∞ Tk f exists for
each f ∈ V. Define T : V → W by
T f = lim Tk f
k→∞
for f ∈ V. Prove that T is a bounded linear map from V to W.
[This result states that the pointwise limit of a sequence of bounded linear maps
on a Banach space is a bounded linear map.]
18 Suppose V is a normed vector space and B is a subset V such that
sup| ϕ( f )| < ∞
f ∈B

for every ϕ ∈ V 0. Prove that supk f k < ∞.


f ∈B

19 Suppose T : V → W is a linear map from a Banach space V to a Banach space


W such that
ϕ ◦ T ∈ V 0 for all ϕ ∈ W 0.
Prove that T is a bounded linear map.
Chapter 7
L p Spaces

Fix a measure space ( X, S , µ) and a positive number p. We begin this chapter by


looking at the vector space of measurable functions f : X → F such that
Z
| f | p dµ < ∞.

Important results called Hölder’s inequality and Minkowski’s inequality help us


investigate this vector space. A useful class of Banach spaces appears when we
identify functions that differ only on a set of measure 0 and require p ≥ 1.

The main building of the Swiss Federal Institute of Technology (ETH Zürich).
Hermann Minkowski (1864–1909) taught at this university from 1896 to 1902.
During this time, Albert Einstein (1879–1955) was a student in several of
Minkowski’s mathematics classes. Minkowski later created mathematics that
helped explain Einstein’s special theory of relativity.
CC-BY-SA Roland zh

193
194 Chapter 7 L p Spaces

7A L p (µ)
Hölder’s Inequality
Our next major goal is to define an important class of vector spaces that generalize the
vector spaces L1 (µ) and `1 introduced in the last two bullet points of Example 6.32.
We begin this process with the definition below. The terminology p-norm introduced
below is convenient, even though it is not necessarily a norm.

7.1 Definition k f kp

Suppose that ( X, S , µ) is a measure space, 0 < p < ∞, and f : X → F is


S -measurable. Then the p-norm of f is denoted by k f k p and is defined by
Z 1/p
k f kp = | f | p dµ .

Also, k f k∞ , which is called the essential supremum of f , is defined by



k f k∞ = inf t > 0 : µ({ x ∈ X : | f ( x )| > t}) = 0 .

The exponent 1/p appears in the definition of the p-norm k f k p because we want
the equation kα f k p = |α| k f k p to hold for all α ∈ F.
For 0 < p < ∞, the p-norm k f k p does not change if f changes on a set of
µ-measure 0. By using the essential supremum rather than the supremum in the defi-
nition of k f k∞ , we arrange for the ∞-norm k f k∞ to enjoy this same property. Think
of k f k∞ as the smallest that you can make the supremum of | f | after modifications
on sets of measure 0.

7.2 Example p-norm for counting measure


Suppose µ is counting measure on Z+. If a = ( a1 , a2 , . . .) is a sequence in F and
0 < p < ∞, then
∞ 1/p
k ak p = ∑ | ak | p and k ak∞ = sup{| ak | : k ∈ Z+}.
k =1

Note that for counting measure, the essential supremum and the supremum are the
same because in this case there are no sets of measure 0 other than the empty set.

Now we can define our generalization of L1 (µ), which was defined in the second-
to-last bullet point of Example 6.32.

7.3 Definition L p (µ)

Suppose ( X, S , µ) is a measure space and 0 < p ≤ ∞. The Lebesgue space


L p (µ), sometimes denoted L p ( X, S , µ), is defined to be the set of S -measurable
functions f : X → F such that k f k p < ∞.
Section 7A L p (µ) 195

7.4 Example `p
When µ is counting measure on Z+, the set L p (µ) is often denoted by ` p (pro-
nounced little el-p). Thus if 0 < p < ∞, then

` p = {( a1 , a2 , . . .) : each ak ∈ F and ∑ | ak | p < ∞}
k =1

and
`∞ = {( a1 , a2 , . . .) : each ak ∈ F and sup | ak | < ∞}.
k ∈Z+

Inequality 7.5(a) below provides an easy proof that L p (µ) is closed under addition.
Soon we will prove Minkowski’s inequality (7.14), which provides an important
improvement of 7.5(a) when p ≥ 1 but is more complicated to prove.

7.5 L p (µ) is a vector space

Suppose ( X, S , µ) is a measure space and 0 < p < ∞. Then


p p p
(a) k f + gk p ≤ 2 p (k f k p + k gk p )

and

(b) kα f k p = |α| k f k p

for all f , g ∈ L p (µ) and all α ∈ F. Furthermore, with the usual operations of
addition and scalar multiplication of functions, L p (µ) is a vector space.

Proof Suppose f , g ∈ L p (µ). If x ∈ X, then


| f ( x ) + g( x )| p ≤ (| f ( x )| + | g( x )|) p
≤ (2 max{| f ( x )|, | g( x )|}) p
≤ 2 p (| f ( x )| p + | g( x )| p ).
Integrating both sides of the inequality above with respect to µ gives the desired
inequality
p p p
k f + gk p ≤ 2 p (k f k p + k gk p ).
This inequality implies that if k f k p < ∞ and k gk p < ∞, then k f + gk p < ∞. Thus
L p (µ) is closed under addition.
The proof that
kα f k p = |α| k f k p
follows easily from the definition of k·k p . This equality implies that L p (µ) is closed
under scalar multiplication.
Because L p (µ) contains the constant function 0 and is closed under addition and
scalar multiplication, L p (µ) is a subspace of F X and thus is a vector space.
196 Chapter 7 L p Spaces

What we call the dual exponent in the definition below is often called the conjugate
exponent or the conjugate index. However, the terminology dual exponent conveys
more meaning because of results (7.25 and 7.26) that we will see in the next section.

7.6 Definition dual exponent; p0

For 1 ≤ p ≤ ∞, the dual exponent of p is denoted by p0 and is the element of


[1, ∞] such that
1 1
+ 0 = 1.
p p

7.7 Example dual exponents

10 = ∞, ∞0 = 1, 20 = 2, 40 = 4/3, (4/3)0 = 4

The result below is a key tool in proving Hölder’s inequality (7.9).

7.8 Young’s inequality

Suppose 1 < p < ∞. Then


0
ap bp
ab ≤ + 0
p p

for all a ≥ 0 and b ≥ 0.

Proof Fix b > 0 and define a function


William Henry Young (1863–1942)
f : (0, ∞) → R by
published what is now called
0 Young’s inequality in 1912.
ap bp
f ( a) = + 0 − ab.
p p

Thus f 0 ( a) = a p−1 − b. Hence f is decreasing on the interval 0, b1/( p−1) and f is

increasing on the interval b1/( p−1) , ∞ . Thus f has a global minimum at b1/( p−1) .

A tiny bit of arithmetic [use p/( p − 1) = p0 ] shows that f b1/( p−1) = 0. Thus
f ( a) ≥ 0 for all a ∈ (0, ∞), which implies the desired inequality.

The important result below furnishes a key tool that is used in the proof of
Minkowski’s inequality (7.14).

7.9 Hölder’s inequality

Suppose ( X, S , µ) is a measure space, 1 ≤ p ≤ ∞, and f , h : X → F are


S -measurable. Then
k f h k1 ≤ k f k p k h k p 0 .
Section 7A L p (µ) 197

Proof Suppose 1 < p < ∞, leaving the cases p = 1 and p = ∞ as exercises for
the reader.
First consider the special case where k f k p = k hk p0 = 1. Young’s inequality (7.8)
tells us that 0
| f ( x )| p |h( x )| p
| f ( x )h( x )| ≤ +
p p0
for all x ∈ X. Integrating both sides of the inequality above with respect to µ shows
that k f hk1 ≤ 1 = k f k p khk p0 , completing the proof in this special case.
If k f k p = 0 or khk p0 = 0, then
Hölder’s inequality was proved in
k f hk1 = 0 and the desired inequal- 1889 by Otto Hölder (1859–1937).
ity holds. Similarly, if k f k p = ∞ or
khk p0 = ∞, then the desired inequality clearly holds. Thus we assume that
0 < k f k p < ∞ and 0 < khk p0 < ∞.
Now define S -measurable functions f 1 , h1 : X → F by
f h
f1 = and h1 = .
k f kp k h k p0
Then k f 1 k p = 1 and k h1 k p0 = 1. By the result for our special case, we have
k f 1 h1 k1 ≤ 1, which implies that k f hk1 ≤ k f k p khk p0 .

The next result gives a key containment among Lebesgue spaces with respect to a
finite measure. Note the crucial role that Hölder’s inequality plays in the proof.

7.10 Lq (µ) ⊂ L p (µ) if p < q and µ( X ) < ∞

Suppose ( X, S , µ) is a finite measure space and 0 < p < q < ∞. Then

k f k p ≤ µ( X )(q− p)/( pq) k f kq

for all f ∈ Lq (µ). Furthermore, Lq (µ) ⊂ L p (µ).


q
Proof Fix f ∈ Lq (µ). Let r = p. Thus r > 1. A short calculation shows that
q
r0= Now Hölder’s inequality (7.9) with p replaced by r and f replaced by | f | p
q− p .
and h replaced by the constant function 1 gives
Z Z 1/r Z 0 1/r0
| f | p dµ ≤ (| f | p )r dµ 1r dµ
Z  p/q
= µ( X )(q− p)/q | f |q dµ .

Now raise both sides of the inequality above to the power 1p , getting
Z 1/p Z 1/q
| f | p dµ ≤ µ( X )(q− p)/( pq) | f |q dµ ,

which is the desired inequality.


The inequality above shows that f ∈ L p (µ). Thus Lq (µ) ⊂ L p (µ).
198 Chapter 7 L p Spaces

7.11 Example L p ( E)
We adopt the common convention that if E is a Borel (or Lebesgue measurable)
subset of R and 0 < p ≤ ∞, then L p ( E) means L p (λ E ), where λ E denotes
Lebesgue measure λ restricted to the Borel (or Lebesgue measurable) subsets of R
that are contained in E.
With this convention, 7.10 implies that
if 0 < p < q < ∞, then Lq ([0, 1]) ⊂ L p ([0, 1]) and k f k p ≤ k f kq
for f ∈ Lq ([0, 1]). See Exercises 12 and 13 in this section for related results.

Minkowski’s Inequality
The next result is used as a tool to prove Minkowski’s inequality (7.14). Once again,
note the crucial role that Hölder’s inequality plays in the proof.

7.12 formula for k f k p

Suppose ( X, S , µ) is a measure space, 1 ≤ p < ∞, and f ∈ L p (µ). Then


n Z o
0
k f k p = sup f h dµ : h ∈ L p (µ) and khk p0 ≤ 1 .

Proof If k f k p = 0, then both sides of the equation in the conclusion of this result
equal 0. Thus we assume that k f k p 6= 0.
0
Hölder’s inequality (7.9) implies that if h ∈ L p (µ) and khk p0 ≤ 1, then
Z Z

f h dµ ≤ | f h| dµ ≤ k f k p k hk p0 ≤ k f k p .
 R 0
Thus sup f h dµ : h ∈ L p (µ) and khk p0 ≤ 1 ≤ k f k p .
To prove the inequality in the other direction, define h : X → F by
f ( x ) | f ( x )| p−2
h( x ) = p/p0
(set h( x ) = 0 when f ( x ) = 0).
k f kp
R p
Then f h dµ = k f k p and khk p0 = 1, as you should verify (use p − p0 = 1). Thus
 R 0
k f k p ≤ sup f h dµ : h ∈ L p (µ) and khk p0 ≤ 1 , as desired.

7.13 Example a point with infinite measure


Suppose X is a set with exactly one element b and µ is the measure such that
µ(∅) = 0 and µ({b}) = ∞. Then L1 (µ) consists only of the 0 function. Thus if
p = ∞ and f is the function whose value at b equals 1, then k f k∞ = 1 but the right
side of the equation in 7.12 equals 0. Thus 7.12 can fail when p = ∞.

Example 7.13 shows that we cannot take p = ∞ in 7.12. However, if µ is a


σ-finite measure, then 7.12 holds even when p = ∞ (see Exercise 9).
Section 7A L p (µ) 199

The next result, which is called Minkowski’s inequality, is an improvement for


p ≥ 1 of the inequality 7.5(a).

7.14 Minkowski’s inequality

Suppose ( X, S , µ) is a measure space, 1 ≤ p ≤ ∞, and f , g ∈ L p (µ). Then

k f + gk p ≤ k f k p + k gk p .

Proof Assume that 1 ≤ p < ∞ (the case p = ∞ is left as an exercise for the reader).
Inequality 7.5(a) implies that f + g ∈ L p (µ).
0
Suppose h ∈ L p (µ) and k hk p0 ≤ 1. Then
Z Z Z

( f + g)h dµ ≤ | f h| dµ + | gh| dµ ≤ (k f k p + k gk p )khk p0
≤ k f k p + k gk p ,
where the second inequality comes from Hölder’s inequality (7.9). Now take the
0
supremum of the left side of the inequality above over the set of h ∈ L p (µ) such
that khk p0 ≤ 1. By 7.12, we get k f + gk p ≤ k f k p + k gk p , as desired.

EXERCISES 7A
1 Suppose µ is a measure. Prove that
k f + gk∞ ≤ k f k∞ + k gk∞ and kα f k∞ = |α| k f k∞
for all f , g ∈ L∞ (µ) and all α ∈ F. Conclude that with the usual operations of
addition and scalar multiplication of functions, L∞ (µ) is a vector space.
2 Suppose a ≥ 0, b ≥ 0, and 1 < p < ∞. Prove that
0
ap bp
ab = + 0
p p
0
if and only if a p = b p [compare to Young’s inequality (7.8)].
3 Suppose a1 , . . . , an are nonnegative numbers. Prove that
( a1 + · · · + a n )5 ≤ n4 ( a1 5 + · · · + a n 5 ).
4 Prove Hölder’s inequality (7.9) in the cases p = 1 and p = ∞.
5 Suppose that ( X, S , µ) is a measure space, 1 < p < ∞, f ∈ L p (µ), and
0
h ∈ L p (µ). Prove that Hölder’s inequality (7.9) is an equality if and only if
there exist nonnegative numbers a and b, not both 0, such that
0
a| f ( x )| p = b|h( x )| p
for almost every x ∈ X.
200 Chapter 7 L p Spaces

6 Suppose ( X, S , µ) is a measure space, f ∈ L1 (µ), and h ∈ L∞ (µ). Prove that


k f hk1 = k f k1 khk∞ if and only if

|h( x )| = khk∞

for almost every x ∈ X such that f ( x ) 6= 0.


7 Suppose ( X, S , µ) is a measure space and f , h : X → F are S -measurable.
Prove that
k f h kr ≤ k f k p k h k q
1 1
for all positive numbers p, q, r such that p + q = 1r .

8 Suppose ( X, S , µ) is a measure space and n ∈ Z+. Prove that

k f 1 f 2 · · · f n k 1 ≤ k f 1 k p1 k f 2 k p2 · · · k f n k p n

for all positive numbers p1 , . . . , pn such that p1 + 1


p2 +···+ 1
pn = 1 and all
1
S -measurable functions f 1 , f 2 , . . . , f n : X → F.
9 Show that the formula in 7.12 holds for p = ∞ if µ is a σ-finite measure.
10 Suppose 0 < p < q ≤ ∞.
(a) Prove that ` p ⊂ `q .
(b) Prove that k( a1 , a2 , . . .)k p ≥ k( a1 , a2 , . . .)kq for every sequence a1 , a2 , . . .
of elements of F.
\
11 Show that ` p 6 = `1 .
p >1
\
12 Show that L p ([0, 1]) 6= L∞ ([0, 1]).
p<∞
[
13 Show that L p ([0, 1]) 6= L1 ([0, 1]).
p >1

14 Suppose p, q ∈ (0, ∞], with p 6= q. Prove that neither of the sets L p (R) and
Lq (R) is a subset of the other.
15 Show that there exists f ∈ L2 (R) such that f ∈
/ L p (R) for all p ∈ (0, ∞] \ {2}.

16 Suppose ( X, S , µ) is a finite measure space. Prove that

lim k f k p = k f k∞
p→∞

for every S -measurable function f : X → F.


17 Suppose µ is a measure, 0 < p ≤ ∞, and f ∈ L p (µ). Prove that for every
ε > 0, there exists a simple function g ∈ L p (µ) such that k f − gk p < ε.
[This exercise extends 3.44.]
Section 7A L p (µ) 201

18 Suppose 0 < p < ∞ and f ∈ L p (R). Prove that for every ε > 0, there exists a
step function g ∈ L p (R) such that k f − gk p < ε.
[This exercise extends 3.47.]
19 Suppose 0 < p < ∞ and f ∈ L p (R). Prove that for every ε > 0, there
exists a continuous function g : R → R such that k f − gk p < ε and the set
{ x ∈ R : g( x ) 6= 0} is bounded.
[This exercise extends 3.48.]
20 Suppose ( X, S , µ) is a measure space, 1 < p < ∞, and f , g ∈ L p (µ). Prove
that Minkowski’s inequality (7.14) is an equality if and only if there exist
nonnegative numbers a and b, not both 0, such that

a f ( x ) = bg( x )

for almost every x ∈ X.


21 Suppose ( X, S , µ) is a measure space and f , g ∈ L1 (µ). Prove that

k f + g k1 = k f k1 + k g k1

if and only if f ( x ) g( x ) ≥ 0 for almost every x ∈ X.


22 Suppose ( X, S , µ) and (Y, T , ν) are σ-finite measure spaces and 0 < p < ∞.
Prove that if f ∈ L p (µ × ν), then

[ f ] x ∈ L p (ν) for almost every x ∈ X

and
[ f ]y ∈ L p (µ) for almost every y ∈ Y,
where [ f ] x and [ f ]y are the cross sections of f as defined in 5.7.
23 Suppose 1 ≤ p < ∞ and f ∈ L p (R).
(a) For t ∈ R, define f t : R → R by f t ( x ) = f ( x − t). Prove that the function
t 7→ k f − f t k p is bounded and uniformly continuous on R.
(b) For t > 0, define f t : R → R by f t ( x ) = f (tx ). Prove that

limk f − f t k p = 0.
t →1

24 Suppose 1 ≤ p < ∞ and f ∈ L p (R). Prove that


Z b+t
1
lim | f − f (b)| p = 0
t ↓0 2t b−t

for almost every b ∈ R.


202 Chapter 7 L p Spaces

7B L p (µ)
Definition of L p (µ)
Suppose ( X, S , µ) is a measure space and 1 ≤ p ≤ ∞. If there exists a nonempty set
E ∈ S such that µ( E) = 0, then kχ Ek p = 0 even though χ E 6= 0; thus k·k p is not a
norm on L p (µ). The standard way to deal with this problem is to identify functions
that differ only on a set of µ-measure 0. To help make this process more rigorous, we
introduce the following definitions.

7.15 Definition Z (µ); f˜

Suppose ( X, S , µ) is a measure space and 0 < p ≤ ∞.


• Z (µ) denotes the set of S -measurable functions from X to F that equal 0
almost everywhere.
• For f ∈ L p (µ), let f˜ be the subset of L p (µ) defined by

f˜ = { f + z : z ∈ Z (µ)}.

The set Z (µ) is clearly closed under scalar multiplication. Also, Z (µ) is closed
under addition because the union of two sets with µ-measure 0 is a set with µ-
measure 0. Thus Z (µ) is a subspace of L p (µ), as we had noted in the third bullet
point of Example 6.32.
Note that if f , F ∈ L p (µ), then f˜ = F̃ if and only if f ( x ) = F ( x ) for almost
every x ∈ X.

7.16 Definition L p (µ)

Suppose µ is a measure and 0 < p ≤ ∞.


• Let L p (µ) denote the collection of subsets of L p (µ) defined by

L p (µ) = { f˜ : f ∈ L p (µ)}.

• For f˜, g̃ ∈ L p (µ) and α ∈ F, define f˜ + g̃ and α f˜ by

f˜ + g̃ = ( f + g)˜ and α f˜ = (α f )˜.

The last bullet point in the definition above requires a bit of care to verify that it
makes sense. The potential problem is that if Z (µ) 6= {0}, then f˜ is not uniquely
represented by f . Thus suppose f , F, g, G ∈ L p (µ) and f˜ = F̃ and g̃ = G̃. For
the definition of addition in L p (µ) to make sense, we must verify that ( f + g)˜ =
( F + G )˜. This verification is left to the reader, as is the similar verification that the
scalar multiplication defined in the last bullet point above makes sense.
You might want to think of elements of L p (µ) as equivalence classes of functions
in L p (µ), where two functions are equivalent if they agree almost everywhere.
Section 7B L p (µ) 203

Mathematicians often pretend that ele-


Note the subtle typographic
ments of L p (µ) are functions, where two
difference between L p (µ) and
functions are considered to be equal if
L p (µ). An element of the
they differ only on a set of µ-measure 0.
calligraphic L p (µ) is a function; an
This fiction is harmless provided that the
element of the italic L p (µ) is a set
operations you perform with such “func-
of functions, any two of which agree
tions” produce the same results if the func-
almost everywhere.
tions are changed on a set of measure 0.

7.17 Definition k·k p on L p (µ)

Suppose µ is a measure and 0 < p ≤ ∞. Define k·k p on L p (µ) by

k f˜k p = k f k p

for f ∈ L p (µ).

Note that if f , F ∈ L p (µ) and f˜ = F̃, then k f k p = k F k p . Thus the definition


above makes sense.
In the result below, the addition and scalar multiplication on L p (µ) come from
7.16 and the norm comes from 7.17.

7.18 L p (µ) is a normed vector space

Suppose µ is a measure and 1 ≤ p ≤ ∞. Then L p (µ) is a vector space and k·k p


is a norm on L p (µ).

The proof of the result above is left to the reader, who will surely use Minkowski’s
inequality (7.14) to verify the triangle inequality. Note that the additive identity of
L p (µ) is 0̃, which equals Z (µ).
For readers familiar with quotients of
If µ is counting measure on Z+, then
vector spaces: you may recognize that
L p (µ) is the quotient space L p (µ) = L p (µ) = ` p
L p ( µ ) / Z ( µ ).
because counting measure has no
For readers who want to learn about quo- sets of measure 0 other than the
tients of vector spaces: see a textbook for empty set.
a second course in linear algebra.
In the next definition, note that if E is a Borel set then 2.95 implies L p ( E) using
Borel measurable functions equals L p ( E) using Lebesgue measurable functions.

7.19 Definition L p ( E) for E ⊂ R

If E is a Borel (or Lebesgue measurable) subset of R and 0 < p ≤ ∞, then


L p ( E) means L p (λ E ), where λ E denotes Lebesgue measure λ restricted to the
Borel (or Lebesgue measurable) subsets of R that are contained in E.
204 Chapter 7 L p Spaces

L p (µ) Is a Banach Space


The proof of the next result does all the hard work we need to prove that L p (µ) is a
Banach space. However, we state the next result in terms of L p (µ) instead of L p (µ)
so that we can work with genuine functions. Moving to L p (µ) will then be easy (see
7.24).

7.20 Cauchy sequences in L p (µ) converge

Suppose ( X, S , µ) is a measure space and 1 ≤ p ≤ ∞. Suppose f 1 , f 2 , . . . is a


sequence of functions in L p (µ) such that for every ε > 0, there exists n ∈ Z+
such that
k f j − fk k p < ε
for all j ≥ n and k ≥ n. Then there exists f ∈ L p (µ) such that

lim k f k − f k p = 0.
k→∞

Proof The case p = ∞ is left as an exercise for the reader. Thus assume 1 ≤ p < ∞.
It suffices to show that limm→∞ k f km − f k p = 0 for some f ∈ L p (µ) and some
subsequence f k1 , f k2 , . . . (see Exercise 14 of Section 6A, whose proof does not require
the positive definite property of a norm).
Thus dropping to a subsequence (but not relabeling) and setting f 0 = 0, we can
assume that

∑ k f k − f k−1 k p < ∞.
k =1
Define functions g1 , g2 , . . . and g from X to [0, ∞] by
m ∞
gm ( x ) = ∑ | f k (x) − f k−1 (x)| and g(x) = ∑ | f k (x) − f k−1 (x)|.
k =1 k =1
Minkowski’s inequality (7.14) implies that
m
7.21 k gm k p ≤ ∑ k f k − f k −1 k p .
k =1
Clearly limm→∞ gm ( x ) = g( x ) for every x ∈ X. Thus the Monotone Convergence
Theorem (3.11) and 7.21 imply
Z Z ∞ p
p
7.22 g dµ = lim gm p dµ ≤ ∑ k f k − f k−1 k p < ∞.
m→∞
k =1
Thus g( x ) < ∞ for almost every x ∈ X.
Because every infinite series of real numbers that converges absolutely also con-
verges, for almost every x ∈ X we can define f ( x ) by
∞  m 
f (x) = ∑ m→∞

f k ( x ) − f k−1 ( x ) = lim f k ( x ) − f k−1 ( x ) = lim f m ( x ).
m→∞
k =1 k =1
In particular, limm→∞ f m ( x ) exists for almost every x ∈ X. Define f ( x ) to be 0 for
those x ∈ X for which the limit does not exist.
Section 7B L p (µ) 205

We now have a function f that is the pointwise limit (almost everywhere) of


f 1 , f 2 , . . .. The definition of f shows that | f ( x )| ≤ g( x ) for almost every x ∈ X.
Thus 7.22 shows that f ∈ L p (µ).
To show that limk→∞ k f k − f k p = 0, suppose ε > 0 and let n ∈ Z+ be such that
k f j − f k k p < ε for all j ≥ n and k ≥ n. Suppose k ≥ n. Then
Z 1/p
k fk − f k p = | f k − f | p dµ
Z 1/p
≤ lim inf | f k − f j | p dµ
j→∞

= lim infk f k − f j k p
j→∞

≤ ε,

where the second line above comes from Fatou’s Lemma (Exercise 17 in Section 3A).
Thus limk→∞ k f k − f k p = 0, as desired.

The proof that we have just completed contains within it the proof of a useful
result that is worth stating separately. A sequence can converge in p-norm without
converging pointwise anywhere (see, for example, Exercise 12). However, the next
result guarantees that some subsequence converges pointwise almost everywhere.

7.23 convergent sequences in L p have pointwise convergent subsequences

Suppose ( X, S , µ) is a measure space and 1 ≤ p ≤ ∞. Suppose f ∈ L p (µ)


and f 1 , f 2 , . . . is a sequence of functions in L p (µ) such that lim k f k − f k p = 0.
k→∞
Then there exists a subsequence f k1 , f k2 , . . . such that

lim f km ( x ) = f ( x )
m→∞

for almost every x ∈ X.

Proof Suppose f k1 , f k2 , . . . is a subsequence such that



∑ k f km − f km−1 k p < ∞.
m =2

An examination of the proof of 7.20 shows that lim f km ( x ) = f ( x ) for almost


m→∞
every x ∈ X.

7.24 L p (µ) is a Banach space

Suppose µ is a measure and 1 ≤ p ≤ ∞. Then L p (µ) is a Banach space.

Proof This result follows immediately from 7.20 and the appropriate definitions.
206 Chapter 7 L p Spaces

Duality
Recall that the dual space of a normed vector space V is denoted by V 0 and is defined
to be the Banach space of bounded linear functionals on V (see 6.71).
In the statement and proof of the next result, an element of an L p space is denoted
by a symbol that makes it look like a function rather than like a collection of functions
that agree except on a set of measure 0. However, because integrals and L p -norms
are unchanged when functions change only on a set of measure 0, this notational
convenience causes no problems.

0 0
7.25 natural map of L p (µ) into L p (µ) preserves norms

0
Suppose µ is a measure and 1 < p ≤ ∞. For h ∈ L p (µ), define ϕh : L p (µ) → F
by Z
ϕh ( f ) = f h dµ.
0 0
Then h 7→ ϕh is a one-to-one linear map from L p (µ) to L p (µ) . Furthermore,
0
k ϕh k = khk p0 for all h ∈ L p (µ).
0
Proof Suppose h ∈ L p (µ) and f ∈ L p (µ). Then Hölder’s inequality (7.9) tells us
that f h ∈ L1 (µ) and that
k f h k1 ≤ k h k p 0 k f k p .
Thus ϕh , as defined above, is a bounded linear map from L p (µ) to F. Also, the map
0 0
h 7→ ϕh is clearly a linear map of L p (µ) into L p (µ) . Now 7.12 (with the roles of
p and p0 reversed) shows that

k ϕh k = sup{| ϕh ( f )| : f ∈ L p (µ) and k f k p ≤ 1} = khk p0.


0
If h1 , h2 ∈ L p (µ) and ϕh1 = ϕh2 , then

kh1 − h2 k p0 = k ϕh1 −h2 k = k ϕh1 − ϕh2 k = k0k = 0,


0 0
which implies h1 = h2 . Thus h 7→ ϕh is a one-to-one map from L p (µ) to L p (µ) .

The result in 7.25 fails for some measures µ if p = 1. However, if µ is a σ-finite


measure, then 7.25 holds even if p = 1 (see Exercise 14).
0
Is the range of the map h 7→ ϕh in 7.25 all of L p (µ) ? The next result provides
an affirmative answer to this question in the special case of ` p for 1 ≤ p < ∞.
We will deal with this question for more general measures later (see 9.42; also see
Exercise 25 in Section 8B).
When thinking of ` p as a normed vector space, as in the next result, unless stated
otherwise you should always assume that the norm on ` p is the usual norm k·k p that
is associated with L p (µ), where µ is counting measure on Z+. In other words, if
1 ≤ p < ∞, then
∞ 1/p
k( a1 , a2 , . . .)k p = ∑ | ak | p .
k =1
Section 7B L p (µ) 207

0
7.26 dual space of ` p can be identified with ` p
0
Suppose 1 ≤ p < ∞. For b = (b1 , b2 , . . .) ∈ ` p , define ϕb : ` p → F by

ϕb ( a) = ∑ a k bk ,
k =1

0
where a = ( a1 , a2 , . . .). Then b 7→ ϕb is a one-to-one linear map from ` p onto
0 0
` p . Furthermore, k ϕb k = kbk p0 for all b ∈ ` p .

Proof For k ∈ Z+, let ek ∈ ` p be the sequence in which each term is 0 except that
the term is 1; thus ek = (0, . . . , 0, 1, 0, . . .).
kth
0
Suppose ϕ ∈ ` p . Define a sequence b = (b1 , b2 , . . .) of numbers in F by
bk = ϕ ( e k ) .
Suppose a = ( a1 , a2 , . . .) ∈ ` p . Then

a= ∑ ak ek ,
k =1

where the infinite sum converges in the norm of ` p (the proof would fail here if we
allowed p to be ∞). Because ϕ is a bounded linear functional on ` p , applying ϕ to
both sides of the equation above shows that

ϕ( a) = ∑ a k bk .
k =1
0
We still need to prove that b ∈ ` p . To do this, for n ∈ Z+ let µn be counting
measure on {1, 2, . . . , n}. We can think of L p (µn ) as a subspace of ` p by identi-
fying each ( a1 , . . . , an ) ∈ L p (µn ) with ( a1 , . . . , an , 0, 0, . . .) ∈ ` p . Restricting the
linear functional ϕ to L p (µn ) gives the linear functional on L p (µn ) that satisfies the
following equation:
n
ϕ | L p ( µ n ) ( a1 , . . . , a n ) = ∑ a k bk .
k =1
Now 7.25 (also see Exercise 14 for the case where p = 1) gives
k(b1 , . . . , bn )k p0 = k ϕ| L p (µn ) k
≤ k ϕ k.
Because limn→∞ k(b1 , . . . , bn )k p0 = kbk p0 , the inequality above implies the in-
0
equality kbk p0 ≤ k ϕk. Thus b ∈ ` p , which implies that ϕ = ϕb , completing the
proof.

The previous result does not hold when p = ∞. In other words, the dual space
of `∞ cannot be identified with `1 . However, see Exercise 15, which shows that the
dual space of a natural subspace of `∞ can be identified with `1 .
208 Chapter 7 L p Spaces

EXERCISES 7B
1 Suppose n > 1 and 0 < p < 1. Prove that if k·k is defined on Fn by
1/p
k( a1 , . . . , an )k = | a1 | p + · · · + | an | p ,

then k·k is not a norm on Fn .


2 (a) Suppose 1 ≤ p < ∞. Prove that there is a countable subset of ` p whose
closure equals ` p .
(b) Prove that there does not exist a countable subset of `∞ whose closure
equals `∞ .
3 (a) Suppose 1 ≤ p < ∞. Prove that there is a countable subset of L p (R)
whose closure equals L p (R).
(b) Prove that there does not exist a countable subset of L∞ (R) whose closure
equals L∞ (R).
4 Suppose ( X, S , µ) is a σ-finite measure space and 1 ≤ p ≤ ∞. Prove that
if f : X → F is an S -measurable function such that f h ∈ L1 (µ) for every
0
h ∈ L p (µ), then f ∈ L p (µ).
5 (a) Prove that if µ is a measure, 1 < p < ∞, and f , g ∈ L p (µ) are such that

f + g

k f k p = k gk p = ,
2 p

then f = g.
(b) Give an example to show that the result in part (a) can fail if p = 1.
(c) Give an example to show that the result in part (a) can fail if p = ∞.
6 Suppose ( X, S , µ) is a measure space and 0 < p < 1. Show that
p p p
k f + gk p ≤ k f k p + k gk p
for all S -measurable functions f , g : X → F.
7 Prove that L p (µ), with addition and scalar multiplication as defined in 7.16 and
norm defined as in 7.17, is a normed vector space. In other words, prove 7.18.
8 Prove 7.20 for the case p = ∞.
9 Prove that 7.20 also holds for p ∈ (0, 1).
10 Prove that 7.23 also holds for p ∈ (0, 1).
11 Suppose 1 ≤ p ≤ ∞. Prove that

{( a1 , a2 , . . .) ∈ ` p : ak 6= 0 for every k ∈ Z+}


is not an open subset of ` p .
Section 7B L p (µ) 209

12 Show that there exists a sequence f 1 , f 2 , . . . of functions in L1 ([0, 1]) such that
lim k f k k1 = 0 but
k→∞
sup{ f k ( x ) : k ∈ Z+ } = ∞
for every x ∈ [0, 1].
[This exercise shows that the conclusion of 7.23 cannot be improved to conclude
that limk→∞ f k ( x ) = f ( x ) for almost every x ∈ X.]
13 Suppose ( X, S , µ) is a measure space, 1 ≤ p ≤ ∞, f ∈ L p (µ), and f 1 , f 2 , . . .
is a sequence in L p (µ) such that limk→∞ k f k − f k p = 0. Show that if
g : X → F is a function such that limk→∞ f k ( x ) = g( x ) for almost every
x ∈ X, then f ( x ) = g( x ) for almost every x ∈ X.
14 (a) Give an example of a measure µ such that 7.25 fails for p = 1.
(b) Show that if µ is a σ-finite measure, then 7.25 holds for p = 1.
15 Let
c0 = {( a1 , a2 , . . .) ∈ `∞ : lim ak = 0}.
k→∞
Give c0 the norm that it inherits as a subspace of `∞ .
(a) Prove that c0 is a Banach space.
(b) Prove that the dual space of c0 can be identified with `1 .

16 Suppose 1 ≤ p ≤ 2.
(a) Prove that if w, z ∈ C, then

|w + z| p + |w − z| p |w + z| p + |w − z| p
≤ |w| p + |z| p ≤ .
2 2 p −1

(b) Prove that if µ is a measure and f , g ∈ L p (µ), then


p p p p
k f + gk p + k f − gk p p p k f + gk p + k f − gk p
≤ k f k p + k gk p ≤ .
2 2 p −1

17 Suppose 2 ≤ p < ∞.
(a) Prove that if w, z ∈ C, then

|w + z| p + |w − z| p |w + z| p + |w − z| p
p 1
≤ |w| p + |z| p ≤ .
2 − 2

(b) Prove that if µ is a measure and f , g ∈ L p (µ), then


p p p p
k f + gk p + k f − gk p p p k f + gk p + k f − gk p
p 1
≤ k f k p + k gk p ≤ .
2 − 2

[The inequalities in the two previous exercises are called Clarkson’s inequalities.
They were discovered by James Clarkson in 1936.]
210 Chapter 7 L p Spaces

18 Suppose ( X, S , µ) is a measure space, 1 ≤ p, q ≤ ∞, and h : X → F is an


S -measurable function such that h f ∈ Lq (µ) for every f ∈ L p (µ). Prove that
f 7→ h f is a continuous linear map from L p (µ) to Lq (µ).

A Banach space is called reflexive if the canonical isometry of the Banach space
into its double dual space is surjective (see Exercise 20 in Section 6D for the defi-
nitions of the double dual space and the canonical isometry).
19 Prove that if 1 < p < ∞, then ` p is reflexive.
20 Prove that `1 is not reflexive.
21 Show that with the natural identifications, the canonical isometry of c0 into its
double dual space is the inclusion map of c0 into `∞ (see Exercise 15 for the
definition of c0 and an identification of its dual space).
22 Suppose 1 ≤ p < ∞ and V, W are Banach spaces. Show that V × W is a
Banach space if the norm on V × W is defined by
1/p
k( f , g)k = k f k p + k gk p

for f ∈ V and g ∈ W.
Chapter 8
Hilbert Spaces

Normed vector spaces and Banach spaces, which were introduced in Chapter 6,
capture the notion of distance. In this chapter we introduce inner product spaces,
which capture the notion of angle. The concept of orthogonality, which corresponds
to right angles in the familiar context of R2 or R3 , plays a particularly important role
in inner product spaces.
Just as a Banach space is defined to be a normed vector space in which every
Cauchy sequence converges, a Hilbert space is defined to be an inner product space
that is a Banach space. Hilbert spaces are named in honor of David Hilbert (1862–
1943), who helped develop parts of the theory in the early twentieth century.
In this chapter, we will see a clean description of the bounded linear functionals
on a Hilbert space. We will also see that every Hilbert space has an orthonormal
basis, which make Hilbert spaces look much like standard Euclidean spaces but with
infinite sums replacing finite sums.

The Mathematical Institute at the University of Göttingen, Germany. This building


was opened in 1930, when Hilbert was near the end of his career there. Other
prominent mathematicians who taught at the University of Göttingen and made
major contributions to mathematics include Richard Courant (1888–1972), Richard
Dedekind (1831–1916), Gustav Lejeune Dirichlet (1805–1859), Carl Friedrich
Gauss (1777–1855), Hermann Minkowski (1864–1909), Emmy Noether (1882–1935),
and Bernhard Riemann (1826–1866).
CC-BY-SA Daniel Schwen

211
212 Chapter 8 Hilbert Spaces

8A Inner Product Spaces


Inner Products
If p = 2, then the dual exponent p0 also equals 2. In this special case Hölder’s
inequality (7.9) implies that if µ is a measure, then
Z

f g dµ ≤ k f k2 k gk2

for all f , gR∈ L2 (µ). Thus we can associate with each pair of functions f , g ∈ L2 (µ)
a number f g dµ. An inner product is almost a generalization of this pairing, with a
slight twist to get a closer connection to the L2 (µ)-norm.
If g = f and F = R, then the left side of the inequality above is k f k22 . However,
if g = f and F = C, then the left side of the inequality above need not equal k f k22 .
Instead, we should take g = f to get k f k22 above.
RThe observations above suggest that we should consider the pairing that takes f , g
to f g dµ. Then pairing f with itself gives k f k22 .
Now we are ready R to define inner products, which abstract the key properties of
the pairing f , g 7→ f g dµ on L2 (µ), where µ is a measure.

8.1 Definition inner product; inner product space

An inner product on a vector space V is a function that takes each ordered pair
f , g of elements of V to a number h f , gi ∈ F and has the following properties:
• positivity
h f , f i ∈ [0, ∞) for all f ∈ V;
• definiteness
h f , f i = 0 if and only if f = 0;
• linearity in first slot
h f + g, hi = h f , hi + h g, hi and hα f , gi = αh f , gi for all f , g, h ∈ V and
all α ∈ F;
• conjugate symmetry
h f , gi = h g, f i for all f , g ∈ V.
A vector space with an inner product on it is called an inner product space. The
terminology real inner product space indicates that F = R; the terminology
complex inner product space indicates that F = C.

If F = R, then the complex conjugate above can be ignored and the conjugate
symmetry property above can be rewritten more simply as h f , gi = h g, f i for all
f , g ∈ V.
Although most mathematicians define an inner product as above, many physicists
use a definition that requires linearity in the second slot instead of the first slot.
Section 8A Inner Product Spaces 213

8.2 Example inner product spaces

• For n ∈ Z+, define an inner product on Fn by

h( a1 , . . . , an ), (b1 , . . . , bn )i = a1 b1 + · · · + an bn

for ( a1 , . . . , an ), (b1 , . . . , bn ) ∈ Fn . When thinking of Fn as an inner product


space, we always mean this inner product unless the context indicates some other
inner product.
• Define an inner product on `2 by

h( a1 , a2 , . . .), (b1 , b2 , . . .)i = ∑ a k bk
k =1

for ( a1 , a2 , . . .), (b1 , b2 , . . .) ∈ `2 . Hölder’s inequality (7.9), as applied to count-


ing measure on Z+ and taking p = 2, implies that the infinite sum above
converges absolutely and hence converges to an element of F. When thinking
of `2 as an inner product space, we always mean this inner product unless the
context indicates some other inner product.
• Define an inner product on C ([0, 1]), which is the vector space of continuous
functions from [0, 1] to F, by
Z 1
h f , gi = fg
0

for f , g ∈ C ([0, 1]). The definiteness requirement for an inner product is


R1
satisfied because if f : [0, 1] → F is a continuous function such that 0 | f |2 = 0,
then the function f is identically 0.
• Suppose ( X, S , µ) is a measure space. Define an inner product on L2 (µ) by
Z
h f , gi = f g dµ

for f , g ∈ L2 (µ). Hölder’s inequality (7.9) with p = 2 implies that the integral
above makes sense as an element of F. When thinking of L2 (µ) as an inner
product space, we always mean this inner product unless the context indicates
some other inner product.
Here we use L2 (µ) rather than L2 (µ) because the definiteness requirement fails
on L2 (µ) if there exist nonempty sets E ∈ S such that µ( E) = 0 (consider
hχ E, χ Ei to see the problem).
The first two bullet points in this example are special cases of L2 (µ), taking µ to
be counting measure on either {1, . . . , n} or Z+.
214 Chapter 8 Hilbert Spaces

As we will see, even though the main examples of inner product spaces are L2 (µ)
spaces, working with the inner product structure is often cleaner and simpler than
working with measures and integrals.

8.3 basic properties of an inner product

Suppose V is an inner product space. Then

(a) h0, gi = h g, 0i = 0 for every g ∈ V;


(b) h f , g + hi = h f , gi + h f , hi for all f , g, h ∈ V;
(c) h f , αgi = αh f , gi for all α ∈ F and f , g ∈ V.

Proof

(a) For g ∈ V, the function f 7→ h f , gi is a linear map from V to F. Because


every linear map takes 0 to 0, we have h0, gi = 0. Now the conjugate symmetry
property of an inner product implies that
h g, 0i = h0, gi = 0 = 0.

(b) Suppose f , g, h ∈ V. Then

h f , g + hi = h g + h, f i = h g, f i + hh, f i = h g, f i + hh, f i = h f , gi + h f , hi.

(c) Suppose α ∈ F and f , g ∈ V. Then

h f , αgi = hαg, f i = αh g, f i = αh g, f i = αh f , gi,


as desired.

If F = R, then parts (b) and (c) of 8.3 imply that for f ∈ V, the function
g 7→ h f , gi is a linear map from V to R. However, if F = C and f 6= 0, then
the function g 7→ h f , gi is not a linear map from V to C because of the complex
conjugate in part (c) of 8.3.

Cauchy–Schwarz Inequality and Triangle Inequality


Now we can define the norm associated with each inner product. We use the word
norm (which will turn out to be correct) even though it is not yet clear that all the
properties required of a norm are satisfied.

8.4 Definition norm associated with an inner product; k·k

Suppose V is an inner product space. For f ∈ V, define the norm of f , denoted


k f k, by q
k f k = h f , f i.
Section 8A Inner Product Spaces 215

8.5 Example norms on inner product spaces


In each of the following examples, the inner product is the standard inner product
as defined in Example 8.2.
• If n ∈ Z+ and ( a1 , . . . , an ) ∈ Fn , then
q
k( a1 , . . . , an )k = | a1 |2 + · · · + | an |2 .

Thus the norm on Fn associated with the standard inner product is the usual
Euclidean norm.
• If ( a1 , a2 , . . .) ∈ `2 , then
 ∞ 1/2
k( a1 , a2 , . . .)k = ∑ | a k |2 .
k =1

Thus the norm associated with the inner product on `2 is just the standard norm
k·k2 on `2 as defined in Example 7.2.
• If µ is a measure and f ∈ L2 (µ), then
Z 1/2
kfk = | f |2 dµ .

Thus the norm associated with the inner product on L2 (µ) is just the standard
norm k·k2 on L2 (µ) as defined in 7.17.

The definition of an inner product (8.1) implies that if V is an inner product space
and f ∈ V, then
• k f k ≥ 0;
• k f k = 0 if and only if f = 0.
The proof of the next result illustrates a frequently used property of the norm on
an inner product space: working with the square of the norm is often easier than
working directly with the norm.

8.6 homogeneity of the norm

Suppose V is an inner product space, f ∈ V, and α ∈ F. Then

k α f k = | α | k f k.

Proof We have

kα f k2 = hα f , α f i = αh f , α f i = ααh f , f i = |α|2 k f k2 .

Taking square roots now gives the desired equality.


216 Chapter 8 Hilbert Spaces

The next definition plays a crucial role in the study of inner product spaces.

8.7 Definition orthogonal

Two elements of an inner product space are called orthogonal if their inner
product equals 0.

In the definition above, the order of the two elements of the inner product space
does not matter because h f , gi = 0 if and only if h g, f i = 0. Instead of saying that
f and g are orthogonal, sometimes we say that f is orthogonal to g.

8.8 Example orthogonal elements of an inner product space


• In C3 , (2, 3, 5i ) and (6, 1, −3i ) are orthogonal because
h(2, 3, 5i ), (6, 1, −3i )i = 2 · 6 + 3 · 1 + 5i · (3i ) = 12 + 3 − 15 = 0.

• The elements of L2 (−π, π ] represented by sin(3t) and cos(8t) are orthogo-
nal because
Z π h cos(5t) cos(11t) it=π
sin(3t) cos(8t) dt = − = 0,
−π 10 22 t=−π

where dt denotes integration with respect to Lebesgue measure on (−π, π ].

Exercise 8 asks you to prove that if a and b are nonzero elements in R2 , then

h a, bi = k ak kbk cos θ,
where θ is the angle between a and b (thinking of a as the vector whose initial point is
the origin and whose end point is a, and similarly for b). Thus two elements of R2 are
orthogonal if and only if the cosine of the angle between them is 0, which happens if
and only if the vectors are perpendicular in the usual sense of plane geometry. Thus
you can think of the word orthogonal as a fancy word meaning perpendicular.

Law professor Richard Friedman presenting a case before the U.S.


Supreme Court in 2010:

Mr. Friedman: I think that issue is entirely orthogonal to the issue here
because the Commonwealth is acknowledging—
Chief Justice Roberts: I’m sorry. Entirely what?
Mr. Friedman: Orthogonal. Right angle. Unrelated. Irrelevant.
Chief Justice Roberts: Oh.
Justice Scalia: What was that adjective? I liked that.
Mr. Friedman: Orthogonal.
Chief Justice Roberts: Orthogonal.
Mr. Friedman: Right, right.
Justice Scalia: Orthogonal, ooh. (Laughter.)
Justice Kennedy: I knew this case presented us a problem. (Laughter.)
Section 8A Inner Product Spaces 217

The next theorem is over 2500 years old, although it was not originally stated in
the context of inner product spaces.

8.9 Pythagorean Theorem

Suppose f and g are orthogonal elements of an inner product space. Then

k f + g k2 = k f k2 + k g k2 .

Proof We have
k f + gk2 = h f + g, f + gi
= h f , f i + h f , gi + h g, f i + h g, gi
= k f k2 + k g k2 ,
as desired.

Exercise 3 shows that whether or not the converse of the Pythagorean Theorem
holds depends upon whether F = R or F = C.
Suppose f and g are elements of an inner product space V,
with g 6= 0. Frequently it is useful to write f as some number c
times g plus an element h of V that is orthogonal to g. The figure
here suggests that such a decomposition should be possible. To
find the appropriate choice for c, note that if f = cg + h for
some c ∈ F and some h ∈ V with hh, gi = 0, then we must
have
h f , gi = hcg + h, gi = ck gk2 ,
h f , gi Here
which implies that c = , which then implies that
k g k2 f = cg + h,
h f , gi where h is
h= f− g. Hence we are led to the following result.
k g k2 orthogonal to g.

8.10 orthogonal decomposition

Suppose f and g are elements of an inner product space, with g 6= 0. Then there
exists h ∈ V such that
h f , gi
hh, gi = 0 and f = g + h.
k g k2

h f , gi
Proof Set h = f − g. Then
k g k2
D h f , gi E h f , gi
hh, gi = f − 2
g, g = h f , gi − h g, gi = 0,
k gk k g k2
giving the first equation in the conclusion. The second equation in the conclusion
follows immediately from the definition of h.
218 Chapter 8 Hilbert Spaces

The orthogonal decomposition 8.10 is the main ingredient in our proof of the next
result, which is one of the most important inequalities in mathematics.

8.11 Cauchy–Schwarz inequality

Suppose f and g are elements of an inner product space. Then

|h f , gi| ≤ k f k k gk,

with equality if and only if one of f , g is a scalar multiple of the other.

Proof If g = 0, then both sides of the desired inequality equal 0. Thus we can
assume g 6= 0. Consider the orthogonal decomposition
h f , gi
f = g+h
k g k2
given by 8.10, where h is orthogonal to g. The Pythagorean Theorem (8.9) implies

2
h f , g i 2
kfk = g + k h k2
k g k2
|h f , gi|2
= + k h k2
k g k2
|h f , gi|2
8.12 ≥ .
k g k2
Multiplying both sides of this inequality by k gk2 and then taking square roots gives
the desired inequality.
The proof above shows that the Cauchy–Schwarz inequality is an equality if and
only if 8.12 is an equality. This happens if and only if h = 0. But h = 0 if and only
if f is a scalar multiple of g (see 8.10). Thus the Cauchy–Schwarz inequality is an
equality if and only if f is a scalar multiple of g or g is a scalar multiple of f (or
both; the phrasing has been chosen to cover cases in which either f or g equals 0).

8.13 Example Cauchy–Schwarz inequality for Fn


Applying the Cauchy–Schwarz inequality with the standard inner product on Fn
to (| a1 |, . . . , | an |) and (|b1 |, . . . , |bn |) gives the inequality
q q
| a1 b1 | + · · · + | an bn | ≤ | a1 |2 + · · · + | an |2 |b1 |2 + · · · + |bn |2

for all ( a1 , . . . , an ), (b1 , . . . , bn ) ∈ Fn .


Thus we have a new and clean proof
The inequality in this example was
of Hölder’s inequality (7.9) for the spe-
first proved by Cauchy in 1821.
cial case where µ is counting measure on
{1, . . . , n} and p = p0 = 2.
Section 8A Inner Product Spaces 219

8.14 Example Cauchy–Schwarz inequality for L2 (µ)


Suppose µ is a measure and f , g ∈ L2 (µ). Applying the Cauchy–Schwarz
inequality with the standard inner product on L2 (µ) to | f | and | g| gives the inequality
Z Z 1/2 Z 1/2
| f g| dµ ≤ | f |2 dµ | g|2 dµ .

The inequality above is equivalent to


In 1859 Viktor Bunyakovsky
Hölder’s inequality (7.9) for the special
0 (1804–1889), who had been
case where p = p = 2. However,
Cauchy’s student in Paris, first
the proof of the inequality above via the
proved integral inequalities like the
Cauchy–Schwarz inequality still depends
one above. Similar discoveries by
upon Hölder’s inequality to show that the
Hermann Schwarz (1843–1921) in
definition of the standard inner product
1885 attracted more attention and
on L2 (µ) makes sense. See Exercise 18
led to the name of this inequality.
in this section for a derivation of the in-
equality above that is truly independent of Hölder’s inequality.

If we think of the norm determined by an


inner product as a length, then the triangle in-
equality has the geometric interpretation that the
length of each side of a triangle is less than the
sum of the lengths of the other two sides.

8.15 triangle inequality

Suppose f and g are elements of an inner product space. Then

k f + g k ≤ k f k + k g k,

with equality if and only if one of f , g is a nonnegative multiple of the other.

Proof We have
k f + gk2 = h f + g, f + gi
= h f , f i + h g, gi + h f , gi + h g, f i
= h f , f i + h g, gi + h f , gi + h f , gi
= k f k2 + k gk2 + 2 Reh f , gi
8.16 ≤ k f k2 + k gk2 + 2|h f , gi|
8.17 ≤ k f k2 + k g k2 + 2k f k k g k
= (k f k + k gk)2 ,

where 8.17 follows from the Cauchy–Schwarz inequality (8.11). Taking square roots
of both sides of the inequality above gives the desired inequality.
220 Chapter 8 Hilbert Spaces

The proof above shows that the triangle inequality is an equality if and only if we
have equality in 8.16 and 8.17. Thus we have equality in the triangle inequality if
and only if
8.18 h f , g i = k f k k g k.
If one of f , g is a nonnegative multiple of the other, then 8.18 holds, as you should
verify. Conversely, suppose 8.18 holds. Then the condition for equality in the Cauchy–
Schwarz inequality (8.11) implies that one of f , g is a scalar multiple of the other.
Clearly 8.18 forces the scalar in question to be nonnegative, as desired.

Applying the previous result to the inner product space L2 (µ), where µ is a
measure, gives a new proof of Minkowski’s inequality (7.14) for the case p = 2.
Now we can prove that what we have been calling a norm on an inner product
space is indeed a norm.

8.19 k·k is a norm

Suppose V is an inner product space and k f k is defined as usual by


q
kfk = hf, fi

for f ∈ V. Then k·k is a norm on V.

Proof The definition of an inner product implies that k·k satisfies the positive defi-
nite requirement for a norm. The homogeneity and triangle inequality requirements
for a norm are satisfied because of 8.6 and 8.15.
The next result has the geometric in-
terpretation that the sum of the squares
of the lengths of the diagonals of a
parallelogram equals the sum of the
squares of the lengths of the four sides.

8.20 parallelogram equality

Suppose f and g are elements of an inner product space. Then

k f + g k2 + k f − g k2 = 2k f k2 + 2k g k2 .

Proof We have
k f + gk2 + k f − gk2 = h f + g, f + gi + h f − g, f − gi
= k f k2 + k gk2 + h f , gi + h g, f i
+ k f k2 + k gk2 − h f , gi − h g, f i
= 2k f k2 + 2k g k2 ,
as desired.
Section 8A Inner Product Spaces 221

EXERCISES 8A
1 Let V denote the vector space of bounded continuous functions from R to F.
Let r1 , r2 , . . . be a list of the rational numbers. For f , g ∈ V, define

f (r k ) g (r k )
h f , gi = ∑ 2k
.
k =1

Show that h·, ·i is an inner product on V.


2 Prove that if f , g ∈ L2 (µ), where µ is a σ-finite measure, then
Z Z
1
k f k2 k gk2 − |h f , gi|2 = | f ( x ) g(y) − g( x ) f (y)|2 dµ(y) dµ( x ).
2

3 Suppose f and g are elements of an inner product space and

k f + g k2 = k f k2 + k g k2 .

(a) Prove that if F = R, then f and g are orthogonal.


(b) Give an example to show that if F = C, then f and g can satisfy the
equation above without being orthogonal.

4 Find a, b ∈ R3 such that a is a scalar multiple of (1, 6, 3), b is orthogonal to


(1, 6, 3), and (5, 4, −2) = a + b.
5 Prove that  
1 1 1 1
16 ≤ ( a + b + c + d) + + +
a b c d
for all positive numbers a, b, c, d, with equality if and only if a = b = c = d.
6 Prove that the square of the average of each finite list of real numbers containing
at least two distinct real numbers is less than the average of the squares of the
numbers in that list.
7 Suppose f and g are elements of an inner product space and k f k ≤ 1 and
k gk ≤ 1. Prove that
q q
1 − k f k2 1 − k gk2 ≤ 1 − |h f , gi|.

8 Suppose a and b are nonzero elements of R2 . Prove that

h a, bi = k ak kbk cos θ,

where θ is the angle between a and b (thinking of a as the vector whose initial
point is the origin and whose end point is a, and similarly for b).
Hint: Draw the triangle formed by a, b, and a − b; then use the law of cosines.
222 Chapter 8 Hilbert Spaces

9 The angle between two vectors (thought of as arrows with initial point at the
origin) in R2 or R3 can be defined geometrically. However, geometry is not as
clear in Rn for n > 3. Thus the angle between two nonzero vectors a, b ∈ Rn
is defined to be
h a, bi
arccos ,
k ak kbk
where the motivation for this definition comes from the previous exercise. Ex-
plain why the Cauchy–Schwarz inequality is needed to show that this definition
makes sense.
10 (a) Suppose f and g are elements of a real inner product space. Prove that f
and g have the same norm if and only if f + g is orthogonal to f − g.
(b) Use part (a) to show that the diagonals of a parallelogram are perpendicular
to each other if and only if the parallelogram is a rhombus.
11 Suppose f and g are elements of an inner product space. Prove that k f k = k gk
if and only if ks f + tgk = kt f + sgk for all s, t ∈ R.
12 Suppose f and g are elements of an inner product space and k f k = k gk = 1
and h f , gi = 1. Prove that f = g.
13 Suppose f and g are elements of a real inner product space. Prove that

k f + g k2 − k f − g k2
h f , gi = .
4
14 Suppose f and g are elements of a complex inner product space. Prove that

k f + gk2 − k f − gk2 + k f + igk2 i − k f − igk2 i


h f , gi = .
4
15 Suppose f , g, h are elements of an inner product space. Prove that

k h − f k2 + k h − g k2 k f − g k2
kh − 21 ( f + g)k2 = − .
2 4
16 Prove that a norm satisfying the parallelogram equality comes from an inner
product. In other words, show that if V is a normed vector space whose norm
k·k satisfies the parallelogram equality, then there is an inner product h·, ·i on
V such that k f k = h f , f i1/2 for all f ∈ V.
17 Let λ denote Lebesgue measure on [1, ∞).
(a) Prove that if f : [1, ∞) → [0, ∞) is Borel measurable, then
Z ∞ 2 Z ∞ 2
f ( x ) dλ( x ) ≤ x2 f ( x ) dλ( x ).
1 1

(b) Describe the set of Borel measurable functions f : [1, ∞) → [0, ∞) such
that the inequality in part (a) is an equality.
Section 8A Inner Product Spaces 223

18 Suppose µ is a measure. For f , g ∈ L2 (µ), define h f , gi by


Z
h f , gi = f g dµ.

(a) Using the inequality



| f ( x ) g( x )| ≤ 1
2 | f ( x )|2 + | g( x )|2 ,

verify that the integral above makes sense and the map sending f , g to h f , gi
defines an inner product on L2 (µ) (without using Hölder’s inequality).
(b) Show that the Cauchy–Schwarz inequality implies that

k f g k1 ≤ k f k2 k g k2

for all f , g ∈ L2 (µ) (again, without using Hölder’s inequality).

19 Suppose V1 , . . . , Vm are inner product spaces. Show that the equation

h( f 1 , . . . , f m ), ( g1 , . . . , gm )i = h f 1 , g1 i + · · · + h f m , gm i

defines an inner product on V1 × · · · × Vm .


[Each of the inner product spaces V1 , . . . , Vm may have a different inner product,
even though the same inner product notation is used on all these spaces.]
20 Suppose V is an inner product space. Make V × V an inner product space
as in the exercise above. Prove that the function that takes an ordered pair
( f , g) ∈ V × V to the inner product h f , gi ∈ F is a continuous function from
V × V to F.
21 Suppose 1 ≤ p ≤ ∞.
(a) Show the norm on ` p comes from an inner product if and only if p = 2.
(b) Show the norm on L p (R) comes from an inner product if and only if p = 2.

22 Use inner products to prove Apollonius’s identity:


In a triangle with sides of length a, b, and c, let d
be the length of the line segment from the midpoint
of the side of length c to the opposite vertex. Then

a2 + b2 = 21 c2 + 2d2 .
224 Chapter 8 Hilbert Spaces

8B Orthogonality
Orthogonal Projections
The previous section developed inner product spaces following a standard linear
algebra approach. Linear algebra focuses mainly on finite-dimensional vector spaces.
Many interesting results about infinite-dimensional inner product spaces require an
additional hypothesis, which we now introduce.

8.21 Definition Hilbert space

A Hilbert space is an inner product space that is a Banach space with the norm
determined by the inner product.

8.22 Example Hilbert spaces

• Suppose µ is a measure. Then L2 (µ) with its usual inner product is a Hilbert
space (by 7.24).
• As a special case of the first bullet point, if n ∈ Z+ then taking µ to be counting
measure on {1, . . . , n} shows that Fn with its usual inner product is a Hilbert
space.
• As another special case of the first bullet point, taking µ to be counting measure
on Z+ shows that `2 with its usual inner product is a Hilbert space.
• Every closed subspace of a Hilbert space is a Hilbert space [by 6.16(b)].

8.23 Example not Hilbert spaces

• The inner product space `1 , where h( a1 , a2 , . . .), (b1 , b2 , . . .)i = ∑∞


k=1 ak bk , is
not a Hilbert space because the associated norm is not complete on `1 .
• The inner product space C ([0, 1]) of continuous F-valued functions on the inter-
R1
val [0, 1], where h f , gi = 0 f g, is not a Hilbert space because the associated
norm is not complete on C ([0, 1]).

The next definition makes sense in the context of normed vector spaces.

8.24 Definition distance from a point to a set

Suppose U is a nonempty subset of a normed vector space V and f ∈ V. The


distance from f to U, denoted distance( f , U ), is defined by

distance( f , U ) = inf{k f − gk : g ∈ U }.

Notice that distance( f , U ) = 0 if and only if f ∈ U.


Section 8B Orthogonality 225

8.25 Definition convex set

• A subset of a vector space is called convex if the subset contains the line
segment connecting each pair of points in it.

• More precisely, suppose V is a vector space and U ⊂ V. Then U is called


convex if

(1 − t) f + tg ∈ U for all t ∈ [0, 1] and all f , g ∈ U.

Convex subset of R2 . Nonconvex subset of R2 .

8.26 Example convex sets

• Every subspace of a vector space is convex, as you should verify.


• If V is a normed vector space, f ∈ V, and r > 0, then the open ball centered at
f with radius r is convex, as you should verify.

The next example shows that the distance from an element of a Banach space to a
closed subspace is not necessarily attained by some element of the closed subspace.
After this example, we will prove that this behavior cannot happen in a Hilbert space.

8.27 Example no closest element to a closed subspace of a Banach space


In the Banach space C ([0, 1]) with norm k gk = sup| g|, let
[0, 1]
n Z 1 o
U= g ∈ C ([0, 1]) : g = 0 and g(1) = 0 .
0

Then U is a closed subspace of C ([0, 1]).


Let f ∈ C ([0, 1]) be defined by f ( x ) = 1 − x. For k ∈ Z+, let

1 xk x−1
gk ( x ) = −x+ + .
2 2 k+1

Then gk ∈ U and limk→∞ k f − gk k = 12 , which implies that distance( f , U ) ≤ 12 .


R1
If g ∈ U, then 0 ( f − g) = 12 and ( f − g)(1) = 0. These conditions imply that
k f − gk > 12 .
Thus distance( f , U ) = 12 but there does not exist g ∈ U such that k f − gk = 21 .
226 Chapter 8 Hilbert Spaces

In the next result, we use for the first time the hypothesis that V is a Hilbert space.

8.28 distance to a closed convex set is attained in a Hilbert space

• The distance from an element of a Hilbert space to a nonempty closed convex


set is attained by a unique element of the nonempty closed convex set.
• More specifically, suppose V is a Hilbert space, f ∈ V, and U is a nonempty
closed convex subset of V. Then there exists a unique g ∈ U such that

k f − gk = distance( f , U ).

Proof First we prove the existence of an element of U that attains the distance to f .
To do this, suppose g1 , g2 , . . . is a sequence of elements of U such that
8.29 lim k f − gk k = distance( f , U ).
k→∞

Then for j, k ∈ Z+ we have


k g j − gk k2 = k( f − gk ) − ( f − g j )k2

= 2k f − gk k2 + 2k f − g j k2 − k2 f − ( gk + g j )k2
gk + g j
2
= 2k f − gk k2 + 2k f − g j k2 − 4 f −
2
2
8.30 ≤ 2k f − gk k2 + 2k f − g j k2 − 4 distance( f , U ) ,
where the second equality comes from the parallelogram equality (8.20) and the
last line holds because the convexity of U implies that ( gk + g j )/2 ∈ U. Now the
inequality above and 8.29 imply that g1 , g2 , . . . is a Cauchy sequence. Thus there
exists g ∈ V such that
8.31 lim k gk − gk = 0.
k→∞
Because U is a closed subset of V and each gk ∈ U, we know that g ∈ U. Now 8.29
and 8.31 imply that
k f − gk = distance( f , U ),
which completes the existence proof of the existence part of this result.
To prove the uniqueness part of this result, suppose g and g̃ are elements of U
such that
8.32 k f − gk = k f − g̃k = distance( f , U ).
Then
2
k g − g̃k2 ≤ 2k f − gk2 + 2k f − g̃k2 − 4 distance( f , U )
8.33 = 0,
where the first line above follows from 8.30 (with g j replaced by g and gk replaced
by g̃) and the last line above follows from 8.32. Now 8.33 implies that g = g̃,
completing the proof of uniqueness.
Section 8B Orthogonality 227

Example 8.27 showed that the existence part of the previous result can fail in a
Banach space. Exercise 13 shows that the uniqueness part can also fail in a Banach
space. These observations highlight the advantages of working in a Hilbert space.

8.34 Definition orthogonal projection; PU

Suppose U is a nonempty closed convex subset of a Hilbert space V. The


orthogonal projection of V onto U is the function PU : V → V defined by setting
PU ( f ) equal to the unique element of U that is closest to f .

The definition above makes sense because of 8.28. We will often use the notation
PU f instead of PU ( f ). To test your understanding of the definition above, make sure
that you can show that if U is a nonempty closed convex subset of a Hilbert space V,
then
• PU f = f if and only if f ∈ U;
• PU ◦ PU = PU .

8.35 Example orthogonal projection onto closed unit ball


Suppose U is the closed unit ball { g ∈ V : k gk ≤ 1} in a Hilbert space V. Then

f
 if k f k ≤ 1,
PU f = f

 if k f k > 1,
kfk
as you should verify.

8.36 Example orthogonal projection onto a closed subspace


Suppose U is the closed subspace of `2 consisting of the elements of `2 whose
even coordinates are all 0:

U = {( a1 , 0, a3 , 0, a5 , 0, . . .) : each ak ∈ F and ∑ |a2k−1 |2 < ∞}.
k =1

Then for b = (b1 , b2 , b3 , b4 , b5 , b6 , . . .) ∈ `2 , we have

PU b = (b1 , 0, b3 , 0, b5 , 0, . . .),

as you should verify.


Note that in this example the function PU is a linear map from `2 to `2 (unlike the
behavior in Example 8.35).
Also, notice that b − PU b = (0, b2 , 0, b4 , 0, b6 , . . .) and thus b − PU b is orthogonal
to every element of U.

The next result shows that the properties stated in the last two paragraphs of the
example above hold whenever U is a closed subspace of a Hilbert space.
228 Chapter 8 Hilbert Spaces

8.37 orthogonal projection onto closed subspace

Suppose U is a closed subspace of a Hilbert space V and f ∈ V. Then

(a) f − PU f is orthogonal to g for every g ∈ U;


(b) if h ∈ U and f − h is orthogonal to g for every g ∈ U, then h = PU f ;
(c) PU : V → V is a linear map;
(d) k PU f k ≤ k f k, with equality if and only if f ∈ U.

Proof The figure below illustrates (a). To prove (a), suppose g ∈ U. Then for all
α ∈ F we have
k f − PU f k2 ≤ k f − PU f + αgk2
= h f − PU f + αg, f − PU f + αgi
= k f − PU f k2 + |α|2 k gk2 + 2 Re αh f − PU f , gi.
Let α = −th f − PU f , gi for t > 0. A tiny bit of algebra applied to the inequality
above implies
2|h f − PU f , gi|2 ≤ t|h f − PU f , gi|2 k gk2
for all t > 0. Thus h f − PU f , gi = 0, completing the proof of (a).
To prove (b), suppose h ∈ U and f − h is orthogonal to g for every g ∈ U. If
g ∈ U, then h − g ∈ U and hence f − h is orthogonal to h − g. Thus

k f − h k2 ≤ k f − h k2 + k h − g k2

= k( f − h) + (h − g)k2

= k f − g k2 ,
f − PU f is orthogonal to each element of U.
where the first equality above follows from the Pythagorean Theorem (8.9). Thus
k f − hk ≤ k f − gk
for all g ∈ U. Hence h is the element of U that minimizes the distance to f , which
implies that h = PU f , completing the proof of (b).
To prove (c), suppose f 1 , f 2 ∈ V. If g ∈ U, then (a) implies that h f 1 − PU f 1 , gi =
h f 2 − PU f 2 , gi = 0, and thus
h( f 1 + f 2 ) − ( PU f 1 + PU f 2 ), gi = 0.
The equation above and (b) now imply that
PU ( f 1 + f 2 ) = PU f 1 + PU f 2 .
The equation above and the equation PU (α f ) = αPU f for α ∈ F (whose proof is left
to the reader) show that PU is a linear map, proving (c).
The proof of (d) is left as an exercise for the reader.
Section 8B Orthogonality 229

Orthogonal Complements

8.38 Definition orthogonal complement; U ⊥

Suppose U is a subset of an inner product space V. The orthogonal complement


of U is denoted by U ⊥ and is defined by

U ⊥ = { h ∈ V : h g, hi = 0 for all g ∈ U }.

In other words, the orthogonal complement of a subset U of an inner product


space V is the set of elements of V that are orthogonal to every element of U.

8.39 Example orthogonal complement


Suppose U is the set of elements of `2 whose even coordinates are all 0:

U = {( a1 , 0, a3 , 0, a5 , 0, . . .) : each ak ∈ F and ∑ |a2k−1 |2 < ∞}.
k =1
Then U ⊥ is the set of elements of `2 whose odd coordinates are all 0:

U ⊥ = {0, a2 , 0, a4 , 0, a6 , . . .) : each ak ∈ F and ∑ |a2k |2 < ∞},
k =1
as you should verify.

8.40 properties of orthogonal complement

Suppose U is a subset of an inner product space V. Then

(a) U ⊥ is a closed subspace of V;


(b) U ∩ U ⊥ ⊂ {0};
(c) if W ⊂ U, then U ⊥ ⊂ W ⊥ ;

(d) U = U⊥;
(e) U ⊂ (U ⊥ )⊥ .

Proof To prove (a), suppose h1 , h2 , . . . is a sequence in U ⊥ that converges to some


h ∈ V. If g ∈ U, then
|h g, hi| = |h g, h − hk i| ≤ k gk kh − hk k for each k ∈ Z+ ;
hence h g, hi = 0, which implies that h ∈ U ⊥ . Thus U ⊥ is closed. The proof of (a)
is completed by showing that U ⊥ is a subspace of V, which is left to the reader.
To prove (b), suppose g ∈ U ∩ U ⊥ . Then h g, gi = 0, which implies that g = 0,
proving (b).
To prove (e), suppose g ∈ U. Thus h g, hi = 0 for all h ∈ U ⊥ , which implies that
g ∈ (U ⊥ )⊥ . Hence U ⊂ (U ⊥ )⊥ , proving (e).
The proofs of (c) and (d) are left to the reader.
230 Chapter 8 Hilbert Spaces

The results in the rest of this subsection have as a hypothesis that V is a Hilbert
space. These results do not hold when V is only an inner product space.

8.41 orthogonal complement of the orthogonal complement

Suppose U is a subspace of a Hilbert space V. Then

U = (U ⊥ ) ⊥ .

Proof Applying 8.40(a) to U ⊥ , we see that (U ⊥ )⊥ is a closed subspace of V. Now


taking closures of both sides of the inclusion U ⊂ (U ⊥ )⊥ [8.40(e)] shows that
U ⊂ (U ⊥ ) ⊥ .
To prove the inclusion in the other direction, suppose f ∈ (U ⊥ )⊥ . Because
f ∈ (U ⊥ )⊥ and PU f ∈ U ⊂ (U ⊥ )⊥ (by the previous paragraph), we see that

f − PU f ∈ (U ⊥ )⊥ .

Also,
f − PU f ∈ U ⊥
by 8.37(a) and 8.40(d). Hence

f − PU f ∈ U ⊥ ∩ (U ⊥ )⊥ .

Now 8.40(b) (applied to U ⊥ in place of U) implies that f − PU f = 0, which implies


that f ∈ U. Thus (U ⊥ )⊥ ⊂ U, completing the proof.

As a special case, the result above implies that if U is a closed subspace of a


Hilbert space V, then U = (U ⊥ )⊥ .
Another special case of the result above is sufficiently useful to deserve stating
separately, as we do in the next result.

8.42 necessary and sufficient condition for a subspace to be dense

Suppose U is a subspace of a Hilbert space V. Then

U = V if and only if U ⊥ = {0}.

Proof First suppose U = V. Then using 8.40(d), we have



U⊥ = U = V ⊥ = {0}.

To prove the other direction, now suppose U ⊥ = {0}. Then 8.41 implies that

U = (U ⊥ )⊥ = {0}⊥ = V,

completing the proof.


Section 8B Orthogonality 231

The next result states that if U is a


closed subspace of a Hilbert space V,
then V is the direct sum of U and U ⊥ ,
often written V = U ⊕ U ⊥ , although
we do not need to use this terminology
or notation further.
The key point to keep in mind is
that the next result shows that the pic-
ture here represents what happens in
general for a closed subspace U of a
Hilbert space V: every element of V
can be uniquely written as an element
of U plus an element of U ⊥ .

8.43 orthogonal decomposition

Suppose U is a closed subspace of a Hilbert space V. Then every element f ∈ V


can be uniquely written in the form

f = g + h,

where g ∈ U and h ∈ U ⊥ . Furthermore, g = PU f and h = f − PU f .

Proof Suppose f ∈ V. Then

f = PU f + ( f − PU f ),

where PU f ∈ U [by definition of PU f as the element of U that is closest to f ] and


f − PU f ∈ U ⊥ [by 8.37(a)]. Thus we have the desired decomposition of f as the
sum of an element of U and an element of U ⊥ .
To prove the uniqueness of this decomposition, suppose

f = g1 + h 1 = g2 + h 2 ,

where g1 , g2 ∈ U and h1 , h2 ∈ U ⊥ . Then g1 − g2 = h2 − h1 ∈ U ∩ U ⊥ , which


implies that g1 = g2 and h1 = h2 , as desired.

In the next definition, the function I depends upon the vector space V. Thus a
notation such as IV might be more precise. However, the domain of I should always
be clear from the context.

8.44 Definition identity map; I

Suppose V is a vector space. The identity map I is the linear map from V to V
defined by I f = f for f ∈ V.

The next result highlights the close relationship between orthogonal projections
and orthogonal complements.
232 Chapter 8 Hilbert Spaces

8.45 range and null space of orthogonal projections

Suppose U is a closed subspace of a Hilbert space V. Then

(a) range PU = U and null PU = U ⊥ ;


(b) range PU ⊥ = U ⊥ and null PU ⊥ = U;
(c) PU ⊥ = I − PU .

Proof The definition of PU f as the closest point in U to f implies range PU ⊂ U.


Because PU g = g for all g ∈ U, we also have U ⊂ range PU . Thus range PU = U.
If f ∈ null PU , then f ∈ U ⊥ [by 8.37(a)]. Thus null PU ⊂ U ⊥ . Conversely, if
f ∈ U ⊥ , then 8.37(b) (with h = 0) implies that PU f = 0; hence U ⊥ ⊂ null PU .
Thus null PU = U ⊥ , completing the proof of (a).
Replace U by U ⊥ in (a), getting range PU ⊥ = U ⊥ and null PU ⊥ = (U ⊥ )⊥ = U
(where the last equality comes from 8.41), completing the proof of (b).
Finally, if f ∈ U, then

PU ⊥ f = 0 = f − PU f = ( I − PU ) f ,

where the first equality above holds because null PU ⊥ = U [by (b)].
If f ∈ U ⊥ , then

PU ⊥ f = f = f − PU f = ( I − PU ) f ,

where the second equality above holds because null PU = U ⊥ [by (a)].
The last two displayed equations show that PU ⊥ and I − PU agree on U and agree
on U ⊥ . Because PU ⊥ and I − PU are both linear maps and because each element of
V equals some element of U plus some element of U ⊥ (by 8.43), this implies that
PU ⊥ = I − PU , completing the proof of (c).

8.46 Example PU ⊥ = I − PU
Suppose U is the closed subspace of L2 (R) defined by

U = { f ∈ L2 (R) : f ( x ) = 0 for almost every x < 0}.

Then, as you should verify,

U ⊥ = { f ∈ L2 (R) : f ( x ) = 0 for almost every x ≥ 0}.

Furthermore, you should also verify that if f ∈ L2 (R), then

PU f = f χ[0, ∞) and PU ⊥ f = f χ(−∞, 0).

Thus PU ⊥ f = f (1 − χ[0, ∞)) = ( I − PU ) f and hence PU ⊥ = I − PU , as asserted in


8.45(c).
Section 8B Orthogonality 233

Riesz Representation Theorem


Suppose h is an element of a Hilbert space V. Define ϕ : V → F by ϕ( f ) = h f , hi
for f ∈ V. The properties of an inner product imply that ϕ is a linear functional.
The Cauchy–Schwarz inequality (8.11) implies that | ϕ( f )| ≤ k f k khk for all f ∈ V,
which implies that ϕ is a bounded linear functional on V. The next result states that
every bounded linear functional on V arises in this fashion.
To motivate the proof of the next result, note that if ϕ is as in the paragraph above,
then null ϕ = {h}⊥ . Thus h ∈ (null ϕ)⊥ [by 8.40(e)]. Hence in the proof of the
next result, to find h we start with an element of (null ϕ)⊥ and then multiply it by a
scalar to make everything come out right.

8.47 Riesz Representation Theorem

Suppose ϕ is a bounded linear functional on a Hilbert space V. Then there exists


a unique h ∈ V such that
ϕ( f ) = h f , hi
for all f ∈ V. Furthermore, k ϕk = khk.

Proof If ϕ = 0, take h = 0. Thus we can assume ϕ 6= 0. Hence null ϕ is a closed


subspace of V not equal to V (see 6.52). The subspace (null ϕ)⊥ is not {0} (by
8.42). Thus there exists g ∈ (null ϕ)⊥ with k gk = 1. Let
h = ϕ( g) g.
Taking the norm of both sides of the equation above, we get khk = | ϕ( g)|. Thus
8.48 ϕ(h) = | ϕ( g)|2 = khk2 .
Now suppose f ∈ V. Then
   
ϕ( f ) ϕ( f )
h f , hi = f − h, h + h, h
k h k2 k h k2
 
ϕ( f )
8.49 = h, h
k h k2

= ϕ ( f ),
ϕ( f )
where 8.49 holds because f − h ∈ null ϕ (by 8.48) and h is orthogonal to all
k h k2
elements of null ϕ.
We have now proved the existence of h ∈ V such that ϕ( f ) = h f , hi for all
f ∈ V. To prove uniqueness, suppose h̃ ∈ V has the same property. Then
hh − h̃, h − h̃i = hh − h̃, hi − hh − h̃, h̃i = ϕ(h − h̃) − ϕ(h − h̃) = 0,
which implies that h = h̃, which proves uniqueness.
The Cauchy–Schwarz inequality implies that | ϕ( f )| = |h f , hi| ≤ k f k khk for
all f ∈ V, which implies that k ϕk ≤ khk. Because ϕ(h) = hh, hi = k hk2 , we also
have k ϕk ≥ khk. Thus k ϕk = khk, completing the proof.
234 Chapter 8 Hilbert Spaces

Suppose that µ is a measure and


Frigyes Riesz (1880–1956) proved
1 < p ≤ ∞. In 7.25 we considered the
0  0 8.47 in 1907.
natural map of L p (µ) into L p (µ) , and
we showed that this maps preserves norms. In the special case where p = p0 = 2,
the Riesz Representation Theorem (8.47) shows that this map is surjective. In other
words, if ϕ is a bounded linear functional on L2 (µ), then there exists h ∈ L2 (µ) such
that Z
ϕ( f ) = f h dµ

for all f ∈ L2 (µ) (take h to be the complex conjugate of the function given by 8.47).
Hence we can identify the dual of L2 (µ) with L2 (µ). In 9.42 we will deal with other
values of p. Also see Exercise 25 in this section.

EXERCISES 8B
1 Show that each of the inner product spaces in Example 8.23 is not a Hilbert
space.
2 Prove or disprove: The inner product space in Exercise 1 in Section 8A is a
Hilbert space.
3 Suppose V1 , V2 , . . . are Hilbert spaces. Let
n ∞ o
V = ( f 1 , f 2 , . . .) ∈ V1 × V2 × · · · : ∑ k f k k2 < ∞ .
k =1

Show that the equation



h( f 1 , f 2 , . . .), ( g1 , g2 , . . .)i = ∑ h f k , gk i
k =1

defines an inner product on V that makes V a Hilbert space.


[Each of the Hilbert spaces V1 , V2 , . . . may have a different inner product, even
though the same notation is used for the norm and inner product on all these
Hilbert spaces.]
4 Suppose V is a real Hilbert space. The complexification of V is the complex
vector space VC defined by VC = V × V, but we write a typical element of VC
as f + ig instead of ( f , g). Addition and scalar multiplication are defined on
VC by
( f 1 + ig1 ) + ( f 2 + ig2 ) = ( f 1 + f 2 ) + i ( g1 + g2 )
and
(α + βi )( f + ig) = (α f − βg) + (αg + β f )i
for f 1 , f 2 , f , g1 , g2 , g ∈ V and α, β ∈ R. Show that

h f 1 + ig1 , f 2 + ig2 i = h f 1 , f 2 i + h g1 , g2 i + (h g1 , f 2 i − h f 1 , g2 i)i


defines an inner product on VC that makes VC into a complex Hilbert space.
Section 8B Orthogonality 235

5 Prove that if V is a normed vector space, f ∈ V, and r > 0, then the open ball
B( f , r ) centered at f with radius r is convex.
6 (a) Suppose V is an inner product space and B is the open unit ball in V (thus
B = { f ∈ V : k f k < 1}). Prove that if U is a subset of V such that
B ⊂ U ⊂ B, then U is convex.
(b) Give an example to show that the result in part (a) can fail if the phrase
inner product space is replaced by Banach space.
7 Suppose V is a normed vector space and U is a closed subset of V. Prove that
U is convex if and only if
f +g
∈ U for all f , g ∈ U.
2
8 Prove that if U is a convex subset of a normed vector space, then U is also
convex.
9 Prove that if U is a convex subset of a normed vector space, then the interior of
U is also convex.
[The interior of U is the set { f ∈ U : B( f , r ) ⊂ U for some r > 0}.]
10 Suppose V is a Hilbert space, U is a nonempty closed convex subset of V, and
g ∈ U is the unique element of U with smallest norm (obtained by taking f = 0
in 8.28). Prove that
Reh g, hi ≥ k gk2
for all h ∈ U.
11 Suppose V is a Hilbert space. A closed half-space of V is a set of the form

{ g ∈ V : Reh g, hi ≥ c}
for some h ∈ V and some c ∈ R. Prove that every closed convex subset of V is
the intersection of all the closed half-spaces that contain it.
12 Give an example of a nonempty closed subset U of the Hilbert space `2 and
a ∈ `2 such that there does not exist b ∈ U with k a − bk = distance( a, U ).
[By 8.28, U cannot be a convex subset of `2 .]
13 In the real Banach space R2 with norm defined by k( x, y)k∞ = max{| x |, |y|},
give an example of a closed convex set U ⊂ R2 and z ∈ R2 such that there
exist infinitely many choices of w ∈ U with kz − wk = distance(z, U ).
14 Suppose f and g are elements of an inner product space. Prove that h f , gi = 0
if and only if
k f k ≤ k f + αgk
for all α ∈ F.
15 Suppose U is a closed subspace of a Hilbert space V and f ∈ V. Prove that
k PU f k ≤ k f k, with equality if and only if f ∈ U.
[This exercise asks you to prove 8.37(d).]
236 Chapter 8 Hilbert Spaces

16 Suppose V is a Hilbert space and P : V → V is a linear map such that P2 = P


and k P f k ≤ k f k for every f ∈ V. Prove that there exists a closed subspace U
of V such that P = PU .
17 Suppose U is a subspace of a Hilbert space V. Suppose also that W is a Banach
space and S : U → W is a bounded linear map. Prove that there exists a bounded
linear map T : V → W such that T |U = S and k T k = kSk.
[If W = F, then this result is just the Hahn–Banach Theorem (6.69) for Hilbert
spaces. The result here is stronger because it allows W to be an arbitrary
Banach space instead of requiring W to be F. Also, the proof in this Hilbert
space context does not require use of Zorn’s Lemma or the Axiom of Choice.]
18 Suppose U and W are subspaces of a Hilbert space V. Prove that U = W if and
only if U ⊥ = W ⊥ .
19 Suppose U and W are closed subspaces of a Hilbert space. Prove that PU PW = 0
if and only if h f , gi = 0 for all f ∈ U and all g ∈ W.
20 Verify the assertions in Example 8.46.
21 Show that every inner product space is a subspace of some Hilbert space.
Hint: See Exercise 13 in Section 6C.
22 Prove that if T is a bounded linear operator on a Hilbert space V and the
dimension of range T is 1, then there exist g, h ∈ V such that
T f = h f , gih
for all f ∈ V.
23 (a) Give an example of a Banach space V and a bounded linear functional ϕ
on V such that | ϕ( f )| < k ϕk k f k for all f ∈ V \ {0}.
(b) Show there does not exist an example in part (a) where V is a Hilbert space.
24 (a) Suppose ϕ and ψ are bounded linear functionals on a Hilbert space V such
that k ϕ + ψk = k ϕk + kψk. Prove that one of ϕ, ψ is a scalar multiple of
the other.
(b) Give an example to show that part (a) can fail if the hypothesis that V is a
Hilbert space is replaced by the hypothesis that V is a Banach space.
25 (a) Suppose that µ is a finite measure, 1 ≤ p ≤ 2, and ϕ is a bounded
0
linear functional
R on L p (µ). Prove that there exists h ∈ L p (µ) such that
ϕ( f ) = f h dµ for every f ∈ L p (µ).
(b) Same as part (a), but with the hypothesis that µ is a finite measure replaced
by the hypothesis that µ is a measure.
[See 7.25, which along with this exercise shows that we can identify the dual of
0
L p (u) with L p (µ) for 1 < p ≤ 2. See 9.42 for an extension to all p ∈ (1, ∞).]
26 Prove that if V is a infinite-dimensional Hilbert space, then the Banach space
B(V ) is nonseparable.
Section 8C Orthonormal Bases 237

8C Orthonormal Bases
Bessel’s Inequality
Recall that a family {ek }k∈Γ in a set V is a function e from a set Γ to V, with the
value of the function e at k ∈ Γ denoted by ek (see 6.53).

8.50 Definition orthonormal family

A family {ek }k∈Γ in an inner product space is called an orthonormal family if


(
0 if j 6= k,
h e j , ek i =
1 if j = k

for all j, k ∈ Γ.

In other words, a family {ek }k∈Γ is an orthonormal family if e j and ek are orthog-
onal for all distinct j, k ∈ Γ and kek k = 1 for all k ∈ Γ.

8.51 Example orthonormal families

• For k ∈ Z+, let ek be the element of `2 all of whose coordinates are 0 except for
the kth coordinate, which is 1:
ek = (0, . . . , 0, 1, 0, . . .).
Then {ek }k∈Z+ is an orthonormal family in `2 . In this case, our family is a
sequence; thus we can call {ek }k∈Z+ an orthonormal sequence.
• More generally, suppose Γ is a nonempty set. The Hilbert space L2 (µ), where
µ is counting measure on Γ, is often denoted by `2 (Γ). For k ∈ Γ, define a
function ek : Γ → F by
(
1 if j = k,
ek ( j ) =
0 if j 6= k.

Then {ek }k∈Γ is an orthonormal family in `2 (Γ).


• For k ∈ Z, define ek : (−π, π ] → R by
 1
 √ sin( kt ) if k > 0,


 π
ek (t) = √12π if k = 0,



 √1 cos(kt)
π
if k < 0.

Then {ek }k∈Z is an orthonormal family in L2 (−π, π ] , as you should verify
(see Exercise 1 for useful formulas that will help with this verification).
This orthonormal family {ek }k∈Z leads to the classical theory of Fourier series,
as we will see in more depth in Chapter 11.
238 Chapter 8 Hilbert Spaces

• For k a nonnegative integer, define ek : [0, 1) → F by


  
1 if x ∈ n2−k 1 , 2nk for some odd integer n,
ek ( x ) =
−1 if x ∈  n−1 , n  for some even integer n.
2k 2k

The figure below shows the graphs


This orthonormal family was
of e0 , e1 , e2 , and e3 . The pattern of
invented by Hans Rademacher
these graphs should convince you that
(1892–1969).
{ek }k∈{0,1,...} is an orthonormal fam-

ily in L2 [0, 1) .

The graph of e0 . The graph of e1 . The graph of e2 . The graph of e3 .

• Now we modify the example in the previous bullet point by translating the
functions in the previous bullet point by arbitrary integers. Specifically, for k a
nonnegative integer and m ∈ Z, define ek,m : R → F by
  

 1 if x ∈ m + n2−k 1 , m + 2nk for some odd integer n ∈ [1, 2k ],

  
ek,m ( x ) = −1 if x ∈ m + n−k 1 , m + nk for some even integer n ∈ [1, 2k ],
 2 2



0 / [m, m + 1).
if x ∈

Then {ek,m }(k,m)∈{0, 1, ...}×Z is an orthonormal family in L2 (R).


This example illustrates the usefulness of considering families that are not
sequences. Although {0, 1, . . .} × Z is a countable set and hence we could
rewrite {ek,m }(k,m)∈{0, 1, ...}×Z as a sequence, doing so would be awkward and
would be less clean than the ek,m notation.

The next result gives our first indication of why orthonormal families are so useful.

8.52 finite orthonormal families

Suppose Ω is a finite set and {e j } j∈Ω is an orthonormal family in an inner product


space. Then
2

∑ α j e j = ∑ | α j |2
j∈Ω j∈Ω

for every family {α j } j∈Ω in F.


Section 8C Orthonormal Bases 239

Proof Suppose {α j } j∈Ω is a family in F. Standard properties of inner products


show that
2 D E

∑ α j e j = ∑ α j e j , ∑ αk ek
j∈Ω j∈Ω k∈Ω

= ∑ α j αk h e j , ek i
j,k∈Ω

= ∑ | α j |2 ,
j∈Ω

as desired.

Suppose Ω is a finite set and {e j } j∈Ω is an orthonormal family in an inner product


space. The result above implies that if ∑ j∈Ω α j e j = 0, then α j = 0 for every j ∈ Ω.
Linear algebra, and algebra more generally, deals with sums of only finitely many
terms. However, in analysis we often want to sum infinitely many terms. For example,
earlier we defined the infinite sum of a sequence g1 , g2 , . . . in a normed vector space
to be the limit as n → ∞ of the partial sums ∑nk=1 gk if that limit exists (see 6.40).
The next definition captures a more powerful method of dealing with infinite sums.
The sum defined below is called an unordered sum because the set Γ is not assumed
to come with any ordering. A finite unordered sum is defined in the obvious way.

8.53 Definition unordered sum; ∑k∈Γ f k

Suppose { f k }k∈Γ is a family in a normed vector space V. The unordered sum


∑k∈Γ f k is said to converge if there exists g ∈ V such that for every ε > 0, there
exists a finite subset Ω of Γ such that


g − ∑ f j < ε
j∈Ω0

for all finite sets Ω0 with Ω ⊂ Ω0 ⊂ Γ. If this happens, we set ∑k∈Γ f k = g. If


there is no such g ∈ V, then ∑k∈Γ f k is left undefined.

Exercises at the end of this section ask you to develop basic properties of unordered
sums, including the following:
• Suppose { ak }k∈Γ is a family in R and ak ≥ 0 for each k ∈ Γ. Then the unordered
sum ∑k∈Γ ak converges if and only if
n o
sup ∑ a j : Ω is a finite subset of Γ < ∞.
j∈Ω

Furthermore, if ∑k∈Γ ak converges then it equals the supremum above. If


∑k∈Γ ak does not converge, then the supremum above is ∞ and we write
∑k∈Γ ak = ∞ (this notation should be used only when ak ≥ 0 for each k ∈ Γ).
240 Chapter 8 Hilbert Spaces

• Suppose { ak }k∈Γ is a family in R. Then the unordered sum ∑k∈Γ ak converges


if and only if ∑k∈Γ | ak | < ∞. Thus convergence of an unordered summation in
R is the same as absolute convergence. As we are about to see, the situation in
more general Hilbert spaces is quite different.
Now we can extend 8.52 to infinite sums.

8.54 linear combinations of an orthonormal family

Suppose {ek }k∈Γ is an orthonormal family in a Hilbert space V. Suppose {αk }k∈Γ
is a family in F. Then

(a) the unordered sum ∑ αk ek converges ⇐⇒ ∑ |αk |2 < ∞.


k∈Γ k∈Γ

Furthermore, if ∑k∈Γ αk ek converges, then


2

(b) ∑ α k ek = ∑ | α k |2 .
k∈Γ k∈Γ

Proof First suppose ∑k∈Γ αk ek converges, with ∑k∈Γ αk ek = g. Suppose ε > 0.


Then there exists a finite set Ω ⊂ Γ such that


g − ∑ α j e j < ε
j∈Ω0

for all finite sets Ω0 with Ω ⊂ Ω0 ⊂ Γ. If Ω0 is a finite set with Ω ⊂ Ω0 ⊂ Γ, then


the inequality above implies that


k gk − ε < ∑ α j e j < k gk + ε,
j∈Ω0

which (using 8.52) implies that


 1/2
k gk − ε < ∑ | α j |2 < k gk + ε.
j∈Ω0

1/2
Thus k gk = ∑k∈Γ |αk |2 , completing the proof of one direction of (a) and the
proof of (b).
To prove the other direction of (a), now suppose ∑k∈Γ |αk |2 < ∞. Thus there
exists an increasing sequence Ω1 ⊂ Ω2 ⊂ · · · of finite subsets of Γ such that for
each m ∈ Z+,
1
8.55 ∑ | α j |2 <
j∈Ω0 \Ωm
m2

for every finite set Ω0 such that Ωm ⊂ Ω0 ⊂ Γ. For each m ∈ Z+, let

gm = ∑ αj ej .
j∈Ωm
Section 8C Orthonormal Bases 241

If n > m, then 8.52 implies that

1
k gn − gm k2 = ∑ | α j |2 < .
j∈Ωn \Ωm
m2

Thus g1 , g2 , . . . is a Cauchy sequence and hence converges to some element g of V.


Temporarily fixing m ∈ Z+ and taking the limit of the equation above as n → ∞,
we see that
1
k g − gm k ≤ .
m
2
To show that ∑k∈Γ αk ek = g, suppose ε > 0. Let m ∈ Z+ be such that m < ε.
Suppose Ω0 is a finite set with Ωm ⊂ Ω0 ⊂ Γ. Then


g − ∑ α j e j ≤ k g − gm k + gm − ∑ α j e j
j∈Ω0 j∈Ω0
1

≤ + ∑ αj ej
m j∈Ω \Ω
0
m

1  1/2
2
= +
m ∑ j | α |
j∈Ω0 \Ω m

< ε,

where the third line comes from 8.52 and the last line comes from 8.55. Thus
∑k∈Γ αk ek = g, completing the proof.

8.56 Example a convergent unordered sum need not converge absolutely


Suppose {ek }k∈Z+ is the orthogonal family in `2 defined by setting ek equal to
the sequence that is 0 everywhere except for a 1 in the kth slot. Then by 8.54, the
unordered sum
∑ 1k ek
k ∈Z+

converges in `2 (because ∑k∈Z+ k12 < ∞) even though ∑k∈Z+ k 1k ek k = ∞. Note


that ∑k∈Z+ 1k ek = (1, 21 , 13 , . . .) ∈ `2 .

Now we prove an important inequality.

8.57 Bessel’s inequality

Suppose {ek }k∈Γ is an orthonormal family in an inner product space V and f ∈ V.


Then
∑ |h f , ek i|2 ≤ k f k2 .
k∈Γ
242 Chapter 8 Hilbert Spaces

Proof Suppose Ω is a finite subset of Γ.


Bessel’s inequality is named in
Then
honor of Friedrich Bessel
 
(1784–1846), who discovered this
f = ∑ h f , e j ie j + f− ∑ h f , e j ie j ,
inequality in 1828 in the special
j∈Ω j∈Ω
case of the trigonometric
where the first sum above is orthogonal orthonormal family given by the
to the term in parentheses above (as you third bullet point in Example 8.51.
should verify).
Applying the Pythagorean Theorem (8.9) to the equation above gives
2 2

k f k2 = ∑ h f , e j i e j + f − ∑ h f , e j ie j
j∈Ω j∈Ω
2

≥ ∑ h f , e j ie j
j∈Ω

= ∑ |h f , e j i|2 ,
j∈Ω

where the last equality follows from 8.52. Because the inequality above holds for
every finite set Ω ⊂ Γ, we conclude that k f k2 ≥ ∑k∈Γ |h f , ek i|2 , as desired.

Recall that the span of a family {ek }k∈Γ in a vector space is the set of finite sums
of the form
∑ αj ej ,
j∈Ω

where Ω is a finite subset of Γ and {α j } j∈Ω is a family in F (see 6.54). Bessel’s


inequality now allows us to prove the following beautiful result showing that the
closure of the span of an orthonormal family is a set of infinite sums.

8.58 closure of the span of an orthonormal family

Suppose {ek }k∈Γ is an orthonormal family in a Hilbert space V. Then


n o
(a) span {ek }k∈Γ = ∑ αk ek : {αk }k∈Γ is a family in F and ∑ |αk |2 < ∞ .
k∈Γ k∈Γ

Furthermore,

(b) f = ∑ h f , ek i ek
k∈Γ

for every f ∈ span {ek }k∈Γ .

Proof The right side of (a) above makes sense because of 8.54(a). Furthermore, the
right side of (a) above is a subspace of V because `2 (Γ) [which equals L2 (µ), where
µ is counting measure on Γ] is closed under addition and scalar multiplication by 7.5.
Section 8C Orthonormal Bases 243

Suppose first {αk }k∈Γ is a family in F and ∑k∈Γ |αk |2 < ∞. Let ε > 0. Then
there is a finite subset Ω of Γ such that

∑ | α j |2 < ε2 .
j∈Γ\Ω

The inequality above and 8.54(b) imply that




∑ αk ek − ∑ α j e j < ε.
k∈Γ j∈Ω

The definition of the closure (see 6.7) now implies that ∑k∈Γ αk ek ∈ span {ek }k∈Γ ,
showing that the right side of (a) is contained in the left side of (a).
To prove the inclusion in the other direction, now suppose f ∈ span {ek }k∈Γ . Let

8.59 g= ∑ h f , ek i ek ,
k∈Γ

where the sum above converges by Bessel’s inequality (8.57) and by 8.54(a). The
direction of the inclusion that we just proved implies that g ∈ span {ek }k∈Γ . Thus

8.60 g − f ∈ span {ek }k∈Γ .

Equation 8.59 implies that h g, e j i = h f , e j i for each j ∈ Γ, as you should verify


(which will require using the Cauchy–Schwarz inequality if done rigorously). Hence

h g − f , ek i = 0 for every k ∈ Γ.

This implies that


⊥ ⊥
g − f ∈ span{e j } j∈Γ = span{e j } j∈Γ ,

where the equality above comes from 8.40(d). Now 8.60 and the inclusion above
imply that f = g [see 8.40(b)], which along with 8.59 implies that f is in the right
side of (a), completing the proof of (a).
The equations f = g and 8.59 also imply (b).

Parseval’s Identity
Note that 8.52 implies that every orthonormal family in an inner product space is
linearly independent (see 6.54 to review the definition of linearly independent and
basis). Linear algebra deals mainly with finite-dimensional vector spaces, but infinite-
dimensional vector spaces frequently appear in analysis. The notion of a basis is not
so useful when doing analysis with infinite-dimensional vector spaces because the
definition of span does not take advantage of the possibility of summing an infinite
number of elements.
However, 8.58 tells us that taking the closure of the span of an orthonormal
family can capture the sum of infinitely many elements. Thus we make the following
definition.
244 Chapter 8 Hilbert Spaces

8.61 Definition orthonormal basis

An orthonormal family {ek }k∈Γ in a Hilbert space V is called an orthonormal


basis of V if
span {ek }k∈Γ = V.

In addition to requiring orthonormality (which implies linear independence), the


definition above differs from the definition of a basis by considering the closure of
the span rather than the span. An important point to keep in mind is that despite the
terminology, an orthonormal basis is not necessarily a basis in the sense of 6.54. In
fact, if Γ is an infinite set and {ek }k∈Γ is an orthonormal basis of V, then {ek }k∈Γ is
not a basis of V (see Exercise 9).

8.62 Example orthonormal bases

• For n ∈ Z+ and k ∈ {1, . . . , n}, let ek be the element of Fn all of whose


coordinates are 0 except the kth coordinate, which is 1:

ek = (0, . . . , 0, 1, 0, . . . , 0).

Then {ek }k∈{1,...,n} is an orthonormal basis of Fn .


  
• Let e1 = √1 , √1 , √1 , e2 = − √1 , √1 , 0 , and e3 = √1 , √1 , − √2 . Then
3 3 3 2 2 6 6 6
{ek }k∈{1,2,3} is an orthonormal basis of F3 , as you should verify.
• The first three bullet points in 8.51 are examples of orthonormal families that are
orthonormal bases. The exercises ask you to verify that we have an orthonormal
basis in the first and second bullet points of 8.51. For the third bullet point
(trigonometric functions), see Exercise 11 in Section 10D or see Chapter 11.

The next result shows why orthonormal bases are so useful—a Hilbert space with
an orthonormal basis {ek }k∈Γ behaves like `2 (Γ).

8.63 Parseval’s identity

Suppose {ek }k∈Γ is an orthonormal basis of a Hilbert space V and f , g ∈ V. Then

(a) f = ∑ h f , ek i ek ;
k∈Γ

(b) h f , gi = ∑ h f , ek ih g, ek i;
k∈Γ

(c) k f k2 = ∑ |h f , ek i|2 .
k∈Γ

Proof The equation in (a) follows immediately from 8.58(b) and the definition of an
orthonormal basis.
Section 8C Orthonormal Bases 245

To prove (b), note that


Equation (c) is called Parseval’s
D E identity in honor of Marc-Antoine
h f , g i = ∑ h f , ek i ek , g Parseval (1755–1836), who
k∈Γ
discovered a special case in 1799.
= ∑ h f , ek ihek , gi
k∈Γ

= ∑ h f , ek ih g, ek i,
k∈Γ

where the first equation follows from (a) and the second equation follows from the
definition of an unordered sum and the Cauchy–Schwarz inequality.
Equation (c) follows from setting g = f in (b). An alternative proof: equation (c)
follows from 8.54(b) and the equation f = ∑ h f , ek iek from (a).
k∈Γ

Gram–Schmidt Process and Existence of Orthonormal Bases

8.64 Definition separable

A normed vector space is called separable if it has a countable subset whose


closure equals the whole space.

8.65 Example separable normed vector spaces

• Suppose n ∈ Z+. Then Fn with the usual Hilbert space norm is separable
because the closure of the countable set

{(c1 , . . . , cn ) ∈ Fn : each c j is rational}

equals Fn (in case F = C: to say that a complex number is rational in this


context means that both the real and imaginary parts of the complex number are
rational numbers in the usual sense).
• The Hilbert space `2 is separable because the closure of the countable set

[
{(c1 , . . . , cn , 0, 0, . . .) ∈ `2 : each c j is rational}
n =1

is `2 .
• The Hilbert spaces L2 ([0, 1]) and L2 (R) are separable, as Exercise 13 asks you
to verify [hint: consider finite linear combinations with rational coefficients of
functions of the form χ(c, d), where c and d are rational numbers].
246 Chapter 8 Hilbert Spaces

A moment’s thought about the definition of closure (see 6.7) shows that a normed
vector space V is separable if and only if there exists a countable subset C of V such
that every open ball in V contains at least one element of C.

8.66 Example nonseparable normed vector spaces

• Suppose Γ is an uncountable set. Then √the Hilbert space `2 (Γ) is not separable.
To see this, note that kχ{ j} − χ{k}k = 2 for all j, k ∈ Γ with j 6= k. Hence
n √  o
2
B χ{k} , 2 :k∈Γ

is an uncountable collection of disjoint open balls in `2 (Γ); no countable set can


have at least one element in each of these balls.
• The Banach space L∞ ([0, 1]) is not separable. Here kχ[0, s] − χ[0, t]k = 1 for all
s, t ∈ [0, 1] with s 6= t. Thus
n  o
B χ[0, t], 12 : t ∈ [0, 1]

is an uncountable collection of disjoint open balls in L∞ ([0, 1]).

We present two proofs of the existence of orthonormal bases of Hilbert spaces.


The first proof works only for separable Hilbert spaces, but it gives a useful algorithm,
called the Gram–Schmidt process, for constructing orthonormal sequences. The
second proof works for all Hilbert spaces, but it uses a result that depends upon the
Axiom of Choice.
Which proof should you read? In practice, the Hilbert spaces you will encounter
will almost certainly be separable. Thus the first proof suffices, and it has the
additional benefit of introducing you to a widely used algorithm. The second proof
uses an entirely different approach and has the advantage of applying to separable
and nonseparable Hilbert spaces. For maximum learning, read both proofs!

8.67 existence of orthonormal bases for separable Hilbert spaces

Every separable Hilbert space has an orthonormal basis.

Proof Suppose V is a separable Hilbert space and { f 1 , f 2 , . . .} is a countable subset


of V whose closure equals V. We will inductively define an orthonormal sequence
{ek }k∈Z+ such that

8.68 span{ f 1 , . . . , f n } ⊂ span{e1 , . . . , en }

for each n ∈ Z+. This will imply that span{ek }k∈Z+ = V, which will mean that
{ek }k∈Z+ is an orthonormal basis of V.
To get started with the induction, set e1 = f 1 /k f 1 k (we can assume that f 1 6= 0).
Section 8C Orthonormal Bases 247

Now suppose n ∈ Z+ and e1 , . . . , en have been chosen so that {ek }k∈{1,...,n} is


an orthonormal family in V and 8.68 holds. If f k ∈ span{e1 , . . . , en } for every
k ∈ Z+, then {ek }k∈{1,...,n} is an orthonormal basis of V (completing the proof) and
the process should be stopped. Otherwise, let m be the smallest positive integer such
that
8.69 / span{e1 , . . . , en }.
fm ∈
Define en+1 by
f m − h f m , e1 i e1 − · · · − h f m , e n i e n
8.70 e n +1 = .
k f m − h f m , e1 i e1 − · · · − h f m , e n i e n k
Clearly ken+1 k = 1 (8.69 guaran-
Jørgen Gram (1850–1916) and
tees there is no division by 0). If
Erhard Schmidt (1876–1959)
k ∈ {1, . . . , n}, then the equation above
popularized this process that
implies that hen+1 , ek i = 0. Thus
constructs orthonormal sequences.
{ek }k∈{1,...,n+1} is an orthonormal fam-
ily in V. Also, 8.68 and the choice of m
as the smallest positive integer satisfying
8.69 imply that
span{ f 1 , . . . , f n+1 } ⊂ span{e1 , . . . , en+1 },
completing the induction and completing the proof.

Before considering nonseparable Hilbert spaces, we take a short detour to illustrate


how the Gram–Schmidt process used in the previous proof can be used to find closest
elements to subspaces. We begin with a result connecting the orthogonal projection
onto a closed subspace with an orthonormal basis of that subspace.

8.71 orthogonal projection in terms of an orthonormal basis

Suppose that U is a closed subspace of a Hilbert space V and {ek }k∈Γ is an


orthonormal basis of U. Then

PU f = ∑ h f , ek i ek
k∈Γ

for all f ∈ V.

Proof Let f ∈ V. If k ∈ Γ, then


8.72 h f , ek i = h f − PU f , ek i + h PU f , ek i = h PU f , ek i,
where the last equality follows from 8.37(a). Now
PU f = ∑ h PU f , ek iek = ∑ h f , ek iek ,
k∈Γ k∈Γ

where the first equality follows from Parseval’s identity [8.63(a)] as applied to U and
its orthonormal basis {ek }k∈Γ , and the second equality follows from 8.72.
248 Chapter 8 Hilbert Spaces

8.73 Example best approximation


Find the polynomial g of degree at most 10 that minimizes
Z 1 q 2

| x | − g( x ) dx.
−1

Solution We will work in the real Hilbert space L2 ([−1, 1]) with the usual inner
R1
product h g, hi = −1 gh. For k ∈ {0, 1, . . . , 10}, let f k ∈ L2 ([−1, 1]) be defined by
f k ( x ) = x k . Let U be the subspace of L2 ([−1, 1]) defined by

U = span{ f k }k∈{0, ..., 10} .

Apply the Gram–Schmidt process from the proof of 8.67 to { f k }k∈{0, ..., 10} , pro-
ducing an orthonormal basis {ek }k∈{0,...,10} of U, which is a closed subspace of
L2 ([−1, 1]) (see Exercise 8). The point here is that {ek }k∈{0, ..., 10} can be computed
explicitly and exactly by using 8.70 and evaluating some integrals (using software √ that
can do exact√ rational arithmetic will make the process easier), getting e 0 ( x ) = 1/ 2,
e1 ( x ) = 6x/2, . . . up to

42
e10 ( x ) = (−63 + 3465x2 − 30030x4 + 90090x6 − 109395x8 + 46189x10 ).
512
p
Define f ∈ L2 ([−1, 1]) by f ( x ) = | x |. Because U is the subspace of
L2 ([−1, 1]) consisting of polynomials of degree at most 10 and PU f equals the
element of U closest to f (see 8.34), the formula in 8.71 tells us that the solution g to
our minimization problem is given by the formula
10
g= ∑ h f , ek i ek .
k =0

Using the explicit expressions for e0 , . . . , e10 and again evaluating some integrals,
this gives

693 + 15015x2 − 64350x4 + 139230x6 − 138567x8 + 51051x10


g( x ) = .
2944
The figure
p here shows the graph of
f ( x ) = | x | (red) and the graph of
its closest polynomial g (blue) of de-
gree at most 10; here closest means as
measured in the norm of L2 ([−1, 1]).
The approximation of f by g is
pretty good, especially considering
that f is not differentiable at 0 and thus
a Taylor series expansion for f does
not make sense.
Section 8C Orthonormal Bases 249

Recall that a subset Γ of a set V can be thought of as a family in V by considering


{e f } f ∈Γ , where e f = f . With this convention, a subset Γ of an inner product space
V is an orthonormal subset of V if k f k = 1 for all f ∈ Γ and h f , gi = 0 for all
f , g ∈ Γ with f 6= g.
The next result characterizes the orthonormal bases as the maximal elements
among the collection of orthonormal subsets of a Hilbert space. Recall that a set
Γ ∈ A in a collection of subsets of a set V is a maximal element of A if there does
not exist Γ0 ∈ A such that Γ $ Γ0 (see 6.55).

8.74 orthonormal bases as maximal elements

Suppose V is a Hilbert space, A is the collection of all orthonormal subsets of V,


and Γ is an orthonormal subset of V. Then Γ is an orthonormal basis of V if and
only if Γ is a maximal element of A.

Proof First suppose Γ is an orthonormal basis of V. Parseval’s identity [8.63(a)]


implies that the only element of V that is orthogonal to every element of Γ is 0. Thus
there does not exist an orthonormal subset of V that strictly contains Γ. In other
words, Γ is a maximal element of A.
To prove the other direction, suppose now that Γ is a maximal element of A. Let
U denote the span of Γ. Then
U ⊥ = {0}
because if f is a nonzero element of U ⊥ , then Γ ∪ { f /k f k} is an orthonormal subset
of V that strictly contains Γ. Hence U = V (by 8.42), which implies that Γ is an
orthonormal basis of V.

Now we are ready to prove that every Hilbert space has an orthonormal basis.
Before reading the next proof, you may want to review the definition of a chain (6.58),
which is a collection of sets such that for each pair of sets in the collection, one of
them is contained in the other. You should also review Zorn’s Lemma (6.60), which
gives a way to show that a collection of sets contains a maximal element.

8.75 existence of orthonormal bases for all Hilbert spaces

Every Hilbert space has an orthonormal basis.

Proof Suppose V is a Hilbert space. Let A be the collection of all orthonormal


subsets of V. Suppose C ⊂ A is a chain. Let L be the union of all the sets in C . If
f ∈ L, then k f k = 1 because f is an element of some orthonormal subset of V that
is contained in C .
If f , g ∈ L with f 6= g, then there exist orthonormal subsets Ω and Γ in C such
that f ∈ Ω and g ∈ Γ. Because C is a chain, either Ω ⊂ Γ or Γ ⊂ Ω. Either way,
there is an orthonormal subset of V that contains both f and g. Thus h f , gi = 0.
We have shown that L is an orthonormal subset of V; in other words, L ∈ A.
Thus Zorn’s Lemma (6.60) implies that A has a maximal element. Now 8.74 implies
that V has an orthonormal basis.
250 Chapter 8 Hilbert Spaces

Riesz Representation Theorem, Revisited


Now that we know that every Hilbert space has an orthonormal basis, we can give a
completely different proof of the Riesz Representation Theorem (8.47) than the proof
we gave earlier.
Note that the new proof below of the Riesz Representation Theorem gives the
formula 8.77 for h in terms of an orthonormal basis. One interesting feature of this
formula is that h is uniquely determined by ϕ and thus h does not depend upon the
choice of an orthonormal basis. Hence despite its appearance, the right side of 8.77
is independent of the choice of an orthonormal basis.

8.76 Riesz Representation Theorem

Suppose ϕ is a bounded linear functional on a Hilbert space V and {ek }k∈Γ is an


orthonormal basis of V. Let

8.77 h= ∑ ϕ ( ek ) ek .
k∈Γ

Then

8.78 ϕ( f ) = h f , hi
1/2
for all f ∈ V. Furthermore, k ϕk = ∑k∈Γ | ϕ(ek )|2 .

Proof First we must show that the sum defining h makes sense. To do this, suppose
Ω is a finite subset of Γ. Then
   1/2

∑ | ϕ(e j )|2 = ϕ ∑ ϕ(e j )e j ≤ k ϕk ∑ ϕ(e j )e j = k ϕk ∑ | ϕ(e j )|2 ,
j∈Ω j∈Ω j∈Ω j∈Ω
 1/2
where the last equality follows from 8.52. Dividing by ∑ j∈Ω | ϕ(e j )|2 gives
 1/2
∑ | ϕ(e j )|2 ≤ k ϕ k.
j∈Ω

Because the inequality above holds for every finite subset Ω of Γ, we conclude that
∑ | ϕ(ek )|2 ≤ k ϕk2 .
k∈Γ
Thus the sum defining h makes sense (by 8.54) in equation 8.77.
Now 8.77 shows that hh, e j i = ϕ(e j ) for each j ∈ Γ. Thus if f ∈ V then
 
ϕ( f ) = ϕ ∑ h f , ek iek = ∑ h f , ek i ϕ(ek ) = ∑ h f , ek ihh, ek i = h f , hi,
k∈Γ k∈Γ k∈Γ
where the first and last equalities follow from 8.63 and the second equality follows
from the boundedness/continuity of ϕ. Thus 8.78 holds.
Finally, the Cauchy–Schwarz inequality, equation 8.78, and the equation ϕ(h) =
1/2
hh, hi show that k ϕk = khk = ∑k∈Γ | ϕ(ek )|2 .
Section 8C Orthonormal Bases 251

EXERCISES 8C
1 Verify that the family {ek }k∈Z as defined in the
 third bullet point of Example
8.51 is an orthonormal family in L2 (−π, π ] . The following formulas should
help:

sin( x + y) + sin( x − y)
(sin x )(cos y) = ,
2
cos( x − y) − cos( x + y)
(sin x )(sin y) = ,
2
cos( x + y) + cos( x − y)
(cos x )(cos y) = .
2

2 Suppose { ak }k∈Γ is a family in R and ak ≥ 0 for each k ∈ Γ. Prove the


unordered sum ∑k∈Γ ak converges if and only if
n o
sup ∑ a j : Ω is a finite subset of Γ < ∞.
j∈Ω

Furthermore, prove that if ∑k∈Γ ak converges then it equals the supremum above.
3 Suppose {ek }k∈Γ is an orthonormal family in an inner product space V. Prove
that if f ∈ V, then {k ∈ Γ : h f , ek i 6= 0} is a countable set.
4 Suppose { f k }k∈Γ and { gk }k∈Γ are families in a normed vector space such that
∑k∈Γ f k and ∑k∈Γ gk converge. Prove that ∑k∈Γ ( f k + gk ) converges and

∑ ( f k + gk ) = ∑ fk + ∑ gk .
k∈Γ k∈Γ k∈Γ

5 Suppose { f k }k∈Γ is a family in a normed vector space such that ∑k∈Γ f k con-
verges. Prove that if c ∈ F, then ∑k∈Γ (c f k ) converges and

∑ (c f k ) = c ∑ fk .
k∈Γ k∈Γ

6 Suppose { ak }k∈Γ is a family in R. Prove that the unordered sum ∑k∈Γ ak


converges if and only if ∑k∈Γ | ak | < ∞.

7 Suppose { f k }k∈Z+ is a family in a normed vector space. Prove that the un-
ordered sum ∑k∈Z+ f k converges if and only if the usual ordered sum ∑∞
k =1 f p ( k )
converges for every injective function p : Z+ → Z+.
8 Explain why 8.58 implies that if Γ is a finite set and {ek }k∈Γ is an orthonormal
family in a Hilbert space V, then span{ek }k∈Γ is a closed subspace of V.
9 Suppose V is an infinite-dimensional Hilbert space. Prove that there does not
exist a basis of V that is an orthonormal family.
252 Chapter 8 Hilbert Spaces

10 (a) Show that the orthonormal family given in the first bullet point of Exam-
ple 8.51 is an orthonormal basis of `2 .
(b) Show that the orthonormal family given in the second bullet point of Exam-
ple 8.51 is an orthonormal basis of `2 (Γ).
(c) Show that the orthonormal family given in the fourth
 bullet point of Exam-
ple 8.51 is not an orthonormal basis of L2 [0, 1) .
(d) Show that the orthonormal family given in the fifth bullet point of Exam-
ple 8.51 is not an orthonormal basis of L2 (R).
11 Suppose µ is a σ-finite measure on ( X, S) and ν is a σ-finite measure on (Y, T ).
Suppose also that {e j } j∈Ω is an orthonormal basis of L2 (µ) and { f k }k∈Γ is an
orthonormal basis of L2 (ν). For j ∈ Ω and k ∈ Γ, define g j,k : X × Y → F by

g j,k ( x, y) = e j ( x ) f k (y).

Prove that { g j,k } j∈Ω, k∈Γ is an orthonormal basis of L2 (µ × ν).

12 Prove the converse of Parseval’s identity. More specifically, prove that if {ek }k∈Γ
is an orthonormal family in a Hilbert space V and

k f k2 = ∑ |h f , ek i|2
k∈Γ

for every f ∈ V, then {ek }k∈Γ is an orthonormal basis of V.


13 (a) Show that the Hilbert space L2 ([0, 1]) is separable.
(b) Show that the Hilbert space L2 (R) is separable.
(c) Show that the Banach space `∞ is not separable.
14 Prove that every subspace of a separable normed vector space is separable.
15 Suppose V is an infinite-dimensional Hilbert space. Prove that there does not
exist a translation invariant measure on the Borel subsets of V that assigns
positive but finite measure to each open ball in V.
[A subset of V is called a Borel set if it is in the smallest σ-algebra containing
all the open subsets of V. A measure µ on the Borel subsets of V is called
translation invariant if µ( f + E) = µ( E) for every f ∈ V and every Borel set
E of V.]
Z 1
16 Find the polynomial g of degree at most 4 that minimizes x5 − g( x ) 2 dx.
0

17 Prove that each orthonormal family in a Hilbert space can be extended to


an orthonormal basis of the Hilbert space. Specifically, suppose {e j } j∈Ω is
an orthonormal family in a Hilbert space V. Prove that there exists a set Γ
containing Ω and an orthonormal basis { f k }k∈Γ of V such that f j = e j for every
j ∈ Ω.
18 Prove that every vector space has a basis.
Section 8C Orthonormal Bases 253

19 Find the polynomial g of degree at most 4 such that


 Z 1
1
f 2 = fg
0

for every polynomial f of degree at most 4.

Exercises 20–25 are for readers familiar with analytic functions.


20 Suppose G is a nonempty open subset of C. The Bergman space L2a ( G ) is
defined to be the set of analytic functions f : G → C such that
Z
| f |2 dλ2 < ∞,
G

where λ2 is the usual Lebesgue measure R on R2 , which is identified with C. For


f , h ∈ L2a ( G ), define h f , hi to be G f h dλ2 .

(a) Show that L2a ( G ) is a Hilbert space.


(b) Show that if w ∈ G, then f 7→ f (w) is a bounded linear functional on
L2a ( G ).

21 Let D denote the open unit disk in C; thus


D = { z ∈ C : | z | < 1}.
(a) Find an orthonormal basis of L2a (D).
(b) Suppose f ∈ L2a (D) has Taylor series

f (z) = ∑ ak zk
k =0

for z ∈ D. Find a formula for k f k in terms of a0 , a1 , a2 , . . . .


(c) Suppose w ∈ D. By the previous exercise and the Riesz Representation
Theorem (8.47 and 8.76), there exists Γw ∈ L2a (D) such that
f (w) = h f , Γw i for all f ∈ L2a (D).
Find an explicit formula for Γw .

22 Suppose G is the annulus defined by


G = { z ∈ C : 1 < | z | < 2}.
(a) Find an orthonormal basis of L2a ( G ).
(b) Suppose f ∈ L2a ( G ) has Laurent series

f (z) = ∑ ak zk
k=−∞

for z ∈ G. Find a formula for k f k in terms of . . . , a−1 , a0 , a1 , . . . .


254 Chapter 8 Hilbert Spaces

23 Prove that if f ∈ L2a (D \ {0}), then f has a removable singularity at 0 (meaning


that f can be extended to a function that is analytic on D).
24 The Dirichlet space D is defined to be the set of analytic functions f : D → C
such that Z
| f 0 |2 dλ2 < ∞.
D
R
For f , g ∈ D , define h f , gi to be f (0) g(0) + D f 0 g0 dλ2 .
(a) Show that D is a Hilbert space.
(b) Show that if w ∈ D, then f 7→ f (w) is a bounded linear functional on D .
(c) Find an orthonormal basis of D .
(d) Suppose f ∈ D has Taylor series

f (z) = ∑ ak zk
k =0

for z ∈ D. Find a formula for k f k in terms of a0 , a1 , a2 , . . . .


(e) Suppose w ∈ D. Find an explicit formula for Γw ∈ D such that

f (w) = h f , Γw i for all f ∈ D .

25 (a) Prove that the Dirichlet space D is contained in the Bergman space L2a (D).
(b) Prove that there exists a function f ∈ L2a (D) such that f is uniformly
continuous on D and f ∈ / D.
Chapter 9
Real and Complex Measures

A measure is a countably additive function from a σ-algebra to [0, ∞]. In this chapter,
we consider countably additive functions from a σ-algebra to either R or C. The first
section of this chapter shows that these functions, called real measures or complex
measures, form an interesting Banach space with an appropriate norm.
The second section of this chapter focuses on decomposition theorems that help
us understand real and complex measures. These results will lead to a proof that the
0
dual space of L p (µ) can be identified with L p (µ).

Dome in the main building of the University of Vienna, where Johann Radon
(1887–1956) was a student and then later a faculty member. The Radon–Nikodym
Theorem, which will be proved in this chapter using Hilbert space techniques,
provides information analogous to differentiation for measures.
CC-BY-SA Hubertl

255
256 Chapter 9 Real and Complex Measures

9A Total Variation
Properties of Real and Complex Measures
Recall that a measurable space is a pair ( X, S), where S is a σ-algebra on X. Recall
also that a measure on ( X, S) is a countably additive function from S to [0, ∞] that
takes ∅ to 0. Countably additive functions that take values in R or C give us new
objects called real measures or complex measures.

9.1 Definition real and complex measures

Suppose ( X, S) is measurable space.


• A function ν : S → F is called countably additive if

[  ∞
ν Ek = ∑ ν(Ek )
k =1 k =1

for every disjoint sequence E1 , E2 , . . . of sets in S .


• A real measure on ( X, S) is a countably additive function ν : S → R.
• A complex measure on ( X, S) is a countably additive function ν : S → C.

The word measure can be ambiguous


The terminology nonnegative
in the mathematical literature. The most
measure would be more appropriate
common use of the word measure is as
than positive measure because the
we defined it in Chapter 2 (see 2.54).
function µ : S → F defined by
However, some mathematicians use the
µ( E) = 0 for every E ∈ S is a
word measure to include what are here
positive measure. However, we will
called real and complex measures; they
stick with tradition and use the
then use the phrase positive measure to
phrase positive measure.
refer to what we defined as a measure in
2.54. To help relieve this ambiguity, in this chapter we usually use the phrase
(positive) measure to refer to measures as defined in 2.54. Putting positive in paren-
theses helps reinforce the idea that it is optional while distinguishing such measures
from real and complex measures.

9.2 Example real and complex measures


• Let λ denote Lebesgue measure on [−1, 1]. Define ν on the Borel subsets of
[−1, 1] by  
ν( E) = λ E ∩ [0, 1] − λ E ∩ [−1, 0) .
Then ν is a real measure.
• If µ1 and µ2 are finite (positive) measures, then µ1 − µ2 is a real measure and
α1 µ1 + α2 µ2 is a complex measure for all α1 , α2 ∈ C.
• If ν is a complex measure, then Re ν and Im ν are real measures.
Section 9A Total Variation 257

Note that every real measure is a complex measure. Note also that by definition,
∞ is not an allowable value for a real or complex measure. Thus a (positive) measure
µ on ( X, S) is a real measure if and only if µ( X ) < ∞.
Some authors use the terminology signed measure instead of real measure; some
authors allow a real measure to take on the value ∞ or −∞ (but not both, because the
expression ∞ − ∞ must be avoided). However, real measures as defined here serve
us better because we need to avoid ±∞ when considering the Banach space of real
or complex measures on a measurable space (see 9.18).
For (positive) measures, we had to make µ(∅) = 0 part of the definition to avoid
the function µ that assigns ∞ to all sets, including the empty set. But ∞ is not an
allowable value for real or complex measures. Thus ν(∅) = 0 is a consequence of
our definition rather than part of the definition, as shown in the next result.

9.3 absolute convergence for a disjoint union

Suppose ν is a complex measure on a measurable space ( X, S). Then

(a) ν(∅) = 0;

(b) ∑ |ν(Ek )| < ∞ for every disjoint sequence E1 , E2 , . . . of sets in S .
k =1

Proof To prove (a), note that ∅, ∅, . . . is a disjoint sequence of sets in S whose


union equals ∅. Thus

ν(∅) = ∑ ν ( ∅ ).
k =1
The right side of the equation above makes sense as an element of R or C only when
ν(∅) = 0, which proves (a).
To prove (b), suppose E1 , E2 , . . . is a disjoint sequence of sets in S . First suppose
ν is a real measure. Thus
 [ 
ν Ek = ∑ ν(Ek ) = ∑ |ν(Ek )|
{k : ν( Ek )>0} {k : ν( Ek )>0} {k : ν( Ek )>0}

and  [ 
−ν Ek = − ∑ ν( Ek ) = ∑ |ν( Ek )|.
{k : ν( Ek )<0} {k : ν( Ek )<0} {k : ν( Ek )<0}

Because ν( E) ∈ R for every E ∈ S , the right side of the last two displayed equations
is finite. Thus ∑∞
k=1 | ν ( Ek )| < ∞, as desired.
Now consider the case where ν is a complex measure. Then
∞ ∞ 
∑ |ν(Ek )| ≤ ∑ |(Re ν)( Ek )| + |(Im ν)( Ek )| < ∞,
k =1 k =1

where the last inequality follows from applying the result for real measures to the
real measures Re ν and Im ν.
258 Chapter 9 Real and Complex Measures

The next definition provides an important class of examples of real and complex
measures.

9.4 measure determined by an L1 -function

Suppose µ is a (positive) measure on a measurable space ( X, S) and h ∈ L1 (µ).


Define ν : S → F by Z
ν( E) = h dµ.
E
Then ν is a real measure on ( X, S) if F = R and is a complex measure on ( X, S)
if F = C.

Proof Suppose E1 , E2 , . . . is a disjoint sequence of sets in S . Then



[  Z ∞  ∞ Z ∞
9.5 ν Ek = ∑ χE (x)h(x) dµ(x) =
k
∑ χ E h dµ =
k
∑ ν(Ek ),
k =1 k =1 k =1 k =1

where the first equality holds because the sets E1 , E2 , . . . are disjoint and the second
equality follows from the inequality
m

∑ χ Ek ( x )h( x ) ≤ |h( x )|,
k =1

which along with the assumption that h ∈ L1 (µ) allows us to interchange the integral
and limit of the partial sums by the Dominated Convergence Theorem (3.31).
The countable additivity shown in 9.5 means ν is a real or complex measure.

The next definition simply gives a notation for the measure defined in the previous
result. In the notation that we are about to define, the symbol d has no separate
meaning—it functions to separate h and µ.

9.6 Definition h dµ

Suppose µ is a (positive) measure on a measurable space ( X, S) and h ∈ L1 (µ).


Then h dµ is the real or complex measure on ( X, S) defined by
Z
(h dµ)( E) = h dµ.
E

Note that if a function h ∈ L1 (µ) takes values in [0, ∞), then h dµ is a finite
(positive) measure.
The next result shows some basic properties of complex measures. No proofs
are given because the proofs are the same as the proofs of the corresponding results
for (positive) measures. Specifically, see the proofs of 2.57, 2.61, 2.59, and 2.60.
Because complex measures cannot take on the value ∞, we do not need to worry
about hypotheses of finite measure that are required of the (positive) measure versions
of all but part (c).
Section 9A Total Variation 259

9.7 properties of complex measures

Suppose ν is a complex measure on a measurable space ( X, S). Then

(a) ν( E \ D ) = ν( E) − ν( D ) for all D, E ∈ S with D ⊂ E;


(b) ν( D ∪ E) = ν( D ) + ν( E) − ν( D ∩ E) for all D, E ∈ S ;

[ 
(c) ν Ek = lim ν( Ek )
k→∞
k =1
for all increasing sequences E1 ⊂ E2 ⊂ · · · of sets in S ;

\ 
(d) ν Ek = lim ν( Ek )
k→∞
k =1
for all decreasing sequences E1 ⊃ E2 ⊃ · · · of sets in S .

Total Variation Measure


We use the terminology total variation measure below even though we have not
yet shown that the object being defined is a measure. Soon we will justify this
terminology (see 9.11).

9.8 Definition total variation measure

Suppose ν is a complex measure on a measurable space ( X, S). The total


variation measure is the function |ν| : S → [0, ∞] defined by

|ν|( E) = sup |ν( E1 )| + · · · + |ν( En )| : n ∈ Z+ and E1 , . . . , En

are disjoint sets in S such that E1 ∪ · · · ∪ En ⊂ E .

To start getting familiar with the definition above, you should verify that if ν is a
complex measure on ( X, S) and E ∈ S , then
• |ν( E)| ≤ |ν|( E);
• |ν|( E) = ν( E) if ν is a finite (positive) measure;
• |ν|( E) = 0 if and only if ν( A) = 0 for every A ∈ S such that A ⊂ E.
The next result states that for real measures, we can consider only n = 2 in the
definition of the total variation measure.

9.9 total variation measure of a real measure

Suppose ν is a real measure on a measurable space ( X, S) and E ∈ S . Then

|ν|( E) = sup{|ν( A)| + |ν( B)| : A, B are disjoint sets in S and A ∪ B ⊂ E}.
260 Chapter 9 Real and Complex Measures

Proof Suppose that n ∈ Z+ and E1 , . . . , En are disjoint sets in S such that


E1 ∪ · · · ∪ En ⊂ E. Let
[ [
A= Ek and B= Ek .
{k : ν( Ek )>0} {k : ν( Ek )<0}

Then A, B are disjoint sets in S and A ∪ B ⊂ E. Furthermore,

|ν( A)| + |ν( B)| = |ν( E1 )| + · · · + |ν( En )|.

Thus in the supremum that defines |ν|( E), we can take n = 2.

The next result could be rephrased as stating that if h ∈ L1 (µ), then the total
variation measure of the measure h dµ is the measure |h| dµ. In the statement below,
the notation dν = h dµ means the same as ν = h dµ; the notation dν is commonly
used when considering expressions involving measures of the form h dµ.

9.10 total variation measure of h dµ

Suppose µ is a (positive) measure on a measurable space ( X, S), h ∈ L1 (µ),


and dν = h dµ. Then Z
|ν|( E) = |h| dµ
E
for every E ∈ S .

Proof Suppose that E ∈ S. If E1 , . . . , En is a disjoint sequence in S such that


E1 ∪ · · · ∪ En ⊂ E, then
n Z
n n Z Z

∑ |ν(Ek )| = ∑ Ek
h dµ ≤ ∑ |h| dµ ≤ |h| dµ.
k =1 k =1 k=1 Ek E
R
The inequality above implies that |ν|( E) ≤ E |h| dµ.
To prove the inequality in the other direction, first suppose F = R; thus h is a
real-valued function and ν is a real measure. Let

A = { x ∈ E : h ( x ) > 0} and B = { x ∈ E : h ( x ) < 0}.

Then A and B are disjoint sets in S and A ∪ B ⊂ E. We have


Z Z Z
|ν( A)| + |ν( B)| = h dµ − h dµ = |h| dµ.
A B E
R
Thus |ν|( E) ≥ E |h| dµ, completing the proof in the case F = R.
Now suppose F = C; thus ν is a complex measure. Let ε > 0. There exists a
simple function g ∈ L1 (µ) such that k g − hk1 < ε (by 3.44). There exist disjoint
sets E1 , . . . , En ∈ S and c1 , . . . , cn ∈ C such that E1 ∪ · · · ∪ En ⊂ E and
n
g| E = ∑ ck χ E . k
k =1
Section 9A Total Variation 261

Now
n n Z

∑ |ν(Ek )| = ∑ Ek
h dµ
k =1 k =1
n Z n Z

≥ ∑ g dµ − ∑ ( g − h ) dµ
k =1 Ek k =1 Ek
n n Z

= ∑ |ck |µ(Ek ) − ∑ Ek
( g − h) dµ
k =1 k =1
Z Zn

= | g| dµ − ∑ ( g − h) dµ
E k =1 Ek
Z n Z

E
| g| dµ − ∑ | g − h| dµ
k=1 Ek
Z
≥ |h| dµ − 2ε.
E
R
The inequality above implies that |ν|( E)R ≥ E |h| dµ − 2ε. Because ε is an arbitrary
positive number, this implies |ν|( E) ≥ E |h| dµ, completing the proof.

Now we justify the terminology total variation measure.

9.11 total variation measure is a measure

Suppose ν is a complex measure on a measurable space ( X, S). Then the total


variation function |ν| is a (positive) measure on ( X, S).

Proof The definition of |ν| and 9.3(a) imply that |ν|(∅) = 0.


To show that |ν| is countably additive, suppose A1 , A2 , . . . are disjoint sets in S .
Fix m ∈ Z+. For each k ∈ {1, . . . , m}, suppose E1,k , . . . , Enk ,k are disjoint sets in S
such that
9.12 E1,k ∪ . . . ∪ Enk ,k ⊂ Ak .
Then { Ej,k : 1 ≤ k S
≤ m and 1 ≤ j ≤ nk } is a disjoint collection of sets in S that
are all contained in ∞k=1 Ak . Hence
m nk ∞
[ 
∑ ∑ |ν(Ej,k )| ≤ |ν| Ak .
k =1 j =1 k =1

Taking the supremum of the left side of the inequality above over all choices of { Ej,k }
satisfying 9.12 shows that
m [∞ 
∑ |ν|( Ak ) ≤ |ν| Ak .
k =1 k =1

Because the inequality above holds for all m ∈ Z+, we have


∞ [∞ 
∑ |ν|( Ak ) ≤ |ν| Ak .
k =1 k =1
262 Chapter 9 Real and Complex Measures

To prove the inequality above in the other


S
direction, suppose E1 , . . . , En ∈ S are
disjoint sets such that E1 ∪ · · · ∪ En ⊂ ∞k=1 Ak . Then
∞ ∞ n
∑ |ν|( Ak ) ≥ ∑ ∑ |ν(Ej ∩ Ak )|
k =1 k =1 j =1

n ∞
= ∑ ∑ |ν(Ej ∩ Ak )|
j =1 k =1

n ∞

≥ ∑ ∑ ν( Ej ∩ Ak )
j =1 k =1

n
= ∑ |ν(Ej )|,
j =1

where the first line above follows from the definition of |ν|( Ak ) and the last line
above follows from the countable additivity of ν. S 
The inequality above and the definition of |ν| ∞ k=1 Ak imply that

∞ ∞
[ 
∑ |ν|( Ak ) ≥ |ν| Ak ,
k =1 k =1

completing the proof.

The Banach Space of Measures


In this subsection, we make the set of complex or real measures on a measurable
space into a vector space and then into a Banach space.

9.13 Definition addition and scalar multiplication of measures

Suppose ( X, S) is a measurable space. For complex measures ν, µ on ( X, S)


and α ∈ F, define complex measures ν + µ and αν on ( X, S) by

(ν + µ)( E) = ν( E) + µ( E) and (αν)( E) = α ν( E) .

You should verify that if ν, µ, and α are as above, then ν + µ and αν are complex
measures on ( X, S). You should also verify that these natural definitions of addition
and scalar multiplication make the set of complex (or real) measures on a measurable
space ( X, S) into a vector space. We now introduce notation for this vector space.

9.14 Definition MF (S)

Suppose ( X, S) is a measurable space. Then MF (S) denotes the vector space


of real measures on ( X, S) if F = R and denotes the vector space of complex
measures on ( X, S) if F = C.
Section 9A Total Variation 263

We use the terminology total variation norm below even though we have not yet
shown that the object being defined is a norm (especially because it is not obvious
that kνk < ∞ for every complex measure ν). Soon we will justify this terminology.

9.15 Definition total variation norm of a complex measure

Suppose ν is a complex measure on a measurable space ( X, S). The total


variation norm of ν, denoted kνk, is defined by

kνk = |ν|( X ).

9.16 Example total variation norm


• If µ is a finite (positive) measure, then kµk = µ( X ), as you should verify.
• If µ is a (positive) measure, h ∈ L1 (µ), and dν = h dµ, then kνk = khk1 (as
follows from 9.10).
The next result implies that if ν is a complex measure on a measurable space
( X, S), then |ν|( E) < ∞ for every E ∈ S .

9.17 total variation norm is finite

Suppose ( X, S) is a measurable space and ν ∈ MF (S). Then kνk < ∞.

Proof First consider the case where F = R. Thus ν is a real measure on ( X, S). To
begin this proof by contradiction, suppose kνk = |ν|( X ) = ∞.
We inductively choose a decreasing sequence E0 ⊃ E1 ⊃ E2 ⊃ · · · of sets in S
as follows: Start by choosing E0 = X. Now suppose n ≥ 0 and En ∈ S has been
chosen with |ν|( En ) = ∞ and |ν( En )| ≥ n. Because |ν|( En ) = ∞, 9.9 implies that
there exists A ∈ S such that A ⊂ En and |ν( A)| ≥ n + 1 + |ν( En )|, which implies
that
|ν( En \ A)| = |ν( En ) − ν( A)| ≥ |ν( A)| − |ν( En )| ≥ n + 1.
Now
|ν|( A) + |ν|( En \ A) = |ν|( En ) = ∞
because the total variation measure |ν| is a (positive) measure (by 9.11). The equation
above shows that at least one of |ν|( A) and |ν|( En \ A) is ∞. Let En+1 = A
if |ν|( A) = ∞ and let En+1 = En \ A if |ν|( A) < ∞. Thus En ⊃ En+1 ,
|ν|( En+1 ) = ∞, and |ν( En+1 )| T
≥ n +1.
Now 9.7(d) implies that ν ∞ n=1 En = limn→∞ ν ( En ). However, | ν ( En )| ≥ n
for each n ∈ Z+, and thus the limit in the last equation does not exist (in R). This
contradiction completes the proof in the case where ν is a real measure.
Consider now the case where F = C; thus ν is a complex measure on ( X, S).
Then
|ν|( X ) ≤ |Re ν|( X ) + |Im ν|( X ) < ∞,
where the last inequality follows from applying the real case to Re ν and Im ν.
264 Chapter 9 Real and Complex Measures

The previous result tells us that if ( X, S) is a measurable space, then kνk < ∞
for all ν ∈ MF (S). This implies (as the reader should verify) that the total variation
norm k·k is a norm on MF (S). The next result shows that this norm makes MF (S)
into a Banach space (in other words, every Cauchy sequence in this norm converges).

9.18 the set of real or complex measures on ( X, S) is a Banach space

Suppose ( X, S) is a measurable space. Then MF (S) is a Banach space with the


total variation norm.

Proof Suppose ν1 , ν2 , . . . is a Cauchy sequence in MF (S). For each E ∈ S , we


have

|νj ( E) − νk ( E)| = |(νj − νk )( E)|


≤ |νj − νk |( E)
≤ kνj − νk k.
Thus ν1 ( E), ν2 ( E), . . . is a Cauchy sequence in F and hence converges. Thus we can
define a function ν : S → F by

ν( E) = lim νj ( E).
j→∞

To show that ν ∈ MF (S), we must verify that ν is countably additive. To do this,


suppose E1 , E2 , . . . is a disjoint sequence of sets in S . Let ε > 0. Let m ∈ Z+ be
such that

9.19 kνj − νk k ≤ ε for all j, k ≥ m.

If n ∈ Z+ is such that

9.20 ∑ |νm (Ek )| ≤ ε
k=n

[such an n exists by applying 9.3(b) to νm ] and if j ≥ m, then


∞ ∞ ∞
∑ |νj (Ek )| ≤ ∑ |(νj − νm )(Ek )| + ∑ |νm (Ek )|
k=n k=n k=n

≤ ∑ |νj − νm |(Ek ) + ε
k=n

[ 
= |νj − νm | Ek + ε
k=n

9.21 ≤ 2ε,
where the second line uses 9.20, the third line uses the countable additivity of the
measure |νj − νm | (see 9.11), and the fourth line uses 9.19.
Section 9A Total Variation 265

If ε and n are as in the paragraph above, then


[
∞  n −1 ∞
[  n −1

ν Ek − ∑ ν( Ek ) = lim νj Ek − lim ∑ νj (Ek )
j→∞ j→∞
k =1 k =1 k =1 k =1


= lim ∑ νj ( Ek )
j→∞
k=n
≤ 2ε,
where the second line uses the countable additivity of the measure νj and the third line
S 
uses 9.21. The inequality above implies that ν ∞ ∞
k=1 Ek = ∑k=1 ν ( Ek ), completing
the proof that ν ∈ MF (S).
We still need to prove that limk→∞ kν − νk k = 0. To do this, suppose ε > 0. Let
m ∈ Z+ be such that

9.22 kνj − νk k ≤ ε for all j, k ≥ m.


Suppose k ≥ m. Suppose also that E1 , . . . , En ∈ S are disjoint subsets of X. Then
n n
∑ |(ν − νk )(E` )| = jlim
→∞
∑ |(νj − νk )(E` )| ≤ ε,
`=1 `=1

where the last inequality follows from 9.22 and the definition of the total variation
norm. The inequality above implies that kν − νk k ≤ ε, completing the proof.

EXERCISES 9A
1 Prove or give a counterexample: If ν is a real measure on a measurable
space ( X, S) and A, B ∈ S are such that ν( A) ≥ 0 and ν( B) ≥ 0, then
ν( A ∪ B) ≥ 0.
2 Suppose ν is a real measure on ( X, S). Define µ : S → [0, ∞) by

µ( E) = |ν( E)|.

Prove that µ is a (positive) measure on ( X, S) if and only if the range of ν is


contained in [0, ∞) or the range of ν is contained in (−∞, 0].
3 Suppose ν is a complex measure on a measurable space ( X, S). Prove that
|ν|( X ) = ν( X ) if and only if ν is a (positive) measure.
4 Suppose ν is a complex measure on a measurable space ( X, S). Prove that if
E ∈ S then
n∞
|ν|( E) = sup ∑ |ν( Ek )| : E1 , E2 , . . . is a disjoint sequence in S
k =1

[ o
such that E = Ek .
k =1
266 Chapter 9 Real and Complex Measures

5 Suppose µ is a (positive) measure on a measurable space ( X, S) and h is a


nonnegative function in L1 (µ). Let ν be the (positive) measure on ( X, S)
defined by dν = h dµ. Prove that
Z Z
f dν = f h dµ

for all S-measurable functions f : X → [0, ∞].


6 Suppose ( X, S , µ) is a (positive) measure space. Prove that
{h dµ : h ∈ L1 (µ)}
is a closed subspace of MF (S).
7 (a) Suppose B is the collection of Borel subsets of R. Show that the Banach
space MF (B) is not separable.
(b) Give an example of a measurable space ( X, S) such that the Banach space
MF (S) is infinite-dimensional and separable.
8 Suppose t > 0 and λ is Lebesgue measure on the σ-algebra of Borel subsets of
[0, t]. Suppose h : [0, t] → C is the function defined by
h( x ) = cos x + i sin x.
Let ν be the complex measure defined by dν = h dλ.
(a) Show that kνk = t.
(b) Show that if E1 , E2 , . . . is a sequence of disjoint Borel subsets of [0, t], then

∑ |ν(Ek )| < t.
k =1
[This exercise shows that the supremum in the definition of |ν|([0, t]) is not
attained, even if countably many disjoint sets are allowed.]
9 Give an example to show that 9.9 can fail if the hypothesis that ν is a real
measure is replaced by the hypothesis that ν is a complex measure.
10 Suppose ( X, S) is a measurable space with S 6= {∅, X }. Prove that the total
variation norm on MF (S) does not come from an inner product. In other
words, show that there does not exist an inner product h·, ·i on MF (S) such
that kνk = hν, νi1/2 for all ν ∈ MF (S), where k·k is the usual total variation
norm on MF (S).
11 For ( X, S) a measurable space and b ∈ X, define a finite (positive) measure δb
on ( X, S) by (
1 if b ∈ E,
δb ( E) =
0 if b ∈
/E
for E ∈ S .
(a) Show that if b, c ∈ X, then kδb + δc k = 2.
(b) Give an example of a measurable space ( X, S) and b, c ∈ X with b 6= c
such that kδb − δc k 6= 2.
Section 9B Decomposition Theorems 267

9B Decomposition Theorems
Hahn Decomposition Theorem
The next result shows that a real measure on a measurable space ( X, S) decomposes
X into two disjoint measurable sets such that every measurable subset of one of these
two sets has nonnegative measure and every measurable subset of the other set has
nonpositive measure.
The decomposition in the result below is not unique because a subset D of X with
|ν|( D ) = 0 could be shifted from A to B or from B to A. However, Exercise 1 at
the end of this section shows that the Hahn decomposition is almost unique.

9.23 Hahn Decomposition Theorem

Suppose ν is a real measure on a measurable space ( X, S). Then there exist sets
A, B ∈ S such that

(a) A ∪ B = X and A ∩ B = ∅;
(b) ν( E) ≥ 0 for every E ∈ S with E ⊂ A;
(c) ν( E) ≤ 0 for every E ∈ S with E ⊂ B.

9.24 Example Hahn decomposition


Suppose µ is a (positive) measure on a measurable space ( X, S), h ∈ L1 (µ) is
real valued, and dν = h dµ. Then a Hahn decomposition of the real measure ν is
obtained by setting

A = { x ∈ X : h ( x ) ≥ 0} and B = { x ∈ X : h ( x ) < 0}.

Proof of 9.23 Let


a = sup{ν( E) : E ∈ S}.
Thus a ≤ kνk < ∞, where the last inequality comes from 9.17. For each j ∈ Z+, let
A j ∈ S be such that

1
9.25 ν( A j ) ≥ a − .
2j
Temporarily fix k ∈ Z+. We will show by induction on n that if n ∈ Z+ with
n ≥ k, then
n
[  n
1
9.26 ν Aj ≥ a − ∑ j .
j=k j=k 2

To get started with the induction, note that if n = k then 9.26 holds because in this
case 9.26 becomes 9.25. Now for the induction step, assume that n ≥ k and that 9.26
holds. Then
268 Chapter 9 Real and Complex Measures

+1
n[  n
[  n
 [  
ν Aj = ν A j + ν ( A n +1 ) − ν A j ∩ A n +1
j=k j=k j=k
 n
1  1 
≥ a− ∑ + a− −a
j=k 2j 2n +1
n +1
1
= a− ∑ 2j
,
j=k

where the first line follows from 9.7(b) and the second line follows from 9.25 and
9.26. We have now verified that 9.26 holds if n is replaced by n + 1, completing the
proof by induction of 9.26.
The sequence of sets Ak , Ak ∪ Ak+1 , Ak ∪ Ak+1 ∪ Ak+2 , . . . is increasing. Thus
taking the limit as n → ∞ of both sides of 9.26 and using 9.7(c) gives

[  1
9.27 ν Aj ≥ a − .
j=k
2k −1

Now let
∞ [
\ ∞
A= Aj.
k =1 j = k
S S
The sequence of sets ∞ ∞
j=1 A j , j=2 A j , . . . is decreasing. Thus 9.27 and 9.7(d) imply
that ν( A) ≥ a. The definition of a now implies that

ν( A) = a.

Suppose E ∈ S and E ⊂ A. Then ν( A) = a ≥ ν( A \ E). Thus we have


ν( E) = ν( A) − ν( A \ E) ≥ 0, which proves (b).
Let B = X \ A; thus (a) holds. Suppose E ∈ S and E ⊂ B. Then we have
ν( A ∪ E) ≤ a = ν( A). Thus ν( E) = ν( A ∪ E) − ν( A) ≤ 0, which proves (c).

Jordan Decomposition Theorem


You should think of two complex or positive measures on a measurable space ( X, S)
as being singular with respect to each other if the two measures live on different sets.
Here is the formal definition.

9.28 Definition singular measures

Suppose ν and µ are complex or positive measures on a measurable space ( X, S).


Then ν and µ are called singular with respect to each other, denoted ν ⊥ µ, if
there exist sets A, B ∈ S such that
• A ∪ B = X and A ∩ B = ∅;
• ν( E) = ν( E ∩ A) and µ( E) = µ( E ∩ B) for all E ∈ S .
Section 9B Decomposition Theorems 269

9.29 Example singular measures


Suppose λ is Lebesgue measure on the σ-algebra B of Borel subsets of R.
• Define positive measures ν, µ on (R, B) by
ν( E) = | E ∩ (−∞, 0)| and µ( E) = | E ∩ (2, 3)|
for E ∈ B . Then ν ⊥ µ because ν lives on (−∞, 0) and µ lives on [0, ∞).
Neither ν nor µ is singular with respect to λ.
• Let r1 , r2 , . . . be a list of the rational numbers. Suppose w1 , w2 , . . . is a bounded
sequence of complex numbers. Define a complex measure ν on (R, B) by
wk
ν( E) = ∑ 2k
{k ∈Z+ : r ∈ E}
k

for E ∈ B . Then ν ⊥ λ because ν lives on Q and λ lives on R \ Q.


The hard work for proving the next result has already been done in proving the
Hahn Decomposition Theorem (9.23).

9.30 Jordan Decomposition Theorem

• Every real measure is the difference of two finite (positive) measures that are
singular with respect to each other.
• More precisely, suppose ν is a real measure on a measurable space ( X, S).
Then there exist unique finite (positive) measures ν+ and ν− on ( X, S) such
that

9.31 ν = ν+ − ν− and ν+ ⊥ ν− .

Furthermore,
|ν| = ν+ + ν− .

Proof Let X = A ∪ B be a Hahn decomposition of ν as in 9.23. Define functions


ν+ : S → [0, ∞) and ν− : S → [0, ∞) by
ν+ ( E) = ν( E ∩ A) and ν − ( E ) = − ν ( E ∩ B ).
The countable additivity of ν implies ν+ and ν− are finite (positive) measures on
( X, S), with ν = ν+ − ν− and ν+ ⊥ ν− .
The definition of the total vari-
Camille Jordan (1838–1922) is also
ation measure and 9.31 imply that
known for certain matrices that are
|ν| = ν+ + ν− , as you should verify.
0 except along the diagonal and the
The equations ν = ν+ − ν− and
line above it.
|ν| = ν+ + ν− imply that
|ν| + ν |ν| − ν
ν+ = and ν− = .
2 2
Thus the finite (positive) measures ν+ and ν− are uniquely determined by ν and the
conditions in 9.31.
270 Chapter 9 Real and Complex Measures

Lebesgue Decomposition Theorem


The next definition captures the notion of one measure having more sets of measure 0
than another measure.

9.32 Definition absolutely continuous; 

Suppose ν is a complex measure on a measurable space ( X, S) and µ is a


(positive) measure on ( X, S). Then ν is called absolutely continuous with respect
to µ, denoted ν  µ, if

ν( E) = 0 for every set E ∈ S with µ( E) = 0.

9.33 Example absolute continuity


The reader should verify all the following examples:
• If µ is a (positive) measure and h ∈ L1 (µ), then h dµ  µ.
• If ν is a real measure, then ν+  |ν| and ν−  |ν|.
• If ν is a complex measure, then ν  |ν|.
• If ν is a complex measure, then Re ν  |ν| and Im ν  |ν|.
• Every measure on a measurable space ( X, S) is absolutely continuous with
respect to counting measure on ( X, S).

The next result should help you think that absolute continuity and singularity are
two extreme possibilities for the relationship between two complex measures.

9.34 absolutely continuous and singular implies 0 measure

Suppose µ is a (positive) measure on a measurable space ( X, S). Then the only


complex measure on ( X, S) that is both absolutely continuous and singular with
respect to µ is the 0 measure.

Proof Suppose ν is a complex measure on ( X, S) such that ν  µ and ν ⊥ µ. Thus


there exist sets A, B ∈ S such that A ∪ B = X, A ∩ B = ∅, and ν( E) = ν( E ∩ A)
and µ( E) = µ( E ∩ B) for every E ∈ S .
Suppose E ∈ S . Then

µ( E ∩ A) = µ ( E ∩ A) ∩ B = µ(∅) = 0.

Because ν  µ, this implies that ν( E ∩ A) = 0. Thus ν( E) = 0. Hence ν is the 0


measure.

Our next result states that a (positive) measure on a measurable space ( X, S)


determines a decomposition of each complex measure on ( X, S) as the sum of the
two extreme types of complex measures (absolute continuity and singularity).
Section 9B Decomposition Theorems 271

9.35 Lebesgue Decomposition Theorem

Suppose µ is a (positive) measure on a measurable space ( X, S).

• Every complex measure on ( X, S) is the sum of a complex measure


absolutely continuous with respect to µ and a complex measure singular
with respect to µ.
• More precisely, suppose ν is a complex measure on ( X, S). Then there exist
unique complex measures νa and νs on ( X, S) such that ν = νa + νs and

νa  µ and νs ⊥ µ.

Proof Let
b = sup{|ν|( B) : B ∈ S and µ( B) = 0}.
For each k ∈ Z+, let Bk ∈ S be such that
1
|ν|( Bk ) ≥ b − k and µ( Bk ) = 0.
Let

[
B= Bk .
k =1
Then µ( B) = 0 and |ν|( B) = b.
Let A = X \ B. Define complex measures νa and νs on ( X, S) by
νa ( E) = ν( E ∩ A) and νs ( E) = ν( E ∩ B).
Clearly ν = νa + νs .
If E ∈ S , then
µ ( E ) = µ ( E ∩ A ) + µ ( E ∩ B ) = µ ( E ∩ A ),
where the last equality holds because µ( B) = 0. The equation above implies that
νs ⊥ µ.
To prove that νa  µ, suppose E ∈ S and µ( E) = 0. Then µ( B ∪ E) = 0 and
hence
b ≥ |ν|( B ∪ E) = |ν|( B) + |ν|( E \ B) = b + |ν|( E \ B),
which implies that |ν|( E \ B) = 0. Thus

νa ( E) = ν( E ∩ A) = ν( E \ B) = 0, The construction of νa and νs shows


that if ν is a positive (or real)
which implies that νa  µ. measure, then so are νa and νs .
We have now proved all parts of this result except the uniqueness of the Lebesgue
decomposition. To prove the uniqueness, suppose ν1 and ν2 are complex measures
on ( X, S) such that ν1  µ, ν2 ⊥ µ, and ν = ν1 + ν2 . Then
ν1 − νa = νs − ν2 .
The left side of the equation above is absolutely continuous with respect to µ and the
right side is singular with respect to µ. Thus both sides are both absolutely continuous
and singular with respect to µ. Thus 9.34 implies that ν1 = νa and ν2 = νs .
≈ 100e Chapter 9 Real and Complex Measures

Radon–Nikodym Theorem
If µ is a (positive) measure, h ∈ L1 (µ),
The result below was first proved by
and dν = h dµ, then ν  µ. The next
Radon and Otto Nikodym
result gives the important converse—if µ
(1887–1974).
is σ-finite, then every complex measure
that is absolutely continuous with respect to µ is of the form h dµ for some h ∈ L1 (µ).
The hypothesis that µ is σ-finite cannot be deleted.

9.36 Radon–Nikodym Theorem

Suppose µ is a (positive) σ-finite measure on a measurable space ( X, S). Suppose


ν is a complex measure on ( X, S) such that ν  µ. Then there exists h ∈ L1 (µ)
such that dν = h dµ.

Proof First consider the case where both µ and ν are finite (positive) measures.
Define ϕ : L2 (ν + µ) → R by
Z
9.37 ϕ( f ) = f dν.

To show that ϕ is well defined, first note that if f ∈ L2 (ν + µ), then


Z Z 1/2
9.38 | f | dν ≤ | f | d( ν + µ ) ≤ ν ( X ) + µ ( X ) k f k L2 (ν+µ) < ∞,

where the middle inequality follows from RHölder’s inequality (7.9) applied to the
functions 1 and f . Now 9.38 shows that f dν makes sense for f ∈ L2 (ν + µ).
Furthermore, if two functions in L2 (ν + µ) differ only on a set of (ν + µ)-measure
0, then they differ only on a set of ν-measure 0. Thus ϕ as defined in 9.37 makes
sense as a linear functionalRon L2 (ν + µ).
Because | ϕ( f )| ≤ | f | dν, 9.38
The clever idea of using Hilbert
shows that ϕ is a bounded linear func-
2 space techniques in this proof comes
tional on L (ν + µ). The Riesz Represen-
from John von Neumann
tation Theorem (8.47) now implies that
2 (1903–1957).
there exists g ∈ L (ν + µ) such that
Z Z
f dν = f g d( ν + µ )

for all f ∈ L2 (ν + µ). Hence


Z Z
9.39 f (1 − g) dν = f g dµ

for all f ∈ L2 (ν + µ).


If f equals the characteristic function of { x ∈ X : g( x ) ≥ 1}, then the left side
of 9.39 is less than or equal to 0 Rand the right side of 9.39 is greater than or equal to
0; hence both sides are 0. Thus f g dµ = 0, which implies (with this choice of f )
that µ({ x ∈ X : g( x ) ≥ 1}) = 0.
Section 9B Decomposition Theorems 273

Similarly, if f equals the characteristic function of { x ∈ X : g( x ) < 0}, then the


left side of 9.39 is greater than or equal to 0R and the right side of 9.39 is less than
or equal to 0; hence both sides are 0. Thus f g dµ = 0, which implies (with this
choice of f ) that µ({ x ∈ X : g( x ) < 0}) = 0.
Because ν  µ, the two previous paragraphs imply that
ν({ x ∈ X : g( x ) ≥ 1}) = 0 and ν({ x ∈ X : g( x ) < 0}) = 0.

Thus we can modify g (for example by redefining g to be 12 on the two sets appearing
above; both those sets have ν-measure 0 and µ-measure 0) and from now on we can
assume that 0 ≤ g( x ) < 1 for all x ∈ X and that 9.39 holds for all f ∈ L2 (ν + µ).
Hence we can define h : X → [0, ∞) by
g( x )
h( x ) = .
1 − g( x )
Suppose E ∈ S . For each k ∈ Z+, let
 Taking f = χ E/(1 − g) in 9.39
R
 χ E( x ) χ (x)
if 1−Eg( x) ≤ k, would give ν( E) = E h dµ, but this
f k (x) = 1 − g ( x )
function f might not be in
0 otherwise. L2 (ν + µ) and thus we need to be a
bit more careful.
Then f k ∈ L2 (ν + µ). Now 9.39 implies
Z Z
f k (1 − g) dν = f k g dµ.

Taking the limit as k → ∞ and using the Monotone Convergence Theorem (3.11)
shows that
Z Z
9.40 1 dν = h dµ.
E E

Thus dν = h dµ, completing the proof in the case where both ν and µ are (positive)
finite measures [note that h ∈ L1 (µ) because h is a nonnegative function and we can
take E = X in the equation above].
Now relax the assumption on µ to the hypothesis that µ is a σ-finite measure.
Thus
S∞
there exists an increasing sequence X1 ⊂ X2 ⊂ · · · of sets in S such that
k=1 Xk = X and µ ( Xk ) < ∞ for each k ∈ Z . For k ∈ Z , let νk and µk denote
+ +
the restrictions of ν and µ to the σ-algebra on Xk consisting of those sets in S that
are subsets of Xk . Then νk  µk . Thus by the case we have already proved, there
exists a nonnegative function hk ∈ L1 (µk ) such that dνk = hk dµk . If j < k, then
Z Z
h j dµ = ν( E) = hk dµ
E E

for every set E ∈ S with E ⊂ X j ; thus µ({ x ∈ X j : h j ( x ) 6= hk ( x )}) = 0. Hence


there exists an S -measurable function h : X → [0, ∞) such that

µ { x ∈ Xk : h( x ) 6= hk ( x )} = 0

for every k ∈ Z+. The Monotone Convergence Theorem (3.11) can now be used to
show that 9.40 holds for every E ∈ S . Thus dν = h dµ, completing the proof in the
case where ν is a (positive) finite measure.
274 Chapter 9 Real and Complex Measures

Now relax the assumption on ν to the assumption that ν is a real measure. The
measure ν equals one-half the difference of the two positive (finite) measures |ν| + ν
and |ν| − ν, each of which is absolutely continuous with respect to µ. By the case
proved in the previous paragraph, there exist h+ , h− ∈ L1 (µ) such that

d(|ν| + ν) = h+ dµ and d(|ν| − ν) = h− dµ.

Taking h = 12 (h+ − h− ), we have dν = h dµ, completing the proof in the case


where ν is a real measure.
Finally, if ν is a complex measure, apply the result in the previous paragraph to the
real measures Re ν, Im ν, producing hRe , hIm ∈ L1 (µ) such that d(Re ν) = hRe dµ
and d(Im ν) = hIm dµ. Taking h = hRe + ihIm , we have dν = h dµ, completing
the proof in the case where ν is a complex measure.

The function h provided by the Radon–Nikodym Theorem is unique up to changes


on sets with µ-measure 0. If we think of h as an element of L1 (µ) instead of L1 (µ),
then the choice of h is unique.

When dν = h dµ, the notation h = dµ is used by some authors, and h is called
the Radon–Nikodym derivative of ν with respect to µ.
The next result is a nice consequence of the Radon–Nikodym Theorem.

9.41 if ν is a complex measure, then dν = h d|ν| for some h with |h( x )| = 1

(a) Suppose ν is a real measure on a measurable space ( X, S). Then there exists
an S -measurable function h : X → {−1, 1} such that dν = h d|ν|.
(b) Suppose ν is a complex measure on a measurable space ( X, S). Then there
exists an S -measurable function h : X → {z ∈ C : |z| = 1} such that
dν = h d|ν|.

Proof Because ν  |ν|, the Radon–Nikodym Theorem (9.36) tells us that there
exists h ∈ L1 (|ν|) (with h real valued if ν is a real measure) such that dν = h d|ν|.
Now 9.10 implies that d|ν| = |h| d|ν|, which implies that |h| = 1 almost everywhere
(with respect to |ν|). Refine h to be 1 on the set { x ∈ X : |h( x )| 6= 1}, which gives
the desired result.

We could have proved part (a) of the result above by taking h = χ A − χ B in the
Hahn Decomposition Theorem (9.23).
Conversely, we could give a new proof of the Hahn Decomposition Theorem by
using part (a) of the result above and taking

A = { x ∈ X : h ( x ) = 1} and B = { x ∈ X : h ( x ) = −1}.

We could also give a new proof of the Jordan Decomposition Theorem (9.30) by
using part (a) of the result above and taking

ν + = χ { x ∈ X : h ( x ) = 1} d | ν | and ν − = χ { x ∈ X : h ( x ) = −1} d | ν | .
Section 9B Decomposition Theorems 275

Dual Space of L p (µ)


Recall that the dual space of a normed vector space V is the Banach space of bounded
linear functionals on V; the dual space of V is denoted by V 0. Recall also that if
1 ≤ p ≤ ∞, then the dual exponent p0 is defined by the equation 1p + p10 = 1.
0
The dual space of ` p can be identified with ` p for 1 ≤ p < ∞, as we saw in 7.26.
We are now ready to prove the analogous result for an arbitrary (positive) measure,
0
identifying the dual space of L p (µ) with L p (µ) [with the mild restriction that µ is
σ-finite if p = 1]. In the special case where µ is counting measure on Z+, this new
result reduces to the previous result about ` p .
For 1 < p < ∞, the next result differs from 7.25 by only one word, with “to” in
7.25 changed to “onto” below. Thus we already know (and will use in the proof)
0 0
that the map h 7→ ϕh is a one-to-one linear map from L p (µ) to L p (µ) and that
0
k ϕh k = khk p0 for all h ∈ L p (µ). The new aspect of the result below is the assertion
0
that every bounded linear functional on L p (µ) is of the form ϕh for some h ∈ L p (µ).
The key tool we use in proving this new assertion is the Radon–Nikodym Theorem.

0
9.42 dual space of L p (µ) is L p (µ)

Suppose µ is a (positive) measure and 1 ≤ p < ∞ [with the additional hypothesis


0
that µ is a σ-finite measure if p = 1]. For h ∈ L p (µ), define ϕh : L p (µ) → F by
Z
ϕh ( f ) = f h dµ.

0 0
Then h 7→ ϕh is a one-to-one linear map from L p (µ) onto L p (µ) . Further-
0
more, k ϕh k = k hk p0 for all h ∈ L p (µ).

Proof The case p = 1 is left to the reader as an exercise. Thus assume that
1 < p < ∞.
Suppose µ is a (positive) measure on a measurable space ( X, S) and ϕ is a
0
bounded linear functional on L p (µ); in other words, suppose ϕ ∈ L p (µ) .
Consider first the case where µ is a finite (positive) measure. Define a function
ν : S → F by
ν ( E ) = ϕ ( χ E ).
If E1 , E2 , . . . are disjoint sets in S , then

[     ∞  ∞ ∞
ν Ek = ϕ χS∞
k =1 Ek
=ϕ ∑ χE
k
= ∑ ϕ(χ E ) =
k
∑ ν(Ek ),
k =1 k =1 k =1 k =1

where the infinite sum in the third term converges in the L p (µ)-norm to χS∞ E , and
k =1 k
the third equality holds because ϕ is a continuous linear functional. The equation
above shows that ν is countably additive. Thus ν is a complex measure on ( X, S)
[and is a real measure if F = R].
276 Chapter 9 Real and Complex Measures

If E ∈ S and µ( E) = 0, then χ E is the 0 element of L p (µ), which implies that


ϕ(χ E) = 0, which means that ν( E) = 0. Hence ν  µ. By the Radon–Nikodym
Theorem (9.36), there exists h ∈ L1 (µ) such that dν = h dµ. Hence
Z Z
ϕ(χ E) = ν( E) = h dµ = χ Eh dµ
E

for every E ∈ S . The equation above, along with the linearity of ϕ, implies that
Z
9.43 ϕ( f ) = f h dµ for every simple S -measurable function f : X → F.

Because every bounded S -measurable function is the uniform limit on X of a


sequence of simple S -measurable functions (see 2.89), we can conclude from 9.43
that
Z
9.44 ϕ( f ) = f h dµ for every f ∈ L∞ (µ).

For k ∈ Z+, let


Ek = { x ∈ X : 0 < |h( x )| ≤ k}
and define f k ∈ L p (µ) by
( 0
h( x ) |h( x )| p −2 if x ∈ Ek ,
9.45 f k (x) =
0 otherwise.

Now
Z Z 1/p
0 0
|h| p χ E dµ = ϕ( f k ) ≤ k ϕk k f k k p = k ϕk |h| p χ E dµ ,
k k

where the first equality follows from 9.44 and 9.45, and the last equality follows from
0
9.45 [which implies that | f k ( x )| p = |h( x )| p χ E ( x ) for x ∈ X]. After dividing by
k
R 0
1/p
|h| p χ E dµ , the inequality between the first and last terms above becomes
k

khχ E k p0 ≤ k ϕk.
k

Taking the limit as k → ∞ shows, via the Monotone Convergence Theorem (3.11),
that
k h k p 0 ≤ k ϕ k.
0
Thus h ∈ L p (µ). Because each f ∈ L p (µ) can be approximated in the L p (µ) norm
by functions in L∞ (µ), 9.44 now shows that ϕ = ϕh , completing the proof in the
case where µ is a finite (positive) measure.
Now relax the assumption that µ is a finite (positive) measure to the hypothesis
that µ is a (positive) measure. For E ∈ S , let S E = { A ∈ S : A ⊂ E} and let µ E
be the (positive) measure on ( E, S E ) defined by µ E ( A) = µ( A) for A ∈ S E . We
can identify L p (µ E ) with the subspace of functions in L p (µ) that vanish (almost
everywhere) outside E. With this identification, let ϕ E = ϕ| L p (µE ) . Then ϕ E is a
bounded linear functional on L p (µ E ) and k ϕ E k ≤ k ϕk.
Section 9B Decomposition Theorems 277

If E ∈ S and µ( E) < ∞, then the finite measure case that we have already proved
0
as applied to ϕ E implies that there exists a unique h E ∈ L p (µ E ) such that
Z
9.46 ϕ( f ) = f h E dµ for all f ∈ L p (µ E ).
E
If D, E ∈ S and D ⊂ E with µ( E) < ∞, then h D ( x ) = h E ( x ) for almost every
x ∈ D (use the uniqueness part of the result).
For each k ∈ Z+, there exists f k ∈ L p (µ) such that
9.47 k f k k p ≤ 1 and | ϕ( f k )| > k ϕk − 1k .
The Dominated Convergence Theorem (3.31) implies that

lim f k χ{x ∈ X : | f (x)| > 1 } − f k p = 0
n→∞ k n

for each k ∈ Z+. Thus we can replace f k by f k χ{x ∈ X : | f 1 for sufficiently


k ( x )| > n}
large n and still have 9.47 hold. In other words, for each k ∈ Z+,
we can assume that
there exists nk ∈ Z+ such that for each x ∈ X, either | f k ( x )| > 1/nk or f k ( x ) = 0.
Set Dk = { x ∈ X : | f k ( x )| > 1/nk }. Then µ( Dk ) < ∞ [because f k ∈ L p (µ)]
and
9.48 f k ( x ) = 0 for all x ∈ X \ Dk .
For k ∈ Z+, let Ek = D1 ∪ · · · ∪ Dk . Because E1 ⊂ E2 ⊂ · · · , we see that if
j < k, then h Ej ( x ) = h Ek ( x ) for almost every x ∈ Ej . Also, 9.47 and 9.48 imply
that
9.49 lim kh Ek k p0 = lim k ϕ Ek k = k ϕk.
k→∞ k→∞
S∞
Let E = k=1 Ek . Let h be the function that equals hk almost everywhere on Ek
for each k ∈ Z+ and equals 0 on X \ E. The Monotone Convergence Theorem and
9.49 show that
k h k p 0 = k ϕ k.
If f ∈ L p (µ E ), then limk→∞ k f − f χ E k p = 0 by the Dominated Convergence
k
Theorem. Thus if f ∈ L p (µ E ), then
Z Z
9.50 ϕ( f ) = lim ϕ( f χ E ) = lim f χ E h dµ = f h dµ,
k→∞ k k→∞ k

where the first equality follows from the continuity of ϕ, the second equality follows
from 9.46 as applied to each Ek [valid because µ( Ek ) < ∞], and the third equality
follows from the Dominated Convergence Theorem.
If D is an S -measurable subset of X \ E with µ( D ) < ∞, then k h D k p0 = 0
because otherwise we would have kh + h D k p0 > k hk p0 and the linear functional on
L p (µ) induced by h + h D would have norm larger than k ϕk even though itR agrees
with ϕ on L p (µ E∪ D ). Because k h D k p0 = 0, we see from 9.50 that ϕ( f ) = f h dµ
for all f ∈ L p (µ E∪ D ).
Every element of L p (µ) can be approximated in norm by elements of L p (µ E )
R of X \ E with finite measure. Thus the previous
plus functions that live on subsets
paragraph implies that ϕ( f ) = f h dµ for all f ∈ L p (µ), completing the proof.
278 Chapter 9 Real and Complex Measures

EXERCISES 9B
1 Suppose ν is a real measure on a measurable space ( X, S). Prove that the Hahn
decomposition of ν is almost unique, in the sense that if A, B and A0 , B0 are
pairs satisfying the Hahn Decomposition Theorem (9.23), then

|ν|( A \ A0 ) = |ν|( A0 \ A) = |ν|( B \ B0 ) = |ν|( B0 \ B) = 0.

2 Suppose µ is a (positive) measure and g, h ∈ L1 (µ). Prove that g dµ ⊥ h dµ if


and only if g( x )h( x ) = 0 for almost every x ∈ X.
3 Suppose ν and µ are complex measures on a measurable space ( X, S). Show
that the following are equivalent:
(a) ν ⊥ µ.
(b) |ν| ⊥ |µ|.
(c) Re ν ⊥ µ and Im ν ⊥ µ.

4 Suppose ν and µ are complex measures on a measurable space ( X, S). Prove


that if ν ⊥ µ, then |ν + µ| = |ν| + |µ| and kν + µk = kνk + kµk.
5 Suppose ν and µ are finite (positive) measures. Prove that ν ⊥ µ if and only if
k ν − µ k = k ν k + k µ k.
6 Suppose µ is a complex or positive measure on a measurable space ( X, S).
Prove that
{ν ∈ MF (S) : ν ⊥ µ}
is a closed subspace of MF (S).
7 Use the Cantor set to prove that there exists a (positive) measure ν on (R, B)
such that ν ⊥ λ and ν(R) 6= 0 but ν({ x }) = 0 for every x ∈ R; here λ
denotes Lebesgue measure on the σ-algebra B of Borel subsets of R.
[The second bullet point in Example 9.29 does not provide an example of the
desired behavior because in that example, ν({rk }) 6= 0 for all k ∈ Z+ with
wk 6= 0.]
8 Suppose ν is a real measure on a measurable space ( X, S). Prove that

ν+ ( E) = sup{ν( D ) : D ∈ S and D ⊂ E}

and
ν− ( E) = − inf{ν( D ) : D ∈ S and D ⊂ E}.

9 Suppose µ is a (positive) finite measure on a measurable space ( X, S) and h is a


nonnegative function in L1 (µ). Thus h dµ  dµ. Find a reasonable condition
on h that is equivalent to the condition dµ  h dµ.
Section 9B Decomposition Theorems 279

10 Suppose µ is a (positive) measure on a measurable space ( X, S) and ν is a


complex measure on ( X, S). Show that the following are equivalent:
(a) ν  µ.
(b) |ν|  µ.
(c) Re ν  µ and Im ν  µ.

11 Suppose µ is a (positive) measure on a measurable space ( X, S) and ν is a real


measure on ( X, S). Show that ν  µ if and only if ν+  µ and ν−  µ.
12 Suppose µ is a (positive) measure on a measurable space ( X, S). Prove that

{ν ∈ MF (S) : ν  µ}

is a closed subspace of MF (S).


13 Give an example to show that the Radon–Nikodym Theorem (9.36) can fail if
the σ-finite hypothesis is eliminated.
14 Suppose µ is a (positive) σ-finite measure on a measurable space ( X, S) and ν
is a complex measure on ( X, S). Show that the following are equivalent:
(a) ν  µ.
(b) for every ε > 0, there exists δ > 0 such that |ν( E)| < ε for every set
E ∈ S with µ( E) < δ.
(c) for every ε > 0, there exists δ > 0 such that |ν|( E) < ε for every set
E ∈ S with µ( E) < δ.

15 Prove 9.42 [with the extra hypothesis that µ is a σ-finite (positive) measure] in
the case where p = 1.
16 Explain where the proof of 9.42 fails if p = ∞.
17 Prove that if µ is a (positive) measure and 1 < p < ∞, then L p (µ) is reflexive.
[See the definition before Exercise 19 in Section 7B for the meaning of reflexive.]
18 Prove that L1 (R) is not reflexive.
Chapter 10
Linear Maps on Hilbert Spaces

A special tool called the adjoint helps provide insight into the behavior of linear maps
on Hilbert spaces. This chapter begins with a study of the adjoint and its connection
to the null space and range of a linear map.
Then we discuss various issues connected with the invertibility of operators on
Hilbert spaces. These issues lead to the spectrum, which is a set of numbers that
gives important information about an operator.
This chapter then looks at special classes of operators on Hilbert spaces: self-
adjoint operators, normal operators, isometries, unitary operators, integral operators,
and compact operators.
Even on infinite-dimensional Hilbert spaces, compact operators display many
characteristics expected from finite-dimensional linear algebra. We will see that
the powerful Spectral Theorem for compact operators greatly resembles the finite-
dimensional version. Also, we develop the Singular Value Decomposition for an
arbitrary compact operator, again quite similar to the finite-dimensional result.

The Botanical Garden at Uppsala University (the oldest university in Sweden,


founded in 1477), where Erik Fredholm (1866–1927) was a student. The theorem
called the Fredholm Alternative, which we prove in this chapter, states that a
compact operator minus a nonzero scalar multiple of the identity operator
is injective if and only if it is surjective.
CC-BY-SA Per Enström

280
Section 10A Adjoints and Invertibility 281

10A Adjoints and Invertibility


Adjoints of Linear Maps on Hilbert Spaces
The next definition provides a key tool for studying linear maps on Hilbert spaces.

10.1 Definition adjoint; T ∗

Suppose V and W are Hilbert spaces and T : V → W is a bounded linear map.


The adjoint of T is the function T ∗ : W → V such that

h T f , gi = h f , T ∗ gi

for every f ∈ V and every g ∈ W.

To see why the definition above makes


The word adjoint has two unrelated
sense, fix g ∈ W. Consider the linear
meanings in linear algebra. We need
functional on V defined by f 7→ h T f , gi.
only the meaning defined above.
This linear functional is bounded because

|h T f , gi| ≤ k T f k k gk ≤ k T k k gk k f k

for all f ∈ V; thus the linear functional f 7→ h T f , gi has norm at most k T k k gk. By
the Riesz Representation Theorem (8.47), there exists a unique element of V (with
norm at most k T k k gk) such that this linear functional is given by taking the inner
product with it. We call this unique element T ∗ g. In other words, T ∗ g is the unique
element of V such that

10.2 h T f , gi = h f , T ∗ gi

for every f ∈ V. Furthermore,

10.3 k T ∗ gk ≤ k T kk gk.

In 10.2, notice that the inner product on the left is the inner product in W and the
inner product on the right is the inner product in V.

10.4 Example multiplication operators


Suppose ( X, S , µ) is a measure space and h ∈ L∞ (µ). Define the multiplication
operator Mh : L2 (µ) → L2 (µ) by

Mh f = f h.

Then Mh is a bounded linear map and k Mh k ≤ khk∞ . Because


Z
The complex conjugates that appear
h Mh f , g i = f hg dµ = h f , Mh gi in this example are unnecessary (but
they do no harm) if F = R.
for all f , g ∈ L2 (µ), we have Mh ∗ = Mh .
282 Chapter 10 Linear Maps on Hilbert Spaces

10.5 Example linear maps induced by integration


Suppose ( X, S , µ) and (Y, T , ν) are σ-finite measure spaces and K ∈ L2 (µ × ν).
Define a linear map IK : L2 (ν) → L2 (µ) by
Z
10.6 (IK f )( x ) = K ( x, y) f (y) dν(y)
Y

for f ∈ L2 (ν) and x ∈ X. To see that this definition makes sense, first note that
there are no worrisome measurability issues because for each x ∈ X, the function
y 7→ K ( x, y) is a T -measurable function on Y (see 5.9).
Suppose f ∈ L2 (ν). Use the Cauchy–Schwarz inequality (8.11) or Hölder’s
inequality (7.9) to show that
Z Z 1/2
10.7 |K ( x, y)| | f (y)| dν(y) ≤ |K ( x, y)|2 dν(y) k f k L2 ( ν ) .
Y Y
for every x ∈ X. Squaring both sides of the inequality above and then integrating on
X with respect to µ gives
Z Z 2 Z Z 
|K ( x, y)| | f (y)| dν(y) dµ( x ) ≤ |K ( x, y)|2 dν(y) dµ( x ) k f k2L2 (ν)
X Y X Y

= kK k2L2 (µ×ν) k f k2L2 (ν) ,


where the last line holds by Tonelli’s Theorem (5.28). The inequality above implies
that the integral on the left side of 10.7 is finite for µ-almost every x ∈ X. Thus
the integral in 10.6 makes sense for µ-almost every x ∈ X. Now the last inequality
above shows that
Z
kIK f k2L2 (µ) = |(IK f )( x )|2 dµ( x ) ≤ kK k2L2 (µ×ν) k f k2L2 (ν) .
X

Thus IK is a bounded linear map from L2 (ν) to L2 (µ) and


10.8 kIK k ≤ kK k L2 (µ×ν) .
Define K ∗ : Y × X → F by
K ∗ (y, x ) = K ( x, y),
and note that K ∗ ∈ L2 (ν × µ). Thus IK∗ : L2 (µ) → L2 (ν) is a bounded linear map.
Using Tonelli’s Theorem (5.28) and Fubini’s Theorem (5.32), we have
Z Z
hIK f , gi = K ( x, y) f (y) dν(y) g( x ) dµ( x )
X Y
Z Z
= f (y) K ( x, y) g( x ) dµ( x ) dν(y)
Y X
Z
= f (y)IK∗ g(y) dν(y) = h f , IK∗ gi
Y

for all f ∈ L2 (ν) and all g ∈ L2 (µ). Thus


10.9 (IK )∗ = IK∗ .
Section 10A Adjoints and Invertibility 283

10.10 Example linear maps induced by matrices


As a special case of the previous example, suppose m, n ∈ Z+, µ is counting
measure on {1, . . . , m}, ν is counting measure on {1, . . . , n}, and K is an m-by-n
matrix with entry K (i, j) ∈ F in row i, column j. In this case, the linear map
IK : L2 (ν) → L2 (µ) induced by integration is given by the equation
n
(IK f )(i ) = ∑ K(i, j) f ( j)
j =1

for f ∈ L2 ( ν ).
If we identify L2 ( ν )
and L2 (µ) with Fn and Fm and then think of
n m
elements of F and F as column vectors, then the equation above shows that the
linear map IK : Fn → Fm is simply matrix multiplication by K.
In this setting, K ∗ is called the conjugate transpose of K because the n-by-m
matrix K ∗ is obtained by interchanging the rows and the columns of K and then
taking the complex conjugate of each entry.
The previous example now shows that
m n 1/2
kIK k ≤ ∑ ∑ |K (i, j)|2 .
i =1 j =1

Furthermore, the previous example shows that the adjoint of the linear map of
multiplication by the matrix K is the linear map of multiplication by the conjugate
transpose matrix K ∗, a result that may be familiar to you from linear algebra.

If T is a bounded linear map from a Hilbert space V to a Hilbert space W, then


the adjoint T ∗ has been defined as a function from W to V. We now show that the
adjoint T ∗ is linear and bounded. Recall that B(V, W ) denotes the Banach space of
bounded linear maps from V to W.

10.11 T ∗ is a bounded linear map

Suppose V and W are Hilbert spaces and T ∈ B(V, W ). Then

T ∗ ∈ B(W, V ), ( T ∗ )∗ = T, and k T ∗ k = k T k.

Proof Suppose g1 , g2 ∈ W. Then


h f , T ∗ ( g1 + g2 )i = h T f , g1 + g2 i = h T f , g1 i + h T f , g2 i
= h f , T ∗ g1 i + h f , T ∗ g2 i
= h f , T ∗ g1 + T ∗ g2 i
for all f ∈ V. Thus T ∗ ( g1 + g2 ) = T ∗ g1 + T ∗ g2 .
Suppose α ∈ F and g ∈ W. Then
h f , T ∗ (αg)i = h T f , αgi = αh T f , gi = αh f , T ∗ gi = h f , αT ∗ gi
for all f ∈ V. Thus T ∗ (αg) = αT ∗ g.
We have now shown that T ∗ : W → V is a linear map. From 10.3, we see that T ∗
is bounded. In other words, T ∗ ∈ B(W, V ).
284 Chapter 10 Linear Maps on Hilbert Spaces

Because T ∗ ∈ B(W, V ), its adjoint ( T ∗ )∗ : V → W is defined. Suppose f ∈ V.


Then

h( T ∗ )∗ f , gi = h g, ( T ∗ )∗ f i = h T ∗ g, f i = h f , T ∗ gi = h T f , gi

for all g ∈ W. Thus ( T ∗ )∗ f = T f , and hence ( T ∗ )∗ = T.


From 10.3, we see that k T ∗ k ≤ k T k. Applying this inequality with T replaced
by T ∗ we have
k T ∗ k ≤ k T k = k( T ∗ )∗ k ≤ k T ∗ k.
Because the first and last terms above are the same, the first inequality must be an
equality. In other words, we have k T ∗ k = k T k.

Parts (a) and (b) of the next result show that if V and W are real Hilbert spaces,
then the function T 7→ T ∗ from B(V, W ) to B(W, V ) is a linear map. However,
if V and W are nonzero complex Hilbert spaces, then T 7→ T ∗ is not a linear map
because of the complex conjugate in (b).

10.12 properties of the adjoint

Suppose V, W, and U are Hilbert spaces. Then

(a) (S + T )∗ = S∗ + T ∗ for all S, T ∈ B(V, W );


(b) (αT )∗ = α T ∗ for all α ∈ F and all T ∈ B(V, W );
(c) I ∗ = I, where I is the identity operator on V;
(d) (S ◦ T )∗ = T ∗ ◦ S∗ for all T ∈ B(V, W ) and S ∈ B(W, U ).

Proof

(a) The proof of (a) is left to the reader as an exercise.


(b) Suppose α ∈ F and T ∈ B(V, W ). If f ∈ V and g ∈ W, then

h f , (αT )∗ gi = hαT f , gi = αh T f , gi = αh f , T ∗ gi = h f , α T ∗ gi.

Thus (αT )∗ g = α T ∗ g, as desired.


(c) If f , g ∈ V, then
h f , I ∗ g i = h I f , g i = h f , g i.
Thus I ∗ g = g, as desired.
(d) Suppose T ∈ B(V, W ) and S ∈ B(W, U ). If f ∈ V and g ∈ U, then

h f , (S ◦ T )∗ gi = h(S ◦ T ) f , gi = hS( T f ), gi = h Tg, S∗ gi = h f , T ∗ (S∗ g)i.

Thus (S ◦ T )∗ g = T ∗ (S∗ g) = ( T ∗ ◦ S∗ )( g). Hence (S ◦ T )∗ = T ∗ ◦ S∗, as


desired.
Section 10A Adjoints and Invertibility 285

Null Spaces and Ranges in Terms of Adjoints


The next result shows the relationship between the null space and the range of a linear
map and its adjoint. The orthogonal complement of each subset of a Hilbert space is
closed [see 8.40(a)]. However, the range of a bounded linear map on a Hilbert space
need not be closed (see Example 10.15 or Exercises 9 and 10 for examples). Thus in
parts (b) and (d) of the result below, we must take the closure of the range.

10.13 null space and range of T ∗

Suppose V and W are Hilbert spaces and T ∈ B(V, W ). Then

(a) null T ∗ = (range T )⊥ ;


(b) range T ∗ = (null T )⊥ ;
(c) null T = (range T ∗ )⊥ ;
(d) range T = (null T ∗ )⊥ .

Proof We begin by proving (a). Let g ∈ W. Then

g ∈ null T ∗ ⇐⇒ T ∗ g = 0
⇐⇒ h f , T ∗ gi = 0 for all f ∈ V
⇐⇒ h T f , gi = 0 for all f ∈ V
⇐⇒ g ∈ (range T )⊥ .

Thus null T ∗ = (range T )⊥ , proving (a).


If we take the orthogonal complement of both sides of (a), we get (d), where we
have used 8.41. Replacing T with T ∗ in (a) gives (c), where we have used 10.11.
Finally, replacing T with T ∗ in (d) gives (b).

As a corollary of the result above, we have the following result, which gives a
useful way to determine whether or not a linear map has a dense range.

10.14 necessary and sufficient condition for dense range

Suppose V and W are Hilbert spaces and T ∈ B(V, W ). Then T has dense range
if and only if T ∗ is injective.

Proof From 10.13(d) we see that T has dense range if and only if (null T ∗ )⊥ = W,
which happens if and only if null T ∗ = {0}, which happens if and only if T ∗ is
injective.

The advantage of using the result above is that to determine whether or not a
bounded linear map T between Hilbert spaces has a dense range, we need only
determine whether or not 0 is the only solution to the equation T ∗ g = 0. The next
example illustrates this procedure.
286 Chapter 10 Linear Maps on Hilbert Spaces

10.15 Example Volterra operator


The Volterra operator is the linear map V : L2 ([0, 1]) → L2 ([0, 1]) defined by
Z x
(V f )( x ) = f (y) dy
0

for f ∈ L2 ([0, 1]) and x ∈ [0, 1]; here dy means dλ(y), where λ is the usual
Lebesgue measure on the interval [0, 1].
To show that V is a bounded linear map from L2 ([0, 1]) to L2 ([0, 1]), let K be the
function on [0, 1] × [0, 1] defined by
(
1 if x > y,
K ( x, y) =
0 if x ≤ y.

In other words, K is the characteris-


Vito Volterra (1860–1940) was a
tic function of the triangle below the
pioneer in developing functional
diagonal of the unit square. Clearly
2 analytic techniques to study integral
K ∈ L (λ × λ) and V = IK as defined
equations.
in 10.6. Thus V is a bounded linear map
2 2 1
from L ([0, 1]) to L ([0, 1]) and kV k ≤ √ (by 10.8).
2
Because V ∗ = IK∗ (by 10.9) and K ∗ is the characteristic function of the closed
triangle above the diagonal of the unit square, we see that
Z 1 Z 1 Z x
10.16 (V ∗ f )( x ) = f (y) dy = f (y) dy − f (y) dy
x 0 0

for f ∈ L2 ([0, 1]) and x ∈ [0, 1].


Now we can show that V ∗ is injective. To do this, suppose f ∈ L2 ([0, 1]) and

V f = 0. Differentiating both sides of 10.16 with respect to x and using the Lebesgue
Differentiation Theorem (4.19), we conclude that f = 0. Hence V ∗ is injective. Thus
the Volterra operator V has dense range (by 10.14).
Although range V is dense in L2 ([0, 1]), it does not equal L2 ([0, 1]) (because
every element of range V is a continuous function on [0, 1] that vanishes at 0). Thus
the Volterra operator V has dense but not closed range in L2 ([0, 1]).

Invertibility of Operators
Linear maps from a vector space to itself are so important that they get a special name
and special notation.

10.17 Definition operator; B(V )

• An operator is a linear map from a vector space to itself.


• If V is a normed vector space, then B(V ) denotes the normed vector space
of bounded operators on V. In other words, B(V ) = B(V, V ).
Section 10A Adjoints and Invertibility 287

10.18 Definition invertible; T −1

• An operator T on a vector space V is called invertible if T is a one-to-one


and surjective linear map of V onto V.
• Equivalently, an operator T : V → V is invertible if and only if there exists
an operator T −1 : V → V such that T −1 ◦ T = T ◦ T −1 = I.

The second bullet point above is equivalent to the first bullet point because if
a linear map T : V → V is one-to-one and surjective, then the inverse function
T −1 : V → V is automatically linear (as you should verify).
Also, if V is a Banach space and T is a bounded operator on V that is invertible,
then the inverse T −1 is automatically bounded, as follows from the Bounded Inverse
Theorem (6.83).
The next result shows that inverses and adjoints work well together. In the proof,
we use the common convention of writing composition of linear maps with the same
notation as multiplication. In other words, if S and T are linear maps such that S ◦ T
makes sense, then from now on

ST = S ◦ T.

10.19 inverse of the adjoint equals adjoint of the inverse

A bounded operator T on a Hilbert space is invertible if and only if T ∗ is invertible.


Furthermore, if T is invertible, then ( T ∗ )−1 = ( T −1 )∗.

Proof First suppose T is invertible. Taking the adjoint of all three sides of the
equation T −1 T = TT −1 = I, we get

T ∗ ( T −1 )∗ = ( T −1 )∗ T ∗ = I,

which implies that T ∗ is invertible and ( T ∗ )−1 = ( T −1 )∗.


Now suppose T ∗ is invertible. Then by the direction just proved, ( T ∗ )∗ is invert-
ible. Because ( T ∗ )∗ = T, this implies that T is invertible, completing the proof.

Norms work well with the composition of linear maps, as shown in the next result.

10.20 norm of a composition of linear maps

Suppose U, V, W are normed vector spaces, T ∈ B(U, V ), and S ∈ B(V, W ).


Then
kST k ≤ kSk k T k.

Proof If f ∈ U, then

k(ST )( f )k = kS( T f )k ≤ kSk k T f k ≤ kSk k T k k f k.

Thus kST k ≤ kSk k T k, as desired.


288 Chapter 10 Linear Maps on Hilbert Spaces

Unlike linear maps from one vector space to a different vector space, operators on
the same vector space can be composed with each other and raised to powers.

10.21 Definition Tk

Suppose T is an operator on a vector space V.


• For k ∈ Z+, the operator T k is defined by T k = |TT {z
· · · T}.
k times

• T 0 is defined to be the identity operator I : V → V.

You should verify that powers of an operator satisfy the usual arithmetic rules:
T j T k = T j+k and ( T j )k = T jk for j, k ∈ Z+. Also, if V is a normed vector space
and T ∈ B(V ), then
k T k k ≤ k T kk
for every k ∈ Z+, as follows from using induction on 10.20.
Recall that if z ∈ C with |z| < 1, then the formula for the sum of a geometric
series shows that

1
= ∑ zk .
1−z k =0
The next result shows that this formula carries over to operators on Banach spaces.

10.22 operators in the open unit ball centered at the identity are invertible

If T is a bounded operator on a Banach space and k T k < 1, then I − T is


invertible and

( I − T ) −1 = ∑ Tk .
k =0

Proof Suppose T is a bounded operator on a Banach space V and k T k < 1. Then


∞ ∞
1
∑ k T k k ≤ ∑ k T kk = 1 − k T k < ∞.
k =0 k =0

Hence 6.47 and 6.41 imply that the infinite sum ∑∞ k


k=0 T converges in B(V ). Now

∞ n
10.23 ( I − T) ∑ T k = lim ( I − T ) ∑ T k = nlim ( I − T n+1 ) = I,
n→∞ →∞
k =0 k =0

where the last equality holds because k T n+1 k ≤ k T kn+1 and k T k < 1. Similarly,
 ∞  n
10.24 ∑ Tk ( I − T ) = lim ∑ T k ( I − T ) = nlim ( I − T n+1 ) = I.
n→∞ →∞
k =0 k =0

Equations 10.23 and 10.24 imply that I − T is invertible and ( I − T )−1 = ∑∞ k


k =0 T .
Section 10A Adjoints and Invertibility 289

Now we use the previous result to show that the set of invertible operators on a
Banach space is open.

10.25 invertible operators form an open set

Suppose V is a Banach space. Then { T ∈ B(V ) : T is invertible} is an open


subset of B(V ).

Proof Suppose T ∈ B(V ) is invertible. Suppose S ∈ B(V ) and

1
k T − Sk < .
k T −1 k
Then
k I − T −1 Sk = k T −1 T − T −1 Sk ≤ k T −1 k k T − Sk < 1.
Hence 10.22 implies that I − ( I − T −1 S) is invertible; in other words, T −1 S is
invertible.
Now S = T ( T −1 S). Thus S is the product of two invertible operators, which
implies that S is invertible with S−1 = ( T −1 S)−1 T −1 .
We have shown that every element of the open ball of radius k T −1 k−1 centered at
T is invertible. Thus the set of invertible elements of B(V ) is open.

10.26 Definition left invertible; right invertible

Suppose T is a bounded operator on a Banach space V.


• T is called left invertible if there exists S ∈ B(V ) such that ST = I.
• T is called right invertible if there exists S ∈ B(V ) such that TS = I.

One of the wonderful theorems of linear algebra states that left invertibility and
right invertibility and invertibility are all equivalent to each other for operators on
a finite-dimensional vector space. The next example shows that this result fails on
infinite-dimensional Hilbert spaces.

10.27 Example left invertibility is not equivalent to right invertibility


Define the right shift T : `2 → `2 and the left shift S : `2 → `2 by

T ( a1 , a2 , a3 , . . .) = (0, a1 , a2 , a3 , . . .)

and
S ( a1 , a2 , a3 , . . . ) = ( a2 , a3 , a4 , . . . ).
Because ST = I, we see that T is left invertible and S is right invertible. However, T
is neither invertible nor right invertible because it is not surjective, and S is neither
invertible nor left invertible because it is not injective.
290 Chapter 10 Linear Maps on Hilbert Spaces

The result 10.29 below gives equivalent conditions for an operator on a Hilbert
space to be left invertible. On finite-dimensional vector spaces, left invertibility
is equivalent to injectivity. The example below shows that this fails on infinite-
dimensional Hilbert spaces. Thus we cannot eliminate the closed range requirement
in part (c) of 10.29.

10.28 Example injective but not left invertible


Define T : `2 → `2 by
 a a 
2 3
T ( a1 , a2 , a3 , . . . ) = a1 , , , . . . .
2 3
Then T is an injective bounded operator on `2 .
Suppose S is an operator on `2 such that ST = I. For n ∈ Z+, let en ∈ `2 be the
vector with 1 in the nth -slot and 0 elsewhere. Then

Sen = S(nTen ) = n(ST )(en ) = nen .

The equation above implies that S is unbounded. Thus T is not left invertible, even
though T is injective.

10.29 left invertibility

Suppose V is a Hilbert space and T ∈ B(V ). Then the following are equivalent:

(a) T is left invertible.


(b) there exists α ∈ (0, ∞) such that k f k ≤ αk T f k for all f ∈ V.
(c) T is injective and has closed range.
(d) T ∗ T is invertible.

Proof First suppose (a) holds. Thus there exists S ∈ B(V ) such that ST = I. If
f ∈ V, then
k f k = kS( T f )k ≤ kSk k T f k.
Thus (b) holds with α = kSk, proving that (a) implies (b).
Now suppose (b) holds. Thus there exists α ∈ (0, ∞) such that

10.30 k f k ≤ αk T f k for all f ∈ V.

The inequality above shows that if f ∈ V and T f = 0, then f = 0. Thus T is


injective. To show that T has closed range, suppose f 1 , f 2 , . . . is a sequence in V
such T f 1 , T f 2 , . . . converges in V to some g ∈ V. Thus the sequence T f 1 , T f 2 , . . . is
a Cauchy sequence in V. The inequality 10.30 then implies that f 1 , f 2 , . . . is a Cauchy
sequence in V. Thus f 1 , f 2 , . . . converges in V to some f ∈ V, which implies that
T f = g. Hence g ∈ range T, completing the proof that T has closed range, and
completing the proof that (b) implies (c).
Section 10A Adjoints and Invertibility 291

Now suppose (c) holds, so T is injective and has closed range. We want to prove
that (a) holds. Let R : range T → V be the inverse of the one-to-one linear function
f 7→ T f that maps V onto range T. Because range T is a closed subspace of V and
thus is a Banach space [by 6.16(b)], the Bounded Inverse Theorem (6.83) implies
that R is a bounded linear map. Let P denote the orthogonal projection of V onto the
closed subspace range T. Define S : V → V by
Sg = R( Pg).
Then for each g ∈ V, we have
kSgk = k R( Pg)k ≤ k Rk k Pgk ≤ k Rkk gk,
where the last inequality comes from 8.37(d). The inequality above implies that S is
a bounded operator on V. If f ∈ V, then

S( T f ) = R P( T f ) = R( T f ) = f .
Thus ST = I, which means that T is left invertible, completing the proof that (c)
implies (a).
At this stage of the proof we know that (a), (b), and (c) are equivalent. To prove
that one of these implies (d), suppose (b) holds. Squaring the inequality in (b), we
see that if f ∈ V, then

k f k2 ≤ α2 k T f k2 = α2 h T ∗ T f , f i ≤ α2 k T ∗ T f k k f k,
which implies that
k f k ≤ α2 k T ∗ T f k.
In other words, (b) holds with T replaced by T ∗ T (and α replaced by α2 ). By the
equivalence we already proved between (a) and (b), we conclude that T ∗ T is left
invertible. Thus there exists S ∈ B(V ) such that S( T ∗ T ) = I. Taking adjoints of
both sides of the last equation shows that ( T ∗ T )S∗ = I. Thus T ∗ T is also right
invertible, which implies that T ∗ T is invertible. Thus (b) implies (d).
Finally, suppose (d) holds, so T ∗ T is invertible. Hence there exists S ∈ B(V )
such that I = S( T ∗ T ) = (ST ∗ ) T. Thus T is left invertible, showing that (d) implies
(a), completing the proof that (a), (b), (c), and (d) are equivalent.

You may be familiar with the finite-dimensional result that right invertibility is
equivalent to surjectivity. The next result shows that this equivalency also holds on
infinite-dimensional Hilbert spaces.

10.31 right invertibility

Suppose V is a Hilbert space and T ∈ B(V ). Then the following are equivalent:

(a) T is right invertible.


(b) T is surjective.

(c) TT ∗ is invertible.
292 Chapter 10 Linear Maps on Hilbert Spaces

Proof Taking adjoints shows that an operator is right invertible if and only if its
adjoint is left invertible. Thus the equivalence of (a) and (c) in this result follows
immediately from the equivalence of (a) and (d) in 10.29 applied to T ∗ instead of T.
Suppose (a) holds, so T is right invertible. Hence there exists S ∈ B(V ) such
that TS = I. Thus T (S f ) = f for every f ∈ V, which implies that T is surjective,
completing the proof that (a) implies (b).
To prove that (b) implies (a), suppose T is surjective. Define R : (null T )⊥ → V
by R = T |(null T )⊥ . Clearly R is injective because

null R = (null T )⊥ ∩ (null T ) = {0}.

If f ∈ V, then f = g + h for some g ∈ null T and some h ∈ (null T )⊥ (by


8.43); thus T f = Th = Rh, which implies that range T = range R. Because T is
surjective, this implies that range R = V. In other words, R is a continuous injective
linear map of (null T )⊥ onto V. The Bounded Inverse Theorem (6.83) now implies
that R−1 : V → (null T )⊥ is a bounded linear map on V. We have TR−1 = I. Thus
T is right invertible, completing the proof that (b) implies (a).

EXERCISES 10A
1 Define T : `2 → `2 by T ( a1 , a2 , . . .) = (0, a1 , a2 , . . .). Find a formula for T ∗.
2 Suppose V is a Hilbert space, U is a closed subspace of V, and T : U → V is
defined by T f = f . Describe the linear operator T ∗ : V → U.
3 Suppose V and W are Hilbert spaces and g ∈ V, h ∈ W. Define T ∈ B(V, W )
by T f = h f , gih. Find a formula for T ∗.
4 Suppose V and W are Hilbert spaces and T ∈ B(V, W ) has finite-dimensional
range. Prove that T ∗ also has finite-dimensional range.
5 Prove or give a counterexample: If V is a Hilbert space and T : V → V is a
bounded linear map such that dim null T < ∞, then dim null T ∗ < ∞.
6 Suppose T is a bounded linear map from a Hilbert space V to a Hilbert space W.
Prove that k T ∗ T k = k T k2 .
[This formula for k T ∗ T k leads to the important subject of C ∗ -algebras.]
7 Suppose V is a Hilbert space and Inv(V ) is the set of invertible bounded oper-
ators on V. Think of Inv(V ) as a metric space with the metric it inherits as a
subset of B(V ). Show that T 7→ T −1 is a continuous function from Inv(V ) to
Inv(V ).
8 Suppose T is a bounded operator on a Hilbert space.
(a) Prove that T is left invertible if and only if T ∗ is right invertible.
(b) Prove that T is invertible if and only if T is both left and right invertible.
Section 10A Adjoints and Invertibility 293

9 Suppose b1 , b2 , . . . is a bounded sequence in F. Define a bounded linear map


T : `2 → `2 by
T ( a1 , a2 , . . .) = ( a1 b1 , a2 b2 , . . .).
(a) Find a formula for T ∗.
(b) Show that T is injective if and only if bk 6= 0 for every k ∈ Z+.
(c) Show that T has dense range if and only if bk 6= 0 for every k ∈ Z+.
(d) Show that T has closed range if and only if
inf{|bk | : k ∈ Z+ and bk 6= 0} > 0.
(e) Show that T is invertible if and only if
inf{|bk | : k ∈ Z+} > 0.
10 Suppose h ∈ L∞ (R) and Mh : L2 (R) → L2 (R) is the bounded operator
defined by Mh f = f h.
(a) Show that Mh is injective if and only if |{ x ∈ R : h( x ) = 0}| = 0.
(b) Find a necessary and sufficient condition (in terms of h) for Mh to have
dense range.
(c) Find a necessary and sufficient condition (in terms of h) for Mh to have
closed range.
(d) Find a necessary and sufficient condition (in terms of h) for Mh to be
invertible.

11 (a) Prove or give a counterexample: If T is a bounded operator on a Hilbert


space such that T and T ∗ are both injective, then T is invertible.
(b) Prove or give a counterexample: If T is a bounded operator on a Hilbert
space such that T and T ∗ are both surjective, then T is invertible.
12 Define T : `2 → `2 by T ( a1 , a2 , a3 , . . . ) = ( a2 , a3 , a4 , . . .). Suppose α ∈ F.
(a) Prove that T − αI is injective if and only if |α| ≥ 1.
(b) Prove that T − αI is invertible if and only if |α| > 1.
(c) Prove that T − αI is surjective if and only if |α| 6= 1.
(d) Prove that T − αI is left invertible if and only if |α| > 1.

13 Suppose V is a Hilbert space.


(a) Show that { T ∈ B(V ) : T is left invertible} is an open subset of B(V ).
(b) Show that { T ∈ B(V ) : T is right invertible} is an open subset of B(V ).

14 Suppose T is a bounded operator on a Hilbert space V.


(a) Prove that T is invertible if and only if T has a unique left inverse. In
other words, prove that T is invertible if and only if there exists a unique
S ∈ B(V ) such that ST = I.
(b) Prove that T is invertible if and only if T has a unique right inverse. In
other words, prove that T is invertible if and only if there exists a unique
S ∈ B(V ) such that TS = I.
294 Chapter 10 Linear Maps on Hilbert Spaces

10B Spectrum
Spectrum of an Operator
The following definitions play key roles in operator theory.

10.32 Definition spectrum; sp( T ); eigenvalue

Suppose T is a bounded operator on a Banach space V.


• A number α ∈ F is called an eigenvalue of T if T − αI is not injective.
• A nonzero vector f ∈ V is called an eigenvector of T corresponding to an
eigenvalue α ∈ F if
Tf = αf.

• The spectrum of T is denoted sp( T ) and is defined by

sp( T ) = {α ∈ F : T − αI is not invertible}.

If T − αI is not injective, then T − αI is not invertible. Thus the set of eigenvalues


of a bounded operator T is contained in the spectrum of T. If V is a finite-dimensional
Banach space and T ∈ B(V ), then T − αI is not injective if and only if T − αI is
not invertible. Thus if T is an operator on a finite-dimensional Banach space, then
the spectrum of T equals the set of eigenvalues of T.
However, on infinite-dimensional Banach spaces, the spectrum of an operator does
not necessarily equal the set of eigenvalues, as shown in the next example.

10.33 Example eigenvalues and spectrum


Verifying all the assertions in this example should help solidify your understanding
of the definition of the spectrum.
• Suppose b1 , b2 , . . . is a bounded sequence in F. Define a bounded linear map
T : `2 → `2 by
T ( a1 , a2 , . . .) = ( a1 b1 , a2 b2 , . . .).
Then the set of eigenvalues of T equals {bk : k ∈ Z+} and the spectrum of T
equals the closure of {bk : k ∈ Z+}.
• Suppose h ∈ L∞ (R). Define a bounded linear map Mh : L2 (R) → L2 (R) by
Mh f = f h.
Then α ∈ F is an eigenvalue of Mh if and only if |{t ∈ R : h(t) = α}| > 0.
Also, α ∈ sp( Mh ) if and only if |{t ∈ R : |h(t) − α| < ε}| > 0 for all ε > 0.
• Define the right shift T : `2 → `2 and the left shift S : `2 → `2 by
T ( a1 , a2 , a3 , . . .) = (0, a1 , a2 , a3 , . . .) and S( a1 , a2 , a3 , . . .) = ( a2 , a3 , a4 , . . .).
Then T has no eigenvalues, and sp( T ) = {α ∈ F : |α| ≤ 1}. Also, the set of
eigenvalues of S is the open set {α ∈ F : |α| < 1}, and the spectrum of S is the
closed set {α ∈ F : |α| ≤ 1}.
Section 10B Spectrum 295

If α is an eigenvalue of an operator T ∈ B(V ) and f is an eigenvector of T


corresponding to α, then

k T f k = k α f k = | α | k f k,
which implies that |α| ≤ k T k. The next result states that the same inequality holds
for elements of sp( T ).

10.34 T − αI is invertible for |α| large

Suppose T is a bounded operator on a Banach space. Then

(a) sp( T ) ⊂ {α ∈ F : |α| ≤ k T k};


(b) T − αI is invertible for all α ∈ F with |α| > k T k;
(c) lim k( T − αI )−1 k = 0.
|α|→∞

Proof We begin by proving (b). Suppose α ∈ F and |α| > k T k. Then


 T
10.35 T − αI = −α I − .
α
Because k T/αk < 1, the equation above and 10.22 imply that T − αI is invertible,
completing the proof of (b).
Using the definition of spectrum, (a) now follows immediately from (b).
To prove (c), again suppose α ∈ F and |α| > k T k. Then 10.35 and 10.22 imply

1 Tk
( T − αI )−1 = − ∑ .
α k =0 αk

Thus

1 k T kk
k( T − αI )−1 k ≤ ∑ k
|α| k =0 | α |

1 1
=
|α| 1 − kT k
|α|

1
= .
|α| − k T k
The inequality above implies (c), completing the proof.

The set of eigenvalues of a bounded operator on a Hilbert space can be any


bounded subset of F, even a nonmeasurable set (see Exercise 3). In contrast, the next
result shows that the spectrum of a bounded operator is a closed subset of F. This
result provides one indication that the spectrum of an operator may be a more useful
set than the set of eigenvalues.
296 Chapter 10 Linear Maps on Hilbert Spaces

10.36 spectrum is closed

The spectrum of a bounded operator on a Banach space is a closed subset of F.

Proof Suppose T is a bounded operator on a Banach space V. Suppose α1 , α2 , . . .


is a sequence in sp( T ) that converges to some α ∈ F. Thus each T − αn I is not
invertible and
lim ( T − αn I ) = T − αI.
n→∞
The set of noninvertible elements of B(V ) is a closed subset of B(V ) (by 10.25).
Hence the equation above implies that T − αI is not invertible. In other words,
α ∈ sp( T ), which implies that sp( T ) is closed.

Our next result provides the key tool used in proving that the spectrum of a
bounded operator on a nonzero complex Hilbert space is nonempty (see 10.38). The
statement of the next result and the proofs of the next two results use a bit of basic
complex analysis. Because sp( T ) is a closed subset of C (by 10.36), C \ sp( T ) is
an open subset of C and thus it makes sense to ask whether the function in the result
below is analytic.
To keep things simple, the next two results are stated for complex Hilbert spaces.
See Exercise 6 for the analogous results for complex Banach spaces.

10.37 analyticity of ( T − αI )−1

Suppose T is a bounded operator on a complex Hilbert space V. Then the function




α 7→ ( T − αI )−1 f , g

is analytic on C \ sp( T ) for every f , g ∈ V.

Proof Suppose β ∈ C \ sp( T ). Then for α ∈ C with |α − β| < 1/k( T − βI )−1 k,


we see from 10.22 that I − (α − β)( T − βI )−1 is invertible and
 −1 ∞ k
I − (α − β)( T − βI )−1 = ∑ (α − β)k ( T − βI )−1 .
k =0

Multiplying both sides of the equation above by ( T − βI )−1 and using the equation
A−1 B−1 = ( BA)−1 for invertible operators A and B, we get
∞  k +1
( T − αI )−1 = ∑ (α − β)k ( T − βI )−1 .
k =0
Thus for f , g ∈ V, we have

∞ D  k +1 E
( T − αI )−1 f , g = ∑ ( T − βI )−1 f , g (α − β)k .
k =0


The equation above shows that the function α 7→ ( T − αI )−1 f , g has a power se-
ries expansion as powers of α − β for α near β. Thus this function is analytic near β.
Section 10B Spectrum 297

A major result in finite-dimensional


The spectrum of a bounded operator
linear algebra states that every operator
on a nonzero real Hilbert space can
on a nonzero finite-dimensional complex
be the empty set. This can happen
vector space has an eigenvalue. We have
even in finite dimensions, where an
seen examples showing that this result
operator on R2 might have no
does not extend to bounded operators on
eigenvalues. Thus the restriction in
complex Hilbert spaces. However, the
the next result to the complex case
next result is an excellent substitute. Al-
cannot be removed.
though a bounded operator on a nonzero
complex Hilbert space need not have an eigenvalue, the next result shows that for
each such operator T, there exists α ∈ C such that T − αI is not invertible.

10.38 spectrum is nonempty

The spectrum of a bounded operator on a complex nonzero Hilbert space is a


nonempty subset of C.

Proof Suppose T ∈ B(V ), where V is a complex Hilbert space with V 6= {0}, and
sp( T ) = ∅. Let f ∈ V with f 6= 0. Take g = T −1 f in 10.37. Because sp( T ) = ∅,
10.37 implies that the function


α 7→ ( T − αI )−1 f , T −1 f

is analytic on all of C. The value of the function above at α = 0 equals the average
value of the function on each circle in C centered at 0 (because analytic functions
satisfy the mean value property). But 10.34(c) implies that this function has limit 0
as |α| → ∞. Thus taking the average over large circles, we see that the value of the
function above at α = 0 is 0. In other words,

−1
T f , T −1 f = 0.

Hence T −1 f = 0. Applying T to both sides of the equation T −1 f = 0 shows that


f = 0, which contradicts our assumption that f 6= 0. This contradiction means that
our assumption that sp( T ) = ∅ was false, completing the proof.

10.39 Definition p( T )

Suppose T is an operator on a vector space V and p is a polynomial with coeffi-


cients in F:
p(z) = b0 + b1 z + · · · + bn zn .
Then p( T ) is the operator on V defined by

p( T ) = b0 I + b1 T + · · · + bn T n .

You should verify that if p and q are polynomials with coefficients in F and T is
an operator, then
( pq)( T ) = p( T ) q( T ).
298 Chapter 10 Linear Maps on Hilbert Spaces

The next result provides a nice way to compute the spectrum of a polynomial
applied to an operator. For example, this result implies that if T is a bounded operator
on a complex Banach space, then the spectrum of T 2 consists of the squares of all
numbers in the spectrum of T.
As with the previous result, the next result fails on real Banach spaces. As you
can see, the proof below uses factorization of a polynomial with complex coefficients
as the product of polynomials with degree 1, which is not necessarily possible when
restricting to the field of real numbers.

10.40 Spectral Mapping Theorem

Suppose T is a bounded operator on a complex Banach space and p is a


polynomial with complex coefficients. Then
 
sp p( T ) = p sp( T ) .

Proof If p is a constant polynomial, then both sides of the equation above consist
of the set containing just that constant. Thus we can assume that p is a nonconstant
polynomial. 
First suppose α ∈ sp p( T ) . Thus p( T ) − αI is not invertible. By the Funda-
mental Theorem of Algebra, there exist c, β 1 , . . . β n ∈ C with c 6= 0 such that

10.41 p(z) − α = c(z − β 1 ) · · · (z − β n )

for all z ∈ C. Thus

p( T ) − αI = c( T − β 1 I ) · · · ( T − β n I ).

The left side of the equation above is not invertible. Hence T − β k I is not invertible
for some k ∈ {1, . . . , n}. Thus β k ∈ sp( T ). Now 10.41
 implies p( β k ) = α. Hence
α ∈ p sp( T ) , completing the proof that sp p( T ) ⊂ p sp( T ) .
To prove the inclusion in the other direction, now suppose β ∈ sp( T ). The
polynomial z 7→ p(z) − p( β) has a zero at β. Hence there exists a polynomial q
with degree 1 less than the degree of p such that

p(z) − p( β) = (z − β)q(z)

for all z ∈ C. Thus

10.42 p( T ) − p( β) I = ( T − βI )q( T )

and

10.43 p( T ) − p( β) I = q( T )( T − βI ).

Because T − βI is not invertible, T − βI is not surjective or T − βI is not injective.


If T − βI is not surjective, then 10.42 shows that p( T ) − p( β) I is not surjective. If
T − βI is not injective, then 10.43 shows that p( T ) − p( β) I is not injective.
 Either
way, we see that p( T ) − p( β) I is not invertible. Thus p( β) ∈ sp p( T ) , completing
 
the proof that sp p( T ) ⊃ p sp( T ) .
Section 10B Spectrum 299

Self-adjoint Operators
In this subsection, we look at a nice special class of bounded operators.

10.44 Definition self-adjoint

A bounded operator T on a Hilbert space is called self-adjoint if T ∗ = T.

The definition of the adjoint implies that a bounded operator T on a Hilbert space
V is self-adjoint if and only if h T f , gi = h f , Tgi for all f , g ∈ V. See Exercise 7 for
an interesting result regarding this last condition.

10.45 Example self-adjoint operators

• Suppose b1 , b2 , . . . is a bounded sequence in F. Define a bounded operator


T : `2 → `2 by
T ( a1 , a2 , . . .) = ( a1 b1 , a2 b2 , . . .).
Then T ∗ : `2 → `2 is the operator defined by
T ∗ ( a1 , a2 , . . .) = ( a1 b1 , a2 b2 , . . .).
Hence T is self-adjoint if and only if bk ∈ R for all k ∈ Z+.
• More generally, suppose ( X, S , µ) is a σ-finite ∞
2
 measure space and h ∈∗L (µ).
Define a bounded operator Mh ∈ B L (µ) by Mh f = f h. Then Mh = Mh .
Thus Mh is self-adjoint if and only if µ({ x ∈ X : h( x ) ∈
/ R}) = 0.
• Suppose n ∈ Z+, K is an n-by-n matrix, and IK : Fn → Fn is the operator of
matrix multiplication by K (thinking of elements of Fn as column vectors). Then
(IK )∗ is the operator of multiplication by the conjugate transpose of K, as shown
in Example 10.10. Thus IK is a self-adjoint operator if and only if the matrix K
equals its conjugate transpose.
• More generally, suppose ( X, S , µ) is a σ-finite measure space, K ∈ L2 (µ × µ),
and IK is the integral operator on L2 (µ) defined in Example 10.5. Define
K ∗ : X × X → F by K ∗ (y, x ) = K ( x, y). Then (IK )∗ is the integral operator
induced by K ∗, as shown in Example 10.5. Thus if K ∗ = K, or in other words if
K ( x, y) = K (y, x ) for all ( x, y) ∈ X × X, then IK is self-adjoint.
• Suppose U is a closed subspace of a Hilbert space V. Recall that PU denotes the
orthogonal projection of V onto U (see Section 8B). We have
h PU f , gi = h PU f , PU g + ( I − PU ) gi
= h PU f , PU gi
= h f − ( I − PU ) f , PU gi
= h f , PU gi,
where the second and fourth equalities above hold because of 8.37(a). The
equation above shows that PU is a self-adjoint operator.
300 Chapter 10 Linear Maps on Hilbert Spaces

For real Hilbert spaces, the next result requires the additional hypothesis that T
is self-adjoint. To see that this extra hypothesis cannot be eliminated, consider the
operator T : R2 → R2 defined by T ( x, y) = (−y, x ). Then, T 6= 0, but with the
standard inner product on R2 , we have h T f , f i = 0 for all f ∈ R2 (which you can
verify either algebraically or by thinking of T as counterclockwise rotation by a right
angle).

10.46 h T f , f i = 0 for all f implies T = 0

Suppose V is a Hilbert space, T ∈ B(V ), and h T f , f i = 0 for all f ∈ V.

(a) If F = C, then T = 0.
(b) If F = R and T is self-adjoint, then T = 0.

Proof First suppose F = C. If g, h ∈ V, then

h T ( g + h ), g + h i − h T ( g − h ), g − h i
h Tg, hi =
4
h T ( g + ih), g + ihi − h T ( g − ih), g − ihi
+ i,
4
as can be verified by computing the right side. Our hypothesis that h T f , f i = 0
for all f ∈ V implies that the right side above equals 0. Thus h Tg, hi = 0 for all
g, h ∈ V. Taking h = Tg, we can conclude that T = 0, which completes the proof
of (a).
Now suppose F = R and T is self-adjoint. Then

h T ( g + h ), g + h i − h T ( g − h ), g − h i
10.47 h Tg, hi = ;
4
this is proved by computing the right side using the equation

h Th, gi = hh, Tgi = h Tg, hi,

where the first equality holds because T is self-adjoint and the second equality holds
because we are working in a real Hilbert space. Each term on the right side of 10.47
is of the form h T f , f i for appropriate f . Thus h Tg, hi = 0 for all g, h ∈ V. This
implies that T = 0 (take h = Tg), completing the proof of (b).

Some insight into the adjoint can be obtained by thinking of the operation T 7→ T ∗
on B(V ) as analogous to the operation z 7→ z on C. Under this analogy, the
self-adjoint operators (characterized by T ∗ = T) correspond to the real numbers
(characterized by z = z). The first two bullet points in Example 10.45 illustrate this
analogy, as we saw that a multiplication operator on L2 (µ) is self-adjoint if and only
if the multiplier is real-valued almost everywhere.
The next two results deepen the analogy between the self-adjoint operators and
the real numbers. First we see this analogy reflected in the behavior of h T f , f i, and
then we see this analogy reflected in the spectrum of T.
Section 10B Spectrum 301

10.48 self-adjoint characterized by h T f , f i

Suppose T is a bounded operator on a complex Hilbert space V. Then T is


self-adjoint if and only if
hT f , f i ∈ R
for all f ∈ V.

Proof Let f ∈ V. Then

h T f , f i − h T f , f i = h T f , f i − h f , T f i = h T f , f i − h T ∗ f , f i = h( T − T ∗ ) f , f i.
If h T f , f i ∈ R for every f ∈ V, then the left side of the equation above equals 0, so
h( T − T ∗ ) f , f i = 0 for every f ∈ V. This implies that T − T ∗ = 0 [by 10.46(a)].
Hence T is self-adjoint.
Conversely, if T is self-adjoint, then the right side of the equation above equals
0, so h T f , f i = h T f , f i for every f ∈ V. This implies that h T f , f i ∈ R for every
f ∈ V, as desired.

10.49 self-adjoint operators have real spectrum

Suppose T is a bounded self-adjoint operator on a Hilbert space. Then


sp( T ) ⊂ R.

Proof The desired result holds if F = R because the spectrum of every operator on
a real Hilbert space is, by definition, contained in R.
Thus we assume that T is a bounded operator on a complex Hilbert space V.
Suppose α, β ∈ R, with β 6= 0. If f ∈ V, then


k T − (α + βi ) I f k k f k ≥ T − (α + βi ) I f , f

= h T f , f i − α k f k2 − β k f k2 i

≥ | β | k f k2 ,
where the first inequality comes from the Cauchy–Schwarz inequality (8.11) and the
last inequality holds because h T f , f i − αk f k2 ∈ R (by 10.48).
The inequality above implies that
1 
kfk ≤ k T − (α + βi ) I f k
| β|
for all f ∈ V. Now the equivalence of (a) and (b) in 10.29 shows that T − (α + βi ) I
is left invertible.
Because T is self-adjoint, the adjoint of T − (α + βi ) I is T − (α − βi ) I, which
is left invertible by the same argument as above (just replace β by − β). Hence
T − (α + βi ) I is right invertible (because its adjoint is left invertible). Because the
operator T − (α + βi ) I is both left and right invertible, it is invertible. In other words,
α + βi ∈ / sp( T ). Thus sp( T ) ⊂ R, as desired.
302 Chapter 10 Linear Maps on Hilbert Spaces

We showed that a bounded operator on a complex nonzero Hilbert space has a


nonempty spectrum. That result can fail on real Hilbert spaces (where by definition
the spectrum is contained in R). For example, the operator T on R2 defined by
T ( x, y) = (−y, x ) has empty spectrum. However, the previous result and 10.38 can
be used to show that every self-adjoint operator on a nonzero real Hilbert space has
nonempty spectrum (see Exercise 9 for the details).
Although the spectrum of every self-adjoint operator is nonempty, it is not true that
every self-adjoint operator
 has an eigenvalue. For example, the self-adjoint operator
Mx ∈ B L2 ([0, 1]) defined by ( Mx f )( x ) = x f ( x ) has no eigenvalues.

Normal Operators
Now we consider another nice special class of operators.

10.50 Definition normal operator

A bounded operator T on a Hilbert space is called normal if it commutes with its


adjoint. In other words, T is normal if

T ∗ T = TT ∗ .

Clearly every self-adjoint operator is normal, but there exist normal operators that
are not self-adjoint, as shown in the next example.

10.51 Example normal operators



• Suppose µ is a positive measure, h ∈ L∞ (µ), and Mh ∈ B L2 (µ) is the
multiplication operator defined by Mh f = f h. Then Mh ∗ = Mh , which means
that Mh is self-adjoint if h is real valued. If F = C, then h can be complex
valued and Mh is not necessarily self-adjoint. However,
M h ∗ M h = M| h | 2 = M h M h ∗
and thus Mh is a normal operator even when h is complex valued.
• Suppose T is the operator on F2 whose matrix with respect to the standard basis
is  
2 −3
.
3 2
Then T is not self-adjoint because the matrix above is not equal to its conjugate
transpose. However, T ∗ T = 13I and TT ∗ = 13I, as you should verify. Because
T ∗ T = TT ∗, we conclude that T is a normal operator.

10.52 Example an operator that is not normal


Suppose T is the right shift on `2 ; thus T ( a1 , a2 , . . .) = (0, a1 , a2 , . . .). Then T ∗
is the left shift: T ∗ ( a1 , a2 , . . .) = ( a2 , a3 , . . .). Hence T ∗ T is the identity operator
on `2 and TT ∗ is the operator ( a1 , a2 , a3 , . . .) 7→ (0, a2 , a3 , . . .). Thus T ∗ T 6= TT ∗,
which means that T is not a normal operator.
Section 10B Spectrum 303

10.53 normal in terms of norms

Suppose T is a bounded operator on a Hilbert space V. Then T is normal if and


only if
kT f k = kT∗ f k
for all f ∈ V.

Proof If f ∈ V, then
k T f k2 − k T ∗ f k2 = h T f , T f i − h T ∗ f , T ∗ f i = h( T ∗ T − TT ∗ ) f , f i.
If T is normal, then the right side of the equation above equals 0, which implies that
the left side also equals 0 and hence k T f k = k T ∗ f k.
Conversely, suppose k T f k = k T ∗ f k for all f ∈ V. Then the left side of the
equation above equals 0, which implies that the right side also equals 0 for all f ∈ V.
Because T ∗ T − TT ∗ is self-adjoint, 10.46 now implies that T ∗ T − TT ∗ = 0. Thus
T is normal, completing the proof.

Each complex number can be written in the form a + bi, where a and b are real
numbers. Part (a) of the next result gives the analogous result for bounded operators
on a complex Hilbert space, with self-adjoint operators playing the role of real
numbers. We could call the operators A and B in part (a) the real and imaginary parts
of the operator T. Part (b) below shows that normality depends upon whether these
real and imaginary parts commute.

10.54 operator is normal if and only if its real and imaginary parts commute

Suppose T is a bounded operator on a complex Hilbert space V.

(a) There exist unique self-adjoint operators A, B on V such that T = A + iB.


(b) T is normal if and only if AB = BA, where A, B are as in part (a).

Proof Suppose T = A + iB, where A and B are self-adjoint. Then T ∗ = A − iB.


Adding these equations for T and T ∗ and then dividing by 2 produces a formula for
A; subtracting the equation for T ∗ from the equation for T and then dividing by 2i
produces a formula for B. Specifically, we have
T + T∗ T − T∗
A= and B = ,
2 2i
which proves the uniqueness part of (a). The existence part of (a) is proved by
defining A and B by the equations above and noting that A and B as defined above
are self-adjoint and T = A + iB.
To prove (b), verify that if A and B are defined as in the equations above, then
T ∗ T − TT ∗
AB − BA = .
2i
Thus AB = BA if and only if T is normal.
304 Chapter 10 Linear Maps on Hilbert Spaces

An operator on a finite-dimensional vector space is left invertible if and only if


it is right invertible. We have seen that this result fails for bounded operators on
infinite-dimensional Hilbert spaces. However, the next result shows that we recover
this equivalency for normal operators.

10.55 invertibility for normal operators

Suppose V is a Hilbert space and T ∈ B(V ) is normal. Then the following are
equivalent:

(a) T is invertible.
(b) T is left invertible.
(c) T is right invertible.
(d) T is surjective.
(e) T is injective and has closed range.
(f) T ∗ T is invertible.
(g) TT ∗ is invertible.

Proof Because T is normal, (f) and (g) are clearly equivalent. From 10.29, we know
that (f), (b), and (e) are equivalent to each other. From 10.31, we know that (g),
(c), and (d) are equivalent to each other. Thus (b), (c), (d), (e), (f), and (g) are all
equivalent to each other.
Clearly (a) implies (b).
Suppose (b) holds. We already know that (b) and (c) are equivalent; thus T is left
invertible and T is right invertible. Hence T is invertible, proving that (b) implies (a)
and completing the proof that (a) through (g) are all equivalent to each other.

The next result shows that a normal operator and its adjoint have the same eigen-
vectors, with eigenvalues that are complex conjugates of each other. This result can
fail for operators that are not normal. For example, 0 is an eigenvalue of the left shift
on `2 but its adjoint the right shift has no eigenvectors and no eigenvalues.

10.56 T normal and T f = α f implies T ∗ f = α f

Suppose T is a normal operator on a Hilbert space V, α ∈ F, and f ∈ V. Then α


is an eigenvalue of T with eigenvector f if and only if α is an eigenvalue of T ∗
with eigenvector f .

Proof Because ( T − αI )∗ = T ∗ − αI and T is normal, T − αI commutes with its


adjoint. Thus T − αI is normal. Hence 10.53 implies that

k( T − αI ) f k = k( T ∗ − αI ) f k.

Thus ( T − αI ) f = 0 if and only if ( T ∗ − αI ) f = 0, as desired.


Section 10B Spectrum 305

Because every self-adjoint operator is normal, the following result also holds for
self-adjoint operators.

10.57 orthogonal eigenvectors for normal operators

Eigenvectors of a normal operator corresponding to distinct eigenvalues are


orthogonal.

Proof Suppose α and β are distinct eigenvalues of a normal operator T, with


corresponding eigenvectors f and g. Then 10.56 implies that T ∗ f = α f . Thus
( β − α)h g, f i = h βg, f i − h g, α f i = h Tg, f i − h g, T ∗ f i = 0.
Because α 6= β, the equation above implies that h g, f i = 0, as desired.

Isometries and Unitary Operators

10.58 Definition isometry; unitary operator

Suppose T is a bounded operator on a Hilbert space V.


• T is called an isometry if k T f k = k f k for every f ∈ V.
• T is called unitary if T ∗ T = TT ∗ = I.

10.59 Example isometries and unitary operators

• Suppose T ∈ B(`2 ) is the right shift defined by


T ( a1 , a2 , a3 , . . .) = (0, a1 , a2 , a3 , . . .).
Then T is an isometry but is not a unitary operator because TT ∗ 6= I (as is clear
without even computing T ∗ because T is not surjective).

• Suppose T ∈ B `2 (Z) is the right shift defined by
( T f )(n) = f (n − 1)
for f : Z → F with ∑∞ 2
k=−∞ | f ( k )| < ∞. Then T is an isometry and is unitary.

• Suppose b1 , b2 , . . . is a bounded sequence in F. Define T ∈ B(`2 ) by


T ( a1 , a2 , . . .) = ( a1 b1 , a2 b2 , . . .).
Then T is an isometry if and only if T is unitary if and only if |bk | = 1 for all
k ∈ Z+.
• More generally, suppose ( X, S , µ) is a σ-finite measure space and h ∈ L∞ (µ).
Define Mh ∈ B L2 (µ) by Mh f = f h. Then T is an isometry if and only if T
is unitary if and only if µ({ x ∈ X : |h( x )| 6= 1}) = 0.
306 Chapter 10 Linear Maps on Hilbert Spaces

By definition, isometries preserve norms. The equivalence of (a) and (b) in the
following result shows that isometries also preserve inner products.

10.60 isometries preserve inner products

Suppose T is a bounded operator on a Hilbert space V. Then the following are


equivalent:

(a) T is an isometry.
(b) h T f , Tgi = h f , gi for all f , g ∈ V.
(c) T ∗ T = I.
(d) { Tek }k∈Γ is an orthonormal family for every orthonormal family {ek }k∈Γ
in V.
(e) { Tek }k∈Γ is an orthonormal family for some orthonormal basis {ek }k∈Γ
of V.

Proof If f ∈ V, then
k T f k2 − k f k2 = h T f , T f i − h f , f i = h( T ∗ T − I ) f , f i.
Thus k T f k = k f k for all f ∈ V if and only if the right side of the equation above
is 0 for all f ∈ V. Because T ∗ T − I is self-adjoint, this happens if and only if
T ∗ T − I = 0 (by 10.46). Thus (a) is equivalent to (c).
If T ∗ T = I, then h T f , Tgi = h T ∗ T f , gi = h f , gi for all f , g ∈ V. Thus (c)
implies (b).
Taking g = f in (b), we see that (b) implies (a). Hence we now know that (a), (b),
and (c) are equivalent to each other.
To prove that (b) implies (d), suppose (b) holds. If {ek }k∈Γ is an orthonormal
family in V, then h Te j , Tek i = he j , ek i for all j, k ∈ Γ, and thus { Tek }k∈Γ is an
orthonormal family in V. Hence (b) implies (d).
Because V has an orthonormal basis (see 8.67 or 8.75), (d) implies (e).
Finally, suppose (e) holds. Thus { Tek }k∈Γ is an orthonormal family for some
orthonormal basis {ek }k∈Γ of V. Suppose f ∈ V. Then by 8.63(a) we have
f = ∑ h f , e j ie j ,
j∈Γ

which implies that


T∗ T f = ∑ h f , e j iT ∗ Te j .
j∈Γ
Thus if k ∈ Γ, then
h T ∗ T f , ek i = ∑ h f , e j ihT ∗ Te j , ek i = ∑ h f , e j ihTe j , Tek i = h f , ek i,
j∈Γ j∈Γ

where the last equality holds because h Te j , Tek i equals 1 if j = k and equals 0
otherwise. Because the equality above holds for every ek in the orthonormal basis
{ek }k∈Γ , we conclude that T ∗ T f = f . Thus (e) implies (c), completing the proof.
Section 10B Spectrum 307

The equivalence between (a) and (c) in the previous result shows that every unitary
operator is an isometry.
Next we have a result giving conditions that are equivalent to being a unitary
operator. Notice that parts (d) and (e) of the previous result refer to orthonormal
families, but parts (f) and (g) of the following result refer to orthonormal bases.

10.61 unitary operators and their adjoints are isometries

Suppose T is a bounded operator on a Hilbert space V. Then the following are


equivalent:

(a) T is unitary.
(b) T is a surjective isometry.
(c) T and T ∗ are both isometries.
(d) T ∗ is unitary.
(e) T is invertible and T −1 = T ∗ .
(f) { Tek }k∈Γ is an orthonormal basis of V for every orthonormal basis {ek }k∈Γ
of V.

(g) { Tek }k∈Γ is an orthonormal basis of V for some orthonormal basis {ek }k∈Γ
of V.

Proof The equivalence of (a), (d), and (e) follows easily from the definition of
unitary.
The equivalence of (a) and (c) follows from the equivalence in 10.60 of (a) and (c).
To prove that (a) implies (b), suppose (a) holds, so T is unitary. As we have
already noted, this implies that T is an isometry. Also, the equation TT ∗ = I implies
that T is surjective. Thus (b) holds, proving that (a) implies (b).
Now suppose (b) holds, so T is a surjective isometry. Because T is surjective and
injective, T is invertible. The equation T ∗ T = I [which follows from the equivalence
in 10.60 of (a) and (c)] now implies that T −1 = T ∗. Thus (b) implies (e). Hence at
this stage of the proof, we know that (a), (b), (c), (d), and (e) are all equivalent to
each other.
To prove that (b) implies (f), suppose (b) holds, so T is a surjective isometry.
Suppose {ek }k∈Γ is an orthonormal basis of V. The equivalence in 10.60 of (a) and (d)
implies that { Tek }k∈Γ is an orthonormal family. Because {ek }k∈Γ is an orthonormal
basis of V and T is surjective, the closure of the span of { Tek }k∈Γ equals V. Thus
{ Tek }k∈Γ is an orthonormal basis of V, which proves that (b) implies (f).
Obviously (f) implies (g).
Now suppose (g) holds. The equivalence in 10.60 of (a) and (e) implies that T
is an isometry, which implies that the range of T is closed. Because { Tek }k∈Γ is an
orthonormal basis of V, the closure of the range of T equals V. Thus T is a surjective
isometry, proving that (g) implies (b) and completing the proof that (a) through (g)
are all equivalent to each other.
308 Chapter 10 Linear Maps on Hilbert Spaces

The equations T ∗ T = TT ∗ = I are analogous to the equation |z|2 = 1 for z ∈ C.


We now extend this analogy to the behavior of the spectrum of a unitary operator.

10.62 spectrum of a unitary operator

Suppose T is a unitary operator on a Hilbert space. Then

sp( T ) ⊂ {α ∈ F : |α| = 1}.

Proof Suppose α ∈ F with |α| 6= 1. Then


( T − αI )∗ ( T − αI ) = ( T ∗ − αI )( T − αI )

= (1 + |α|2 ) I − (αT ∗ + αT )
 αT ∗ + αT 
10.63 = (1 + | α |2 ) I − .
1 + | α |2
Looking at the last term in parentheses above, we have
αT ∗ + αT 2| α |

10.64 ≤ < 1,
1 + | α |2 1 + | α |2
where the last inequality holds because |α| 6= 1. Now 10.64, 10.63, and 10.22 imply
that ( T − αI )∗ ( T − αI ) is invertible. Thus T − αI is left invertible. Because T − αI
is normal, this implies that T − αI is invertible (see 10.55). Hence α ∈ / sp( T ). Thus
sp( T ) ⊂ {α ∈ F : |α| = 1}, as desired.

As a special case of the next result, we can conclude (without doing any calcula-
tions!) that the spectrum of the right shift on `2 is {α ∈ F : |α| ≤ 1}.

10.65 spectrum of an isometry

Suppose T is an isometry on a Hilbert space and T is not unitary. Then

sp( T ) = {α ∈ F : |α| ≤ 1}.

Proof Because T is an isometry but is not unitary, we know that T is not surjective
[by the equivalence of (a) and (b) in 10.61]. In particular, T is not invertible. Thus
T ∗ is not invertible.
Suppose α ∈ F with |α| < 1. Because T ∗ T = I, we have
T ∗ ( T − αI ) = I − αT ∗ .
The right side of the equation above is invertible (by 10.22). If T − αI were invertible,
then the equation above would imply T ∗ = ( I − αT ∗ )( T − αI )−1 , which would
make T ∗ invertible as the product of invertible operators. However, the paragraph
above shows T ∗ is not invertible. Thus T − αI is not invertible. Hence α ∈ sp( T ).
Thus {α ∈ F : |α| < 1} ⊂ sp( T ). Because sp( T ) is closed (see 10.36), this
implies {α ∈ F : |α| ≤ 1} ⊂ sp( T ). The inclusion in the other direction follows
from 10.34(a). Thus sp( T ) = {α ∈ F : |α| ≤ 1}.
Section 10B Spectrum 309

EXERCISES 10B
1 Verify all the assertions in Example 10.33.
2 Suppose T is a bounded operator on a Hilbert space V.
(a) Prove that sp(S−1 TS) = sp( T ) for all bounded invertible operators S on V.
(b) Prove that sp( T ∗ ) = {α : α ∈ sp( T )}.
1
(c) Prove that if T is invertible, then sp( T −1 ) = α : α ∈ sp( T ) .

3 Suppose E is a bounded subset of F. Show that there exists a Hilbert space V


and T ∈ B(V ) such that the set of eigenvalues of T equals E.
4 Suppose E is a nonempty closed bounded subset of F. Show that there exists
T ∈ B(`2 ) such that sp( T ) = E.
5 Give an example of a bounded operator T on a normed vector space such that
for every α ∈ F, the operator T − αI is not invertible.
6 Suppose T is a bounded operator on a complex nonzero Banach space V.
(a) Prove that the function

α 7→ ϕ ( T − αI )−1 f

is analytic on C \ sp( T ) for every f ∈ V and every ϕ ∈ V 0.


(b) Prove that sp( T ) 6= ∅.

7 Prove that if T is an operator on a Hilbert space V such that h T f , gi = h f , Tgi


for all f , g ∈ V, then T is a bounded operator.
8 Suppose P is a bounded operator on a Hilbert space V such that P2 = P. Prove
that P is self-adjoint if and only if there exists a closed subspace U of V such
that P = PU .
9 Suppose V is a real Hilbert space and T ∈ B(V ). The complexification of T is
the function TC : VC → VC defined by

TC ( f + ig) = T f + iTg

for f , g ∈ V (see Exercise 4 in Section 8B for the definition of VC ).


(a) Show that TC is a bounded operator on the complex Hilbert space VC and
k TC k = k T k.
(b) Show that TC is invertible if and only if T is invertible.
(c) Show that ( TC )∗ = ( T ∗ )C .
(d) Show that T is self-adjoint if and only if TC is self-adjoint.
(e) Use the previous parts of this exercise and 10.49 and 10.38 to show that if
T is self-adjoint and V 6= {0}, then sp( T ) 6= ∅.
310 Chapter 10 Linear Maps on Hilbert Spaces

10 Suppose T is a bounded operator on a Hilbert space V such that h T f , f i ≥ 0


for all f ∈ V. Prove that sp( T ) ⊂ [0, ∞).
11 Suppose P is a bounded operator on a Hilbert space V such that P2 = P. Prove
that P is self-adjoint if and only if P is normal.
12 Prove that a normal operator on a separable Hilbert space has at most countably
many eigenvalues.
13 Prove or give a counterexample: If T is a normal operator onp
a Hilbert space and
T = A + iB, where A and B are self-adjoint, then k T k = k Ak2 + k Bk2 .

A number α ∈ F is called an approximate eigenvalue of a bounded operator T


on a Hilbert space V if
inf{k( T − αI ) f k : f ∈ V and k f k = 1} = 0.
14 Suppose T is a normal operator on a Hilbert space and α ∈ F. Prove that
α ∈ sp( T ) if and only if α is an approximate eigenvalue of T.
15 Suppose T is a normal operator on a Hilbert space.
(a) Prove that if α is an eigenvalue of T, then |α|2 is an eigenvalue of T ∗ T.
(b) Prove that if α ∈ sp( T ), then |α|2 ∈ sp( T ∗ T ).

16 Suppose {ek }k∈Z+ is an orthonormal basis of a Hilbert space V. Suppose also


that T is a normal operator on V and ek is an eigenvector of T for every k ≥ 2.
Prove that e1 is an eigenvector of T.
17 Prove that if T is a self-adjoint operator on a Hilbert space, then k T n k = k T kn
for every n ∈ Z+.
18 Prove that if T is a normal operator on a Hilbert space, then k T n k = k T kn for
every n ∈ Z+.
19 Suppose T is an invertible operator on a Hilbert space. Prove that T is unitary if
and only if k T k = k T −1 k = 1.
20 Suppose T is a bounded operator on a complex Hilbert space, with T = A + iB,
where A and B are self-adjoint (see 10.54). Prove that T is unitary if and only if
T is normal and A2 + B2 = I.
[If z = x + yi, where x, y ∈ R, then |z| = 1 if and only if x2 + y2 = 1. Thus
this exercise strengthens the analogy between the unit circle in the complex
plane and the unitary operators.]
21 Suppose T is a unitary operator on a complex Hilbert space such that T − I is
invertible. Prove that
i ( T + I )( T − I )−1
is a self-adjoint operator.
[The function z 7→ i (z + 1)(z − 1)−1 maps {z ∈ C : |z| = 1} \ {1} to R.
Thus this exercise provides another useful illustration of the analogies showing
unitary ≈ {z ∈ C : |z| = 1} and self-adjoint ≈ R.]
Section 10B Spectrum 311

22 Suppose T is a self-adjoint operator on a complex Hilbert space. Prove that

( T + iI )( T − iI )−1

is a unitary operator.
[The function z 7→ (z + i )(z − i )−1 maps R to {z ∈ C : |z| = 1} \ {1}.
Thus this exercise provides another useful illustration of the analogies showing
(a) unitary ⇐⇒ {z ∈ C : |z| = 1}; (b) self-adjoint ⇐⇒ R.]

For T a bounded operator on a Banach space, define e T by



Tk
eT = ∑ k!
.
k =0

23 (a) Prove that if T is a bounded operator on a Banach space V, then the infinite
sum above converges in B(V ) and ke T k ≤ ekT k .
(b) Prove that if S, T are bounded operators on a Banach space V such that
ST = TS, then eS e T = eS+T .
(c) Prove that if T is a self-adjoint operator on a complex Hilbert space, then
eiT is unitary.

A bounded operator T on a Hilbert space is called a partial isometry if

k T f k = k f k for all f ∈ (null T )⊥ .

24 Suppose ( X, S , µ) is a σ-finite measure space and h ∈ L∞ (µ). As usual, let


Mh ∈ B L2 (µ) denote the multiplication operator defined by Mh f = f h.
Prove that Mh is a partial isometry if and only if there exists a set E ∈ S such
that h = χ E .
25 Suppose T is an isometry on a Hilbert space. Prove that T ∗ is a partial isometry.
26 Suppose T is a bounded operator on a Hilbert space V. Prove that T is a partial
isometry if and only if T ∗ T = PU for some closed subspace U of V.
312 Chapter 10 Linear Maps on Hilbert Spaces

10C Compact Operators


The Ideal of Compact Operators
A rich theory describes the behavior of compact operators, which we now define.

10.66 Definition compact operator

• An operator T on a Hilbert space V is called compact if for every bounded


sequence f 1 , f 2 , . . . in V, the sequence T f 1 , T f 2 , . . . has a convergent
subsequence.
• The collection of compact operators on V is denoted by C(V ).

The next result provides a large class of examples of compact operators. We will
see more examples after proving a few more results.

10.67 bounded operators with finite-dimensional range are compact

If T is a bounded operator on a Hilbert space and range T is finite-dimensional,


then T is compact.

Proof Suppose T is a bounded operator on a Hilbert space V and range T is


finite-dimensional. Suppose e1 , . . . , em is an orthonormal basis of range T (a finite
orthonormal basis of range T exists because the Gram–Schmidt process applied to
any basis of range T produces an orthonormal basis; see the proof of 8.67).
Now suppose f 1 , f 2 , . . . is a bounded sequence in V. For each n ∈ Z+, we have
T f n = h T f n , e1 i e1 + · · · + h T f n , e m i e m .
The Cauchy–Schwarz inequality shows that |h T f n , e j i| ≤ k T k sup k f k k for every
k ∈Z+
n ∈ Z+ and j ∈ {1, . . . , m}. Thus there exists a subsequence f n1 , f n2 , . . . such that
limk→∞ h T f nk , e j i exists in F for each j ∈ {1, . . . , m}. The equation displayed above
now implies that limk→∞ T f nk exists in V. Thus T is compact.

Not every bounded operator is compact. For example, the identity map on an
infinite-dimensional Hilbert space is not compact (to see this, consider an orthonormal
sequence, which does not have a convergent subsequence because √ the distance
between any two distinct elements of the orthonormal sequence is 2).

10.68 compact operators are bounded

Every compact operator on a Hilbert space is a bounded operator.

Proof We show that if T is an operator that is not bounded, then T is not compact.
To do this, suppose V is a Hilbert space and T is an operator on V that is not bounded.
Thus there exists a bounded sequence f 1 , f 2 , . . . in V such that limn→∞ k T f n k = ∞.
Hence no subsequence of T f 1 , T f 2 , . . . converges, which means T is not compact.
Section 10C Compact Operators 313

If V is a Hilbert space, then a two-


If V is finite-dimensional, then the
sided ideal of B(V ) is a subspace of
only two-sided ideals of B(V ) are
B(V ) that is closed under multiplication
{0} and B(V ). In contrast, if V is
on either side by bounded operators on V.
infinite-dimensional, then the next
The next result states that the set of com-
result shows that B(V ) has a closed
pact operators on V is a two-sided ideal
two-sided ideal that is neither {0}
of B(V ) that is closed in the topology on
nor V.
B(V ) that comes from the norm.

10.69 C(V ) is a closed two-sided ideal of B(V )

Suppose V is a Hilbert space.


(a) C(V ) is a closed subspace of B(V ).
(b) If T ∈ C(V ) and S ∈ B(V ), then ST ∈ C(V ) and TS ∈ C(V ).

Proof Suppose f 1 , f 2 , . . . is a bounded sequence in V.


To prove that C(V ) is closed under addition, suppose S, T ∈ C(V ). Because
S is compact, S f 1 , S f 2 , . . . has a convergent subsequence S f n1 , S f n2 , . . .. Because
T is compact, some subsequence of T f n1 , T f n2 , . . . converges. Thus we have a
subsequence of (S + T ) f 1 , (S + T ) f 2 , . . . that converges. Hence S + T ∈ C(V ).
The proof that C(V ) is closed under scalar multiplication is easier and is left to
the reader. Thus we now know that C(V ) is a subspace of B(V ).
To show that C(V ) is closed in B(V ), suppose T ∈ B(V ) and there is a sequence
T1 , T2 , . . . in C(V ) such that limm→∞ k T − Tm k = 0. To show that T is compact, we
need to show that T f n1 , T f n2 , . . . is a Cauchy sequence for some increasing sequence
of positive integers n1 < n2 < · · · .
Because T1 is compact, there is an infinite set Z1 ⊂ Z+ with k T1 f j − T1 f k k < 1
for all j, k ∈ Z1 . Let n1 be the smallest element of Z1 .
Now suppose m ∈ Z+ with m > 1 and an infinite set Zm−1 ⊂ Z+ and
nm−1 ∈ Zm−1 have been chosen. Because Tm is compact, there is an infinite set
Zm ⊂ Zm−1 with
k Tm f j − Tm f k k < m1
for all j, k ∈ Zm . Let nm be the smallest element of Zm such that nm > nm−1 .
Thus we produce an increasing sequence n1 < n2 < · · · of positive integers and
a decreasing sequence Z1 ⊃ Z2 ⊃ · · · of infinite subsets of Z+.
If m ∈ Z+ and j, k ≥ m, then
k T f n j − T f nk k ≤ k T f n j − Tm f n j k + k Tm f n j − Tm f nk k + k Tm f nk − T f nk k
1
≤ k T − Tm k(k f n j k + k f nk k) + m.
We can make the first term on the last line above as small as we want by choosing
m large (because limm→∞ k T − Tm k = 0 and the sequence f 1 , f 2 , . . . is bounded).
Thus T f n1 , T f n2 , . . . is a Cauchy sequence, as desired, completing the proof of (a).
To prove (b), suppose T ∈ C(V ) and S ∈ B(V ). Hence some subsequence of
T f 1 , T f 2 , . . . converges, and applying S to that subsequence gives another convergent
sequence. Thus ST ∈ C(V ). Similarly, S f 1 , S f 2 , . . . is a bounded sequence, and thus
T (S f 1 ), T (S f 2 ), . . . has a convergent subsequence; thus TS ∈ C(V ).
≈ 100π Chapter 10 Linear Maps on Hilbert Spaces

The previous result now allows us to see many new examples of compact operators.

10.70 compact integral operators

Suppose ( X, S , µ) is a σ-finite measure space, K ∈ L2 (µ × µ), and IK is the


integral operator on L2 (µ) defined by
Z
(IK f )( x ) = K ( x, y) f (y) dµ(y)
X

for f ∈ L2 (µ) and x ∈ X. Then IK is a compact operator.

Proof Example 10.5 shows that IK is a bounded operator on L2 (µ).


First consider the case where there exist g, h ∈ L2 (µ) such that
10.71 K ( x, y) = g( x )h(y)
for almost every ( x, y) ∈ X × X. In that case, if f ∈ L2 (µ) then
Z
(IK f )( x ) = g( x )h(y) f (y) dµ(y) = h f , hi g( x )
X

for almost every x ∈ X. Thus IK f = h f , hi g. In other words, IK has a one-


dimensional range in this case (or a zero-dimensional range if g = 0). Hence 10.67
implies that IK is compact.
Now consider the case where K is a finite sum of functions of the form given by
the right side of 10.71. Then because the set of compact operators on V is closed
under addition [by 10.69(a)], the operator IK is compact in this case.
Next, consider the case of K ∈ L2 (µ × µ) such that K is the limit in L2 (µ × µ)
of a sequence of functions K1 , K2 , . . ., each of which is of the form discussed in the
previous paragraph. Then
kIK − IKn k = kIK−Kn k ≤ kK − Kn k2 ,
where the inequality above comes from 10.8. Thus IK = limn→∞ IKn . By the
previous paragraph, each IKn is compact. Because the set of compact operators is a
closed subset of B(V ) [by 10.69(a)], we conclude that IK is compact.
We finish the proof by showing that the case considered in the previous paragraph
includes all K ∈ L2 (µ × µ). To do this, suppose F ∈ L2 (µ × µ) is orthogonal to all
the elements of L2 (µ × µ) of the form considered in the previous paragraph. Thus
Z Z Z
0= g( x )h(y) F ( x, y) d(µ×µ)( x, y) = g( x ) h(y) F ( x, y) dµ(y) dµ( x )
X×X X X

for all g, h ∈ L2 (µ) where we have used Tonelli’s Theorem, Fubini’s Theorem, and
Hölder’s inequality (with p = 2). For fixed h ∈ L2 (µ), the right side above equalling
0 for all g ∈ L2 (µ) implies that
Z
h(y) F ( x, y) dµ(y) = 0
X
for almost every x ∈ X. Now F ( x, y) = 0 for almost every ( x, y) ∈ X × X [because
the equation above holds for all h ∈ L2 (µ)], which by 8.42 completes the proof.
Section 10C Compact Operators 315

As a special case of the previous result, we can now see that the Volterra operator
V : L2 ([0, 1]) → L2 ([0, 1]) defined by
Z x
(V f )( x ) = f
0
is compact. This holds because, as shown in Example 10.15, the Volterra operator is
an integral operator of the type considered in the previous result.
R x The Volterra operator is injective [because differentiating both sides of the equation
0 f = 0 with respect to x and using the Lebesgue Differentiation Theorem (4.19)
shows that f = 0]. Thus the Volterra operator is an example of a compact operator
with infinite-dimensional range. The next example provides another class of compact
operators that do not necessarily have finite-dimensional range.

10.72 Example compact multiplication operators on `2


Suppose b1 , b2 , . . . is a sequence in F such that limn→∞ bn = 0. Define a bounded
linear map T : `2 → `2 by
T ( a1 , a2 , . . .) = ( a1 b1 , a2 b2 , . . .)
and for n ∈ Z+, define a bounded linear map Tn : `2 → `2 by
Tn ( a1 , a2 , . . .) = ( a1 b1 , a2 b2 , . . . , an bn , 0, 0, . . .).
Note that each Tn is a bounded operator with finite-dimensional range and thus is
compact (by 10.67). The condition limn→∞ bn = 0 implies that limn→∞ Tn = T.
Thus T is compact because C(V ) is a closed subset of B(V ) [by 10.69(a)].

The next result states that an operator is compact if and only if its adjoint is
compact.

10.73 T compact ⇐⇒ T ∗ compact

Suppose T is a bounded operator on a Hilbert space. Then T is compact if and


only if T ∗ is compact.

Proof First suppose T is compact. We want to prove that T ∗ is compact. To do


this, suppose f 1 , f 2 , . . . is a bounded sequence in V. Because TT ∗ is compact [by
10.69(b)], some subsequence TT ∗ f n1 , TT ∗ f n2 , . . . converges. Now


k T ∗ f n j − T ∗ f n k k2 = T ∗ ( f n j − f n k ), T ∗ ( f n j − f n k )


= TT ∗ ( f n j − f nk ), f n j − f nk

≤ k TT ∗ ( f n j − f nk )k k f n j − f nk k.
The inequality above implies that T ∗ f n1 , T ∗ f n2 , . . . is a Cauchy sequence and hence
converges. Thus T ∗ is a compact operator, completing the proof that if T is compact,
then T ∗ is compact.
Now suppose T ∗ is compact. By the result proved in the paragraph above, ( T ∗ )∗
is compact. Because ( T ∗ )∗ = T (see 10.11), we conclude that T is compact.
316 Chapter 10 Linear Maps on Hilbert Spaces

Spectrum of Compact Operator and Fredholm Alternative


We noted earlier that the identity map on an infinite-dimensional Hilbert space is not
compact. The next result shows that much more is true.

10.74 no infinite-dimensional closed subspace in range of compact operator

The range of each compact operator on a Hilbert space contains no infinite-


dimensional closed subspaces.

Proof Suppose T is a bounded operator on a Hilbert space V and U is an infinite-


dimensional closed subspace contained in range T. We want to show that T is not
compact.
Because T is a continuous operator, T −1 (U ) is a closed subspace of V. Let
S = T | T −1 (U ) . Thus S is a surjective bounded linear map from the Hilbert space
T −1 (U ) onto the Hilbert space U [here T −1 (U ) and U are Hilbert spaces by 6.16(b)].
The Open Mapping Theorem (6.81) implies S maps the open unit ball of T −1 (U ) to
an open subset of U. Thus there exists r > 0 such that
10.75 { g ∈ U : k gk < r } ⊂ { T f : f ∈ T −1 (U ) and k f k < 1}.
Because U is an infinite-dimensional Hilbert space, there exists an orthonormal
sequence e1 , e2 , . . . in U, as can be seen by applying the Gram–Schmidt process (see
ren
the proof of 8.67) to any linearly independent sequence in U. Each is in the left
2
− 1
side of 10.75. Thus for each n ∈ Z , there exists f n ∈ T (U ) such that k f n k < 1
+
ren
and T f n = . The sequence f 1 , f 2 , . . . is bounded, but the sequence T f 1 , T f 2 , . . .
2 √
re rek 2r
j
has no convergent subsequence because − = for j 6= k. Thus T is
2 2 2
not compact, as desired.

Suppose T is a compact operator on an infinite-dimensional Hilbert space. The


result above implies that T is not surjective. In particular, T is not invertible. Thus
we have the following result.

10.76 compact implies not invertible on infinite-dimensional Hilbert spaces

If T is a compact operator on an infinite-dimensional Hilbert space, then


0 ∈ sp( T ).

Although 10.74 shows that if T is compact then range T contains no infinite-


dimensional closed subspaces, the next result shows that the situation differs drasti-
cally for T − αI if α ∈ F \ {0}.
The proof of the next result makes use of the restriction of T − αI to the closed
⊥
subspace null( T − αI ) . As motivation for considering this restriction, recall that
each f ∈ V can be written uniquely as f = g + h, where g ∈ null( T − αI ) and
⊥
h ∈ null( T − αI ) (see 8.43). Thus ( T − αI ) f = ( T − αI )h, which implies that
⊥ 
range( T − αI ) = ( T − αI ) null( T − αI ) .
Section 10C Compact Operators 317

10.77 closed range

If T is a compact operator on a Hilbert space, then T − αI has closed range for


every α ∈ F with α 6= 0.

Proof Suppose T is a compact operator on a Hilbert space V and α ∈ F is such that


α 6= 0.

10.78 Claim: there exists r > 0 such that


⊥
k f k ≤ r k( T − αI ) f k for all f ∈ null( T − αI ) .

To prove the claim above, suppose it is false. Then for each n ∈ Z+, there exists
⊥
f n ∈ null( T − αI ) such that

k f n k = 1 and k( T − αI ) f n k < n1 .

Because T is compact, there exists a subsequence T f n1 , T f n2 , . . . such that

10.79 lim T f nk = g
k→∞

for some g ∈ V. Subtracting the equation

10.80 lim ( T − αI ) f nk = 0
k→∞

from 10.79 and then dividing by α shows that

lim f nk = α1 g.
k→∞
⊥
The equation above implies k gk = |α|; hence g 6= 0. Each f nk ∈ null( T − αI ) ;
⊥
hence we also conclude that g ∈ null( T − αI ) . Applying T − αI to both sides of
the equation above and using 10.80 shows that g ∈ null( T − αI ). Thus g is a nonzero
element of both null( T − αI ) and its orthogonal complement. This contradiction
completes the proof of the claim in 10.78.
To show that range( T − αI ) is closed, suppose h1 , h2 , . . . is a sequence in
range( T − αI ) that converges to some h ∈ V. For each n ∈ Z+, there exists
⊥
f n ∈ null( T − αI ) such that ( T − αI ) f n = hn . Because h1 , h2 , . . . is a Cauchy
sequence, 10.78 shows that f 1 , f 2 , . . . is also a Cauchy sequence. Thus there exists
f ∈ V such that limn→∞ f n = f , which implies h = ( T − αI ) f ∈ range( T − αI ).
Hence range( T − αI ) is closed.

Suppose T is a compact operator on a Hilbert space V and f ∈ V and α ∈ F \ {0}.


An immediate consequence (often useful when investigating integral equations) of
the result above and 10.13(d) is that the equation

Tg − αg = f

has a solution g ∈ V if and only if h f , hi = 0 for every h ∈ V such that T ∗ h = αh.


318 Chapter 10 Linear Maps on Hilbert Spaces

10.81 Definition geometric multiplicity

• The geometric multiplicity of an eigenvalue α of an operator T is defined to


be the dimension of null( T − αI ).
• In other words, the geometric multiplicity of an eigenvalue α of T is the
dimension of the subspace consisting of 0 and all the eigenvectors of T
corresponding to α.

There exist compact operators for which the eigenvalue 0 has infinite geometric
multiplicity. The next result shows that this cannot happen for nonzero eigenvalues.

10.82 nonzero eigenvalues of compact operators have finite multiplicity

Suppose T is a compact operator on a Hilbert space and α ∈ F with α 6= 0. Then


null( T − αI ) is finite-dimensional.
f
Proof Suppose f ∈ null( T − αI ). Then f = T α . Hence f ∈ range T.
Thus we have shown that null( T − αI ) ⊂ range T. Because T is continuous,
null( T − αI ) is closed. Thus 10.74 implies that null( T − αI ) is finite-dimensional.

The next lemma is used in our proof of the Fredholm Alternative (10.85). Note
that this lemma implies that every injective operator on a finite-dimensional vector
space is surjective (because a finite-dimensional vector space cannot have an infinite
chain of strictly decreasing subspaces—the dimension decreases by at least 1 in each
step). Also, see Exercise 10 for the analogous result implying that every surjective
operator on a finite-dimensional vector space is injective.

10.83 injective but not surjective

If T is an injective but not surjective operator on a vector space, then

range T % range T 2 % range T 3 % · · · .

Proof Suppose T is an injective but not surjective operator on a vector space V.


Suppose n ∈ Z+. If g ∈ V, then

T n+1 g = T n ( Tg) ∈ range T n .

Thus range T n ⊃ range T n+1 .


To show that the last inclusion is not an equality, note that because T is not
surjective, there exists f ∈ V such that

10.84 / range T.
f ∈

Now T n f ∈ range T n . However, T n f ∈ / range T n+1 because if g ∈ V and


n
T f =T n + 1 g, then T f = T ( Tg), which would imply that f = Tg (because T n
n n

is injective), which would contradict 10.84. Thus range T n % range T n+1 .


Section 10C Compact Operators 319

Compact operators behave, in some respects, like operators on a finite-dimensional


vector space. For example, the following important theorem should be familiar to you
in the finite-dimensional context (where the choice of α = 0 need not be excluded).

10.85 Fredholm Alternative

Suppose T is a compact operator on a Hilbert space and α ∈ F with α 6= 0. Then


the following are equivalent:

(a) α ∈ sp( T ).
(b) α is an eigenvalue of T.
(c) T − αI is not surjective.

Proof Clearly (b) implies (a) and (c) implies (a).


To prove that (a) implies (b), suppose α ∈ sp( T ) but α is not an eigenvalue of T.
Thus T − αI is injective but T − αI is not surjective. Thus 10.83 applied to T − αI
shows that
10.86 range( T − αI ) % range( T − αI )2 % range( T − αI )3 % · · · .
If n ∈ Z+, then the Binomial Theorem and 10.69 show that
( T − αI )n = S + (−α)n I
for some compact operator S. Now 10.77 shows that range( T − αI )n is a closed
subspace of the Hilbert space on which T operates. Thus 10.86 implies that for each
n ∈ Z+, there exists
⊥
10.87 f n ∈ range( T − αI )n ∩ range( T − αI )n+1
such that k f n k = 1.
Now suppose j, k ∈ Z+ with j < k. Then
10.88 T f j − T f k = ( T − αI ) f j − ( T − αI ) f k − α f k + α f j .

Because f j and f k are both in range( T − αI ) j , the first two terms on the right side of
10.88 are in range( T − αI ) j+1 . Because j + 1 ≤ k, the third term in 10.88 is also in
range( T − αI ) j+1 . Now 10.87 implies that the last term in 10.88 is orthogonal to
the sum of the first three terms. Thus 10.88 leads to the inequality
k T f j − T f k k ≥ k α f j k = | α |.
The inequality above implies that T f 1 , T f 2 , . . . has no convergent subsequence, which
contradicts the compactness of T. This contradiction means the assumption that α is
not an eigenvalue of T was false, completing the proof that (a) implies (b).
At this stage, we know that (a) and (b) are equivalent and that (c) implies (a). To
prove that (a) implies (c), suppose α ∈ sp( T ). Thus α ∈ sp( T ∗ ). Applying the
equivalence of (a) and (b) to T ∗, we conclude that α is an eigenvalue of T ∗. Thus
applying 10.13(d) to T − αI shows that T − αI is not surjective, completing the proof
that (a) implies (c).
320 Chapter 10 Linear Maps on Hilbert Spaces

The previous result traditionally has the word alternative in its name because it
can be rephrased as follows:
If T is a compact operator on a Hilbert space V and α ∈ F \ {0}, then
exactly one of the following holds:

1. the equation T f = α f has a nonzero solution f ∈ V;


2. the equation g = T f − α f has a solution f ∈ V for every g ∈ V.
The next example shows the power of the Fredholm Alternative. In this example,
we want to show that V − αI is invertible for all α ∈ F \ {0}. The verification
that V − αI is injective is straightforward. Showing that V − αI is surjective would
require more work. However, the Fredholm Alternative tells us, with no further work,
that V − αI is invertible.

10.89 Example spectrum of the Volterra operator


We want to show that the spectrum of the Volterra operator V is {0} (see Example
10.15 for the definition of V ). The Volterra operator V is compact (see the comment
after the proof of 10.70). Thus 0 ∈ sp(V ), by 10.76.
Suppose α ∈ F \ {0}. To show that α ∈ / sp(V ), we need only show that α is not
an eigenvalue of V (by 10.85). Thus suppose f ∈ L2 ([0, 1]) and V f = α f . Hence
Z x
10.90 f = α f (x)
0

for almost every x ∈ [0, 1]. The left side of 10.90 is a continuous function of x and
thus so is the right side, which implies that f is continuous. The continuity of f
now implies that the left side of 10.90 has a continuous derivative, and thus f has a
continuous derivative.
Now differentiate both sides of 10.90 with respect to x, getting

f (x) = α f 0 (x)

for all x ∈ (0, 1). Standard calculus shows that the equation above implies that

f ( x ) = ce x/α

for some constant c. However, 10.90 implies that the continuous function f must
satisfy the equation f (0) = 0. Thus c = 0, which implies f = 0.
The conclusion of the last paragraph shows that α is not an eigenvalue of V . The
Fredholm Alternative (10.85) now shows that α ∈ / sp(V ). Thus sp(V ) = {0}.

If α is an eigenvalue of an operator T on a finite-dimensional Hilbert space,


then α is an eigenvalue of T ∗. This result does not hold for bounded operators on
infinite-dimensional Hilbert spaces.
However, suppose T is a compact operator on a Hilbert space and α is a nonzero
eigenvalue of T. Thus α ∈ sp( T ), which implies that α ∈ sp( T ∗ ) (because a
bounded operator is invertible if and only if its adjoint is invertible). The Fredholm
Alternative (10.85) now shows that α is an eigenvalue of T ∗. Thus compactness
allows us to recover the finite-dimensional result (except for the case α = 0).
Section 10C Compact Operators 321

Our next result states that if T is a compact operator and α 6= 0, then null( T − αI )
and null( T ∗ − αI ) have the same dimension (denoted dim). This result about
the dimensions of spaces of eigenvectors is easier to prove in finite dimensions.
Specifically, suppose S is an operator on a finite-dimensional Hilbert space V (you
can think of S = T − αI). Then
dim null S = dim V − dim range S = dim(range S)⊥ = dim null S∗,
where the justification for each step should be familiar to you from finite-dimensional
linear algebra. This finite-dimensional proof does not work in infinite dimensions
because the expression dim V − dim range S could be of the form ∞ − ∞.
Although the dimensions of the two null spaces in the result below are the same,
even in finite dimensions the two null spaces are not necessarily equal to each other
(but we do have equality of the two null spaces when T is normal; see 10.56).
Note that both dimensions in the result below are finite (by 10.82 and 10.73).

10.91 null spaces of T − αI and T ∗ − αI have same dimensions

Suppose T is a compact operator on a Hilbert space and α ∈ F with α 6= 0. Then

dim null( T − αI ) = dim null( T ∗ − αI ).

Proof Suppose dim null( T − αI ) < dim null( T ∗ − αI ). Because null( T ∗ − αI )


⊥
equals range( T − αI ) , there is a bounded injective linear map
⊥
R : null( T − αI ) → range( T − αI )
that is not surjective. Let V denote the Hilbert space on which T operates, and let P
be the orthogonal projection of V onto null( T − αI ). Define a linear map S : V → V
by
S = T + RP.
Because RP is a bounded operator with finite-dimensional range, S is compact. Also,
S − αI = ( T − αI ) + RP.
Every element of range( T − αI ) is orthogonal to every element of range RP.
Suppose f ∈ V and (S − αI ) f = 0. The equation above shows that ( T − αI ) f = 0
and RP f = 0. Because f ∈ null( T − α f ), we see that P f = f , which then implies
that R f = RP f = 0, which then implies that f = 0 (because R is injective). Hence
S − αI is injective.
⊥
However, because R maps onto a proper subset of range( T − αI ) , we see that
S − αI is not surjective, which contradicts the equivalence of (b) and (c) in 10.85. This
contradiction means the assumption that dim null( T − αI ) < dim null( T ∗ − αI )
was false. Hence we have proved that
10.92 dim null( T − αI ) ≥ dim null( T ∗ − αI )
for every compact operator T and every α ∈ F \ {0}.
Now apply the conclusion of the previous paragraph to T ∗ (which is compact by
10.73) and α, getting 10.92 with the inequality reversed, completing the proof.
322 Chapter 10 Linear Maps on Hilbert Spaces

The spectrum of an operator on a finite-dimensional Hilbert space is a finite set,


consisting just of the eigenvalues of the operator. The spectrum of a compact operator
on an infinite-dimensional Hilbert space can be an infinite set. However, our next
result implies that if a compact operator has infinite spectrum, then that spectrum
consists of 0 and a sequence in F with limit 0.

10.93 spectrum of a compact operator

Suppose T is a compact operator on a Hilbert space. Then

{α ∈ sp( T ) : |α| ≥ δ}

is a finite set for every δ > 0.

Proof Fix δ > 0. Suppose there exist distinct α1 , α2 , . . . in sp( T ) with |αn | ≥ δ
for every n ∈ Z+. The Fredholm Alternative (10.85) implies that each αn is an
eigenvalue of T. For n ∈ Z+, let

Un = null ( T − α1 I ) · · · ( T − αn I ) .

and let U0 = {0}. Because T is continuous, each Un is a closed subspace of the


Hilbert space on which T operates. Furthermore, Un−1 ⊂ Un for each n ∈ Z+
because operators of the form T − α j I and T − αk I commute with each other.
If n ∈ Z+ and g is an eigenvector of T corresponding to the eigenvalue αn , then
g ∈ Un but g ∈/ Un−1 because

( T − α1 I ) · · · ( T − αn−1 I ) g = (αn − α1 ) · · · (αn − αn−1 ) g 6= 0.

In other words, we have


U1 $ U2 $ U3 $ · · · .
Thus for each n ∈ Z+, there exists

10.94 en ∈ Un ∩ (Un−1 ⊥ )

such that ken k = 1.


Now suppose j, k ∈ Z+ with j < k. Then

10.95 Te j − Tek = ( T − α j I )e j − ( T − αk I )ek + α j e j − αk ek .

Because j ≤ k − 1, the first three terms on the right side of 10.95 are in Uk−1 . Now
10.94 implies that the last term in 10.95 is orthogonal to the sum of the first three
terms. Thus 10.95 leads to the inequality

k Te j − Tek k ≥ kαk ek k = |αk | ≥ δ.

The inequality above implies that Te1 , Te2 , . . . has no convergent subsequence, which
contradicts the compactness of T. This contradiction means that the assumption that
sp( T ) contains infinitely many elements with absolute value at least δ was false.
Section 10C Compact Operators 323

EXERCISES 10C
1 Prove that if T is a compact operator on a Hilbert space V and e1 , e2 , . . . is an
orthonormal sequence in V, then limn→∞ Ten = 0.

2 Prove that if T is a compact operator on L2 ([0, 1]), then lim nk T ( x n )k2 = 0,
n→∞
where x n means the element of L2 ([0, 1]) defined by x 7→ x n .
3 Suppose T is a compact operator on a Hilbert space V and f 1 , f 2 , . . . is a
sequence in V such that limn→∞ h f n , gi = 0 for every g ∈ V. Prove that
limn→∞ k T f n k = 0.

4 Suppose h ∈ L∞ (R). Define Mh ∈ B L2 (R) by Mh f = f h. Prove that if
khk∞ > 0, then Mh is not compact.
5 Suppose (b1 , b2 , . . .) ∈ `∞ . Define T : `2 → `2 by

T ( a1 , a2 , . . .) = ( a1 b1 , a2 b2 , . . .).

Prove that T is compact if and only if lim bn = 0.


n→∞

6 Suppose T is a bounded operator on a Hilbert space V. Prove that if there exists


an orthonormal basis {ek }k∈Γ of V such that

∑ kTek k2 < ∞,
k∈Γ

then T is compact.
7 Suppose T is a bounded operator on a Hilbert space V. Prove that if {ek }k∈Γ
and { f j } j∈Ω are orthonormal bases of V, then

∑ kTek k2 = ∑ kT f j k2 .
k∈Γ j∈Ω

8 Suppose T is a bounded operator on a Hilbert space. Prove that T is compact if


and only if T ∗ T is compact.
9 Prove that if T is a compact operator on an infinite-dimensional Hilbert space,
then k I − T k ≥ 1.
10 Show that if T is a surjective but not injective operator on a vector space V, then

null T $ null T 2 $ null T 3 $ · · · .

11 Suppose T is a compact operator on a Hilbert space and α ∈ F \ {0}.


(a) Prove that range( T − αI )m−1 = range( T − αI )m for some m ∈ Z+.
(b) Prove that null( T − αI )n−1 = null( T − αI )n for some n ∈ Z+.
(c) Show that the smallest positive integer m that works in (a) equals the
smallest positive integer n that works in (b).
324 Chapter 10 Linear Maps on Hilbert Spaces

12 Prove that if f : [0, 1] → F is a continuous function, then there exists a continu-


ous function g : [0, 1] → F such that
Z x
f ( x ) = g( x ) + g
0
for all x ∈ [0, 1].
13 Suppose S is a bounded invertible operator on a Hilbert space V and T is a
compact operator on V.
(a) Prove that S + T has closed range.
(b) Prove that S + T is injective if and only if S + T is surjective.
(c) Prove that null(S + T ) and null(S∗ + T ∗ ) are finite-dimensional.
(d) Prove that dim null(S + T ) = dim null(S∗ + T ∗ ).
(e) Prove that there exists R ∈ B(V ) such that range R is finite-dimensional
and S + T + R is invertible.

14 Suppose T is a compact operator on a Hilbert space V. Prove that range T is a


separable subspace of V.
15 Suppose T is a compact operator on a Hilbert space V and e1 , e2 , . . . is an
orthonormal basis of range T. Let Pn denote the orthogonal projection of V
onto span{e1 , . . . , en }.
(a) Prove that lim k T − Pn T k = 0.
n→∞
(b) Prove that an operator on a Hilbert space V is compact if and only if it is the
limit in B(V ) of a sequence of bounded operators with finite-dimensional
range.

16 Prove that if T is a compact operator on a Hilbert space V, then there exists a


sequence S1 , S2 , . . . of invertible operators on V such that lim k T − Sn k = 0.
n→∞

17 Suppose T is a bounded operator on a Hilbert space such that p( T ) is compact


for some nonzero polynomial p with coefficients in F. Prove that sp( T ) is a
countable set.

Suppose T is a bounded operator on a Hilbert space. The algebraic multiplicity


of an eigenvalue α of T is defined to be the dimension of the subspace

[
null( T − αI )n .
n =1
As an easy example, if T is the left shift as defined in the next exercise, then the
eigenvalue 0 of T has geometric multiplicity 1 but algebraic multiplicity ∞.
The definition above of algebraic multiplicity is equivalent on finite-dimensional
spaces to the common definition involving the multiplicity of a root of the charac-
teristic polynomial. However, the definition used here is cleaner (no determinants
needed) and has the advantage of working on infinite-dimensional Hilbert spaces.
Section 10C Compact Operators 325

18 Suppose T ∈ B(`2 ) is defined by T ( a1 , a2 , a3 , . . .) = ( a2 , a3 , a4 , . . .). Suppose


also that α ∈ F and |α| < 1.
(a) Show that the geometric multiplicity of α as an eigenvalue of T equals 1.
(b) Show that the algebraic multiplicity of α as an eigenvalue of T equals ∞.

19 Prove that the geometric multiplicity of an eigenvalue of a normal operator on a


Hilbert space equals the algebraic multiplicity of that eigenvalue.
20 Prove that every nonzero eigenvalue of a compact operator on a Hilbert space
has finite algebraic multiplicity.
21 Prove that if T is a compact operator on a Hilbert space and α is a nonzero
eigenvalue of T, then the algebraic multiplicity of α as an eigenvalue of T equals
the algebraic multiplicity of α as an eigenvalue of T ∗.
22 Prove that if V is a separable Hilbert space, then the Banach space C(V ) is
separable.
326 Chapter 10 Linear Maps on Hilbert Spaces

10D Spectral Theorem for Compact


Operators
Orthonormal Bases Consisting of Eigenvectors
We begin this section with the following useful lemma.

10.96 T ∗ T − k T k2 I is not invertible

If T is a bounded operator on a nonzero Hilbert space, then k T k2 ∈ sp( T ∗ T ).

Proof Suppose T is a bounded operator on a nonzero Hilbert space V. Let f 1 , f 2 , . . .


be a sequence in V such that k f n k = 1 for each n ∈ Z+ and

10.97 lim k T f n k = k T k.
n→∞

Then

T T f n − k T k2 f n 2 = k T ∗ T f n k2 − 2k T k2 h T ∗ T f n , f n i + k T k4

= k T ∗ T f n k2 − 2k T k2 k T f n k2 + k T k4

10.98 ≤ 2k T k4 − 2k T k2 k T f n k2 ,

where the last line holds because k T ∗ T f n k ≤ k T ∗ k k T f n k ≤ k T k2 . Now 10.97


and 10.98 imply that
lim ( T ∗ T − k T k2 I ) f n = 0.
n→∞

Because k f n k = 1 for each n ∈ Z+, the equation above implies that T ∗ T − k T k2 I


is not invertible, as desired.

The next result indicates one way in which self-adjoint compact operators behave
like self-adjoint operators on finite-dimensional Hilbert spaces.

10.99 every self-adjoint compact operator has an eigenvalue.

Suppose T is a self-adjoint compact operator on a nonzero Hilbert space. Then


either k T k or −k T k is an eigenvalue of T.

Proof Because T is self-adjoint, 10.96 states that T 2 − k T k2 I is not invertible. Now


 
T 2 − k T k2 I = T − k T k I T + k T k I .

Thus T − k T k I and T + k T k I cannot both be invertible. Hence k T k ∈ sp( T ) or


−k T k ∈ sp( T ). Because T is compact, 10.85 now implies that k T k or −k T k is an
eigenvalue of T, as desired, or that k T k = 0, which means that T = 0, in which case
0 is an eigenvalue of T.
Section 10D Spectral Theorem for Compact Operators 327

If T is an operator on a vector space V and U is a subspace of V, then T |U is a


linear map from U to V. For T |U to be an operator (meaning that it is a linear map
from a vector space to itself), we need T (U ) ⊂ U. Thus we are led to the following
definition.

10.100 Definition invariant subspace

Suppose T is an operator on a vector space V. A subspace U of V is called an


invariant subspace for T if T f ∈ U for every f ∈ U.

10.101 Example invariant subspaces


You should verify each of the assertions below.
• For b ∈ [0, 1], the subspace
{ f ∈ L2 ([0, 1]) : f (t) = 0 for almost every t ∈ [0, b]}
is an invariant subspaceR for the Volterra operator V : L2 ([0, 1]) → L2 ([0, 1])
x
defined by (V f )( x ) = 0 f .
• Suppose T is an operator on a Hilbert space V and f ∈ V with f 6= 0. Then
span{ f } is an invariant subspace for T if and only if f is an eigenvector of T.
• Suppose T is an operator on a Hilbert space V. Then {0}, V, null T, and range T
are invariant subspaces for T.
• If T is a bounded operator on a Hilbert space and U is an invariant subspace for
T, then U is an invariant subspace for T.

If T is a compact operator on a Hilbert


The most important open question in
space and U is an invariant subspace for
operator theory is the invariant
T, then T |U is a compact operator on U,
subspace problem, which asks
as follows from the definitions.
whether every bounded operator on
If U is an invariant subspace for a
a Hilbert space with dimension
self-adjoint operator T, then T |U is self-
greater than 1 has a closed invariant
adjoint because
subspace other than {0} and V.

h( T |U ) f , gi = h T f , gi = h f , Tgi = h f , ( T |U ) gi
for all f , g ∈ U. The next result shows that a bit more is true.

10.102 U invariant for self-adjoint T implies U ⊥ invariant for T

Suppose U is an invariant subspace for a self-adjoint operator T. Then

(a) U ⊥ is also an invariant subspace for T;


(b) T |U ⊥ is a self-adjoint operator on U ⊥.
328 Chapter 10 Linear Maps on Hilbert Spaces

Proof To prove (a), suppose f ∈ U ⊥. If g ∈ U, then


h T f , gi = h f , Tgi = 0,
where the first equality holds because T is self-adjoint and the second equality holds
because Tg ∈ U and f ∈ U ⊥. Because the equation above holds for all g ∈ U, we
conclude that T f ∈ U ⊥. Thus U ⊥ is an invariant subspace for T, proving (a).
By part (a), we can think of T |U ⊥ as an operator on U ⊥. To prove (b), suppose
h ∈ U ⊥. If f ∈ U ⊥, then
h f , ( T |U ⊥ )∗ hi = h T |U ⊥ f , hi = h T f , hi = h f , Thi = h f , T |U ⊥ hi.
Because ( T |U ⊥ )∗ h and T |U ⊥ h are both in U ⊥ and the equation above holds for all
f ∈ U ⊥, we conclude that ( T |U ⊥ )∗ h = T |U ⊥ h, proving (b).

Operators for which there exists an orthonormal basis consisting of eigenvectors


may be the easiest operators to understand. The next result states that any such
operator must be self-adjoint in the case of a real Hilbert space and normal in the
case of a complex Hilbert space.

10.103 orthonormal basis of eigenvectors implies self-adjoint or normal

Suppose T is a bounded operator on a Hilbert space V and there is an orthonormal


basis of V consisting of eigenvectors of T.

(a) If F = R, then T is self-adjoint.


(b) If F = C, then T is normal.

Proof Suppose {e j } j∈Γ is an orthonormal basis of V such that e j is an eigenvector


of T for each j ∈ Γ. Thus there exists a family {α j } j∈Γ in F such that
10.104 Te j = α j e j
for each j ∈ Γ. If k ∈ Γ and f ∈ V, then
D   E
h f , T ∗ ek i = h T f , ek i = T ∑ h f , e j i e j , ek
j∈Γ
D E
= ∑ α j h f , e j i e j , ek = α k h f , e k i = h f , α k e k i.
j∈Γ

The equation above implies that


10.105 T ∗ ek = αk ek .
To prove (a), suppose F = R. Then 10.105 and 10.104 imply T ∗ ek = αk ek = Tek
for each k ∈ K. Hence T ∗ = T, completing the proof of (a).
To prove (b), now suppose F = C. If k ∈ Γ, then 10.105 and 10.104 imply that
( T ∗ T )(ek ) = T ∗ (αk ek ) = |αk |2 ek = T (αk ek ) = ( TT ∗ )(ek ).
Because the equation above holds for all k ∈ Γ, we conclude that T ∗ T = TT ∗. Thus
T is normal, completing the proof of (b).
Section 10D Spectral Theorem for Compact Operators 329

The next result is one of the major highlights of the theory of compact operators
on Hilbert spaces. The result as stated below applies to both real and complex Hilbert
spaces. In the case of a real Hilbert space, the result below can be combined with
10.103(a) to produce the following result: A compact operator on a real Hilbert
space is self-adjoint if and only if there is an orthonormal basis of the Hilbert space
consisting of eigenvectors of the operator.

10.106 Spectral Theorem for self-adjoint compact operators

Suppose T is a self-adjoint compact operator on a Hilbert space V. Then

(a) there is an orthonormal basis of V consisting of eigenvectors of T;


(b) there is a countable set Ω, an orthonormal family {ek }k∈Ω in V, and a family
{αk }k∈Ω in R \ {0} such that

Tf = ∑ α k h f , ek i ek
k∈Ω

for every f ∈ V.

Proof Let U denote the span of all the eigenvectors of T. Then U is an invariant
subspace for T. Hence U ⊥ is also an invariant subspace for T and T |U ⊥ is a self-
adjoint operator on U ⊥ (by 10.102). However, T |U ⊥ has no eigenvalues, because
all the eigenvectors of T are in U. Because all self-adjoint compact operators on a
nonzero Hilbert space have an eigenvalue (by 10.99), this implies that U ⊥ = {0}.
Hence U = V (by 8.42).
For each eigenvalue α of T, there is an orthonormal basis of null( T − αI ) consist-
ing of eigenvectors corresponding to the eigenvalue α. The union (over all eigenvalues
α of T) of all these orthonormal bases is an orthonormal family in V because eigen-
vectors corresponding to distinct eigenvalues are orthogonal (see 10.57). The previous
paragraph tells us that the closure of the span of this orthonormal family is V (here
we are using the set itself as the index set). Hence we have an orthonormal basis of
V consisting of eigenvectors of T, completing the proof of (a).
By part (a) of this result, there is an orthonormal basis {ek }k∈Γ of V and a family
{αk }k∈Γ in R such that Tek = αk ek for each k ∈ Γ (even if F = C, the eigenvalues
of T are in R by 10.49) . Thus if f ∈ V, then
 
T f = T ∑ h f , ek iek = ∑ h f , ek i Tek = ∑ αk h f , ek iek .
k∈Γ k∈Γ k∈Γ

Letting Ω = {k ∈ Γ : αk 6= 0}, we can rewrite the equation above as

Tf = ∑ αk h f , ek i ek
k∈Ω

for every f ∈ V. The set Ω is countable because T has only countably many
eigenvalues (by 10.93) and each nonzero eigenvalue can appear only finitely many
times in the sum above (by 10.82), completing the proof of (b).
330 Chapter 10 Linear Maps on Hilbert Spaces

A normal compact operator on a nonzero real Hilbert space might have no eigen-
values [consider, for example the normal operator T of counterclockwise rotation by
a right angle on R2 defined by T ( x, y) = (−y, x )]. However, the next result shows
that normal compact operators on complex Hilbert spaces behave better. The key idea
in proving this result is that on a complex Hilbert space, the real and imaginary parts
of a normal compact operator are commuting self-adjoint compact operators, which
then allows us to apply the Spectral Theorem for self-adjoint compact operators.

10.107 Spectral Theorem for normal compact operators

Suppose T is a compact operator on a complex Hilbert space V. Then there is an


orthonormal basis of V consisting of eigenvectors of T if and only if T is normal.

Proof One direction of this result has already been proved as part (b) of 10.103.
To prove the other direction, suppose T is a normal compact operator. We can
write
T = A + iB,
where A and B are self-adjoint operators and, because T is normal, AB = BA (see
10.54). Because A = ( T + T ∗ )/2 and B = ( T − T ∗ )/(2i ), the operators A and B
are both compact.
If α ∈ R and f ∈ null( A − αI ), then

( A − αI )( B f ) = A( B f ) − αB f = B( A f ) − αB f = B ( A − αI ) f = B(0) = 0
and thus B f ∈ null( A − αI ). Hence null( A − αI ) is an invariant subspace for B.
Applying the Spectral Theorem for self-adjoint compact operators [10.106(a)] to
B|null( A−αI ) shows that for each eigenvalue α of A, there is an orthonormal basis of
null( A − αI ) consisting of eigenvectors of B. The union (over all eigenvalues α of
A) of all these orthonormal bases is an orthonormal family in V (use the set itself as
the index set) because eigenvectors of A corresponding to distinct eigenvalues of A
are orthogonal (see 10.57). The Spectral Theorem for self-adjoint compact operators
[10.106(a)] as applied to A tells us that the closure of the span of this orthonormal
family is V. Hence we have an orthonormal basis of V, each of whose elements is an
eigenvector of A and an eigenvector of B.
If f ∈ V is an eigenvector of both A and B, then there exist α, β ∈ R such
that A f = α f and B f = β f . Thus T f = ( A + iB)( f ) = (α + βi ) f ; hence f is
an eigenvector of T. Thus the orthonormal basis of V constructed in the previous
paragraph is an orthonormal basis consisting of eigenvectors of T, completing the
proof.

The following example shows the power of the Spectral Theorem for normal
compact operators. Finding the eigenvalues and eigenvectors of the normal compact
operator V − V ∗ in the next example leads us to an orthonormal basis of L2 ([0, 1]).
Easy calculus shows that the family {ek }k∈Z , where ek is defined as in 10.112, is
an orthonormal family in L2 ([0, 1]). The hard part of showing that {ek }k∈Z is an
orthonormal basis of L2 ([0, 1]) is to show that the closure of the span of this family
is L2 ([0, 1]). However, the Spectral Theorem for normal compact operators (10.107)
provides this information with no further work required.
Section 10D Spectral Theorem for Compact Operators 331

10.108 Example an orthonormal basis of eigenvectors


Suppose V : L2 ([0, 1]) → L2 ([0, 1]) is the Volterra operator defined by
Z x
(V f )( x ) = f.
0
The operator V is compact (see the paragraph after the proof of 10.70), but it is not
normal. Because V is compact, so is V ∗ (by 10.73). Hence V − V ∗ is compact. Also,
(V − V ∗ )∗ = V ∗ − V = −(V − V ∗ ). Because every operator commutes with its
negative, we conclude that V − V ∗ is a compact normal operator. Because we want
to apply the Spectral Theorem, for the rest of this example we will take F = C.
If f ∈ L2 ([0, 1]) and x ∈ [0, 1], then the formula for V ∗ given by 10.16 shows
that
 Z x Z 1
10.109 (V − V ∗ ) f ( x ) = 2 f− f.
0 0
The right side of the equation above is a continuous function of x whose value at
x = 0 is the negative of its value at x = 1.
Differentiating both sides of the equation above and using the Lebesgue Differen-
tiation Theorem (4.19) shows that
0
(V − V ∗ ) f ( x ) = 2 f ( x )
for almost every x ∈ [0, 1]. If f ∈ null(V − V ∗ ), then differentiating both sides
of the equation (V − V ∗ ) f = 0 shows that 2 f ( x ) = 0 for almost every x ∈ [0, 1];
hence f = 0, and we conclude that V − V ∗ is injective (so 0 is not an eigenvalue).
Suppose f is an eigenvector of V − V ∗ with eigenvalue α. Thus f is in the range
of V − V ∗, which by 10.109 implies that f is continuous on [0, 1], which by 10.109
again implies that f is continuously differentiable on (0, 1). Differentiating both
sides of the equation (V − V ∗ ) f = α f gives
2 f (x) = α f 0 (x)
for all x ∈ (0, 1). Hence the function whose value at x is e−(2/α) x f ( x ) has derivative
0 everywhere on (0, 1) and thus is a constant function. In other words,
10.110 f ( x ) = ce(2/α) x
for some constant c 6= 0. Because f ∈ range(V − V ∗ ), we have f (0) = − f (1),
which with the equation above implies that there exists k ∈ Z such that
10.111 2/α = i (2k + 1)π.
Replacing 2/α in 10.110 with the value of 2/α derived in 10.111 shows that for
k ∈ Z, we should define ek ∈ L2 ([0, 1]) by
10.112 ek ( x ) = ei(2k+1)πx .
Clearly {ek }k∈Z is an orthonormal family in L2 ([0, 1]) [the orthogonality can be
verified by a straightforward calculation or by using 10.57]. The paragraph above
and the Spectral Theorem for compact normal operators (10.107) imply that this
orthonormal family is an orthonormal basis of L2 ([0, 1]).
332 Chapter 10 Linear Maps on Hilbert Spaces

Singular Value Decomposition


The next result provides an important generalization of 10.106(b) to arbitrary compact
operators that need not be self-adjoint or normal. This generalization requires two
orthonormal families, as compared to the single orthonormal family in 10.106(b).

10.113 Singular Value Decomposition

Suppose T is a compact operator on a Hilbert space V. Then there exist a


countable set Ω, orthonormal families {ek }k∈Ω and {hk }k∈Ω in V, and a family
{sk }k∈Ω of positive numbers such that

10.114 Tf = ∑ sk h f , ek i hk
k∈Ω

for every f ∈ V.

Proof If α is an eigenvalue of T ∗ T, then ( T ∗ T ) f = α f for some f 6= 0 and

α k f k2 = h α f , f i = h T ∗ T f , f i = h T f , T f i = k T f k2 .

Thus α ≥ 0. Hence all eigenvalues of T ∗ T are nonnegative.


Apply 10.106(b) and the conclusion of the paragraph above to the self-adjoint
compact operator T ∗ T, getting a countable set Ω, an orthonormal family {ek }k∈Ω in

V, and a family {sk }k∈Ω of positive numbers (take sk = αk ) such that

10.115 (T∗ T ) f = ∑ s k 2 h f , ek i ek
k∈Ω

for every f ∈ V. The equation above implies that ( T ∗ T )e j = s j 2 e j for each j ∈ Ω.


For k ∈ Ω, let
Te
hk = k .
sk
For j, k ∈ Ω, we have
1 1 sj
h h j , hk i = h Te j , Tek i = h T ∗ Te j , ek i = he j , ek i.
s j sk s j sk sk

The equation above implies that {hk }k∈Ω is an orthonormal family in V.


If f ∈ span {ek }k∈Ω , then
 
T f = T ∑ h f , ek iek = ∑ h f , ek i Tek = ∑ sk h f , ek ihk ,
k∈Ω k∈Ω k∈Ω

showing that 10.114 holds for such f .


⊥
If f ∈ span {ek }k∈Ω , then 10.115 shows that ( T ∗ T ) f = 0, which implies
that T f = 0 (because 0 = h T ∗ T f , f i = k T f k2 ); thus both sides of 10.114 are 0.
Hence the two sides of 10.114 agree for f in a closed subspace of V and for f in
the orthogonal complement of that closed subspace, which by linearity implies that
the two sides of 10.114 agree for all f ∈ V.
Section 10D Spectral Theorem for Compact Operators 333

An expression of the form 10.114 is called a singular value decomposition of


the compact operator T. The orthonormal families {ek }k∈Ω and {hk }k∈Ω in the
singular value decomposition are not uniquely determined by T. However, the
positive numbers {sk }k∈Ω are uniquely determined as positive square roots of positive
eigenvalues of T ∗ T. These positive numbers can be placed in decreasing order
(because if there are infinitely many of them, then they form a sequence with limit 0,
by 10.93). This procedure leads to the definition of singular values given below.
Suppose T is a compact operator. Recall that the geometric multiplicity of a
positive eigenvalue α of T ∗ T is defined to be dim √null( T ∗ T − αI ) [see 10.81]. This
geometric multiplicity is the number of times that α appears in the family {sk }k∈Ω
corresponding to a singular value decomposition of T. By 10.82, this geometric
multiplicity is finite.
Now we can define the singular values of a compact operator T, where we are
careful to repeat the square root of each positive eigenvalue of T ∗ T as many times as
its geometric multiplicity.

10.116 Definition singular values

• Suppose T is a compact operator on a Hilbert space. The singular values


of T, denoted s1 ( T ) ≥ s2 ( T ) ≥ s3 ( T ) ≥ · · · , are the positive square roots
of the positive eigenvalues of T ∗ T, arranged in decreasing order with each
singular value s repeated as many times as the geometric multiplicity of s2
as an eigenvalue of T ∗ T.
• If T ∗ T has only finitely many positive eigenvalues, then define sn ( T ) = 0
for all n ∈ Z+ for which sn ( T ) is not defined by the first bullet point.

10.117 Example singular values on a finite-dimensional Hilbert space


Define T : F4 → F4 by

T (z1 , z2 , z3 , z4 ) = (0, 3z1 , 2z2 , −3z4 ).

A calculation shows that

( T ∗ T )(z1 , z2 , z3 , z4 ) = (9z1 , 4z2 , 0, 9z4 ).

Thus the eigenvalues of T ∗ T are 9, 4, 0 and

dim( T ∗ T − 9I ) = 2 and dim( T ∗ T − 4I ) = 1.

Taking square roots of the positive eigenvalues of T ∗ T and then adjoining an infinite
string of 0’s shows that the singular values of T are 3 ≥ 3 ≥ 2 ≥ 0 ≥ 0 ≥ · · · .
Note that −3 and 0 are the only eigenvalues of T. Thus in this case, the list of
eigenvalues of T did not pick up the number 2 that appears in the definition (and
hence the behavior) of T, but the list of singular values of T does include 2.

If T is a compact operator, then the first singular value s1 ( T ) equals k T k, as you


are asked to verify in Exercise 12.
334 Chapter 10 Linear Maps on Hilbert Spaces

10.118 Example singular values of V − V ∗


Let V denote the Volterra operator and let T = V − V ∗. In Example 10.108, we
saw that if ek is defined by 10.112 then {ek }k∈Z is an orthonormal basis of L2 ([0, 1])
and
2
Tek = e
i (2k + 1)π k
for each k ∈ Z, where the eigenvalue shown above corresponding to ek comes from
10.111. Now 10.56 implies that
−2
T ∗ ek = e
i (2k + 1)π k
for each k ∈ Z. Hence
4
10.119 T ∗ Tek = e
(2k + 1)2 π 2 k
for each k ∈ Z. After taking positive square roots of the eigenvalues, we see that the
equation above shows that the singular values of T are
2 2 2 2 2 2
≥ ≥ ≥ ≥ ≥ ≥ ···,
π π 3π 3π 5π 5π
where the first two singular values above come from taking k = −1 and k = 0 in
10.119, the next two singular values above come from taking k = −2 and k = 1,
the next two singular values above come from taking k = −3 and k = 2, and so on.
Each singular value of T appears twice in the list of singular values above because
each eigenvalue of T ∗ T has geometric multiplicity 2.

For n ∈ Z+, the singular value sn ( T ) of a compact operator T tells us how well
we can approximate T by operators whose range has dimension less than n (see
Exercise 15).
The next result makes an important connection between K ∈ L2 (µ × µ) and the
singular values of the integral operator associated with K.

10.120 sum of squares of singular values of integral operator

Suppose µ is a σ-finite measure and K ∈ L2 (µ × µ). Then


∞ 2
kK k2L2 (µ×µ) = ∑ sn (IK ) .
n =1

Proof Consider a singular value decomposition

10.121 IK ( f ) = ∑ sk h f , ek i hk
k∈Ω

of the compact operator IK . Extend {e j } j∈Ω to an orthonormal basis {e j } j∈Γ of


L2 (µ), and extend { hk }k∈Ω to an orthonormal basis { hk }k∈Γ0 of L2 (µ).
Section 10D Spectral Theorem for Compact Operators 335

Let X denote the set on which the measure µ lives. For j ∈ Γ and k ∈ Γ0 , define
g j,k : X × X → F by
g j,k ( x, y) = e j (y)hk ( x ).
Then { g j,k } j∈Γ, k∈Γ0 is an orthonormal basis of L2 (µ × µ), as you should verify. Thus

kK k2L2 (µ×µ) = ∑ |hK, g j,k i|2


j∈Γ, k∈Γ0

Z Z 2

= ∑ K ( x, y)e j (y)hk ( x ) dµ(y) dµ( x )
j∈Γ, k∈Γ0

Z 2

= ∑ (IK e j )( x )hk ( x ) dµ( x )
j∈Γ, k∈Γ0

Z 2

10.122 = ∑ s j h j ( x )hk ( x ) dµ( x )
j∈Ω, k∈Γ0

10.123 = ∑ sj2
j∈Ω

∞ 2
= ∑ sn (IK ) ,
n =1

where 10.122 holds because 10.121 shows that IK e j = s j h j for j ∈ Ω and IK e j = 0


for j ∈ Γ \ Ω; 10.123 holds because {hk }k∈Γ0 is an orthonormal family.

Now we can give a spectacular application of the previous result.

1 1 1 π2
10.124 Example 2
+ 2 + 2 +··· =
1 3 5 8
Define K : [0, 1] × [0, 1] → R by

1
 if x > y,
K ( x, y) = 0 if x = y,


−1 if x < y.

Letting µ be Lebesgue measure on [0, 1], we note that IK is the normal compact
operator V − V ∗ examined in Example 10.118.
Clearly kK k L2 (µ×µ) = 1. Using the list of singular values for IK obtained in
Example 10.118, the formula in 10.120 tells us that

4
1=2 ∑ (2k + 1)2 π 2
.
k =0

Thus
1 1 1 π2
+ + + · · · = .
12 32 52 8
336 Chapter 10 Linear Maps on Hilbert Spaces

EXERCISES 10D
1 Prove that if T is a compact operator on a nonzero Hilbert space, then k T k2 is
an eigenvalue of T ∗ T.
2 Prove that if T is a self-adjoint operator on a nonzero Hilbert space V, then

k T k = sup{|h T f , f i| : f ∈ V and k f k = 1}.

3 Suppose T is a bounded operator on a Hilbert space V and U is a closed subspace


of V. Prove that the following are equivalent:
(a) U is an invariant subspace for T.
(b) U ⊥ is an invariant subspace for T ∗ .
(c) TPU = PU TPU .

4 Suppose T is a bounded operator on a Hilbert space V and U is a closed subspace


of V. Prove that the following are equivalent:
(a) U and U ⊥ are invariant subspaces for T.
(b) U and U ⊥ are invariant subspaces for T ∗ .
(c) TPU = PU T.

5 Suppose T is a bounded operator on a nonseparable normed vector space V.


Prove that T has a closed invariant subspace other than {0} and V.
6 Suppose T is an operator on a Banach space V with dimension greater than 2.
Prove that T has an invariant subspace other than {0} and V.
[For this exercise, T is not assumed to be bounded and the invariant subspace is
not required to be closed.]
7 Suppose T is a self-adjoint compact operator on a Hilbert space that has only
finitely many distinct eigenvalues. Prove that T has finite-dimensional range.
8 (a) Prove that if T is a self-adjoint compact operator on a Hilbert space, then
there exists a self-adjoint compact operator S such that S3 = T.
(b) Prove that if T is a normal compact operator on a complex Hilbert space,
then there exists a normal compact operator S such that S2 = T.
9 Suppose T is a compact normal operator on a nonzero Hilbert space V. Prove
that there is a subspace of V with dimension 1 or 2 that is an invariant subspace
for T.
[If F = C, the desired result follows immediately from the Spectral Theorem for
compact normal operators. Thus you can assume that F = R.]
10 Suppose T is a self-adjoint compact operator on a Hilbert space and k T k ≤ 14 .
Prove that there exists a self-adjoint compact operator S such that S2 + S = T.
Section 10D Spectral Theorem for Compact Operators 337

 
11 For k ∈ Z, define gk ∈ L2 (−π, π ] and hk ∈ L2 (−π, π ] by

gk ( t ) = √1 eit/2 eikt and hk (t) = √1 eikt ;


2π 2π

here we are assuming that F = C.

 10.108 to show that { gk }k∈Z is an ortho-


(a) Use the conclusion of Example
normal basis of L2 (−π, π ] .
 in part (a) to show that {hk }k∈Z is an orthonormal basis of
(b) Use the result
L2 (−π, π ] .
(c) Use the result in part (b) to show that the orthonormal family in thethird
bullet point of Example 8.51 is an orthonormal basis of L2 (−π, π ] .

12 Suppose T is a compact operator on a Hilbert space. Prove that s1 ( T ) = k T k.


13 Suppose T is a compact operator on a Hilbert space and n ∈ Z+. Prove that
dim range T < n if and only if sn ( T ) = 0.
14 Suppose T is a compact operator on a Hilbert space V with singular value
decomposition

Tf = ∑ sk (T )h f , ek ihk
k =1

for all f ∈ V. For n ∈ Z+, define Tn : V → V by


n
Tn f = ∑ sk (T )h f , ek ihk .
k =1

Prove that limn→∞ k T − Tn k = 0.


[This exercise gives another proof, in addition to the proof suggested by Exercise
15 in Section 10C, that an operator on a Hilbert space is compact if and only if
it is the limit of bounded operators with finite-dimensional range.]
15 Suppose T is a compact operator on a Hilbert space V and n ∈ Z+. Prove that

inf{k T − Sk : S ∈ B(V ) and dim range S < n} = sn ( T ).

16 Suppose T is a compact operator on a Hilbert space V and n ∈ Z+. Prove that

sn ( T ) = inf{k T |U ⊥ k : U is a subspace of V with dim U < n}.

17 Suppose T is a compact operator on a Hilbert space with singular value decom-


position
T f = ∑ sk h f , ek i hk
k∈Ω
for all f ∈ V. Prove that

T∗ f = ∑ sk h f , hk i ek
k∈Ω

for all f ∈ V.
338 Chapter 10 Linear Maps on Hilbert Spaces

18 Suppose that T is an operator on a finite-dimensional Hilbert space V with


dim V = n.
(a) Prove that T is invertible if and only if sn ( T ) 6= 0.
(b) Suppose T is invertible and T has a singular value decomposition

T f = s1 ( T )h f , e1 ih1 + · · · + sn ( T )h f , en ihn

for all f ∈ V. Show that

h f , h1 i h f , hn i
T −1 f = e +···+ en
s1 ( T ) 1 sn ( T )

for all f ∈ V.

19 Suppose T is a compact operator on a Hilbert space V. Prove that


∞ 2
∑ kTek k2 = ∑ sn ( T )
k∈Γ n =1

for every orthonormal basis {ek }k∈Γ of V.



1
20 Use the result of Example 10.124 to evaluate ∑ 2
.
n =1 n

21 Suppose T is a normal compact operator. Prove that the following are equivalent:
• range T is finite-dimensional.
• sp( T ) is a finite set.
• sn ( T ) = 0 for some n ∈ Z.

22 Find the singular values of the Volterra operator.


[Your answer, when combined with Exercise 12, should show that the norm of
the Volterra operator is π2 . This appearance of π can be surprising because the
definition of the Volterra operator does not involve π.]
Chapter 11
Fourier Analysis

This chapter uses Hilbert space theory to motivate the introduction of Fourier coeffi-
cients and Fourier series. The classical setting applies these concepts to functions
defined on bounded intervals of the real line. However, the theory becomes easier and
cleaner when we instead use a modern approach by considering functions defined on
the unit circle of the complex plane.
The first section of this chapter shows how consideration of Fourier series leads us
to harmonic functions and a solution to the Dirichlet problem. In the second section
of this chapter, convolution becomes a major tool for the L p theory.
The third section of this chapter changes the context to functions defined on the
real line. Many of the techniques introduced in the first two sections of the chapter
transfer easily to provide results about the Fourier transform on the real line. The
highlights of our treatment of the Fourier transform are the Fourier Inversion Formula
and the extension of the Fourier transform to a unitary operator on L2 (R).
The vast field of Fourier analysis cannot be completely covered in a single chapter.
Thus this chapter gives readers just a taste of the subject. Readers who go on from
this chapter to one of the many book-length treatments of Fourier analysis will then
already be familiar with the terminology and techniques of the subject.

The Giza pyramids, near where the Battle of Pyramids took place in 1798 during
Napoleon’s invasion of Egypt. Joseph Fourier (1768–1830) was one of the scientific
advisors to Napoleon in Egypt. While in Egypt as part of Napoleon’s invading force,
Fourier began thinking about the mathematical theory of heat propagation, which
eventually led to what we now call Fourier series and the Fourier transform.
CC-BY-SA Ricardo Liberato

339
340 Chapter 11 Fourier Analysis

11A Fourier Series and Poisson Integral


Fourier Coefficients and Riemann–Lebesgue Lemma
For k ∈ Z, suppose ek : (−π, π ] → R is defined by
 1
 √ sin( kt ) if k > 0,


 π
11.1 ek (t) = √12π if k = 0,



 √1 cos(kt) if k < 0.
π

The classical theory of Fourier series features {ek }k∈Z as an orthonormal basis of
L2 (−π, π ] . The trigonometric formulas displayed in Exercise 1 in Section 8C can

be used to show that {ek }k∈Z is indeed an orthonormal family in L2 (−π, π ] .

To show that {ek }k∈Z is an orthonormal basis of L2 (−π, π ] requires more
work. One slick possibility is to note that the Spectral Theorem for compact operators
produces orthonormal bases; an appropriate choice of a compact normal operator 
can then be used to show that {ek }k∈Z is an orthonormal basis of L2 (−π, π ] [see
Exercise 11(c) in Section 10D].
In this chapter we take a cleaner approach to Fourier series by working on the unit
circle in the complex plane instead of on the interval (−π, π ]. The map

11.2 t 7→ eit = cos t + i sin t


can be used to identify the interval (−π, π ] with the unit circle; thus the two ap-
proaches are equivalent. However, the calculations are easier in the unit circle context.
In addition, we will see that the unit circle context provides the huge benefit of
making a connection with harmonic functions.
We begin by introducing notation for the open unit disk and the unit circle in the
complex plane.

11.3 Definition D; ∂D

• D denotes the open unit disk in the complex plane:

D = { w ∈ C : | w | < 1}.

• ∂D is the unit circle in the complex plane:

∂D = {z ∈ C : |z| = 1}.

The function given in 11.2 is a one-to-one map of (−π, π ] onto ∂D. We use
this map to define a σ-algebra on ∂D by transferring the Borel subsets of (−π, π ]
to subsets of ∂D that we will call the measurable subsets of ∂D. We also transfer
Lebesgue measure on the Borel subsets of (π, π ] to a measure called σ on the
measurable subsets of ∂D, except that for convenience we normalize by dividing
by 2π so that the measure of ∂D is 1 rather than 2π. We are now ready to give the
formal definitions.
Section 11A Fourier Series and Poisson Integral 341

11.4 Definition measurable subsets of ∂D; σ

• A subset E of ∂D is measurable if {t ∈ (−π, π ] : eit ∈ E} is a Borel subset


of R.
• σ is the measure on the measurable subsets of ∂D obtained by transferring
Lebesgue measure from (−π, π ] to ∂D, normalized so that σ(∂D) = 1. In
other words, if E ⊂ ∂D is measurable, then

|{t ∈ (−π, π ] : eit ∈ E}|


σ( E) = .

Our definition of the measure σ on ∂D allows us to transfer integration on ∂D to


the familiar context of integration on (−π, π ]. Specifically,
Z Z Z π
dt
f dσ = f (z) dσ(z) = f (eit )
∂D ∂D −π 2π
for all measurable functions f : ∂D → C such that any of these integrals is defined.
Throughout this chapter, we assume that the scalar field F is the complex field C.
Furthermore, L p (∂D) is defined as follows.

11.5 Definition L p (∂D)

For 1 ≤ p ≤ ∞, define L p (∂D) to mean the complex version (F = C) of L p (σ).

Note that if z = eit for some t ∈ R, then z = e−it = 1z and zn = eint and
zn = e−int for all n ∈ Z. These observations make the proof of the next result
much simpler than the proof of the corresponding result for the trigonometric family
defined by 11.1.
In the statement of the next result, zn means the function on ∂D defined by z 7→ zn .

11.6 orthonormal family in L2 (∂D)

{zn }n∈Z is an orthonormal family in L2 (∂D).

Proof If n ∈ Z, then
Z Z
hzn , zn i = |zn |2 dσ(z) = 1 dσ = 1.
∂D ∂D
If m, n ∈ Z with m 6= n, then
Z π Z π
dt dt ei (m−n)t i t=π
hzm , zn i = eimt e−int = ei (m−n)t = = 0,
−π 2π −π 2π i (m − n)2π t=−π
as desired.

In the next section, we improve the result above by showing that {zn }n∈Z is an
orthonormal basis of L2 (∂D) (see 11.30).
342 Chapter 11 Fourier Analysis

Hilbert space theory tells us that if f is in the closure in L2 (∂D) of span{zn }n∈Z ,
then
f = ∑ h f , zn izn ,
n∈Z

where the infinite sum above converges as an unordered sum in the norm of L2 (∂D)
(see 8.58). The inner product h f , zn i above equals
Z
f (z)zn dσ(z).
∂D

Because |zn | = 1 for every z ∈ ∂D, the integral above makes sense not only for
f ∈ L2 (∂D) but also for f in the larger space L1 (∂D). Thus we make the following
definition.

11.7 Definition Fourier coefficient

Suppose f ∈ L1 (∂D).

• For n ∈ Z, the nth Fourier coefficient of f is denoted fˆ(n) and is defined by


Z Z π
dt
fˆ(n) = f (z)zn dσ(z) = f (eit )e−int .
∂D −π 2π

• The Fourier series of f is the formal sum



∑ fˆ(n)zn .
n=−∞

As we will see, Fourier analysis helps describe the sense in which the Fourier
series of f represents f .

11.8 Example Fourier coefficients

• Suppose h is an analytic function on an open set that contains D. Then h has a


power series representation

h(z) = ∑ an zn ,
n =0

where the sum on the right converges uniformly on D to h. Because uniform


convergence on ∂D implies convergence in L2 (∂D), 8.58(b) and 11.6 now imply
that (
an if n ≥ 0,
(h|∂D )ˆ(n) =
0 if n < 0
for all n ∈ Z. In other words, for functions analytic on an open set containing
D, the Fourier series is the same as the Taylor series.
Section 11A Fourier Series and Poisson Integral 343

• Suppose f : ∂D → R is defined by
1
f (z) = .
|3 − z |2
Then for z ∈ ∂D we have
1
f (z) =
(3 − z)(3 − z)
1 z 3 
= +
8 3−z 3−z
1  3z 1 
= +
8 1 − 3z 1 − 3z

1  z ∞ zn ∞
(z)n 
8 3 n∑ ∑ 3n
= n
+
=0 3 n =0

1 ∞ zn
= ∑ 3| n | ,
8 n=− ∞

where the infinite sums above converge uniformly on ∂D. Thus we see that
1 1
fˆ(n) = · |n|
8 3
for all n ∈ Z.

We begin with some simple algebraic properties of Fourier coefficients, whose


proof is left to the reader.

11.9 algebraic properties of Fourier coefficients

Suppose f , g ∈ L1 (∂D) and n ∈ Z. Then

(a) ( f + g)ˆ(n) = fˆ(n) + ĝ(n);

(b) (α f )ˆ(n) = α fˆ(n) for all α ∈ C;

(c) | fˆ(n)| ≤ k f k1 .

Parts (a) and (b) above could be restated by saying that for each n ∈ Z, the
function f 7→ fˆ(n) is a linear functional from L1 (∂D) to C. Part (c) could be
restated by saying that this linear functional has norm at most 1.
Part (c) above implies that the set of Fourier coefficients { fˆ(n)}n∈Z is bounded
for each f ∈ L1 (∂D). The Fourier coefficients of the functions in Example 11.8
have the stronger property that limn→±∞ fˆ(n) = 0. The next result shows that this
stronger conclusion holds for all functions in L1 (∂D).
344 Chapter 11 Fourier Analysis

11.10 Riemann–Lebesgue Lemma

Suppose f ∈ L1 (∂D). Then lim fˆ(n) = 0.


n→±∞

Proof Suppose ε > 0. There exists g ∈ L2 (∂D) such that k f − gk1 < ε (by 3.44).
By 11.6 and Bessel’s inequality (8.57), we have

∑ | ĝ(n)|2 ≤ k gk22 < ∞.
n=−∞

Thus there exists M ∈ Z+ such that | ĝ(n)| < ε for all n ∈ Z with |n| ≥ M. Now if
n ∈ Z and |n| ≥ M, then

| fˆ(n)| ≤ | fˆ(n) − ĝ(n)| + | ĝ(n)|

< |( f − g)ˆ(n)| + ε

≤ k f − g k1 + ε

< 2ε.

Thus lim fˆ(n) = 0.


n→±∞

Poisson Kernel
Suppose f : ∂D → C is continuous and z ∈ ∂D. For this fixed z ∈ ∂D, the Fourier
series

∑ fˆ(n)zn
n=−∞

is a series of complex numbers. It would be nice if f (z) = ∑∞ ˆ n


n=−∞ f ( n ) z , but this
is not necessarily true because the series ∑n=−∞ fˆ(n)z might not converge, as you
∞ n

can see in Exercise 11.


Various techniques exist for trying to assign some meaning to a series of complex
numbers that does not converge. In one such technique, called Abel summation, the
nth -term of the series is multiplied by r n and then the limit is taken as r ↑ 1. For
example, if the nth -term of the divergent series

1−1+1−1+···

is multiplied by r n for r ∈ [0, 1), we get a convergent series whose sum equals 1+r r .
Taking the limit of this sum as r ↑ 1 then gives 12 as the value of the Abel sum of the
series above.
The next definition can be motivated by applying a similar technique to the Fourier
series ∑∞ ˆ n
n=−∞ f ( n ) z . Here we have a series of complex numbers whose terms are
indexed by Z rather than by Z+. Thus we use r |n| rather than r n because we want
these multipliers to have limit 0 as n → ±∞ for each r ∈ [0, 1) (and to have limit 1
as r ↑ 1 for each n ∈ Z).
Section 11A Fourier Series and Poisson Integral 345

11.11 Definition Pr f

For f ∈ L1 (∂D) and 0 ≤ r < 1, define Pr f : ∂D → C by



(Pr f )(z) = ∑ r |n| fˆ(n)zn .
n=−∞

No convergence problems arise in the series above because


|n|
r fˆ(n)zn ≤ k f k1 r |n|

for each z ∈ ∂D, which implies that



1+r
∑ |r |n| fˆ(n)zn | ≤ k f k1 < ∞.
n=−∞ 1−r

Thus for each r ∈ [0, 1), the partial sums of the series above converge uniformly on
∂D, which implies that Pr f is a continuous function from ∂D to C (for r = 0 and
n = 0, interpret the expression 00 to be 1).
Let’s unravel the formula in 11.11. If f ∈ L1 (∂D), 0 ≤ r < 1, and z ∈ ∂D, then

(Pr f )(z) = ∑ r |n| fˆ(n)zn
n=−∞

∞ Z
= ∑ r |n| f (w)wn dσ(w)zn
n=−∞ ∂D

Z  ∞ 
11.12 = f (w) ∑ r |n| (zw)n dσ(w),
∂D n=−∞

where interchanging the sum and integral above is justified by the uniform conver-
gence of the series on ∂D. To evaluate the sum in parentheses in the last line above,
let ζ ∈ ∂D (think of ζ = zw in the formula above). Thus (ζ )−n = (ζ )n and
∞ ∞ ∞
∑ r |n| ζ n = ∑ (rζ )n + ∑ (rζ )n
n=−∞ n =0 n =1

1 rζ
= +
1 − rζ 1 − rζ

(1 − rζ ) + (1 − rζ )rζ
=
|1 − rζ |2

1 − r2
11.13 = .
|1 − rζ |2
Motivated by the formula above, we now make the following definition. Notice
that 11.11 uses calligraphic P , while the next definition uses italic P.
346 Chapter 11 Fourier Analysis

11.14 Definition Pr (ζ ); Poisson kernel

• For 0 ≤ r < 1, define Pr : ∂D → (0, ∞) by

1 − r2
Pr (ζ ) = .
|1 − rζ |2

• The family of functions { Pr }r∈[0,1) is called the Poisson kernel on D.

Combining 11.12 and 11.13 now gives the following result.

11.15 integral formula for Pr f

If f ∈ L1 (∂D), 0 ≤ r < 1, and z ∈ ∂D, then


Z Z
1 − r2
(Pr f )(z) = f (w) Pr (zw) dσ(w) = f (w) dσ(w).
∂D ∂D |1 − rzw|2

The terminology approximate identity is sometimes used to describe the three


properties for the Poisson kernel given in the next result.

11.16 properties of Pr

(a) Pr (ζ ) > 0 for all r ∈ [0, 1) and all ζ ∈ ∂D.


Z
(b) Pr (ζ ) dσ(ζ ) = 1 for each r ∈ [0, 1).
∂D
Z
(c) lim Pr (ζ ) dσ(ζ ) = 0 for each δ > 0.
r ↑1 {ζ ∈∂D:|1−ζ |≥δ}

Proof Part (a) follows immediately from the definition of Pr (ζ ) given in 11.14.
Part (b) follows from integrating the series representation for Pr given by 11.13
termwise and noting that
Z Z π
dt eint it=π
ζ n dσ(ζ ) = eint = = 0 for all n ∈ Z \ {0};
∂D −π 2π in2π t=−π
R
for n = 0, we have ∂D ζ n dσ(ζ ) = 1.
To prove part (c), suppose δ > 0. If ζ ∈ ∂D, |1 − ζ | ≥ δ, and 1 − r < 2δ , then
|1 − rζ | = |1 − ζ − (r − 1)ζ |
≥ |1 − ζ | − (1 − r )
> 2δ .
Thus as r ↑ 1, the denominator in the definition of Pr (ζ ) is uniformly bounded away
from 0 on {ζ ∈ ∂D : |1 − ζ | ≥ δ} and the numerator goes to 0. Thus the integral of
Pr over {ζ ∈ ∂D : |1 − ζ | ≥ δ} goes to 0 as r ↑ 1.
Section 11A Fourier Series and Poisson Integral 347

Here is the intuition behind the proof of the


next result: Parts (a) and (b) of the previous re-
sult and 11.15 mean that (Pr f )(z) is a weighted
average of f . Part (c) of the previous result says
that for r close to 1, most of the weight in this
weighted average is concentrated near z. Thus
(Pr f )(z) → f (z) as r ↑ 1.
The figure here transfers the context from
∂D to (−π, π ]. The area under both curves
is 2π [corresponding to 11.16(b)] and Pr (eit )
becomes more concentrated near t = 0 as r ↑ 1
[corresponding to 11.16(c)]. See Exercise 3 for
the formula for Pr (eit ).
One more ingredient is needed for the next
proof: If h ∈ L1 (∂D) and z ∈ ∂D, then
The graphs of P 1 (eit ) [red] and
Z Z 2
11.17 h(zw) dσ(w) = h(ζ ) dσ(ζ ). P 3 (eit ) [blue] on (−π, π ].
4
∂D ∂D
The equation above holds because the measure σ is rotation and reflection invariant.
In other words, σ({w ∈ ∂D : h(zw) ∈ E}) = σ({ζ ∈ ∂D : h(ζ ) ∈ E}) for all
measurable E ⊂ ∂D.

11.18 if f is continuous, then limk f − Pr f k∞ = 0


r ↑1

Suppose f : ∂D → C is continuous. Then Pr f converges uniformly to f on ∂D


as r ↑ 1.

Proof Suppose ε > 0. Because f is uniformly continuous on ∂D, there exists δ > 0
such that
| f (z) − f (w)| < ε for all z, w ∈ ∂D with |z − w| < δ.
If z ∈ ∂D, then
Z

| f (z) − (Pr f )(z)| = f (z) − f (w) Pr (zw) dσ(w)
∂D
Z 

= f (z) − f (w) Pr (zw) dσ(w)
∂D
Z
≤ε Pr (zw) dσ(w)
{w∈∂D : |z−w|<δ}
Z
+ 2k f k ∞ Pr (zw) dσ(w)
{w∈∂D : |z−w|≥δ}
Z
≤ ε + 2k f k ∞ Pr (ζ ) dσ(ζ ),
{ζ ∈∂D : |1−ζ |≥δ}
where we have used 11.17, 11.16(a), and 11.16(b); the last line uses the equality
|z − w| = |1 − ζ |, which holds when ζ = zw. Now 11.16(c) shows that the value
of the last integral above has uniform (with respect to z ∈ ∂D) limit 0 as r ↑ 1.
348 Chapter 11 Fourier Analysis

Solution to Dirichlet Problem on Disk


As a bonus to our investigation into Fourier series, the previous result provides the
solution to the Dirichlet problem on the unit disk. To state the Dirichlet problem, we
first need a few definitions. As usual, we identify C with R2 . Thus for x, y ∈ R, we
can think of w = x + yi ∈ C or w = ( x, y) ∈ R2 . Hence
D = {w ∈ C : |w| < 1} = {( x, y) ∈ R2 : x2 + y2 < 1}.
For a function f : G → C on an open subset G of C (or an open subset G of
R2 ), the partial derivatives D1 f and D2 f are defined as in 5.46 except that now we
allow f to be a complex-valued function. Clearly D j f = D j (Re f ) + iD j (Im f ) for
j = 1, 2.

11.19 Definition harmonic function

A function u : G → C on an open subset G of R2 is called harmonic if


 
D1 ( D1 f ) (w) + D2 ( D2 f ) (w) = 0

for all w ∈ G. The left side of the equation above is called the Laplacian of f at
w and is often denoted by (∆ f )(w).

11.20 Example harmonic functions

• If f : G → C is an analytic function on an open set G ⊂ C, then the functions


Re f , Im f , f , and f are all harmonic functions on G, as is usually discussed
near the beginning of a course on complex analysis.
• If ζ ∈ ∂D, then the function
1 − | w |2
w 7→
|1 − ζw|2
is harmonic on C \ {ζ } (see Exercise 7).
• The function u : C \ {0} → R defined by u(w) = log|w| is harmonic on
C \ {0}, as you should verify. However, there does not exist a function f
analytic on C \ {0} such that u = Re f .

The Dirichlet problem asks to extend a continuous function on the boundary of an


open subset of R2 to a function that is harmonic on the open set and continuous on
the closure of the open set. Here is a more formal statement:
Dirichlet problem on G: Suppose G ⊂ R2 is an open set and
11.21 f : ∂G → C is a continuous function. Find a continuous function
u : G → C such that u|G is harmonic and u|∂G = f .
For some open sets G ⊂ R2 , there exist continuous functions f on ∂G whose
Dirichlet problem has no solution. However, the situation on the open unit disk D is
much nicer, as we will soon see.
Section 11A Fourier Series and Poisson Integral 349

The function u defined in the result below is called the Poisson integral of f on D.

11.22 Poisson integral is harmonic

Suppose f ∈ L1 (∂D). Define u : D → C by

u(rz) = (Pr f )(z)

for r ∈ [0, 1) and z ∈ ∂D. Then u is harmonic on D.

Proof If w ∈ D, then w = rz for some r ∈ [0, 1) and some z ∈ ∂D. Thus


u(w) = (Pr f )(z)
∞ ∞
= ∑ fˆ(n)(rz)n + ∑ fˆ(−n)(rz)n
n =0 n =1

∞ ∞
= ∑ fˆ(n)wn + ∑ fˆ(−n)wn .
n =0 n =1

Every function that has a power series representation on D is analytic on D. Thus


the equation above shows that u is the sum of an analytic function and the complex
conjugate of an analytic function. Hence u is harmonic.

11.23 Poisson integral solves Dirichlet problem on unit disk

Suppose f : ∂D → C is continuous. Define u : D → C by


(
(Pr f )(z) if 0 ≤ r < 1 and z ∈ ∂D,
u(rz) =
f (z) if r = 1 and z ∈ ∂D.

Then u is continuous on D, u|D is harmonic, and u|∂D = f .

Proof Suppose ζ ∈ ∂D. To prove that u is continuous at ζ, we need to show that


if w ∈ D is close to ζ, then u(w) is close to u(ζ ). Because u|∂D = f and f is
continuous on ∂D, we do not need to worry about the case where w ∈ ∂D. Thus
assume w ∈ D. We can write w = rz, where r ∈ [0, 1) and z ∈ ∂D. Now
|u(ζ ) − u(w)| = | f (ζ ) − (Pr f )(z)|
≤ | f (ζ ) − f (z)| + | f (z) − (Pr f )(z)|.
If w is close to ζ, then z is also close to ζ, and hence by the continuity of f the first
term in the last line above is small. Also, if w is close to ζ, then r is close to 1, and
hence by 11.18 the second term in the last line above is small. Thus if w is close to ζ,
then u(w) is close to u(ζ ), as desired.
The function u|D is harmonic on D (and hence continuous on D) by 11.22.
The definition of u immediately implies that u|∂D = f .
350 Chapter 11 Fourier Analysis

Fourier Series of Smooth Functions


The Fourier series of a continuous function on ∂D need not converge pointwise (see
Exercise 11). However, in this subsection we will see that Fourier series behave well
for functions that are twice continuously differentiable.
First we need to define what we mean
The idea here is that we transfer a
for a function on ∂D to be differen-
function defined on ∂D to (−π, π ],
tiable. The formal definition is given be-
take the usual derivative there, then
low, along with the introduction of the no-
transfer back to ∂D.
tation f˜ for the transfer of f to (−π, π ]
and f [k] for the transfer back to ∂D of the kth -derivative of f˜.

11.24 Definition f˜; k times continuously differentiable; f [k]

Suppose f : ∂D → C is a complex-valued function on ∂D and k ∈ Z+ ∪ {0}.


• Define f˜ : R → C by f˜(t) = f (eit ).
• f is called k times continuously differentiable if f˜ is k times differentiable
everywhere on R and its kth -derivative f˜(k) : R → C is continuous.

• If f is k times continuously differentiable, then f [k] : ∂D → C is defined by

f [k] (eit ) = f˜(k) (t)

for t ∈ R. Here f˜(0) is defined to be f˜, which means that f [0] = f .

Note that the function f˜ defined above is periodic on R because f˜(t + 2π ) = f˜(t)
for all t ∈ R. Thus all derivatives of f˜ are also periodic on R.

11.25 Example Suppose n ∈ Z and f : ∂D → C is defined by f (z) = zn . Then


˜f : R → C is defined by f˜(t) = eint .
If k ∈ Z+, then f˜(k) (t) = ik nk eint . Thus f [k] (z) = ik nk zn for z ∈ ∂D.
Our next result gives a formula for the Fourier coefficients of a derivative.

11.26 Fourier coefficients of differentiable functions

Suppose k ∈ Z+ and f : ∂D → C is k times continuously differentiable. Then



f [k] ˆ(n) = ik nk fˆ(n)

for every n ∈ Z.

Proof First suppose n = 0. By the Fundamental Theorem of Calculus, we have


Z π Z π it=π
 dt dt
f [ k ] ˆ(0) = f [k] (eit ) = f˜(k) (t) = f˜(k−1) (t) = 0,
−π 2π −π 2π t=−π
which is the desired result for n = 0.
Section 11A Fourier Series and Poisson Integral 351

Now suppose n ∈ Z \ {0}. Then


Z
 π dt
f [ k ] ˆ( n ) = f˜(k) (t)e−int
−π 2π
it=π Z π
1 ˜(k−1) dt
= f (t)e−int + in f˜(k−1) (t)e−int
2π t=−π −π 2π

= in f [k−1] ˆ(n),

where the second equality above follows from integration by parts.


Iterating the equation above now produces the desired result.

Now we can prove the beautiful result that a twice continuously differentiable func-
tion on ∂D equals its Fourier series, with uniform convergence of the Fourier series.
This conclusion holds with the weaker hypothesis that the function is continuously
differentiable, but the proof is easier with the hypothesis used here.

11.27 Fourier series of twice continuously differentiable functions converge

Suppose f : ∂D → C is twice continuously differentiable. Then



f (z) = ∑ fˆ(n)zn
n=−∞
M
for all z ∈ ∂D. Furthermore, the partial sums ∑ fˆ(n)zn converge uniformly
n=−K
on ∂D to f as K, M → ∞.

Proof If n ∈ Z \ {0}, then



| f [2] ˆ(n)| k f [2] k 1
11.28 | fˆ(n)| = 2
≤ ,
n n2
where the equality above follows from 11.26 and the inequality above follows
from 11.9(c). Now 11.28 implies that
∞ ∞
11.29 ∑ | fˆ(n)zn | = ∑ | fˆ(n)| < ∞
n=−∞ n=−∞

for all z ∈ ∂D. The inequality above implies that ∑∞ ˆ n


n=−∞ f ( n ) z converges and that
the partial sums converge uniformly on ∂D.
Furthermore, for each ∈ ∂D we have
∞ ∞
f (z) = lim ∑ r |n| fˆ(n)zn = ∑ fˆ(n)zn ,
r ↑1 n=−∞ n=−∞

where the first equality holds by 11.18 and 11.11, and the second equality holds by
the Dominated Convergence Theorem (use counting measure on Z) and 11.29.
352 Chapter 11 Fourier Analysis

In 1923 Andrey Kolmogorov (1903–1987) published a proof that there exists


a function in L1 (∂D) whose Fourier series diverges almost everywhere on ∂D.
Kolmogorov’s result and the result in Exercise 11 probably led most mathematicians
to suspect that there exists a continuous function on ∂D whose Fourier series diverges
almost everywhere. However, in 1966 Lennart Carleson (1928–) showed that if
f ∈ L2 (∂D) (and in particular if f is continuous on ∂D), then the Fourier series of f
converges to f almost everywhere.

EXERCISES 11A
1 Prove that ( f )ˆ(n) = fˆ(−n) for all f ∈ L1 (∂D) and all n ∈ Z.
2 Suppose 1 ≤ p ≤ ∞ and n ∈ Z.
(a) Show that the function f 7→ fˆ(n) is a bounded linear functional on L p (∂D)
with norm 1.
(b) Find all f ∈ L p (∂D) such that k f k p = 1 and | fˆ(n)| = 1.

3 Show that if 0 ≤ r < 1 and t ∈ R, then


1 − r2
Pr (eit ) = .
1 − 2r cos t + r2
4 Suppose f ∈ L1 (∂D), z ∈ ∂D, and f is continuous at z. Prove that
lim(Pr f )(z) = f (z).
r ↑1

[Here L1 (∂D) means the complex version of L1 (σ). The result in this exercise
differs from 11.18 because here we are assuming continuity only at a single
point and we are not even assuming that f is bounded, as compared to 11.18,
which assumed continuity at all points of ∂D.]
5 Suppose f ∈ L1 (∂D), z ∈ ∂D, lim f (eit z) = a, and lim f (eit z) = b. Prove
t ↓0 t ↑0
that
a+b
lim(Pr f )(z) = .
r ↑1 2
[If a 6= b, then f is said to have a jump discontinuity at z.]
6 Prove that for each p ∈ [1, ∞), there exists f ∈ L1 (∂D) such that

∑ | fˆ(n)| p = ∞.
n=−∞

7 Suppose ζ ∈ ∂D. Show that the function


1 − | w |2
w 7→
|1 − ζw|2
is harmonic on C \ {ζ } by finding an analytic function on C \ {ζ } whose real
part is the function above.
Section 11A Fourier Series and Poisson Integral 353

8 Suppose f : ∂D → R is the function defined by

f ( x, y) = x4 y
for ( x, y) ∈ R2 with x2 + y2 = 1. Find a polynomial u of two variables x, y
such that u is harmonic on R2 and u|∂D = f .
[Of course, u|D is the Poisson integral of f . However, here you are asked to
find an explicit formula for u in closed form, without involving or computing
an integral. It may help to think of f as defined by f (z) = (Re z)4 (Im z) for
z ∈ ∂D.]
9 Find a formula (in closed form, not as an infinite sum) for Pr f , where f is the
function in the second bullet point of Example 11.8.
10 Suppose f : ∂D → C is three times continuously differentiable. Prove that

f [1] ( z ) = i ∑ n fˆ(n)zn
n=−∞

for all z ∈ ∂D.


11 Let C (∂D) denote the Banach space of continuous function from ∂D to C, with
the supremum norm. For M ∈ Z+, define a linear functional ϕ M : C (∂D) → C
by
M
ϕM ( f ) = ∑ fˆ(n).
n=− M

Thus ϕ M ( f ) is a partial sum of the Fourier series ∑ fˆ(n)zn , evaluated at
z = 1. n=−∞

(a) Show that


Z π
sin( M + 21 )t dt
ϕM ( f ) = f (eit )
−π sin 2t 2π
for every f ∈ C (∂D) and every M ∈ Z+.
(b) Show that
Z π 1
sin( M + 2 )t dt
lim = ∞.
M→∞ −π sin 2t 2π
(c) Show that lim M→∞ k ϕ M k = ∞. M
(d) Show that there exists f ∈ C (∂D) such that lim ∑ fˆ(n) does not
M→∞ n=− M
exist (as an element of C).
[Because the sum in part (d) is a partial sum of the Fourier series evaluated at
z = 1, part (d) shows that the Fourier series of a continuous function on ∂D
need not converge pointwise on ∂D.
The family of functions (one for each M ∈ Z+ ) on ∂D defined by
sin( M + 12 )t
eit 7→
sin 2t
is called the Dirichlet kernel.]
354 Chapter 11 Fourier Analysis

12 Define f : ∂D → R by

1
 if Im z > 0,
f (z) = −1 if Im z < 0,


0 if Im z = 0.

(a) Show that if n ∈ Z, then


(
2i
− nπ if n is odd,
fˆ(n) =
0 if n is even.

(b) Show that


2 2r Im z
(Pr f )(z) = arctan
π 1 − r2
for every r ∈ [0, 1) and every z ∈ ∂D.
(c) Verify that limr↑1 (Pr f )(z) = f (z) for every z ∈ ∂D.
(d) Prove that Pr f does not converge uniformly to f on ∂D.
Section 11B Fourier Series and L p of Unit Circle ≈ 113π

11B Fourier Series and L p of Unit Circle


The last paragraph of the previous section mentioned the result that the Fourier series
of a function in L2 (∂D) converges pointwise to the function almost everywhere. This
terrific result had been an open question until 1966. Its proof is not included in this
book, partly because the proof is difficult and partly because pointwise convergence
has turned out to be less useful than norm convergence.
Thus we begin this section with the easy proof that the Fourier series converges
in the norm of L2 (∂D). The remainder of this section then concentrates on issues
connected with norm convergence.

Orthonormal Basis for L2 of Unit Circle


We already showed that {zn }n∈Z is an orthonormal family in L2 (∂D) (see 11.6).
Now we show that {zn }n∈Z is an orthonormal basis of L2 (∂D).

11.30 orthonormal basis of L2 (∂D)

The family {zn }n∈Z is an orthonormal basis of L2 (∂D).


⊥
Proof Suppose f ∈ span{zn }n∈Z . Thus h f , zn i = 0 for all n ∈ Z. In other
words, fˆ(n) = 0 for all n ∈ Z.
Suppose ε > 0. Let g : ∂D → C be a twice continuously differentiable function
such that k f − gk2 < ε. [To prove the existence of g ∈ L2 (∂D) with this property,
first approximate f by step functions as in 3.47, but use the L2 -norm instead of the
L1 -norm. Then approximate the characteristic function of an interval as in 3.48, but
again use the L2 -norm and round the corners of the graph in the proof of 3.48 to get a
twice continuously differentiable function.]
Now
k f k2 ≤ k f − g k2 + k g k2
 1/2
= k f − g k2 + ∑ | ĝ(n)|2
n∈Z
 1/2
= k f − g k2 + ∑ |( g − f )ˆ(n)|2
n∈Z

≤ k f − g k2 + k g − f k2

< 2ε,
where the second line above follows from 11.27, the third line above holds because
fˆ(n) = 0 for all n ∈ Z, and the fourth line above follows from Bessel’s inequality
(8.57).
Because the inequality above holds for all ε > 0, we conclude that f = 0. We
⊥
have now shown that span{zn }n∈Z = {0}. Hence span{zn }n∈Z = L2 (∂D)
by 8.42, which implies that {zn }n∈Z is an orthonormal basis of L2 (∂D).
356 Chapter 11 Fourier Analysis

Now the convergence of the Fourier series of f ∈ L2 (∂D) to f follows immedi-


ately from standard Hilbert space theory [see 8.63(a)] and the previous result. Thus
with no further proof needed, we have the following important result.

11.31 convergence of Fourier series in the norm of L2 (∂D)

Suppose f ∈ L2 (∂D). Then



f = ∑ fˆ(n)zn ,
n=−∞

where the infinite sum converges to f in the norm of L2 (∂D).

The next example is a spectacular ap-


Euler’s proof, which would not be
plication of Hilbert space theory and the
considered sufficiently rigorous by
orthonormal basis {zn }n∈Z of L2 (∂D).
today’s standards, was quite
The evaluation of ∑∞ 1
n=1 n2 had been an different from the technique used in
open question until Euler discovered in
2 the example below.
1734 that this infinite sum equals π6 .

1 1 1 π2
11.32 Example + + + · · · =
12 22 32 6
R π dt
Define f ∈ L (∂D) by f (e ) = t for t ∈ (−π, π ]. Then fˆ(0) = −π t 2π
2 it = 0.
For n ∈ Z \ {0}, we have
Z π
dt
fˆ(n) = te−int
−π 2π
Z π
teint it=π 1 dt
= + e−int
−i2πn t=−π in −π 2π
(−1)n i
= ,
n
where the second line above follows from integration by parts. The equation above
implies that
∞ ∞
1
11.33 ∑ | fˆ(n)|2 = 2 ∑ 2
.
n=−∞ n =1 n
Also,
Z π
dt π2
11.34 k f k22 = t2 = .
−π 2π 3
Parseval’s identity [8.63(c)] implies that the left side of 11.33 equals the left side of
11.34. Setting the right side of 11.33 equal to the right side of 11.34 shows that

1 π2
∑ 2
=
6
.
n =1 n
Section 11B Fourier Series and L p of Unit Circle 357

Convolution on Unit Circle


Recall that Z
11.35 (Pr f )(z) = f (w) Pr (zw) dσ(w)
∂D
for f ∈ L1 (∂D), 0 ≤ r < 1, and z ∈ ∂D (see 11.15). The kind of integral formula
that appears in the result above is so useful that it gets a special name and notation.

11.36 Definition convolution; f ∗ g

Suppose f , g ∈ L1 (∂D). The convolution of f and g is denoted f ∗ g and is the


function defined by
Z
( f ∗ g)(z) = f (w) g(zw) dσ(w)
∂D

for those z ∈ ∂D for which the integral above makes sense.

Thus 11.35 states that Pr f = f ∗ Pr . Here f ∈ L1 (∂D) and Pr ∈ L∞ (∂D); hence


there is no problem with the integral in the definition of f ∗ Pr being defined for all
z ∈ ∂D. See Exercise 11 for an interpretation of convolution when the functions are
transferred to the real line.
The definition above of the convolution of two functions allows both functions to
be in L1 (∂D). The product of two functions in L1 (∂D) is not, in general, in L1 (∂D).
Thus it is not obvious that the convolution of two functions in L1 (∂D) is defined
anywhere. However, the next result shows that all is well.

11.37 convolution of two functions in L1 (∂D) is in L1 (∂D)

If f , g ∈ L1 (∂D), then ( f ∗ g)(z) is defined for almost every z ∈ ∂D. Further-


more, f ∗ g ∈ L1 (∂D) and k f ∗ gk1 ≤ k f k1 k gk1 .

Proof Suppose f , g ∈ L1 (∂D). The function (w, z) 7→ f (w) g(zw) is a measur-


able function on ∂D × ∂D, as you are asked to show in Exercise 4. Now Tonelli’s
Theorem (5.28) and 11.17 imply that
Z Z Z Z
| f (w) g(zw)| dσ(w) dσ(z) = | f (w)| | g(zw)| dσ(z) dσ(w)
∂D ∂D ∂D ∂D
Z
= | f (w)|k gk1 dσ(w)
∂D

= k f k1 k g k1 .
R
The equation above implies that ∂D | f (w) g(zw)| dσ(w) < ∞ for almost every
z ∈ ∂D. Thus ( f ∗ g)(z) is defined for almost every z ∈ ∂D.
The equation above also implies that k f ∗ gk1 ≤ k f k1 k gk1 .

Soon we will apply convolution results to Poisson integrals. However, first we


need to extend the previous result by bounding k f ∗ gk p when g ∈ L p (∂D).
358 Chapter 11 Fourier Analysis

11.38 L p -norm of a convolution

Suppose 1 ≤ p ≤ ∞, f ∈ L1 (∂D), and g ∈ L p (∂D). Then

k f ∗ g k p ≤ k f k1 k g k p .

Proof We use the following result to estimate the norm in L p (∂D):

If F : ∂D → C is measurable and 1 ≤ p ≤ ∞, then


Z
0
11.39 k F k p = sup{ | Fh| dσ : h ∈ L p (∂D) and khk p0 = 1}.
∂D

Hölder’s inequality (7.9) shows that the left side of the equation above is greater
than or equal to the right side. The inequality in the other direction almost follows
from 7.12, but 7.12 would require the hypothesis that f ∈ L p (∂D) (and we want the
equation above to hold even if k f k p = ∞). To get around this problem, apply 7.12
to truncations of F and use the Monotone Convergence Theorem (3.11); the details
of verifying 11.39 are left to the reader.
0
Suppose h ∈ L p (∂D) and khk p0 = 1. Then
Z Z Z 
|( f ∗ g)(z)h(z)| dσ(z) ≤ | f (w) g(zw)| dσ(w)|h(z)| dσ(z)
∂D ∂D ∂D
Z Z
= | f (w)| | g(zw)h(z)| dσ(z) dσ(w)
∂D ∂D
Z
≤ | f (w)|k gk p khk p0 dσ(w)
∂D

11.40 = k f k1 k g k p ,

where the second line above follows from Tonelli’s Theorem (5.28) and the third line
follows from Hölder’s inequality (7.9) and 11.17. Now 11.39 (with F = f ∗ g) and
11.40 imply that k f ∗ gk p ≤ k f k1 k gk p .

Order does not matter in convolutions, as we now prove.

11.41 convolution is commutative

Suppose f , g ∈ L1 (∂D). Then f ∗ g = g ∗ f .

Proof Suppose z ∈ ∂D is such that ( f ∗ g)(z) is defined. Then


Z Z
( f ∗ g)(z) = f (w) g(zw) dσ(w) = f (zζ ) g(ζ ) dσ(ζ ) = ( g ∗ f )(z),
∂D ∂D

where the second equality follows from making the substitution ζ = zw (which
implies that w = zζ); the invariance of the integral under this substitution is explained
in connection with 11.17.
Section 11B Fourier Series and L p of Unit Circle 359

Now we come to a major result, stating that for p ∈ [1, ∞), the Poisson integrals
of functions in L p (∂D) converge in the norm of L p (∂D). This result fails for p = ∞
[see, for example, Exercise 12(d) in Section 11A].

11.42 if f ∈ L p (∂D), then Pr f converges to f in L p (∂D)

Suppose 1 ≤ p < ∞ and f ∈ L p (∂D). Then limk f − Pr f k p = 0.


r ↑1

Proof Suppose ε > 0. Let g : ∂D → C be a continuous function on ∂D such that

k f − gk p < ε.

By 11.18, there exists R ∈ [0, 1) such that

k g − Pr g k ∞ < ε

for all r ∈ ( R, 1). If r ∈ ( R, 1), then

k f − Pr f k p ≤ k f − gk p + k g − Pr gk p + kPr g − Pr f k p
< ε + k g − Pr gk∞ + kPr ( g − f )k p
< 2ε + k Pr ∗ ( g − f )k p
≤ 2ε + k Pr k1 k g − f k p
< 3ε,

where the third line above is justified by 11.41, the fourth line above is justified by
11.38, and the last line above is justified by the equation k Pr k1 = 1, which follows
from 11.16(a) and 11.16(b). The last inequality implies that limk f − Pr f k p = 0.
r ↑1

As a consequence of the result above, we can now prove that functions in L1 (∂D),
and thus functions in L p (∂D) for every p ∈ [1, ∞], are uniquely determined by
their Fourier coefficients. Specifically, if g, h ∈ L1 (∂D) and ĝ(n) = ĥ(n) for every
n ∈ Z, then applying the result below to g − h shows that g = h.

11.43 functions are determined by their Fourier coefficients

Suppose f ∈ L1 (∂D) and fˆ(n) = 0 for every n ∈ Z. Then f = 0.

Proof Because Pr f is defined in terms of Fourier coefficients (see 11.11), we know


that Pr f = 0 for all r ∈ [0, 1). Because Pr f → f in L1 (∂D) as r ↑ 1 [by 11.42]),
this implies that f = 0.

Our next result shows that multiplication of Fourier coefficients corresponds to


convolution of the corresponding functions.
360 Chapter 11 Fourier Analysis

11.44 Fourier coefficients of a convolution

Suppose f , g ∈ L1 (∂D). Then

( f ∗ g)ˆ(n) = fˆ(n) ĝ(n)

for every n ∈ Z.

Proof First note that if w ∈ ∂D and n ∈ Z, then


Z Z
11.45 g(zw)zn dσ(z) = g(ζ )ζ n wn dσ(ζ ) = wn ĝ(n),
∂D ∂D
where the first equality comes from the substitution ζ = zw (equivalent to z = ζw),
which is justified by the rotation invariance of σ.
Now
Z
( f ∗ g )ˆ( n ) = ( f ∗ g)(z)zn dσ(z)
∂D
Z Z
= zn f (w) g(zw) dσ(w) dσ(z)
∂D ∂D
Z Z
= f (w) g(zw)zn dσ(z) dσ(w)
∂D ∂D
Z
= f (w)wn ĝ(n) dσ(w)
∂D

= fˆ(n) ĝ(n),
where the interchange of integration order in the third equality is justified by the same
steps used in the proof of 11.37 and the fourth equality above is justified by 11.45.

The next result could be proved by appropriate uses of Tonelli’s Theorem and
Fubini’s Theorem. However, the slick proof technique used in the proof below should
be useful in dealing with some of the exercises.

11.46 convolution is associative

Suppose f , g, h ∈ L1 (∂D). Then ( f ∗ g) ∗ h = f ∗ ( g ∗ h).

Proof Suppose n ∈ Z. Using 11.44 twice, we have



( f ∗ g) ∗ h ˆ(n) = ( f ∗ g)ˆ(n)ĥ(n) = fˆ(n) ĝ(n)ĥ(n).
Similarly,

f ∗ ( g ∗ h) ˆ(n) = fˆ(n)( g ∗ h)ˆ(n) = fˆ(n) ĝ(n)ĥ(n).
Hence ( f ∗ g) ∗ h and f ∗ ( g ∗ h) have the same Fourier coefficients. Because
functions in L1 (∂D) are determined by their Fourier coefficients (see 11.43), this
implies that ( f ∗ g) ∗ h = f ∗ ( g ∗ h).
Section 11B Fourier Series and L p of Unit Circle 361

EXERCISES 11B
1 Show that the family {ek }k∈Z of trigonometric
 functions defined by 11.1 is an
orthonormal basis of L2 (−π, π ] .
2 Use the result of Exercise 12(a) in Section 11A to show that
1 1 1 π2
1+ + + + · · · = .
32 52 72 8

1
3 Use techniques similar to Example 11.32 to evaluate
n 4∑.
n =1
[If you feel industrious, you may also want to evaluate ∑∞ 6
n=1 1/n . Similar
∞ k
techniques work to evaluate ∑n=1 1/n for each positive even integer k. You can
become famous if you figure out how to evaluate ∑∞ 3
n=1 1/n , which currently is
an open question.]
4 Suppose f , g : ∂D → C are measurable functions. Prove that the function
(w, z) 7→ f (w) g(zw) is a measurable function from ∂D × ∂D to C.
[Here the σ-algebra on ∂D × ∂D is the usual product σ-algebra as defined in
5.2.]
5 Where does the proof of 11.42 fail when p = ∞?
6 Suppose f ∈ L1 (∂D). Prove that f is real valued (almost everywhere) if and
only if fˆ(−n) = fˆ(n) for every n ∈ Z.

7 Suppose f ∈ L1 (∂D). Show that f ∈ L2 (∂D) if and only if ∑ | fˆ(n)|2 < ∞.
n=−∞

8 Suppose f ∈ L2 (∂D). Prove that | f (z)| = 1 for almost every z ∈ ∂D if and


only if (

1 if n = 0,
∑ fˆ(k) fˆ(k − n) =
k=−∞ 0 if n 6= 0
for all n ∈ Z.
9 For this exercise, for each r ∈ [0, 1) think of Pr as an operator on L2 (∂D).
(a) Show that Pr is a self-adjoint compact operator for each r ∈ [0, 1).
(b) For each r ∈ [0, 1), find all eigenvalues and eigenvectors of Pr .
(c) Prove or disprove: limr↑1 k I − Pr k = 0.

10 Suppose f ∈ L1 (∂D). Define T : L2 (∂D) → L2 (∂D) by Tg = f ∗ g.


(a) Show that T is a compact operator on L2 (∂D).
(b) Prove that T is injective if and only if fˆ(n) 6= 0 for every n ∈ Z.
(c) Find a formula for T ∗ .
(d) Prove: T is self-adjoint if and only if all Fourier coefficients of f are real.
(e) Show that T is a normal operator.
362 Chapter 11 Fourier Analysis

11 Show that if f , g ∈ L1 (∂D) then


Z π
( f ∗ g )˜( t ) = f˜( x ) g̃(t − x ) dx,
−π

for those t ∈ R such that ( f ∗ g)(eit ) makes sense; here ( f ∗ g)˜, f˜, and g̃
denote the transfers to the real line as defined in 11.24.
0
12 Suppose 1 ≤ p ≤ ∞. Prove that if f ∈ L p (∂D) and g ∈ L p (∂D), then f ∗ g
is a continuous function on ∂D.
13 Suppose g ∈ L1 (∂D) is such that ĝ(n) 6= 0 for infinitely many n ∈ Z. Prove
that if f ∈ L1 (∂D) and f ∗ g = g, then f = 0.
14 Show that there exists a two-sided sequence . . . , b−2 , b−1 , b0 , b1 , b2 , . . . such
that lim bn = 0 but there does not exist f ∈ L1 (∂D) with fˆ(n) = bn for all
n→±∞
n ∈ Z.
15 Prove that if f , g ∈ L2 (∂D), then

( f g )ˆ( n ) = ∑ fˆ(k ) ĝ(n − k)
k=−∞

for every n ∈ Z.
16 Suppose f ∈ L1 (∂D). Prove that Pr (Ps f ) = Prs f for all r, s ∈ [0, 1).
17 Suppose p ∈ [1, ∞] and f ∈ L p (∂D). Prove that if 0 ≤ r < s < 1, then

kPr f k p ≤ kPs f k p .

18 R π If f : R → R is a continuously differentiable
Prove Wirtinger’s inequality:
2π-periodic function and −π f (t) dt = 0, then
Z π 2 Z π 2
f (t) dt ≤ f 0 (t) dt,
−π −π

with equality if and only if f (t) = a sin(t) + b cos(t) for some constants a, b.
Section 11C Fourier Transform 363

11C Fourier Transform


Fourier Transform on L1 (R)
We now switch from consideration of functions defined on the unit circle ∂D to
consideration of functions defined on the real line R. Instead of dealing with Fourier
coefficients andRFourier series, we now
R deal with Fourier transforms.

Recall that −∞ f ( x ) dx means R f dλ, where λ denotes Lebesgue measure
on R, and similarly if a dummy variable other than x is used (see 3.39). Similarly,
L p (R) means L p (λ) (the version that allows the functions to be complex valued).
R∞ 1/p
Thus in this section, k f k p = −∞ | f ( x )| p dx for 1 ≤ p < ∞.

11.47 Definition Fourier transform

For f ∈ L1 (R), the Fourier transform of f is the function fˆ : R → C defined by


Z ∞
fˆ(t) = f ( x )e−2πitx dx.
−∞

We use the same notation fˆ for the Fourier transform as we did for Fourier
coefficients. The analogies that we will see between the two concepts makes using
the same notation reasonable. The context should make it clear whether this notation
refers to Fourier transforms (when we are working with functions defined on R)
or whether the notation refers to Fourier coefficients (when we are working with
functions defined on ∂D).
The factor 2π that appears in the exponent in the definition above of the Fourier
transform is a normalization factor. Without this normalization, we would lose the
beautiful result that k fˆk2 = k f k2 (see 11.82). Another possible normalization,
which is used by some books, is to define the Fourier transform of f at t to be
Z ∞
dx
f ( x )e−itx √ .
−∞ 2π
There is no right or wrong way to do the normalization—pesky π’s will pop up
somewhere regardless of the normalization or lack of normalization. However, the
choice made in 11.47 seems to cause fewer problems than other choices.

11.48 Example Fourier transforms

(a) Suppose b ≤ c. If t ∈ R, then


Z c
(χ[b, c])ˆ(t) = e−2πitx dx
b
 −2πict 
i e
 − e−2πibt
if t 6= 0,
= 2πt


c−b if t = 0.
364 Chapter 11 Fourier Analysis

(b) Suppose f ( x ) = e−2π | x| for x ∈ R. If t ∈ R, then


Z ∞
fˆ(t) = e−2π | x| e−2πitx dx
−∞
Z 0 Z ∞
= e2πx e−2πitx dx + e−2πx e−2πitx dx
−∞ 0

1 1
= +
2π (1 − it) 2π (1 + it)
1
= .
π ( t2 + 1)
Recall that the Riemann–Lebesgue Lemma on the unit circle ∂D states that if
f ∈ L1 (∂D), then limn→±∞ fˆ(n) = 0 (see 11.10). Now we come to the analogous
result in the context of the real line.

11.49 Riemann–Lebesgue Lemma

Suppose f ∈ L1 (R). Then fˆ is uniformly continuous on R. Furthermore,

k fˆk∞ ≤ k f k1 and lim fˆ(t) = 0.


t→±∞

Proof Because |e−2πitx | = 1 for all t ∈ R and all x ∈ R, the definition of the
Fourier transform implies that if t ∈ R then
Z ∞
| fˆ(t)| ≤ | f ( x )| dx = k f k1 .
−∞

Thus k fˆk∞ ≤ k f k1 .
If f is the characteristic function of a bounded interval, then the formula in
Example 11.48(a) shows that fˆ is uniformly continuous on R and limt→±∞ fˆ(t) = 0.
Thus the same result holds for finite linear combinations of such functions. Such
finite linear combinations are called step functions (see 3.46).
Now consider arbitrary f ∈ L1 (R). There exists a sequence f 1 , f 2 , . . . of step
functions in L1 (R) such that limk→∞ k f − f k k1 = 0 (by 3.47). Thus

lim k fˆ − fˆk k∞ = 0.
k→∞

In other words, the sequence fˆ1 , fˆ2 , . . . converges uniformly on R to fˆ. Because the
uniform limit of uniformly continuous functions is uniformly continuous, we can
conclude that fˆ is uniformly continuous on R. Furthermore, the uniform limit of
functions on R each of which has limit 0 at ±∞ also has limit 0 at ±∞, completing
the proof.

The next result gives a condition that forces the Fourier transform of a function to
be continuously differentiable. This result also gives a formula for the derivative of
the Fourier transform. See Exercise 8 for a formula for the nth derivative.
Section 11C Fourier Transform 365

11.50 derivative of a Fourier transform

Suppose f ∈ L1 (R). Define g : R → C by g( x ) = x f ( x ). If g ∈ L1 (R), then


fˆ is a continuously differentiable function on R and

( fˆ)0 (t) = −2πi ĝ(t)

for all t ∈ R.

Proof Fix t ∈ R. Then


Z ∞  e−2πisx − 1 
fˆ(t + s) − fˆ(t)
lim = lim f ( x )e−2πitx dx
s →0 s s →0 − ∞ s
Z ∞  e−2πisx − 1 
= f ( x )e−2πitx lim dx
−∞ s →0 s
Z ∞
= −2πi x f ( x )e−2πitx dx
−∞

= −2πi ĝ(t),

where the second equality is justified by using the inequality |eiθ − 1| ≤ θ (valid
for all θ ∈ R, as the reader should verify) to show that |(e−2πisx − 1)/s| ≤ 2π | x |
for all s ∈ R \ {0} and all x ∈ R; the hypothesis that x f ( x ) ∈ L1 (R) and the
Dominated Convergence Theorem (3.31) then allow for the interchange of the limit
and the integral that is used in the second equality above.
The equation above shows that fˆ is differentiable and that ( fˆ)0 (t) = −2πi ĝ(t)
for all t ∈ R. Because ĝ is continuous on R (by 11.49), we can also conclude that fˆ
is continuously differentiable.

2
11.51 Example e−πx equals its Fourier transform
2
Suppose f ∈ L1 (R) is defined by f ( x ) = e−πx . Then the function g : R → C
2
defined by g( x ) = x f ( x ) = xe−πx is in L1 (R). Hence 11.50 implies that if t ∈ R
then
Z ∞
2
( fˆ)0 (t) = −2πi xe−πx e−2πitx dx
−∞
 i x =∞ Z ∞
2 2
= ie−πx e−2πitx − 2πt e−πx e−2πitx dx
x =−∞ −∞

11.52 = −2πt fˆ(t),

where the second equality follows from integration by parts (if you are nervous about
doing an integration by parts from −∞ to ∞, change each integral to be the limit as
M → ∞ of the integral from − M to M).
366 Chapter 11 Fourier Analysis

2
Note that f 0 (t) = −2πte−πt = −2πt f (t). Combining this equation with 11.52
shows that
 fˆ 0 f (t)( fˆ)0 (t) − f 0 (t) fˆ(t) f (t) fˆ(t) − f (t) fˆ(t)
(t) = 2 = −2πt 2 =0
f f (t) f (t)

for all t ∈ R. Thus fˆ/ f is a constant function. In other words, there exists c ∈ C
such that fˆ = c f . To evaluate c, note that
Z ∞
2
11.53 fˆ(0) = e−πx dx = 1 = f (0),
−∞

where the integral above is evaluated by writing its square as the integral times the
same integral but using y instead of x for the dummy variable and then converting to
polar coordinates (dx dy = r dr dθ).
Clearly 11.53 implies that c = 1. Thus fˆ = f .

The next result gives a formula for the Fourier transform of a derivative. See
Exercise 9 for a formula for the Fourier transform of the nth derivative.

11.54 Fourier transform of a derivative

Suppose f ∈ L1 (R) is a continuously differentiable function and f 0 ∈ L1 (R).


If t ∈ R, then
( f 0 )ˆ(t) = 2πit fˆ(t).

Proof Suppose ε > 0. Because f and f 0 are in L1 (R), there exists a ∈ R such that
Z ∞
| f 0 ( x )| dx < ε and | f ( a)| < ε.
a

Now if b > a then


Z b Z ∞

| f (b)| = f 0 ( x ) dx + f ( a) ≤ | f 0 ( x )| dx + | f ( a)| < 2ε.
a a

Hence limx→∞ f ( x ) = 0. Similarly, limx→−∞ f ( x ) = 0.


If t ∈ R, then
Z ∞
( f 0 )ˆ( t ) = f 0 ( x )e−2πitx dx
−∞
i x =∞ Z ∞
= f ( x )e−2πitx + 2πit f ( x )e−2πitx dx
x =−∞ −∞

= 2πit fˆ(t),
where the second equality comes from integration by parts and the third equality
holds because we showed in the paragraph above that limx→±∞ f ( x ) = 0.

The next result gives formulas for the Fourier transforms of some algebraic
transformations of a function. Proofs of these formulas are left to the reader.
Section 11C Fourier Transform 367

11.55 Fourier transforms of translations, rotations, and dilations

Suppose f ∈ L1 (R), b ∈ R, and t ∈ R.

(a) If g( x ) = f ( x − b) for all x ∈ R, then ĝ(t) = e−2πibt fˆ(t).

(b) If g( x ) = e2πibx f ( x ) for all x ∈ R, then ĝ(t) = fˆ(t − b).



(c) If b 6= 0 and g( x ) = f (bx ) for all x ∈ R, then ĝ(t) = 1
|b|
fˆ t
b .

11.56 Example Fourier transform of a rotation of an exponential function


Suppose y > 0, x ∈ R, and h(t) = e−2πy|t| e2πixt . To find the Fourier transform
of h, first consider the function g defined by g(t) = e−2πy|t| . By 11.48(b) and
11.55(c), we have
1 1 1 y
11.57 ĝ(t) =    = .
y π t +1 2 π t + y2
2
y
Now 11.55(b) implies that
1 y
11.58 ĥ(t) = ;
π ( t − x )2 + y2
note that x is a constant in the definition of h, which has t as the variable, but x is the
variable in 11.55(b)—this slightly awkward permutation of variables is done in this
example to make a later reference to 11.58 come out cleaner.

The next result will be immensely useful later in this section.

11.59 integral of a function times a Fourier transform

Suppose f , g ∈ L1 (R). Then


Z ∞ Z ∞
fˆ(t) g(t) dt = f (t) ĝ(t) dt.
−∞ −∞

Proof Both integrals in the equation above make sense because f , g ∈ L1 (R) and
fˆ, ĝ ∈ L∞ (R) (by 11.49). Using the definition of the Fourier transform, we have
Z ∞ Z ∞ Z ∞
fˆ(t) g(t) dt = g(t) f ( x )e−2πitx dx dt
−∞ −∞ −∞
Z ∞ Z ∞
= f (x) g(t)e−2πitx dt dx
−∞ −∞
Z ∞
= f ( x ) ĝ( x ) dx,
−∞
where Tonelli’s Theorem and Fubini’s Theorem justify the second equality. Changing
the dummy variable x to t in the last expression gives the desired result.
368 Chapter 11 Fourier Analysis

Convolution on R
Our next big goal is to prove the Fourier Inversion Formula. This remarkable formula,
discovered by Fourier, states that if f ∈ L1 (R) and fˆ ∈ L1 (R), then
Z ∞
11.60 f (x) = fˆ(t)e2πixt dt
−∞

for almost every x ∈ R. We will eventually prove this result (see 11.76), but first we
need to develop some tools that will be used in the proof. To motivate these tools, we
look at the right side of the equation above for fixed x ∈ R and see what we would
need to prove that it equals f ( x ).
To get from the right side of 11.60 to an expression involving f rather than fˆ, we
should be tempted to use 11.59. However, we cannot use 11.59 because the function
t 7→ e2πixt is not in L1 (R), which is a hypothesis needed for 11.59. Thus we throw
in a convenient convergence factor, fixing y > 0 and considering the integral
Z ∞
11.61 fˆ(t)e−2πy|t| e2πixt dt.
−∞

The convergence factor above is a good choice because for fixed y > 0 the function
t 7→ e−2πy|t| is in L1 (R), and limy↓0 e−2πy|t| = 1 for every t ∈ R (which means
that 11.61 may be a good approximation to 11.60 for y close to 0).
Now let’s be rigorous. Suppose f ∈ L1 (R). Fix y > 0 and x ∈ R. Define
h : R → C by h(t) = e−2πy|t| e2πixt . Then h ∈ L1 (R) and
Z ∞ Z ∞
fˆ(t)e−2πy|t| e2πixt dt = fˆ(t)h(t) dt
−∞ −∞
Z ∞
= f (t)ĥ(t) dt
−∞
Z ∞
1 y
11.62 = f (t) dt,
π −∞ ( x − t )2 + y2
where the second equality comes from 11.59 and the third equality comes from 11.58.
We will come back to the specific formula in 11.62
R ∞later, but for now we use 11.62 as
motivation for study of expressions of the form −∞ f (t) g( x − t) dt. Thus we have
been led to the following definition.

11.63 Definition convolution; f ∗ g

Suppose f , g : R → C are measurable functions. The convolution of f and g is


denoted f ∗ g and is the function defined by
Z ∞
( f ∗ g)( x ) = f (t) g( x − t) dt
−∞

for those x ∈ R for which the integral above makes sense.


Section 11C Fourier Transform 369

Here we are using the same terminology and notation as was used for the convolu-
tion of functions on the unit circle. Recall that if F, G ∈ L1 (∂D), then
Z π
ds
( F ∗ G )(eiθ ) = F (eis ) G (ei(θ −s) )
−π 2π
for θ ∈ R (see 11.36). The context should always indicate whether f ∗ g denotes
convolution on the unit circle or convolution on the real line. The formal similarities
between the two notions of convolution make many of the proofs transfer in either
direction from one context to the other.
If f , g ∈ L1 (R), then f ∗ g is defined
If 1 ≤ p ≤ ∞, f ∈ L p (R), and
for almost every x ∈ R, and furthermore 0
k f ∗ gk1 ≤ k f k1 k gk1 (as you should g ∈ L p (R), then Hölder’s
verify by translating the proof of 11.37 to inequality (7.9) and the translation
the context of R). invariance of Lebesgue measure
If p ∈ (1, ∞], then neither L1 (R) nor imply ( f ∗ g)( x ) is defined for all
L p (R) is a subset of the other [unlike the x ∈ R and k f ∗ gk∞ ≤ k f k p k gk p0
inclusion L p (∂D) ⊂ L1 (∂D)]. Thus we (more is true; with these hypothesis,
do not yet know that f ∗ g makes sense f ∗ g is a uniformly continuous
for f ∈ L1 (R) and g ∈ L p (R). However, function on R, as you are asked to
the next result shows that all is well. show in Exercise 10).

11.64 L p -norm of a convolution

Suppose 1 ≤ p ≤ ∞, f ∈ L1 (R), and g ∈ L p (R). Then ( f ∗ g)( x ) is defined


for almost every x ∈ R. Furthermore,

k f ∗ g k p ≤ k f k1 k g k p .

Proof First consider the case where f ( x ) ≥ 0 and g( x ) ≥ 0 for almost every
x ∈ R. Thus ( f ∗ g)( x ) is defined for each x ∈ R, although its value might equal ∞.
Apply the proof of 11.38 to the context of R, concluding that k f ∗ gk p ≤ k f k1 k gk p
[which implies that ( f ∗ g)( x ) < ∞ for almost every x ∈ R].
Now consider arbitrary f ∈ L1 (R), and g ∈ L p (R). Apply the case of the
previous paragraph to | f | and | g| to get the desired conclusions.

The next proof, as is the case for several other proofs in this section, asks the
reader to transfer the proof of the analogous result from the context of the unit circle
to the context of the real line. This should require only minor adjustments of a proof
from one of the two previous sections. The best way to learn this material is to write
out for yourself the required proof in the context of the real line.

11.65 convolution is commutative

Suppose f , g : R → C are measurable functions and x ∈ R is such that


( f ∗ g)( x ) is defined. Then ( f ∗ g)( x ) = ( g ∗ f )( x ).

Proof Adjust the proof of 11.41 to the context of R.


370 Chapter 11 Fourier Analysis

Our next result shows that multiplication of Fourier transforms corresponds to


convolution of the corresponding functions.

11.66 Fourier transform of a convolution

Suppose f , g ∈ L1 (R). Then

( f ∗ g)ˆ(t) = fˆ(t) ĝ(t)

for every t ∈ R.

Proof Adjust the proof of 11.44 to the context of R.

Poisson Kernel on Upper Half-Plane


As usual, we identify R2 with C, as illustrated in the following definition. We will
see that the upper half-plane plays a role in the context of R similar to the role that
the open unit disk plays in the context of ∂D.

11.67 Definition H; upper half-plane

• H denotes the open upper half-plane in R2 :

H = {( x, y) ∈ R2 : y > 0} = {z ∈ C : Im z > 0}.

• ∂H is identified with the real line:

∂H = {( x, y) ∈ R2 : y = 0} = {z ∈ C : Im z = 0} = R.

Recall that we defined a family of functions on ∂D called the Poisson kernel on D


(see 11.14, where the family is called the Poisson kernel on D because 0 ≤ r < 1 and
ζ ∈ ∂D implies rζ ∈ D). Now we are ready to define a family of functions on R that
is called the Poisson kernel on H [because x ∈ R and y > 0 implies ( x, y) ∈ H].
The following definition is motivated by 11.62. The notation Pr for the Poisson
kernel on D and Py for the Poisson kernel on H is potentially ambiguous (what is
P1/2 ?), but the intended meaning should always be clear from the context.

11.68 Definition Py ; Poisson kernel

• For y > 0, define Py : R → (0, ∞) by

1 y
Py ( x ) = .
π x + y2
2

• The family of functions { Py }y>0 is called the Poisson kernel on H.


Section 11C Fourier Transform 371

The properties of the Poisson kernel on H listed in the result below should be
compared to the corresponding properties (see 11.16) of the Poisson kernel on D.

11.69 properties of Py

(a) Py ( x ) > 0 for all y > 0 and all x ∈ R.


Z ∞
(b) Py ( x ) dx = 1 for each y > 0.
−∞
Z
(c) lim Py ( x ) dx = 0 for each δ > 0.
y ↓0 { x ∈R:| x |≥δ}

Proof Part (a) follows immediately from the definition of Py ( x ) given in 11.68.
Parts (b) and (c) follow from explicitly evaluating the integrals, using the result
that for each y > 0, an anti-derivative of Py ( x ) (as a function of x) is π1 arctan yx .

If p ∈ [1, ∞] and f ∈ L p (R) and y > 0, then f ∗ Py makes sense because


0
Py ∈ L p (R). Thus the following definition makes sense.

11.70 Definition Py f

For f ∈ L p (R) for some p ∈ [1, ∞] and for y > 0, define Py f : R → C by


Z ∞ Z ∞
1 y
(Py f )( x ) = f (t) Py ( x − t) dt = f (t) dt
−∞ π −∞ ( x − t )2 + y2

for x ∈ R. In other words, Py f = f ∗ Py .

The next result is analogous to 11.18,


When Napoleon appointed Fourier
except that now we need to include in the
to an administrative position in
hypothesis that our function is uniformly
1806, Siméon Poisson (1781–1840)
continuous and bounded (those conditions
was appointed to the professor
follow automatically from continuity in
position at École Polytechnique
the context of the unit circle).
vacated by Fourier. Poisson
For the proof of the result below, you
published over 300 mathematical
should use the properties in 11.69 instead
papers during his lifetime.
of the corresponding properties in 11.16.

11.71 if f is uniformly continuous and bounded, then limk f − Py f k∞ = 0


y ↓0

Suppose f : R → C is uniformly continuous and bounded. Then Py f converges


uniformly to f on R as y ↓ 0.

Proof Adjust the proof of 11.18 to the context of R.


372 Chapter 11 Fourier Analysis

The function u defined in the result below is called the Poisson integral of f on H.

11.72 Poisson integral is harmonic

Suppose f ∈ L p (R) for some p ∈ [1, ∞]. Define u : H → C by

u( x, y) = (Py f )( x )

for x ∈ R and y > 0. Then u is harmonic on H.

Proof First we consider the case where f is real valued. For x ∈ R and y > 0, let
z = x + iy. Then
y 1
= − Im
( x − t )2 + y2 z−t
for t ∈ R. Thus Z
1 ∞ 1
u( x, y) = − Im f (t) dt.
π −∞ z−t
R∞ 1
The function z 7→ − −∞ f (t) z− t dt is analytic on H; its derivative is the function
R∞ 1
z 7→ −∞ f (t) (z−t)2 dt (justification for this statement is in the next paragraph).
In other words, we can differentiate (with respect to z) under the integral sign in
the expression above. Because u is the imaginary part of an analytic function, u is
harmonic on H, as desired.
To justify the differentiation under
R ∞ the integral sign, fix z ∈ H and define a
1
function g : H → C by g(z) = − −∞ f (t) z− t dt. Then
Z ∞ Z ∞
g(z) − g(w) 1 z−w
− f (t) dt = f (t) dt.
z−w −∞ ( z − t )2 −∞ ( z − t )2 ( w − t )
z−w 0
As w → z, the function t 7→ ( z − t )2 ( w − t )
goes to 0 in the norm of L p (R). Thus
Hölder’s Rinequality (7.9) and the equation above imply that g0 (z) exists and that

g0 (z) = −∞ f (t) (z−1t)2 dt, as desired.

We have now solved the Dirichlet problem on the half-space for uniformly contin-
uous, bounded functions on R (see 11.21 for the statement of the Dirichlet problem).

11.73 Poisson integral solves Dirichlet problem on half-plane

Suppose f : R → C is uniformly continuous and bounded. Define u : H → C


by (
(Py f )( x ) if x ∈ R and y > 0,
u( x, y) =
f (x) if x ∈ R and y = 0.

Then u is continuous on H, u|H is harmonic, and u|R = f .

Proof Adjust the proof of 11.23 to the context of R; now you will need to use 11.71
and 11.72 instead of the corresponding results for the unit circle.
Section 11C Fourier Transform 373

The next result, which states that the


Poisson and Fourier are two of the
Poisson integrals of functions in L p (R)
72 mathematicians/scientists whose
converge in the norm of L p (R), will be
names are prominently inscribed on
a major tool in proving the Fourier Inver-
the Eiffel Tower in Paris.
sion Formula and other results later in this
section.
For the result below, the proof of the corresponding result on the unit circle (11.42)
does not transfer to the context of R (because the inequality k·k p ≤ k·k∞ fails in the
context of R).

11.74 if f ∈ L p (R), then Py f converges to f in L p (R)

Suppose 1 ≤ p < ∞ and f ∈ L p (R). Then limk f − Py f k p = 0.


y ↓0

Proof If y > 0 and x ∈ R, then


Z ∞

| f ( x ) − (Py f )( x )| = f ( x ) − f ( x − t) Py (t) dt
−∞
Z ∞ 

= f ( x ) − f ( x − t) Py (t) dt
−∞
Z ∞ 1/p
11.75 ≤ f ( x ) − f ( x − t) p Py (t) dt ,
−∞
where the inequality comes from applying 7.10 to the measure Py dt (note that the
measure of R with respect to this measure is 1).
Define h : R → [0, ∞) by
Z ∞
h(t) = f ( x ) − f ( x − t) p dx.
−∞

Then h is a bounded function that is uniformly continuous on R [by Exercise 23(a) in


Section 7A]. Furthermore, h(0) = 0.
Raising both sides of 11.75 to the pth power and then integrating over R with
respect to x, we have
Z ∞ Z ∞
p
k f − Py f k p ≤ f ( x ) − f ( x − t) p Py (t) dt dx
−∞ −∞
Z ∞ Z ∞
= Py (t) f ( x ) − f ( x − t) p dx dt
−∞ −∞
Z ∞
= Py (−t)h(t) dt
−∞

= (Py h)(0).
Now 11.71 implies that lim(Py h)(0) = h(0) = 0. Hence the last inequality above
y ↓0
implies that limk f − Py f k p = 0.
y ↓0
374 Chapter 11 Fourier Analysis

Fourier Inversion Formula


Now we can prove the remarkable Fourier Inversion Formula.

11.76 Fourier Inversion Formula

Suppose f ∈ L1 (R) and fˆ ∈ L1 (R). Then


Z ∞
f (x) = fˆ(t)e2πixt dt
−∞

for almost every x ∈ R. In other words,

f ( x ) = ( fˆ)ˆ(− x )

for almost every x ∈ R.

Proof Equation 11.62 states that


Z ∞
11.77 fˆ(t)e−2πy|t| e2πixt dt = (Py f )( x )
−∞

for every x ∈ R and every y > 0.


Because fˆ ∈ L1 (R), the Dominated Convergence Theorem (3.31) implies that for
every x ∈ R, the left side of 11.77 has limit ( fˆ)ˆ(− x ) as y ↓ 0.
Because f ∈ L1 (R), 11.74 implies that limy↓0 k f − Py f k1 = 0. Now 7.23 im-
plies that there is a sequence of positive numbers y1 , y2 , . . . such that limn→∞ yn = 0
and limn→∞ (Pyn f )( x ) = f ( x ) for almost every x ∈ R.
Combining the results in the two previous paragraphs and equation 11.77 shows
that f ( x ) = ( fˆ)ˆ(− x ) for almost every x ∈ R.

The Fourier transform of a function in L1 (R) is a uniformly continuous function on


R (by 11.49). Thus the Fourier Inversion Formula (11.76) implies that if f ∈ L1 (R)
and fˆ ∈ L1 (R), then f can be modified on a set of measure zero to become a
uniformly continuous function on R.
The Fourier Inversion Formula now allows us to calculate the Fourier transform
of Py for each y > 0.

11.78 Example Fourier transform of Py


Suppose y > 0. Define f : R → (0, 1] by

f (t) = e−2πy|t| .

Then fˆ = Py by 11.57. Hence both f and fˆ are in L1 (R). Thus we can apply the
Fourier Inversion Formula (11.76), concluding that

11.79 ( Py )ˆ( x ) = ( fˆ)ˆ( x ) = f (− x ) = e−2πy| x|


for all x ∈ R.
Section 11C Fourier Transform 375

Now we can prove that the map on L1 (R) defined by f 7→ fˆ is one-to-one.

11.80 functions are determined by their Fourier transforms

Suppose f ∈ L1 (R) and fˆ(t) = 0 for every t ∈ R. Then f = 0.

Proof Because fˆ = 0, we also have ( fˆ)ˆ = 0. The Fourier Inversion Formula


(11.76) now implies that f = 0.

The next result could be proved directly using the definition of convolution and
Tonelli’s/Fubini’s Theorems. However, the following cute proof deserves to be seen.

11.81 convolution is associative

Suppose f , g, h ∈ L1 (R). Then ( f ∗ g) ∗ h = f ∗ ( g ∗ h).

Proof The Fourier transform of ( f ∗ g) ∗ h and the Fourier transform of f ∗ ( g ∗ h)


both equal fˆĝĥ (by 11.66). Because the Fourier transform is a one-to-one mapping
on L1 (R) [see 11.80], this implies that ( f ∗ g) ∗ h = f ∗ ( g ∗ h).

Extending Fourier Transform to L2 (R)


We now prove that the map f 7→ fˆ preserves L2 (R) norms on L1 (R) ∩ L2 (R).

11.82 Plancherel’s Theorem: Fourier transform preserves L2 (R) norms

Suppose f ∈ L1 (R) ∩ L2 (R). Then k fˆk2 = k f k2 .

Proof First consider the case where fˆ ∈ L1 (R) in addition to the hypothesis that
f ∈ L1 (R) ∩ L2 (R). Define g : R → C by g( x ) = f (− x ). Then ĝ(t) = fˆ(t) for
all t ∈ R, as is easy to verify. Now
Z ∞
k f k22 = f ( x ) f ( x ) dx
−∞
Z ∞
= f (− x ) f (− x ) dx
−∞
Z ∞
11.83 = ( fˆ)ˆ( x ) g( x ) dx
−∞
Z ∞
11.84 = fˆ( x ) ĝ( x ) dx
−∞
Z ∞
= fˆ( x ) fˆ( x ) dx
−∞

= k fˆk22 ,
376 Chapter 11 Fourier Analysis

where 11.83 holds by the Fourier Inversion Formula (11.76) and 11.84 follows from
11.59. The equation above shows that our desired result holds in the case when
fˆ ∈ L1 (R).
Now consider arbitrary f ∈ L1 (R) ∩ L2 (R). If y > 0, then f ∗ Py ∈ L1 (R) by
11.64. If x ∈ R, then

( f ∗ Py )ˆ( x ) = fˆ( x )( Py )ˆ( x )

11.85 = fˆ( x )e−2πy| x| ,

where the first equality above comes from 11.66 and the second equality comes from
11.79. The equation above shows that ( f ∗ Py )ˆ ∈ L1 (R). Thus we can apply the
first case to f ∗ Py , concluding that

k f ∗ Py k2 = k( f ∗ Py )ˆk2 .

As y ↓ 0, the left side of the equation above converges to k f k2 [by 11.74]. As y ↓ 0,


the right side of the equation above converges to k fˆk2 [by the explicit formula for
f ∗ Py given in 11.85 and the Monotone Convergence Theorem (3.11)]. Thus the
equation above implies that k fˆk2 = k f k2 .

Because L1 (R) ∩ L2 (R) is dense in L2 (R), Plancherel’s Theorem (11.82) allows


us to extend the map f 7→ fˆ uniquely to a bounded linear map from L2 (R) to L2 (R)
(see Exercise 14 in Section 6C). This extension is called the Fourier transform on
L2 (R); it gets its own notation, as shown below.

11.86 Definition Fourier transform on L2 (R)

The Fourier transform F on L2 (R) is the bounded operator on L2 (R) such that
F f = fˆ for all f ∈ L1 (R) ∩ L2 (R).

For f ∈ L1 (R) ∩ L2 (R), we can use either fˆ or F f to denote the Fourier


transform of f . But if f ∈ L1 (R) \ L2 (R), we will use only the notation fˆ, and if
f ∈ L2 (R) \ L1 (R), we will use only the notation F f .
Suppose f ∈ L2 (R) \ L1 (R) and t ∈ R. Do not make the mistake of thinking
that (F f )(t) equals Z ∞
f ( x )e−2πitx dx.
−∞

Indeed, the integral above makes no sense because f ( x )e−2πitx = | f ( x )| and
/ L1 (R). Instead of defining F f via the equation above, F f must be defined as the
f ∈
limit in L2 (R) of ( f 1 )ˆ, ( f 2 )ˆ, . . ., where f 1 , f 2 , . . . is a sequence in L1 (R) ∩ L2 (R)
such that
k f − f n k2 → 0 as n → ∞.
For example, one could take f n = f χ[−n, n] because k f − f χ[−n, n]k2 → 0 as n → ∞
by the Dominated Convergence Theorem (3.31).
Section 11C Fourier Transform 377

Because F is obtained by continuously extending [in the norm of L2 (R)] the


Fourier transform from L1 (R) ∩ L2 (R) to L2 (R), we know that kF f k2 = k f k2 for
all f ∈ L2 (R). In other words, F is an isometry on L2 (R). The next result shows
that even more is true.

11.87 properties of the Fourier transform on L2 (R)

(a) F is a unitary operator on L2 (R).


(b) F 4 = I.
(c) sp(F ) = {1, i, −1, −i }.

Proof First we prove (b). Suppose f ∈ L1 (R) ∩ L2 (R). If y > 0, then Py ∈ L1 (R)
and hence 11.64 implies that

11.88 f ∗ Py ∈ L1 (R) ∩ L2 (R).

Also,

11.89 ( f ∗ Py )ˆ ∈ L1 (R) ∩ L2 (R),

as follows from the equation ( f ∗ Py )ˆ = fˆ · ( Py )ˆ [see 11.66] and the observation


that fˆ ∈ L∞ (R), ( Py )ˆ ∈ L1 (R) [see 11.49 and 11.79] and the observation that
fˆ ∈ L2 (R), ( Py )ˆ ∈ L∞ (R) [see 11.82 and 11.49].
Now the Fourier Inversion Formula (11.76) as applied to f ∗ Py (which is valid by
11.88 and 11.89) implies that

F 4 ( f ∗ Py ) = f ∗ Py .

Taking the limit in L2 (R) of both sides of the equation above as y ↓ 0, we have
F 4 f = f (by 11.74), completing the proof of (b).
Plancherel’s Theorem (11.82) tells us that F is an isometry on L2 (R). Part (a)
implies that F is surjective. Because a surjective isometry is unitary (see 10.61), we
conclude that F is unitary, completing the proof of (a).
The Spectral Mapping Theorem [see 10.40—take p(z) = z4 ] and (b) imply that
4
α = 1 for each α ∈ sp( T ). In other words, sp( T ) ⊂ {1, i, −1, −i }. However, 1, i,
−1, −i are all eigenvalues of F (see Example 11.51 and Exercises 2, 3, and 4) and
thus are all in sp( T ). Hence sp( T ) = {1, i, −1, −i }, completing the proof of (c).

EXERCISES 11C
1 Suppose f ∈ L1 (R). Prove that k fˆk∞ = k f k1 if and only if there exists
ζ ∈ ∂D and t ∈ R such that ζ f ( x )e−itx ≥ 0 for almost every x ∈ R.
2
2 Suppose f ( x ) = xe−πx for all x ∈ R. Show that fˆ = −i f .
378 Chapter 11 Fourier Analysis

2 2
3 Suppose f ( x ) = 4πx2 e−πx − e−πx for all x ∈ R. Show that fˆ = − f .

4 Find f ∈ L1 (R) such that f 6= 0 and fˆ = i f .


5 Prove that if p is a polynomial on R with complex coefficients and f : R → C
2
is defined by f ( x ) = p( x )e−πx , then there exists a polynomial q on R with
2
complex coefficients such that deg q = deg p and fˆ(t) = q(t)e−πt for all
t ∈ R.
6 Suppose (
xe−2πx if x > 0,
f (x) =
0 if x ≤ 0.
1
Show that fˆ(t) = for all t ∈ R.
4π 2 (1 + it)2
7 Prove the formulas in 11.55 for the Fourier transforms of translations, rotations,
and dilations.
8 Suppose f ∈ L1 (R) and n ∈ Z+. Define g : R → C by g( x ) = x n f ( x ). Prove
that if g ∈ L1 (R), then fˆ is n times continuously differentiable on R and

( fˆ)(n) (t) = (−2πi )n ĝ(t)


for all t ∈ R.
9 Suppose n ∈ Z+ and f ∈ L1 (R) is n times continuously differentiable and
f (n) ∈ L1 (R). Prove that if t ∈ R, then

( f (n) )ˆ(t) = (2πit)n fˆ(t).


0
10 Suppose 1 ≤ p ≤ ∞, f ∈ L p (R), and g ∈ L p (R). Prove that f ∗ g is a
uniformly continuous function on R.
11 Suppose f ∈ L∞ (R), x ∈ R, and f is continuous at x. Prove that

lim(Py f )( x ) = f ( x ).
y ↓0

12 Suppose p ∈ [1, ∞] and f ∈ L p (R). Prove that Py (Py0 f ) = Py+y0 f for all
y, y0 > 0.
13 Suppose p ∈ [1, ∞] and f ∈ L p (R). Prove that if 0 < y < y0 , then

kPy f k p ≥ kPy0 f k p .

14 Suppose f ∈ L1 (R).

(a) Prove that ( f )ˆ(t) = fˆ(−t) for all t ∈ R.


(b) Prove that f ( x ) ∈ R for almost every x ∈ R if and only if fˆ(t) = fˆ(−t)
for all t ∈ R.
Section 11C Fourier Transform 379

4
15 Define f ∈ L1 (R) by f ( x ) = e− x χ[0, ∞)( x ). Show that fˆ ∈
/ L1 ( R ).

16 Suppose f ∈ L1 (R) and fˆ ∈ L1 (R). Prove that f ∈ L2 (R) and fˆ ∈ L2 (R).


17 Prove there exists a continuous function g : R → R such that lim g(t) = 0
t→±∞
/ { fˆ : f ∈ L1 (R)}.
and g ∈

18 Prove that if f ∈ L1 (R), then k fˆk2 = k f k2 .


[This exercise slightly improves Plancherel’s Theorem (11.82) because here we
have the weaker hypothesis that f ∈ L1 (R) instead of f ∈ L1 (R) ∩ L2 (R).
Because of Plancherel’s Theorem, here you need only prove that if f ∈ L1 (R)
and k f k2 = ∞, then k fˆk2 = ∞.]
19 Suppose y > 0. Define on operator T on L2 (R) by T f = f ∗ Py .

(a) Show that T is a self-adjoint operator on L2 (R).


(b) Show that sp( T ) = [0, 1].
[Because the spectrum of each compact operator is a countable set (by 10.93),
part (b) above implies that T is not a compact operator. This conclusion differs
from the situation on the unit circle—see Exercise 9 in Section 11B.]
20 Prove that if f ∈ L1 (R) and g ∈ L2 (R), then F ( f ∗ g) = fˆ · F g.
21 Prove that f , g ∈ L2 (R), then ( f g)ˆ = (F f ) ∗ (F g).
Chapter 12
Probability Measures

Probability theory has become increasingly important in multiple parts of science.


Getting deeply into probability theory requires a full book, not just a chapter. For
readers who intend to pursue further studies in probability theory, this chapter gives
you a good head start. For readers not intending to delve further into probability
theory, this chapter gives you a taste of the subject.
Modern probability theory makes major use of measure theory. As we will see, a
probability measure is simply a measure such that the measure of the whole space
equals 1. Thus a thorough understanding of the chapters of this book dealing with
measure theory and integration provides a solid foundation for probability theory.
However, probability theory is not simply the special case of measure theory where
the whole space has measure 1. The questions that probability theory investigates
differ from the questions natural to measure theory. For example, the probability
notions of independent sets and independent random variables, which are introduced
in this chapter, do not arise in measure theory.
Even when concepts in probability theory have the same meaning as well-known
concepts in measure theory, the terminology and notation can be quite different. Thus
one goal of this chapter is to introduce the vocabulary of probability theory. This
difference in vocabulary between probability theory and measure theory occurred
because the two subjects had different historical developments, only coming together
in the first half of the twentieth century.

Dice used in games of chance. The beginning of probability theory can be traced to
correspondence in 1654 between Pierre de Fermat (1601–1665) and Blaise Pascal
(1623–1662) about how to distribute fairly money bet on an unfinished game of dice.
CC-BY-SA Alexander Dreyer

380
Chapter 12 Probability Measures 381

Probability Spaces
We begin with an intuitive and nonrigorous motivation. Suppose we pick a real
number at random from the interval (0, 1), with each real number having an equal
probability of being chosen (whatever that means). What is the probability that the
9
chosen number is in the interval ( 10 , 1)? The only reasonable answer to this question
1
is 10 . More generally, if I1 , I2 , . . . is a disjoint sequence of open intervals contained
S
in (0, 1), then the probability that our randomly chosen real number is in ∞ n=1 In

should be ∑n=1 `( In ), where `( I ) denotes the length of an interval I. Still more
generally, if A is a Borel subset of (0, 1), then the probability that our random number
is in A should be the Lebesgue measure of A.
With the paragraph above as motivation, we are now ready to define a probability
measure. We will use the notation and terminology common in probability theory
instead of the conventions of measure theory.
In particular, the set in which everything takes place is now called Ω instead of
the usual X in measure theory. The σ-algebra on Ω is called F instead of S , which
we have used in previous chapters. Our measure is now called P instead of µ. This
new notation and terminology can be disorienting when first encountered. However,
reading this chapter should help you become comfortable with this notation and
terminology, which are standard in probability theory.

12.1 Definition probability measure

Suppose F is a σ-algebra on a set Ω.


• A probability measure on (Ω, F ) is a measure P on (Ω, F ) such that
P(Ω) = 1.
• Ω is called the sample space.
• An event is an element of F (F need not be mentioned if it is clear from the
context).
• If A is an event, then P( A) is called the probability of A.
• If P is a probability measure on (Ω, F ), then the triple (Ω, F , P) is called a
probability space.

12.2 Example probability measures

• Suppose n ∈ Z+ and Ω is a set containing exactly n elements. Let F denote


the collection of all subsets of Ω. Then
counting measure on Ω
n
is a probability measure on (Ω, F ).
382 Chapter 12 Probability Measures

• As a more specific example of


This example illustrates the common
the previous item, suppose that
practice in probability theory of
Ω = {40, 41, . . . , 49} and P =
using lower case ω to denote a
(counting measure on Ω)/10. Let
typical element of upper case Ω.
A = {ω ∈ Ω : ω is even} and
B = {ω ∈ Ω : ω is prime}. Then P( A) [which is the probability that an
element of this sample space Ω is even] is 12 and P( B) [which is the probability
3
that an element of this sample space Ω is prime] is 10 .
• Let λ denote Lebesgue measure on the interval [0, 1]. Then λ is a probability
measure on ([0, 1], B), where B denotes the σ-algebra of Borel subsets of [0, 1].
• Let λ denote Lebesgue measure on R, and let B denote the σ-algebra of Borel
2
subsets of R. Define h : R → (0, ∞) by h( x ) = √1 e− x /2 . Then h dλ is a

probability measure on (R, B) [see 9.6 for the definition of h dλ].

In measure theory, we used the notation χ A to denote the characteristic function


of a set A. In probability theory, this function has a different name and different
notation, as we see in the next definition.

12.3 Definition indicator function; 1 A

If Ω is a set and A ⊂ Ω, then the indicator function of A is the function


1 A : Ω → R defined by
(
1 if ω ∈ A,
1 A (ω ) =
0 if ω ∈
/ A.

The next definition gives the replacement in probability theory for measure theory’s
phrase almost every.

12.4 Definition almost surely

Suppose (Ω, F , P) is a probability space. An event A is said to happen almost


surely if the probability of A is 1, or equivalently if P(Ω \ A) = 0.

12.5 Example almost surely


Let P denote Lebesgue measure on the interval [0, 1]. If ω ∈ [0, 1], then ω is
almost surely an irrational number (because the set of rational numbers has Lebesgue
measure 0).
This example shows that an event having probability 1 (equivalent to happening
almost surely) does not mean that the event definitely happens. Conversely, an event
having probability 0 does not mean that the event is impossible. Specifically, if a real
number is chosen at random from [0, 1] using Lebesgue measure as the probability,
then the probability that the number is rational is 0, but that event can still happen.
Chapter 12 Probability Measures 383

The following result is frequently useful in probability theory. A careful reading of


the proof of this result, as our first proof in this chapter, should give you good practice
using some of the notation and terminology commonly used in probability theory.
This proof also illustrates the point that having a good understanding of measure
theory and integration can often be extremely useful in probability theory—here we
use the Monotone Convergence Theorem.

12.6 Borel–Cantelli Lemma

Suppose (Ω, F , P) is a probability space and A1 , A2 , . . . is a sequence of events


such that ∑∞
n=1 P ( An ) < ∞. Then

P({ω ∈ Ω : ω ∈ An for infinitely many n ∈ Z+ }) = 0.

Proof Let A = {ω ∈ Ω : ω ∈ An for infinitely many n ∈ Z+ }. Then


∞ [
\ ∞
A= An .
m =1 n = m

Thus A ∈ F , and hence P( A) makes sense.


The Monotone Convergence Theorem (3.11) implies that
Z  ∞  ∞ Z ∞
∑ 1 An
Ω n =1
dP = ∑ 1 An dP = ∑ P( An ) < ∞.
n =1 Ω n =1

Thus ∑∞
n=1 1 An is almost surely finite. Hence P ( A ) = 0.

Independent Events and Independent Random Variables


The notion of independent events, which we now define, is one of the key concepts
that distinguishes probability theory from measure theory.

12.7 Definition independent events

Suppose (Ω, F , P) is a probability space.


• Two events A and B are called independent if

P ( A ∩ B ) = P ( A ) · P ( B ).

• More generally, a family of events { Ak }k∈Γ is called independent if

P( Ak1 ∩ · · · ∩ Ak n ) = P( Ak1 ) · · · P( Ak n )

whenever k1 , . . . , k n are distinct elements of Γ.

The next two examples should help develop your intuition about independent
events.
384 Chapter 12 Probability Measures

12.8 Example independent events: coin tossing


Suppose Ω = { H, T }4 , where H and T are symbols that you can think of as
denoting “heads” and “tails”. Thus elements of Ω are 4-tuples of the form
ω = ( ω1 , ω2 , ω3 , ω4 ) ,
where each ω j is H or T. Let F be the collection of all subsets of Ω, and let
P = (counting measure on Ω)/16, as we expect from a fair coin toss.
Let
A = { ω ∈ Ω : ω1 = ω2 = ω3 = H } and B = { ω ∈ Ω : ω4 = H } .
Then A contains two elements and thus P( A) = 18 , corresponding to probability 18
that the first three coin tosses are all heads. Also, B contains eight elements and thus
P( B) = 12 , corresponding to probability 21 that the fourth coin toss is heads.
Now
1
P( A ∩ B) = 16 = P ( A ) · P ( B ),
where the first equality holds because A ∩ B consists of only the one element
( H, H, H, H ) and the second equality holds because P( A) = 18 and P( B) = 12 .
The equation above shows that A and B are independent events.
If we toss a fair coin many times, we expect that about half the time it will be
heads. Thus some people mistakenly believe that if the first three tosses of a fair
coin are heads, then the fourth toss should have a higher probability of being tails,
to balance out the previous heads. However, the coin cannot remember that it had
three heads in a row, and thus the fourth coin toss has probability 12 of being heads
regardless of the results of the three previous coin tosses. The independence of the
events A and B above captures the notion that the results of a fair coin toss do not
depend upon previous results.

12.9 Example independent events: product probability space


Suppose (Ω1 , F1 , P1 ) and (Ω2 , F2 , P2 ) are probability spaces. Then
(Ω1 × Ω2 , F1 ⊗ F2 , P1 × P2 ),
as defined in Chapter 5, is also a probability space.
If A ∈ F1 and B ∈ F2 , then ( A × Ω2 ) ∩ (Ω1 × B) = A × B. Thus

( P1 × P2 ) ( A × Ω2 ) ∩ (Ω1 × B) = ( P1 × P2 )( A × B)
= P1 ( A) · P2 ( B)
= ( P1 × P2 )( A × Ω2 ) · ( P1 × P2 )(Ω1 × B),
where the second equality follows from the definition of the product measure, and
the third equality holds because of the definition of the product measure and because
P1 and P2 are probability measures.
The equation above shows that the events A × Ω2 and Ω1 × B are independent
events in F1 ⊗ F2 .
Chapter 12 Probability Measures 385

Compare the next result to the Borel–Cantelli Lemma (12.6).

12.10 relative of Borel–Cantelli Lemma

Suppose (Ω, F , P ) is a probability space and { An }n∈Z+ is an independent


family of events such that ∑∞
n=1 P ( An ) = ∞. Then

P({ω ∈ Ω : ω ∈ An for infinitely many n ∈ Z+ }) = 1.

Proof Let A = {ω ∈ Ω : ω ∈ An for infinitely many n ∈ Z+ }. Then


∞ \
[ ∞
12.11 Ω\A = ( Ω \ A n ).
m =1 n = m

If m, M ∈ Z+ are such that m ≤ M, then


M
\  M
P (Ω \ An ) = ∏ P(Ω \ An )
n=m n=m

M 
= ∏ 1 − P( An )
n=m

M
12.12 ≤ e − ∑n=m P( An ) ,
where the first line holds because the family {Ω \ An }n∈Z+ is independent (see
Exercise 4) and the third line holds because 1 − t ≤ e−t for all t ≥ 0.
Because ∑∞ n=1 P ( An ) = ∞, by choosing M large we can make the right side of
12.12 as close to 0 as we wish. Thus
\∞ 
P (Ω \ An ) = 0
n=m

for all m ∈ Z+. Now 12.11 implies that P(Ω \ A) = 0. Thus we conclude that
P( A) = 1, as desired.

For the rest of this chapter, assume that F = R. Thus, for example, if (Ω, F , P) is
a probability space, then L1 ( P) will always
R refer to the vector space of real-valued
F -measurable functions on Ω such that Ω | f | dP < ∞.

12.13 Definition random variable; expectation; EX

Suppose (Ω, F , P) is a probability space.


• A random variable on (Ω, F ) is a measurable function from Ω to R.
• If X ∈ L1 ( P), then the expectation (sometimes called the expected value)
of the random variable X is denoted EX and is defined by
Z
EX = X dP.

386 Chapter 12 Probability Measures

If F is clear from the context, the phrase “random variable on Ω” can be used
instead of the more precise phrase “random variable on (Ω, F )”. If both Ω and F
are clear from the context, then the phrase “random variable” has no ambiguity and
is often used.
Because P(Ω) = 1, the expectation EX of a random variable X ∈ L1 ( P) can be
thought of as the average or mean value of X.
The next definition illustrates a convention often used in probability theory: the
variable is often omitted when describing a set. Thus, for example, { X ∈ U } means
{ω ∈ Ω : X (ω ) ∈ U }, where U is a subset of R. Also, probabilists often also omit
the set brackets, as we do for the first time in the second bullet point below, when
appropriate parentheses are nearby.

12.14 Definition independent random variables

Suppose (Ω, F , P) is a probability space.


• Two random variables X and Y are called independent if { X ∈ U } and
{Y ∈ V } are independent events for all Borel sets U, V in R.
• More generally, a family of random variables { Xk }k∈Γ is called independent
if { Xk ∈ Uk }k∈Γ is independent for all families of Borel sets {Uk }k∈Γ in R.

12.15 Example independent random variables

• Suppose (Ω, F , P) is a probability space and A, B ∈ F . Then 1 A and 1B are


independent random variables if and only if A and B are independent events, as
you should verify.
• Suppose Ω = { H, T }4 is the sample space of four coin tosses, with Ω and P as
in Example 12.8. Define random variables X and Y by
X (ω1 , ω2 , ω3 , ω4 ) = number of ω1 , ω2 , ω3 that equal H
and
Y (ω1 , ω2 , ω3 , ω4 ) = number of ω3 , ω4 that equal H.
1
Then X and Y are not independent random variables because P( X = 3) = 8
and P(Y = 0) = 14 but P({ X = 3} ∩ {Y = 0}) = P(∅) = 0 6= 18 · 41 .
• Suppose (Ω1 , F1 , P1 ) and (Ω2 , F2 , P2 ) are probability spaces, Z1 is a random
variable on Ω1 , and Z2 is a random variable on Ω2 . Define random variables X
and Y on Ω1 × Ω2 by
X (ω1 , ω2 ) = Z1 (ω1 ) and Y (ω1 , ω2 ) = Z2 (ω2 ).

Then X and Y are independent random variables on Ω1 × Ω2 (with respect to the


probability measure P1 × P2 ), as you should verify.

If X is a random variable and f : R → R is Borel measurable, then f ◦ X is a


random variable (by 2.44). For example, if X is a random variable, then X 2 and e X
are random variables. The next result states that compositions preserve independence.
Chapter 12 Probability Measures 387

12.16 functions of independent random variables are independent

Suppose (Ω, F , P ) is a probability space, X and Y are independent random


variables, and f , g : R → R are Borel measurable. Then f ◦ X and g ◦ Y are
independent random variables.

Proof Suppose U, V are Borel subsets of R. Then


 
P { f ◦ X ∈ U } ∩ { g ◦ Y ∈ V } = P { X ∈ f −1 (U )} ∩ {Y ∈ g−1 (V )}
 
= P X ∈ f −1 (U ) · P Y ∈ g −1 ( V )

= P ( f ◦ X ∈ U ) · P ( g ◦ Y ∈ V ),
where the second equality holds because X and Y are independent random variables.
The equation above shows that f ◦ X and g ◦ Y are independent random variables.

If X, Y ∈ L1 ( P), then clearly E( X + Y ) = E( X ) + E(Y ). The next result gives


a nice formula for the expectation of XY when X and Y are independent. This
formula has sometimes been called the dream equation of calculus students.

12.17 expectation of product of independent random variables

Suppose (Ω, F , P ) is a probability space and X and Y are independent random


variables in L2 ( P). Then

E( XY ) = EX · EY.

Proof First consider the case where X and Y are each simple functions, taking
on only finitely many values. Thus there are distinct numbers a1 , . . . , a M ∈ R and
distinct numbers b1 , . . . , b N ∈ R such that
X = a 1 1{ X = a1 } + · · · + a M 1{ X = a M } and Y = b1 1{Y =b1 } + · · · + b N 1{Y =bN } .
Now
M N M N
XY = ∑∑ a j b k 1 { X = a j } 1 {Y = b k } = ∑ ∑ a j bk 1{X=aj }∩{Y=bk } .
j =1 k =1 j =1 k =1

Thus
M N 
E( XY ) = ∑ ∑ a j bk P { X = a j } ∩ {Y = b k }
j =1 k =1

 M  N 
= ∑ a j P( X = a j ) ∑ b k P (Y = b k )
j =1 k =1

= EX · EY,
where the second equality above comes from the independence of X and Y. The last
equation gives the desired conclusion in the case where X and Y are simple functions.
388 Chapter 12 Probability Measures

Now consider arbitrary independent random variables X and Y in L2 ( P). Let


f 1 , f 2 , . . . be a sequence of Borel measurable simple functions from R to R that
approximate the identity function on R (the function t 7→ t) in the sense that
limn→∞ f n (t) = t for every t ∈ R and | f n (t)| ≤ t for all t ∈ R and all n ∈ Z+
(see 2.89, taking f to be the identity function, for construction of this sequence). The
random variables f n ◦ X and f n ◦ Y are independent (by 12.17). Thus the result in
the first paragraph of this proof shows that

E ( f n ◦ X )( f n ◦ Y ) = E( f n ◦ X ) · E( f n ◦ Y )

for each n ∈ Z+. The limit as n → ∞ of the right side of the equation above equals
EX · EY [by the Dominated Convergence Theorem (3.31)]. The limit as n → ∞
of the left side of the equation above equals E( XY ) [use Hölder’s inequality (7.9)].
Thus the equation above implies that E( XY ) = EX · EY.

Variance and Standard Deviation


The variance and standard deviation of a random variable, defined below, measure
how much a random variable differs from its expectation.

12.18 Definition variance; standard deviation; σ( X )

Suppose (Ω, F , P) is a probability space and X ∈ L2 ( P) is a random variable.



• The variance of X is defined to be E ( X − EX )2 .
• The standard deviation of X is denoted σ( X ) and is defined by
q 
σ( X ) = E ( X − EX )2 .

In other words, the standard deviation of X is the square root of the variance
of X.
2
The notation σ2 ( X ) means σ( X ) . Thus σ2 ( X ) is the variance of X.

12.19 Example variance and standard deviation of an indicator function


Suppose (Ω, F , P) is a probability space and A ∈ F is an event. Then

σ2 (1 A ) = E (1 A − E1 A )2

= E (1 A − P( A))2
= E(1 A − 2P( A) · 1 A + P( A)2 )
2 2
= P( A) − 2 P( A) + P( A)

= P( A) · 1 − P( A) .
q 
Thus σ(1 A ) = P( A) · 1 − P( A) .
Chapter 12 Probability Measures 389

The next result gives a formula for the variance of a random variable. This formula
is often more convenient to use than the formula that defines the variance.

12.20 variance formula

Suppose (Ω, F , P) is a probability space and X ∈ L2 ( P) is a random variable.


Then
σ2 ( X ) = E( X 2 ) − ( EX )2 .

Proof We have

σ2 ( X ) = E ( X − EX )2

= E X 2 − 2( EX ) X + ( EX )2

= E( X 2 ) − 2( EX )2 + ( EX )2

= E( X 2 ) − ( EX )2 ,

as desired.

Our next result is called Chebyshev’s inequality. It states, for example (take t = 2
below) that the probability that a random variable X differs from its average by more

than twice its standard deviation is at most 14 . Note that P | X − EX | ≥ tσ( X ) is

shorthand for P {ω ∈ Ω : | X (ω ) − EX | ≥ tσ( X )} .

12.21 Chebyshev’s inequality

Suppose (Ω, F , P) is a probability space and X ∈ L2 ( P) is a random variable.


Then
 1
P | X − EX | ≥ tσ( X ) ≤ 2
t
for all t > 0.

Proof Suppose t > 0. Then


 
P | X − EX | ≥ tσ( X ) = P | X − EX |2 ≥ t2 σ2 ( X )

1 
≤ E ( X − EX )2
t2 σ 2 ( X )
1
= ,
t2
where the second line above comes from applying Markov’s inequality (4.1) with
h = | X − EX |2 and c = t2 σ2 ( X ).

The next result gives a beautiful formula for the variance of the sum of independent
random variables.
390 Chapter 12 Probability Measures

12.22 variance of sum of independent random variables

Suppose (Ω, F , P) is a probability space and X1 , . . . , Xn ∈ L2 ( P) are indepen-


dent random variables. Then

σ 2 ( X1 + · · · + X n ) = σ 2 ( X1 ) + · · · + σ 2 ( X n ) .

Proof Using the variance formula given by 12.20, we have


 n   n 2   n  2
σ2 ∑ kX = E ∑ kX − E ∑ kX
k =1 k =1 k =1
 n     n 2
=E ∑ Xk 2 + 2E ∑ X j Xk − ∑ EXk
k =1 1≤ j < k ≤ n k =1
n n    
= ∑ E(Xk2 ) − ∑ (EXk )2 + 2 ∑ E ( X j Xk ) − 2 ∑ EX j · EXk
k =1 k =1 1≤ j < k ≤ n 1≤ j < k ≤ n
n
= ∑ σ 2 ( Xk ),
k =1
where the last equality uses 12.20, 12.17, and the hypothesis that X1 , . . . , Xn are
independent random variables.

Conditional Probability and Bayes’ Theorem


The conditional probability PB ( A) that we are about to define should be interpreted
to mean the probability that ω will be in A given that ω ∈ B. Because ω is in A ∩ B
if and only if ω ∈ B and ω ∈ A, and because we expect probabilities to multiply, it
is reasonable to expect that
P( B) · PB ( A) = P( A ∩ B).
Thus we are led to the following definition.

12.23 Definition conditional probability; PB

Suppose (Ω, F , P) is a probability space and B is an event with P( B) > 0.


Define PB : F → [0, 1] by

P( A ∩ B)
PB ( A) = .
P( B)

If A ∈ F , then PB ( A) is called the conditional probability of A given B.

You should verify that with B as above, PB is a probability measure on (Ω, F ). If


A ∈ F , then PB ( A) = P( A) if and only if A and B are independent events.
We now present two versions of what is called Bayes’ Theorem. You should
do a web search and read about the many uses of these results, including some
controversial applications.
Chapter 12 Probability Measures 391

12.24 Bayes’ Theorem, first version

Suppose (Ω, F , P) is a probability space and A, B are events with positive


probability. Then
P ( B) · P( A)
PB ( A) = A .
P( B)

Proof We have
P( A ∩ B) P( A ∩ B) · P( A) P ( B) · P( A)
PB ( A) = = = A .
P( B) P( A) · P( B) P( B)

Plaque honoring Thomas


Bayes in Tunbridge Wells,
England.
CC-BY-SA Alexander Dreyer

12.25 Bayes’ Theorem, second version

Suppose (Ω, F , P) is a probability space, B is an event with positive probability,


and A1 , . . . , An are pairwise disjoint events, each with positive probability, such
that A1 ∪ · · · ∪ An = Ω. Then
PA ( B) · P( Ak )
PB ( Ak ) = n k
∑ j=1 PA j ( B) · P( A j )
for each k ∈ {1, . . . , n}.

Proof Consider the denominator of the expression above. We have


n n
12.26 ∑ PAj ( B) · P( A j ) = ∑ P( A j ∩ B) = P( B).
j =1 j =1

Now suppose k ∈ {1, . . . , n}. Then


PAk ( B) · P( Ak ) PA ( B) · P( Ak )
PB ( Ak ) = = n k ,
P( B) ∑ j=1 PA j ( B) · P( A j )
where the first equality comes from the first version of Bayes’s Theorem (12.24) and
the second equality comes from 12.26.
392 Chapter 12 Probability Measures

Distribution and Density Functions of Random Variables


For the rest of this chapter, let B denote the σ-algebra of Borel subsets of R.
Each random variable X determines a probability measure PX on (R, B) and a
function X̃ : R → [0, 1] as in the next definition.

12.27 Definition probability distribution and distribution function; PX ; X̃

Suppose (Ω, F , P) is a probability space and X is a random variable.


• The probability distribution of X is the probability measure PX defined on
(R, B) by 
PX ( B) = P( X ∈ B) = P X −1 ( B) .

• The distribution function of X is the function X̃ : R → [0, 1] defined by



X̃ (s) = PX (−∞, s] = P( X ≤ s).

You should verify that the probability distribution PX as defined above is indeed a
probability measure on (R, B). Note that the distribution function X̃ depends upon
the probability measure P as well as the random variable X, even though P is not
included in the notation X̃ (because P is usually clear from the context).

12.28 Example probability distribution and distribution function of an indicator function


Suppose (Ω, F , P) is a probability space and A ∈ F is an event. Then you should
verify that 
P1 A = 1 − P( A) δ0 + P( A)δ1 ,
where for t ∈ R the measure δt on (R, B) is defined by
(
1 if t ∈ B,
δt ( B) =
0 if t ∈/ B.

The distribution function of 1 A is the function (1 A )˜ : R → [0, 1] given by



0
 if s < 0,
(1 A )˜(s) = 1 − P( A) if 0 ≤ s < 1,


1 if s ≥ 1,

as you should verify.

One direction of the next result states that every probability distribution is a right-
continuous increasing function, with limit 0 at −∞ and limit 1 at ∞. The other
direction of the next result states that every function with those properties is the
distribution function of some random variable on some probability space. The proof
shows that we can take the sample space to be (0, 1), the σ-algebra to be the Borel
subsets of (0, 1), and the probability measure to be Lebesgue measure on (0, 1).
Chapter 12 Probability Measures 393

Your understanding of the proof of the next result should be enhanced by Exercise
13, which asserts that if the function H : R → (0, 1) appearing in the next result is
continuous and injective, then the random variable X : (0, 1) → R in the proof is the
inverse function of H.

12.29 characterization of distribution functions

Suppose H : R → [0, 1] is a function. Then there exists a probability space


(Ω, F , P) and a random variable X on (Ω, F ) such that H = X̃ if and only if
the following conditions are all satisfied:

(a) s < t ⇒ H (s) ≤ H (t) (in other words, H is an increasing function);


(b) lim H (t) = 0;
t→−∞

(c) lim H (t) = 1;


t→∞

(d) lim H (t) = H (s) for every s ∈ R (in other words, H is right continuous).
t↓s

Proof First suppose H = X̃ for some probability space (Ω, F , P) and some random
variable X on (Ω, F ). Then (a) holds because s < t implies (−∞, s] ⊂ (−∞, t].
Also, (b) and (d) follow from 2.60. Furthermore, (c) follows from 2.59, completing
the proof in this direction.
To prove the other direction, now suppose that H satisfies (a) through (d). Let
Ω = (0, 1), let F be the collection of Borel subsets of the interval (0, 1), and let P
be Lebesgue measure on F . Define a random variable X by
12.30 X (ω ) = sup{t ∈ R : H (t) < ω }
for ω ∈ (0, 1). Clearly X is an increasing function and thus is measurable (in other
words, X is indeed a random variable).

Suppose s ∈ R. If ω ∈ 0, H (s) , then

X (ω ) ≤ X H (s) = sup{t ∈ R : H (t) < H (s)} ≤ s,
where the first inequality holds because X is an increasing function and the last
inequality holds because H is an increasing function. Hence

12.31 0, H (s) ⊂ { X ≤ s}.
If ω ∈ (0, 1) and X (ω ) ≤ s, then H (t) ≥ ω for all t > s (by 12.30). Thus
H (s) = lim H (t) ≥ ω,
t↓s

where the equality


 above comes from (d). Rewriting the inequality above,
 we have
ω ∈ 0, H (s) . Thus we have shown that { X ≤ s} ⊂ 0, H (s) , which when

combined with 12.31 shows that { X ≤ s} = 0, H (s) . Hence
 
X̃ (s) = P( X ≤ s) = P 0, H (s) = H (s),
as desired.
394 Chapter 12 Probability Measures

In the definition below and in the following discussion, λ denotes Lebesgue


measure on R, as usual.

12.32 Definition density function

Suppose X is a random variable on some probability space. If there exists


h ∈ L1 (R) such that Z s
X̃ (s) = h dλ
−∞
for all s ∈ R, then h is called the density function of X.

If there is a density function of a random variable X, then it is unique [up to


changes on sets of Lebesgue measure 0, which is already taken into account because
we are thinking of density functions as elements of L1 (R) instead of elements of
L1 (R)]; see Exercise 6 in Chapter 4.
If X is a random variable that has a density function h, then the distribution
function X̃ is differentiable almost everywhere (with respect to Lebesgue measure)
and X̃ 0 (s) = h(s) for almost every s ∈ R (by the second version of the Lebesgue
Differentiation Theorem; see 4.19). Because X̃ is an increasing function, this implies
that h(s) ≥ 0 for almost every s ∈ R. In other words, we can assume that a density
function is nonnegative.
In the definition above of a density function, we started with a probability space
and a random variable on it. Often in probability theory, the procedure goes in the
other direction.
R∞ Specifically, we can start with a nonnegative function h ∈ L1 (R)
such that −∞ h dλ = 1. We use h to define a probability measure on (R, B) and then
consider the identity random variable X on R. The function h that we started with is
then the density function of X. The following result formalizes this procedure and
gives formulas for the mean and standard deviation in terms of the density function h.

12.33 mean and variance of random variable generated by density function


R∞
Suppose h ∈ L1 (R) is such that −∞ h dλ = 1 and h( x ) ≥ 0 for almost every
x ∈ R. Let P be the probability measure on (R, B) defined by
Z
P( B) = h dλ.
B

Let X be the random variable on (R, B) defined by X ( x ) = x for each x ∈ R.


Then h is the density function of X. Furthermore, if X ∈ L1 ( P) then
Z ∞
EX = xh( x ) dλ( x ),
−∞

and if X ∈ L2 ( P) then
Z ∞ Z ∞ 2
σ2 ( X ) = x2 h( x ) dλ( x ) − xh( x ) dλ( x ) .
−∞ −∞
Chapter 12 Probability Measures 395

Rs
Proof The equation X̃ (s) = −∞ h dλ holds by the definitions of X̃ and P. Thus h
is the density function of X. R∞ R∞
Our definition of P to equal h dλ implies that −∞ f dP = −∞ f h dλ for all
f ∈ L1 ( P) [see Exercise 5 in Section 9A]. Thus the formula for the mean EX
follows immediately from the definition of EX, and the formula for the variance
σ2 ( X ) follows from 12.20.

The following example illustrates the result above with a few especially useful
choices of the density function h.

12.34 Example density functions

• Suppose h = 1[0,1] . This density function h is called the uniform density on


[0, 1]. In this case, P( B) = λ( B ∩ [0, 1]) for each Borel set B ⊂ R. For the
corresponding random variable X ( x ) = x for x ∈ R, the distribution function
X̃ is given by the formula

0 if s ≤ 0,

X̃ (s) = s if 0 < s < 1,


1 if s ≥ 1.

1 1
The formulas in 12.33 show that EX = 2 and σ( X ) = √ .
2 3

• Suppose α > 0 and (


0 if x < 0,
h( x ) =
αe−αx if x ≥ 0.
This density function h is called the exponential density on [0, ∞). For the
corresponding random variable X ( x ) = x for x ∈ R, the distribution function
X̃ is given by the formula
(
0 if s < 0,
X̃ (s) =
1 − e−αs if s ≥ 0.

1
The formulas in 12.33 show that EX = α and σ( X ) = α1 .
• Suppose
1 2
h( x ) = √ e− x /2

for x ∈ R. This density function is called the standard normal density. For
the corresponding random variable X ( x ) = x for x ∈ R, we have X̃ (0) = 12 .
For general s ∈ R, no formula exists for X̃ (s) in terms of elementary functions.
However, the formulas in 12.33 show that EX = 0 and (with the help of some
calculus) σ( X ) = 1.
396 Chapter 12 Probability Measures

Weak Law of Large Numbers


Families of random variables all of which look the same in terms of their distribution
functions get a special name, as we see in the next definition.

12.35 Definition identically distributed; i.i.d.

Suppose (Ω, F , P) is a probability space.


• A family of random variables on (Ω, F ) is called identically distributed if
all the random variables in the family have the same distribution function.
• More specifically, a family { Xk }k∈Γ of random variables on (Ω, F ) is called
identically distributed if

P ( X j ≤ s ) = P ( Xk ≤ s )

for all j, k ∈ Γ.
• A family of random variables that is independent and identically distributed
is said to be independent identically distributed, often abbreviated as i.i.d.

12.36 Example family of random variables for decimal digits is i.i.d.


Consider the probability space ([0, 1], B , P), where B is the collection of Borel
subsets of the interval [0, 1] and P is Lebesgue measure on ([0, 1], B). For k ∈ Z+,
define a random variable Xk : [0, 1] → R by

Xk (ω ) = kth -digit in decimal expansion of ω,

where for those numbers ω that have two different decimal expansions we use the
one that does not end in an infinite string of 9s.
Notice that P( Xk ≤ π ) = 0.4 for every k ∈ Z+. More generally, the family
{ Xk }k∈Z+ is identically distributed, as you should verify.
The family { Xk }k∈Z+ is also independent, as you should verify. Thus { Xk }k∈Z+
is an i.i.d. family of random variables.

Identically distributed random variables have the same expectation and the same
standard deviation, as the next result shows.

12.37 identically distributed random variables have same mean and variance

Suppose (Ω, F , P) is a probability space and { Xk }k∈Γ is an identically dis-


tributed family of random variables in L2 ( P). Then

EX j = EXk and σ ( X j ) = σ ( Xk )

for all j, k ∈ Γ.
Chapter 12 Probability Measures 397

Proof Suppose j ∈ Z+. Let f 1 , f 2 , . . . be the sequence of simple functions converg-


ing pointwise to X j as constructed in the proof of 2.89. The Dominated Convergence
Theorem (3.31) implies that EX j = limn→∞ E f n . Because of how each f n is con-
structed, each E f n depends only on n and the numbers P(c ≤ X j < d) for c < d.
However,
  
P(c ≤ X j < d) = lim P X j ≤ d − m1 − P X j ≤ c − m1
m→∞

for c < d. Because { Xk }k∈Γ is an identically distributed family, the numbers above
on the right are independent of j. Thus EX j = EXk for all j, k ∈ Z+.
Apply the result from the paragraph above to the identically distributed family
{ Xk 2 }k∈Γ and use 12.20 to conclude that σ( X j ) = σ( Xk ) for all j, k ∈ Γ.

The next result has the nicely intuitive interpretation that if we repeat a random
process many times, then the probability that the average of our results differs from
our expected average by more than any fixed positive number ε has limit 0 as we
increase the number of repetitions of the process.

12.38 Weak Law of Large Numbers

Suppose (Ω, F , P) is a probability space and { Xk }k∈Z+ is an i.i.d. family of


random variables in L2 ( P), each with expectation µ. Then
 n  

lim P
n→∞
1
n ∑ Xk − µ ≥ ε = 0
k =1

for all ε > 0.

Proof Because the random variables { Xk }k∈Z+ all have the same expectation and
same standard deviation, by 12.37 there exist µ ∈ R and s ∈ [0, ∞) such that
EXk = µ and σ ( Xk ) = s
for all k ∈ Z+. Thus
1 n  1 n  1 2 n  s2
12.39 E ∑ Xk = µ and σ2 ∑ Xk = σ ∑ X k = ,
n k =1 n k =1
n2 k =1
n
where the last equality follows from 12.22 (this is where we use the independent part
of the hypothesis).
Now suppose ε > 0. In the special case where s = 0, all the Xk are almost surely
equal to the same constant
√ function and the desired result clearly holds. Thus we
assume s > 0. Let t = nε/s and apply Chebyshev’s inequality (12.21) with this
value of t to the random variable n1 ∑nk=1 Xk , using 12.39 to get
 1 n   s2

P ∑
n k =1
Xk − µ ≥ ε ≤ 2 .

Taking the limit as n → ∞ of both sides of the inequality above gives the desired
result.
398 Chapter 12 Probability Measures

EXERCISES 12
1 Suppose (Ω, F , P) is a probability space and A ∈ F . Prove that A and Ω \ A
are independent if and only if P( A) = 0 or P( A) = 1.
2 Suppose P is Lebesgue measure on [0, 1]. Give an example of two disjoint Borel
subsets sets A and B of [0, 1] such that P( A) = P( B) = 21 , [0, 21 ] and A are
independent, and [0, 12 ] and B are independent.
3 Suppose (Ω, F , P) is a probability space and A, B ∈ F . Prove that the follow-
ing are equivalent:
• A and B are independent events.
• A and Ω \ B are independent events.
• Ω \ A and B are independent events.
• Ω \ A and Ω \ B are independent events.

4 Suppose (Ω, F , P) is a probability space and { Ak }k∈Γ is a family of events.


Prove the family { Ak }k∈Γ is independent if and only if the family {Ω \ Ak }k∈Γ
is independent.
5 Give an example of a probability space (Ω, F , P) and events A, B1 , B2 such
that A and B1 are independent, A and B2 are independent, but A and B1 ∪ B2
are not independent.
6 Give an example of a probability space (Ω, F , P) and events A1 , A2 , A3 such
that A1 and A2 are independent, A1 and A3 are independent, and A2 and A3
are independent, but the family A1 , A2 , A3 is not independent.
7 Suppose (Ω, F , P) is a probability space, A ∈ F , and B1 ⊂ B2 ⊂ · · · is an
increasing sequence of events such that A and Bn are independent events for
S
each n ∈ Z+. Show that A and ∞ n=1 Bn are independent.

8 Suppose (Ω, F , P) is a probability space and { At }t∈R is an independent family


of events such that P( At ) < 1 for each t ∈ R. Prove that there exists a sequence
T
t1 , t2 , . . . in R such that P ∞n=1 Atn = 0.

9 Suppose (Ω, F , P) is a probability space and B1 , . . . , Bn ∈ F are such that


P( B1 ∩ · · · ∩ Bn ) > 0. Prove that

P( A ∩ B1 ∩ · · · ∩ Bn ) = P( B1 ) · PB1 ( B2 ) · · · PB1 ∩···∩ Bn−1 ( Bn ) · PB1 ∩···∩ Bn ( A)

for every event A ∈ F .


10 Suppose (Ω, F , P) is a probability space and A ∈ F is an event such that
0 < P( A) < 1. Prove that

P( B) = PA ( B) · P( A) + PΩ\ A ( B) · P(Ω \ A)

for every event B ∈ F .


Chapter 12 Probability Measures 399

11 Give an example of a probability space (Ω, F , P) and X, Y ∈ L2 ( P) such


that σ2 ( X + Y ) = σ2 ( X ) + σ2 (Y ) but X and Y are not independent random
variables.
12 Prove that if X and Y are random variables (possibly on two different probability
spaces) and X̃ = Ỹ, then PX = PY .
13 Suppose H : R → (0, 1) is a continuous one-to-one function satisfying condi-
tions (a) through (d) of 12.29. Show that the function X : (0, 1) → R produced
in the proof of 12.29 is the inverse function of H.
14 Suppose (Ω, F , P) is a probability space and X is a random variable. Prove
that the following are equivalent:
• X̃ is a continuous function on R.
• X̃ is a uniformly continuous function on R.
• P( X = t) = 0 for every t ∈ R.
• ( X̃ ◦ X )˜(s) = s for all s ∈ R.
(
0 if x < 0,
15 Suppose α > 0 and h( x ) = 2 −
α xe αx if x ≥ 0.
Let P = h dλ and let X be the random variable defined by X ( x ) = x for x ∈ R.
R∞
(a) Verify that −∞ h dλ = 1.
(b) Find a formula for the distribution function X̃.
(c) Find a formula (in terms of α) for EX.
(d) Find a formula (in terms of α) for σ( X ).

16 Suppose B isthe σ-algebra of Borel subsets of [0, 1) and P is Lebesgue measure


on [0, 1], B . Let {ek }k∈Z+ be the family of functions defined by the fourth
bullet point of Example 8.51 (notice that k = 0 is excluded). Show that the
family {ek }k∈Z+ is an i.i.d.
17 Suppose B is the σ-algebra of Borel subsets of (−π, π ] and P is Lebesgue
measure on (−π, π ], B divided by 2π. Let {ek }k∈Z\{0} be the family of
trigonometric functions defined by the third bullet point of Example 8.51 (notice
that k = 0 is excluded).
(a) Show that {ek }k∈Z\{0} is not an independent family of random variables.
(b) Show that {ek }k∈Z\{0} is an identically distributed family.
Photo Credits

• page iv: photos by Carrie Heeter and Sheldon Axler


• page 1: digital sculpture created by Doris Fiebig; all rights reserved
• page 13: photo by Petar Milošević; Creative Commons Attribution-Share Alike
4.0 International license; verified on 9 July 2019 at https://
commons.wikimedia.org/wiki/File:Mausoleum_of_Galla_Placidia_ceiling_mosaics.jpg

• page 19: photo by Fagairolles 34; Creative Commons Attribution-Share Alike


4.0 International license; verified on 9 July 2019 at https://
commons.wikimedia.org/wiki/File:Saint-Affrique_pont.JPG
• page 44: photo by Sheldon Axler
• page 51: photo by James Mitchell; Creative Commons Attribution-Share Alike
2.0 Generic license; verified on 9 July 2019 at https://
commons.wikimedia.org/wiki/File:Beauvais_Cathedral_SE_exterior.jpg
• page 68: photo by A. Savin; Creative Commons Attribution-Share Alike 3.0
Unported license; verified on 9 July 2019 at https://
commons.wikimedia.org/wiki/File:Moscow_05-2012_Mokhovaya_05.jpg
• page 73: photo by Giovanni Dall’Orto; permission to use verified on 11 July
2019 at https://en.wikipedia.org/wiki/Maria_Gaetana_Agnesi
• page 101: photo by Rafa Esteve; Creative Commons Attribution-Share Alike 4.0
International license; verified on 9 July 2019 at https://
commons.wikimedia.org/wiki/File:Trinity_College_-_Great_Court_02.jpg
• page 102: photo by A. Savin; Creative Commons Attribution-Share Alike 3.0
Unported license; verified on 9 July 2019 at https://
commons.wikimedia.org/wiki/File:Spb_06-2012_University_Embankment_01.jpg

• page 116: photo by Lucarelli; Creative Commons Attribution-Share Alike 3.0


Unported license; verified on 9 July 2019 at https://
commons.wikimedia.org/wiki/File:Palazzo_Carovana_Pisa.jpg
• page 146: photo by Petar Milošević; Creative Commons Attribution-Share Alike
3.0 Unported license; verified on 9 July 2019 at https://
en.wikipedia.org/wiki/Lviv
• page 152: photo by NonOmnisMoriar; Creative Commons Attribution-Share
Alike 3.0 Unported license; verified on 9 July 2019 at https://
commons.wikimedia.org/wiki/File:Ecole_polytechnique_former_building_entrance.jpg
400
Photo Credits 401

• page 193: photo by Roland zh; Creative Commons Attribution-Share Alike 3.0
Unported license; verified on 9 July 2019 at https://
commons.wikimedia.org/wiki/File:ETH_Zürich_-_Polyterasse_2011-04-11_19-03-06_ShiftN.jpg

• page 211: photo by Daniel Schwen; Creative Commons Attribution-Share Alike


2.5 Generic license; verified on 9 July 2019 at https://
commons.wikimedia.org/wiki/File:Mathematik_Göttingen.jpg
• page 255: photo by Hubertl; Creative Commons Attribution-Share Alike 4.0
International license; verified on 11 July 2019 at https://
en.wikipedia.org/wiki/University_of_Vienna
• page 280: copyright Per Enström; Creative Commons Attribution-Share Alike
4.0 International license; verified on 9 July 2019 at https://
commons.wikimedia.org/wiki/File:Botaniska_trädgården,_Uppsala_II.jpg
• page 339: photo by Ricardo Liberato, Creative Commons Attribution-Share
Alike 2.0 Generic license; verified on 9 July 2019 at https://
commons.wikimedia.org/wiki/File:All_Gizah_Pyramids.jpg

• page 380: photo by Alexander Dreyer, Creative Commons Attribution-Share


Alike 3.0 Unported license; verified on 9 July 2019 at https://
commons.wikimedia.org/wiki/File:Dices6-5.png
• page 391: photo by Simon Harriyott, Creative Commons Attribution 2.0
Generic license; verified on 9 July 2019 at https://
commons.wikimedia.org/wiki/File:Bayes_tab.jpg
Bibliography

The chapters of this book on linear operators on Hilbert spaces, Fourier analysis, and
probability only introduce these huge subjects, all of which have a vast literature.
This bibliography gives book recommendations for readers who want to go into these
topics in more depth.

• Sheldon Axler, Linear Algebra Done Right, third edition, Springer, 2015.
• Leo Breiman, Probability, Society for Industrial and Applied Mathematics,
1992.
• John B. Conway, A Course in Functional Analysis, second edition, Springer,
1990.
• Ronald G. Douglas, Banach Algebra Techniques in Operator Theory, second
edition, Springer, 1998.
• Rick Durrett, Probability: Theory and Examples, fifth edition, Cambridge
University Press, 2019.
• Paul R. Halmos, A Hilbert Space Problem Book, second edition, Springer,
1982.
• Yitzhak Katznelson, An Introduction to Harmonic Analysis, second edition,
Dover, 1976.
• T. W. Körner, Fourier Analysis, Cambridge University Press, 1988.
• Michael Reed and Barry Simon, Functional Analysis, Academic Press, 1980.
• Walter Rudin, Functional Analysis, second edition, McGraw–Hill, 1991.
• Barry Simon, A Comprehensive Course in Analysis, American Mathematical
Society, 2015.
• Elias M. Stein and Rami Shakarchi, Fourier Analysis, Princeton University
Press, 2003.

402
Notation Index

\, 19 dν, 260

1 A , 382 dµ , 274
3 ∗ I, 103
E, 149
| A|, 14 [ E] a , 118
Rb
f , 5, 95 [ E]b , 118
R ab R
a f ( x ) dx, 6 E f dµ, 88
a + bi, 155 EX, 385

B , 51 F, 159
B( f , r ), 148 F , 376
B( f , r ), 148 k f k, 163, 214
Bn , 137 f˜, 202, 350
Bn , 141 f + , 81
B(V ), 286 f − , 81
B(V, W ), 167 k f k1 , 95, 97
B( x, δ), 136 f −1 ( A), 29
[ f ] a , 119
C, 56 [ f ]b , 119
C, 155 h f , gi, 212
c0 , 177, 209 fˆ(n), 342
χ E , 31 fˆ(t), 363
Cn , 160 f I , 115
C(V ), 312 f [k] , 350
R
f dµ, 74, 81, 156
D, 253, 340 n
F , 160
∂D, 340
k f k p , 194
d, 75
k f˜k p , 203
D1 f , D2 f , 142
k f k∞ , 194
∆, 348
F X , 160
dim, 321
distance( f , U ), 224 g0 , 110
403
404 Notation Index

∑∞k=1 gk , 166 |ν|, 259


graph( T ), 179 kνk, 263
R
g( x ) dµ( x ), 125 νa , 271
null T, 172
H, 370 ν− , 269
h dµ, 258 ν  µ, 270
h∗ , 104 ν ⊥ µ, 268
ν+ , 269
I, 231
νs , 271
IK , 282
Im z, 155 p0 , 196
inf, 2 PB , 390
A
Pr , 346
K ∗ , 282
Pr , 345
∑k∈Γ f k , 239
p( T ), 297
`1 , 96 PU , 227
L1 (∂D), 352 PX , 392
L1 ( µ ), 95, 156, 161 Py , 370
L1 ( R ), 97 Py , 371
`2 ( Γ ), 237
Q, 15
Λ, 58
λ, 63 R, 2
λn , 139 Re z, 155
L( f , [ a, b]), 4 Rn , 136
L( f , P), 74
L( f , P, [ a, b]), 2 σ, 341
`( I ), 14 σ( X ), 388
`∞ , 177, 195 sn ( T ), 333
`p, 195 span{ek }k∈Γ , 174
L p (∂D), 341 sp( T ), 294
L p ( E ), 203 S ⊗ T , 117
L p (µ), 194 sup, 2
A
L p (µ), 202
k T k, 167
MF (S), 262 T ∗ , 281
Mh , 281 T −1 , 287
µ × ν, 127 t + A, 16
Notation Index 405

tA, 23
TC , 309
T k , 288

U ⊥ , 229
U f , 133
U ( f , [ a, b]), 4
U ( f , P, [ a, b]), 2

V 0 , 180
V 00 , 183
V , 286
VC , 234

X̃, 392
k( x1 , . . . , xn )k∞ , 136
R R
X Y f ( x, y ) dν ( y ) dµ ( x ), 126

Z, 5
|z|, 155
z, 158
Z (µ), 202
Z+ , 5
Index

Abel summation, 344 Bergman space, 253


absolute value, 155 Bessel’s inequality, 241
absolutely continuous, 270 Bessel, Friedrich, 242
additivity of outer measure, 47–48, 50, Borel measurable function, 33
60 Borel set, 28, 36, 137, 252
adjoint of a linear map, 281 Borel, Émile, 19
a. e., 90 Borel–Cantelli Lemma, 383
Agnesi, Maria Gaetana, 73 bounded, 136
algebra, 120 linear functional, 173
algebraic multiplicity, 324 linear map, 167
almost every, 90 bounded below, 192
almost surely, 382 Bounded Convergence Theorem, 89
Apollonius’s identity, 223 Bounded Inverse Theorem, 188
approximate eigenvalue, 310 Bunyakovsky, Viktor, 219
approximate identity, 346
approximation of Borel set, 48, 52 Cantor function, 58
approximation of function Cantor set, 56, 278
by continuous functions, 98 Carleson, Lennart, 352
by dilations, 100 Cauchy sequence, 151, 165
by simple functions, 96 Cauchy, Augustin-Louis, 152
by step functions, 97 Cauchy–Schwarz inequality, 218
by translations, 100 chain, 176
area paradox, 24 characteristic function, 31
area under graph, 134 Chebyshev’s inequality, 106, 389
average, 386 Chebyshev, Pafnuty, 106
Axiom of Choice, 22–23, 176, 183, Clarkson’s inequalities, 209
236 Clarkson, James, 209
closed, 136
bad Borel set, 113 ball, 148
Baire Category Theorem, 184 half-space, 235
Baire’s Theorem, 184–185 set, 149
Baire, René-Louis, 184 Closed Graph Theorem, 189
Banach space, 165 closure, 149
Banach, Stefan, 146 compact operator, 312
Banach–Steinhaus Theorem, 189, 192 complete metric space, 151
basis, 174 complex conjugate, 158
Bayes’ Theorem, 391 complex inner product space, 212
Bayes, Thomas, 391 complex measure, 256
Bengal, iv complex number, 155
406
Index 407

complexification, 234, 309 exponent, 196


conditional probability, 390 of ` p , 207
conjugate symmetry, 212 of L p (µ), 275
conjugate transpose, 283 space, 180
continuous function, 150
continuous linear map, 169 École Polytechnique Paris, 152, 371
continuously differentiable, 350 Egorov’s Theorem, 63
convex set, 225 Egorov, Dmitri, 63, 68
convolution Eiffel Tower, 373
on ∂D, 357 eigenvalue, 294
on R, 368 eigenvector, 294
countable additivity, 25 Einstein, Albert, 193
countable subadditivity, 17, 43 essential supremum, 194
countably additive, 256 ETH Zürich, 193
Courant, Richard, 211 Euler, Leonhard, 155, 356
cross section event, 381
of function, 119 expectation, 385
of rectangle, 118 exponential density, 395
of set, 118 Extension Lemma, 178

family, 174
Darboux, Gaston, 6
Fatou’s Lemma, 86
Dedekind, Richard, 211
Fermat, Pierre de, 380
dense, 184
finite measure, 123
density function, 394
finite subadditivity, 17
density of a set, 112
finite subcover, 18
derivative, 110
finite-dimensional, 174
differentiable, 110 Fourier coefficient, 342
dilation, 60, 140 Fourier Inversion Formula, 374
Dini’s Theorem, 71 Fourier series, 342
Dirac measure, 84 Fourier transform
Dirichlet kernel, 353 on L1 (R), 363
Dirichlet problem, 348 on L2 (R), 376
on half-plane, 372 Fourier, Joseph, 339, 371, 373
on unit disk, 349 Fredholm Alternative, 319
Dirichlet space, 254 Fredholm, Erik, 280
Dirichlet, Gustav Lejeune, 211 Fubini’s Theorem, 132
discontinuous linear functional, 177 Fubini, Guido, 116
distance Fundamental Theorem of Calculus,
from point to set, 224 110
to closed convex set in Hilbert
space, 226 Gauss, Carl Friedrich, 211
to closed subspace of Banach geometric multiplicity, 318
space not attained, 225 Giza pyramids, 339
distribution function, 392 Gödel, Kurt, 179
Dominated Convergence Theorem, 92 Gram, Jørgen, 247
dual Gram–Schmidt process, 246
double, 183 graph, 135, 179
408 Index

Hahn Decomposition Theorem, 267 invariant subspace, 327


Hahn, Hans, 179 inverse image, 29
Hahn–Banach Theorem, 179, 236 invertible, 287
Hamel basis, 175 isometry, 305
Hardy, Godfrey H., 101 iterated integral, 126, 129
Hardy–Littlewood maximal function,
104 Jordan Decomposition Theorem, 269
Hardy–Littlewood maximal inequality, Jordan, Camille, 269
105 jump discontinuity, 352
best constant, 107
harmonic function, 348 kernel, 172
Heine, Eduard, 19 Kolmogorov, Andrey, 352
Heine–Borel Theorem, 19
L1 -norm, 95
Hilbert space, 224 Laplacian, 348
Hilbert, David, 211 Lebesgue Decomposition Theorem,
Hölder, Otto, 197 271
Hölder’s inequality, 196 Lebesgue Density Theorem, 113
Lebesgue Differentiation Theorem,
i.i.d., 396
108, 111, 112
identically distributed, 396
Lebesgue measurable function, 69
identity map, 231
Lebesgue measurable set, 52
imaginary part, 155 Lebesgue measure
increasing function, 34 on R, 51, 55
independent events, 383 on Rn , 139
independent random variables, 386 Lebesgue space, 95, 194
indicator function, 382 Lebesgue, Henri, 40, 87
infinite sum in normed vector space, left invertible, 289
166 left shift, 289, 294
infinite-dimensional, 174 length of open interval, 14
inner measure, 61 lim inf, 86
inner product, 212 limit, 149, 165
inner product space, 212 linear functional, 172
integral linear map, 167
iterated, 126 linearly independent, 174
of characteristic function, 75 Littlewood, John, 101
of complex-valued function, 156 lower Lebesgue sum, 74
of linear combination of charac- lower Riemann integral, 4, 85
teristic functions, 80 lower Riemann sum, 2
of nonnegative function, 74 Luzin’s Theorem, 66, 68
of real-valued function, 81 Luzin, Nikolai, 66, 68
of simple function, 76 Lviv, 146
on a subset, 88 Lwów, 146
on small sets, 90
with respect to counting measure, Markov’s inequality, 102
76 Markov, Andrei, 102, 106
integration operator, 282, 314 maximal element, 175
interior, 184 mean value, 386
Index 409

measurable projection, 227, 232, 247


function, 31, 37, 156 orthonormal
rectangle, 117 basis, 244
set, 27, 55 family, 237
space, 27 sequence, 237
measure, 41 subset, 249
of decreasing intersection, 44, 46 outer measure, 14
of increasing union, 43
of intersection of two sets, 45 parallelogram equality, 220
of union of three sets, 46 Parseval’s identity, 244
of union of two sets, 45 Parseval, Marc-Antoine, 245
measure space, 42 partial derivative, 142
Melas, Antonios, 107 partial derivatives, order does not mat-
metric, 147 ter, 143
metric space, 147 partial isometry, 311
Minkowski’s inequality, 199 partition, 2
Minkowski, Hermann, 193, 211 Pascal, Blaise, 380
monotone class, 122 photo credits, 400–401
Monotone Class Theorem, 122 Plancherel’s Theorem, 375
Monotone Convergence Theorem, 78 p-norm, 194
Moon, 44 pointwise convergence, 62
Moscow State University, 68 Poisson integral
multiplication operator, 281 on unit disk, 349
on upper half-plane, 372
Napoleon, 339, 371
Poisson kernel, 346, 370
Nikodym, Otto, 272
Poisson, Siméon, 371, 373
Noether, Emmy, 211
positive measure, 41, 256
norm, 163
Principle of Uniform Boundedness,
coming from inner product, 214
190
normal, 302
probability distribution, 392
normed vector space, 163
null space, 172 probability measure, 381
of T ∗ , 285 probability of a set, 381
probability space, 381
open product
ball, 148 of Banach spaces, 188
cover, 18 of Borel sets, 138
cube, 136 of measures, 127
set, 136, 137, 148 of open sets, 137
unit ball, 141 of scalar and vector, 159
unit disk, 253, 340 of σ-algebras, 117
Open Mapping Theorem, 186 Pythagorean Theorem, 217
operator, 286
orthogonal Rademacher, Hans, 238
complement, 229 Radon, Johann, 255
decomposition, 217, 231 Radon–Nikodym derivative, 274
elements of inner product space, Radon–Nikodym Theorem, 272
216 Ramanujan, Srinivasa, 101
410 Index

random variable, 385 step function, 97


range St. Petersburg University, 102
dense, 285 subspace, 160
of T ∗ , 285 dense, 230
real inner product space, 212 sum, unordered, 239
real measure, 256 Supreme Court, 216
real part, 155 Swiss Federal Institute of Technology,
rectangle, 117 193
reflexive, 210
region under graph, 133 Tonelli’s Theorem, 129, 131
Riemann integrable, 5 Tonelli, Leonida , 116
Riemann integral, 5, 93 total variation measure, 259
Riemann sum, 2 total variation norm, 263
translation of set, 16, 60
Riemann, Bernhard, 1, 211
triangle inequality, 163, 219
Riemann–Lebesgue Lemma, 344, 364
Trinity College, Cambridge, 101
Riesz Representation Theorem, 233,
two-sided ideal, 313
250
Riesz, Frigyes, 234 unbounded linear functional, 177
right invertible, 289 uniform convergence, 62
right shift, 289, 294 uniform density, 395
unit circle, 340
sample space, 381 unit disk, 253, 340
scalar multiplication, 159 unitary, 305
Schmidt, Erhard, 247 University of Göttingen, 211
Schwarz, Hermann, 219 University of Vienna, 255
Scuola Normale Superiore di Pisa, 116 upper half-plane, 370
self-adjoint, 299 upper Riemann integral, 4
separable, 183, 245 upper Riemann sum, 2
Shakespeare, William, 70 Uppsala University, 280
σ-algebra, 26
smallest, 28 variance, 388
σ-finite measure, 123 vector space, 159
signed measure, 257 Vitali Covering Lemma, 103
Simon, Barry, 69 Vitali, Giuseppe, 13
simple function, 65 Volterra operator, 286, 315, 320, 331,
singular measures, 268 334
Singular Value Decomposition, 332 Volterra, Vito, 286
singular values, 333 volume of unit ball, 141
span, 174 von Neumann, John, 272
S -partition, 74
Weak Law of Large Numbers, 397
Spectral Mapping Theorem, 298
Wirtinger’s inequality, 362
Spectral Theorem, 329–330
spectrum, 294 Young’s inequality, 196
standard deviation, 388 Young, William Henry, 196
standard normal density, 395
standard representation, 79 Zorn’s Lemma, 176, 183, 236
Steinhaus, Hugo, 189 Zorn, Max, 176
Colophon:
Notes on Typesetting

• This book was typeset in pdfLATEX by the author, who wrote the LATEX code to
implement the book’s design. The pdfLATEX software was developed by Hàn Thế
Thành.
• The LATEX software used for this book was written by Leslie Lamport. The TEX
software, which forms the base for LATEX, was written by Donald Knuth.
• The main text font in this book is Nimbus Roman No. 9 L, created by URW as a
legal clone of Times, which was designed for the British newspaper The Times
in 1931.
• The math fonts in this book are various versions of Pazo Math, URW Palladio,
and Computer Modern. Pazo Math was created by Diego Puga; URW Palladio
is a legal clone of Palatino, which was created by Hermann Zapf; Computer
Modern was created by Donald Knuth.
• The san serif font used for chapter titles, section titles, and subsection titles is
NimbusSanL.
• The figures in the book were produced by Mathematica, using Mathematica
code written by the author. Mathematica was created by Stephen Wolfram.
• The Mathematica package MaTeX, written by Szabolcs Horvát, was used to
place LATEX-generated labels in the Mathematica figures.
• The LATEX package graphicx, written by David Carlisle and Sebastian Rahtz,
was used to integrate into the manuscript photos and the figures produced by
Mathematica.
• The LATEX package multicol, written by Frank Mittelbach, was used to get around
LATEX’s limitation that two-column format must start on a new page (needed for
the Notation Index and the Index).
• The LATEX packages Tik Z, written by Till Tantau, and tcolorbox, written by
Thomas Sturm, were used to produce the definition boxes and result boxes.
• The LATEX package color, written by David Carlisle, was used to add appropriate
color to various design elements, such as used on the first page of each chapter.
• The LATEX package wrapfig, written by Donald Arseneau, was used to wrap text
around the comment boxes.
• The LATEX package microtype, written by Robert Schlicht, was used to reduce
hyphenation and produce more pleasing right justification.
411

You might also like