Casarotto
Casarotto
Casarotto
CHAD CASAROTTO
Abstract. This paper will explore the basics of discrete-time Markov chains
used to prove the Ergodic Theorem. Definitions and basic theorems will allow
us to prove the Ergodic Theorem without any prior knowledge of Markov
chains, although some knowledge about Markov chains will allow the reader
better insight about the intuitions behind the provided theorems. Even for
those familiar with Markov chains, the provided definitions will be important
in providing the uses for the various notations used in this paper.
Contents
1. Basic Definitions and Properties of Markov Chains 1
2. Stopping Times and the Strong Markov Property 3
3. Recurrence and Transience 4
4. Communication Classes and Recurrence 5
5. The Strong Law of Large Numbers and the Ergodic Theorem 6
References 7
Ti = inf{n 1 | Xn = i}.
This leads to the natural definition of the length of the rth excursion to i as
(r) (r1) (r1)
(r) Ti Ti if Ti <
Si =
0 otherwise.
The following theorem shows how the Markov property holds at stopping times.
Theorem 2.3. Let T be a stopping time of (Xn )n0 which is Markov(, P ). Then
given T < and XT = i, (Xl )lT is Markov(i , P ) and independent of Xk ,
0 k < T.
Proof. First, we already have that (Xl )lT is Markov(i , P ) by Theorem 1.4, so
we just need to show the independence condition. Let the event A = {XT =
i0 , . . . , XT +n = in } and the event B be any event determined by X0 , . . . , XT . It is
important to notice that the event B {T = m} is determined by X0 , . . . , Xm . We
get that
Definition 3.4. Another intuitive and useful term is the return probability to i,
defined as
fi = Pi (Ti < ).
Lemma 3.5. Pi (Vi > r) = (fi )r for r = 0, 1, 2, . . . .
Proof. First, we know that our claim is necessarily true when r = 0. Thus, we can
(r)
use induction and the fact that if X0 = i then {Vi > r} = {Ti < } to conclude
that
(r+1)
Pi (Vi > r + 1) = Pi (Ti < )
(r) (r+1)
= Pi (Ti < and Si < )
(r+1) (r) (r)
= Pi (Si < | Ti < )Pi (Ti < )
= fi (fi )r = (fi )r+1
using Lemma 3.2, so our claim is true for all r.
MARKOV CHAINS AND THE ERGODIC THEOREM 5
Theorem 3.6. The following two cases hold and show that any state is either
recurrent or transient:
P
(1) if Pi (Ti < ) = 1, then i is recurrent and P n=0 Pi (Xn = i) = ;
(2) if Pi (Ti < ) < 1, then i is transient and n=0 Pi (Xn = i) < .
Proof. If Pi (Ti < ) = fi = 1 by Lemma 3.5, then
Pi (Vi = ) = lim Pi (Vi > r) = lim 1r = 1
r r
so i is recurrent and
X
Pi (Xn = i) = Ei (Vi ) = .
n=0
In the other case, fi = Pi (Ti < ) < 1 then using our fact about Vi
X
X n1
X X
Pi (Xn = i) = Ei (Vi ) = nPi (Vi = n) = Pi (Vi = n)
n=0 n=1 n=1 r=0
X
X X X 1
= Pi (Vi = n) = Pi (Vi > r) = (fi )r = <
r=0 n=r+1 r=0 r=0
1 fi
so Pi (Vi = ) = 0 and i is transient.
Thus, for some pii1 pi1 i2 pin1 j > 0 for some states i1 , i2 , . . . , in1 .
() Take some i1 , i2 , . . . , in1 such that
0 < pii1 pi1 i2 pin1 j Pi (Xn = j) Pi (Xn = j for some n 0).
Now it is clear from the proven inequality that i j, j k i k. Also, it is
true that i i for any state i and that i j j i. Thus, is an equivalence
relation on I.
Proof. Take any distinct pair of states i, j C and suppose that i is transient.
Then there exist n, m 0 such that Pi (Xn = j) > 0 and Pj (Xm = i) > 0, and for
all r 0
Pi (Xn+r+m = i) Pi (Xn = j)Pj (Xr = j)Pj (Xm = i).
This implies that
X 1 X
Pj (Xr = j) Pi (Xn+r+m = i) <
r=0
Pi (Xn = j)Pj (Xm = i) r=0
by Theorem 3.6. So any arbitrary j is transient, again by Theorem 3.6, so the
whole of C is transient. The only way for this not to be true is if all states in C
are recurrent.
This theorem shows us that recurrence and transience is a class property, and
we will refer to it in the future as such.
Theorem 4.5. Suppose P is irreducible and recurrent. Then for all i I we have
P (Ti < ) = 1.
Proof. By Theorem 2.3 we have
X
P (Ti < ) = Pj (Ti < )P (X0 = j)
jI
The Ergodic theorem examines the long-run value of the ratio Vi (n)/n, which is
the proportion of time spent in state i before step n.
Theorem 5.3 (Ergodic Theorem). Let P be irreducible and let be any distribu-
tion. If (Xn )0nN is Markov(, P ) then
Vi (n) 1
P as n = 1
n mi
where mi = Ei (Ti ), i.e. the expected return time to state i.
Proof. First, we will consider the case where P is transient. We then know that
with probability 1, Vi < , so
Vi (n) Vi 1
0= .
n n mi
Now, we will consider the case that P is recurrent, and we will fix a state i. Let
T = Ti , and we have Pi (T < ) = 1 by Theorem 3.6 and (Xl )lT is Markov(i , P )
and independent of Xk , 0 k < T by Theorem 2.3. Since the long-run proportion
of time spent in state i is the same for (Xl )lT as for (Xn )n0 , we need to only
consider the case where = i .
(1) (2)
By Lemma 3.2, we know that the non-negative, random variables Si , Si , . . .
(r)
are independent and identically distributed with Ei (Si ) = Ei (Ti ) = mi . We know
(Vi (n)1) (1) (Vi (n)1)
Ti = Si + . . . + Si n 1,
where the left-hand side of the inequality is the step of the last visit to i before step
n. In addition,
(V (n)) (1) (V (n))
Ti i = Si + . . . + Si i n,
where the left-hand side of the inequality is the step of the first visit to i after step
n 1. So we can squeeze the value of n/Vi (n) using the following inequality:
(1) (V (n)1) (1) (Vi (n))
Si+ . . . + Si i n S + . . . + Si
(5.4) i
(Vi (n)) Vi (n) (Vi (n))
Now we can use the Strong Law of Large Numbers to get
!
(1) (n)
Si + . . . + Si
P mi as n = 1,
n
and since P is recurrent,
P (Vi (n) as n ) = 1.
Thus, if we let n in (5.4), we get
n
P mi as n = 1,
Vi (n)
which implies that
Vi (n) 1
P as n = 1,
n mi
References
[1] Norris, James R. Markov Chains. University of Cambridge Press. 1998.