0% found this document useful (0 votes)
2 views5 pages

Feb-04 (1)

The document discusses the behavior of simple symmetric random walks on different dimensions, stating that they are recurrent in 1 and 2 dimensions but transient in 3 or more. It also defines stationary distributions in Markov chains, explaining their existence, uniqueness, and the conditions under which they can be found. Additionally, it introduces the concepts of positive and null recurrence for recurrent states in Markov chains.

Uploaded by

lkljclk
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
2 views5 pages

Feb-04 (1)

The document discusses the behavior of simple symmetric random walks on different dimensions, stating that they are recurrent in 1 and 2 dimensions but transient in 3 or more. It also defines stationary distributions in Markov chains, explaining their existence, uniqueness, and the conditions under which they can be found. Additionally, it introduces the concepts of positive and null recurrence for recurrent states in Markov chains.

Uploaded by

lkljclk
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd

This is equivalent to rotating the coordinate axes through an angle of ⇡/4.

First,
P(Yn+1 Yn = (1, 1)) = P(An+1 An = 1) = P(Xn+1 Xn = (1, 0)) = 1/4,
similarly for the other three possibilities for Yn+1 Yn , namely ( 1, 1), (1, 1), and ( 1, 1).
Therefore, Yn+1 Yn takes values in (1, 1), ( 1, 1), (1, 1), and ( 1, 1), all with proba-
bility 1/4. It follows that the coordinates of (Yn )n 0 , which are (Un )n 0 and (Vn )n 0 , are
independent symmetric random walks on Z. Also, Xn = (0, 0) if and only if Yn = (0, 0).

Simple Symmetric Random walk on Z3 . Consider S = Z3 and for any a, b, c 2 Z,


1
P(a,b,c),(a±1,b,c) = P(a,b,c),(a,b±1,c) = P(a,b,c),(a,b,c±1) = .
6
In this case,
X (2n)! ⇣ 1 ⌘2n
(P 2n )(0,0,0),(0,0,0) =
i!i!j!j!k!k! 6
i+j+k=n
⇣ 1 ⌘2n X (2n)!
=
6 (i!j!k!)2
i+j+k=n
⇣ 1 ⌘2n ✓2n◆ X ✓ n! ◆2
=
2 n 3n i!j!k!
i+j+k=n
⇣ 1 ⌘2n ✓2n◆ n n! o X n!
 max n :i+j+k =n n
2 n 3 i!j!k! 3 i!j!k!
i+j+k=n

⇣ 1 ⌘2n 2n ◆ n n! o
= max n :i+j+k =n
2 n 3 i!j!k!
⇣ 1 ⌘2n ✓2n◆ n!
= (assuming n is divisible by 3)
2 n 3n n ! 3
3
(2n)!
=  Cn 3/2 (by Stirling’s formula),
n n 3
12 n! 3 !
where C is a constant that does not depend on n. So, the return probabilities are summable.
Therefore, the MC is transient.
Remark 1.32. One can show that 2n-step return probability simple symmetric random walk on Zd
behaves approximately like Cd n d/2 .
To summarize, we have shown

Theorem 1.33 (Pólya, 1921). The simple symmetric random walk


on Zd is recurrent if d = 1, 2 and transient if d 3.
George Pólya 1887-1985

1.5 Stationary Distribution


Definition 1.34. Consider a MC with transition
P matrix P on state space S. Let ⇡ be probability
vector on S, i.e., ⇡(x) 0 for all x 2 S and x ⇡(x) = 1. We call ⇡ a stationary distribution if
⇡P = ⇡ (⇡ is a row vector).

17
That is, X
⇡(x)p(x, y) = ⇡(y) for all y 2 S. (11)
x

(a) Why is ⇡ called stationary distribution? If the initial distribution of X0 is ⇡, i.e., X0 ⇠ ⇡, then
Xn ⇠ ⇡P n = ⇡ for any n 1. That means X0 and Xn have the same marginal distribution
for all n.
Also, suppose that starting from an initial distribution µ, Xn converges in distribution for
some probability distribution ⇡ on S, which is equivalent to saying that Pµ (Xn = y) ! ⇡(y)
for each y. Then
⇡ = lim µP n = (lim µP n 1 )P = ⇡P,
n n

i.e., ⇡ must satisfy the stationary equation ⇡ = ⇡P .

(b) For finite state MC, the stationary distribution always exists, but it may not be unique. For
example, S = {0, 1}, P0,0 = 1 = P1,1 . Then any ⇡ = (↵, 1 ↵) for all ↵ 2 [0, 1] is a stationary
distribution.

(c) For infinite MC, stationary distribution may not even exist. Consider SSRW on Z. If ⇡P = ⇡,
then for all i 2 Z,
1
⇡(i 1) + ⇡(i + 1) = ⇡(i).
2
This implies ⇡(i + 1) ⇡(i) = ⇡(i) ⇡(i 1) for all i, i.e, ⇡ : Z ! R+ has constant slope, or
⇡(i) = c for all i. Then c = 0, a contradiction.

Example 1.35. Consider a MC on S = {1, 2} with transition matrix



1 ↵ ↵
P = ,
1

where ↵, 2 (0, 1). The stationary equations are

⇡(1)(1 ↵) + ⇡(2) = ⇡(1) (12)


⇡(1)↵ + ⇡(2)(1 ) = ⇡(2) (13)

(12) yields

↵⇡(1) = ⇡(2) ) ⇡(1) = c/↵ and ⇡(2) = c/ for some constant c.

Since ⇡(1) + ⇡(2) = 1, we obtain


✓ ◆
1 1 ↵
c + =1)c= .
↵ ↵+

Therefore, ✓ ◆

(⇡(1), ⇡(2)) = , .
↵+ ↵+

The following lemma (proof omitted) provides a situation where the stationary distribution exists
and is unique.

18
Lemma 1.36. An irreducible finite state MC has a unique stationary distribution.

Proof (existence part, under a stronger assumption p(x, y) > 0 for all x, y 2 S.)
Since (P I)1 = 0, rank(P I) < |S|. So, there exists ⇡ 6⌘ 0 such that

⇡(P I) = 0, i.e., ⇡P = ⇡.

Now we would like to show that the entries of ⇡ are either all non-negative or all non-positive. If
we can show that then by taking negative of ⇡ if necessary, we can assume that ⇡(x) 0 for all
x 2 S. Then we normalize ⇡ so that it becomes a probability vector, i.e., define
⇡(x)
¯ (x) = P
⇡ .
y2S ⇡(y)

Clearly, ⇡
¯P = ⇡
¯ , as desired. We now prove the claim. Suppose, if possible, ⇡ contains both positive
and negative entries. Then
X X X XX
|⇡(y)| = | ⇡(x)p(x, y)| < |⇡(x)|p(x, y)
y2S y2S x2S y2S x2S
X X X
= |⇡(x)| p(x, y) = |⇡(x)|,
x2S y2S x2S

which is a contradiction. The crucial strict inequality in the middle appears due to the hypotheses
that ⇡ has both positive and negative entries and the transition probabilities are strictly positive. 2

Consider an irreducible Markov chain on state space S = {1, 2, 3} with transition matrix

1 2 3
2 3
1 0.8 0.1 0.1
P = 2 4 0.2 0.6 0.2 5.
3 0.3 0.3 0.4

Let us compute the stationary distribution ⇡. We need to solve the equation ⇡P = ⇡ or ⇡(P I) = 0,
I being the identity matrix.
2 3
0.2 0.1 0.1
(⇡(1) ⇡(2) ⇡(3)) 4 0.2 0.4 0.2 5 = (0 0 0).
0.3 0.3 0.6

Writing out the linear equations, we have

0.2⇡(1) + 0.2⇡(2) + 0.3⇡(3) = 0


0.1⇡(1) 0.4⇡(2) + 0.3⇡(3) = 0
0.1⇡(1) + 0.2⇡(2) 0.6⇡(3) = 0.

These three equations are not ‘linearly independent’ as summing them yields 0 = 0. To get rid of
the redundancy, we remove one of them, say the third one to have
2 3
0.2 0.1
(⇡(1) ⇡(2) ⇡(3)) 4 0.2 0.45 = (0 0)
0.3 0.3

19
However, we have an additional equation in the form ⇡(1) + ⇡(2) + ⇡(3) = 1, which we add to the
above matrix equation as
2 3
0.2 0.1 1
(⇡(1) ⇡(2) ⇡(3)) 4 0.2 0.4 15 = (0 0 1).
0.3 0.3 1

Therefore, we conclude that


2 3 1 2 3 1
0.2 0.1 1 0.2 0.1 1
4
(⇡(1) ⇡(2) ⇡(3)) = (0 0 1) 0.2 0.4 15 4
= the last row of 0.2 0.4 15 .
0.3 0.3 1 0.3 0.3 1

Below we record this strategy as an algorithm for computing the stationary distribution of a finite
irreducible MC:

Step 1. Compute P I and replace the last column by all 1’s. Call the new matrix A.

Step 2. Compute A 1.

Step 3. ⇡ is the last row of A 1.

Finding stationary distributions beyond irreducibility in finite MC.

Lemma 1.37 (stationary probability vanishes on transient states). Let ⇡ be a stationary distribu-
tion of a (possibility infinite) MC. If x is transient, then ⇡(x) = 0.

Proof.
⇣ [
1 ⌘
⇡(x) = P⇡ (Xn = x)  P⇡ {Xm = x} .
m=n

Taking n ! 1, we have
⇣ [
1 ⌘ ⇣\
1 [
1 ⌘
⇡(x)  lim P⇡ {Xm = x} = P⇡ {Xm = x} = P⇡ (Xn = x infinitely often) = 0.
n
m=n n=1 m=n

t
u
The above lemma shows that a stationary distribution is always supported only on recurrent
states. This rules out the possibility of existence of stationary distribution for the biased (p 6= 1/2)
simple random walk on Z as in that case every state is transient.
We now describe, through the following example, how to find all stationary distributions of a
finite MC with more than one recurrent class (along with some transient classes).
Consider a MC on S = {1, 2, 3, 4, 5} with the transition diagram. It has two recurrent classes
(closed, irreducible sets) {1, 2} and {4, 5} whereas 3 is a transient state. The transition matrix is
given by
Note that the 2 ⇥ 2 submatrices P1 and P2 are themselves irreducible transition matrices on
{1, 2} and {4, 5}. By Lemma 1.37, ⇡(3) = 0. The stationary equation becomes

(⇡(1) ⇡(2) 0 ⇡(4) ⇡(5))P = (⇡(1) ⇡(2) 0 ⇡(4) ⇡(5)).

20
This implies that

(⇡(1) ⇡(2))P1 = (⇡(1) ⇡(2)) and (⇡(4) ⇡(5))P2 = (⇡(4) ⇡(5)).

By Example 1.35,
✓ ◆ ✓ ◆
1 0.6 3 5
(⇡(1) ⇡(2)) = = and
0.6 + 1 0.6 + 1 8 8
✓ ◆ ✓ ◆
0.3 0.8 3 8
(⇡(4) ⇡(5)) = = .
0.8 + 0.3 0.8 + 0.3 11 11
Note that both (⇡(1) ⇡(2) 0 0 0) and (0 0 0 ⇡(4) ⇡(5)) are stationary distributions of P .
Therefore, any convex combination of them ↵(⇡(1) ⇡(2) 0 0 0) + (1 ↵)(0 0 0 ⇡(4) ⇡(5)) for
any ↵ 2 [0, 1] is also a stationary distribution. So, for any ↵ 2 [0, 1],
✓ ◆
3↵ 5↵ 3(1 ↵) 8(1 ↵)
0
8 8 11 11

is a stationary distribution for P .

1.5.1 Positive and null recurrence

Definition 1.38. For a recurrent state y, we have Py (Ty < 1) = 1. If we also have Ey [Ty ] < 1,
then y is called a positive recurrent state. Otherwise, if Ey [Ty ] = 1 is called a null recurrent state.

21

You might also like