Skip to article frontmatterSkip to article content

8.2 Markov Processes

transitioning with short-term memory

Dept. of Electrical and Systems Engineering
University of Pennsylvania

Binder

Lecture notes

1Reading

Material related to this page, as well as additional exercises, can be found in Section 4.9 and Chapter 10 of LAA 5th5^{th} edition, and ALA 9.3.

2Learning Objectives

By the end of this page, you should know:

  • what is a Markov chain model using a weather prediction example
  • what is a probability vector and a transition matrix (regular)
  • when does a Markov chain converge to a unique probability vector
  • how the eigenvalue and eigenvector of the transition matrix relates to the convergence of the Markov chain

3Weather Prediction: Introduction

We will spend this section on Markov Chains, which are a widely used linear iterative model to describe a wide variety of situations in biology, business, chemistry, engineering, physics, and elsewhere.

In each case, the model is used to describe an experiment or measurement that is performed many times in the same way. The outcome of an experiment can be one of several known possible outcomes, and importantly, the outcome of one experiment depends only on the experiment conducted immediately before it. Before introducing a formal model for Markov chains, let’s look at an example.

4Convergence in Markov Chains

Let’s try to understand why the convergence in Example 1 happens, and then we’ll look at some interesting applications of Markov chains.

Our starting point is a general definition of a probability vector.

In general, a Markov chain is given by the first order linear iterative system

x(k+1)=Px(k)(MC)\vv x(k+1) = P \vv x(k) \quad (\text{MC})

whose initial state x(0)\vv x(0) is a probability vector. The entries of the transition matrix PP must satisfy

0pij1 and p1j++pnj=1.(TM)0 \leq p_{ij} \leq 1 \ \text{and} \ p_{1j} + \cdots + p_{nj} = 1. \quad (\text{TM})

for all i,j=1,,ni,j=1,\ldots,n. The entry pijp_{ij} is the transition probability that the system will switch from state jj to state ii. Because this covers all possible transitions, this means each column sums to 1. Under these conditions, we can guarantee that if x(k)\vv x(k) is a probability vector, so is x(k+1)=Px(k)\vv x(k+1) = P\vv x(k). To see this, note that 1P=[1p11pn]=[11]=1\vv 1^{\top} P = \bm \vv 1^{\top} \vv p_1 & \cdots & 1^{\top} \vv p_n\em = \bm 1 & \cdots & 1 \em = \vv 1^{\top} so that 1x(k+1)=1Px(k)=1x(k)=1\vv 1^{\top} \vv x(k+1) = \vv 1^{\top} P\vv x(k) = \vv 1^{\top} \vv x(k) = 1. That x(k+1)\vv x(k+1) is entrywise non-negative follows from PP and x(k)\vv x(k) being entry-wise non-negative.

Next, let’s investigate convergence properties. We first need to impose a very mild technical condition on the transition matrix PP, namely we assume that it is regular.

The long-term behavior of a Markov chain with regular transition matrix PP is governed by the Perron-Frobenius theorem, which we state next. The proof is quite involved, so we won’t cover it, but if you’re curious, check out the end of ALA 9.3.

This is a very exciting development! It tells us that we can understand the long-term behavior of a regular Markov chain by solving for the eigenvector x\vv x^* associated with the eigenvalue λ1=1\lambda_1=1 of PP.

Returning to our weather prediction example, we compute the steady state probability vector x\vv x^* by just solving (PI)v=0(P-I)\vv v=\vv 0:

(PI)v=[.3.2.3.2][v1v2]=0=>v1=23v2v=[231] (P-I)\vv v = \bm -.3 & .2 \\ .3 & -.2 \em \bm v_1 & v_2 \em = \vv 0 => v_1 = \frac{2}{3} v_2 \Rightarrow \vv v = \bm \frac{2}{3} \\ 1 \em

and then normalizing v\vv v so that its entries add up to 1:

x=11+23[231]=[2535]=[0.40.6] \vv x^* = \frac{1}{1+\frac{2}{3}} \bm \frac{2}{3} \\ 1 \em = \bm \frac{2}{5} \\ \frac{3}{5} \em = \bm 0.4 \\ 0.6 \em

This special eigenvector x\vv x^* tells us that no matter the initial state x(0)\vv x(0), the long term behavior is that we are in State 1 (sunny) 40% of days and State 2 (cloudy) 60% of days.

Binder