We will spend this section on Markov Chains, which are a widely used linear iterative model to describe a wide variety of situations in biology, business, chemistry, engineering, physics, and elsewhere.
In each case, the model is used to describe an experiment or measurement that is performed many times in the same way. The outcome of an experiment can be one of several known possible outcomes, and importantly, the outcome of one experiment depends only on the experiment conducted immediately before it. Before introducing a formal model for Markov chains, let’s look at an example.
for all i,j=1,…,n. The entry pij is the transition probability that the system will switch from state j to state i. Because this covers all possible transitions, this means each column sums to 1. Under these conditions, we can guarantee that if x(k) is a probability vector, so is x(k+1)=Px(k). To see this, note that 1⊤P=[1⊤p1⋯1⊤pn]=[1⋯1]=1⊤ so that 1⊤x(k+1)=1⊤Px(k)=1⊤x(k)=1. That x(k+1) is entrywise non-negative follows from P and x(k) being entry-wise non-negative.
Next, let’s investigate convergence properties. We first need to impose a very mild technical condition on the transition matrix P, namely we assume that it is regular.
The long-term behavior of a Markov chain with regular transition matrix P is governed by the Perron-Frobenius theorem, which we state next. The proof is quite involved, so we won’t cover it, but if you’re curious, check out the end of ALA 9.3.
This is a very exciting development! It tells us that we can understand the long-term behavior of a regular Markov chain by solving for the eigenvector x∗ associated with the eigenvalue λ1=1 of P.
Returning to our weather prediction example, we compute the steady state probability vector x∗ by just solving (P−I)v=0:
This special eigenvector x∗ tells us that no matter the initial statex(0), the long term behavior is that we are in State 1 (sunny) 40% of days and State 2 (cloudy) 60% of days.