Skip to main content

Markov Chain Behavior

As we learn more about Markov chains (especially the infinite ones), we would like to find out the behavior of the chain. One of the interesting behavior is the stationary distribution.

Stationary Distribution

Sometimes we find there exists π\overline{\pi} satisfied πP=π\overline{\pi} P = \overline{\pi} where PP is the transition matrix. We define such π\overline{\pi} is called stationary distribution. From the equation, we not hard to conclude that once MC reach the distribution, it will never leave.

  • Assume πP=π\overline{\pi} P = \overline{\pi}, then we have πP=πPP=πPP=πP(n)=π\overline{\pi} P = \overline{\pi} P P = \overline{\pi} P \cdots P = \overline{\pi} P^{(n)} = \overline{\pi}.

We define a MC with probability distribution π\overline{\pi} satisfies πipij=πjpji,i,jS\pi_i p_{ij} = \pi_j p_{ji}, \forall i,j \in S is called reversible.

  • If a MC is reversible, then such π\overline{\pi} is the stationary distribution. (converse is not true)

But we actually may not have such stationary distribution when following conditions are satisfied:

  • i,jS,limnpij(n)=0\forall i,j \in S,\lim_{n \to \infty} p_{ij}^{(n)} = 0
  • MC is irreducible with some i,jSi,j\in S such that limnpij(n)=0\lim_{n \to \infty} p_{ij}^{(n)} = 0

Finite MC convergence to stationary distribution

Since PP is a transition matrix, we would like to use some linear algebra ways to find the limit behavior of the MC, i.e. limnP(n)\lim_{n \to \infty} P^{(n)}. We would like to use Jordan decomposition to represent PP where Q\exists Q such that P=QDQ1P = QDQ^{-1} where D=Q1PQD = Q^{-1}PQ. Then we have limnP(n)=limnQDnQ1\lim_{n \to \infty} P^{(n)} = \lim_{n \to \infty} QD^nQ^{-1}

And we can use Perron-Frobenius theorem to find the stationary distribution since PP is a stochastic matrix with all of entries are strictly positive, that is, all eigenvalues of PP has absolute value less than 1. The columns of QQ are right eigenvectors of PP and the rows of Q1Q^{-1} are left eigenvectors of PP.

Theorem 1

According to above, we have the fact/theorem: If PP is a transition matrix such that for some n,P(n)n, P^{(n)} has all entries strictly positive, then PP statisfies:

  • The left eigenvector can be chosen to have all nonnegative entries.
  • The eigenvalue 1 is simple and all other eigenvalues have absolute value less than 1.
  • The first row of Q1Q^{-1} is the stationary distribution.

Theorem 2

To find all entries strictly positive P(n)P^{(n)}, we can use the following theorem: If PP is irreducible and aperiodic, then M>0\exists M > 0 such that for all nM,P(n)n \geq M, P^{(n)} has all entries strictly positive.

But what if PP is not irreducible and aperiodic? We divide into following cases:

PP is not irreducible or aperiodic

Since PP is reducible with recurrent classes R1,,RrR_1, \ldots, R_r transient classes T1,,TsT_1, \ldots, T_s. Then we will have rr stationary distributions π1,,πr\pi_1, \ldots, \pi_r corresponding to recurrent classes R1,,RrR_1, \ldots, R_r. Let PkP_k be submatrix of PP corresponding to recurrent classes RkR_k. Then we have iRk,limnPij(n)=πk(j)jRk\forall i \in R_k, \lim_{n \to \infty} P^{(n)}_{ij} = \pi_k(j) \forall j \in R_k and limnPij(n)=0jRk\lim_{n \to \infty} P^{(n)}_{ij} = 0 \forall j \notin R_k.

if state ii is transient, then limnPij(n)=0\lim_{n \to \infty} P^{(n)}_{ij} = 0 for any other transient state jj, and limnPij(n)=αk(i)πk(j)\lim_{n \to \infty} P^{(n)}_{ij} = \alpha_k(i) \pi_k(j) for any recurrent state jj.

That is, we have limnνPij(n)=αk(i)πk(j)\lim_{n \to \infty} \nu P^{(n)}_{ij} = \alpha_k(i) \pi_k(j) exists, where ν\nu is initial probability distribution.

PP is irreducible but periodic

Since the chain periodic, we can split the state into dd classes A1,,AdA_1, \ldots, A_d such that each state class has the same period. Then PP will have dd eigenvalue of absolute value 1, dd complex number zz with zd=1z^d = 1. That is, given initial probability distribution ν\nu, we have the stationary distribution limn1dk=1dνP(n+k)=π\lim_{n \to \infty} \frac{1}{d} \sum_{k=1}^d \nu P^{(n + k)} = \pi. Therefore, the limit and the stationary distribution are not the same.

Therefore, we formally have the following theorem:

If PP is irreducible and aperiodic, then there exists a unique stationary distribution π\pi such that limnνP(n)=π\lim_{n \to \infty} \nu P^{(n)} = \pi with initial probability distribution ν\nu.

Countable MC convergence to stationary distribution

If MC is countable, irreducible and aperiodic, the positive recurrent states is similar to finite MC. limnP(n)=π\lim_{n \to \infty} P^{(n)} = \pi.

  • A irreducible, aperiodic and positive recurrent MC called ergodic.

Furthermore, If a stationary distribution exists, then the chain is positive recurrent.

Success-run chain

For a MC with transition probability px,x+1=px,px,0=qxp_{x,x+1} = p_x, p_{x,0} = q_{x} where px+qx=1p_{x} + q_{x} = 1, we call it a success-run chain. It can be shown that XnX_n is:

  • Transient if limnj=0n1pj>0\lim_{n \to \infty} \prod_{j = 0}^{n-1} p_{j} > 0.
  • Null recurrent if limnj=0n1pj=0\lim_{n \to \infty} \prod_{j = 0}^{n-1} p_{j} = 0.
  • Positive recurrent if nj=0n1pj<\sum_{n \to \infty} \prod_{j = 0}^{n-1} p_{j} < \infty.

Mean Recurrence Time

Let XnX_n be an irreducible and positive recurrent MC with transition probability matrix PP. Consider this MC start from ii the amount of time spent in state jj up to and including time nn: Y(j,n)=i=0n1Xi=jY(j,n) = \sum_{i=0}^n 1_{X_i = j}. Then we have limn1n+1E[Y(j,n)X0=i]=limn1n+1k=0nP(Xk=jX0=i)=π(j)\lim_{n \to \infty} \frac{1}{n + 1} E[Y(j,n)|X_0 = i] = \lim_{n \to \infty} \frac{1}{n + 1} \sum_{k=0}^n P(X_k = j|X_0 = i) = \pi(j).

Again, assume X0=iX_0 = i, Let τi\tau_i be the first time that MC returns to state ii. Then we have E[τi]=1π(i)E[\tau_i] = \frac{1}{\pi(i)} by law of large number. Here, E[τi]E[\tau_i] is called the mean recurrence time of state ii.

Consider an ergodic MC with stationary distribution π\pi. Let NjN_j be the number of visits to jj between consecutive visits to ii. Then we have E[Nj]=π(j)π(i)E[N_j] = \frac{\pi(j)}{\pi(i)} and E[τi]=1π(i)E[\tau_i] = \frac{1}{\pi(i)}.