CSU | Hayward | |
---|---|---|
Statistics | Department |
Let
For example, p01 = P{X(2)=1|X(1)=0}= a is the probability of making a transition from state 0 to state 1 in one step. Also, for any starting step m = 1, 2 ..., the homogeneity and Markov properties imply, respectively, that
and that
It is not difficult to show that if P has all positive elements (0 < a, b < 1), then the probability structure of the chain stabilizes to a limit as the number of stages increasesand to find what that limit must be. Our purpose here is to sketch a proof of these results.
The two-step transition probabilities are found by taking into account whether the value of the chain at the intermediate step is 0 or 1. For example,
p01(2) = P{X(3)=1|X(1)=0} = p00 p01 + p01 p11 = a(2 a b).
The probabilities of two-step transitions from 0 to 0, 1 to 0, and 1 to 1 are found similarly. Thus, the two-step transition matrix is P2 (obtained by ordinary matrix multiplication). By induction (see the section below), the n-step transition matrix Pn can be expressed as the sum of two matrices:
The first term in this sum does not depend on n. The second depends on n only via the exponent in the constant (scalar) multiplier; this term vanishes as n approaches infinity provided that |D| = |1 a b| < 1, which is guaranteed by our original restriction that all elements of P be positive. This result not only shows that Pn converges to a limit but also that the convergence takes place at a relatively rapid "geometric" rate.
Technical notes: Here we focus on cases were P has all positive elements. However, the P-matrix for a so-called ergodic two-state chain either has all positive elements or has one of a and b (but not both) equal to 1. This is equivalent to saying that P2 has all positive elements. The condition |D| < 1 required for convergence above includes not only ergodic chains, but also the "absorbing" cases (either a or b, not both, equal to 0), which have degenerate limiting distributions.
The upper-right element of the matrix Pn is the n-step
transition probability
p01(n) = P{X(n+m)=1|X(m)=0}.
We have just seen that, as n approaches infinity, p01(n)
approaches a/(a + b).
Similarly, looking at the lower-right element of Pn,
we find that p11(n) also approaches
a/(a + b).
In the long run, such a Markov chain "forgets" where it started.
Regardless of the starting value of the process, in the long run we have simply
p = P(X(¥) = 1)
= a/(a + b).
Similarly, looking at the elements in the first column of Pn,
we have the long-run probability
l
= (p*, p) = (b/ (a + b), a/(a + b)).An alternative way to obtain this limiting distribution, when it exists, is to solve the matrix equation lP = l. This matrix equation yields two linear equations:
p
* = p*(1 a) + pb and p = p*a + p(1 b).These equations are co-linear. Together with the requirement that p + p* = 1, either one of them yields p = a/(a + b) and p* = b/(a + b). These are the same results we obtained from the limiting argument above.
For example, let a = 0.0123 and b = 0.6016, so that 1/(a + b ) = 1 / 0.6139 = 1.6289, and the limit is
This shows the limit of a particular matrix, but also illustrates the principle that when Pn approaches a limit, each row of the limit P¥ is the same as the solution l = (p*, p ) of the matrix equation lP = l.
The proof of the expression for Pn as the sum of two matrices is by induction, using simple algebra. The initial step is to verify that this expression simplifies to P when n = 1. The induction step is to multiply this expression for Pn by P, and to verify that the result simplifies to
|
The treatment of two-state Markov chains in this brief appendix has necessarily been sketchy. Many beginning probability texts cover finite Markov chains in some detail. One that gives a careful proof of the limit theorem for a k-state chain is W. Feller: An Introduction to Probability Theory and its Applications, Vol. 1 (3rd ed.), 1950, Wiley, New York.