A discrete-time stochastic process is made up of for , where takes on values in a finite set
The possible values for are the states of the system
We can describe the probability of a process as or more expressively, as an initial probability distribution and the transition probabilities
The Markov property states that transitions are reliant only on the current state, not on previous states,
A process is time-homogeneous if the transition probabilities are not functions of
A time-homogeneous Markov chain has each transition probability defined as for some function . These are of special interest.
We can define this kind of process with an initial distribution and transition probabilities . is nicely expressed as a transition matrix .
We call a stochastic matrix, since and
If we express as a vector , we can write
Large-Time Behavior
We are interested in seeing what happens with . We first observe that for many systems, for any starting .
We call an invariant probability distribution for if (also called a stationary, equilibrium, or steady-state distribution)
Definitionally, is a left eigenvector of with eigenvalue
Question: How do we know exists?
The key in this process is that diagonalizing gives us an eigenvalue of and other eigenvalues with absolute value . This means that repeated powers leaves us with only the eigenvector corresponding to .
Now, for any stochastic matrix, we have as a right eigenvector with eigenvalue , which means there must be a left eigenvector with eigenvalue . So we must show,
- The left eigenvector can be chosen to have nonnegative entries (and therefore can be normalized to be a distribution)
- The eigenvalue is simple (no multiplicity), and all other eigenvalues have absolute value
We cannot always diagonalize , but the Jordon decomposition is good enough and has the same effect
Perron-Frobenius Theorem: If has all positive entries, then is a simple eigenvalue, its left eigenvector can be chosen as positive, and all other eigenvalues have absolute value
This does not cover certain stochastic matrices with entries that do satisfy the limit behavior. If satisfies Perron-Frobenius, then satisfies our needed conditions
Now our question is which kinds of stochastic matrices have some with all positive entries
Cases where this is not true,
- Simple random walk with reflecting boundary: In this case, every step alternates between even and odd states. The matrix will basically “oscillate” instead of converging.
- Simple random walk with absorbing boundary: In this case, the boundaries absorb all processes, while the middle (transient) states will go to .
- Two non-interacting chains: We call this reducible
Definition: Two states communicate if there exist and , meaning a process starting in one state can reach the other
is an equivalence relation
- Reflexive:
- Symmetric:
- Transitive:
These equivalence relations partition our state space into disjoint communication classes
If there is only one communication class, then the chain is irreducible. Matrices with invariant distributions are irreducible. But we also see irreducible chains that do not have invariant distributions (our reflecting boundary chain).
Some communication classes are transient, meaning a process will leave them and never return with probability
For each recurrent class , the submatrix of obtained from only considering the rows and columns for states in is also a stochastic matrix. We can analyze the large-time behavior of by only considering this submatrix.
Definition: The period of a state is the greatest common divisor of
Theorem: If then
Definition: An irreducible matrix is aperiodic if and therefore has all entries positive for some . According to Perron-Frobenius, there exists a unique invariant probability vector , such that for any .
Reducible or Periodic Markov Chains
If is reducible with recurrent classes and transient classes , then each recurrent class acts as a smaller Markov chain. Therefore, there are invariant probability vectors , and the eigenvalue has multiplicity .
If , if or otherwise
Therefore, for each transient state
Let be the probability that the chain starting in state ends up in recurrent class . We see that . exists but depends on .
How do we understand the powers of ?
If is irreducible but has period , then the state space splits into sets . In general, will have simple eigenvalues with absolute value , the solutions to . will cycle through these distributions, which will average to . In this case, does not represent the limit of , but the average time spent in .
Return Times
Let be an irreducible Markov chain with transition matrix . Consider the amount of time spent in state up to and including time ,
We can relate this quantity to , since
Assume and let be the first time after that the Markov chain is in ,
The th return to state is given as the sum of independent and identical random variables . Therefore, for large , we have
In other words, in steps we find about visits. And since we know that in steps we expect about visits, we have
Suppose has some transient states and let be the submatrix of which includes only the rows and columns for the transient states,
is called a substochastic matrix, since it has row sums . We already know , implying its eigenvalues have absolute value strictly less than . Therefore, has no and we can define , which is called the fundamental matrix.
For a transient state , . Suppose is another transient state, , which is also the entry of the matrix
We can show that
So
We can do a lot with this. Observe how the expected number of steps until the chain enters a recurrent class, assuming , is the sum of the th row of
To find the expected number of steps the irreducible Markov chain takes to go from any state to another state , we rearrange such that is the first site, and modify it to be an absorbing site.
Note that the above procedure is only going to make sense on an irreducible chain (I don’t think this quantity would always be defined otherwise).
What if there are multiple recurrent classes? Starting at a given transient state , what is the probability that the Markov chain ends up in a particular recurrent class?
To simplify this problem, we treat recurrent classes as single points with . This lets us order our transition matrix as,
Denote be the probability that the chain starting at transient state ends up in recurrent state . With two recurrent states, and if . For any transient state ,
We can write this as , where is an matrix,
Reversibility
^202837
Markov chains define a direction for time. Can we reverse the direction of a chain?
Let’s label this “flipped” probability
We see that conditioning on the entire future leads to the same result,
However, this probability is still a function of . We can take the initial condition to be the invariant measure, and now the Markov chain is time-homogeneous. So we see must be equal to
Definition: If for some then and is reversible (and is the invariant measure)