Stationary Distribution
Long-Run Behavior
The central question of Markov chain theory: if we run the chain for a very long time, does the distribution of settle down to a limit, independent of where the chain started? The answer is yes, under the right conditions, and the limiting distribution is the stationary distribution . This distribution governs the long-run fraction of time spent in each state and is the key to computing steady-state performance metrics in communications systems.
Definition: Stationary (Invariant) Distribution
Stationary (Invariant) Distribution
A probability distribution on the state space is a stationary distribution (or invariant distribution) for the DTMC with transition matrix if
i.e., for all , with and .
If , then for all : the distribution does not change over time.
The equation says that is a left eigenvector of for eigenvalue 1. Since is stochastic, eigenvalue 1 always exists (the all-ones vector is a right eigenvector). The question is whether a non-negative left eigenvector with entries summing to 1 exists and is unique.
Stationary distribution
A probability vector satisfying . If the chain starts in this distribution, it stays in it forever.
Related: Stationary (Invariant) Distribution
Theorem: Existence and Uniqueness of the Stationary Distribution
Let be an irreducible DTMC on a countable state space .
-
If the chain is positive recurrent (i.e., the expected return time for some, hence all, states ), then there exists a unique stationary distribution given by
where is the first return time.
-
If the chain is transient or null recurrent (), then no stationary distribution exists.
For a finite irreducible chain, positive recurrence is automatic, so a unique stationary distribution always exists.
The stationary probability of state is inversely proportional to the mean return time: states that take longer to revisit are less likely to be occupied in steady state. For finite chains, every irreducible chain is positive recurrent because there are only finitely many states to wander among.
Positive recurrence implies existence
Define and set . The key step is to show .
Consider the chain started at state . Between successive visits to , the chain visits each state an expected number of times equal to . Call this . One can show and , which implies . Setting yields a probability distribution satisfying .
Uniqueness
Suppose is another stationary distribution. Then for any state :
As , by the ergodic theorem (Theorem below), . Hence . So .
Transient/null recurrent case
If for all , any candidate would need for all (since ), but is impossible. So no stationary distribution exists.
Example: Stationary Distribution of the Two-State Chain
Find the stationary distribution of the two-state chain from Example ETwo-State Markov Chain (Sunny/Rainy Weather) with , where , .
Set up the equations
We need with . Writing out:
Both equations simplify to: .
Solve with normalization
From and :
Interpret
In the long run, it is sunny 62.5% of the time and rainy 37.5%. Note that because : the chain transitions out of rain faster than out of sun.
Example: Doubly Stochastic Chains
A DTMC on has a doubly stochastic transition matrix (both rows and columns sum to 1). Find the stationary distribution.
Guess and verify
Let , the uniform distribution. Then:
since (columns sum to 1). So .
Conclusion
Any doubly stochastic irreducible chain has the uniform stationary distribution. This is exploited in algorithm design: random walks on graphs with doubly stochastic transition matrices sample nodes uniformly.
Theorem: Convergence Theorem (Ergodic Theorem for Markov Chains)
Let be an irreducible, aperiodic, positive recurrent DTMC with stationary distribution . Then for any initial distribution:
In matrix form: , where is the matrix with every row equal to .
A chain satisfying all three conditions (irreducible, aperiodic, positive recurrent) is called ergodic.
After sufficiently many steps, the chain "forgets" its initial state and settles into the stationary distribution. Every row of becomes approximately , regardless of the starting state.
Coupling argument (sketch)
The elegant proof uses coupling: construct two copies of the chain, , with and , evolving under the same transition matrix but coupled so they "meet" (enter the same state) as quickly as possible.
Meeting time
Irreducibility and aperiodicity guarantee that for any two starting states, there exists such that for all . This ensures the coupling time is finite almost surely.
After meeting
Once , both chains evolve identically: for all . Since for all (stationarity), we get
Hence as .
What Happens When the Chain is Periodic?
If the chain is irreducible and positive recurrent but has period , then does not converge to . Instead, is 0 unless , and . The time-averaged version still holds:
So the stationary distribution still governs long-run averages, even without pointwise convergence.
Computing and Visualizing the Stationary Distribution
Enter a transition matrix and see its stationary distribution computed as the left eigenvector for eigenvalue 1. Also displays the row distribution of converging to (for ergodic chains).
Parameters
Track the distribution starting from this state
Historical Note: Perron-Frobenius and the Spectral Theory of Non-negative Matrices
Early 20th centuryThe existence and uniqueness of the stationary distribution is intimately connected to the Perron-Frobenius theorem (Oskar Perron, 1907; Georg Frobenius, 1912). For a positive (or irreducible non-negative) matrix, this theorem guarantees a unique largest eigenvalue that is real and positive, with a corresponding eigenvector that has all positive entries. For stochastic matrices, this largest eigenvalue is exactly 1, and the corresponding left eigenvector (normalized) is the stationary distribution. Google's PageRank algorithm is arguably the most famous modern application of Perron-Frobenius theory applied to Markov chains.
Quick Check
For a doubly stochastic irreducible chain on 5 states, what is ?
Cannot determine without knowing
A doubly stochastic chain on states has uniform stationary distribution for all .
Common Mistake: Stationary Distribution vs Limiting Distribution
Mistake:
Assuming that if is a stationary distribution, then . This conflates existence of with convergence.
Correction:
Convergence requires aperiodicity in addition to irreducibility and positive recurrence. A periodic chain has a stationary distribution but oscillates and does not converge pointwise.
Key Takeaway
An irreducible, positive recurrent DTMC has a unique stationary distribution with . Adding aperiodicity gives the convergence theorem: . For finite irreducible chains, positive recurrence is automatic β only aperiodicity needs checking.