State Evolution
What State Evolution Delivers
The payoff for introducing the Onsager term is that AMP admits a one-dimensional deterministic description of its per-iteration performance in the large-system limit. This description β state evolution (SE) β is a scalar recursion whose fixed points predict the terminal MSE, the convergence speed, and the sharp phase-transition boundary between recoverable and unrecoverable signals.
The analytical power is considerable: we can predict AMP's MSE without running AMP, plot phase diagrams by sweeping , and design denoisers by minimising a scalar function.
Theorem: State Evolution (Bayati--Montanari)
Consider AMP with denoiser on the model of Definition DLinear Observation Model, with at fixed . Assume is Lipschitz and the prior has finite second moment. Then, for every iteration :
(a) Gaussianity of pseudo-data. In the sense of empirical distribution convergence,
(b) Scalar recursion. The variance evolves as with , , independent.
Part (a) says that AMP behaves as if at every iteration we were denoising the signal observed in an AWGN channel of variance . Part (b) is Pythagoras: the effective noise variance at the next iteration equals the physical noise plus a rescaling () of the scalar MSE achieved by the denoiser at the previous iteration.
Conditioning on the iterates (cavity argument)
Denote by the -algebra generated by . By the orthogonality of Gaussian projections, conditioning on leaves the complementary random projection Gaussian. This is the same cavity-method argument used in the TAP analysis of the SK model.
Onsager cancels the bias
Writing as a sum of independent terms conditioned on yields a linear combination of Gaussian increments plus a systematic bias . The Onsager term in the AMP residual is constructed to cancel this bias exactly, leaving a pure Gaussian.
Variance bookkeeping
The residual has entries with variance (the factor comes from ). Hence is a Gaussian of per-entry variance . The denoiser is then applied, producing error per coordinate. Plugging this back through the residual update gives the claimed recursion.
Passage to the limit
Bayati--Montanari (2011) rigorously show that, for Lipschitz , the empirical distribution of converges almost surely to the joint distribution in Wasserstein-2 sense. This upgrades the heuristic cavity argument into a theorem.
Definition: State-Evolution Fixed Point
State-Evolution Fixed Point
A fixed point of state evolution is a value with Define . A fixed point is stable if and unstable otherwise.
When has multiple fixed points, AMP generically converges to the stable fixed point encountered first by the iteration β typically the one with the largest . This is why AMP can miss the global optimum even when state evolution admits a smaller fixed point.
Theorem: AMP MSE Equals State-Evolution Fixed Point
Under the hypotheses of Theorem TState Evolution (Bayati--Montanari), let be the limit of the SE recursion. Then almost surely.
The terminal MSE of AMP is determined entirely by the one scalar β a remarkable dimensionality reduction from an -dimensional distribution to a single number.
Convergence of the recursion
is monotone (a mixture of MSE curves, each monotone in input noise) and bounded, so is monotone in and converges to . Since is continuous, the limit is a fixed point.
Passing limits inside
By Lipschitz continuity of , the expectation is continuous in . Combined with the empirical-distribution convergence from Theorem TState Evolution (Bayati--Montanari), we obtain the stated equality almost surely.
Connection to the Replica Prediction
If is the MMSE denoiser for the prior matched to effective noise , the state-evolution fixed point coincides with the replica-symmetric prediction of the Bayes-optimal MMSE, which is conjectured (and in many cases proved, e.g.\ Reeves--Pfister 2016) to be the fundamental information-theoretic limit of the Bayesian CS problem.
In other words: when AMP converges, AMP matched to the prior is asymptotically Bayes-optimal. This elevates AMP from a heuristic iterative solver to a provably optimal estimator in the proportional asymptotic regime.
Example: State Evolution, Noiseless LASSO, and the Donoho--Tanner Curve
Take and a sparse signal with i.i.d. entries from . Using AMP with soft-thresholding tuned optimally, derive the state-evolution fixed-point equation and identify the boundary between successful and failed recovery.
Scalar MMSE of the soft-threshold
Let be the scalar MSE of soft-thresholding on the Bernoulli--Gaussian signal with threshold . Direct computation (integrating the Gaussian density against the piecewise-linear soft-threshold) gives a closed-form expression depending on and only.
Normalised SE recursion
With the recursion becomes Scaling: for a function independent of . The recursion is then , a geometric decrease iff .
Phase transition
Optimising over for each yields the Donoho--Tanner phase boundary i.e., the curve of below which and AMP converges to (exact recovery) and above which the iteration has a nonzero stable fixed point (recovery fails). This curve matches exactly the Donoho--Tanner combinatorial geometry result on phase transitions of recovery.
State Evolution vs Empirical AMP
Compares the scalar state-evolution prediction (deterministic recursion) against the empirical per-iteration MSE from a single run of AMP at finite . As grows the two curves lock onto each other.
Parameters
Phase Transition Diagram
State-evolution phase diagram in the plane for noiseless sparse recovery with a Bernoulli--Gaussian signal. Colour = predicted terminal MSE of AMP with optimally tuned soft-thresholding. The curve separates the "recovery" and "failure" regions β the Donoho--Tanner phase boundary.
Parameters
State-Evolution Cobweb Diagram
Common Mistake: State Evolution Is an Asymptotic Statement
Mistake:
Reading off the SE prediction for a specific finite- problem and being surprised when AMP's empirical MSE differs by 20% or more. Worse, tuning assuming SE holds exactly at .
Correction:
State evolution predicts in the limit . For deviations of 5--15% are normal. For - the agreement is typically tight (<2%). Always sanity-check by running AMP at two values of and confirming the empirical trajectories are converging to the SE prediction as grows.
State Evolution as a Diagnostic Tool
Beyond its role as a performance predictor, state evolution is useful as a diagnostic. If AMP's empirical MSE diverges from the SE prediction, one of the assumptions has been violated β typically: (i) the matrix is not i.i.d.\ (Section 20.4); (ii) is too small; (iii) the denoiser is not Lipschitz. Each produces a characteristic signature that an experienced practitioner learns to recognise.
Key Takeaway
State evolution reduces the analysis of AMP β a random, high-dimensional, nonlinear dynamical system β to a one-dimensional deterministic recursion . This is the most striking analytical feature of AMP and the reason phase transitions and Bayes-optimality can be computed in closed form.