Chapter Summary
Chapter Summary
Key Points
- 1.
AMP = ISTA + Onsager correction. The iteration , differs from proximal gradient by a single additive term — the Onsager correction — that removes the bias introduced by feedback through . It is the defining feature of AMP.
- 2.
Gaussianity of pseudo-data. In the large- limit with i.i.d. Gaussian , the denoiser input is distributed as the true signal plus independent AWGN of variance . This justifies calling a denoiser and underpins all downstream analysis.
- 3.
State evolution. AMP's per-iteration MSE is governed by the scalar recursion . Fixed points of this recursion determine the terminal MSE; their stability determines convergence. This one-dimensional description is the most striking analytical feature of AMP.
- 4.
Phase transitions and Bayes-optimality. Plotting SE fixed points over reveals the Donoho--Tanner curve separating recovery from failure. With the MMSE denoiser matched to the prior, the AMP fixed point coincides with the replica-symmetric prediction of the Bayes MMSE — i.e., AMP is asymptotically Bayes-optimal.
- 5.
Denoiser design is the central knob. Soft-thresholding yields LASSO (with a calibration formula); posterior means yield Bayes-optimal estimation; learned neural denoisers yield D-AMP and deep unfolding. All fit in the same AMP scaffold provided the denoiser is Lipschitz and its divergence can be computed.
- 6.
AMP fails for structured matrices. The Onsager coefficient is calibrated to the Marchenko--Pastur spectrum. Sub-sampled DFT, Hadamard, ill-conditioned, or Kronecker-structured matrices break the calibration and produce divergence. Damping stabilises AMP but slows convergence and breaks state evolution; the principled fix is OAMP/VAMP (Chapter 21).
Looking Ahead
Chapter 21 generalises the AMP framework to right-rotationally-invariant sensing matrices via OAMP (orthogonal AMP) and VAMP (vector AMP). The key idea is to replace the simple transpose with an LMMSE estimator and enforce orthogonality between the linear and prior-denoising steps. This restores the Gaussianity-of-pseudo-data property for a much larger class of matrices — including structured sensing operators typical of communications and imaging — at the cost of an inversion per iteration (amenable to Kronecker factorisation when the structure is known). Chapter 21 also introduces GAMP for generalised linear models (non-Gaussian likelihoods, e.g., 1-bit compressed sensing) and LAMP/LISTA for learned message passing via deep unfolding.