Chapter Summary
Chapter 3 Summary: Bayesian Inverse Problems
Key Points
- 1.
Bayes' theorem combines the likelihood and prior to yield the full posterior — the complete solution to the Bayesian inverse problem. MAP equals variational regularization; MMSE equals the posterior mean.
- 2.
For Gaussian priors with Gaussian noise, the posterior is Gaussian with mean — exactly the Tikhonov solution with having a clear probabilistic interpretation as the noise-to-signal variance ratio.
- 3.
Sparsity-promoting priors — Laplace (yields LASSO MAP), Bernoulli-Gaussian (standard RF scene model), spike-and-slab (ideal but intractable), and horseshoe (adaptive, near-minimax) — assign high prior probability to sparse scenes and produce sparser reconstructions than Gaussian priors.
- 4.
Sparse Bayesian Learning (SBL) uses ARD per-component precisions updated via EM. The algorithm automatically drives irrelevant (pruning those components) without pre-specifying sparsity level . SBL provides posterior uncertainty on the active support that LASSO cannot.
- 5.
Gaussian measures on Hilbert spaces provide discretization-invariant priors via trace-class covariance operators. The Cameron-Martin theorem characterizes which MAP estimates live in function space. Stuart's theorem guarantees existence, uniqueness, and stability of the posterior as a measure — the Bayesian analogue of Tikhonov well-posedness.
- 6.
Posterior credible intervals and variance maps quantify reconstruction confidence. The posterior contracts at the minimax-optimal rate when the prior regularity matches the truth.
- 7.
The pCN sampler achieves dimension-independent MCMC acceptance rates by proposing in the Cameron-Martin space. HMC scales as using gradients. Calibration checks (empirical coverage vs nominal level) are mandatory before deploying UQ in safety-critical applications.
Looking Ahead
Chapter 4 turns from probabilistic formulations to efficient algorithms: given the Bayesian and variational problems formulated in Chapters 2-3, how do we compute MAP estimates efficiently for unknowns?
- Fast algorithms for structured operators — the NUFFT, Kronecker products, and FFT-based convolutions that make imaging-scale optimization tractable.
- GPU acceleration — CuPy and PyTorch forward/adjoint implementations enabling ISTA, FISTA, and ADMM at scale.
- Automatic differentiation — computing Jacobians through iterative solvers, enabling gradient-based hyperparameter optimization and the deep-unfolding algorithms of §Computed Tomography: The Canonical Inverse Problem.
- Convergence diagnostics — primal residuals, dual residuals, the discrepancy principle, and warm-starting strategies for practical imaging solvers.