EM-GAMP for RF Imaging
The Hyperparameter Problem in Bayesian Inference
The VAMP algorithm from Chapter 18 achieves near-optimal reconstruction when the prior distribution is correctly specified — but this requires knowing the sparsity rate , the signal variance , and the noise variance .
In any real RF imaging deployment, these parameters are unknown and vary with scene statistics and hardware noise floor. Manual tuning by cross-validation requires running the reconstruction many times and does not scale.
EM-GAMP solves this by treating the signal as a latent variable and the hyperparameters as unknowns to be estimated by the Expectation-Maximization (EM) algorithm. The result is a self-tuning Bayesian algorithm that converges to near-oracle performance from arbitrary initialization.
Historical Note: Origins of Expectation-Maximization
1977–2014The EM algorithm was formally unified by Dempster, Laird, and Rubin in their landmark 1977 JRSS-B paper, though special cases had been used for decades (e.g., Baum–Welch for HMMs, 1970). The key insight was the Q-function — the expected complete-data log-likelihood — which EM monotonically maximizes, guaranteeing non-decreasing marginal likelihood.
The integration of EM with approximate message passing for sparse recovery was developed independently by Vila & Schniter (EM-GM-GAMP, 2013) and Kamilov et al. (parametric GAMP, 2014), who proved consistency of the EM parameter estimates in the large-system limit.
Definition: The EM-GAMP Model
The EM-GAMP Model
The EM-GAMP model couples the RF imaging observation model with parameterized priors:
where is the hyperparameter vector.
The complete-data likelihood (treating as known) is:
EM maximizes the marginal likelihood using GAMP's posterior statistics as a surrogate for the intractable posterior .
EM-GAMP Algorithm
Complexity: — same order as GAMPIn practice, (interleaved EM) works well and is the most common choice. Interleaving avoids running GAMP to convergence at each EM step, which is expensive and unnecessary.
Theorem: Closed-Form EM M-Step Updates
Let , , and be the posterior statistics from the GAMP E-step. Then the M-step maximizes:
under the factored Gaussian approximation . The closed-form solutions are:
The noise variance update is a bias-corrected residual: the raw squared residual underestimates the true noise power because GAMP's estimate is not exactly the signal. The correction term accounts for the uncertainty in .
Take the derivative of with respect to and set to zero.
For and , treat as a weighted Gaussian mixture; the updates are standard mixture-model MLE.
The correction term in comes from .
Setup the Q-function
The Q-function decomposes as:
Optimize over noise variance
Setting :
The second equality uses .
Optimize over prior parameters
For Bernoulli-Gaussian :
Summing over and differentiating gives the stated updates.
EM-GAMP Hyperparameter Convergence
Traces of estimated , , and versus EM outer iteration (left axis). The NMSE (dB) is shown on the right axis. Dashed horizontal lines mark the true parameter values.
Observe: EM converges within 10–20 iterations to near-oracle NMSE regardless of initialization. Select all to see joint estimation; the noise and sparsity estimates interact — at low SNR they can trade off against each other.
Parameters
Example: EM-GAMP for Sparse RF Scene with Unknown Parameters
Setup: scene voxels, measurements (), Bernoulli-Gaussian prior with , , (SNR 20 dB).
Initial guesses: , , (all overestimated by 5–50×).
After running EM-GAMP for 25 outer iterations (with 1 GAMP step per EM step), report the estimated parameters and the reconstruction NMSE.
Iteration 0 (initialization)
With , the denoiser applies a weak threshold — most coefficients survive — giving poor sparsity promotion. The residual is large.
Approximate NMSE dB: essentially no reconstruction.
Iterations 1–10 (rapid convergence)
The M-step updates downward (many are small) and downward (as the fit improves). After 10 steps: , . NMSE dB.
Iterations 10–25 (fine-tuning)
Parameters converge to , , . NMSE stabilizes at dB.
Comparison with oracle
Oracle GAMP (true parameters known) achieves NMSE dB. EM-GAMP gap: 0.1 dB — essentially negligible.
The EM self-tuning eliminates the need for cross-validation while incurring only a 20–25% increase in runtime (the EM M-step is negligible cost).
Common Mistake: EM Can Converge to Local Optima at Low SNR
Mistake:
The EM objective function is generally non-convex in . Multiple distinct parameter settings can explain the same data:
- High noise + low sparsity (many small scatterers) vs.
- Low noise + high sparsity (few strong scatterers).
At SNR dB, EM may converge to the wrong mode, producing a poor estimate.
Correction:
Practical remedies: (1) Multiple restarts: run EM from 3–5 random initializations; select the solution with highest marginal likelihood. (2) Warm start: initialize from the residual energy of a simple matched filter. (3) Physical constraints: bound and .
Common Mistake: Running Full GAMP to Convergence at Each EM Step Is Wasteful
Mistake:
A common implementation error is to run GAMP to full convergence at each EM outer iteration before updating . This is correct in theory but wasteful in practice: GAMP iterates in the wrong parameter regime at early EM iterations.
Correction:
Use interleaved EM: one GAMP iteration per EM update. This is equivalent to gradient ascent on the EM objective and converges at the same rate as full EM for well-behaved problems. The implementation simply nests the EM M-step inside the GAMP loop.
EM-GAMP Calibration in Practice
In real RF imaging systems, the noise variance varies with:
- Transmit power and propagation path loss (accounted for in the link budget).
- Receiver noise figure (typically 3–10 dB above thermal noise floor).
- Clutter — returns from unwanted scene elements not in the model.
The sparsity rate depends on the scene type:
- Urban scenes: –.
- Cluttered environments: –.
EM-GAMP adapts to these variations automatically, but its adaptation rate is limited: for rapidly varying clutter, the single-snapshot EM assumption fails. A sliding-window EM that averages M-step updates across consecutive frames improves stability.
- •
Noise floor estimation requires at least measurements (well-posed EM)
- •
EM convergence typically requires 15–30 outer iterations for SNR > 10 dB
- •
Interleaved EM adds < 5% overhead to GAMP runtime (M-step is )
EM Objective Landscape:
Contour plot of the approximate marginal log-likelihood as a function of sparsity and noise variance , with the true parameter marked as a star. The landscape is generally unimodal at high SNR but develops spurious local maxima at low SNR.
Observe how the objective elongates into a ridge (the noise-sparsity tradeoff) as the SNR decreases.
Parameters
Quick Check
In EM-GAMP, the M-step update for (sparsity) is:
Correct. The posterior inclusion probability is the soft indicator of support. Averaging over all components gives the expected fraction of nonzero elements, which is the MLE of the Bernoulli parameter .
Key Takeaway
EM-GAMP wraps GAMP in an EM loop: the E-step runs GAMP to compute posterior means , variances , and inclusion probabilities ; the M-step updates , , and in closed form. Interleaving (one GAMP iteration per EM step) is efficient and achieves near-oracle NMSE — typically within 0.1 dB — from arbitrary initialization, eliminating manual tuning entirely.
EM-Based Hyperparameter Estimation for RF Imaging
The parametric GAMP framework — of which EM-GAMP is the principal instance — was developed to provide consistent hyperparameter estimates alongside the signal estimate. The analysis shows that in the large-system limit ( with ), the EM-GAMP parameter estimates converge to the true values whenever the true parameters are identifiable. This result provides the theoretical foundation for the self-tuning RF imaging pipeline described in this chapter.