Mismatch Analysis

What Happens When the Prior Is Wrong?

The BG-MMSE denoiser is Bayes-optimal when the prior is correctly specified β€” i.e., the scene is truly Bernoulli-Gaussian with the assumed parameters (ρ,Οƒc2)(\rho, \sigma_c^2). In practice, the prior is always wrong: the true scene may have a different sparsity, a different amplitude distribution, or structure that no parametric prior captures (extended targets, spatial correlations).

How robust is OAMP to such mismatch? Can we still use state evolution to predict performance under a mismatched denoiser? This section answers both questions.

Definition:

Mismatched State Evolution

Let p0p_0 be the true prior and p~0\tilde{p}_0 be the assumed prior used to design the denoiser Ξ·~t\tilde{\eta}_t. The mismatched state evolution for OAMP is:

v1t=F(v2tβˆ’1;A),v_1^t = F(v_2^{t-1}; \mathbf{A}),

v2t=Ep0[∣η~t(C0+v1t Z)βˆ’C0∣2],v_2^t = \mathbb{E}_{p_0}\bigl[|\tilde{\eta}_t(C_0 + \sqrt{v_1^t}\,Z) - C_0|^2\bigr],

where the expectation is over C0∼p0C_0 \sim p_0 (the true prior) and Z∼CN(0,1)Z \sim \mathcal{CN}(0, 1). The LMMSE step is unchanged (it does not depend on the prior).

The key point: state evolution is still valid under mismatch. It tracks the actual algorithm, not the intended algorithm. But the MSE is degraded because Ξ·~t\tilde{\eta}_t is not optimal for p0p_0.

Mismatched state evolution enables offline analysis of robustness: sweep over possible true priors p0p_0 while keeping the denoiser fixed, and plot the resulting MSE.

Theorem: MSE Degradation Under Prior Mismatch

Let v2βˆ—v_2^* be the fixed-point MSE of OAMP with the matched (Bayes-optimal) denoiser for the true prior p0p_0, and let v~2βˆ—\tilde{v}_2^* be the fixed-point MSE with the mismatched denoiser Ξ·~\tilde{\eta} designed for p~0\tilde{p}_0. Then

v~2βˆ—β‰₯v2βˆ—,\tilde{v}_2^* \geq v_2^*,

with equality if and only if Ξ·~\tilde{\eta} equals the Bayes-optimal denoiser for p0p_0 almost everywhere.

Moreover, the MSE degradation is bounded by

v~2βˆ—βˆ’v2βˆ—β‰€sup⁑v1β‰₯0[mse(Ξ·~,v1;p0)βˆ’mmse(p0,v1)]β‹…(1βˆ’βˆ‚Fβˆ‚v2∣v2βˆ—)βˆ’1,\tilde{v}_2^* - v_2^* \leq \sup_{v_1 \geq 0} \bigl[ \text{mse}(\tilde{\eta}, v_1; p_0) - \text{mmse}(p_0, v_1)\bigr] \cdot \bigl(1 - \frac{\partial F}{\partial v_2} \big|_{v_2^*}\bigr)^{-1},

where the supremum is over noise levels, and the second factor accounts for the amplification through the LMMSE step.

The mismatched denoiser is suboptimal at each iteration β€” it removes less noise than the Bayes-optimal denoiser. This excess MSE feeds back into the LMMSE step, which sees a larger prior variance, producing a noisier output, creating a vicious cycle. The bound quantifies how much the per-step suboptimality is amplified by the iterative loop.

Example: Sparsity Mismatch in BG-MMSE

The true scene is Bernoulli-Gaussian with sparsity ρ0=0.10\rho_0 = 0.10 and variance Οƒc2=1\sigma_c^2 = 1. The denoiser uses an assumed sparsity ρ~\tilde{\rho}. Compute the OAMP fixed-point NMSE via mismatched state evolution for ρ~∈{0.01,0.05,0.10,0.15,0.20,0.30}\tilde{\rho} \in \{0.01, 0.05, 0.10, 0.15, 0.20, 0.30\}.

Parameters: Ξ΄=0.4\delta = 0.4, SNR=25 dB\text{SNR} = 25\,\text{dB}, N1=N2=32N_1 = N_2 = 32 (Kronecker sensing with partial DFT).

OAMP Performance Under Prior Mismatch

Explore how OAMP's NMSE degrades when the assumed sparsity or signal variance differs from the truth. The dashed line shows the Bayes-optimal NMSE.

Parameters
0.1
1
0.4
25

Minimax and Robust Denoisers

When no reliable prior information is available, one can use a minimax denoiser that minimizes the worst-case MSE over a class of signals (e.g., the β„“2\ell_2 ball of radius RR). The minimax denoiser for AWGN is the James-Stein estimator:

Ξ·JS(r)=(1βˆ’(Nβˆ’2)vβˆ₯rβˆ₯2)+ r,\eta_{\text{JS}}(\mathbf{r}) = \Bigl(1 - \frac{(N-2)v} {\|\mathbf{r}\|^2}\Bigr)_+\,\mathbf{r},

which shrinks toward zero by a data-dependent factor.

Minimax denoisers are conservative: they sacrifice performance when the prior is favorable in exchange for robustness when the prior is unfavorable. In RF imaging, this tradeoff is usually not worthwhile β€” the scene statistics are often well-characterized from training data, and a learned denoiser outperforms minimax by a large margin.

πŸ”§Engineering Note

Online Parameter Learning via EM

Rather than choosing prior parameters offline, they can be learned from the measurements using an EM (expectation- maximization) approach interleaved with the OAMP iterations:

  • E-step: Run one OAMP iteration with current parameters (ρ~,Οƒ~c2)(\tilde{\rho}, \tilde{\sigma}_c^2).
  • M-step: Update ρ~\tilde{\rho} and Οƒ~c2\tilde{\sigma}_c^2 using the posterior statistics from the denoiser output.

This is essentially EM-GAMP (Chapter 19) applied to the OAMP framework. It converges in 5–10 outer EM iterations for typical RF imaging problems, adding minimal overhead.

Practical Constraints
  • β€’

    EM converges to a local maximum of the marginal likelihood; multiple initializations may be needed

  • β€’

    For very low SNR or very high undersampling, the EM landscape may have spurious local optima

Common Mistake: Signal Variance Mismatch Can Be Worse Than Sparsity Mismatch

Mistake:

Carefully tuning the sparsity parameter ρ\rho but using a default signal variance Οƒc2=1\sigma_c^2 = 1 without checking whether it matches the actual signal power.

Correction:

The BG-MMSE denoiser depends on both ρ\rho and Οƒc2\sigma_c^2. A 10x error in Οƒc2\sigma_c^2 can cause 5–8 dB of MSE degradation, comparable to a 5x error in ρ\rho. Always estimate both parameters, either from training data or via EM within OAMP.

Quick Check

Mismatched state evolution for OAMP:

Still accurately predicts the empirical MSE of the mismatched algorithm

Overestimates the MSE because it assumes worst-case mismatch

Is invalid because the orthogonality condition requires a matched prior

Key Takeaway

OAMP is moderately robust to prior mismatch: a 2x error in the sparsity parameter costs about 1–2 dB, while a 10x error can cost 5–8 dB. Underestimating sparsity is more harmful than overestimating it. Mismatched state evolution remains valid and enables offline robustness analysis. When prior parameters are uncertain, EM-based learning or learned denoisers provide the most robust reconstructions.