Practical OAMP for Structured Sensing Matrices
The LMMSE Step Is the Bottleneck
OAMP and VAMP are attractive in theory β state evolution holds for the RRI class, convergence is fast, and the fixed point is Bayes-optimal. The price is the LMMSE step, which in the generic dense case costs to set up and per application. For imaging problems with on the order of to , this is prohibitive.
The good news is that imaging sensing matrices are almost never generic. They carry physical structure β delays, angles of arrival, subcarrier-antenna Kronecker products β and this structure can be exploited to make the LMMSE solve essentially free. In this section we develop three practical tools: the Kronecker solve, the Hutchinson trace estimator for divergence, and a mismatch analysis that tells us how robust OAMP is to getting the prior wrong.
Definition: Kronecker-Structured Sensing Matrix
Kronecker-Structured Sensing Matrix
A sensing matrix is Kronecker-structured if
so that and . Equivalently, the signal can be reshaped into an matrix , and the measurement is
This appears in RF imaging as a product of an angular dictionary and a delay dictionary, in MIMO-OFDM as a product of space and subcarrier operators, and more generally whenever the sensing process factors across two physical dimensions.
The key identity is , which propagates the Kronecker structure through products, inverses, and SVDs.
Theorem: Efficient LMMSE for Kronecker Sensing
Let with SVDs . Then the LMMSE filter
can be applied to any vector in time
which is dominated by two small matrix products and a diagonal scaling in the SVD basis. For , the cost is rather than .
The singular values of are the pairwise products , and the right singular basis is . So the LMMSE inverse is diagonal in the tensor-product basis, and the application reduces to small per-factor transforms.
Compute the product's SVD
Using , the eigenvalues of are with eigenvectors .
Diagonalize the regularized inverse
In the tensor basis, becomes a diagonal matrix with entries . Applying it is element-wise in this basis.
Translate back to the original basis
Applying (acting row-wise and column-wise on ) costs and respectively. The sandwich contributes a symmetric cost along the signal-side factors. Summing gives the stated total.
Definition: Hutchinson Trace Estimator
Hutchinson Trace Estimator
For any matrix , the trace admits the stochastic identity
where the entries of are i.i.d. with zero mean and unit variance (e.g., Rademacher or Gaussian). Averaging independent realizations gives an unbiased estimator
whose variance scales as for Rademacher probes.
The estimator does not require access to the entries of β only the ability to compute matrix-vector products . This is decisive when is the composition with a matrix-free LMMSE solver.
Hutchinson Estimator for Denoiser Divergence
Complexity: calls to the denoiser plus inner products. Since the denoiser is separable, each call is , giving total β negligible compared to the linear step when .For smooth denoisers (soft-thresholding, MMSE on Bernoulli-Gaussian) a closed-form divergence is available and preferred. The Hutchinson estimator is the method of last resort for black-box / learned denoisers where no closed form exists.
Example: Variance of the Hutchinson Estimator
For a diagonal matrix with Rademacher probes, compute the mean and variance of the single-probe estimator . What does this tell us about choosing ?
Expand the quadratic form
. For Rademacher probes , so . For diagonal the cross terms vanish, giving , exactly, with zero variance.
Handle the general case
For non-diagonal , when (Rademacher independence), so . The variance is .
Choose $K$ for target accuracy
To achieve relative error with high probability, Chebyshev's inequality gives . For "balanced" Jacobians this is typically β cheap enough to use at every iteration.
Theorem: Mismatch Penalty for OAMP
Let OAMP be run with a denoiser matched to an assumed prior , while the true prior is . In the large-system limit, the MSE fixed point satisfies
where the penalty is non-negative and equals zero if and only if almost everywhere. A first-order expansion in the KL discrepancy gives
where is the fixed-point effective noise variance.
Assumption mismatch costs MSE, and it costs MSE linearly in the squared score-function error. Overconfident sparsity (assuming lower than truth) is typically more damaging than underconfident sparsity β a folklore observation made precise by this bound.
Fixed-point equation with mismatched denoiser
At the OAMP fixed point, the effective MSE satisfies where is the MSE of the assumed-prior denoiser evaluated under the true data distribution.
Compare with the Bayes-optimal fixed point
For the Bayes fixed point, the same equation holds with . Subtract: by the optimality of the MMSE estimator.
First-order expansion
Expanding the denoiser difference around the true posterior and using the I-MMSE identity (Guo-Shamai-VerdΓΊ) gives the closed-form bound in terms of score-function discrepancy. Details are standard in the mismatched-estimation literature.
OAMP Mismatch Analysis
Explore the MSE penalty of running OAMP with an assumed Bernoulli-Gaussian sparsity when the true sparsity is . Overestimating sparsity (too small ) is typically worse than underestimating it.
Parameters
Kronecker LMMSE Speedup
Compare the wall-clock cost of a dense LMMSE solve against the factorized Kronecker solve as a function of the problem dimension. The asymptotic speedup is visible already for modest .
Parameters
Calibrating the Assumed Prior in Practice
In deployed imaging systems the true prior is never perfectly known. Two strategies mitigate mismatch:
- EM tuning (see section 21.4): alternate OAMP steps with maximum-likelihood updates of , refining the assumed prior from the data.
- Conservative default: if is uncertain, choose it on the larger side. Underestimating sparsity degrades performance gracefully; overestimating it can cause premature thresholding and support errors.
Either strategy costs some Bayes-optimality but buys robustness against distributional drift between calibration and deployment.
Kronecker-Structured OAMP for RF Imaging
The CommIT group's RF-imaging pipeline uses an OAMP iteration in which the LMMSE step is implemented by exploiting the angular delay subcarrier factorization of the physical sensing operator. Together with a Hutchinson estimator for the divergence of a learned denoiser, this makes OAMP feasible at imaging scales () where a dense solve would be infeasible. The resulting algorithm is the backbone of the unrolled-network reconstruction in Book 2 Chapter 27.
Common Mistake: Kronecker Row/Column Conventions
Mistake:
Writing without checking the vectorization convention (column-major vs row-major). A single index flip will silently swap the two Kronecker factors and produce a correct-looking but completely wrong result.
Correction:
The standard identity is under column-major vectorization. NumPy / PyTorch default to row-major, so the identity becomes in those libraries. Always write out the first three indices by hand for a test case before committing the code.
When the Structure Is Only Approximate
In practice the sensing operator is rarely exactly Kronecker. A realistic imaging pipeline has small perturbations β mutual coupling between antennas, phase noise across subcarriers, calibration errors β that break the separability. Two approaches keep things tractable:
- Absorb the perturbation into the prior by modeling it as correlated additional noise, which keeps the LMMSE solve factorized.
- Use a Kronecker preconditioner for an iterative CG solve of the exact LMMSE system. CG typically converges in 5-20 inner iterations when the preconditioner captures the dominant spectral structure.
Neither approach is uniformly best β the trade-off depends on how severely the structure is broken. A healthy implementation allows the user to switch between them.
Quick Check
If is with singular values and is with singular values , what is the largest singular value of ?
The singular values of are the pairwise products of the component singular values. The largest product is .
Hutchinson trace estimator
An unbiased randomized estimator of that requires only matrix-vector products with . Used in OAMP to estimate the denoiser divergence when no closed-form is available.
Related: Divergence-free estimator