Unrolled OAMP with ProxNet
From Iterations to Layers
Every iterative reconstruction algorithm computes a sequence of updates . The key observation of algorithm unrolling (also called deep unfolding) is that a fixed number of such iterations can be viewed as a feedforward neural network with layers. Each layer performs one iteration, and the algorithm parameters (step sizes, thresholds, penalty weights, denoiser functions) become learnable parameters optimised via backpropagation through the entire -layer graph.
This paradigm inherits the interpretability of iterative algorithms (each layer has a known function) while gaining the adaptability of neural networks (parameters are tuned from data). For RF imaging, unrolled OAMP is the natural choice: OAMP already handles the Kronecker-structured from Ch 17, and unrolling lets us learn the denoiser and step sizes end-to-end.
Definition: Algorithm Unrolling
Algorithm Unrolling
Algorithm unrolling (or deep unfolding) converts a -iteration algorithm into a -layer neural network by:
- Truncating the iteration to steps.
- Parameterising each iteration's operators with learnable parameters .
- Training the parameters end-to-end by minimising
where is the output of the -th layer and is a task loss (e.g., MSE).
Unlike generic deep networks, unrolled networks have strong inductive bias: the architecture encodes the structure of the forward model . This dramatically reduces the number of learnable parameters compared to a generic U-Net and improves sample efficiency.
Definition: OAMP Iteration (Review)
OAMP Iteration (Review)
The Orthogonal AMP (OAMP) algorithm from Ch 17 iterates two steps:
Linear Estimation (LE):
Nonlinear Estimation (NLE):
where satisfies the divergence constraint: . This ensures in the large-system limit, regardless of the structure of .
The orthogonalisation step is what makes OAMP suitable for RF imaging: standard AMP diverges for Kronecker-structured or partial-DFT sensing matrices, while OAMP's linear estimator works for any right-unitarily invariant matrix.
Definition: Learned OAMP (Unrolled OAMP with ProxNet)
Learned OAMP (Unrolled OAMP with ProxNet)
Learned OAMP unrolls iterations of OAMP into a feedforward neural network.
Layer ():
- Linear module (LMMSE): Compute the linear estimate using the known and the current noise estimate :
- Orthogonalisation (fixed): The analytical orthogonalisation formula from Ch 17.3 ensures the divergence-free condition.
- ProxNet denoiser: Replace the hand-designed with a neural network , typically a small CNN or DnCNN with parameters .
Training: Minimise the end-to-end loss:
where the expectation is over training data .
Key advantages over pure deep learning:
- Physics-informed: The linear module and orthogonalisation encode the forward model exactly.
- Fewer parameters: Only the denoiser is learned (-- parameters per layer, vs. -- for end-to-end networks).
- Interpretable: Each layer has a clear role (LMMSE + denoiser), and the state evolution provides a theoretical performance prediction.
Definition: ProxNet --- Learned Denoiser for OAMP
ProxNet --- Learned Denoiser for OAMP
ProxNet replaces the fixed denoiser in OAMP with a learned neural network :
where is the estimated noise variance at layer (provided by the state evolution). The network is typically a small U-Net or DnCNN that takes as input and produces a denoised estimate.
The divergence of must be computed for the state evolution update:
In practice, Monte Carlo divergence estimation (Hutchinson's trace estimator) provides an efficient single-backward-pass approximation using a random probe vector : .
Definition: Kronecker-Structured OAMP for RF Imaging
Kronecker-Structured OAMP for RF Imaging
When the sensing matrix has Kronecker structure (common in 2D imaging with separable Tx/Rx arrays and frequency grids), the OAMP linear estimator decomposes as:
where is the LMMSE estimator for : .
This reduces the computational cost from to (for square images with ), and from to when are partial DFT matrices (computed via FFT).
The state evolution for Kronecker-structured OAMP involves the joint singular value distribution of and .
Kronecker structure arises naturally in phased-array imaging (azimuth elevation), MIMO radar (transmit receive), and separable OFDM frequency grids. Exploiting this structure is essential for scaling unrolled OAMP to practical problem sizes.
Unrolled OAMP with ProxNet
Complexity: Per layer: for Kronecker-FFT LMMSE + for CNN denoiser. Total: .The Kronecker LMMSE step dominates the per-layer cost. For a image with layers, the total cost is comparable to 10 FFTs --- orders of magnitude faster than generic matrix inversions.
Theorem: State Evolution for Unrolled OAMP
In the large-system limit ( with ), the effective noise at each OAMP layer is characterised by a scalar state evolution:
where is the per-component MSE of the denoiser at noise level , and is the measurement noise variance.
The state evolution is exact for right-unitarily invariant matrices (which include Haar-distributed unitary, partial DFT, and Kronecker products of such matrices).
The orthogonalisation step in OAMP "Gaussianises" the residual, so the denoiser always sees signal-plus-Gaussian-noise. The state evolution tracks the noise variance through the layers, and the denoiser's performance at each noise level determines the next layer's noise level. This creates a virtuous cycle: better denoisers produce lower noise, which makes the next denoiser's job easier.
Gaussianity of the linear estimate
For right-unitarily invariant , the residual is approximately Gaussian with variance in the large-system limit. This follows from the free probability analysis of the OAMP linear estimator.
Denoiser MSE determines next variance
The NLE step produces . The error has per-component variance . The LE step at the next layer introduces .
Layer-Wise vs End-to-End Training
Two training strategies for unrolled OAMP:
Layer-wise training: Train each layer independently to minimise . Advantage: each sub-problem is small and converges quickly. Disadvantage: layers do not cooperate; early layers cannot anticipate later layers' needs.
End-to-end training: Train all layers jointly to minimise the final output loss . Advantage: globally optimal parameters; layers specialise (e.g., early layers do coarse recovery, late layers refine). Disadvantage: vanishing gradients for large ; higher memory cost.
Practical recommendation: Initialise with layer-wise pre-training (3--5 epochs per layer), then fine-tune end-to-end. Use gradient checkpointing for to manage memory.
Unrolled OAMP vs Classical OAMP
Compare the reconstruction performance of unrolled OAMP-ProxNet with classical OAMP using soft-thresholding and BM3D denoisers. The plot shows NMSE (dB) versus layer/iteration index.
Adjust the SNR and number of layers to see how the learned denoiser provides consistent gains, especially at moderate SNR where the prior mismatch of hand-designed denoisers is most costly.
Parameters
Example: Learned OAMP vs Hand-Tuned OAMP for RF Imaging
Setup: MIMO RF imaging with transmit antennas, receive antennas, subcarriers. Scene: mixture of point scatterers and extended targets. SNR = 20 dB. Training set: 10,000 synthetic scenes. All methods use iterations/layers.
Compare the reconstruction quality.
Performance comparison
| Method | NMSE (dB) | Params | Inference time |
|---|---|---|---|
| OAMP + soft threshold | 0 | 0.08 s | |
| OAMP + BM3D | 0 | 2.1 s | |
| Learned OAMP (ProxNet) | 55K | 0.12 s | |
| Pure U-Net (no physics) | 1.2M | 0.05 s | |
| ADMM-Net (unrolled ADMM) | 85K | 0.10 s |
Analysis
Learned OAMP gains 2.8 dB over OAMP+BM3D and 5.6 dB over soft thresholding. It outperforms a pure U-Net by 3.4 dB with 22 fewer parameters, demonstrating the value of encoding physics in the architecture.
Learned OAMP also beats ADMM-Net by 1.6 dB, suggesting the orthogonalisation step provides a better inductive bias than the ADMM splitting for the Kronecker-structured operator .
Backpropagation Through OAMP Layers
Gradients flow back through each layer in the standard manner:
- Through the ProxNet denoiser : standard backpropagation through the CNN.
- Through the orthogonalisation: involves and the LMMSE weight --- both differentiable.
- Through the LMMSE step: .
If is imperfectly known (calibration errors), gradients can flow through the LMMSE step to refine , enabling joint calibration and reconstruction.
Memory optimisation: For large , use gradient checkpointing --- store only every -th intermediate result and recompute the rest during backpropagation.
Unrolled OAMP with ProxNet for RF Imaging
The CommIT group developed the unrolled OAMP-ProxNet architecture specifically for RF imaging with Kronecker-structured sensing matrices. The key innovations are:
- Kronecker-LMMSE integration: exploiting the separable structure to reduce the per-layer LMMSE cost from to .
- Noise-level-aware ProxNet: the denoiser receives from the state evolution as a conditioning input, enabling a single network to handle the decreasing noise schedule across layers.
- State-evolution-guided training: the state evolution prediction is used as an auxiliary loss to regularise the learned noise variances, improving stability for small training sets.
The architecture achieves state-of-the-art performance on simulated MIMO-OFDM imaging scenes with 3--6 dB improvement over hand-tuned OAMP while using 10--100 fewer parameters than pure deep learning approaches.
ProxNet Layer Architecture
Visualise the structure of a single ProxNet layer within the unrolled OAMP network. The diagram shows how data flows from the LMMSE linear estimate through the CNN denoiser, with the noise level from state evolution conditioning the denoiser.
Adjust the number of convolutional channels and layers in the ProxNet to see how parameter count and receptive field change.
Parameters
Unrolled OAMP Processing an RF Image
Common Mistake: Learned OAMP May Not Generalise to Unseen Matrix Structures
Mistake:
Training learned OAMP on a single sensing matrix and expecting it to work on a different matrix at test time.
Correction:
Learned OAMP is typically trained for a specific measurement matrix (or family of matrices). If the test-time matrix differs:
- Different SNR: Moderate degradation; mitigated by training with a range of SNR values.
- Different : Performance can degrade significantly if the singular value distribution changes.
- Different scene statistics: The learned denoiser may perform poorly on out-of-distribution scenes.
Best practices: Train on diverse scenes, include as a denoiser input (noise-adaptive), validate on held-out data with different matrix realisations.
Quick Check
What is the primary advantage of unrolled OAMP over running classical OAMP for iterations with a fixed denoiser?
Unrolled OAMP uses fewer total FLOPs per iteration
Layer-wise learnable parameters and ProxNet enable faster per-layer convergence
Unrolled OAMP does not require a forward model
Unrolled OAMP always converges to the global minimum
With fixed parameters, every iteration uses the same denoiser and step size. Unrolling allows each layer to use parameters optimised for its position in the network: aggressive early denoising and conservative late refinement.
Algorithm Unrolling
Converting a -iteration algorithm into a -layer feedforward neural network where the algorithm parameters become learnable, trained end-to-end via backpropagation.
Related: Deep Unfolding, ProxNet
Deep Unfolding
Synonym for algorithm unrolling. The term emphasises the "unfolding" of a recursive iteration into a feedforward graph.
Related: Algorithm Unrolling
ProxNet
A learned neural network denoiser that replaces the fixed proximal operator in an unrolled optimisation algorithm, typically a small CNN conditioned on the current noise level.
Related: Algorithm Unrolling
Key Takeaway
Unrolled OAMP with ProxNet converts OAMP iterations into a trainable neural network where the LMMSE step encodes exactly and the learned CNN denoiser adapts to the signal prior. Kronecker structure reduces the per-layer cost to . The state evolution provides both a noise schedule for the denoiser and theoretical performance predictions.