Physics-Informed Post-Processing
Giving the Network an Explicit Map of the Physics
The MF→U-Net pipeline uses only once (to form ) and discards it. MoDL re-applies at each CG step but only for data consistency — the denoiser still operates without knowledge of the PSF.
A complementary strategy is physics-informed post-processing: augmenting the U-Net input with additional channels that encode the sensing geometry explicitly. Instead of learning to infer the PSF from training data alone, the network is handed the relevant physics on a plate:
- The PSF image (per-pixel energy of the Gram matrix)
- The residual back-projection
- The noise level map
- Geometry embeddings encoding array positions, frequencies, or Tx–Rx pairs
These inputs give the U-Net explicit information about the forward model without requiring iterative computations at test time.
Definition: Physics-Informed Post-Processing Network
Physics-Informed Post-Processing Network
A physics-informed post-processing network augments the matched-filter image with additional physics-derived input channels:
The network input is a multi-channel tensor where each channel encodes a different aspect of the measurement physics. The architecture may remain a standard U-Net, ResNet, or attention-based model, but the input representation is enriched.
This approach occupies a middle ground between pure post-processing (Section 20.1) and full MoDL unrolling (Section 20.2). It incorporates physics without requiring differentiable forward/adjoint operators during inference, making it compatible with non-differentiable legacy simulators.
Example: Residual-Feedback Post-Processing
Consider a two-pass architecture:
Pass 1:
Pass 2:
Show that the second input to pass 2 is the negative gradient of the data-fidelity term evaluated at , and explain the connection to MoDL.
Identify the data-fidelity gradient
The data-fidelity term is . Its gradient with respect to is
Evaluated at , this is .
Interpret as a learned gradient-descent step
The second input to the network is exactly the negative gradient of data fidelity at . Pass 2 receives:
- The current estimate
- The descent direction for improving data consistency
The network learns to combine these into an improved estimate, effectively performing one step of learned gradient descent with an adaptive step size embedded in .
Connection to MoDL
Stacking such passes yields a cascade network:
This is a precursor to MoDL: MoDL replaces the single gradient step with a CG solve that takes multiple conjugate gradient steps to minimise the data-consistency objective exactly. Both architectures share the principle of injecting physics-derived information (the gradient or residual) into the network at each stage.
Theorem: Conditioning on PSF Reduces the Deconvolution Task
Let and denote reconstruction networks with and without PSF conditioning. Under a Gaussian scene prior and noise , the MMSE estimator given achieves strictly lower MSE than the MMSE estimator given only , unless is a multiple of .
If the network does not know the PSF, it must estimate it implicitly from the input image — a harder task. When is provided as an additional channel, the network can focus on denoising and regularisation, delegating the geometry-dependent deconvolution to a simple closed-form computation.
Data-processing inequality
Conditioning on strictly more information cannot increase the MMSE. Since contains strictly more information about than alone (unless is a deterministic function of for all scenes), the conditional MMSE
Strict inequality for non-trivial PSF
When , the PSF diagonal carries information about the per-pixel SNR not present in : pixel in has SNR , which varies with . Knowing this map enables the network to weight contributions from high-SNR pixels more strongly.
Historical Note: Physics-Informed Neural Networks in Imaging — A Brief History
2016–2024The idea of embedding physics into neural networks for imaging predates the deep learning era. In X-ray computed tomography, filtered back-projection (FBP) — a physics-derived preprocessing step — has been combined with learned post-processors since at least 2017 (Jin et al.).
The term "physics-informed neural network" (PINN) was popularised by Raissi et al. (2019) in the context of PDE-constrained learning, but its spirit in imaging is older: the DeepMedic (2017) and Cascade Net (Schlemper et al., 2018) architectures for MRI both embed measurement operators into the network graph.
For RF imaging specifically, the challenge is that the sensing operator varies with array geometry, frequency, and scene position — making it more like a family of forward models than a single fixed one. Geometry-conditioned networks that take array positions as part of their input represent the current frontier, bridging learned reconstruction with the array signal processing tradition.
Conditioning on Sensing Geometry for Generalisation
A physics-informed network that takes only the matched-filter image and the PSF diagonal still cannot generalise to entirely new sensing geometries without retraining, because the full off-diagonal structure of matters for deconvolution and is not captured by alone.
Geometry-conditioned networks take an additional input that encodes the sensing configuration:
- Explicit antenna positions: as a fixed-length embedding computed from the array geometry.
- Frequency set: the set of OFDM subcarrier frequencies used.
- PSF slice: a small representative region of near the scene centre (cheaper to compute than the full Gram matrix).
These embeddings are fed into the network via feature-wise linear modulation (FiLM) or as additional input channels. The result is a single network that generalises across operator families, adapting its deconvolution strategy on-the-fly.
This is the architecture principle behind conditional reconstruction networks for RF imaging — a research direction currently explored by the CommIT group.
Physics-Informed Post-Processing with PSF Conditioning
Compare reconstruction quality as physics-derived channels are added to the network input. "MF only" uses just . "MF + PSF" adds the PSF diagonal as a second channel. "MF + PSF + residual" further adds the residual gradient .
For the physical sensing matrix, each additional channel measurably reduces the reconstruction error — especially near bright point targets where sidelobe artefacts are strongest.
Parameters
Common Mistake: Computing the Full Gram Matrix Instead of Just What Is Needed
Mistake:
Naively computing the full Gram matrix at inference time to provide the PSF as a physics channel to the network.
Correction:
For realistic scene sizes ( voxels), has million complex entries, which is prohibitively expensive to form and store.
In practice, only the PSF of a central point reflector is needed: this is the column of corresponding to the scene centre, computed as — a single matrix-vector product. This single column captures the dominant sidelobe structure for shift-invariant (or slowly varying) PSFs.
For the PSF diagonal: (column-wise squared norms of ) — computed in one pass in time.
Quick Check
Why does providing the PSF diagonal as an additional input channel help a post-processing network?
It allows the network to compute the full inverse of
It tells the network the per-pixel SNR, enabling geometry-aware weighting
It forces the network to produce measurement-consistent output
It replaces the need for a data-consistency layer
The SNR at pixel in the matched-filter image is approximately . By knowing , the network can weight high-SNR pixels more strongly and rely less on its learned prior for well-illuminated scene regions.
Key Takeaway
-
Physics-informed post-processing augments the U-Net input with physics-derived channels: PSF diagonal, residual gradient, noise level map, and geometry embeddings.
-
Residual feedback gives the network an explicit data-consistency correction direction, bridging pure post-processing and iterative MoDL.
-
By the data-processing inequality, conditioning on the PSF strictly reduces the MMSE (unless ).
-
Geometry-conditioned networks use array position and frequency embeddings to generalise across operator configurations without retraining.
-
The full Gram matrix is never needed: the PSF diagonal is , computed in .