Physics-Informed Post-Processing

Giving the Network an Explicit Map of the Physics

The MF→U-Net pipeline uses A\mathbf{A} only once (to form c^BP\hat{\mathbf{c}}^{\text{BP}}) and discards it. MoDL re-applies A\mathbf{A} at each CG step but only for data consistency — the denoiser still operates without knowledge of the PSF.

A complementary strategy is physics-informed post-processing: augmenting the U-Net input with additional channels that encode the sensing geometry explicitly. Instead of learning to infer the PSF from training data alone, the network is handed the relevant physics on a plate:

  • The PSF image diag(G)\operatorname{diag}(\mathbf{G}) (per-pixel energy of the Gram matrix)
  • The residual back-projection AH(yAc^0)\mathbf{A}^{H}(\mathbf{y} - \mathbf{A}\hat{\mathbf{c}}_0)
  • The noise level map σ2diag(G)\sigma^2\operatorname{diag}(\mathbf{G})
  • Geometry embeddings encoding array positions, frequencies, or Tx–Rx pairs

These inputs give the U-Net explicit information about the forward model without requiring iterative computations at test time.

,

Definition:

Physics-Informed Post-Processing Network

A physics-informed post-processing network augments the matched-filter image with additional physics-derived input channels:

c^=fθ ⁣(c^BPMF image,    diag(G)PSF diagonal,    AH(yAc^BP)residual gradient,    ).\hat{\mathbf{c}} = f_\theta\!\bigl( \underbrace{\hat{\mathbf{c}}^{\text{BP}}}_{\text{MF image}},\;\; \underbrace{\operatorname{diag}(\mathbf{G})}_{\text{PSF diagonal}},\;\; \underbrace{\mathbf{A}^{H}(\mathbf{y} - \mathbf{A}\hat{\mathbf{c}}^{\text{BP}})}_{\text{residual gradient}},\;\; \ldots \bigr).

The network input is a multi-channel tensor where each channel encodes a different aspect of the measurement physics. The architecture may remain a standard U-Net, ResNet, or attention-based model, but the input representation is enriched.

This approach occupies a middle ground between pure post-processing (Section 20.1) and full MoDL unrolling (Section 20.2). It incorporates physics without requiring differentiable forward/adjoint operators during inference, making it compatible with non-differentiable legacy simulators.

Example: Residual-Feedback Post-Processing

Consider a two-pass architecture:

Pass 1: c^0=fθ1(c^BP)\hat{\mathbf{c}}_0 = f_{\theta_1}(\hat{\mathbf{c}}^{\text{BP}})

Pass 2: c^=fθ2 ⁣(c^0,  AH(yAc^0))\hat{\mathbf{c}} = f_{\theta_2}\!\bigl(\hat{\mathbf{c}}_0,\; \mathbf{A}^{H}(\mathbf{y} - \mathbf{A}\hat{\mathbf{c}}_0)\bigr)

Show that the second input to pass 2 is the negative gradient of the data-fidelity term 12yAc2\frac{1}{2}\|\mathbf{y} - \mathbf{A}\mathbf{c}\|^2 evaluated at c^0\hat{\mathbf{c}}_0, and explain the connection to MoDL.

Theorem: Conditioning on PSF Reduces the Deconvolution Task

Let fθblind(c^BP)f_\theta^{\text{blind}}(\hat{\mathbf{c}}^{\text{BP}}) and fθinformed(c^BP,diag(G))f_\theta^{\text{informed}}(\hat{\mathbf{c}}^{\text{BP}}, \operatorname{diag}(\mathbf{G})) denote reconstruction networks with and without PSF conditioning. Under a Gaussian scene prior cCN(0,σx2I)\mathbf{c} \sim \mathcal{CN}(\mathbf{0}, \sigma_x^2\mathbf{I}) and noise wCN(0,σ2I)\mathbf{w} \sim \mathcal{CN}(\mathbf{0}, \sigma^2\mathbf{I}), the MMSE estimator given (c^BP,diag(G))(\hat{\mathbf{c}}^{\text{BP}}, \operatorname{diag}(\mathbf{G})) achieves strictly lower MSE than the MMSE estimator given only c^BP\hat{\mathbf{c}}^{\text{BP}}, unless G\mathbf{G} is a multiple of I\mathbf{I}.

If the network does not know the PSF, it must estimate it implicitly from the input image — a harder task. When G\mathbf{G} is provided as an additional channel, the network can focus on denoising and regularisation, delegating the geometry-dependent deconvolution to a simple closed-form computation.

Historical Note: Physics-Informed Neural Networks in Imaging — A Brief History

2016–2024

The idea of embedding physics into neural networks for imaging predates the deep learning era. In X-ray computed tomography, filtered back-projection (FBP) — a physics-derived preprocessing step — has been combined with learned post-processors since at least 2017 (Jin et al.).

The term "physics-informed neural network" (PINN) was popularised by Raissi et al. (2019) in the context of PDE-constrained learning, but its spirit in imaging is older: the DeepMedic (2017) and Cascade Net (Schlemper et al., 2018) architectures for MRI both embed measurement operators into the network graph.

For RF imaging specifically, the challenge is that the sensing operator A\mathbf{A} varies with array geometry, frequency, and scene position — making it more like a family of forward models than a single fixed one. Geometry-conditioned networks that take array positions as part of their input represent the current frontier, bridging learned reconstruction with the array signal processing tradition.

,

Conditioning on Sensing Geometry for Generalisation

A physics-informed network that takes only the matched-filter image and the PSF diagonal still cannot generalise to entirely new sensing geometries without retraining, because the full off-diagonal structure of G\mathbf{G} matters for deconvolution and is not captured by diag(G)\operatorname{diag}(\mathbf{G}) alone.

Geometry-conditioned networks take an additional input that encodes the sensing configuration:

  • Explicit antenna positions: {(si,rj)}i,j\{(\mathbf{s}_{i}, \mathbf{r}_{j})\}_{i,j} as a fixed-length embedding computed from the array geometry.
  • Frequency set: the set of OFDM subcarrier frequencies used.
  • PSF slice: a small representative region of G\mathbf{G} near the scene centre (cheaper to compute than the full Gram matrix).

These embeddings are fed into the network via feature-wise linear modulation (FiLM) or as additional input channels. The result is a single network that generalises across operator families, adapting its deconvolution strategy on-the-fly.

This is the architecture principle behind conditional reconstruction networks for RF imaging — a research direction currently explored by the CommIT group.

Physics-Informed Post-Processing with PSF Conditioning

Compare reconstruction quality as physics-derived channels are added to the network input. "MF only" uses just c^BP\hat{\mathbf{c}}^{\text{BP}}. "MF + PSF" adds the PSF diagonal diag(G)\operatorname{diag}(\mathbf{G}) as a second channel. "MF + PSF + residual" further adds the residual gradient AH(yAc^BP)\mathbf{A}^{H}(\mathbf{y} - \mathbf{A}\hat{\mathbf{c}}^{\text{BP}}).

For the physical sensing matrix, each additional channel measurably reduces the reconstruction error — especially near bright point targets where sidelobe artefacts are strongest.

Parameters
20

Common Mistake: Computing the Full Gram Matrix Instead of Just What Is Needed

Mistake:

Naively computing the full N×NN \times N Gram matrix G=AHA\mathbf{G} = \mathbf{A}^{H}\mathbf{A} at inference time to provide the PSF as a physics channel to the network.

Correction:

For realistic scene sizes (N=64×64=4096N = 64 \times 64 = 4096 voxels), G\mathbf{G} has N216.8N^2 \approx 16.8 million complex entries, which is prohibitively expensive to form and store.

In practice, only the PSF of a central point reflector is needed: this is the column of G\mathbf{G} corresponding to the scene centre, computed as AHAeN/2\mathbf{A}^{H}\mathbf{A}\mathbf{e}_{N/2} — a single matrix-vector product. This single column captures the dominant sidelobe structure for shift-invariant (or slowly varying) PSFs.

For the PSF diagonal: diag(G)=Acol2\operatorname{diag}(\mathbf{G}) = \|\mathbf{A}\|_{\text{col}}^2 (column-wise squared norms of A\mathbf{A}) — computed in one pass in O(MN)\mathcal{O}(MN) time.

Quick Check

Why does providing the PSF diagonal diag(G)\operatorname{diag}(\mathbf{G}) as an additional input channel help a post-processing network?

It allows the network to compute the full inverse of G\mathbf{G}

It tells the network the per-pixel SNR, enabling geometry-aware weighting

It forces the network to produce measurement-consistent output

It replaces the need for a data-consistency layer

Key Takeaway

  1. Physics-informed post-processing augments the U-Net input with physics-derived channels: PSF diagonal, residual gradient, noise level map, and geometry embeddings.

  2. Residual feedback AH(yAc^)\mathbf{A}^{H}(\mathbf{y} - \mathbf{A}\hat{\mathbf{c}}) gives the network an explicit data-consistency correction direction, bridging pure post-processing and iterative MoDL.

  3. By the data-processing inequality, conditioning on the PSF strictly reduces the MMSE (unless GI\mathbf{G} \propto \mathbf{I}).

  4. Geometry-conditioned networks use array position and frequency embeddings to generalise across operator configurations without retraining.

  5. The full Gram matrix is never needed: the PSF diagonal is diag(G)=Acol2\operatorname{diag}(\mathbf{G}) = \|\mathbf{A}\|_{\text{col}}^2, computed in O(MN)\mathcal{O}(MN).