Exercises

ex20-01-mf-decomposition

Easy

Given a sensing matrix A∈CMΓ—N\mathbf{A} \in \mathbb{C}^{M \times N} and measurements y=Ac+w\mathbf{y} = \mathbf{A}\mathbf{c} + \mathbf{w}, write down the matched-filter image c^BP\hat{\mathbf{c}}^{\text{BP}} and show that it decomposes into a signal term and a noise term. Identify the Gram matrix and characterise the covariance of the noise term.

ex20-02-dc-idempotent

Easy

Verify that the hard data-consistency layer DC(c^)=c^βˆ’AH(Ac^βˆ’y)\text{DC}(\hat{\mathbf{c}}) = \hat{\mathbf{c}} - \mathbf{A}^{H}(\mathbf{A}\hat{\mathbf{c}} - \mathbf{y}) is idempotent when AAH=I\mathbf{A}\mathbf{A}^{H} = \mathbf{I}, i.e., DC(DC(c^))=DC(c^)\text{DC}(\text{DC}(\hat{\mathbf{c}})) = \text{DC}(\hat{\mathbf{c}}).

ex20-03-loss-estimator

Easy

A post-processing network fΞΈf_\theta is trained with the MSE loss L(ΞΈ)=E[βˆ₯fΞΈ(c^BP)βˆ’cβˆ₯2]\mathcal{L}(\theta) = \mathbb{E}[\|f_\theta(\hat{\mathbf{c}}^{\text{BP}}) - \mathbf{c}\|^2].

(a) State the optimal network fΞΈβˆ—f_{\theta^*} in the limit of infinite data and network capacity.

(b) For a 1D binary scene c∈{βˆ’1,+1}c \in \{-1, +1\} with equal probability and measurement y=c+ny = c + n with n∼N(0,Οƒ2)n \sim \mathcal{N}(0, \sigma^2), compute fΞΈβˆ—(y)=E[c∣y]f_{\theta^*}(y) = \mathbb{E}[c \mid y].

ex20-04-gram-random

Medium

Let A∈CMΓ—N\mathbf{A} \in \mathbb{C}^{M \times N} have i.i.d. entries Aij∼CN(0,1/M)A_{ij} \sim \mathcal{CN}(0, 1/M). Show that:

(a) E[Gii]=1\mathbb{E}[G_{ii}] = 1 for all ii. (b) E[Gij]=0\mathbb{E}[G_{ij}] = 0 for iβ‰ ji \neq j. (c) Var⁑(Gij)=1/M\operatorname{Var}(G_{ij}) = 1/M for iβ‰ ji \neq j.

Conclude that Gβ‰ˆI\mathbf{G} \approx \mathbf{I} for large MM and state the implication for MFβ†’U-Net performance.

ex20-05-dc-mri

Medium

In MRI, the sensing operator is A=PΞ©F\mathbf{A} = \mathbf{P}_\Omega\mathbf{F} where F∈CNΓ—N\mathbf{F} \in \mathbb{C}^{N \times N} is the DFT matrix and PΞ©\mathbf{P}_\Omega selects MM rows.

(a) Show that AAH=IM\mathbf{A}\mathbf{A}^{H} = \mathbf{I}_M (orthonormal rows). (b) Write the explicit form of the hard DC layer for this operator. (c) Interpret the DC layer as a k-space replacement operation.

ex20-06-modl-cg

Medium

The MoDL data-consistency step solves c^k=(AHA+Ξ»kI)βˆ’1(AHy+Ξ»kzk)\hat{\mathbf{c}}_k = (\mathbf{A}^{H}\mathbf{A} + \lambda_k\mathbf{I})^{-1}(\mathbf{A}^{H}\mathbf{y} + \lambda_k\mathbf{z}_k).

(a) Show that this is the solution to the regularised least-squares problem min⁑cβˆ₯Acβˆ’yβˆ₯2+Ξ»kβˆ₯cβˆ’zkβˆ₯2\min_{\mathbf{c}} \|\mathbf{A}\mathbf{c} - \mathbf{y}\|^2 + \lambda_k\|\mathbf{c} - \mathbf{z}_k\|^2.

(b) For Ξ»kβ†’0\lambda_k \to 0, what does the solution approach? For Ξ»kβ†’βˆž\lambda_k \to \infty, what does it approach?

ex20-07-perceptual-pseudo-metric

Medium

Show that the perceptual loss Lperc(c^,c)\mathcal{L}_{\text{perc}}(\hat{\mathbf{c}}, \mathbf{c}) is not a true metric on the image space by providing a counterexample where Lperc=0\mathcal{L}_{\text{perc}} = 0 but c^β‰ c\hat{\mathbf{c}} \neq \mathbf{c}. Explain the practical implication for RF imaging.

ex20-08-transfer-bound

Medium

A U-Net fΞΈf_\theta is trained on data from sensing matrix A1\mathbf{A}_{1} with Gram matrix G1\mathbf{G}_1. At deployment, the sensing matrix is A2\mathbf{A}_{2} with Gram matrix G2\mathbf{G}_2. Derive an upper bound on the deployment reconstruction error in terms of the training error and βˆ₯G1βˆ’G2βˆ₯\|\mathbf{G}_1 - \mathbf{G}_2\|. Assume fΞΈf_\theta is Lipschitz with constant LfL_f.

ex20-09-sidelobe-correlation

Hard

Consider a scene with a single strong reflector at position i0i_0: c=c ei0\mathbf{c} = c\,\mathbf{e}_{i_0} with c∈Cc \in \mathbb{C}. The matched-filter image is c^BP=Gc+w~\hat{\mathbf{c}}^{\text{BP}} = \mathbf{G}\mathbf{c} + \tilde{\mathbf{w}}.

(a) Write the expression for the sidelobe at pixel j≠i0j \neq i_0 in c^BP\hat{\mathbf{c}}^{\text{BP}}.

(b) Show that the covariance between the sidelobe at pixel jj and the back-projected noise at pixel jj is Οƒ2 Gji0Gji0βˆ—\sigma^2\,G_{ji_0}G_{ji_0}^*.

(c) Explain why this correlation makes it impossible for the U-Net to distinguish sidelobes from real features at pixel jj using c^BP\hat{\mathbf{c}}^{\text{BP}} alone.

ex20-10-modl-convergence

Hard

Consider the MoDL iteration with a fixed denoiser DΞΈ(c)=proxΞ»R(c)\mathcal{D}_\theta(\mathbf{c}) = \text{prox}_{\lambda R}(\mathbf{c}) for a convex regulariser RR. Show that the MoDL iteration:

c^k+1=(AHA+ΞΌI)βˆ’1(AHy+μ proxΞ»R(c^k))\hat{\mathbf{c}}_{k+1} = (\mathbf{A}^{H}\mathbf{A} + \mu\mathbf{I})^{-1}(\mathbf{A}^{H}\mathbf{y} + \mu\,\text{prox}_{\lambda R}(\hat{\mathbf{c}}_k))

is equivalent to a proximal gradient step, and give conditions under which the iteration converges to the minimiser of 12βˆ₯Acβˆ’yβˆ₯2+Ξ»R(c)\frac{1}{2}\|\mathbf{A}\mathbf{c} - \mathbf{y}\|^2 + \lambda R(\mathbf{c}).

ex20-11-unet-receptive-field

Hard

A U-Net for post-processing has LL encoder levels, each performing 2Γ—2\times downsampling followed by two 3Γ—33 \times 3 convolutions. Derive the receptive field diameter rLr_L. For L=4L = 4, what is the maximum sidelobe range that can be suppressed by this network, and what does this imply for long-range sidelobes in SAR?

ex20-12-physics-channel-benefit

Hard

Consider a linear Gaussian model: c∼CN(0,Οƒc2I)\mathbf{c} \sim \mathcal{CN}(\mathbf{0}, \sigma_c^2\mathbf{I}) and w∼CN(0,Οƒ2I)\mathbf{w} \sim \mathcal{CN}(\mathbf{0}, \sigma^2\mathbf{I}).

(a) Compute the MMSE estimator c^blind=E[c∣c^BP]\hat{\mathbf{c}}^{\text{blind}} = \mathbb{E}[\mathbf{c} \mid \hat{\mathbf{c}}^{\text{BP}}] and its MSE when G\mathbf{G} is known.

(b) Compute the MMSE estimator c^informed=E[c∣c^BP,G]\hat{\mathbf{c}}^{\text{informed}} = \mathbb{E}[\mathbf{c} \mid \hat{\mathbf{c}}^{\text{BP}}, \mathbf{G}] and its MSE when both c^BP\hat{\mathbf{c}}^{\text{BP}} and G\mathbf{G} are given.

(c) Show that the informed estimator achieves lower MSE whenever G≠cI\mathbf{G} \neq c\mathbf{I} for any scalar cc.

ex20-13-modl-optimal-lambda

Challenge

In MoDL with a linear denoiser DΞΈ(c)=Dc\mathcal{D}_\theta(\mathbf{c}) = \mathbf{D}\mathbf{c} (e.g., a linear Wiener denoiser), find the optimal step size Ξ»βˆ—\lambda^* that minimises the one-step reconstruction MSE

MSE(Ξ»)=E[βˆ₯(AHA+Ξ»I)βˆ’1(AHy+Ξ»Dc0)βˆ’cβˆ₯2],\text{MSE}(\lambda) = \mathbb{E}\bigl[\|(\mathbf{A}^{H}\mathbf{A} + \lambda\mathbf{I})^{-1}(\mathbf{A}^{H}\mathbf{y} + \lambda\mathbf{D}\mathbf{c}_{0}) - \mathbf{c}\|^2\bigr],

where c0=c+Ο΅\mathbf{c}_{0} = \mathbf{c} + \boldsymbol{\epsilon} is a noisy initial estimate with ϡ∼CN(0,σϡ2I)\boldsymbol{\epsilon} \sim \mathcal{CN}(\mathbf{0}, \sigma_\epsilon^2\mathbf{I}). Express Ξ»βˆ—\lambda^* in terms of the eigenvalues of AHA\mathbf{A}^{H}\mathbf{A} and D\mathbf{D}.

ex20-14-gan-hallucination

Challenge

Construct a formal example showing that a GAN-trained reconstruction network can produce a hallucinated target. Consider a two-pixel scene c∈{(1,0)⊀, (0,1)⊀}\mathbf{c} \in \{(1,0)^\top,\, (0,1)^\top\} with equal probability and a single measurement y=c1+c2+ny = c_1 + c_2 + n with n∼N(0,Οƒ2)n \sim \mathcal{N}(0, \sigma^2).

(a) Show that the measurement provides no information about which pixel is active. (b) Compare the MSE-trained and GAN-trained network outputs. (c) Explain which is more dangerous for radar target detection and why.

ex20-15-geometry-generalisation

Challenge

A physics-informed U-Net takes inputs (c^BP,diag⁑(G))(\hat{\mathbf{c}}^{\text{BP}}, \operatorname{diag}(\mathbf{G})) and is trained on a family of MIMO sensing operators {A(Ξ±)}\{\mathbf{A}(\boldsymbol{\alpha})\} parameterised by array geometry Ξ±\boldsymbol{\alpha} (e.g., antenna positions). Using PAC-Bayes theory, derive a generalisation bound on the expected reconstruction error for a new geometry Ξ±βˆ—\boldsymbol{\alpha}^* not seen during training.

Express the bound in terms of: training error, number of training geometries TT, network complexity (number of parameters PP), and scene dimension NN.