Exercises

ex21-01-prox-denoiser

Easy

Show that prox⁑λβˆ₯β‹…βˆ₯1(v)\operatorname{prox}_{\lambda\|\cdot\|_1}(\mathbf{v}) is soft-thresholding, and interpret it as a MAP denoiser for the model v=x+n\mathbf{v} = \mathbf{x} + \mathbf{n} where n∼N(0,Ξ»I)\mathbf{n} \sim \mathcal{N}(\mathbf{0}, \lambda\mathbf{I}) and x\mathbf{x} has a Laplace prior.

ex21-02-pnp-pgd-iter

Easy

Write out the PnP-PGD iteration for the problem min⁑x12βˆ₯yβˆ’Axβˆ₯2+Ξ»R(x)\min_\mathbf{x} \tfrac{1}{2}\|\mathbf{y} - \mathbf{A}\mathbf{x}\|^2 + \lambda R(\mathbf{x}) with step size Ξ±\alpha and denoiser DΟƒ\mathcal{D}_\sigma. What is the relationship between Οƒ\sigma and Ξ»\lambda?

ex21-03-nonexpansive

Easy

Verify that soft-thresholding SΟ„\mathcal{S}_\tau is non-expansive: βˆ₯SΟ„(a)βˆ’SΟ„(b)βˆ₯≀βˆ₯aβˆ’bβˆ₯\|\mathcal{S}_\tau(\mathbf{a}) - \mathcal{S}_\tau(\mathbf{b})\| \leq \|\mathbf{a} - \mathbf{b}\| for all a,b\mathbf{a}, \mathbf{b}.

ex21-04-pnp-fourier

Easy

Derive the efficient PnP-ADMM c\mathbf{c}-update for a partial Fourier sensing matrix A=PΞ©F\mathbf{A} = \mathbf{P}_\Omega\mathbf{F} (F\mathbf{F} is unitary DFT, PΞ©\mathbf{P}_\Omega selects rows in Ξ©\Omega). Show the update requires two FFTs.

ex21-05-drunet-conditioning

Easy

Explain why the noise-level input Οƒ\sigma in DRUNet is critical for PnP applications, and describe two failure modes when Οƒ\sigma is misspecified.

ex21-06-red-gradient

Medium

For the RED regulariser RRED(x)=12xT(xβˆ’DΟƒ(x))R_\text{RED}(\mathbf{x}) = \tfrac{1}{2}\mathbf{x}^T(\mathbf{x} - \mathcal{D}_\sigma(\mathbf{x})), compute βˆ‡RRED\nabla R_\text{RED} without assuming Jacobian symmetry. Show that the result involves the Jacobian JD(x)\mathbf{J}_\mathcal{D}(\mathbf{x}).

ex21-07-noise-schedule

Medium

For PnP-ADMM with K=30K = 30 iterations, design a noise schedule {Οƒk}\{\sigma_k\} starting at Οƒ1=0.196\sigma_1 = 0.196 (i.e., 50/25550/255) and ending at Οƒ30=0.004\sigma_{30} = 0.004 (i.e., 1/2551/255). Compare geometric and cosine schedules.

ex21-08-firmly-nonexpansive

Medium

Prove that the proximal operator of any proper, lower semicontinuous, convex function ff is firmly non-expansive: βˆ₯prox⁑f(a)βˆ’prox⁑f(b)βˆ₯2+βˆ₯(Iβˆ’prox⁑f)(a)βˆ’(Iβˆ’prox⁑f)(b)βˆ₯2≀βˆ₯aβˆ’bβˆ₯2.\|\operatorname{prox}_f(\mathbf{a}) - \operatorname{prox}_f(\mathbf{b})\|^2 + \|(\mathbf{I} - \operatorname{prox}_f)(\mathbf{a}) - (\mathbf{I} - \operatorname{prox}_f)(\mathbf{b})\|^2 \leq \|\mathbf{a} - \mathbf{b}\|^2.

ex21-09-icnn-architecture

Medium

Prove that the following ICNN is convex in its input x\mathbf{x}: z1=ReLU⁑(U1x+b1),zβ„“=ReLU⁑(Wβ„“(z)zβ„“βˆ’1+Uβ„“x+bβ„“),\mathbf{z}_1 = \operatorname{ReLU}(\mathbf{U}_1\mathbf{x} + \mathbf{b}_1), \quad \mathbf{z}_\ell = \operatorname{ReLU}(\mathbf{W}_\ell^{(z)}\mathbf{z}_{\ell-1} + \mathbf{U}_\ell\mathbf{x} + \mathbf{b}_\ell), Φθ(x)=wTzD+c,\Phi_\theta(\mathbf{x}) = \mathbf{w}^T\mathbf{z}_D + c, under the constraints Wβ„“(z)β‰₯0\mathbf{W}_\ell^{(z)} \geq 0 (element-wise) and wβ‰₯0\mathbf{w} \geq 0.

ex21-10-red-fixed-point

Medium

Show that the RED-GD fixed point satisfies the optimality condition of the RED objective: AH(Axβˆ—βˆ’y)+Ξ»(xβˆ—βˆ’DΟƒ(xβˆ—))=0\mathbf{A}^{H}(\mathbf{A}\mathbf{x}^* - \mathbf{y}) + \lambda(\mathbf{x}^* - \mathcal{D}_\sigma(\mathbf{x}^*)) = \mathbf{0}.

ex21-11-pnp-complex

Hard

Extend PnP-ADMM to complex-valued RF imaging. For y=Ac+w\mathbf{y} = \mathbf{A}\mathbf{c} + \mathbf{w} with c∈CQ\mathbf{c} \in \mathbb{C}^Q:

  1. Derive the complex-valued ADMM updates.
  2. Describe how to apply a real DRUNet to the complex intermediate image.
  3. Show that non-expansiveness of the real denoiser implies non-expansiveness on the complex representation.

ex21-12-spectral-norm

Hard

A 2-layer network has weight matrices W1∈RmΓ—n\mathbf{W}_1 \in \mathbb{R}^{m \times n} and W2∈RpΓ—m\mathbf{W}_2 \in \mathbb{R}^{p \times m} and ReLU activations.

  1. Show the Lipschitz constant satisfies L≀βˆ₯W2βˆ₯βˆ₯W1βˆ₯L \leq \|\mathbf{W}_2\|\|\mathbf{W}_1\|.
  2. Describe a per-iteration power method to estimate βˆ₯Wβ„“βˆ₯\|\mathbf{W}_\ell\|.
  3. Explain why spectral normalisation preserves denoising quality better than weight clipping.

ex21-13-red-convergence

Hard

Show that RED gradient descent converges to a global minimum when RREDR_\text{RED} is convex and the total objective F(x)=12βˆ₯yβˆ’Axβˆ₯2+Ξ»RRED(x)F(\mathbf{x}) = \tfrac{1}{2}\|\mathbf{y} - \mathbf{A}\mathbf{x}\|^2 + \lambda R_\text{RED}(\mathbf{x}) is ΞΌ\mu-strongly convex. What step size is required?

ex21-14-pnp-divergence

Hard

Construct an explicit 2D example where PnP-PGD diverges. Use a linear denoiser D(x)=cx\mathcal{D}(\mathbf{x}) = c\mathbf{x} with c>1c > 1 and show that the iterates grow without bound for a specific step size Ξ±\alpha.

ex21-15-pnp-vs-oamp-theory

Challenge

In the large-system limit with Q,Mβ†’βˆžQ, M \to \infty at fixed ratio Ξ΄=M/Q\delta = M/Q, OAMP with the exact BG prior achieves the MMSE estimator for Bernoulli-Gaussian signals. Explain why PnP cannot match this guarantee in general, and identify a condition under which PnP-ADMM would achieve the MMSE.

ex21-16-icnn-expressivity

Challenge

Design an ICNN Φθ ⁣:R2β†’R\Phi_\theta \colon \mathbb{R}^2 \to \mathbb{R} that approximates the TV regulariser RTV(x)=∣x1βˆ’x2∣R_\text{TV}(\mathbf{x}) = |x_1 - x_2| for x=(x1,x2)\mathbf{x} = (x_1, x_2). Describe the architecture, constraints, and prove it is convex. Quantify the approximation error.

ex21-17-score-red-connection

Challenge

Using Tweedie's formula DΟƒ(v)=v+Οƒ2βˆ‡vlog⁑pΟƒ(v)\mathcal{D}_\sigma(\mathbf{v}) = \mathbf{v} + \sigma^2\nabla_\mathbf{v}\log p_\sigma(\mathbf{v}), show that the RED gradient descent step at x\mathbf{x} is equivalent to an approximate score-function gradient ascent step in the image distribution p(x)p(\mathbf{x}). Under what approximation does this hold?

ex21-18-pnp-rf-imaging-experiment

Medium

An RF imaging system has M=512M = 512 complex measurements of a Q=1024Q = 1024-voxel scene, modelled as y=Ac+w\mathbf{y} = \mathbf{A}\mathbf{c} + \mathbf{w} with A∈C512Γ—1024\mathbf{A} \in \mathbb{C}^{512 \times 1024} drawn i.i.d. Gaussian and Οƒ2=0.01βˆ₯Acβˆ₯2/M\sigma^2 = 0.01\|\mathbf{A}\mathbf{c}\|^2/M (SNR = 20 dB).

For PnP-ADMM with K=40K = 40, ρ=1\rho = 1, and a geometric noise schedule from Οƒ1=0.05\sigma_1 = 0.05 to Οƒ40=0.005\sigma_{40} = 0.005:

  1. Compute the ADMM penalty parameter ρ\rho that makes Οƒ1\sigma_1 consistent with Ξ»/ρ\sqrt{\lambda/\rho} for Ξ»=0.05\lambda = 0.05.
  2. Estimate the per-iteration cost in flops (assume Q=Nlog⁑NQ = N\log N for the AHA\mathbf{A}^{H}\mathbf{A} solve via FFT).
  3. Compare with LASSO (FISTA, 200 iterations, same sensing matrix).