Spectral Regularization Methods
Spectral Methods — Filtering in the SVD Domain
The pseudoinverse amplifies noise because it divides each SVD coefficient by , which tends to zero. The simplest remedy is to filter these divisions: replace with a bounded function that approximates for large but damps the dangerous small- components.
This leads to the general class of spectral regularization methods:
Different choices of the filter function yield different methods: truncated SVD, Tikhonov, Landweber, and many others. This unified view is powerful: it allows comparative analysis, qualification comparison, and generalisation to new methods.
Definition: Spectral Filter Function
Spectral Filter Function
A spectral filter is a family of functions satisfying:
-
Boundedness: For each , there exists such that for all .
-
Pointwise convergence: For each fixed ,
The spectral regularization associated with is
This framework unifies many seemingly different methods under a single umbrella. The filter function encodes exactly how each method treats the different singular value components.
Definition: Truncated SVD (TSVD)
Truncated SVD (TSVD)
The truncated SVD retains only the first singular components:
The filter function is the sharp cutoff:
The regularization parameter is the truncation level ; smaller means more regularization.
TSVD has infinite qualification: it achieves the minimax optimal convergence rate for any source condition order , provided is chosen optimally. The optimal truncation level satisfies approximately where bounds the source condition. Components with are dominated by noise and should be discarded.
Definition: Tikhonov Regularization
Tikhonov Regularization
The Tikhonov regularized solution is defined as the minimizer of the functional
where is the regularization parameter.
The first term is the data fidelity (how well explains the data). The second term is the penalty (how large is). The parameter controls the trade-off between fitting the data and keeping the solution small.
The Tikhonov filter function is
More generally, one can replace with for a penalty operator (e.g., a derivative operator to penalise oscillations). This is generalised Tikhonov regularization. The non-quadratic generalization (replacing by ) is treated in Section 2.6.
Theorem: Normal Equation for Tikhonov Regularization
The Tikhonov minimizer satisfies the normal equation
The operator is boundedly invertible for all , so
Compute the gradient
The functional is Fréchet differentiable with
Setting gives the normal equation.
Invertibility
For any ,
Thus is strictly positive definite and invertible with .
Theorem: SVD Form of Tikhonov Regularization
In terms of the singular system of , the Tikhonov solution is
The Tikhonov filter function is
The factor acts as a smooth low-pass filter: it passes components with and damps those with .
Diagonalise in the SVD basis
In the SVD basis, is diagonal with entries , and has components . The normal equation becomes
yielding the stated formula.
Theorem: Convergence Rate for Tikhonov Regularization
Let with (source condition of order ), and let . Then:
(a) For and :
(b) For , the rate saturates at regardless of how smooth is.
Tikhonov regularization has qualification .
Saturation for occurs because the Tikhonov filter can only approximate to first order near . It cannot fully exploit the rapid decay of for very smooth solutions. Higher-order iterated Tikhonov and TSVD overcome this limitation.
Bias–variance decomposition
$
Bound the bias
Using the source condition, . The bias term is
For : , giving bias .
Bound the variance
\sigma/(\sigma^2+\alpha) \leq 1/(2\sqrt{\alpha})$).
Optimise and conclude
Setting bias variance: gives , so and the error is .
For : the bias is (the sup saturates at ), and optimising gives and rate .
Theorem: Convergence Rate for Truncated SVD
Let with (source condition of order ). Choose the truncation level such that . Then
This rate is minimax optimal for all .
Bias–variance decomposition
$
Bound the bias
Using the source condition:
Bound the variance and optimise
\sigma_K^{2\mu} E^2 + K\delta^2/\sigma_K^2\sigma_K \sim (\delta/E)^{1/(2\mu+1)}\blacksquare$
Definition: Landweber Iteration as Spectral Regularization
Landweber Iteration as Spectral Regularization
The Landweber iteration for is
where is a step-size parameter. After iterations, the spectral filter function is
The iteration count plays the role of : early stopping provides regularization. As , and we recover the pseudoinverse (which diverges for noisy data).
Landweber iteration is computationally attractive because it requires only applications of and — no matrix factorizations. This is crucial for large-scale imaging problems where is only available as a matrix-free operator (e.g., a fast forward model). It has infinite qualification, matching TSVD in theoretical performance, but may converge slowly in practice. The first Landweber iterate is , which is simply the backprojection image — the standard starting point in SAR and CT imaging.
Example: Bayesian Interpretation of Tikhonov Regularization
Show that the Tikhonov solution is the maximum a posteriori (MAP) estimate under Gaussian assumptions:
- Likelihood:
- Prior:
Identify the relationship between and the noise/prior variances.
Write the posterior
By Bayes' theorem, the log-posterior is (up to constants):
Maximise
Maximising is equivalent to minimising
This is exactly the Tikhonov functional with
Interpretation
The regularization parameter encodes the signal-to-noise ratio:
- High noise ( large) large strong regularization (trust the prior more).
- Confident prior ( small) large the solution is pulled toward zero.
This Bayesian viewpoint motivates more sophisticated priors (sparsity, total variation) that lead to the variational methods of Section 2.6.
Spectral Filter Functions Compared
Compares the spectral filter functions of three regularization methods applied to the same ill-posed problem.
Top-left: The filter vs. singular value index . The ideal (pseudoinverse) filter is (shown dashed), which diverges. Each method approximates this curve for large but deviates for small ones.
Top-right: SVD coefficients of exact data, noisy data, and regularized reconstruction. The regularized coefficients should follow exact data for small and suppress noise for large .
Bottom: The reconstruction compared to the true solution.
Truncated SVD gives a sharp cutoff (all-or-nothing); Tikhonov gives a smooth roll-off; Landweber (with iterations) gives a polynomial filter that gradually engages higher frequencies.
Parameters
Tikhonov Reconstruction — Bias–Variance Trade-Off
Sweep the regularization parameter to observe the bias–variance trade-off directly.
Top panel: True solution (blue), reconstruction (red), noisy data (gray dots).
Bottom-left: Reconstruction error vs. (log scale), showing the characteristic U-shape with a minimum at the optimal .
Bottom-right: The Tikhonov filter function for each singular value component.
For very small the reconstruction is noisy (variance-dominated); for very large it is over-smoothed (bias-dominated). The optimal sits at the bottom of the U-curve.
Parameters
Spectral Regularization Methods Compared
| Property | Truncated SVD | Tikhonov | Landweber |
|---|---|---|---|
| Filter | Sharp cutoff at | Smooth roll-off | Polynomial |
| Qualification | Infinite | Infinite | |
| Parameter | Truncation level | Iteration count | |
| Computation | Full SVD | Normal equation solve | Matrix-free per step |
| Large-scale problems | Not suitable | Moderate | Ideal |
| Sensitivity to parameter | High (near gaps) | Smooth | Moderate (semi-convergence) |
Common Mistake: TSVD Is Sensitive to Singular Value Clustering
Mistake:
Choosing the truncation level for TSVD based solely on the noise level, without examining the singular value distribution.
Correction:
When singular values form clusters separated by gaps, truncating within a cluster can lead to artefacts. The optimal truncation should respect the natural gaps in the singular value spectrum.
For example, if and (a tight cluster), truncating at discards a component almost as strong as those retained. A better strategy is to look for a gap in the singular value spectrum and truncate there, or to use Tikhonov regularization which provides a smooth transition.
In imaging practice, the singular value spectrum often has a knee point where the decay steepens — this is typically the optimal truncation region.
Common Mistake: Tikhonov Cannot Exploit High-Smoothness Solutions
Mistake:
Applying standard Tikhonov regularization to an inverse problem where the true solution is very smooth (), expecting to achieve faster convergence rates than .
Correction:
Standard Tikhonov has qualification : for solutions satisfying a source condition of order , the rate saturates at regardless of how much smoothness is available.
For such problems, use iterated Tikhonov (apply the normal equation times, increasing qualification to ), TSVD, or Landweber iteration with early stopping — all of which have infinite qualification and can exploit arbitrary smoothness.
Quick Check
For Tikhonov regularization with parameter , what fraction of the signal energy is passed for a component with ?
100%
50%
25%
0%
The Tikhonov filter is (50%). The transition occurs at , which is exactly our value. Components with pass nearly unattenuated; those with are heavily damped.
Key Takeaway
All spectral regularization methods modify the pseudoinverse by replacing with a bounded filter . Truncated SVD uses a sharp cutoff (infinite qualification; optimal but sensitive to the spectrum). Tikhonov uses a smooth roll-off at (finite qualification ; closed-form solution via the normal equation). Landweber iteration uses a polynomial filter with the iteration count as regularization parameter (infinite qualification; only needs matrix-vector products — ideal for large-scale imaging). The Bayesian interpretation identifies Tikhonov as MAP estimation with a Gaussian prior: .