Exercises
ex-ch09-01
EasyLet and be jointly WSS with cross-correlation and observation auto-correlation . Starting from the orthogonality principle, derive the Wiener-Hopf equation satisfied by the non-causal filter :
Write the estimation error and set .
Use WSS to collapse absolute-time indices into lag differences.
Write the error and impose orthogonality
The LMMSE estimate is . Orthogonality requires for every .
Collapse using WSS
Expanding the expectation: and . The equation becomes for all .
Remark
This is a discrete convolution equation on the full integer line. Taking the DTFT gives directly.
ex-ch09-02
EasyCompute the non-causal Wiener filter and the resulting MMSE for the signal-plus-noise model , where and are independent, zero-mean, WSS with PSDs and respectively.
Independence plus zero means gives and .
Cross and auto spectra
, so . Similarly .
Filter and MMSE
The non-causal Wiener filter is , and the MMSE is .
Operational reading
The filter is a frequency-selective attenuator: it passes frequencies where and suppresses those with .
ex-ch09-03
MediumLet be an AR(1) process with , , where is white with variance , observed in independent white noise with variance . Derive closed-form expressions for , , and the non-causal Wiener filter .
.
PSD of the AR(1) signal
Taking the transfer function acting on white , .
Observation PSD
.
Non-causal Wiener filter
.
Operational reading
For close to 1 the signal PSD concentrates near DC, and the Wiener filter becomes a low-pass filter. For near 0 the signal is nearly white and the filter collapses to the scalar MMSE gain .
ex-ch09-04
MediumProve that the non-causal MMSE can be written as Interpret the integrand.
MMSE ; apply Parseval.
Time-domain MMSE
From orthogonality, .
Parseval
and . Substituting yields .
Interpretation
The integrand is the residual PSD after projecting the desired signal onto the observation spectrum at each frequency. When (perfect coherence), the integrand vanishes and the estimate is exact.
ex-ch09-05
MediumState the Paley-Wiener condition for a PSD and explain why it is required for spectral factorization with causal and minimum-phase.
The condition is on the integrability of .
The condition
(equivalently, ).
Why it is needed
Spectral factorization constructs with Fourier coefficients of . These coefficients exist and decay only when . Without this, no causal with a causal inverse exists.
Practical consequence
A PSD that vanishes on a positive-measure set (spectral null) violates Paley-Wiener; the process is then deterministic in the forward direction and cannot be represented as the output of a causal stable filter driven by white noise.
ex-ch09-06
MediumPerform the spectral factorization of , more concretely of . Identify the minimum-phase factor and the innovations variance .
Express in : .
Find roots of ; pair inside/outside the unit circle.
Rewrite in $z$
.
Factor in terms of unit-circle roots
. Equivalently, .
Extract minimum-phase factor
Take (zero inside unit disk at , causal and stable), and . Check: .
ex-ch09-07
MediumShow that the innovations process , obtained by passing through the whitening filter , is white and has variance . Why are innovations useful for deriving the causal Wiener filter?
Compute .
Whiten
Let . In frequency, , constant. So is white with variance .
Why it helps
Because is white, projecting onto is a sequence of independent scalar projections. The causal Wiener filter is then the causal part of the cross-spectrum, normalized by : .
Remark
The bracketed operation keeps only causal (non-negative index) Fourier coefficients; it is the formal projection onto the past.
ex-ch09-08
HardDerive the causal Wiener filter for the AR(1)-in-white-noise model of Exercise ex-ch09-03 in closed form. Compute the causal MMSE and compare with the non-causal MMSE.
Spectral-factor .
Find the unique and gain such that .
Combine and factor
. The numerator is a second-order symmetric polynomial; write it as with and . Matching coefficients gives and .
Causal filter
After cancelling the common causal pole, one obtains , where arises from the operation. This is a first-order IIR smoother.
Comparison
The causal MMSE is strictly larger than the non-causal one unless (i.e. the signal is white). The gap is quantified by the Kolmogorov-Szego formula applied to the prediction-error spectrum.
ex-ch09-09
MediumState the Kolmogorov-Szego formula for the one-step prediction error variance of a purely non-deterministic WSS process, and interpret it as a geometric mean.
The formula involves .
Formula
For a purely non-deterministic WSS process with PSD a.e. satisfying Paley-Wiener, the minimum one-step prediction-error variance is .
Geometric-mean interpretation
The arithmetic mean of is (signal power). The geometric mean is . The ratio is the achievable prediction-gain: , with equality iff the process is white.
Consequence
Concentrating PSD energy (colored spectra) lowers : predictable structure reduces the innovations floor.
ex-ch09-10
EasyA random process has , , . Fit a second-order linear predictor by solving the Yule-Walker equations.
Yule-Walker: , with , .
Set up the system
.
Solve
Determinant . , .
Prediction-error variance
.
ex-ch09-11
HardConsider the two-sided noncausal smoother for with independent zero-mean WSS . Prove that the non-causal MMSE is strictly smaller than the causal MMSE whenever , and equal only in pathological cases.
Express both MMSEs via the prediction-error spectrum of the observation after scaling.
MMSE gap
One can show (Kailath-Sayed-Hassibi, Ch. 7) that , i.e. the energy of the anti-causal part of the whitened cross-spectrum.
Strict positivity
The right-hand side is non-negative and vanishes iff , that is, iff the whitened cross-correlation is purely causal. For the additive-noise model this fails unless almost everywhere.
Operational reading
Access to the future strictly improves estimation whenever signal and noise overlap in frequency; the smoothing gap is a direct measure of that overlap.
ex-ch09-12
MediumDerive the frequency response of the -step Wiener predictor of , , for a purely non-deterministic WSS with innovations representation .
The -step predictor keeps the causal part of times .
Setup
. The MMSE -step predictor of from equals the projection onto , namely .
Transfer function
Reindexing, the predictor has transfer function, in -domain, , which reduces to , where keeps non-negative powers of .
Prediction error
, which grows with and saturates at .
ex-ch09-13
MediumDerive the MMSE of the one-step predictor for an AR() process , white with variance .
For an AR() process the optimal predictor is exact in the recursion.
Predictor
The true recursion shows that the AR() predictor uses only the past samples and is optimal.
MMSE
The prediction error is , hence . This is also what the Kolmogorov-Szego formula yields: .
ex-ch09-14
EasyConsider a narrowband desired signal with PSD concentrated in , observed in independent wideband white noise of variance . Sketch the Wiener filter magnitude response and the in-band / out-of-band gains.
Use and approximate in the two regimes.
In-band
For , so (signal passes).
Out-of-band
For , so (noise suppressed).
Transition
Near the band edge the filter smoothly transitions; its sharpness depends on the roll-off of . The Wiener filter behaves as a signal-matched low-pass filter.
ex-ch09-15
MediumShow that the Wiener filter is a contraction: , where and are the spectra of and respectively. Interpret.
since in the additive-noise case.
Bound on $|H(f)|$
For with independent, .
Power inequality
. Since (adding noise adds variance), in general this bounds by the observation power. When additionally one also gets .
Interpretation
The Wiener filter is never a gain; it attenuates uniformly and preserves no noise power at frequencies where the signal is absent.
ex-ch09-16
MediumThe LMS algorithm updates a tap vector by , where . Identify the deterministic counterpart of LMS that converges to the Wiener solution, and state the standard step-size condition for mean convergence.
The deterministic (expected-value) update is steepest descent on the MSE cost.
Use the eigenvalues of .
Deterministic counterpart
Taking expectations yields , which is steepest descent on . Its fixed point solves , the Wiener-Hopf normal equations.
Step-size condition
Convergence in the mean requires . Slow-converging modes are dictated by , so the condition number controls the misadjustment-speed tradeoff.
ex-ch09-17
HardConsider the \emph{smoothing} problem: estimate from for fixed lag . Derive the frequency response of the optimal lag- smoother and show that as it converges to the non-causal Wiener filter.
Use causal Wiener with desired signal and then shift.
Reduce to causal
Estimating from data up to is equivalent (up to a delay) to causally estimating from . Apply causal Wiener with cross-spectrum .
Smoother formula
.
Limit $L\to\infty$
As the bracketed causal projection absorbs the entire spectrum , yielding , which is the non-causal Wiener filter.
Interpretation
Every extra unit of allowable delay lets the smoother exploit more of the anti-causal cross-correlation. The curve of MMSE vs. is the canonical smoothing tradeoff.
ex-ch09-18
MediumShow that the Wiener-Hopf normal equations for an FIR filter of length reduce to the matrix equation , with Hermitian Toeplitz. Discuss when is singular and how to regularize.
Truncate the infinite Wiener-Hopf equation to lags .
Finite truncation
Imposing for and and restricting orthogonality to : , i.e. with Hermitian Toeplitz.
Singularity
is singular iff lies in a proper subspace (e.g. a deterministic sum of sinusoids with fewer than components). Then the estimator is underdetermined on the null space.
Regularization
Use diagonal loading (ridge), or drop toward a shorter filter. Diagonal loading is the MAP solution under a Gaussian prior on .