Exercises
ex-ch12-01
EasyLet and . Find and verify the tower property by computing both directly and via .
For a uniform on , the mean is .
Find $\mathbb{E}[X|Y]$
, so .
Tower check
. Direct: .
ex-ch12-02
EasyProve that .
Use the law of total variance.
Variance is always non-negative.
Apply the law of total variance
. Since , we get .
ex-ch12-03
EasyLet and be independent with and . What is ? What is the MMSE?
What does independence imply for the conditional expectation?
Apply independence
. The MMSE . Observing gives no information about .
ex-ch12-04
MediumLet be jointly Gaussian with , , , , and . (a) Find given . (b) Find the MSE of the LMMSE estimator. (c) Is the LMMSE equal to the MMSE here?
Use the scalar LMMSE formula: .
Part (a)
.
Part (b)
.
Part (c)
Yes. For jointly Gaussian variables, LMMSE = MMSE.
ex-ch12-05
MediumLet equiprobably and with independent of . (a) Find . (b) Find the LMMSE estimator . (c) Compare the MSE of both estimators numerically for .
For (a), use Bayes' rule to get the posterior, then compute the posterior mean.
For (b), compute and .
Part (a): MMSE
, as derived in Example EMMSE Estimation of a Binary Signal in Gaussian Noise.
Part (b): LMMSE
(since ). . So .
Part (c): Compare
For : MMSE estimate . LMMSE estimate . The MMSE estimator uses the nonlinear to better exploit the binary structure of .
ex-ch12-06
MediumProve the orthogonality principle: for any measurable with .
Use the tower property and the 'pulling out known' property.
Apply iterated conditioning
.
Now .
ex-ch12-07
MediumLet and be i.i.d. with mean and variance , independent of . Let (with if ). Use the law of total variance to find .
Condition on . What are and ?
Conditional moments
and .
Apply total variance
.
This is Wald's identity for the variance of a random sum.
ex-ch12-08
MediumDerive the LMMSE estimator of given where , , are zero-mean with , , , , .
Write and form , then invert.
Set up
, .
Invert $\mathbf{C}_{YY}$
. .
Compute the estimator
.
MSE .
ex-ch12-09
HardShow that for any estimator , the MSE can be decomposed as
Interpret each term.
Add and subtract inside the square.
Use the orthogonality principle to show the cross-term vanishes.
Decompose
.
Cross-term vanishes
is a function of . By the orthogonality principle, for any . So the cross-term is zero.
Conclude
.
The first term is the irreducible error. The second is the penalty for using instead of the optimal . Setting eliminates the penalty.
ex-ch12-10
HardLet where , independent of , and is known. Derive the LMMSE estimator of and its MSE matrix.
Compute and .
Use the matrix inversion lemma to get a dual form.
Compute covariances
, .
LMMSE estimator
.
By the matrix inversion lemma: .
MSE matrix
.
ex-ch12-11
HardProve the conditional Jensen's inequality: if is convex and , then
Use the supporting hyperplane characterization of convexity.
For convex , there exist such that and at .
Supporting line
Since is convex, for each there exist constants such that for all , with equality at .
Condition and apply
Take . Then .
ex-ch12-12
HardLet and where independent of . (a) Find . (b) Find . (c) Is the LMMSE equal to the MMSE? Justify your answer.
.
Part (a)
.
Part (b)
. .
Part (c)
No. The joint distribution of is not Gaussian (both marginals are exponential). The conditional has a truncated distribution on , so is a nonlinear function of . LMMSE MMSE.
ex-ch12-13
MediumA sensor measures temperature with additive noise: where (in Celsius) and independent of . (a) Find the LMMSE estimate of given . (b) What fraction of the total variance of is explained by ?
This is Gaussian, so LMMSE = MMSE.
Part (a)
.
Part (b)
. Fraction explained: where , so .
ex-ch12-14
ChallengeLet in and suppose we observe a noisy linear combination where is known and is independent. Show that the MMSE estimate of given is
and find the MSE matrix.
Compute and .
Since everything is Gaussian, LMMSE = MMSE.
Cross-covariance
.
Observation variance
.
LMMSE = MMSE
.
MSE matrix: . This is a rank-one update to the prior covariance.
ex-ch12-15
Challenge(Nonlinear MMSE vs. LMMSE gap) Let and where independent of . (a) Show that depends on nonlinearly (find it numerically if needed). (b) Compute the LMMSE estimate. What is ? (c) Compare the MSE of both estimators via Monte Carlo simulation.
Note: .
For , by symmetry.
Part (b): LMMSE
, , . So . Therefore for all ! The LMMSE estimate ignores entirely because and are uncorrelated (despite being dependent).
Parts (a) and (c)
The MMSE estimator uses the posterior . This is a bimodal distribution (symmetric around 0 when ), so by symmetry as well. Both estimators give the same result here β not because LMMSE is good, but because the MMSE itself cannot distinguish from given . The MMSE error is , same as the LMMSE error.