Rao--Blackwell, Lehmann--Scheffe, and the MVUE
A Procedure for Building the MVUE
The previous section gave us sufficient statistics --- lossless compressors for . This section converts them into estimators. The Rao--Blackwell theorem is the mechanism: it takes any unbiased estimator and projects it onto the sufficient -algebra to produce a better (smaller-variance) unbiased estimator. When the sufficient statistic is additionally complete, Lehmann--Scheffe certifies the projected estimator as the MVUE --- unique over all unbiased competitors. This is the textbook recipe every receiver designer walks through (often without naming the theorems).
Theorem: Rao--Blackwell Theorem
Let be an unbiased estimator of with finite second moment, and let be a sufficient statistic for . Define Then is: (i) a statistic (does not depend on ); (ii) unbiased: ; (iii) variance-improving: , with equality iff almost surely.
The conditional expectation is an orthogonal projection: it is the closest function of to the original estimator. Projection never increases the norm, which is exactly the variance reduction. The sufficiency assumption is what guarantees the conditional expectation does not depend on --- so the projected estimator is an honest statistic, computable from the data alone.
Why $\tilde{g}$ does not depend on $\theta$
Because is sufficient, the conditional distribution does not depend on . Hence neither does . We may write without a subscript.
Unbiasedness by tower
By the tower property, .
Variance reduction by Jensen (or conditional variance identity)
Using the law of total variance, , with equality iff the conditional variance is zero a.s., i.e., is a.s. a function of .
Historical Note: Rao (1945), Blackwell (1947), Scheffe (1950)
1945--1950C. R. Rao's 1945 paper introduced the inequality and the conditioning trick for a single statistic. D. Blackwell's 1947 note showed the same procedure works for any convex loss function and clarified its orthogonal-projection character. E. L. Lehmann and H. Scheffe closed the circle in 1950 by adding the completeness requirement, producing the uniqueness statement: there is at most one unbiased estimator that is a function of the complete sufficient statistic, and it is the MVUE. These three short papers (Rao's is four pages) gave point estimation its modern skeleton.
Theorem: Lehmann--Scheffe Theorem
Let be a complete sufficient statistic for the family . If is an unbiased estimator of that is a function of , then is the unique (almost-surely) minimum-variance unbiased estimator (MVUE).
Rao--Blackwell says: for any unbiased competitor , the projected estimator is a function of and has weakly smaller variance. Completeness of says: there is at most one unbiased function of --- any two would differ by a function of with zero mean, which completeness kills. Combining: the unbiased function of is the MVUE.
Any two unbiased functions of $T$ agree a.s.
Suppose and are both unbiased for . Then satisfies for all . Completeness forces a.s., so a.s.
Rao--Blackwell shows every unbiased estimator projects here
Let be any unbiased estimator of (not necessarily a function of ). By Rao--Blackwell, is an unbiased function of . By Step 1, a.s. Hence for every unbiased . So is MVUE, and it is unique.
Recipe for Constructing the MVUE
Complexity: Depends on step 6 — closed form in the exponential family, Monte Carlo elsewhere.In practice steps 4--6 collapse when the family is a textbook exponential family: the natural sufficient statistic is automatically complete, and an unbiased function of is often visible by inspection (e.g., rescaling the sample mean).
Example: MVUE of Amplitude in AWGN via Rao--Blackwell
Observe , i.i.d. , unknown. Starting from the naive unbiased estimator (for ), construct the MVUE via Rao--Blackwell.
Naive estimator is unbiased
. Its variance is --- bad, because it ignores the other samples.
Sufficient statistic: $T = \mathbf{s}^T\mathbf{Y}$
From EFactorization: Signal Amplitude in AWGN, is sufficient and, as the natural sufficient statistic of an exponential family with one-dimensional natural parameter image, it is complete.
Compute $\tilde{g}(T)$
Under , is jointly Gaussian with , , , , (the cross-term). The Gaussian conditional mean formula gives The -dependence cancels (as it must, by sufficiency). Hence with variance , matching the CRB --- the MVUE is efficient.
Example: MVUE of from a Gaussian Sample
Given i.i.d. with both unknown, find the MVUE of .
Complete sufficient statistic
From EGaussian as an Exponential Family, is complete sufficient for .
Unbiased function of $T$
The Bessel-corrected variance is a function of (both and are components of ), and it is unbiased for .
Conclude by Lehmann--Scheffe
By Lehmann--Scheffe, is the MVUE of . Its variance exceeds the CRB --- there is no efficient unbiased estimator of in this model. Efficiency is strictly stronger than being MVUE.
Key Takeaway
The MVUE workflow. (1) find a complete sufficient statistic; (2) construct any unbiased function of it. That function is the MVUE. No calculus of variations needed --- sufficiency and completeness do the work. The subtlety that efficient MVUE but not conversely is the reason we need this machinery beyond the CRB.
Efficient vs. MVUE vs. Unbiased
| Property | Unbiased | MVUE | Efficient |
|---|---|---|---|
| Bias | |||
| Variance | Anything | Minimum among unbiased | Equals CRB for all |
| Always exists? | Usually | Not always; needs complete sufficient statistic | Rare — requires affine score |
| Certificate | Compute mean | Lehmann--Scheffe | Score linear in |
| Implication chain | Needed for MVUE | efficient | MVUE (if unbiased) |
Common Mistake: Sufficient Does Not Imply Complete
Mistake:
Treating any sufficient statistic as a green light to apply Lehmann--Scheffe.
Correction:
Lehmann--Scheffe needs completeness, not just sufficiency. Example: in the location family with (a discrete two-point parameter set), the sample mean is sufficient but not complete --- the parameter image is too small. Without completeness, multiple unbiased functions of can exist, and none is "the" MVUE. Completeness is usually secured by the full-dimensionality condition in the exponential family theorem.
Common Mistake: Sometimes No MVUE Exists
Mistake:
Assuming the MVUE always exists and that the workflow will terminate.
Correction:
There are families with no uniformly minimum-variance unbiased estimator: different are minimized by different estimators, and no single function dominates the others. This typically happens outside the exponential family, or with nuisance parameters that destroy completeness. In such cases one resorts to locally minimum variance, asymptotic criteria, or Bayesian/risk-based approaches.
Rao--Blackwellization in Action
Start with a noisy naive unbiased estimator for the amplitude , and apply Rao--Blackwell with respect to . Compare the sampling distributions and variances as you sweep and SNR.
Parameters
Least Squares as the BLUE: An MVUE Within a Class
Under a linear observation model with zero-mean noise of covariance , the weighted least-squares estimator is the best linear unbiased estimator (BLUE) --- the MVUE within the class of linear estimators, regardless of the noise distribution (Gauss--Markov theorem). When the noise is Gaussian, this is also the global MVUE. In channel estimation, this is why LS is the reference point even when the channel covariance is unknown: MMSE improves on it by exploiting priors, but LS is optimal when the prior is absent.
- •
Requires invertible (identifiability)
- •
Variance floor: — the vector CRB in Gaussian case
- •
Non-Gaussian noise: linear class optimality, but nonlinear estimators may beat it
Quick Check
Under what condition does Rao--Blackwell give zero variance reduction?
The original estimator is already a function of the sufficient statistic
The sufficient statistic has the same dimension as the data
The parameter is a scalar
The estimator is efficient
If is a function of , then almost surely, so conditioning does nothing. The conditional variance is zero.
Quick Check
Which hypothesis in Lehmann--Scheffe guarantees uniqueness of the MVUE?
Sufficiency of
Completeness of
Unbiasedness of
The exponential-family form
Completeness says the only unbiased estimator of zero that is a function of is zero itself; this is what forces uniqueness of any unbiased function of .
Why This Matters: Channel Estimation: LS is the BLUE, MMSE is Beyond Unbiased
Classical pilot-based channel estimation returns the least-squares estimator --- the BLUE by Gauss--Markov, and MVUE in the Gaussian case. MMSE channel estimators trade off unbiasedness for smaller MSE by exploiting the channel covariance ; they live in a different optimality framework (Bayesian, Chapter 7). Both sides of this story are already visible in the machinery of this chapter.