LS and MMSE Channel Estimation
The Estimation Problem
After receiving , the base station performs channel estimation. We work with the sufficient statistic obtained by correlating with user ''s pilot:
Defining :
We estimate from this observation. Two classical estimators exist: LS (requires no statistical knowledge) and MMSE (requires knowledge of ).
Definition: Least-Squares (LS) Channel Estimator
Least-Squares (LS) Channel Estimator
The least-squares estimator minimizes with respect to . Since the observation model is linear, the LS solution is:
Properties:
- Unbiased:
- Error covariance:
- MSE:
- Requires no statistical knowledge of
- Does NOT exploit spatial correlation structure
The LS estimator treats each antenna independently, ignoring the correlation structure encoded in . Its MSE scales as β proportional to the number of antennas, reflecting that each of channel coefficients is estimated independently.
Definition: MMSE Channel Estimator
MMSE Channel Estimator
Assume with known covariance . The MMSE estimator (= posterior mean for jointly Gaussian signals) is:
Properties:
- Error covariance:
- MSE: always
- The estimated channel (the estimate is distributed as a zero-mean Gaussian with covariance equal to the reduction in uncertainty)
- The estimate and error are uncorrelated: (orthogonality principle)
Theorem: MMSE Estimator MSE via Eigendecomposition
Let be the eigendecomposition of the spatial covariance matrix, where with .
Then the MMSE estimation MSE is:
As , . As , .
Each eigendirection is estimated independently with a per-mode SNR . Low-energy eigenmodes (small ) contribute little signal and much noise β their estimation is dominated by the prior, which shrinks the estimate to near zero. High-energy eigenmodes (large ) are estimated with high SNR and nearly perfect accuracy.
Use the eigendecomposition to diagonalize the matrix inverse in the error covariance expression.
The error covariance is . Apply the matrix inversion lemma to simplify.
In the eigenbasis, each eigenvalue contributes to the MSE.
Eigendecompose the error covariance
Substituting the eigendecomposition into :
Factor out unitary matrices
Since is unitary, .
Substituting and using :
Simplify each diagonal entry
Since all matrices inside are diagonal with entries :
Taking the trace: .
LS vs. MMSE Channel Estimation MSE
Compare the normalized MSE of the LS estimator () and the MMSE estimator () as functions of . Use sliders to adjust SNR, pilot length, and channel rank.
Parameters
Theorem: MMSE vs. LS MSE Gap
Let be the effective channel rank. Then:
where is the largest eigenvalue of . The MSE gap grows with the pilot SNR and the ratio . At high SNR, LS is times worse than MMSE asymptotically.
MMSE knows the channel lives in the -dimensional subspace spanned by the eigenvectors of . It suppresses the 'empty' dimensions, receiving only signal there. LS, being unaware of this structure, spreads its resources uniformly across all dimensions, wasting effort on directions where there is only noise.
At high SNR, for the nonzero eigenvalues.
Compare with \\text{MSE}_k^\\text{LS} = N_t\\sigma^2/(p_u\\tau_p).
High-SNR MMSE MSE
For :
High-SNR LS MSE
$
Take the ratio
r_k = N_tr_k \ll N_t\blacksquare$
Key Takeaway
MMSE exploits correlation; LS does not. When the channel has low effective rank (as occurs with spatial correlation and limited angular spread), MMSE outperforms LS by the factor at high SNR. In a massive MIMO array with 128 antennas and a typical urban rank of 10β20, this is a 6β11 dB improvement in estimation quality.
Example: MMSE Estimation with the One-Ring Covariance Model
A ULA with antennas, half-wavelength spacing, serves a user at angle with angular spread . The one-ring model gives (using the approximation from Ch. 2). At pilot SNR dB, compare the LS and MMSE MSE.
Determine effective rank
For a ULA with , half-wavelength spacing, and , the effective rank is approximately:
So roughly dominant eigenvalues concentrate most of the channel energy.
Compute LS MSE
With pilot SNR :
(normalized by , assuming )
Compute MMSE MSE
With eigenvalues of magnitude :
Compute the gain
$
The MMSE estimator is roughly 24 dB better because it knows the channel lives in a 4-dimensional subspace of the 64-dimensional space.
Common Mistake: MMSE Requires Accurate β LS Does Not
Mistake:
MMSE always beats LS, so just use MMSE.
Correction:
MMSE requires accurate knowledge of the spatial covariance matrix . In practice, must be estimated from data (typically using long-term sample averaging over many coherence intervals). If is estimated incorrectly β e.g., due to limited averaging samples, user mobility, or angular spread mismatch β the "MMSE' estimator can actually perform worse than LS in some scenarios. Furthermore, the matrix inversion has complexity , which may be prohibitive for very large arrays.
Rule of thumb: Use MMSE when is reliably estimated and hardware complexity allows. Use LS or regularized LS when only rough statistical knowledge is available.
LS vs. MMSE Channel Estimator Comparison
| Property | LS Estimator | MMSE Estimator |
|---|---|---|
| Required prior knowledge | None | , |
| MSE expression | ||
| Estimation bias | Unbiased | Biased (shrinks toward prior mean ) |
| Exploits spatial correlation | No | Yes (via eigenbasis) |
| Complexity per user | (matrix inversion) | |
| MSE at high pilot SNR | ||
| MSE at low pilot SNR | (prior dominates) |
MMSE via Matrix Inversion Lemma
By the Woodbury matrix identity, the MMSE estimator has an equivalent form:
(valid when is invertible). This form is sometimes preferred numerically when has well-conditioned eigenvalues. The second form with is preferred when is rank-deficient (as is typical when the effective rank ).
Estimating in Practice
The covariance matrix evolves on a timescale much longer than the channel itself (seconds to minutes, vs. milliseconds for ). It can be estimated using a long-term sample average of the instantaneous outer products:
where is the pilot observation in coherence interval . Statistical accuracy requires observations (typically for reasonable estimation quality). For , this means averaging over coherence intervals β feasible for stationary or slow-moving users but challenging for vehicular UEs.
- β’
For N_t = 64: at least 640 coherence intervals needed for 10% Frobenius norm error in R_k estimate
- β’
Covariance estimation overhead not counted in pilot overhead but represents long-term cost
- β’
Structured covariance models (one-ring, Kronecker) reduce estimation to few parameters
Quick Check
A user's channel covariance has effective rank with equal eigenvalues, and . At high pilot SNR, the MMSE MSE is approximately what fraction of the LS MSE?
1 (they are equal)
At high SNR, \\text{MSE}^\\text{MMSE} \\approx r_k\\sigma^2/(p_u\\tau_p) and \\text{MSE}^\\text{LS} = N_t\\sigma^2/(p_u\\tau_p), so the ratio is .
MMSE Estimator
The minimum mean squared error estimator is the conditional mean . For jointly Gaussian signals, it equals the LMMSE estimator and takes the linear form .
Related: Least-Squares (LS) Channel Estimator, Spatial Covariance Matrix