Prerequisites & Notation
Before You Begin
This chapter sits at the intersection of estimation theory, random matrix theory, and empirical Bayes. The classical CRLB-centred picture that dominates Chapters 4–7 is not wrong — it is simply incomplete. The point is that in modern applications (massive MIMO, compressed sensing, covariance estimation, high-dimensional inference) the dimension of the parameter vector grows in lockstep with the number of observations . Classical consistency arguments, which hold fixed and let , collapse in this regime, and the behaviour of the MLE can swing from "efficient" to "spectacularly wrong". The reader should be comfortable with the following before continuing.
- Maximum likelihood estimation and the Cramér–Rao lower bound(Review ch04)
Self-check: Can you state the CRLB for a vector parameter, write down the Fisher information matrix, and explain when the MLE achieves the bound?
- Linear MMSE estimation and the Wiener filter(Review ch05)
Self-check: Can you derive the LMMSE estimator for the Gaussian linear model and compute its MSE?
- Compressed sensing and LASSO at a conceptual level(Review ch17)
Self-check: Can you state the -minimisation programme and recognise why an penalty promotes sparsity?
- Random matrix theory essentials
Self-check: Do you know what the Marchenko–Pastur law says about the eigenvalues of when has i.i.d. entries?
- Convex optimisation (unconstrained and penalised)
Self-check: Can you recognise a convex problem, write the KKT conditions for a quadratic-plus- objective, and describe proximal-gradient iteration?
- Bayesian decision theory(Review ch06)
Self-check: Can you compute a Bayes risk, define admissibility, and explain the relationship between a minimax estimator and a least-favourable prior?
Notation for This Chapter
Symbols used throughout Chapter 22. The ratio is the single most important parameter — the entire chapter can be read as a study of how estimation behaves as this ratio departs from zero.
| Symbol | Meaning | Introduced |
|---|---|---|
| Ambient dimension of the parameter vector | s01 | |
| Number of observations (rows of / samples) | s01 | |
| Aspect ratio ; the proportional-asymptotics regime fixes | s01 | |
| Design / sensing / measurement matrix in | s01 | |
| Regularization parameter (ridge / LASSO penalty) | s02 | |
| Ridge estimator | s02 | |
| LASSO estimator | s02 | |
| James–Stein estimator | s03 | |
| Frequentist risk | s03 | |
| Minimax risk over a parameter class | s04 | |
| Sparsity level — number of non-zero components of | s04 | |
| Least-favourable prior attaining the minimax risk | s04 |