Prerequisites & Notation
Before You Begin
This chapter sits between three separate communities: (i) the massive MIMO system-design literature of Parts I-IV of this book, (ii) the model-based estimation and inference theory of Book FSI, and (iii) the modern deep-learning toolbox. The reader is expected to be comfortable with all three viewpoints and, crucially, with the tension between them: when the closed-form MMSE estimator exists it beats any black-box network, and when it does not the learned alternative is the best tool we have.
- Linear MMSE channel estimation with known spatial covariance(Review ch03)
Self-check: Can you write the MMSE estimator and identify the two ingredients that make it optimal?
- FDD CSI feedback overhead, Type I / Type II codebooks, JSDM(Review ch08)
Self-check: Can you explain why FDD massive MIMO feedback overhead scales with the product of the number of antennas and the number of quantization bits per coefficient?
- Deep unfolding: turning an iterative algorithm into a trainable network by making each iteration a layer(Review ch18)
Self-check: Can you describe how unrolling iterations of ISTA produces a -layer network with learnable step sizes and soft-threshold parameters?
- Approximate message passing (AMP) and orthogonal AMP (OAMP)(Review ch20)
Self-check: Can you state the AMP iteration and explain the role of the Onsager correction term?
- Mutual information and rate-distortion as information-theoretic primitives(Review ch13)
Self-check: Can you write the rate-distortion function and identify the optimization variable?
- Stochastic gradient descent, backpropagation, standard DL architectures (MLP, CNN, Transformer)
Self-check: Can you implement a minimal training loop in PyTorch or JAX with a custom loss, without relying on a high-level trainer?
- Markov decision processes, Bellman equation, policy gradients
Self-check: Can you write the Bellman equation and explain why policy gradient methods optimize a stochastic policy rather than a deterministic one?
Notation for This Chapter
Symbols introduced or specialized in this chapter. Customizable symbols use tokens. Machine-learning-specific symbols (network weights, loss functions, learning rate) are written in raw LaTeX because they are not part of the massive MIMO notation registry. See the NGlobal Notation Table master table.
| Symbol | Meaning | Introduced |
|---|---|---|
| True channel matrix (target of estimation / feedback / prediction) | s01 | |
| Estimated or reconstructed channel matrix (output of the learned network) | s01 | |
| Pilot matrix used for uplink training | s01 | |
| Pilot observation vector at the BS | s01 | |
| Neural network with trainable parameters | s01 | |
| Training loss (NMSE, cross-entropy, negative log-likelihood, etc.) | s01 | |
| Normalized mean-squared error | s01 | |
| Latent code in the CSI feedback encoder-decoder | s02 | |
| Feedback payload size in bits per channel instance | s02 | |
| Rate-distortion function: minimum bits per channel sample to achieve NMSE | s02 | |
| Optimal beam index at time slot within a predefined codebook | s03 | |
| , , | MDP state, action, and reward at time | s04 |
| Stochastic policy parametrized by (PPO actor network) | s04 | |
| Value function of policy at state | s04 | |
| Per-user SNR (linear scale) | s01 | |
| Number of users sharing the resource block | s04 | |
| Number of BS transmit antennas | s01 |