Chapter Summary

Chapter Summary

Key Points

  • 1.

    The AEP: 1nlogP(X)H(X)-\frac{1}{n}\log P(\mathbf{X}) \to H(X) in probability. For i.i.d. sequences, the per-symbol log-probability concentrates around the entropy. This partitions Xn\mathcal{X}^n into a typical set of 2nH\approx 2^{nH} likely sequences and an atypical set with negligible total probability.

  • 2.

    Strong typicality controls per-symbol frequencies. A sequence is strongly typical if P^(a)P(a)ϵP(a)|\hat{P}(a) - P(a)| \leq \epsilon P(a) for all aa. Unlike weak typicality, strong typicality guarantees marginal consistency: jointly typical pairs have individually typical components.

  • 3.

    The Joint Typicality Lemma: true pairs are typical, independent pairs are not. Pr((X,Y)Tϵ(n))1\Pr((\mathbf{X}, \mathbf{Y}) \in \mathcal{T}_\epsilon^{(n)}) \to 1 for true pairs, but Pr((X~,Y)Tϵ(n))2nI(X;Y)\Pr((\tilde{\mathbf{X}}, \mathbf{Y}) \in \mathcal{T}_\epsilon^{(n)}) \doteq 2^{-nI(X;Y)} for independent pairs. This exponential gap drives all coding theorems.

  • 4.

    The Packing Lemma ensures reliable decoding. If R<I(X;Y)R < I(X;Y), no incorrect codeword is jointly typical with the output. This is the achievability engine for channel coding.

  • 5.

    The Covering Lemma ensures source representation. If R>I(X;X^)R > I(X;\hat{X}), at least one codeword is jointly typical with the source. This is the achievability engine for lossy source coding.

  • 6.

    The method of types gives exact exponential bounds. Type class probability: Pn(TQ)2nD(QP)P^n(T_Q) \doteq 2^{-nD(Q\|P)}. Type class size: TQ2nH(Q)|T_Q| \doteq 2^{nH(Q)}. Number of types: polynomial in nn. Sanov's theorem governs large deviations.

  • 7.

    The proof pattern is reusable. Random codebook + typicality decoding/encoding + packing/covering lemma = achievability proof. This architecture works for DMCs, Gaussian channels, MACs, broadcast channels, and beyond. By Part IV, you will construct these proofs for channels you have never seen.

Looking Ahead

Chapter 4 develops the method of types in full depth, leading to error exponents — the exponential rate at which error probability decreases with blocklength. These exponents quantify not just whether reliable communication is possible (capacity), but how fast reliability improves (the speed of convergence). This is the natural quantitative refinement of the qualitative coding theorems we have just established.