Chapter Summary
Chapter Summary
Key Points
- 1.
The lossy compression problem. When lossless compression is too expensive (continuous sources, large alphabets), we allow controlled distortion and ask: what is the minimum rate? The distortion measure quantifies reconstruction quality, and the rate-distortion pair parametrizes the fundamental tradeoff.
- 2.
The rate-distortion function. is the information-theoretic minimum rate for distortion . It is convex and non-increasing, ranging from (lossless) to 0 (). For binary sources with Hamming distortion: .
- 3.
The rate-distortion theorem: covering as the dual of packing. The achievability proof generates a random codebook and uses the covering lemma to show that codewords suffice to cover the typical set when . The converse uses data processing and convexity. Shannon's separation theorem unifies source and channel coding: a source is transmissible iff .
- 4.
Gaussian rate-distortion: the 6 dB rule. For : , equivalently . Each bit halves the distortion (6 dB). For parallel Gaussian sources, reverse waterfilling allocates more distortion to weak components β the dual of channel waterfilling.
- 5.
Wyner-Ziv: side information at the decoder. When the decoder has correlated side information , binning reduces the required rate to . For Gaussian sources, this equals β side information at the encoder is worthless. For general sources, there is a strict loss.
- 6.
Successive refinement and scalable coding. A source is successively refinable if layered encoding achieves at every quality level. Gaussian sources are always successively refinable; this is the theoretical basis for scalable video coding (SVC, HEVC scalability).
- 7.
Transform coding: theory meets practice. The pipeline of transform (DCT/wavelet) + quantize + entropy code approaches for Gaussian sources. ECSQ operates within 1.53 dB of the bound. Modern codecs (HEVC, VVC) close much of this gap through prediction, variable-size transforms, and context-adaptive arithmetic coding.
Looking Ahead
Chapter 7 addresses distributed source coding: compressing correlated sources that cannot communicate with each other. The Slepian-Wolf theorem (lossless) and Berger-Tung bounds (lossy) extend the ideas of this chapter to multi-terminal settings, where the interplay between sources adds a new dimension to the rate-distortion tradeoff.