Chapter Summary
Chapter Summary
Key Points
- 1.
Differential entropy extends entropy to continuous RVs. Unlike discrete entropy, it can be negative, depends on coordinates, and has no direct operational meaning. But mutual information is well-defined and coordinate-invariant.
- 2.
The Gaussian maximizes differential entropy under a variance constraint: , with equality iff . This is the foundation of all Gaussian channel results.
- 3.
For Gaussian vectors, . The covariance determinant measures the uncertainty volume. Hadamard's inequality follows as a corollary.
- 4.
The entropy power inequality (EPI): . Entropy powers are superadditive for independent summands. This proves Gaussian noise is worst-case and underpins converse proofs for Gaussian channels and broadcast channels.
- 5.
AWGN capacity is . Achieved by Gaussian input . The achievability uses the maximum entropy property; the converse uses the EPI.
- 6.
Differential entropy equals quantization entropy minus quantization cost: . This explains why can be negative and connects continuous information theory to practical quantization.
Looking Ahead
Chapter 3 introduces the concept of typicality — the key technical tool for proving coding theorems. The asymptotic equipartition property (AEP) shows that for long i.i.d. sequences, only about sequences out of carry appreciable probability. This "concentration of measure" phenomenon is the engine behind both source coding and channel coding.