Connections Between Discrete and Continuous Entropy
Bridging the Discrete and Continuous Worlds
We have treated discrete and continuous entropy as separate concepts. But in practice, every continuous signal is eventually quantized — by an ADC, by a digital representation, by finite-precision arithmetic. The natural question is: what happens to entropy when we quantize? This section establishes the precise relationship between discrete entropy of a quantized variable and the differential entropy of the underlying continuous variable. The connection resolves the puzzle of why differential entropy can be negative — and it has practical implications for quantization theory and analog-to-digital conversion.
Definition: Quantized Random Variable
Quantized Random Variable
Let be a continuous random variable with PDF . The -quantized version of is the discrete random variable , which rounds down to the nearest multiple of . The PMF of is:
for small .
Theorem: Quantization and Differential Entropy
For a continuous random variable with PDF :
Equivalently: for small .
The discrete entropy diverges as because finer quantization creates more possible values. But the excess entropy beyond bits (the cost of specifying which bin) converges to the differential entropy.
The point is that differential entropy is "entropy up to an additive constant" that depends on the quantization resolution. This explains why can be negative: it means the quantized entropy grows slower than .
Approximate the PMF
For small : .
Identify the limit
\Delta \to 0-\int f_X(x) \log f_X(x),dx = h(X)H(X^\Delta) + \log\Delta \to h(X)$.
Example: Quantizing a Gaussian
A Gaussian source is uniformly quantized with step size . Approximately how many bits per sample does the quantized representation require?
Apply the quantization formula
.
For (fine quantization): bits per sample.
Comparison with ADC bits
An -bit uniform ADC over range has . For (covers ), bits gives .
bits — close to but slightly less, because the source entropy is less than the ADC's full dynamic range.
Entropy of a Quantized Gaussian
Observe how the discrete entropy of a quantized Gaussian grows as the quantization step decreases. The curve converges to the differential entropy .
Parameters
Standard deviation of the Gaussian source
Logarithm of the quantization step size
Definition: Rényi Entropy
Rényi Entropy
The Rényi entropy of order () of a discrete RV is:
As : (Shannon entropy).
As : (Hartley entropy).
As : (min-entropy).
Rényi entropy plays a role in source coding with different error criteria, guessing problems, and one-shot information theory (Chapter 26). The parameter controls which events are weighted more: emphasizes high-probability events, emphasizes low-probability events.
Rényi entropy
A one-parameter family of entropy measures generalizing Shannon entropy. . Converges to Shannon entropy as . Used in one-shot information theory and security analysis.
Related: Entropy
Key Takeaway
Differential entropy equals the entropy of quantization minus the quantization cost. Precisely: . This explains why can be negative (the quantization entropy grows slower than ) and why mutual information — which is a difference of entropies — is always well-defined and non-negative for continuous variables.
Quick Check
If and we quantize with step , approximately how many bits does require?
bits
bits
bit
bit
bit. The quantized variable has about equally likely levels (range divided by step ), so bits. This matches bits.
Capacity of the Gaussian MIMO Broadcast Channel
The maximum entropy property of the Gaussian distribution is central to characterizing the capacity region of the Gaussian MIMO broadcast channel. Weingarten, Steinberg, and Shamai proved that the capacity region is achieved by dirty-paper coding (DPC) with Gaussian signaling, establishing the connection between the continuous entropy results of this chapter and the multiuser capacity theory of Chapter 15. The proof relies heavily on the entropy power inequality and the vector version of the "Gaussian is worst noise" result.