Gaussian Rate-Distortion and Reverse Waterfilling
The 6 dB Rule
The Gaussian source is the most important continuous source in practice β it models thermal noise, quantization error, and many natural signals. Its rate-distortion function has a beautiful closed form that yields the famous "6 dB per bit" rule: each additional bit of rate halves the distortion. For vector Gaussian sources, the optimal strategy is "reverse waterfilling" β the dual of the waterfilling solution for channel capacity. Understanding these results is essential for anyone working with quantization, compression, or signal processing.
Theorem: Rate-Distortion Function for Gaussian Source
For a Gaussian source with squared-error distortion : Equivalently, the distortion-rate function is .
Each bit of rate halves the distortion: going from to bits reduces by a factor of 4 (= 6 dB). This is the information-theoretic version of the "6 dB per bit" rule. At zero rate, the best we can do is estimate by its mean (zero), giving distortion . At infinite rate, distortion approaches zero.
Optimal test channel
The optimal test channel is where is independent of ? No β we need to be a function of plus noise. The correct test channel is: where independent of , and . This gives .
Mutual information
With and :
Optimality of Gaussian test channel
The Gaussian test channel is optimal because: (i) among all distributions with variance , the Gaussian maximizes the differential entropy (this does not help here since we're minimizing ); (ii) the key is that the Gaussian maximizes for fixed . Since , maximizing minimizes .
Example: Gaussian Rate-Distortion in Practice
A speech signal is modeled as Gaussian with . Compute the rate needed for SNR values of 20 dB, 40 dB, and 60 dB, where SNR .
Rate computation
.
- SNR = 20 dB (= 100): bits/sample.
- SNR = 40 dB (= 10000): bits/sample.
- SNR = 60 dB (= ): bits/sample.
Every 20 dB of SNR costs 3.32 bits. Equivalently, 1 bit 6.02 dB of SNR.
Practical comparison
Telephone-quality speech (8 kHz, 8 bits/sample = 64 kbps) achieves about 48 dB SNR. The rate-distortion limit for 48 dB at 8 kHz would be kbps. Modern speech codecs (Opus, EVS) achieve near-transparent quality at 16β24 kbps, approaching but not reaching the R-D bound (they exploit speech-specific structure).
Theorem: Reverse Waterfilling for Vector Gaussian Sources
Let where is diagonal (parallel independent Gaussian sources). The rate-distortion function under total squared-error distortion is: where and the reverse waterfilling level is chosen so that .
Reverse waterfilling allocates more distortion to components with smaller variance. If a component has variance , it gets (no bits allocated β the component is "drowned" in distortion). Components with get , with the excess captured by the code.
This is the opposite of channel waterfilling, where we pour power into strong subchannels. Here we pour distortion into weak components. The connection to transform coding is immediate: apply the KLT (eigendecomposition) to decorrelate, then reverse-waterfill on the eigenvalues.
Independent optimization
Since the components are independent, and . The optimization decomposes into minimizing subject to and .
KKT conditions
Lagrangian: . , giving for active components (). Inactive components () are "shut off" β they contribute zero rate and take their full variance as distortion.
Reverse Waterfilling for Parallel Gaussian Sources
Visualize the reverse waterfilling solution for a set of parallel Gaussian sources. Adjust the total distortion budget and observe how distortion is allocated across components. Compare with the forward waterfilling for channel capacity.
Parameters
The 6 dB Rule in System Design
The 6 dB/bit rule () is the fundamental benchmark for all quantization and compression system design. A practical quantizer that achieves SNR dB for some constant is said to be " dB from the R-D bound." The best practical schemes:
- Uniform quantizer + entropy coding: dB (Gish-Pierce, 1968)
- Lloyd-Max quantizer (no entropy coding): dB for large
- Lattice quantizer + entropy coding: dB (approaching the Gaussian R-D bound)
These gaps quantify the "price" of structured (implementable) coding versus random coding.
Common Mistake: Forward vs. Reverse Waterfilling
Mistake:
Confusing forward waterfilling (channel capacity, pour power into strong subchannels) with reverse waterfilling (rate-distortion, pour distortion into weak components).
Correction:
The two are duals:
- Channel (forward): Fill from the bottom up with power. Strong subchannels get more power. Weak subchannels may be shut off.
- Source (reverse): Fill from the top down with distortion. Weak components get more distortion. Weak components may be completely discarded.
The reversal comes from maximizing vs. minimizing: channel capacity maximizes rate (use strong subchannels), rate-distortion minimizes rate (sacrifice weak components).
Reverse Waterfilling: Pouring Distortion into Weak Components
Key Takeaway
For a Gaussian source, β the "6 dB per bit" rule. For parallel Gaussian sources, reverse waterfilling allocates more distortion to weak components and zero bits to components with variance below the waterfilling level . This is the dual of channel waterfilling and the foundation of transform coding: apply the KLT to decorrelate, then reverse-waterfill across the transform coefficients.