Converse for the Degraded Broadcast Channel
Why the Converse Requires New Tools
In Section 15.2, we proved the converse for the discrete degraded BC using Fano's inequality and single-letterization. That argument establishes the capacity region in terms of some auxiliary variable and distribution . For the discrete case, the optimization over is a finite-dimensional problem.
For the Gaussian BC, however, we need to show that the optimal is jointly Gaussian — i.e., that the Gaussian superposition coding scheme is optimal among all possible coding strategies. This is the hard part, and it requires the entropy power inequality (EPI).
The argument, due to Bergmans (1973), is one of the most elegant applications of the EPI in information theory. The key idea is that Gaussian noise is the "worst" noise (it maximizes conditional entropy for a given power), which translates into Gaussian inputs being the "best" for the broadcast channel.
Definition: Entropy Power
Entropy Power
The entropy power of a continuous random variable with differential entropy is defined as:
Equivalently, is the variance of a Gaussian random variable with the same differential entropy as . If , then .
The entropy power provides a "Gaussian equivalent" for any continuous distribution. The EPI says that this equivalent behaves super-additively under addition of independent random variables.
Theorem: The Entropy Power Inequality
For independent continuous random variables and in :
or equivalently,
Equality holds if and only if both and are Gaussian.
The EPI says that adding independent random variables together produces "at least as much entropy" as adding independent Gaussians with the same individual entropy powers. Since the Gaussian maximizes entropy for a given variance, the EPI captures the idea that Gaussian noise is the "most entropic" perturbation.
A useful corollary: if , then for any independent of :
with equality iff is Gaussian.
Proof sketch (via Fisher information)
The standard proof uses the de Bruijn identity and Fisher information. We sketch the main steps:
-
De Bruijn identity: For with density and independent: , where is the Fisher information.
-
Fisher information inequality (FII): For independent : .
-
Combining the de Bruijn identity with the FII and integrating over yields the EPI. The details are in Chapter 2, Section 2.4.
The proof is non-trivial; what matters for us is the consequence: Gaussian inputs are optimal for the Gaussian BC.
Theorem: Bergmans' Converse for the Gaussian BC
For the Gaussian broadcast channel with and power , any achievable rate pair must satisfy
for some . Thus the superposition coding region is exactly the capacity region.
The converse shows that no coding scheme — however clever — can achieve rates outside the superposition coding region. The key tool is the entropy power inequality, which proves that Gaussian inputs are the best choice for and .
The argument proceeds in two steps: first, Fano's inequality gives rate constraints in terms of differential entropies; second, the EPI converts these into the Gaussian expressions.
Step 1: Fano's inequality
From the general converse for the degraded BC (Section 15.2):
where . The challenge is to bound each term without assuming or is Gaussian.
Step 2: The EPI-based bound for the weak user
Focus on a single time index (we drop the subscript ). We need to show that for some appropriate related to the conditional variance of given .
Let and . Define .
Then:
Since :
- (Gaussian maximizes entropy).
- . By the EPI: where the last step uses (the Gaussian lower bound on entropy power).
Actually, we use the upper bound more carefully: the conditional entropy is lower bounded using the EPI, which gives us an upper bound on .
Step 3: Completing the converse
After applying the EPI to bound from below and the maximum-entropy property to bound from above, we obtain:
For the strong user, . Since :
Averaging over and using concavity of (Jensen's inequality), we can reduce to a single with the power constraint . The result matches the superposition coding region.
The Role of the EPI in the Converse
The EPI is used in one specific place in Bergmans' converse: to lower-bound , the conditional entropy at the weak receiver. Without the EPI, we could not rule out the possibility that a non-Gaussian choice of might yield a higher than the Gaussian choice.
The EPI essentially says: "adding Gaussian noise is the worst thing that can happen to your mutual information." Since the weak user suffers from more noise (), this worst-case noise argument is what forces the Gaussian optimality.
This is a recurring theme in Gaussian information theory: the extremal properties of the Gaussian distribution (maximum entropy, worst-case noise, EPI) are the tools that convert achievability results (which work for any distribution) into tight capacity characterizations.
Example: Verifying the Converse for a Specific Rate Pair
For the Gaussian BC with , , , verify that the rate pair is outside the capacity region.
Find the boundary
On the capacity region boundary, and .
Setting :
So , giving , hence , so .
Check the strong user's rate
At :
The maximum compatible with is , which is less than . Therefore lies outside the capacity region — confirming the converse.
Preview: MAC-BC Duality
There is a remarkable duality between the MAC and BC capacity regions for Gaussian channels: the capacity region of the Gaussian BC with total power equals the capacity region of the "dual" Gaussian MAC with individual power constraints that sum to .
This duality, formalized by Vishwanath, Jindal, and Goldsmith (2003), means that computing the BC capacity region (hard, because the converse requires the EPI) can be reduced to computing the MAC capacity region (easier, because the MAC converse uses standard Fano arguments).
The duality extends to the MIMO case and is the computational engine behind practical beamforming design for the MIMO broadcast channel. We develop this fully in Chapter 16.
Quick Check
Under what condition does equality hold in the entropy power inequality ?
When and have the same variance
When both and are Gaussian
When and are identically distributed
When or is zero (deterministic)
Equality in the EPI holds if and only if both and are Gaussian. This is the key fact that makes Gaussian inputs optimal for the Gaussian BC.
Entropy Power
For a continuous random variable , the entropy power is . It equals the variance of a Gaussian with the same differential entropy. The entropy power inequality states for independent .
Related: Entropy Power Inequality (EPI)
Entropy Power Inequality (EPI)
A fundamental inequality in information theory: for independent continuous random variables and . Equality holds iff both are Gaussian. The EPI is the key tool for proving converse results in Gaussian broadcast and interference channel problems.
Related: Entropy Power
MIMO Broadcast Channel Capacity via Dirty Paper Coding
The scalar Gaussian BC converse by Bergmans relies on the EPI, which does not have a natural matrix extension for the MIMO case. The capacity region of the MIMO BC was established through a different route: Weingarten, Steinberg, and Shamai (with contributions from Caire and Shamai's earlier work on the MIMO BC) proved that dirty paper coding (DPC) achieves the capacity region.
The proof uses the MAC-BC duality (the capacity region of the MIMO BC equals a transformed version of the MIMO MAC region) and the channel enhancement technique. This result resolved one of the major open problems in network information theory and provided the theoretical foundation for MU-MIMO precoding in 4G/5G systems.
Caire's work on the iterative water-filling algorithm for computing the DPC rate region is the practical bridge between the theoretical capacity result and implementable beamforming designs (see Book telecom, Chapter 17).