Converse for the Degraded Broadcast Channel

Why the Converse Requires New Tools

In Section 15.2, we proved the converse for the discrete degraded BC using Fano's inequality and single-letterization. That argument establishes the capacity region in terms of some auxiliary variable UU and distribution p(u,x)p(u, x). For the discrete case, the optimization over p(u,x)p(u, x) is a finite-dimensional problem.

For the Gaussian BC, however, we need to show that the optimal (U,X)(U, X) is jointly Gaussian — i.e., that the Gaussian superposition coding scheme is optimal among all possible coding strategies. This is the hard part, and it requires the entropy power inequality (EPI).

The argument, due to Bergmans (1973), is one of the most elegant applications of the EPI in information theory. The key idea is that Gaussian noise is the "worst" noise (it maximizes conditional entropy for a given power), which translates into Gaussian inputs being the "best" for the broadcast channel.

Definition:

Entropy Power

The entropy power of a continuous random variable XX with differential entropy h(X)h(X) is defined as:

N(X)=12πee2h(X).N(X) = \frac{1}{2\pi e} e^{2h(X)}.

Equivalently, N(X)N(X) is the variance of a Gaussian random variable with the same differential entropy as XX. If XN(0,σ2)X \sim \mathcal{N}(0, \sigma^2), then N(X)=σ2N(X) = \sigma^2.

The entropy power provides a "Gaussian equivalent" for any continuous distribution. The EPI says that this equivalent behaves super-additively under addition of independent random variables.

Theorem: The Entropy Power Inequality

For independent continuous random variables XX and YY in R\mathbb{R}:

N(X+Y)N(X)+N(Y),N(X + Y) \geq N(X) + N(Y),

or equivalently,

e2h(X+Y)e2h(X)+e2h(Y).e^{2h(X+Y)} \geq e^{2h(X)} + e^{2h(Y)}.

Equality holds if and only if both XX and YY are Gaussian.

The EPI says that adding independent random variables together produces "at least as much entropy" as adding independent Gaussians with the same individual entropy powers. Since the Gaussian maximizes entropy for a given variance, the EPI captures the idea that Gaussian noise is the "most entropic" perturbation.

A useful corollary: if ZN(0,N)Z \sim \mathcal{N}(0, N), then for any XX independent of ZZ:

h(X+Z)12log(2πe(N(X)+N))h(X + Z) \geq \frac{1}{2}\log(2\pi e(N(X) + N))

with equality iff XX is Gaussian.

Theorem: Bergmans' Converse for the Gaussian BC

For the Gaussian broadcast channel with N1<N2N_1 < N_2 and power PP, any achievable rate pair (R1,R2)(R_{1}, R_{2}) must satisfy

R212log ⁣(1+αP(1α)P+N2),R_{2} \leq \frac{1}{2}\log\!\left(1 + \frac{\alpha P}{(1-\alpha)P + N_2}\right), R112log ⁣(1+(1α)PN1),R_{1} \leq \frac{1}{2}\log\!\left(1 + \frac{(1-\alpha)P}{N_1}\right),

for some α[0,1]\alpha \in [0,1]. Thus the superposition coding region is exactly the capacity region.

The converse shows that no coding scheme — however clever — can achieve rates outside the superposition coding region. The key tool is the entropy power inequality, which proves that Gaussian inputs are the best choice for UU and XX.

The argument proceeds in two steps: first, Fano's inequality gives rate constraints in terms of differential entropies; second, the EPI converts these into the Gaussian expressions.

,

The Role of the EPI in the Converse

The EPI is used in one specific place in Bergmans' converse: to lower-bound h(Y2U)h(Y_2 | U), the conditional entropy at the weak receiver. Without the EPI, we could not rule out the possibility that a non-Gaussian choice of (U,X)(U, X) might yield a higher I(U;Y2)I(U; Y_2) than the Gaussian choice.

The EPI essentially says: "adding Gaussian noise is the worst thing that can happen to your mutual information." Since the weak user suffers from more noise (N2>N1N_2 > N_1), this worst-case noise argument is what forces the Gaussian optimality.

This is a recurring theme in Gaussian information theory: the extremal properties of the Gaussian distribution (maximum entropy, worst-case noise, EPI) are the tools that convert achievability results (which work for any distribution) into tight capacity characterizations.

Example: Verifying the Converse for a Specific Rate Pair

For the Gaussian BC with P=10P = 10, N1=1N_1 = 1, N2=5N_2 = 5, verify that the rate pair (R1,R2)=(1.5,0.5)(R_{1}, R_{2}) = (1.5, 0.5) is outside the capacity region.

Preview: MAC-BC Duality

There is a remarkable duality between the MAC and BC capacity regions for Gaussian channels: the capacity region of the Gaussian BC with total power PP equals the capacity region of the "dual" Gaussian MAC with individual power constraints that sum to PP.

This duality, formalized by Vishwanath, Jindal, and Goldsmith (2003), means that computing the BC capacity region (hard, because the converse requires the EPI) can be reduced to computing the MAC capacity region (easier, because the MAC converse uses standard Fano arguments).

The duality extends to the MIMO case and is the computational engine behind practical beamforming design for the MIMO broadcast channel. We develop this fully in Chapter 16.

Quick Check

Under what condition does equality hold in the entropy power inequality N(X+Y)N(X)+N(Y)N(X + Y) \geq N(X) + N(Y)?

When XX and YY have the same variance

When both XX and YY are Gaussian

When XX and YY are identically distributed

When XX or YY is zero (deterministic)

Entropy Power

For a continuous random variable XX, the entropy power is N(X)=12πee2h(X)N(X) = \frac{1}{2\pi e} e^{2h(X)}. It equals the variance of a Gaussian with the same differential entropy. The entropy power inequality states N(X+Y)N(X)+N(Y)N(X+Y) \geq N(X) + N(Y) for independent X,YX, Y.

Related: Entropy Power Inequality (EPI)

Entropy Power Inequality (EPI)

A fundamental inequality in information theory: e2h(X+Y)e2h(X)+e2h(Y)e^{2h(X+Y)} \geq e^{2h(X)} + e^{2h(Y)} for independent continuous random variables XX and YY. Equality holds iff both are Gaussian. The EPI is the key tool for proving converse results in Gaussian broadcast and interference channel problems.

Related: Entropy Power

🎓CommIT Contribution(2006)

MIMO Broadcast Channel Capacity via Dirty Paper Coding

H. Weingarten, Y. Steinberg, S. Shamai (Shitz), G. CaireIEEE Transactions on Information Theory

The scalar Gaussian BC converse by Bergmans relies on the EPI, which does not have a natural matrix extension for the MIMO case. The capacity region of the MIMO BC was established through a different route: Weingarten, Steinberg, and Shamai (with contributions from Caire and Shamai's earlier work on the MIMO BC) proved that dirty paper coding (DPC) achieves the capacity region.

The proof uses the MAC-BC duality (the capacity region of the MIMO BC equals a transformed version of the MIMO MAC region) and the channel enhancement technique. This result resolved one of the major open problems in network information theory and provided the theoretical foundation for MU-MIMO precoding in 4G/5G systems.

Caire's work on the iterative water-filling algorithm for computing the DPC rate region is the practical bridge between the theoretical capacity result and implementable beamforming designs (see Book telecom, Chapter 17).

MIMObroadcast channeldirty paper codingcapacity regionView Paper →