The AWGN Channel

Why the Gaussian Channel?

Every wireless, wired, and optical communication system ultimately faces additive noise. The central limit theorem tells us that the aggregate of many small independent disturbances converges to a Gaussian distribution β€” and this is precisely what we observe in practice: thermal noise in receivers, shot noise in photodetectors, and aggregate interference in dense networks are all well-modeled as Gaussian.

The Gaussian channel is therefore not just a mathematical convenience β€” it is the canonical model for communication under noise. Its capacity formula, C=12log⁑(1+SNR)C = \frac{1}{2}\log(1 + \text{SNR}), is arguably the single most important equation in all of communication theory. It tells every engineer, before writing a single line of code, the absolute limit of what is achievable.

Definition:

The Scalar AWGN Channel

The additive white Gaussian noise (AWGN) channel is defined by

Yi=Xi+Zi,i=1,2,…,n,Y_i = X_i + Z_{i}, \quad i = 1, 2, \ldots, n,

where:

  • Xi∈RX_i \in \mathbb{R} is the channel input at time ii,
  • Zi∼N(0,N)Z_{i} \sim \mathcal{N}(0, N) are i.i.d. Gaussian noise samples, independent of the input,
  • Yi∈RY_i \in \mathbb{R} is the channel output.

The encoder maps a message m∈{1,…,2nR}m \in \{1, \ldots, 2^{nR}\} to a codeword x(m)=(x1(m),…,xn(m))\mathbf{x}(m) = (x_1(m), \ldots, x_n(m)) subject to the average power constraint:

1nβˆ‘i=1nxi2(m)≀PforΒ allΒ m.\frac{1}{n} \sum_{i=1}^n x_i^2(m) \leq P \quad \text{for all } m.

AWGN channel

Additive White Gaussian Noise channel: Y=X+ZY = X + Z with Z∼N(0,N)Z \sim \mathcal{N}(0, N) i.i.d. and an average power constraint on the input. The most fundamental continuous-alphabet channel model in information theory.

Related: Signal-to-noise ratio (SNR), Channel capacity

Signal-to-noise ratio (SNR)

The ratio of average signal power to noise power: SNR=P/N\text{SNR} = P/N. In decibels, SNRdB=10log⁑10(P/N)\text{SNR}_{\text{dB}} = 10 \log_{10}(P/N).

Related: AWGN channel

Why an Average Power Constraint?

In practice, transmitters have a finite energy budget (battery, power amplifier limits). The average power constraint 1nβˆ‘xi2≀P\frac{1}{n}\sum x_i^2 \leq P models the total energy per symbol. We could also impose a peak power constraint ∣xiβˆ£β‰€A|x_i| \leq A, but the average constraint is more natural for information-theoretic analysis and yields cleaner results. The peak constraint leads to a harder optimization β€” the capacity-achieving input distribution becomes discrete (Smith, 1971) rather than Gaussian.

Theorem: Capacity of the Scalar AWGN Channel

The capacity of the AWGN channel Y=X+ZY = X + Z with Z∼N(0,N)Z \sim \mathcal{N}(0, N) and average power constraint E[X2]≀P\mathbb{E}[X^2] \leq P is

C=12log⁑(1+SNR)bits per channel use,C = \frac{1}{2}\log\bigl(1 + \text{SNR}\bigr) \quad \text{bits per channel use},

where SNR=P/N\text{SNR} = P/N. The capacity-achieving input distribution is X∼N(0,P)X \sim \mathcal{N}(0, P).

The formula says that doubling the SNR buys you roughly one extra bit per channel use (at high SNR). Intuitively, the Gaussian input spreads energy as "evenly" as possible across the signal space β€” any other distribution with the same power produces less entropy at the output, hence less mutual information.

,

Key Takeaway

The AWGN capacity C=12log⁑(1+SNR)C = \frac{1}{2}\log(1 + \text{SNR}) is the single most important formula in communication theory. The Gaussian input is optimal because it maximizes the output entropy under a power constraint β€” a direct consequence of the entropy maximization property of the Gaussian distribution.

Historical Note: Shannon's 1948 Paper and the Gaussian Channel

Shannon derived the Gaussian channel capacity in his landmark 1948 paper "A Mathematical Theory of Communication." What is remarkable is that Shannon not only gave the formula but also proved both achievability (via random coding with Gaussian codebooks) and the converse β€” all in the same paper that invented the field.

The result was initially met with skepticism: how could one transmit reliably at any positive rate over a noisy channel? The key insight was that coding over long blocks concentrates the noise around a thin shell, and the number of distinguishable signal spheres grows exponentially with the block length. It took nearly 50 years for practical codes (turbo codes, LDPC codes) to approach Shannon's limit within a fraction of a dB.

The Sphere-Packing Picture

There is a beautiful geometric interpretation of the AWGN capacity. Consider transmission of nn symbols:

  • The codeword x(m)\mathbf{x}(m) has energy βˆ₯x(m)βˆ₯2≀nP\|\mathbf{x}(m)\|^2 \leq nP, so all codewords lie in a sphere of radius nP\sqrt{nP} in Rn\mathbb{R}^n.
  • The noise vector ZnZ^{n} concentrates (with high probability) on a thin shell of radius β‰ˆnN\approx \sqrt{nN}.
  • The received vector y=x(m)+Zn\mathbf{y} = \mathbf{x}(m) + Z^{n} lies in a sphere of radius β‰ˆn(P+N)\approx \sqrt{n(P+N)}.

For reliable decoding, the "noise spheres" centered at different codewords must not overlap. The number of non-overlapping noise spheres that fit is

Vol(receivedΒ sphere)Vol(noiseΒ sphere)β‰ˆ(P+NN)n/2=2n2log⁑(1+P/N).\frac{\text{Vol}(\text{received sphere})}{\text{Vol}(\text{noise sphere})} \approx \left(\frac{P+N}{N}\right)^{n/2} = 2^{\frac{n}{2}\log(1 + P/N)}.

Taking 1nlog⁑\frac{1}{n}\log of this count gives exactly C=12log⁑(1+SNR)C = \frac{1}{2}\log(1 + \text{SNR}).

Sphere-Packing Interpretation of AWGN Capacity

Sphere-Packing Interpretation of AWGN Capacity
Codewords (black dots) lie in a signal sphere of radius nP\sqrt{nP}. Each is surrounded by a noise sphere of radius nN\sqrt{nN}. The capacity counts how many non-overlapping noise spheres fit inside the received sphere of radius n(P+N)\sqrt{n(P+N)}.

Sphere-Packing Interpretation of AWGN Capacity

Codewords (blue dots) packed inside a signal sphere, each surrounded by a noise sphere. The capacity counts how many non-overlapping noise spheres fit inside the received sphere. Watch the packing ratio match the Shannon formula C=12log⁑(1+SNR)C = \frac{1}{2}\log(1 + \text{SNR}).

Definition:

The Complex AWGN Channel

The complex AWGN channel models passband communication via complex baseband:

Y=gX+Z,Z∼CN(0,N0),Y = gX + Z, \quad Z \sim \mathcal{CN}(0, N_0),

where g∈Cg \in \mathbb{C} is the (known, deterministic) channel gain and the power constraint is E[∣X∣2]≀Es\mathbb{E}[|X|^2] \leq E_s.

The capacity is

C(SNR)=log⁑(1+SNR)bits per complex symbol,C(\text{SNR}) = \log(1 + \text{SNR}) \quad \text{bits per complex symbol},

where SNR=∣g∣2Es/N0\text{SNR} = |g|^2 E_s / N_0.

The factor-of-two difference from the real case (12log⁑\frac{1}{2}\log vs. log⁑\log) arises because each complex symbol carries two real dimensions.

In wireless communications, the standard convention uses the complex model with SNR=∣g∣2Es/N0\text{SNR} = |g|^2 E_s/N_0. The capacity in bits/s is C=Wlog⁑(1+SNR)C = W \log(1 + \text{SNR}), where WW is the bandwidth in Hz.

Example: Computing AWGN Capacity

A wireless link operates at SNR=20\text{SNR} = 20 dB with bandwidth W=10W = 10 MHz. What is the maximum achievable data rate?

Example: Required SNR for a Target Rate

What minimum SNR\text{SNR} (in dB) is needed to achieve a spectral efficiency of R=4R = 4 bits/s/Hz on a complex AWGN channel?

AWGN Channel Capacity vs. SNR

Explore how the AWGN capacity grows logarithmically with SNR\text{SNR}. At low SNR, capacity grows approximately linearly; at high SNR, each 3 dB increase adds roughly 1 bit/s/Hz.

Parameters
30

Quick Check

For an AWGN channel with SNR=7\text{SNR} = 7 (linear, not dB), what is the capacity in bits per real channel use?

12log⁑2(8)=1.5\frac{1}{2}\log_2(8) = 1.5 bits

log⁑2(8)=3\log_2(8) = 3 bits

12log⁑2(7)β‰ˆ1.40\frac{1}{2}\log_2(7) \approx 1.40 bits

log⁑2(7)β‰ˆ2.81\log_2(7) \approx 2.81 bits

Common Mistake: Real vs. Complex AWGN: The Factor of Two

Mistake:

Using C=log⁑2(1+SNR)C = \log_2(1 + \text{SNR}) for a real-valued AWGN channel (or 12log⁑2(1+SNR)\frac{1}{2}\log_2(1+\text{SNR}) for a complex channel).

Correction:

Real AWGN: C=12log⁑2(1+SNR)C = \frac{1}{2}\log_2(1 + \text{SNR}) bits per real symbol. Complex AWGN: C=log⁑2(1+SNR)C = \log_2(1 + \text{SNR}) bits per complex symbol. The complex channel has two real dimensions, hence the factor of two. When computing bits/s, both give Cbits/s=Wlog⁑2(1+SNR)C_{\text{bits/s}} = W\log_2(1 + \text{SNR}) because the complex symbol rate is half the real sample rate.

Common Mistake: SNR in dB vs. Linear

Mistake:

Plugging SNRdB\text{SNR}_{\text{dB}} directly into the capacity formula: C=12log⁑(1+20)C = \frac{1}{2}\log(1 + 20) when SNR=20\text{SNR} = 20 dB.

Correction:

Always convert to linear scale first: SNR=1020/10=100\text{SNR} = 10^{20/10} = 100. Then C=12log⁑(1+100)β‰ˆ3.34C = \frac{1}{2}\log(1 + 100) \approx 3.34 bits (real) β€” very different from 12log⁑(21)β‰ˆ2.18\frac{1}{2}\log(21) \approx 2.18 bits.

Theorem: MMSE Lower Bound via Differential Entropy

For jointly distributed continuous random variables XX and YY, the minimum mean-square error (MMSE) satisfies

E[∣Xβˆ’x^(Y)∣2]β‰₯12Ο€e 22h(X∣Y),\mathbb{E}\bigl[|X - \hat{x}(Y)|^2\bigr] \geq \frac{1}{2\pi e}\, 2^{2h(X|Y)},

where x^(Y)=E[X∣Y]\hat{x}(Y) = \mathbb{E}[X|Y] is the optimal estimator. Equality holds if and only if XX given YY is Gaussian.

This theorem connects two seemingly different worlds: estimation theory (MMSE) and information theory (entropy). It says that the conditional entropy places a fundamental lower bound on how well you can estimate XX from YY. The Gaussian case is the "hardest to estimate" for a given entropy β€” any other conditional distribution with the same entropy is easier to estimate.

Example: Verifying the MMSE Bound for Jointly Gaussian Variables

Let (X,Y)(X, Y) be jointly Gaussian with zero mean, Var(X)=Οƒx2\text{Var}(X) = \sigma_x^2, Var(Y)=Οƒy2\text{Var}(Y) = \sigma_y^2, and correlation coefficient ρ\rho. Verify that the MMSE lower bound holds with equality.

Why This Matters: AWGN Capacity and Spectral Efficiency in 5G NR

The AWGN capacity formula C=Wlog⁑2(1+SNR)C = W\log_2(1 + \text{SNR}) is the benchmark against which every practical modulation and coding scheme is measured. In 5G NR, adaptive modulation and coding (AMC) selects the highest-rate modulation-coding scheme (MCS) that the current SNR\text{SNR} can support. The Shannon limit tells us how close each MCS comes to the theoretical maximum. Modern LDPC and polar codes in 5G NR operate within 1–2 dB of the AWGN capacity for long block lengths.

See Book telecom, Ch. 14 for the detailed treatment of AMC and link adaptation, and Book telecom, Ch. 12 for the coding schemes that approach this limit.

Quick Check

If we double the transmit power PP (keeping noise NN fixed), by how much does the AWGN capacity increase?

It exactly doubles

It increases by 12\frac{1}{2} bit per real channel use

It increases by 12log⁑2(2P/N)βˆ’12log⁑2(P/N)=12log⁑2(2)=0.5\frac{1}{2}\log_2(2P/N) - \frac{1}{2}\log_2(P/N) = \frac{1}{2}\log_2(2) = 0.5 bits only when P≫NP \gg N

It does not change