Exercises
ex-ch26-01
EasyCompute the capacity and dispersion of the BEC with erasure probability . Determine the blocklength needed for rate bits/use at .
BEC capacity: . BEC dispersion: .
Use and solve for .
Capacity and dispersion
bits/use. bits.
Required blocklength
. . .
ex-ch26-02
EasyShow that the information density of the BSC with crossover probability at the capacity-achieving uniform input is: (in bits, with ).
when and when .
The output distribution under uniform input is for all .
Output distribution
With : . Similarly .
Information density
When : . When : .
Verification
. Correct.
ex-ch26-03
EasyFor the AWGN channel at dB, compute the maximum coding rate at blocklength and error probability . How close is this to the Shannon capacity?
At SNR = 20 dB (= 100 linear): bits/use.
Dispersion: in nats.
Capacity
bits/use.
Dispersion
nats. bits.
Normal approximation
bits/use. This is of capacity. At high SNR with , the gap is modest.
ex-ch26-04
EasyVerify that the AWGN dispersion satisfies as and as (in nats).
Take limits directly.
Low SNR limit
As : .
High SNR limit
As : .
Interpretation
At low SNR, the channel is nearly deterministic (almost all noise, very little signal), so the information density has small variance. At high SNR, the variance saturates at 1/2, which is the variance of when .
ex-ch26-05
EasyA binary communication system uses a BSC with . How many channel uses are needed to transmit information bits with error probability ?
BSC(0.05): bits/use, bits.
Need . Iterate or solve approximately.
Channel parameters
bits/use. . Actually: bits.
Required blocklength
. Need .
Try : . Bits: . Works.
Try : . Bits: . Too short.
Approximately channel uses.
ex-ch26-06
MediumDerive the dispersion of the BSC from the information density. Specifically, show that: (in the base matching the logarithm used for capacity).
The information density takes two values: (no error) and (error).
Use .
Information density values
With base-2 logarithm and uniform input: with probability , with probability .
Variance
. Note and , so and .
.
ex-ch26-07
MediumShow that the RCU bound for the AWGN channel with Gaussian codebook gives:
where is the sphere-packing exponent, in the limit of large . Explain why this recovers the error exponent at rates below capacity.
The confusion probability is related to the probability that a random codeword falls inside a sphere around the received signal.
For Gaussian codes, this involves the chi-squared distribution.
Confusion probability
For Gaussian codebook and ML decoding, the probability that a random codeword has higher likelihood than the true codeword given output is: .
Since and , this involves the distribution of vs .
Large-$n$ asymptotics
For large , by the law of large numbers and Cramer's theorem: where is the sphere-packing exponent. The RCU bound becomes: which gives where is the random coding exponent.
ex-ch26-08
MediumFor the AWGN channel, compute the SNR penalty (in dB) for operating at vs at , for dB. Plot the penalty as a function of SNR.
At each SNR, compute and find the SNR needed to achieve this rate at .
The penalty is in dB.
At SNR = 0 dB
bit/use, bits. bits/use. SNR for : . dB.
At SNR = 10 dB
bits/use, bits. bits/use. SNR for : (18.5 dB). dB. Note: this is large because is small at high SNR.
Summary table
| SNR (dB) | (dB) | ||
|---|---|---|---|
| 0 | 1.00 | 0.67 | 1.8 |
| 5 | 1.46 | 1.08 | 2.7 |
| 10 | 3.46 | 3.08 | 8.5 |
| 20 | 6.66 | 6.21 | large |
The penalty increases with SNR because flattens out.
ex-ch26-09
MediumProve that the Neyman-Pearson function satisfies as , provided is fixed.
This is Stein's lemma. Use the fact that the log-likelihood ratio concentrates around .
Optimal test
The Neyman-Pearson test compares to a threshold . Under : a.s. by SLLN.
Type-II error analysis
Under : a.s. The threshold must satisfy , so (CLT correction).
Exponential decay
Under : by Cramer's theorem. Therefore .
ex-ch26-10
MediumFor a Rayleigh fading channel where is known at the receiver, , and :
(a) Compute the capacity .
(b) Show that the dispersion includes a fading term: .
The channel is conditionally AWGN given . Use the law of total variance.
, so involves the second moment of the log-exponential.
Capacity
(in nats), where is the exponential integral. For : nats bits/use.
Dispersion decomposition
By the law of total variance: .
The first term is the average AWGN dispersion across fading realizations. The second term is the fading dispersion — the randomness of capacity across fading states. For Rayleigh fading, this second term dominates at low SNR.
ex-ch26-11
MediumConsider the Z-channel with , , . Compute the capacity-achieving input distribution, the capacity, and the dispersion for .
The capacity-achieving input is where satisfies a transcendental equation.
The information density has three possible values depending on .
Capacity-achieving input
The capacity is . For , the optimal and bits/use.
Dispersion computation
The information density takes values depending on : : with prob : with prob : with prob
. For : bits.
ex-ch26-12
MediumA URLLC system operates over a quasi-static Rayleigh fading channel (one fading realization per codeword). Show that the outage probability lower-bounds the error probability at any blocklength:
and that no finite-blocklength code can beat the outage probability in this regime.
In quasi-static fading, the channel realization is fixed for the entire codeword.
Conditioned on , the channel is AWGN with SNR .
Conditional analysis
Given , the channel is AWGN with capacity . If , then no code can achieve error probability on this realization (by the converse of the channel coding theorem for the AWGN channel).
Unconditional bound
.
A tighter argument using the meta-converse shows , confirming that outage dominates at all finite blocklengths.
Implications
In quasi-static fading, diversity is essential: without it, the error probability is fundamentally limited by the outage probability, regardless of code length. The normal approximation is not useful here because the fading "dispersion" is infinite (the variance of does not shrink with when is constant over the block).
ex-ch26-13
HardProve the achievability part of the normal approximation using the Berry-Esseen theorem. Specifically, show that for a DMC with capacity , dispersion , and third absolute moment :
for a universal constant .
Start with the RCU bound and bound the confusion probability using the CLT + Berry-Esseen.
The Berry-Esseen theorem gives .
RCU bound simplification
From the RCU bound: there exists a code with error if . This is equivalent to where is chosen to control the error.
Berry-Esseen application
Set where . By Berry-Esseen: . The confusion probability .
Rate bound
. Dividing by : . The precise third-order term involves .
ex-ch26-14
HardDerive the meta-converse bound for the BSC at blocklength and error probability . Compute the exact maximum code size by evaluating numerically.
For the BSC with uniform , the function can be computed exactly using the binomial distribution.
The likelihood ratio between and is a function of the Hamming weight.
Likelihood ratio
For BSC() with transmitted and received : where is the Hamming distance.
Neyman-Pearson test
Under : . Under : . The optimal test accepts when for some threshold . Set so that .
Meta-converse bound
.
For , , : find such that , then evaluate .
Numerically: gives . . . bits.
ex-ch26-15
HardShow that for a channel with zero dispersion (), the finite-blocklength rate converges to capacity at rate instead of . Give an example of such a channel.
A channel has if the information density is deterministic: with probability 1.
The noiseless binary channel has .
Deterministic information density
If , then with probability 1 for each . The cumulative information density deterministically. There is no CLT correction; the convergence to capacity is determined by the term in the finite-blocklength expansion.
Example: noiseless channel
The noiseless binary channel () has bit/use and bit deterministically under uniform input. So . At any finite : for any .
BEC as intermediate case
The BEC has (the erasure creates randomness in ), but some channels approach behavior. Geometrically, means the typical set has extremely sharp boundaries, so random coding works almost immediately.
ex-ch26-16
HardFor the two-user Gaussian MAC with and , compute the full dispersion matrix at . Verify that the (3,3) entry equals the point-to-point dispersion at SNR .
The three information densities are: , , .
and are conditionally independent given .
Individual rate densities
is the AWGN information density at SNR . nats.
Sum-rate density
is the AWGN information density at SNR . nats.
Cross-correlations and matrix
and are computed from the joint distribution of . Since (chain rule), the correlations are non-zero.
The full matrix has entries computable from the moments of . The (3,3) entry is nats, confirming it equals .
ex-ch26-17
HardSaddlepoint approximation. The normal approximation can be refined using the saddlepoint method. For the AWGN channel, the cumulant generating function of the information density is:
(a) Compute for the real AWGN channel with Gaussian input.
(b) Show that the saddlepoint approximation gives: where is the saddlepoint satisfying a specific equation.
For Gaussian variables, the cumulant generating function involves the moment generating function of chi-squared distributions.
The saddlepoint approximation is more accurate than the normal approximation at the tails.
CGF for AWGN
The information density for AWGN with Gaussian input is a quadratic form in Gaussian variables. Its CGF is: (up to constants depending on the SNR parametrization).
Saddlepoint equation
The saddlepoint satisfies where is the threshold in the hypothesis test. This gives a more accurate approximation than the CLT because it accounts for the skewness of the information density.
Practical impact
The saddlepoint approximation matches the exact RCU/MC bounds to within 0.01 bits/use for , compared to 0.05-0.1 bits/use for the normal approximation. This makes it the tool of choice for precise finite-blocklength analysis in system design.
ex-ch26-18
ChallengeResearch-level. Derive the second-order coding rate for the Gaussian MIMO channel with known at the receiver. Show that the dispersion is:
where are the per-stream SNRs after water-filling, and are the singular values of .
After SVD of , the MIMO channel decomposes into parallel AWGN channels.
The total information density is the sum of per-stream information densities, which are independent.
The dispersion of a sum of independent variables is the sum of individual dispersions.
SVD decomposition
Write and transform to parallel channels: for , with water-filling power .
Per-stream dispersion
Each parallel channel has AWGN dispersion in nats.
Total dispersion
Since the streams are independent (after unitary transformation): . The normal approximation gives: .
ex-ch26-19
ChallengeOpen problem flavor. The meta-converse for the fading MAC with users and blocklength is not fully characterized when grows with (the many-access regime). Formulate the problem for active users, each sending bits, and derive an achievability bound using random Gaussian codebooks.
In the many-access regime, the traditional MAC capacity region is not the right benchmark.
Use the per-user error probability (PUPE) as the performance metric.
The achievable scales as .
System model
active users, each with bits, blocklength , over AWGN MAC. Each user transmits where is drawn from a Gaussian codebook of size .
Per-user achievability
Using the RCU bound for the Gaussian MAC: the PUPE satisfies where .
Required energy-per-bit
Setting : . For and : which goes to 0.
However, the per-user energy must satisfy , giving and .
ex-ch26-20
ChallengeImplementation project. Write a simulation that computes the exact RCU bound and meta-converse for the BSC with at blocklengths and error probabilities .
(a) Plot the exact bounds alongside the normal approximation.
(b) Quantify the gap between the normal approximation and the exact bounds.
(c) Overlay the performance of the best known codes (e.g., BCH, Reed-Muller, polar) from the literature.
The BSC allows exact computation via binomial sums.
For the meta-converse, optimize over the output distribution (or use the uniform output, which is often optimal for the BSC).
RCU computation
For BSC with uniform codebook: where is the confusion probability at Hamming distance .
Meta-converse computation
with : find threshold such that , then .
Expected results
The normal approximation is within 0.5 bits of the exact bounds for . For , the gap can be 1-2 bits. Best known codes (BCH for small , polar for large ) are typically within 1-2 dB of the meta-converse.