Achievability and Converse Bounds
Beyond the Normal Approximation: Exact Bounds
The normal approximation is an asymptotic expansion: it becomes accurate as grows. But for very short blocklengths (-), the remainder can be significant. We need non-asymptotic bounds that are tight for any .
Polyanskiy, Poor, and Verdu provided two such bounds: the random coding union (RCU) bound for achievability and the meta-converse ( bound) for the converse. Together, these bounds sandwich to within a fraction of a bit for most channels of practical interest.
The key insight is to replace typicality-based arguments (which are inherently asymptotic) with hypothesis testing arguments (which work for any ).
RCU Bound and Meta-Converse: Tight Sandwich
Definition: Random Coding Union (RCU) Bound
Random Coding Union (RCU) Bound
For a DMC with input distribution , the average error probability of a random code with codewords drawn i.i.d. from satisfies:
where is the transmitted codeword and received signal, and is an independent codeword. Equivalently:
where is drawn independently from the same codebook distribution.
The RCU bound implies the existence of an -code.
The RCU bound is a refinement of the classical random coding bound. Instead of using the union bound over all incorrect codewords (which is loose), it uses the exact probability that any incorrect codeword has higher information density than the correct one. This is tight because the dominant error event is typically a single confusion, not multiple.
Theorem: RCU Achievability Bound
For any DMC and any input distribution , there exists an -code with:
For the AWGN channel with and i.i.d. inputs, this simplifies to:
where the expectation is over the random information density .
The RCU bound says: "draw a random code, compute the probability that any wrong codeword looks better than the right one, and average over the channel randomness." This is tight because it captures the exact pairwise confusion probability, not an upper bound on it.
Error event decomposition
The ML decoder makes an error if there exists such that , where is the transmitted message. By the union bound: . The RCU improves this by taking the min with 1: .
Averaging over the random code
Since all codewords are drawn i.i.d., the inner sum becomes times the probability for a single independent codeword . Averaging over the transmitted codeword and the channel output yields the RCU bound.
Existence argument
The RCU bound is an average over the random code ensemble. Therefore, there exists at least one deterministic code achieving error probability no worse than .
Definition: The Function (Hypothesis Testing)
The Function (Hypothesis Testing)
For two probability distributions and on the same alphabet, the minimum type-II error at significance level is:
where the minimum is over all randomized tests . By the Neyman-Pearson lemma, the optimal test is the likelihood ratio test:
where and are chosen so that .
Theorem: The Meta-Converse ( Bound)
For any DMC and any -code:
for every codeword in the code, where the supremum is over all output distributions .
For the average error probability formulation:
where is the conditional error probability for message and .
The meta-converse connects coding to hypothesis testing. The idea is: if a code can reliably distinguish codewords, then each codeword-output distribution must be "far" from the background distribution . The function measures this distance. The larger is, the harder it is for all codewords to be far from , which gives the converse.
The point is that this bound works for any , not just asymptotically. It replaces Fano's inequality with a hypothesis-testing argument that is tight to second order.
Hypothesis testing formulation
Consider testing vs for each message . The decoder's decision region acts as a test: it accepts with probability (correct decoding) and has type-II error .
Union of decision regions
Since the decision regions are disjoint and cover the output space: . Therefore: .
Bound on $M$
By Jensen's inequality (or direct averaging): , giving the bound on . The supremum over makes the bound as tight as possible.
RCU Achievability vs Meta-Converse
Compare the RCU bound (achievability), the meta-converse (converse), and the normal approximation. For many channels, the gap between RCU and meta-converse is remarkably small, even at short blocklengths.
Parameters
Common Mistake: Confusing Dispersion with Error Exponent
Mistake:
Believing that the error exponent and the channel dispersion capture the same information about finite-blocklength performance.
Correction:
The error exponent describes how decays at a fixed rate below capacity as . The dispersion describes the rate penalty at a fixed as varies. They answer different questions:
- Error exponent: "At rate , how fast does go to zero?"
- Normal approximation: "At error probability , what rate can we achieve at blocklength ?"
For system design, the normal approximation is usually more useful because the designer specifies and wants to know the achievable rate, not the other way around.
Common Mistake: Trusting the Normal Approximation at Very Short Blocklengths
Mistake:
Using for , where the CLT approximation of the information density may be inaccurate.
Correction:
For -, use the exact RCU and meta-converse bounds, which do not rely on CLT. The normal approximation has an error term that can be 0.1-0.3 bits/use at , which is significant relative to the term. The Berry-Esseen refinement helps: .
Example: Computing the RCU Bound for BSC
For the BSC with crossover probability and blocklength , compute the RCU bound on the maximum code size at error probability .
Setup
With uniform input and BSC, the information density for a pair with bit errors is: (in nats) where is the binary entropy function.
RCU bound evaluation
The RCU bound is where .
For a BSC, depends on the number of errors in the true pair: (probability that a random codeword has at most disagreements with ).
The RCU bound is then: .
Numerical result
Evaluating numerically with , : bits, bits. Normal approximation: . Exact RCU: (slightly tighter than the normal approximation).
How Close Do Practical Codes Get to the PPV Bounds?
Modern channel codes approach the finite-blocklength bounds remarkably well. For the AWGN channel at and :
- Polar codes (SCL decoding, list size 32): within 0.5 dB of the RCU bound
- LDPC codes (5G NR base graph): within 0.7 dB of the RCU bound at , but degrade faster at shorter blocklengths
- Turbo codes: within 0.3 dB at , but with higher decoding complexity
- Tail-biting convolutional codes (used in LTE control channels): within 1 dB at
The PPV bounds serve as the ultimate benchmark: if a code is within 0.5 dB of the meta-converse, there is very little room for improvement at that blocklength.
- ā¢
5G NR uses polar codes for control channels (n = 32-1024) and LDPC for data (n = 256-8448)
- ā¢
Decoding latency scales with list size in SCL: L=8 is practical, L=32 is costly
- ā¢
At n < 64, algebraic codes (Reed-Muller, BCH) can outperform capacity-approaching codes
Key Takeaway
RCU + meta-converse = tight sandwich. The random coding union bound (achievability) and the meta-converse (converse) together characterize to within a fraction of a bit for most channels. They replace the asymptotic achievability-converse pair (random coding + Fano) with non-asymptotic hypothesis-testing arguments that work at any blocklength. For practical system design, the normal approximation is accurate for .
Random coding union bound
A non-asymptotic achievability bound that upper-bounds the error probability of a random code by the probability that any incorrect codeword has higher information density than the correct one.
Related: Meta-converse
Meta-converse
A non-asymptotic converse bound that lower-bounds the minimum error probability using hypothesis testing between the channel output distribution and a reference distribution. Also called the bound.
Related: Random coding union bound