The Finite Blocklength Regime
Beyond the Shannon Limit: Why Blocklength Matters
Shannon's channel coding theorem guarantees that rates arbitrarily close to capacity are achievable with vanishing error probability --- provided the blocklength . In classical broadband communications with large packets (), this asymptotic result is an excellent approximation.
However, URLLC services in 5G NR demand end-to-end latencies of 1 ms or less, which restricts the blocklength to -- channel uses. At such short blocklengths, the Shannon limit is overly optimistic: there is a significant penalty for operating at finite . This section develops the precise characterisation of that penalty through the normal approximation of Polyanskiy, Poor, and Verd'{u} (2010).
Definition: Information Density
Information Density
Consider a memoryless channel with transition kernel . For input and output , the information density is the random variable
where the logarithm is base 2 (so the units are bits). Its expectation is the mutual information:
The information density captures the per-sample information content, which fluctuates from one channel use to the next.
Definition: Channel Dispersion
Channel Dispersion
The channel dispersion of a memoryless channel, at the capacity-achieving input distribution , is the variance of the information density:
For the real AWGN channel with SNR :
For the complex AWGN channel:
Channel dispersion is the information-theoretic analogue of variance in the central limit theorem. A high-dispersion channel pays a larger finite-blocklength penalty because the accumulated information is "noisier" across uses.
Theorem: Normal Approximation (Polyanskiy-Poor-Verd'u)
Interpreting the Dispersion Penalty
The normal approximation reveals a clean structure:
- Capacity is the asymptotic rate, approached only as .
- Dispersion penalty decreases as --- convergence to capacity is slow. Halving the gap requires quadrupling the blocklength.
- Reliability cost: since increases as , demanding higher reliability (smaller ) further reduces the achievable rate.
For URLLC with and , the rate loss compared to Shannon capacity can exceed 30%.
Example: Rate Loss at Short Blocklength over AWGN
Consider a complex AWGN channel at SNR dB. Compute the maximum achievable rate using the normal approximation for: (a) , ; (b) , ; (c) , . Express results as a fraction of the Shannon capacity .
Shannon capacity and channel dispersion
At (i.e., 10 dB):
The complex AWGN dispersion is
Case (a): $n = 100$, $\varepsilon = 10^{-3}$
We have , so
This is of Shannon capacity.
Case (b): $n = 100$, $\varepsilon = 10^{-5}$
Now :
which is of capacity. Tightening reliability from to costs an additional of capacity.
Case (c): $n = 1000$, $\varepsilon = 10^{-5}$
94.4%nO(1/\sqrt{n})$ convergence.
Finite Blocklength: Rate Penalty as Error Probability Varies
Rate vs Blocklength under Normal Approximation
Explore how the maximum achievable rate varies with blocklength for different target error probabilities . The plot shows the normal approximation curves alongside the Shannon capacity (dashed horizontal line). Observe that (i) convergence to is slow (), (ii) stricter reliability requirements (smaller ) shift the curve downward, and (iii) at short blocklengths (), the rate penalty can exceed 20--40% of capacity.
Parameters
Theorem: Finite Blocklength under Quasi-Static Fading
Quick Check
The channel dispersion of a complex AWGN channel at high SNR () behaves approximately as . What does this imply about the finite blocklength penalty at high SNR?
The penalty becomes independent of SNR, so increasing SNR only shifts the capacity but does not reduce the rate gap
The penalty vanishes at high SNR because the channel becomes deterministic
The penalty grows without bound because increases linearly with SNR
The penalty depends on only through the ratio , which vanishes
At high SNR, bits/use, which is a constant. Thus the absolute rate penalty saturates. However, since grows as , the relative penalty (as a fraction of ) decreases with SNR.
Achievability and Converse Bounds
The normal approximation is sandwiched between two non-asymptotic bounds:
-
Random Coding Union (RCU) bound (achievability): for any codebook size and optimal ML decoder,
where is an independent codeword drawn from the same distribution.
-
Meta-converse bound (converse): the error probability of any code is bounded below via a binary hypothesis testing formulation:
where is the minimum type-II error probability at level .
For the AWGN channel, these bounds are remarkably tight: the gap between achievability and converse is less than 0.5 dB even at .
Quick Check
To halve the gap between and the Shannon capacity (at a fixed ), how must the blocklength change?
Double it ()
Quadruple it ()
Square it ()
It depends on and cannot be determined
The gap is , which scales as . To halve this gap, we need , giving . This slow convergence is a fundamental challenge for URLLC, where latency constraints severely limit .
Historical Note: From Shannon to Polyanskiy: 62 Years to the Finite Blocklength Answer
1948--2010Shannon's 1948 paper established channel capacity as the supremum of achievable rates with vanishing error probability as . But the question "how fast does the achievable rate approach ?" proved remarkably difficult.
Strassen (1962) showed that the convergence rate is and identified the channel dispersion for the DMC, but the result attracted limited attention at the time. Hayashi (2009) independently developed second-order asymptotics using the information spectrum method.
The breakthrough came with Polyanskiy, Poor, and Verdú (2010), who provided tight achievability (random coding union bound) and converse (meta-converse) bounds that pinned down to within fractions of a dB for the AWGN channel. Their normal approximation became the standard engineering formula for short-packet design and directly enabled the analytical framework for 5G URLLC.
The finite blocklength paradigm fundamentally changed how wireless engineers think about latency: it showed that reliability at short blocklength has a precise, quantifiable cost in rate, governed by the channel dispersion.
Common Mistake: Using Shannon Capacity to Design Short-Packet Systems
Mistake:
"The AWGN channel at dB has capacity bit/channel use, so I can reliably transmit at rate bits/use with a code of blocklength ."
Correction:
At and target error probability , the normal approximation gives:
For the AWGN channel at dB, the dispersion is bit/channel use, and . Therefore:
The actual achievable rate is 0.574 bits/use, not 0.9. The 0.426 bits/use gap is the dispersion penalty, which is enormous at short blocklength. Designing at would result in error probability , not . The Shannon limit is only a useful design target when .
Channel Dispersion
The variance of the information density under the capacity-achieving input distribution. Denoted , it governs the penalty in the normal approximation. For the AWGN channel at SNR : .
Related: Channel Dispersion, Normal Approximation (Polyanskiy-Poor-Verd'u)
Normal Approximation (Polyanskiy-Poor-Verdú)
The second-order asymptotic expansion . Accurate to within 0.5 dB for on the AWGN channel.
Related: Normal Approximation (Polyanskiy-Poor-Verd'u), Channel Dispersion
Blocklength
The number of channel uses allocated to encoding a single message. In OFDM-based systems, equals the number of resource elements assigned to the codeword. Shorter blocklength reduces latency but increases the rate penalty .
Related: Information Density
Key Takeaway
The dispersion penalty is the price of low latency. The normal approximation shows that short-packet communication operates fundamentally below Shannon capacity. The gap scales as : halving the gap requires quadrupling the blocklength. For URLLC with and , the rate loss can exceed 40% of capacity — a fact that Shannon's asymptotic theory cannot predict.