Cutoff Rate and Practical PEP Bounds

Below the Cutoff, Decoders Have Work Exponential in Rate

The Gallager random-coding exponent of §4 treats the decoder as a maximum-metric decoder — the receiver computes the metric for every codeword in the codebook and picks the largest. That is conceptually clean but operationally expensive: the decoder's work is exponential in the blocklength times the rate. Real BICM decoders — LDPC belief propagation, sequential decoders, list decoders — are not of this type.

For sequential decoders on tree codes, Massey-Fano-Jelinek-Zigangirov showed in the 1960s that the practical rate limit is not the capacity but the cutoff rate R0=E0(1)R_0 = E_0(1): at rates R<R0R < R_0, the expected decoder work per decoded bit is finite; at rates R>R0R > R_0, the expected work diverges. For list decoders with list size growing polynomially in NN, the same threshold separates "polynomial work" from "exponential work". The cutoff rate is therefore the practical rate limit of any non-ML decoder on a binary-input channel, and its BICM specialisation R0BICMR_0^{\rm BICM} is the quantity every receiver designer cares about.

Beyond the cutoff rate, §5 also revisits the pairwise error probability (PEP) of a BICM codeword pair on a fading channel from Ch. 6. The classical Caire-Taricco-Biglieri 1998 union bound uses a Bhattacharyya-factor product over distinct bit positions; the improved bound of Martinez, Guillén i Fàbregas, and Caire 2006 replaces this with the GMI-tuned saddle-point at each distance, tightening the bound by 1–2 dB at moderate SNR. Both bounds share the same exponential order (same diversity, same coding gain), but the prefactor in the Martinez-Fàbregas-Caire bound is sharper — a direct benefit of the mismatched-decoding framework.

This section defines R0R_0 precisely, develops its BICM specialisation, compares it to the CM cutoff rate, and derives the saddle-point PEP bound. We close with engineering and pitfall remarks on what the cutoff rate does and does not mean for practical decoders.

,

Definition:

Gallager Cutoff Rate R0R_0

The cutoff rate of a memoryless channel p(yx)p(y\mid x) with input distribution qq is the value of Gallager's E0E_0 at ρ=1\rho = 1: R0(q)  =  E0(ρ=1,q)  =  log ⁣y ⁣(xq(x)p(yx))2dy.R_0(q) \;=\; E_0(\rho = 1, q) \;=\; -\log \!\int_y \!\Big( \sum_x q(x) \sqrt{p(y\mid x)} \Big)^{2}\, dy. Equivalently, for binary-input channels with uniform input, R0  =  1log2(1+β),β=yp(y0)p(y1),R_0 \;=\; 1 - \log_2(1 + \beta), \quad \beta = \sum_y \sqrt{p(y\mid 0) p(y\mid 1)}, where β\beta is the Bhattacharyya parameter of the channel. R0R_0 has units of bits per channel use under log2\log_2, nats under ln\ln.

For BICM with the product bit metric, the cutoff rate per symbol is R0BICM  =  =0L1[1log2(1+β)],β=ypW(y0)pW(y1).R_0^{\rm BICM} \;=\; \sum_{\ell = 0}^{L-1} [\, 1 - \log_2(1 + \beta_\ell)\, ], \qquad \beta_\ell = \sum_y \sqrt{p_{W_\ell}(y\mid 0) p_{W_\ell}(y\mid 1)}. This is just the parallel-channel cutoff-rate sum over the LL bit channels — a direct consequence of the parallel-channel decomposition of §1.

Operational meaning. For any decoder with polynomial complexity in the blocklength NN (sequential, list, threshold), the rate R0R_0 is the threshold above which the expected complexity grows exponentially in RR0R - R_0. Below R0R_0, complexity is polynomial — or even linear for a well-tuned sequential decoder. This makes R0R_0 the practical rate limit of non-ML decoders.

Three comments on R0R_0:

  1. R0R_0 is always \le capacity: R0=E0(1)E0(0)1=CR_0 = E_0(1) \le E_0'(0) \cdot 1 = C by convexity of E0(ρ)E_0(\rho) with E0(0)=0E_0(0) = 0. At low SNR R0CR_0 \approx C; at high SNR the gap grows but remains O(C)O(C).
  2. Gallager thought R0R_0 was the "true" practical capacity until the invention of LDPC decoding in the late 1990s. LDPC+BP is an exception: it can operate above the cutoff rate with polynomial complexity, because its complexity depends on the graph sparsity, not on a search over codeword candidates. This is part of why LDPC
    • BICM became standard.
  3. Even so, R0BICMR_0^{\rm BICM} remains a useful sanity check: when a 5G NR LDPC decoder is reported to operate within 0.3–0.5 dB of Shannon, one can verify that the corresponding rate is above R0R_0 — confirming that the decoder is extracting information beyond the Gallager threshold.
,

Theorem: R0BICMR0CMR_0^{\mathrm{BICM}} \le R_0^{\mathrm{CM}} and R0IR_0 \le I

For any memoryless channel and constellation X\mathcal{X} with labelling μ\mu:

(i) Cutoff rate is \le capacity: R0BICM(μ)    CBICM(μ),R0CM    CCM.R_0^{\rm BICM}(\mu) \;\le\; C_{\rm BICM}(\mu), \qquad R_0^{\rm CM} \;\le\; C_{\rm CM}. Both with equality only at zero SNR (trivial channel), and the gap grows as a fraction of capacity with SNR — reaching 30\sim 3050%50\% of capacity at moderate SNR for QAM.

(ii) BICM cutoff rate \le CM cutoff rate: R0BICM(μ)    R0CM,R_0^{\rm BICM}(\mu) \;\le\; R_0^{\rm CM}, with equality at zero SNR and strict inequality for M16M \ge 16 at positive SNR. The ordering mirrors the capacity ordering of Ch. 5.

R0=E0(1)R_0 = E_0(1) is a point evaluation of the Gallager function at ρ=1\rho = 1; since E0BICM(ρ)E0CM(ρ)E_0^{\rm BICM}(\rho) \le E_0^{\rm CM}(\rho) pointwise (Thm. \le CM Exponent" data-ref-type="theorem">TBICM Random-Coding Exponent \le CM Exponent), the ρ=1\rho = 1 values inherit the ordering. The capacity-vs-cutoff ordering is standard from Gallager's convexity argument.

,

Three PEP Bounds at Fixed Hamming Distance dHd_H

Compare three different bounds on the pairwise error probability P(cc)P(\mathbf{c} \to \mathbf{c}') where c\mathbf{c} and c\mathbf{c}' differ in dHd_H bit positions, for BICM on AWGN with 16-QAM and Gray labelling:

  1. Caire-Taricco-Biglieri 1998 union bound (Ch. 6): a product of Bhattacharyya factors, one per distinct bit position.
  2. Chernoff/union with scaling s=1s = 1: the GMI-style bound at the fixed decoder scaling.
  3. Martinez-Fàbregas-Caire 2006 saddle-point bound: optimise the scaling ss at each dHd_H to tighten the bound.

The saddle-point bound is 1–2 dB tighter than the union bound at moderate SNR for the same dHd_H — a non-trivial improvement that makes PEP-based union bounds on BICM much more accurate than the classical ones. At high SNR and low SNR all three bounds converge to the same exponential order.

Parameters
10

Theorem: Martinez-Fàbregas-Caire Saddle-Point PEP Bound

Let c,c{0,1}N\mathbf{c}, \mathbf{c}' \in \{0,1\}^{N} be two codewords differing in dHd_H bit positions, transmitted over a BICM system with the product bit metric. The pairwise error probability satisfies P(cc)    mins>0  n:cncnEY,B ⁣[(pW(YBˉ)pW(YB))s],P(\mathbf{c} \to \mathbf{c}') \;\le\; \min_{s > 0}\; \prod_{n: c_n \ne c'_n} \mathbb{E}_{Y, B_\ell}\!\left[\, \Big(\frac{p_{W_\ell}(Y\mid \bar B_\ell)}{p_{W_\ell}(Y\mid B_\ell)}\Big)^{s} \right], where Bˉ=1B\bar B_\ell = 1 - B_\ell is the complementary bit at position \ell, and the product runs over the dHd_H distinct bit positions. The minimisation over s>0s > 0 is the saddle-point optimisation.

At s=1/2s = 1/2 this reduces to the Bhattacharyya-factor product of the Caire-Taricco-Biglieri 1998 bound; optimising over ss gives the tighter Martinez-Fàbregas-Caire 2006 bound. The improvement is 1\sim 122 dB at moderate dHd_H and moderate SNR; at large dHd_H the saddle-point approaches s=1/2s = 1/2 and the improvement vanishes.

The classical union bound uses a Chernoff-type tilt at s=1/2s = 1/2 — the "neutral" choice that symmetrically weights the two codewords. The saddle-point argument optimises the tilt per distance: for short distances (where the decoder's "surprise" is largest) the optimal ss is smaller, placing more weight on the transmitted codeword; for long distances, ss approaches 1/21/2 and the bound reduces to the Bhattacharyya form. The GMI framework naturally provides this tilt optimisation as part of the same machinery used for capacity.

,

Example: Saddle-Point vs Union Bound: 16-QAM Gray at 8 dB, dH=10d_H = 10

For BICM with 16-QAM Gray labelling on AWGN at SNR=8\text{SNR} = 8 dB and Hamming distance dH=10d_H = 10, compute (i) the CTB union-bound PEP, (ii) the MFC saddle-point PEP, and (iii) compare the two in dB of SNR margin.

,

Common Mistake: Cutoff Rate Is Not Capacity

Mistake:

Quoting R0BICMR_0^{\rm BICM} as the "BICM capacity" or assuming that LDPC+BP decoders cannot operate at rates above R0R_0.

Correction:

The cutoff rate R0R_0 is not the capacity. It is a tighter rate threshold for a specific class of decoders — namely those whose decoding complexity grows with the number of candidate codewords examined per information bit (sequential, list, threshold decoders). For these decoders, above R0R_0 the expected complexity diverges exponentially in RR0R - R_0.

LDPC + belief propagation is NOT in this class. BP complexity is proportional to the number of edges in the factor graph, which is O(N)O(N) independent of rate (for fixed graph density). BP therefore can operate above the cutoff rate — in fact, at rates within 0.3\sim 0.3 dB of capacity in 5G NR, well above R0BICMR_0^{\rm BICM}. The invention of LDPC in the 1990s is what retired the cutoff rate as "the practical capacity"; the practical capacity is now (nearly) the Shannon capacity.

So when you see R0R_0 quoted in a paper, ask:

  1. What decoder architecture is the paper about?
  2. If the answer is "LDPC+BP", then R0R_0 is a historical curiosity — informative but not operational.
  3. If the answer is "sequential" or "list" or "polar-SCL with small list", then R0R_0 is the tight operational limit below which complexity stays polynomial.
🔧Engineering Note

5G NR LDPC Decoders Operate Above the Cutoff Rate

Every 5G NR LDPC decoder operates above the Gallager cutoff rate of the BICM channel. For QPSK at 3 dB, R0BICM1.63R_0^{\rm BICM} \approx 1.63 bits/symbol (Example ECutoff Rate of BICM-QPSK at 3 dB); the typical 5G NR QPSK operating rate is 1.7\sim 1.7 bits/symbol (code rate 0.85, QPSK spectral efficiency 2, giving 1.7). LDPC's density-evolution capacity is within 0.3\sim 0.3 dB of Shannon, which is well above R0R_0.

This is why the cutoff rate is, in the 5G design process, a sanity check, not a design constraint. The design constraint is the density-evolution threshold of the chosen LDPC base graph on the BICM channel, which determines the achievable rate under BP decoding. That threshold is computed on the mixture-channel LLR histogram, and is the quantity a 5G LDPC designer actually optimises.

The cutoff-rate sanity check is useful when diagnosing simulation behaviour: if a decoder is reported to operate at RR0R \gg R_0 with low error probability, this confirms the decoder is extracting information beyond the Gallager threshold — a hallmark of modern iterative / message-passing decoders. If instead the decoder is reported to work only at RR0R \le R_0, it is likely a sequential or list-type decoder and the cutoff rate is doing real operational work.

Practical Constraints
  • R0BICMR_0^{\rm BICM} is below practical 5G NR LDPC operating rates

  • Density-evolution threshold (not cutoff rate) is the design target for LDPC

  • For sequential decoders (polar SCL with small list), the cutoff rate is tight

  • R0R_0 is useful as a diagnostic when comparing decoder architectures

📋 Ref: 3GPP TS 38.212, §5.1
,

Common Mistake: Saddle-Point Is a Provable Upper Bound, Not an Exact PEP

Mistake:

Interpreting the Martinez-Fàbregas-Caire saddle-point bound as an exact PEP expression or as a lower bound on PEP.

Correction:

Both the CTB union bound and the MFC saddle-point bound are upper bounds on the pairwise error probability. The saddle-point tightens the Chernoff exponent but does not achieve equality; at finite blocklength the actual PEP is still below both bounds, often by 1122 dB at the SNRs of interest. To match simulation, one typically uses the MFC bound plus a small correction term (or fits directly to the simulated PEP curve).

In particular, the MFC bound is a union bound in spirit: it is tight on the codeword-pair-level event, but a codeword-error event is a union over many pairs, and the union bound can loosen further as the list of pairs grows. For low error rates the MFC bound is tight; near the waterfall region the union bound component dominates and the bound loosens by 1–3 dB.

Historical Note: Martinez-Fàbregas-Caire 2006: Sharpening the BICM PEP

2006

The 2006 IEEE Transactions on Information Theory paper of Alfonso Martinez, Albert Guillén i Fàbregas, and Giuseppe Caire — "Error probability analysis of bit-interleaved coded modulation" — is the companion analysis paper to the 2008 Foundations & Trends monograph. It derives the saddle-point PEP bound of Thm. TMartinez-Fàbregas-Caire Saddle-Point PEP Bound and shows that it tightens the classical Caire-Taricco-Biglieri 1998 union bound by 1\sim 122 dB at moderate SNR.

The paper's contribution is twofold. First, it exhibits the exact saddle-point s(dH)s^\star(d_H) as a function of Hamming distance — a numerical knob that the 1998 paper had not optimised. Second, it demonstrates that the improved bound tracks simulation far more closely than the union bound — to within a few tenths of a dB at Pe105P_e \sim 10^{-5}. This makes the MFC bound the practical PEP bound used in modern BICM analyses; the CTB union bound is retained for its simplicity (single evaluation at s=1/2s = 1/2) when rough estimates suffice.

Methodologically, the 2006 paper is the first concrete demonstration that the mismatched-decoding GMI framework (then being developed for the 2008 monograph) pays numerical dividends beyond the information-theoretic cleanup. The ss-optimisation that gives the exact achievable rate in §3 also gives the sharpest PEP bound in §5 — the framework is internally coherent.

Why This Matters: From Cutoff Rate to BICM-ID: Closing the Gap

Chapter 7's takeaway is that BICM — at one-shot demapping — has a rate gap to CM of 0.3\lesssim 0.3 dB on Gray-QAM, an exponent ratio of 0.85\sim 0.85, and a PEP improvement of 1\sim 1 dB via the MFC saddle-point. All three quantities can be further improved by iterative demapping + decoding, where the binary decoder's extrinsic LLRs are fed back to the demapper for a second pass through the per-symbol likelihood computation.

BICM-ID — the subject of Chapter 8 — closes the product-metric mismatch by making the demapper posterior conditioned on the current LLRs of the other bits. In the limit of perfect extrinsic information, BICM-ID's demapper computes the full joint symbol likelihood; i.e., the iterative system eliminates the mismatched- decoding penalty entirely. The penalty for this gain is complexity (multiple demapper passes) and the need for set-partition labelling rather than Gray (because iterative feedback needs structural separation between bits). Chapter 8 develops the EXIT-chart analysis and the code-design implications.

Forward reference: the GMI saddle-point of this chapter and the EXIT chart of Ch. 8 are two sides of the same coin. The GMI parameterises single-pass BICM rate; the EXIT chart parameterises iterative BICM-ID rate. Both are mismatched-decoding quantities; both reduce to CM capacity as the mismatch vanishes (at high SNR for GMI, at full iteration for EXIT).

Quick Check

The cutoff rate R0BICMR_0^{\mathrm{BICM}} is

the practical rate limit of sequential / list decoders, above which decoder complexity grows exponentially in RR0R - R_0

the BICM capacity minus a fixed gap of 3 dB

the rate above which the error probability is always 1

the rate at which the BICM mismatched decoder achieves the same PEP as the CM matched decoder

Quick Check

The Martinez-Fàbregas-Caire saddle-point PEP bound is tighter than the Caire-Taricco-Biglieri union bound because

it optimises the Chernoff tilt ss at each Hamming distance rather than fixing s=1/2s = 1/2 (Bhattacharyya)

it uses a different channel model (block fading instead of i.i.d.)

it accounts for BICM-ID iterative feedback

it assumes Gray labelling and a specific constellation

Cutoff Rate R0R_0

The Gallager function evaluated at ρ=1\rho = 1: R0=E0(1)=logy(xq(x)p(yx))2dyR_0 = E_0(1) = -\log \int_y (\sum_x q(x)\sqrt{p(y\mid x)})^2 dy. For binary-input channels with uniform input, R0=1log2(1+β)R_0 = 1 - \log_2(1 + \beta) where β\beta is the Bhattacharyya parameter. Operationally, the rate threshold above which sequential and list decoders have exponential expected complexity.

Related: Gallager Function E0(ρ)E_0(\rho), Bhattacharyya Parameter β\beta, Sequential Decoder

Bhattacharyya Parameter β\beta

For a binary-input channel, β=yp(y0)p(y1)\beta = \sum_y \sqrt{p(y\mid 0) p(y\mid 1)} — a scalar in [0,1][0, 1] characterising the discriminability of the two inputs. β=0\beta = 0: perfectly discriminable (zero-SNR noise limit is β=1\beta = 1, noiseless limit is β=0\beta = 0). The Bhattacharyya parameter controls the Chernoff PEP bound at s=1/2s = 1/2 and the cutoff rate via R0=1log2(1+β)R_0 = 1 - \log_2(1 + \beta).

Related: Gallager Cutoff Rate R0R_0, Chernoff Bound, Martinez-Fàbregas-Caire Saddle-Point PEP Bound

Saddle-Point PEP Bound (Martinez-Fàbregas-Caire 2006)

The PEP bound P(cc)minsnE[(pW(ybˉ)/pW(yb))s]P(\mathbf{c}\to\mathbf{c}') \le \min_s \prod_n \mathbb{E}[ (p_{W_\ell}(y\mid \bar b)/p_{W_\ell}(y\mid b))^s], where the minimum is over the decoder scaling ss and the product runs over distinct bit positions. Tightens the classical s=1/2s = 1/2 Bhattacharyya bound by 1\sim 122 dB at moderate SNR; reduces to the classical bound at s=1/2s^\star = 1/2 for long distances.

Related: Martinez-Fàbregas-Caire Saddle-Point PEP Bound, Chernoff Bound, BICM as LL Parallel Binary Channels, Union Bound