BICM Pairwise Error Probability
From Capacity to Error Probability
Chapter 5 told us what rate a BICM transmitter can support — the BICM capacity . The punchline there was that with a Gray labelling, sits within a few tenths of a bit of the CM capacity over essentially every SNR of practical interest. Capacity, however, is only half of the design story. What the hardware engineer actually has to hit is a target bit-error rate — for broadband data, for voice, perhaps for optical backhaul — at some specified SNR. Two BICM systems with the same capacity can reach those targets at wildly different SNRs, depending on how the labelling, the binary code, and the interleaver interact.
We now turn to the error-probability side of the Caire–Taricco–Biglieri analysis. The backbone of this analysis is the pairwise error probability (PEP): the probability that a maximum-likelihood decoder prefers a wrong codeword over the transmitted one when those are the only two candidates. From the PEP we assemble union bounds on the codeword and bit error probabilities. The technique — Chernoff bounding applied to the log-likelihood metric — is the same one we used for TCM in Ch. 2. The novelty is that for BICM, the metric is a BIT metric, not a symbol metric, and that is what makes the labelling play a role through the average squared distance rather than the minimum distance.
Definition: BICM Codeword Pair and Hamming Distance
BICM Codeword Pair and Hamming Distance
Let and be two distinct BICM codewords (at the output of the binary encoder, before mapping). Their Hamming distance is the number of bit positions at which they disagree. For a binary linear code, the distribution of over all nonzero codewords is captured by the weight enumerator , where is the number of codewords of Hamming weight . The smallest with is the minimum Hamming distance (or simply when the context is clear).
The interleaver permutes before mapping, so the two codewords at the encoder output correspond to two constellation-symbol sequences at the mapper output. For the error-probability analysis we model the interleaver as placing the differing bit positions at DIFFERENT symbol times — the fully-interleaved assumption of s03.
Definition: Pairwise Error Probability
Pairwise Error Probability
Given a channel and a maximum-likelihood (ML) decoder, the pairwise error probability (PEP) between two codewords and is The codeword error probability is upper-bounded by the union over all competing codewords, and the bit error probability is upper-bounded by the same sum, weighted by the number of input-information-bit errors each PEP event causes.
The PEP is a two-codeword metric. At high SNR it dominates the error probability because only the nearest competitor matters; at low SNR the union bound over-counts and becomes loose (s04).
Definition: BICM Bit Metric and Its Likelihood Ratio
BICM Bit Metric and Its Likelihood Ratio
For a BICM receiver, the ML decoder operates on the bit metric where is the subset of constellation points whose -th label bit equals . The log-likelihood ratio for bit position is . A binary ML decoder on the interleaved coded stream then computes for each candidate codeword and picks the maximiser.
In the high-SNR regime the "max-log" approximation is common; it reduces the bit metric to a nearest-subset computation. All PEP bounds in this section hold for the exact bit metric; the max-log version introduces a small additional loss that is characterised in [?fabregas-martinez-caire-2008].
Definition: Average Squared Intra-Subset Euclidean Distance
Average Squared Intra-Subset Euclidean Distance
For labelling and bit position , define the average squared intra-subset Euclidean distance That is, is the mean over all pairs of constellation points that differ in bit position of their squared Euclidean distance. The averaged-over-positions quantity is This quantity replaces the minimum squared distance of uncoded modulation analysis in the BICM PEP exponent.
Notice that is an AVERAGE, not a minimum. This is the fundamental operational difference between BICM PEP analysis and uncoded symbol-error analysis, and it is exactly what makes Gray labelling attractive on AWGN: Gray spreads pairs of "bit-different" symbols uniformly over all Euclidean distances, so the average is dominated by close neighbours and does not explode.
Theorem: BICM Pairwise Error Probability on AWGN
Consider a BICM system with labelling , constellation of size with unit average energy per symbol, on an AWGN channel with . For two BICM codewords at Hamming distance , the PEP under ML decoding of the bit metric is bounded by Equivalently, the error exponent per unit grows linearly in with slope .
The point is that the binary code contributes its Hamming distance and the labelling/constellation contributes its average squared distance . These multiply in the exponent — exactly the same structure as the TCM free-distance bound in Ch. 2, where the trellis contributed its number of error events and the constellation contributed . The difference is only that for BICM the "per-event distance" averages over many label pairs rather than taking the minimum.
Start from the union bound for ML on bit metrics: .
Use the Bhattacharyya bound: , where is the Bhattacharyya factor of the -th bit channel.
Under the fully-interleaved assumption the bit-position indices are i.i.d. uniform over , so the product becomes where .
Upper-bound by using the max over label-pair distances inside the subset sum.
Step 1: Express ML decision via bit-metric score
Under the interleaver , the -th coded bit is mapped to bit position of symbol transmitted at time . The ML decision on vs reduces to the sign of since bits where contribute zero. An error occurs when this sum is positive given was sent.
Step 2: Chernoff bound on the sum
By the Chernoff inequality with exponent (the Bhattacharyya-optimal choice for equiprobable hypotheses), where we used independence of noise samples and defined Here is the Bhattacharyya factor of the -th BICM bit channel; it is a number in that captures the "distinguishability" of the two bit values given .
Step 3: Averaging over interleaver positions
Because the interleaver is uniformly random, each is independent and uniform over . Taking expectation over the interleaver, The PEP depends on the binary code only through , and on the labelling/constellation/noise only through the scalar .
Step 4: Bounding $\bar\beta$ by the average distance
For each bit channel on AWGN, A standard Jensen-type bound (used throughout coded-modulation PEP analysis — see [?proakis-2008]) yields Averaging over and using convexity gives after normalisation by .
Step 5: Assemble and tighten via Q-function
Substituting back into yields the exponential bound stated in the theorem. The tighter Q-function form follows from the observation that the dominant pair in the bit-metric error event is Gaussian-noise-dominated with effective distance , which then goes through the standard Gaussian tail .
Key Takeaway
The BICM AWGN PEP factorises into code and constellation contributions. The Hamming distance of the binary code and the average squared distance of the labelled constellation enter multiplicatively in the exponent. Pick a code with large and a labelling with large — AND these can be designed independently, which is the whole point of the BICM architecture.
BICM Pairwise Error Probability on AWGN
Exact -form PEP from Thm. 1 as a function of (in dB), for two Gray-labelled QAM sizes and Hamming distances in the range . The slope at high SNR is the same for all (one-sided AWGN exponent); the shift is dB. Use the plot to read off the SNR required for PEP — this is your design SNR for the nearest competitor.
Parameters
Pairwise Error: Euclidean Distance Depends on Labelling
Example: Computing for Gray-Labelled 16-QAM
Compute for each bit position of a unit-average-energy 16-QAM constellation with Gray labelling, and the overall . Compare with .
Setup: 16-QAM with unit average energy
The 16-QAM constellation has points at scaled by so that the average energy is one. The minimum distance is . Gray labelling assigns the four label bits so that each Gray-adjacent pair of constellation points differs in exactly one label bit.
Bit 0 (in-phase LSB)
Bit 0 partitions the constellation into the left half () and right half (). For each on the left, contains four mirror points on the right; the squared distances are — or, computing properly, (A direct enumeration over the pairs gives the same number.)
Bits 1, 2, 3 (in-phase MSB, quadrature LSB, quadrature MSB)
By symmetry of the Gray labelling, bits 2 and 3 play the same role for the quadrature axis that bits 0 and 1 play for the in-phase axis, so The MSB splits the constellation into outer and inner rings; the mean separation is larger there.
Averaging over bit positions
d_{\min}^{2} = 0.4d_H \cdot 1.0/(4 N_0)2.5\timesd_{\min}^{2}$ in the exponent. This numerical separation is exactly why BICM does not suffer as much on AWGN as a naive "treat each bit as an independent channel" argument would predict.
Common Mistake: Is an AVERAGE, Not the Minimum
Mistake:
A common mistake when reading the BICM PEP bound for the first time is to conclude, by analogy with uncoded QAM, that the exponent is governed by and that labelling only matters insofar as it affects of the underlying constellation.
Correction:
The BICM exponent involves , not . Two labellings on the same constellation (same ) can produce dramatically different . For 16-QAM, the natural binary labelling has smaller than Gray's, because bit positions with close neighbours get more weight. On AWGN, this difference can be dB. On fading (s03), a completely different story unfolds — there, MIN distance over label pairs matters via .
Pattern: This Is the Same Argument as TCM PEP
The proof of Thm. 1 — Chernoff bound on a sum of log-likelihoods, followed by Bhattacharyya averaging — is the same pattern we used in Ch. 2 to bound the TCM pairwise error along a trellis error event. The only change is the decoding metric: in TCM we use the symbol metric ; in BICM we use the bit metric . Because the bit metric averages over a subset of symbols rather than pointing to a single one, the Bhattacharyya factor becomes a subset-wise average, and the constellation contribution to the exponent becomes rather than .
This is a useful principle to internalise: every coded-modulation PEP bound has the same skeleton — Chernoff, Bhattacharyya, sum-over-error- event-positions. What changes between schemes is only which distance metric lives inside the Bhattacharyya factor.
Quick Check
The BICM AWGN PEP bound of Thm. 1 scales, for large , as with exponent equal to:
From Thm. 1: the exponent is , factorising into code () and constellation/labelling () contributions.
Pairwise Error Probability (PEP)
The probability that a maximum-likelihood decoder prefers a specific wrong codeword over the transmitted one , given only those two are candidates. On AWGN the PEP is a Q-function of the Euclidean distance between the codewords; for BICM, the Euclidean distance is replaced by the bit-metric Bhattacharyya factor.
Related: Bhattacharyya Factor, Chernoff Bound, BICM Codeword Pair and Hamming Distance
Bhattacharyya Factor
For a binary-input channel with transition densities and , the Bhattacharyya factor is . It is the Chernoff-optimal exponent for hypothesis testing between equiprobable binary inputs. For a BICM bit channel, depends on the labelling and the noise level.
Related: Chernoff Bound, Pattern: This Is the Same Argument as TCM PEP, Bit Metric