Pairwise Error Probability and the Coding-Gain Criterion

From Distance to Design Criterion

We have seen that the gap to capacity in the bandwidth-limited regime is mostly a coding deficit. The question is: how do we design the constellation points XβŠ‚RN\mathcal{X} \subset \mathbb{R}^N to recover it?

The answer, at high SNR, comes from the pairwise error probability: the probability that the receiver chooses a wrong codeword x^\hat{\mathbf{x}} when the true one is x\mathbf{x}. This probability depends only on the Euclidean distance βˆ₯xβˆ’x^βˆ₯\|\mathbf{x} - \hat{\mathbf{x}}\| and the noise variance β€” not on the structure of the constellation. Summing over all wrong codewords via the union bound then translates a geometric property of the point set into an asymptotic error exponent. The design criterion that falls out is the starting point for Ungerboeck's TCM in Chapter 2 and for the space-time coding criteria in Chapter 10.

,

Definition:

Pairwise Error Probability on AWGN

Let x,x^∈RN\mathbf{x}, \hat{\mathbf{x}} \in \mathbb{R}^N be two candidate codewords and let y=x+w\mathbf{y} = \mathbf{x} + \mathbf{w} with w∼N(0,(N0/2)IN)\mathbf{w} \sim \mathcal{N}(\mathbf{0}, (N_0/2) \mathbf{I}_N). The pairwise error probability (PEP) is the probability that the maximum-likelihood receiver, given the choice between x\mathbf{x} and x^\hat{\mathbf{x}}, prefers x^\hat{\mathbf{x}}:

P(xβ†’x^)β€…β€Š=β€…β€ŠPr⁑{βˆ₯yβˆ’x^βˆ₯2≀βˆ₯yβˆ’xβˆ₯2}β€…β€Š=β€…β€ŠQ ⁣(βˆ₯Ξ”βˆ₯2Οƒ),P(\mathbf{x} \to \hat{\mathbf{x}}) \;=\; \Pr\{\|\mathbf{y} - \hat{\mathbf{x}}\|^2 \le \|\mathbf{y} - \mathbf{x}\|^2\} \;=\; Q\!\left(\frac{\|\boldsymbol{\Delta}\|}{2 \sigma}\right),

where Ξ”=xβˆ’x^\boldsymbol{\Delta} = \mathbf{x} - \hat{\mathbf{x}} is the error vector and Οƒ=N0/2\sigma = \sqrt{N_0/2}. Equivalently,

P(xβ†’x^)β€…β€Š=β€…β€ŠQ ⁣(βˆ₯Ξ”βˆ₯22N0).P(\mathbf{x} \to \hat{\mathbf{x}}) \;=\; Q\!\left(\sqrt{\frac{\|\boldsymbol{\Delta}\|^2}{2N_0}}\right).

The PEP depends on x\mathbf{x} and x^\hat{\mathbf{x}} only through βˆ₯Ξ”βˆ₯\|\boldsymbol{\Delta}\|. This is a distinctive feature of the AWGN channel: the direction of the error vector is irrelevant, and only its length matters. Over fading channels this is no longer true, and the PEP will acquire additional structure (see Chapter 10).

,

Theorem: Pairwise Error Probability on the AWGN Channel

For x,x^∈RN\mathbf{x}, \hat{\mathbf{x}} \in \mathbb{R}^N transmitted over the AWGN channel with noise variance Οƒ2=N0/2\sigma^2 = N_0/2 per real dimension, the ML pairwise error probability is

P(xβ†’x^)β€…β€Š=β€…β€ŠQ ⁣ ⁣(βˆ₯Ξ”βˆ₯2Οƒ)β€…β€Šβ‰€β€…β€Š12exp⁑ ⁣(βˆ’βˆ₯Ξ”βˆ₯28Οƒ2)β€…β€Š=β€…β€Š12exp⁑ ⁣(βˆ’βˆ₯Ξ”βˆ₯24N0),P(\mathbf{x} \to \hat{\mathbf{x}}) \;=\; Q\!\!\left(\frac{\|\boldsymbol{\Delta}\|}{2\sigma}\right) \;\le\; \tfrac{1}{2} \exp\!\left(-\frac{\|\boldsymbol{\Delta}\|^2}{8\sigma^2}\right) \;=\; \tfrac{1}{2} \exp\!\left(-\frac{\|\boldsymbol{\Delta}\|^2}{4 N_0}\right),

where the upper bound uses the Chernoff inequality Q(x)≀12eβˆ’x2/2Q(x) \le \tfrac{1}{2} e^{-x^2/2}.

Project the noise onto the direction Ξ”\boldsymbol{\Delta}: the component along Ξ”\boldsymbol{\Delta} is zero-mean Gaussian with variance Οƒ2\sigma^2. The ML detector picks x^\hat{\mathbf{x}} when this component exceeds βˆ₯Ξ”βˆ₯/2\|\boldsymbol{\Delta}\|/2 β€” a one-dimensional Gaussian tail. This is why only βˆ₯Ξ”βˆ₯\|\boldsymbol{\Delta}\| enters the PEP: the (Nβˆ’1)(N-1) orthogonal noise components cannot cause the pair error.

,

Pairwise Geometry: Decision Boundary and Noise Circle

Two constellation points x,x^\mathbf{x}, \hat{\mathbf{x}} at Euclidean distance dd in R2\mathbb{R}^2, with a noise cloud centered on the transmitted point. The decision boundary is the perpendicular bisector; a pair error occurs whenever the noise pushes y\mathbf{y} across it. As the distance dd grows, the boundary moves away and the tail probability Q(d/2Οƒ)Q(d / 2 \sigma) shrinks exponentially.

Parameters
2
8

Definition:

Minimum Distance and Asymptotic Coding Gain

Let XβŠ‚RN\mathcal{X} \subset \mathbb{R}^N be a finite constellation of MM equiprobable signal vectors of average energy EsE_s per 2D dimension. Its minimum Euclidean distance is

dE,min⁑2(X)β€…β€Š=β€…β€Šmin⁑x,x^∈Xxβ‰ x^βˆ₯xβˆ’x^βˆ₯2.d_{\rm E, \min}^2(\mathcal{X}) \;=\; \min_{\substack{\mathbf{x}, \hat{\mathbf{x}} \in \mathcal{X} \\ \mathbf{x} \ne \hat{\mathbf{x}}}} \|\mathbf{x} - \hat{\mathbf{x}}\|^2.

The normalized minimum distance is dE,min⁑2/Esd_{\rm E, \min}^2 / E_s, which is invariant under scaling of the constellation. The asymptotic coding gain of X\mathcal{X} over a baseline constellation Xref\mathcal{X}_{\rm ref} of the same spectral efficiency η\eta is

Ξ³casy(X,Xref)β€…β€Š=β€…β€Š10log⁑10dE,min⁑2(X)/Es(X)dE,min⁑2(Xref)/Es(Xref)[dB].\gamma_c^{\rm asy}(\mathcal{X}, \mathcal{X}_{\rm ref}) \;=\; 10 \log_{10} \frac{d_{\rm E, \min}^2(\mathcal{X}) / E_s(\mathcal{X})} {d_{\rm E, \min}^2(\mathcal{X}_{\rm ref}) / E_s(\mathcal{X}_{\rm ref})} \quad [\text{dB}].

The asymptotic coding gain depends on the constellation only through the ratio dE,min⁑2/Esd_{\rm E,\min}^2 / E_s. Rescaling the constellation does not change Ξ³casy\gamma_c^{\rm asy} β€” this is a sensible invariance: if you double every point, you double the distances and quadruple the energy, leaving the ratio alone.

,

Theorem: Union Bound on Error Probability

For a constellation X\mathcal{X} transmitted over AWGN with ML decoding, the codeword-error probability satisfies the union bound

Peβ€…β€Šβ‰€β€…β€Š1Mβˆ‘x∈Xβˆ‘x^β‰ xQ ⁣ ⁣(βˆ₯xβˆ’x^βˆ₯2Οƒ).P_e \;\le\; \frac{1}{M} \sum_{\mathbf{x} \in \mathcal{X}} \sum_{\hat{\mathbf{x}} \ne \mathbf{x}} Q\!\!\left(\frac{\|\mathbf{x} - \hat{\mathbf{x}}\|}{2\sigma}\right).

At high SNR the sum is dominated by pairs at minimum distance. Let Kmin⁑K_{\min} be the average multiplicity of nearest neighbors (per codeword). Then

Peβ€…β€Šβ‰ˆβ€…β€ŠKmin⁑ Q ⁣ ⁣(dE,min⁑2Οƒ)β€…β€Š=β€…β€ŠKmin⁑ Q ⁣ ⁣(dE,min⁑22N0),P_e \;\approx\; K_{\min}\, Q\!\!\left(\frac{d_{\rm E, \min}}{2\sigma}\right) \;=\; K_{\min}\, Q\!\!\left(\sqrt{\frac{d_{\rm E, \min}^2}{2 N_0}}\right),

which makes dE,min⁑2d_{\rm E, \min}^2 the dominant design parameter at high SNR.

Every wrong codeword contributes a term in the union bound, but at high SNR the smallest distances dominate because the QQ function is exponentially decreasing. The minimum-distance term has the largest multiplier Kmin⁑K_{\min} per codeword, and everything else is exponentially smaller. At low SNR the full distance spectrum matters, and the union bound can be quite loose β€” a recurring theme in code performance analysis.

,

Example: BPSK and QPSK Have the Same Asymptotic Coding Gain

Compute dE,min⁑2/Esd_{\rm E,\min}^2 / E_s for BPSK (points ±1\pm 1) and QPSK (points (±1,±1)/2(\pm 1, \pm 1)/\sqrt{2}). Interpret the result.

Hamming Distance β‰  Euclidean Distance

Binary coding is dominated by the Hamming distance dHd_H of the code: the minimum number of coordinate positions in which two codewords differ. Coded modulation is dominated by the Euclidean distance dE,min⁑d_{\rm E, \min} of the transmitted constellation: the minimum β„“2\ell_2 distance between two transmitted signal vectors.

These are not the same thing. Two binary codewords differing in many positions may map to two QAM sequences that happen to be close in signal space, if the QAM labeling is poorly chosen. The Ungerboeck insight (Chapter 2) is precisely that the code must be designed to maximize dE,min⁑d_{\rm E,\min} of the transmitted signal, not dHd_H of the binary label β€” and that this requires a coordinated choice of code and labeling.

,

Common Mistake: Union bound is loose at low SNR

Mistake:

Assuming that the union-bound expression Peβ‰ˆKmin⁑Q(dE,min⁑/(2Οƒ))P_e \approx K_{\min} Q(d_{\rm E,\min}/(2\sigma)) is accurate at all SNRs.

Correction:

The union bound is asymptotically tight as SNR β†’βˆž\to \infty. At low SNR (where the error probability is large), the bound can exceed 1, and the full distance spectrum contributes meaningfully. Improved bounds (Gallager, Poltyrev, Divsalar) exist but are more intricate; in practice the first-order union-bound expression guides design but must be validated by simulation at the operating SNR of interest.

πŸ”§Engineering Note

The Role of dE,min⁑d_{\rm E, \min} in Modern Standards

Design tables for DVB-S2X, 5G NR LDPC codes, and Wi-Fi 6/7 explicitly report the minimum Euclidean distance between coded QAM sequences, not just the minimum Hamming distance of the binary LDPC code. For high-rate QAM constellations, the effective Euclidean distance multiplier of the code-plus-labeling combination is the quantity that sets the operating SNR. Gray labeling achieves close to the best effective dE,min⁑d_{\rm E,\min} among binary labelings; set-partitioning labeling (Ungerboeck) does better when coupled with a matched code.

Practical Constraints
  • β€’

    5G NR uses QPSK / 16-QAM / 64-QAM / 256-QAM with Gray labels in the base tables; 1024-QAM was added in Rel-17 for IAB

  • β€’

    DVB-S2X uses APSK with specially-designed labels that are approximately Gray on the amplitude bits and rotationally structured on the phase bits

  • β€’

    LDPC designers tune the degree distribution to match a given QAM order so as to maximize effective dE,min⁑d_{\rm E, \min} at the target error rate

Quick Check

If you double the minimum Euclidean distance of a constellation at fixed average energy, by roughly how many dB does the high-SNR error probability improve, asymptotically?

3 dB

6 dB

10 dB

It depends on the multiplicity Kmin⁑K_{\min}.

Key Takeaway

At high SNR, the minimum Euclidean distance is the design parameter. The pairwise error probability depends only on βˆ₯Ξ”βˆ₯\|\boldsymbol{\Delta}\|, and the union bound shows that Peβ‰ˆKmin⁑Q(dE,min⁑/(2Οƒ))P_e \approx K_{\min} Q(d_{\rm E,\min}/(2\sigma)). The coding-gain design criterion on AWGN is therefore: maximize dE,min⁑2/Esd_{\rm E,\min}^2 / E_s at the target spectral efficiency. This is the criterion that Ungerboeck's TCM optimizes; it will generalize to a determinant-and-rank criterion on fading MIMO channels (Chapter 10).

Pairwise error probability (PEP)

The probability that the ML detector, given a binary choice between two candidate codewords, picks the wrong one under AWGN: P(xβ†’x^)=Q(βˆ₯Ξ”βˆ₯/(2Οƒ))P(\mathbf{x} \to \hat{\mathbf{x}}) = Q(\|\boldsymbol{\Delta}\|/(2\sigma)), where Ξ”\boldsymbol{\Delta} is the error vector.

Related: Union Bound, Minimum Euclidean distance, Ml Detection

Minimum Euclidean distance

The smallest β„“2\ell_2 distance between any two distinct signal vectors in a finite constellation. Denoted dE,min⁑d_{\rm E,\min}; the normalized quantity dE,min⁑2/Esd_{\rm E,\min}^2 / E_s determines asymptotic coding gain.

Related: Pairwise Error Probability on AWGN, Coding Gain

Historical Note: Massey's 1974 ZΓΌrich Paper β€” The Seeds of Coded Modulation

1974

At the 1974 ZΓΌrich Seminar on Digital Communications, James Massey gave a talk that proved seminal: he argued that the conventional separation of coding (over a binary channel) from modulation (over a Gaussian channel) was responsible for a large part of the gap to capacity in bandwidth-limited systems, and that a joint design of code and modulator could close much of that gap. The argument rested on the observation that the Hamming-distance criterion of classical coding theory was the wrong criterion when the downstream channel was AWGN with a higher-order constellation. This insight took eight years to become an engineering reality β€” Ungerboeck's TCM paper in 1982 β€” but Massey's talk is the moment the coded-modulation program was first announced.

Why This Matters: On AWGN We Optimize dmin⁑2d_{\min}^2; On Fading We Will Optimize Rank and Determinant

On AWGN, the PEP is Q(dE,min⁑/(2Οƒ))Q(d_{\rm E,\min} / (2\sigma)): one number, the minimum distance. On a Rayleigh fading MIMO channel, the PEP depends on the full singular-value spectrum of the error matrix Ξ”\boldsymbol{\Delta}, and the analog of dE,min⁑d_{\rm E,\min} is a pair of numbers: the rank of Ξ”\boldsymbol{\Delta} (diversity order) and the product of its nonzero eigenvalues (coding gain). Chapter 10 will rederive the fading PEP from scratch; the structural parallel with what we have just done is worth keeping in mind.