Iterative Demapping: Closing the BICM Loop
The Open Loop in One-Shot BICM
Look back at the BICM receiver of Chapter 5. A channel observation enters the demapper, which produces soft per-bit LLRs , one per label bit. The deinterleaver sends these LLRs to a SISO decoder, which produces per-bit a posteriori LLRs and decisions. At that point the receiver stops. The decoder's a posteriori output — which contains information the demapper did not have — is DISCARDED before it can help the demapper do a better job on the same observation.
This is exactly the suboptimality argument that motivated the whole Chapter 5 discussion of . The BICM demapper treats each label bit as independent, because it has no reason to believe the bits are correlated; but the binary code MAKES them correlated, and the decoder knows that correlation. If the demapper could exploit what the decoder knows, its per-bit LLRs would be sharper, and the decoder would then produce sharper outputs, and so on. That feedback loop is BICM with Iterative Decoding (BICM-ID), introduced independently by Li–Ritcey and by ten Brink–Speidel–Yan in 1997–1998. The point is simple: nothing in the demapper computation prevents us from incorporating a priori bit probabilities into the sums. Doing so gives a refined LLR; passing the refinement through the decoder gives an even better one; iterating closes the gap to CM capacity.
What is new is not any single computation — the demapper with a priori and the SISO decoder both existed before BICM-ID — but the ARCHITECTURE: the extrinsic information exchanged between the two boxes, iterated to a fixed point. The analysis of that iteration is the content of this chapter.
Definition: Demapper with A Priori Information
Demapper with A Priori Information
Let be the label of a transmitted constellation symbol , and let be the vector of a-priori LLRs supplied to the demapper (one per label bit, from the decoder in the previous iteration). The demapper with a priori computes, for each bit position , where is the Bernoulli probability induced by the a-priori LLR, and is the -th label bit of .
Two limiting cases check the definition: if all , the a-priori probabilities are uniform and the formula reduces to the one-shot BICM LLR of Ch. 5. If for all , the sums collapse to a single term and becomes the CM max-log LLR against the known symbol — the demapper has become a genie.
Definition: Extrinsic Information
Extrinsic Information
At any SISO box — demapper or decoder — the input is a vector of a-priori LLRs and a set of channel observations, and the output is a vector of a-posteriori LLRs . The extrinsic LLR at bit position is Equivalently, is the log-likelihood ratio of given ALL OTHER evidence — the channel observations and the a-priori LLRs of all bits . The extrinsic component is the "new" information the box has produced; subtracting is what prevents the iterative scheme from feeding a box its own previous output back to itself, which would create a positive-feedback loop.
The extrinsic/a-priori decomposition is one of the great conceptual inventions of iterative coding theory. It works because LLRs add under independence, so subtracting the a-priori from the a-posteriori yields the contribution of everything BUT that a-priori. In BICM-ID the decoder extrinsic LLRs become the demapper's a-priori LLRs in the next iteration, and vice versa — a strict separation that turns a feedback system into an iteration between two well-defined maps.
BICM-ID Iteration (Single Pass)
Complexity: Per iteration: demapper is with constellation points (or with max-log and smart subset indexing); SISO decoder depends on the code (BCJR: for -state convolutional; LDPC BP: per iteration). Total cost is the per-iteration cost times , typically 5–10 outer iterations in practice.The structure is a serial concatenation: demapper (inner) and decoder (outer). The interleaver randomises the bit order between them and is essential for making the extrinsic-information assumption of independence approximately valid. Without a long enough interleaver, bits correlated inside a code constraint get mapped to neighbouring constellation positions and the extrinsic information becomes contaminated by its own past — the classical iterative-decoder convergence failure.
BICM-ID Receiver Block Diagram
Theorem: Fixed-Point Rate of a Converged BICM-ID Receiver
Consider BICM-ID with a binary linear code of rate and a labelling on a memoryless channel with input SNR. Assume the Gaussian-LLR model holds (s02) and the interleaver is long. Let be a fixed point of the iteration — i.e., a point satisfying and . If , the receiver achieves arbitrarily low BER and the effective information rate through the combined channel is . If the only fixed point strictly below is attracting, the iteration stalls there and the BER is bounded away from zero.
A fixed point is where the two maps cross. The iteration walks from — no a priori, no extrinsic — up the staircase until it hits a crossing. If the crossing is , the code's bits are all perfectly known and the BER is zero. If it is anything less, the bits are only partially known and there is an error floor. All the BICM-ID design story is about arranging the two curves so that the FIRST crossing encountered on the way up is itself.
A fixed point is invariant under one round-trip of the iteration: after demapper + deinterleaver + decoder + interleaver, the MI values are unchanged.
The BER is a strictly decreasing function of the a-posteriori MI, which at a fixed point equals (well, once LLR variances are added — see s02).
At the a-posteriori LLR has infinite variance, hence the BER is zero.
Step 1: Fixed-point characterisation
Writing one round-trip of the BICM-ID iteration as the composition , the fixed points are solutions of . Equivalently, they are the crossings of the demapper curve and the INVERTED decoder curve on the -plane.
Step 2: BER at a fixed point
Under the Gaussian-LLR model, the a-posteriori LLR of a bit at the decoder output has mean and variance , with determined by the a-priori MI and channel observation MI. The bit error rate is , a strictly decreasing function of . At , and BER .
Step 3: Achievability of rate $R$
The iteration passes exactly bits per channel use from transmitter to receiver with vanishing BER; the overhead is zero because the iteration itself does not require any rate-costly signalling. The fixed point at therefore achieves rate at the operating SNR, which is the information-theoretic rate of the iterative scheme.
Why Extrinsic — Not A Posteriori
The subtraction is not a mere bookkeeping convention. It is the MECHANISM that prevents BICM-ID from becoming a runaway positive-feedback system. If the demapper received from the decoder (the full a-posteriori), it would effectively be counting the channel observation twice: once directly through , and a second time through the a-priori term, which the decoder computed FROM that same . The "new" estimate would merely re-emphasise the old one, and after a few iterations every bit would be declared confident and wrong in equal measure. Passing only the extrinsic part keeps each iteration honest: the demapper learns from code structure (via the decoder's extrinsic), not from its own prior LLR.
This is the same principle that makes turbo decoding work, and it generalises to turbo equalisation, iterative MIMO detection, and iterative source-channel decoding. Memorise it.
Common Mistake: Feeding the A-Posteriori Back (Positive Feedback)
Mistake:
In a first implementation, it is tempting to simplify the BICM-ID receiver by feeding the decoder's a-posteriori LLRs back to the demapper — after all, these are "the best estimates we have" of the coded bits. Surely using better a priori can only help?
Correction:
The a-posteriori LLR for bit contains the CHANNEL observation (through the demapper) plus the code-induced cross-information from the other bits. Feeding it back as a-priori to the demapper makes the demapper multiply by , which already contains — the channel is now counted twice. The iteration no longer tracks MI correctly: confidence inflates without justification, LLRs diverge to , and decisions become arbitrary. The correct quantity to exchange is the EXTRINSIC part , which removes the double-counted contribution. Every EXIT-chart analysis, every density-evolution proof, every convergence theorem in this chapter relies on extrinsic feedback. Posterior feedback is a bug.
Example: Demapper-with-A-Priori for QPSK with Gray Labelling
Consider QPSK with Gray labelling : the four constellation points are with labels . The channel is AWGN, , . The decoder supplies a-priori LLRs . Compute the demapper extrinsic LLR for bit 0.
Factor the bit subsets
Bit 0 is the I-axis LSB of Gray QPSK: , . Bit 1 is the Q-axis LSB, independent. Thus the a-priori factors across axes:
Separate I and Q contributions
Because and the label bits decouple across axes, the demapper-with-a-priori LLR for bit 0 reduces to Note that (the Q-axis a priori) drops out because the Q-axis subsets are identical on both sides of the bit-0 partition — Gray decoupling.
Extrinsic is the full LLR (for Gray QPSK)
The input a-priori for bit 0 was , not included in the demapper's bit-0 computation (it would be double-counting), so . Equivalently, for Gray QPSK the demapper extrinsic is just the channel LLR, independent of a priori: BICM-ID on Gray QPSK is pointless because the two label bits are already independent bit channels. This is exactly the sense in which Gray is "one-shot optimal" and BICM-ID offers nothing extra — a pattern that generalises (s04) to all labelings that fully decouple bit channels.
Example: Demapper-with-A-Priori for 16-QAM with Set-Partition Labelling
Consider 16-QAM with the set-partition labelling : the MSB bit splits the constellation into two 8-point QAM lattices of large minimum distance, and successively finer bits resolve inside each subset. Suppose the decoder supplies an a-priori LLR for the MSB (perfect knowledge that bit 0 = 0). Describe what the demapper LLRs for bits 1, 2, 3 become, and contrast with the case .
Zero a priori: one-shot BICM demapper
With , the demapper sums over all 16 points when computing for any bit . For the LSB of the I-axis under SP, pairs of neighbours at minimum distance alternate between bit-0 = 0 and bit-0 = 1, so the two subsets and overlap heavily — the demapper cannot tell "bit 1 = 0" from "bit 1 = 1" with high confidence. This is exactly why SP labelling has lower BICM capacity than Gray (Ch. 5).
Perfect a priori on bit 0: constellation shrinks to 8 points
With , the demapper sums ONLY over the 8 points with — the 8-QAM sub-constellation of larger minimum distance. For a fixed label bit , the two subsets and are now separated by the SP sub-partition's minimum distance, which is a factor – larger than the original. The demapper LLRs become much sharper.
The BICM-ID gain mechanism for SP
This is the quantitative version of the "SP wins under iteration" story. SP concentrates bit-discrimination information in the MSB; learning the MSB via iteration unlocks larger minimum distances for the remaining bits. Gray distributes discrimination evenly, so learning one bit does not unlock much for the others — and BICM-ID gains little. Section 4 quantifies this via EXIT curve steepness.
Quick Check
In BICM-ID, what is exchanged between the demapper and the decoder on each iteration?
The channel LLRs (demapper computes them once and decoder reuses them every iteration).
The full a-posteriori LLRs from each box.
The extrinsic LLRs: a-posteriori minus a-priori at each box's output.
The hard decisions after each decoder pass.
Correct. The extrinsic decomposition enforces that each box's output reflects only what that box has LEARNED, not what it was told. This is the fundamental mechanism of iterative decoding.
Key Takeaway
BICM-ID is a two-box serial iteration with extrinsic-only exchange. The demapper with a priori sharpens per-bit LLRs using the code's cross-bit correlations (via decoder extrinsic), and the SISO decoder sharpens a-posteriori LLRs using the refined per-bit LLRs (via demapper extrinsic). Subtracting the a-priori at each output is what keeps the iteration from double-counting. Gray labelling gets nothing from the loop — it fully decouples bits already. Set-partition labelling gets a lot — it concentrates bit information and the loop unlocks the concentration.
Why This Matters: BICM-ID in DVB-S2X Very-Low-SNR Modes
The DVB-S2X standard extends DVB-S2 with a set of very-low-SNR MODCODs (down to QPSK at rate , an operating near dB) for sub-1 m satellite terminals in tropical-scintillation environments. At these SNRs, the one-shot BICM receiver leaves several tenths of a dB on the table relative to CM capacity — enough that the standard includes an optional iterative demapping mode. The receiver performs 3–5 passes between the APSK demapper and the LDPC decoder before declaring a codeword. In practice this closes about 0.3 dB of the gap on the VL-SNR modes and pushes the link budget just over the rain-fade edge that the system was designed for.
Extrinsic Information
At any soft-in/soft-out decoding box, the extrinsic LLR for a bit is the a-posteriori LLR minus the a-priori LLR for that same bit. It represents the box's learning — the information contributed by the box from sources other than its own input for that bit. In BICM-ID, extrinsic LLRs are the only quantity exchanged between demapper and decoder, to prevent positive feedback.
Related: A Posteriori Llr, Siso Decoding, Turbo Principle
BICM-ID (Bit-Interleaved Coded Modulation with Iterative Decoding)
A BICM receiver architecture that iterates between a demapper-with- a-priori and a soft-in/soft-out decoder, exchanging extrinsic information through a bit interleaver. Introduced by Li and Ritcey (1997) and ten Brink–Speidel–Yan (1998), BICM-ID can close most of the gap between BICM capacity and CM capacity when combined with a set-partition (or similar) labelling.
Related: The Open Loop in One-Shot BICM, Iterative Decoding, Exit Chart