Mapping by Set Partitioning and Distance Maximization
From Partition to Mapping: Turning Geometry into Bits
In Β§Set Partitioning of Signal Constellations we saw that an Ungerboeck partition of a -point constellation produces levels of cosets, each level at least doubling the intra-subset MSED. That is a property of the geometry. To build a coded-modulation scheme we need a mapping from binary sequences to constellation points β a rule that tells the encoder which point to transmit given the output of the convolutional code.
The punchline is this: if we label each constellation point with a binary address such that the first bits identify the level- coset, then a convolutional code acting on the low-order bits (the coset-selecting bits) automatically enjoys a large intra-sequence Euclidean distance, because flipping a low-order bit moves us between points of the same coset at a high partition level. The high-order bits (uncoded) pick a specific point inside the coset.
This is the entire recipe: partition β label β separate coded from uncoded bits β push the code to the bits that select the smallest coset. The rest of the chapter spells out the consequences.
Definition: Trellis-Coded Modulation (TCM)
Trellis-Coded Modulation (TCM)
A trellis-coded modulation (TCM) scheme transmits information bits per channel use from a constellation of size (one redundancy bit per symbol) according to the following encoder:
- Of the information bits per symbol, are fed into a rate- convolutional encoder whose -bit output selects a level- coset of an Ungerboeck partition.
- The remaining uncoded information bits select one of the points inside the chosen coset.
The free Euclidean distance of the TCM scheme is the minimum Euclidean distance between any two distinct code sequences it can produce.
The spectral efficiency is bits/symbol β same as an uncoded -point constellation. The constellation has been doubled to , but no additional bandwidth is used. This is the "coding gain without bandwidth expansion" feature Ungerboeck emphasized.
Definition: Parallel Transitions
Parallel Transitions
In the TCM trellis, two edges between the same pair of states that differ only in the uncoded bits are called parallel transitions. They correspond to distinct constellation points lying in the same level- coset.
The contribution of parallel transitions to the free Euclidean distance is at least , because two parallel-transition sequences differ in exactly one symbol and the two competing symbols lie in the same coset at level .
Parallel transitions exist only when , i.e., when at least one information bit is left uncoded. They can dominate the free distance when the convolutional code is strong enough that non-parallel error events are farther away β more on this in s03.
Definition: Subset Decoding (Within-Coset Nearest-Point Rule)
Subset Decoding (Within-Coset Nearest-Point Rule)
Given a received sample and a coset , the subset decoder outputs
In TCM, subset decoding is performed as a preprocessing step for every coset at every time step: the Viterbi algorithm then operates on a trellis whose branch metric at each transition is the squared Euclidean distance to the subset-decoded point of the corresponding coset.
Subset decoding decouples the uncoded-bit decision from the trellis search, reducing the effective branch-metric computation to "find nearest point in coset" ( candidates) followed by Viterbi over the much smaller coset trellis.
Definition: Free Euclidean Distance
Free Euclidean Distance
The free (squared) Euclidean distance of a TCM scheme is the minimum squared Euclidean distance between any two distinct infinite code sequences and produced by the encoder:
Since the TCM trellis is linear in the state transitions (not in the symbol values), can be computed by examining all finite-length error events β paths that diverge from a common state at time and first remerge at time .
The free distance plays the role the minimum Hamming distance plays in classical convolutional codes β it governs the asymptotic error performance on AWGN via the union bound. The bit-error probability at high SNR is approximately , where is the number of error events achieving (the "error coefficient").
Parallel Transitions
Two or more trellis edges connecting the same pair of states. In TCM they are labeled with distinct points from the same level- coset, and the distance between parallel edges equals the intra-subset MSED at that level. Parallel transitions bound the free distance from above: .
Related: Parallel Transitions, Free Euclidean Distance
Subset Decoding
The preprocessing step in TCM where, for each coset and each received sample, the nearest point within that coset is found and its squared distance to the received sample is used as the branch metric in the Viterbi decoder. Reduces decoding complexity from to branches per time step, times the in-coset search cost.
Related: Subset Decoding (Within-Coset Nearest-Point Rule), Parallel Transitions
Free (Euclidean) Distance
The minimum squared Euclidean distance between any two distinct code sequences in a TCM scheme. Analogue of free Hamming distance for binary convolutional codes. On AWGN, determines the asymptotic coding gain via .
Related: Free Euclidean Distance
Theorem: Lower Bound on Free Euclidean Distance via Partition Levels
Consider a TCM scheme built from a rate- convolutional code driving an Ungerboeck partition of a -point constellation. Let denote the intra-subset MSED at level of the partition and let be the minimum Hamming distance of the convolutional code when its output labels are read as binary addresses of the level- cosets. Then
The first term bounds error events that traverse parallel transitions (single-symbol errors within a coset); the second bounds error events that stay on distinct coset sequences for at least time steps.
Two error mechanisms compete. Parallel transitions give an error at minimum-distance . Coset-sequence errors accumulate Euclidean distance along each time step where the two coset sequences differ β each such time step contributes at least (the distance between the two level-1 cosets) and the two sequences differ in at least time steps by the Hamming-distance property of the convolutional code. The free Euclidean distance is then the minimum of the two mechanisms. A good Ungerboeck design balances them: pick and the convolutional code so that .
Split the analysis into two cases: error events of length 1 (single-symbol errors) and error events of length .
For length-1 events, the two competing symbols must lie in the same coset, so their distance is .
For longer events, at every time step at which the two coset sequences differ, the Euclidean displacement is at least .
Case 1 β single-symbol error events
Suppose two code sequences and differ in exactly one symbol, say at time , and are identical elsewhere. Identical state transitions at time mean the coset index is the same for both sequences at time : the only way they can differ at is via the uncoded bits, i.e., they take parallel transitions. The two transmitted points lie in the same level- coset, so
Case 2 β multi-symbol error events
Now suppose and differ at two or more time steps, in a contiguous block (the error event). The coset sequences and must therefore differ at at least one time step inside β otherwise the state trajectories would coincide and so would the (otherwise identical uncoded-bit decisions). Because the convolutional code has minimum Hamming distance , the binary-label sequences of and differ in at least time steps (bit-by-bit equivalent of symbol index difference). At every such time step, and lie in different level-1 cosets (they differ in the highest-order coset bit), so .
Case 2 β accumulated distance
Summing over the at-least- differing time steps, The non-differing coset time steps contribute (they may still differ in uncoded bits inside a coset, but never less than zero).
Combine the two cases
The free Euclidean distance is the minimum over all non-zero error events, which is at most the minimum of the two case bounds: When the bound is tight, which it is for all of Ungerboeck's tabulated TCM codes, the design problem reduces to balancing the two terms by the choice of and the convolutional code.
BER of Ungerboeck 8-PSK TCM vs. Uncoded QPSK
Bit-error rate of Ungerboeck's 8-PSK TCM family at various numbers of trellis states () compared to uncoded QPSK (same 2 bit/symbol spectral efficiency) on AWGN. Select the number of states to highlight. Observe that adding states (longer memory) improves coding gain, but with diminishing returns: 4-state gives ~3 dB, 8-state ~3.6 dB, 16-state ~4.1 dB, 64-state ~4.6 dB. The gap to the ultimate coding-gain ceiling ( dB, set by the partition geometry) is never closed by trellis memory alone.
Parameters
Example: Free Euclidean Distance of the 4-State 8-PSK TCM
Consider Ungerboeck's canonical 4-state 8-PSK TCM: information bits per symbol, coded bit, rate- convolutional code with generator polynomials (in systematic feedback form β the encoder has 2 memory cells, so and ). The code output selects a level-2 coset of the 8-PSK partition (4 cosets of 2 points each, with ). The 1 uncoded bit selects a point inside the coset, with distance to its coset partner.
Compute a lower bound on using TLower Bound on Free Euclidean Distance via Partition Levels, then state the actual value from Ungerboeck's Table I (1982).
Read off the partition distances
For unit-energy 8-PSK, we computed in TMSED Doubling Along the 8-PSK Ungerboeck Partition: The coset-selection bit chooses between a level-1 coset (since , a rate- code produces a 2-bit output that indexes one of 4 level-2 cosets β but the first output bit indexes the level-1 coset, which is what the level-1 distance refers to). The parallel-transition distance is .
Wait β we must be careful. The rate- convolutional code has two output bits per symbol; together they specify 4 level-2 cosets. When two code paths diverge, their output-bit sequences differ; the Hamming distance of those sequences is what the convolutional code's measures. But within any one time step the two output labels may differ in either the level-1 bit or the level-2 bit (or both). The minimum Euclidean distance at a single differing time step is therefore (the smaller of and ).
Compute $d_{\rm free}^{(H)}$ for the rate-1/2, $K=3$ convolutional code
The rate- convolutional code with has constraint length and free Hamming distance .
(More precisely: one derives this from the code's state-transition diagram or from standard tables β see Convolutional Codes over BPSK" data-ref-type="interactive_plot">πBER of Rate- Convolutional Codes over BPSK. A length-3 error event is dominant.)
Lower bound
Applying TLower Bound on Free Euclidean Distance via Partition Levels, Parallel transitions dominate: a single-symbol error between two parallel edges gives , which is smaller than the accumulated distance of any coset-sequence error event.
Actual value from Ungerboeck's Table I
Ungerboeck (1982, Table I) tabulates for the 4-state 8-PSK TCM β confirming that parallel transitions indeed achieve the minimum free Euclidean distance. Since the uncoded QPSK baseline has (unit energy), the asymptotic coding gain is This is the famous "3 dB from the simplest TCM" figure. Larger trellises (8, 16, 64 states) break the tie by designing the code so that β parallel transitions no longer dominate, and the gain rises toward 6 dB.
BER of Rate- Convolutional Codes over BPSK
Bit-error rate (coded) of rate- binary convolutional codes of constraint length over BPSK/AWGN. This is the workhorse building block of TCM: the convolutional code drives the coset-selection bits. Notice how quickly adding constraint length pays off β adding one memory cell roughly doubles the decoder complexity but gains dB.
This plot is the binary precursor to the TCM BER plot in this section: combining one of these convolutional codes with an Ungerboeck 8-PSK partition yields the corresponding row of Ungerboeck's Table I.
Parameters
Common Mistake: Hamming Distance and Euclidean Distance Are Not the Same Currency
Mistake:
Confusing the free Hamming distance of the underlying convolutional code with the free Euclidean distance of the TCM scheme.
Correction:
- measures how many bit positions two distinct coset sequences must differ in β a property of the convolutional code alone.
- measures squared Euclidean distance in the signal space β a property of the convolutional code and the partition together.
These two scale differently with code memory: doubling the states of a binary convolutional code roughly adds to ; doubling the states of a TCM code typically adds to (about 3 dB worth), but only until parallel transitions become the bottleneck. Once they do, increasing state count buys nothing until is also increased.
Common Mistake: Mixing Up Coded and Uncoded Bits in the TCM Encoder
Mistake:
Believing that all information bits must go through the convolutional encoder.
Correction:
Only of the bits are coded. The remaining ride uncoded through the encoder β they select a specific point within the coset chosen by the code. This is the whole point of the partition trick: high-order uncoded bits live in a large intra-subset MSED, so they do not need protection. Forcing all bits through the code wastes trellis resources on bits that are already geometrically robust.
Quick Check
In the 4-state 8-PSK TCM of EFree Euclidean Distance of the 4-State 8-PSK TCM, the free Euclidean distance is . Which type of error event achieves this minimum?
A long error event with differing coset positions.
A single-symbol error along parallel transitions (within one coset).
An uncoded-bit-only error in the rightmost flat region of the trellis.
A catastrophic error event from a poorly chosen generator polynomial.
From the lower bound , the minimum is achieved by parallel transitions (which give exactly ). The longer error event would give , so it is not the dominant error mechanism. This is a sign that the 4-state code is parallel-transition-limited: to improve further one must either go to a level-3 partition (splitting each antipodal pair) or increase the trellis memory and the code's .
Quick Check
A TCM scheme transmits 3 bits/symbol at baud rate . The constellation size is doubled to 16-QAM, and of the 3 information bits are coded by a rate- convolutional code. What is the spectral efficiency and what is the rate of bandwidth expansion compared to uncoded 8-PSK at the same bit rate?
3 bit/symbol, bandwidth expansion factor (no expansion).
3 bit/symbol, bandwidth expansion factor .
2 bit/symbol, bandwidth expansion factor .
4 bit/symbol, bandwidth expansion factor .
Spectral efficiency is bits/symbol β the added redundancy is absorbed by enlarging the constellation from 8 to 16 points at the same baud rate. This is exactly Ungerboeck's signature feature: coding gain without bandwidth expansion. Compared to uncoded 8-PSK at the same bit rate, the bandwidth is identical; only the constellation is denser.
Hamming Distance (binary codes) vs. Euclidean Distance (TCM)
| Property | Binary convolutional code | TCM |
|---|---|---|
| Distance currency | Hamming () | Squared Euclidean () |
| Relevant minimum distance | ||
| Code operates on | Bits | Signal points (via coset indices) |
| Soft-decision gain over hard | dB | Built-in (decoder is inherently soft) |
| Spectral efficiency | Reduced by rate | Unchanged β constellation expanded instead |
| Dominant error mechanism | Always a -length event | Either parallel transitions or long events β whichever is smaller |
Key Takeaway
A good TCM design balances parallel transitions against longer error events. The free distance is . A design with a very strong code but small (few coded bits) is parallel-transition-limited β the geometry, not the code, sets the ceiling. A design with a weak code but many coded bits is Hamming-distance-limited. Ungerboeck's tables pick so that the two terms are close to equal at each state count.