Polar Codes β Channel Polarization
The Polarization Phenomenon
Polar codes, introduced by Erdal Ar\i kan in 2009, are the first (and so far only) family of codes with an explicit construction that provably achieves the capacity of any binary-input symmetric memoryless channel, with encoding and decoding complexity .
The idea is beautifully simple: apply a specific linear transformation to copies of a channel . The resulting "virtual" channels polarize β as grows, each virtual channel becomes either nearly perfect (capacity ) or nearly useless (capacity ). The fraction of good channels approaches the capacity of the original channel. Send information on the good channels, send known (frozen) bits on the bad ones.
Definition: The Polar Transform
The Polar Transform
Start with independent copies of a binary-input DMC with capacity .
The polar transform applies the matrix , where is the basic kernel, is the -fold Kronecker product, and is a bit-reversal permutation.
The input bits are transformed to channel inputs (mod 2 for binary, or over for the Gaussian channel).
Definition: Bit-Channels (Virtual Channels)
Bit-Channels (Virtual Channels)
After the polar transform, define the -th bit-channel as
i.e., the channel from to the outputs and all previous inputs (which are assumed known at the decoder for successive cancellation).
The symmetric capacity of the -th bit-channel is .
Theorem: Channel Polarization Theorem
For any binary-input symmetric DMC with capacity , the bit-channels polarize as : for any ,
That is, the fraction of "good" bit-channels approaches the capacity , and the fraction of "bad" bit-channels approaches .
The polar transform is a recursive process. At each level, two copies of a channel are combined into a "better" channel and a "worse" channel . The total capacity is preserved: . But the capacity splits unevenly, and repeated application drives all channels to the extremes. This is the polarization phenomenon β and it is the information-theoretic equivalent of distillation.
The basic polarization step
Given two independent copies of , define:
- (the "worse" channel)
- (the "better" channel)
By the chain rule: .
Moreover, (data processing inequality for , side information for ).
Recursive application
Apply the same splitting recursively: and . After levels, we have bit-channels. The sequence of capacities forms a bounded martingale.
Convergence
By the martingale convergence theorem, converges almost surely. Since the process preserves total capacity and drives individual channels toward 0 or 1, the limiting distribution must be a Bernoulli mixture: fraction at 1 and fraction at 0.
Polar code
A capacity-achieving code based on channel polarization. Information bits are sent on the "good" virtual channels (high mutual information), while "frozen" known bits are sent on the bad channels. Encoding is via the Kronecker structure; decoding is via successive cancellation.
Related: Turbo code, LDPC code
Frozen bits
In a polar code, the bits assigned to unreliable (low-capacity) bit-channels. These are set to predetermined values (usually 0) known to both encoder and decoder.
Related: Polar code
Successive Cancellation (SC) Decoding
Complexity: using the recursive structure of the polar transform. The LLR computation at step 1a decomposes into a butterfly network of stages.SC decoding is sequential (bit-by-bit), which limits throughput. CRC-aided SC list (CA-SCL) decoding maintains candidate paths in parallel and selects the one passing a CRC check, achieving near-ML performance at the cost of complexity.
Channel Polarization Tree
Channel Polarization Visualization
Observe how the bit-channel capacities evolve as grows. For a BEC with erasure probability , the capacities split recursively toward 0 and 1. Adjust and the channel quality to see the polarization effect.
Parameters
Example: Polar Code for the BEC
Design a polar code of length for the BEC with erasure probability . Which bits are frozen?
Compute bit-channel erasure probabilities
For the BEC, the polarization recursion is exact: (worse channel), (better channel).
Starting from : Level 1: , . Level 2: , , , . Level 3 (8 channels): .
Select frozen and information bits
The channel capacity is , so we should have information bits.
Rank by reliability (lowest erasure probability): Bit 8: , Bit 6: , Bit 4: , Bit 7: .
Frozen set: (worst 4 channels). Information set: (best 4 channels).
Why This Matters: Polar Codes in 5G NR
Polar codes were adopted for the 5G NR control channels (PDCCH, PUCCH, PBCH) in 2016, making them the first provably capacity-achieving code family deployed in a commercial standard. The 5G implementation uses CRC-aided SC list (CA-SCL) decoding with list size , which provides near-ML performance at short to moderate block lengths (up to 1024 bits).
The choice of polar codes for control channels reflects their excellent performance at short block lengths and low rates β exactly the regime where control information operates. For long blocks (data), LDPC codes remain preferred due to their higher throughput and easier parallelization.
See Book telecom, Ch. 24 for the full treatment of 5G NR channel coding.
Historical Note: ArΔ±kan's Breakthrough
Erdal Ar\i kan published the polar coding paper in 2009, and it was immediately recognized as a breakthrough β the first explicit, provably capacity-achieving construction with polynomial complexity. The elegance of the proof (martingale convergence!) and the simplicity of the construction (just a Kronecker product of a matrix) were remarkable.
Ar\i kan received the IEEE Richard W. Hamming Medal in 2010, the Shannon Award in 2018, and the Mustafa Prize in 2019 for this work. The rapid adoption of polar codes in 5G NR (less than 10 years from publication to commercial deployment) is exceptional in the history of coding theory.
Quick Check
In a polar code of length for a channel with capacity , approximately how many information bits are transmitted?
4
8
12
16
The number of information bits is approximately . The remaining 4 bit-channels are frozen (set to known values). As grows, the fraction of information bits approaches exactly .