The mod- Scheme: Erez–Zamir Capacity
Why Lattices Should Achieve AWGN Capacity
Shannon's 1948 proof that a Gaussian codebook achieves the AWGN capacity is famously non-constructive. The codewords are drawn i.i.d. from a Gaussian distribution, so the code has no algebraic structure whatsoever — no linearity, no symmetry, no efficient encoder or decoder. For half a century, the open question was: can a structured codebook (one with algebraic linearity, so that encoder and decoder are as cheap as matrix-vector products) achieve the same capacity?
The point is that lattices are the natural structured codebook for a Gaussian channel. Every lattice operation is linear. Every lattice codeword is an integer combination of basis vectors, and the lattice "universe" matches the Gaussian noise support. So lattices should work — but the naïve scheme (take a dense lattice, truncate to a power constraint, decode to the nearest lattice point) leaves a gap that nobody could close for 56 years.
Erez and Zamir, in their 2004 IEEE Trans. IT paper, finally closed it. Their scheme has four moving parts:
- A good fine lattice (provides the coding gain);
- A good coarse lattice (provides the shaping gain);
- An MMSE scalar at the receiver (the single scalar that Shannon's proof doesn't need but lattices do);
- A random dither , uniform on and known at both ends, that converts the input to an unconstrained AWGN input via the crypto-lemma.
The "proof pattern" of inflating a bounded input to look like an unbounded Gaussian via dither + modulo is the trick that recurs throughout lattice coding theory — in multiple-access, broadcast, interference, and compute-and-forward. Master it here, and the rest of Part IV falls into place.
Definition: MMSE Coefficient
MMSE Coefficient
For the AWGN channel with power-constrained input and noise variance per dimension, the MMSE coefficient is It is the unique scalar that minimises when is zero-mean with second moment and uncorrelated with the noise.
is strictly less than : the MMSE estimator shrinks the received signal towards zero because the noise adds energy that the estimator can discount. The parameter is the noise share of the total energy. The specific choice is what makes the residual noise after scaling have the smallest possible variance — a factor of smaller than the raw channel noise.
Definition: Dither Vector
Dither Vector
A dither is a random vector uniformly distributed on the Voronoi region of the shaping lattice, drawn independently of the message and the channel noise, and known to both transmitter and receiver (e.g., via a shared pseudo-random seed).
The dither is the ingredient that makes the transmitted signal unconditionally uniform on , regardless of which codeword was sent.
The dither is not transmitted; it is a pre-agreed random seed. In practice the shared randomness is generated by a pseudo-random sequence seeded with a public frame index (as in scrambling codes in 3GPP and DVB). The dither's only purpose is analytic: it lets us apply the crypto-lemma and reduce the bounded-input AWGN problem to an unbounded-input lattice-decoding problem.
Theorem: Crypto-Lemma (Zamir–Feder)
Let be an arbitrary (possibly deterministic) vector, and let be uniform on and independent of . Then is uniformly distributed on and independent of .
Adding an independent uniform dither on and then reducing mod- spreads any input to the uniform distribution — because any rigid translation of a uniform distribution on a fundamental region gives a uniform distribution on the (translated, then-modulo-reduced) fundamental region, and the modulo reduction maps back into the canonical Voronoi cell. This is the lattice analogue of the one-time pad.
Condition on . Show that is uniform on , a translated Voronoi region.
Use shift-invariance of Lebesgue measure and the fact that is a fundamental region, so still tiles under .
Apply the modulo reduction: is uniform on , and this holds for every conditioning — hence independence.
Conditional distribution
Fix any . Then has the distribution of shifted by : uniform on the set . This is another fundamental region of , because any translate of a fundamental region is again a fundamental region.
Modulo reduction is a bijection fundamental-region → $\mathcal{V}(\Lambda)$
The map is a bijection from any fundamental region of onto the canonical Voronoi cell (it pushes each coset representative to its canonical representative). Moreover the map is Lebesgue-measure-preserving because translation by a lattice vector preserves Lebesgue measure.
Result
Therefore is uniform on — and this holds for every . Since the conditional distribution of given is the same uniform distribution for every value of , the two random variables are independent.
Erez–Zamir Achievability: mod- vs Shannon vs Uncoded QAM
Rate curves as a function of SNR (dB) per real dimension: (i) the AWGN capacity ; (ii) the rate achieved by the Erez–Zamir mod- scheme with optimal — which coincides with at every SNR; (iii) the rate of uncoded -QAM at a fixed target symbol error probability of , showing the gap to capacity that classical truncated-cube constellations leave on the table. For small the gap is dB at high SNR; Erez–Zamir closes it entirely. Increasing asymptotically recovers the capacity but at exponentially growing complexity, whereas the lattice scheme achieves capacity with polynomial .
Parameters
Erez–Zamir mod- Encoder and Decoder
Complexity: Encoder: for mod- (if has a fast quantiser). Decoder: dominated by step 5 — the closest point, which is a CLP problem (s05).Steps 3–4 together reduce the coded AWGN channel to an unconstrained lattice AWGN channel: receiving point , decoding the closest -point without any power constraint. The Erez–Zamir theorem below shows that this reduction loses nothing — the capacity of the unconstrained lattice channel at effective SNR equals the AWGN capacity.
Theorem: Erez–Zamir (2004): Lattice Codes Achieve AWGN Capacity
For every (bits per real dimension) and every , there exist nested lattices (for some ) and a dither such that the mod- scheme (Algorithm above) with MMSE coefficient achieves rate with average error probability under a power constraint .
The proof has three moves, each earning a "good" lattice. First, the dither + mod- + MMSE reduces the bounded-input AWGN channel to an unbounded lattice channel with effective noise — a mixture of dither and scaled noise, whose second moment is exactly (the MMSE error power). Second, by the Loeliger random-lattice averaging, there exists a fine lattice whose decoding error probability on a pure Gaussian channel at SNR vanishes at rates up to the Poltyrev capacity . Third, for a Voronoi-shaping lattice with (Rogers, Poltyrev), this rate equals exactly. The three moves compose to close the full capacity gap.
Step 1: use the crypto-lemma to show that the transmit signal is uniform on and independent of the message.
Step 2: compute the effective noise after the decoder's mod- reduction. Show that , and that .
Step 3: use the Loeliger random-lattice averaging to argue existence of a good whose decoding error probability on the effective Gaussian-like noise vanishes at rates up to the Poltyrev capacity.
Step 4: optimise to drive the normalised second moment ; substitute to recover the Shannon formula.
Step 1: transmit distribution is uniform
By the crypto-lemma (Theorem above), the transmit signal is uniform on and statistically independent of the message . Its per-dimension second moment is second moment of , satisfying the power constraint with equality.
Step 2: equivalent unconstrained channel
At the decoder, , where we used by and a absorption into the mod. Define the effective noise . Its per-dimension second moment is The last equality uses : .
Step 3: the effective channel is 'almost Gaussian'
The decoder sees with . If were purely Gaussian, this would be the unconstrained lattice channel (Poltyrev channel) at effective noise power , whose capacity is by Poltyrev's analysis. The dither component is not Gaussian — it is uniform on a scaled Voronoi region — but a key lemma (Erez–Zamir's MMSE-inflation lemma, also known as the "Linder–Zamir" / Zamir–Feder moment inequality) shows that for any lattice with finite second moment, Gaussian decoding of through the - noise achieves the same asymptotic error performance as through a pure Gaussian of the same variance. The key step is Lemma 6 of Erez–Zamir: the cumulant-generating function of is dominated by that of a Gaussian of variance .
Step 4: random-lattice averaging supplies $\Lambda_c$
By the Loeliger random-lattice averaging argument (Ch. 15 Minkowski–Hlawka, Theorem 15.x), for any rate there exists a fine lattice with and decoding error probability (at effective noise variance ) vanishing exponentially in . The Poltyrev exponent is strictly positive whenever we are below the Poltyrev capacity, so ensemble averaging gives us what we need.
Step 5: optimise $\Lambda_s$ to close the shaping gap
Rogers (1959) and Poltyrev (1994) proved that there exist lattices with normalised second moment as . (The cube has , so the shaping gap is dB — Ch. 4 and s03 of this chapter.) Substituting and (the second-moment identity ), the rate bound in Step 4 becomes The last equality used to cancel. Hence every is achievable by the mod- scheme with a suitable .
Erez–Zamir mod- Scheme: Achieving AWGN Capacity with Lattices
Example: mod- at Rate bits per Real Dimension
For a nested lattice code with and , the rate is bits per real dimension. By the Erez–Zamir theorem, what is the minimum SNR at which this rate is achievable in the limit of large ? Compare to the SNR required by an uncoded 16-QAM (rate ) constellation at error probability , and to a cubic- shaping ( scaled cube) lattice code.
Shannon limit for $R = 2$
The AWGN capacity equals when , i.e., dB. This is the absolute minimum SNR at which can be reliably communicated, irrespective of coding scheme.
Erez–Zamir asymptotic
The Erez–Zamir theorem tells us that any nested lattice pair with a Voronoi-optimal shaping lattice (i.e., ) and a Poltyrev-good coding lattice (i.e., decoding error probability vanishing exponentially) achieves at dB as . The scheme closes the gap completely.
Uncoded 16-QAM baseline
Uncoded 16-QAM requires dB for symbol error probability (using the Q-function approximation). The gap to the Shannon limit is dB, split into a coding gap and a shaping gap. The shaping gap alone (going from cubic shaping to Voronoi shaping) is dB asymptotically. The remaining dB is coding gain, recoverable by choosing a dense (e.g., in dimension 8 gives dB; gives dB; random lattices in dimension recover the full dB).
Practical implication
In practice, reaching within dB of the Shannon limit at requires -ish dimensional lattices, which is done via LDPC-lattice constructions (e.g., LDPC-sigma-mapping by Sommer–Feder–Shalvi 2008). This is the structured-lattice counterpart of the LDPC-with-BICM of Ch. 9, and it is the construction that Erez–Zamir proved sharply optimal.
Why the MMSE Coefficient Matters: 56 Years of Progress
From Shannon's 1948 paper until Erez–Zamir's 2004 paper, it was known that un-scaled lattice decoding (the so-called "inflated lattice" argument of de Buda, 1989, and Urbanke–Rimoldi, 1998) could achieve only , about dB short of the Shannon capacity at high SNR. The gap was the absence of the MMSE scalar ; without it, the effective noise after mod- has variance , not , and the shaping- gain factor appears as a residual loss.
The Erez–Zamir insight was subtle but decisive: scaling by before the modulo operation is the receiver's analogue of the MMSE-DFE, and it converts the "un-inflated" lattice channel into a capacity-achieving one. The MMSE factor is a fixed scalar determined by SNR; in practical receivers it requires only AGC-level calibration.
- •
Receiver must know SNR to compute
- •
Transmitter and receiver must share the dither (pseudo-random seed)
- •
Sensitive to SNR mismatch at low SNR: a dB error in estimate can cost dB
Common Mistake: MMSE Scaling Is Not the Same as Zero-Forcing
Mistake:
Confusing the Erez–Zamir MMSE scalar with a zero-forcing or matched-filter receiver. A zero-forcing receiver would use (no scaling); a matched-filter receiver would use signal-to-total power ratio, depending on how "total" is defined. Neither choice achieves the AWGN capacity with lattice codes.
Correction:
The MMSE scalar is the exact scalar that minimises the per-dimension expected-square error when and with . It is the orthogonality-principle solution. Using introduces an information-theoretic loss of exactly dB at high SNR (the shaping-gain ceiling reappearing as a residual). The MMSE is what makes the effective noise after modulo reduction have variance rather than , closing precisely this dB gap.
Capacity Gap = Coding Gap + Shaping Gap (Additive in dB)
Once the Erez–Zamir theorem is in hand, it sharpens an earlier intuition of Forney–Ungerboeck into a quantitative statement. For any finite-dimensional lattice code , write the rate gap to capacity as These two terms are additive (in dB) and independent: depends only on , only on . For the designer this means: pick any good fine lattice you like (for its algebraic nice-ness and decoder complexity), pick any good coarse lattice you like (for its second-moment closeness to a ball), and the overall gap is the sum of the two. This is the operational payoff of the Erez–Zamir proof — it transforms "achieving capacity" into two decoupled engineering problems, neither of which is easy but both of which are well-understood.
Quick Check
In the Erez–Zamir mod- scheme with MMSE coefficient , the effective noise after modulo reduction has per-dimension second moment equal to:
(the raw channel noise variance)
(the transmit power)
Correct. The effective noise has variance after substituting . This factor is exactly the shrinkage that closes the dB shaping-gain gap.
Key Takeaway
The Erez–Zamir mod- scheme achieves the AWGN capacity with a structured codebook, by combining a good fine lattice (via Loeliger's random lattices), a good coarse lattice (with normalised second moment ), an MMSE scalar , and a shared dither uniform on . The crypto-lemma turns the bounded-input channel into an unconstrained lattice channel; the MMSE factor shrinks the effective noise by ; and optimising kills the dB shaping gap. The capacity gap of any finite lattice code decomposes additively (in dB) into a coding gap and a shaping gap — two independent knobs for the designer.