The Shaping Gain and the Ceiling
The dB Shannon Tax
We now derive the number dB β the single most famous number in signal shaping. Where does it come from? Two observations, connected by one inequality:
- Gaussian is optimal. The maximum-entropy distribution at fixed energy on is the isotropic Gaussian. A finite, uniformly distributed constellation cannot match its entropy at the same energy.
- Cubes are bad boundaries. A bounded-support distribution with the same entropy as a Gaussian has higher energy. For a cube the penalty is . For a sphere the penalty vanishes as .
Combining these gives the claim: the best possible shaping gain (over cubic QAM) is , and it is achieved asymptotically by spherical constellations. This is the part of the gap to Shannon capacity that we can never close with coding alone β we have to shape the marginal distribution. But it is also upper bounded by a single universal dB number. Once you have your dB, you are done with shaping forever. The rest of the gap is coding.
Theorem: Shaping Gain Ceiling:
For any lattice , with equality only in the limit for the -dimensional ball. Consequently the shaping gain satisfies i.e.\ dB.
The proof is a one-liner combining two classical facts: the Gaussian maximum-entropy theorem (the most random-looking distribution at fixed energy is Gaussian) and the fact that the differential entropy of a bounded-support uniform distribution equals , where is the support volume. A uniform distribution on has the same entropy as the Gaussian if and only if their variances are related by , i.e.\ .
Differential entropy of uniform on a region of volume equals (in nats per dimension, dividing by ).
Differential entropy of a Gaussian with per-dimension variance equals .
Apply the Gaussian max-entropy bound.
Entropy of a uniform distribution on $\mathcal{V}(\Lambda_s)$
Let be uniform on . Its differential entropy (per dimension, in nats) is Its per-dimension energy is .
Gaussian maximum-entropy
Among all random vectors with per-dimension variance , the isotropic Gaussian maximises differential entropy, with value per dimension. Hence
Combine
Substituting on the right and on the left: The terms cancel, leaving , i.e.\ .
Equality and the shaping gain bound
Equality holds when the uniform distribution on matches the Gaussian β which, by the entropy-power inequality, is possible only as with approaching a ball. Dividing into the constant gives dB.
Shaping Gain vs Dimension
The per-dimension shaping gain of the -ball (the infimum over lattices at dimension ) climbs from dB at to the dB ceiling as . The curve grows slowly: you need to get the first dB, and another factor of in dimension to approach dB. This is why practical shaping stays in moderate dimensions (shell mapping uses ) and why the last dB of shaping gain is almost impossibly expensive.
Parameters
Theorem: Normalised Second Moment of the -Ball
The -dimensional ball of volume has normalised second moment As , by Stirling's approximation.
The ball is the "roundest" possible region at a given volume, so it achieves the smallest possible second moment per unit volume. The fact that converges to the Gaussian limit says that a high-dimensional uniform distribution on a ball is indistinguishable from a Gaussian in its moments β the entropy gap closes.
Integrate in spherical coordinates.
Use the volume of an -radius ball.
Stirling: gives the limit.
Second moment of a ball of radius $R$
By symmetry, where is the surface area of the unit -sphere. Dividing by gives the per-dimension second moment .
Normalisation
Using , we have . Therefore
Asymptotic limit
Using Stirling's approximation , one obtains to leading order, so as .
Reading the Ultimate Gap
A canonical way to read Shannon's capacity formula is as a prescription: if you are dB below capacity at some rate , then decomposes as
The three components are independent. An uncoded QAM constellation at moderate rate sits at dB from capacity. A good code closes about dB of that (the coding side); the last dB is the shaping tax, which no code β however long or cleverly designed β can close. Shaping is the necessary complement to coding, not an optional refinement.
Example: Shaping Gain of
Compute the shaping gain and express it as a fraction of the ultimate ceiling .
Second moment of $E_8$
ConwayβSloane (Ch. 2) report , based on exact integration over the Voronoi cell.
Shaping gain
, i.e.
dB.
Fraction of the ceiling
The ceiling is (ratio), i.e.\ dB. gets to of the available shaping gain in just dimensions. For comparison, the Leech lattice gets to , and (standard QAM) gets to .
Common Mistake: Shaping and Coding Contributions Are Orthogonal, Not Interchangeable
Mistake:
Claiming that a stronger code can compensate for missing shaping, or that a better shaping strategy reduces the need for coding.
Correction:
The total gap to Shannon capacity equals (in dB) the sum of the coding-gap and shaping-gap terms. A code improves the coding term but leaves the shaping term untouched; a shaping scheme improves the shaping term but leaves the coding term untouched. The two cannot substitute for each other. Even a capacity-achieving binary code on a uniform QAM constellation is dB short of Shannon β the shaping tax is unavoidable without shaping.
Common Mistake: The Ceiling Is Asymptotic
Mistake:
Assuming a practical shaping scheme at or dimensions can recover the full dB.
Correction:
The ceiling is attained only in the limit . At (Gosset lattice ) the best shaping gain is dB; at (Leech lattice ) it is dB. The last half-decibel of shaping gain is prohibitively expensive in complexity, which is why all practical systems accept dB and move on.
Why This Matters: Probabilistic Shaping: Recovering dB with a Different Mechanism
Modern standards (DVB-S2X, 5G, optical coherent) recover the same shaping gain through a different mechanism: probabilistic shaping. Instead of confining lattice points to a bounded region, the constellation is left fixed and the input distribution is made non-uniform β inner points appear more frequently than outer points. BΓΆcherer, Steiner, and Schulte (2015) showed that this is equivalent to spherical shaping in the high-rate regime. Chapter 19 will return to this in detail, with the probabilistic-amplitude-shaping (PAS) architecture of modern coherent optical links as the main example. The connection to this chapter is direct: the information-theoretic ceiling applies to probabilistic shaping as well, and from the same max-entropy argument.
Quick Check
Why is a universal upper bound on shaping gain?
Because the Gaussian is the maximum-entropy distribution at fixed variance, and the -ball achieves this limit as .
Because the AWGN channel capacity is .
Because is the Euler number times .
Because the -ball has volume .
The shaping gain is proportional to the ratio of "support volume" to "energy". The Gaussian distribution β which maximises entropy at fixed variance β provides the tightest such ratio in infinite dimensions. Bounded-support distributions (like uniform on a lattice's Voronoi region) cannot exceed this ratio, and the -ball approaches it asymptotically. The numerical value is the dB-ratio between the Gaussian's optimum and the cube's .
Key Takeaway
The shaping gain ceiling is dB. This number comes from a single inequality: the differential entropy of a uniform distribution on a bounded region is at most the differential entropy of a Gaussian at the same variance. The ceiling is attained in the limit by the -ball; all finite-dimensional constellations fall short. In practice, dB is readily attainable at ; the last half-decibel costs enormous complexity. The next section surveys the two practical shaping schemes β shell mapping and trellis shaping β that attempt to approach the ceiling.