Channels with State Known at the Decoder

Side Information at the Decoder

What if the state is known at the decoder instead of (or in addition to) the encoder? This scenario arises naturally in fading channels where the receiver estimates the channel (CSIR) through pilot symbols.

The answer turns out to be much simpler than the encoder-side case: the decoder simply uses the state as additional observations, and the capacity is max⁑PXI(X;Y,S)\max_{P_X} I(X; Y, S). There is no need for binning or auxiliary variables β€” the state directly improves the decoder's ability to distinguish codewords.

The asymmetry between encoder-side and decoder-side state information is one of the recurring themes of multiuser information theory.

Theorem: Capacity with State Known at the Decoder

The capacity of a DMC with state (PY∣X,S,PS)(P_{Y|X,S}, P_S) where the state SnS^n is known at the decoder (but not at the encoder) is

C=max⁑PXI(X;Y,S)=max⁑PXI(X;Y∣S),C = \max_{P_X} I(X; Y, S) = \max_{P_X} I(X; Y | S),

where XβŠ₯SX \perp S and the second equality follows from the independence of XX and SS.

The decoder knows (Yn,Sn)(Y^n, S^n) and uses both for decoding. Since the encoder does not know SS, the best it can do is choose PXP_X to maximize the mutual information as if SS were part of the channel output. The condition I(X;Y,S)=I(X;Y∣S)I(X; Y, S) = I(X; Y|S) holds because XβŠ₯SX \perp S.

Example: Gaussian Channel with State at the Decoder

For Y=X+S+ZY = X + S + Z with S∼N(0,Q)S \sim \mathcal{N}(0, Q), Z∼N(0,N)Z \sim \mathcal{N}(0, N), and state known at the decoder only, compute the capacity.

Capacity with Different State Information Configurations

ConfigurationCapacity (Gaussian)Comment
No state info12log⁑(1+P/(Q+N))\frac{1}{2}\log(1 + P/(Q+N))Interference treated as noise
Causal at encodermax⁑α12log⁑Pβˆ’Ξ±2Q(1βˆ’Ξ±)2Q+N+1\max_\alpha \frac{1}{2}\log\frac{P - \alpha^2 Q}{(1-\alpha)^2 Q + N} + 1Partial cancellation possible
Non-causal at encoder (DPC)12log⁑(1+P/N)\frac{1}{2}\log(1 + P/N)Costa: interference eliminated
At decoder only (CSIR)12log⁑(1+P/N)\frac{1}{2}\log(1 + P/N)State subtracted at decoder
At both (full CSI)12log⁑(1+(P+Q)2/N)\frac{1}{2}\log(1 + (\sqrt{P} + \sqrt{Q})^2/N)Coherent combining of XX and SS

The Asymmetry of Side Information

The comparison table reveals a subtle asymmetry:

  • Decoder state is always helpful (it can only improve decoding).
  • Encoder state (causal) helps partially but not completely.
  • Encoder state (non-causal) achieves the same as decoder state for the Gaussian channel, but via a completely different mechanism (binning vs. direct observation).
  • Full CSI (state at both) is strictly better than either alone, because the encoder can coherently add to the state signal.

The Gaussian case is special because DPC and CSIR give the same result. For discrete channels, the Gel'fand-Pinsker capacity can exceed the CSIR capacity β€” the encoder can exploit the state in ways that simply observing it at the decoder cannot.

Capacity Under Different State Information Configurations

Compare the channel capacity for the Gaussian state channel Y=X+S+ZY = X + S + Z under different CSI configurations. Adjust the interference power QQ to see how the gap between configurations changes.

Parameters
10
1
50

Common Mistake: Full CSI Is Not Just DPC + CSIR

Mistake:

Assuming that having state at both encoder and decoder gives capacity 12log⁑(1+P/N)\frac{1}{2}\log(1 + P/N) (same as one-sided).

Correction:

When state is known at both sides, the encoder can coherently align XX with SS, achieving C=12log⁑(1+(P+Q)2/N)C = \frac{1}{2}\log(1 + (\sqrt{P} + \sqrt{Q})^2/N). This is strictly larger than 12log⁑(1+P/N)\frac{1}{2}\log(1 + P/N) when Q>0Q > 0. The encoder treats SS as a free power source, allocating XX to constructively add to SS. This is the principle behind energy harvesting and simultaneous wireless information and power transfer.

Quick Check

For Y=X+S+ZY = X + S + Z with SS known only at the decoder, P=5P = 5, Q=100Q = 100, N=2N = 2, the capacity is:

12log⁑(1+5/102)β‰ˆ0.034\frac{1}{2}\log(1 + 5/102) \approx 0.034 bits

12log⁑(1+5/2)=12log⁑(3.5)β‰ˆ0.91\frac{1}{2}\log(1 + 5/2) = \frac{1}{2}\log(3.5) \approx 0.91 bits

12log⁑(1+105/2)β‰ˆ2.88\frac{1}{2}\log(1 + 105/2) \approx 2.88 bits

Depends on QQ