Examples of DMC Capacity
Computing Capacity for Important Channels
The channel coding theorem tells us , but actually computing the maximum requires solving an optimization problem. For channels with symmetry structure, the solution simplifies dramatically. We now compute the capacity of the most important DMC examples, each illustrating a different technique.
BSC vs. BEC Capacity Comparison
Definition: Binary Symmetric Channel (BSC)
Binary Symmetric Channel (BSC)
The binary symmetric channel BSC() has with transition law:
where is modulo-2 addition and is independent of . The parameter is the crossover probability β the probability that a bit is flipped.
Theorem: Capacity of the BSC
The capacity of the BSC() is:
where is the binary entropy function. The capacity-achieving input distribution is (uniform).
The maximum output entropy is bit (achieved when is uniform). The noise entropy is (regardless of the input distribution). So is maximized by making , which requires .
Compute mutual information
X \sim \text{Bernoulli}(1/2)Y = X \oplus Z\text{Bernoulli}(1/2)H(Y) = 1$.
Definition: Binary Erasure Channel (BEC)
Binary Erasure Channel (BEC)
The binary erasure channel BEC() has , :
With probability , the input passes through perfectly; with probability , it is erased (the decoder sees "?" and knows the bit was lost but not which bit it was).
Theorem: Capacity of the BEC
The capacity of the BEC() is:
The capacity-achieving input distribution is (uniform).
A fraction of the bits arrive perfectly, and a fraction are lost. So the effective throughput is bits per channel use. The beauty of the BEC is that it separates "noise" from "loss" β when a bit arrives, it arrives perfectly. This makes the BEC the ideal channel for understanding erasure-based coding (LDPC codes, fountain codes).
Compute conditional entropy
Y = 0X = 0H(X|Y=0) = 0Y = 1X = 1H(X|Y=1) = 0Y = ?X\text{Bernoulli}(P_X(1))H(X|Y=?) = H(X)H(X|Y) = \epsilon \cdot H(X)$.
Maximize mutual information
$
Definition: Symmetric and Strongly Symmetric Channels
Symmetric and Strongly Symmetric Channels
A DMC with transition matrix (where ) is:
-
Weakly symmetric: Every row of is a permutation of the first row. (Each input "sees" the same noise pattern, just shuffled.)
-
Strongly symmetric: Additionally, every column of is a permutation of the first column.
For a strongly symmetric channel:
achieved by the uniform input distribution .
Both the BSC and the BEC are strongly symmetric. The additive noise channel over is also strongly symmetric with capacity .
Theorem: Capacity of Additive Noise Channels
For a discrete additive noise channel over (addition in the finite field of order ), where has PMF independent of :
The capacity-achieving input distribution is .
The maximum output entropy is (achieved when is uniform, which happens when is uniform since is a convolution). The noise consumes bits of the output entropy. The BSC is the special case , : .
Strongly symmetric structure
The transition matrix has . Each row is a cyclic shift of the noise PMF, and each column sums to 1 with the same set of values (permuted). This is strongly symmetric.
Example: Capacity of the Z-Channel
The Z-channel has with:
- (0 always passes through)
- , (1 may be flipped to 0)
Compute the capacity and the capacity-achieving input distribution.
Compute mutual information
Let . Then:
Optimize over $\alpha$
Taking the derivative with respect to and setting to zero:
This gives (not uniform!).
The capacity is: .
Note: the Z-channel is NOT symmetric, so the optimal input is not uniform. For : (noiseless). For : .
DMC Capacity Comparison
Compare the capacity of the BSC, BEC, and Z-channel as a function of the noise/erasure parameter. The BSC has the lowest capacity for a given because bit errors are worse than erasures.
Parameters
Upper limit of the noise parameter range
BSC Capacity vs. Crossover Probability
The capacity of the BSC as a function of the crossover probability . At : perfect channel (). At : completely noisy ().
Parameters
Toggle display of the binary entropy function
Comparison of Important DMC Channels
| Channel | Capacity | Optimal | Symmetric? |
|---|---|---|---|
| BSC() | Bernoulli(1/2) | Strongly symmetric | |
| BEC() | Bernoulli(1/2) | Strongly symmetric | |
| Z-channel() | Bernoulli(), | Not symmetric | |
| Additive | Uniform on | Strongly symmetric |
Historical Note: The BEC: A Pedagogical Powerhouse
1955The binary erasure channel was introduced by Elias in 1955 as a simplified model for packet-based communication. Its clean structure (no bit errors, only losses) makes it the ideal setting for understanding LDPC codes (density evolution is exact on the BEC), polar codes (polarization is easiest to analyze on the BEC), and fountain codes (rateless codes for the BEC).
In modern systems, the BEC models packet erasures in internet communication, where packets either arrive perfectly (via checksums) or are declared lost. This is the operational regime of HTTP, TCP/IP, and streaming protocols.
Common Mistake: BEC Capacity is Always Higher than BSC Capacity
Mistake:
Assuming the BSC and BEC have similar capacities for the same parameter value (e.g., BSC(0.1) and BEC(0.1)). Students sometimes treat and as equivalent.
Correction:
For the same parameter: while for . The BEC always has higher capacity because erasures are "kinder" than errors: the decoder knows which bits were lost (and can request retransmission), while with errors it does not even know which bits are wrong.
Key Takeaway
The BSC capacity is , the BEC capacity is , and additive noise channels over have capacity . For symmetric channels, the uniform input is optimal. For asymmetric channels (like the Z-channel), the optimal input must be computed via optimization. Erasures are less harmful than errors: for the same parameter value.