Slepian-Wolf Coding
Why Distributed Source Coding?
Consider two sensors measuring correlated quantities β say, temperature at two nearby locations. Each sensor must compress its data independently (they cannot communicate with each other), but a central base station receives both compressed streams and decodes them jointly. The fundamental question is: does the inability to coordinate encoding cost us in compression efficiency?
The remarkable answer, due to Slepian and Wolf (1973), is no β at least in terms of the achievable rate region. Separate encoding with joint decoding achieves exactly the same rates as if the two encoders could cooperate. This is one of the most surprising results in information theory, and the proof technique β random binning β is as fundamental to multiuser source coding as random coding is to channel coding.
Slepian-Wolf Rate Region Animation
Definition: Distributed Source Coding Setup
Distributed Source Coding Setup
Let be i.i.d. copies of a pair of correlated discrete random variables with finite alphabets and .
Encoder 1 observes and produces an index .
Encoder 2 observes and produces an index .
The joint decoder receives both indices and produces estimates .
A rate pair is achievable if for every and all sufficiently large , there exist encoders and a decoder such that .
The Slepian-Wolf rate region is the closure of the set of all achievable rate pairs.
Random binning
A proof technique where each source sequence is randomly assigned to one of bins. The decoder uses joint typicality to identify the correct sequences within their respective bins. Random binning is the source coding analogue of random coding for channel coding.
Related: Random coding, Slepian-Wolf coding, Joint typicality decoding
Slepian-Wolf rate region
The set of rate pairs for which lossless distributed source coding is achievable: , , .
Theorem: Slepian-Wolf Theorem
The Slepian-Wolf rate region for lossless distributed compression of is the set of rate pairs satisfying:
The point is that separate encoding achieves the same rate region as joint encoding β the only requirement is that decoding be performed jointly.
Think of it this way: if Encoder 2 sends at rate , then the decoder knows perfectly, and Encoder 1 only needs to send at rate . The surprise is that Encoder 1 can achieve rate even though it does not know β it just needs to send enough information for the decoder (who does know ) to resolve the ambiguity. Random binning achieves this: each bin contains roughly sequences, and the correlation with is enough to pick out the right one.
Achievability via random binning
Fix . We prove that any rate pair in the interior of the Slepian-Wolf region is achievable.
Random binning: Independently and uniformly assign each sequence to a bin , and each to a bin .
Encoding: Encoder 1 sends . Encoder 2 sends .
Decoding: Given bin indices , the decoder looks for a unique pair such that:
- (in the correct -bin),
- (in the correct -bin),
- (jointly typical).
Error analysis
An error occurs if:
- : β the true pair is not jointly typical.
- : There exists with and .
- : There exists with and .
- : There exist , in the correct bins with .
By the AEP, .
Bounding the error events
For : The number of sequences that are jointly typical with is at most . Each falls in the same bin as with probability . So:
This vanishes if .
Similarly, if .
For : The number of jointly typical pairs is at most , each pair falls in the correct bins with probability , so:
This vanishes if .
Converse via Fano's inequality
Suppose a rate pair is achievable with . By Fano's inequality:
where . Then:
The last step uses the i.i.d. property. Dividing by and letting gives . By symmetry, . The sum-rate bound follows from:
The Slepian-Wolf Surprise
The Slepian-Wolf theorem is genuinely surprising. Consider the corner point . Encoder 1 compresses to bits per symbol β this is the conditional entropy, the rate you would need if you knew . But Encoder 1 does not know ! The trick is that Encoder 1 does not need to know to compress at this rate β it just needs to send enough information for the decoder, who does know after decoding Encoder 2's message, to resolve the remaining uncertainty.
Intuitively, what happens is that random binning creates a many-to-one mapping. The bin index, combined with the side information at the decoder, is enough to uniquely identify among the bin members. This is possible because the correlation between and effectively "narrows down" the search within the bin.
Historical Note: Slepian and Wolf's Landmark Paper
1973David Slepian and Jack Wolf published their result in 1973, but the ideas had been circulating in Bell Labs for several years before that. The result was initially met with some skepticism β how could separate encoders do as well as joint encoders? The answer lies in the power of joint decoding and the structure that random binning introduces. The paper also introduced the concept of a rate region, which became the standard way to characterize multiuser information-theoretic limits.
The practical significance of Slepian-Wolf coding was not realized for decades. It was not until the early 2000s, when Pradhan and Ramchandran showed that LDPC syndrome codes could implement Slepian-Wolf coding efficiently, that the result found practical applications in distributed sensor networks and distributed video coding.
Example: Slepian-Wolf for Binary Symmetric Sources
Let and where is independent of (so is a doubly symmetric binary source with parameter ). Compute the Slepian-Wolf rate region and sketch it.
Compute the relevant entropies
Since and :
- bit
- bit (since is also )
- (knowing leaves only the uncertainty of )
- (by symmetry)
Write the rate region
The Slepian-Wolf region is:
This is a pentagon in the plane. The corner points are:
- : Encoder 2 sends at full rate, Encoder 1 only sends the "innovation" relative to
- : the symmetric corner
- Any point on the line between these corners
When (perfect correlation), the region degenerates: , so one encoder can be silent. When (independence), the region is β no compression gain from joint decoding.
Slepian-Wolf Rate Region Explorer
Explore how the Slepian-Wolf rate region changes with the correlation parameter for the doubly symmetric binary source. The shaded region shows achievable rate pairs.
Parameters
BSC parameter controlling correlation between X and Y
Definition: Random Binning Construction
Random Binning Construction
A random binning scheme at rate assigns each sequence to a bin uniformly and independently at random.
The bin assignment partitions into bins, each containing approximately sequences.
The encoder sends the bin index , which requires bits. The decoder must determine which sequence within the bin was the true source output.
Definition: Syndrome-Based Slepian-Wolf Coding
Syndrome-Based Slepian-Wolf Coding
For binary sources, Slepian-Wolf coding has an elegant implementation via syndrome coding. Let be the parity-check matrix of a linear code.
Encoder: Given , compute the syndrome and send (which requires bits if is ).
Decoder: Given , find the in the coset that is most likely given .
This is equivalent to decoding the error pattern (where is any member of the coset) using as side information. The rate is , and reliable decoding requires .
The syndrome approach reveals a deep duality between channel coding and source coding: the parity-check matrix of a good channel code is a good source coding binning matrix, and vice versa.
Quick Check
For the Slepian-Wolf problem with , which of the following is true about the corner point ?
Encoder 1 must know to achieve this rate
Encoder 1 compresses as if it had access to , even though it does not
This point requires joint encoding of
This point is not achievable β only the interior of the region is
Correct! The random binning argument shows that the same conditional rate is achievable even without side information at the encoder. The decoder uses to resolve the bin ambiguity.
Common Mistake: Confusing Slepian-Wolf with Source Coding with Side Information
Mistake:
Students often confuse the Slepian-Wolf problem (both sources must be reconstructed losslessly) with source coding with side information at the decoder (only needs to be reconstructed, is available at the decoder for free). In the latter, the minimum rate is simply .
Correction:
In Slepian-Wolf, both sources are encoded and both must be reconstructed. The rate region involves constraints on , , and . Source coding with side information at the decoder is the special case where is "free" (the decoder already has ), giving the corner point .
Common Mistake: Random Binning is a Proof Technique, Not a Construction
Mistake:
Treating random binning as a practical coding scheme. In the proof, the bin assignments are random and known to both encoder and decoder. A specific realization of the binning works with high probability, but finding good deterministic binning schemes is a separate (hard) problem.
Correction:
Random binning proves existence of good codes. Practical implementations use structured codes (e.g., LDPC syndromes) that approximate the random binning behavior. The transition from random to structured codes is analogous to going from random codes to turbo/LDPC codes in channel coding.
LDPC-Based Slepian-Wolf Coding in Practice
The most practical implementations of Slepian-Wolf coding use LDPC syndrome codes. The idea, introduced by Pradhan and Ramchandran (2003) in the DISCUS framework, is to use the parity-check matrix of a capacity-approaching LDPC code for binning. The encoder sends the syndrome , and the decoder uses belief propagation with the side information to recover .
This approach achieves rates within a fraction of a dB of the Slepian-Wolf limit for symmetric binary sources, and is the basis for distributed video coding standards like DISCOVER.
- β’
LDPC codes must be designed for the specific correlation structure
- β’
Belief propagation convergence requires careful graph design
- β’
Rate adaptation is needed when correlation statistics are unknown
Random Binning Visualization
Visualize how random binning partitions source sequences into bins and how joint typicality with the side information resolves bin ambiguity. The plot shows bin assignments and the decoding region for a simple binary example.
Parameters
Length of source sequences
Rate of the binning scheme (bits per symbol)
Crossover probability of the BSC correlation model
Quick Check
In the Slepian-Wolf problem, the sum-rate constraint is . Why is this not simply ?
Because the sources are correlated, joint decoding exploits the correlation to reduce the total rate
Because the encoders communicate with each other
Because the decoder uses feedback to the encoders
Correct! Joint decoding allows the decoder to exploit the statistical dependence between and . The joint entropy is less than by exactly the mutual information.
Why This Matters: Slepian-Wolf Coding in Sensor Networks
Distributed sensor networks are the canonical application of Slepian-Wolf coding. Sensors measure correlated physical quantities (temperature, humidity, pressure) and must transmit their measurements to a fusion center over rate-limited links. The Slepian-Wolf theorem guarantees that the total rate needed is determined by the joint entropy, not the sum of individual entropies β even though the sensors cannot coordinate their compression.
In modern IoT systems, this principle is used in compressed sensing for distributed data acquisition and in cooperative communication protocols where relays compress overheard signals. See Book telecom, Ch. 12 for practical code constructions that approach the Slepian-Wolf limit.
Key Takeaway
The Slepian-Wolf theorem shows that separate encoding of correlated sources incurs no rate penalty compared to joint encoding, provided decoding is performed jointly. The achievable rate region is the set of satisfying , , . The proof uses random binning for achievability and Fano's inequality for the converse.