Source Coding with a Helper
When Side Information is Rate-Limited
In the Wyner-Ziv problem (Chapter 6), the decoder has full access to side information . In the Slepian-Wolf problem (Chapter 7), both sources are encoded at positive rates. But what about the intermediate case: one encoder compresses , while a helper who observes correlated can send a rate-limited description to the decoder?
This is the source coding with a helper problem. The helper cannot send in full β it has a rate budget β but even a partial description of can help the decoder reconstruct . The natural question is: what rate pair is needed to achieve distortion ?
This problem connects to Wyner's common information, a measure of the shared structure between and that is more nuanced than mutual information.
Distributed Video Coding Pipeline
Definition: Source Coding with a Helper
Source Coding with a Helper
Let with distortion measure .
- Main encoder: observes , sends index
- Helper encoder: observes , sends index
- Decoder: produces
The rate-distortion region is the closure of rate-distortion triples for which for all and sufficiently large .
When , this reduces to standard rate-distortion. When , the decoder has full side information and we recover the Wyner-Ziv problem. The interesting regime is .
Definition: Wyner's Common Information
Wyner's Common Information
The Wyner common information between and is defined as:
where the minimum is over all random variables such that and are conditionally independent given : .
Equivalently, where has the smallest entropy among all common causes that render and independent.
Wyner's common information satisfies . It equals when is jointly Gaussian, and equals when one variable is a deterministic function of the other. Unlike mutual information, common information captures the "shared randomness" that cannot be split β it measures the cost of generating from independent components via a common random variable.
Wyner's common information
A measure of shared structure between two random variables, defined as the minimum rate needed to generate from conditionally independent components. It is at least as large as mutual information and captures aspects of dependence that mutual information misses.
Related: Mutual Information, Common randomness, Source Coding with a Helper
Theorem: Rate Region for Source Coding with a Helper (Lossless Case)
For lossless reconstruction of (, is Hamming distance), the minimum rate for the main encoder as a function of the helper rate is:
In particular:
- At : (no help)
- At : (full Slepian-Wolf benefit)
The transition between these extremes depends on the joint distribution .
The helper sends a compressed version of at rate . The decoder uses as side information to decode , needing rate . The question is: which minimizes subject to the rate constraint ?
Intuitively, the helper should focus on sending the parts of that are most informative about . This is precisely the optimization that Wyner's common information addresses: the helper extracts the "common" component between and and sends it to the decoder.
Achievability
Fix satisfying the rate and Markov constraints. The helper uses a Wyner-Ziv-like code to send at rate . The main encoder uses Slepian-Wolf coding with side information at the decoder, achieving rate .
Converse
By Fano's inequality and the Markov chain structure, the minimum rate for the main encoder given any helper description of rate is lower bounded by where is the best -rate description of for predicting .
Example: Helper for the Doubly Symmetric Binary Source
Let and where , . The helper observes and can send at rate . What is the minimum rate for lossless reconstruction of as a function of ?
Boundary values
- At : bit
- At : bits
The full Slepian-Wolf benefit is achieved when bits.
Intermediate regime
For , the helper sends a partial description of . The optimal strategy is to use a "soft" version of that preserves the most information about .
For the DSBS, the rate function is piecewise linear between the two extremes, achieved by time-sharing between the corner points of the Slepian-Wolf region.
Historical Note: Two Notions of Common Information
1975There are actually two distinct notions of "common information" in information theory, both introduced in 1975. Wyner's common information (discussed here) is the minimum rate to generate the joint distribution from conditionally independent components. Gacs and Korner's common information is the maximum rate of "common randomness" that can be extracted from the joint distribution.
These two quantities can be very different: for the DSBS, Wyner's common information is (close to 1 for small ), while Gacs-Korner's common information is 0 (there is no nontrivial common part). The gap between them reflects different operational settings β generation vs. extraction of shared randomness.
Quick Check
In source coding with a helper, what happens when the helper's rate exceeds ?
Further increasing continues to reduce
The minimum saturates at
The system becomes equivalent to joint encoding
Correct! Once the helper sends enough to resolve (given the decoder also has 's encoding), the full Slepian-Wolf benefit is achieved.
Common Mistake: Confusing Common Information with Mutual Information
Mistake:
Treating Wyner's common information as if it equals the mutual information . While always holds, they can differ significantly β especially for discrete sources.
Correction:
Wyner's common information measures the minimum rate to make and conditionally independent, while mutual information measures the total statistical dependence. For jointly Gaussian variables, they coincide. For discrete variables, can be much larger than .
Key Takeaway
Source coding with a helper bridges the gap between standard compression () and Slepian-Wolf coding (). The optimal helper strategy extracts the "common information" between and , as formalized by Wyner's common information. The rate savings from even a small helper rate can be substantial when the sources are strongly correlated.