Independence
Why Independence Matters
Independence is the structural assumption that makes most of information theory work. The i.i.d. (independent and identically distributed) model for sources and channels allows us to factor joint probabilities, write channel capacities as single-letter expressions, and prove coding theorems via the law of large numbers. When independence fails β correlated fading, bursty interference, memory in the channel β the analysis becomes markedly harder and often requires the Markov and mixing tools developed in Chapter 13.
Independence is also the most commonly over-assumed property in engineering. Verifying that a model truly has independent components, rather than merely treating them as independent for mathematical convenience, is a critical modelling skill.
Definition: Independence of Events
Independence of Events
A collection of events is mutually independent (or simply independent) if for every finite subset : Two events and are independent if .
The collection is pairwise independent if every pair satisfies but the higher-order product conditions are not required. Mutual independence implies pairwise independence, but not conversely.
Theorem: Equivalent Characterization of Independence
When , two events and are independent if and only if That is, knowing occurred provides no information about .
Forward implication
If (independent), then . Dividing by :
Reverse implication
If , multiply both sides by :
Theorem: Independence Is Preserved Under Complementation
If and are independent, then so are and , and , and and .
Show $A^c$ and $B$ are independent
We have (since and partition ). Therefore:
Cascading the result
Applying the same argument to the pair shows and are independent. Similarly and are independent.
Example: Pairwise Independence Does Not Imply Mutual Independence
Toss two fair coins. Let , , . Show that , , are pairwise independent but not mutually independent.
Sample space and probabilities
, each with probability . , , (since ).
Pairwise independence
. β . β . β
Failure of mutual independence
. But . The three events are pairwise independent but not mutually independent.
Common Mistake: Pairwise Independence Is NOT Mutual Independence
Mistake:
In simulations and modelling, it is tempting to verify independence only for pairs of events and conclude that all events in the collection are independent. The example above (three two-coin events) shows this is false with three events; analogous constructions exist for any number of events.
Correction:
Mutual independence requires the product rule to hold for every finite subset, not just pairs. For events, there are conditions beyond the pairwise conditions. All must be checked.
Independence: Pairwise vs. Mutual
| Property | Pairwise Independent | Mutually Independent |
|---|---|---|
| Definition | for all | for all finite |
| Implications | Does NOT imply mutual independence | Implies pairwise independence |
| Number of conditions | equations | equations (all subsets of size ) |
| Used in practice | Weaker, easier to verify | Required for most probabilistic analysis |
| Example counterexample | Two fair coins + same-face event | N/A (no gap in reverse direction) |
Independence Checker: vs.
Set the probabilities of three events , , defined on a two-coin experiment and verify whether each pair satisfies the product rule for independence.
Parameters
The i.i.d. Assumption in Shannon Theory
Shannon's channel coding theorem assumes the channel is memoryless: consecutive uses of the channel are statistically independent. Under this assumption the capacity per channel use is a single-letter expression . The i.i.d. source coding theorem similarly requires independent symbols. These independence assumptions are the reason capacity results look so clean.
In practice, wireless channels are NOT memoryless: multipath creates frequency-selective fading (correlated across subcarriers) and Doppler creates time-selective fading (correlated across symbols). Engineers work around this via interleaving (reordering symbols to break correlation before decoding) and OFDM (converting a frequency-selective channel into many parallel flat-fading sub-channels, each approximately memoryless).
- β’
LTE/5G NR use OFDM with cyclic prefix to create approximately i.i.d. sub-channel model
- β’
Interleaver depth must exceed the coherence time to achieve near-independence
- β’
When coherence bandwidth channel bandwidth, frequency diversity approaches the i.i.d. bound
Independent Events
Events are mutually independent if for every finite subset . Intuitively, knowledge of any subset of events provides no information about the remaining events.
Related: Pairwise Independence, Conditional Independence, Discrete-Time i.i.d. Gaussian Noise
Quick Check
Two events and both have positive probability and are disjoint (). Are they independent?
Yes, because they have no overlap.
No, because .
Only if .
Impossible to determine without more information.
Independence requires . But disjointness gives , while by assumption. So disjoint events with positive probability are DEPENDENT: knowing occurred tells you cannot have occurred.
Key Takeaway
Independence means no information flows between events. iff : observing leaves the probability of unchanged. Disjointness is the opposite extreme β the most dependent possible relationship. Mutual independence is strictly stronger than pairwise independence and requires conditions for events. In information theory, independence is the assumption that makes entropy additive: .