Independence of Random Variables
Why Independence Is the Central Assumption
Independence is the single most important structural assumption in probability and its applications. The entire machinery of information theory β i.i.d. sources, memoryless channels, random codebooks β is built on independence. When it holds, computations simplify dramatically: joint distributions factor, expectations of products equal products of expectations, and the variance of a sum equals the sum of variances. When it fails, the analysis becomes harder, but the deviations from independence are often the most interesting part of the problem.
Definition: Independence of Random Variables
Independence of Random Variables
Random variables and are independent if for all :
Equivalently:
- Discrete case: for all .
- Continuous case: for all .
A collection is (mutually) independent if the joint CDF (or PMF, or PDF) factors into a product of marginals for every subset of the collection.
Pairwise independence does not imply mutual independence. The same subtlety we encountered for events in Chapter 2 carries over to random variables.
Theorem: Functions of Independent RVs Are Independent
If and are independent random variables and are Borel-measurable functions, then and are independent.
CDF factorization
For any :
Since and are Borel sets and are independent, the probability factors:
Theorem: Product of Expectations
If and are independent and , then
More generally, for any measurable with :
Joint factorization
$
Example: Poisson Splitting Property
A coin is tossed times where . Each toss independently lands heads with probability . Let = number of heads and = number of tails. Show that and are independent with and .
Joint PMF calculation
Since and given , :
The conditional term is zero unless , in which case it equals .
Simplify
$
Conclusion
The joint PMF factors into a product of marginals, so and are independent. Each marginal is a Poisson PMF with the indicated parameter. This is the Poisson splitting property β a hallmark result that makes Poisson processes so tractable.
Independent random variables
Random variables whose joint CDF (or PMF/PDF) factors as a product of marginals. Independence means knowing the value of one provides no information about the other.
Related: Joint probability density function, Conditional expectation
Quick Check
If for and zero otherwise, are and independent?
Yes, because the joint PDF factors as .
No, because they are both exponential.
Cannot determine without computing the CDF.
The joint PDF factors as where each marginal is .
Historical Note: The Formalization of Independence
1933The concept of independence has been used informally since the earliest work on games of chance. But its rigorous mathematical definition β as the factorization of a joint distribution β was established by Kolmogorov in his 1933 monograph Grundbegriffe der Wahrscheinlichkeitsrechnung. Kolmogorov's formalization made it possible to state precisely when two random variables "have nothing to do with each other" and to derive consequences such as the strong law of large numbers and the central limit theorem.
Key Takeaway
Independence of random variables means the joint distribution factors as a product of marginals. It implies uncorrelatedness, but the converse is false except for jointly Gaussian RVs. Independence is the key structural assumption that makes most of information theory and performance analysis tractable.