The Weak Law of Large Numbers
Why the Law of Large Numbers Matters
The sample mean is the most basic statistical estimator. Every Monte Carlo simulation, every sample average in a communication receiver, every training loss in machine learning relies on the principle that averaging many independent copies of a random quantity produces something close to the true mean. The Weak Law of Large Numbers makes this precise: converges to in probability. The proof is a clean application of Chebyshev's inequality β and the simplicity of the argument is part of its beauty.
Theorem: Weak Law of Large Numbers (WLLN)
Let be i.i.d. random variables with mean and finite variance . Then the sample mean converges to in probability:
that is, for every :
The variance of is , which shrinks to zero. Chebyshev's inequality translates vanishing variance into vanishing tail probability. The more samples we average, the tighter the distribution of concentrates around .
Compute the variance of the sample mean
Since the are i.i.d.:
Apply Chebyshev's inequality
By Chebyshev: for any ,
Take the limit
As :
By the squeeze theorem, .
Alternative Proof via Characteristic Functions
The WLLN can also be proved using characteristic functions, as Caire does in the course. The Ch.F of is . Taylor-expanding and using the limit , we get , which is the Ch.F of the constant . By the Levy continuity theorem, , and since the limit is a constant, convergence in distribution upgrades to convergence in probability.
This proof only requires (no finite variance needed), so it is strictly more general than the Chebyshev proof above.
Example: Empirical Frequency of a Biased Coin
Let be i.i.d. with . How many coin tosses are needed so that the empirical frequency is within of the true probability with probability at least ?
Set up the Chebyshev bound
We need . By Chebyshev:
Solve for $n$
We need , giving .
This is a conservative bound β Chebyshev is loose. The CLT (Section 11.4) will give the much tighter estimate .
WLLN in Action: Sample Mean Trajectories
Watch multiple independent trajectories of converge to as grows. Choose the underlying distribution and observe how the convergence rate depends on the variance.
Parameters
Bernoulli: p; Exponential: lambda; Uniform: b (on [0,b]); Gaussian: sigma
Why This Matters: Monte Carlo Simulation in Communications
Every bit error rate (BER) simulation in wireless communications is a direct application of the WLLN. We transmit symbols through a simulated channel, count the errors , and report as the estimated error probability. The WLLN guarantees as .
The practical question is: how large must be? For (a typical target in 5G), the Chebyshev bound suggests , but the CLT (Section 11.4) gives the more useful rule of thumb: we need roughly errors, so symbol transmissions.
See full treatment in The Linear MMSE Estimator
Confidence Intervals for Monte Carlo BER Estimates
The Chebyshev-based WLLN bound is overly conservative for practical Monte Carlo design. In practice, we use the CLT approximation:
where for a 95% confidence interval. For and a relative accuracy of 10%, this requires β feasible but not cheap. This is why importance sampling and other variance reduction techniques are essential in practice.
- β’
For , direct Monte Carlo becomes impractical (> samples)
- β’
Importance sampling can reduce the required sample count by orders of magnitude
Quick Check
The Chebyshev-based proof of the WLLN requires which condition on the i.i.d. sequence?
Finite mean only
Finite variance
Finite fourth moment
The distribution must be continuous
Correct. Chebyshev's inequality uses , which requires .
Weak Law of Large Numbers
States that for i.i.d. with finite mean . "Weak" refers to convergence in probability, as opposed to the "strong" law which gives almost sure convergence.
Related: Convergence in Probability, Strong Law of Large Numbers
Common Mistake: Chebyshev's Bound Is Loose β Do Not Use It for Design
Mistake:
Using the WLLN's Chebyshev bound to determine the required sample size in a real system.
Correction:
The Chebyshev bound is distribution-free and therefore very conservative. For system design, use the CLT normal approximation (Section 11.4) or, for small error probabilities, the Chernoff/Hoeffding exponential bounds (FSP Ch. 9). The Chebyshev bound is a proof tool, not a design tool.