Chapter Summary
Chapter 10 Summary: Probability Inequalities
Key Points
- 1.
Markov's inequality () is the simplest tail bound, using only the first moment of a non-negative RV. Every other inequality in this chapter is a corollary.
- 2.
Chebyshev's inequality () follows from applying Markov to . It uses two moments but is still loose for light-tailed distributions.
- 3.
The Chernoff bound () leverages the full MGF via the exponential tilt trick. It is exponentially tight and captures the correct decay rate for most practical distributions.
- 4.
Hoeffding's inequality provides exponential concentration for sums of bounded independent RVs. It is the workhorse for non-asymptotic sample complexity bounds in statistics and learning theory.
- 5.
Jensen's inequality ( for convex ) connects convexity with expectation. It implies in information theory and for fading channels.
- 6.
The hierarchy of tightness: Markov (loosest, ) Chebyshev () Chernoff/Hoeffding (exponential in or ). Choose the bound that matches the available information.
Looking Ahead
Chapter 11 builds on these inequalities to establish the different modes of convergence of random variables: almost sure, in probability, in , and in distribution. Chebyshev gives the simplest proof of the Weak Law of Large Numbers, while the Chernoff/Hoeffding bounds connect to the Strong Law and large deviations theory.