Chapter Summary
Chapter Summary
Key Points
- 1.
The Lebesgue integral, which partitions the range of a function instead of the domain, generalizes the Riemann integral and provides the rigorous foundation for expectation of arbitrary random variables — discrete, continuous, mixed, or singular.
- 2.
A -algebra specifies which subsets of are "measurable," a measure assigns non-negative values to them, and a measurable function (random variable) is one whose preimages are in .
- 3.
Conditional expectation is a -measurable random variable, not a number. It is the best approximation of using only the information in , and it satisfies the tower property, linearity, Jensen's inequality, and the "taking out what is known" rule.
- 4.
Martingales are adapted sequences where — modeling fair games. The martingale convergence theorem and the optional stopping theorem are powerful tools for proving convergence results and computing stopping-time expectations.
- 5.
The Radon-Nikodym theorem states that implies the existence of a density . This unifies PDFs (), PMFs (), and likelihood ratios () under a single framework, extending hypothesis testing and estimation to settings without classical densities.
Looking Ahead
The measure-theoretic tools in this chapter form the language of advanced probability and information theory. The conditional expectation framework is the foundation for martingale methods in filtering (Kalman filter, particle filters), the Radon-Nikodym theorem underpins detection theory for continuous-time processes (Book FSI), and the convergence theorems (MCT, DCT) are used throughout information theory (Book ITA) whenever we exchange limits and expectations. These are the tools that separate engineering intuition from mathematical proof.