Chapter Summary
Chapter Summary
Key Points
- 1.
The decision problem. A binary hypothesis test is specified by two densities on an observation space and a decision rule . Performance is captured by two scalars --- false-alarm and detection (or equivalently miss ) --- that trade off through the choice of decision region.
- 2.
Bayes-optimal rule. With priors and costs , the rule minimising Bayes risk is the LRT with threshold . For 0-1 costs this is the MAP rule: decide the hypothesis with the larger posterior.
- 3.
The likelihood ratio is sufficient. captures all information in relevant to the decision. Every Bayes-optimal rule, every Neyman-Pearson rule, every ML rule is a threshold test on (or its monotone transform, the LLR). Compute in log-domain for numerical stability.
- 4.
Neyman-Pearson lemma. Without priors, fix and maximise : the solution is the (possibly randomised) LRT with threshold chosen to make . Proved via a variational argument that every other rule wastes budget.
- 5.
ROC geometry. Sweeping the LRT threshold traces the ROC curve in space. The ROC is monotone, lies above the diagonal, is concave, and has slope equal to the LRT threshold at each operating point. AUC summarises separability in a single scalar.
- 6.
Bhattacharyya and Chernoff bounds. The MAP error satisfies (Bhattacharyya) and more generally for every (Chernoff), where is concave and vanishes at the endpoints.
- 7.
Chernoff information is the error exponent. is the exponential rate at which decays with the number of i.i.d. observations. For equal-variance Gaussians, with the normalised mean separation, and the Chernoff-optimal tilt is .
- 8.
Connection to coding and information theory. The same exponent structure drives random-coding error bounds (ITA Ch. 4) and CFAR radar detection. The LLR representation is the universal interchange format used in LDPC/turbo decoders (CC book) and in message passing (FSI Part V).
Looking Ahead
Chapter 2 applies these ideas to detection of known signals in additive Gaussian noise, deriving the matched filter as the LRT test statistic and obtaining the celebrated BPSK error probability . Chapter 3 introduces composite hypothesis testing via the GLRT, relaxing the assumption of fully-known densities. Chapter 4 extends to colored Gaussian noise via prewhitening, and to continuous-time detection via the signal-space viewpoint. The likelihood ratio we have introduced here will reappear in every subsequent chapter --- as a message in a factor graph, as a posterior in Bayesian estimation, as a pairwise comparison in coded-system error analysis.