Prerequisites & Notation

Before You Begin

This chapter assumes familiarity with discrete entropy and its properties from Chapter 1. The method of types (Chapter 4) provides the error exponent perspective but is not required for the main results.

  • Discrete entropy H(X)H(X) and its properties (Ch 1)(Review ch01)

    Self-check: Can you prove that H(X)≀log⁑∣X∣H(X) \leq \log |\mathcal{X}| with equality iff XX is uniform?

  • KL divergence D(Pβˆ₯Q)D(P \| Q) and the information inequality (Ch 1)(Review ch01)

    Self-check: Can you state and prove D(Pβˆ₯Q)β‰₯0D(P \| Q) \geq 0?

  • Conditional entropy and chain rule (Ch 1)(Review ch01)

    Self-check: Can you write H(X1,…,Xn)=βˆ‘i=1nH(Xi∣Xiβˆ’1,…,X1)H(X_1, \ldots, X_n) = \sum_{i=1}^n H(X_i | X_{i-1}, \ldots, X_1)?

  • Basic probability: expectation, Jensen's inequality

    Self-check: Can you apply Jensen's inequality to prove that E[log⁑f(X)]≀log⁑E[f(X)]\mathbb{E}[\log f(X)] \leq \log \mathbb{E}[f(X)] for concave log⁑\log?

Notation for This Chapter

Symbols introduced in this chapter.

SymbolMeaningIntroduced
X\mathcal{X}Source alphabet (finite, ∣X∣=m|\mathcal{X}| = m)s01
D\mathcal{D}Code alphabet (typically {0,1}\{0, 1\} for binary codes)s01
c:Xβ†’Dβˆ—c : \mathcal{X} \to \mathcal{D}^*Source code: maps each source symbol to a codewords01
β„“(x)\ell(x) or β„“i\ell_iCodeword length for symbol xx (or xix_i)s01
L=E[β„“(X)]L = \mathbb{E}[\ell(X)]Expected codeword lengths01
Lβˆ—L^*Minimum achievable expected codeword lengths02
H∞H_\infty or H(X)H(\mathbb{X})Entropy rate of a stationary process {Xi}\{X_i\}s05
Ο€\piStationary distribution of a Markov chains05
c(x)c(\mathbf{x})Block code: maps length-nn sequences to binary stringss01