Exercises

ex-ch04-01

Easy

Let X={a,b,c}\mathcal{X} = \{a, b, c\} and n=3n = 3. List all types in P3(X)\mathcal{P}_3(\mathcal{X}), compute TQ|T_Q| for each type QQ, and verify that QTQ=33=27\sum_Q |T_Q| = 3^3 = 27.

ex-ch04-02

Easy

Show that for a binary alphabet X={0,1}\mathcal{X} = \{0, 1\} and the uniform distribution P=(1/2,1/2)P = (1/2, 1/2), the type class TQT_Q with Q=(k/n,1k/n)Q = (k/n, 1 - k/n) has Pn(TQ)=(nk)2n.P^n(T_Q) = \binom{n}{k} 2^{-n}. Verify that k=0nPn(TQk)=1\sum_{k=0}^n P^n(T_{Q_k}) = 1.

ex-ch04-03

Easy

Let P=(0.7,0.3)P = (0.7, 0.3) on X={0,1}\mathcal{X} = \{0, 1\}. Compute D(QP)D(Q \| P) for Q=(0.5,0.5)Q = (0.5, 0.5) and interpret the result using the type class probability formula.

ex-ch04-04

Easy

For the set E={QP({0,1}):Q(1)0.8}\mathcal{E} = \{Q \in \mathcal{P}(\{0,1\}) : Q(1) \geq 0.8\} and true distribution P=(0.5,0.5)P = (0.5, 0.5), find D(EP)D^*(\mathcal{E} \| P) and state the asymptotic probability Pn(P^XE)P^n(\hat{P}_{\mathbf{X}} \in \mathcal{E}).

ex-ch04-05

Medium

Prove that D(QP)=0D(Q \| P) = 0 if and only if Q=PQ = P, using only the type class probability formula Pn(TQ)2nD(QP)P^n(T_Q) \doteq 2^{-nD(Q \| P)} and the fact that Pn(TP)P^n(T_P) is the maximum over all type class probabilities.

ex-ch04-06

Medium

Let X1,,XnX_1, \ldots, X_n be i.i.d. Bernoulli(pp) with p=0.2p = 0.2. Use Sanov's theorem to find the exponential rate at which Pr[P^X(1)0.5]\Pr[\hat{P}_{\mathbf{X}}(1) \geq 0.5] decays. Compare with the Chernoff bound.

ex-ch04-07

Medium

Show that the source coding error exponent Es(R)E_s(R) is a convex function of RR for R>H(P)R > H(P), and that Es(R)D(UP)E_s(R) \to D(U \| P) as RlogXR \to \log |\mathcal{X}|, where UU is the uniform distribution on X\mathcal{X}.

ex-ch04-08

Medium

For a BSC with crossover probability pp, the random coding exponent with the optimal (uniform) input distribution is Er(R)=max0ρ1[E0(ρ)ρR]E_r(R) = \max_{0 \leq \rho \leq 1} \left[E_0(\rho) - \rho R\right] where E0(ρ)=1(1+ρ)log ⁣[p1/(1+ρ)+(1p)1/(1+ρ)]E_0(\rho) = 1 - (1+\rho)\log\!\left[p^{1/(1+\rho)} + (1-p)^{1/(1+\rho)}\right] (in bits). Show that Er(C)=0E_r(C) = 0 and Er(0)=E0(1)E_r(0) = E_0(1).

ex-ch04-09

Medium

Consider a ternary symmetric channel with X=Y={0,1,2}\mathcal{X} = \mathcal{Y} = \{0, 1, 2\} and transition probabilities W(yx)=12ϵW(y|x) = 1-2\epsilon if y=xy = x and W(yx)=ϵW(y|x) = \epsilon otherwise. Find the channel capacity and the zero-rate exponent Er(0)=E0(1)E_r(0) = E_0(1) for ϵ=0.1\epsilon = 0.1.

ex-ch04-10

Medium

Prove that for any DMC, Er(R)Esp(R)E_r(R) \geq E_{sp}(R) cannot hold — i.e., the sphere-packing exponent is always an upper bound: Er(R)Esp(R)E_r(R) \leq E_{sp}(R). (You do not need to prove the sphere-packing bound; just explain why the achievability exponent cannot exceed the converse exponent.)

ex-ch04-11

Hard

Derive the source coding error exponent for a binary source with P(0)=pP(0) = p, P(1)=1pP(1) = 1-p in closed form. Show that: Es(R)=D(QP)E_s(R) = D(Q^* \| P) where Q=(q,1q)Q^* = (q^*, 1-q^*) with H(q)=RH(q^*) = R and qp<1qp|q^* - p| < |1-q^* - p| (i.e., qq^* is the root of H(q)=RH(q) = R on the same side as pp).

ex-ch04-12

Hard

(Conditional type class size.) Let X=Y={0,1}\mathcal{X} = \mathcal{Y} = \{0, 1\}, n=6n = 6, and x=(0,0,0,1,1,1)\mathbf{x} = (0, 0, 0, 1, 1, 1) (type Q=(1/2,1/2)Q = (1/2, 1/2)). A conditional type VV is defined by V(00)=2/3V(0|0) = 2/3, V(10)=1/3V(1|0) = 1/3, V(01)=1/3V(0|1) = 1/3, V(11)=2/3V(1|1) = 2/3. Enumerate all yTVx\mathbf{y} \in T_{V|\mathbf{x}} and verify TVx2nH(VQ)|T_{V|\mathbf{x}}| \leq 2^{nH(V|Q)}.

ex-ch04-13

Hard

(Sanov for hypothesis testing.) Consider testing H0:P=(0.5,0.5)H_0: P = (0.5, 0.5) vs H1:P=(0.3,0.7)H_1: P = (0.3, 0.7) on binary sequences of length nn. A type-based test rejects H0H_0 if D(P^XP0)>γD(\hat{P}_\mathbf{X} \| P_0) > \gamma.

(a) Find the type-I error exponent as a function of γ\gamma. (b) Find the type-II error exponent as a function of γ\gamma. (c) Find the optimal γ\gamma that maximizes the type-II error exponent subject to the type-I error exponent being at least α0=0.1\alpha_0 = 0.1.

ex-ch04-14

Hard

(Method of types proof of the channel coding theorem.) Use the method of types to prove that for a DMC with capacity C>0C > 0, any rate R<CR < C is achievable. Specifically, show that a random codebook with 2nR2^{nR} codewords drawn i.i.d. from the capacity-achieving distribution QQ^* has average error probability Pe0\overline{P}_e \to 0 as nn \to \infty.

ex-ch04-15

Hard

(I-projection onto an exponential family.) Let PP be a distribution on X={1,,m}\mathcal{X} = \{1, \ldots, m\} and E={Q:EQ[X]μ}\mathcal{E} = \{Q : \mathbb{E}_Q[X] \geq \mu\} for some μ>EP[X]\mu > \mathbb{E}_P[X]. Show that the I-projection Q=argminQED(QP)Q^* = \arg\min_{Q \in \mathcal{E}} D(Q \| P) is an exponential tilting of PP: Q(x)=P(x)eλxxP(x)eλxQ^*(x) = \frac{P(x) e^{\lambda^* x}}{\sum_{x'} P(x') e^{\lambda^* x'}} for some λ0\lambda^* \geq 0.

ex-ch04-16

Challenge

(Gap between ErE_r and EspE_{sp} below the critical rate.) For the BEC with erasure probability ϵ\epsilon, compute Er(R)E_r(R) and Esp(R)E_{sp}(R) in closed form. Show that: (a) The critical rate is Rcr=0R_{cr} = 0 (i.e., the random coding exponent is tight for all positive rates). (b) Therefore Er(R)=Esp(R)E_r(R) = E_{sp}(R) for all 0<R<C=1ϵ0 < R < C = 1 - \epsilon. This makes the BEC one of the few channels where the reliability function is known exactly.

ex-ch04-17

Challenge

(Csiszár–Körner proof of the converse for channel coding.) Using the method of types, prove that for a DMC, any sequence of codes with rate R>CR > C has Pe(n)1P_e^{(n)} \to 1. Specifically, show that for any codebook {x1,,xM}\{\mathbf{x}_1, \ldots, \mathbf{x}_M\} with M>2nCM > 2^{nC}: Pe(n)12nCMϵnP_e^{(n)} \geq 1 - \frac{2^{nC}}{M} - \epsilon_n where ϵn0\epsilon_n \to 0.

Hint: This does not require Fano's inequality — only the type counting arguments from this chapter.

ex-ch04-18

Challenge

(Duality between source and channel exponents.) The source coding exponent at rate RR is Es(R)=minQ:H(Q)RD(QP)E_s(R) = \min_{Q : H(Q) \geq R} D(Q \| P), and the simple channel coding exponent (at zero rate, for maximum likelihood decoding with two codewords) is Ec=minVD(VWQ)E_c = \min_V D(V \| W | Q) subject to a confusion constraint.

Show that both exponents can be expressed as I-projections on the probability simplex, and explain the geometric duality: source coding projects onto a superlevel set of entropy, while channel coding projects onto a sublevel set of mutual information.