Exercises
ex-ch20-01
EasyCompute the cumulant generating function and rate function for , i.e., for .
for .
Take the log and compute the Legendre transform.
CGF
for .
Rate function
. Setting the derivative to zero: . Substituting: for . Note , confirming the rate function vanishes at the mean.
ex-ch20-02
EasyCompute the rate function for .
.
CGF
, defined for all .
Legendre transform
. Setting derivative to zero: , so . for , with .
ex-ch20-03
MediumLet be i.i.d. . Using Cramér's theorem, compute the exact rate function and determine the exponential rate of .
The rate function for Bernoulli() is .
Apply the Bernoulli rate function
With :
Compute
Using natural log: nats. So .
ex-ch20-04
MediumShow that is convex and satisfies with where .
The Legendre transform of a convex function is convex.
Use to show non-negativity.
Convexity
is a supremum of affine (hence convex) functions of , so is convex.
Non-negativity
.
$I(\mu) = 0$
At : . Since is convex and , the supremum is attained at : .
ex-ch20-05
MediumProve that if is -sub-Gaussian, then for all .
Apply the Chernoff technique to the sub-Gaussian MGF bound.
Chernoff bound
For any : .
Optimize
Minimize over : . Substituting: .
ex-ch20-06
MediumShow that if is -sub-Gaussian and is -sub-Gaussian, and are independent, then is -sub-Gaussian.
Use independence to factor the MGF.
Factor MGFs
.
Conclude
is -sub-Gaussian. Sub-Gaussian parameters add (like variances), and the effective parameter is the square root of the sum.
ex-ch20-07
MediumLet . Show that is sub-exponential but not sub-Gaussian.
for . This blows up at .
MGF of $X^2$
for .
Not sub-Gaussian
A sub-Gaussian MGF bound is finite for all , but for . So is not sub-Gaussian.
Sub-exponential
For : (by for ). So is sub-exponential with parameters .
ex-ch20-08
HardProve Hoeffding's lemma: if and , then .
Use convexity of : bound by the chord between and .
After taking expectation and using , show the bound reduces to where .
Convexity bound
For : . Taking expectation with : .
Reparametrize
Let , . Then where .
Bound $g(u)$
, (using the reparametrization), and . The last step uses . By Taylor: .
ex-ch20-09
HardLet be i.i.d. with . Compare the upper bounds on from: (a) Markov's inequality, (b) Chebyshev's inequality, (c) Hoeffding's inequality, (d) the exact Cramér rate function. Evaluate numerically for .
Markov: .
Chebyshev: .
Markov
. (Useless.)
Chebyshev
.
Hoeffding
.
Exact Cramér
nats. . The Cramér bound is about 3.7x tighter than Hoeffding in the exponent.
ex-ch20-10
HardProve Sanov's theorem for the binary alphabet with and where .
Use the type probability bound: .
Sum over all types in and use the polynomial-in- bound on the number of types.
Upper bound
.
Lower bound
where . Using the lower type probability bound: in the exponent.
Conclude
Both bounds give rate .
ex-ch20-11
MediumVerify Stein's lemma for the specific case and with . Show the optimal Type II error exponent (with Type I error constrained) is .
The LLR is .
Compute KL divergence
. Wait — more carefully: , so .
Verify with NP test
The NP test rejects when . Under , . . Choosing (to fix ): for large , confirming the exponent .
ex-ch20-12
HardCompute the Chernoff information for and with .
.
Chernoff exponent
.
Optimize
Take the derivative with respect to and set to zero. This yields a transcendental equation that generally requires numerical solution. For , : numerical optimization gives and nats.
ex-ch20-13
Challenge(Gärtner-Ellis) Let be a stationary ergodic Markov chain on with transition matrix . Show that the limiting CGF exists (where ) and compute it.
The MGF of can be expressed using matrix products involving the tilted transition matrix.
The tilted matrix is .
Tilted transition matrix
Define with entries : .
Spectral radius
where is the spectral radius (largest eigenvalue) of . Therefore: .
Compute eigenvalue
The characteristic polynomial of is . The largest root gives in closed form.
ex-ch20-14
MediumUsing the matrix Bernstein inequality, determine how many pilot transmissions are needed so that with probability , where , , and the antenna dimension is .
The summands are zero-mean with (requires truncation or sub-exponential argument).
The matrix variance is .
Setup
Let . For Gaussian vectors, is sub-exponential, so with truncation at level , the bounded matrix Bernstein inequality applies.
Apply bound
. Setting and requiring the RHS : suffices (with depending on the truncation level).
ex-ch20-15
EasyShow that a Rademacher random variable () is 1-sub-Gaussian.
Compute directly.
Compute MGF
.
Bound by Gaussian MGF
Using (since ). Therefore , confirming is 1-sub-Gaussian.
ex-ch20-16
Hard(Cramér's theorem - lower bound via tilting) Let be i.i.d. with finite CGF on , and let . Using the exponential tilting measure where , prove the lower bound .
Under , has mean and finite variance .
Use the CLT under the tilted measure to show .
Change of measure
.
Restrict to interval
.
CLT under tilted measure
Under , has mean and variance . So . Taking , the CLT term is , and letting : .
ex-ch20-17
Challenge(Connection to hypothesis testing) Using Cramér's theorem, derive Stein's lemma for the special case of testing vs with . Show that the optimal Type II error exponent is .
The NP test thresholds at some level .
Under , the event is a large deviation.
NP test structure
The LLR is . This is a monotone function of (decreasing since ), so the NP test reduces to for some threshold .
Type I constraint determines $\gamma$
. For , this goes to zero. The exact value of affects only the sub-exponential term.
Type II error via Cramér
. Under , has mean and we need . By Cramér's theorem with the Bernoulli rate function: where . As (optimal threshold): .
ex-ch20-18
EasyShow that with equality iff , using Jensen's inequality.
Write and apply Jensen's to the concave .
Apply Jensen's
.
Equality condition
Equality in Jensen's holds iff is constant a.s. under , i.e., .