Exercises
ex-ch02-01
EasyCompute the differential entropy of for .
The PDF is on .
Direct computation
.
For : . For : .
ex-ch02-02
EasyShow that for any constant (translation invariance).
The PDF of is .
Proof
. Substituting : .
ex-ch02-03
EasyCompute for .
Differential entropy depends on the variance, not the mean.
Apply the formula
bits.
Note: for any .
ex-ch02-04
EasyLet and . Compute using the scaling property.
.
Apply scaling and translation
.
Alternatively: , so . Consistent.
ex-ch02-05
MediumProve that for continuous random variables is invariant under invertible transformations: if and where are invertible, then .
Write as a KL divergence.
KL divergence between continuous distributions is invariant under invertible maps.
Via KL divergence
.
Under the invertible map , both the joint density and the product of marginals transform with the same Jacobian factor , which cancels in the ratio:
.
Therefore .
ex-ch02-06
MediumShow that for the exponential distribution with parameter : (a) , and (b) among all non-negative distributions with mean , the exponential uniquely maximizes .
For (b), use the same KL divergence technique as the Gaussian case.
The cross-entropy term depends on only through .
(a) Compute
.
(b) Maximum entropy proof
Let be any PDF on with , and let .
.
.
Therefore .
ex-ch02-07
MediumCompute the mutual information for the AWGN channel where and are independent.
.
.
Compute
, so .
.
.
ex-ch02-08
MediumProve that for a random vector : for any invertible matrix .
Use the change-of-variables formula for densities.
Change of variables
Let . The PDF transforms as .
.
ex-ch02-09
MediumShow that for the Gaussian vector , the differential entropy can be written as , where are the eigenvalues of .
.
Apply the formula .
Eigenvalue decomposition
where are the components in the eigenbasis. The total entropy is the sum of entropies along independent principal components.
ex-ch02-10
HardDerive the capacity of the parallel Gaussian channel: for , where are independent, and the total power constraint is .
The capacity is the sum of individual capacities, optimized over power allocation.
Use Lagrange multipliers — the constraint is linear in the powers .
The resulting power allocation is waterfilling.
Sum of capacities
Since the channels are independent: .
Waterfilling
Maximize subject to and .
The Lagrangian is .
KKT conditions give: where is chosen so that .
Capacity
.
The waterfilling interpretation: fill water to level over a "landscape" with heights . Channels with noise above the water level get zero power — they are too noisy to be useful.
ex-ch02-11
HardProve that for independent with :
This follows from .
Alternatively, use the EPI.
Via mutual information
.
Therefore .
Interpretation
Adding an independent signal to Gaussian noise can only increase the total entropy — the signal "spreads" the distribution further.
ex-ch02-12
HardThe de Bruijn identity states that if is independent of :
where is the Fisher information. Verify this for (deterministic) and for .
For : .
Fisher information of is .
Case $X = 0$
, so and .
, so . In nats: . Verified.
Case $X \sim \ntn{gauss}(0, \sigma^2)$
.
.
.
. Converting to bits: matches. Verified.
ex-ch02-13
HardProve the maximum entropy under covariance constraint for complex random vectors: among all distributions on with covariance matrix , the circularly symmetric complex Gaussian uniquely maximizes differential entropy, achieving .
Follow the same KL divergence proof as the real case.
Complex Gaussian PDF: .
KL divergence approach
Let be the PDF of . For any with the same covariance:
.
The cross-entropy uses .
Therefore .
ex-ch02-14
MediumLet and be independent. Compute (the conditional differential entropy of the input given the noisy output).
is Gaussian with known mean and variance (MMSE estimation).
The MMSE estimate is with error variance .
Conditional distribution
.
The conditional variance does not depend on .
Conditional entropy
.
Check: . Consistent.
ex-ch02-15
Challenge(Costa's EPI strengthening) Show that for independent and , the function is concave in . Deduce the EPI as a special case.
Use the de Bruijn identity: .
Show using the Fisher information inequality.
Sketch
Define . Using de Bruijn's identity:
, where is in nats.
, so .
.
One shows using the Cramér-Rao-type bound where is the entropy power of . This gives , proving concavity.
EPI as corollary
Concavity of gives: is not the right bound. Instead, The EPI follows from concavity along with a scaling argument.