Exercises
ex-ch29-01
EasyA binary source with is to be communicated for a classification task where (the goal is to recover exactly). What is the minimum rate needed to achieve classification accuracy ?
This is equivalent to a rate-distortion problem with Hamming distortion .
Use the binary rate-distortion function.
Formulate as rate-distortion
Classification accuracy means error probability . The distortion is Hamming: with .
Compute the rate
The binary rate-distortion function is \R(D) = \H_{b}(p) - \H_{b}(D) for . With and : \R(0.05) = \H_{b}(0.3) - \H_{b}(0.05) = 0.881 - 0.286 = 0.595 bits.
ex-ch29-02
EasyVerify that linear JSCC achieves the distortion-rate bound for a Gaussian source over an AWGN channel with dB and .
Compute and compare to .
Compute the linear JSCC distortion
(20 dB). .
Compute the separation bound
bits. . They match, confirming optimality.
ex-ch29-03
EasyExplain why the "cliff effect" occurs in digital communication systems but not in analog/JSCC systems. Illustrate with a concrete example.
Think about what happens when the channel SNR drops below the code rate.
Digital system
A digital system uses a fixed-rate channel code designed for . The code rate is . When the actual , , and the channel code fails completely (BER ). The reconstruction goes from near-perfect to garbage.
Analog/JSCC system
An analog system transmits a scaled version of the source. When SNR drops, the noise increases but the received signal still contains information about the source. The MMSE estimator adapts to the noise level, producing a noisier but still useful reconstruction. Example: at dB, MSE = 0.01. At dB, MSE = 0.5 (still better than pure noise, MSE = 1.0).
ex-ch29-04
EasyFor a -dimensional Gaussian source where a feature extractor selects the top principal components, what is the maximum rate savings of semantic over classical rate-distortion?
Compare active components vs. active components.
Rate comparison
At high rate (low distortion), and . For the same per-component distortion, the ratio is . With , : semantic communication needs of the rate.
ex-ch29-05
EasyShow that for a Gaussian source over AWGN with bandwidth expansion (), repetition coding is suboptimal. What is the optimal scheme?
Repetition coding gives . The optimal scheme exploits all channel uses.
Repetition coding
Transmit for (repeat times). The receiver averages: , giving MSE .
Optimal: separate coding
The separation bound gives where . So D^* = \\sigma_S^2/(1+\\text{SNR})^\\rho. For , (1+\\text{SNR})^\\rho > 1+\\rho\\text{SNR}, so separate coding strictly outperforms repetition. Repetition is optimal only for (matched bandwidth).
ex-ch29-06
MediumFormulate the rate-utility function for a remote inference problem: the source is with , and the goal is to estimate where with . Show that .
The goal is an -dimensional linear function of .
Only the components of in the column space of matter.
Reduce to relevant components
Let A = U_A \\Sigma_A V_A^\\top (SVD). The goal depends on only through V_A^\\top S \\in \\mathbb{R}^m. Define \\tilde{S} = V_A^\\top S \\sim \\mathcal{N}(0, V_A^\\top \\Sigma_S V_A).
Rate-distortion of the relevant components
The rate-utility function satisfies (distortion of is amplified by ). For the Gaussian case: where are eigenvalues of V_A^\\top \\Sigma_S V_A and are singular values of .
Upper bound
Bounding each term: . This shows the rate scales with the task dimension , not the source dimension .
ex-ch29-07
MediumA DeepJSCC system trained at dB is deployed at dB. Estimate the PSNR degradation compared to a system trained at the deployment SNR, using the Gaussian source/AWGN channel analogy.
For the Gaussian case, the MSE scales as . A mismatched system may not achieve the MMSE.
Optimal performance at each SNR
At 10 dB (): , PSNR dB. At 5 dB (): , PSNR dB.
Mismatched system at 5 dB
A system trained at 10 dB has learned a representation optimized for . At dB, the noise is higher than expected. The encoder's implicit power allocation may not be optimal, and the decoder's learned denoiser under-compensates. Empirically, the PSNR degradation is typically 0.5-2 dB compared to a matched system, depending on the architecture and training strategy.
Mitigation
Train with SNR drawn from a range (e.g., dB) and condition the decoder on the estimated SNR. This "SNR-adaptive" DeepJSCC loses dB vs. per-SNR training.
ex-ch29-08
MediumProve that the perception-distortion tradeoff is non-trivial for any non-degenerate source. Specifically, show that for a Gaussian source with MMSE reconstruction , the distribution .
The MMSE estimator is a contraction: .
Variance of the MMSE estimator
For observed through with : .
Compare distributions
for any . Therefore . The MMSE reconstruction is "narrower" than the source — it produces less variability.
Tradeoff
To achieve (perfect perception), we must add noise to to restore the variance: where . But this increases MSE from to . This is the perception-distortion tradeoff.
ex-ch29-09
HardDerive the optimal bandwidth ratio for transmitting a -dimensional Gaussian source over an AWGN channel to minimize end-to-end MSE, when using linear JSCC. Show that (matched bandwidth) and the MSE is , regardless of .
For (compression), linear JSCC must project to a subspace.
For (expansion), repetition is suboptimal but linear JSCC cannot beat the separation bound.
Linear JSCC with arbitrary $\rho$
The encoder is where with and power constraint \\text{tr}(A \\Sigma_S A^\\top) \\leq kP. The decoder is the LMMSE estimator.
MSE calculation
MSE = \\text{tr}(\\Sigma_S) - \\text{tr}(\\Sigma_S A^\\top (A\\Sigma_S A^\\top + \\sigma^2 I_k)^{-1} A \\Sigma_S). For i.i.d. source () and (projection): MSE .
Optimize over $\rho$
For (): MSE . With power constraint , i.e., : MSE . This is minimized at , giving MSE .
ex-ch29-10
HardConsider a semantic communication system for remote classification. The source belongs to one of classes with equal probability, and the receiver must determine the class. The channel is AWGN with and bandwidth ratio . What is the minimum to achieve classification error ?
The task requires transmitting bits reliably.
Use the channel coding theorem with rate .
Information requirement
The classification task requires bits of information. The encoder maps the source to channel symbols. The channel can carry bits reliably, where .
Minimum bandwidth ratio
For reliable transmission: . Therefore . For , (CIFAR), dB: .
Comparison with reconstruction
Reconstruction at PSNR 30 dB requires - (depending on the source). The semantic is smaller — this is the semantic communication gain. At error probability , add for the finite-blocklength correction.
ex-ch29-11
HardProve that for a block-fading channel where the fading coefficient is constant within a block of symbols, analog JSCC (uncoded transmission) achieves a strictly lower expected distortion than any fixed-rate digital scheme, when is finite.
The digital scheme fails (outage) when . The analog scheme always produces a useful estimate.
Digital scheme
A digital scheme at rate succeeds when , i.e., . In outage (probability ), distortion . In success, distortion . Expected distortion: .
Analog scheme
The analog scheme transmits , receiving . MMSE decoding gives distortion . Expected: .
Comparison
For Rayleigh fading, where is the exponential integral. The digital distortion has a discontinuity (outage) while the analog distortion is smooth. For any fixed rate , there exists a range of SNR where , because the analog scheme avoids the catastrophic outage penalty.
ex-ch29-12
MediumThe FID between two Gaussian distributions and is given by . Compute the FID between a source and the MMSE reconstruction distribution where .
Both distributions are zero-mean with proportional covariances.
Compute the FID
, so . , . . .
Substitute $D = \sigma^2/(1+\ntn{snr})$
. . At high SNR: — it decreases as . The FID quantifies the perception gap: even the optimal MMSE estimator produces reconstructions with lower variance than the source.
ex-ch29-13
ChallengeDesign a hybrid semantic communication scheme with a "base layer" (rate , sufficient for any task) and a "semantic layer" (rate , optimized for a specific classification task). For a Gaussian source with dimensions and task dimension , find the optimal rate allocation subject to total rate that minimizes weighted distortion .
The base layer encodes all dimensions at low fidelity. The semantic layer refines the task-relevant dimensions.
This is a successive refinement problem.
Base layer
The base layer encodes at rate , achieving MSE by reverse water-filling. With i.i.d. source: .
Semantic layer
The semantic layer encodes the residual of the task-relevant components at rate . The residual variance per component is (what the base layer did not capture). After semantic refinement: .
Optimize
Minimize subject to . Setting the derivative to zero gives the optimal allocation. For (task-focused): allocate most rate to the task dimensions, . For (MSE-focused): , .
ex-ch29-14
MediumShow that the rate-utility function is convex and non-increasing in .
Use a time-sharing argument between two achievable pairs and .
Monotonicity
is the minimum rate for utility . If , the constraint set for is a subset of the constraint set for , so . Therefore is non-increasing. Wait — actually, higher utility requires more rate (more information about ), so should be non-decreasing. Let me reconsider. Since we maximize utility, higher requires more rate: for . So is non-decreasing (like with replaced by ).
Convexity
Let be optimal mappings achieving and . Time-sharing: use with probability and with probability . Rate: (MI is convex in for mixture). Utility: . Therefore .
ex-ch29-15
ChallengeConsider a multi-task semantic communication system that must simultaneously serve tasks with utility functions . Formulate the multi-objective rate-utility region and show that it is convex. Discuss the implications for system design when tasks share a single encoder.
The utility region is the set of achievable tuples at a given rate .
Use a time-sharing argument.
Define the utility region
At rate , the utility region is: \\mathcal{U}(R) = \\{(u_1, \\ldots, u_L) : \\exists P_{\\hat{S}|S}, \I(S;\\hat{S}) \\leq R, \\mathbb{E}[U_\\ell(\\hat{S}, G_\\ell)] \\geq u_\\ell \\forall \\ell\\}.
Convexity via time-sharing
If and , the convex combination is achieved by time-sharing the two encoding schemes. Rate: . Utility: \\lambda u_\\ell^{(a)} + (1-\\lambda)u_\\ell^{(b)}. So is convex.
System design implications
A shared encoder must operate on the Pareto boundary of . With tasks, the optimal representation is a compromise — it preserves information about all goals . As (many diverse tasks), the optimal encoder approaches the Shannon encoder that preserves all information in , recovering the classical separation architecture. This is the formal version of the universality-efficiency tradeoff.