Exercises
ex-ch04-01
EasyState the alignment condition for the -user finite-field interference channel, and explain in one sentence why it enables total DoF.
The condition is a matrix equation involving , , .
Condition
for every , while has full column rank.
Why $K/2$
All unintended transmissions at each receiver align into a common -dimensional subspace (the nullspace of ), freeing the other half of the signal space for the intended message.
ex-ch04-02
EasyFor coded matrix multiplication with partitions, compute the recovery threshold and compare to the "naive" replication scheme ( if all data is replicated on every worker).
Recovery threshold for polynomial codes: .
Polynomial-code threshold
.
Replication comparison
Full replication would have and (one worker's response suffices). At , polynomial codes achieve β better straggler tolerance at lower storage.
ex-ch04-03
EasyCompute the global caching gain for users, files, cache size . What is the coded delivery rate?
Gain ; rate .
Gain
. Coded caching is more efficient than uncoded.
Coded rate
.
ex-ch04-04
EasyUsing the classical PIR capacity formula, compute for .
.
Plug in
.
Interpretation
For every bit of downloaded data from the 5 databases, the user receives about bits of useful file content β nearly the theoretical maximum of .
ex-ch04-05
MediumConsider a 3-user finite-field interference channel with signal dimension . Compute the maximum achievable per-user DoF via zero-forcing alone (no alignment), and via IA. Quantify the gap.
Zero-forcing at the receiver uses .
IA achieves roughly per user in the limit.
Zero-forcing
With uniform per-user DoF , each receiver must null interferers of dimension each, requiring , i.e., .
IA
per user in the limit. At finite one gets or depending on rounding.
Gap
IA gains a factor of over zero-forcing in the sum-DoF: vs. . At , the practical gain is about .
ex-ch04-06
MediumDesign a concrete coded-matrix-multiplication scheme for and workers over . Specify the storage and decoding. What is the straggler tolerance?
Use polynomial codes with evaluation points .
Storage
Worker stores and , for mod 7.
Computation
where , degree 3.
Recovery threshold
. Master can tolerate straggler.
Decoding
Collect any responses and interpolate the degree-3 polynomial via Lagrange interpolation; the four coefficients are the four output blocks.
ex-ch04-07
MediumProve that the global caching gain cannot be exceeded by any centralized scheme using information-theoretic arguments. Sketch the cut-set bound.
Cut: all users on one side, server on the other.
Each user has cache ; server must fill in the rest.
Cut-set
For demand with all-distinct files, the users collectively need files' worth of information. They already cache files' worth; the server broadcast must supply the remaining files β exactly the Maddah-Ali / Niesen rate.
Why the bound is tight
The global gain cannot be exceeded because every broadcast can satisfy at most users simultaneously (each bit of the broadcast lies in at most user "cache-aligned" subspaces).
Caveat
The converse is strictly for centralized placement; decentralized placement has a slightly relaxed converse (see Maddah-Ali / Niesen Β§VI).
ex-ch04-08
MediumFor classical PIR with databases and files, write out a concrete scheme achieving the capacity .
Split each file into 3 chunks.
Each database returns 1 or 2 chunks; total download = 4 chunks; useful = 3 (one file).
File split
, .
Scheme (user wants $W_1$)
- DB 1: returns .
- DB 2: returns .
- DB 3: returns .
- Need additionally for cancellation; DB 3 also returns .
Total: 4 chunks downloaded. Rate . β
Privacy check
Each database sees one or two chunk-labels of each file; the choice of which labels depends on a random permutation unknown to any single database. With the permutation uniform, each database's query is statistically independent of the desired index .
ex-ch04-09
MediumArgue why finite-field IA for coded matrix multiplication has zero failure probability, while a generic random-encoding scheme has strictly positive failure probability over small fields.
Polynomial codes use Vandermonde structure.
Polynomial code (deterministic)
Any submatrix of a Vandermonde matrix (with distinct evaluation points) is invertible. Zero failure probability.
Generic random encoding
With i.i.d. uniform coefficients in , a submatrix is singular with probability . For this failure probability exceeds β noticeable.
Practical upshot
Production systems use polynomial codes, not random encoding. The wireless-IA literature fascination with generic randomized constructions does not carry over to the algebraic, code-based distributed computing setting.
ex-ch04-10
MediumGive an explicit 2-user finite-field interference channel over with generic channel matrices, and verify the alignment condition for a scheme achieving DoF per user with signal dimensions.
With , IA reduces to zero-forcing; choose to null one cross-channel.
Setup
Let , , over .
Precoders and projections
, . The interferer at receiver 1 is . Choose β null of the interferer.
Verification
β. , full rank β. Similarly at receiver 2. Per-user DoF: . Sum DoF: , matching the no-interference upper bound.
ex-ch04-11
HardProve that the Maddah-Ali / Niesen coded-caching scheme is exactly optimal within the class of uncoded-placement schemes. What happens with coded placement β is the gain strictly larger?
Yu, Maddah-Ali, Avestimehr 2018 showed a gap for small with coded placement.
Uncoded-placement optimality
Within uncoded placement (each user caches file chunks in plain), the MN scheme achieves the cut-set bound exactly (see Yu/Maddah-Ali/Avestimehr 2018). Any scheme reaching this bound is optimal.
Coded placement
Yu / Maddah-Ali / Avestimehr (2018) showed a marginal improvement with coded placement for small . The gap vanishes as , so asymptotically the MN bound is tight. At the coded- placement rate can be up to smaller.
Takeaway
For the production regimes of interest (), the MN bound is tight and coded placement offers negligible advantage. Uncoded placement is therefore the right default in Chapter 7's shuffling analysis.
ex-ch04-12
HardExtend the polynomial-code construction of Β§4.2 to -private coded matrix multiplication (any colluding workers learn nothing about ). What is the recovery threshold?
Add random-coefficient terms to the encoding polynomial.
Construction
Replace the polynomial-code polynomial by a randomized version , where are fresh random matrices.
Recovery threshold
. The added terms preserve privacy (any evaluations are uniform random) at the cost of more worker responses for the master to decode.
Tradeoff
The privacy parameter adds linearly to the recovery threshold. For , one extra worker in exchange for perfect secrecy against any single worker. This is the construction used in Chapter 11 for ByzSecAgg.
ex-ch04-13
HardShow that the classical PIR capacity approaches as . Interpret the result operationally.
as .
Limit
As , , so .
Operational meaning
With a huge library, the PIR download rate per byte of useful content is . With 2 databases this is ; with 5 databases, . More databases less overhead β the marginal cost of privacy vanishes as the system scales.
Operational vs. asymptotic
Practical PIR deployments care about finite- behavior, where the capacity formula has a small correction. Chapter 13 develops the coded-storage variant where the effective file size grows and the correction becomes visible.
ex-ch04-14
HardCompare the recovery threshold of polynomial-coded matrix multiplication with the PIR download rate in terms of their achievability via finite-field IA. What are the structural parallels and differences?
Parallels
Both constructions express a target computation (matrix product; file retrieval) as a polynomial evaluation, and decode via Lagrange interpolation on a subset of evaluations. Both achieve optimality from finite-field IA: cross-channel alignments reduce to Vandermonde invertibility.
Differences
Coded matrix multiplication aligns many output blocks (the entries of the product) into a low-degree polynomial; PIR aligns file interferers (the non-requested files) into a small number of independent directions. The first is a source-coding-style counting argument; the second is a privacy-preserving download.
Synthesis
Both problems are instances of finite-field IA in the "many-to-one" configuration β many messages, one receiver (master or user). The alignment reduces the effective per-unit overhead from (trivial) to (IA-optimal) β exactly the DoF gain of Section 4.1 in another disguise.
ex-ch04-15
ChallengeConsider a hybrid system combining distributed matrix multiplication with coded caching: workers compute , but intermediate results can be cached at the workers to speed up future products. Conjecture a joint tradeoff between the matrix-multiplication recovery threshold , the caching gain , and the per-worker storage .
Think about what's reused across different products.
The caching gain applies to the repeat-product traffic.
Structure
For a single product, polynomial codes give (Β§4.2). For a stream of products, cached intermediate values (the sub-products or their polynomial combinations) can be reused across products.
Conjectured tradeoff
The effective recovery threshold for a stream of products with cache size scales as in the many-products limit β achieving a reduction in the per-product straggler-tolerance requirement. The total per-worker storage is .
Research status
The precise tradeoff is open. The CCG (coded-computing with caching) literature (2019βpresent) has partial achievability results but no matching converse. This is a genuine open problem at the intersection of coded computing and coded caching β a natural research direction extending the IA machinery of this chapter.