Exercises
ex-ch05-01
EasyFor a polynomial code with , what are the per-worker storage , the recovery threshold , and the degree of the product polynomial ?
, , .
Plug in
; ; .
ex-ch05-02
EasyA polynomial code uses workers with . What is the maximum number of stragglers the master can tolerate?
Straggler tolerance = .
Compute
. Straggler tolerance β almost half of all workers can fail.
ex-ch05-03
EasyState why the polynomial code requires the field to satisfy .
Distinct evaluation points must come from .
Reason
has nonzero elements. The polynomial code needs distinct nonzero , so , i.e., . Practically we round up to a prime power .
ex-ch05-04
EasyCompare the recovery threshold of the polynomial code and MDS-coded replication for .
Polynomial: . MDS: .
Both
Polynomial: . MDS: . MDS has lower but higher per-worker storage ( for MDS vs. for polynomial). Different tradeoffs.
ex-ch05-05
MediumConstruct a polynomial code for over with workers and evaluation points . Write out worker 5's stored matrices and response.
Encoding: , .
Encoding for worker 5
; .
Computation
where has degree .
Recovery threshold
. Master can tolerate stragglers.
ex-ch05-06
MediumProve that the encoding polynomial in Β§5.2 has the desired output blocks as the coefficients of (for , ).
Multiply by .
Multiplication
.
Distinct exponents
The exponents for are distinct (since uniquely encodes the residue mod ). Hence each appears as a distinct coefficient, and Lagrange interpolation recovers all of them independently.
ex-ch05-07
MediumFor a polynomial code with , , simulate (on paper or in your head) which workers respond when straggle. Verify that decoding still succeeds.
Polynomial codes are MDS-like β any of suffice.
Identify responders
Responders: β exactly workers. The master does not need to know which workers will straggle in advance.
Decoding
With distinct evaluation points , Lagrange interpolation succeeds because any Vandermonde submatrix is invertible.
Result
The full output is recovered exactly. The straggler tolerance is exactly the number of failed workers β at the edge of the budget.
ex-ch05-08
MediumShow that the decoder complexity of polynomial codes can be reduced from (naive Lagrange) to using FFT-based interpolation. When does the FFT speedup matter in practice?
FFT polynomial interpolation is over a field with appropriate roots of unity.
Naive Lagrange
For each output coefficient, compute Lagrange basis in time. Sum over for coefficients: total.
FFT speedup
Using the polynomial-evaluation FFT and inverse-FFT algorithms, polynomial interpolation runs in provided the field has a primitive -th root of unity. For , this is roughly a speedup over naive Lagrange.
When it matters
The FFT speedup is significant for or so. Below that, the naive approach is simpler and the constant factors of FFT overhead dominate. Modern coded-computing libraries provide both algorithms and switch based on .
ex-ch05-09
MediumCompute the storage requirement of an MDS-coded matrix multiplication scheme that achieves for . Compare with the polynomial code's .
MDS schemes typically store per worker.
MDS storage
. Each worker stores of and of β total of each input.
Polynomial-code storage
. Same per-worker storage as MDS in this case.
Comparison
At , the storage budgets are identical. The difference is the recovery threshold: MDS achieves vs. polynomial's . MDS is better for straggler tolerance at this storage. The catch is that MDS requires a different polynomial structure that is sensitive to the partition shape; polynomial codes generalize more cleanly to .
ex-ch05-10
MediumA polynomial-code system runs at on workers with i.i.d. exponential task times of rate . Compute the expected wall-clock time per matrix multiplication (assuming responses needed).
Expected wait for of exponentials: .
Recovery threshold
.
Order-statistic mean
. Compute: , . .
Compare with no redundancy
With (no redundancy), β about slower. The polynomial-code redundancy buys a factor-3 speedup at this point.
ex-ch05-11
HardProve that the -private polynomial code of Β§5.4 leaks no information to any size- coalition. State the privacy guarantee precisely in mutual-information terms.
Coalition's view = evaluations of , of .
Each polynomial has random coefficients absorbing the observations.
Privacy claim
For any with , .
Argument for $\mathbf{A}$
Coalition's view of : evaluations . The polynomial has "data" coefficients plus random . The Vandermonde linear system has random degrees of freedom on the right-hand side, fully absorbing the observations. Hence the observations are uniform over the masking space, independent of .
Argument for $\mathbf{B}$
Identical: the random terms in absorb the observations. By symmetry, no information about leaks.
Joint privacy
Since the masks for and are drawn independently, joint independence follows directly.
ex-ch05-12
HardSketch the -private polynomial-code construction for matrix-vector multiplication ( instead of ), with row partitions of . What is the recovery threshold?
in the matrix-vector case.
Setup
, (no partitioning since is a vector). Worker computes where has degree .
Recovery threshold
. With this matches the polynomial- code (since ), and each privacy unit costs extra response (vs. in the matrix- matrix case because does not need its own masking).
Storage
Per-worker storage (in normalized terms β vector is small relative to matrix). Privacy adds to the storage.
ex-ch05-13
HardConsider the MatDot (entangled polynomial) variant of Section 5.4: it achieves recovery threshold at storage β better than polynomial codes on . Why is this not a contradiction with the lower-bound converse of Β§5.3?
Read the statement of the converse carefully.
MatDot does not satisfy the polynomial code's correctness requirement β the encoded polynomials have specific relationships.
Resolve the apparent contradiction
The Β§5.3 converse applies to schemes where and are independent linear combinations of . MatDot introduces entanglement β the encoding of depends on properties of as well, breaking the independence assumption.
Where the converse applies
The converse holds for "separable" coded schemes (each worker stores a separable function of the inputs). MatDot uses a joint encoding that does not fit this template.
Lesson
Information-theoretic lower bounds depend on the scheme class. Tight bounds for one class can be beaten by extending to a strictly larger class. The coded-computing literature has many such "partial converse" results, and characterizing the full achievable region remains an active research area.
ex-ch05-14
HardA production system uses a polynomial code over the reals with . The Vandermonde decoder matrix has condition number at integer evaluation points, completely unusable. Propose three distinct strategies to fix this without changing the recovery threshold.
Chebyshev points reduce the condition number to .
Real number system β modular arithmetic β finite field.
Decoupling encoding from decoding via random rotations.
Strategy 1: Chebyshev evaluation points
Replace with Chebyshev nodes . Condition number drops to β still large but tractable in 128-bit floats.
Strategy 2: Switch to finite-field arithmetic
Compute the matrix product modulo a large prime (e.g., ), then lift back to reals. Avoids floating-point conditioning entirely. Cost: modular reduction of large matrices.
Strategy 3: Random rotations / orthonormal bases
Replace the Vandermonde encoding with a random orthogonal matrix . Provides recovery threshold with condition number (orthonormal rows). Cost: random encoding loses the explicit Lagrange decoder structure.
Production choice
Most production deployments use Strategy 2 (finite- field arithmetic), often with chosen to match the natural data type (e.g., for int32 matrices). Strategy 1 is the right choice when finite-field arithmetic is too slow.
ex-ch05-15
ChallengeOpen problem. For coded matrix multiplication with non- symmetric storage budgets ( across workers), the optimal recovery threshold is not known in full generality. Sketch a candidate construction that combines polynomial codes with unequal per-worker storage, and discuss what the recovery-threshold formula might look like.
Asymmetric polynomial codes use different evaluation points and exponents per worker.
Thinking of the problem as a weighted Vandermonde matrix may help.
Candidate construction
Assign each worker a budget (number of evaluation points it can store). Encode as polynomials with degrees adapted to the per-worker budgets β workers with more budget store more coefficients. The recovery threshold becomes the smallest such that for any -subset .
Conjectured optimal $K$
For storage budgets , the optimal is the smallest integer such that (i.e., the workers with the smallest budgets collectively store enough). This is similar to the "fractional cover" formulation in coded caching.
Status
The conjecture is consistent with known special cases (symmetric storage recovers ). A matching converse is open. This is one of the open problems of Chapter 18 β a research direction at the intersection of coded computing and combinatorial optimization.