Optimality and the Lower Bound

Why We Need a Converse for Polynomial Codes

Section 5.2 established that polynomial codes achieve K=pqK = pq. The next question is whether any cleverer scheme, at the same storage level ΞΌ=1/p+1/q\mu = 1/p + 1/q, could achieve K<pqK < pq. A rigorous answer requires a converse: a proof that no such scheme exists, i.e., a lower bound that matches the achievability. If we do not have the converse, we do not know whether polynomial codes are the final word or merely a good construction waiting to be beaten.

This section develops the converse argument. The structure mirrors Chapter 2's cut-set converse: identify the cut, apply an entropy bound, symmetrize, normalize. The point is that the four-step recipe transfers from generic coded shuffling to the specific matrix-multiplication setting β€” an instance of the golden thread from Chapter 2.

Theorem: Matching Lower Bound: Kβ‰₯pqK \geq pq

Consider any distributed matrix-multiplication scheme with per-worker storage ΞΌ=1/p+1/q\mu = 1/p + 1/q and independent, uniform A∈FqmΓ—d\mathbf{A} \in \mathbb{F}_q^{m \times d}, B∈FqmΓ—dβ€²\mathbf{B} \in \mathbb{F}_q^{m \times d'}. Suppose the scheme's decoder must succeed on every size-KK subset of worker responses. Then Kβ€…β€Šβ‰₯β€…β€Špq.K \;\geq\; pq. The polynomial-code construction of Β§5.2 achieves this bound with equality, so the construction is information-theoretically optimal.

The master must reconstruct pqpq matrix blocks; each worker response is a single Fq(d/p)Γ—(dβ€²/q)\mathbb{F}_q^{(d/p) \times (d'/q)}- valued observation. Without structure, pqpq observations are needed. The proof shows that the storage constraint ΞΌ=1/p+1/q\mu = 1/p + 1/q does not enable the scheme to do better β€” every coding scheme at this storage level has at least pqpq-dimensional ambiguity in pqβˆ’1pq - 1 responses.

,

Key Takeaway

Polynomial codes are information-theoretically optimal at storage ΞΌ=1/p+1/q\mu = 1/p + 1/q. No scheme can simultaneously have smaller recovery threshold and the same storage. The result is tight and matches the cut-set converse. This is a genuine achievability-converse closure β€” one of the rare instances in the coded-computing literature.

What 'Optimal' Actually Means Here

The optimality statement is subtle. Polynomial codes achieve the minimum KK at their specific storage level ΞΌ=1/p+1/q\mu = 1/p + 1/q. The tradeoff is two-dimensional: you can buy a smaller KK by increasing ΞΌ\mu. For instance:

  • ΞΌ=1/pq\mu = 1/pq (fractional storage, each worker stores one block pair): K=pqK = pq β€” uncoded minimum storage.
  • ΞΌ=1\mu = 1 (full replication): K=1K = 1 β€” any single response reconstructs.

Polynomial codes sit between these extremes at ΞΌ=1/p+1/q\mu = 1/p + 1/q. The shape of the tradeoff curve Kβˆ—(ΞΌ)K^*(\mu) is not fully characterized for all ΞΌ\mu, but piecewise-linear interpolation gives a good upper bound on the achievable KK. Section 5.4 compares polynomial codes with MDS schemes that occupy different points on this curve.

Example: Converse at p=q=2p = q = 2

For p=q=2p = q = 2 (so pq=4pq = 4), show explicitly that any scheme using 3 workers' responses to reconstruct ATB\mathbf{A}^T \mathbf{B} must have error probability bounded away from zero.

Storage vs. Recovery Threshold: the Kβˆ—(ΞΌ)K^*(\mu) Frontier

Plot the achievable (K,mu)(K, \\mu) operating points for three schemes: (i) uncoded disjoint storage, (ii) MDS-coded replication, (iii) polynomial codes with varying (p,q)(p, q). The frontier shows how KK decreases as mu\\mu increases. Each scheme occupies a specific point; polynomial codes dominate at their specific mu=1/p+1/q\\mu = 1/p + 1/q.

Parameters
20

Number of workers

16

Range of partition products to plot

Common Mistake: Optimality Is Per-Storage, Not Global

Mistake:

Claim polynomial codes are "the" optimal coded-computing scheme for matrix multiplication.

Correction:

Polynomial codes are optimal at their specific storage level ΞΌ=1/p+1/q\mu = 1/p + 1/q. At lower storage (e.g., ΞΌ=1/(pq)\mu = 1/(pq)), different schemes might have different KK; at higher storage (e.g., ΞΌ=1/2\mu = 1/2), MDS-coded replication achieves K=p+qβˆ’1≀pqK = p + q - 1 \leq pq. The right statement is: polynomial codes achieve the minimum KK at their per-worker storage level. The full frontier across all ΞΌ\mu is piecewise and requires multiple constructions.

πŸ”§Engineering Note

Converse-Matching Achievability in Production

The fact that polynomial codes match the converse is not merely aesthetic β€” it tells system architects when to stop optimizing. If a hypothetical "better" scheme promises K=pqβˆ’1K = pq - 1 at the same storage, the information-theoretic bound tells us the promise is impossible. Engineering time spent looking for one is wasted. This is perhaps the single most practical consequence of the converse proof: it saves teams from pursuing chimeras. The same lesson will reappear for PIR (Chapter 13 β€” Sun-Jafar capacity is tight) and secure aggregation (Chapter 10 β€” Caire et al.'s optimality result).

Practical Constraints
  • β€’

    At ΞΌ=1/p+1/q\mu = 1/p + 1/q, K=pqK = pq is tight β€” no cleverer scheme possible

  • β€’

    Moving to higher storage (ΞΌ>1/p+1/q\mu > 1/p + 1/q) can reduce KK further

  • β€’

    Matching converse means no 'secret' gain at this storage

Historical Note: The Converse Got Cleaner Over Time

2017–2019

Yu, Maddah-Ali, and Avestimehr's original 2017 paper established the Kβ‰₯pqK \geq pq bound via a counting argument specialized to matrix multiplication. Dutta, Cadambe, and Grover's follow-on work (2019) gave a cleaner cut-set-based proof that generalizes to other linear computations (matrix-vector products, batched linear algebra). The modern treatment β€” which we present in this section β€” uses the Chapter 2 output-entropy recipe, making the converse a case study in the "cut, bound, symmetrize, normalize" template.

,

Why This Matters: The Same Converse Drives Chapters 6, 10, 13

The four-step converse recipe applied here β€” cut, entropy bound, symmetrize, normalize β€” drives the converses in multiple later chapters. Chapter 6's gradient-coding converse, Chapter 10's secure-aggregation optimality (the Caire et al. CommIT contribution), and Chapter 13's Sun-Jafar PIR capacity are all instances of the same template, specialized to different computation tasks and threat models. Reading Section 5.3 carefully pays dividends many times over.

Quick Check

Which step of the cut-set recipe introduces the storage constraint ΞΌ\mu in the polynomial-code converse?

Identifying the cut

Applying the output-entropy bound

Symmetrizing over all equivalent cuts

Normalizing by the output size