IA in Coded-Caching Delivery
Why Caching and Distributed Computing Are the Same Problem
Coded caching (Maddah-Ali / Niesen 2014) and coded distributed computing (Li / Maddah-Ali / Yu / Avestimehr 2018, our Chapter 2) are two faces of the same information-theoretic coin. In coded caching, users with local memory request files from a central server; the server uses a single broadcast to satisfy all demands. In coded computing, workers store partial intermediate values; the master uses an all-to-all shuffle to collect the missing parts. Both schemes use interference alignment to cram more information into fewer broadcasts: the user-side cache content is precisely what allows the "interference" of unwanted file fragments to be cancelled.
This section makes the connection precise. The cross-reference is intentionally heavy: readers of Book CC will recognize the Maddah-Ali / Niesen scheme of Β§4.3, and readers of Chapter 2 of this book will recognize the same achievability counting. The unifying principle is finite-field IA β the topic of Section 4.1 β applied to the broadcast / shuffle channel.
Definition: Coded-Caching Problem
Coded-Caching Problem
A coded-caching system has:
- A library of files , each of size bits.
- A server connected to users via a noiseless shared broadcast link.
- Per-user cache of size bits (), populated in a placement phase when user demands are unknown.
- A delivery phase in which user requests file for a random demand vector .
The delivery rate is the number of bits the server must broadcast (normalized by ) to satisfy demand . The worst-case rate is . The information-theoretic question is: given cache size , what is the minimum achievable ?
Coded Caching
A two-phase scheme in which user caches are filled (placement) before demands are known, and a single broadcast satisfies all requests (delivery). Coded broadcasts exploit the cached side information to reduce the delivery rate.
Cache Memory
The size of each user's local cache, measured in number-of-files units (). Larger allows more side information to be exploited, reducing the delivery rate. The Maddah-Ali / Niesen tradeoff characterizes the optimal exchange.
Theorem: Maddah-Ali / Niesen Coded-Caching Tradeoff
For the symmetric coded-caching system with users and files (assume for clarity), the centralized coded-caching scheme achieves worst-case delivery rate which improves over the uncoded baseline by a multiplicative factor of β the global caching gain. The scheme uses finite-field IA in the delivery phase: each broadcast XOR combines user demands into a single transmission, and each user's cache content cancels the unintended interferers.
Without caching, the server needs one broadcast per demanded file: if uncached portions are sent in the clear. With coded delivery, the server XORs together requests at a time; each user already has all but its own desired chunk and so can solve for the missing chunk by subtracting (XORing out) the others using its local cache. The "alignment" is in the cache placement: it is chosen centrally so that the right chunks meet at the right users.
Placement
Split each file into subfiles indexed by subsets of size . User caches all subfiles whose contains β total subfiles per file, or scaled correctly to bits as required.
Delivery β XOR coding
For each subset of size , broadcast , where is file 's subfile indexed by subset .
Total number of broadcasts: . Each broadcast is one subfile chunk, so the rate is . Algebraic simplification gives .
User $k$ decodes
For each broadcast indexed by , user has cached for all , so it can XOR them out and isolate β the desired subfile. Repeating across all recovers all subfiles of .
Key Takeaway
The global caching gain is finite-field IA in disguise. Each broadcast satisfies users simultaneously by XOR-aligning their unintended portions inside their local caches. The factor in the gain is the number of users β not the cache size β confirming that the coding (alignment), not the storage, is what delivers the multiplicative improvement.
Example: Users, Files,
Three users, three files , each cache holds file. Design a coded-caching scheme. Compare the delivery rate to the uncoded baseline.
Placement
. Split each file into subfiles: , similarly and . User caches . Total per-user storage: subfiles = file, satisfying .
Delivery β three XOR broadcasts
Suppose users 1, 2, 3 demand . The server broadcasts:
- (for users 1 and 2 who hold )
- (for users 1 and 3)
- (for users 2 and 3)
Total: 3 broadcasts of subfile size, i.e., delivery rate file-equivalent.
Comparison
Uncoded delivery: send each requested file in its entirety minus what is cached, giving rate . The coded scheme cuts this to β a reduction. Each user decodes its desired file by XOR-subtracting the cached portions from each broadcast.
Operational meaning
The single-broadcast XOR aligns three demands into one transmission. The "interference" β the two unintended chunks per broadcast β is precisely what the recipient's cache cancels.
Coded Caching MemoryβRate Tradeoff
Plot the delivery rate vs. the per-user cache size for coded vs. uncoded caching, for various numbers of users . The coded curve achieves the global gain over the uncoded baseline. As grows, the gain widens β the operational signature of the IA-style alignment in the delivery broadcasts.
Parameters
Number of users sharing the broadcast link
Total library size in number of files
One Broadcast Satisfying Three Users
Why This Matters: Coded Caching Coded Shuffling
The Maddah-Ali / Niesen scheme is mathematically equivalent (up to relabeling) to the Li-Maddah-Ali-Avestimehr coded- shuffling scheme of Chapter 2. In both, the global gain is , achieved by the same XOR-aligned broadcast construction. Chapter 7 of this book formalizes the equivalence and uses the connection to handle non-uniform demand distributions in distributed-ML data shuffling β a result where the CommIT group has made several contributions (Wan / Tuninetti / Caire 2021).
Common Mistake: Coded Caching Requires Coordinated Placement
Mistake:
Treat each user as caching independently β e.g., each user caches its top- favorite files based on local statistics.
Correction:
The global caching gain depends critically on coordinated placement: the server (or a designated controller) must assign each user a specific subset of subfiles, chosen so that the delivery-phase XOR alignment works. Decentralized placement (Maddah-Ali / Niesen 2014, Β§VI) recovers most of the gain at a small rate cost, but completely uncoordinated placement forfeits the alignment and reverts to uncoded performance.
Coded Caching in Production CDNs
Major content delivery networks (Akamai, Cloudflare, Netflix Open Connect) currently use uncoded caching at the edge nodes. The reasons are practical: video chunks are large relative to per-edge memory, demand statistics shift on short timescales, and the IA-style coordination among edge nodes is hard to deploy without a tightly-controlled cluster. Several research-track deployments (Cisco's coded-caching pilot, MIT's TICS testbed) have demonstrated the predicted gains in laboratory settings, but production adoption is waiting for tighter integration with QUIC / HTTP/3 transport. Wireless-edge (5G / 6G) deployments are expected to be more receptive because spectrum is precious and coordination overhead is amortized over many users.
- β’
Production CDNs typically use uncoded LRU/LFU eviction
- β’
Coded-caching gains require coordinated placement, hard at scale
- β’
5G / 6G edge caching standards (3GPP TR 23.748) leave room for coded variants
Historical Note: Coded Caching: A Decade of Impact
2014βpresentMohammad Ali Maddah-Ali and Urs Niesen's 2014 paper introduced coded caching as an information-theoretic problem distinct from cache-eviction policies. Their main result β a multiplicative global caching gain of β was the first polynomial-in- improvement over uncoded caching at any non-trivial cache size. The paper opened a decade of follow-on work: decentralized variants, hierarchical networks, demand-private caching (with relevant CommIT contributions in Chapters 7 and 15 of this book), and extensions to wireless interference networks. The "coded" half of the modern coded-computing field traces directly to this paper.
Quick Check
For users with cache size , the global caching gain over the uncoded baseline is approximately:
. The coded delivery rate is of the uncoded rate at this point.