Cache-Aided PIR
From Side Information to Caches
Section §15.1 considered side info as a set of complete files. In practice, the user's cache often contains partial file content — a generic cache is some function of the library, not necessarily full files.
Cache-aided PIR generalizes side-info PIR to arbitrary cache contents. The cache is populated in a placement phase before the user knows ; during delivery, the PIR protocol exploits whatever the cache contains.
The general capacity is open. We focus on two important regimes: (i) uncoded caching (the cache stores raw file segments), and (ii) MAN-style coded caching where multiple users share the placement design (relevant for §15.3, the CommIT contribution).
Definition: Cache-Aided PIR Protocol
Cache-Aided PIR Protocol
A cache-aided PIR protocol consists of:
-
Placement phase (offline, before user knows ): databases or system generates a cache content of size bits, where is the cache fraction. The cache is delivered to the user.
-
Delivery phase (online, after user picks ): user sends queries to databases; databases respond with answers. User combines cache and answers to decode .
Privacy: .
Cache-aided PIR rate (delivery rate): where is the delivery-phase download. Note: this excludes the placement-phase transmission (which is amortized across queries).
Theorem: Cache-Aided PIR with Uncoded Prefetching (Wei–Banawan–Ulukus 2019)
For PIR with files, databases, and user cache of size bits populated via uniformly-random uncoded prefetching (each bit independently cached with probability ), the delivery-phase PIR capacity is where .
Operational interpretation: the capacity is a -binomial mixture of the rates (the "effective " weighted by the probability that exactly of the other files are in the cache).
Achievability
Random uncoded prefetching means each bit of each file is cached independently with probability . The user knows which bits they have. The PIR protocol uses the side-info structure of §15.1 conditional on the cache realization.
Converse
The cut-set converse (Wei et al.) shows that no scheme can do better than the binomial mixture under uniform random uncoded prefetching.
Limits
: (classical Sun-Jafar). : (everything cached; no download needed). : monotonically interpolates.
Example: Cache-Aided PIR at
Compute the cache-aided PIR rate for .
$\gamma = 0$
.
$\gamma = 0.25$
Binomial average over weighted by Bin. Computation: .
$\gamma = 0.5$
Bin centered at . Effective . .
$\gamma = 0.75$
Bin centered at . Effective . .
$\gamma = 1$
(everything cached).
Operational
Even modest caches (, i.e., cache fraction) give non-trivial rate gains. At , rate is close to . Cache-aided PIR is a strong engineering pattern.
Cache-Aided PIR: Cache Size vs. Rate
Plot the cache-aided PIR rate as a function of the cache fraction for fixed and . The curve is monotone in , starting from the classical Sun-Jafar value at and reaching at . The convexity reflects the binomial structure.
Parameters
Cache-Aided vs. Classical PIR — Operational Cases
| Setting | Cache Fraction | PIR Rate | Use Case |
|---|---|---|---|
| Cold start | (classical) | First retrieval, no cache | |
| Browser cache | Modest improvement | ||
| CDN edge cache | Significant improvement | ||
| Pinned/dedicated cache | Near-zero download cost | ||
| Cache hit | (file in cache) | (no PIR query needed) | Local response only |
Definition: Cache-Aided PIR with Unknown Caches
Cache-Aided PIR with Unknown Caches
In this variant, the databases do not know the user's cache content (or the placement strategy). The user must perform PIR-style queries that work for any realization of , with privacy of against the databases.
The achievable rate matches the public-cache case for uniformly-random caches: . This is because the user can encode the cache realization implicitly into the query structure.
For non-uniform / structured caches (e.g., Zipf-distributed popularity), the unknown-cache variant may incur a rate cost. The exact gap is open.
Common Mistake: Cache Fraction Counted in Different Units
Mistake:
Confuse the cache fraction (fraction of bits in the cache) with the side-info parameter (number of complete files in the cache).
Correction:
These are different parameterizations of the same idea. In §15.1 (side info), the user has complete files: , so . In §15.2 (cache), each bit is independently cached with probability : on average, but no file is stored in full. The capacity formulas differ! Side info gives (effective- reduction). Cache-aided gives a binomial mixture. They coincide only at the extremes ( or or or ).
Bridge to Coded Caching (Book CC)
Cache-aided PIR borrows the placement-then- delivery structure from coded caching (Maddah-Ali & Niesen 2014; see Book CC). The key difference: in coded caching, the user's demand is public; in cache-aided PIR, the demand is private.
Coded caching uses a clever placement to enable coded delivery (broadcasting a XOR of multiple users' missing pieces). Cache-aided PIR borrows this idea but adds the privacy constraint, complicating the delivery design.
Section §15.3 explores the natural extension: cached coded delivery with multiple users and demand privacy. This is the CommIT group's contribution (Wan-Tuninetti-Caire 2021).
Deploying Cache-Aided PIR
Practical guidelines for cache-aided PIR:
-
Cache placement strategy: uniformly- random caching achieves the rate . Popularity-driven (Zipf) caching may improve average-case rate but loses worst-case privacy guarantees.
-
Cache size budget: typical browser caches: to . Edge caches: . Pinned caches: +. The delivery rate scales correspondingly.
-
Privacy of cache content: the user's cache content is private from the databases — this is automatic if the placement was done long ago. Live cache updates require care to prevent leak via update timing.
-
Cache hit handling: pre-check the cache before issuing a PIR query. Cache hit → return locally (no PIR). Cache miss → run PIR with the remaining cache as side info.
-
Composition with coded caching: see §15.3 for the multi-user case with coded delivery + privacy.
- •
Cache fraction: , typical –
- •
Uniform-random placement: achieves capacity formula
- •
Cache hit: skip PIR, respond locally
- •
Multi-user + privacy: see §15.3 (CommIT)
Key Takeaway
Cache-aided PIR uses user-side caches to reduce delivery-phase download. The capacity for uniformly-random uncoded prefetching is a -binomial mixture of effective- rates: with . The rate is monotone in and reaches at . Multi-user extensions (Section §15.3) connect to coded caching and the CommIT group's demand-privacy work.
Quick Check
For cache-aided PIR with , the rate is approximately:
(same as )
(cache covers everything)
Cannot be computed without specifying the cache content.
The binomial mixture is centered at - (effective -), giving . The full mixture averages around -.