Chapter Summary

Chapter Summary

Key Points

  • 1.

    Multi-access coded caching generalizes MAN: each user accesses LL caches (not just 1). Effective per-user memory LML \cdot M if placement is designed so accessed caches are disjoint.

  • 2.

    HKD cyclic wrap-around scheme (2017) achieves rate RMA(L)=K(1LM/N)/(1+KLM/N)R_\text{MA}(L) = K(1 - LM/N)/(1 + KLM/N) — MAN rate with effective memory LMLM. Gain over single-access: factor LL.

  • 3.

    Cyclic topology matches femtocell / Wi-Fi mesh deployments. In dense urban areas, users naturally see multiple APs — multi-access is the natural model.

  • 4.

    Resolvable combinatorial designs extend HKD to non-cyclic topologies. Kirkman triples, affine planes, and similar structures enable multi-access schemes with richer parameter regimes.

  • 5.

    CommIT contribution (Wan-Caire 2023) on multi-tier multi-access caching provides a theoretical grounding for 5G MEC deployments. Practical operator collaborations (Orange, TIM, Ericsson) are in research pilot stage.

  • 6.

    Tradeoff: coordination vs gain. Multi-access requires coordinated placement across AP cluster. Random (decentralized) placement wastes effective memory. Standards (3GPP Rel-17/18) enable the required coordination.

  • 7.

    Stadium example: L=4L = 4, K=1000K = 1000 fans, μ=0.1\mu = 0.1: rate reduction 55×\sim 55\times over single-access — enormous practical value for dense-venue Wi-Fi.

Looking Ahead

Chapter 18 closes Part IV (Extensions and Applications). Part V (Ch 19-22) moves to advanced topics and open problems: coded caching meets ISAC (Ch 19), online coded caching (Ch 20), video streaming with adaptive bitrate (Ch 21), and open problems in the field (Ch 22). Part V frames the CommIT research agenda for the next decade.