The Θ(M/N)\Theta(M/N) Scaling Law (Ji-Caire-Molisch)

The Headline Result

The central theorem of D2D caching theory β€” due to Ji, Caire, and Molisch (IEEE Trans. IT, 2016) β€” is that the per-user throughput scales as Θ(M/N)\Theta(M/N) independent of the network size nn. This is strikingly different from every ad-hoc capacity result before it: in classical Gupta-Kumar analysis, per-user throughput vanishes as nn grows. Caching reverses this: the effective throughput is determined entirely by the memory ratio M/NM/N, not by the number of users.

This result is a CommIT contribution and one of the most fundamental in coded-caching theory. It provides the theoretical justification for D2D-based content delivery as a scalable alternative to infrastructure.

Theorem: Ji-Caire-Molisch Scaling Law

Consider a D2D caching network with nn users uniformly distributed in a unit-area region, per-user cache M=ΞΌβ‹…NM = \mu \cdot N files from a library of NN files. Under i.i.d. uniform demands and a protocol-model interference constraint, the per-user throughput satisfies Tn(M)β€…β€Š=β€…β€ŠΞ˜(ΞΌ),nβ†’βˆž,T_n(M) \;=\; \Theta(\mu), \quad n \to \infty, provided the per-user cache has constant ΞΌ\mu (i.e., M=Θ(N)M = \Theta(N)).

Intuition in three moves:

  1. With nMn M total cached copies and nn users demanding random files, each file is cached ∼nM/N\sim nM/N times on average.
  2. The probability that a random user's demand is within D2D range of a cached copy is Θ(1)\Theta(1) (since neighbors are Θ(1)\Theta(1) many and each has M/NM/N fraction of the library).
  3. Spatial reuse allows Θ(n)\Theta(n) simultaneous D2D transmissions. Aggregate throughput Θ(n)\Theta(n); per-user Θ(1)\Theta(1)... but scaled by the hit probability M/NM/N, giving Θ(M/N)\Theta(M/N).

The miracle: throughput is independent of nn because both the supply (cached copies) and demand scale with nn.

πŸŽ“CommIT Contribution(2016)

Fundamental Limits of Caching in Wireless D2D Networks

M. Ji, G. Caire, A. F. Molisch β€” IEEE Transactions on Information Theory, vol. 62, no. 2

The Ji-Caire-Molisch 2016 paper is the foundational CommIT contribution for D2D caching theory. Its key results:

  1. Per-user throughput scales as Θ(M/N)\Theta(M/N), independent of nn. This is a constant scaling β€” the first throughput result in ad-hoc wireless that doesn't decrease with network size.
  2. Caching converts the local D2D network into a globally distributed library: the aggregate cache has effective size ∼nM/N\sim nM/N copies of each file, ensuring near-certain local hit.
  3. Order-optimality: the achievable rate matches a cut-set lower bound up to constants, so the scaling is the correct asymptotic answer.

The result has been extended in many directions: to coded multicasting (Ch 11), to D2D with privacy (Ch 12), and to hybrid D2D/infrastructure networks. It is one of the most-cited coded-caching theory papers.

coded-cachingd2dcommitscaling-lawView Paper β†’

Per-User Throughput Scaling

Per-user throughput vs network size nn on log-log axes. D2D + caching: flat at μ\mu (constant, Θ(M/N)\Theta(M/N)). D2D without caching (Gupta-Kumar): Θ(1/nlog⁑n)\Theta(1/\sqrt{n \log n}), decreasing. Infrastructure (cellular, no caching): Θ(1/n)\Theta(1/n), decreasing faster. The D2D+caching advantage grows with nn.

Parameters
0.3
500

D2D Local Exchange and Θ(M/N)\Theta(M/N) Scaling

Six users clustered in a D2D range. User 1 requests file W4W_4; neighbor user 4 serves it directly via short-range D2D link. No base station involved. Per-user throughput scales as Θ(M/N)\Theta(M/N), independent of the total network size nn β€” the Ji-Caire-Molisch scaling result.

Example: Scaling at Urban Scale

Compare per-user throughput for two urban scenarios: (a) 100 users in a 1 kmΒ² area, ΞΌ=0.1\mu = 0.1. (b) 10,000 users in a 1 kmΒ² area (100x denser), same ΞΌ\mu. For each, give the scaling-law prediction.

Key Takeaway

D2D caching achieves Θ(M/N)\Theta(M/N) per-user throughput independent of nn. More users = proportionally more aggregate cache + proportionally more demand + spatial reuse of links. All three scale together, giving flat per-user performance. This is the Ji-Caire-Molisch 2016 result and the theoretical foundation for 6G D2D caching architectures.

Common Mistake: The Scaling Is Model-Dependent

Mistake:

Quoting Θ(M/N)\Theta(M/N) as a universal D2D + caching rate without stating the model assumptions.

Correction:

The scaling law holds under specific assumptions: (i) random geometric graph / protocol interference model, (ii) random uniform demands, (iii) fixed memory ratio M/NM/N (not fixed MM), (iv) nβ†’βˆžn \to \infty asymptotic regime.

For finite nn, the constant factor matters. For Zipf demands, the bound is replaced by a popularity-aware version (Ji-Caire-Molisch 2017 extensions). For physical interference models (SINR-based), constants change. Don't overclaim.

Caching Changes the Scaling Game

The classical wisdom (Gupta-Kumar 2000): "wireless ad-hoc per-user rate Θ(1/n)\Theta(1/\sqrt{n}); does not scale." This led to a decade of pessimism about scaling.

The coded caching perspective overturns this: if you store content at users (memory is cheap), you convert transmit traffic into retrieval traffic. Each user already has some of what others need. The network's delivery demand is correspondingly reduced. Per-user rate becomes Θ(M/N)\Theta(M/N), constant in nn.

This is a recurring theme: caching changes the scaling. In D2D networks, it turns a vanishing throughput into a constant one. In infrastructure, it turns a 1/K1/K serial bound into a multiplicative 1+KM/N1 + KM/N gain. Whenever memory is abundant at users, caching rescales the fundamental limits.