Exercises

ex-ch20-e01

Easy

Define regret. What does "sublinear" mean for RT\mathcal{R}_T?

ex-ch20-e02

Easy

For T=10000T = 10000, K=20K = 20, N=100N = 100: compute the regret bound for adversarial online coded caching.

ex-ch20-e03

Medium

Show that LRU has Θ(T)\Theta(T) regret against the offline coded MAN baseline.

ex-ch20-e04

Medium

Derive the optimal FTPL learning rate Ξ·\eta for our setting.

ex-ch20-e05

Medium

For i.i.d. Zipf-Ξ±=0.8\alpha = 0.8 demand, estimate the convergence time for empirical popularity estimation to within 1% KL.

ex-ch20-e06

Hard

Prove Theorem 20.2.2 (logarithmic regret under i.i.d.).

ex-ch20-e07

Medium

Compare adversarial and stochastic regret bounds at T=106T = 10^6 rounds. Which is tighter? When would you expect each to dominate?

ex-ch20-e08

Hard

Design an adaptive algorithm for piecewise-stationary demand with SS change points over TT rounds. What is the optimal regret?

ex-ch20-e09

Medium

Why does FTPL work (even though the action space is discrete)?

ex-ch20-e10

Hard

Design a coded LRU hybrid: how would you combine LRU's admission with MAN-style delivery structure?

ex-ch20-e11

Medium

The sliding window for non-stationary learning has trade-off: narrow vs wide. Derive the optimal window size.

ex-ch20-e12

Medium

Explain how change-point detection (CUSUM) can improve regret.

ex-ch20-e13

Hard

How does online coded caching interact with demand privacy (Ch 12)? Can both be achieved simultaneously?

ex-ch20-e14

Hard

In a real CDN, can we bound the "coded-caching regret" if we use LRU for admission and coded delivery after? Analyze.

ex-ch20-e15

Hard

State one important open problem in online coded caching.