We consider a set-valued online prediction problem in the context of network caching. Assume that multiple users are connected to several caches via a bipartite network. At any time slot, each user requests an arbitrary file chosen from a large catalog. A user's request at a slot is met if the requested file is cached in at least one of the caches connected to the user. Our objective is to predict, prefetch, and optimally distribute the files on the caches to maximize the total number of cache hits in an online setting. The problem is non-trivial due to the non-convex and non-smooth nature of the objective function. In this paper, we propose $\texttt{LeadCache}$ - an online caching policy based on the Follow-the-Perturbed-Leader paradigm. We show that the policy is regret-optimal up to a factor of $\tilde{O}(n^{3/8}),$ where $n$ is the number of users. We design two efficient implementations of the $\texttt{LeadCache}$ policy, one based on Pipage rounding and the other based on Madow's sampling, each of which makes precisely one call to an LP-solver per iteration. With a Strong-Law-type assumption, we show that the total number of file fetches under $\texttt{LeadCache}$ remains almost surely finite over an infinite horizon. Finally, we derive a tight regret lower bound using results from graph coloring. We conclude that the learning-based $\texttt{LeadCache}$ policy decisively outperforms the known caching policies both theoretically and empirically.
翻译:在网络缓存的背景下, 我们考虑一个设定价值的在线预测问题 。 假设多个用户通过双边网络连接到多个缓存。 在任何时间段, 每个用户都会要求从大型目录中选择任意的文件 。 如果请求的文件在至少一个与用户连接的缓存中被缓存到至少一个缓存中, 用户的要求得到满足 。 我们的目标是预测、 预发并优化地分配缓存上的文件, 以在网络设置中最大限度地增加缓存点击的总数 。 问题在于目标功能的非convex 和非mooth 性质, 问题在于非三端 。 在本文中, 我们提议 $\ t{ liadCachase} 。 如果请求的用户在至少一个缓存中隐藏在与用户连接的缓存中, 则满足了一个在线缓存政策 。 我们的目标是预测、 预略 并优化缓存到 $$$@ $ (n_ 3/8} 美元,, 美元是用户人数。 我们设计了两个高效的 $\ text deal deal deal deal deal) 政策, max max 。