For traffic routing platforms, the choice of which route to recommend to a user depends on the congestion on these routes -- indeed, an individual's utility depends on the number of people using the recommended route at that instance. Motivated by this, we introduce the problem of Congested Bandits where each arm's reward is allowed to depend on the number of times it was played in the past $\Delta$ timesteps. This dependence on past history of actions leads to a dynamical system where an algorithm's present choices also affect its future pay-offs, and requires an algorithm to plan for this. We study the congestion aware formulation in the multi-armed bandit (MAB) setup and in the contextual bandit setup with linear rewards. For the multi-armed setup, we propose a UCB style algorithm and show that its policy regret scales as $\tilde{O}(\sqrt{K \Delta T})$. For the linear contextual bandit setup, our algorithm, based on an iterative least squares planner, achieves policy regret $\tilde{O}(\sqrt{dT} + \Delta)$. From an experimental standpoint, we corroborate the no-regret properties of our algorithms via a simulation study.
翻译:对于交通路线平台,选择向用户推荐的路线取决于这些路线的堵塞情况。 事实上, 个人的效用取决于在此情况下使用推荐路线的人数。 受此驱动, 我们引入了Congestested Banits问题, 允许每个手臂的奖赏取决于过去$\ Delta$ 时间步数。 对过去行动历史的依赖导致一个动态系统, 算法目前的选择也会影响其未来的回报, 并且需要一种算法来为此进行规划。 我们研究多臂土匪(MAB) 设置和背景强盗设置的拥挤意识配方, 并获得线性奖赏。 对于多臂的设置, 我们提出一个UCB风格算法, 并显示其政策后悔等级为$tilde{O}( sqrt{K ktta} ) 。 对于线性背景土匪设置, 我们的算法, 以迭接式最小平方平方图设计师为基础, 实现政策上 $\ tdede{ {g\\\\\ drtqal adal devoal sal sq) a advial develop viewal develop as.