Machine learning models often encounter distribution shifts when deployed in the real world. In this paper, we focus on adaptation to label distribution shift in the online setting, where the test-time label distribution is continually changing and the model must dynamically adapt to it without observing the true label. Leveraging a novel analysis, we show that the lack of true label does not hinder estimation of the expected test loss, which enables the reduction of online label shift adaptation to conventional online learning. Informed by this observation, we propose adaptation algorithms inspired by classical online learning techniques such as Follow The Leader (FTL) and Online Gradient Descent (OGD) and derive their regret bounds. We empirically verify our findings under both simulated and real world label distribution shifts and show that OGD is particularly effective and robust to a variety of challenging label shift scenarios.
翻译:机器学习模型在现实世界中部署时往往会遇到分配变化。 在本文中,我们侧重于适应在线环境中标签分配变化的适应性,即测试时间标签分布在不断变化,模型必须动态适应,而不观察真实标签。我们利用新颖的分析,表明缺乏真实标签并不妨碍对预期测试损失的估计,从而能够减少在线标签的适应性向常规在线学习转变。我们从这一观察中了解到,我们提出了由传统在线学习技术(如Collow The Leader(FTL)和在线梯子源(ODD))启发的适应性算法,并得出了他们的遗憾界限。我们通过经验验证了模拟和真实世界标签分布变化下的结果,并表明OGD对于各种具有挑战性的标签转换情景特别有效且有力。