Conformal inference is a powerful tool for quantifying the uncertainty around predictions made by black-box models (e.g. neural nets, random forests). Formally, this methodology guarantees that if the training and test data are exchangeable (e.g. i.i.d.) then we can construct a prediction set $C$ for the target $Y$ such that $P(Y \in C) = 1-\alpha$ for any target level $\alpha$. In this article, we extend this methodology to an online prediction setting where the distribution generating the data is allowed to vary over time. To account for the non-exchangeability, we develop a protective layer that lies on top of conformal inference and gradually re-calibrates its predictions to adapt to the observed changes in the environment. Our methods are highly flexible and can be used in combination with any predictive algorithm that produces estimates of the target or its conditional distribution and without any assumptions on the size or type of the distribution shift. We test our techniques on two real-world datasets aimed at predicting stock market volatility and COVID-19 case counts and find that they are robust and adaptive to real-world distribution shifts.
翻译:正式地说,这一方法保证,如果培训和测试数据可以互换(例如,d.),那么我们可以为目标Y美元设计一套C$的预测,这样P(Y)=1-alpha$,任何目标水平为1美元/阿尔法元。在本条中,我们将这一方法推广到一个在线预测环境,允许数据生成的分布随着时间的变化而变化。为了说明非交换性,我们开发了一个保护层,该层位于一致性推断之上,并逐步调整其预测,以适应观察到的环境变化。我们的方法非常灵活,可以结合任何预测算法,对目标或其有条件分布进行估计,对分布变化的规模或类型不作任何假设。我们测试了两种真实世界数据设置的技术,目的是预测股票市场的波动和COVID-19的分布变化,发现它们具有适应性,并发现它们能够适应真实世界的变化。