Current evaluations of Continual Learning (CL) methods typically assume that there is no constraint on training time and computation. This is an unrealistic assumption for any real-world setting, which motivates us to propose: a practical real-time evaluation of continual learning, in which the stream does not wait for the model to complete training before revealing the next data for predictions. To do this, we evaluate current CL methods with respect to their computational costs. We hypothesize that under this new evaluation paradigm, computationally demanding CL approaches may perform poorly on streams with a varying distribution. We conduct extensive experiments on CLOC, a large-scale dataset containing 39 million time-stamped images with geolocation labels. We show that a simple baseline outperforms state-of-the-art CL methods under this evaluation, questioning the applicability of existing methods in realistic settings. In addition, we explore various CL components commonly used in the literature, including memory sampling strategies and regularization approaches. We find that all considered methods fail to be competitive against our simple baseline. This surprisingly suggests that the majority of existing CL literature is tailored to a specific class of streams that is not practical. We hope that the evaluation we provide will be the first step towards a paradigm shift to consider the computational cost in the development of online continual learning methods.
翻译:目前对持续学习方法的评估通常假定培训时间和计算不受限制,这是对任何现实世界环境的一种不切实际的假设,激励我们提出:对持续学习进行实际实时评价,在这种评价中,流不等待模型完成培训,然后才披露下一个预测数据。为此,我们评估当前CL方法的计算成本。我们假设,在这一新的评价模式下,计算要求CL方法在分布不一的流上可能表现不佳。我们在CLOC上进行了广泛的实验,这是一个大型数据集,包含3 900万个标有地理定位标签的时标图像。我们显示,在这一评价中,简单的基线超越了最新的CL方法,质疑现有方法在现实环境中的适用性。此外,我们探索文献中常用的各种CL方法,包括记忆抽样战略和正规化方法。我们发现,所有考虑过的方法都无法与我们简单的基准相比具有竞争力。这令人惊讶地表明,我们现有的CL文献大多数都适合特定类别的流流模式,而这种模式的在线发展不是实际的计算方法。我们希望,我们的第一个步骤能够提供一种成本。