Current evaluations of Continual Learning (CL) methods typically assume that there is no constraint on training time and computation. This is an unrealistic assumption for any real-world setting, which motivates us to propose: a practical real-time evaluation of continual learning, in which the stream does not wait for the model to complete training before revealing the next data for predictions. To do this, we evaluate current CL methods with respect to their computational costs. We conduct extensive experiments on CLOC, a large-scale dataset containing 39 million time-stamped images with geolocation labels. We show that a simple baseline outperforms state-of-the-art CL methods under this evaluation, questioning the applicability of existing methods in realistic settings. In addition, we explore various CL components commonly used in the literature, including memory sampling strategies and regularization approaches. We find that all considered methods fail to be competitive against our simple baseline. This surprisingly suggests that the majority of existing CL literature is tailored to a specific class of streams that is not practical. We hope that the evaluation we provide will be the first step towards a paradigm shift to consider the computational cost in the development of online continual learning methods.
翻译:持续学习(CL)方法的当前评估通常假定训练时间和计算没有限制。这是任何实际环境下的不切实际的假设,这促使我们提出:一种实用的在线真实时间评估连续学习的方法,在这种方法中,流不会等待模型完成训练就在进行下一次数据预测。为此,我们评估目前使用的CL方法,并根据其计算成本进行评估。我们在CLOC上进行了大量实验,该数据集包含3900万个时间戳的图像和地理位置标签。我们展示了一个简单的基线在这种评估下优于最先进的CL方法,质疑了现有方法在实际情况下的适用性。此外,我们探索了文献中常用的各种CL组件,包括内存采样策略和正则化方法。我们发现,所有考虑的方法都无法与我们的简单基线相竞争。这出乎意料地表明,现有的大多数CL文献都是针对一类非实际应用的流的。我们希望提供的评估将成为考虑计算成本开发在线连续学习方法的首要步骤。