Online continual learning for image classification studies the problem of learning to classify images from an online stream of data and tasks, where tasks may include new classes (class incremental) or data nonstationarity (domain incremental). One of the key challenges of continual learning is to avoid catastrophic forgetting (CF), i.e., forgetting old tasks in the presence of more recent tasks. Over the past few years, many methods and tricks have been introduced to address this problem, but many have not been fairly and systematically compared under a variety of realistic and practical settings. To better understand the relative advantages of various approaches and the settings where they work best, this survey aims to (1) compare state-of-the-art methods such as MIR, iCARL, and GDumb and determine which works best at different experimental settings; (2) determine if the best class incremental methods are also competitive in domain incremental setting; (3) evaluate the performance of 7 simple but effective trick such as "review" trick and nearest class mean (NCM) classifier to assess their relative impact. Regarding (1), we observe iCaRL remains competitive when the memory buffer is small; GDumb outperforms many recently proposed methods in medium-size datasets and MIR performs the best in larger-scale datasets. For (2), we note that GDumb performs quite poorly while MIR -- already competitive for (1) -- is also strongly competitive in this very different but important setting. Overall, this allows us to conclude that MIR is overall a strong and versatile method across a wide variety of settings. For (3), we find that all 7 tricks are beneficial, and when augmented with the "review" trick and NCM classifier, MIR produces performance levels that bring online continual learning much closer to its ultimate goal of matching offline training.
翻译:在线持续学习图像分类研究学习将图像从在线数据和任务流中分类的问题,其中任务可能包括新的类别(类递增)或数据不静止(域递增),持续学习的关键挑战之一是避免灾难性的忘记(CF),即忘记了最近的任务中的旧任务。在过去几年里,为解决这一问题采用了许多方法和技巧,但许多方法和技巧没有在各种现实和实际的环境下得到公平和系统的比较。为了更好地了解各种方法的相对优势和它们工作最佳的环境,本调查旨在(1) 比较诸如MIR、 ICARL 和 GDumb 等最先进的总体方法,并确定哪些方法在不同的实验环境中最有效;(2) 确定最佳的级递增方法在域递增的设置中是否也具有竞争力;(3) 评估7种简单而有效的诀窍(NCMMMM) 的性能表现,当记忆缓冲器小的时候,我们发现iCAR的种类仍然具有竞争力;(3) GDumb 超越了总体的设置,确定哪些方法在不同实验环境中最有竞争力;(2) 高的MIR 进行最有竞争力的MSD 。