Most continual learning (CL) algorithms have focused on tackling the stability-plasticity dilemma, that is, the challenge of preventing the forgetting of previous tasks while learning new ones. However, they have overlooked the impact of the knowledge transfer when the dataset in a certain task is biased - namely, when some unintended spurious correlations of the tasks are learned from the biased dataset. In that case, how would they affect learning future tasks or the knowledge already learned from the past tasks? In this work, we carefully design systematic experiments using one synthetic and two real-world datasets to answer the question from our empirical findings. Specifically, we first show through two-task CL experiments that standard CL methods, which are unaware of dataset bias, can transfer biases from one task to another, both forward and backward, and this transfer is exacerbated depending on whether the CL methods focus on the stability or the plasticity. We then present that the bias transfer also exists and even accumulate in longer sequences of tasks. Finally, we propose a simple, yet strong plug-in method for debiasing-aware continual learning, dubbed as Group-class Balanced Greedy Sampling (BGS). As a result, we show that our BGS can always reduce the bias of a CL model, with a slight loss of CL performance at most.
翻译:大多数持续学习(CL)算法都着重解决稳定性和可塑性的难题,即在学习新任务时如何防止忘记先前任务中学到的知识。然而,它们忽视了当某个任务的数据集存在偏差时,知识转移的影响——也就是说,当从偏倚的数据集中学习到某些意外的伪相关性时,它们会如何影响未来任务的学习或已经从以前任务中学到的知识? 在本文中,我们使用一个合成数据集和两个真实数据集设计了系统实验,从实证研究结果中回答这个问题。具体而言,我们首先通过双任务CL实验表明,标准CL方法(不知道数据集误差)可以将一个任务中的偏差转移到另一个任务中,双向转移,这种转移会因CL方法集中于稳定性或可塑性而加剧。然后我们证明误差转移也存在于更长的任务序列中,甚至会积累。最后,我们提出了一种简单而强大的插件方法:Greedy Sampling(BGS),它可以使持续学习意识到偏倚,实现去偏倚。结果表明,我们的BGS可以始终降低CL模型的偏差,最多只会略微损失CL性能。