Continual learning in environments with shifting data distributions is a challenging problem with several real-world applications. In this paper we consider settings in which the data distribution(task) shifts abruptly and the timing of these shifts are not known. Furthermore, we consider a semi-supervised task-agnostic setting in which the learning algorithm has access to both task-segmented and unsegmented data for offline training. We propose a novel approach called mixture of Basismodels (MoB) for addressing this problem setting. The core idea is to learn a small set of basis models and to construct a dynamic, task-dependent mixture of the models to predict for the current task. We also propose a new methodology to detect observations that are out-of-distribution with respect to the existing basis models and to instantiate new models as needed. We test our approach in multiple domains and show that it attains better prediction error than existing methods in most cases while using fewer models than other multiple model approaches. Moreover, we analyze the latent task representations learned by MoB and show that similar tasks tend to cluster in the latent space and that the latent representation shifts at the task boundaries when tasks are dissimilar.
翻译:在不断变化的数据分布环境中不断学习是一个具有挑战性的问题,涉及若干现实世界应用。在本文中,我们考虑了数据分布(任务)突变以及这些转变的时机不为人知的各种环境。此外,我们考虑的是半监督的任务 -- -- 不可知环境,学习算法在这种环境中既可以获得任务分解的数据,也可以获得非分解的数据,以便进行离线培训。我们提出了一种新颖的方法,称为基础模型(MoB)的混合方法,以解决这一问题。核心思想是学习一套小型的基础模型,并构建一种动态的、任务独立的模型组合,以预测当前任务。我们还提出了一种新方法,用以检测在现有基础模型方面超越分配的观测结果,并在需要时对新模型进行即时即时即时的速化。我们在多个领域测试我们的方法,并表明在多数情况下,它比现有方法的预测错误要好,同时使用比其他多种模式方法更少的模式。此外,我们分析了教育部学到的潜在任务表现方式,并表明类似的任务往往集中在潜在空间的集群,任务在任务不同时,任务边界上的潜在代表性变化。