Real world learning scenarios involve a nonstationary distribution of classes with sequential dependencies among the samples, in contrast to the standard machine learning formulation of drawing samples independently from a fixed, typically uniform distribution. Furthermore, real world interactions demand learning on-the-fly from few or no class labels. In this work, we propose an unsupervised model that simultaneously performs online visual representation learning and few-shot learning of new categories without relying on any class labels. Our model is a prototype-based memory network with a control component that determines when to form a new class prototype. We formulate it as an online Gaussian mixture model, where components are created online with only a single new example, and assignments do not have to be balanced, which permits an approximation to natural imbalanced distributions from uncurated raw data. Learning includes a contrastive loss that encourages different views of the same image to be assigned to the same prototype. The result is a mechanism that forms categorical representations of objects in nonstationary environments. Experiments show that our method can learn from an online stream of visual input data and is significantly better at category recognition compared to state-of-the-art self-supervised learning methods.
翻译:现实世界的学习设想方案涉及在样本中不固定地分配有相继依存关系的班级,这与标准机器学习模式的提取样本的模型不同,而不受固定的、典型的统一分布;此外,现实世界的相互作用要求从少数类或无类标签中实时学习。在这项工作中,我们提议了一个不受监督的模式,既进行在线视觉表现学习,又在不依赖任何类标签的情况下对新类别进行微小的学习。我们的模型是一个基于原型的记忆网络,其控制部分决定何时形成一个新的类原型。我们把它设计成一个在线高斯混合模型,其中各组成部分仅以一个新的实例在网上创建,任务并不需要平衡,这样就可以接近从未精确的原始数据中自然不平衡的分布。学习包括一种对比性损失,鼓励对同一图像的不同观点被分配到同一原型。结果是一种机制,在不固定的环境下对物体进行绝对的表达。实验表明,我们的方法可以从在线的视觉输入数据流中学习,而且比状态的自我监督学习方法要好得多。