Real world learning scenarios involve a nonstationary distribution of classes with sequential dependencies among the samples, in contrast to the standard machine learning formulation of drawing samples independently from a fixed, typically uniform distribution. Furthermore, real world interactions demand learning on-the-fly from few or no class labels. In this work, we propose an unsupervised model that simultaneously performs online visual representation learning and few-shot learning of new categories without relying on any class labels. Our model is a prototype-based memory network with a control component that determines when to form a new class prototype. We formulate it as an online mixture model, where components are created with only a single new example, and assignments do not have to be balanced, which permits an approximation to natural imbalanced distributions from uncurated raw data. Learning includes a contrastive loss that encourages different views of the same image to be assigned to the same prototype. The result is a mechanism that forms categorical representations of objects in nonstationary environments. Experiments show that our method can learn from an online stream of visual input data and its learned representations are significantly better at category recognition compared to state-of-the-art self-supervised learning methods.
翻译:现实世界的学习情景涉及在样本中不固定地分配有相继依存关系的班级,这与标准机器学习模式相比,在样本中不以固定的、典型的统一分布方式独立地绘制样本。此外,现实世界的相互作用要求从少数类或无类标签中实时学习。在这项工作中,我们提议了一个不受监督的模式,即同时进行在线视觉表现学习和不依赖任何类标签就新类别进行微小的学习。我们的模型是一个基于原型的记忆网络,其控制组件决定何时形成一个新的类原型。我们把它设计成一个在线混合模型,其组件仅用一个单一的新例子创建,而任务不一定要平衡,这样可以近似从未精确的原始数据中自然不平衡分布。学习包括一种对比性损失,鼓励对同一图像的不同观点被分配到同一原型。结果是一种机制,在非类环境中对物体进行直截面的描述。实验表明,我们的方法可以从在线的视觉输入数据流中学习,其所学表现在类别识别上要好得多。