This paper proposes Mutual Information Regularized Assignment (MIRA), a pseudo-labeling algorithm for unsupervised representation learning inspired by information maximization. We formulate online pseudo-labeling as an optimization problem to find pseudo-labels that maximize the mutual information between the label and data while being close to a given model probability. We derive a fixed-point iteration method and prove its convergence to the optimal solution. In contrast to baselines, MIRA combined with pseudo-label prediction enables a simple yet effective clustering-based representation learning without incorporating extra training techniques or artificial constraints such as sampling strategy, equipartition constraints, etc. With relatively small training epochs, representation learned by MIRA achieves state-of-the-art performance on various downstream tasks, including the linear/k-NN evaluation and transfer learning. Especially, with only 400 epochs, our method applied to ImageNet dataset with ResNet-50 architecture achieves 75.6% linear evaluation accuracy.
翻译:本文提出“相互信息规范化任务”(MIRA),这是一种假标签算法,用于在信息最大化的启发下进行不受监督的代理学习;我们将在线假标签作为一种优化问题,以寻找假标签,使标签和数据之间的相互信息最大化,同时接近某一模型概率;我们得出固定点迭代方法,并证明它与最佳解决方案趋同;与基线不同,MIRA与假标签预测相结合,可以进行简单而有效的集群代表学习,而不必包括额外的培训技术或人工限制,如抽样战略、设备设计限制等。随着培训步子相对较小,MIRA所学的代表性在包括线性/k-NN评价和转移学习在内的各种下游任务上达到最先进的业绩,包括线性/k-NN评价和转移学习。特别是只有400个环,我们用于带有ResNet-50架构的图像网络数据集的方法实现了75.6%线性评价准确度。