Contrastive representation learning has shown to be effective to learn representations from unlabeled data. However, much progress has been made in vision domains relying on data augmentations carefully designed using domain knowledge. In this work, we propose i-Mix, a simple yet effective domain-agnostic regularization strategy for improving contrastive representation learning. We cast contrastive learning as training a non-parametric classifier by assigning a unique virtual class to each data in a batch. Then, data instances are mixed in both the input and virtual label spaces, providing more augmented data during training. In experiments, we demonstrate that i-Mix consistently improves the quality of learned representations across domains, including image, speech, and tabular data. Furthermore, we confirm its regularization effect via extensive ablation studies across model and dataset sizes. The code is available at https://github.com/kibok90/imix.
翻译:然而,在利用域知识仔细设计的数据扩增方面,在愿景领域已经取得了很大进展。在这项工作中,我们提议采用i-Mix,这是一个简单而有效的域-不可知的正规化战略,用于改进对比代表性学习。我们把对比性学习描绘成对非参数分类师的培训,方法是分批为每组数据指定一个独特的虚拟类。然后,输入和虚拟标签空间的数据实例混杂在一起,在培训期间提供了更多的数据。在实验中,我们证明i-Mix始终在提高不同领域(包括图像、语音和表格数据)的学习代表性质量。此外,我们通过在模型和数据集大小之间开展广泛的扩张研究,确认其正规化效果。该代码可在https://github.com/kibok90/imix查阅。