We demonstrate that self-learning techniques like entropy minimization and pseudo-labeling are simple and effective at improving performance of a deployed computer vision model under systematic domain shifts. We conduct a wide range of large-scale experiments and show consistent improvements irrespective of the model architecture, the pre-training technique or the type of distribution shift. At the same time, self-learning is simple to use in practice because it does not require knowledge or access to the original training data or scheme, is robust to hyperparameter choices, is straight-forward to implement and requires only a few adaptation epochs. This makes self-learning techniques highly attractive for any practitioner who applies machine learning algorithms in the real world. We present state-of-the-art adaptation results on CIFAR10-C (8.5% error), ImageNet-C (22.0% mCE), ImageNet-R (17.4% error) and ImageNet-A (14.8% error), theoretically study the dynamics of self-supervised adaptation methods and propose a new classification dataset (ImageNet-D) which is challenging even with adaptation.
翻译:我们证明,自学技术,例如最小化和假标签,在系统化域变换下改进部署的计算机视觉模型的性能是简单而有效的。我们进行了广泛的大规模实验,并显示出一致的改进,而不论模型结构、培训前技术或分布式变换类型如何。与此同时,自学在实践上是简单的,因为它不需要知识或原始培训数据或办法的获取,对超参数选择是强有力的,是直接向前走的,只需要几个适应方法。这使得自学技术对任何在现实世界中应用机器学习算法的执业者都具有高度吸引力。我们在CIFAR10-C(8.5%错误)、图像网络-C(22.0% mCE)、图像网-R(17.4%错误)和图像网-A(14.8%错误)上展示了最先进的适应结果,从理论上研究自我监督的适应方法的动态,并提出新的分类数据集(IMageNet-D),即使适应也具有挑战性。