While self-learning methods are an important component in many recent domain adaptation techniques, they are not yet comprehensively evaluated on ImageNet-scale datasets common in robustness research. In extensive experiments on ResNet and EfficientNet models, we find that three components are crucial for increasing performance with self-learning: (i) using short update times between the teacher and the student network, (ii) fine-tuning only few affine parameters distributed across the network, and (iii) leveraging methods from robust classification to counteract the effect of label noise. We use these insights to obtain drastically improved state-of-the-art results on ImageNet-C (22.0% mCE), ImageNet-R (17.4% error) and ImageNet-A (14.8% error). Our techniques yield further improvements in combination with previously proposed robustification methods. Self-learning is able to reduce the top-1 error to a point where no substantial further progress can be expected. We therefore re-purpose the dataset from the Visual Domain Adaptation Challenge 2019 and use a subset of it as a new robustness benchmark (ImageNet-D) which proves to be a more challenging dataset for all current state-of-the-art models (58.2% error) to guide future research efforts at the intersection of robustness and domain adaptation on ImageNet scale.
翻译:虽然自学方法是最近许多领域适应技术的重要组成部分,但还没有对在稳健性研究中常见的图像网络规模数据集进行全面评估。在ResNet和高效网络模型的广泛实验中,我们发现三个组成部分对于提高自学性能至关重要:(一) 使用教师和学生网络之间的短暂更新时间,(二) 微调在整个网络中分布的少数偏差参数,以及(三) 从稳健分类到抵消标签噪音效应的杠杆化方法。我们利用这些洞察力获得大幅改进的图像网络-C(22.0% mCE)、图像网络-R(17.4%错误)和图像网络-A(14.8%错误)的最新结果,我们发现三个组成部分对于提高自学绩效至关重要:(一) 利用教师和学生网络之间的短暂更新时间,(二) 将头一误差降低到无法预期进一步取得任何重大进展的点。因此,我们重新启用了2019年视觉域适应挑战的数据集,并使用其中的一组新的稳健性基准(ImageNet-D),这在目前域域网的精确性模型上证明更具有挑战性的数据性。