Modern deep convolutional networks (CNNs) are often criticized for not generalizing under distributional shifts. However, several recent breakthroughs in transfer learning suggest that these networks can cope with severe distribution shifts and successfully adapt to new tasks from a few training examples. In this work we study the interplay between out-of-distribution and transfer performance of modern image classification CNNs for the first time and investigate the impact of the pre-training data size, the model scale, and the data preprocessing pipeline. We find that increasing both the training set and model sizes significantly improve the distributional shift robustness. Furthermore, we show that, perhaps surprisingly, simple changes in the preprocessing such as modifying the image resolution can significantly mitigate robustness issues in some cases. Finally, we outline the shortcomings of existing robustness evaluation datasets and introduce a synthetic dataset SI-Score we use for a systematic analysis across factors of variation common in visual data such as object size and position.
翻译:现代深层革命网络(CNNs)往往被批评为没有在分布式转换下加以概括化;然而,最近在转移学习方面的一些突破表明,这些网络能够应对严重的分布式转换,并成功地适应几个培训实例中的新任务;在这项工作中,我们首次研究现代图像分类CNN的分布和传输性能之间的相互作用,并调查培训前数据规模、模型规模和数据处理前管道的影响;我们发现,增加培训组和模型规模大大改善了分布式转换的稳健性;此外,我们发现,或许令人惊讶的是,预处理过程中的简单变化,如修改图像分辨率,在某些情况下可以大大减轻稳健问题;最后,我们概述了现有稳健性评价数据集的缺点,并采用了合成数据集SI-S-S-Score,我们用来系统分析视觉数据中常见的差异因素,如对象大小和位置。