Contrastive self-supervised learning (CSL) has managed to match or surpass the performance of supervised learning in image and video classification. However, it is still largely unknown if the nature of the representation induced by the two learning paradigms is similar. We investigate this under the lens of adversarial robustness. Our analytical treatment of the problem reveals intrinsic higher sensitivity of CSL over supervised learning. It identifies the uniform distribution of data representation over a unit hypersphere in the CSL representation space as the key contributor to this phenomenon. We establish that this increases model sensitivity to input perturbations in the presence of false negatives in the training data. Our finding is supported by extensive experiments for image and video classification using adversarial perturbations and other input corruptions. Building on the insights, we devise strategies that are simple, yet effective in improving model robustness with CSL training. We demonstrate up to 68% reduction in the performance gap between adversarially attacked CSL and its supervised counterpart. Finally, we contribute to robust CSL paradigm by incorporating our findings in adversarial self-supervised learning. We demonstrate an average gain of about 5% over two different state-of-the-art methods in this domain.
翻译:在图像和视频分类中,自我监督的自我监督学习(CSL)已经成功地匹配或超过了监督性学习的绩效。然而,如果两种学习模式所引发的表述方式的性质相似,则基本上还不清楚。我们从对抗性强力的角度对此进行调查。我们对问题的分析处理表明,CSL对监督性学习具有内在的高度敏感性。我们确定,在CSL代表空间中,对一个单位超视距的数据表示的统一分布是这一现象的关键促成者。我们确定,在培训数据中存在虚假的负差时,这提高了输入干扰的模型敏感性。我们的调查结果得到使用对抗性干扰和其他输入腐败进行的大量图像和视频分类试验的支持。我们根据这些洞察,制定了简单但有效的战略,改进CSL培训的模型稳健性。我们通过将我们的调查结果纳入对立性自我监督的学习,证明CSLSL在两种不同领域的平均收益为5%。