Self-supervised learning methods have shown impressive results in downstream classification tasks. However, there is limited work in understanding and interpreting their learned representations. In this paper, we study the representation space of several state-of-the-art self-supervised models including SimCLR, SwaV, MoCo V2 and BYOL. Without the use of class label information, we first discover discriminative features that are highly active for various subsets of samples and correspond to unique physical attributes in images. We show that, using such discriminative features, one can compress the representation space of self-supervised models up to 50% without affecting downstream linear classification significantly. Next, we propose a sample-wise Self-Supervised Representation Quality Score (or, Q-Score) that can be computed without access to any label information. Q-Score, utilizes discriminative features to reliably predict if a given sample is likely to be mis-classified in the downstream classification task achieving AUPRC of 0.91 on SimCLR and BYOL trained on ImageNet-100. Q-Score can also be used as a regularization term to remedy low-quality representations leading up to 8% relative improvement in accuracy on all 4 self-supervised baselines on ImageNet-100, CIFAR-10, CIFAR-100 and STL-10. Moreover, through heatmap analysis, we show that Q-Score regularization enhances discriminative features and reduces feature noise, thus improving model interpretability.
翻译:自我监督的学习方法在下游分类任务中显示了令人印象深刻的成果。然而,在理解和解释其所学的表述方面,工作有限,但了解和解释这些自监督模型的学习方式的工作有限。在本文件中,我们研究一些最先进的自我监督模型(包括SimCLR、SwaV、Moco V2和BYOL)的代表性空间,这些模型包括:SimCLR、SwaVV、MoCo V2和BYOL。不使用阶级标签信息,我们首先发现对各种抽样组极为活跃的歧视性特征,与图像中独特的物理特征相适应。我们表明,使用这种歧视性特征可以将自我监督模型的展示空间压缩到50 %的表示空间,而不影响下游线性分类的描述方式。我们还可以将自我监督自我监督模型的表达空间压缩至50 %,同时将自我监督性特征的自我监督性特征(或Q-S-S-S-S-SQ-S-S-Srmalizal laimal-laimal laimal laimal laim-laimal laim-laimal laimal laimal-S-S-Siltical laim-I-Siltical-Silvial laut laual laut laute)