We propose a unifying view to analyze the representation quality of self-supervised learning (SSL) models without access to supervised labels, while being agnostic to the architecture, learning algorithm or data manipulation used during training. We argue that representations can be evaluated through the lens of expressiveness and learnability. We propose to use the Intrinsic Dimension (ID) to assess expressiveness and introduce Cluster Learnability (CL) to assess learnability. CL is measured as the learning speed of a KNN classifier trained to predict labels obtained by clustering the representations with K-means. We thus combine CL and ID into a single predictor: CLID. Through a large-scale empirical study with a diverse family of SSL algorithms, we find that CLID better correlates with in-distribution model performance than other competing recent evaluation schemes. We also benchmark CLID on out-of-domain generalization, where CLID serves as a predictor of the transfer performance of SSL models on several classification tasks, yielding improvements with respect to the competing baselines.
翻译:我们提出统一的观点,分析自我监督学习(SSL)模式的代表性质量,而没有获得受监督的标签,同时对培训期间使用的架构、学习算法或数据操纵进行不可知性分析,我们主张可以通过表达性和可学习性的角度来评估这些模式的代表性;我们提议使用Intrinsic 维度(ID)来评估表现性,并采用集群可学习性(CL)来评估学习性;CLL的衡量方法是,受过训练的KNN分类器的学习速度,该分类器通过将表述与K手段相结合来预测获得的标签。我们因此将CL和ID合并成一个单一的预测器:CLID。我们通过大规模的经验性研究与不同系列的SLS算法相比,我们发现CLID与分配模式的性表现比其他相互竞争的最近评估计划要好。我们还将CLID作为CLID的外部普及性基准,在那里作为SLI模型在若干分类任务上的转移性表现的预测器,从而改进了竞争基线。