We present Self-Classifier -- a novel self-supervised end-to-end classification learning approach. Self-Classifier learns labels and representations simultaneously in a single-stage end-to-end manner by optimizing for same-class prediction of two augmented views of the same sample. To guarantee non-degenerate solutions (i.e., solutions where all labels are assigned to the same class) we propose a mathematically motivated variant of the cross-entropy loss that has a uniform prior asserted on the predicted labels. In our theoretical analysis, we prove that degenerate solutions are not in the set of optimal solutions of our approach. Self-Classifier is simple to implement and scalable. Unlike other popular unsupervised classification and contrastive representation learning approaches, it does not require any form of pre-training, expectation-maximization, pseudo-labeling, external clustering, a second network, stop-gradient operation, or negative pairs. Despite its simplicity, our approach sets a new state of the art for unsupervised classification of ImageNet; and even achieves comparable to state-of-the-art results for unsupervised representation learning. Code is available at https://github.com/elad-amrani/self-classifier.
翻译:我们提出自定义解析器 -- -- 一种全新的自我监督端到端分类学习方法。自定义解析器以单阶段端到端的方式同时学习标签和表达方式,通过优化同一样本中两种强化观点的同级预测,对同一样本中两种强化观点进行最佳预测。为了保证非变性解决方案(即所有标签都指定在同一类的解决方案),我们提议了一个具有数学动机的跨物种损失变异器,该变异器先前在预测的标签上坚持统一。在理论分析中,我们证明退化的解决方案并不是我们的方法的最佳解决方案组合。自定义解析器简单易执行和可缩放。不同于其他流行的未经监督的分类和对比代表性学习方法,它并不要求任何形式的预培训、预期-质化、伪标签、外部集群、第二个网络、断层操作或负对等。尽管其简单,但我们的方法为不超超超级图像网络分类设置了新的艺术状态;甚至实现可与州级/变校正的自我学习结果可比的版本。