Costly, noisy, and over-specialized, labels are to be set aside in favor of unsupervised learning if we hope to learn cheap, reliable, and transferable models. To that end, spectral embedding, self-supervised learning, or generative modeling have offered competitive solutions. Those methods however come with numerous challenges \textit{e.g.} estimating geodesic distances, specifying projector architectures and anti-collapse losses, or specifying decoder architectures and reconstruction losses. In contrast, we introduce a simple explainable alternative -- coined \textbf{DIET} -- to learn representations from unlabeled data, free of those challenges. \textbf{DIET} is blatantly simple: take one's favorite classification setup and use the \textbf{D}atum \textbf{I}nd\textbf{E}x as its \textbf{T}arget class, \textit{i.e. each sample is its own class}, no further changes needed. \textbf{DIET} works without a decoder/projector network, is not based on positive pairs nor reconstruction, introduces no hyper-parameters, and works out-of-the-box across datasets and architectures. Despite \textbf{DIET}'s simplicity, the learned representations are of high-quality and often on-par with the state-of-the-art \textit{e.g.} using a linear classifier on top of DIET's learned representation reaches $71.4\%$ on CIFAR100 with a Resnet101, $52.5\%$ on TinyImagenet with a Resnext50.
翻译:成本、 噪音和超专业性, 标签将被搁置, 以有利于不受监督的学习。 如果我们希望学习廉价、 可靠和可转移的模式。 为此, 光谱嵌入、 自我监督的学习或基因化模型提供了竞争性的解决方案 。 然而, 这些方法带来很多挑战 : extit{ e. g.} 估计大地距离, 指定投影建筑和反折叠损失, 或者指定解码结构和重建损失 。 相反, 我们引入一个简单的可解释的替代方案 { commond\ textf{ DIETET} -- 来学习未标注的数据 。\ textbeplod 嵌入、 自定义和自定义的 Rextal- diversal_ reports a textffrlation_ dismessional_ discoal_ discoildal_ discorrupal_ discoal_ discoal_ dismologs a.