This paper proposes an unsupervised method for learning a unified representation that serves both discriminative and generative purposes. While most existing unsupervised learning approaches focus on a representation for only one of these two goals, we show that a unified representation can enjoy the mutual benefits of having both. Such a representation is attainable by generalizing the recently proposed \textit{closed-loop transcription} framework, known as CTRL, to the unsupervised setting. This entails solving a constrained maximin game over a rate reduction objective that expands features of all samples while compressing features of augmentations of each sample. Through this process, we see discriminative low-dimensional structures emerge in the resulting representations. Under comparable experimental conditions and network complexities, we demonstrate that these structured representations enable classification performance close to state-of-the-art unsupervised discriminative representations, and conditionally generated image quality significantly higher than that of state-of-the-art unsupervised generative models. Source code can be found at https://github.com/Delay-Xili/uCTRL.
翻译:本文建议了一种不加监督的方法,用于学习既符合歧视目的又符合基因目的的统一代表制。 虽然大多数现有的未经监督的学习方法只侧重于这两个目标中的一个代表制,但我们表明,一个统一的代表制可以享受两个目标的相互利益。 通过将最近提出的称为CTRL(CTRL)的框架推广到无人监督的环境,这种代表制是可以实现的。这需要解决一个限制性的极限游戏,即降低费率目标,即扩大所有样本的特征,同时压缩每个样本的增殖特征。通过这一过程,我们看到由此产生的代表制中出现了歧视性的低维结构。在相似的实验条件和网络复杂性下,我们证明这些结构化代表制使得能够进行接近最先进的、不受监督的歧视性代表制的分类,并且有条件生成的图像质量大大高于最先进的非受监督的基因化模型。源代码见https://github.com/Delay-Xili/uCTRL。