We prove identifiability of a broad class of deep latent variable models that (a) have universal approximation capabilities and (b) are the decoders of variational autoencoders that are commonly used in practice. Unlike existing work, our analysis does not require weak supervision, auxiliary information, or conditioning in the latent space. Recently, there has been a surge of works studying identifiability of such models. In these works, the main assumption is that along with the data, an auxiliary variable $u$ (also known as side information) is observed as well. At the same time, several works have empirically observed that this doesn't seem to be necessary in practice. In this work, we explain this behavior by showing that for a broad class of generative (i.e. unsupervised) models with universal approximation capabilities, the side information $u$ is not necessary: We prove identifiability of the entire generative model where we do not observe $u$ and only observe the data $x$. The models we consider are tightly connected with autoencoder architectures used in practice that leverage mixture priors in the latent space and ReLU/leaky-ReLU activations in the encoder. Our main result is an identifiability hierarchy that significantly generalizes previous work and exposes how different assumptions lead to different "strengths" of identifiability. For example, our weakest result establishes (unsupervised) identifiability up to an affine transformation, which already improves existing work. It's well known that these models have universal approximation capabilities and moreover, they have been extensively used in practice to learn representations of data.
翻译:我们证明,一系列深层潜伏变量模型的可辨识性很大,这些模型(a)具有普遍近似能力,(b)是实践中常用的变异自动校对器的解码器。与现有的工作不同,我们的分析并不需要薄弱的监督、辅助信息或潜伏空间的调节。最近,研究这些模型可辨识性的工程激增。在这些工程中,主要假设是,在数据的同时,还观察到一个辅助变量$u(也称为侧面信息)。与此同时,一些工作经验显示,在实践中,这似乎并不必要。在这项工作中,我们通过表明,对于具有普遍近似能力的宽度(即无超强)模型,侧信息是不必要的。我们证明了整个归真模型的可辨识性,我们没有看到美元,而只是观察数据元值。我们认为,这些模型与实践中使用的自动校正结构结构已经紧密地连接起来,在暗层空间中的混合型变现变现模型和变现性结构中,我们现有的变现的变现性系统化结果是如何在一般工作结果中学会了。