Autoencoders exhibit impressive abilities to embed the data manifold into a low-dimensional latent space, making them a staple of representation learning methods. However, without explicit supervision, which is often unavailable, the representation is usually uninterpretable, making analysis and principled progress challenging. We propose a framework, called latent responses, which exploits the locally contractive behavior exhibited by variational autoencoders to explore the learned manifold. More specifically, we develop tools to probe the representation using interventions in the latent space to quantify the relationships between latent variables. We extend the notion of disentanglement to take the learned generative process into account and consequently avoid the limitations of existing metrics that may rely on spurious correlations. Our analyses underscore the importance of studying the causal structure of the representation to improve performance on downstream tasks such as generation, interpolation, and inference of the factors of variation.
翻译:自动编码器在将数据元体嵌入低维潜层空间方面表现出令人印象深刻的能力,使它们成为代议制学习方法的主轴。然而,如果没有明确的监督(通常没有这种监督),这种表述通常无法解释,因此分析和原则性进展具有挑战性。我们提出了一个框架,称为潜在反应,利用变异自动编码器展示的当地契约行为探索所学的元体。更具体地说,我们开发工具,利用潜潜层空间的干预来探测代议,以量化潜伏变量之间的关系。我们扩大了分解概念,以考虑到所学的基因化过程,从而避免现有指标的局限性,因为现有指标可能依赖虚假的相关性。我们的分析强调必须研究代议制的因果结构,以改进下游任务的绩效,例如生成、内插和变异因素的推断。