Recent years have seen growing interest in learning disentangled representations, in which distinct features, such as size or shape, are represented by distinct neurons. Quantifying the extent to which a given representation is disentangled is not straightforward; multiple metrics have been proposed. In this paper, we identify two failings of existing metrics, which mean they can assign a high score to a model which is still entangled, and we propose two new metrics, which redress these problems. We then consider the task of compositional generalization. Unlike prior works, we treat this as a classification problem, which allows us to use it to measure the disentanglement ability of the encoder, without depending on the decoder. We show that performance on this task is (a) generally quite poor, (b) correlated with most disentanglement metrics, and (c) most strongly correlated with our newly proposed metrics.
翻译:近年来,越来越多的研究者对学习解缠表示感兴趣,其中不同的特征,如大小或形状,由不同的神经元表示。量化给定表示解缠的程度并不简单;已经提出了多种度量方法。在本文中,我们确定了现有度量方法的两个缺点,这意味着它们可能会给一个仍然被缠绕的模型分配一个高分数,并提出了两种新的度量方法,用于纠正这些问题。然后,我们考虑了组合推广的任务。与之前的工作不同,我们将其视为一个分类问题,这使得我们可以使用它来衡量编码器的解缠能力,而不依赖于解码器。我们证明了这项任务的表现普遍很差,与大多数解缠度量相关,并且与我们新提出的度量最为相关。