Learning to disentangle and represent factors of variation in data is an important problem in AI. While many advances are made to learn these representations, it is still unclear how to quantify disentanglement. Several metrics exist, however little is known on their implicit assumptions, what they truly measure and their limits. As a result, it is difficult to interpret results when comparing different representations. In this work, we survey supervised disentanglement metrics and thoroughly analyze them. We propose a new taxonomy in which all metrics fall into one of three families: intervention-based, predictor-based and information-based. We conduct extensive experiments, where we isolate representation properties to compare all metrics on many aspects. From experiment results and analysis, we provide insights on relations between disentangled representation properties. Finally, we provide guidelines on how to measure disentanglement and report the results.
翻译:学会解剖和代表数据差异因素是AI中的一个重要问题。虽然在了解这些表述方面有许多进展,但仍不清楚如何量化脱解。虽然存在一些衡量标准,但对其隐含的假设、其真正计量和局限性却知之甚少。因此,在比较不同表述时很难解释结果。在这项工作中,我们调查了分解指标并深入分析了这些指标。我们提出了一个新的分类方法,其中所有指标都属于三个组中的一个:以干预为基础的、以预测为基础的和信息为基础的。我们进行了广泛的实验,在实验中我们分离了代表性属性,以比较许多方面的所有衡量标准。从实验结果和分析中,我们提供了关于分解的代表性属性之间关系的洞察力。最后,我们就如何衡量分解并报告结果提供了指导方针。