Probabilistic generative models provide a flexible and systematic framework for learning the underlying geometry of data. However, model selection in this setting is challenging, particularly when selecting for ill-defined qualities such as disentanglement or interpretability. In this work, we address this gap by introducing a method for ranking generative models based on the training dynamics exhibited during learning. Inspired by recent theoretical characterizations of disentanglement, our method does not require supervision of the underlying latent factors. We evaluate our approach by demonstrating the need for disentanglement metrics which do not require labels\textemdash the underlying generative factors. We additionally demonstrate that our approach correlates with baseline supervised methods for evaluating disentanglement. Finally, we show that our method can be used as an unsupervised indicator for downstream performance on reinforcement learning and fairness-classification problems.
翻译:概率基因模型为了解数据的基本几何提供了灵活和系统的框架。然而,在这种环境下选择模型具有挑战性,特别是在选择诸如分解或可解释性等定义不清的特性时。在这项工作中,我们采用基于学习期间所显示的培训动态的基因模型排序方法来弥补这一差距。受最近对分解的理论描述的启发,我们的方法不需要监督潜在的潜在因素。我们通过证明需要分解的参数而来评估我们的方法,这些参数不需要标签和文字来描述基本的遗传因素。我们还表明,我们的方法与评估分解的基线监督方法相关。最后,我们表明,我们的方法可以用作加强学习和公平分类问题的下游业绩的不受监督的指标。