Disentangled Representation Learning (DRL) aims to learn a model capable of identifying and disentangling the underlying factors hidden in the observable data in representation form. The process of separating underlying factors of variation into variables with semantic meaning benefits in learning explainable representations of data, which imitates the meaningful understanding process of humans when observing an object or relation. As a general learning strategy, DRL has demonstrated its power in improving the model explainability, controlability, robustness, as well as generalization capacity in a wide range of scenarios such as computer vision, natural language processing, data mining etc. In this article, we comprehensively review DRL from various aspects including motivations, definitions, methodologies, evaluations, applications and model designs. We discuss works on DRL based on two well-recognized definitions, i.e., Intuitive Definition and Group Theory Definition. We further categorize the methodologies for DRL into four groups, i.e., Traditional Statistical Approaches, Variational Auto-encoder Based Approaches, Generative Adversarial Networks Based Approaches, Hierarchical Approaches and Other Approaches. We also analyze principles to design different DRL models that may benefit different tasks in practical applications. Finally, we point out challenges in DRL as well as potential research directions deserving future investigations. We believe this work may provide insights for promoting the DRL research in the community.
翻译:代表学习(DRL)旨在学习一种模型,能够识别和分解以代表形式显示的可观察数据中隐藏的基本因素。将变化的基本因素分解为具有语义含义的变量的过程有助于学习可解释的数据表述,这模仿了人类在观察物体或关系时有意义的理解过程。作为一项一般学习战略,DRL展示了它在改进模型解释、可控性、稳健性以及广泛情景中的一般化能力,如计算机视觉、自然语言处理、数据挖掘等。在文章中,我们从动机、定义、方法、评价、应用和模型设计等各个方面全面审查DRL。我们讨论基于两个公认的定义的关于DRL的工作,即直观定义和组论定义。我们进一步将DRL的方法分为四个组,即传统统计方法、虚拟自动编码方法、吉列式自动编码网络基础方法、吉拉尔杰性社区分析方法和其他方法。我们还将DRL系统的未来研究方向分析作为我们未来的研究方向。