Deep generative modeling using flows has gained popularity owing to the tractable exact log-likelihood estimation with efficient training and synthesis process. However, flow models suffer from the challenge of having high dimensional latent space, the same in dimension as the input space. An effective solution to the above challenge as proposed by Dinh et al. (2016) is a multi-scale architecture, which is based on iterative early factorization of a part of the total dimensions at regular intervals. Prior works on generative flow models involving a multi-scale architecture perform the dimension factorization based on static masking. We propose a novel multi-scale architecture that performs data-dependent factorization to decide which dimensions should pass through more flow layers. To facilitate the same, we introduce a heuristic based on the contribution of each dimension to the total log-likelihood which encodes the importance of the dimensions. Our proposed heuristic is readily obtained as part of the flow training process, enabling the versatile implementation of our likelihood contribution based multi-scale architecture for generic flow models. We present such implementations for several state-of-the-art flow models and demonstrate improvements in log-likelihood score and sampling quality on standard image benchmarks. We also conduct ablation studies to compare the proposed method with other options for dimension factorization.
翻译:利用流动的深基因模型由于以高效的培训和合成过程对精密原样进行精确的逻辑类比估计而越来越受欢迎。然而,流动模型面临着具有与输入空间同样维度的高维潜层的挑战。Dinh 等人(2016年)提出的上述挑战的有效解决办法是一个多尺度的结构,其基础是定期对总维度的一部分进行迭接性早期乘数,定期对部分总维度进行迭接性考虑。以前涉及多尺度结构的基因类流模型的工程,在静态遮罩的基础上,实现维度因素化。我们建议建立一个新的多尺度结构,对哪些维度进行数据依赖的因子化,以决定哪些维度应穿过更多的流程层。同样地,我们引入基于每个维度对总日志类比的贡献的狂妄主义。我们所提议的超自然系是流动培训过程的一部分,使得基于多尺度的多维度结构能够以适应性的方式落实我们的可能性贡献。我们为若干状态的流程模型提出这样的实施情况,以显示对哪些维度的维度应贯穿于更多的流程层层。为同一层。为了便利,我们还展示了对正数值进行对比的方法进行对比,并展示了对象学质量选择的方法,以其他标准进行对比。