Normalizing flows provide an elegant method for obtaining tractable density estimates from distributions by using invertible transformations. The main challenge is to improve the expressivity of the models while keeping the invertibility constraints intact. We propose to do so via the incorporation of localized self-attention. However, conventional self-attention mechanisms don't satisfy the requirements to obtain invertible flows and can't be naively incorporated into normalizing flows. To address this, we introduce a novel approach called Attentive Contractive Flow (ACF) which utilizes a special category of flow-based generative models - contractive flows. We demonstrate that ACF can be introduced into a variety of state of the art flow models in a plug-and-play manner. This is demonstrated to not only improve the representation power of these models (improving on the bits per dim metric), but also to results in significantly faster convergence in training them. Qualitative results, including interpolations between test images, demonstrate that samples are more realistic and capture local correlations in the data well. We evaluate the results further by performing perturbation analysis using AWGN demonstrating that ACF models (especially the dot-product variant) show better and more consistent resilience to additive noise.
翻译:常规自留机制不能满足获取不可逆流的要求,也不能天真地纳入流流的正常化。为了解决这个问题,我们引入了一种新型方法,称为 " 惯性缩压流(ACF) ",它使用一种特殊的流基基因模型类别 -- -- 缩放流。我们通过插接和播放方式证明,可以将ACF引入各种艺术流模型的状态中。这证明不仅能够提高这些模型的代表性(改善每微分度的位数),而且还能大大加快这些模型的趋同程度。为了解决这个问题,我们引入了一种新型方法,即 " 惯性缩缩放流流(ACF) ",它利用一种特殊的流基基因模型类别 -- -- 缩放流。我们进一步评估结果,利用ACCF-N进行更精确化分析,以显示ACTF的耐变异性,以显示ACFM的耐受性。