Efficient unsupervised training and inference in deep generative models remains a challenging problem. One basic approach, called Helmholtz machine, involves training a top-down directed generative model together with a bottom-up auxiliary model used for approximate inference. Recent results indicate that better generative models can be obtained with better approximate inference procedures. Instead of improving the inference procedure, we here propose a new model which guarantees that the top-down and bottom-up distributions can efficiently invert each other. We achieve this by interpreting both the top-down and the bottom-up directed models as approximate inference distributions and by defining the model distribution to be the geometric mean of these two. We present a lower-bound for the likelihood of this model and we show that optimizing this bound regularizes the model so that the Bhattacharyya distance between the bottom-up and top-down approximate distributions is minimized. This approach results in state of the art generative models which prefer significantly deeper architectures while it allows for orders of magnitude more efficient approximate inference.
翻译:在深基因模型中,一个称为Helmholtz 机器的基本方法涉及培训自上而下定向基因模型和用于近似推理的自下而上辅助模型。最近的结果显示,较佳的基因模型可以通过更近似推理程序获得。我们在此提出一个新的模型,以保证自上而下和自下而上分布能够有效地相互反射。我们通过将自上而下和自下而上向上分布模型解释为近似推导分布,以及将模型分布确定为这两种模型的几何平均值,来实现这一目标。我们为这一模型的可能性提出了一个较低的范围,并且我们表明,优化这一约束性模型可以使自下而上和自上而下的近似分布之间的距离最小化。这一方法的结果是艺术基因模型的状态,这些模型更偏好于更深层次的结构,同时允许有更高效的推论。