To measure the difference between two probability distributions, referred to as the source and target, respectively, we exploit both the chain rule and Bayes' theorem to construct conditional transport (CT), which is constituted by both a forward component and a backward one. The forward CT is the expected cost of moving a source data point to a target one, with their joint distribution defined by the product of the source probability density function (PDF) and a source-dependent conditional distribution, which is related to the target PDF via Bayes' theorem. The backward CT is defined by reversing the direction. The CT cost can be approximated by replacing the source and target PDFs with their discrete empirical distributions supported on mini-batches, making it amenable to implicit distributions and stochastic gradient descent-based optimization. When applied to train a generative model, CT is shown to strike a good balance between mode-covering and mode-seeking behaviors and strongly resist mode collapse. On a wide variety of benchmark datasets for generative modeling, substituting the default statistical distance of an existing generative adversarial network with CT is shown to consistently improve the performance. PyTorch code is provided.
翻译:为了衡量分别称为源和目标的两种概率分布之间的差异,我们利用链规则和Bayes的理论来建造由前方部分和后向部分组成的有条件运输(CT)。前方CT是将源数据点移动到目标点的预期成本,其联合分布由源概率密度函数(PDF)的产物和依赖源的有条件分布确定,这与目标PDF通过Bayes的理论体的目标PDF有关。后方CT的定义是通过颠倒方向来定义的。CT成本可以通过替换源和目标PDF(CT)来近似,代之以其支持的离散实验性分布,使其适合隐含的分布和基于梯度梯度的基底部优化。在用于培训一种归正模型时,CT显示在模式覆盖和模式寻求行为之间保持良好的平衡,并强烈抵制模式崩溃。关于基因化模型的多种基准数据集,以取代现有CT对称网络的默认统计距离,从而持续改进性能。