The idea of slicing divergences has been proven to be successful when comparing two probability measures in various machine learning applications including generative modeling, and consists in computing the expected value of a `base divergence' between one-dimensional random projections of the two measures. However, the topological, statistical, and computational consequences of this technique have not yet been well-established. In this paper, we aim at bridging this gap and derive various theoretical properties of sliced probability divergences. First, we show that slicing preserves the metric axioms and the weak continuity of the divergence, implying that the sliced divergence will share similar topological properties. We then precise the results in the case where the base divergence belongs to the class of integral probability metrics. On the other hand, we establish that, under mild conditions, the sample complexity of a sliced divergence does not depend on the problem dimension. We finally apply our general results to several base divergences, and illustrate our theory on both synthetic and real data experiments.
翻译:在比较包括基因模型模型在内的各种机器学习应用中的两种概率计量方法时,切片差异的概念已证明是成功的,它包括计算两种计量方法的一维随机预测之间的“基差”的预期值。然而,这一技术的表面、统计和计算后果尚未完全确定。在本文件中,我们的目标是缩小这一差距,并得出切片概率差异的各种理论属性。首先,我们表明切片保留了度轴和差分的薄弱连续性,意味着切片差异将具有相似的地貌特性。然后,我们精确地确定在基本差异属于整体概率指标类别的情况下的结果。另一方面,我们确定,在轻微条件下,切片差异的抽样复杂性并不取决于问题层面。我们最后将我们的总体结果应用于几个基本差异,并展示我们关于合成和真实数据实验的理论。