Generative networks have experienced great empirical successes in distribution learning. Many existing experiments have demonstrated that generative networks can generate high-dimensional complex data from a low-dimensional easy-to-sample distribution. However, this phenomenon can not be justified by existing theories. The widely held manifold hypothesis speculates that real-world data sets, such as natural images and signals, exhibit low-dimensional geometric structures. In this paper, we take such low-dimensional data structures into consideration by assuming that data distributions are supported on a low-dimensional manifold. We prove statistical guarantees of generative networks under the Wasserstein-1 loss. We show that the Wasserstein-1 loss converges to zero at a fast rate depending on the intrinsic dimension instead of the ambient data dimension. Our theory leverages the low-dimensional geometric structures in data sets and justifies the practical power of generative networks. We require no smoothness assumptions on the data distribution which is desirable in practice.
翻译:许多现有实验显示,基因网络可以从低维简单到简单分布的低维分布中产生高维的复杂数据。然而,现有理论无法证明这种现象有理。广泛持有的多重假设推断,真实世界数据集,如自然图像和信号,具有低维几何结构。在本文中,我们考虑到这种低维数据结构,假设数据分布是在低维的多元上得到支持。我们证明,在瓦森斯坦-1损失中,基因网络的统计保障是有效的。我们表明,瓦瑟斯坦-1损失以快速递归为零,取决于内在维度,而不是环境数据维度。我们的理论利用数据集中的低维几何结构,并证明基因网络的实际力量是合理的。我们不需要对实际需要的数据分布进行平稳的假设。</s>