Numerous invariant (or equivariant) neural networks have succeeded in handling invariant data such as point clouds and graphs. However, a generalization theory for the neural networks has not been well developed, because several essential factors for the theory, such as network size and margin distribution, are not deeply connected to the invariance and equivariance. In this study, we develop a novel generalization error bound for invariant and equivariant deep neural networks. To describe the effect of invariance and equivariance on generalization, we develop a notion of a \textit{quotient feature space}, which measures the effect of group actions for the properties. Our main result proves that the volume of quotient feature spaces can describe the generalization error. Furthermore, the bound shows that the invariance and equivariance significantly improve the leading term of the bound. We apply our result to specific invariant and equivariant networks, such as DeepSets (Zaheer et al. (2017)), and show that their generalization bound is considerably improved by $\sqrt{n!}$, where $n!$ is the number of permutations. We also discuss the expressive power of invariant DNNs and show that they can achieve an optimal approximation rate. Our experimental result supports our theoretical claims.
翻译:许多变异(或等异)神经网络成功地处理了诸如点云和图形等变异性数据。 但是,神经网络的概括理论没有很好地发展,因为数种理论的基本因素,例如网络大小和差幅分布,没有与变异和等异性密切相连。在本研究报告中,我们为变异和等异性深中线网络开发了一种新的概括错误。为了描述变异和等异性对一般化的影响,我们发展了一个概念,即测量组合动作对属性的效果的\textit{qquative space}。我们的主要结果证明,数位参数空间的量可以描述一般化错误。此外,约束表明变异性和变异性能大大改善了约束的起始期。我们把结果应用到特定的变异性和变异性网络中,如DeepSet(Zaheer等人(2017年)),并表明其一般化约束度大大改进了 $(qualticalal) 和我们理论性主张的最佳比率。