Explicit antisymmetrization of a two-layer neural network is a potential candidate for a universal function approximator for generic antisymmetric functions, which are ubiquitous in quantum physics. However, this strategy suffers from a sign problem, namely, due to near exact cancellation of positive and negative contributions, the magnitude of the antisymmetrized function may be significantly smaller than that before antisymmetrization. We prove that the severity of the sign problem is directly related to the smoothness of the activation function. For smooth activation functions (e.g., $\tanh$), the sign problem of the explicitly antisymmetrized two-layer neural network deteriorates super-polynomially with respect to the system size. On the other hand, for rough activation functions (e.g., ReLU), the deterioration rate of the sign problem can be tamed to be at most polynomial with respect to the system size. Finally, the cost of a direct implementation of antisymmetrized two-layer neural network scales factorially with respect to the system size. We describe an efficient algorithm for approximate evaluation of such a network, of which the cost scales polynomially with respect to the system size and inverse precision.
翻译:双层神经网络的显微反对称功能是通用反对称功能近似功能的可能的候选条件,这些功能在量子物理学中普遍存在。然而,这一战略存在一个标志性问题,即由于几乎完全取消正和负贡献,反对称功能的规模可能大大小于抗对称功能之前。我们证明信号问题的严重性与激活功能的顺利性直接相关。对于顺利启动功能(例如$\tanh$)而言,明显反对称双层神经网络的标志问题使系统规模的超双层神经网络恶化。另一方面,对于粗化的激活功能(例如RELU),信号问题的恶化速度可能大大小于抗正对称功能之前。我们证明,信号问题的严重性与系统规模的平稳性直接实施两层神经网络的顺利性功能(例如$\tanh$)的成本。对于系统规模的精确度,我们用系统规模的精确度来描述系统规模的精确度的精确度,我们用这种系统规模的精确度来描述一个系统规模的精确性算法。