Recent years have witnessed a hot wave of deep neural networks in various domains; however, it is not yet well understood theoretically. A theoretical characterization of deep neural networks should point out their approximation ability and complexity, i.e., showing which architecture and size are sufficient to handle the concerned tasks. This work takes one step on this direction by theoretically studying the approximation and complexity of deep neural networks to invariant functions. We first prove that the invariant functions can be universally approximated by deep neural networks. Then we show that a broad range of invariant functions can be asymptotically approximated by various types of neural network models that includes the complex-valued neural networks, convolutional neural networks, and Bayesian neural networks using a polynomial number of parameters or optimization iterations. We also provide a feasible application that connects the parameter estimation and forecasting of high-resolution signals with our theoretical conclusions. The empirical results obtained on simulation experiments demonstrate the effectiveness of our method.
翻译:近些年来,在不同领域出现了深神经网络的热潮;然而,在理论上还不能很好地理解。深神经网络的理论特征应该指出其近似能力和复杂性,即显示哪些结构和规模足以处理相关任务。这项工作朝这个方向迈出了一步,从理论上研究深神经网络的近似和复杂程度,使其与无变化的功能相适应。我们首先证明,深神经网络可以普遍地接近于无变化的功能。然后,我们表明,各种类型的神经网络模型可以同时接近广泛的不变化功能,这些模型包括复杂价值的神经网络、同源神经网络以及使用多数值参数或优化迭代法的海湾神经网络。我们还提供了一种可行的应用,将高分辨率信号的参数估计和预测与我们的理论结论联系起来。从模拟实验中得出的实验结果证明了我们的方法的有效性。