Recent theoretical studies proved that deep neural network (DNN) estimators obtained by minimizing empirical risk with a certain sparsity constraint can attain optimal convergence rates for regression and classification problems. However, the sparsity constraint requires to know certain properties of the true model, which are not available in practice. Moreover, computation is difficult due to the discrete nature of the sparsity constraint. In this paper, we propose a novel penalized estimation method for sparse DNNs, which resolves the aforementioned problems existing in the sparsity constraint. We establish an oracle inequality for the excess risk of the proposed sparse-penalized DNN estimator and derive convergence rates for several learning tasks. In particular, we prove that the sparse-penalized estimator can adaptively attain minimax convergence rates for various nonparametric regression problems. For computation, we develop an efficient gradient-based optimization algorithm that guarantees the monotonic reduction of the objective function.
翻译:最近的理论研究证明,通过将经验风险降到最低程度而获得的深度神经网络(DNN)的测算器,具有一定的宽度限制,可以达到回归和分类问题的最佳趋同率;然而,聚度限制要求了解真实模型的某些特性,而实际上并不具备这些特性;此外,由于偏度限制的离散性质,计算起来很困难;在本文件中,我们建议对稀疏的DNN提出一种新颖的受处罚的估计方法,解决在宽度限制中存在的上述问题;我们为拟议的稀释式DNN测算器的过重风险建立甲骨不平等,并为几项学习任务得出趋同率;特别是,我们证明,稀有的受惩罚的测算器能够适应各种非对等回归问题达到微量的趋同率;关于计算,我们开发一种高效的梯度优化算法,保证目标函数的单体减小。