Deep neural networks have achieved tremendous success due to their representation power and adaptation to low-dimensional structures. Their potential for estimating structured regression functions has been recently established in the literature. However, most of the studies require the input dimension to be fixed and consequently ignore the effect of dimension on the rate of convergence and hamper their applications to modern big data with high dimensionality. In this paper, we bridge this gap by analyzing a $k^{th}$ order nonparametric interaction model in both growing dimension scenarios ($d$ grows with $n$ but at a slower rate) and in high dimension ($d \gtrsim n$). In the latter case, sparsity assumptions and associated regularization are required in order to obtain optimal rates of convergence. A new challenge in diverging dimension setting is in calculation mean-square error, the covariance terms among estimated additive components are an order of magnitude larger than those of the variances and they can deteriorate statistical properties without proper care. We introduce a critical debiasing technique to amend the problem. We show that under certain standard assumptions, debiased deep neural networks achieve a minimax optimal rate both in terms of $(n, d)$. Our proof techniques rely crucially on a novel debiasing technique that makes the covariances of additive components negligible in the mean-square error calculation. In addition, we establish the matching lower bounds.
翻译:深心神经网络由于其代表力和对低维结构的适应性而取得了巨大成功,它们估计结构回归功能的潜力最近已在文献中确立。然而,大多数研究都要求固定输入维度,从而忽视维度对趋同率的影响,阻碍其应用高维度的现代大数据。在本文中,我们通过分析美元值的顺序非对称互动模型,在不断增长的维度假设中弥补这一差距,因为美元以美元增长,但以较慢的速度增长)和高维度(美元)为单位。在后一种情况下,为了达到最佳趋同率,必须假设和相关的规范化。不同维度设定的新挑战是计算平均值差差差差差差差的误差,估计添加成分的共变性术语在幅度上大于差异值的大小,而且它们可以不经适当注意地恶化统计属性。我们引入了一种关键的降低偏差的技术来修正问题。我们在某些标准假设下,在深度神经网络中,偏差的深度神经网络要达到最优度最佳化的最佳速度,以便取得最佳的趋同性的方法。