Deep learning has gained huge empirical successes in large-scale classification problems. In contrast, there is a lack of statistical understanding about deep learning methods, particularly in the minimax optimality perspective. For instance, in the classical smooth decision boundary setting, existing deep neural network (DNN) approaches are rate-suboptimal, and it remains elusive how to construct minimax optimal DNN classifiers. Moreover, it is interesting to explore whether DNN classifiers can circumvent the curse of dimensionality in handling high-dimensional data. The contributions of this paper are two-fold. First, based on a localized margin framework, we discover the source of suboptimality of existing DNN approaches. Motivated by this, we propose a new deep learning classifier using a divide-and-conquer technique: DNN classifiers are constructed on each local region and then aggregated to a global one. We further propose a localized version of the classical Tsybakov's noise condition, under which statistical optimality of our new classifier is established. Second, we show that DNN classifiers can adapt to low-dimensional data structures and circumvent the curse of dimensionality in the sense that the minimax rate only depends on the effective dimension, potentially much smaller than the actual data dimension. Numerical experiments are conducted on simulated data to corroborate our theoretical results.
翻译:深层次的学习在大规模分类问题上取得了巨大的经验性成功。相反,对于深层次的学习方法缺乏统计上的理解,特别是在小型最佳化观点方面。例如,在古典的平稳决定边界设置中,现有的深神经网络(DNN)方法不优于率,而且仍然难以构建小型最大最佳DNN分类器。此外,我们进一步提出传统Tsybakov噪声状态的本地化版本,在这种状态下,我们新分类器的统计优化性得到确立。第二,我们表明DNN分类器只能适应低层次的数据结构,并避免现有DNNNN方法的亚优性。为此,我们提出一个新的深层次的学习分类器,使用差异与共性技术:DNNNG分类器在每一个地方上建构,然后汇总成一个全球分类器。我们进一步提议一个传统Tsybakov噪声状态的本地化版本,在这种状态下,我们的新分类器的统计性最佳性是双重的。第二,我们表明DNNG分类器只能适应低层次的数据结构结构,并且绕过我们实际的理论级化程度,这取决于所学系的层次上进行的数据。