Deep learning has powered recent successes of artificial intelligence (AI). However, the deep neural network, as the basic model of deep learning, has suffered from issues such as local traps and miscalibration. In this paper, we provide a new framework for sparse deep learning, which has the above issues addressed in a coherent way. In particular, we lay down a theoretical foundation for sparse deep learning and propose prior annealing algorithms for learning sparse neural networks. The former has successfully tamed the sparse deep neural network into the framework of statistical modeling, enabling prediction uncertainty correctly quantified. The latter can be asymptotically guaranteed to converge to the global optimum, enabling the validity of the down-stream statistical inference. Numerical result indicates the superiority of the proposed method compared to the existing ones.
翻译:深层学习使人工智能(AI)最近取得了成功。然而,深层神经网络,作为深层学习的基本模式,受到了本地陷阱和误差等问题的影响。在本文中,我们为稀疏深习提供了一个新框架,上述问题得到了连贯处理。特别是,我们为稀疏深学习奠定了理论基础,并提出了学习稀薄神经网络的先导算法。前者成功地将稀疏深神经网络引入了统计模型框架,使预测的不确定性得以正确量化。后者可能同样得到保证,以便与全球最佳结合,使下流统计推论的有效性得以实现。数字结果表明拟议方法与现有方法相比具有优势。