One main obstacle for the wide use of deep learning in medical and engineering sciences is its interpretability. While neural network models are strong tools for making predictions, they often provide little information about which features play significant roles in influencing the prediction accuracy. To overcome this issue, many regularization procedures for learning with neural networks have been proposed for dropping non-significant features. Unfortunately, the lack of theoretical results casts doubt on the applicability of such pipelines. In this work, we propose and establish a theoretical guarantee for the use of the adaptive group lasso for selecting important features of neural networks. Specifically, we show that our feature selection method is consistent for single-output feed-forward neural networks with one hidden layer and hyperbolic tangent activation function. We demonstrate its applicability using both simulation and data analysis.
翻译:在医学和工程科学中广泛使用深层学习的一个主要障碍是其可解释性。虽然神经网络模型是作出预测的有力工具,但它们往往很少提供关于哪些特征在影响预测的准确性方面起着重要作用的信息。为了解决这一问题,提出了许多与神经网络学习的正规化程序,以降低非重要特征。不幸的是,缺乏理论结果使人对这种管道的适用性产生怀疑。在这项工作中,我们提议并建立一个理论保证机制,以便利用适应性拉索组选择神经网络的重要特征。具体地说,我们表明,我们的特征选择方法对于单输出向前的神经网络具有一种隐藏层和超偏向相切的激活功能。我们用模拟和数据分析来证明其适用性。我们用模拟和数据分析来证明其适用性。