Despite their success with unstructured data, deep neural networks are not yet a panacea for structured tabular data. In the tabular domain, their efficiency crucially relies on various forms of regularization to prevent overfitting and provide strong generalization performance. Existing regularization techniques include broad modelling decisions such as choice of architecture, loss functions, and optimization methods. In this work, we introduce Tabular Neural Gradient Orthogonalization and Specialization (TANGOS), a novel framework for regularization in the tabular setting built on latent unit attributions. The gradient attribution of an activation with respect to a given input feature suggests how the neuron attends to that feature, and is often employed to interpret the predictions of deep networks. In TANGOS, we take a different approach and incorporate neuron attributions directly into training to encourage orthogonalization and specialization of latent attributions in a fully-connected network. Our regularizer encourages neurons to focus on sparse, non-overlapping input features and results in a set of diverse and specialized latent units. In the tabular domain, we demonstrate that our approach can lead to improved out-of-sample generalization performance, outperforming other popular regularization methods. We provide insight into why our regularizer is effective and demonstrate that TANGOS can be applied jointly with existing methods to achieve even greater generalization performance.
翻译:尽管在无结构化数据方面取得了成功,但深神经网络还不是结构化表格数据的万灵丹。在表格领域,其效率关键依赖于各种形式的正规化,以防止过度配置和提供强有力的概括性业绩。现有的正规化技术包括广泛的建模决定,如结构选择、损失功能和优化方法。在这项工作中,我们引入了土形神经梯度梯度梯度分级和专业化(TANGOS),这是在基于潜在单位属性的表格设置中实现正规化的新框架。在特定输入特征方面,激活的梯度表示神经元如何关注这一特征,并常常被用于解释深网络的预测。在TANGOS中,我们采取不同的做法,将神经属性直接纳入培训,以鼓励完全联网网络中潜在属性的整形或分化和专业化。我们的正规化鼓励神经元关注分散、不重叠的投入特征和结果,并形成一系列多样化和专门的潜在单位。在表格领域,我们表明我们的方法可以导致改进超模范总化的常规化绩效,我们可以通过其他的普通化方法来展示。</s>