Tabular data are ubiquitous in real world applications. Although many commonly-used neural components (e.g., convolution) and extensible neural networks (e.g., ResNet) have been developed by the machine learning community, few of them were effective for tabular data and few designs were adequately tailored for tabular data structures. In this paper, we propose a novel and flexible neural component for tabular data, called Abstract Layer (AbstLay), which learns to explicitly group correlative input features and generate higher-level features for semantics abstraction. Also, we design a structure re-parameterization method to compress the learned AbstLay, thus reducing the computational complexity by a clear margin in the reference phase. A special basic block is built using AbstLays, and we construct a family of Deep Abstract Networks (DANets) for tabular data classification and regression by stacking such blocks. In DANets, a special shortcut path is introduced to fetch information from raw tabular features, assisting feature interactions across different levels. Comprehensive experiments on seven real-world tabular datasets show that our AbstLay and DANets are effective for tabular data classification and regression, and the computational complexity is superior to competitive methods. Besides, we evaluate the performance gains of DANet as it goes deep, verifying the extendibility of our method. Our code is available at https://github.com/WhatAShot/DANet.
翻译:尽管机器学习界开发了许多常用的神经元件(例如,变化)和扩展神经网络(例如,ResNet),但其中很少对表格数据有效,也很少设计适合表格数据结构。在本文中,我们为表格数据提议了一个新颖和灵活的神经元件,称为“摘要图”(AbstLay),它学习明确组合相关输入特性,为语义抽象生成更高层次的特征。此外,我们还设计了结构重新参数化方法,以压缩所学的AbstLay和扩展神经网络(例如,ResNet),从而在参考阶段将计算的复杂性降低一个明确的比值。用AbstLays建造了一个特殊的基本元件,而我们建造了一个“深摘要网络”系列,用于表格数据分类和回归。在DANet中,引入了一条特殊的快捷路径,从原始表格特征中获取信息,协助不同层次的特征互动。在七个真实世界的表格数据设置上进行全面实验,从而在参考阶段将计算中将计算的复杂性降低。在AbstL和DADA中, 我们的升级方法是我们现有的变压方法。