The necessity of deep learning for tabular data is still an unanswered question addressed by a large number of research efforts. The recent literature on tabular DL proposes several deep architectures reported to be superior to traditional "shallow" models like Gradient Boosted Decision Trees. However, since existing works often use different benchmarks and tuning protocols, it is unclear if the proposed models universally outperform GBDT. Moreover, the models are often not compared to each other, therefore, it is challenging to identify the best deep model for practitioners. In this work, we start from a thorough review of the main families of DL models recently developed for tabular data. We carefully tune and evaluate them on a wide range of datasets and reveal two significant findings. First, we show that the choice between GBDT and DL models highly depends on data and there is still no universally superior solution. Second, we demonstrate that a simple ResNet-like architecture is a surprisingly effective baseline, which outperforms most of the sophisticated models from the DL literature. Finally, we design a simple adaptation of the Transformer architecture for tabular data that becomes a new strong DL baseline and reduces the gap between GBDT and DL models on datasets where GBDT dominates.
翻译:对表格数据进行深层次学习的必要性仍然是许多研究努力解决的一个尚未解决的问题。最近关于表格DL的文献建议了一些深层结构,据报告,这些结构优于传统的“浅”模型,如“渐进推进式决策树”等传统“浅水”模型。然而,由于现有工作往往使用不同的基准和调制程序,因此不清楚拟议的模型是否普遍优于GBDT。此外,这些模型往往没有相互比较,因此,确定从业人员的最佳深水模型是困难的。在这项工作中,我们从彻底审查最近为表格数据开发的DL模型的主要组别开始。我们仔细调整和评估了广泛的数据集,并揭示了两项重要结论。首先,我们表明GBDT和DL模型之间的选择高度依赖数据,仍然没有普遍优于数据的解决方案。第二,我们证明简单的ResNet型结构是一个令人惊讶的有效基准,它比DL文献中的大多数复杂模型都差。最后,我们设计一个简单的变换器结构,用于表格数据,成为新的强大的DL基线,并缩小DDDG模型之间的支配地位。