Deep Learning (DL) is considered the state-of-the-art in computer vision, speech recognition and natural language processing. Until recently, it was also widely accepted that DL is irrelevant for learning tasks on tabular data, especially in the small sample regime where ensemble methods are acknowledged as the gold standard. We present a new end-to-end differentiable method to train a standard FFNN. Our method, \textbf{Muddling labels for Regularization} (\texttt{MLR}), penalizes memorization through the generation of uninformative labels and the application of a differentiable close-form regularization scheme on the last hidden layer during training. \texttt{MLR} outperforms classical NN and the gold standard (GBDT, RF) for regression and classification tasks on several datasets from the UCI database and Kaggle covering a large range of sample sizes and feature to sample ratios. Researchers and practitioners can use \texttt{MLR} on its own as an off-the-shelf \DL{} solution or integrate it into the most advanced ML pipelines.
翻译:深学习( DL) 被视为计算机视觉、语音识别和自然语言处理方面的最先进艺术。 直到最近, 人们也广泛接受 DL 与表格数据学习任务无关, 特别是在小型样本制度中, 共用方法被承认为金标准。 我们为培训标准的 FFNN 提供了一种新的端到端的不同方法。 我们的方法,\ textbf{ mudling 标签用于常规化 (\ textt{MLR}), 惩罚通过生成非信息规范标签和在培训期间对最后一个隐藏层应用不同的近身规范化计划实现记忆化。\ textt{MLR} 超越经典NNN 和黄金标准(GBDT, RF), 用于回归和分类任务。 我们的方法,\ textbf{Mdlegle, 覆盖大量样本大小和样本比例特征。 研究人员和从业者可以使用\ textt{ML}, 或将其自己整合到最先进的ML 。