Deep learning models have gained great popularity in statistical modeling because they lead to very competitive regression models, often outperforming classical statistical models such as generalized linear models. The disadvantage of deep learning models is that their solutions are difficult to interpret and explain, and variable selection is not easily possible because deep learning models solve feature engineering and variable selection internally in a nontransparent way. Inspired by the appealing structure of generalized linear models, we propose a new network architecture that shares similar features as generalized linear models, but provides superior predictive power benefiting from the art of representation learning. This new architecture allows for variable selection of tabular data and for interpretation of the calibrated deep learning model, in fact, our approach provides an additive decomposition in the spirit of Shapley values and integrated gradients.
翻译:深层学习模式在统计模型中已获得极大支持,因为它们导致极具竞争力的回归模型,往往优于典型的统计模型,如通用线性模型。深层学习模型的缺点是其解决方案难以解释和解释,而且由于深层学习模型以不透明的方式解决内部地貌工程和可变选择的问题,选择变量不易。在通用线性模型的吸引力结构的启发下,我们提议一个新的网络架构,其特征与通用线性模型相似,但提供了从代表性学习艺术中受益的超强预测力。 这一新架构允许对列表数据进行变量选择,并解释经过校准的深层学习模型。 事实上,我们的方法提供了符合Shapley价值和集成梯度精神的添加分解。