Deployment of machine learning models in real high-risk settings (e.g. healthcare) often depends not only on model's accuracy but also on its fairness, robustness and interpretability. Generalized Additive Models (GAMs) have a long history of use in these high-risk domains, but lack desirable features of deep learning such as differentiability and scalability. In this work, we propose a neural GAM (NODE-GAM) and neural GA$^2$M (NODE-GA$^2$M) that scale well to large datasets, while remaining interpretable and accurate. We show that our proposed models have comparable accuracy to other non-interpretable models, and outperform other GAMs on large datasets. We also show that our models are more accurate in self-supervised learning setting when access to labeled data is limited.
翻译:在真实的高风险环境中部署机器学习模型(例如医疗保健)往往不仅取决于模型的准确性,而且取决于其公平性、稳健性和可解释性。通用Additive模型(GAMs)在这些高风险领域有很长的使用历史,但缺乏深层学习的可取特征,如差异性和可缩放性。在这项工作中,我们提议建立一个神经GAM(NODE-GAM)和神经GA$2$M(NODE-GA$2$M),该模型在与大型数据集相适应的同时,仍然可以解释和准确。我们表明,我们提议的模型与其他非可解释模型具有可比性,在大型数据集上优于其他GAMs。我们还表明,当使用标签数据的机会有限时,我们的模型在自我监督的学习环境中更加精确。