Low-order functional ANOVA (fANOVA) models have been rediscovered in the machine learning (ML) community under the guise of inherently interpretable machine learning. Explainable Boosting Machines or EBM (Lou et al. 2013) and GAMI-Net (Yang et al. 2021) are two recently proposed ML algorithms for fitting functional main effects and second-order interactions. We propose a new algorithm, called GAMI-Tree, that is similar to EBM, but has a number of features that lead to better performance. It uses model-based trees as base learners and incorporates a new interaction filtering method that is better at capturing the underlying interactions. In addition, our iterative training method converges to a model with better predictive performance, and the embedded purification ensures that interactions are hierarchically orthogonal to main effects. The algorithm does not need extensive tuning, and our implementation is fast and efficient. We use simulated and real datasets to compare the performance and interpretability of GAMI-Tree with EBM and GAMI-Net.
翻译:在机器学习界,以内在可解释的机器学习为幌子,重新发现了低顺序功能 ANOVA(fANOVA)模型;可解释的推动机或EBM(Lou等人,2013年)和GAMI-Net(Yang等人,2021年)是最近提出的两个ML算法,用于适应功能性主要效应和二级互动。我们提出了一个新的算法,称为GAMI-Tree,与EBM相似,但具有一些能提高性能的特性。它使用基于模型的树木作为基础学习者,并采用新的互动过滤方法,以更好地捕捉基本互动。此外,我们的迭代培训方法与一个具有更好的预测性能的模型汇合,而嵌入式净化则确保相互作用在等级或高度上与主要效果相交。这一算法不需要广泛的调整,我们的实施是迅速有效的。我们使用模拟和真实的数据集,将GAMI-Tree与EBM和GAMI-Net的性能和可解释性作比较。