With increasing interest in explaining machine learning (ML) models, the first part of this two-part study synthesizes recent research on methods for explaining global and local aspects of ML models. This study distinguishes explainability from interpretability, local from global explainability, and feature importance versus feature relevance. We demonstrate and visualize different explanation methods, how to interpret them, and provide a complete Python package (scikit-explain) to allow future researchers to explore these products. We also highlight the frequent disagreement between explanation methods for feature rankings and feature effects and provide practical advice for dealing with these disagreements. We used ML models developed for severe weather prediction and sub-freezing road surface temperature prediction to generalize the behavior of the different explanation methods. For feature rankings, there is substantially more agreement on the set of top features (e.g., on average, two methods agree on 6 of the top 10 features) than on specific rankings (on average, two methods only agree on the ranks of 2-3 features in the set of top 10 features). On the other hand, two feature effect curves from different methods are in high agreement as long as the phase space is well sampled. Finally, a lesser-known method, tree interpreter, was found comparable to SHAP for feature effects, and with the widespread use of random forests in geosciences and computational ease of tree interpreter, we recommend it be explored in future research.
翻译:随着对解释机器学习模型的兴趣日益浓厚,这两份研究报告的第一部分综合了最近关于解释ML模型全球和地方方面的方法的研究。本研究报告将解释与可解释性、局部性与全球解释性、特征重要性和特征相关性区分开来。我们展示和想象不同的解释方法,如何解释这些模型,并提供完整的Python软件包(缩略图解析),以便未来的研究人员能够探索这些产品。我们还强调地物排名和地物效应解释方法之间经常存在分歧,并为处理这些分歧提供实际建议。我们利用为严酷天气预测和次冻结道路表面温度预测而开发的ML模型来概括不同解释方法的行为。关于地物排名,对地貌特征集(例如,平均说,在10强的地貌特征中,有两种方法在10强的特征中达成一致),比具体等级(平均说,两种方法只就前10个特征组特征的2-3级地貌特征达成一致,为处理这些差异提供了实际建议。另一方面,两种不同方法的地貌效果曲线是高度一致的,因为长期以来,可比较的SHAFrical-rography rode 和最终发现,对树的概率分析结果是比较容易理解方法。