Interpretable machine learning has been focusing on explaining final models that optimize performance. The current state-of-the-art is the Shapley additive explanations (SHAP) that locally explains variable impact on individual predictions, and it is recently extended for a global assessment across the dataset. Recently, Dong and Rudin proposed to extend the investigation to models from the same class as the final model that are "good enough", and identified a previous overclaim of variable importance based on a single model. However, this method does not directly integrate with existing Shapley-based interpretations. We close this gap by proposing a Shapley variable importance cloud that pools information across good models to avoid biased assessments in SHAP analyses of final models, and communicate the findings via novel visualizations. We demonstrate the additional insights gain compared to conventional explanations and Dong and Rudin's method using criminal justice and electronic medical records data.
翻译:解释性机器学习一直侧重于解释最佳性能的最终模型。 目前的先进技术是“ Shapley ” 添加解释(SHAP ), 该解释在当地解释了对个别预测的可变影响,最近该解释范围扩大到整个数据集的全球评估。 最近, Dong 和 Rudin 提议将调查范围扩大到与“足够好”的最后模型同一类的模型。 并基于一个单一模型,确定了先前一个具有可变重要性的多称。 但是,这一方法并没有直接与现有的“ Shapley ” 解释相结合。 我们缩小了这一差距,提出一个“ Shapley ” 变式重要云, 将信息汇集在对最终模型的分析中避免偏差,并通过新颖的可视化来传播结果。 我们展示了与传统解释以及使用刑事司法和电子病历数据的方式相比获得的更多洞察力。