State-of-the-art machine learning models often learn spurious correlations embedded in the training data. This poses risks when deploying these models for high-stake decision-making, such as in medical applications like skin cancer detection. To tackle this problem, we propose Reveal to Revise (R2R), a framework entailing the entire eXplainable Artificial Intelligence (XAI) life cycle, enabling practitioners to iteratively identify, mitigate, and (re-)evaluate spurious model behavior with a minimal amount of human interaction. In the first step (1), R2R reveals model weaknesses by finding outliers in attributions or through inspection of latent concepts learned by the model. Secondly (2), the responsible artifacts are detected and spatially localized in the input data, which is then leveraged to (3) revise the model behavior. Concretely, we apply the methods of RRR, CDEP and ClArC for model correction, and (4) (re-)evaluate the model's performance and remaining sensitivity towards the artifact. Using two medical benchmark datasets for Melanoma detection and bone age estimation, we apply our R2R framework to VGG, ResNet and EfficientNet architectures and thereby reveal and correct real dataset-intrinsic artifacts, as well as synthetic variants in a controlled setting. Completing the XAI life cycle, we demonstrate multiple R2R iterations to mitigate different biases. Code is available on https://github.com/maxdreyer/Reveal2Revise.
翻译:最先进的机器学习模型往往会学习嵌入在训练数据中的假相关性。当将这些模型部署于高风险的决策制定中,例如皮肤癌检测等医疗应用时,这会带来风险。为了解决这个问题,我们提出了“揭示到修订”(R2R)框架,涵盖了整个可解释人工智能(XAI)生命周期,使从业人员能够迭代地识别、缓解和(重新)评估具有最少人类干预的虚假模型行为。 在第一步(1)中,R2R通过找到指标中的异常值或通过检查模型学习的潜在概念来揭示模型的缺陷。其次(2),检测到的相关因素会在输入数据中被识别并在其上执行空间定位,然后用于(3)修改模型行为。具体而言,我们应用RRR、CDEP和ClArC的方法进行模型修正,然后(4)(重新)评估模型的性能和对问题的剩余敏感性。利用Melanoma检测和骨龄估计两个医疗基准数据集,我们将我们的R2R框架应用于VGG、ResNet和EfficientNet架构,并在控制实验中揭示和纠正了真实的数据集内在相关因素及其合成变体。通过完成XAI生命周期,我们展示了多个R2R迭代,以缓解不同的偏见。代码可在https://github.com/maxdreyer/Reveal2Revise上获得。