Early detection of melanoma is crucial for preventing severe complications and increasing the chances of successful treatment. Existing deep learning approaches for melanoma skin lesion diagnosis are deemed black-box models, as they omit the rationale behind the model prediction, compromising the trustworthiness and acceptability of these diagnostic methods. Attempts to provide concept-based explanations are based on post-hoc approaches, which depend on an additional model to derive interpretations. In this paper, we propose an inherently interpretable framework to improve the interpretability of concept-based models by incorporating a hard attention mechanism and a coherence loss term to assure the visual coherence of concept activations by the concept encoder, without requiring the supervision of additional annotations. The proposed framework explains its decision in terms of human-interpretable concepts and their respective contribution to the final prediction, as well as a visual interpretation of the locations where the concept is present in the image. Experiments on skin image datasets demonstrate that our method outperforms existing black-box and concept-based models for skin lesion classification.
翻译:医学图像中基于概念的连贯解释及其在皮损诊断中的应用
早期发现恶性黑色素瘤对于预防严重并发症和增加治疗成功率至关重要。现有的基于深度学习的恶性黑色素瘤皮损诊断方法被认为是黑匣子模型,因为它们省略了模型预测的原因,影响了这些诊断方法的可信度和可接受度。提供基于概念的解释的尝试基于事后方法,这些方法依赖于附加模型来推导解释。在本文中,我们提出了一种天然可解释的框架,通过将硬注意机制和连贯损失项纳入概念编码器,以确保概念激活的视觉连贯性,而无需监督附加注释,从而改善概念模型的可解释性。所提出的框架通过人类可解释概念及其对最终预测的贡献以及概念在图像中存在的位置的视觉解释来解释其决策。在皮肤图像数据集上进行的实验表明,我们的方法优于现有的黑匣子和基于概念的皮损分类模型。