In this study, we take a departure and explore an explainability-driven strategy to data auditing, where actionable insights into the data at hand are discovered through the eyes of quantitative explainability on the behaviour of a dummy model prototype when exposed to data. We demonstrate this strategy by auditing two popular medical benchmark datasets, and discover hidden data quality issues that lead deep learning models to make predictions for the wrong reasons. The actionable insights gained from this explainability driven data auditing strategy is then leveraged to address the discovered issues to enable the creation of high-performing deep learning models with appropriate prediction behaviour. The hope is that such an explainability-driven strategy can be complimentary to data-driven strategies to facilitate for more responsible development of machine learning algorithms for computer vision applications.
翻译:在此研究中,我们先走一步,探索数据审计的可解释性战略,通过数据暴露时对模拟模型原型的行为进行量化解释,发现对手头数据可采取行动的洞察力。我们通过审计两个受欢迎的医学基准数据集,发现隐藏的数据质量问题,从而导致深层次学习模型为错误的原因作出预测。然后,利用从这一可解释性数据审计战略中获得的可采取行动的洞察力,解决已发现的问题,以便能够创建具有适当预测行为的高性能深层学习模型。希望这种可解释性战略能够与数据驱动战略相辅相成,从而推动更负责地开发计算机视觉应用的机器学习算法。