Machine learning (ML) holds great promise for improving healthcare, but it is critical to ensure that its use will not propagate or amplify health disparities. An important step is to characterize the (un)fairness of ML models - their tendency to perform differently across subgroups of the population - and to understand its underlying mechanisms. One potential driver of algorithmic unfairness, shortcut learning, arises when ML models base predictions on improper correlations in the training data. However, diagnosing this phenomenon is difficult, especially when sensitive attributes are causally linked with disease. Using multi-task learning, we propose the first method to assess and mitigate shortcut learning as a part of the fairness assessment of clinical ML systems, and demonstrate its application to clinical tasks in radiology and dermatology. Finally, our approach reveals instances when shortcutting is not responsible for unfairness, highlighting the need for a holistic approach to fairness mitigation in medical AI.
翻译:机器学习(ML)对于改善医疗保健有很大的希望,但关键是要确保其使用不会传播或扩大健康差异。一个重要步骤是描述ML模式的(不)公平性 — — 它们在人口分组之间表现不同的趋势 — — 并理解其基本机制。算法不公平、捷径学习的一个潜在驱动因素是,ML模型根据培训数据中的不适当相关性作出预测。然而,分析这一现象是困难的,特别是当敏感属性与疾病有因果关系时。我们通过多任务学习,提出第一个评估和减少捷径学习的方法,作为临床ML系统公平评估的一部分,并展示其在放射学和皮肤学临床任务中的应用。最后,我们的方法揭示了在短处并不对不公平负有责任的例子,突出了在医疗护理中采取全面方法来减少公平性的必要性。