Auditing machine learning-based (ML) healthcare tools for bias is critical to preventing patient harm, especially in communities that disproportionately face health inequities. General frameworks are becoming increasingly available to measure ML fairness gaps between groups. However, ML for health (ML4H) auditing principles call for a contextual, patient-centered approach to model assessment. Therefore, ML auditing tools must be (1) better aligned with ML4H auditing principles and (2) able to illuminate and characterize communities vulnerable to the most harm. To address this gap, we propose supplementing ML4H auditing frameworks with SLOGAN (patient Severity-based LOcal Group biAs detectioN), an automatic tool for capturing local biases in a clinical prediction task. SLOGAN adapts an existing tool, LOGAN (LOcal Group biAs detectioN), by contextualizing group bias detection in patient illness severity and past medical history. We investigate and compare SLOGAN's bias detection capabilities to LOGAN and other clustering techniques across patient subgroups in the MIMIC-III dataset. On average, SLOGAN identifies larger fairness disparities in over 75% of patient groups than LOGAN while maintaining clustering quality. Furthermore, in a diabetes case study, health disparity literature corroborates the characterizations of the most biased clusters identified by SLOGAN. Our results contribute to the broader discussion of how machine learning biases may perpetuate existing healthcare disparities.
翻译:为预防病人伤害,特别是对于不成比例地面临健康不平等的社区而言,基于偏见的检查机(ML)保健工具对于预防病人伤害至关重要。一般框架越来越容易用于衡量不同群体之间ML的公平差距。然而,健康(ML4H)审计原则要求对模型评估采取以病人为中心的背景方法。因此,ML审计工具必须(1) 更好地与ML4H审计原则保持一致,(2) 能够揭示和辨别最容易受到伤害的社区。为弥补这一差距,我们提议用SLOGAN(以病人为主的Local Group biAs检测N)补充ML4H审计框架,这是在临床预测任务中捕捉地方偏见的自动工具。SLOGAN(LM4H)审计原则要求采用基于病人为主的、以病人为主、以病人为主、以病人为主的偏见检测能力与LOGAN的检测能力。我们调查SLOGAN现有子群中的偏差和其他组合技术。平均而言,SLOGAN在病人群体中找出超过75%的公平差距差距差距,同时维持我们所查明的SLOAAN的病例分析质量。