Transparency in Machine Learning (ML), attempts to reveal the working mechanisms of complex models. Transparent ML promises to advance human factors engineering goals of human-centered AI in the target users. From a human-centered design perspective, transparency is not a property of the ML model but an affordance, i.e. a relationship between algorithm and user; as a result, iterative prototyping and evaluation with users is critical to attaining adequate solutions that afford transparency. However, following human-centered design principles in healthcare and medical image analysis is challenging due to the limited availability of and access to end users. To investigate the state of transparent ML in medical image analysis, we conducted a systematic review of the literature. Our review reveals multiple severe shortcomings in the design and validation of transparent ML for medical image analysis applications. We find that most studies to date approach transparency as a property of the model itself, similar to task performance, without considering end users during neither development nor evaluation. Additionally, the lack of user research, and the sporadic validation of transparency claims put contemporary research on transparent ML for medical image analysis at risk of being incomprehensible to users, and thus, clinically irrelevant. To alleviate these shortcomings in forthcoming research while acknowledging the challenges of human-centered design in healthcare, we introduce the INTRPRT guideline, a systematic design directive for transparent ML systems in medical image analysis. The INTRPRT guideline suggests formative user research as the first step of transparent model design to understand user needs and domain requirements. Following this process produces evidence to support design choices, and ultimately, increases the likelihood that the algorithms afford transparency.
翻译:机械学习(ML)中的透明度,试图揭示复杂模型的工作机制。透明的 ML承诺在目标用户中推进以人为中心的AI的人类因素工程目标。从以人为中心的设计角度看,透明度不是ML模型的属性,而是一种开销,即算法和用户之间的关系;结果,与用户的迭代原型和评价工作对于实现具有透明度的适当解决方案至关重要。然而,在保健和医疗图像分析中遵循以人为本的设计原则,由于终端用户的可用性和获取途径有限,因此具有挑战性。为了调查医学图像分析中透明的ML的状态,我们最终对文献进行了系统审查。我们的审查表明,在设计和验证透明的ML用于医学图像分析应用程序的设计和验证方面存在多重严重缺陷。我们发现,大多数迄今为止的研究将透明度作为模型本身的一种属性,与任务性业绩相似,而没有在开发或评价期间考虑终端用户。此外,用户模型研究的缺乏,对透明性主张的验证使透明的ML用于医学图像分析的当代研究在风险下进行,因此,临床上也不相关,我们从临床角度对文献进行了审查。我们提出的系统设计模型设计设计过程中的这些缺陷分析。