Classification models learn to generalize the associations between data samples and their target classes. However, researchers have increasingly observed that machine learning practice easily leads to systematic errors in AI applications, a phenomenon referred to as AI blindspots. Such blindspots arise when a model is trained with training samples (e.g., cat/dog classification) where important patterns (e.g., black cats) are missing or periphery/undesirable patterns (e.g., dogs with grass background) are misleading towards a certain class. Even more sophisticated techniques cannot guarantee to capture, reason about, and prevent the spurious associations. In this work, we propose ESCAPE, a visual analytic system that promotes a human-in-the-loop workflow for countering systematic errors. By allowing human users to easily inspect spurious associations, the system facilitates users to spontaneously recognize concepts associated misclassifications and evaluate mitigation strategies that can reduce biased associations. We also propose two statistical approaches, relative concept association to better quantify the associations between a concept and instances, and debias method to mitigate spurious associations. We demonstrate the utility of our proposed ESCAPE system and statistical measures through extensive evaluation including quantitative experiments, usage scenarios, expert interviews, and controlled user experiments.
翻译:分类模型学习泛化数据样本与目标类别之间的关联。然而研究人员越来越发现在人工智能应用中机器学习很容易导致系统性误差,这种现象称为人工智能“盲点”。当模型使用缺少重要模式(例如黑猫)或边缘/不必要模式(例如有草地背景的狗)的训练样本(例如猫/狗分类)进行训练时,就会出现此类盲点。即使采用更为复杂的技术,也无法保证捕捉、推理和防止虚假联想。在这项工作中,我们提出了一种名为ESCAPE的视觉分析系统,它因采用“人在环流”工作流程而可以对抗系统性误差。通过允许用户轻松检查虚假联想,该系统可帮助用户自我学习有关与误分类相关的概念,并评估可减少偏差关联的缓解策略。我们还提出了两种统计方法:相对概念关联用于更好地量化概念与实例之间的关联,以及Debias方法用于减轻虚假联想。我们通过广泛的评估来演示所提出的ESCAPE系统和统计测量工具的效用,包括定量实验、使用场景、专家访谈和受控用户实验。