As machine learning (ML) systems become increasingly widespread, it is necessary to audit these systems for biases prior to their deployment. Recent research has developed algorithms for effectively identifying intersectional bias in the form of interpretable, underperforming subsets (or slices) of the data. However, these solutions and their insights are limited without a tool for visually understanding and interacting with the results of these algorithms. We propose Visual Auditor, an interactive visualization tool for auditing and summarizing model biases. Visual Auditor assists model validation by providing an interpretable overview of intersectional bias (bias that is present when examining populations defined by multiple features), details about relationships between problematic data slices, and a comparison between underperforming and overperforming data slices in a model. Our open-source tool runs directly in both computational notebooks and web browsers, making model auditing accessible and easily integrated into current ML development workflows. An observational user study in collaboration with domain experts at Fiddler AI highlights that our tool can help ML practitioners identify and understand model biases.
翻译:随着机器学习(ML)系统日益普及,有必要对这些系统进行审计,以确定其部署前的偏差。最近的研究已经开发了各种算法,以有效识别数据可解释、业绩不佳的子集(或切片)形式的交叉偏差。然而,这些解决方案及其洞察力是有限的,没有视觉理解和与这些算法结果互动的工具。我们提议了视觉审计员,这是一个用于审计和总结模型偏差的交互式可视化工具。视觉审计员通过提供可解释的交叉偏差(在审查由多重特征界定的人口时存在的偏差)、问题数据切片之间关系的细节以及模型中表现不佳和表现不佳的数据切片之间的比较来协助模型验证。我们的开放源工具直接运行在计算笔记本和网络浏览器上,使示范审计便于使用并容易融入目前的ML开发工作流程。与Fidellr AI的域专家合作进行的观察用户研究强调,我们的工具可以帮助ML从业者识别和理解模型偏差。