Labelled "ground truth" datasets are routinely used to evaluate and audit AI algorithms applied in high-stakes settings. However, there do not exist widely accepted benchmarks for the quality of labels in these datasets. We provide empirical evidence that quality of labels can significantly distort the results of algorithmic audits in real-world settings. Using data annotators typically hired by AI firms in India, we show that fidelity of the ground truth data can lead to spurious differences in performance of ASRs between urban and rural populations. After a rigorous, albeit expensive, label cleaning process, these disparities between groups disappear. Our findings highlight how trade-offs between label quality and data annotation costs can complicate algorithmic audits in practice. They also emphasize the need for development of consensus-driven, widely accepted benchmarks for label quality.
翻译:“地面真相”数据集被例行用于评价和审计在高取量环境中应用的AI算法,然而,对于这些数据集中的标签质量,还没有得到广泛接受的基准。我们提供了经验证据,证明标签质量可以大大扭曲现实世界环境中的算法审计结果。使用大赦国际公司在印度通常雇用的数据批注员,我们表明地面真相数据的忠实性可能导致城市和农村人口在ASR绩效上的虚假差异。经过严格但昂贵的标签清理程序,这些群体之间的差别消失了。我们的调查结果突出表明,标签质量和数据批注成本之间的取舍会如何使实际的算法审计复杂化。他们还强调,需要为标签质量制定协商一致的、普遍接受的基准。