Biomedical image analysis algorithm validation depends on high-quality annotation of reference datasets, for which labeling instructions are key. Despite their importance, their optimization remains largely unexplored. Here, we present the first systematic study of labeling instructions and their impact on annotation quality in the field. Through comprehensive examination of professional practice and international competitions registered at the MICCAI Society, we uncovered a discrepancy between annotators' needs for labeling instructions and their current quality and availability. Based on an analysis of 14,040 images annotated by 156 annotators from four professional companies and 708 Amazon Mechanical Turk (MTurk) crowdworkers using instructions with different information density levels, we further found that including exemplary images significantly boosts annotation performance compared to text-only descriptions, while solely extending text descriptions does not. Finally, professional annotators constantly outperform MTurk crowdworkers. Our study raises awareness for the need of quality standards in biomedical image analysis labeling instructions.
翻译:生物医学图像分析算法的验证取决于参考数据集的高质量说明,这些数据集的标签指示是关键。尽管这些数据集很重要,但其优化在很大程度上尚未探索。在这里,我们提出首次系统研究标签指示及其对实地批注质量的影响。通过全面审查专业做法和在MICCAI协会注册的国际竞争,我们发现在批注员对标签指示的需求与目前质量和可用性之间存在差异。根据对来自四家专业公司的156名批注员和708名亚马逊机械土耳其(MTurk)人群工人使用不同信息密度水平的指示所作的14 040张图象的分析,我们进一步发现,与只用文字描述相比,包括模范图像极大地提高了说明性能,而只是扩展文字描述。最后,专业批注员不断超越MTurk人群工人的功能。我们的研究提高了对生物医学图像分析标记指示质量标准必要性的认识。