Modern machine learning and computer science conferences are experiencing a surge in the number of submissions that challenges the quality of peer review as the number of competent reviewers is growing at a much slower rate. To curb this trend and reduce the burden on reviewers, several conferences have started encouraging or even requiring authors to declare the previous submission history of their papers. Such initiatives have been met with skepticism among authors, who raise the concern about a potential bias in reviewers' recommendations induced by this information. In this work, we investigate whether reviewers exhibit a bias caused by the knowledge that the submission under review was previously rejected at a similar venue, focusing on a population of novice reviewers who constitute a large fraction of the reviewer pool in leading machine learning and computer science conferences. We design and conduct a randomized controlled trial closely replicating the relevant components of the peer-review pipeline with $133$ reviewers (master's, junior PhD students, and recent graduates of top US universities) writing reviews for $19$ papers. The analysis reveals that reviewers indeed become negatively biased when they receive a signal about paper being a resubmission, giving almost 1 point lower overall score on a 10-point Likert item ($\Delta = -0.78, \ 95\% \ \text{CI} = [-1.30, -0.24]$) than reviewers who do not receive such a signal. Looking at specific criteria scores (originality, quality, clarity and significance), we observe that novice reviewers tend to underrate quality the most.
翻译:现代机器学习和计算机科学会议正在经历挑战同行审议质量的提交材料数量剧增,因为主管审查者的数量正在以慢得多的速度增长,从而挑战同行审议质量。为遏制这一趋势并减轻审查者的负担,一些会议已经开始鼓励甚至要求作者宣布其以前提交文件的历史。这类举措在作者中受到怀疑,他们对这一信息引发的对审查者建议中潜在偏差的关切提出怀疑。在这项工作中,我们调查审查者是否表现出偏见,因为知道所审查的提交材料以前在类似地点被拒绝了,重点是在领导机器学习和计算机科学会议方面占审查者人才库大部分的新审查者群体。我们设计并进行随机化的控制性试验,与133美元的审查者(主任、初级博士学生和最近美国高级大学毕业生)一起,对19美元的文件进行书面审查。分析表明,当他们收到关于重新提交文件的信号时,审查者确实存在负面偏差,在10美元基点(D=xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx