The Competition on Software Verification (SV-COMP) is a large computational experiment benchmarking many different software verification tools on a vast collection of C and Java benchmarks. Such experimental research should be reproducible by researchers independent from the team that performed the original experiments. In this reproduction report, we present our recent attempt at reproducing SV-COMP 2023: We chose a meaningful subset of the competition and re-ran it on the competition organiser's infrastructure, using the scripts and tools provided in the competition's archived artifacts. We see minor differences in tool scores that appear explainable by the interaction of small runtime fluctuations with the competition's scoring rules, and successfully reproduce the overall ranking within our chosen subset. Overall, we consider SV-COMP 2023 to be reproducible.
翻译:软件核查竞争(SV-COMP)是一个庞大的计算实验,它用大量C和爪哇基准集衡量许多不同的软件核查工具。这种实验研究应该由独立于最初实验团队的研究人员复制。在本复制报告中,我们介绍了我们最近试图复制SV-COMP 2023:我们选择了竞争的一个有意义的子集,并用竞争档案文物提供的脚本和工具重新将它重新用于竞争组织者的基础设施。我们看到工具分数略有差异,这些差异似乎可以通过小运行时间波动与竞争评分规则的相互作用来解释,并且成功地复制了我们所选择的子集的总排名。总的来说,我们认为SV-COMP 2023是可以复制的。</s>