Background: Recent advancements in large language models have motivated the practical use of such models in code generation and program synthesis. However, little is known about the effects of such tools on code readability and visual attention in practice. Objective: In this paper, we focus on GitHub Copilot to address the issues of readability and visual inspection of model generated code. Readability and low complexity are vital aspects of good source code, and visual inspection of generated code is important in light of automation bias. Method: Through a human experiment (n=21) we compare model generated code to code written completely by human programmers. We use a combination of static code analysis and human annotators to assess code readability, and we use eye tracking to assess the visual inspection of code. Results: Our results suggest that model generated code is comparable in complexity and readability to code written by human pair programmers. At the same time, eye tracking data suggests, to a statistically significant level, that programmers direct less visual attention to model generated code. Conclusion: Our findings highlight that reading code is more important than ever, and programmers should beware of complacency and automation bias with model generated code.
翻译:目标:在本文件中,我们侧重于GitHub Copilot, 以解决模型生成代码的可读性和视觉检查问题。可读性和低复杂性是良好源代码的重要方面,而对生成代码的直观检查在自动化偏向方面很重要。方法:通过人类实验(n=21),我们将生成的代码与人类程序员完全编写的代码进行比较。我们使用静态代码分析和人文识别器来评估代码可读性,我们使用眼睛跟踪来评估代码的可视检查。结果:我们的结果显示,生成的代码在复杂性和可读性上与人对配程序程序员编写的代码相仿。与此同时,眼跟踪数据表明,从统计上看,程序员对生成的代码的视觉关注较少。结论:我们的调查结果突出表明,阅读代码比以往更加重要,程序员应当对生成的代码保持自满和自动化的偏向。