Codex, a large language model (LLM) trained on a variety of codebases, exceeds the previous state of the art in its capacity to synthesize and generate code. Although Codex provides a plethora of benefits, models that may generate code on such scale have significant limitations, alignment problems, the potential to be misused, and the possibility to increase the rate of progress in technical fields that may themselves have destabilizing impacts or have misuse potential. Yet such safety impacts are not yet known or remain to be explored. In this paper, we outline a hazard analysis framework constructed at OpenAI to uncover hazards or safety risks that the deployment of models like Codex may impose technically, socially, politically, and economically. The analysis is informed by a novel evaluation framework that determines the capacity of advanced code generation techniques against the complexity and expressivity of specification prompts, and their capability to understand and execute them relative to human ability.
翻译:代码x是一个在各种代码库中受过训练的大型语言模型(LLM),它超越了以前在综合和生成代码能力方面的先进水平。虽然代码x提供了大量好处,但可能生成如此规模的代码的模型存在重大局限性、一致性问题、被滥用的可能性以及提高技术领域进展速度的可能性,而这些技术领域本身可能具有破坏稳定的影响或滥用潜力。然而,这种安全影响尚不清楚或有待探讨。在本文件中,我们概述了在开放国际公司建立的危险分析框架,以发现在技术上、社会上、政治上和经济上使用代码等模型可能造成的危害或安全风险。这一分析参考了一个新的评估框架,它决定了高级代码生成技术的能力,以对抗规格提示的复杂性和清晰度,以及它们相对于人的能力来理解和执行这些技术的能力。