Security risks from AI have motivated calls for international agreements that guardrail the technology. However, even if states could agree on what rules to set on AI, the problem of verifying compliance might make these agreements infeasible. To help clarify the difficulty of verifying agreements on AI$\unicode{x2013}$and identify actions that might reduce this difficulty$\unicode{x2013}$this report examines the case study of verification in nuclear arms control. We review the implementation, track records, and politics of verification across three types of nuclear arms control agreements. Then, we consider implications for the case of AI, especially AI development that relies on thousands of highly specialized chips. In this context, the case study suggests that, with certain preparations, the foreseeable challenges of verification would be reduced to levels that were successfully managed in nuclear arms control. To avoid even worse challenges, substantial preparations are needed: (1) developing privacy-preserving, secure, and acceptably priced methods for verifying the compliance of hardware, given inspection access; and (2) building an initial, incomplete verification system, with authorities and precedents that allow its gaps to be quickly closed if and when the political will arises.
翻译:智能体可能产生的安全风险已经引起国际社会对智能体监管协议的呼吁。但是,即使各国能够就智能体监管问题达成协议,验证其实施可能会使这些协议变得不可行。为了帮助阐明在智能体监管协议中验证的困难以及确定可能减轻这种困难的行动,本报告研究了在核武器管制领域中进行验证的案例研究。我们回顾了核武器管制协议的三种类型的实施、记录和政治,然后考虑对智能体案例的影响,特别是依靠数以千计的高度专业化芯片进行开发的智能体。在这种情况下,案例研究表明,通过一定的准备工作,可以将预见到的验证挑战降至能够在核武器管制中成功处理的水平。为避免更严重的挑战,需要进行大量的准备工作:(1)开发保护隐私、安全和价格合理的验证硬件的方法,假设检查机制已经实施;(2)建立一个初步的、不完整的验证系统,设立主管部门和先例,以便在政治意愿出现时尽快弥补其缺陷。