As artificial intelligence (AI) becomes more powerful and widespread, the AI alignment problem - how to ensure that AI systems pursue the goals that we want them to pursue - has garnered growing attention. This article distinguishes two types of alignment problems depending on whose goals we consider, and analyzes the different solutions necessitated by each. The direct alignment problem considers whether an AI system accomplishes the goals of the entity operating it. In contrast, the social alignment problem considers the effects of an AI system on larger groups or on society more broadly. In particular, it also considers whether the system imposes externalities on others. Whereas solutions to the direct alignment problem center around more robust implementation, social alignment problems typically arise because of conflicts between individual and group-level goals, elevating the importance of AI governance to mediate such conflicts. Addressing the social alignment problem requires both enforcing existing norms on their developers and operators and designing new norms that apply directly to AI systems.
翻译:随着人工智能(AI)变得更加强大和广泛,AI调整问题----如何确保AI系统追求我们想要追求的目标----引起了越来越多的关注。这一条根据我们考虑的目标区分两类匹配问题,并分析了每个目标所需的不同解决办法。直接调整问题考虑了AI系统是否实现了运营该系统的实体的目标。相反,社会调整问题考虑了AI系统对较大群体的影响,或对更广泛的社会的影响。特别是,它也考虑了该系统是否对其他人造成外在影响。虽然直接匹配问题的解决办法集中在更强有力的实施上,但社会调整问题通常产生于个人和群体一级目标之间的冲突,强调AI治理对于调解此类冲突的重要性。解决社会调整问题需要既对其开发者和操作者实施现行规范,又设计直接适用于AI系统的新规范。