This paper addresses the question of how to align AI systems with human values and situates it within a wider body of thought regarding technology and value. Far from existing in a vacuum, there has long been an interest in the ability of technology to 'lock-in' different value systems. There has also been considerable thought about how to align technologies with specific social values, including through participatory design-processes. In this paper we look more closely at the question of AI value alignment and suggest that the power and autonomy of AI systems gives rise to opportunities and challenges in the domain of value that have not been encountered before. Drawing important continuities between the work of the fairness, accountability, transparency and ethics community, and work being done by technical AI safety researchers, we suggest that more attention needs to be paid to the question of 'social value alignment' - that is, how to align AI systems with the plurality of values endorsed by groups of people, especially on the global level.
翻译:本文探讨了如何使AI系统与人类价值相一致,并将其置于关于技术和价值的更广泛的思考范围内的问题。远非在真空中存在,长期以来人们一直关心的是技术是否有能力“锁定”不同的价值体系。对于如何使技术与特定社会价值相一致,包括采用参与性设计-过程,也进行了大量思考。本文更仔细地探讨了AI价值一致性问题,并建议AI系统的力量和自主性在价值领域带来了以前未曾遇到的机遇和挑战。在公平、问责制、透明度和道德界的工作与技术AI安全研究人员正在开展的工作之间,我们建议需要更多地注意“社会价值调整”问题,即如何使AI系统与人群所认可的价值观的多元性相一致,特别是在全球一级。