If future AI systems are to be reliably safe in novel situations, they will need to incorporate general principles guiding them to robustly recognize which outcomes and behaviours would be harmful. Such principles may need to be supported by a binding system of regulation, which would need the underlying principles to be widely accepted. They should also be specific enough for technical implementation. Drawing inspiration from law, this article explains how negative human rights could fulfil the role of such principles and serve as a foundation both for an international regulatory system and for building technical safety constraints for future AI systems.
翻译:如果未来的AI系统在新的情况下要可靠地安全,就需要纳入指导这些系统的一般性指导原则,以有力地确认哪些结果和行为有害,可能需要有一个具有约束力的监管制度来支持这些原则,这种制度需要得到广泛接受的基本原则,它们也应足够具体,以便技术执行,从法律中得到启发,这一条解释了消极的人权如何能发挥这些原则的作用,并成为国际监管制度的基础,为未来的AI系统建立技术安全限制的基础。