Machine Ethics (ME) is concerned with the design of Artificial Moral Agents (AMAs), i.e. autonomous agents capable of reasoning and behaving according to moral values. Previous approaches have treated values as labels associated with some actions or states of the world, rather than as integral components of agent reasoning. It is also common to disregard that a value-guided agent operates alongside other value-guided agents in an environment governed by norms, thus omitting the social dimension of AMAs. In this blue sky paper, we propose a new AMA paradigm grounded in moral and social psychology, where values are instilled into agents as context-dependent goals. These goals intricately connect values at individual levels to norms at a collective level by evaluating the outcomes most incentivized by the norms in place. We argue that this type of normative reasoning, where agents are endowed with an understanding of norms' moral implications, leads to value-awareness in autonomous agents. Additionally, this capability paves the way for agents to align the norms enforced in their societies with respect to the human values instilled in them, by complementing the value-based reasoning on norms with agreement mechanisms to help agents collectively agree on the best set of norms that suit their human values. Overall, our agent model goes beyond the treatment of values as inert labels by connecting them to normative reasoning and to the social functionalities needed to integrate value-aware agents into our modern hybrid human-computer societies.
翻译:机器伦理学(ME)关注人工道德媒介的设计,即能够理性和按照道德价值观行事的自主代理人;以往的做法将价值观视为与世界上某些行动或国家有关的标签,而不是代理推理的有机组成部分;同样常见的情况是,忽视了价值指导代理人与其他价值指导代理人在受规范制约的环境中一起运作,从而忽略了AMAs的社会层面;在这份蓝天文件中,我们提出了一个新的基于道德和社会心理学的AMA模式,其中价值观被注入代理人,成为依背景而定的目标;这些目标通过评估现有规范最激励的结果,将个人层面的价值观与集体层面的规范紧密地联系在一起;我们认为,这种规范推理,即代理人具有对规范的道德影响的理解,从而导致自主代理人的价值观意识;此外,这一能力为代理人使其社会上执行的规范与灌输的人类价值观相一致铺平了道路,通过补充基于价值观的理论推理,将规范与现有规范中最激励的规范结合起来,帮助代理人集体地将自己作为社会层面的规范的规范纳入到符合其要求的规范的规范之中;我们认为,使代理人能够集体地商定其社会理念,从而同意将人类的规范的推理。