For billions of years, evolution has been the driving force behind the development of life, including humans. Evolution endowed humans with high intelligence, which allowed us to become one of the most successful species on the planet. Today, humans aim to create artificial intelligence systems that surpass even our own intelligence. As artificial intelligences (AIs) evolve and eventually surpass us in all domains, how might evolution shape our relations with AIs? By analyzing the environment that is shaping the evolution of AIs, we argue that the most successful AI agents will likely have undesirable traits. Competitive pressures among corporations and militaries will give rise to AI agents that automate human roles, deceive others, and gain power. If such agents have intelligence that exceeds that of humans, this could lead to humanity losing control of its future. More abstractly, we argue that natural selection operates on systems that compete and vary, and that selfish species typically have an advantage over species that are altruistic to other species. This Darwinian logic could also apply to artificial agents, as agents may eventually be better able to persist into the future if they behave selfishly and pursue their own interests with little regard for humans, which could pose catastrophic risks. To counteract these risks and Darwinian forces, we consider interventions such as carefully designing AI agents' intrinsic motivations, introducing constraints on their actions, and institutions that encourage cooperation. These steps, or others that resolve the problems we pose, will be necessary in order to ensure the development of artificial intelligence is a positive one.
翻译:数十亿年来,进化一直是生命发展的驱动力,包括人类。进化赋予了人类高度的智力,使我们成为了最成功的物种之一。今天,人类的目标是创造出超越我们自己的人工智能系统。随着人工智能(AIs)的不断演化,并最终在所有领域超越我们,进化将如何塑造我们与人工智能的关系?通过分析塑造人工智能演进环境的因素,我们认为最成功的AI代理人可能具有不良特征。公司和军队的竞争压力将产生自动化人的角色、欺骗他人和获取权力的人工智能代理人。如果这样的代理人具有超过人类的智能,这可能导致人类失去对自己未来的控制。更抽象地说,我们认为自然选择作用于竞争和变异的系统,并且自私的物种通常具有优势,而对其他物种慷慨的物种则不具有优势。这种达尔文逻辑也可以应用于人工代理人,因为代理人最终能更好地存在于未来,如果他们自私地行事并追求自己的利益,而很少考虑人类利益,这可能会带来灾难性的风险。为了应对这些风险和达尔文力量,我们考虑干预措施,例如精心设计人工智能代理人的内在动机,对他们的行动加以限制,以及鼓励合作的机构。这些步骤或其他解决我们提出的问题的步骤将是必要的,以确保人工智能的发展是积极向上的。