One of the most important lessons from the success of deep learning is that learned representations tend to perform much better at any task compared to representations we design by hand. Yet evolution of evolvability algorithms, which aim to automatically learn good genetic representations, have received relatively little attention, perhaps because of the large amount of computational power they require. The recent method Evolvability ES allows direct selection for evolvability with little computation. However, it can only be used to solve problems where evolvability and task performance are aligned. We propose Quality Evolvability ES, a method that simultaneously optimizes for task performance and evolvability and without this restriction. Our proposed approach Quality Evolvability has similar motivation to Quality Diversity algorithms, but with some important differences. While Quality Diversity aims to find an archive of diverse and well-performing, but potentially genetically distant individuals, Quality Evolvability aims to find a single individual with a diverse and well-performing distribution of offspring. By doing so Quality Evolvability is forced to discover more evolvable representations. We demonstrate on robotic locomotion control tasks that Quality Evolvability ES, similarly to Quality Diversity methods, can learn faster than objective-based methods and can handle deceptive problems.
翻译:从深层学习的成功中,最重要的教训之一是,与我们手头设计的演示相比,学习到的表述在任何任务上效果都好得多。然而,旨在自动学习良好遗传表现的演进性演算法的演变相对没有受到多少关注,这或许是因为它们需要大量的计算能力。最近的方法“可变性ES”允许直接选择可变性,而没有多少计算。然而,它只能用于解决可变性和任务性能一致的问题。我们提出了质量可变性ES,这是一种既能优化任务性能和可变性,又不受这种限制的方法。我们提议的方法“可变性能”对质量多样性的演算法有着相似的动机,但也有一些重要的不同之处。虽然质量多样性的目的是找到多样化和业绩良好的个人档案,但可能具有基因上的遥远者,但质量可变性只能用来找到一个具有多样化和工作性能的后代分布的单一个人。我们提出“可变性能”是质量的,这种方法既能优化,又没有这种限制。我们提议的“可变性”方法具有相似的动机,但是也有一些重要的差异。虽然“可变性,但是质量多样性的目的是找到“可变性,但是,我们可以更快地处理“可变性”的方法可以处理“可变性的方法。