Collective behavior is widespread across the animal kingdom. To date, however, the developmental and mechanistic foundations of collective behavior have not been formally established. What learning mechanisms drive the development of collective behavior in newborn animals? Here, we used deep reinforcement learning and curiosity-driven learning -- two learning mechanisms deeply rooted in psychological and neuroscientific research -- to build newborn artificial agents that develop collective behavior. Like newborn animals, our agents learn collective behavior from raw sensory inputs in naturalistic environments. Our agents also learn collective behavior without external rewards, using only intrinsic motivation (curiosity) to drive learning. Specifically, when we raise our artificial agents in natural visual environments with groupmates, the agents spontaneously develop ego-motion, object recognition, and a preference for groupmates, rapidly learning all of the core skills required for collective behavior. This work bridges the divide between high-dimensional sensory inputs and collective action, resulting in a pixels-to-actions model of collective animal behavior. More generally, we show that two generic learning mechanisms -- deep reinforcement learning and curiosity-driven learning -- are sufficient to learn collective behavior from unsupervised natural experience.
翻译:集体行为在整个动物王国广泛存在。 但是,到目前为止,集体行为的发展和机械基础还没有正式建立。 是什么学习机制驱动新生动物的集体行为发展? 在这里,我们使用了深强化学习和好奇心驱动的学习 -- -- 两个深深植根于心理和神经科学研究的学习机制 -- -- 来培养形成集体行为的新生人工剂。像新生动物一样,我们的代理人从自然环境中的原始感官投入中学习集体行为。我们的代理人还学习没有外部奖励的集体行为,只使用内在动机(精度)来推动学习。具体地说,当我们在自然视觉环境中与群友一起培养我们的人工剂时,这些代理人自发地发展自我感动、物体识别和偏好集体行为所需的所有核心技能,迅速学习集体行为所需的所有核心技能。这项工作弥合了高维感官投入和集体行动之间的鸿沟,从而形成集体动物行为的像素-行动模式。 更一般地说,我们显示两种普通学习机制 -- 深度强化学习和由好奇心驱动的学习 -- -- 足以从不受控制的自然经验中学习集体行为。