Animals are equipped with a rich innate repertoire of sensory, behavioral and motor skills, which allows them to interact with the world immediately after birth. At the same time, many behaviors are highly adaptive and can be tailored to specific environments by means of learning. In this work, we use mathematical analysis and the framework of meta-learning (or 'learning to learn') to answer when it is beneficial to learn such an adaptive strategy and when to hard-code a heuristic behavior. We find that the interplay of ecological uncertainty, task complexity and the agents' lifetime has crucial effects on the meta-learned amortized Bayesian inference performed by an agent. There exist two regimes: One in which meta-learning yields a learning algorithm that implements task-dependent information-integration and a second regime in which meta-learning imprints a heuristic or 'hard-coded' behavior. Further analysis reveals that non-adaptive behaviors are not only optimal for aspects of the environment that are stable across individuals, but also in situations where an adaptation to the environment would in fact be highly beneficial, but could not be done quickly enough to be exploited within the remaining lifetime. Hard-coded behaviors should hence not only be those that always work, but also those that are too complex to be learned within a reasonable time frame.
翻译:动物拥有丰富的感官、 行为和运动技能, 让他们在出生后立即与世界互动。 同时, 许多行为具有高度的适应性, 可以通过学习适应特定环境。 在这项工作中, 我们使用数学分析和元学习框架( 或“ 学习学习” ) 来回答是否有益于学习这种适应性战略, 以及何时硬化一种超常行为 。 我们发现, 生态不确定性、 任务复杂性和代理人一生的相互作用, 对代理人的元学摊销性贝耶斯推断作用有着至关重要的影响 。 存在两种制度: 一种是, 元学习产生一种学习算法, 执行依赖任务的信息整合, 第二种是, 元学习印印记一种超常或“ 硬码” 行为 。 进一步的分析显示, 非适应性行为不仅最适合个人之间稳定的环境方面, 而且在一种对环境的适应事实上非常有益的情况下, 也不可能很快地进行适应, 并且不会在这种复杂的生命周期内完成这些复杂的工作。