Online learning via Bayes' theorem allows new data to be continuously integrated into an agent's current beliefs. However, a naive application of Bayesian methods in non stationary environments leads to slow adaptation and results in state estimates that may converge confidently to the wrong parameter value. A common solution when learning in changing environments is to discard/downweight past data; however, this simple mechanism of "forgetting" fails to account for the fact that many real-world environments involve revisiting similar states. We propose a new framework, Bayes with Adaptive Memory (BAM), that takes advantage of past experience by allowing the agent to choose which past observations to remember and which to forget. We demonstrate that BAM generalizes many popular Bayesian update rules for non-stationary environments. Through a variety of experiments, we demonstrate the ability of BAM to continuously adapt in an ever-changing world.
翻译:通过 Bayes 理论的在线学习使得新的数据能够不断融入代理商的当前信仰。 但是,在非固定环境中天真地应用Bayesian方法导致适应速度缓慢,导致国家估算结果可能自信地与错误的参数值趋同。 在变化环境中学习时的一个共同解决办法是抛弃/降低过去的数据;然而,这种简单的“忘记”机制未能考虑到许多现实世界环境涉及重新审视类似状态这一事实。 我们提出了一个新框架,即“有适应记忆的Bayes ” (BAM ), 利用过去的经验,让代理商选择过去哪些观察可以记住,哪些可以忘记。 我们证明BAM将许多受欢迎的Bayesian 更新非静止环境的规则普遍化。 通过各种实验,我们展示了BAM 在一个不断变化的世界中不断适应的能力。