Open-world novelty occurs when the rules of an environment can change abruptly, such as when a game player encounters "house rules". To address open-world novelty, game playing agents must be able to detect when novelty is injected, and to quickly adapt to the new rules. We propose a model-based reinforcement learning approach where game state and rules are represented as knowledge graphs. The knowledge graph representation of the state and rules allows novelty to be detected as changes in the knowledge graph, assists with the training of deep reinforcement learners, and enables imagination-based re-training where the agent uses the knowledge graph to perform look-ahead.
翻译:当环境规则突然改变时,例如游戏玩家遇到“内部规则”时,就会出现开放世界的新规则。为了解决开放世界的新规则问题,游戏玩家必须能够发现新规则注入时,并能迅速适应新规则。我们提出了一个基于模型的强化学习方法,其中游戏状态和规则以知识图表形式出现。 国家和规则的知识图显示允许将新规则作为知识图的变化被检测出来,帮助培训深层强化学习者,并在游戏玩家使用知识图进行外观分析时,进行基于想象的再培训。