The recent phenomenal success of language models has reinvigorated machine learning research, and large sequence models such as transformers are being applied to a variety of domains. One important problem class that has remained relatively elusive however is purposeful adaptive behavior. Currently there is a common perception that sequence models "lack the understanding of the cause and effect of their actions" leading them to draw incorrect inferences due to auto-suggestive delusions. In this report we explain where this mismatch originates, and show that it can be resolved by treating actions as causal interventions. Finally, we show that in supervised learning, one can teach a system to condition or intervene on data by training with factual and counterfactual error signals respectively.
翻译:最近语言模型的惊人成功振兴了机器学习研究,并且正在将变压器等大型序列模型应用于多个领域。一个重要的问题类别仍然相对难以找到,但有目的的适应行为却是一个重要的问题类别。目前有一种普遍的看法,即序列模型“不了解其行动的原因和影响”导致它们得出不正确的推论,其原因是自发性错觉。在本报告中,我们解释这种不匹配的起源,并表明通过将行动视为因果关系干预可以解决这个问题。最后,我们表明,在监督下的学习中,我们可以教一个系统,通过分别用事实错误信号和反事实错误信号的培训来决定或干预数据。