Current deep learning approaches have shown good in-distribution generalization performance, but struggle with out-of-distribution generalization. This is especially true in the case of tasks involving abstract relations like recognizing rules in sequences, as we find in many intelligence tests. Recent work has explored how forcing relational representations to remain distinct from sensory representations, as it seems to be the case in the brain, can help artificial systems. Building on this work, we further explore and formalize the advantages afforded by 'partitioned' representations of relations and sensory details, and how this inductive bias can help recompose learned relational structure in newly encountered settings. We introduce a simple architecture based on similarity scores which we name Compositional Relational Network (CoRelNet). Using this model, we investigate a series of inductive biases that ensure abstract relations are learned and represented distinctly from sensory data, and explore their effects on out-of-distribution generalization for a series of relational psychophysics tasks. We find that simple architectural choices can outperform existing models in out-of-distribution generalization. Together, these results show that partitioning relational representations from other information streams may be a simple way to augment existing network architectures' robustness when performing out-of-distribution relational computations.
翻译:目前深层的学习方法在分布上表现良好,但在分布上表现良好,但与分布上一般化斗争。我们在许多情报测试中发现,在涉及抽象关系的任务中尤其如此,例如承认序列规则等抽象关系,正如我们在许多情报测试中发现的那样。最近的工作探索了如何迫使关系表达方式与感觉表达方式保持区别,这在大脑中似乎是这样的情况,有助于人工系统。在这项工作的基础上,我们进一步探索并正式确定“分离”关系和感官细节的表达方式所提供的优势,以及这种进化性偏见如何有助于在新遇到的环境下重新配置学习到的关系结构。我们引入了一个基于相似性分数的简单结构,我们称之为构成关系网络(CoRelNet) 。我们利用这一模式,调查了一系列确保抽象关系得到学习并有别于感官数据的感性偏差偏差,并探索其对一系列关系心理物理学任务分配外一般化的影响。我们发现,简单的建筑选择可以超越在分配上普遍化的现有模式。这些结果共同表明,在从现有信息流到结构的稳健化关系中,分化关系关系结构可以从现有结构向其他结构演变演变为简单。