Compositional and relational learning is a hallmark of human intelligence, but one which presents challenges for neural models. One difficulty in the development of such models is the lack of benchmarks with clear compositional and relational task structure on which to systematically evaluate them. In this paper, we introduce an environment called ConceptWorld, which enables the generation of images from compositional and relational concepts, defined using a logical domain specific language. We use it to generate images for a variety of compositional structures: 2x2 squares, pentominoes, sequences, scenes involving these objects, and other more complex concepts. We perform experiments to test the ability of standard neural architectures to generalize on relations with compositional arguments as the compositional depth of those arguments increases and under substitution. We compare standard neural networks such as MLP, CNN and ResNet, as well as state-of-the-art relational networks including WReN and PrediNet in a multi-class image classification setting. For simple problems, all models generalize well to close concepts but struggle with longer compositional chains. For more complex tests involving substitutivity, all models struggle, even with short chains. In highlighting these difficulties and providing an environment for further experimentation, we hope to encourage the development of models which are able to generalize effectively in compositional, relational domains.
翻译:人类智慧的特征是构成和关系学习,但它是人类智力的标志,但这种知识是神经模型的挑战。这种模型的开发困难之一是缺乏具有明确组成和关系任务结构的基准,因此无法系统地评估这些模型。在本文中,我们引入了一个称为概念世界的环境,它能够从组成和关系概念中生成图像,这种概念概念以逻辑领域特定语言界定。我们用它为各种组成结构生成图像:2x2方形、笔记、序列、涉及这些对象的场景和其他更复杂的概念。我们进行实验,测试标准神经结构是否有能力将与构成参数的关系概括化,因为这些参数的构成深度增加和处于替代状态。我们比较了标准神经网络,例如MLP、CNN和ResNet,以及包括WReN和PrediNet等最新的最新关系网络,以多级图像分类为背景。对于简单的问题,所有模型都概括了概念,但都与较长期的构成链条相争。我们进行了试验,比较了更复杂的测试,这些模型都与构成的深度相交织关系,甚至与短链条的模型都能够有效地促进发展。