Abstract reasoning ability is fundamental to human intelligence. It enables humans to uncover relations among abstract concepts and further deduce implicit rules from the relations. As a well-known abstract visual reasoning task, Raven's Progressive Matrices (RPM) are widely used in human IQ tests. Although extensive research has been conducted on RPM solvers with machine intelligence, few studies have considered further advancing the standard answer-selection (classification) problem to a more challenging answer-painting (generating) problem, which can verify whether the model has indeed understood the implicit rules. In this paper we aim to solve the latter one by proposing a deep latent variable model, in which multiple Gaussian processes are employed as priors of latent variables to separately learn underlying abstract concepts from RPMs; thus the proposed model is interpretable in terms of concept-specific latent variables. The latent Gaussian process also provides an effective way of extrapolation for answer painting based on the learned concept-changing rules. We evaluate the proposed model on RPM-like datasets with multiple continuously-changing visual concepts. Experimental results demonstrate that our model requires only few training samples to paint high-quality answers, generate novel RPM panels, and achieve interpretability through concept-specific latent variables.
翻译:抽象推理能力是人类智力的基础。它使人类能够发现抽象概念之间的关系,并从关系中进一步推断出隐含的规则。作为众所周知的抽象抽象直观推理任务,Raven的累进矩阵(RPM)在人类智商测试中被广泛使用。虽然对机器智能的RPM解答器进行了广泛的研究,但很少有研究考虑进一步将标准答案选择(分类)问题推向更具挑战性的答案绘制(生成)问题,以核实模型是否确实理解了隐含的规则。在本文中,我们的目标是通过提出一种深潜潜变模型来解决后一种问题。在这个模型中,多个高斯进程被用作潜在变量的前身,分别从RPM中学习基本抽象概念;因此,拟议的模型可以按概念特定的潜在变量加以解释。潜伏高斯进程还提供了一种有效的外推法,用以根据所学的概念变化规则对答案画出答案。我们用多种不断改变的视觉概念来评估RPMM的模型的拟议模型模式。实验结果表明,我们的模型只需要很少的培训样本来绘制高清晰度的变量。