Relational machine learning programs like those developed in Inductive Logic Programming (ILP) offer several advantages: (1) The ability to model complex relationships amongst data instances; (2) The use of domain-specific relations during model construction; and (3) The models constructed are human-readable, which is often one step closer to being human-understandable. However, these ILP-like methods have not been able to capitalise fully on the rapid hardware, software and algorithmic developments fuelling current developments in deep neural networks. In this paper, we treat relational features as functions and use the notion of generalised composition of functions to derive complex functions from simpler ones. We formulate the notion of a set of $\text{M}$-simple features in a mode language $\text{M}$ and identify two composition operators ($\rho_1$ and $\rho_2$) from which all possible complex features can be derived. We use these results to implement a form of "explainable neural network" called Compositional Relational Machines, or CRMs, which are labelled directed-acyclic graphs. The vertex-label for any vertex $j$ in the CRM contains a feature-function $f_j$ and a continuous activation function $g_j$. If $j$ is a "non-input" vertex, then $f_j$ is the composition of features associated with vertices in the direct predecessors of $j$. Our focus is on CRMs in which input vertices (those without any direct predecessors) all have $\text{M}$-simple features in their vertex-labels. We provide a randomised procedure for constructing and learning such CRMs. Using a notion of explanations based on the compositional structure of features in a CRM, we provide empirical evidence on synthetic data of the ability to identify appropriate explanations; and demonstrate the use of CRMs as 'explanation machines' for black-box models that do not provide explanations for their predictions.
翻译:引入逻辑编程( ILP) 所开发的那种关系机学习程序具有若干优势:(1) 模拟数据实例之间复杂关系的能力;(2) 在模型构建期间使用特定域关系;(3) 所建模型是人类可读的,这往往离人类可理解性更近一步。 然而, 这些类似 ILP 的方法无法充分利用快速硬件、 软件和算法的发展, 以刺激深层神经网络的当前发展。 在本文中, 我们把关系特性作为功能, 并使用一般化的函数构成概念, 从更简单的函数中产生复杂的函数。 我们将一组美元( text{M} ) 的简单特性概念放在模式语言 $\ text{M} 中, 并指定两个组成操作者(rho_ 1美元 和 $_ 美元) 。 我们用这些结果来实施一种“ 可解释性神经网络” 的形式, 称为“ 直流化神经机”, 或 CRMD, 以美元 美元 的直流化数据解释功能为 。