Inductive Logic Programming (ILP) is a form of machine learning (ML) which in contrast to many other state of the art ML methods typically produces highly interpretable and reusable models. However, many ILP systems lack the ability to naturally learn from any noisy or partially misclassified training data. We introduce the relaxed learning from failures approach to ILP, a noise handling modification of the previously introduced learning from failures (LFF) approach which is incapable of handling noise. We additionally introduce the novel Noisy Popper ILP system which implements this relaxed approach and is a modification of the existing Popper system. Like Popper, Noisy Popper takes a generate-test-constrain loop to search its hypothesis space wherein failed hypotheses are used to construct hypothesis constraints. These constraints are used to prune the hypothesis space, making the hypothesis search more efficient. However, in the relaxed setting, constraints are generated in a more lax fashion as to avoid allowing noisy training data to lead to hypothesis constraints which prune optimal hypotheses. Constraints unique to the relaxed setting are generated via hypothesis comparison. Additional constraints are generated by weighing the accuracy of hypotheses against their sizes to avoid overfitting through an application of the minimum description length. We support this new setting through theoretical proofs as well as experimental results which suggest that Noisy Popper improves the noise handling capabilities of Popper but at the cost of overall runtime efficiency.
翻译:引导逻辑编程(ILP)是一种机器学习的形式(ML),它与其他许多最先进的ML方法相比,通常会产生高度可解释和可再使用的模型。然而,许多ILP系统缺乏自然地从任何吵闹或部分分类错误的培训数据中自然学习的能力。我们向ILP引入了从失败中轻松学习的方法,这是以前从失败中学习的方法(LFF)的一个噪音处理修改,无法处理噪音。我们还引入了新颖的Noisy Popper ILP系统(ML),它实施这种宽松的方法,是对现有流行器系统的修改。像Popper一样,Noisy Popper采用生成的生成测试-测试-螺旋环来搜索其假设空间,而其中没有使用失败的假设来构建假设限制。我们使用这些限制来利用假设空间,使假设搜索更有效率。然而,在宽松的环境下,限制的产生更为宽松的方式是,避免让噪音培训数据导致假设的制约,而这种优化的假设则是通过假设的比较而产生独特的限制。额外的限制,因为通过权衡来调整这个假设的精确度的精确性,因此,我们没有根据试验的理论的精确的精确性来评估的精确性结果,因此而提出。