Most AI systems are black boxes generating reasonable outputs for given inputs. Some domains, however, have explainability and trustworthiness requirements that cannot be directly met by these approaches. Various methods have therefore been developed to interpret black-box models after training. This paper advocates an alternative approach where the models are transparent and explainable to begin with. This approach, EVOTER, evolves rule-sets based on simple logical expressions. The approach is evaluated in several prediction/classification and prescription/policy search domains with and without a surrogate. It is shown to discover meaningful rule sets that perform similarly to black-box models. The rules can provide insight to the domain, and make biases hidden in the data explicit. It may also be possible to edit them directly to remove biases and add constraints. EVOTER thus forms a promising foundation for building trustworthy AI systems for real-world applications in the future.
翻译:大多数AI系统都是为特定投入产生合理产出的黑盒,但有些领域具有解释性和可信赖性要求,这些要求无法直接由这些方法满足,因此,已经制定了各种方法来解释培训后的黑盒模型。本文主张一种替代方法,即模型透明和可以解释。EVOTER(EVOTER)这个方法基于简单的逻辑表达方式发展规则集。这个方法在若干预测/分类和处方/政策搜索域中加以评价,并且没有替代功能。它显示它发现了与黑盒模型相似的有意义的规则集。规则可以提供对域的洞察力,使数据中隐藏的偏差变得明确。也可以直接编辑它们,以消除偏见和增加限制。EVOTER因此为今后建立可靠的用于现实世界应用的AI系统奠定了良好的基础。