Decision trees are a popular choice of explainable model, but just like neural networks, they suffer from adversarial examples. Existing algorithms for fitting decision trees robust against adversarial examples are greedy heuristics and lack approximation guarantees. In this paper we propose ROCT, a collection of methods to train decision trees that are optimally robust against user-specified attack models. We show that the min-max optimization problem that arises in adversarial learning can be solved using a single minimization formulation for decision trees with 0-1 loss. We propose such formulations in Mixed-Integer Linear Programming and Maximum Satisfiability, which widely available solvers can optimize. We also present a method that determines the upper bound on adversarial accuracy for any model using bipartite matching. Our experimental results demonstrate that the existing heuristics achieve close to optimal scores while ROCT achieves state-of-the-art scores.
翻译:决策树是人们广泛选择的可解释的模式,但和神经网络一样,它们也遭受对抗性的例子。现有的设计决策树的算法是贪婪的惯性,缺乏近似保证。在本文中,我们提出了一套对用户指定的攻击模式最有力的决策树培训方法。我们表明,对抗性学习中产生的微量最大优化问题可以通过对0-1损失的决策树的单一最小化配方来解决。我们提出了混合- Integer线性规划和最大可满足性中的这种配方,这些配方是广泛存在的解决方案可以优化的。我们还提出了一个方法,用以确定任何模型使用双面匹配的对抗性准确度上限。我们的实验结果表明,现有的超量性在达到最优分数的同时,ROCT也取得了最先进的分数。