The Right to Explanation and the Right to be Forgotten are two important principles outlined to regulate algorithmic decision making and data usage in real-world applications. While the right to explanation allows individuals to request an actionable explanation for an algorithmic decision, the right to be forgotten grants them the right to ask for their data to be deleted from all the databases and models of an organization. Intuitively, enforcing the right to be forgotten may trigger model updates which in turn invalidate previously provided explanations, thus violating the right to explanation. In this work, we investigate the technical implications arising due to the interference between the two aforementioned regulatory principles, and propose the first algorithmic framework to resolve the tension between them. To this end, we formulate a novel optimization problem to generate explanations that are robust to model updates due to the removal of training data instances by data deletion requests. We then derive an efficient approximation algorithm to handle the combinatorial complexity of this optimization problem. We theoretically demonstrate that our method generates explanations that are provably robust to worst-case data deletion requests with bounded costs in case of linear models and certain classes of non-linear models. Extensive experimentation with real-world datasets demonstrates the efficacy of the proposed framework.
翻译:解释权和被遗忘权是规范逻辑决策和数据在现实世界应用中的使用的两个重要原则。虽然解释权允许个人要求对算法决定提出可操作的解释,但被遗忘的权利赋予了他们要求从一个组织的所有数据库和模型中删除数据的权利。自然,执行被遗忘的权利可能会触发模型更新,反过来又使先前提供的解释无效,从而侵犯解释权。在这项工作中,我们调查了上述两项管理原则之间的干扰所产生的技术影响,并提出了解决它们之间紧张关系的第一个算法框架。为此,我们提出了一个新的优化问题,以产生对由于删除数据而删除培训数据实例而导致的模型更新非常有力的解释。然后我们得出一个高效的近似算法来处理这一优化问题的组合复杂性。我们理论上证明,我们的方法对最坏的数据删除请求的解释非常可靠,在线性模型和某些非线性模型中具有约束性成本。与现实世界数据集进行广泛的实验,展示了拟议框架的功效。