Recent advancements in explainable machine learning provide effective and faithful solutions for interpreting model behaviors. However, many explanation methods encounter efficiency issues, which largely limit their deployments in practical scenarios. Real-time explainer (RTX) frameworks have thus been proposed to accelerate the model explanation process by learning a one-feed-forward explainer. Existing RTX frameworks typically build the explainer under the supervised learning paradigm, which requires large amounts of explanation labels as the ground truth. Considering that accurate explanation labels are usually hard to obtain due to constrained computational resources and limited human efforts, effective explainer training is still challenging in practice. In this work, we propose a COntrastive Real-Time eXplanation (CoRTX) framework to learn the explanation-oriented representation and relieve the intensive dependence of explainer training on explanation labels. Specifically, we design a synthetic strategy to select positive and negative instances for the learning of explanation. Theoretical analysis show that our selection strategy can benefit the contrastive learning process on explanation tasks. Experimental results on three real-world datasets further demonstrate the efficiency and efficacy of our proposed CoRTX framework.
翻译:最近的可解释的机器学习进展为解释模型行为提供了有效而忠实的解决方案。然而,许多解释方法都遇到效率问题,这在很大程度上限制了在实际情况下的部署。因此,提出了实时解释(RTX)框架,以通过学习一元前方解释来加速示范解释过程。现有的RTX框架通常在监督的学习模式下建立解释者,这需要大量的解释标签作为地面真理。考虑到由于计算资源有限和人力有限,准确的解释标签通常难以获得,有效的解释培训在实践中仍然具有挑战性。我们在此工作中提议建立一个实时实时解释(RTX)框架,以学习面向解释的表述,并减轻解释者在解释标签方面的密集依赖。具体地说,我们设计了一个综合战略,以选择积极的和消极的例子来学习解释。理论分析表明,我们的选择战略可以有利于在解释任务上对比性学习过程。三个真实世界数据集的实验结果进一步展示了我们提议的CORTX框架的效率和效力。</s>