Research in human-centered AI has shown the benefits of systems that can explain their predictions. Methods that allow an AI to take advice from humans in response to explanations are similarly useful. While both capabilities are well-developed for transparent learning models (e.g., linear models and GA$^2$Ms), and recent techniques (e.g., LIME and SHAP) can generate explanations for opaque models, little attention has been given to advice methods for opaque models. This paper introduces LIMEADE, the first general framework that translates both positive and negative advice (expressed using high-level vocabulary such as that employed by post-hoc explanations) into an update to an arbitrary, underlying opaque model. We demonstrate the generality of our approach with case studies on seventy real-world models across two broad domains: image classification and text recommendation. We show our method improves accuracy compared to a rigorous baseline on the image classification domains. For the text modality, we apply our framework to a neural recommender system for scientific papers on a public website; our user study shows that our framework leads to significantly higher perceived user control, trust, and satisfaction.
翻译:以人为中心的大赦国际的研究显示了能够解释其预测的系统的好处。允许大赦国际根据解释听取人类意见的方法也同样有用。虽然两种能力都为透明学习模式(如线性模型和GA$2兆元)开发良好,而且最近技术(如LIME和SHAP)可以为不透明的模型提供解释,但对不透明的模型的咨询方法很少注意。本文介绍了LIMEADE,这是第一个既正面又负面的建议(使用高层次词汇(如事后解释所使用的词汇))来更新任意的、潜在的不透明模型。我们展示了我们对70个现实世界模型的案例研究在两个广泛领域(如图象分类和文本建议)的一般做法。我们展示了我们的方法比形象分类领域的严格基线提高了准确性。关于文本模式,我们将我们的框架应用于公共网站上科学论文的神经建议系统;我们的用户研究表明,我们的框架可以大大提高对用户的认知控制、信任和满意度。