Machine Learning (ML) models are increasingly used to make critical decisions in real-world applications, yet they have become more complex, making them harder to understand. To this end, researchers have proposed several techniques to explain model predictions. However, practitioners struggle to use these explainability techniques because they often do not know which one to choose and how to interpret the results of the explanations. In this work, we address these challenges by introducing TalkToModel: an interactive dialogue system for explaining machine learning models through conversations. Specifically, TalkToModel comprises of three key components: 1) a natural language interface for engaging in conversations, making ML model explainability highly accessible, 2) a dialogue engine that adapts to any tabular model and dataset, interprets natural language, maps it to appropriate explanations, and generates text responses, and 3) an execution component that constructs the explanations. We carried out extensive quantitative and human subject evaluations of TalkToModel. Overall, we found the conversational system understands user inputs on novel datasets and models with high accuracy, demonstrating the system's capacity to generalize to new situations. In real-world evaluations with humans, 73% of healthcare workers (e.g., doctors and nurses) agreed they would use TalkToModel over baseline point-and-click systems for explainability in a disease prediction task, and 85% of ML professionals agreed TalkToModel was easier to use for computing explanations. Our findings demonstrate that TalkToModel is more effective for model explainability than existing systems, introducing a new category of explainability tools for practitioners. Code & demo released here: https://github.com/dylan-slack/TalkToModel.
翻译:机器学习( ML) 模型越来越多地被用于在现实世界应用中做出关键决定, 但是它们已经变得更加复杂, 更难理解。 为此, 研究人员提出了几种技术来解释模型预测。 然而, 实践者很难使用这些解释技术, 因为他们往往不知道要选择谁和如何解释解释解释的结果。 在这项工作中, 我们通过引入 TalkToModel: 一个互动式对话系统来应对这些挑战。 具体地说, TalkToModel包含三个关键组成部分:(1) 一个用于参与对话的自然语言用户界面, 使得 ML 模型的可理解性非常容易理解;(2) 一个对话引擎, 适应任何表格模型和数据集, 解释自然语言, 绘制适当的解释, 并生成文本回应; 3 一个执行组件, 构建解释解释解释解释结果。 我们对TalkTodel TalkTodel: 我们进行了广泛的定量和人文主题评估。 总体来说, 我们发现谈话系统理解用户对新式的模型和模型和模型的投入非常准确, 显示系统对新情况的普及能力。 在现实世界评价中, 73%的医疗保健工作者的M- 将使用商定的基准 。</s>