As AI is more and more pervasive in everyday life, humans have an increasing demand to understand its behavior and decisions. Most research on explainable AI builds on the premise that there is one ideal explanation to be found. In fact, however, everyday explanations are co-constructed in a dialogue between the person explaining (the explainer) and the specific person being explained to (the explainee). In this paper, we introduce a first corpus of dialogical explanations to enable NLP research on how humans explain as well as on how AI can learn to imitate this process. The corpus consists of 65 transcribed English dialogues from the Wired video series \emph{5 Levels}, explaining 13 topics to five explainees of different proficiency. All 1550 dialogue turns have been manually labeled by five independent professionals for the topic discussed as well as for the dialogue act and the explanation move performed. We analyze linguistic patterns of explainers and explainees, and we explore differences across proficiency levels. BERT-based baseline results indicate that sequence information helps predicting topics, acts, and moves effectively
翻译:由于大赦国际在日常生活中越来越普遍,人类越来越需要了解其行为和决定。大多数关于可解释的AI的研究都基于以下前提:有一个理想的解释可以找到。但事实上,日常的解释是在解释者(解释者)和向解释者(解释者)解释的具体人员(解释者)之间的对话中共同构建的。在本文中,我们引入了第一批对话解释材料,以便国家语言方案研究人类如何解释以及可以如何学会模仿这一过程。该材料包括65个从Wired视频系列 \ emph{5 levels 中转录的英语对话,向5个不同熟练程度的解释者解释了13个专题。所有1550个对话都由5个独立专业人员手动标注,用于讨论的专题以及对话行为和解释动作。我们分析了解释者和解释者的语言模式,并探讨了不同熟练程度的差异。基于ERT的基线结果显示,序列信息有助于预测专题、行为和有效移动。