We propose a general framework called Text Modular Networks(TMNs) for building interpretable systems that learn to solve complex tasks by decomposing them into simpler ones solvable by existing models. To ensure solvability of simpler tasks, TMNs learn the textual input-output behavior (i.e., language) of existing models through their datasets. This differs from prior decomposition-based approaches which, besides being designed specifically for each complex task, produce decompositions independent of existing sub-models. Specifically, we focus on Question Answering (QA) and show how to train a next-question generator to sequentially produce sub-questions targeting appropriate sub-models, without additional human annotation. These sub-questions and answers provide a faithful natural language explanation of the model's reasoning. We use this framework to build ModularQA, a system that can answer multi-hop reasoning questions by decomposing them into sub-questions answerable by a neural factoid single-span QA model and a symbolic calculator. Our experiments show that ModularQA is more versatile than existing explainable systems for DROP and HotpotQA datasets, is more robust than state-of-the-art blackbox (uninterpretable) systems, and generates more understandable and trustworthy explanations compared to prior work.
翻译:我们提出一个名为文本模块网络(TMNs)的一般框架,用于建立可解释的系统,通过将这些系统分解成现有模型可以溶解的更简单的任务,从而学会解决复杂的任务。为了确保更简单的任务的可溶性,TMNs通过数据集学习现有模型的文本输入-输出行为(即语言)。这不同于先前基于分解的方法,前者是专门为每个复杂任务设计的,其产生分解与现有子模型无关。具体地说,我们侧重于问题解答模型(QA),并展示如何训练下一个问题生成者,以便按顺序生成针对适当的子模型的子问题,而无需额外的人文说明。这些子问题和答案为模型的推理提供了忠实的自然语言解释。我们使用这个框架来构建模块QA,这个系统可以解解析多点推理问题,将其分为一个可解析的线性事实模型(QA)模型和一个象征性的解析器。我们的实验显示,比当前更易懂和可理解性的数据系统更难理解性,我们之前的解算法是更难易变的、更难易变的模型。