Answering complex questions often requires multi-step reasoning in order to obtain the final answer. Most research into decompositions of complex questions involves open-domain systems, which have shown success in using these decompositions for improved retrieval. In the machine reading setting, however, work to understand when decompositions are helpful is understudied. We conduct experiments on decompositions in machine reading to unify recent work in this space, using a range of models and datasets. We find that decompositions can be helpful in the few-shot case, giving several points of improvement in exact match scores. However, we also show that when models are given access to datasets with around a few hundred or more examples, decompositions are not helpful (and can actually be detrimental). Thus, our analysis implies that models can learn decompositions implicitly even with limited data.
翻译:回答复杂的问题往往需要多步推理才能获得最终答案。对复杂问题分解的多数研究涉及开放域系统,这些系统在利用这些分解系统改进检索方面已经证明是成功的。然而,在机器阅读环境中,对在分解有助于了解何时分解的工作研究不足。我们用一系列模型和数据集对机器阅读中的分解进行实验,以统一这个空间最近的工作。我们发现分解在微小的案例中会有所帮助,在精确匹配的分数中给出几个改进点。然而,我们也表明当模型获得大约几百个或更多例子的数据集时,分解是没有帮助的(而且实际上可能有害 ) 。 因此,我们的分析表明,即使数据有限,模型也可以隐含分解。