The premises of an argument give evidence or other reasons to support a conclusion. However, the amount of support required depends on the generality of a conclusion, the nature of the individual premises, and similar. An argument whose premises make its conclusion rationally worthy to be drawn is called sufficient in argument quality research. Previous work tackled sufficiency assessment as a standard text classification problem, not modeling the inherent relation of premises and conclusion. In this paper, we hypothesize that the conclusion of a sufficient argument can be generated from its premises. To study this hypothesis, we explore the potential of assessing sufficiency based on the output of large-scale pre-trained language models. Our best model variant achieves an F1-score of .885, outperforming the previous state-of-the-art and being on par with human experts. While manual evaluation reveals the quality of the generated conclusions, their impact remains low ultimately.
翻译:论据的前提提供了证据或其他理由来支持结论。然而,所需支持的数量取决于结论的一般性、个别房地的性质以及相似之处。在论证质量研究中,认为有理由作出合理结论的论点就足够了。以前的工作将充足程度评估作为一个标准文本分类问题处理,而不是房舍的固有关系和结论的模型。在本文件中,我们假设可以从其前提中得出充分论据的结论。研究这一假设,我们探讨根据大规模预先培训的语文模型的产出评估充足程度的可能性。我们的最佳模型模型模型模型取得了885个F1核心,超过了以前的工艺水平,与人类专家相当。虽然人工评估揭示了所得出的结论的质量,但其影响最终仍然很小。