We examine how well the state-of-the-art (SOTA) models used in legal reasoning support abductive reasoning tasks. Abductive reasoning is a form of logical inference in which a hypothesis is formulated from a set of observations, and that hypothesis is used to explain the observations. The ability to formulate such hypotheses is important for lawyers and legal scholars as it helps them articulate logical arguments, interpret laws, and develop legal theories. Our motivation is to consider the belief that deep learning models, especially large language models (LLMs), will soon replace lawyers because they perform well on tasks related to legal text processing. But to do so, we believe, requires some form of abductive hypothesis formation. In other words, while LLMs become more popular and powerful, we want to investigate their capacity for abductive reasoning. To pursue this goal, we start by building a logic-augmented dataset for abductive reasoning with 498,697 samples and then use it to evaluate the performance of a SOTA model in the legal field. Our experimental results show although these models can perform well on tasks related to some aspects of legal text processing, they still fall short in supporting abductive reasoning tasks.
翻译:我们研究了用于法律推理的最先进(SOTA)模型在支持附加推理任务方面的表现。附加推理是一种逻辑推理形式,通过一组观察结果提出假设,并使用该假设解释这些观察结果。能够提出这样的假设对于律师和法学者来说非常重要,因为它们有助于阐述逻辑论点,解释法律,并发展法律理论。我们的动机在于考虑一个信念,即深度学习模型,特别是大型语言模型(LLM)将很快取代律师,因为它们在与法律文本处理相关的任务上表现良好。但是,为了做到这一点,我们认为需要某种形式的附加假设形成。换句话说,虽然LLM变得越来越流行和强大,我们也想研究它们在附加推理方面的能力。为了追求这一目标,我们首先构建了一个附加推理的逻辑增强数据集,含有498,697个样本,然后使用它来评估法律领域SOTA模型的性能。我们的实验结果表明,尽管这些模型在与一些法律文本处理方面的任务上表现良好,但它们在支持附加推理任务方面仍然存在欠缺。