Inspired by the inductive transfer learning on computer vision, many efforts have been made to train contextualized language models that boost the performance of natural language processing tasks. These models are mostly trained on large general-domain corpora such as news, books, or Wikipedia.Although these pre-trained generic language models well perceive the semantic and syntactic essence of a language structure, exploiting them in a real-world domain-specific scenario still needs some practical considerations to be taken into account such as token distribution shifts, inference time, memory, and their simultaneous proficiency in multiple tasks. In this paper, we focus on the legal domain and present how different language model strained on general-domain corpora can be best customized for multiple legal document reviewing tasks. We compare their efficiencies with respect to task performances and present practical considerations.
翻译:在计算机视觉感化传导学习的启发下,已作出许多努力来培训背景化语言模型,以促进自然语言处理任务的执行,这些模型大多在诸如新闻、书籍或维基百科等大型一般性公司中接受培训。 虽然这些经过事先训练的通用语言模型非常清楚一种语言结构的语义和综合本质,但在现实世界特定领域的情景中加以利用,仍然需要考虑一些实际因素,例如象征性的分布转换、推论时间、记忆以及它们在多重任务中的同时熟练程度。在本论文中,我们侧重于法律领域,并介绍了如何对一般语言公司进行不同培训的不同语言模型能够最好地适应多重法律文件审查任务。我们比较了这些模型在任务性表现和提出实际考虑方面的效率。