In recent years, Question Answering systems have become more popular and widely used by users. Despite the increasing popularity of these systems, the their performance is not even sufficient for textual data and requires further research. These systems consist of several parts that one of them is the Answer Selection component. This component detects the most relevant answer from a list of candidate answers. The methods presented in previous researches have attempted to provide an independent model to undertake the answer-selection task. An independent model cannot comprehend the syntactic and semantic features of questions and answers with a small training dataset. To fill this gap, language models can be employed in implementing the answer selection part. This action enables the model to have a better understanding of the language in order to understand questions and answers better than previous works. In this research, we will present the "BAS" (BERT Answer Selection) that uses the BERT language model to comprehend language. The empirical results of applying the model on the TrecQA Raw, TrecQA Clean, and WikiQA datasets demonstrate that using a robust language model such as BERT can enhance the performance. Using a more robust classifier also enhances the effect of the language model on the answer selection component. The results demonstrate that language comprehension is an essential requirement in natural language processing tasks such as answer-selection.
翻译:近些年来,问题解答系统越来越受欢迎,用户广泛使用。尽管这些系统的普及程度越来越高,但其性能甚至不足以满足文本数据,需要进一步研究。这些系统包括几个部分,其中一个部分是答案选择部分。本组成部分从候选人回答清单中检测最相关的答案。以前研究中介绍的方法试图提供一个独立模型来进行答案选择任务。独立模型无法用一个小的培训数据集来理解问答的合成和语义特征。为填补这一空白,可以使用语言模型来实施答案选择部分。这一行动使模型能够更好地了解语言,以便更好地了解问题和答案。在这项研究中,我们将介绍“BASS”(BERT回答选择),使用“BERT语言模型”来理解语言。应用TrecQA Raw、TrecQA Clean和WikikQA数据集的经验结果显示,使用强大的语言模型,如BERT等语言模型可以提高语言选择部分的性能。使用更坚实的“BART”的“BARELA答案,也可以用更牢固的“BARILA”来显示基本语言选择语言的答案。