Text-based machine comprehension (MC) systems have a wide-range of applications, and standard corpora exist for developing and evaluating approaches. There has been far less research on spoken question answering (SQA) systems. The SQA task considered in this paper is to extract the answer from a candidate$\text{'}$s spoken response to a question in a prompt-response style language assessment test. Applying these MC approaches to this SQA task rather than, for example, off-topic response detection provides far more detailed information that can be used for further downstream processing. One significant challenge is the lack of appropriately annotated speech corpora to train systems for this task. Hence, a transfer-learning style approach is adopted where a system trained on text-based MC is evaluated on an SQA task with non-native speakers. Mismatches must be considered between text documents and spoken responses; non-native spoken grammar and written grammar. In practical SQA, ASR systems are used, necessitating an investigation of the impact of ASR errors. We show that a simple text-based ELECTRA MC model trained on SQuAD2.0 transfers well for SQA. It is found that there is an approximately linear relationship between ASR errors and the SQA assessment scores but grammar mismatches have minimal impact.
翻译:以文字为基础的机器理解(MC)系统的应用范围很广,在开发和评价方法方面存在着标准公司,对口答(SQA)系统的研究要少得多,本文考虑的SQA的任务是从候选人中提取对一个问题的口头回答,即:在迅速答复式语言评估测试中,对一个问题的答复是$\text{}美元;在SQA任务中应用这些MC方法,而不是以非主题的语法和书面语法。在实际应用SQA中,ASR系统提供了更详细得多的信息,可用于进一步下游处理。一个重大挑战是缺乏适当的附加说明的演讲公司来培训这项任务的系统。因此,在对基于文字的MC任务进行评估时,采用了一种转让学习方式,即与非母语发言人一起对SQA任务进行评价的系统。必须在文本文件和口语法答复中考虑错误;使用非本地语言语法和书面语法。在SQA实际应用中,需要调查ASR错误的影响。我们表明,基于文字的E-ECTRAMC在SQA中发现了一个典型的SQA关系,但SA-AMA是在那里的典型的SAQATRAC 。