Building a dialogue system that can communicate naturally with humans is a challenging yet interesting problem of agent-based computing. The rapid growth in this area is usually hindered by the long-standing problem of data scarcity as these systems are expected to learn syntax, grammar, decision making, and reasoning from insufficient amounts of task-specific dataset. The recently introduced pre-trained language models have the potential to address the issue of data scarcity and bring considerable advantages by generating contextualized word embeddings. These models are considered counterpart of ImageNet in NLP and have demonstrated to capture different facets of language such as hierarchical relations, long-term dependency, and sentiment. In this short survey paper, we discuss the recent progress made in the field of pre-trained language models. We also deliberate that how the strengths of these language models can be leveraged in designing more engaging and more eloquent conversational agents. This paper, therefore, intends to establish whether these pre-trained models can overcome the challenges pertinent to dialogue systems, and how their architecture could be exploited in order to overcome these challenges. Open challenges in the field of dialogue systems have also been deliberated.
翻译:建立能够自然地与人类沟通的对话系统是一个具有挑战性但有趣的代理计算问题。该领域的快速增长通常受到长期数据稀缺问题的阻碍,因为这些系统预计将学习语法、语法、决策以及因任务特定数据集数量不足而产生的推理。最近引进的经过培训的语言模型有可能解决数据稀缺问题,并通过产生背景化的词嵌入而带来相当大的优势。这些模型被认为是NLP中图像网络的对应模型,并证明能够捕捉语言的不同方面,如等级关系、长期依赖性和情绪。在这个简短的调查文件中,我们讨论了在预先培训的语言模型领域最近取得的进展。我们还审议了如何利用这些语言模型的优势来设计更具参与性和更雄辩的对话工具。因此,本文件打算确定这些经过培训的模式是否能够克服与对话系统有关的挑战,以及如何利用这些模型的结构来克服这些挑战。对话系统领域的公开挑战也已经得到了讨论。