Slot filling is one of the critical tasks in modern conversational systems. The majority of existing literature employs supervised learning methods, which require labeled training data for each new domain. Zero-shot learning and weak supervision approaches, among others, have shown promise as alternatives to manual labeling. Nonetheless, these learning paradigms are significantly inferior to supervised learning approaches in terms of performance. To minimize this performance gap and demonstrate the possibility of open-domain slot filling, we propose a Self-supervised Co-training framework, called SCot, that requires zero in-domain manually labeled training examples and works in three phases. Phase one acquires two sets of complementary pseudo labels automatically. Phase two leverages the power of the pre-trained language model BERT, by adapting it for the slot filling task using these sets of pseudo labels. In phase three, we introduce a self-supervised cotraining mechanism, where both models automatically select highconfidence soft labels to further improve the performance of the other in an iterative fashion. Our thorough evaluations show that SCot outperforms state-of-the-art models by 45.57% and 37.56% on SGD and MultiWoZ datasets, respectively. Moreover, our proposed framework SCot achieves comparable performance when compared to state-of-the-art fully supervised models.
翻译:槽位填充是现代对话系统中关键的任务之一。大多数现有文献采用监督学习方法,需要为每个新领域进行标签化训练数据。零样本学习和弱监督方法等已被证明是手动标注的替代方案。然而,这些学习方法在性能方面明显劣于监督学习方法。为了尽量缩小这种性能差距并证明实现开放域的槽位填充的可能性,我们提出了一个自我监督协同训练框架SCot,不需要领域内手动标记的训练样本,分三个阶段工作。第一阶段自动获取两组互补伪标签。第二阶段利用预训练的语言模型BERT的强大功能,通过使用这些伪标签将其适应于槽位填充任务。在第三阶段中,我们引入一个自我监督协同训练机制,两个模型均自动选择高置信软标签,以迭代方式进一步提高对方的性能。我们全面的评估表明,SCot在SGD和MultiWoZ数据集上分别比最先进的模型提高了45.57%和37.56%的性能。此外,与最先进的完全监督模型相比,我们提出的框架SCot实现了可比较的性能。