Since we can leverage a large amount of unlabeled data without any human supervision to train a model and transfer the knowledge to target tasks, self-supervised learning is a de-facto component for the recent success of deep learning in various fields. However, in many cases, there is a discrepancy between a self-supervised learning objective and a task-specific objective. In order to tackle such discrepancy in Text-to-SQL task, we propose a novel self-supervised learning framework. We utilize the task-specific properties of Text-to-SQL task and the underlying structures of table contents to train the models to learn useful knowledge of the \textit{header-column} alignment task from unlabeled table data. We are able to transfer the knowledge to the supervised Text-to-SQL training with annotated samples, so that the model can leverage the knowledge to better perform the \textit{header-span} alignment task to predict SQL statements. Experimental results show that our self-supervised learning framework significantly improves the performance of the existing strong BERT based models without using large external corpora. In particular, our method is effective for training the model with scarce labeled data. The source code of this work is available in GitHub.
翻译:由于我们可以在没有任何人力监督的情况下利用大量未贴标签的数据来培训模型并将知识转让给目标任务,自我监督的学习是最近不同领域深层学习成功的一个实际组成部分。 但是,在许多情况下,自我监督的学习目标和任务特定目标之间存在差异。 为了解决文本到SQL任务中的这种差异,我们提议了一个全新的自我监督的学习框架。我们利用文本到SQL任务的任务特性和表格内容的基本结构来培训模型,从未贴标签的表格数据中学习关于\ textit{header-clunn}协调的有用知识。我们能够将知识转让给受监督的文本到SQL培训,并配有附加说明的样本,以便模型能够利用知识更好地执行 textitit{header-span}协调任务,预测SQL的报表。实验结果显示,我们自我监督的学习框架大大改进了基于BERTERT的强大模型的性能,而没有使用大号数据源码。 我们的GiH培训是使用大型数据源。