Recent research showed promising results on combining pretrained language models (LMs) with canonical utterance for few-shot semantic parsing. The canonical utterance is often lengthy and complex due to the compositional structure of formal languages. Learning to generate such canonical utterance requires significant amount of data to reach high performance. Fine-tuning with only few-shot samples, the LMs can easily forget pretrained knowledge, overfit spurious biases, and suffer from compositionally out-of-distribution generalization errors. To tackle these issues, we propose a novel few-shot semantic parsing method -- SeqZero. SeqZero decomposes the problem into a sequence of sub-problems, which correspond to the sub-clauses of the formal language. Based on the decomposition, the LMs only need to generate short answers using prompts for predicting sub-clauses. Thus, SeqZero avoids generating a long canonical utterance at once. Moreover, SeqZero employs not only a few-shot model but also a zero-shot model to alleviate the overfitting. In particular, SeqZero brings out the merits from both models via ensemble equipped with our proposed constrained rescaling. SeqZero achieves SOTA performance of BART-based models on GeoQuery and EcommerceQuery, which are two few-shot datasets with compositional data split.
翻译:最近的研究显示,在将预先培训的语言模型(LMS)与少数语义解析的粗话相结合方面,取得了令人乐观的结果。由于正式语言的构成结构,这种粗话的表达往往冗长而复杂。学习产生这种粗话需要大量的数据才能达到高性能。通过微调,LMS可以很容易忘记事先培训的知识,过分适应虚假偏见,并且会受到组成上超出分布通用错误的影响。为了解决这些问题,我们建议一种新颖的少数语义解析方法 -- -- SeqZero。SeqZero将问题分为一系列子问题,这与正式语言的子问题相符,需要大量的数据才能达到高性能。根据微调,LMSMs只需用提示来生成简短的答案,就可以预测子语言。因此,Seqegozeerro避免一次产生长期的粗话解说。此外,Seqeqzero不仅使用少数张的模型,而且还将问题溶解说成一个小的分解方法。Seqereal Q的模型可以使Seqreareal的模型与Seq Stareareal 之间的两个数据都超越了。