Vast efforts have been devoted to creating high-performance few-shot learners, i.e., large-scale pretrained language models (PLMs) that perform well with little downstream task training data. Training PLMs has incurred significant cost, but utilizing the few-shot learners is still challenging due to their enormous size. This work focuses on a crucial question: How to make effective use of these few-shot learners? We propose LMTurk, a novel approach that treats few-shot learners as crowdsourcing workers. The rationale is that crowdsourcing workers are in fact few-shot learners: They are shown a few illustrative examples to learn about a task and then start annotating. LMTurk employs few-shot learners built upon PLMs as workers. We show that the resulting annotations can be utilized to train models that solve the task well and are small enough to be deployable in practical scenarios. Active learning is integrated into LMTurk to reduce the amount of queries made to PLMs, minimizing the computational cost of running PLM inference passes. Altogether, LMTurk is an important step towards making effective use of current PLMs.
翻译:大量的努力都致力于创造高性能的少见学习者,即大型的预先培训语言模型(PLMs),这些模型在下游任务培训数据很少的情况下表现良好。培训PLMs花费了大量费用,但利用少见学习者仍因其规模巨大而具有挑战性。这项工作侧重于一个关键问题:如何有效利用这些少见学习者?我们建议LMTurk,这是一种将少见学习者作为众包工人对待的新办法。其理由是,众包工人实际上是少见的学习者:他们展示了几个示例,以了解一项任务,然后开始作说明。LMTurk雇用了几个在PLMS上建立的少见学习者作为工人。我们表明,由此产生的说明可用于培训能够很好地解决任务的模型,而且小到在实际情况下可以部署的模型。积极学习被纳入LMTurk,以减少对PLMS的查询数量,最大限度地减少使用PLM的推价的计算成本。加之,LMTurk是有效利用目前PLMMs的重要一步。