Pretrained language models have demonstrated extraordinary capabilities in language generation. However, real-world tasks often require controlling the distribution of generated text in order to mitigate bias, promote fairness, and achieve personalization. Existing techniques for controlling the distribution of generated text only work with quantified distributions, which require pre-defined categories, proportions of the distribution, or an existing corpus following the desired distributions. However, many important distributions, such as personal preferences, are unquantified. In this work, we tackle the problem of generating text following arbitrary distributions (quantified and unquantified) by proposing Nano, a few-shot human-in-the-loop training algorithm that continuously learns from human feedback. Nano achieves state-of-the-art results on single topic/attribute as well as quantified distribution control compared to previous works. We also show that Nano is able to learn unquantified distributions, achieves personalization, and captures differences between different individuals' personal preferences with high sample efficiency.
翻译:然而,现实世界的任务往往要求控制生成文本的分布,以减少偏见、促进公平和实现个性化。控制生成文本的分布的现有技术仅涉及量化的分布,这需要预先界定的类别、分配比例或按照预期分布后的现有内容。然而,许多重要分布,如个人偏好等,都没有量化。在这项工作中,我们通过提出纳米(Nano),即几个镜头的“人对流”培训算法,不断从人类反馈中学习。纳米(Nano)在单一议题/归属上取得最新的最新结果,并与以往的作品相比,实现量化的分配控制。我们还表明,纳米(Nano)能够学习未经量化的分布,实现个性化,并捕捉不同个人偏好之间的差异,并具有高样本效率。