Controlling the generative model to adapt a new domain with limited samples is a difficult challenge and it is receiving increasing attention. Recently, methods based on meta-learning have shown promising results for few-shot domain adaptation. However, meta-learning-based methods usually suffer from the problem of overfitting, which results in a lack of diversity in the generated texts. To avoid this problem, in this study, a novel framework based on reinforcement learning (RL) is proposed. In this framework, to increase the sample utilization of RL and decrease its sample requirement, maximum likelihood estimation learning is incorporated into the RL process. When there are only a few in-domain samples available, experimental results on five target domains in two few-shot configurations show that this framework performs better than baselines.
翻译:控制基因化模型以适应具有有限样本的新领域是一项困难的挑战,而且日益受到注意。最近,基于元学习的方法为几发域适应工作展现出有希望的结果。然而,基于元学习的方法通常存在过大的问题,造成生成文本缺乏多样性。为了避免这一问题,本研究报告提出了一个基于强化学习的新框架。在这个框架内,为了增加RL的样本利用并减少其样本要求,将最大的可能性估计学习纳入RL进程。当仅有少量的域内样本可用时,五个目标领域的实验结果在两个微小的组合中显示,这一框架的运行优于基线。