We address the product question generation task. For a given product description, our goal is to generate questions that reflect potential user information needs that are either missing or not well covered in the description. Moreover, we wish to cover diverse user information needs that may span a multitude of product types. To this end, we first show how the T5 pre-trained Transformer encoder-decoder model can be fine-tuned for the task. Yet, while the T5 generated questions have a reasonable quality compared to the state-of-the-art method for the task (KPCNet), many of such questions are still too general, resulting in a sub-optimal global question diversity. As an alternative, we propose a novel learning-to-diversify (LTD) fine-tuning approach that allows to enrich the language learned by the underlying Transformer model. Our empirical evaluation shows that, using our approach significantly improves the global diversity of the underlying Transformer model, while preserves, as much as possible, its generation relevance.
翻译:我们处理产品问题的生成任务。对于给定的产品描述,我们的目标是提出反映潜在用户信息需求的问题,这些需求要么缺失,要么在描述中未充分涵盖。此外,我们希望涵盖各种用户信息需求,这些需求可能涉及多种产品类型。为此,我们首先展示T5预先培训的变异器编码器编码器编码器编码器(Docoder)模型如何能为任务进行微调。然而,尽管T5产生的问题与最先进的任务方法(KPCNet)相比质量合理,但其中许多问题仍然过于泛泛,导致次优化的全球问题多样化。作为一种替代办法,我们提议采用新的学习变异器模型(LTD)微调方法,以丰富基本变异器模型所学的语言。我们的经验评估表明,利用我们的方法极大地改进了基本变异器模型的全球多样性,同时尽可能保持其生成相关性。