The emergency of Pre-trained Language Models (PLMs) has achieved tremendous success in the field of Natural Language Processing (NLP) by learning universal representations on large corpora in a self-supervised manner. The pre-trained models and the learned representations can be beneficial to a series of downstream NLP tasks. This training paradigm has recently been adapted to the recommendation domain and is considered a promising approach by both academia and industry. In this paper, we systematically investigate how to extract and transfer knowledge from pre-trained models learned by different PLM-related training paradigms to improve recommendation performance from various perspectives, such as generality, sparsity, efficiency and effectiveness. Specifically, we propose an orthogonal taxonomy to divide existing PLM-based recommender systems w.r.t. their training strategies and objectives. Then, we analyze and summarize the connection between PLM-based training paradigms and different input data types for recommender systems. Finally, we elaborate on open issues and future research directions in this vibrant field.
翻译:预先培训语言模式(PLM)的紧急情况在自然语言处理领域取得了巨大成功,通过自我监督的方式学习了对大型公司的普遍代表性,预先培训的模式和学到的表述可以有益于一系列下游国家语言模式的任务。这种培训模式最近已经适应建议领域,并被认为是学术界和工业界的一种有希望的方法。在本文件中,我们系统地调查如何从各种与个人语言处理有关的培训模式所学的预先培训模式中提取和转让知识,以便从一般性、分散性、效率和有效性等不同角度改进建议绩效。具体地说,我们建议一种正方位分类学,以区分现有的基于PLM的推荐系统的培训战略和目标。然后,我们分析和总结基于PLM培训模式与建议系统的不同输入数据类型之间的联系。最后,我们阐述了这一充满活力的领域的公开问题和未来研究方向。