In recent years, Multi-task Learning (MTL) has yielded immense success in Recommender System (RS) applications. However, current MTL-based recommendation models tend to disregard the session-wise patterns of user-item interactions because they are predominantly constructed based on item-wise datasets. Moreover, balancing multiple objectives has always been a challenge in this field, which is typically avoided via linear estimations in existing works. To address these issues, in this paper, we propose a Reinforcement Learning (RL) enhanced MTL framework, namely RMTL, to combine the losses of different recommendation tasks using dynamic weights. To be specific, the RMTL structure can address the two aforementioned issues by (i) constructing an MTL environment from session-wise interactions and (ii) training multi-task actor-critic network structure, which is compatible with most existing MTL-based recommendation models, and (iii) optimizing and fine-tuning the MTL loss function using the weights generated by critic networks. Experiments on two real-world public datasets demonstrate the effectiveness of RMTL with a higher AUC against state-of-the-art MTL-based recommendation models. Additionally, we evaluate and validate RMTL's compatibility and transferability across various MTL models.
翻译:近年来,多任务学习(MTL)在建议系统应用方面取得了巨大成功,然而,目前以MTL为基础的建议模式往往忽视了用户项目互动的会话模式,因为它们主要是根据项目数据集构建的。此外,平衡多重目标始终是这一领域的一个挑战,通常通过现有工程中的线性估算来避免。为了解决这些问题,我们在本文件中提议加强学习(RL)强化的MTL框架,即RMTL,用动态权重合并不同建议任务的损失。具体而言,RMTL结构可以通过以下方式解决上述两个问题:(一) 从会话互动中构建一个MTL环境;(二) 培训多任务行为者-气候网络结构,这与大多数现有的基于MTL的建议模式相一致,以及(三) 利用批评网络产生的权重优化和调整MTL损失功能。两个现实世界公共数据集的实验展示了RMTL的有效性,用更高的AML环境环境环境,与基于州-MT的兼容性模型和跨州-MTMT的ML建议转让。</s>