In this work, we aim to learn multi-level user intents from the co-interacted patterns of items, so as to obtain high-quality representations of users and items and further enhance the recommendation performance. Towards this end, we develop a novel framework, Hierarchical User Intent Graph Network, which exhibits user intents in a hierarchical graph structure, from the fine-grained to coarse-grained intents. In particular, we get the multi-level user intents by recursively performing two operations: 1) intra-level aggregation, which distills the signal pertinent to user intents from co-interacted item graphs; and 2) inter-level aggregation, which constitutes the supernode in higher levels to model coarser-grained user intents via gathering the nodes' representations in the lower ones. Then, we refine the user and item representations as a distribution over the discovered intents, instead of simple pre-existing features. To demonstrate the effectiveness of our model, we conducted extensive experiments on three public datasets. Our model achieves significant improvements over the state-of-the-art methods, including MMGCN and DisenGCN. Furthermore, by visualizing the item representations, we provide the semantics of user intents.
翻译:在这项工作中,我们的目标是从共同互动的项目模式中学习多层次用户的意图,以便获得用户和项目的高质量代表,并进一步提高建议性能。为此,我们开发了一个新颖的框架,即 " 高端用户输入图网络 ",它显示用户在分级图表结构中的意图,从细微的图形结构到粗糙的图意图。特别是,我们通过反复执行两种操作来获得多层次用户的意图:1) 内部汇总,它从共同互动的项目图表中提取与用户意图相关的信号;和2) 跨层次汇总,它构成更高层次的超级节点,通过在较低层次收集节点表示来模拟粗略的用户意图。然后,我们改进用户和项目表示,作为所发现意图的分布,而不是简单的原有特征。为了显示我们模型的有效性,我们在三个公共数据集中进行了广泛的实验。我们的模型在州-艺术项目中实现了显著的改进,包括MCN和MCD图像,我们通过视觉-G 提供SIMG和SDIS 。