Sequential recommendation aims to model users' evolving preferences based on their historical interactions. Recent advances leverage Transformer-based architectures to capture global dependencies, but existing methods often suffer from high computational overhead, primarily due to discontinuous memory access in temporal encoding and dense attention over long sequences. To address these limitations, we propose FuXi-$γ$, a novel sequential recommendation framework that improves both effectiveness and efficiency through principled architectural design. FuXi-$γ$ adopts a decoder-only Transformer structure and introduces two key innovations: (1) An exponential-power temporal encoder that encodes relative temporal intervals using a tunable exponential decay function inspired by the Ebbinghaus forgetting curve. This encoder enables flexible modeling of both short-term and long-term preferences while maintaining high efficiency through continuous memory access and pure matrix operations. (2) A diagonal-sparse positional mechanism that prunes low-contribution attention blocks using a diagonal-sliding strategy guided by the persymmetry of Toeplitz matrix. Extensive experiments on four real-world datasets demonstrate that FuXi-$γ$ achieves state-of-the-art performance in recommendation quality, while accelerating training by up to 4.74$\times$ and inference by up to 6.18$\times$, making it a practical and scalable solution for long-sequence recommendation. Our code is available at https://github.com/Yeedzhi/FuXi-gamma.
翻译:序列推荐旨在根据用户的历史交互行为建模其动态变化的偏好。近期研究利用基于Transformer的架构捕捉全局依赖关系,但现有方法常面临高计算开销问题,主要源于时间编码中的非连续内存访问以及对长序列的密集注意力计算。为克服这些限制,我们提出FuXi-$\\gamma$,一种新颖的序列推荐框架,通过原则性的架构设计同时提升推荐效果与计算效率。FuXi-$\\gamma$采用仅解码器的Transformer结构,并引入两项关键创新:(1)指数幂时间编码器:受艾宾浩斯遗忘曲线启发,采用可调指数衰减函数编码相对时间间隔。该编码器能够灵活建模短期与长期偏好,同时通过连续内存访问和纯矩阵运算保持高效率。(2)对角稀疏位置机制:利用托普利茨矩阵的对称性指导对角滑动策略,剪枝低贡献度的注意力模块。在四个真实数据集上的大量实验表明,FuXi-$\\gamma$在推荐质量上达到最先进水平,同时训练速度最高提升4.74倍,推理速度最高提升6.18倍,为长序列推荐提供了实用且可扩展的解决方案。代码已开源:https://github.com/Yeedzhi/FuXi-gamma。