The goal of sequential event prediction is to estimate the next event based on a sequence of historical events, with applications to sequential recommendation, user behavior analysis and clinical treatment. In practice, the next-event prediction models are trained with sequential data collected at one time and need to generalize to newly arrived sequences in remote future, which requires models to handle temporal distribution shift from training to testing. In this paper, we first take a data-generating perspective to reveal a negative result that existing approaches with maximum likelihood estimation would fail for distribution shift due to the latent context confounder, i.e., the common cause for the historical events and the next event. Then we devise a new learning objective based on backdoor adjustment and further harness variational inference to make it tractable for sequence learning problems. On top of that, we propose a framework with hierarchical branching structures for learning context-specific representations. Comprehensive experiments on diverse tasks (e.g., sequential recommendation) demonstrate the effectiveness, applicability and scalability of our method with various off-the-shelf models as backbones.
翻译:相继事件预测的目标是根据一系列历史事件来估计下一个事件,同时应用顺序建议、用户行为分析和临床治疗。在实践中,下一项活动的预测模型经过一次收集的顺序数据的培训,需要概括到遥远的未来新到的顺序,这就要求模型处理从培训到测试的时间分布变化。在本文中,我们首先从数据生成的角度揭示一个负面结果,即由于潜在背景混乱,即历史事件和下一项事件的共同原因,现有最有可能估计的办法来分配转移不会成功。然后,我们设计一个新的学习目标,以后门调整为基础,并进一步利用变式推论,使它能够用于顺序学习问题。此外,我们提出了一个框架,以分级分级结构来学习特定背景的表述。关于不同任务(例如顺序建议)的全面实验表明我们的方法的有效性、适用性和可缩放性,并以各种离式模型作为骨干。