We present a variational method for online state estimation and parameter learning in state-space models (SSMs), a ubiquitous class of latent variable models for sequential data. As per standard batch variational techniques, we use stochastic gradients to simultaneously optimize a lower bound on the log evidence with respect to both model parameters and a variational approximation of the states' posterior distribution. However, unlike existing approaches, our method is able to operate in an entirely online manner, such that historic observations do not require revisitation after being incorporated and the cost of updates at each time step remains constant, despite the growing dimensionality of the joint posterior distribution of the states. This is achieved by utilizing backward decompositions of this joint posterior distribution and of its variational approximation, combined with Bellman-type recursions for the evidence lower bound and its gradients. We demonstrate the performance of this methodology across several examples, including high-dimensional SSMs and sequential Variational Auto-Encoders.
翻译:我们提出了在州空间模型中进行在线状态估计和参数学习的变式方法,这是一系列相继数据的潜在潜在变量模型的无处不在的类别。按照标准的批量变异技术,我们使用随机梯度同时优化日志证据的下限,既包括模型参数,也包括各州后方分布的变异近似值。然而,与现有方法不同,我们的方法能够完全在线运作,因此历史观测在纳入后不需要重新审视,而每个步骤的更新费用保持不变,尽管各州联合后方分布的维度日益提高。这是通过利用这种联合远端分布及其变异近度的后向分解,结合Bellman型对证据下界及其梯度的回溯。我们通过多个实例,包括高维度 SMMs和连续自动电解码器等,展示了这一方法的绩效。