Non-linear state-space models, also known as general hidden Markov models, are ubiquitous in statistical machine learning, being the most classical generative models for serial data and sequences in general. The particle-based, rapid incremental smoother PaRIS is a sequential Monte Carlo (SMC) technique allowing for efficient online approximation of expectations of additive functionals under the smoothing distribution in these models. Such expectations appear naturally in several learning contexts, such as likelihood estimation (MLE) and Markov score climbing (MSC). PARIS has linear computational complexity, limited memory requirements and comes with non-asymptotic bounds, convergence results and stability guarantees. Still, being based on self-normalised importance sampling, the PaRIS estimator is biased. Our first contribution is to design a novel additive smoothing algorithm, the Parisian particle Gibbs PPG sampler, which can be viewed as a PaRIS algorithm driven by conditional SMC moves, resulting in bias-reduced estimates of the targeted quantities. We substantiate the PPG algorithm with theoretical results, including new bounds on bias and variance as well as deviation inequalities. Our second contribution is to apply PPG in a learning framework, covering MLE and MSC as special examples. In this context, we establish, under standard assumptions, non-asymptotic bounds highlighting the value of bias reduction and the implicit Rao--Blackwellization of PPG. These are the first non-asymptotic results of this kind in this setting. We illustrate our theoretical results with numerical experiments supporting our claims.
翻译:非线性状态-空间模型,也称为一般隐藏的马尔科夫模型,在统计机学习中无处不在,这是最典型的序列数据和整个序列的最典型遗传模型。基于粒子的快速增量光滑PARIS(PARIS)技术是连续的蒙特卡洛(SMC)技术,允许在这些模型的平滑分布下高效在线近似添加功能功能的预期值。这些预期自然出现在若干学习环境中,如可能性估计(MLE)和马克夫得分攀升(MSC)。巴黎的计算方法具有线性复杂性、记忆要求有限,并带有非无线性界限、趋同结果和稳定性保证。不过,基于自我标准化重要性取样,PARIS天平滑滑度测量器是有偏差的。我们的第一个贡献是设计新的添加式平滑动算法,即巴黎粒子GPGPG样本取样器,可以被视为由有条件的SMC动作驱动的PRIS算法,从而得出了目标数量的偏差性估计值。我们用理论结果来证实PGUC的算法的计算结果,包括新的偏差性和差异和差异性缩性缩性缩缩缩缩图。我们在IMGPGPGLI的缩缩缩框架之下,我们用了这个标准模型的缩缩缩缩缩缩缩缩缩缩的模型的模型,我们用了一个模型的模型的缩缩缩缩缩缩缩缩缩缩缩缩缩的模型。