In recent years particle filters have being used as components in systems optimized end-to-end with gradient descent. However, the resampling step in a particle filter is not differentiable, which biases gradients and interferes with optimization. To remedy this problem, several differentiable variants of resampling have been proposed, all of which modify the behavior of the particle filter in significant and potentially undesirable ways. In this paper, we show how to obtain unbiased estimators of the gradient of the marginal likelihood by only modifying messages used in backpropagation, leaving the standard forward pass of a particle filter unchanged. Our method is simple to implement, has a low computational overhead, does not introduce additional hyperparameters, and extends to derivatives of higher orders. We call it stop-gradient resampling, since it can easily be implemented with automatic differentiation libraries using the stop-gradient operator instead of explicitly modifying the backward messages.
翻译:近些年来,粒子过滤器一直被用作以梯度下降为顶端至顶端优化系统中的部件。 但是,粒子过滤器中的再抽样步骤是无法区分的,它会偏向梯度和干扰优化。 为了解决这个问题,提出了几种不同的再抽样变种,所有这些变种都以重要和潜在不可取的方式改变了粒子过滤器的行为。在本文中,我们展示了如何通过只修改后向回传中所使用的信息来获得对微值可能性梯度的不公正性估计器,使粒子过滤器的标准前传不变。我们的方法简单易执行,计算间接偏低,不会引入额外的双参数,而是扩展到更高订单的衍生物。我们称之为中位再抽样,因为它很容易通过自动区分图书馆使用静态操作器而不是明确修改后端信息来实施。