Score matching (SM) is a convenient method for training flexible probabilistic models, which is often preferred over the traditional maximum-likelihood (ML) approach. However, these models are less interpretable than normalized models; as such, training robustness is in general difficult to assess. We present a critical study of existing variational SM objectives, showing catastrophic failure on a wide range of datasets and network architectures. Our theoretical insights on the objectives emerge directly from their equivalent autoencoding losses when optimizing variational autoencoder (VAE) models. First, we show that in the Fisher autoencoder, SM produces far worse models than maximum-likelihood, and approximate inference by Fisher divergence can lead to low-density local optima. However, with important modifications, this objective reduces to a regularized autoencoding loss that resembles the evidence lower bound (ELBO). This analysis predicts that the modified SM algorithm should behave very similarly to ELBO on Gaussian VAEs. We then review two other FD-based objectives from the literature and show that they reduce to uninterpretable autoencoding losses, likely leading to poor performance. The experiments verify our theoretical predictions and suggest that only ELBO and the baseline objective robustly produce expected results, while previously proposed SM methods do not.
翻译:计分匹配(SM)是培训弹性概率模型的方便方法,通常比传统的最大相似度(ML)方法更可取。然而,这些模型比常规模型更难解释;然而,这些模型比常规模型更难解释;因此,培训的稳健性一般难以评估。我们对现有差异性SM目标进行严格研究,显示大量数据集和网络结构存在灾难性故障。我们在对目标的理论见解中发现,在优化变异自动编码模型时,它们的目标与等值自动编码损失直接相关。首先,我们显示,在Fisher 自动编码器(VAE)中,SM生成的模型比最大相似性模型要差得多,而渔业差异的大致推论则可能导致低密度地方选择。然而,如果进行重大修改,这个目标将降低为与证据较低约束范围(ELBOBO)相似的正常自动编码损失。这项分析预测预测,在最佳变异性自动编码(VAusian VAEE)模型中,修改的SM算法应该非常相似。我们随后从文献中审查另外两个基于FD的目标目标的目标,并表明它们可能会导致无法进行精确的预期的实验结果。