The kernel exponential family is a rich class of distributions, which can be fit efficiently and with statistical guarantees by score matching. Being required to choose a priori a simple kernel such as the Gaussian, however, limits its practical applicability. We provide a scheme for learning a kernel parameterized by a deep network, which can find complex location-dependent local features of the data geometry. This gives a very rich class of density models, capable of fitting complex structures on moderate-dimensional problems. Compared to deep density models fit via maximum likelihood, our approach provides a complementary set of strengths and tradeoffs: in empirical studies, the former can yield higher likelihoods, whereas the latter gives better estimates of the gradient of the log density, the score, which describes the distribution's shape.
翻译:内核指数式家庭是分布的丰富类别,可以高效地适应,并通过得分匹配获得统计保障。 但是,由于需要先验地选择一个简单的内核,例如高森人,这限制了其实际适用性。 我们为学习深网络所设定的内核参数提供了一个计划,这个网络可以找到数据几何中复杂地因位置而异的地方特征。 这提供了非常丰富的密度模型,能够在中度问题上安装适当的复杂结构。 与最有可能的深度密度模型相比,我们的方法提供了一套互补的优势和权衡:在经验研究中,前者可以产生更高的可能性,而后者则对描述分布形状的日志密度梯度(分数)作出更好的估计。