Probabilistic circuits (PCs) are a family of generative models which allows for the computation of exact likelihoods and marginals of its probability distributions. PCs are both expressive and tractable, and serve as popular choices for discrete density estimation tasks. However, large PCs are susceptible to overfitting, and only a few regularization strategies (e.g., dropout, weight-decay) have been explored. We propose HyperSPNs: a new paradigm of generating the mixture weights of large PCs using a small-scale neural network. Our framework can be viewed as a soft weight-sharing strategy, which combines the greater expressiveness of large models with the better generalization and memory-footprint properties of small models. We show the merits of our regularization strategy on two state-of-the-art PC families introduced in recent literature -- RAT-SPNs and EiNETs -- and demonstrate generalization improvements in both models on a suite of density estimation benchmarks in both discrete and continuous domains.
翻译:概率电路(PCs)是一组基因模型,可以计算其概率分布的确切可能性和边缘。PCs既可表达又可移动,并可作为离散密度估计任务的流行选择。然而,大型PCs很容易被过度安装,而且只探索了少数正规化战略(例如辍学、体重下降)。我们提议超超标准PNs:一种使用小型神经网络生成大型PCs混合重量的新模式。我们的框架可被视为一种软的权重共享战略,将大型模型的更大清晰度与小型模型的更普通化和记忆-足迹特性结合起来。我们展示了我们在最近文献中引入的关于两个最先进的PC家庭 -- -- RAT-SPNs 和 Einets -- -- 的正规化战略的优点,并展示了在离散和连续领域对一套密度估计基准进行总体化的两种模型的改进。