Predictive coding (PC) accounts of perception now form one of the dominant computational theories of the brain, where they prescribe a general algorithm for inference and learning over hierarchical latent probabilistic models. Despite this, they have enjoyed little export to the broader field of machine learning, where comparative generative modelling techniques have flourished. In part, this has been due to the poor performance of models trained with PC when evaluated by both sample quality and marginal likelihood. By adopting the perspective of PC as a variational Bayes algorithm under the Laplace approximation, we identify the source of these deficits to lie in the exclusion of an associated Hessian term in the PC objective function, which would otherwise regularise the sharpness of the probability landscape and prevent over-certainty in the approximate posterior. To remedy this, we make three primary contributions: we begin by suggesting a simple Monte Carlo estimated evidence lower bound which relies on sampling from the Hessian-parameterised variational posterior. We then derive a novel block diagonal approximation to the full Hessian matrix that has lower memory requirements and favourable mathematical properties. Lastly, we present an algorithm that combines our method with standard PC to reduce memory complexity further. We evaluate models trained with our approach against the standard PC framework on image benchmark datasets. Our approach produces higher log-likelihoods and qualitatively better samples that more closely capture the diversity of the data-generating distribution.
翻译:认知的预测编码(PC)账户现在构成了大脑的主要计算理论之一,其中,它们规定了用于对等级潜潜伏概率模型进行推论和学习的一般算法,尽管如此,它们很少向更广泛的机器学习领域出口,比较的基因建模技术已经发展起来。部分原因在于,在抽样质量和边缘可能性的评价下,通过抽样质量和边际可能性来评估时,经过计算机培训的模型表现不佳。通过采用PC作为拉普尔近距离下的变异巴耶斯算法,我们发现这些缺陷的根源在于:在PC目标功能中排除一个相关的赫萨语术语,否则就会使概率景观的清晰度正规化,防止近似后方图像的过度不确定性。为了纠正这一点,我们做出了三项主要贡献:我们首先建议一个简单的蒙特卡洛估计证据较低的范围,该范围依赖于从赫萨氏比差度的变异差场采集的样本。我们接着从一个新颖的基调调调调调基点到全赫萨马基体矩阵,该词的记忆要求较低,并且有利于数学特性的分布。最后,我们用一个经过训练的更精准的模型来将我们的标准模型与我们的标准模型与我们的标准模型结合起来。</s>