Comparing Bayesian neural networks (BNNs) with different widths is challenging because, as the width increases, multiple model properties change simultaneously, and, inference in the finite-width case is intractable. In this work, we empirically compare finite- and infinite-width BNNs, and provide quantitative and qualitative explanations for their performance difference. We find that when the model is mis-specified, increasing width can hurt BNN performance. In these cases, we provide evidence that finite-width BNNs generalize better partially due to the properties of their frequency spectrum that allows them to adapt under model mismatch.
翻译:将贝叶斯神经网络(BNNs)与不同宽度进行比较具有挑战性,因为随着宽度的增加,多重模型属性会同时发生变化,而有限宽度的推论是难以解决的。在这项工作中,我们从经验上比较了有限和无限宽度的BNNs,并提供了其性能差异的定量和定性解释。我们发现,当模型被错误指定时,增加宽度会损害BNN的性能。在这些案例中,我们提供了证据,证明有限宽度的BNNs由于其频谱的特性而部分地比较了优异,从而使它们能够在模型不匹配的情况下适应。