This paper investigates two prominent probabilistic neural modeling paradigms: Bayesian Neural Networks (BNNs) and Mixture Density Networks (MDNs) for uncertainty-aware nonlinear regression. While BNNs incorporate epistemic uncertainty by placing prior distributions over network parameters, MDNs directly model the conditional output distribution, thereby capturing multimodal and heteroscedastic data-generating mechanisms. We present a unified theoretical and empirical framework comparing these approaches. On the theoretical side, we derive convergence rates and error bounds under H\"older smoothness conditions, showing that MDNs achieve faster Kullback-Leibler (KL) divergence convergence due to their likelihood-based nature, whereas BNNs exhibit additional approximation bias induced by variational inference. Empirically, we evaluate both architectures on synthetic nonlinear datasets and a radiographic benchmark (RSNA Pediatric Bone Age Challenge). Quantitative and qualitative results demonstrate that MDNs more effectively capture multimodal responses and adaptive uncertainty, whereas BNNs provide more interpretable epistemic uncertainty under limited data. Our findings clarify the complementary strengths of posterior-based and likelihood-based probabilistic learning, offering guidance for uncertainty-aware modeling in nonlinear systems.
翻译:本文研究了两种主流的概率神经建模范式:用于不确定性感知非线性回归的贝叶斯神经网络(BNNs)和混合密度网络(MDNs)。BNNs通过在网络参数上设置先验分布来纳入认知不确定性,而MDNs则直接对条件输出分布进行建模,从而捕捉多模态和异方差的数据生成机制。我们提出了一个统一的理论与实证框架来比较这两种方法。在理论方面,我们在Hölder光滑性条件下推导了收敛速率和误差界,表明MDNs由于其基于似然的特性,在Kullback-Leibler(KL)散度收敛上更快,而BNNs则因变分推断引入额外的近似偏差。在实证方面,我们在合成非线性数据集和一个放射影像基准(RSNA儿科骨龄挑战赛)上评估了两种架构。定量和定性结果表明,MDNs能更有效地捕捉多模态响应和自适应不确定性,而BNNs在数据有限时提供更具可解释性的认知不确定性。我们的发现阐明了基于后验和基于似然的概率学习的互补优势,为非线性系统中的不确定性感知建模提供了指导。