In this paper, we introduce a novel combination of Bayesian Models (BMs) and Neural Networks (NNs) for making predictions with a minimum expected risk. Our approach combines the best of both worlds, the data efficiency and interpretability of a BM with the speed of a NN. For a BM, making predictions with the lowest expected loss requires integrating over the posterior distribution. In cases for which exact inference of the posterior predictive distribution is intractable, approximation methods are typically applied, e.g. Monte Carlo (MC) simulation. The more samples, the higher the accuracy -- but at the expense of increased computational cost. Our approach removes the need for iterative MC simulation on the CPU at prediction time. In brief, it works by fitting a NN to synthetic data generated using the BM. In a single feed-forward pass of the NN, it gives a set of point-wise approximations to the BM's posterior predictive distribution for a given observation. We achieve risk minimized predictions significantly faster than standard methods with a negligible loss on the testing dataset. We combine this approach with Active Learning (AL) to minimize the amount of data required for fitting the NN. This is done by iteratively labeling more data in regions with high predictive uncertainty of the NN.
翻译:在本文中,我们引入了一种新型的Bayesian模型(BMS)和神经网络(NNs)组合,以作出预测并降低预期风险。我们的方法结合了两个世界的最佳数据、数据效率和BB的可解释性以及NN的速度。对于一个BM,用预期损失最小的预测需要结合后视分布。在精确推断后视预测分布是棘手的案例中,通常会采用近似方法,例如Monte Carlo(MC)模拟。更多的样本,准确性越高,但以更高的计算成本为代价。我们的方法消除了在预测时间对CPU进行反复的MC模拟的需要。简而言之,它将NN与使用BM生成的合成数据相匹配。在NNN的单向前传中,它给BM的后视预测分布提供了一套点性近似近似的近似值。我们的风险比标准方法要小得多,在测试CPU时损失微不足道。我们把NF的方法与高的标签的这一高数值合并起来。我们把NFIL的方法与DL的更新数据与DLAdisal的更新所需的高数值合并起来。