Neural networks (NNs) have demonstrated their potential in a wide range of applications such as image recognition, decision making or recommendation systems. However, standard NNs are unable to capture their model uncertainty which is crucial for many safety-critical applications including healthcare and autonomous vehicles. In comparison, Bayesian neural networks (BNNs) are able to express uncertainty in their prediction via a mathematical grounding. Nevertheless, BNNs have not been as widely used in industrial practice, mainly because of their expensive computational cost and limited hardware performance. This work proposes a novel FPGA-based hardware architecture to accelerate BNNs inferred through Monte Carlo Dropout. Compared with other state-of-the-art BNN accelerators, the proposed accelerator can achieve up to 4 times higher energy efficiency and 9 times better compute efficiency. Considering partial Bayesian inference, an automatic framework is proposed, which explores the trade-off between hardware and algorithmic performance. Extensive experiments are conducted to demonstrate that our proposed framework can effectively find the optimal points in the design space.
翻译:神经网络(NNs)在图像识别、决策或建议系统等广泛应用中显示了其潜力,但标准NNs无法捕捉其模型不确定性,而模型不确定性对于许多安全关键应用,包括保健和自主车辆,至关重要。相比之下,Bayesian神经网络(BNNs)能够通过数学基础在预测中表达不确定性。然而,BNes没有在工业实践中被广泛使用,主要是因为其计算成本昂贵,硬件性能有限。这项工作提出了基于FPGA的新型硬件结构,以加速通过Monte Carlo 抛出得出的BNS。与其他先进的BNNN加速器相比,拟议的加速器可以达到4倍的能源效率和9倍的承受效率。考虑到部分Bayesian的推论,提出了一个自动框架,探讨硬件和算法性能之间的取舍。进行了广泛的实验,以证明我们提议的框架能够有效地找到设计空间的最佳点。