Deep neural networks (DNNs) are often coupled with physics-based models or data-driven surrogate models to perform fault detection and health monitoring of systems in the low data regime. These models serve as digital twins to generate large quantities of data to train DNNs which would otherwise be difficult to obtain from the real-life system. However, such models can exhibit parametric uncertainty that propagates to the generated data. In addition, DNNs exhibit uncertainty in the parameters learnt during training. In such a scenario, the performance of the DNN model will be influenced by the uncertainty in the physics-based model as well as the parameters of the DNN. In this article, we quantify the impact of both these sources of uncertainty on the performance of the DNN. We perform explicit propagation of uncertainty in input data through all layers of the DNN, as well as implicit prediction of output uncertainty to capture the former. Furthermore, we adopt Monte Carlo dropout to capture uncertainty in DNN parameters. We demonstrate the approach for fault detection of power lines with a physics-based model, two types of input data and three different neural network architectures. We compare the performance of such uncertainty-aware probabilistic models with their deterministic counterparts. The results show that the probabilistic models provide important information regarding the confidence of predictions, while also delivering an improvement in performance over deterministic models.
翻译:深度神经网络(DNNs)通常与基于物理的模型或数据驱动的代理模型耦合,以在低数据范畴内执行故障检测和健康监测。这些模型作为数字孪生,生成大量的数据来训练DNN,否则要从实际生活系统中获取这些数据很困难。然而,这样的模型可能表现出参数不确定性,导致生成的数据受到影响。此外,DNN在训练过程中学习的参数也表现出不确定性。在这种情况下,DNN模型的性能将受到基于物理的模型的不确定性以及DNN参数的影响。在本文中,我们量化了这两种不确定性对DNN性能的影响。我们通过所有DNN层进行输入数据的显式不确定性传播,并隐式预测输出不确定性以捕获前者。此外,采用蒙特卡罗dropout方法捕获DNN参数的不确定性。我们将该方法演示为物理模型下电力线的故障检测,使用两种类型的输入数据和三种不同的神经网络架构。我们将具有不确定性感知的概率模型与其确定性对应物的性能进行比较。结果表明,概率模型提供了有关预测置信度的重要信息,并且在性能方面比确定性模型有所提升。