Deep neural networks (DNNs), are widely used in many industries such as image recognition, supply chain, medical diagnosis, and autonomous driving. However, prior work has shown the high accuracy of a DNN model does not imply high robustness (i.e., consistent performances on new and future datasets) because the input data and external environment (e.g., software and model configurations) for a deployed model are constantly changing. Hence, ensuring the robustness of deep learning is not an option but a priority to enhance business and consumer confidence. Previous studies mostly focus on the data aspect of model variance. In this article, we systematically summarize DNN robustness issues and formulate them in a holistic view through two important aspects, i.e., data and software configuration variances in DNNs. We also provide a predictive framework to generate representative variances (counterexamples) by considering both data and configurations for robust learning through the lens of search-based optimization.
翻译:深神经网络(DNN)在许多行业被广泛使用,如图像识别、供应链、医疗诊断和自主驱动等,但是,先前的工作表明DNN模型的高度准确性并不意味着高度稳健(即新的和未来数据集的一致性能),因为投入数据和外部环境(如软件和模型配置)对部署模型的影响不断变化。因此,确保深层学习的稳健性不是提高商业和消费者信心的一个选项,而是一个优先事项。以往的研究主要侧重于模型差异的数据方面。在本条中,我们系统地总结DNNN稳健问题,并通过两个重要方面,即数字NNNS的数据和软件配置差异,从整体角度来制定这些问题。我们还提供了一个预测框架,通过基于搜索的优化镜头,为强健学习而考虑数据和配置,从而产生代表性差异(对应特征)。