Due to the increasing usage of machine learning (ML) techniques in security- and safety-critical domains, such as autonomous systems and medical diagnosis, ensuring correct behavior of ML systems, especially for different corner cases, is of growing importance. In this paper, we propose a generic framework for evaluating security and robustness of ML systems using different real-world safety properties. We further design, implement and evaluate VeriVis, a scalable methodology that can verify a diverse set of safety properties for state-of-the-art computer vision systems with only blackbox access. VeriVis leverage different input space reduction techniques for efficient verification of different safety properties. VeriVis is able to find thousands of safety violations in fifteen state-of-the-art computer vision systems including ten Deep Neural Networks (DNNs) such as Inception-v3 and Nvidia's Dave self-driving system with thousands of neurons as well as five commercial third-party vision APIs including Google vision and Clarifai for twelve different safety properties. Furthermore, VeriVis can successfully verify local safety properties, on average, for around 31.7% of the test images. VeriVis finds up to 64.8x more violations than existing gradient-based methods that, unlike VeriVis, cannot ensure non-existence of any violations. Finally, we show that retraining using the safety violations detected by VeriVis can reduce the average number of violations up to 60.2%.
翻译:由于在安保和安全关键领域越来越多地使用机器学习技术(ML),例如自主系统和医疗诊断,确保ML系统的正确行为,特别是针对不同角落的病例,因此越来越重要。在本文件中,我们提出一个通用框架,用于评估使用不同现实世界安全特性的ML系统的安全和稳健性;我们进一步设计、实施和评价VeriVisis,这是一个可推广的方法,可用于核查最新计算机视觉系统的各种安全特性,只有黑匣接入。VeriVis利用不同的投入空间减少技术对不同安全特性进行有效的核查。VeriVis利用15个最先进的计算机视觉系统,包括10个深神经网络(DNNNS),例如Incepion-V3和Nvidia的Dave自驾驶系统,使用数千个神经元和5个商业第三方视觉API,包括谷歌视觉和Clarifai,用于12种安全特性的API。 Verivis Visivis能够成功地核查当地安全特性,平均为60.7%左右,无法使用VER-VA级的不透明图像。我们最后发现任何侵犯情况。