There has been a concurrent significant improvement in the medical images used to facilitate diagnosis and the performance of machine learning techniques to perform tasks such as classification, detection, and segmentation in recent years. As a result, a rapid increase in the usage of such systems can be observed in the healthcare industry, for instance in the form of medical image classification systems, where these models have achieved diagnostic parity with human physicians. One such application where this can be observed is in computer vision tasks such as the classification of skin lesions in dermatoscopic images. However, as stakeholders in the healthcare industry, such as insurance companies, continue to invest extensively in machine learning infrastructure, it becomes increasingly important to understand the vulnerabilities in such systems. Due to the highly critical nature of the tasks being carried out by these machine learning models, it is necessary to analyze techniques that could be used to take advantage of these vulnerabilities and methods to defend against them. This paper explores common adversarial attack techniques. The Fast Sign Gradient Method and Projected Descent Gradient are used against a Convolutional Neural Network trained to classify dermatoscopic images of skin lesions. Following that, it also discusses one of the most popular adversarial defense techniques, adversarial training. The performance of the model that has been trained on adversarial examples is then tested against the previously mentioned attacks, and recommendations to improve neural networks robustness are thus provided based on the results of the experiment.
翻译:最近几年来,用于促进诊断的医学图象和机器学习技术的运用,以完成分类、检测和分化等任务的医疗图象也同时有了显著改善,因此,在保健行业可以观察到使用这种系统的情况迅速增加,例如医疗图象分类系统,这些模型已经与人类医生实现了诊断等同;可以观察到的其中一个应用是计算机视觉任务,如皮肤图象中的皮肤损伤分类;然而,由于保健行业的利益攸关方,如保险公司,继续大量投资于机器学习基础设施,了解这类系统中的脆弱性变得日益重要;由于这些机器学习模型所执行的任务非常关键,有必要分析可以利用这些弱点和防患于未然的方法来加以利用的技术; 快速信号梯度法和预测源梯度法是用来对付受过训练的神经神经神经网络,此后,还讨论了最受欢迎的对抗性攻击试验技术,并据此对一个经过训练的对抗性攻击试验后,对一个经过训练的对抗性攻击试验的实验性建议进行了改进。