Smart healthcare systems are gaining popularity with the rapid development of intelligent sensors, the Internet of Things (IoT) applications and services, and wireless communications. However, at the same time, several vulnerabilities and adversarial attacks make it challenging for a safe and secure smart healthcare system from a security point of view. Machine learning has been used widely to develop suitable models to predict and mitigate attacks. Still, the attacks could trick the machine learning models and misclassify outputs generated by the model. As a result, it leads to incorrect decisions, for example, false disease detection and wrong treatment plans for patients. In this paper, we address the type of adversarial attacks and their impact on smart healthcare systems. We propose a model to examine how adversarial attacks impact machine learning classifiers. To test the model, we use a medical image dataset. Our model can classify medical images with high accuracy. We then attacked the model with a Fast Gradient Sign Method attack (FGSM) to cause the model to predict the images and misclassify them inaccurately. Using transfer learning, we train a VGG-19 model with the medical dataset and later implement the FGSM to the Convolutional Neural Network (CNN) to examine the significant impact it causes on the performance and accuracy of the machine learning model. Our results demonstrate that the adversarial attack misclassifies the images, causing the model's accuracy rate to drop from 88% to 11%.
翻译:随着智能传感器的迅速发展,智能智能保健体系的迅速发展,Tings Internet(IoT)应用程序和服务以及无线通信的迅速发展,智能保健体系日益受到欢迎。然而,与此同时,一些弱点和对抗性攻击使得从安全角度对一个安全可靠的智能保健体系提出了挑战。机器学习被广泛用来开发适当的模型来预测和减轻攻击。不过,这些攻击可能欺骗机器学习模型,错误地分类模型生成的结果。因此,它导致错误的决定,例如,错误的疾病检测和病人治疗计划。在本文中,我们处理对抗性攻击的类型及其对智能保健系统的影响。我们提出了一个模型来检查对抗性攻击如何影响机的分类师。测试模型时,我们使用一个医疗图像数据集。我们的模型可以非常精确地对医疗图像进行分类,以快速渐进的信号方法攻击模型(FGSM)导致模型预测图像并错误地分类这些图像。我们用一个VCNGG-19模型来研究医疗数据集的种类及其对智能保健系统的影响。我们提出了一个模型,然后将FGSMSM 应用FSM到革命性模型来影响机器的精确度网络,从而展示了我们的重要的精确度。