In recent years, machine learning algorithms have been applied widely in various fields such as health, transportation, and the autonomous car. With the rapid developments of deep learning techniques, it is critical to take the security concern into account for the application of the algorithms. While machine learning offers significant advantages in terms of the application of algorithms, the issue of security is ignored. Since it has many applications in the real world, security is a vital part of the algorithms. In this paper, we have proposed a mitigation method for adversarial attacks against machine learning models with an autoencoder model that is one of the generative ones. The main idea behind adversarial attacks against machine learning models is to produce erroneous results by manipulating trained models. We have also presented the performance of autoencoder models to various attack methods from deep neural networks to traditional algorithms by using different methods such as non-targeted and targeted attacks to multi-class logistic regression, a fast gradient sign method, a targeted fast gradient sign method and a basic iterative method attack to neural networks for the MNIST dataset.
翻译:近年来,机器学习算法在卫生、交通和自主汽车等各个领域广泛应用。随着深层学习技术的迅速发展,在应用算法时必须考虑到安全问题。虽然机器学习在应用算法方面有很大的优势,但安全问题却被忽视。由于机器学习在现实世界中有许多应用,安全是算法的重要组成部分。在本文中,我们提出了一个减缓对机器学习模型进行对抗性攻击的方法,这种模型是自动编码模型的基因化模型之一。对机器学习模型进行对抗性攻击的主要想法是操纵经过训练的模型,产生错误的结果。我们还介绍了自动编码模型的性能,从深神经网络到传统算法的各种攻击方法,采用不同的方法,如非定向和有针对性的攻击到多级后勤回归、快速梯度标志方法、定向快速梯度标志方法、对内线网络进行MNIST数据集的基本迭代方法攻击。