The reliability of a learning model is key to the successful deployment of machine learning in various industries. Creating a robust model, particularly one unaffected by adversarial attacks, requires a comprehensive understanding of the adversarial examples phenomenon. However, it is difficult to describe the phenomenon due to the complicated nature of the problems in machine learning. Consequently, many studies investigate the phenomenon by proposing a simplified model of how adversarial examples occur and validate it by predicting some aspect of the phenomenon. While these studies cover many different characteristics of the adversarial examples, they have not reached a holistic approach to the geometric and analytic modeling of the phenomenon. This paper propose a formal framework to study the phenomenon in learning theory and make use of complex analysis and holomorphicity to offer a robust learning rule for artificial neural networks. With the help of complex analysis, we can effortlessly move between geometric and analytic perspectives of the phenomenon and offer further insights on the phenomenon by revealing its connection with harmonic functions. Using our model, we can explain some of the most intriguing characteristics of adversarial examples, including transferability of adversarial examples, and pave the way for novel approaches to mitigate the effects of the phenomenon.
翻译:学习模式的可靠性是在不同行业成功部署机器学习的关键。创建一个强大的模型,特别是不受敌对式攻击影响的模型,要求全面理解对抗性实例现象。然而,由于机器学习问题的复杂性,很难描述这种现象。因此,许多研究通过提出一个简化模型来调查这一现象,说明对抗性实例是如何发生的,并通过预测这一现象的某些方面加以验证。虽然这些研究涵盖了对抗性实例的许多不同特征,但它们尚未对这一现象的几何和分析性模型采取全面的方法。本文提出了一个正式框架,以研究理论学现象,并利用复杂的分析和全貌性为人工神经网络提供强有力的学习规则。在复杂分析的帮助下,我们可以不遗余力地在对这一现象进行几何与分析性观点之间移动,并通过揭示其与调理功能的联系来进一步洞察该现象。我们可以用我们的模型来解释敌对性实例的一些最令人感兴趣的特征,包括对抗性实例的可转移性,并铺平减轻新现象影响的新办法。