As more and more artificial intelligence (AI) technologies move from the laboratory to real-world applications, the open-set and robustness challenges brought by data from the real world have received increasing attention. Data augmentation is a widely used method to improve model performance, and some recent works have also confirmed its positive effect on the robustness of AI models. However, most of the existing data augmentation methods are heuristic, lacking the exploration of their internal mechanisms. We apply the explainable artificial intelligence (XAI) method, explore the internal mechanisms of popular data augmentation methods, analyze the relationship between game interactions and some widely used robustness metrics, and propose a new proxy for model robustness in the open-set environment. Based on the analysis of the internal mechanisms, we develop a mask-based boosting method for data augmentation that comprehensively improves several robustness measures of AI models and beats state-of-the-art data augmentation approaches. Experiments show that our method can be widely applied to many popular data augmentation methods. Different from the adversarial training, our boosting method not only significantly improves the robustness of models, but also improves the accuracy of test sets. Our code is available at \url{https://github.com/Anonymous_for_submission}.
翻译:随着越来越多的人工智能(AI)技术从实验室转向现实世界应用,来自真实世界的数据带来的开放性和稳健性挑战日益受到越来越多的关注。数据增强是用来改进模型性能的一种广泛使用的方法,最近的一些工作也证实了其对AI模型的稳健性的积极影响。然而,大多数现有的数据增强方法都是超强的,缺乏对内部机制的探索。我们采用可解释的人工智能(XAI)方法,探索流行数据增强方法的内部机制,分析游戏互动和一些广泛使用的稳健度度量度指标之间的关系,并提出开放环境模型稳健性的新替代物。我们根据对内部机制的分析,开发了一种基于面具的增强数据增强方法,全面改进了AI模型的若干稳健性计量,并击败了最新数据增强方法。实验表明,我们的方法可以广泛应用于许多流行的数据增强方法。不同于对口培训,我们的增强方法不仅大大改进了模型的稳健性,而且还提高了测试组的准确性。我们的数据代码可在http://Angur_com}