Adversarial training (i.e., training on adversarially perturbed input data) is a well-studied method for making neural networks robust to potential adversarial attacks during inference. However, the improved robustness does not come for free but rather is accompanied by a decrease in overall model accuracy and performance. Recent work has shown that, in practical robot learning applications, the effects of adversarial training do not pose a fair trade-off but inflict a net loss when measured in holistic robot performance. This work revisits the robustness-accuracy trade-off in robot learning by systematically analyzing if recent advances in robust training methods and theory in conjunction with adversarial robot learning can make adversarial training suitable for real-world robot applications. We evaluate a wide variety of robot learning tasks ranging from autonomous driving in a high-fidelity environment amenable to sim-to-real deployment, to mobile robot gesture recognition. Our results demonstrate that, while these techniques make incremental improvements on the trade-off on a relative scale, the negative side-effects caused by adversarial training still outweigh the improvements by an order of magnitude. We conclude that more substantial advances in robust learning methods are necessary before they can benefit robot learning tasks in practice.
翻译:Aversarial培训(即关于对抗性扰动输入数据的培训)是使神经网络在推论期间对潜在对抗性攻击具有活力的一种研究周密的方法,但是,增强的稳健性并非免费的,而是伴随着总体模型准确性和性能的下降。最近的工作表明,在实际的机器人学习应用中,对抗性培训的效果不是一种公平的权衡,但在以整体机器人性能衡量时造成净损失。这项工作通过系统分析强健的培训方法和理论的最新进展以及对抗性机器人学习的理论能否使对抗性培训适合现实世界机器人应用。我们评估了范围广泛的各种机器人学习任务,从在适合模拟到实际部署的高纤维环境中自主驾驶到移动式机器人手势的承认。我们的结果表明,虽然这些技术使交易在相对规模上取得了逐步的改进,但对抗性培训造成的负面副作用仍然超过其规模的改进。我们的结论是,在采用较强的机器人学习方法之前,更实质性的进展是必要的。