Recent studies have shown that, like traditional machine learning, federated learning (FL) is also vulnerable to adversarial attacks. To improve the adversarial robustness of FL, federated adversarial training (FAT) methods have been proposed to apply adversarial training locally before global aggregation. Although these methods demonstrate promising results on independent identically distributed (IID) data, they suffer from training instability on non-IID data with label skewness, resulting in degraded natural accuracy. This tends to hinder the application of FAT in real-world applications where the label distribution across the clients is often skewed. In this paper, we study the problem of FAT under label skewness, and reveal one root cause of the training instability and natural accuracy degradation issues: skewed labels lead to non-identical class probabilities and heterogeneous local models. We then propose a Calibrated FAT (CalFAT) approach to tackle the instability issue by calibrating the logits adaptively to balance the classes. We show both theoretically and empirically that the optimization of CalFAT leads to homogeneous local models across the clients and better convergence points.
翻译:最近的研究显示,与传统的机器学习一样,联合学习(FL)也容易受到对抗性攻击。为了提高FL的对抗性强力,提议在全球汇总之前在当地应用联合对抗性训练方法。虽然这些方法显示独立分布相同的(IID)数据取得了有希望的结果,但它们在非IID数据方面受到培训不稳定,标签偏差导致自然精确度下降。这往往妨碍FAT在现实应用中应用FAT,因为客户之间的标签分布往往被扭曲。在本文中,我们研究了FAT问题,并揭示了培训不稳定和自然准确性退化问题的根源之一:扭曲的标签导致非同级的班级概率和混杂的本地模型。我们随后建议用校准FAT(CalFAT)方法解决不稳定问题,通过适应性校准对对日志来平衡各个班级。我们从理论上和实证上都表明,CalFAT的优化导致客户之间的当地模型趋同性以及更好的趋同点。