This work focuses on the problem of hyper-parameter tuning (HPT) for robust (i.e., adversarially trained) models, with the twofold goal of i) establishing which additional HPs are relevant to tune in adversarial settings, and ii) reducing the cost of HPT for robust models. We pursue the first goal via an extensive experimental study based on 3 recent models widely adopted in the prior literature on adversarial robustness. Our findings show that the complexity of the HPT problem, already notoriously expensive, is exacerbated in adversarial settings due to two main reasons: i) the need of tuning additional HPs which balance standard and adversarial training; ii) the need of tuning the HPs of the standard and adversarial training phases independently. Fortunately, we also identify new opportunities to reduce the cost of HPT for robust models. Specifically, we propose to leverage cheap adversarial training methods to obtain inexpensive, yet highly correlated, estimations of the quality achievable using state-of-the-art methods (PGD). We show that, by exploiting this novel idea in conjunction with a recent multi-fidelity optimizer (taKG), the efficiency of the HPT process can be significantly enhanced.
翻译:本文主要研究针对鲁棒(即对抗性训练)模型的超参数调整(HPT)问题,旨在确定在对抗环境下需要调整哪些额外的超参数,并降低鲁棒模型的超参数调整成本。通过对3种在对抗性鲁棒性文献中广泛采用的新模型进行广泛的实验研究,我们实现了第一个目标。我们的研究发现,由于两个主要原因,即需要平衡标准和对抗训练的额外超参数以及需要独立调整标准和对抗训练阶段的超参数,HPT问题的复杂性,已经众所周知的昂贵,会在对抗环境中恶化。幸运的是,我们也发现了一些新的机会,以降低鲁棒模型的HPT成本。具体来说,我们建议利用廉价的对抗训练方法,以获得成本低廉但高度相关的使用最新方法(PGD)可实现的质量估计。我们展示了通过将这种新颖的想法与最近的多精度优化器(taKG)相结合,可以显著提高HPT过程的效率。