As acquiring manual labels on data could be costly, unsupervised domain adaptation (UDA), which transfers knowledge learned from a rich-label dataset to the unlabeled target dataset, is gaining increasing popularity. While extensive studies have been devoted to improving the model accuracy on target domain, an important issue of model robustness is neglected. To make things worse, conventional adversarial training (AT) methods for improving model robustness are inapplicable under UDA scenario since they train models on adversarial examples that are generated by supervised loss function. In this paper, we present a new meta self-training pipeline, named SRoUDA, for improving adversarial robustness of UDA models. Based on self-training paradigm, SRoUDA starts with pre-training a source model by applying UDA baseline on source labeled data and taraget unlabeled data with a developed random masked augmentation (RMA), and then alternates between adversarial target model training on pseudo-labeled target data and finetuning source model by a meta step. While self-training allows the direct incorporation of AT in UDA, the meta step in SRoUDA further helps in mitigating error propagation from noisy pseudo labels. Extensive experiments on various benchmark datasets demonstrate the state-of-the-art performance of SRoUDA where it achieves significant model robustness improvement without harming clean accuracy. Code is available at https://github.com/Vision.
翻译:由于获取关于数据的手工标签可能成本高昂,不受监督的域适应(UDA)将丰富标签数据集获得的知识转让给未贴标签的目标数据集,因此越来越受欢迎。虽然对改进目标域的模型准确性进行了广泛的研究,但忽略了模型稳健性这一重要问题。更糟糕的是,在UDA设想下,用于改进模型稳健性的常规对抗性培训(AT)方法在UDA设想下是不适用的,因为它们就受监督的损失函数生成的对抗性实例进行示范。在本文中,我们介绍了一个新的自培训管道,名为SROUDA, 用于改进UDA模式的对抗性强性强性。在自我培训模式的基础上,SROUDA开始对源模型进行预先培训,将UDA的基线应用于源代码标签数据基准,并将无标签的无标签数据与开发的随机掩码增强值(RMA)相拖动,然后在假标签目标目标数据模型培训模式和元性调整源模型之间进行替代。自我培训使AT直接纳入UDA,而SROUDA的元性步骤则是在SROUDA的代谢性精确性改进标准的元性升级,从而进一步展示了显著的升级的精确性测试,从而进一步展示了SROUDUDUDA的精确性测试。