While directly fine-tuning (FT) large-scale, pretrained models on task-specific data is well-known to induce strong in-distribution task performance, recent works have demonstrated that different adaptation protocols, such as linear probing (LP) prior to FT, can improve out-of-distribution generalization. However, the design space of such adaptation protocols remains under-explored and the evaluation of such protocols has primarily focused on distribution shifts. Therefore, in this work, we evaluate common adaptation protocols across distributions shifts and machine learning safety metrics (e.g., anomaly detection, calibration, robustness to corruptions). We find that protocols induce disparate trade-offs that were not apparent from prior evaluation. Further, we demonstrate that appropriate pairing of data augmentation and protocol can substantially mitigate this trade-off. Finally, we hypothesize and empirically see that using hardness-promoting augmentations during LP and then FT with augmentations may be particularly effective for trade-off mitigation.
翻译:虽然直接微调(FT)大规模任务特定数据的预先培训模型众所周知,能够产生很强的分配任务绩效,但最近的工作表明,不同的适应协议,如FT之前的线性勘测(LP),可以改进分配外的概括性,然而,这类适应协议的设计空间仍然探索不足,而这类协议的评价主要侧重于分配转移。因此,在这项工作中,我们评价了分布转移和机器学习安全度量(例如异常检测、校准、对腐败的稳健性)之间的共同适应协议。 我们发现,协议产生了不同程度的取舍,而以前的评估并不明显。此外,我们证明,适当的数据增强和协议配对可以大大减轻这种取舍。最后,我们虚伪和实证地发现,在LP期间使用硬性促进增强增强作用,然后采用增强作用的FT可能对于贸易减缓特别有效。