Achieving transferability of targeted attacks is reputed to be remarkably difficult. Currently, state-of-the-art approaches are resource-intensive because they necessitate training model(s) for each target class with additional data. In our investigation, we find, however, that simple transferable attacks which require neither additional data nor model training can achieve surprisingly high targeted transferability. This insight has been overlooked until now, mainly due to the widespread practice of unreasonably restricting attack optimization to a limited number of iterations. In particular, we, for the first time, identify that a simple logit loss can yield competitive results with the state of the arts. Our analysis spans a variety of transfer settings, especially including three new, realistic settings: an ensemble transfer setting with little model similarity, a worse-case setting with low-ranked target classes, and also a real-world attack against the Google Cloud Vision API. Results in these new settings demonstrate that the commonly adopted, easy settings cannot fully reveal the actual properties of different attacks and may cause misleading comparisons. We also show the usefulness of the simple logit loss for generating targeted universal adversarial perturbations in a data-free and training-free manner. Overall, the aim of our analysis is to inspire a more meaningful evaluation on targeted transferability. Code is available at https://github.com/ZhengyuZhao/Targeted-Tansfer
翻译:据认为,实现有针对性的袭击的可转移性非常困难。目前,最先进的方法需要大量的资源,因为它们需要为每个目标类别制定培训模式,并需要更多数据。然而,在我们的调查中,我们发现,不需要额外数据或示范培训的简单可转移袭击可能达到惊人的高可转移性。这种洞察力至今一直被忽视,主要是因为不合理地将袭击优化限制在数量有限的迭代上的普遍做法。特别是,我们第一次发现,简单登录丢失可以带来与艺术状态的竞争结果。我们的分析涉及各种转让环境,特别是包括三个新的、现实的设置:一个基本转让环境,几乎没有模式相似,一个更糟糕的可转移环境,低级目标类别,以及一个真实世界对谷歌Cloud Right Vision API的攻击。这些新环境的结果表明,通常采用的简易环境无法充分揭示不同袭击的实际性质,并可能导致误导性比较。我们还显示简单的逻辑损失对于以无损率生成有针对性的全球对立性透透度评估的实用性。在无数据/底层进行无目标性分析时,可以进行无目标的无目标的无目标的《准则》。