Attack vectors that compromise machine learning pipelines in the physical world have been demonstrated in recent research, from perturbations to architectural components. Building on this work, we illustrate the self-obfuscation attack: attackers target a pre-processing model in the system, and poison the training set of generative models to obfuscate a specific class during inference. Our contribution is to describe, implement and evaluate a generalized attack, in the hope of raising awareness regarding the challenge of architectural robustness within the machine learning community.
翻译:在最近的研究中,从扰动到建筑构件,我们以这项工作为基础,展示了自我模糊攻击:攻击者瞄准系统中的预处理模型,毒害一套基因模型训练,在推论期间混淆某个特定类别。我们的贡献是描述、实施和评价一场普遍攻击,希望提高人们对机器学习界建筑坚固性挑战的认识。