Generally, regularization-based continual learning models limit access to the previous task data to imitate the real-world setting which has memory and privacy issues. However, this introduces a problem in these models by not being able to track the performance on each task. In other words, current continual learning methods are vulnerable to attacks done on the previous task. We demonstrate the vulnerability of regularization-based continual learning methods by presenting simple task-specific training time adversarial attack that can be used in the learning process of a new task. Training data generated by the proposed attack causes performance degradation on a specific task targeted by the attacker. Experiment results justify the vulnerability proposed in this paper and demonstrate the importance of developing continual learning models that are robust to adversarial attack.
翻译:一般而言,基于正规化的持续学习模式限制人们利用先前的任务数据,以模仿具有记忆和隐私问题的真实世界环境,然而,这给这些模式带来了一个问题,因为无法跟踪每项任务的业绩;换言之,目前的持续学习方法很容易受到前一项任务的攻击;我们通过提供可用于新任务的学习过程的简单的任务特定培训时间对抗性攻击,来证明基于正规化的持续学习方法的脆弱性;拟议攻击所产生的培训数据导致攻击者针对特定任务的工作业绩退化;实验结果证明本文件中提议的脆弱程度是正当的,并表明开发对对抗性攻击具有活力的持续学习模式的重要性。