The ability to continuously process and retain new information like we do naturally as humans is a feat that is highly sought after when training neural networks. Unfortunately, the traditional optimization algorithms often require large amounts of data available during training time and updates wrt. new data are difficult after the training process has been completed. In fact, when new data or tasks arise, previous progress may be lost as neural networks are prone to catastrophic forgetting. Catastrophic forgetting describes the phenomenon when a neural network completely forgets previous knowledge when given new information. We propose a novel training algorithm called training by explaining in which we leverage Layer-wise Relevance Propagation in order to retain the information a neural network has already learned in previous tasks when training on new data. The method is evaluated on a range of benchmark datasets as well as more complex data. Our method not only successfully retains the knowledge of old tasks within the neural networks but does so more resource-efficiently than other state-of-the-art solutions.
翻译:持续处理和保留新信息的能力,就像我们自然的人类一样,是培养神经网络时高度追求的一门成就。 不幸的是,传统的优化算法往往需要大量培训时间和更新过程中可获得的数据。 培训过程完成后,新的数据很困难。 事实上,当出现新的数据或任务时,由于神经网络容易被灾难性地遗忘,先前的进展可能会丧失。 灾难性的遗忘描述了神经网络在提供新信息时完全忘记先前知识时出现的现象。 我们提出一种新的培训算法,称为培训,解释我们如何利用图层智慧的“相关性促进”来保留在培训新数据时在以往任务中已经学到的信息。该方法在一系列基准数据集以及更为复杂的数据上被评估。 我们的方法不仅成功地保留了神经网络中旧任务的知识,而且比其他最先进的解决方案更节省资源。