In this paper we want to present the results of empirical verification of some issues concerning the methods for overcoming catastrophic forgetting in neural networks. First, in the introduction, we will try to describe in detail the problem of catastrophic forgetting and methods for overcoming it for those who are not yet familiar with this topic. Then we will discuss the essence and limitations of the WVA method which we presented in previous papers. Further, we will touch upon the issues of applying the WVA method to gradients or optimization steps of weights, choosing the optimal attenuation function in this method, as well as choosing the optimal hyper-parameters of the method depending on the number of tasks in sequential training of neural networks.
翻译:在本文中,我们想介绍有关克服神经网络中灾难性遗忘的方法的一些问题的经验性核查结果。首先,在导言中,我们将试图详细说明灾难性遗忘问题和尚未熟悉这一专题的人克服这一问题的方法。然后我们将讨论我们在前几份文件中介绍的WVA方法的本质和局限性。此外,我们将谈到将WVA方法应用于梯度或最大重量级步骤、选择这一方法的最佳减肥功能以及根据神经网络连续培训的任务数量选择最佳方法的超参数等问题。