Most modern neural networks for classification fail to take into account the concept of the unknown. Trained neural networks are usually tested in an unrealistic scenario with only examples from a closed set of known classes. In an attempt to develop a more realistic model, the concept of working in an open set environment has been introduced. This in turn leads to the concept of incremental learning where a model with its own architecture and initial trained set of data can identify unknown classes during the testing phase and autonomously update itself if evidence of a new class is detected. Some problems that arise in incremental learning are inefficient use of resources to retrain the classifier repeatedly and the decrease of classification accuracy as multiple classes are added over time. This process of instantiating new classes is repeated as many times as necessary, accruing errors. To address these problems, this paper proposes the Classification Confidence Threshold approach to prime neural networks for incremental learning to keep accuracies high by limiting forgetting. A lean method is also used to reduce resources used in the retraining of the neural network. The proposed method is based on the idea that a network is able to incrementally learn a new class even when exposed to a limited number samples associated with the new class. This method can be applied to most existing neural networks with minimal changes to network architecture.
翻译:用于分类的现代神经网络大多没有考虑到未知概念的概念。 受过训练的神经网络通常在一个不现实的情景中测试,仅以一组封闭的已知阶级为例。 为了试图开发一个更现实的模式,已经引入了在开放的一组环境中工作的概念。 这反过来又导致一种递增学习的概念,在测试阶段,一个具有自身架构和初步训练有素的数据集的模型可以识别未知的类别,如果发现新类的证据,则可以自动更新; 增量学习中产生的一些问题是, 利用资源对分类器反复进行再培训效率低下, 以及随着多类被长期添加而降低分类精度。 这种即时新类的即刻化过程会视需要重复重复, 累积错误。 为了解决这些问题, 本文建议对初级神经网络进行分类信任阈值处理, 以便通过限制遗忘来保持高的封闭状态。 一种精干的方法还用来减少神经网络再培训中使用的资源。 所提议的方法基于这样的理念,即网络能够逐步学习一个新的类, 即使当接触到一个有限的网络时, 也应用到与新类中最小的样本。