When we can not assume a large amount of annotated data , active learning is a good strategy. It consists in learning a model on a small amount of annotated data (annotation budget) and in choosing the best set of points to annotate in order to improve the previous model and gain in generalization. In deep learning, active learning is usually implemented as an iterative process in which successive deep models are updated via fine tuning, but it still poses some issues. First, the initial batch of annotated images has to be sufficiently large to train a deep model. Such an assumption is strong, especially when the total annotation budget is reduced. We tackle this issue by using an approach inspired by transfer learning. A pre-trained model is used as a feature extractor and only shallow classifiers are learned during the active iterations. The second issue is the effectiveness of probability or feature estimates of early models for AL task. Samples are generally selected for annotation using acquisition functions based only on the last learned model. We introduce a novel acquisition function which exploits the iterative nature of AL process to select samples in a more robust fashion. Samples for which there is a maximum shift towards uncertainty between the last two learned models predictions are favored. A diversification step is added to select samples from different regions of the classification space and thus introduces a representativeness component in our approach. Evaluation is done against competitive methods with three balanced and imbalanced datasets and outperforms them.
翻译:当我们无法承担大量附加说明的数据时,积极学习是一种良好的战略。它包括学习少量附加说明数据(说明预算)的模型,以及选择最佳的一组点数作为注释,以便改进以前的模型,并普遍化。在深层次学习中,积极学习通常是一个迭接过程,通过微调更新连续的深层模型的概率或特征估计,但仍然构成一些问题。首先,最初一批附加说明的图像必须足够大,以培养一个深层次模型。这种假设是强有力的,特别是在总注解预算减少时。我们采用转让学习所启发的方法解决这个问题。先行培训的模型被用作特征提取器,在积极的迭代过程中只学习浅色分类器。第二个问题是,通过微调更新连续的深层模型的概率或特征估计的有效性,但通常只根据最后学习的模型选择样本来进行注解。我们引入了一种新的获取功能,即利用AL进程的迭代性质来选择更加稳健的样本。我们采用的方法是采用最稳健的模型,然后采用最稳的模型,然后采用最稳的模型,然后采用最稳的模型,然后是不同的模型,然后采用不同的模型,然后采用不同的模型,然后采用不同的变式,然后采用不同的模型,然后采用不同的变式,然后采用不同的变式方法。