Active learning is popular approach for reducing the amount of data in training deep neural network model. Its success hinges on the choice of an effective acquisition function, which ranks not yet labeled data points according to their expected informativeness. In uncertainty sampling, the uncertainty that the current model has about a point's class label is the main criterion for this type of ranking. This paper proposes a new approach to uncertainty sampling in training a Convolutional Neural Network (CNN). The main idea is to use feature representation extracted extracted by the CNN as data for training a Sum-Product Network (SPN). Since SPNs are typically used for estimating the distribution of a dataset, they are well suited to the task of estimating class probabilities that can be used directly by standard acquisition functions such as max entropy and variational ratio. Moreover, we enhance these acquisition functions by weights calculated with the help of the SPN model; these weights make the acquisition function more sensitive to the diversity of conceivable class labels for data points. The effectiveness of our method is demonstrated in an experimental study on the MNIST, Fashion-MNIST and CIFAR-10 datasets, where we compare it to the state-of-the-art methods MC Dropout and Bayesian Batch.
翻译:积极学习是减少深神经网络模型培训中数据数量的流行方法,其成功取决于选择有效的获取功能,该功能通常用于估计数据集的分布,因此,根据预期的信息性,将尚未标出数据点排在标签上。在不确定性抽样中,当前模型对某一点等级标签的不确定性是这类排名的主要标准。本文提出在培训进化神经网络时采用新的不确定性抽样方法。主要想法是利用CNN提取的特征代表作为数据点培训数据点的数据。由于SPN通常用于估计数据集的分布,因此,SPN非常适合估算可以通过标准获取功能直接使用的类别概率,例如最大诱变率和变异率。此外,我们用SPN模型计算出的权重来增强这些获取功能;这些权重使获取功能对数据点方面可想象的等级标签的多样性更加敏感。我们的方法在对MNIST、Fashin-MINST和CIFAR-10MS进行实验性研究时展示了效力,我们将它与BAR-MS-BAR-MS的状态进行比较。