Active learning is typically used to label data, when the labeling process is expensive. Several active learning algorithms have been theoretically proved to perform better than their passive counterpart. However, these algorithms rely on some assumptions, which themselves contain some specific parameters. This paper adresses the problem of adaptive active learning in a nonparametric setting with minimal assumptions. We present a novel algorithm that is valid under more general assumptions than the previously known algorithms, and that can moreover adapt to the parameters used in these assumptions. This allows us to work with a larger class of distributions, thereby avoiding to exclude important densities like gaussians. Our algorithm achieves a minimax rate of convergence, and therefore performs almost as well as the best known non-adaptive algorithms.
翻译:当标签过程昂贵时,主动学习通常用于标签数据。 几个主动学习算法在理论上被证明比被动对应算法表现更好。 但是,这些算法依赖某些假设,而这些假设本身就包含某些具体参数。 本文在非参数化的环境下,用最低假设来形容适应性积极学习的问题。 我们提出了一个比先前已知算法更一般的假设有效的新奇算法, 并且可以适应这些假设中使用的参数。 这使我们能够用更大规模的分布法来工作, 从而避免排除象大便士这样的重要密度。 我们的算法实现了微缩的趋同率, 因而几乎实现了最著名的非适应性算法。