The acquisition of labels for supervised learning can be expensive. In order to improve the sample-efficiency of neural network regression, we study active learning methods that adaptively select batches of unlabeled data for labeling. We present a framework for constructing such methods out of (network-dependent) base kernels, kernel transformations and selection methods. Our framework encompasses many existing Bayesian methods based on Gaussian Process approximations of neural networks as well as non-Bayesian methods. Additionally, we propose to replace the commonly used last-layer features with sketched finite-width Neural Tangent Kernels, and to combine them with a novel clustering method. To evaluate different methods, we introduce an open-source benchmark consisting of 15 large tabular regression data sets. Our proposed method outperforms the state-of-the-art on our benchmark, scales to large data sets, and works out-of-the-box without adjusting the network architecture or training code. We provide open-source code that includes efficient implementations of all kernels, kernel transformations, and selection methods, and can be used for reproducing our results.
翻译:为了提高神经网络回归的样本效率,我们研究主动学习方法,以适应性方式选择一组无标签数据进行标签。我们提出了一个框架,用(依赖网络的)基内核、内核变换和选择方法构建这些方法。我们的框架包括许多基于高斯进程神经网络近似和非拜耶方法的现有巴伊西亚方法。此外,我们提议用草图定定型的内脏和内核中枢取代常用的最后一层特征,并将之与新颖的集群方法相结合。为了评估不同方法,我们采用了由15个大表式回归数据集组成的开放源基准。我们的拟议方法超越了我们基准的状态、尺度到大型数据集,并在不调整网络结构或培训代码的情况下从框中工作。我们提供了开放源代码,其中包括高效率地实施所有内核、内核变和选择方法,并可用于复制我们的成果。