Despite the vast body of literature on Active Learning (AL), there is no comprehensive and open benchmark allowing for efficient and simple comparison of proposed samplers. Additionally, the variability in experimental settings across the literature makes it difficult to choose a sampling strategy, which is critical due to the one-off nature of AL experiments. To address those limitations, we introduce OpenAL, a flexible and open-source framework to easily run and compare sampling AL strategies on a collection of realistic tasks. The proposed benchmark is augmented with interpretability metrics and statistical analysis methods to understand when and why some samplers outperform others. Last but not least, practitioners can easily extend the benchmark by submitting their own AL samplers.
翻译:尽管关于主动学习(AL)的大量文献已经存在,但仍缺乏一个全面且开放的基准,可以轻松比较所提出的采样器。此外,文献中实验设置的变异性使得在选择采样策略时很难做出决策,这是由于AL实验的一次性性质所必需的。为解决这些限制,我们引入了OpenAL,这是一个灵活且开源的框架,可轻松在一组现实任务上运行和比较采样AL策略。所提出的基准被扩展为具有可解释性度量和统计分析方法,以了解何时和为什么某些采样器优于其他采样器。最后,实践者可以通过提交自己的AL采样器轻松扩展该基准。