The bias-variance trade-off is a well-known problem in machine learning that only gets more pronounced the less available data there is. In active learning, where labeled data is scarce or difficult to obtain, neglecting this trade-off can cause inefficient and non-optimal querying, leading to unnecessary data labeling. In this paper, we focus on active learning with Gaussian Processes (GPs). For the GP, the bias-variance trade-off is made by optimization of the two hyperparameters: the length scale and noise-term. Considering that the optimal mode of the joint posterior of the hyperparameters is equivalent to the optimal bias-variance trade-off, we approximate this joint posterior and utilize it to design two new acquisition functions. The first one is a Bayesian variant of Query-by-Committee (B-QBC), and the second is an extension that explicitly minimizes the predictive variance through a Query by Mixture of Gaussian Processes (QB-MGP) formulation. Across six common simulators, we empirically show that B-QBC, on average, achieves the best marginal likelihood, whereas QB-MGP achieves the best predictive performance. We show that incorporating the bias-variance trade-off in the acquisition functions mitigates unnecessary and expensive data labeling.
翻译:在机器学习中,偏差权衡是一个众所周知的问题,它只会更明显地显示现有数据较少。在积极学习中,标签数据稀少或难以获得,忽略这种权衡可能导致低效率和非最佳的查询,导致不必要的数据标签。在本文中,我们侧重于与Gausian process(GPs)的积极学习。在GP中,偏差权衡是通过优化两个超参数(长度尺度和噪音期)实现的。考虑到超参数联合远地点的最佳模式相当于最佳偏差折交易,我们在积极学习中,忽略这一权衡可能导致低效率和非最佳的查询,从而导致不必要的数据标签标签标签标签标签标签标签标签标签标签标签标签标签制度(QB-MGP)的优化,我们用它来设计两个新的获取功能。在本文中,我们侧重于与Gaysian Quarby-Comproach (B-QB-B) 的变换式,而第二个是扩展,通过Mixservering(Q-B-MGP)的查询来明确将预测差异最小化,我们的经验显示最佳的购买率显示B-B的成绩。