Discriminative Feature Feedback is a setting proposed by Dastupta et al. (2018), which provides a protocol for interactive learning based on feature explanations that are provided by a human teacher. The features distinguish between the labels of pairs of possibly similar instances. That work has shown that learning in this model can have considerable statistical and computational advantages over learning in standard label-based interactive learning models. In this work, we provide new robust interactive learning algorithms for the Discriminative Feature Feedback model, with mistake bounds that are significantly lower than those of previous robust algorithms for this setting. In the adversarial setting, we reduce the dependence on the number of protocol exceptions from quadratic to linear. In addition, we provide an algorithm for a slightly more restricted model, which obtains an even smaller mistake bound for large models with many exceptions. In the stochastic setting, we provide the first algorithm that converges to the exception rate with a polynomial sample complexity. Our algorithm and analysis for the stochastic setting involve a new construction that we call Feature Influence, which may be of wider applicability.
翻译:Dastupta等人(2018年)建议设置差异性地物反馈,这是Dastupta等人(2018年)提出的一个设置,它提供了一个基于人类教师提供的特征解释的互动式学习协议。其特征区分了可能相似实例的配对标签。该工作表明,该模型的学习在统计和计算上比标准标签基互动学习模型的学习具有相当大的优势。在这项工作中,我们为差异性地物反馈模型提供了新的强有力的互动式学习算法,其错误范围大大低于先前的强势算法。在对抗性环境中,我们减少了对协议例外从四面形到线形的依赖。此外,我们为一个略为限制性的模型提供了一种算法,该模型除了许多例外外,还会获得更小的错误。在随机环境中,我们提供了第一个与例外率一致的、具有多面抽样复杂性的算法。我们用于随机环境的算法和分析涉及一种新构造,我们称之为“特性影响”,它可能具有更广泛的适用性。