Learning from imprecise labels such as "animal" or "bird", but making precise predictions like "snow bunting" at test time is an important capability when expertly labeled training data is scarce. Contributions by volunteers or results of web crawling lack precision in this manner, but are still valuable. And crucially, these weakly labeled examples are available in larger quantities for lower cost than high-quality bespoke training data. CHILLAX, a recently proposed method to tackle this task, leverages a hierarchical classifier to learn from imprecise labels. However, it has two major limitations. First, it is not capable of learning from effectively unlabeled examples at the root of the hierarchy, e.g. "object". Second, an extrapolation of annotations to precise labels is only performed at test time, where confident extrapolations could be already used as training data. In this work, we extend CHILLAX with a self-supervised scheme using constrained extrapolation to generate pseudo-labels. This addresses the second concern, which in turn solves the first problem, enabling an even weaker supervision requirement than CHILLAX. We evaluate our approach empirically and show that our method allows for a consistent accuracy improvement of 0.84 to 1.19 percent points over CHILLAX and is suitable as a drop-in replacement without any negative consequences such as longer training times.
翻译:从“ 动物” 或“ 鸟类” 等不精确的标签中学习, 但是在测试时作出精确的预测, 当有专家标签的培训数据缺乏时, 这是一种重要的能力。 志愿者的贡献或网络爬动的结果缺乏这种方式的精确性, 但仍然有价值。 最重要的是, 这些标签不准确的例子数量比高质量的培训数据要大, 成本比高质量的公开的培训数据要低。 CHILLAX是最近提出的一个处理这项任务的方法, 利用一个等级分类器从不精确的标签中学习。 但是, 它有两个主要的局限性。 首先, 它无法从结构根部的无标签的例子中有效地学习。 例如“ 目标 ” 。 其次, 精确标签的注释的外推法只能在测试时间进行, 而在测试时间里, 信心外推法已经被用来作为培训数据。 在这项工作中, 我们推广 CHILLAX, 使用一个自上调的办法, 使用有限的外推法来生成假标签。 这解决了第二个问题, 这又解决了第一个问题, 使监督要求比标准更弱得多, 而不是精确性ALILLA 。 我们用一个适当的方法来评估。