Real-world machine learning systems are achieving remarkable performance in terms of coarse-grained metrics like overall accuracy and F-1 score. However, model improvement and development often require fine-grained modeling on individual data subsets or slices, for instance, the data slices where the models have unsatisfactory results. In practice, it gives tangible values for developing such models that can pay extra attention to critical or interested slices while retaining the original overall performance. This work extends the recent slice-based learning (SBL)~\cite{chen2019slice} with a mixture of attentions (MoA) to learn slice-aware dual attentive representations. We empirically show that the MoA approach outperforms the baseline method as well as the original SBL approach on monitored slices with two natural language understanding (NLU) tasks.
翻译:现实世界的机器学习系统在整体精确度和F-1分等粗劣的计量标准方面正在取得显著的成绩,然而,模型改进和发展往往要求对单个数据子集或切片进行细微的模型分析,例如,模型结果不尽人意的数据切片。在实践中,它为开发这些模型提供了有形价值,这些模型可以对关键或感兴趣的切片给予额外关注,同时保留原有的总体绩效。这项工作扩大了最近以切片为基础的学习(SBL),并结合了多种关注(MoA)来学习切片觉双重关注的表达方式。我们从经验上表明,MoA的方法超越了基线方法,以及用两种自然语言理解任务监测切片的原始SBL方法。