State-of-the-art deep learning methods achieve human-like performance on many tasks, but make errors nevertheless. Characterizing these errors in easily interpretable terms gives insight into whether a model is prone to making systematic errors, but also gives a way to act and improve the model. In this paper we propose a method that allows us to do so for arbitrary classifiers by mining a small set of patterns that together succinctly describe the input data that is partitioned according to correctness of prediction. We show this is an instance of the more general label description problem, which we formulate in terms of the Minimum Description Length principle. To discover good pattern sets we propose the efficient and hyperparameter-free Premise algorithm, which through an extensive set of experiments we show on both synthetic and real-world data performs very well in practice; unlike existing solutions it ably recovers ground truth patterns, even on highly imbalanced data over many unique items, or where patterns are only weakly associated to labels. Through two real-world case studies we confirm that Premise gives clear and actionable insight into the systematic errors made by modern NLP classifiers.
翻译:最先进的深层次学习方法在许多任务上取得了人性化的成绩,但还是犯了错误。用容易解释的术语来描述这些错误,可以洞察出一个模型是否容易造成系统性错误,但也可以采取某种行动和改进模型。在本文中,我们提出了一个方法,允许任意分类者这样做,通过挖掘一小套模式,简洁地描述根据预测的正确性而分割的输入数据。我们展示了这个更笼统的标签描述问题的例子,我们用最低描述长度原则来描述。为了发现良好的模式设置,我们提出了高效和无超参数的预言算法,我们通过在合成数据和现实世界数据上展示的一套广泛的实验,在实践中表现得非常好;与现有的解决方案不同,它可以恢复地面的真相模式,即使是在许多独特项目上高度不平衡的数据,或者那些模式与标签关系不大。我们通过两个真实世界的案例研究证实,Premise提供了对现代NLP分类人员系统错误的清晰和可操作的洞察力。