Negative flips are errors introduced in a classification system when a legacy model is replaced with a new one. Existing methods to reduce the negative flip rate (NFR) either do so at the expense of overall accuracy using model distillation, or use ensembles, which multiply inference cost prohibitively. We present a method to train a classification system that achieves paragon performance in both error rate and NFR, at the inference cost of a single model. Our method introduces a generalized distillation objective, Logit Difference Inhibition (LDI), that penalizes changes in the logits between the new and old model, without forcing them to coincide as in ordinary distillation. LDI affords the model flexibility to reduce error rate along with NFR. The method uses a homogeneous ensemble as the reference model for LDI, hence the name Ensemble LDI, or ELODI. The reference model can then be substituted with a single model at inference time. The method leverages the observation that negative flips are typically not close to the decision boundary, but often exhibit large deviations in the distance among their logits, which are reduced by ELODI.
翻译:负翻转是指在以新的模式取代遗留模式时,在分类系统中引入错误。现有的降低负翻转率(NFR)的方法要么以使用模型蒸馏的总体准确性为代价,要么以使用模型蒸馏的总体准确性为代价来降低负翻转率(NFR),或者使用模型组合,这种组合将极高的推论成本乘以极高的推论成本。我们提出了一个方法,以单一模型的推论成本来培训一个在错误率和NFR中都达到参数性能的分类系统。我们的方法引入了一个普遍的蒸馏目标,即Logit差异触发(LDI),它惩罚新式和旧模型之间对日志的修改,而不会迫使它们与普通的蒸馏同步。LDI提供了模型的灵活性,以降低误差率和NFR。该方法使用同质共性共性共性词作为LDI的参考模型,因此名称为Esemble LDI,或ELODI。然后在推论时可以用一个单一模型取代。这种方法利用这一观察,即负翻通常不接近决定边界,但往往显示其日志之间的偏差很大。