When upgrading neural models to a newer version, new errors that were not encountered in the legacy version can be introduced, known as regression errors. This inconsistent behavior during model upgrade often outweighs the benefits of accuracy gain and hinders the adoption of new models. To mitigate regression errors from model upgrade, distillation and ensemble have proven to be viable solutions without significant compromise in performance. Despite the progress, these approaches attained an incremental reduction in regression which is still far from achieving backward-compatible model upgrade. In this work, we propose a novel method, Gated Fusion, that promotes backward compatibility via learning to mix predictions between old and new models. Empirical results on two distinct model upgrade scenarios show that our method reduces the number of regression errors by 62% on average, outperforming the strongest baseline by an average of 25%.
翻译:当将神经模型升级为较新版本时,可以引入遗留版本中未遇到的新错误,称为回归错误。模型升级期间的这种不一致行为往往大于准确性收益的好处,并阻碍采用新模型。为了从模型升级中减少回归错误,蒸馏和组合已证明是可行的解决办法,而没有显著的绩效妥协。尽管取得了这些进展,但这些方法还是逐渐地减少了倒退,这远远没有实现与后向兼容的模式升级。在这项工作中,我们提出了一个新颖的方法,即Gated Fusion,通过学习将旧模型和新模型的预测混在一起,促进后向兼容性。 两种不同的模型升级情景的经验显示,我们的方法平均将回归错误减少62%,比最强的基线平均减少25%。