Machine-learning systems such as self-driving cars or virtual assistants are composed of a large number of machine-learning models that recognize image content, transcribe speech, analyze natural language, infer preferences, rank options, etc. Models in these systems are often developed and trained independently, which raises an obvious concern: Can improving a machine-learning model make the overall system worse? We answer this question affirmatively by showing that improving a model can deteriorate the performance of downstream models, even after those downstream models are retrained. Such self-defeating improvements are the result of entanglement between the models in the system. We perform an error decomposition of systems with multiple machine-learning models, which sheds light on the types of errors that can lead to self-defeating improvements. We also present the results of experiments which show that self-defeating improvements emerge in a realistic stereo-based detection system for cars and pedestrians.
翻译:自驾车或虚拟助理等机学系统由大量机学模型组成,这些模型承认图像内容、录音、分析自然语言、推论偏好、排名选项等。 这些系统中的模型往往是独立开发和培训的,这引起一个明显的关注:改进机学模型能够使整个系统更加糟糕吗?我们肯定地回答这个问题,表明改进模型可能使下游模型的性能恶化,即使这些下游模型经过再培训后也是如此。这种自败改进是系统中模型相互纠缠的结果。我们用多种机学模型对系统进行错误分解,揭示了可能导致自败改进的错误类型。我们还介绍了实验结果,表明在现实的汽车和行人立体探测系统中出现了自我败坏的改进。