In recent years, a huge amount of deep neural architectures have been developed for image classification. It remains curious whether these models are similar or different and what factors contribute to their similarities or differences. To address this question, we aim to design a quantitative and scalable similarity function between neural architectures. We utilize adversarial attack transferability, which has information related to input gradients and decision boundaries that are widely used to understand model behaviors. We conduct a large-scale analysis on 69 state-of-the-art ImageNet classifiers using our proposed similarity function to answer the question. Moreover, we observe neural architecture-related phenomena using model similarity that model diversity can lead to better performance on model ensembles and knowledge distillation under specific conditions. Our results provide insights into why the development of diverse neural architectures with distinct components is necessary.
翻译:近年来,为图像分类开发了大量深层神经结构。 这些模型是否相似或不同,以及哪些因素导致它们的相似或差异,仍然令人好奇。 为了解决这一问题,我们的目标是设计神经结构之间数量和可扩展的相似功能。 我们使用对抗性攻击可转移性,它含有与输入梯度和决定界限有关的信息,被广泛用于理解模型行为。我们利用我们提议的相似功能对69个最先进的图像网络分类器进行大规模分析,以解答问题。此外,我们利用模型多样性能够导致在特定条件下更好地表现模型组合和知识蒸馏的模型相似性来观测与神经结构相关的现象。我们的结果为为什么有必要开发具有不同组成部分的多种神经结构提供了深刻的见解。</s>