Research in the field of Continual Semantic Segmentation is mainly investigating novel learning algorithms to overcome catastrophic forgetting of neural networks. Most recent publications have focused on improving learning algorithms without distinguishing effects caused by the choice of neural architecture.Therefore, we study how the choice of neural network architecture affects catastrophic forgetting in class- and domain-incremental semantic segmentation. Specifically, we compare the well-researched CNNs to recently proposed Transformers and Hybrid architectures, as well as the impact of the choice of novel normalization layers and different decoder heads. We find that traditional CNNs like ResNet have high plasticity but low stability, while transformer architectures are much more stable. When the inductive biases of CNN architectures are combined with transformers in hybrid architectures, it leads to higher plasticity and stability. The stability of these models can be explained by their ability to learn general features that are robust against distribution shifts. Experiments with different normalization layers show that Continual Normalization achieves the best trade-off in terms of adaptability and stability of the model. In the class-incremental setting, the choice of the normalization layer has much less impact. Our experiments suggest that the right choice of architecture can significantly reduce forgetting even with naive fine-tuning and confirm that for real-world applications, the architecture is an important factor in designing a continual learning model.
翻译:持续语义分割领域的研究主要是研究新颖的学习算法,以克服灾难性地遗忘神经网络的问题。大多数最近的出版物都侧重于改进学习算法,而没有因选择神经结构而产生不同的影响。因此,我们研究神经网络结构的选择如何影响在阶级和地域中灾难性地遗忘在阶级和地表偏移的语义分割领域。具体地说,我们把经过研究的CNN与最近提议的变异器和混合结构进行比较,以及选择新颖的正常化层和不同的解密头的影响。我们发现,像ResNet这样的传统CNN具有高度的可塑性,但稳定性较低,而变异结构则更加稳定。当CNN结构的感性偏向性偏向性与混合结构的变异性结构相结合时,这种神经网络结构的稳定性会提高。这些模型的稳定性可以用它们学习与分布变化相适应的一般特征的能力来解释。与不同模式的正常化层实验表明,在模型的适应性和稳定性方面实现最佳贸易。在类级结构设计中,对正统性结构的选择会大大降低我们正统性结构的影响。