In the computer vision community, Convolutional Neural Networks (CNNs), first proposed in the 1980's, have become the standard visual classification model. Recently, as alternatives to CNNs, Capsule Networks (CapsNets) and Vision Transformers (ViTs) have been proposed. CapsNets, which were inspired by the information processing of the human brain, are considered to have more inductive bias than CNNs, whereas ViTs are considered to have less inductive bias than CNNs. All three classification models have received great attention since they can serve as backbones for various downstream tasks. However, these models are far from being perfect. As pointed out by the community, there are two weaknesses in standard Deep Neural Networks (DNNs). One of the limitations of DNNs is the lack of explainability. Even though they can achieve or surpass human expert performance in the image classification task, the DNN-based decisions are difficult to understand. In many real-world applications, however, individual decisions need to be explained. The other limitation of DNNs is adversarial vulnerability. Concretely, the small and imperceptible perturbations of inputs can mislead DNNs. The vulnerability of deep neural networks poses challenges to current visual classification models. The potential threats thereof can lead to unacceptable consequences. Besides, studying model adversarial vulnerability can lead to a better understanding of the underlying models. Our research aims to address the two limitations of DNNs. Specifically, we focus on deep visual classification models, especially the core building parts of each classification model, e.g. dynamic routing in CapsNets and self-attention module in ViTs.
翻译:在计算机视觉界,1980年代首次提出的革命神经网络(CNN)已成为标准的视觉分类模式。最近,作为CNN、Capsule网络(CapsNets)和愿景变异器(VNS)的替代方案,提出了新的建议。CapsNet(受人类大脑信息处理的启发)被认为比CNN更具有感人偏见,而VNT(VNT)被认为比CNN更难以理解。所有三个分类模式都受到极大关注,因为它们可以成为各种下游任务的骨干。然而,这些模式远非完美无缺。正如社区指出的,标准深神经网络(CapsNets)和愿景变异器(VNNT)存在两个弱点。DNNNet的一个局限性是缺乏解释性。即使它们能够达到或超过人类专家在图像分类任务中的表现,但基于DNNNW的决定却难以理解。但是在许多现实世界应用中,个人决定需要加以解释。DNNF的另一项限制是建立对抗性脆弱性模型的弱点模型。具体和视觉变异性模型的每个弱点都能够更精确地解释。我们研究其内在的内在的内在的内在的内变现变变变变的弱点。