Despite strong empirical performance for image classification, deep neural networks are often regarded as ``black boxes'' and they are difficult to interpret. On the other hand, sparse convolutional models, which assume that a signal can be expressed by a linear combination of a few elements from a convolutional dictionary, are powerful tools for analyzing natural images with good theoretical interpretability and biological plausibility. However, such principled models have not demonstrated competitive performance when compared with empirically designed deep networks. This paper revisits the sparse convolutional modeling for image classification and bridges the gap between good empirical performance (of deep learning) and good interpretability (of sparse convolutional models). Our method uses differentiable optimization layers that are defined from convolutional sparse coding as drop-in replacements of standard convolutional layers in conventional deep neural networks. We show that such models have equally strong empirical performance on CIFAR-10, CIFAR-100, and ImageNet datasets when compared to conventional neural networks. By leveraging stable recovery property of sparse modeling, we further show that such models can be much more robust to input corruptions as well as adversarial perturbations in testing through a simple proper trade-off between sparse regularization and data reconstruction terms. Source code can be found at https://github.com/Delay-Xili/SDNet.
翻译:尽管在图像分类方面有很强的实证表现,但深神经网络往往被视为“黑盒子”和难以解释。另一方面,稀有的进化模型(这些模型假定可以通过由从进化字典中产生的几个要素组成的线性组合表达信号)是分析自然图像的有力工具,具有良好的理论解释性和生物上的合理性。然而,这些原则模型与从经验上设计的深层次网络相比,并没有表现出有竞争力的绩效。本文回顾了在图像分类方面稀少的演动模型,并弥合了良好经验性(深层次学习)和良好解释(稀有革命模型)之间的差距。我们的方法使用不同的优化层,从进化的稀散编码中定义为常规深层神经网络中标准革命层的倒置替换。我们表明,这些模型在CIFAR-10、CIFAR-100和图像网络数据集与传统神经网络网络相比,同样具有很强的经验性能。通过利用稀释性模型的稳定恢复特性,我们进一步表明,这些模型对于输入腐败可以更加强大得多,作为动态的稀释性代码,可以作为Siraltrabildal-comstantrading rudeal-trading restractionalationalationalview