Deep neural networks (DNNs) and, in particular, convolutional neural networks (CNNs) have brought significant advances in a wide range of modern computer application problems. However, the increasing availability of large amounts of datasets as well as the increasing available computational power of modern computers lead to a steady growth in the complexity and size of DNN and CNN models, and thus, to longer training times. Hence, various methods and attempts have been developed to accelerate and parallelize the training of complex network architectures. In this work, a novel CNN-DNN architecture is proposed that naturally supports a model parallel training strategy and that is loosely inspired by two-level domain decomposition methods (DDM). First, local CNN models, that is, subnetworks, are defined that operate on overlapping or nonoverlapping parts of the input data, for example, sub-images. The subnetworks can be trained completely in parallel. Each subnetwork outputs a local decision for the given machine learning problem which is exclusively based on the respective local input data. Subsequently, an additional DNN model is trained which evaluates the local decisions of the local subnetworks and generates a final, global decision. With respect to the analogy to DDM, the DNN can be interpreted as a coarse problem and hence, the new approach can be interpreted as a two-level domain decomposition. In this paper, solely image classification problems using CNNs are considered. Experimental results for different 2D image classification problems are provided as well as a face recognition problem, and a classification problem for 3D computer tomography (CT) scans. The results show that the proposed approach can significantly accelerate the required training time compared to the global model and, additionally, can also help to improve the accuracy of the underlying classification problem.
翻译:深神经网络(DNNs),特别是神经神经网络(CNNNs)的深度神经网络(CNN-DNNs)在广泛的现代计算机应用问题中取得了显著进步。然而,由于大量数据集的可用性不断增加以及现代计算机的计算能力不断增加,使得DNN和CNN模式的复杂性和规模稳步增加,从而导致培训时间延长。因此,已经开发了各种方法和尝试来加速和平行复杂网络结构的培训。在此工作中,提出了一个新的CNN-DNNNT结构,它自然地支持一个模型平行培训战略,并且受到两级域域分解方法(DDM)的启发。首先,本地CNNNM模式,即子网络,其定义是运行输入数据中的重叠部分或非重叠,例如子图像。子网络可以完全平行地培训。每个子网络输出给机器学习问题的本地决定,它完全基于各自的输入数据。随后,另一个DNNNM模型被培训用来评估当地面图像的加速度, 其深度的深度的图像结果也可以被解读为两个域域域域网络和最终的图像。在解释, 将DM的图像上,可以解释为一种新的图像, 。在解释, 将它作为两个类中,可以被解释为一个新的的,可以解释到一个新的的,可以用来解释到一个新的文件,可以用来解释到一个新的的,可以解释到一个新的文件,可以用来解释到一个新的文件,可以解释到一个新的的。