Ensemble learning has gain attention in resent deep learning research as a way to further boost the accuracy and generalizability of deep neural network (DNN) models. Recent ensemble training method explores different training algorithms or settings on multiple sub-models with the same model architecture, which lead to significant burden on memory and computation cost of the ensemble model. Meanwhile, the heurtsically induced diversity may not lead to significant performance gain. We propose a new prespective on exploring the intrinsic diversity within a model architecture to build efficient DNN ensemble. We make an intriguing observation that pruning and quantization, while both leading to efficient model architecture at the cost of small accuracy drop, leads to distinct behavior in the decision boundary. To this end, we propose Heterogeneously Compressed Ensemble (HCE), where we build an efficient ensemble with the pruned and quantized variants from a pretrained DNN model. An diversity-aware training objective is proposed to further boost the performance of the HCE ensemble. Experiemnt result shows that HCE achieves significant improvement in the efficiency-accuracy tradeoff comparing to both traditional DNN ensemble training methods and previous model compression methods.
翻译:强化的学习已经引起人们的注意,因为人们痛恨深深层的学习研究,认为这是进一步提高深神经网络(DNN)模型的准确性和可概括性的一种方法。最近的混合培训方法探索了具有相同模型结构的多个子模型的不同培训算法或设置,这给共同模型的记忆和计算成本带来沉重负担。与此同时,由杂乱引起的多样性可能不会带来显著的绩效收益。我们提议了一种新的尊重,以探索模型结构内在的多样性,以建立高效的DNNN联合体。我们提出了探索内在多样性的新建议,以建立高效的DNNN(DNN)联合体。我们做了一个令人感兴趣的观察,即精密和定量化,同时以小精度下降为代价导致高效的模型结构,从而导致决定界限上的不同行为。为此,我们提议了高度折叠合的集合模型(HCEE),我们在那里建立一个高效的组合组合组合,与未经训练过的DNNNM模型中精细的变体。我们提议了一个多样性培训目标,以进一步提升HCE 元组合体的性运行和量化的传统贸易结果。