Semantic segmentation has been a major topic in research and industry in recent years. However, due to the computation complexity of pixel-wise prediction and backpropagation algorithm, semantic segmentation has been demanding in computation resources, resulting in slow training and inference speed and large storage space to store models. Existing schemes that speed up segmentation network change the network structure and come with noticeable accuracy degradation. However, neural network quantization can be used to reduce computation load while maintaining comparable accuracy and original network structure. Semantic segmentation networks are different from traditional deep convolutional neural networks (DCNNs) in many ways, and this topic has not been thoroughly explored in existing works. In this paper, we propose a new quantization framework for training and inference of segmentation networks, where parameters and operations are constrained to 8-bit integer-based values for the first time. Full quantization of the data flow and the removal of square and root operations in batch normalization give our framework the ability to perform inference on fixed-point devices. Our proposed framework is evaluated on mainstream semantic segmentation networks like FCN-VGG16 and DeepLabv3-ResNet50, achieving comparable accuracy against floating-point framework on ADE20K dataset and PASCAL VOC 2012 dataset.
翻译:最近几年来,静默分解一直是研究和工业领域的一个主要议题,然而,由于对像素预测和反反光演算法的复杂性进行计算,静脉分解在计算资源方面一直要求很高,导致培训速度和推断速度缓慢,储存模型的储存空间大;现有的加速分解网络的现有办法改变了网络结构,并导致网络结构出现明显精确度下降;然而,神经网络分解可以用来减少计算负荷,同时保持可比的准确性和原始的网络结构;静脉分解网络在许多方面不同于传统的深层神经网络(DCNNN),而现有工作尚未彻底探讨这个专题;在本文件中,我们提出了一个新的分解网络培训和推断的定量框架,其中参数和操作首次受8位整数基值的限制;数据流动的全面四分解以及分解成批正常状态中的平方和根操作,使我们的框架有能力对固定点装置进行推断。我们提议的框架是在主流分解网络上,如FCN-CN-Reset数据准确度框架,2012-VAGA16 和GAGADL 等主流分级数据框架。