With the rising popularity of intelligent mobile devices, it is of great practical significance to develop accurate, realtime and energy-efficient image Super-Resolution (SR) inference methods. A prevailing method for improving the inference efficiency is model quantization, which allows for replacing the expensive floating-point operations with efficient fixed-point or bitwise arithmetic. To date, it is still challenging for quantized SR frameworks to deliver feasible accuracy-efficiency trade-off. Here, we propose a Fully Quantized image Super-Resolution framework (FQSR) to jointly optimize efficiency and accuracy. In particular, we target on obtaining end-to-end quantized models for all layers, especially including skip connections, which was rarely addressed in the literature. We further identify training obstacles faced by low-bit SR networks and propose two novel methods accordingly. The two difficulites are caused by 1) activation and weight distributions being vastly distinctive in different layers; 2) the inaccurate approximation of the quantization. We apply our quantization scheme on multiple mainstream super-resolution architectures, including SRResNet, SRGAN and EDSR. Experimental results show that our FQSR using low bits quantization can achieve on par performance compared with the full-precision counterparts on five benchmark datasets and surpass state-of-the-art quantized SR methods with significantly reduced computational cost and memory consumption.
翻译:随着智能移动设备越来越受欢迎,开发准确、实时和节能图像超分辨率(SR)的推断方法具有重大的实际意义。提高推论效率的常用方法是模型量化,这样可以以高效固定点或微误算取代昂贵的浮点作业。迄今为止,对于量化的SR框架来说,提供可行的准确性-效率权衡交易仍然具有挑战性。在这里,我们提议一个充分量化的图像超分辨率框架(FQSR),以共同优化效率和准确性。特别是,我们的目标是获得所有层次的端到端的定量模型,特别是空连接,这在文献中很少提及。我们进一步找出低位SR网络面临的培训障碍,并据此提出两种新颖的方法。造成两种差异的原因是:(1) 在不同层次的启动和加权分布非常独特;(2) 夸度的准确近近近度。我们将我们的量化计划应用于多种主流超级分辨率架构,包括SRNet、SRIAN和QEDSR的升级模型。我们用低比值的存储率和低位存储率 实验结果可以显著地显示我们州-QSR的精确度基准业绩。