Deep convolutional neural networks (DCNNs) have shown dominant performance in the task of super-resolution (SR). However, their heavy memory cost and computation overhead significantly restrict their practical deployments on resource-limited devices, which mainly arise from the floating-point storage and operations between weights and activations. Although previous endeavors mainly resort to fixed-point operations, quantizing both weights and activations with fixed coding lengths may cause significant performance drop, especially on low bits. Specifically, most state-of-the-art SR models without batch normalization have a large dynamic quantization range, which also serves as another cause of performance drop. To address these two issues, we propose a new quantization scheme termed PArameterized Max Scale (PAMS), which applies the trainable truncated parameter to explore the upper bound of the quantization range adaptively. Finally, a structured knowledge transfer (SKT) loss is introduced to fine-tune the quantized network. Extensive experiments demonstrate that the proposed PAMS scheme can well compress and accelerate the existing SR models such as EDSR and RDN. Notably, 8-bit PAMS-EDSR improves PSNR on Set5 benchmark from 32.095dB to 32.124dB with 2.42$\times$ compression ratio, which achieves a new state-of-the-art.
翻译:深相神经网络(DCNNS)在超分辨率(SR)任务中表现突出。然而,它们的大量记忆成本和计算间接费用大大限制了其在资源有限的装置上的实际部署,这主要来自在重量和激活之间浮动点存储和操作。虽然以前的努力主要依靠固定点操作,但将加权数和固定编码长度的激活量进行量化,可能会导致显著的性能下降,特别是在低位上。具体地说,大多数没有批量正常化的先进SR模型具有很大的动态量化范围,这也成为另一个性能下降的原因。为了解决这两个问题,我们提出了一个新的量化计划,称为“Parammeterizized Max 比例(PAMS) ” (PAMMS ), 将可训练的细化参数应用到调整的四分数范围上限上。最后,引入了结构化知识转移(SKT) 损失来微调四分化网络。新的实验表明,拟议的PMS 计划可以很好地压缩和加速现有的SR模型,如EDSR和RDM5 的P-RM-RMS-RMS-S-S-S-S-S-S-S-S-S-S-S-S-Sirg-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-T-S-S-S-S-S-S-S-S-S-S-S-S-S-T-T-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-