In recent years, image and video delivery systems have begun integrating deep learning super-resolution (SR) approaches, leveraging their unprecedented visual enhancement capabilities while reducing reliance on networking conditions. Nevertheless, deploying these solutions on mobile devices still remains an active challenge as SR models are excessively demanding with respect to workload and memory footprint. Despite recent progress on on-device SR frameworks, existing systems either penalize visual quality, lead to excessive energy consumption or make inefficient use of the available resources. This work presents NAWQ-SR, a novel framework for the efficient on-device execution of SR models. Through a novel hybrid-precision quantization technique and a runtime neural image codec, NAWQ-SR exploits the multi-precision capabilities of modern mobile NPUs in order to minimize latency, while meeting user-specified quality constraints. Moreover, NAWQ-SR selectively adapts the arithmetic precision at run time to equip the SR DNN's layers with wider representational power, improving visual quality beyond what was previously possible on NPUs. Altogether, NAWQ-SR achieves an average speedup of 7.9x, 3x and 1.91x over the state-of-the-art on-device SR systems that use heterogeneous processors (MobiSR), CPU (SplitSR) and NPU (XLSR), respectively. Furthermore, NAWQ-SR delivers an average of 3.2x speedup and 0.39 dB higher PSNR over status-quo INT8 NPU designs, but most importantly mitigates the negative effects of quantization on visual quality, setting a new state-of-the-art in the attainable quality of NPU-based SR.
翻译:近年来,图像和视频传送系统已开始整合深层次学习超分辨率(SR)方法,利用其前所未有的视觉增强能力,同时减少对网络条件的依赖;然而,在移动设备上部署这些解决方案仍然是一项积极的挑战,因为移动设备模型在工作量和记忆足迹方面要求过高;尽管最近在安装视频SR框架方面取得了进展,但现有系统要么损害视觉质量,导致能源消耗过多,或者对现有资源的利用效率低下;这项工作是NAWQ-SR,这是高效执行甚高分辨率模型的新框架;通过新的混合精度量化技术和运行时间神经图像编码,NAWQ-SR利用现代移动设备多精度能力,以尽量减少延缩,同时满足用户指定的质量限制;此外,NAWQ-SR有选择地调整运行时的算精度精度,使SRDNNNNNNPN的层拥有更广泛的代表力量,提高视觉质量,超过以前可能的水平。