Deep learning-based super-resolution (SR) has gained tremendous popularity in recent years because of its high image quality performance and wide application scenarios. However, prior methods typically suffer from large amounts of computations and huge power consumption, causing difficulties for real-time inference, especially on resource-limited platforms such as mobile devices. To mitigate this, we propose a compiler-aware SR neural architecture search (NAS) framework that conducts depth search and per-layer width search with adaptive SR blocks. The inference speed is directly taken into the optimization along with the SR loss to derive SR models with high image quality while satisfying the real-time inference requirement. Instead of measuring the speed on mobile devices at each iteration during the search process, a speed model incorporated with compiler optimizations is leveraged to predict the inference latency of the SR block with various width configurations for faster convergence. With the proposed framework, we achieve real-time SR inference for implementing 720p resolution with competitive SR performance (in terms of PSNR and SSIM) on GPU/DSP of mobile platforms (Samsung Galaxy S21).
翻译:近些年来,深层学习超分辨率(SR)因其高图像质量性能和广泛应用情景而获得极大支持。然而,以往方法通常会受到大量计算和大量电耗的影响,造成实时推断困难,特别是在诸如移动设备等资源有限的平台上。为此,我们提议建立一个编译-感知SR神经结构搜索(NAS)框架,进行深度搜索和与适应性SR区块进行每层宽度搜索。推断速度直接与SR损失一起优化,在满足实时推断要求的同时,得出高图像性能模型。在搜索过程中,使用一个与编译器优化相结合的速度模型,而不是在每次迭代中测量移动设备的速度,而是用来预测SR区块与各种宽度配置的推断,以便更快地趋同。我们利用拟议框架,在移动平台的GPU/DSP(Samsung S21)上,通过具有竞争力的SR性能(PSNR和SSIM)执行720p分辨率的实时SR推论。我们用实时推论得出了实时推论。在移动平台的GPU/DSP(Sung S.21)。