Modern single image super-resolution (SISR) system based on convolutional neural networks (CNNs) achieves fancy performance while requires huge computational costs. The problem on feature redundancy is well studied in visual recognition task, but rarely discussed in SISR. Based on the observation that many features in SISR models are also similar to each other, we propose to use shift operation to generate the redundant features (i.e., ghost features). Compared with depth-wise convolution which is time-consuming on GPU-like devices, shift operation can bring a practical inference acceleration for CNNs on common hardwares. We analyze the benefits of shift operation on SISR task and make the shift orientation learnable based on Gumbel-Softmax trick. Besides, a clustering procedure is explored based on pre-trained models to identify the intrinsic filters for generating intrinsic features. The ghost features will be derived by moving these intrinsic features along a specific orientation. Finally, the complete output features are constructed by concatenating the intrinsic and ghost features together. Extensive experiments on several benchmark models and datasets demonstrate that both the non-compact and lightweight SISR models embedded with the proposed method can achieve a comparable performance to that of their baselines with a large reduction of parameters, FLOPs and GPU inference latency. For instance, we reduce the parameters by 46%, FLOPs by 46% and GPU inference latency by 42% of $\times2$ EDSR network with basically lossless performance.
翻译:以进化神经网络(CNNs)为基础的现代单一图像超分辨率系统(SISR)在需要大量计算成本的情况下,可以取得高性能,而需要大量计算成本。关于功能冗余的问题在视觉识别任务中研究得周密,但在SISR中很少讨论。基于SISSR模型中的许多特征彼此相似的观察,我们建议使用转换操作来产生冗余特征(即幽灵特征)。与在GPU类似设备上耗时的深度变动相比,转换操作可以为CNN在通用硬件上带来实际的加速。我们分析了SISSR任务上变换操作的好处,并使基于 Gumbel-Softmax 技巧的变向方向可以学习。此外,根据预先培训的模型来探索一个集成程序,以确定生成内在特征的内在过滤器(即幽灵特征)。通过将这些内在特征与特定方向相推导出。最后,完整的输出功能特征是通过将46美元内在和幽灵特性相融合,对一些基准模型和数据设置进行广泛的实验,我们通过降低性能和光性能基准参数来降低其基础性能模型,从而降低性能基准基准基准基准和光值的SISL模型,从而降低其基本的性能模型和光值基准值。