Lighter and faster models are crucial for the deployment of video super-resolution (VSR) on resource-limited devices, e.g., smartphones and wearable devices. In this paper, we develop Residual Sparsity Connection Learning (RSCL), a structured pruning scheme, to reduce the redundancy of convolution kernels and obtain a compact VSR network with a minor performance drop. However, residual blocks require the pruned filter indices of skip and residual connections to be the same, which is tricky for pruning. Thus, to mitigate the pruning restrictions of residual blocks, we design a Residual Sparsity Connection (RSC) scheme by preserving the feature channels and only operating on the important channels. Moreover, for the pixel-shuffle operation, we design a special pruning scheme by grouping several filters as pruning units to guarantee the accuracy of feature channel-space conversion after pruning. In addition, we introduce Temporal Finetuning (TF) to reduce the pruning error amplification of hidden states with temporal propagation. Extensive experiments show that the proposed RSCL significantly outperforms recent methods quantitatively and qualitatively. Codes and models will be released.
翻译:更亮和更快的模型对于在资源有限的装置(例如智能手机和可磨损装置)上部署超分辨率视频(VSR)至关重要。 在本文中,我们开发了残余分级连接学习(RSCL)方案,这是一个结构化的剪裁方案,以减少卷发内核的冗余,并获得一个小型性能下降的紧凑VSR网络。然而,残留区块要求螺旋过量和剩余连接的经处理过滤的过滤指数相同,这在剪裁时很困难。因此,为了减轻残余区块的剪裁限制,我们设计了一个残留分连接(RSC)方案,办法是通过维护特性频道,并只在重要频道上操作。此外,对于像素- Sheffle 操作,我们设计了一个特殊的剪裁方案,将几个过滤器分组成一个运行单位,以保证特性频道-空间转换的准确性能。此外,我们引入了温度微调调整(TFF),以减少隐藏状态的剪裁断误差,我们设计了时间传播。广泛的实验显示拟议的 RCL 将大大超出最近的质量和定量模型。