The video super-resolution (VSR) method based on the recurrent convolutional network has strong temporal modeling capability for video sequences. However, the temporal receptive field of different recurrent units in the unidirectional recurrent network is unbalanced. Earlier reconstruction frames receive less spatio-temporal information, resulting in fuzziness or artifacts. Although the bidirectional recurrent network can alleviate this problem, it requires more memory space and fails to perform many tasks with low latency requirements. To solve the above problems, we propose an end-to-end information prebuilt recurrent reconstruction network (IPRRN), consisting of an information prebuilt network (IPNet) and a recurrent reconstruction network (RRNet). By integrating sufficient information from the front of the video to build the hidden state needed for the initially recurrent unit to help restore the earlier frames, the information prebuilt network balances the input information difference at different time steps. In addition, we demonstrate an efficient recurrent reconstruction network, which outperforms the existing unidirectional recurrent schemes in all aspects. Many experiments have verified the effectiveness of the network we propose, which can effectively achieve better quantitative and qualitative evaluation performance compared to the existing state-of-the-art methods.
翻译:以循环变迁网络为基础的超分辨率(VSR)视频方法具有强大的视频序列时间模型能力。然而,单向经常性网络中不同经常性单位的瞬间可接受空间是不平衡的。早期重建框架接收的信息较少时空信息,导致模糊或人工制品。虽然双向经常性网络可以缓解这一问题,但它需要更多的记忆空间,并且无法执行许多低延迟要求的任务。为了解决上述问题,我们提议建立一个端对端信息预建的经常性重建网络(IPRRN),由信息预建网络(IPNet)和经常性重建网络(RRNet)组成。通过整合视频前方的充分信息,建立最初的经常性单位所需的隐藏状态,帮助恢复早期框架,信息预建网络可以平衡不同时间步骤的输入信息差异。此外,我们展示了一个高效的经常性重建网络,它超越了现有的单向经常性计划的各个方面。许多实验已经核实了我们提议的网络的有效性,可以有效地实现质量和定量评估方法,与现有状态相比较。