Real-time video deblurring still remains a challenging task due to the complexity of spatially and temporally varying blur itself and the requirement of low computational cost. To improve the network efficiency, we adopt residual dense blocks into RNN cells, so as to efficiently extract the spatial features of the current frame. Furthermore, a global spatio-temporal attention module is proposed to fuse the effective hierarchical features from past and future frames to help better deblur the current frame. Another issue needs to be addressed urgently is the lack of a real-world benchmark dataset. Thus, we contribute a novel dataset (BSD) to the community, by collecting paired blurry/sharp video clips using a co-axis beam splitter acquisition system. Experimental results show that the proposed method (ESTRNN) can achieve better deblurring performance both quantitatively and qualitatively with less computational cost against state-of-the-art video deblurring methods. In addition, cross-validation experiments between datasets illustrate the high generality of BSD over the synthetic datasets. The code and dataset are released at https://github.com/zzh-tech/ESTRNN.
翻译:由于空间和时间的模糊性复杂,而且计算成本低,因此实时视频破碎仍然是一项艰巨的任务。为了提高网络效率,我们将残留的稠密区块引入RNN的细胞,以便有效地提取当前框架的空间特征。此外,还提议了一个全球时空关注模块,以整合过去和将来框架的有效等级特征,帮助更好地破碎当前框架。另一个需要紧急解决的问题是缺乏真实世界的基准数据集。因此,我们通过使用共同轴波断裂器获取系统收集配对的模糊/沙普视频剪片,为社区贡献了一个新的数据集。实验结果表明,拟议的方法(ERTNNN)可以在数量和质量上实现更好的分解性能,用较低的计算成本来配合当前框架的当前框架。此外,各数据集之间的交叉校验实验表明了BSD在合成数据集方面的高通用性。代码和数据元集在 https://giuth/strechetal上发布。