Spherical videos, also known as \ang{360} (panorama) videos, can be viewed with various virtual reality devices such as computers and head-mounted displays. They attract large amount of interest since awesome immersion can be experienced when watching spherical videos. However, capturing, storing and transmitting high-resolution spherical videos are extremely expensive. In this paper, we propose a novel single frame and multi-frame joint network (SMFN) for recovering high-resolution spherical videos from low-resolution inputs. To take advantage of pixel-level inter-frame consistency, deformable convolutions are used to eliminate the motion difference between feature maps of the target frame and its neighboring frames. A mixed attention mechanism is devised to enhance the feature representation capability. The dual learning strategy is exerted to constrain the space of solution so that a better solution can be found. A novel loss function based on the weighted mean square error is proposed to emphasize on the super-resolution of the equatorial regions. This is the first attempt to settle the super-resolution of spherical videos, and we collect a novel dataset from the Internet, MiG Panorama Video, which includes 204 videos. Experimental results on 4 representative video clips demonstrate the efficacy of the proposed method. The dataset and code are available at https://github.com/lovepiano/SMFN_For_360VSR.
翻译:球形视频,也称为 ang{360} (panorama) (panorama) 视频,可以用各种虚拟现实设备查看,如计算机和头顶显示器等。 它们吸引了许多人的兴趣, 因为观看球形视频时可以体验到惊人的沉浸状态。 但是, 捕捉、 储存和传输高分辨率球形视频非常昂贵。 在本文中, 我们提议建立一个新颖的单一框架和多框架联合网络( SMFN), 用于从低分辨率输入中恢复高分辨率球形视频。 为了利用像素级别的跨框架一致性, 使用可变的调动来消除目标框架特征图及其相邻框架之间的运动差异。 设计了一个混合关注机制, 以加强功能代表能力。 双重学习策略是为了限制解决方案的空间, 以便找到更好的解决方案。 提议了一个基于加权中方错误的新式损失功能, 以强调赤道区域的超级分辨率。 这是第一次尝试解决超分辨率视频的超级分辨率, 并且我们在互联网上收集新版数据集的目标框图和相框的相图图集, 将MGPanama/ prealmama 演示结果 。