In this paper, we describe a method to capture nearly entirely spherical (360 degree) depth information using two adjacent frames from a single spherical video with motion parallax. After illustrating a spherical depth information retrieval using two spherical cameras, we demonstrate monocular spherical stereo by using stabilized first-person video footage. Experiments demonstrated that the depth information was retrieved on up to 97% of the entire sphere in solid angle. At a speed of 30 km/h, we were able to estimate the depth of an object located over 30 m from the camera. We also reconstructed the 3D structures (point cloud) using the obtained depth data and confirmed the structures can be clearly observed. We can apply this method to 3D structure retrieval of surrounding environments such as 1) previsualization, location hunting/planning of a film, 2) real scene/computer graphics synthesis and 3) motion capture. Thanks to its simplicity, this method can be applied to various videos. As there is no pre-condition other than to be a 360 video with motion parallax, we can use any 360 videos including those on the Internet to reconstruct the surrounding environments. The cameras can be lightweight enough to be mounted on a drone. We also demonstrated such applications.
翻译:在本文中,我们描述一种方法,从一个球形视频中用运动抛光器用两个相邻框架来捕捉几乎完全球形(360度)深度信息。在用两个球形摄像头来说明球形深度信息检索后,我们用稳定的第一人视频镜头演示单球球形立体立体声。实验显示,深度信息以固态角度从整个球体的97%处检索到。以30公里/小时的速度,我们得以从相机中估计出一个位于30米以上的物体的深度。我们还利用所获得的深度数据重建了3D结构(点云),并确认了结构可以明显观察。我们可以将这种方法应用于周围环境的三维结构检索,例如:1)预视化、定位搜索/规划电影、2)真实场/计算机图形合成和3动作捕获。由于简单,这种方法可以应用到各种视频。除了使用360米的视频外,我们没有其他先决条件。我们可以使用任何360个视频,包括互联网上的那些视频来重建周围环境。我们还可以将这些摄像头安装得足够轻的无人机。