Video reconstruction from a single motion-blurred image is a challenging problem, which can enhance the capabilities of existing cameras. Recently, several works addressed this task using conventional imaging and deep learning. Yet, such purely-digital methods are inherently limited, due to direction ambiguity and noise sensitivity. Some works proposed to address these limitations using non-conventional image sensors, however, such sensors are extremely rare and expensive. To circumvent these limitations with simpler means, we propose a hybrid optical-digital method for video reconstruction that requires only simple modifications to existing optical systems. We use a learned dynamic phase-coding in the lens aperture during the image acquisition to encode the motion trajectories, which serve as prior information for the video reconstruction process. The proposed computational camera generates a sharp frame burst of the scene at various frame rates from a single coded motion-blurred image, using an image-to-video convolutional neural network. We present advantages and improved performance compared to existing methods, using both simulations and a real-world camera prototype. We extend our optical coding also to video frame interpolation and present robust and improved results for noisy videos.
翻译:从单一的移动模糊图像中重建视频是一个具有挑战性的问题,可以提高现有相机的能力。最近,一些作品利用传统成像和深层学习来应对这项任务。然而,由于方向模糊和噪音敏感性,这类纯数字方法本身是有限的。一些提议使用非常规图像传感器来解决这些局限性的工程非常罕见和昂贵。为了以更简单的方式绕过这些限制,我们提议了一种混合光学数字方法,用于视频重建,只需要对现有光学系统进行简单的修改。我们在获取图像时,使用在镜头孔中学习的动态分层编码来编码这些移动轨迹,作为视频重建过程的先前信息。拟议的计算相机利用一个图像到视频的动态神经网络,以各种框架速率生成一个锐利的画面。我们利用模拟和现实世界相机原型,提出与现有方法相比的优势和更好的性能。我们还将我们的光学编码扩展到视频框内插图,并展示振动视频的可靠和改进的结果。