Event cameras such as DAVIS can simultaneously output high temporal resolution events and low frame-rate intensity images, which own great potential in capturing scene motion, such as optical flow estimation. Most of the existing optical flow estimation methods are based on two consecutive image frames and can only estimate discrete flow at a fixed time interval. Previous work has shown that continuous flow estimation can be achieved by changing the quantities or time intervals of events. However, they are difficult to estimate reliable dense flow , especially in the regions without any triggered events. In this paper, we propose a novel deep learning-based dense and continuous optical flow estimation framework from a single image with event streams, which facilitates the accurate perception of high-speed motion. Specifically, we first propose an event-image fusion and correlation module to effectively exploit the internal motion from two different modalities of data. Then we propose an iterative update network structure with bidirectional training for optical flow prediction. Therefore, our model can estimate reliable dense flow as two-frame-based methods, as well as estimate temporal continuous flow as event-based methods. Extensive experimental results on both synthetic and real captured datasets demonstrate that our model outperforms existing event-based state-of-the-art methods and our designed baselines for accurate dense and continuous optical flow estimation.
翻译:DAVIS等事件摄像机可以同时输出高时间分辨率事件和低框架强度图像,这些图像具有捕捉现场运动的巨大潜力,例如光流估计。现有的光学流估计方法大多以两个连续的图像框架为基础,只能在固定的时间间隔内估计离流。先前的工作表明,通过改变事件的数量或时间间隔可以实现连续流量估计。然而,它们很难估计可靠的密集流量,特别是在没有发生任何事件的区域。在本文件中,我们建议从一个带有事件流的图像中,建立一个新的深层次的基于学习的密集和连续的光流估计框架,这个框架有利于对高速运动的准确认识。具体地说,我们首先提出一个事件模拟聚合和关联模块,以便从两种不同的数据模式中有效地利用内部运动。然后,我们提出一个带有光流预测双向培训的迭代更新网络结构。因此,我们的模型可以估计可靠的密集流量,作为基于两个框架的方法,以及作为基于事件的方法估计的时间持续流。在合成和真实的采集的数据集上,广泛的实验结果表明,我们的模型已经超越了我们所设计的精确的动态基线和连续的模型。