Scene Dynamic Recovery (SDR) by inverting distorted Rolling Shutter (RS) images to an undistorted high frame-rate Global Shutter (GS) video is a severely ill-posed problem, particularly when prior knowledge about camera/object motions is unavailable. Commonly used artificial assumptions on motion linearity and data-specific characteristics, regarding the temporal dynamics information embedded in the RS scanlines, are prone to producing sub-optimal solutions in real-world scenarios. To address this challenge, we propose an event-based RS2GS framework within a self-supervised learning paradigm that leverages the extremely high temporal resolution of event cameras to provide accurate inter/intra-frame information. % In this paper, we propose to leverage the event camera to provide inter/intra-frame information as the emitted events have an extremely high temporal resolution and learn an event-based RS2GS network within a self-supervised learning framework, where real-world events and RS images can be exploited to alleviate the performance degradation caused by the domain gap between the synthesized and real data. Specifically, an Event-based Inter/intra-frame Compensator (E-IC) is proposed to predict the per-pixel dynamic between arbitrary time intervals, including the temporal transition and spatial translation. Exploring connections in terms of RS-RS, RS-GS, and GS-RS, we explicitly formulate mutual constraints with the proposed E-IC, resulting in supervisions without ground-truth GS images. Extensive evaluations over synthetic and real datasets demonstrate that the proposed method achieves state-of-the-art and shows remarkable performance for event-based RS2GS inversion in real-world scenarios. The dataset and code are available at https://w3un.github.io/selfunroll/.
翻译:采用逆推扭曲的滚动快门(RS)图像到未扭曲的高帧率全局快门(GS)视频,实现场景动态恢复(SDR)是一个严重病态问题,尤其是在缺乏有关摄像机/物体运动的先验知识的情况下。在现实世界的情况下,对于运动线性性和数据特异性的使用人工假设,拟合RS扫描线所嵌入的时间动态信息容易产生次优解。为解决这一挑战,我们提出了一个事件驱动的自我监督RS2GS框架,利用事件相机极高的时间分辨率提供准确的帧内/帧间信息。在自我监督学习范式下,学习一个基于事件的RS2GS网络,其中真实世界事件和RS图像可用于缓解合成数据和真实数据之间的域差。我们提出了一个基于事件的帧内/帧间补偿器(E-IC)来预测任意时间间隔之间的每个像素的动态,包括时间过渡和空间平移。在RS-RS、RS-GS和GS-RS方面探索连接,我们明确给出了与所提出的E-IC的相互约束,从而导致了没有真实GS图像的监督。在合成和真实数据集上进行了广泛的评估,证明了所提出方法的最新技术水平,并展示了在现实世界场景中基于事件的RS2GS反演的显着性能。该数据集和代码可在https://w3un.github.io/selfunroll/上获得。