Active depth sensors like structured light, lidar, and time-of-flight systems sample the depth of the entire scene uniformly at a fixed scan rate. This leads to limited spatio-temporal resolution where redundant static information is over-sampled and precious motion information might be under-sampled. In this paper, we present an efficient bio-inspired event-camera-driven depth estimation algorithm. In our approach, we dynamically illuminate areas of interest densely, depending on the scene activity detected by the event camera, and sparsely illuminate areas in the field of view with no motion. The depth estimation is achieved by an event-based structured light system consisting of a laser point projector coupled with a second event-based sensor tuned to detect the reflection of the laser from the scene. We show the feasibility of our approach in a simulated autonomous driving scenario and real indoor sequences using our prototype. We show that, in natural scenes like autonomous driving and indoor environments, moving edges correspond to less than 10% of the scene on average. Thus our setup requires the sensor to scan only 10% of the scene, which could lead to almost 90% less power consumption by the illumination source. While we present the evaluation and proof-of-concept for an event-based structured-light system, the ideas presented here are applicable for a wide range of depth-sensing modalities like LIDAR, time-of-flight, and standard stereo. Video is available at \url{https://youtu.be/Rvv9IQLYjCQ}.
翻译:活动深度传感器, 如结构化光、 激光雷达和飞行时间系统, 以固定的扫描速率统一取样整个场景的深度。 这导致有限的片段时空分辨率, 多余的静态信息被过度取样, 宝贵的运动信息可能会被少量采样。 在本文中, 我们展示了一种高效的生动事件驱动活动- 相机驱动的深度估计算法。 在我们的方法中, 我们根据事件摄像头所检测到的场景活动, 动态地照亮感兴趣的区域, 并且没有动静地在视野中进行微光区。 深度估计是通过一个基于事件的结构化的光线系统, 由激光点投影仪组成, 加上第二个基于事件的传感器, 来检测激光在现场的反射情况。 我们展示了在模拟自主驾驶场景和室内环境等自然场景中的方法的可行性, 移动的边缘与平均摄像机的10% 。 因此, 我们设置的传感器只能扫描场景的10%, 并且可以导致近90 % 的频率, 系统使用一个结构清晰度范围 。