In this paper, we present a fast monocular depth estimation method for enabling 3D perception capabilities of low-cost underwater robots. We formulate a novel end-to-end deep visual learning pipeline named UDepth, which incorporates domain knowledge of image formation characteristics of natural underwater scenes. First, we adapt a new input space from raw RGB image space by exploiting underwater light attenuation prior, and then devise a least-squared formulation for coarse pixel-wise depth prediction. Subsequently, we extend this into a domain projection loss that guides the end-to-end learning of UDepth on over 9K RGB-D training samples. UDepth is designed with a computationally light MobileNetV2 backbone and a Transformer-based optimizer for ensuring fast inference rates on embedded systems. By domain-aware design choices and through comprehensive experimental analyses, we demonstrate that it is possible to achieve state-of-the-art depth estimation performance while ensuring a small computational footprint. Specifically, with 70%-80% less network parameters than existing benchmarks, UDepth achieves comparable and often better depth estimation performance. While the full model offers over 66 FPS (13 FPS) inference rates on a single GPU (CPU core), our domain projection for coarse depth prediction runs at 51.5 FPS rates on single-board NVIDIA Jetson TX2s. The inference pipelines are available at https://github.com/uf-robopi/UDepth.
翻译:在本文中,我们展示了一种快速单向深度估计方法,使低成本水下机器人的3D感知能力得以实现。我们设计了一个名为UDepth的新式端到端深视学习管道,其中包括自然水下场景图像形成特点的域知识。首先,我们通过在之前利用水下光减色,改造原始 RGB 图像空间的新输入空间,然后设计出一种最差的公式,用于粗略的像素深度预测。随后,我们将这一公式扩大到一个域预测损失,引导UDepkh在超过9K RGB-D培训样本方面进行端到端的学习。UDeptah设计了一个计算性极轻的移动网络2主干线和一个基于变异器的优化器,以确保嵌入系统的快速失灵率。我们通过域觉设计选择和全面实验分析,证明有可能在确保小的计算足迹上实现最先进的深度估算性能。具体地说,UDeptax-80%的网络参数比现有的基准要低70%-80%,UDeptah 实现可比较的深度估计性,而且往往更精确地估计性地表现。尽管完全的模型显示了我们FPSPSA的FPS-PS-PS-PS-PS-PS-PS-PS-PS-PS-PS-PS-PER 的精确的预测率。