LiDAR devices are widely used in autonomous driving scenarios and researches on 3D point cloud achieve remarkable progress over the past years. However, deep learning-based methods heavily rely on the annotation data and often face the domain generalization problem. Unlike 2D images whose domains are usually related to the texture information, the feature extracted from the 3D point cloud is affected by the distribution of the points. Due to the lack of a 3D domain adaptation benchmark, the common practice is to train the model on one benchmark (e.g, Waymo) and evaluate it on another dataset (e.g. KITTI). However, in this setting, there are two types of domain gaps, the scenarios domain, and sensors domain, making the evaluation and analysis complicated and difficult. To handle this situation, we propose LiDAR Dataset with Cross-Sensors (LiDAR-CS Dataset), which contains large-scale annotated LiDAR point cloud under 6 groups of different sensors but with same corresponding scenarios, captured from hybrid realistic LiDAR simulator. As far as we know, LiDAR-CS Dataset is the first dataset focused on the sensor (e.g., the points distribution) domain gaps for 3D object detection in real traffic. Furthermore, we evaluate and analyze the performance with several baseline detectors on the LiDAR-CS benchmark and show its applications.
翻译:在自主驱动情景和对三维点云的研究中,激光雷达装置被广泛用于自主驱动情景和对三维点云的研究,过去几年来取得了显著的进展。然而,深层次的学习方法在很大程度上依赖注释数据,往往面临领域普遍性问题。与通常与纹理信息有关的领域通常相关的二维图像不同,从三维点云中提取的特征受到点分布的影响。由于缺乏三维点云的3D域适应基准,通常的做法是在一个基准(如Waymo)下对模型进行培训,并在另一个数据集(如KITTI)上对其进行评价。然而,在这一设置中,有两种类型的域差距,即情景域域和传感器域,使评价和分析变得复杂和困难。为了处理这种情况,我们建议三维点云中提取的三维点数据集(LiDAR-CSD数据集),其中包含一个大型的三维域域域域内附加说明的激光雷达点云,但有相同的情景,从混合现实的LIDAR模拟数据模型中捕捉到它。据我们所知,有两种类型的域差距,即情景域域域域域域域域域域域域域域域域域域域域域域域,我们用于分析数据库的传播。