An essential prerequisite for unleashing the potential of supervised deep learning algorithms in the area of 3D scene understanding is the availability of large-scale and richly annotated datasets. However, publicly available datasets are either in relative small spatial scales or have limited semantic annotations due to the expensive cost of data acquisition and data annotation, which severely limits the development of fine-grained semantic understanding in the context of 3D point clouds. In this paper, we present an urban-scale photogrammetric point cloud dataset with nearly three billion richly annotated points, which is three times the number of labeled points than the existing largest photogrammetric point cloud dataset. Our dataset consists of large areas from three UK cities, covering about 7.6 km^2 of the city landscape. In the dataset, each 3D point is labeled as one of 13 semantic classes. We extensively evaluate the performance of state-of-the-art algorithms on our dataset and provide a comprehensive analysis of the results. In particular, we identify several key challenges towards urban-scale point cloud understanding. The dataset is available at https://github.com/QingyongHu/SensatUrban.
翻译:在3D场景理解领域,释放监督深层学习算法潜力的一个必要先决条件是提供大规模和大量附加说明的数据集,然而,公开提供的数据集或者处于相对小的空间尺度,或者由于数据获取和数据注释费用昂贵,因此其语义说明有限,这严重限制了在3D点云范围内发展精细精密的语义理解。在本文中,我们提出了一个城市尺度光度测算点云数据集,有近30亿个丰富的附加注释点,比现有最大的光度测点云数据集多三倍。我们的数据集由三个英国城市的大片地区组成,覆盖城市地貌的大约7.6平方公里。在数据集中,每个3D点被标为13个语义类中的一个。我们广泛评价了我们数据集上的最新算法的性能,并对结果进行了全面分析。特别是,我们查明了城市尺度云值理解的几个关键挑战。数据集可在 httpsurgyong/Hguthu/Qing查阅。