Accurate and fast scene understanding is one of the challenging task for autonomous driving, which requires to take full advantage of LiDAR point clouds for semantic segmentation. In this paper, we present a \textbf{concise} and \textbf{efficient} image-based semantic segmentation network, named \textbf{CENet}. In order to improve the descriptive power of learned features and reduce the computational as well as time complexity, our CENet integrates the convolution with larger kernel size instead of MLP, carefully-selected activation functions, and multiple auxiliary segmentation heads with corresponding loss functions into architecture. Quantitative and qualitative experiments conducted on publicly available benchmarks, SemanticKITTI and SemanticPOSS, demonstrate that our pipeline achieves much better mIoU and inference performance compared with state-of-the-art models. The code will be available at https://github.com/huixiancheng/CENet.
翻译:准确和快速的场景理解是自主驾驶的艰巨任务之一,它要求充分利用LIDAR点云云进行语义分解。 在本文中,我们展示了一个基于图像的语义分解网络,名为\ textbf{concise}和\ textbf{cenet}。为了提高学习特征的描述力,减少计算和时间复杂性,我们的CENet将演动与更大的内核体大小而不是MLP、仔细选择的激活功能和多个辅助分解头以及相应的损失功能整合到建筑中。我们根据公开的基准、SmanticKITTI和SenmanticPOSS进行的定量和定性实验表明,我们的输油管比State-of-the-art模型取得更好的 mIoU和推断性能。该代码将在https://github.com/huixiancheng/CENet上查阅。