LiDAR semantic segmentation can provide vehicles with a rich understanding of scene, which is essential to the perception system in robotics and autonomous driving. In this paper, we propose LENet, a lightweight and efficient projection-based LiDAR semantic segmentation network, which has an encoder-decoder architecture. The encoder consists of a set of MSCA module, which is a simple convolutional attention module to capture multi-scale feature maps. The decoder consists of IAC module, which uses bilinear interpolation to upsample the multi-resolution feature maps and a single convolution layer to integrate the previous and current dimensional features. IAC is very lightweight and dramatically reduces the complexity and storage cost. Moreover, we introduce multiple auxiliary segmentation heads to further refine the network accuracy. We have conducted detailed quantitative experiments, which shows how each component contributes to the final performance. We evaluate our approach on well known public benchmarks (SemanticKITTI), which demonstrates our proposed LENet is more lightweight and effective than state-of-the-art semantic segmentation approaches.
翻译:LiDAR 语义分解可以使飞行器对场景有丰富的了解,这对机器人和自主驾驶的感知系统至关重要。在本文中,我们提议使用LENet,即轻量和高效投射的LiDAR 语义分解网络,其结构有一个编码器解码器结构。编码器由一套MSCA模块组成,这是一个简单的共进关注模块,用于捕捉多尺度的地貌图。解码器由IAC模块组成,该模块使用双线间插,将多分辨率地貌图和单一相交层合成,以整合先前和当前维系特征。IAC非常轻量,大大降低了复杂性和存储成本。此外,我们引入了多个辅助分解头,以进一步完善网络的准确性。我们进行了详细的定量实验,展示了每个组成部分如何为最终性能做出贡献。我们评估了我们对众所周知的公共基准(SemanticKTI)采用的方法,该基准显示我们提议的LENet比州级的语义分解法方法更轻、更有效。