A common approach for modeling the environment of an autonomous vehicle are dynamic occupancy grid maps, in which the surrounding is divided into cells, each containing the occupancy and velocity state of its location. Despite the advantage of modeling arbitrary shaped objects, the used algorithms rely on hand-designed inverse sensor models and semantic information is missing. Therefore, we introduce a multi-task recurrent neural network to predict grid maps providing occupancies, velocity estimates, semantic information and the driveable area. During training, our network architecture, which is a combination of convolutional and recurrent layers, processes sequences of raw lidar data, that is represented as bird's eye view images with several height channels. The multi-task network is trained in an end-to-end fashion to predict occupancy grid maps without the usual preprocessing steps consisting of removing ground points and applying an inverse sensor model. In our evaluations, we show that our learned inverse sensor model is able to overcome some limitations of a geometric inverse sensor model in terms of representing object shapes and modeling freespace. Moreover, we report a better runtime performance and more accurate semantic predictions for our end-to-end approach, compared to our network relying on measurement grid maps as input data.
翻译:自主飞行器环境建模的通用方法就是动态占用网格图,将周围分为包含其位置占用和速度状态的单元,每个单元都包含其位置的占用和速度状态。尽管模拟任意形状物体的优点,但使用过的算法依靠的是人工设计的反感应模型和语义信息缺失。因此,我们引入了一个多任务经常性神经网络,以预测电格图,提供占用、速度估计、语义信息和可驾驶区域。在培训期间,我们的网络结构,即由横向和经常性层、原始利达尔数据过程序列组成的组合,以几高通道作为鸟的视觉图像。多任务网络经过端对端到端培训,以预测占用网格图,而不采用通常的预处理步骤,包括拆除地面点和应用反传感器模型。我们的评估表明,我们所学过的反传感器模型能够克服代表物体形状和自由空间建模的几何反感传感器模型的某些局限性。此外,我们报告一个更精确的运行状态和更精确的网络测量方法,作为我们最终的测算方法,用以比较我们的测距网格数据。