We propose DeepIPC, an end-to-end autonomous driving model that handles both perception and control tasks in driving a vehicle. The model consists of two main parts, perception and controller modules. The perception module takes an RGBD image to perform semantic segmentation and bird's eye view (BEV) semantic mapping along with providing their encoded features. Meanwhile, the controller module processes these features with the measurement of GNSS locations and angular speed to estimate waypoints that come with latent features. Then, two different agents are used to translate waypoints and latent features into a set of navigational controls to drive the vehicle. The model is evaluated by predicting driving records and performing automated driving under various conditions in real environments. The experimental results show that DeepIPC achieves the best drivability and multi-task performance even with fewer parameters compared to the other models. Codes are available at https://github.com/oskarnatan/DeepIPC.
翻译:我们提出DeepIPC,这是一种端到端的自主驾驶模型,可处理驾驶车辆中的感知和控制任务。该模型由两个主要部分组成,即感知和控制器模块。感知模块采用RGBD图像执行语义分割和鸟瞰图(BEV)语义映射,并提供它们的编码特征。同时,控制器模块将这些特征与GNSS位置和角速度的测量一起处理,以估计带有潜在特征的路点。然后,使用两个不同的代理将路点和潜在特征转换为一组导航控制以驱动车辆。模型通过预测驾驶记录并在真实环境中执行自动驾驶来进行评估。实验结果表明,DeepIPC即使与其他模型相比参数较少,也能实现最佳的可驾驶性和多任务性能。代码可在https://github.com/oskarnatan/DeepIPC获得。