This paper presents Deep-PANTHER, a learning-based perception-aware trajectory planner for unmanned aerial vehicles (UAVs) in dynamic environments. Given the current state of the UAV, and the predicted trajectory and size of the obstacle, Deep-PANTHER generates multiple trajectories to avoid a dynamic obstacle while simultaneously maximizing its presence in the field of view (FOV) of the onboard camera. To obtain a computationally tractable real-time solution, imitation learning is leveraged to train a Deep-PANTHER policy using demonstrations provided by a multimodal optimization-based expert. Extensive simulations show replanning times that are two orders of magnitude faster than the optimization-based expert, while achieving a similar cost. By ensuring that each expert trajectory is assigned to one distinct student trajectory in the loss function, Deep-PANTHER can also capture the multimodality of the problem and achieve a mean squared error (MSE) loss with respect to the expert that is up to 18 times smaller than state-of-the-art (Relaxed) Winner-Takes-All approaches. Deep-PANTHER is also shown to generalize well to obstacle trajectories that differ from the ones used in training.
翻译:本文展示了在动态环境中无人驾驶航空器(无人驾驶航空器)的基于学习的感知轨迹规划师Deep-Panthery。鉴于无人驾驶航空器的现状以及障碍的预计轨迹和大小,Deep-Panthere生成了多种轨迹,以避免动态障碍,同时最大限度地扩大其在机载相机视野(FOV)中的存在。为了获得一个可计算可移动的实时解决方案,利用基于多式联运优化的专家提供的演示,模拟学习被用来培训深海探索者政策。广泛的模拟显示的重新规划时间比优化专家快两级,规模比优化专家快,但成本也相似。通过确保将每个专家轨迹分配到损失功能中不同的学生轨迹,Deep-Panthery还可以捕捉到问题的多式,并实现与专家相比的平均正方差(MSE)损失,因为专家比基于多式联运优化的专家提供的演示要小到18倍。深度的模拟显示比基于优化的专家要快得多。深Panther还展示了在使用的不同轨迹中普遍使用障碍。