Coverage path planning (CPP) is the task of designing a trajectory that enables a mobile agent to travel over every point of an area of interest. We propose a new method to control an unmanned aerial vehicle (UAV) carrying a camera on a CPP mission with random start positions and multiple options for landing positions in an environment containing no-fly zones. While numerous approaches have been proposed to solve similar CPP problems, we leverage end-to-end reinforcement learning (RL) to learn a control policy that generalizes over varying power constraints for the UAV. Despite recent improvements in battery technology, the maximum flying range of small UAVs is still a severe constraint, which is exacerbated by variations in the UAV's power consumption that are hard to predict. By using map-like input channels to feed spatial information through convolutional network layers to the agent, we are able to train a double deep Q-network (DDQN) to make control decisions for the UAV, balancing limited power budget and coverage goal. The proposed method can be applied to a wide variety of environments and harmonizes complex goal structures with system constraints.
翻译:覆盖路径规划(CPP)是设计一条轨道的任务,使移动剂能够穿越一个感兴趣的地区的每一点。我们提出一种新的方法来控制无人驾驶航空飞行器(UAV),该飞行器携带着一台照相机进行CP飞行任务,有随机的启动位置和在含有禁飞区的环境下着陆位置的多种选择。虽然已提出许多办法来解决类似的CPP问题,但我们利用端到端强化学习(RL)学习一项控制政策,该政策将无人驾驶航空飞行器的各种权力限制加以概括。尽管最近在电池技术方面有所改进,但小型无人驾驶航空飞行器的最大飞行范围仍是一个严重的制约,由于UAV的电力消耗量变化难以预测而加剧。通过地图式输入渠道向代理人提供空间信息,我们可以训练双深的Q-网络(DQQN)对UA进行控制决策,平衡有限的电力预算和覆盖范围目标。拟议方法可以应用于广泛的环境,使复杂的目标结构与系统限制相协调。