In this paper, we consider the problem where a drone has to collect semantic information to classify multiple moving targets. In particular, we address the challenge of computing control inputs that move the drone to informative viewpoints, position and orientation, when the information is extracted using a ``black-box'' classifier, e.g., a deep learning neural network. These algorithms typically lack of analytical relationships between the viewpoints and their associated outputs, preventing their use in information-gathering schemes. To fill this gap, we propose a novel attention-based architecture, trained via Reinforcement Learning (RL), that outputs the next viewpoint for the drone favoring the acquisition of evidence from as many unclassified targets as possible while reasoning about their movement, orientation, and occlusions. Then, we use a low-level MPC controller to move the drone to the desired viewpoint taking into account its actual dynamics. We show that our approach not only outperforms a variety of baselines but also generalizes to scenarios unseen during training. Additionally, we show that the network scales to large numbers of targets and generalizes well to different movement dynamics of the targets.
翻译:在本文中,我们考虑了无人机必须收集语义信息以对多重移动目标进行分类的问题。特别是,当信息使用“黑盒子”的分类器(例如深层的神经网络)提取信息时,我们处理计算控制投入将无人机转向信息性观点、位置和方向的挑战。这些算法通常缺乏观点及其相关产出之间的分析关系,防止其在信息收集计划中的使用。为了填补这一空白,我们建议通过“强化学习”(RL)培训建立一个新的关注型架构,该架构将无人机的下一个观点输出为尽可能多的非分类目标获取证据,同时解释其移动、方向和隐蔽性。然后,我们用一个低级别的MPC控制器将无人机移动到理想的观点,同时考虑到其实际动态。我们表明,我们的方法不仅超越了各种基线,而且还概括了培训过程中看不见的情景。此外,我们展示了网络的规模,将大量目标扩大到大量目标,并概括了目标的不同动态。