In this paper, we consider the problem where a drone has to collect semantic information to classify multiple moving targets. In particular, we address the challenge of computing control inputs that move the drone to informative viewpoints, position and orientation, when the information is extracted using a "black-box" classifier, e.g., a deep learning neural network. These algorithms typically lack of analytical relationships between the viewpoints and their associated outputs, preventing their use in information-gathering schemes. To fill this gap, we propose a novel attention-based architecture, trained via Reinforcement Learning (RL), that outputs the next viewpoint for the drone favoring the acquisition of evidence from as many unclassified targets as possible while reasoning about their movement, orientation, and occlusions. Then, we use a low-level MPC controller to move the drone to the desired viewpoint taking into account its actual dynamics. We show that our approach not only outperforms a variety of baselines but also generalizes to scenarios unseen during training. Additionally, we show that the network scales to large numbers of targets and generalizes well to different movement dynamics of the targets.
翻译:在本文中,我们考虑了无人机必须收集语义信息以对多重移动目标进行分类的问题。特别是,当信息使用“黑箱”分类器(例如深层的神经网络)提取信息时,我们处理计算控制投入将无人机转移到信息性观点、位置和方向的挑战。这些算法通常缺乏观点及其相关产出之间的分析关系,防止其在信息收集计划中的使用。为了填补这一空白,我们建议建立一个通过“强化学习”培训的新的关注型架构,该架构将无人机的下一个观点输出出来,有利于从尽可能多的非分类目标获取证据,同时解释其移动、方向和隐蔽性。然后,我们用一个低级别的MPC控制器将无人机移到理想的观点,同时考虑到其实际动态。我们表明我们的方法不仅超越了各种基线,而且还概括了在培训过程中看不见的情景。此外,我们展示了网络的规模,以大量目标为对象,并概括了目标的不同动态。