In robot sensing scenarios, instead of passively utilizing human captured views, an agent should be able to actively choose informative viewpoints of a 3D object as discriminative evidence to boost the recognition accuracy. This task is referred to as active object recognition. Recent works on this task rely on a massive amount of training examples to train an optimal view selection policy. But in realistic robot sensing scenarios, the large-scale training data may not exist and whether the intelligent view selection policy can be still learned from few object samples remains unclear. In this paper, we study this new problem which is extremely challenging but very meaningful in robot sensing -- Few-shot Active Object Recognition, i.e., to learn view selection policies from few object samples, which has not been considered and addressed before. We solve the proposed problem by adopting the framework of meta learning and name our method "MetaView". Extensive experiments on both category-level and instance-level classification tasks demonstrate that the proposed method can efficiently resolve issues that are hard for state-of-the-art active object recognition methods to handle, and outperform several baselines by large margins.
翻译:在机器人遥感假设中,一个代理人应该能够积极选择3D对象的信息观点,而不是被动地利用人所捕捉到的视角,而不是被动地利用人类所捕捉到的视角,作为提高识别准确度的歧视性证据。这项任务被称为主动对象识别。这项任务被称为主动对象识别。最近关于这项任务的工作依靠大量培训范例来培训最佳的视图选择政策。但在现实的机器人遥感假设中,大规模培训数据可能不存在,智能视图选择政策能否从少数对象样本中学习仍然不清楚。在本文件中,我们研究了在机器人遥感中极具挑战性但非常有意义的新问题 -- -- 少见的主动对象识别,即从少数未经过审议和处理的对象样本中学习选择政策。我们通过采用元学习框架和命名我们的方法“元观察”来解决拟议的问题。在分类和实例层面的分类任务上进行的广泛实验表明,拟议的方法能够有效解决对于最先进的主动对象识别方法难以处理的问题,并且通过大型边缘超越若干基线。