Few-shot object detection (FSOD) aims at learning a detector that can fast adapt to previously unseen objects with scarce annotated examples, which is challenging and demanding. Existing methods solve this problem by performing subtasks of classification and localization utilizing a shared component (e.g., RoI head) in a detector, yet few of them take the preference difference in embedding space of two subtasks into consideration. In this paper, we carefully analyze the characteristics of FSOD and present that a general few-shot detector should consider the explicit decomposition of two subtasks, and leverage information from both of them for enhancing feature representations. To the end, we propose a simple yet effective Adaptive Fully-Dual Network (AFD-Net). Specifically, we extend Faster R-CNN by introducing Dual Query Encoder and Dual Attention Generator for separate feature extraction, and Dual Aggregator for separate model reweighting. Spontaneously, separate decision making is achieved with the R-CNN detector. Besides, for the acquisition of enhanced feature representations, we further introduce Adaptive Fusion Mechanism to adaptively perform feature fusion suitable for the specific subtask. Extensive experiments on PASCAL VOC and MS COCO in various settings show that, our method achieves new state-of-the-art performance by a large margin, demonstrating its effectiveness and generalization ability.
翻译:微小的物体探测(FSOD)旨在学习能够快速适应先前看不见的物体的探测器,该探测器具有挑战性和要求性。现有方法通过在探测器中使用一个共享组件(例如RoI头)进行分类和本地化子任务来解决该问题,但其中很少有人会考虑两个子任务空间的偏好差异。我们仔细分析FSOD的特性,并表明一般的微小探测器应考虑两个子任务的明确分解,并利用它们提供的信息加强地貌表现。最后,我们提出一个简单而有效的全体适应网络(AAFD-Net),具体地说,我们扩大R-CN,方法是采用双Query Ecoder和双引力生成器分别进行地貌提取,并采用双重聚合器分别进行模型再加权。我们与R-CN探测器一道作出单独的决定。此外,为了获取强化的地貌表现,我们进一步引入了适应性聚合机制,以便以适应性的方式进行地貌的、有效的全体网络网络网络(AFA-CO-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-C-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-A-