Object encoding and identification are crucial for many robotic tasks such as autonomous exploration and semantic relocalization. Existing works heavily rely on the tracking of detected objects but have difficulty recalling revisited objects precisely. In this paper, we propose a novel object encoding method, which is named as AirCode, based on a graph of key-points. To be robust to the number of key-points detected, we propose a feature sparse encoding and object dense encoding method to ensure that each key-point can only affect a small part of the object descriptors, leading it to be robust to viewpoint changes, scaling, occlusion, and even object deformation. In the experiments, we show that it achieves superior performance for object identification than the state-of-the-art algorithms and is able to provide reliable semantic relocalization. It is a plug-and-play module and we expect that it will play an important role in various applications.
翻译:对象编码和识别对于自动勘探和语义再定位等许多机器人任务至关重要。 现有的工程非常依赖对已检测到的物体的跟踪, 但很难精确地回忆重新审视的物体。 在本文中, 我们根据一个关键点的图表, 提议了一个新的对象编码方法, 名为 AirCode 。 要对所检测到的关键点数保持稳健, 我们建议了一种特性稀疏编码和对象密度编码方法, 以确保每个关键点只能影响物体的一小部分描述器, 从而能够对视图的改变、 缩放、 隐蔽甚至对象的变形产生强力 。 在实验中, 我们显示它比最先进的算法在物体识别上表现优, 并且能够提供可靠的语义再定位。 这是一个插件和动作模块, 我们期望它将在各种应用中发挥重要作用 。