Affordance detection is a challenging problem with a wide variety of robotic applications. Traditional affordance detection methods are limited to a predefined set of affordance labels, hence potentially restricting the adaptability of intelligent robots in complex and dynamic environments. In this paper, we present the Open-Vocabulary Affordance Detection (OpenAD) method, which is capable of detecting an unbounded number of affordances in 3D point clouds. By simultaneously learning the affordance text and the point feature, OpenAD successfully exploits the semantic relationships between affordances. Therefore, our proposed method enables zero-shot detection and can detect previously unseen affordances without a single annotation example. Intensive experimental results show that OpenAD works effectively on a wide range of affordance detection setups and outperforms other baselines by a large margin. Additionally, we demonstrate the practicality of the proposed OpenAD in real-world robotic applications with a fast inference speed (~100 ms).
翻译:买得起检测是一个具有挑战性的问题,涉及多种机器人应用。传统的买得起检测方法仅限于一套预先定义的买得起标签,因此有可能限制智能机器人在复杂和动态环境中的适应性。在本文中,我们介绍了开放票买得起检测(Open-Voccal Affordance)方法,该方法能够在3D点云中检测到数量无限制的买得起。通过同时学习买得起文本和点特征,OpenAD成功地探索了买不起者之间的语义关系。因此,我们提议的方法可以进行零发光检测,并且可以在没有单一注释的情况下探测以前看不见的买家。密集实验结果显示,OpenAD在广泛的买得起检测装置上有效工作,并且大大超越了其他基线。此外,我们展示了提议的OpenAD在现实世界机器人应用中的实用性,其速度很快(~100米)。</s>