Cooperation among humans makes it easy to execute tasks and navigate seamlessly even in unknown scenarios. With our individual knowledge and collective cognition skills, we can reason about and perform well in unforeseen situations and environments. To achieve a similar potential for a robot navigating among humans and interacting with them, it is crucial for it to acquire the ability for easy, efficient and natural ways of communication and cognition sharing with humans. In this work, we aim to exploit human gestures which is known to be the most prominent modality of communication after the speech. We demonstrate how the incorporation of gestures for communicating spatial understanding can be achieved in a very simple yet effective way using a robot having the vision and listening capability. This shows a big advantage over using only Vision and Language-based Navigation, Language Grounding or Human-Robot Interaction in a task requiring the development of cognition and indoor navigation. We adapt the state-of-the-art modules of Language Grounding and Human-Robot Interaction to demonstrate a novel system pipeline in real-world environments on a Telepresence robot for performing a set of challenging tasks. To the best of our knowledge, this is the first pipeline to couple the fields of HRI and language grounding in an indoor environment to demonstrate autonomous navigation.
翻译:人类之间的合作使得在不为人知的情景下执行任务和无缝飞行变得容易,即便在不为人知的情景下,人类之间的合作也容易执行任务和无缝飞行。凭借我们的个人知识和集体认知技能,我们可以在不可预见的情况和环境中解释和表现良好。为了实现机器人在人类之间进行机器人导航和与人类互动的类似潜力,人类获得与人类进行简单、高效和自然的通信和认知交流和交流的能力至关重要。在这项工作中,我们的目标是利用已知是演讲后最突出的通信模式的人类手势。我们展示如何利用具有视觉和听力能力的机器人,以非常简单而有效的方式将空间理解交流的姿态纳入其中。这显示了在只使用视觉和语言导航、语言定位或人类机器人与人类进行互动方面的巨大优势。在需要发展认知和室内导航的任务中,我们把语言定位和人类机器人的最先进的模块用于展示在现实世界环境中的新型系统管道,我们展示了执行一系列具有挑战性的任务的远程定位机器人。这显示了仅使用视与语言导航、语言定位或人类机器人的巨大优势。我们的知识中,这是在地面上的一种自主导航领域的第一个展示。