Humans have a rich representation of the entities in their environment. Entities are described by their attributes, and entities that share attributes are often semantically related. For example, if two books have "Natural Language Processing" as the value of their `title' attribute, we can expect that their `topic' attribute will also be equal, namely, "NLP". Humans tend to generalize such observations, and infer sufficient conditions under which the `topic' attribute of any entity is "NLP". If robots need to interact successfully with humans, they need to represent entities, attributes, and generalizations in a similar way. This ends in a contextualized cognitive agent that can adapt its understanding, where context provides sufficient conditions for a correct understanding. In this work, we address the problem of how to obtain these representations through human-robot interaction. We integrate visual perception and natural language input to incrementally build a semantic model of the world, and then use inductive reasoning to infer logical rules that capture generic semantic relations, true in this model. These relations can be used to enrich the human-robot interaction, to populate a knowledge base with inferred facts, or to remove uncertainty in the robot's sensory inputs.
翻译:人类在其环境中拥有丰富的实体代表。 实体按其属性来描述, 共有属性的实体往往具有生理联系。 例如, 如果两本书将“ 自然语言处理” 作为其“ 标题” 属性的价值, 我们就可以期望它们的“ 主题” 属性也将是平等的, 即“ NLP ” 。 人类倾向于将这种观察简单化, 并推导出足够的条件, 使任何实体的“ 主题” 属性成为“ NLP ” 。 如果机器人需要成功地与人类互动, 它们需要以类似的方式代表实体、 属性和一般化。 结束于一个背景化的认知媒介, 它可以在环境为正确理解提供足够条件的情况下调整其理解。 在这项工作中, 我们处理如何通过人类- 机器人互动获得这些表达的问题。 我们结合视觉感知和自然语言输入, 逐步建立世界的语义模型, 然后用感性推推逻辑规则, 来捕捉到通用的语义关系, 以类似的方式代表实体、 和一般的语义关系 。 这种关系可以用来在模型中, 使人类感官变变变变变的变的变数据, 。