Language grounding aims at linking the symbolic representation of language (e.g., words) into the rich perceptual knowledge of the outside world. The general approach is to embed both textual and visual information into a common space -the grounded space-confined by an explicit relationship between both modalities. We argue that this approach sacrifices the abstract knowledge obtained from linguistic co-occurrence statistics in the process of acquiring perceptual information. The focus of this paper is to solve this issue by implicitly grounding the word embeddings. Rather than learning two mappings into a joint space, our approach integrates modalities by determining a reversible grounded mapping between the textual and the grounded space by means of multi-task learning. Evaluations on intrinsic and extrinsic tasks show that our embeddings are highly beneficial for both abstract and concrete words. They are strongly correlated with human judgments and outperform previous works on a wide range of benchmarks. Our grounded embeddings are publicly available here.
翻译:语言基础旨在将语言(如文字)的象征性表述与外部世界的丰富认知知识联系起来。一般的做法是将文字和视觉信息都嵌入共同空间――两种模式之间的明确关系将封闭的空间封闭起来。我们争辩说,这种方法牺牲了在获取感知信息过程中从语言共发统计数据中获得的抽象知识。本文件的重点是通过隐含词嵌入词层来解决这一问题。我们的方法不是将两个图解混入一个共同空间,而是通过多任务学习的方式,在文本和封闭空间之间确定一种可逆的、有根有据的绘图。对内在和外在任务的评价表明,我们的嵌入对于抽象和具体语言都非常有益。它们与人类的判断密切相关,并且超越了在广泛基准方面的以往工作。我们基于根基的嵌入在这里可以公开查阅。