Grounding language in contextual information is crucial for fine-grained natural language understanding. One important task that involves grounding contextual modifiers is color generation. Given a reference color "green", and a modifier "bluey", how does one generate a color that could represent "bluey green"? We propose a computational pragmatics model that formulates this color generation task as a recursive game between speakers and listeners. In our model, a pragmatic speaker reasons about the inferences that a listener would make, and thus generates a modified color that is maximally informative to help the listener recover the original referents. In this paper, we show that incorporating pragmatic information provides significant improvements in performance compared with other state-of-the-art deep learning models where pragmatic inference and flexibility in representing colors from a large continuous space are lacking. Our model has an absolute 98% increase in performance for the test cases where the reference colors are unseen during training, and an absolute 40% increase in performance for the test cases where both the reference colors and the modifiers are unseen during training.
翻译:在背景信息中定位语言对于细微的自然语言理解至关重要。 包含背景变异器的重要任务之一是色彩生成。 使用参考颜色“ 绿色” 和修饰符“ bluey ”, 一个人如何产生能代表“ 粗绿色” 的颜色? 我们提议了一个计算实用模型, 将这种颜色生成任务设计成演讲者和听众之间的循环游戏。 在我们的模型中, 一个实用的演讲者关于听众会做出推论的理由, 从而产生一种修改过的颜色, 其内容极其丰富, 以帮助听众恢复原始参考。 在本文中, 我们显示, 与其它最先进的学习模型相比, 与缺少实用的推论和灵活性来代表大连续空间的颜色的模型相比, 实用的信息可以大大改进性能。 我们的模型在测试案例中的性能绝对增加98%, 测试案例的性能在培训期间, 参考颜色和修饰者都是看不见的, 测试案例的性能绝对增加40%。