Current technology for autonomous cars primarily focuses on getting the passenger from point A to B. Nevertheless, it has been shown that passengers are afraid of taking a ride in self-driving cars. One way to alleviate this problem is by allowing the passenger to give natural language commands to the car. However, the car can misunderstand the issued command or the visual surroundings which could lead to uncertain situations. It is desirable that the self-driving car detects these situations and interacts with the passenger to solve them. This paper proposes a model that detects uncertain situations when a command is given and finds the visual objects causing it. Optionally, a question generated by the system describing the uncertain objects is included. We argue that if the car could explain the objects in a human-like way, passengers could gain more confidence in the car's abilities. Thus, we investigate how to (1) detect uncertain situations and their underlying causes, and (2) how to generate clarifying questions for the passenger. When evaluating on the Talk2Car dataset, we show that the proposed model, \acrfull{pipeline}, improves \gls{m:ambiguous-absolute-increase} in terms of $IoU_{.5}$ compared to not using \gls{pipeline}. Furthermore, we designed a referring expression generator (REG) \acrfull{reg_model} tailored to a self-driving car setting which yields a relative improvement of \gls{m:meteor-relative} METEOR and \gls{m:rouge-relative} ROUGE-l compared with state-of-the-art REG models, and is three times faster.
翻译:自主汽车的当前技术 { 自主汽车的当前技术主要侧重于让乘客从A点到B点。 { 然而, 已经显示乘客害怕乘坐自驾驶的汽车。 缓解这一问题的方法之一是允许乘客对汽车发出自然语言指令。 但是, 汽车可以误解签发的命令或可能导致不确定情况的视觉环境。 自驾驶汽车应当检测这些情况, 并与乘客进行互动以解决这些问题。 本文提出了一个模型, 在发布命令并发现视觉物体时可以检测不确定的情况。 可能包括系统生成的描述不确定物体的问题。 我们争论说, 如果汽车能够以人种的方式解释物体, 乘客就可以对汽车的能力有更大的信心。 因此, 我们调查如何(1) 检测不确定的情况及其根本原因, 以及如何为乘客澄清问题。 在对 Talk2car 数据集进行评估时, 我们显示, 拟议的模型,\crefull{ roil} 改进了。 ( 系统: egleg- demotele- demotelemental- reqreal_ legreal_ gr_ legr_ deal_ deal_ legral_ gral_ gral_ gr_ gr_ gr_ gr_ gr_ gr_ gr_ gr_ g_ gr_ =_ lex} lex} lax_ leg_ leg_ leg_ g_ g_ leg_ gr_ g_ g_ gr_ g_ g_ g_ g_ lex_ g_ g_ g_ =) leg_ = = =x_ g_ =x_ =x_ g_ =x_