We present LaLaLoc to localise in environments without the need for prior visitation, and in a manner that is robust to large changes in scene appearance, such as a full rearrangement of furniture. Specifically, LaLaLoc performs localisation through latent representations of room layout. LaLaLoc learns a rich embedding space shared between RGB panoramas and layouts inferred from a known floor plan that encodes the structural similarity between locations. Further, LaLaLoc introduces direct, cross-modal pose optimisation in its latent space. Thus, LaLaLoc enables fine-grained pose estimation in a scene without the need for prior visitation, as well as being robust to dynamics, such as a change in furniture configuration. We show that in a domestic environment LaLaLoc is able to accurately localise a single RGB panorama image to within 8.3cm, given only a floor plan as a prior.
翻译:我们介绍LaLaLoc在无需事先访问的情况下在环境中进行本地化,而且其方式与现场外观的巨大变化(如家具的全面重新排列)非常相适应。具体地说,LaLoc通过房间布局的潜在显示方式进行本地化。LaLoc学习了RGB全色和布局之间共享的丰富嵌入空间,而布局则从一个已知的地平图中推断出不同地点之间的结构相似性。此外,LaLaLoc在其潜藏空间中引入直接的跨式图像优化。因此,LaLoc在无需事先访问的情况下进行精细的图像估计,并能够对动态(如家具配置的改变)保持稳健。我们显示,在一种国内环境中,LaLaLoc能够将单一的RGB全色图像精确地定位到8.3cm范围内,而以前只给出了一个地面图。