Collaborative object localization aims to collaboratively estimate locations of objects observed from multiple views or perspectives, which is a critical ability for multi-agent systems such as connected vehicles. To enable collaborative localization, several model-based state estimation and learning-based localization methods have been developed. Given their encouraging performance, model-based state estimation often lacks the ability to model the complex relationships among multiple objects, while learning-based methods are typically not able to fuse the observations from an arbitrary number of views and cannot well model uncertainty. In this paper, we introduce a novel spatiotemporal graph filter approach that integrates graph learning and model-based estimation to perform multi-view sensor fusion for collaborative object localization. Our approach models complex object relationships using a new spatiotemporal graph representation and fuses multi-view observations in a Bayesian fashion to improve location estimation under uncertainty. We evaluate our approach in the applications of connected autonomous driving and multiple pedestrian localization. Experimental results show that our approach outperforms previous techniques and achieves the state-of-the-art performance on collaboration localization.
翻译:合作对象定位旨在协作估计从多种观点或角度观测到的物体的位置,这是多试剂系统(例如连接车辆)的关键能力。为了实现协作本地化,已经开发了若干基于模型的国家估计和基于学习的本地化方法。鉴于这些方法的绩效令人鼓舞,基于模型的国家估计往往缺乏模型,无法模拟多个物体之间的复杂关系,而基于学习的方法通常无法将来自任意的众多观点的观测结果结合起来,也无法很好地模拟不确定性。在本文件中,我们采用了一种新型的时空图过滤法,将图形学习和基于模型的估算结合起来,以进行多视角传感器聚合,促进协作对象本地化。我们的方法模型复杂的对象关系使用新的光时态图示,并以巴伊斯方式结合多视角观测,以在不确定的情况下改进位置估计。我们评估了我们应用相连接的自主驾驶和多行人本地化的方法。实验结果表明,我们的方法超越了以往的技术,并实现了协作本地化方面的最先进的业绩。