We propose a decentralised view-overlap recognition framework that operates across freely moving cameras without the need of a reference 3D map. Each camera independently extracts, aggregates into a hierarchical structure, and shares feature-point descriptors over time. A view overlap is recognised by view-matching and geometric validation to discard wrongly matched views. The proposed framework is generic and can be used with different descriptors. We conduct the experiments on publicly available sequences as well as new sequences we collected with hand-held cameras. We show that Oriented FAST and Rotated BRIEF (ORB) features with Bags of Binary Words within the proposed framework lead to higher precision and a higher or similar accuracy compared to NetVLAD, RootSIFT, and SuperGlue.
翻译:我们建议一个分散的视图重叠识别框架,在不需要参考 3D 地图的情况下,在自由移动的相机上运行。 每个相机独立地提取、集成成成一个等级结构,并随着时间的推移共享特征描述符。 一种观点重叠通过视觉匹配和几何验证来识别,以错误地丢弃相匹配的观点。 拟议的框架是通用的,可以与不同的描述符一起使用。 我们用手持相机对公开可用的序列和新序列进行实验。 我们显示,在拟议框架内,带有二元字袋的Orented FAST 和旋转 BRIEF (ORB) 特征与 NetVLAD、 RootSIFT 和 SuperGlue 相比更加精确或相似。