Visual localization is the problem of estimating the camera pose of a given image with respect to a known scene. Visual localization algorithms are a fundamental building block in advanced computer vision applications, including Mixed and Virtual Reality systems. Many algorithms used in practice represent the scene through a Structure-from-Motion (SfM) point cloud and use 2D-3D matches between a query image and the 3D points for camera pose estimation. As recently shown, image details can be accurately recovered from SfM point clouds by translating renderings of the sparse point clouds to images. To address the resulting potential privacy risks for user-generated content, it was recently proposed to lift point clouds to line clouds by replacing 3D points by randomly oriented 3D lines passing through these points. The resulting representation is unintelligible to humans and effectively prevents point cloud-to-image translation. This paper shows that a significant amount of information about the 3D scene geometry is preserved in these line clouds, allowing us to (approximately) recover the 3D point positions and thus to (approximately) recover image content. Our approach is based on the observation that the closest points between lines can yield a good approximation to the original 3D points. Code is available at https://github.com/kunalchelani/Line2Point.
翻译:视觉本地化算法是高级计算机视觉应用(包括混合和虚拟现实系统)中的基本构件。许多实际使用的算法通过结构-从运动(SfM)点云来代表场景,并使用2D-3D匹配的查询图像和3D摄像头的3D点进行估计。如最近显示的,通过将稀疏点云的图像转换成图像,可以从SfM点云云层中准确地从SfM点云层中恢复图像细节。为了解决用户生成的内容可能产生的隐私风险,最近有人提议通过随机方向的3D线来取代3D点云层。由此产生的表示法对人类不易感知,并有效地防止点云到图像的翻译。本文显示,关于3D现场测量的大量信息保存在这些线云层中,使我们能够(大约)恢复3D点位置的位置,从而(大约)恢复图像内容。我们的方法基于以下观察,即行间最接近的3D点能够产生原始的3D/CRGO2号。