This paper presents a new approach for synthesizing a novel street-view panorama given an overhead satellite image. Taking a small satellite image patch as input, our method generates a Google's omnidirectional street-view type panorama, as if it is captured from the same geographical location as the center of the satellite patch. Existing works tackle this task as an image generation problem which adopts generative adversarial networks to implicitly learn the cross-view transformations, while ignoring the domain relevance. In this paper, we propose to explicitly establish the geometric correspondences between the two-view images so as to facilitate the cross-view transformation learning. Specifically, we observe that when a 3D point in the real world is visible in both views, there is a deterministic mapping between the projected points in the two-view images given the height information of this 3D point. Motivated by this, we develop a novel Satellite to Street-view image Projection (S2SP) module which explicitly establishes such geometric correspondences and projects the satellite images to the street viewpoint. With these projected satellite images as network input, we next employ a generator to synthesize realistic street-view panoramas that are geometrically consistent with the satellite images. Our S2SP module is differentiable and the whole framework is trained in an end-to-end manner. Extensive experimental results on two cross-view benchmark datasets demonstrate that our method generates images that better respect the scene geometry than existing approaches.
翻译:本文展示了将小街景全景合成为近视卫星图像的一种新方法。 我们的方法将小型卫星图像补丁作为输入, 生成谷歌的全向街道视图类型全景全景全景全景, 仿佛它是从与卫星补丁中心相同的地理位置拍摄的。 现有的作品将这项任务作为一个图像生成问题来应对, 采用基因对抗网络, 暗中学习交叉视图转换, 同时忽略域域相关性 。 在本文中, 我们提议明确建立两景图像之间的几何对应, 以便于交叉视图转型学习。 具体地说, 我们的方法是, 当真实世界中看到3D点全景全景三景全景全景全景全景三维时, 我们观察到的三维景全景三维图像的预测点之间会有一个确定性绘图图。 我们开发的新卫星到街景图像预测模块(S2SP) 明确建立这样的几何对地对应方式, 并且将卫星图像投射到街景观。 我们接下来的卫星图像作为网络输入的输入, 我们用一个更符合现实的图像模型, 将一个更精确的模型转换成一个更精确的模型。