Geometry-aware modules are widely applied in recent deep learning architectures for scene representation and rendering. However, these modules require intrinsic camera information that might not be obtained accurately. In this paper, we propose a Spatial Transformation Routing (STR) mechanism to model the spatial properties without applying any geometric prior. The STR mechanism treats the spatial transformation as the message passing process, and the relation between the view poses and the routing weights is modeled by an end-to-end trainable neural network. Besides, an Occupancy Concept Mapping (OCM) framework is proposed to provide explainable rationals for scene-fusion processes. We conducted experiments on several datasets and show that the proposed STR mechanism improves the performance of the Generative Query Network (GQN). The visualization results reveal that the routing process can pass the observed information from one location of some view to the associated location in the other view, which demonstrates the advantage of the proposed model in terms of spatial cognition.
翻译:在最近的深层学习结构中广泛应用了几何测量模型模块,用于现场展示和制作。然而,这些模块需要无法准确获得的内在相机信息。在本文中,我们提议建立一个空间转换路由机制,用以建模空间属性,而不必事先应用任何几何方法。STR机制将空间转换作为电文传递过程对待,且视景构成与路由可编程神经网络建模。此外,还提议了一个“占用概念绘图”框架,以便为现场融合过程提供可解释的合理性。我们在若干数据集上进行了实验,并表明拟议的斯特克机制改善了Geneuration Query网络(GQN)的性能。可视化结果显示,路由过程可以将观测到的信息从某些视图的一个地点传送到另一个视图中的相关地点,从而展示了拟议模型在空间认知方面的优势。