We present a large-scale synthetic dataset for novel view synthesis consisting of ~300k images rendered from nearly 2000 complex scenes using high-quality ray tracing at high resolution (1600 x 1600 pixels). The dataset is orders of magnitude larger than existing synthetic datasets for novel view synthesis, thus providing a large unified benchmark for both training and evaluation. Using 4 distinct sources of high-quality 3D meshes, the scenes of our dataset exhibit challenging variations in camera views, lighting, shape, materials, and textures. Because our dataset is too large for existing methods to process, we propose Sparse Voxel Light Field (SVLF), an efficient voxel-based light field approach for novel view synthesis that achieves comparable performance to NeRF on synthetic data, while being an order of magnitude faster to train and two orders of magnitude faster to render. SVLF achieves this speed by relying on a sparse voxel octree, careful voxel sampling (requiring only a handful of queries per ray), and reduced network structure; as well as ground truth depth maps at training time. Our dataset is generated by NViSII, a Python-based ray tracing renderer, which is designed to be simple for non-experts to use and share, flexible and powerful through its use of scripting, and able to create high-quality and physically-based rendered images. Experiments with a subset of our dataset allow us to compare standard methods like NeRF and mip-NeRF for single-scene modeling, and pixelNeRF for category-level modeling, pointing toward the need for future improvements in this area.
翻译:我们展示了一个大型合成数据集,用于新视觉合成,由来自近2000年复杂场景的 ~ 300k 图像组成,以高分辨率(1600 x 1600 像素) 进行高质量的射线追踪,从近2000年的复杂场景中提供 ~ 300k 图像。 数据集是比现有合成数据集规模更大的数量级,用于新颖视图合成,从而为培训和评估提供一个巨大的统一基准。 使用4种不同的3D模类来源,我们数据集的场景在摄像视图、照明、形状、材料和纹理等方面呈现着挑战性的变化。 由于我们的数据集过于庞大,无法处理现有方法,因此我们提议采用Sprass Voxel Light Fiel(SVLF) (SVLF), 高效的Voxel光的光场方法, 用于新视觉合成, 能够比NRFSII的性能性能, 而我们的数据设置的直径直到直径直径直径直径直径, 也就是直径直径直径直径直到直径直径直径直径直径直径直的直径直径直径直径直径直径直径, 。