Generative models, as an important family of statistical modeling, target learning the observed data distribution via generating new instances. Along with the rise of neural networks, deep generative models, such as variational autoencoders (VAEs) and generative adversarial network (GANs), have made tremendous progress in 2D image synthesis. Recently, researchers switch their attentions from the 2D space to the 3D space considering that 3D data better aligns with our physical world and hence enjoys great potential in practice. However, unlike a 2D image, which owns an efficient representation (i.e., pixel grid) by nature, representing 3D data could face far more challenges. Concretely, we would expect an ideal 3D representation to be capable enough to model shapes and appearances in details, and to be highly efficient so as to model high-resolution data with fast speed and low memory cost. However, existing 3D representations, such as point clouds, meshes, and recent neural fields, usually fail to meet the above requirements simultaneously. In this survey, we make a thorough review of the development of 3D generation, including 3D shape generation and 3D-aware image synthesis, from the perspectives of both algorithms and more importantly representations. We hope that our discussion could help the community track the evolution of this field and further spark some innovative ideas to advance this challenging task.
翻译:作为统计建模的重要大家庭,有针对性地通过创造新的实例来学习观测到的数据分布。随着神经网络的兴起,深层次的基因模型,如变异自动电算器(VAEs)和基因对抗网络(GANs),在2D图像合成方面取得了巨大进展。最近,研究人员将其注意力从2D空间转向3D空间,因为3D数据与我们的实际世界更加吻合,因而在实践中具有巨大的潜力。然而,与自然界拥有高效代表(即像素电网)的2D图像不同,代表3D数据的自然界可能面临更多挑战。具体地说,我们期望一个理想的3D代表模型能够足够地在细节上建模和外观,并高效地建模高分辨率数据,快速和低记忆成本。然而,现有的3D图像,如点云、模具和最近的神经领域,通常无法同时满足上述要求。我们在这次调查中,彻底地审查了3D新一代模型的开发情况,包括3D模型模型的生成和希望领域。