论文标题
XDGAN:2D空间中的多模式3D形状生成
XDGAN: Multi-Modal 3D Shape Generation in 2D Space
论文作者
论文摘要
由于2D卷积体系结构的效率,2D图像的生成模型最近在质量,分辨率和速度方面取得了巨大进展。但是,由于当前最当前的3D表示依赖于自定义网络组件,因此很难将此进度扩展到3D域。本文解决了一个核心问题:是否可以直接利用2D图像生成模型来生成3D形状?为了回答这个问题,我们提出了XDGAN,这是一种有效而快速的方法,用于将2D图像gan体系结构应用于3D对象几何形状的生成,并结合了其他表面属性,例如颜色纹理和正常属性。具体而言,我们提出了一种新颖的方法,将3D形状转换为紧凑的1通道几何图像,并利用stylegan3和图像到图像转换网络,以在2D空间中生成3D对象。生成的几何图像可以快速转换为3D网格,从而实现实时3D对象合成,可视化和交互式编辑。此外,使用标准2D体系结构可以帮助将更多的2D进步带入3D领域。我们在定量上和定性上都表明,我们的方法在各种任务中非常有效,例如3D形状生成,单视图重建和形状的操作,而与最近的3D生成模型相比,同时更快,更灵活。
Generative models for 2D images has recently seen tremendous progress in quality, resolution and speed as a result of the efficiency of 2D convolutional architectures. However it is difficult to extend this progress into the 3D domain since most current 3D representations rely on custom network components. This paper addresses a central question: Is it possible to directly leverage 2D image generative models to generate 3D shapes instead? To answer this, we propose XDGAN, an effective and fast method for applying 2D image GAN architectures to the generation of 3D object geometry combined with additional surface attributes, like color textures and normals. Specifically, we propose a novel method to convert 3D shapes into compact 1-channel geometry images and leverage StyleGAN3 and image-to-image translation networks to generate 3D objects in 2D space. The generated geometry images are quick to convert to 3D meshes, enabling real-time 3D object synthesis, visualization and interactive editing. Moreover, the use of standard 2D architectures can help bring more 2D advances into the 3D realm. We show both quantitatively and qualitatively that our method is highly effective at various tasks such as 3D shape generation, single view reconstruction and shape manipulation, while being significantly faster and more flexible compared to recent 3D generative models.