Generative 3D Mesh Modeling with Text-to-Texture Generator

Class Conditioned Mesh Generation

Method

Our 3D representation is based on deep marching tetrahedra (DMTet) method. And we use a latent diffusion model for generation.

To embed the class condition, we use concatenation of the class embedding and the latent code as the input of each block of unet. We also use cross attention layer to fuse the class embedding and the latent code.

image1

Classifier guidance is used for the generation process.

Results

Text-to-Texture Generation

Method

We distill the knowledge from 2D diffusion model to optimize the trainable texture.

image1

Results