Timezone: »
Denoising diffusion models (DDMs) have shown promising results in 3D point cloud synthesis. To advance 3D DDMs and make them useful for digital artists, we require (i) high generation quality, (ii) flexibility for manipulation and applications such as conditional synthesis and shape interpolation, and (iii) the ability to output smooth surfaces or meshes. To this end, we introduce the hierarchical Latent Point Diffusion Model (LION) for 3D shape generation. LION is set up as a variational autoencoder (VAE) with a hierarchical latent space that combines a global shape latent representation with a point-structured latent space. For generation, we train two hierarchical DDMs in these latent spaces. The hierarchical VAE approach boosts performance compared to DDMs that operate on point clouds directly, while the point-structured latents are still ideally suited for DDM-based modeling. Experimentally, LION achieves state-of-the-art generation performance on multiple ShapeNet benchmarks. Furthermore, our VAE framework allows us to easily use LION for different relevant tasks: LION excels at multimodal shape denoising and voxel-conditioned synthesis, and it can be adapted for text- and image-driven 3D generation. We also demonstrate shape autoencoding and latent shape interpolation, and we augment LION with modern surface reconstruction techniques to generate smooth 3D meshes. We hope that LION provides a powerful tool for artists working with 3D shapes due to its high-quality generation, flexibility, and surface reconstruction. Project page and code: https://nv-tlabs.github.io/LION.
Author Information
xiaohui zeng (university of Toronto)
Arash Vahdat (NVIDIA Research)
Francis Williams (NVIDIA)

I am a [research scientist at NVIDIA](https://nv-tlabs.github.io/) in NYC working at the intersection of computer vision, machine learning, and computer graphics. My research is a mix of theory and application, aiming to solve practical problems in elegant ways. In particular, I’m very interested in 3d shape representations which can enable deep learning on “real-world” geometric datasets which are often noisy, unlabeled, and consisting of very large inputs. I completed my PhD from NYU in 2021 where I worked in the [Math and Data Group](https://mad.cds.nyu.edu/) and the [Geometric Computing Lab](https://cims.nyu.edu/gcl/). My advisors were [Joan Bruna](https://cims.nyu.edu/~bruna/) and [Denis Zorin](https://cims.nyu.edu/gcl/denis.html). In addition to doing research, I am the creator and maintainer of several [open source projects](https://github.com/fwilliams). These include [NumpyEigen](https://github.com/fwilliams/numpyeigen), [Point Cloud Utils](https://github.com/fwilliams/point-cloud-utils), and [FML](https://github.com/fwilliams/fml). I’m currently looking for motivated interns to work with. Please reach out to me if you would like to chat about potential collaborations at NVIDIA!
Zan Gojcic (NVIDIA)
Or Litany (NVIDIA)
Sanja Fidler (TTI at Chicago)
Karsten Kreis (NVIDIA)
More from the Same Authors
-
2022 Poster: Optimizing Data Collection for Machine Learning »
Rafid Mahmood · James Lucas · Jose M. Alvarez · Sanja Fidler · Marc Law -
2022 : Dynamic-backbone protein-ligand structure prediction with multiscale generative diffusion models »
Zhuoran Qiao · Weili Nie · Arash Vahdat · Thomas Miller · Anima Anandkumar -
2022 : Fast Sampling of Diffusion Models via Operator Learning »
Hongkai Zheng · Weili Nie · Arash Vahdat · Kamyar Azizzadenesheli · Anima Anandkumar -
2022 : How many trained neural networks are needed for influence estimation in modern deep learning? »
Sasha (Alexandre) Doubov · Tianshi Cao · David Acuna · Sanja Fidler -
2022 Spotlight: GET3D: A Generative Model of High Quality 3D Textured Shapes Learned from Images »
Jun Gao · Tianchang Shen · Zian Wang · Wenzheng Chen · Kangxue Yin · Daiqing Li · Or Litany · Zan Gojcic · Sanja Fidler -
2022 : Dynamic-backbone protein-ligand structure prediction with multiscale generative diffusion models »
Zhuoran Qiao · Weili Nie · Arash Vahdat · Thomas Miller · Anima Anandkumar -
2022 Workshop: NeurIPS 2022 Workshop on Score-Based Methods »
Yingzhen Li · Yang Song · Valentin De Bortoli · Francois-Xavier Briol · Wenbo Gong · Alexia Jolicoeur-Martineau · Arash Vahdat -
2022 Poster: EPIC-KITCHENS VISOR Benchmark: VIdeo Segmentations and Object Relations »
Ahmad Darkhalil · Dandan Shan · Bin Zhu · Jian Ma · Amlan Kar · Richard Higgins · Sanja Fidler · David Fouhey · Dima Damen -
2022 Poster: GENIE: Higher-Order Denoising Diffusion Solvers »
Tim Dockhorn · Arash Vahdat · Karsten Kreis -
2022 Poster: GET3D: A Generative Model of High Quality 3D Textured Shapes Learned from Images »
Jun Gao · Tianchang Shen · Zian Wang · Wenzheng Chen · Kangxue Yin · Daiqing Li · Or Litany · Zan Gojcic · Sanja Fidler -
2021 Poster: DIB-R++: Learning to Predict Lighting and Material with a Hybrid Differentiable Renderer »
Wenzheng Chen · Joey Litalien · Jun Gao · Zian Wang · Clement Fuji Tsang · Sameh Khamis · Or Litany · Sanja Fidler -
2021 Poster: A Contrastive Learning Approach for Training Variational Autoencoder Priors »
Jyoti Aneja · Alex Schwing · Jan Kautz · Arash Vahdat -
2021 Poster: Score-based Generative Modeling in Latent Space »
Arash Vahdat · Karsten Kreis · Jan Kautz -
2021 Poster: Controllable and Compositional Generation with Latent-Space Energy-Based Models »
Weili Nie · Arash Vahdat · Anima Anandkumar -
2021 Poster: Don’t Generate Me: Training Differentially Private Generative Models with Sinkhorn Divergence »
Tianshi Cao · Alex Bie · Arash Vahdat · Sanja Fidler · Karsten Kreis -
2020 Poster: NVAE: A Deep Hierarchical Variational Autoencoder »
Arash Vahdat · Jan Kautz -
2020 Spotlight: NVAE: A Deep Hierarchical Variational Autoencoder »
Arash Vahdat · Jan Kautz -
2020 Poster: On the distance between two neural networks and the stability of learning »
Jeremy Bernstein · Arash Vahdat · Yisong Yue · Ming-Yu Liu -
2018 Poster: DVAE#: Discrete Variational Autoencoders with Relaxed Boltzmann Priors »
Arash Vahdat · Evgeny Andriyash · William Macready -
2017 Poster: Toward Robustness against Label Noise in Training Deep Discriminative Neural Networks »
Arash Vahdat