Learning Controllable 3D Diffusion Models from Single-view Images

04/13/2023
by   Jiatao Gu, et al.
35

Diffusion models have recently become the de-facto approach for generative modeling in the 2D domain. However, extending diffusion models to 3D is challenging due to the difficulties in acquiring 3D ground truth data for training. On the other hand, 3D GANs that integrate implicit 3D representations into GANs have shown remarkable 3D-aware generation when trained only on single-view image datasets. However, 3D GANs do not provide straightforward ways to precisely control image synthesis. To address these challenges, We present Control3Diff, a 3D diffusion model that combines the strengths of diffusion models and 3D GANs for versatile, controllable 3D-aware image synthesis for single-view datasets. Control3Diff explicitly models the underlying latent distribution (optionally conditioned on external inputs), thus enabling direct control during the diffusion process. Moreover, our approach is general and applicable to any type of controlling input, allowing us to train it with the same diffusion objective without any auxiliary supervision. We validate the efficacy of Control3Diff on standard image generation benchmarks, including FFHQ, AFHQ, and ShapeNet, using various conditioning inputs such as images, sketches, and text prompts. Please see the project website (<https://jiataogu.me/control3diff>) for video comparisons.

READ FULL TEXT

page 1

page 6

page 8

page 17

page 18

page 19

page 20

page 21

research
06/13/2023

Viewset Diffusion: (0-)Image-Conditioned 3D Generative Models from 2D Data

We present Viewset Diffusion: a framework for training image-conditioned...
research
06/26/2023

Fuzzy-Conditioned Diffusion and Diffusion Projection Attention Applied to Facial Image Correction

Image diffusion has recently shown remarkable performance in image synth...
research
12/22/2022

Scalable Adaptive Computation for Iterative Generation

We present the Recurrent Interface Network (RIN), a neural net architect...
research
12/06/2022

NeRDi: Single-View NeRF Synthesis with Language-Guided Diffusion as General Image Priors

2D-to-3D reconstruction is an ill-posed problem, yet humans are good at ...
research
12/02/2022

DiffRF: Rendering-Guided 3D Radiance Field Diffusion

We introduce DiffRF, a novel approach for 3D radiance field synthesis ba...
research
05/18/2023

UniControl: A Unified Diffusion Model for Controllable Visual Generation In the Wild

Achieving machine autonomy and human control often represent divergent o...
research
02/24/2023

Modulating Pretrained Diffusion Models for Multimodal Image Synthesis

We present multimodal conditioning modules (MCM) for enabling conditiona...

Please sign up or login with your details

Forgot password? Click here to reset