Entropy-driven Sampling and Training Scheme for Conditional Diffusion Generation

06/23/2022
by   Shengming Li, et al.
0

Denoising Diffusion Probabilistic Model (DDPM) is able to make flexible conditional image generation from prior noise to real data, by introducing an independent noise-aware classifier to provide conditional gradient guidance at each time step of denoising process. However, due to the ability of classifier to easily discriminate an incompletely generated image only with high-level structure, the gradient, which is a kind of class information guidance, tends to vanish early, leading to the collapse from conditional generation process into the unconditional process. To address this problem, we propose two simple but effective approaches from two perspectives. For sampling procedure, we introduce the entropy of predicted distribution as the measure of guidance vanishing level and propose an entropy-aware scaling method to adaptively recover the conditional semantic guidance. For training stage, we propose the entropy-aware optimization objectives to alleviate the overconfident prediction for noisy data.On ImageNet1000 256x256, with our proposed sampling scheme and trained classifier, the pretrained conditional and unconditional DDPM model can achieve 10.89 respectively.

READ FULL TEXT

page 2

page 13

page 14

research
03/23/2023

End-to-End Diffusion Latent Optimization Improves Classifier Guidance

Classifier guidance – using the gradients of an image classifier to stee...
research
12/17/2022

DAG: Depth-Aware Guidance with Denoising Diffusion Probabilistic Models

In recent years, generative models have undergone significant advancemen...
research
04/14/2023

Towards Controllable Diffusion Models via Reward-Guided Exploration

By formulating data samples' formation as a Markov denoising process, di...
research
10/21/2022

Conditional Diffusion with Less Explicit Guidance via Model Predictive Control

How much explicit guidance is necessary for conditional diffusion? We co...
research
03/23/2023

MagicFusion: Boosting Text-to-Image Generation Performance by Fusing Diffusion Models

The advent of open-source AI communities has produced a cornucopia of po...
research
11/13/2022

TIER-A: Denoising Learning Framework for Information Extraction

With the development of deep neural language models, great progress has ...
research
08/31/2023

Conditioning Score-Based Generative Models by Neuro-Symbolic Constraints

Score-based and diffusion models have emerged as effective approaches fo...

Please sign up or login with your details

Forgot password? Click here to reset