Differential Diffusion: Giving Each Pixel Its Strength

06/01/2023
by   Eran Levin, et al.
0

Text-based image editing has advanced significantly in recent years. With the rise of diffusion models, image editing via textual instructions has become ubiquitous. Unfortunately, current models lack the ability to customize the quantity of the change per pixel or per image fragment, resorting to changing the entire image in an equal amount, or editing a specific region using a binary mask. In this paper, we suggest a new framework which enables the user to customize the quantity of change for each image fragment, thereby enhancing the flexibility and verbosity of modern diffusion models. Our framework does not require model training or fine-tuning, but instead performs everything at inference time, making it easily applicable to an existing model. We show both qualitatively and quantitatively that our method allows better controllability and can produce results which are unattainable by existing models. Our code is available at: https://github.com/exx8/differential-diffusion

READ FULL TEXT

page 1

page 4

page 5

page 7

page 8

page 9

page 11

research
07/05/2023

DragonDiffusion: Enabling Drag-style Manipulation on Diffusion Models

Despite the ability of existing large-scale text-to-image (T2I) models t...
research
09/19/2023

Forgedit: Text Guided Image Editing via Learning and Forgetting

Text guided image editing on real images given only the image and the ta...
research
12/05/2022

Fine-grained Image Editing by Pixel-wise Guidance Using Diffusion Models

Generative models, particularly GANs, have been utilized for image editi...
research
03/17/2023

DialogPaint: A Dialog-based Image Editing Model

We present DialogPaint, an innovative framework that employs an interact...
research
08/25/2023

Unified Concept Editing in Diffusion Models

Text-to-image models suffer from various safety issues that may limit th...
research
11/26/2018

Collaging on Internal Representations: An Intuitive Approach for Semantic Transfiguration

We present a novel CNN-based image editing method that allows the user t...
research
03/25/2023

Equivariant Similarity for Vision-Language Foundation Models

This study explores the concept of equivariance in vision-language found...

Please sign up or login with your details

Forgot password? Click here to reset