Direct Inversion: Optimization-Free Text-Driven Real Image Editing with Diffusion Models

11/15/2022
by   Adham Elarabawy, et al.
0

With the rise of large, publicly-available text-to-image diffusion models, text-guided real image editing has garnered much research attention recently. Existing methods tend to either rely on some form of per-instance or per-task fine-tuning and optimization, require multiple novel views, or they inherently entangle preservation of real image identity, semantic coherence, and faithfulness to text guidance. In this paper, we propose an optimization-free and zero fine-tuning framework that applies complex and non-rigid edits to a single real image via a text prompt, avoiding all the pitfalls described above. Using widely-available generic pre-trained text-to-image diffusion models, we demonstrate the ability to modulate pose, scene, background, style, color, and even racial identity in an extremely flexible manner through a single target text detailing the desired edit. Furthermore, our method, which we name Direct Inversion, proposes multiple intuitively configurable hyperparameters to allow for a wide range of types and extents of real image edits. We prove our method's efficacy in producing high-quality, diverse, semantically coherent, and faithful real image edits through applying it on a variety of inputs for a multitude of tasks. We also formalize our method in well-established theory, detail future experiments for further improvement, and compare against state-of-the-art attempts.

READ FULL TEXT

page 1

page 2

page 3

page 4

page 6

page 7

research
10/17/2022

Imagic: Text-Based Real Image Editing with Diffusion Models

Text-conditioned image editing has recently attracted considerable inter...
research
12/07/2022

NeRFEditor: Differentiable Style Decomposition for Full 3D Scene Editing

We present NeRFEditor, an efficient learning framework for 3D scene edit...
research
06/13/2023

Paste, Inpaint and Harmonize via Denoising: Subject-Driven Image Editing with Pre-Trained Diffusion Model

Text-to-image generative models have attracted rising attention for flex...
research
04/17/2023

MasaCtrl: Tuning-Free Mutual Self-Attention Control for Consistent Image Synthesis and Editing

Despite the success in large-scale text-to-image generation and text-con...
research
12/12/2022

Towards Practical Plug-and-Play Diffusion Models

Diffusion-based generative models have achieved remarkable success in im...
research
05/26/2023

Negative-prompt Inversion: Fast Image Inversion for Editing with Text-guided Diffusion Models

In image editing employing diffusion models, it is crucial to preserve t...
research
04/12/2023

An Edit Friendly DDPM Noise Space: Inversion and Manipulations

Denoising diffusion probabilistic models (DDPMs) employ a sequence of wh...

Please sign up or login with your details

Forgot password? Click here to reset