Skip to content

ant-research/edicho

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

10 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Edicho: Consistent Image Editing in the Wild

Edicho: Consistent Image Editing in the Wild
Qingyan Bai, Hao Ouyang, Yinghao Xu, Qiuyu Wang, Ceyuan Yang, Ka Leong Cheng, Yujun Shen, Qifeng Chen

Figure: Given two images in the wild, Edicho generates consistent editing versions of them in a zero-shot manner. Our approach achieves precise consistency for editing parts (left), objects (middle), and the entire images (right) by leveraging explicit correspondence.

[Paper] [Project Page]

As a verified need, consistent editing across in-the-wild images remains a technical challenge arising from various unmanageable factors, like object poses, lighting conditions, and photography environments. Edicho, embodying "echoing the editing effect", steps in with a training-free solution based on diffusion models, featuring a fundamental design principle of using explicit image correspondence to direct editing. Specifically, the key components include an attention manipulation module and a carefully refined classifier-free guidance (CFG) denoising strategy, both of which take into account the pre-estimated correspondence. Such an inference-time algorithm enjoys a plug-and-play nature and is compatible to most diffusion-based editing methods, such as ControlNet and BrushNet. Extensive results demonstrate the efficacy of Edicho in consistent cross-image editing under diverse settings.

Figure: Method overview. To achieve consistent editing, we propose a training-free and plug-and-play method that injects the pre-computed correspondence into the pre-trained diffusion models and guides the denoising in the two levels of (a) attention features and (b) noisy latents in classifier-free guidance (CFG).

How to Run

1) Environment setup

2) Prepare pretrained model weights

3) Run inference

Organize your inputs under one folder (e.g., ./images/toy) with images and masks:

  • Images named like input_*.png or input_*.jpg
  • Masks named like mask_*.jpg

Then run:

python infer.py \
  --base_model_path /path/to/checkpoints \
  --brushnet_path Sanster/brushnet_segmentation_mask \
  --in_dir ./images/toy \
  --save_dir_root ./results \
  --prompt_main "Your prompt" \
  --prompt_add "Prompt suffix" \
  --num_inference_steps 50 \
  --seed 42

Outputs will be saved under ./results/<input_folder_name>/ as a concatenated image.

BibTeX

If you find our work helpful for your research, please consider to cite:

@inproceedings{bai2024edicho,
  title     = {Edicho: Consistent Image Editing in the Wild},
  author    = {Bai, Qingyan and Ouyang, Hao and Xu, Yinghao and Wang, Qiuyu and Yang, Ceyuan and Cheng, Ka Leong and Shen, Yujun and Chen, Qifeng},
  booktitle = {arXiv preprint arXiv:2412.21079},
  year      = {2024}
}

About

[ICCV 2025] Edicho: Consistent Image Editing in the Wild

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages