arielreplicate / multi-stage-blended-diffusion

Multi stage image editing of images using stable-diffusion with a mask and a prompt

  • Public
  • 790 runs
  • GitHub
  • Paper
  • License

High-Resolution Image Editing via Multi-Stage Blended Diffusion

Overview of Multi-Stage Blended Diffusion

This repository includes our implementation of Multi-Stage Blended Diffusion, as described in our paper High-Resolution Image Editing via Multi-Stage Blended Diffusion

Our implementation builds on the original implementation of Latent Diffusion, available at (https://github.com/CompVis/latent-diffusion), which is licensed under the MIT license. Specifically multi-scale-blended-diffusion/ldm, multi-scale-blended-diffusion/configs, and multi-scale-blended-diffusion/models are entirely from latent diffusion.

Setup:

  • Install the environment specified in multi-scale-blended-diffusion/environment.yaml:
conda env create -f environment.yaml
  • Download the stable diffusion v1.4 from the huggingface space and copy it to multi-scale-blended-diffusion/models/ldm/stable-diffusion-v1/model.ckpt. This requires login and has to be done manually.
  • Also download RealESRGANx4plus from here and place it in multi-scale-blended-diffusion/RealESRGAN_x4plus.pth:
wget https://github.com/xinntao/Real-ESRGAN/releases/download/v0.1.0/RealESRGAN_x4plus.pth -P multi-scale-blended-diffusion/

Usage

To try our approach for interactive editing, use the multi-scale-blended-diffusion/InteractiveEditing.ipynb notebook.

To run, validate our approach on the examples used in our paper, use multi-scale-blended-diffusion/blended_latent_diffusion.py.