To further reduce VRAM usage, pass `--gradient_checkpointing` and `--use_8bit_adam` flag to use 8 bit adam optimizer from [bitsandbytes](https://github.com/TimDettmers/bitsandbytes). Training takes around 11GB VRAM and 18-20 minutes on Tesla T4 in colab free tier. [![Imagic Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/ShivamShrirao/diffusers/blob/main/examples/imagic/Imagic_Stable_Diffusion.ipynb) # Imagic training example [Imagic](https://arxiv.org/abs/2210.09276) is a method for Text-Based Real Image editing with models like stable diffusion with just one image of a subject. The `train_imagic.py` script shows how to implement the training procedure and adapt it for stable diffusion. Below are examples produced using the colab notebook. | Target Text | Input Image | Edited Image | |-------------|-------------|--------------| |A photo of Barack Obama smiling with a big grin.|![Obama](imgs/obama.jpg)|![Obama](imgs/obama_edited.png)| |A bird spreading wings|![Bird](imgs/bird.jpg)|![Bird](imgs/bird_edited.png)| TODO: Update README, Please refer to the colab notebook for example usage until then. ![Imagic](imgs/imagic_paper.jpeg)