Inpainting
Text-guided image-inpainting
The StableDiffusionInpaintPipeline allows you to edit specific parts of an image by providing a mask and a text prompt. It uses a version of Stable Diffusion, like runwayml/stable-diffusion-inpainting specifically trained for inpainting tasks.
Get started by loading an instance of the StableDiffusionInpaintPipeline:
Copied
import PIL
import requests
import torch
from io import BytesIO
from diffusers import StableDiffusionInpaintPipeline
pipeline = StableDiffusionInpaintPipeline.from_pretrained(
"runwayml/stable-diffusion-inpainting",
torch_dtype=torch.float16,
use_safetensors=True,
variant="fp16",
)
pipeline = pipeline.to("cuda")Download an image and a mask of a dog which youβll eventually replace:
Copied
Now you can create a prompt to replace the mask with something else:
Copied
image
mask_image
prompt
output


Face of a yellow cat, high resolution, sitting on a park bench

A previous experimental implementation of inpainting used a different, lower-quality process. To ensure backwards compatibility, loading a pretrained pipeline that doesnβt contain the new model will still apply the old inpainting method.
Check out the Spaces below to try out image inpainting yourself!
Preserving the Unmasked Area of the Image
Generally speaking, StableDiffusionInpaintPipeline (and other inpainting pipelines) will change the unmasked part of the image as well. If this behavior is undesirable, you can force the unmasked area to remain the same as follows:
Copied
Forcing the unmasked portion of the image to remain the same might result in some weird transitions between the unmasked and masked areas, since the model will typically change the masked and unmasked areas to make the transition more natural.
Last updated