Diffusers
Last updated
Last updated
Diffusers is the go-to library for state-of-the-art pretrained diffusion models for generating images, audio, and even 3D structures of molecules. Whether youβre looking for a simple inference solution or want to train your own diffusion model, Diffusers is a modular toolbox that supports both. The library is designed with a focus on usability over performance, simple over easy, and customizability over abstractions.
There are over 10,000 diffusers
compatible pipelines on the Hub which you can find by filtering at the left of . Diffusion systems are typically composed of multiple components such as text encoder, UNet, VAE, and scheduler. Even though they are not standalone models, the pipeline abstraction makes it easy to use them for inference or training.
You can find diffusion pipelines for many different tasks:
Generating images from natural language text prompts ().
Transforming images using natural language text prompts ().
Generating videos from natural language descriptions ().
You can try out the models directly in the browser if you want to test them out without downloading them, thanks to the in-browser widgets!
All diffusers
pipelines are a line away from being used! To run generation we recommended to always start from the DiffusionPipeline
:
Copied
If you want to load a specific pipeline component such as the UNet, you can do so by:
Copied
All the , , and are fully compatible with the Hub. More specifically, they can be easily loaded from the Hub using the from_pretrained()
method and can be shared with others using the push_to_hub()
method.
For more details, please check out the .
Diffusers .
Diffusers .