Low-Rank Adaptation of Large Language Models (LoRA)
Low-Rank Adaptation of Large Language Models (LoRA)
This is an experimental feature. Its APIs can change in future.
Low-Rank Adaptation of Large Language Models (LoRA) is a training method that accelerates the training of large models while consuming less memory. It adds pairs of rank-decomposition weight matrices (called update matrices) to existing weights, and only trains those newly added weights. This has a couple of advantages:
Previous pretrained weights are kept frozen so the model is not as prone to catastrophic forgetting.
Rank-decomposition matrices have significantly fewer parameters than the original model, which means that trained LoRA weights are easily portable.
LoRA matrices are generally added to the attention layers of the original model. 🧨 Diffusers provides the load_attn_procs() method to load the LoRA weights into a model’s attention layers. You can control the extent to which the model is adapted toward new training images via a
scale
parameter.The greater memory-efficiency allows you to run fine-tuning on consumer GPUs like the Tesla T4, RTX 3080 or even the RTX 2080 Ti! GPUs like the T4 are free and readily accessible in Kaggle or Google Colab notebooks.
💡 LoRA is not only limited to attention layers. The authors found that amending the attention layers of a language model is sufficient to obtain good downstream performance with great efficiency. This is why it’s common to just add the LoRA weights to the attention layers of a model. Check out the Using LoRA for efficient Stable Diffusion fine-tuning blog for more information about how LoRA works!
cloneofsimo was the first to try out LoRA training for Stable Diffusion in the popular lora GitHub repository. 🧨 Diffusers now supports finetuning with LoRA for text-to-image generation and DreamBooth. This guide will show you how to do both.
If you’d like to store or share your model with the community, login to your BOINC AI account (create one if you don’t have one already):
Copied
Text-to-image
Finetuning a model like Stable Diffusion, which has billions of parameters, can be slow and difficult. With LoRA, it is much easier and faster to finetune a diffusion model. It can run on hardware with as little as 11GB of GPU RAM without resorting to tricks such as 8-bit optimizers.
Training
Let’s finetune stable-diffusion-v1-5
on the Pokémon BLIP captions dataset to generate your own Pokémon.
Specify the MODEL_NAME
environment variable (either a Hub model repository id or a path to the directory containing the model weights) and pass it to the pretrained_model_name_or_path
argument. You’ll also need to set the DATASET_NAME
environment variable to the name of the dataset you want to train on. To use your own dataset, take a look at the Create a dataset for training guide.
The OUTPUT_DIR
and HUB_MODEL_ID
variables are optional and specify where to save the model to on the Hub:
Copied
There are some flags to be aware of before you start training:
--push_to_hub
stores the trained LoRA embeddings on the Hub.--report_to=wandb
reports and logs the training results to your Weights & Biases dashboard (as an example, take a look at this report).--learning_rate=1e-04
, you can afford to use a higher learning rate than you normally would with LoRA.
Now you’re ready to launch the training (you can find the full training script here). Training takes about 5 hours on a 2080 Ti GPU with 11GB of RAM, and it’ll create and save model checkpoints and the pytorch_lora_weights
in your repository.
Copied
Inference
Now you can use the model for inference by loading the base model in the StableDiffusionPipeline and then the DPMSolverMultistepScheduler:
Copied
Load the LoRA weights from your finetuned model on top of the base model weights, and then move the pipeline to a GPU for faster inference. When you merge the LoRA weights with the frozen pretrained model weights, you can optionally adjust how much of the weights to merge with the scale
parameter:
💡 A scale
value of 0
is the same as not using your LoRA weights and you’re only using the base model weights, and a scale
value of 1
means you’re only using the fully finetuned LoRA weights. Values between 0
and 1
interpolates between the two weights.
Copied
If you are loading the LoRA parameters from the Hub and if the Hub repository has a base_model
tag (such as this), then you can do:
Copied
DreamBooth
DreamBooth is a finetuning technique for personalizing a text-to-image model like Stable Diffusion to generate photorealistic images of a subject in different contexts, given a few images of the subject. However, DreamBooth is very sensitive to hyperparameters and it is easy to overfit. Some important hyperparameters to consider include those that affect the training time (learning rate, number of training steps), and inference time (number of steps, scheduler type).
💡 Take a look at the Training Stable Diffusion with DreamBooth using 🧨 Diffusers blog for an in-depth analysis of DreamBooth experiments and recommended settings.
Training
Let’s finetune stable-diffusion-v1-5
with DreamBooth and LoRA with some 🐶 dog images. Download and save these images to a directory. To use your own dataset, take a look at the Create a dataset for training guide.
To start, specify the MODEL_NAME
environment variable (either a Hub model repository id or a path to the directory containing the model weights) and pass it to the pretrained_model_name_or_path
argument. You’ll also need to set INSTANCE_DIR
to the path of the directory containing the images.
The OUTPUT_DIR
variables is optional and specifies where to save the model to on the Hub:
Copied
There are some flags to be aware of before you start training:
--push_to_hub
stores the trained LoRA embeddings on the Hub.--report_to=wandb
reports and logs the training results to your Weights & Biases dashboard (as an example, take a look at this report).--learning_rate=1e-04
, you can afford to use a higher learning rate than you normally would with LoRA.
Now you’re ready to launch the training (you can find the full training script here). The script creates and saves model checkpoints and the pytorch_lora_weights.bin
file in your repository.
It’s also possible to additionally fine-tune the text encoder with LoRA. This, in most cases, leads to better results with a slight increase in the compute. To allow fine-tuning the text encoder with LoRA, specify the --train_text_encoder
while launching the train_dreambooth_lora.py
script.
Copied
Inference
Now you can use the model for inference by loading the base model in the StableDiffusionPipeline:
Copied
Load the LoRA weights from your finetuned DreamBooth model on top of the base model weights, and then move the pipeline to a GPU for faster inference. When you merge the LoRA weights with the frozen pretrained model weights, you can optionally adjust how much of the weights to merge with the scale
parameter:
💡 A scale
value of 0
is the same as not using your LoRA weights and you’re only using the base model weights, and a scale
value of 1
means you’re only using the fully finetuned LoRA weights. Values between 0
and 1
interpolates between the two weights.
Copied
If you used --train_text_encoder
during training, then use pipe.load_lora_weights()
to load the LoRA weights. For example:
Copied
If your LoRA parameters involve the UNet as well as the Text Encoder, then passing cross_attention_kwargs={"scale": 0.5}
will apply the scale
value to both the UNet and the Text Encoder.
Note that the use of load_lora_weights() is preferred to load_attn_procs() for loading LoRA parameters. This is because load_lora_weights() can handle the following situations:
LoRA parameters that don’t have separate identifiers for the UNet and the text encoder (such as
"patrickvonplaten/lora_dreambooth_dog_example"
). So, you can just do:Copied
LoRA parameters that have separate identifiers for the UNet and the text encoder such as:
"sayakpaul/dreambooth"
.
You can also provide a local directory path to load_lora_weights() as well as load_attn_procs().
Stable Diffusion XL
We support fine-tuning with Stable Diffusion XL. Please refer to the following docs:
Unloading LoRA parameters
You can call unload_lora_weights() on a pipeline to unload the LoRA parameters.
Fusing LoRA parameters
You can call fuse_lora() on a pipeline to merge the LoRA parameters with the original parameters of the underlying model(s). This can lead to a potential speedup in the inference latency.
Unfusing LoRA parameters
To undo fuse_lora
, call unfuse_lora() on a pipeline.
Working with different LoRA scales when using LoRA fusion
If you need to use scale
when working with fuse_lora()
to control the influence of the LoRA parameters on the outputs, you should specify lora_scale
within fuse_lora()
. Passing the scale
parameter to cross_attention_kwargs
when you call the pipeline won’t work.
To use a different lora_scale
with fuse_lora()
, you should first call unfuse_lora()
on the corresponding pipeline and call fuse_lora()
again with the expected lora_scale
.
Copied
Supporting different LoRA checkpoints from Diffusers
🌍 Diffusers supports loading checkpoints from popular LoRA trainers such as Kohya and TheLastBen. In this section, we outline the current API’s details and limitations.
Kohya
This support was made possible because of the amazing contributors: @takuma104 and @isidentical.
We support loading Kohya LoRA checkpoints using load_lora_weights(). In this section, we explain how to load such a checkpoint from CivitAI in Diffusers and perform inference with it.
First, download a checkpoint. We’ll use this one for demonstration purposes.
Copied
Next, we initialize a ~DiffusionPipeline:
Copied
We then load the checkpoint downloaded from CivitAI:
Copied
If you’re loading a checkpoint in the safetensors
format, please ensure you have safetensors
installed.
And then it’s time for running inference:
Copied
Below is a comparison between the LoRA and the non-LoRA results:
You have a similar checkpoint stored on the BOINC AI Hub, you can load it directly with load_lora_weights() like so:
Copied
Kohya + Stable Diffusion XL
After the release of Stable Diffusion XL, the community contributed some amazing LoRA checkpoints trained on top of it with the Kohya trainer.
Here are some example checkpoints we tried out:
Here is an example of how to perform inference with these checkpoints in diffusers
:
Copied
Kamepan.safetensors
comes from https://civitai.com/models/22279?modelVersionId=118556 .
If you notice carefully, the inference UX is exactly identical to what we presented in the sections above.
Thanks to @isidentical for helping us on integrating this feature.
Known limitations specific to the Kohya LoRAs:
When images don’t looks similar to other UIs, such as ComfyUI, it can be because of multiple reasons, as explained here.
We don’t fully support LyCORIS checkpoints. To the best of our knowledge, our current
load_lora_weights()
should support LyCORIS checkpoints that have LoRA and LoCon modules but not the other ones, such as Hada, LoKR, etc.
TheLastBen
Here is an example:
Copied
Last updated