diff options
author | Xavier <xiaozhisheng950@gmail.com> | 2022-09-20 21:55:16 -0700 |
---|---|---|
committer | GitHub <noreply@github.com> | 2022-09-20 21:55:16 -0700 |
commit | 12fa9509f644ca1bb84c5869fe9f54ed1447cafc (patch) | |
tree | 46f4cd098f38c15fe2a83aef8ed7baa1c0c47427 | |
parent | Update README.md (diff) |
-rw-r--r-- | README.md | 2 |
1 files changed, 2 insertions, 0 deletions
@@ -10,6 +10,8 @@ The implementation makes minimum changes over the official codebase of Textual I ## Usage ### Preparation +First set-up the ```ldm``` enviroment following the instruction from textual inversion repo, or the original Stable Diffusion repo. + To fine-tune a stable diffusion model, you need to obtain the pre-trained stable diffusion models following their [instructions](https://github.com/CompVis/stable-diffusion#stable-diffusion-v1). Weights can be downloaded on [HuggingFace](https://huggingface.co/CompVis). You can decide which version of checkpoint to use, but I use ```sd-v1-4-full-ema.ckpt```. We also need to create a set of images for regularization, as the fine-tuning algorithm of Dreambooth requires that. Details of the algorithm can be found in the paper. Note that in the original paper, the regularization images seem to be generated on-the-fly. However, here I generated a set of regularization images before the training. The text prompt for generating regularization images can be ```photo of a <class>```, where ```<class>``` is a word that describes the class of your object, such as ```dog```. The command is |