aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorXavier <xiaozhisheng950@gmail.com>2022-09-20 21:51:52 -0700
committerGitHub <noreply@github.com>2022-09-20 21:51:52 -0700
commit4628a5293d8f00c605c67c10e6cc4558c381757f (patch)
treec16d2c15ffe825ca4a1a56d98e6e9e2977a9afad
parentUpdate util.py (diff)
Update README.md
-rw-r--r--README.md3
1 files changed, 2 insertions, 1 deletions
diff --git a/README.md b/README.md
index 7a01589..8caf483 100644
--- a/README.md
+++ b/README.md
@@ -5,7 +5,8 @@ This is an implementtaion of Google's [Dreambooth](https://arxiv.org/abs/2208.12
This code repository is based on that of [Textual Inversion](https://github.com/rinongal/textual_inversion). Note that Textual Inversion only optimizes word ebedding, while dreambooth fine-tunes the whole diffusion model.
The implementation makes minimum changes over the official codebase of Textual Inversion. In fact, due to lazyness, some components in Textual Inversion, such as the embedding manager, are not deleted, although they will never be used here.
-
+## Update
+9/20/2022: I just found a way to reduce the GPU memory a bit. Remember that this code is based on Textual Inversion, and TI's code base has [this line](https://github.com/rinongal/textual_inversion/blob/main/ldm/modules/diffusionmodules/util.py#L112), which disable gradient checkpointing in a hard-code way. This is because in TI, the Unet is not optimized. However, in Dreambooth we optimize the Unet, so we can turn on the gradient checkpoint pointing trick, as in the original SD repo [here](https://github.com/CompVis/stable-diffusion/blob/main/ldm/modules/diffusionmodules/util.py#L112). The gradient checkpoint is default to be True in [config] (https://github.com/XavierXiao/Dreambooth-Stable-Diffusion/blob/main/configs/stable-diffusion/v1-finetune_unfrozen.yaml#L47)
## Usage
### Preparation