summaryrefslogtreecommitdiffstats
path: root/training/strategy/lora.py
diff options
context:
space:
mode:
authorVolpeon <git@volpeon.ink>2023-05-06 16:25:36 +0200
committerVolpeon <git@volpeon.ink>2023-05-06 16:25:36 +0200
commit7b04d813739c0b5595295dffdc86cc41108db2d3 (patch)
tree8958b612f5d3d665866770ad553e1004aa4b6fb8 /training/strategy/lora.py
parentUpdate (diff)
downloadtextual-inversion-diff-7b04d813739c0b5595295dffdc86cc41108db2d3.tar.gz
textual-inversion-diff-7b04d813739c0b5595295dffdc86cc41108db2d3.tar.bz2
textual-inversion-diff-7b04d813739c0b5595295dffdc86cc41108db2d3.zip
Update
Diffstat (limited to 'training/strategy/lora.py')
-rw-r--r--training/strategy/lora.py10
1 files changed, 5 insertions, 5 deletions
diff --git a/training/strategy/lora.py b/training/strategy/lora.py
index 3f4dbbc..0c0f633 100644
--- a/training/strategy/lora.py
+++ b/training/strategy/lora.py
@@ -120,11 +120,11 @@ def lora_strategy_callbacks(
120 unet_ = accelerator.unwrap_model(unet, keep_fp32_wrapper=False) 120 unet_ = accelerator.unwrap_model(unet, keep_fp32_wrapper=False)
121 text_encoder_ = accelerator.unwrap_model(text_encoder, keep_fp32_wrapper=False) 121 text_encoder_ = accelerator.unwrap_model(text_encoder, keep_fp32_wrapper=False)
122 122
123 for (token, ids) in zip(placeholder_tokens, placeholder_token_ids): 123 # for (token, ids) in zip(placeholder_tokens, placeholder_token_ids):
124 text_encoder_.text_model.embeddings.save_embed( 124 # text_encoder_.text_model.embeddings.save_embed(
125 ids, 125 # ids,
126 checkpoint_output_dir / f"{slugify(token)}_{step}_{postfix}.bin" 126 # checkpoint_output_dir / f"{slugify(token)}_{step}_{postfix}.bin"
127 ) 127 # )
128 128
129 if not pti_mode: 129 if not pti_mode:
130 lora_config = {} 130 lora_config = {}