summaryrefslogtreecommitdiffstats
path: root/training/strategy/lora.py
diff options
context:
space:
mode:
authorVolpeon <git@volpeon.ink>2023-04-10 13:42:50 +0200
committerVolpeon <git@volpeon.ink>2023-04-10 13:42:50 +0200
commitcda7eba710dfde7b2e67964bcf76cd410c6a4a63 (patch)
tree7fa63edba444d4c8f05e8ef8bc66ca32fbb86bbc /training/strategy/lora.py
parentFix sample gen: models sometimes weren't in eval mode (diff)
downloadtextual-inversion-diff-cda7eba710dfde7b2e67964bcf76cd410c6a4a63.tar.gz
textual-inversion-diff-cda7eba710dfde7b2e67964bcf76cd410c6a4a63.tar.bz2
textual-inversion-diff-cda7eba710dfde7b2e67964bcf76cd410c6a4a63.zip
Update
Diffstat (limited to 'training/strategy/lora.py')
-rw-r--r--training/strategy/lora.py2
1 files changed, 1 insertions, 1 deletions
diff --git a/training/strategy/lora.py b/training/strategy/lora.py
index 0f72a17..73ec8f2 100644
--- a/training/strategy/lora.py
+++ b/training/strategy/lora.py
@@ -146,7 +146,7 @@ def lora_strategy_callbacks(
146 if torch.cuda.is_available(): 146 if torch.cuda.is_available():
147 torch.cuda.empty_cache() 147 torch.cuda.empty_cache()
148 148
149 @on_eval() 149 @torch.no_grad()
150 def on_sample(step): 150 def on_sample(step):
151 unet_ = accelerator.unwrap_model(unet, keep_fp32_wrapper=True) 151 unet_ = accelerator.unwrap_model(unet, keep_fp32_wrapper=True)
152 text_encoder_ = accelerator.unwrap_model(text_encoder, keep_fp32_wrapper=True) 152 text_encoder_ = accelerator.unwrap_model(text_encoder, keep_fp32_wrapper=True)