From 11e6f8f88483e6cfdccd66ad758ae1dfcfc0283b Mon Sep 17 00:00:00 2001 From: Volpeon Date: Sun, 2 Apr 2023 08:42:33 +0200 Subject: Lora: Only register params with grad to optimizer --- training/strategy/ti.py | 1 - 1 file changed, 1 deletion(-) (limited to 'training/strategy/ti.py') diff --git a/training/strategy/ti.py b/training/strategy/ti.py index 677f5a3..c7520ed 100644 --- a/training/strategy/ti.py +++ b/training/strategy/ti.py @@ -209,7 +209,6 @@ def textual_inversion_prepare( text_encoder.text_model.final_layer_norm.requires_grad_(False) text_encoder.text_model.embeddings.position_embedding.requires_grad_(False) text_encoder.text_model.embeddings.token_embedding.requires_grad_(False) - text_encoder.eval() return text_encoder, unet, optimizer, train_dataloader, val_dataloader, lr_scheduler, {} -- cgit v1.2.3-54-g00ecf