summaryrefslogtreecommitdiffstats
path: root/training
diff options
context:
space:
mode:
Diffstat (limited to 'training')
-rw-r--r--training/strategy/lora.py4
-rw-r--r--training/strategy/ti.py1
2 files changed, 0 insertions, 5 deletions
diff --git a/training/strategy/lora.py b/training/strategy/lora.py
index 8905171..209785a 100644
--- a/training/strategy/lora.py
+++ b/training/strategy/lora.py
@@ -139,10 +139,6 @@ def lora_prepare(
139 train_dataloader: DataLoader, 139 train_dataloader: DataLoader,
140 val_dataloader: Optional[DataLoader], 140 val_dataloader: Optional[DataLoader],
141 lr_scheduler: torch.optim.lr_scheduler._LRScheduler, 141 lr_scheduler: torch.optim.lr_scheduler._LRScheduler,
142 lora_rank: int = 4,
143 lora_alpha: int = 32,
144 lora_dropout: float = 0,
145 lora_bias: str = "none",
146 **kwargs 142 **kwargs
147): 143):
148 return accelerator.prepare(text_encoder, unet, optimizer, train_dataloader, val_dataloader, lr_scheduler) + ({},) 144 return accelerator.prepare(text_encoder, unet, optimizer, train_dataloader, val_dataloader, lr_scheduler) + ({},)
diff --git a/training/strategy/ti.py b/training/strategy/ti.py
index 677f5a3..c7520ed 100644
--- a/training/strategy/ti.py
+++ b/training/strategy/ti.py
@@ -209,7 +209,6 @@ def textual_inversion_prepare(
209 text_encoder.text_model.final_layer_norm.requires_grad_(False) 209 text_encoder.text_model.final_layer_norm.requires_grad_(False)
210 text_encoder.text_model.embeddings.position_embedding.requires_grad_(False) 210 text_encoder.text_model.embeddings.position_embedding.requires_grad_(False)
211 text_encoder.text_model.embeddings.token_embedding.requires_grad_(False) 211 text_encoder.text_model.embeddings.token_embedding.requires_grad_(False)
212 text_encoder.eval()
213 212
214 return text_encoder, unet, optimizer, train_dataloader, val_dataloader, lr_scheduler, {} 213 return text_encoder, unet, optimizer, train_dataloader, val_dataloader, lr_scheduler, {}
215 214