From 842f26654bbe7dfd2f45df1fd2660d3f902af8cc Mon Sep 17 00:00:00 2001 From: Volpeon Date: Fri, 17 Feb 2023 14:53:25 +0100 Subject: Remove xformers, switch to Pytorch Nightly --- training/strategy/lora.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) (limited to 'training/strategy/lora.py') diff --git a/training/strategy/lora.py b/training/strategy/lora.py index 4dd1100..ccec215 100644 --- a/training/strategy/lora.py +++ b/training/strategy/lora.py @@ -90,7 +90,7 @@ def lora_strategy_callbacks( def on_checkpoint(step, postfix): print(f"Saving checkpoint for step {step}...") - unet_ = accelerator.unwrap_model(unet) + unet_ = accelerator.unwrap_model(unet, False) unet_.save_attn_procs(checkpoint_output_dir / f"{step}_{postfix}") del unet_ -- cgit v1.2.3-54-g00ecf