From 8f4d212b3833041448678ad8a44a9a327934f74a Mon Sep 17 00:00:00 2001 From: Volpeon Date: Thu, 15 Dec 2022 20:30:59 +0100 Subject: Avoid increased VRAM usage on validation --- infer.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) (limited to 'infer.py') diff --git a/infer.py b/infer.py index efeb24d..420cb83 100644 --- a/infer.py +++ b/infer.py @@ -34,7 +34,7 @@ torch.backends.cudnn.benchmark = True default_args = { "model": "stabilityai/stable-diffusion-2-1", "precision": "fp32", - "ti_embeddings_dir": "embeddings_ti", + "ti_embeddings_dir": "embeddings", "output_dir": "output/inference", "config": None, } @@ -190,7 +190,7 @@ def create_pipeline(model, embeddings_dir, dtype): unet = UNet2DConditionModel.from_pretrained(model, subfolder='unet', torch_dtype=dtype) scheduler = DDIMScheduler.from_pretrained(model, subfolder='scheduler', torch_dtype=dtype) - added_tokens = load_text_embeddings(tokenizer, text_encoder, embeddings_dir) + added_tokens = load_text_embeddings(tokenizer, text_encoder, Path(embeddings_dir)) print(f"Added {len(added_tokens)} tokens from embeddings dir: {added_tokens}") pipeline = VlpnStableDiffusion( -- cgit v1.2.3-70-g09d2