summaryrefslogtreecommitdiffstats
path: root/infer.py
diff options
context:
space:
mode:
Diffstat (limited to 'infer.py')
-rw-r--r--infer.py7
1 files changed, 4 insertions, 3 deletions
diff --git a/infer.py b/infer.py
index 329c60b..13219f8 100644
--- a/infer.py
+++ b/infer.py
@@ -21,7 +21,8 @@ from diffusers import (
21 LMSDiscreteScheduler, 21 LMSDiscreteScheduler,
22 EulerAncestralDiscreteScheduler, 22 EulerAncestralDiscreteScheduler,
23 KDPM2DiscreteScheduler, 23 KDPM2DiscreteScheduler,
24 KDPM2AncestralDiscreteScheduler 24 KDPM2AncestralDiscreteScheduler,
25 UniPCMultistepScheduler
25) 26)
26from transformers import CLIPTextModel 27from transformers import CLIPTextModel
27 28
@@ -29,7 +30,6 @@ from data.keywords import prompt_to_keywords, keywords_to_prompt
29from models.clip.embeddings import patch_managed_embeddings 30from models.clip.embeddings import patch_managed_embeddings
30from models.clip.tokenizer import MultiCLIPTokenizer 31from models.clip.tokenizer import MultiCLIPTokenizer
31from pipelines.stable_diffusion.vlpn_stable_diffusion import VlpnStableDiffusion 32from pipelines.stable_diffusion.vlpn_stable_diffusion import VlpnStableDiffusion
32from schedulers.scheduling_unipc_multistep import UniPCMultistepScheduler
33from util import load_config, load_embeddings_from_dir 33from util import load_config, load_embeddings_from_dir
34 34
35 35
@@ -245,7 +245,8 @@ def create_pipeline(model, dtype):
245 tokenizer=tokenizer, 245 tokenizer=tokenizer,
246 scheduler=scheduler, 246 scheduler=scheduler,
247 ) 247 )
248 pipeline.enable_xformers_memory_efficient_attention() 248 # pipeline.enable_xformers_memory_efficient_attention()
249 pipeline.unet = torch.compile(pipeline.unet)
249 pipeline.enable_vae_slicing() 250 pipeline.enable_vae_slicing()
250 pipeline.to("cuda") 251 pipeline.to("cuda")
251 252