Get backend - 8
Browse files- src/pipeline.py +2 -2
src/pipeline.py
CHANGED
@@ -56,9 +56,9 @@ def load_pipeline() -> Pipeline:
|
|
56 |
|
57 |
pipeline.transformer.to(memory_format=torch.channels_last)
|
58 |
pipeline.transformer = torch.compile(pipeline.transformer, fullgraph=True, mode="max-autotune")
|
59 |
-
quantize_(pipeline.vae, int8_weight_only())
|
60 |
# pipeline.vae = torch.compile(pipeline.vae, mode="max-autotune")
|
61 |
-
|
62 |
|
63 |
PROMPT = 'semiconformity, peregrination, quip, twineless, emotionless, tawa, depickle'
|
64 |
with torch.no_grad():
|
|
|
56 |
|
57 |
pipeline.transformer.to(memory_format=torch.channels_last)
|
58 |
pipeline.transformer = torch.compile(pipeline.transformer, fullgraph=True, mode="max-autotune")
|
59 |
+
# quantize_(pipeline.vae, int8_weight_only())
|
60 |
# pipeline.vae = torch.compile(pipeline.vae, mode="max-autotune")
|
61 |
+
pipeline.set_progress_bar_config(disable=True)
|
62 |
|
63 |
PROMPT = 'semiconformity, peregrination, quip, twineless, emotionless, tawa, depickle'
|
64 |
with torch.no_grad():
|