Update app.py
Browse files
app.py
CHANGED
@@ -16,7 +16,7 @@ if not torch.cuda.is_available():
|
|
16 |
DESCRIPTION += "\n<p>Running on CPU 🥶 This demo may not work on CPU.</p>"
|
17 |
|
18 |
MAX_SEED = np.iinfo(np.int32).max
|
19 |
-
CACHE_EXAMPLES = torch.cuda.is_available() and os.getenv("CACHE_EXAMPLES", "1") == "
|
20 |
MAX_IMAGE_SIZE = int(os.getenv("MAX_IMAGE_SIZE", "1536"))
|
21 |
USE_TORCH_COMPILE = os.getenv("USE_TORCH_COMPILE", "0") == "1"
|
22 |
ENABLE_CPU_OFFLOAD = os.getenv("ENABLE_CPU_OFFLOAD", "0") == "1"
|
@@ -82,7 +82,7 @@ def generate(
|
|
82 |
width=width,
|
83 |
height=height,
|
84 |
guidance_scale=guidance_scale,
|
85 |
-
num_inference_steps=
|
86 |
generator=generator,
|
87 |
num_images_per_prompt=NUM_IMAGES_PER_PROMPT,
|
88 |
use_resolution_binning=use_resolution_binning,
|
|
|
16 |
DESCRIPTION += "\n<p>Running on CPU 🥶 This demo may not work on CPU.</p>"
|
17 |
|
18 |
MAX_SEED = np.iinfo(np.int32).max
|
19 |
+
CACHE_EXAMPLES = torch.cuda.is_available() and os.getenv("CACHE_EXAMPLES", "1") == "1"
|
20 |
MAX_IMAGE_SIZE = int(os.getenv("MAX_IMAGE_SIZE", "1536"))
|
21 |
USE_TORCH_COMPILE = os.getenv("USE_TORCH_COMPILE", "0") == "1"
|
22 |
ENABLE_CPU_OFFLOAD = os.getenv("ENABLE_CPU_OFFLOAD", "0") == "1"
|
|
|
82 |
width=width,
|
83 |
height=height,
|
84 |
guidance_scale=guidance_scale,
|
85 |
+
num_inference_steps=75,
|
86 |
generator=generator,
|
87 |
num_images_per_prompt=NUM_IMAGES_PER_PROMPT,
|
88 |
use_resolution_binning=use_resolution_binning,
|