Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -61,7 +61,7 @@ def generate(
|
|
61 |
if torch.cuda.is_available():
|
62 |
|
63 |
if use_img2img:
|
64 |
-
pipe = AutoPipelineForImage2Image.from_pretrained(model, torch_dtype=torch.float16)
|
65 |
|
66 |
response = requests.get(url)
|
67 |
init_image = Image.open(BytesIO(response.content)).convert("RGB")
|
@@ -106,6 +106,7 @@ def generate(
|
|
106 |
images = pipe(
|
107 |
prompt=prompt,
|
108 |
image=init_image,
|
|
|
109 |
negative_prompt=negative_prompt,
|
110 |
prompt_2=prompt_2,
|
111 |
negative_prompt_2=negative_prompt_2,
|
@@ -114,7 +115,6 @@ def generate(
|
|
114 |
guidance_scale=guidance_scale_base,
|
115 |
num_inference_steps=num_inference_steps_base,
|
116 |
generator=generator,
|
117 |
-
strength=0.5,
|
118 |
output_type="pil",
|
119 |
).images[0]
|
120 |
return images
|
|
|
61 |
if torch.cuda.is_available():
|
62 |
|
63 |
if use_img2img:
|
64 |
+
pipe = AutoPipelineForImage2Image.from_pretrained(model, torch_dtype=torch.float16, variant="fp16", use_safetensors=True)
|
65 |
|
66 |
response = requests.get(url)
|
67 |
init_image = Image.open(BytesIO(response.content)).convert("RGB")
|
|
|
106 |
images = pipe(
|
107 |
prompt=prompt,
|
108 |
image=init_image,
|
109 |
+
strength=0.5,
|
110 |
negative_prompt=negative_prompt,
|
111 |
prompt_2=prompt_2,
|
112 |
negative_prompt_2=negative_prompt_2,
|
|
|
115 |
guidance_scale=guidance_scale_base,
|
116 |
num_inference_steps=num_inference_steps_base,
|
117 |
generator=generator,
|
|
|
118 |
output_type="pil",
|
119 |
).images[0]
|
120 |
return images
|