Spaces:
Paused
Paused
uuu
Browse files
app.py
CHANGED
@@ -3,9 +3,7 @@ import requests
|
|
3 |
import torch
|
4 |
import gradio as gr
|
5 |
import spaces
|
6 |
-
from PIL import Image
|
7 |
from huggingface_hub import login
|
8 |
-
import torchvision.transforms as T
|
9 |
from diffusers.utils import load_image
|
10 |
|
11 |
from models.transformer_sd3 import SD3Transformer2DModel
|
@@ -34,6 +32,23 @@ if not token:
|
|
34 |
raise ValueError("Hugging Face token not found. Set the 'HF_TOKEN' environment variable.")
|
35 |
login(token=token)
|
36 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
37 |
|
38 |
|
39 |
# ----------------------------
|
@@ -42,23 +57,6 @@ login(token=token)
|
|
42 |
@spaces.GPU
|
43 |
def gui_generation(prompt, ref_img, guidance_scale, ipadapter_scale):
|
44 |
|
45 |
-
model_path = 'stabilityai/stable-diffusion-3.5-large'
|
46 |
-
ip_adapter_path = './ip-adapter.bin'
|
47 |
-
image_encoder_path = "google/siglip-so400m-patch14-384"
|
48 |
-
|
49 |
-
transformer = SD3Transformer2DModel.from_pretrained(
|
50 |
-
model_path, subfolder="transformer", torch_dtype=torch.bfloat16
|
51 |
-
)
|
52 |
-
|
53 |
-
pipe = StableDiffusion3Pipeline.from_pretrained(
|
54 |
-
model_path, transformer=transformer, torch_dtype=torch.bfloat16
|
55 |
-
).to("cuda")
|
56 |
-
|
57 |
-
pipe.init_ipadapter(
|
58 |
-
ip_adapter_path=ip_adapter_path,
|
59 |
-
image_encoder_path=image_encoder_path,
|
60 |
-
nb_token=64,
|
61 |
-
)
|
62 |
|
63 |
ref_img = load_image(ref_img.name).convert('RGB')
|
64 |
|
|
|
3 |
import torch
|
4 |
import gradio as gr
|
5 |
import spaces
|
|
|
6 |
from huggingface_hub import login
|
|
|
7 |
from diffusers.utils import load_image
|
8 |
|
9 |
from models.transformer_sd3 import SD3Transformer2DModel
|
|
|
32 |
raise ValueError("Hugging Face token not found. Set the 'HF_TOKEN' environment variable.")
|
33 |
login(token=token)
|
34 |
|
35 |
+
model_path = 'stabilityai/stable-diffusion-3.5-large'
|
36 |
+
ip_adapter_path = './ip-adapter.bin'
|
37 |
+
image_encoder_path = "google/siglip-so400m-patch14-384"
|
38 |
+
|
39 |
+
transformer = SD3Transformer2DModel.from_pretrained(
|
40 |
+
model_path, subfolder="transformer", torch_dtype=torch.bfloat16
|
41 |
+
)
|
42 |
+
|
43 |
+
pipe = StableDiffusion3Pipeline.from_pretrained(
|
44 |
+
model_path, transformer=transformer, torch_dtype=torch.bfloat16
|
45 |
+
).to("cuda")
|
46 |
+
|
47 |
+
pipe.init_ipadapter(
|
48 |
+
ip_adapter_path=ip_adapter_path,
|
49 |
+
image_encoder_path=image_encoder_path,
|
50 |
+
nb_token=64,
|
51 |
+
)
|
52 |
|
53 |
|
54 |
# ----------------------------
|
|
|
57 |
@spaces.GPU
|
58 |
def gui_generation(prompt, ref_img, guidance_scale, ipadapter_scale):
|
59 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
60 |
|
61 |
ref_img = load_image(ref_img.name).convert('RGB')
|
62 |
|