Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -8,7 +8,7 @@ logging.set_verbosity_error() # This suppresses warnings, including cache migra
|
|
8 |
from diffusers import StableDiffusionPipeline
|
9 |
|
10 |
# Load the model
|
11 |
-
model_id = "
|
12 |
pipe = StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.float16)
|
13 |
|
14 |
# Check if CUDA is available and move the model to GPU if it is, otherwise use CPU
|
@@ -18,11 +18,6 @@ pipe.to(device)
|
|
18 |
print(f"Using device: {device}")
|
19 |
|
20 |
|
21 |
-
# Load your fine-tuned model from Hugging Face
|
22 |
-
model_id = "MostafaAly/stable-diffusion-finetuned"
|
23 |
-
pipe = StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.float16)
|
24 |
-
pipe.to("cuda")
|
25 |
-
|
26 |
# Define the function for text-to-image generation
|
27 |
def generate_image(prompt):
|
28 |
image = pipe(prompt).images[0]
|
|
|
8 |
from diffusers import StableDiffusionPipeline
|
9 |
|
10 |
# Load the model
|
11 |
+
model_id = "MostafaAly/stable-diffusion-finetuned"
|
12 |
pipe = StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.float16)
|
13 |
|
14 |
# Check if CUDA is available and move the model to GPU if it is, otherwise use CPU
|
|
|
18 |
print(f"Using device: {device}")
|
19 |
|
20 |
|
|
|
|
|
|
|
|
|
|
|
21 |
# Define the function for text-to-image generation
|
22 |
def generate_image(prompt):
|
23 |
image = pipe(prompt).images[0]
|