File size: 4,975 Bytes
882e052 61b9893 c8f1f54 4fe456a 61b9893 c8f1f54 61b9893 882e052 61b9893 c8f1f54 61b9893 882e052 61b9893 5d4b17c 61b9893 9aeab3c 61b9893 c8f1f54 61b9893 882e052 087c578 61b9893 087c578 61b9893 faa166e e3617aa 61b9893 882e052 61b9893 882e052 61b9893 882e052 faa166e 61b9893 882e052 61b9893 087c578 61b9893 087c578 5d4b17c 087c578 61b9893 087c578 dc45b2e 5d4b17c 61b9893 087c578 5d4b17c 882e052 61b9893 087c578 882e052 c8f1f54 61b9893 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 |
import os
import requests
import gradio as gr
from PIL import Image, ImageDraw, ImageFont
import io
import time
from concurrent.futures import ThreadPoolExecutor
# ===== CONFIGURATION =====
HF_API_TOKEN = os.environ.get("HF_API_TOKEN")
MODEL_NAME = "stabilityai/stable-diffusion-xl-base-1.0"
API_URL = f"https://api-inference.huggingface.co/models/{MODEL_NAME}"
headers = {"Authorization": f"Bearer {HF_API_TOKEN}"}
WATERMARK_TEXT = "SelamGPT"
MAX_RETRIES = 3
TIMEOUT = 60
EXECUTOR = ThreadPoolExecutor(max_workers=2)
# ===== WATERMARK FUNCTION =====
def add_watermark(image_bytes):
"""Add watermark with optimized PNG output"""
try:
image = Image.open(io.BytesIO(image_bytes)).convert("RGB")
draw = ImageDraw.Draw(image)
font_size = 24
try:
font = ImageFont.truetype("Roboto-Bold.ttf", font_size)
except:
font = ImageFont.load_default(font_size)
text_width = draw.textlength(WATERMARK_TEXT, font=font)
x = image.width - text_width - 10
y = image.height - 34
draw.text((x+1, y+1), WATERMARK_TEXT, font=font, fill=(0, 0, 0, 128))
draw.text((x, y), WATERMARK_TEXT, font=font, fill=(255, 255, 255))
# Convert to optimized PNG
img_byte_arr = io.BytesIO()
image.save(img_byte_arr, format='PNG', optimize=True, quality=85)
img_byte_arr.seek(0)
return Image.open(img_byte_arr)
except Exception as e:
print(f"Watermark error: {str(e)}")
return Image.open(io.BytesIO(image_bytes))
# ===== IMAGE GENERATION =====
def generate_image(prompt):
if not prompt.strip():
return None, "⚠️ Please enter a prompt"
def api_call():
return requests.post(
API_URL,
headers=headers,
json={
"inputs": prompt,
"parameters": {
"height": 1024,
"width": 1024,
"num_inference_steps": 30
},
"options": {"wait_for_model": True}
},
timeout=TIMEOUT
)
for attempt in range(MAX_RETRIES):
try:
future = EXECUTOR.submit(api_call)
response = future.result()
if response.status_code == 200:
return add_watermark(response.content), "✔️ Generation successful"
elif response.status_code == 503:
wait_time = (attempt + 1) * 15
print(f"Model loading, waiting {wait_time}s...")
time.sleep(wait_time)
continue
else:
return None, f"⚠️ API Error: {response.text[:200]}"
except requests.Timeout:
return None, f"⚠️ Timeout: Model took >{TIMEOUT}s to respond"
except Exception as e:
return None, f"⚠️ Unexpected error: {str(e)[:200]}"
return None, "⚠️ Failed after multiple attempts. Please try later."
# ===== GRADIO THEME =====
theme = gr.themes.Default(
primary_hue="emerald",
secondary_hue="amber",
font=[gr.themes.GoogleFont("Poppins"), "Arial", "sans-serif"]
)
# ===== GRADIO INTERFACE =====
with gr.Blocks(theme=theme, title="SelamGPT Image Generator") as demo:
gr.Markdown("""
# 🎨 SelamGPT Image Generator
*Powered by Stable Diffusion XL (1024x1024 PNG output)*
""")
with gr.Row():
with gr.Column(scale=3):
prompt_input = gr.Textbox(
label="Describe your image",
placeholder="A futuristic Ethiopian city with flying cars...",
lines=3,
max_lines=5
)
with gr.Row():
generate_btn = gr.Button("Generate Image", variant="primary")
clear_btn = gr.Button("Clear")
gr.Examples(
examples=[
["An ancient Aksumite warrior in cyberpunk armor, 4k detailed"],
["Traditional Ethiopian coffee ceremony in zero gravity"],
["Portrait of a Habesha queen with golden jewelry"]
],
inputs=prompt_input
)
with gr.Column(scale=2):
output_image = gr.Image(
label="Generated Image",
type="pil",
format="png",
height=512
)
status_output = gr.Textbox(
label="Status",
interactive=False
)
generate_btn.click(
fn=generate_image,
inputs=prompt_input,
outputs=[output_image, status_output],
queue=True
)
clear_btn.click(
fn=lambda: [None, ""],
outputs=[output_image, status_output]
)
if __name__ == "__main__":
demo.queue(max_size=2)
demo.launch(server_name="0.0.0.0", server_port=7860) |