Spaces:
Runtime error
Runtime error
Upload 3 files
Browse files- README.md +13 -12
- app.py +49 -0
- requirements.txt +14 -0
README.md
CHANGED
|
@@ -1,12 +1,13 @@
|
|
| 1 |
-
---
|
| 2 |
-
title:
|
| 3 |
-
emoji:
|
| 4 |
-
colorFrom:
|
| 5 |
-
colorTo:
|
| 6 |
-
sdk: gradio
|
| 7 |
-
sdk_version:
|
| 8 |
-
app_file: app.py
|
| 9 |
-
pinned: false
|
| 10 |
-
|
| 11 |
-
|
| 12 |
-
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
title: test flux safetensors to diffusers for Zero GPU
|
| 3 |
+
emoji: π
|
| 4 |
+
colorFrom: indigo
|
| 5 |
+
colorTo: purple
|
| 6 |
+
sdk: gradio
|
| 7 |
+
sdk_version: 4.44.0
|
| 8 |
+
app_file: app.py
|
| 9 |
+
pinned: false
|
| 10 |
+
license: mit
|
| 11 |
+
---
|
| 12 |
+
|
| 13 |
+
Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
|
app.py
ADDED
|
@@ -0,0 +1,49 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import spaces
|
| 2 |
+
|
| 3 |
+
@spaces.GPU
|
| 4 |
+
def dummy_gpu():
|
| 5 |
+
pass
|
| 6 |
+
|
| 7 |
+
import gradio as gr
|
| 8 |
+
import torch
|
| 9 |
+
from pathlib import Path
|
| 10 |
+
from diffusers import FluxPipeline, FluxTransformer2DModel
|
| 11 |
+
from huggingface_hub import hf_hub_download, HfApi
|
| 12 |
+
|
| 13 |
+
IS_TURBO = False
|
| 14 |
+
TEMP_DIR = "./temp"
|
| 15 |
+
repo_id = "camenduru/FLUX.1-dev-diffusers"
|
| 16 |
+
#repo_id = "black-forest-labs/FLUX.1-schnell" # if schnell
|
| 17 |
+
#repo_id = "aoxo/flux.1dev-abliteratedv2" # if dev.abl
|
| 18 |
+
dtype = torch.bfloat16
|
| 19 |
+
#cp = hf_hub_download("John6666/flux1-backup-202502", "ultrarealFineTune_v1.safetensors", repo_type="dataset")
|
| 20 |
+
cp = hf_hub_download("John6666/flux1-backup-202502", "jibMixFlux_v8AccentueightNSFW.safetensors", repo_type="dataset")
|
| 21 |
+
transformer = FluxTransformer2DModel.from_single_file(cp, subfolder="transformer", torch_dtype=dtype, config=repo_id)
|
| 22 |
+
pipe = FluxPipeline.from_pretrained(repo_id, transformer=transformer, torch_dtype=dtype)
|
| 23 |
+
|
| 24 |
+
if IS_TURBO:
|
| 25 |
+
pipe.to("cuda")
|
| 26 |
+
pipe.load_lora_weights(hf_hub_download("ByteDance/Hyper-SD", "Hyper-FLUX.1-dev-8steps-lora.safetensors"), adapter_name="hyper-sd")
|
| 27 |
+
pipe.set_adapters(["hyper-sd"], adapter_weights=[0.125])
|
| 28 |
+
pipe.fuse_lora()
|
| 29 |
+
pipe.unload_lora_weights()
|
| 30 |
+
pipe.to("cpu")
|
| 31 |
+
|
| 32 |
+
def upload_model(repo_id: str="", token: str="", progress=gr.Progress(track_tqdm=True)):
|
| 33 |
+
if not token: return "Token not found."
|
| 34 |
+
pipe.save_pretrained(TEMP_DIR)
|
| 35 |
+
api = HfApi(token=token if token else False)
|
| 36 |
+
api.create_repo(repo_id=repo_id, token=token, private=True, exist_ok=True)
|
| 37 |
+
api.upload_folder(repo_id=repo_id, repo_type="model", folder_path=TEMP_DIR, path_in_repo=".")
|
| 38 |
+
api.upload_file(repo_id=repo_id, repo_type="model", path_or_fileobj=cp, path_in_repo=Path(cp).name)
|
| 39 |
+
return ""
|
| 40 |
+
|
| 41 |
+
with gr.Blocks() as demo:
|
| 42 |
+
repo_id = gr.Textbox(label="Repo ID", value="")
|
| 43 |
+
hf_token = gr.Textbox(label="Your HF write token", value="")
|
| 44 |
+
run_button = gr.Button("Submit", variant="primary")
|
| 45 |
+
info_md = gr.Markdown("<br><br><br>")
|
| 46 |
+
|
| 47 |
+
run_button.click(upload_model, [repo_id, hf_token], [info_md])
|
| 48 |
+
|
| 49 |
+
demo.launch()
|
requirements.txt
ADDED
|
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
huggingface_hub
|
| 2 |
+
#torch==2.4.0
|
| 3 |
+
#torchvision
|
| 4 |
+
#torchaudio
|
| 5 |
+
torchao==0.9.0
|
| 6 |
+
diffusers==0.32.2
|
| 7 |
+
peft
|
| 8 |
+
transformers==4.48.3
|
| 9 |
+
numpy
|
| 10 |
+
scipy
|
| 11 |
+
Pillow
|
| 12 |
+
sentencepiece
|
| 13 |
+
protobuf
|
| 14 |
+
triton
|