Spaces:
Build error
Build error
Update app.py
Browse files
app.py
CHANGED
@@ -13,6 +13,7 @@ HF_TOKEN = os.environ.get("HF_TOKEN")
|
|
13 |
def process_model(
|
14 |
model_id: str,
|
15 |
file_path: str,
|
|
|
16 |
key: str,
|
17 |
value: str,
|
18 |
oauth_token: gr.OAuthToken | None,
|
@@ -24,6 +25,11 @@ def process_model(
|
|
24 |
|
25 |
MODEL_NAME = model_id.split("/")[-1]
|
26 |
|
|
|
|
|
|
|
|
|
|
|
27 |
FILE_NAME = file_path.split("/")[-1]
|
28 |
|
29 |
api.snapshot_download(
|
@@ -33,7 +39,7 @@ def process_model(
|
|
33 |
)
|
34 |
print("Model downloaded successully!")
|
35 |
|
36 |
-
metadata_update = f"python3 llama.cpp/gguf-py/scripts/gguf_set_metadata.py {MODEL_NAME}/{
|
37 |
subprocess.run(metadata_update, shell=True)
|
38 |
print(f"Model metadata {key} updated to {value} successully!")
|
39 |
|
@@ -60,13 +66,15 @@ with gr.Blocks() as demo:
|
|
60 |
|
61 |
file_path = gr.Textbox(lines=1, label="File path")
|
62 |
|
|
|
|
|
63 |
key = gr.Textbox(lines=1, label="Key")
|
64 |
|
65 |
value = gr.Textbox(lines=1, label="Value")
|
66 |
|
67 |
iface = gr.Interface(
|
68 |
fn=process_model,
|
69 |
-
inputs=[model_id, file_path, key, value],
|
70 |
outputs=[
|
71 |
gr.Markdown(label="output"),
|
72 |
gr.Image(show_label=False),
|
@@ -76,6 +84,24 @@ with gr.Blocks() as demo:
|
|
76 |
api_name=False,
|
77 |
)
|
78 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
79 |
|
80 |
def restart_space():
|
81 |
HfApi().restart_space(
|
|
|
13 |
def process_model(
|
14 |
model_id: str,
|
15 |
file_path: str,
|
16 |
+
file_path_dropdown: str,
|
17 |
key: str,
|
18 |
value: str,
|
19 |
oauth_token: gr.OAuthToken | None,
|
|
|
25 |
|
26 |
MODEL_NAME = model_id.split("/")[-1]
|
27 |
|
28 |
+
if file_path_dropdown:
|
29 |
+
FILE_PATH = file_path_dropdown
|
30 |
+
else:
|
31 |
+
FILE_PATH = file_path
|
32 |
+
|
33 |
FILE_NAME = file_path.split("/")[-1]
|
34 |
|
35 |
api.snapshot_download(
|
|
|
39 |
)
|
40 |
print("Model downloaded successully!")
|
41 |
|
42 |
+
metadata_update = f"yes | python3 llama.cpp/gguf-py/scripts/gguf_set_metadata.py {MODEL_NAME}/{FILE_PATH} {key} {value}"
|
43 |
subprocess.run(metadata_update, shell=True)
|
44 |
print(f"Model metadata {key} updated to {value} successully!")
|
45 |
|
|
|
66 |
|
67 |
file_path = gr.Textbox(lines=1, label="File path")
|
68 |
|
69 |
+
file_path_dropdown = gr.Dropdown(["None"], label="File", visible=False)
|
70 |
+
|
71 |
key = gr.Textbox(lines=1, label="Key")
|
72 |
|
73 |
value = gr.Textbox(lines=1, label="Value")
|
74 |
|
75 |
iface = gr.Interface(
|
76 |
fn=process_model,
|
77 |
+
inputs=[model_id, file_path, file_path_dropdown, key, value],
|
78 |
outputs=[
|
79 |
gr.Markdown(label="output"),
|
80 |
gr.Image(show_label=False),
|
|
|
84 |
api_name=False,
|
85 |
)
|
86 |
|
87 |
+
def updateFilePath(model_id: HuggingfaceHubSearch):
|
88 |
+
try:
|
89 |
+
api = HfApi()
|
90 |
+
files = []
|
91 |
+
for file in api.list_repo_tree(
|
92 |
+
repo_id=model_id,
|
93 |
+
recursive=True,
|
94 |
+
):
|
95 |
+
files.append(file.path)
|
96 |
+
|
97 |
+
return gr.update(visible=False), gr.update(visible=True, choices=files)
|
98 |
+
except Exception:
|
99 |
+
return gr.update(visible=True), gr.update(visible=False)
|
100 |
+
|
101 |
+
model_id.change(
|
102 |
+
fn=updateFilePath, inputs=model_id, outputs=[file_path, file_path_dropdown]
|
103 |
+
)
|
104 |
+
|
105 |
|
106 |
def restart_space():
|
107 |
HfApi().restart_space(
|