Nekochu commited on
Commit
6251f1e
·
verified ·
1 Parent(s): 36bc0f1

Update tabbed.py

Browse files
Files changed (1) hide show
  1. tabbed.py +9 -4
tabbed.py CHANGED
@@ -1,6 +1,7 @@
1
  import gradio as gr
2
  import yaml
3
- from huggingface_hub import hf_hub_download
 
4
  from huggingface_hub.utils import LocalEntryNotFoundError
5
  from llama_cpp import Llama
6
 
@@ -24,6 +25,10 @@ while True:
24
 
25
  llm = Llama(model_path=fp, **config["llama_cpp"])
26
 
 
 
 
 
27
 
28
  def user(message, history):
29
  history = history or []
@@ -31,7 +36,7 @@ def user(message, history):
31
  history.append([message, ""])
32
  return "", history
33
 
34
-
35
  def chat(history, system_message, max_tokens, temperature, top_p, top_k, repeat_penalty):
36
  history = history or []
37
 
@@ -59,7 +64,7 @@ def chat(history, system_message, max_tokens, temperature, top_p, top_k, repeat_
59
  # stream the response
60
  yield history, history
61
 
62
-
63
  def rp_chat(history, system_message, max_tokens, temperature, top_p, top_k, repeat_penalty):
64
  history = history or []
65
 
@@ -119,7 +124,7 @@ with gr.Blocks() as demo:
119
  with gr.Row():
120
  with gr.Column():
121
  gr.Markdown(f"""
122
- ### This is the [{config["hub"]["repo_id"]}](https://huggingface.co/{config["hub"]["repo_id"]}) quantized model file [{config["hub"]["filename"]}](https://huggingface.co/{config["hub"]["repo_id"]}/blob/main/{config["hub"]["filename"]})
123
 
124
  <details>
125
  <summary><a href="https://huggingface.co/spaces/Nekochu/Llama-2-13B-novel17-french-GGUF?duplicate=true">Duplicate the Space</a> to skip the queue and run in a private space or to use your own GGUF models, simply update the <a href="https://huggingface.co/spaces/Nekochu/Llama-2-13B-novel17-french-GGUF/blob/main/config.yml">config.yml</a></summary>
 
1
  import gradio as gr
2
  import yaml
3
+ import spaces
4
+ from huggingface_hub import hf_hub_download, SpaceVariable
5
  from huggingface_hub.utils import LocalEntryNotFoundError
6
  from llama_cpp import Llama
7
 
 
25
 
26
  llm = Llama(model_path=fp, **config["llama_cpp"])
27
 
28
+ _space_name = os.getenv('HF_SPACE')
29
+ space_name = SpaceVariable("MODEL_REPO_ID").value
30
+ base_url = f"https://huggingface.co/spaces/{space_name}/blob/main/config.yml"
31
+ duplicate_url = f"https://huggingface.co/spaces/{space_name}?duplicate=true"
32
 
33
  def user(message, history):
34
  history = history or []
 
36
  history.append([message, ""])
37
  return "", history
38
 
39
+ @spaces.GPU
40
  def chat(history, system_message, max_tokens, temperature, top_p, top_k, repeat_penalty):
41
  history = history or []
42
 
 
64
  # stream the response
65
  yield history, history
66
 
67
+ @spaces.GPU
68
  def rp_chat(history, system_message, max_tokens, temperature, top_p, top_k, repeat_penalty):
69
  history = history or []
70
 
 
124
  with gr.Row():
125
  with gr.Column():
126
  gr.Markdown(f"""
127
+ ### This {_space_name} & {space_name} is the [{config["hub"]["repo_id"]}](https://huggingface.co/{config["hub"]["repo_id"]}) quantized model file [{config["hub"]["filename"]}](https://huggingface.co/{config["hub"]["repo_id"]}/blob/main/{config["hub"]["filename"]})
128
 
129
  <details>
130
  <summary><a href="https://huggingface.co/spaces/Nekochu/Llama-2-13B-novel17-french-GGUF?duplicate=true">Duplicate the Space</a> to skip the queue and run in a private space or to use your own GGUF models, simply update the <a href="https://huggingface.co/spaces/Nekochu/Llama-2-13B-novel17-french-GGUF/blob/main/config.yml">config.yml</a></summary>