Spaces:
Running
on
Zero
Running
on
Zero
keitokei1994
commited on
Update app.py
Browse files
app.py
CHANGED
@@ -3,9 +3,9 @@ import gradio as gr
|
|
3 |
from huggingface_hub import hf_hub_download
|
4 |
from llama_cpp_cuda_tensorcores import Llama
|
5 |
|
6 |
-
REPO_ID = "
|
7 |
-
MODEL_NAME = "
|
8 |
-
MAX_CONTEXT_LENGTH =
|
9 |
CUDA = True
|
10 |
SYSTEM_PROMPT = "You are a helpful, smart, kind, and efficient AI assistant. You always fulfill the user's requests to the best of your ability."
|
11 |
TOKEN_STOP = ["<|eot_id|>"]
|
@@ -130,7 +130,7 @@ def clear_chat(chat_history_state, chat_message):
|
|
130 |
|
131 |
def gui(llm_chat):
|
132 |
with gr.Blocks(theme="NoCrypt/miku", css=css) as app:
|
133 |
-
gr.Markdown("#
|
134 |
gr.Markdown(
|
135 |
f"""
|
136 |
### This demo utilizes the repository ID {REPO_ID} with the model {MODEL_NAME}, powered by the LLaMA.cpp backend.
|
|
|
3 |
from huggingface_hub import hf_hub_download
|
4 |
from llama_cpp_cuda_tensorcores import Llama
|
5 |
|
6 |
+
REPO_ID = "keitokei1994/shisa-v1-qwen2-7b-GGUF"
|
7 |
+
MODEL_NAME = "shisa-v1-qwen2-7b.Q8_0.gguf"
|
8 |
+
MAX_CONTEXT_LENGTH = 32768
|
9 |
CUDA = True
|
10 |
SYSTEM_PROMPT = "You are a helpful, smart, kind, and efficient AI assistant. You always fulfill the user's requests to the best of your ability."
|
11 |
TOKEN_STOP = ["<|eot_id|>"]
|
|
|
130 |
|
131 |
def gui(llm_chat):
|
132 |
with gr.Blocks(theme="NoCrypt/miku", css=css) as app:
|
133 |
+
gr.Markdown("# shisa-v1-qwen2-7b.Q8_0.gguf")
|
134 |
gr.Markdown(
|
135 |
f"""
|
136 |
### This demo utilizes the repository ID {REPO_ID} with the model {MODEL_NAME}, powered by the LLaMA.cpp backend.
|