keitokei1994 commited on
Commit
1f5ea77
·
verified ·
1 Parent(s): 2bf5e65

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +4 -4
app.py CHANGED
@@ -3,9 +3,9 @@ import gradio as gr
3
  from huggingface_hub import hf_hub_download
4
  from llama_cpp_cuda_tensorcores import Llama
5
 
6
- REPO_ID = "MaziyarPanahi/Meta-Llama-3-70B-Instruct-GGUF"
7
- MODEL_NAME = "Meta-Llama-3-70B-Instruct.Q3_K_L.gguf"
8
- MAX_CONTEXT_LENGTH = 8192
9
  CUDA = True
10
  SYSTEM_PROMPT = "You are a helpful, smart, kind, and efficient AI assistant. You always fulfill the user's requests to the best of your ability."
11
  TOKEN_STOP = ["<|eot_id|>"]
@@ -130,7 +130,7 @@ def clear_chat(chat_history_state, chat_message):
130
 
131
  def gui(llm_chat):
132
  with gr.Blocks(theme="NoCrypt/miku", css=css) as app:
133
- gr.Markdown("# Llama 3 70B Instruct GGUF")
134
  gr.Markdown(
135
  f"""
136
  ### This demo utilizes the repository ID {REPO_ID} with the model {MODEL_NAME}, powered by the LLaMA.cpp backend.
 
3
  from huggingface_hub import hf_hub_download
4
  from llama_cpp_cuda_tensorcores import Llama
5
 
6
+ REPO_ID = "keitokei1994/shisa-v1-qwen2-7b-GGUF"
7
+ MODEL_NAME = "shisa-v1-qwen2-7b.Q8_0.gguf"
8
+ MAX_CONTEXT_LENGTH = 32768
9
  CUDA = True
10
  SYSTEM_PROMPT = "You are a helpful, smart, kind, and efficient AI assistant. You always fulfill the user's requests to the best of your ability."
11
  TOKEN_STOP = ["<|eot_id|>"]
 
130
 
131
  def gui(llm_chat):
132
  with gr.Blocks(theme="NoCrypt/miku", css=css) as app:
133
+ gr.Markdown("# shisa-v1-qwen2-7b.Q8_0.gguf")
134
  gr.Markdown(
135
  f"""
136
  ### This demo utilizes the repository ID {REPO_ID} with the model {MODEL_NAME}, powered by the LLaMA.cpp backend.