zac commited on
Commit
2ee138e
·
1 Parent(s): 35951e7

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -3
app.py CHANGED
@@ -8,7 +8,7 @@ from huggingface_hub import hf_hub_download #load from huggingfaces
8
 
9
 
10
  llm = Llama(model_path= hf_hub_download(repo_id="TheBloke/orca_mini_3B-GGML", filename="orca-mini-3b.ggmlv3.q4_1.bin"), n_ctx=2048) #download model from hf/ n_ctx=2048 for high ccontext length
11
-
12
  history = []
13
 
14
  pre_prompt = " The user and the AI are having a conversation : "
@@ -36,8 +36,8 @@ def generate_text(input_text, history):
36
 
37
 
38
  demo = gr.ChatInterface(generate_text,
39
- title="Yes Man",
40
- description="Ask Yes Man any question",
41
  examples=["Hello", "Am I cool?", "Are tomatoes vegetables?"],
42
  cache_examples=True,
43
  retry_btn=None,
 
8
 
9
 
10
  llm = Llama(model_path= hf_hub_download(repo_id="TheBloke/orca_mini_3B-GGML", filename="orca-mini-3b.ggmlv3.q4_1.bin"), n_ctx=2048) #download model from hf/ n_ctx=2048 for high ccontext length
11
+ print ("repo_id", repo_id)
12
  history = []
13
 
14
  pre_prompt = " The user and the AI are having a conversation : "
 
36
 
37
 
38
  demo = gr.ChatInterface(generate_text,
39
+ title="LLM on CPU",
40
+ description="Running LLM with https://github.com/abetlen/llama-cpp-python. btw the text streaming thing was the hardest thing to impliment",
41
  examples=["Hello", "Am I cool?", "Are tomatoes vegetables?"],
42
  cache_examples=True,
43
  retry_btn=None,