Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
@@ -8,7 +8,7 @@ from huggingface_hub import hf_hub_download #load from huggingfaces
|
|
8 |
|
9 |
|
10 |
llm = Llama(model_path= hf_hub_download(repo_id="TheBloke/orca_mini_3B-GGML", filename="orca-mini-3b.ggmlv3.q4_1.bin"), n_ctx=2048) #download model from hf/ n_ctx=2048 for high ccontext length
|
11 |
-
|
12 |
history = []
|
13 |
|
14 |
pre_prompt = " The user and the AI are having a conversation : "
|
@@ -36,8 +36,8 @@ def generate_text(input_text, history):
|
|
36 |
|
37 |
|
38 |
demo = gr.ChatInterface(generate_text,
|
39 |
-
title="
|
40 |
-
description="
|
41 |
examples=["Hello", "Am I cool?", "Are tomatoes vegetables?"],
|
42 |
cache_examples=True,
|
43 |
retry_btn=None,
|
|
|
8 |
|
9 |
|
10 |
llm = Llama(model_path= hf_hub_download(repo_id="TheBloke/orca_mini_3B-GGML", filename="orca-mini-3b.ggmlv3.q4_1.bin"), n_ctx=2048) #download model from hf/ n_ctx=2048 for high ccontext length
|
11 |
+
print ("repo_id", repo_id)
|
12 |
history = []
|
13 |
|
14 |
pre_prompt = " The user and the AI are having a conversation : "
|
|
|
36 |
|
37 |
|
38 |
demo = gr.ChatInterface(generate_text,
|
39 |
+
title="LLM on CPU",
|
40 |
+
description="Running LLM with https://github.com/abetlen/llama-cpp-python. btw the text streaming thing was the hardest thing to impliment",
|
41 |
examples=["Hello", "Am I cool?", "Are tomatoes vegetables?"],
|
42 |
cache_examples=True,
|
43 |
retry_btn=None,
|