Update app.py
Browse files
app.py
CHANGED
@@ -68,7 +68,7 @@ generation_config = AutoConfig.from_pretrained(
|
|
68 |
#model = llama_cpp.Llama(model_path = tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf AutoModelForCausalLM.from_pretrained(llm_model, model_file = "tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf", model_type="llama", gpu_layers=0, config = generation_config)
|
69 |
model = LlamaCpp(
|
70 |
model_path="./tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf",
|
71 |
-
chat_format="llama-2"
|
72 |
n_gpu_layers = 0,
|
73 |
temperature=0.75,
|
74 |
max_tokens=500,
|
|
|
68 |
#model = llama_cpp.Llama(model_path = tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf AutoModelForCausalLM.from_pretrained(llm_model, model_file = "tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf", model_type="llama", gpu_layers=0, config = generation_config)
|
69 |
model = LlamaCpp(
|
70 |
model_path="./tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf",
|
71 |
+
chat_format="llama-2",
|
72 |
n_gpu_layers = 0,
|
73 |
temperature=0.75,
|
74 |
max_tokens=500,
|