Update app.py
Browse files
app.py
CHANGED
@@ -67,7 +67,7 @@ generation_config = AutoConfig.from_pretrained(
|
|
67 |
# send additional parameters to model for generation
|
68 |
|
69 |
#model = llama_cpp.Llama(model_path = tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf AutoModelForCausalLM.from_pretrained(llm_model, model_file = "tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf", model_type="llama", gpu_layers=0, config = generation_config)
|
70 |
-
model =
|
71 |
model_path="./tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf",
|
72 |
chat_format="llama-2",
|
73 |
n_gpu_layers = 0,
|
|
|
67 |
# send additional parameters to model for generation
|
68 |
|
69 |
#model = llama_cpp.Llama(model_path = tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf AutoModelForCausalLM.from_pretrained(llm_model, model_file = "tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf", model_type="llama", gpu_layers=0, config = generation_config)
|
70 |
+
model = Llama(
|
71 |
model_path="./tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf",
|
72 |
chat_format="llama-2",
|
73 |
n_gpu_layers = 0,
|