Update app.py
Browse files
app.py
CHANGED
@@ -3,6 +3,7 @@ import os
|
|
3 |
import requests
|
4 |
from llama_cpp import Llama
|
5 |
|
|
|
6 |
llm_name = "MuntasirHossain/Meta-Llama-3-8B-OpenOrca-GGUF"
|
7 |
llm_path = os.path.basename(llm_name)
|
8 |
|
@@ -94,7 +95,7 @@ def generate(prompt, history, max_new_tokens=512): # temperature=0.95, top_p=0.9
|
|
94 |
chatbot = gr.Chatbot(height=500)
|
95 |
with gr.Blocks(theme=gr.themes.Default(primary_hue="sky")) as demo:
|
96 |
gr.HTML("<center><h1>Fine-tuned Meta-Llama-3-8B Chatbot</h1><center>")
|
97 |
-
gr.Markdown("
|
98 |
gr.ChatInterface(
|
99 |
generate,
|
100 |
chatbot=chatbot,
|
|
|
3 |
import requests
|
4 |
from llama_cpp import Llama
|
5 |
|
6 |
+
|
7 |
llm_name = "MuntasirHossain/Meta-Llama-3-8B-OpenOrca-GGUF"
|
8 |
llm_path = os.path.basename(llm_name)
|
9 |
|
|
|
95 |
chatbot = gr.Chatbot(height=500)
|
96 |
with gr.Blocks(theme=gr.themes.Default(primary_hue="sky")) as demo:
|
97 |
gr.HTML("<center><h1>Fine-tuned Meta-Llama-3-8B Chatbot</h1><center>")
|
98 |
+
gr.Markdown("This AI agent is using the MuntasirHossain/Meta-Llama-3-8B-OpenOrca-GGUF model for text-generation. <b>Note</b>: The app is running on a free basic CPU hosted on Hugging Facce Hub. Responses may be slow!")
|
99 |
gr.ChatInterface(
|
100 |
generate,
|
101 |
chatbot=chatbot,
|