Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
@@ -1,25 +1,20 @@
|
|
1 |
import os
|
2 |
-
from transformers import AutoTokenizer, AutoModelForCausalLM
|
3 |
-
from sentence_transformers import SentenceTransformer
|
4 |
-
from datasets import load_dataset, Dataset
|
5 |
-
import faiss
|
6 |
import torch
|
|
|
|
|
|
|
|
|
7 |
import gradio as gr
|
8 |
|
9 |
-
|
10 |
-
# νκ²½ λ³μμμ API ν€ λ‘λ
|
11 |
hf_api_key = os.getenv('HF_API_KEY')
|
12 |
|
13 |
-
# λͺ¨λΈ λ° ν ν¬λμ΄μ
|
14 |
-
tokenizer = AutoTokenizer.from_pretrained(model_id, use_auth_token=hf_api_key)
|
15 |
-
model = AutoModelForCausalLM.from_pretrained(model_id, use_auth_token=hf_api_key)
|
16 |
-
|
17 |
-
|
18 |
-
# λͺ¨λΈ λ° ν ν¬λμ΄μ μ€μ
|
19 |
model_id = "mistralai/Mixtral-8x7B-Instruct-v0.1"
|
20 |
-
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
21 |
model = AutoModelForCausalLM.from_pretrained(
|
22 |
model_id,
|
|
|
23 |
torch_dtype=torch.bfloat16,
|
24 |
device_map="auto",
|
25 |
quantization_config=BitsAndBytesConfig(
|
@@ -68,6 +63,8 @@ def rag_chatbot_interface(prompt: str, k: int = 2):
|
|
68 |
formatted_prompt = format_prompt(prompt, retrieved_documents, k)
|
69 |
return generate(formatted_prompt)
|
70 |
|
|
|
|
|
71 |
# Gradio μΈν°νμ΄μ€ μ€μ
|
72 |
iface = gr.Interface(
|
73 |
fn=rag_chatbot_interface,
|
|
|
1 |
import os
|
|
|
|
|
|
|
|
|
2 |
import torch
|
3 |
+
import faiss
|
4 |
+
from transformers import AutoTokenizer, AutoModelForCausalLM, BitsAndBytesConfig
|
5 |
+
from sentence_transformers import SentenceTransformer
|
6 |
+
from datasets import load_dataset
|
7 |
import gradio as gr
|
8 |
|
9 |
+
# νκ²½ λ³μμμ Hugging Face API ν€ λ‘λ
|
|
|
10 |
hf_api_key = os.getenv('HF_API_KEY')
|
11 |
|
12 |
+
# λͺ¨λΈ ID λ° ν ν¬λμ΄μ μ€μ
|
|
|
|
|
|
|
|
|
|
|
13 |
model_id = "mistralai/Mixtral-8x7B-Instruct-v0.1"
|
14 |
+
tokenizer = AutoTokenizer.from_pretrained(model_id, use_auth_token=hf_api_key)
|
15 |
model = AutoModelForCausalLM.from_pretrained(
|
16 |
model_id,
|
17 |
+
use_auth_token=hf_api_key,
|
18 |
torch_dtype=torch.bfloat16,
|
19 |
device_map="auto",
|
20 |
quantization_config=BitsAndBytesConfig(
|
|
|
63 |
formatted_prompt = format_prompt(prompt, retrieved_documents, k)
|
64 |
return generate(formatted_prompt)
|
65 |
|
66 |
+
SYS_PROMPT = "You are an assistant for answering questions. You are given the extracted parts of a long document and a question. Provide a conversational answer. If you don't know the answer, just say 'I do not know.' Don't make up an answer."
|
67 |
+
|
68 |
# Gradio μΈν°νμ΄μ€ μ€μ
|
69 |
iface = gr.Interface(
|
70 |
fn=rag_chatbot_interface,
|