seawolf2357 commited on
Commit
456ec91
Β·
verified Β·
1 Parent(s): 883f7e7

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +10 -13
app.py CHANGED
@@ -1,25 +1,20 @@
1
  import os
2
- from transformers import AutoTokenizer, AutoModelForCausalLM
3
- from sentence_transformers import SentenceTransformer
4
- from datasets import load_dataset, Dataset
5
- import faiss
6
  import torch
 
 
 
 
7
  import gradio as gr
8
 
9
-
10
- # ν™˜κ²½ λ³€μˆ˜μ—μ„œ API ν‚€ λ‘œλ“œ
11
  hf_api_key = os.getenv('HF_API_KEY')
12
 
13
- # λͺ¨λΈ 및 ν† ν¬λ‚˜μ΄μ € λ‘œλ“œν•  λ•Œ API ν‚€ μ‚¬μš©
14
- tokenizer = AutoTokenizer.from_pretrained(model_id, use_auth_token=hf_api_key)
15
- model = AutoModelForCausalLM.from_pretrained(model_id, use_auth_token=hf_api_key)
16
-
17
-
18
- # λͺ¨λΈ 및 ν† ν¬λ‚˜μ΄μ € μ„€μ •
19
  model_id = "mistralai/Mixtral-8x7B-Instruct-v0.1"
20
- tokenizer = AutoTokenizer.from_pretrained(model_id)
21
  model = AutoModelForCausalLM.from_pretrained(
22
  model_id,
 
23
  torch_dtype=torch.bfloat16,
24
  device_map="auto",
25
  quantization_config=BitsAndBytesConfig(
@@ -68,6 +63,8 @@ def rag_chatbot_interface(prompt: str, k: int = 2):
68
  formatted_prompt = format_prompt(prompt, retrieved_documents, k)
69
  return generate(formatted_prompt)
70
 
 
 
71
  # Gradio μΈν„°νŽ˜μ΄μŠ€ μ„€μ •
72
  iface = gr.Interface(
73
  fn=rag_chatbot_interface,
 
1
  import os
 
 
 
 
2
  import torch
3
+ import faiss
4
+ from transformers import AutoTokenizer, AutoModelForCausalLM, BitsAndBytesConfig
5
+ from sentence_transformers import SentenceTransformer
6
+ from datasets import load_dataset
7
  import gradio as gr
8
 
9
+ # ν™˜κ²½ λ³€μˆ˜μ—μ„œ Hugging Face API ν‚€ λ‘œλ“œ
 
10
  hf_api_key = os.getenv('HF_API_KEY')
11
 
12
+ # λͺ¨λΈ ID 및 ν† ν¬λ‚˜μ΄μ € μ„€μ •
 
 
 
 
 
13
  model_id = "mistralai/Mixtral-8x7B-Instruct-v0.1"
14
+ tokenizer = AutoTokenizer.from_pretrained(model_id, use_auth_token=hf_api_key)
15
  model = AutoModelForCausalLM.from_pretrained(
16
  model_id,
17
+ use_auth_token=hf_api_key,
18
  torch_dtype=torch.bfloat16,
19
  device_map="auto",
20
  quantization_config=BitsAndBytesConfig(
 
63
  formatted_prompt = format_prompt(prompt, retrieved_documents, k)
64
  return generate(formatted_prompt)
65
 
66
+ SYS_PROMPT = "You are an assistant for answering questions. You are given the extracted parts of a long document and a question. Provide a conversational answer. If you don't know the answer, just say 'I do not know.' Don't make up an answer."
67
+
68
  # Gradio μΈν„°νŽ˜μ΄μŠ€ μ„€μ •
69
  iface = gr.Interface(
70
  fn=rag_chatbot_interface,