frankhuangg commited on
Commit
91c5c87
·
verified ·
1 Parent(s): 2b91664

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +10 -27
app.py CHANGED
@@ -1,34 +1,17 @@
1
- import os
2
- from huggingface_hub import login
3
  import gradio as gr
4
  from transformers import AutoModelForCausalLM, AutoTokenizer
5
- import torch
6
 
7
- # 从 Hugging Face Secrets 中读取 Token
8
- token = os.getenv("HUGGINGFACE_API_TOKEN")
9
-
10
- # 登录 Hugging Face 账户
11
- login(token=token)
12
-
13
- # 加载模型和分词器
14
  model_name = "taide/Llama3-TAIDE-LX-8B-Chat-Alpha1"
 
15
  tokenizer = AutoTokenizer.from_pretrained(model_name)
16
- model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.float16, device_map="auto")
17
-
18
- # 定义推理函数
19
- def generate_text(prompt):
20
- inputs = tokenizer(prompt, return_tensors="pt").to("cuda")
21
- outputs = model.generate(**inputs)
22
- return tokenizer.decode(outputs[0], skip_special_tokens=True)
23
 
24
- # 创建 Gradio 界面
25
- interface = gr.Interface(
26
- fn=generate_text,
27
- inputs="text",
28
- outputs="text",
29
- title="Llama3 Chatbot",
30
- description="与 Llama3-TAIDE-LX-8B-Chat-Alpha1 交谈"
31
- )
32
 
33
- # 启动 Gradio 应用
34
- interface.launch(share=True)
 
 
 
 
1
  import gradio as gr
2
  from transformers import AutoModelForCausalLM, AutoTokenizer
 
3
 
4
+ # 載入模型和分詞器
 
 
 
 
 
 
5
  model_name = "taide/Llama3-TAIDE-LX-8B-Chat-Alpha1"
6
+ model = AutoModelForCausalLM.from_pretrained(model_name)
7
  tokenizer = AutoTokenizer.from_pretrained(model_name)
 
 
 
 
 
 
 
8
 
9
+ def chat_function(input_text):
10
+ inputs = tokenizer(input_text, return_tensors="pt")
11
+ output = model.generate(**inputs)
12
+ response = tokenizer.decode(output[0], skip_special_tokens=True)
13
+ return response
 
 
 
14
 
15
+ # 創建 Gradio 介面
16
+ iface = gr.Interface(fn=chat_function, inputs="text", outputs="text")
17
+ iface.launch()