daresearch commited on
Commit
582704c
·
verified ·
1 Parent(s): ae8034c

Create app.py

Browse files
Files changed (1) hide show
  1. app.py +31 -0
app.py ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import gradio as gr
2
+ import torch
3
+ from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
4
+ from peft import PeftModel
5
+
6
+ BASE_MODEL = "meta-llama/Llama-3.3-70B-Instruct"
7
+ ADAPTER = "daresearch/Llama-3.3-70B-ft-exec-roles"
8
+
9
+ # Load the base model
10
+ base_model = AutoModelForCausalLM.from_pretrained(BASE_MODEL, torch_dtype=torch.float16, device_map="auto")
11
+ tokenizer = AutoTokenizer.from_pretrained(BASE_MODEL)
12
+
13
+ # Load the adapter
14
+ model = PeftModel.from_pretrained(base_model, ADAPTER, device_map="auto")
15
+
16
+ # Create a text generation pipeline
17
+ pipe = pipeline("text-generation", model=model, tokenizer=tokenizer)
18
+
19
+ def infer(prompt):
20
+ outputs = pipe(prompt, max_length=128)
21
+ return outputs[0]["generated_text"]
22
+
23
+ iface = gr.Interface(
24
+ fn=infer,
25
+ inputs="text",
26
+ outputs="text",
27
+ title="LoRA-Enhanced Model"
28
+ )
29
+
30
+ if __name__ == "__main__":
31
+ iface.launch()