GGmorello commited on
Commit
e3edf18
·
verified ·
1 Parent(s): 7140313

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +11 -5
app.py CHANGED
@@ -5,15 +5,21 @@ import torch
5
  from peft import PeftConfig, PeftModel
6
  from transformers import LlamaForCausalLM, AutoTokenizer, BitsAndBytesConfig
7
 
8
- # config = PeftConfig.from_pretrained("GGmorello/FLAMES")
9
  model = LlamaForCausalLM.from_pretrained(
10
- "GGmorello/FLAMES-100k",
 
 
 
 
 
 
11
  )
12
- # model = PeftModel.from_pretrained(model, "GGmorello/FLAMES-100k")
13
 
14
 
15
  MAX_SEQ_LEN = 4096
16
- tokenizer = AutoTokenizer.from_pretrained("codellama/CodeLlama-7b-hf")
17
  model.config.pad_token = tokenizer.pad_token = tokenizer.unk_token
18
 
19
 
@@ -32,4 +38,4 @@ def predict(text):
32
 
33
 
34
  demo = gr.Interface(fn=predict, inputs=gr.Text(), outputs=gr.Text())
35
- demo.launch()
 
5
  from peft import PeftConfig, PeftModel
6
  from transformers import LlamaForCausalLM, AutoTokenizer, BitsAndBytesConfig
7
 
8
+ config = PeftConfig.from_pretrained("GGmorello/FLAMES-20k")
9
  model = LlamaForCausalLM.from_pretrained(
10
+ config.base_model_name_or_path,
11
+ quantization_config=BitsAndBytesConfig(
12
+ load_in_4bit=True,
13
+ bnb_4bit_quant_type="nf4",
14
+ bnb_4bit_use_double_quant=True,
15
+ bnb_4bit_compute_dtype=torch.bfloat16,
16
+ ),
17
  )
18
+ model = PeftModel.from_pretrained(model, "GGmorello/FLAMES-20k")
19
 
20
 
21
  MAX_SEQ_LEN = 4096
22
+ tokenizer = AutoTokenizer.from_pretrained(config.base_model_name_or_path)
23
  model.config.pad_token = tokenizer.pad_token = tokenizer.unk_token
24
 
25
 
 
38
 
39
 
40
  demo = gr.Interface(fn=predict, inputs=gr.Text(), outputs=gr.Text())
41
+ demo.launch()