gildead commited on
Commit
0cae964
·
verified ·
1 Parent(s): 3f7b349

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +8 -8
app.py CHANGED
@@ -1,21 +1,21 @@
1
  from huggingface_hub import InferenceClient
2
  import gradio as gr
3
- from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline, AutoModel, BitsAndBytesConfig
4
  from peft import PeftModel
5
  import torch
6
  # load base model
7
  base_model = "mistralai/Mistral-7B-Instruct-v0.1"
8
 
9
- bnb_config = BitsAndBytesConfig(
10
- load_in_4bit= True,
11
- bnb_4bit_quant_type= "nf4",
12
- bnb_4bit_compute_dtype= torch.bfloat16,
13
- bnb_4bit_use_double_quant= False,
14
- )
15
 
16
  model = AutoModelForCausalLM.from_pretrained(
17
  base_model,
18
- quantization_config=bnb_config,
19
  torch_dtype=torch.bfloat16,
20
  device_map="auto",
21
  trust_remote_code=True,
 
1
  from huggingface_hub import InferenceClient
2
  import gradio as gr
3
+ from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline, AutoModel
4
  from peft import PeftModel
5
  import torch
6
  # load base model
7
  base_model = "mistralai/Mistral-7B-Instruct-v0.1"
8
 
9
+ #bnb_config = BitsAndBytesConfig(
10
+ # load_in_4bit= True,
11
+ # bnb_4bit_quant_type= "nf4",
12
+ # bnb_4bit_compute_dtype= torch.bfloat16,
13
+ # bnb_4bit_use_double_quant= False,
14
+ #)
15
 
16
  model = AutoModelForCausalLM.from_pretrained(
17
  base_model,
18
+ # quantization_config=bnb_config,
19
  torch_dtype=torch.bfloat16,
20
  device_map="auto",
21
  trust_remote_code=True,