r1208 commited on
Commit
320473d
·
verified ·
1 Parent(s): bfe1ea4

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +5 -4
app.py CHANGED
@@ -18,14 +18,15 @@ from peft import AutoPeftModelForCausalLM
18
  from transformers import AutoTokenizer
19
  import torch
20
 
21
- attn_implementation = None
22
- if USE_FLASH_ATTENTION:
23
- attn_implementation="flash_attention_2"
 
24
 
25
 
26
  model_id = "r1208/c4ai-command-r-v01-4bit_32r"
27
 
28
- model = AutoPeftModelForCausalLM.from_pretrained(model_id, torch_dtype=torch.bfloat16, attn_implementation=attn_implementation,)
29
  tokenizer = AutoTokenizer.from_pretrained(model_id)
30
 
31
  tokenizer_with_prefix_space = AutoTokenizer.from_pretrained(model_id, add_prefix_space=True)
 
18
  from transformers import AutoTokenizer
19
  import torch
20
 
21
+ # attn_implementation = None
22
+ # USE_FLASH_ATTENTION = False
23
+ # if USE_FLASH_ATTENTION:
24
+ # attn_implementation="flash_attention_2"
25
 
26
 
27
  model_id = "r1208/c4ai-command-r-v01-4bit_32r"
28
 
29
+ model = AutoPeftModelForCausalLM.from_pretrained(model_id, torch_dtype=torch.bfloat16)
30
  tokenizer = AutoTokenizer.from_pretrained(model_id)
31
 
32
  tokenizer_with_prefix_space = AutoTokenizer.from_pretrained(model_id, add_prefix_space=True)