dar-tau commited on
Commit
655bead
1 Parent(s): c9e13f6

Update configs.py

Browse files
Files changed (1) hide show
  1. configs.py +1 -1
configs.py CHANGED
@@ -35,7 +35,7 @@ model_info = {
35
  ),
36
  'CodeLLAMA 70B Instruct (Experimental)': dict(model_path='TheBloke/CodeLlama-70B-Instruct-GPTQ',
37
  token=os.environ['hf_token'],
38
- wait_with_hidden_states=True, dont_cuda=True, device_map='cuda', disable_exllama=True,
39
  original_prompt_template='<s>{prompt}',
40
  interpretation_prompt_template='<s>[INST] [X] [/INST] {prompt}',
41
  layers_format=llama_layers_format
 
35
  ),
36
  'CodeLLAMA 70B Instruct (Experimental)': dict(model_path='TheBloke/CodeLlama-70B-Instruct-GPTQ',
37
  token=os.environ['hf_token'],
38
+ wait_with_hidden_states=True, dont_cuda=True, device_map='cuda', # disable_exllama=True,
39
  original_prompt_template='<s>{prompt}',
40
  interpretation_prompt_template='<s>[INST] [X] [/INST] {prompt}',
41
  layers_format=llama_layers_format