Update configs.py
Browse files- configs.py +1 -1
configs.py
CHANGED
@@ -35,7 +35,7 @@ model_info = {
|
|
35 |
),
|
36 |
'CodeLLAMA 70B Instruct (Experimental)': dict(model_path='TheBloke/CodeLlama-70B-Instruct-GPTQ',
|
37 |
token=os.environ['hf_token'],
|
38 |
-
wait_with_hidden_states=True, dont_cuda=True, device_map='cuda', disable_exllama=True,
|
39 |
original_prompt_template='<s>{prompt}',
|
40 |
interpretation_prompt_template='<s>[INST] [X] [/INST] {prompt}',
|
41 |
layers_format=llama_layers_format
|
|
|
35 |
),
|
36 |
'CodeLLAMA 70B Instruct (Experimental)': dict(model_path='TheBloke/CodeLlama-70B-Instruct-GPTQ',
|
37 |
token=os.environ['hf_token'],
|
38 |
+
wait_with_hidden_states=True, dont_cuda=True, device_map='cuda', # disable_exllama=True,
|
39 |
original_prompt_template='<s>{prompt}',
|
40 |
interpretation_prompt_template='<s>[INST] [X] [/INST] {prompt}',
|
41 |
layers_format=llama_layers_format
|