MekkCyber commited on
Commit
3d2f5ba
·
1 Parent(s): 1e73d78

try with cuda

Browse files
Files changed (1) hide show
  1. app.py +1 -1
app.py CHANGED
@@ -66,7 +66,7 @@ def quantize_model(model_name, quantization_type, group_size=128, auth_token=Non
66
  model = AutoModelForCausalLM.from_pretrained(model_name, device_map="cuda", torch_dtype=torch.bfloat16, quantization_config=quantization_config, use_auth_token=auth_token.token)
67
  else :
68
  quantization_config = TorchAoConfig(quantization_type)
69
- model = AutoModelForCausalLM.from_pretrained(model_name, device_map="cpu", torch_dtype=torch.bfloat16, quantization_config=quantization_config, use_auth_token=auth_token.token)
70
 
71
  return model
72
 
 
66
  model = AutoModelForCausalLM.from_pretrained(model_name, device_map="cuda", torch_dtype=torch.bfloat16, quantization_config=quantization_config, use_auth_token=auth_token.token)
67
  else :
68
  quantization_config = TorchAoConfig(quantization_type)
69
+ model = AutoModelForCausalLM.from_pretrained(model_name, device_map="cuda", torch_dtype=torch.bfloat16, quantization_config=quantization_config, use_auth_token=auth_token.token)
70
 
71
  return model
72