bachvudinh commited on
Commit
dbf9701
·
1 Parent(s): 78bce4d

add move model to device

Browse files
Files changed (1) hide show
  1. app.py +2 -5
app.py CHANGED
@@ -50,7 +50,7 @@ tts = TTSProcessor(device)
50
  use_8bit = False
51
  llm_path = "homebrewltd/Llama3.1-s-instruct-2024-08-19-epoch-3"
52
  tokenizer = AutoTokenizer.from_pretrained(llm_path)
53
- model_kwargs = {"device_map": "auto"}
54
  if use_8bit:
55
  model_kwargs["quantization_config"] = BitsAndBytesConfig(
56
  load_in_8bit=True,
@@ -59,10 +59,7 @@ if use_8bit:
59
  )
60
  else:
61
  model_kwargs["torch_dtype"] = torch.bfloat16
62
- model = AutoModelForCausalLM.from_pretrained(llm_path, **model_kwargs)
63
- pipe = pipeline("text-generation", model=model, tokenizer=tokenizer)
64
- tokenizer = pipe.tokenizer
65
- model = pipe.model
66
  # print(tokenizer.encode("<|sound_0001|>", add_special_tokens=False))# return the audio tensor
67
  # print(tokenizer.eos_token)
68
 
 
50
  use_8bit = False
51
  llm_path = "homebrewltd/Llama3.1-s-instruct-2024-08-19-epoch-3"
52
  tokenizer = AutoTokenizer.from_pretrained(llm_path)
53
+ model_kwargs = {}
54
  if use_8bit:
55
  model_kwargs["quantization_config"] = BitsAndBytesConfig(
56
  load_in_8bit=True,
 
59
  )
60
  else:
61
  model_kwargs["torch_dtype"] = torch.bfloat16
62
+ model = AutoModelForCausalLM.from_pretrained(llm_path, **model_kwargs).to(device)
 
 
 
63
  # print(tokenizer.encode("<|sound_0001|>", add_special_tokens=False))# return the audio tensor
64
  # print(tokenizer.eos_token)
65