Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -35,17 +35,19 @@ translate_token_id = all_special_ids[-6]
|
|
35 |
|
36 |
def _preload_and_load_models():
|
37 |
global vicuna_tokenizer, vicuna_model
|
38 |
-
VICUNA_MODEL_NAME = "EleutherAI/gpt-neo-2.7B" # Or another model
|
|
|
|
|
39 |
vicuna_tokenizer = AutoTokenizer.from_pretrained(VICUNA_MODEL_NAME)
|
40 |
vicuna_model = AutoModelForCausalLM.from_pretrained(
|
41 |
VICUNA_MODEL_NAME,
|
42 |
torch_dtype=torch.float16,
|
43 |
-
|
44 |
-
)
|
45 |
|
46 |
_preload_and_load_models()
|
47 |
|
48 |
-
tts = Text2Speech.from_pretrained("espnet/kan-bayashi_ljspeech_vits")
|
49 |
|
50 |
@spaces.GPU(required=True)
|
51 |
def process_audio(microphone, state, task="transcribe"):
|
|
|
35 |
|
36 |
def _preload_and_load_models():
|
37 |
global vicuna_tokenizer, vicuna_model
|
38 |
+
#VICUNA_MODEL_NAME = "EleutherAI/gpt-neo-2.7B" # Or another model
|
39 |
+
#VICUNA_MODEL_NAME = "lmsys/vicuna-13b-v1.5" # Or another model
|
40 |
+
VICUNA_MODEL_NAME = "lmsys/vicuna-7b-v1.5" # Or another model
|
41 |
vicuna_tokenizer = AutoTokenizer.from_pretrained(VICUNA_MODEL_NAME)
|
42 |
vicuna_model = AutoModelForCausalLM.from_pretrained(
|
43 |
VICUNA_MODEL_NAME,
|
44 |
torch_dtype=torch.float16,
|
45 |
+
# device_map="auto", # or.to('cuda')
|
46 |
+
).to('cuda') # Explicitly move to CUDA after loading
|
47 |
|
48 |
_preload_and_load_models()
|
49 |
|
50 |
+
tts = Text2Speech.from_pretrained("espnet/kan-bayashi_ljspeech_vits").to('cuda')
|
51 |
|
52 |
@spaces.GPU(required=True)
|
53 |
def process_audio(microphone, state, task="transcribe"):
|