Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -19,8 +19,6 @@ TOKEN = os.environ['TOKEN']
|
|
19 |
# Translation function for Nano and Large models
|
20 |
@spaces.GPU
|
21 |
def translate_nano_large(text, model_path):
|
22 |
-
device = "cuda:0" if torch.cuda.is_available() else "cpu"
|
23 |
-
print(f'[INFO] Using device: {device}')
|
24 |
translator = pipeline("translation", model=model_path, token=TOKEN)
|
25 |
translated = translator(
|
26 |
text,
|
@@ -32,7 +30,6 @@ def translate_nano_large(text, model_path):
|
|
32 |
pad_token_id=translator.tokenizer.pad_token_id,
|
33 |
bos_token_id=translator.tokenizer.bos_token_id,
|
34 |
eos_token_id=translator.tokenizer.eos_token_id,
|
35 |
-
device=device,
|
36 |
)
|
37 |
return translated[0]["translation_text"]
|
38 |
|
|
|
19 |
# Translation function for Nano and Large models
|
20 |
@spaces.GPU
|
21 |
def translate_nano_large(text, model_path):
|
|
|
|
|
22 |
translator = pipeline("translation", model=model_path, token=TOKEN)
|
23 |
translated = translator(
|
24 |
text,
|
|
|
30 |
pad_token_id=translator.tokenizer.pad_token_id,
|
31 |
bos_token_id=translator.tokenizer.bos_token_id,
|
32 |
eos_token_id=translator.tokenizer.eos_token_id,
|
|
|
33 |
)
|
34 |
return translated[0]["translation_text"]
|
35 |
|