File size: 901 Bytes
5dc48e3
e623ffc
b12d2a9
5dc48e3
5518a73
b12d2a9
 
5dc48e3
b12d2a9
 
 
5dc48e3
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
import streamlit as st
from transformers import MBartForConditionalGeneration, MBart50TokenizerFast
from transformers import AutoModelForMaskedLM, AutoTokenizer

model = MBartForConditionalGeneration.from_pretrained("facebook/mbart-large-50-one-to-many-mmt")
model = AutoModelForMaskedLM.from_pretrained("alabnii/jmedroberta-base-sentencepiece")
modek.eval()
tokenizer = MBart50TokenizerFast.from_pretrained("facebook/mbart-large-50-one-to-many-mmt", src_large="en_XX")
tokenizer = AutoTokenizer.from_pretrained("alabnii/jmedroberta-base-sentencepiece")

text = st.text_area('Enter the text:')

if text:
    model_inputs = tokenizer(text, return_tensors="pt")
    generated_tokens = model.generate(
        **model_inputs,
        forced_bos_token_id=tokenizer.lang_code_to_id["hi_IN"]
    )
    translation = tokenizer.batch_decode(generated_tokens, skip_special_tokens=True)
    st.json(translation)