rdlf commited on
Commit
052012f
·
verified ·
1 Parent(s): 7549d85

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +19 -10
app.py CHANGED
@@ -1,18 +1,27 @@
1
  from transformers import GPT2LMHeadModel, GPT2Tokenizer
 
 
2
  tokenizer = GPT2Tokenizer.from_pretrained("sberbank-ai/mGPT")
3
  model = GPT2LMHeadModel.from_pretrained("sberbank-ai/mGPT")
4
 
 
5
  text = "Александр Сергеевич Пушкин родился в "
6
- input_ids = tokenizer.encode(text, return_tensors="pt").cuda(device)
 
 
 
 
7
  out = model.generate(
8
- input_ids,
9
- min_length=100,
10
- max_length=100,
11
- eos_token_id=5,
12
- pad_token=1,
13
- top_k=10,
14
- top_p=0.0,
15
- no_repeat_ngram_size=5
16
  )
 
 
17
  generated_text = list(map(tokenizer.decode, out))[0]
18
- print(generated_text)
 
1
  from transformers import GPT2LMHeadModel, GPT2Tokenizer
2
+
3
+ # Load the pre-trained model and tokenizer
4
  tokenizer = GPT2Tokenizer.from_pretrained("sberbank-ai/mGPT")
5
  model = GPT2LMHeadModel.from_pretrained("sberbank-ai/mGPT")
6
 
7
+ # Input text
8
  text = "Александр Сергеевич Пушкин родился в "
9
+
10
+ # Encode the input text
11
+ input_ids = tokenizer.encode(text, return_tensors="pt")
12
+
13
+ # Generate text
14
  out = model.generate(
15
+ input_ids,
16
+ min_length=100,
17
+ max_length=100,
18
+ eos_token_id=5,
19
+ pad_token_id=1,
20
+ top_k=10,
21
+ top_p=0.0,
22
+ no_repeat_ngram_size=5
23
  )
24
+
25
+ # Decode the generated output
26
  generated_text = list(map(tokenizer.decode, out))[0]
27
+ print(generated_text)