Spaces:
Runtime error
Runtime error
vinayakdev
commited on
Commit
•
9bc33fc
1
Parent(s):
f769cbe
Ran locally
Browse files- generator.py +14 -13
generator.py
CHANGED
@@ -1,4 +1,3 @@
|
|
1 |
-
|
2 |
import transformers
|
3 |
from transformers import (
|
4 |
# Text2TextGenerationPipeline,
|
@@ -37,16 +36,16 @@ import streamlit as st
|
|
37 |
def load_model():
|
38 |
hfm = pickle.load(open('t5_model.sav','rb'))
|
39 |
hft = T5TokenizerFast.from_pretrained("t5-base")
|
40 |
-
# tok = att.from_pretrained("mrm8488/elect")
|
41 |
model = pickle.load(open('electra_model.sav','rb'))
|
|
|
42 |
# return hfm, hft,tok, model
|
43 |
-
return hfm, hft, model
|
44 |
|
45 |
-
hfmodel, hftokenizer, model = load_model()
|
46 |
|
47 |
def run_model(input_string, **generator_args):
|
48 |
generator_args = {
|
49 |
-
"
|
50 |
"num_beams": 4,
|
51 |
"length_penalty": 1.5,
|
52 |
"no_repeat_ngram_size": 2,
|
@@ -61,8 +60,8 @@ def run_model(input_string, **generator_args):
|
|
61 |
|
62 |
res = hfmodel.generate(input_ids=inputs['input_ids'], attention_mask=inputs['attention_mask'], **generator_args)
|
63 |
output = hftokenizer.decode(res[0], skip_special_tokens=True)
|
64 |
-
output =
|
65 |
-
output = [o.strip() for o in output[:-1]]
|
66 |
return output
|
67 |
|
68 |
|
@@ -127,17 +126,19 @@ def read_file(filepath_name):
|
|
127 |
|
128 |
def create_string_for_generator(context):
|
129 |
gen_list = gen_question(context)
|
130 |
-
return
|
131 |
|
132 |
def creator(context):
|
133 |
questions = create_string_for_generator(context)
|
|
|
134 |
pairs = []
|
135 |
for ques in questions:
|
136 |
-
|
137 |
-
|
138 |
pair = QA(ques,context)
|
|
|
139 |
pairs.append(pair)
|
140 |
return pairs
|
141 |
-
|
142 |
-
#
|
143 |
-
|
|
|
|
|
1 |
import transformers
|
2 |
from transformers import (
|
3 |
# Text2TextGenerationPipeline,
|
|
|
36 |
def load_model():
|
37 |
hfm = pickle.load(open('t5_model.sav','rb'))
|
38 |
hft = T5TokenizerFast.from_pretrained("t5-base")
|
|
|
39 |
model = pickle.load(open('electra_model.sav','rb'))
|
40 |
+
tok = et.from_pretrained("mrm8488/electra-small-finetuned-squadv2")
|
41 |
# return hfm, hft,tok, model
|
42 |
+
return hfm, hft,tok, model
|
43 |
|
44 |
+
hfmodel, hftokenizer,tok, model = load_model()
|
45 |
|
46 |
def run_model(input_string, **generator_args):
|
47 |
generator_args = {
|
48 |
+
"max_length": 256,
|
49 |
"num_beams": 4,
|
50 |
"length_penalty": 1.5,
|
51 |
"no_repeat_ngram_size": 2,
|
|
|
60 |
|
61 |
res = hfmodel.generate(input_ids=inputs['input_ids'], attention_mask=inputs['attention_mask'], **generator_args)
|
62 |
output = hftokenizer.decode(res[0], skip_special_tokens=True)
|
63 |
+
# output = output.split('</sep>')
|
64 |
+
# output = [o.strip() for o in output[:-1]]
|
65 |
return output
|
66 |
|
67 |
|
|
|
126 |
|
127 |
def create_string_for_generator(context):
|
128 |
gen_list = gen_question(context)
|
129 |
+
return gen_list
|
130 |
|
131 |
def creator(context):
|
132 |
questions = create_string_for_generator(context)
|
133 |
+
questions = questions.split('?')
|
134 |
pairs = []
|
135 |
for ques in questions:
|
136 |
+
if ques[-1] != '?':
|
137 |
+
ques = ques + '?'
|
138 |
pair = QA(ques,context)
|
139 |
+
print(pair)
|
140 |
pairs.append(pair)
|
141 |
return pairs
|
142 |
+
# creator(""""Hello, World!" program by Brian Kernighan (1978).
|
143 |
+
# A "Hello, World!" program is generally a computer program that ignores any input and outputs or displays a message similar to "Hello, World!". A small piece of code in most general-purpose programming languages, this program is used to illustrate a language's basic syntax. "Hello, World!" programs are often the first a student learns to write in a given language,[1] and they can also be used as a sanity check to ensure computer software intended to compile or run source code is correctly installed, and that its operator understands how to use it.
|
144 |
+
# """)
|