Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -41,8 +41,8 @@ def generate_prompt(instruction, input=None):
|
|
41 |
|
42 |
### Response:"""
|
43 |
|
44 |
-
model= load_model(name = 's3nh/pythia-410m-70k-steps-self-instruct-polish')
|
45 |
-
tokenizer = load_tokenizer(name = 's3nh/pythia-410m-70k-steps-self-instruct-polish')
|
46 |
generation_config = create_generator()
|
47 |
|
48 |
|
@@ -71,7 +71,7 @@ def choose_model(name):
|
|
71 |
|
72 |
return load_model(name), load_tokenizer(name)
|
73 |
|
74 |
-
model, tokenizer = gr.Interface(choose_model, [gr.inputs.Dropdown(["s3nh/pythia-1.4b-deduped-16k-steps-self-instruct-polish", "s3nh/pythia-1.4b-deduped-16k-steps-self-instruct-polish"])
|
75 |
|
76 |
io = gr.Interface(
|
77 |
inference,
|
|
|
41 |
|
42 |
### Response:"""
|
43 |
|
44 |
+
# model= load_model(name = 's3nh/pythia-410m-70k-steps-self-instruct-polish')
|
45 |
+
# tokenizer = load_tokenizer(name = 's3nh/pythia-410m-70k-steps-self-instruct-polish')
|
46 |
generation_config = create_generator()
|
47 |
|
48 |
|
|
|
71 |
|
72 |
return load_model(name), load_tokenizer(name)
|
73 |
|
74 |
+
model, tokenizer = gr.Interface(choose_model, [gr.inputs.Dropdown(["s3nh/pythia-1.4b-deduped-16k-steps-self-instruct-polish", "s3nh/pythia-1.4b-deduped-16k-steps-self-instruct-polish"])])
|
75 |
|
76 |
io = gr.Interface(
|
77 |
inference,
|