Upload model
Browse files- config.json +4 -8
- generation_config.json +1 -1
- pytorch_model.bin +2 -2
config.json
CHANGED
@@ -1,6 +1,5 @@
|
|
1 |
{
|
2 |
-
"
|
3 |
-
"_name_or_path": "./model_char_1_fix",
|
4 |
"architectures": [
|
5 |
"VisionEncoderDecoderModel"
|
6 |
],
|
@@ -71,7 +70,7 @@
|
|
71 |
"return_dict": true,
|
72 |
"return_dict_in_generate": false,
|
73 |
"scale_embedding": true,
|
74 |
-
"sep_token_id":
|
75 |
"suppress_tokens": null,
|
76 |
"task_specific_params": null,
|
77 |
"temperature": 1.0,
|
@@ -83,11 +82,10 @@
|
|
83 |
"top_p": 1.0,
|
84 |
"torch_dtype": null,
|
85 |
"torchscript": false,
|
86 |
-
"transformers_version": "4.26.0",
|
87 |
"typical_p": 1.0,
|
88 |
"use_bfloat16": false,
|
89 |
"use_cache": true,
|
90 |
-
"vocab_size":
|
91 |
},
|
92 |
"decoder_start_token_id": 56,
|
93 |
"encoder": {
|
@@ -180,7 +178,6 @@
|
|
180 |
"top_p": 1.0,
|
181 |
"torch_dtype": null,
|
182 |
"torchscript": false,
|
183 |
-
"transformers_version": "4.26.0",
|
184 |
"typical_p": 1.0,
|
185 |
"use_absolute_embeddings": false,
|
186 |
"use_bfloat16": false,
|
@@ -189,8 +186,7 @@
|
|
189 |
"is_encoder_decoder": true,
|
190 |
"model_type": "vision-encoder-decoder",
|
191 |
"pad_token_id": 55,
|
192 |
-
"sep_token_id": 75,
|
193 |
"tie_word_embeddings": false,
|
194 |
"torch_dtype": "float32",
|
195 |
-
"transformers_version":
|
196 |
}
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "model_char_1",
|
|
|
3 |
"architectures": [
|
4 |
"VisionEncoderDecoderModel"
|
5 |
],
|
|
|
70 |
"return_dict": true,
|
71 |
"return_dict_in_generate": false,
|
72 |
"scale_embedding": true,
|
73 |
+
"sep_token_id": null,
|
74 |
"suppress_tokens": null,
|
75 |
"task_specific_params": null,
|
76 |
"temperature": 1.0,
|
|
|
82 |
"top_p": 1.0,
|
83 |
"torch_dtype": null,
|
84 |
"torchscript": false,
|
|
|
85 |
"typical_p": 1.0,
|
86 |
"use_bfloat16": false,
|
87 |
"use_cache": true,
|
88 |
+
"vocab_size": 75
|
89 |
},
|
90 |
"decoder_start_token_id": 56,
|
91 |
"encoder": {
|
|
|
178 |
"top_p": 1.0,
|
179 |
"torch_dtype": null,
|
180 |
"torchscript": false,
|
|
|
181 |
"typical_p": 1.0,
|
182 |
"use_absolute_embeddings": false,
|
183 |
"use_bfloat16": false,
|
|
|
186 |
"is_encoder_decoder": true,
|
187 |
"model_type": "vision-encoder-decoder",
|
188 |
"pad_token_id": 55,
|
|
|
189 |
"tie_word_embeddings": false,
|
190 |
"torch_dtype": "float32",
|
191 |
+
"transformers_version": "4.34.0"
|
192 |
}
|
generation_config.json
CHANGED
@@ -6,5 +6,5 @@
|
|
6 |
"forced_eos_token_id": 2,
|
7 |
"max_length": 60,
|
8 |
"pad_token_id": 55,
|
9 |
-
"transformers_version": "4.
|
10 |
}
|
|
|
6 |
"forced_eos_token_id": 2,
|
7 |
"max_length": 60,
|
8 |
"pad_token_id": 55,
|
9 |
+
"transformers_version": "4.34.0"
|
10 |
}
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9ed5c1b23e6dbda25aa8cb9281ef71b511a301256660d71dd013e5adb57de685
|
3 |
+
size 573856985
|