Python commited on
Commit
ad1b0ec
·
0 Parent(s):
.gitattributes ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ pytorch_model.bin filter=lfs diff=lfs merge=lfs -text
2
+ optimizer.pt filter=lfs diff=lfs merge=lfs -text
3
+ scheduler.pt filter=lfs diff=lfs merge=lfs -text
4
+ spiece.model filter=lfs diff=lfs merge=lfs -text
5
+ training_args.bin filter=lfs diff=lfs merge=lfs -text
config.json ADDED
@@ -0,0 +1,246 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "Python/ACROSS-m2o-eng-base",
3
+ "architectures": [
4
+ "MT5ForConditionalGeneration"
5
+ ],
6
+ "consistency_loss_weight": 2.0,
7
+ "consistency_temperature": 1.0,
8
+ "contrastive": true,
9
+ "contrastive_mini_batch_size": 4,
10
+ "cross_alignment_loss_weight": 1.0,
11
+ "d_ff": 2048,
12
+ "d_kv": 64,
13
+ "d_model": 768,
14
+ "decoder_start_token_id": 0,
15
+ "dropout_rate": 0.1,
16
+ "eos_token_id": 1,
17
+ "feature_pooling_method": "mean_pooling",
18
+ "feed_forward_proj": "gated-gelu",
19
+ "freeze_decoder": false,
20
+ "freeze_mono_encoder": true,
21
+ "initializer_factor": 1.0,
22
+ "is_encoder_decoder": true,
23
+ "layer_norm_epsilon": 1e-06,
24
+ "length_penalty": 0.6,
25
+ "max_length": 84,
26
+ "model_type": "mt5",
27
+ "num_beams": 4,
28
+ "num_decoder_layers": 12,
29
+ "num_heads": 12,
30
+ "num_layers": 12,
31
+ "output_past": true,
32
+ "pad_token_id": 0,
33
+ "queue_size": 4096,
34
+ "relative_attention_num_buckets": 32,
35
+ "similarity_method": "cosine",
36
+ "t5_path_or_url": "output/eng2eng_base_baseline",
37
+ "task_specific_params": {
38
+ "langid_map": {
39
+ "amharic": [
40
+ 39,
41
+ "\u2581<extra_id_64>"
42
+ ],
43
+ "arabic": [
44
+ 8,
45
+ "\u2581<extra_id_95>"
46
+ ],
47
+ "azerbaijani": [
48
+ 11,
49
+ "\u2581<extra_id_92>"
50
+ ],
51
+ "bengali": [
52
+ 46,
53
+ "\u2581<extra_id_57>"
54
+ ],
55
+ "burmese": [
56
+ 37,
57
+ "\u2581<extra_id_66>"
58
+ ],
59
+ "chinese_simplified": [
60
+ 44,
61
+ "\u2581<extra_id_59>"
62
+ ],
63
+ "chinese_traditional": [
64
+ 48,
65
+ "\u2581<extra_id_55>"
66
+ ],
67
+ "czech": [
68
+ 3,
69
+ "\u2581<extra_id_54>"
70
+ ],
71
+ "dutch": [
72
+ 1,
73
+ "\u2581<extra_id_52>"
74
+ ],
75
+ "english": [
76
+ 34,
77
+ "\u2581<extra_id_69>"
78
+ ],
79
+ "french": [
80
+ 14,
81
+ "\u2581<extra_id_89>"
82
+ ],
83
+ "german": [
84
+ 0,
85
+ "\u2581<extra_id_51>"
86
+ ],
87
+ "gujarati": [
88
+ 31,
89
+ "\u2581<extra_id_72>"
90
+ ],
91
+ "hausa": [
92
+ 47,
93
+ "\u2581<extra_id_56>"
94
+ ],
95
+ "hindi": [
96
+ 25,
97
+ "\u2581<extra_id_78>"
98
+ ],
99
+ "igbo": [
100
+ 13,
101
+ "\u2581<extra_id_90>"
102
+ ],
103
+ "indonesian": [
104
+ 5,
105
+ "\u2581<extra_id_98>"
106
+ ],
107
+ "italian": [
108
+ 2,
109
+ "\u2581<extra_id_53>"
110
+ ],
111
+ "japanese": [
112
+ 41,
113
+ "\u2581<extra_id_62>"
114
+ ],
115
+ "kirundi": [
116
+ 4,
117
+ "\u2581<extra_id_99>"
118
+ ],
119
+ "korean": [
120
+ 33,
121
+ "\u2581<extra_id_70>"
122
+ ],
123
+ "kyrgyz": [
124
+ 9,
125
+ "\u2581<extra_id_94>"
126
+ ],
127
+ "marathi": [
128
+ 17,
129
+ "\u2581<extra_id_86>"
130
+ ],
131
+ "nepali": [
132
+ 24,
133
+ "\u2581<extra_id_79>"
134
+ ],
135
+ "oromo": [
136
+ 45,
137
+ "\u2581<extra_id_58>"
138
+ ],
139
+ "pashto": [
140
+ 38,
141
+ "\u2581<extra_id_65>"
142
+ ],
143
+ "persian": [
144
+ 27,
145
+ "\u2581<extra_id_76>"
146
+ ],
147
+ "pidgin": [
148
+ 18,
149
+ "\u2581<extra_id_85>"
150
+ ],
151
+ "portuguese": [
152
+ 43,
153
+ "\u2581<extra_id_60>"
154
+ ],
155
+ "punjabi": [
156
+ 21,
157
+ "\u2581<extra_id_82>"
158
+ ],
159
+ "russian": [
160
+ 40,
161
+ "\u2581<extra_id_63>"
162
+ ],
163
+ "scottish_gaelic": [
164
+ 28,
165
+ "\u2581<extra_id_75>"
166
+ ],
167
+ "serbian_cyrillic": [
168
+ 32,
169
+ "\u2581<extra_id_71>"
170
+ ],
171
+ "serbian_latin": [
172
+ 15,
173
+ "\u2581<extra_id_88>"
174
+ ],
175
+ "sinhala": [
176
+ 35,
177
+ "\u2581<extra_id_68>"
178
+ ],
179
+ "somali": [
180
+ 23,
181
+ "\u2581<extra_id_80>"
182
+ ],
183
+ "spanish": [
184
+ 7,
185
+ "\u2581<extra_id_96>"
186
+ ],
187
+ "swahili": [
188
+ 22,
189
+ "\u2581<extra_id_81>"
190
+ ],
191
+ "tamil": [
192
+ 36,
193
+ "\u2581<extra_id_67>"
194
+ ],
195
+ "telugu": [
196
+ 26,
197
+ "\u2581<extra_id_77>"
198
+ ],
199
+ "thai": [
200
+ 10,
201
+ "\u2581<extra_id_93>"
202
+ ],
203
+ "tigrinya": [
204
+ 20,
205
+ "\u2581<extra_id_83>"
206
+ ],
207
+ "turkish": [
208
+ 19,
209
+ "\u2581<extra_id_84>"
210
+ ],
211
+ "ukrainian": [
212
+ 6,
213
+ "\u2581<extra_id_97>"
214
+ ],
215
+ "urdu": [
216
+ 42,
217
+ "\u2581<extra_id_61>"
218
+ ],
219
+ "uzbek": [
220
+ 12,
221
+ "\u2581<extra_id_91>"
222
+ ],
223
+ "vietnamese": [
224
+ 16,
225
+ "\u2581<extra_id_87>"
226
+ ],
227
+ "welsh": [
228
+ 30,
229
+ "\u2581<extra_id_73>"
230
+ ],
231
+ "yoruba": [
232
+ 29,
233
+ "\u2581<extra_id_74>"
234
+ ]
235
+ }
236
+ },
237
+ "temperature": 0.1,
238
+ "tie_word_embeddings": false,
239
+ "tokenizer_class": "T5Tokenizer",
240
+ "transformers_version": "4.10.0.dev0",
241
+ "use_cache": true,
242
+ "use_consistency_loss": true,
243
+ "use_cross_alignment_loss": true,
244
+ "use_queue_for_contrastive": true,
245
+ "vocab_size": 250112
246
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d7e13c64ad48a61fdf31c793117d9bb2c3d82d5aba578e708a6b53bd6c0ab91
3
+ size 2329703385
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>"}
spiece.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef78f86560d809067d12bac6c09f19a462cb3af3f54d2b8acbba26e1433125d6
3
+ size 4309802
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>", "extra_ids": 0, "additional_special_tokens": null, "special_tokens_map_file": "/home/patrick/.cache/torch/transformers/685ac0ca8568ec593a48b61b0a3c272beee9bc194a3c7241d15dcadb5f875e53.f76030f3ec1b96a8199b2593390c610e76ca8028ef3d24680000619ffb646276", "tokenizer_file": null, "name_or_path": "./"}