franfj commited on
Commit
6413702
·
verified ·
1 Parent(s): 0ec1bc1

Upload folder using huggingface_hub

Browse files
checkpoint-1413/config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "distilbert/distilbert-base-multilingual-cased",
3
+ "activation": "gelu",
4
+ "architectures": [
5
+ "DistilBertForSequenceClassification"
6
+ ],
7
+ "attention_dropout": 0.1,
8
+ "dim": 768,
9
+ "dropout": 0.1,
10
+ "hidden_dim": 3072,
11
+ "initializer_range": 0.02,
12
+ "max_position_embeddings": 512,
13
+ "model_type": "distilbert",
14
+ "n_heads": 12,
15
+ "n_layers": 6,
16
+ "output_past": true,
17
+ "pad_token_id": 0,
18
+ "problem_type": "single_label_classification",
19
+ "qa_dropout": 0.1,
20
+ "seq_classif_dropout": 0.2,
21
+ "sinusoidal_pos_embds": false,
22
+ "tie_weights_": true,
23
+ "torch_dtype": "float32",
24
+ "transformers_version": "4.48.3",
25
+ "vocab_size": 119547
26
+ }
checkpoint-1413/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70a0ae2cb4bac7c2d330dd6fd0b668ebf330686cfec0695eb102b5f10aa8760e
3
+ size 541317368
checkpoint-1413/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92536fd35dc61548236a192091bdfb7c5bd1b60b43cf5a2d6e809e2ff9c97f9f
3
+ size 1082696890
checkpoint-1413/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ecbdb486f45d21770a6a8c506d67ae992ed5d201c96375ba0c67e7a235bf64dc
3
+ size 14244
checkpoint-1413/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a08c30c7856ff414c1e25b4833f35b648e6e81f982b3b1e22b09d20c7b0f5ca0
3
+ size 1064
checkpoint-1413/special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
checkpoint-1413/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-1413/tokenizer_config.json ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": false,
45
+ "cls_token": "[CLS]",
46
+ "do_lower_case": false,
47
+ "extra_special_tokens": {},
48
+ "mask_token": "[MASK]",
49
+ "model_max_length": 512,
50
+ "pad_token": "[PAD]",
51
+ "sep_token": "[SEP]",
52
+ "strip_accents": null,
53
+ "tokenize_chinese_chars": true,
54
+ "tokenizer_class": "DistilBertTokenizer",
55
+ "unk_token": "[UNK]"
56
+ }
checkpoint-1413/trainer_state.json ADDED
@@ -0,0 +1,259 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 3.0,
5
+ "eval_steps": 500,
6
+ "global_step": 1413,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.10615711252653928,
13
+ "grad_norm": 0.822494387626648,
14
+ "learning_rate": 4.823071479122435e-05,
15
+ "loss": 0.6599,
16
+ "step": 50
17
+ },
18
+ {
19
+ "epoch": 0.21231422505307856,
20
+ "grad_norm": 0.9401699304580688,
21
+ "learning_rate": 4.6461429582448696e-05,
22
+ "loss": 0.6465,
23
+ "step": 100
24
+ },
25
+ {
26
+ "epoch": 0.3184713375796178,
27
+ "grad_norm": 0.5120589733123779,
28
+ "learning_rate": 4.469214437367304e-05,
29
+ "loss": 0.6396,
30
+ "step": 150
31
+ },
32
+ {
33
+ "epoch": 0.42462845010615713,
34
+ "grad_norm": 0.8457750678062439,
35
+ "learning_rate": 4.292285916489738e-05,
36
+ "loss": 0.6295,
37
+ "step": 200
38
+ },
39
+ {
40
+ "epoch": 0.5307855626326964,
41
+ "grad_norm": 0.7705994248390198,
42
+ "learning_rate": 4.115357395612173e-05,
43
+ "loss": 0.6223,
44
+ "step": 250
45
+ },
46
+ {
47
+ "epoch": 0.6369426751592356,
48
+ "grad_norm": 0.962993323802948,
49
+ "learning_rate": 3.9384288747346076e-05,
50
+ "loss": 0.6202,
51
+ "step": 300
52
+ },
53
+ {
54
+ "epoch": 0.7430997876857749,
55
+ "grad_norm": 0.8745971322059631,
56
+ "learning_rate": 3.761500353857042e-05,
57
+ "loss": 0.6179,
58
+ "step": 350
59
+ },
60
+ {
61
+ "epoch": 0.8492569002123143,
62
+ "grad_norm": 0.9388220310211182,
63
+ "learning_rate": 3.584571832979476e-05,
64
+ "loss": 0.6164,
65
+ "step": 400
66
+ },
67
+ {
68
+ "epoch": 0.9554140127388535,
69
+ "grad_norm": 0.643945038318634,
70
+ "learning_rate": 3.407643312101911e-05,
71
+ "loss": 0.6156,
72
+ "step": 450
73
+ },
74
+ {
75
+ "epoch": 1.0,
76
+ "eval_accuracy": 0.6478773375875211,
77
+ "eval_f1": 0.6935123042505593,
78
+ "eval_loss": 0.6068633794784546,
79
+ "eval_runtime": 29.9559,
80
+ "eval_samples_per_second": 376.654,
81
+ "eval_steps_per_second": 3.939,
82
+ "step": 471
83
+ },
84
+ {
85
+ "epoch": 1.0615711252653928,
86
+ "grad_norm": 0.9551882743835449,
87
+ "learning_rate": 3.2307147912243455e-05,
88
+ "loss": 0.5978,
89
+ "step": 500
90
+ },
91
+ {
92
+ "epoch": 1.167728237791932,
93
+ "grad_norm": 0.8925243020057678,
94
+ "learning_rate": 3.05378627034678e-05,
95
+ "loss": 0.5925,
96
+ "step": 550
97
+ },
98
+ {
99
+ "epoch": 1.2738853503184713,
100
+ "grad_norm": 1.4399162530899048,
101
+ "learning_rate": 2.8768577494692145e-05,
102
+ "loss": 0.5946,
103
+ "step": 600
104
+ },
105
+ {
106
+ "epoch": 1.3800424628450108,
107
+ "grad_norm": 0.8456072807312012,
108
+ "learning_rate": 2.699929228591649e-05,
109
+ "loss": 0.5872,
110
+ "step": 650
111
+ },
112
+ {
113
+ "epoch": 1.48619957537155,
114
+ "grad_norm": 0.8160797357559204,
115
+ "learning_rate": 2.523000707714084e-05,
116
+ "loss": 0.5868,
117
+ "step": 700
118
+ },
119
+ {
120
+ "epoch": 1.5923566878980893,
121
+ "grad_norm": 1.2058725357055664,
122
+ "learning_rate": 2.3460721868365182e-05,
123
+ "loss": 0.5844,
124
+ "step": 750
125
+ },
126
+ {
127
+ "epoch": 1.6985138004246285,
128
+ "grad_norm": 0.8997677564620972,
129
+ "learning_rate": 2.1691436659589525e-05,
130
+ "loss": 0.5942,
131
+ "step": 800
132
+ },
133
+ {
134
+ "epoch": 1.8046709129511678,
135
+ "grad_norm": 0.6996552348136902,
136
+ "learning_rate": 1.9922151450813872e-05,
137
+ "loss": 0.5934,
138
+ "step": 850
139
+ },
140
+ {
141
+ "epoch": 1.910828025477707,
142
+ "grad_norm": 1.2173207998275757,
143
+ "learning_rate": 1.8152866242038215e-05,
144
+ "loss": 0.5935,
145
+ "step": 900
146
+ },
147
+ {
148
+ "epoch": 2.0,
149
+ "eval_accuracy": 0.654701763715324,
150
+ "eval_f1": 0.690252822388297,
151
+ "eval_loss": 0.6040562987327576,
152
+ "eval_runtime": 29.6452,
153
+ "eval_samples_per_second": 380.601,
154
+ "eval_steps_per_second": 3.98,
155
+ "step": 942
156
+ },
157
+ {
158
+ "epoch": 2.0169851380042463,
159
+ "grad_norm": 0.7898246645927429,
160
+ "learning_rate": 1.6383581033262562e-05,
161
+ "loss": 0.5917,
162
+ "step": 950
163
+ },
164
+ {
165
+ "epoch": 2.1231422505307855,
166
+ "grad_norm": 1.168487787246704,
167
+ "learning_rate": 1.461429582448691e-05,
168
+ "loss": 0.5637,
169
+ "step": 1000
170
+ },
171
+ {
172
+ "epoch": 2.229299363057325,
173
+ "grad_norm": 1.5715628862380981,
174
+ "learning_rate": 1.2845010615711253e-05,
175
+ "loss": 0.5563,
176
+ "step": 1050
177
+ },
178
+ {
179
+ "epoch": 2.335456475583864,
180
+ "grad_norm": 1.6441806554794312,
181
+ "learning_rate": 1.1075725406935598e-05,
182
+ "loss": 0.5399,
183
+ "step": 1100
184
+ },
185
+ {
186
+ "epoch": 2.4416135881104033,
187
+ "grad_norm": 1.376325011253357,
188
+ "learning_rate": 9.306440198159943e-06,
189
+ "loss": 0.5608,
190
+ "step": 1150
191
+ },
192
+ {
193
+ "epoch": 2.5477707006369426,
194
+ "grad_norm": 1.4360628128051758,
195
+ "learning_rate": 7.537154989384289e-06,
196
+ "loss": 0.5552,
197
+ "step": 1200
198
+ },
199
+ {
200
+ "epoch": 2.653927813163482,
201
+ "grad_norm": 1.4048738479614258,
202
+ "learning_rate": 5.767869780608635e-06,
203
+ "loss": 0.55,
204
+ "step": 1250
205
+ },
206
+ {
207
+ "epoch": 2.7600849256900215,
208
+ "grad_norm": 1.1524888277053833,
209
+ "learning_rate": 3.99858457183298e-06,
210
+ "loss": 0.5521,
211
+ "step": 1300
212
+ },
213
+ {
214
+ "epoch": 2.8662420382165603,
215
+ "grad_norm": 1.2835032939910889,
216
+ "learning_rate": 2.229299363057325e-06,
217
+ "loss": 0.5704,
218
+ "step": 1350
219
+ },
220
+ {
221
+ "epoch": 2.9723991507431,
222
+ "grad_norm": 1.3388535976409912,
223
+ "learning_rate": 4.6001415428167025e-07,
224
+ "loss": 0.5608,
225
+ "step": 1400
226
+ },
227
+ {
228
+ "epoch": 3.0,
229
+ "eval_accuracy": 0.6450412124434991,
230
+ "eval_f1": 0.6825208085612366,
231
+ "eval_loss": 0.6194082498550415,
232
+ "eval_runtime": 29.5475,
233
+ "eval_samples_per_second": 381.859,
234
+ "eval_steps_per_second": 3.994,
235
+ "step": 1413
236
+ }
237
+ ],
238
+ "logging_steps": 50,
239
+ "max_steps": 1413,
240
+ "num_input_tokens_seen": 0,
241
+ "num_train_epochs": 3,
242
+ "save_steps": 500,
243
+ "stateful_callbacks": {
244
+ "TrainerControl": {
245
+ "args": {
246
+ "should_epoch_stop": false,
247
+ "should_evaluate": false,
248
+ "should_log": false,
249
+ "should_save": true,
250
+ "should_training_stop": true
251
+ },
252
+ "attributes": {}
253
+ }
254
+ },
255
+ "total_flos": 1.7933568897447936e+16,
256
+ "train_batch_size": 96,
257
+ "trial_name": null,
258
+ "trial_params": null
259
+ }
checkpoint-1413/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:daf31243f710cadd7c3481998fafb12812a09847d81944a77ae6e616b24b8cfd
3
+ size 5304
checkpoint-1413/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "distilbert/distilbert-base-multilingual-cased",
3
+ "activation": "gelu",
4
+ "architectures": [
5
+ "DistilBertForSequenceClassification"
6
+ ],
7
+ "attention_dropout": 0.1,
8
+ "dim": 768,
9
+ "dropout": 0.1,
10
+ "hidden_dim": 3072,
11
+ "initializer_range": 0.02,
12
+ "max_position_embeddings": 512,
13
+ "model_type": "distilbert",
14
+ "n_heads": 12,
15
+ "n_layers": 6,
16
+ "output_past": true,
17
+ "pad_token_id": 0,
18
+ "problem_type": "single_label_classification",
19
+ "qa_dropout": 0.1,
20
+ "seq_classif_dropout": 0.2,
21
+ "sinusoidal_pos_embds": false,
22
+ "tie_weights_": true,
23
+ "torch_dtype": "float32",
24
+ "transformers_version": "4.48.3",
25
+ "vocab_size": 119547
26
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70a0ae2cb4bac7c2d330dd6fd0b668ebf330686cfec0695eb102b5f10aa8760e
3
+ size 541317368
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": false,
45
+ "cls_token": "[CLS]",
46
+ "do_lower_case": false,
47
+ "extra_special_tokens": {},
48
+ "mask_token": "[MASK]",
49
+ "model_max_length": 512,
50
+ "pad_token": "[PAD]",
51
+ "sep_token": "[SEP]",
52
+ "strip_accents": null,
53
+ "tokenize_chinese_chars": true,
54
+ "tokenizer_class": "DistilBertTokenizer",
55
+ "unk_token": "[UNK]"
56
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff