sara-nabhani
commited on
Commit
·
6f99f7a
1
Parent(s):
d705747
Training in progress, step 1400
Browse files
last-checkpoint/optimizer.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 2843370360
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:976cd361702845d04ad65280156780ed8c1b0a9629274afdd72c13e1ad90f179
|
3 |
size 2843370360
|
last-checkpoint/pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1421660981
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a8e602de7d5661528acdeede64816370b29c93a649719dd35411b397bef90abe
|
3 |
size 1421660981
|
last-checkpoint/rng_state.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14575
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e069df12ba9b4de739400b0d7bec09598652ee49792842bf99a2a5652a33bd26
|
3 |
size 14575
|
last-checkpoint/scheduler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 627
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b4735c8454ebaaf0f7e5072d8f53d7dcffa20fd292c672caa239ebbcbcd191bb
|
3 |
size 627
|
last-checkpoint/trainer_state.json
CHANGED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"best_metric": 0.3662892905615983,
|
3 |
"best_model_checkpoint": "/home2/s5432073/language-tech-project/results/ltp-roberta-large-defaultltp-roberta-large-default-2/checkpoint-1200",
|
4 |
-
"epoch":
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -228,11 +228,48 @@
|
|
228 |
"eval_samples_per_second": 616.829,
|
229 |
"eval_steps_per_second": 19.52,
|
230 |
"step": 1200
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
231 |
}
|
232 |
],
|
233 |
"max_steps": 1690,
|
234 |
"num_train_epochs": 10,
|
235 |
-
"total_flos":
|
236 |
"trial_name": null,
|
237 |
"trial_params": null
|
238 |
}
|
|
|
1 |
{
|
2 |
"best_metric": 0.3662892905615983,
|
3 |
"best_model_checkpoint": "/home2/s5432073/language-tech-project/results/ltp-roberta-large-defaultltp-roberta-large-default-2/checkpoint-1200",
|
4 |
+
"epoch": 8.284023668639053,
|
5 |
+
"global_step": 1400,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
228 |
"eval_samples_per_second": 616.829,
|
229 |
"eval_steps_per_second": 19.52,
|
230 |
"step": 1200
|
231 |
+
},
|
232 |
+
{
|
233 |
+
"epoch": 8.28,
|
234 |
+
"learning_rate": 1.8068535825545173e-06,
|
235 |
+
"loss": 0.2721,
|
236 |
+
"step": 1400
|
237 |
+
},
|
238 |
+
{
|
239 |
+
"epoch": 8.28,
|
240 |
+
"eval_f1": 0.36379633008264506,
|
241 |
+
"eval_f1_all": [
|
242 |
+
0.5247524752475247,
|
243 |
+
0.5838509316770186,
|
244 |
+
0.028368794326241138,
|
245 |
+
0.07407407407407407,
|
246 |
+
0.6463654223968567,
|
247 |
+
0.13259668508287292,
|
248 |
+
0.35106382978723405,
|
249 |
+
0.0,
|
250 |
+
0.7543075941289088,
|
251 |
+
0.6308243727598566,
|
252 |
+
0.4,
|
253 |
+
0.5212620027434842,
|
254 |
+
0.0,
|
255 |
+
0.015503875968992248,
|
256 |
+
0.5748709122203098,
|
257 |
+
0.07612456747404846,
|
258 |
+
0.6950959488272921,
|
259 |
+
0.6986899563318777,
|
260 |
+
0.14785992217898833,
|
261 |
+
0.42031523642732044
|
262 |
+
],
|
263 |
+
"eval_loss": 0.3058605194091797,
|
264 |
+
"eval_runtime": 3.1337,
|
265 |
+
"eval_samples_per_second": 605.037,
|
266 |
+
"eval_steps_per_second": 19.147,
|
267 |
+
"step": 1400
|
268 |
}
|
269 |
],
|
270 |
"max_steps": 1690,
|
271 |
"num_train_epochs": 10,
|
272 |
+
"total_flos": 1.045325614845444e+16,
|
273 |
"trial_name": null,
|
274 |
"trial_params": null
|
275 |
}
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1421660981
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a8e602de7d5661528acdeede64816370b29c93a649719dd35411b397bef90abe
|
3 |
size 1421660981
|