sara-nabhani commited on
Commit
e6ab595
·
1 Parent(s): 6f99f7a

Training in progress, step 1600

Browse files
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:976cd361702845d04ad65280156780ed8c1b0a9629274afdd72c13e1ad90f179
3
  size 2843370360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86d6b276cef5daa276527a3756567b7ae93219eb7baa714fbe0fd9959cf62b6d
3
  size 2843370360
last-checkpoint/pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a8e602de7d5661528acdeede64816370b29c93a649719dd35411b397bef90abe
3
  size 1421660981
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e72fe7d37409a3a5152899e9d418267d6c142e81d29311593784f251158fc777
3
  size 1421660981
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e069df12ba9b4de739400b0d7bec09598652ee49792842bf99a2a5652a33bd26
3
- size 14575
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2dfca1e89ec698dcc09dabb346705ad9e84f475664620c7c4ad5a001b381ad43
3
+ size 14639
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b4735c8454ebaaf0f7e5072d8f53d7dcffa20fd292c672caa239ebbcbcd191bb
3
  size 627
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:11baa14531d50c3f9f69ec4ac3bd5c6ff50c2ce50ac93b36482779e67e2514b6
3
  size 627
last-checkpoint/trainer_state.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "best_metric": 0.3662892905615983,
3
- "best_model_checkpoint": "/home2/s5432073/language-tech-project/results/ltp-roberta-large-defaultltp-roberta-large-default-2/checkpoint-1200",
4
- "epoch": 8.284023668639053,
5
- "global_step": 1400,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -265,11 +265,48 @@
265
  "eval_samples_per_second": 605.037,
266
  "eval_steps_per_second": 19.147,
267
  "step": 1400
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
268
  }
269
  ],
270
  "max_steps": 1690,
271
  "num_train_epochs": 10,
272
- "total_flos": 1.045325614845444e+16,
273
  "trial_name": null,
274
  "trial_params": null
275
  }
 
1
  {
2
+ "best_metric": 0.37634159533224454,
3
+ "best_model_checkpoint": "/home2/s5432073/language-tech-project/results/ltp-roberta-large-defaultltp-roberta-large-default-2/checkpoint-1600",
4
+ "epoch": 9.467455621301776,
5
+ "global_step": 1600,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
265
  "eval_samples_per_second": 605.037,
266
  "eval_steps_per_second": 19.147,
267
  "step": 1400
268
+ },
269
+ {
270
+ "epoch": 9.47,
271
+ "learning_rate": 5.607476635514019e-07,
272
+ "loss": 0.2672,
273
+ "step": 1600
274
+ },
275
+ {
276
+ "epoch": 9.47,
277
+ "eval_f1": 0.37634159533224454,
278
+ "eval_f1_all": [
279
+ 0.5393794749403341,
280
+ 0.5859085290482077,
281
+ 0.06896551724137931,
282
+ 0.15384615384615385,
283
+ 0.6564885496183206,
284
+ 0.12972972972972974,
285
+ 0.38000000000000006,
286
+ 0.0,
287
+ 0.7552715654952076,
288
+ 0.6600441501103753,
289
+ 0.41463414634146345,
290
+ 0.5263157894736842,
291
+ 0.0,
292
+ 0.015503875968992248,
293
+ 0.5658263305322129,
294
+ 0.10596026490066227,
295
+ 0.6955848179705655,
296
+ 0.7085201793721972,
297
+ 0.13076923076923078,
298
+ 0.4340836012861736
299
+ ],
300
+ "eval_loss": 0.3051939308643341,
301
+ "eval_runtime": 3.2008,
302
+ "eval_samples_per_second": 592.358,
303
+ "eval_steps_per_second": 18.746,
304
+ "step": 1600
305
  }
306
  ],
307
  "max_steps": 1690,
308
  "num_train_epochs": 10,
309
+ "total_flos": 1.1949029604729552e+16,
310
  "trial_name": null,
311
  "trial_params": null
312
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a8e602de7d5661528acdeede64816370b29c93a649719dd35411b397bef90abe
3
  size 1421660981
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e72fe7d37409a3a5152899e9d418267d6c142e81d29311593784f251158fc777
3
  size 1421660981