sara-nabhani commited on
Commit
d705747
·
1 Parent(s): d732907

Training in progress, step 1200

Browse files
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:655b17b3d6f1e5bdfa8dcd055f9b36da92d2798e7db0e299ca3318174b1e06d0
3
  size 2843370360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:788b338bd5f82c9055c82c6efad563f37343bbedfdb77e083ddba75d5854efad
3
  size 2843370360
last-checkpoint/pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4bdfd57d1e70cd331c7da4da7e930f75fa6690bcbcb63665b2897fc97f39df12
3
  size 1421660981
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6388e9a479aa96e7783654f3cf7f877d8519fea0ecf267a47186c81df6aa2bc3
3
  size 1421660981
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:652d935d5ee7c5076bca6cc7ae23a8afdb8292d537b33dfe49fca52f345bb9c7
3
  size 14575
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1d48ce52911ba342a51be063c340c69db0186afab7796ff01fa7e0ee5b90b1c
3
  size 14575
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:68a6f395aca12d068a355806ad25a39cc1cf9ac8d7df16949ceeaa9f78f9ba22
3
  size 627
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c29fefa34f25d7add4e8494b8c29833a78eba1e513492e2ab6a5d3ed607e7c36
3
  size 627
last-checkpoint/trainer_state.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "best_metric": 0.3556641957175606,
3
- "best_model_checkpoint": "/home2/s5432073/language-tech-project/results/ltp-roberta-large-defaultltp-roberta-large-default-2/checkpoint-1000",
4
- "epoch": 5.9171597633136095,
5
- "global_step": 1000,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -191,11 +191,48 @@
191
  "eval_samples_per_second": 577.521,
192
  "eval_steps_per_second": 18.276,
193
  "step": 1000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
194
  }
195
  ],
196
  "max_steps": 1690,
197
  "num_train_epochs": 10,
198
- "total_flos": 7474403932394472.0,
199
  "trial_name": null,
200
  "trial_params": null
201
  }
 
1
  {
2
+ "best_metric": 0.3662892905615983,
3
+ "best_model_checkpoint": "/home2/s5432073/language-tech-project/results/ltp-roberta-large-defaultltp-roberta-large-default-2/checkpoint-1200",
4
+ "epoch": 7.100591715976331,
5
+ "global_step": 1200,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
191
  "eval_samples_per_second": 577.521,
192
  "eval_steps_per_second": 18.276,
193
  "step": 1000
194
+ },
195
+ {
196
+ "epoch": 7.1,
197
+ "learning_rate": 3.0529595015576325e-06,
198
+ "loss": 0.2783,
199
+ "step": 1200
200
+ },
201
+ {
202
+ "epoch": 7.1,
203
+ "eval_f1": 0.3662892905615983,
204
+ "eval_f1_all": [
205
+ 0.5064599483204134,
206
+ 0.5771643663739021,
207
+ 0.028169014084507046,
208
+ 0.125,
209
+ 0.6408730158730159,
210
+ 0.1222222222222222,
211
+ 0.35789473684210527,
212
+ 0.0,
213
+ 0.7578947368421052,
214
+ 0.6680988184747583,
215
+ 0.4291497975708502,
216
+ 0.5216178521617852,
217
+ 0.0,
218
+ 0.015503875968992248,
219
+ 0.5635864592863677,
220
+ 0.10135135135135136,
221
+ 0.6790622473726757,
222
+ 0.7058823529411764,
223
+ 0.10317460317460318,
224
+ 0.422680412371134
225
+ ],
226
+ "eval_loss": 0.3067641854286194,
227
+ "eval_runtime": 3.0738,
228
+ "eval_samples_per_second": 616.829,
229
+ "eval_steps_per_second": 19.52,
230
+ "step": 1200
231
  }
232
  ],
233
  "max_steps": 1690,
234
  "num_train_epochs": 10,
235
+ "total_flos": 8966928172547544.0,
236
  "trial_name": null,
237
  "trial_params": null
238
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4bdfd57d1e70cd331c7da4da7e930f75fa6690bcbcb63665b2897fc97f39df12
3
  size 1421660981
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6388e9a479aa96e7783654f3cf7f877d8519fea0ecf267a47186c81df6aa2bc3
3
  size 1421660981