iamnguyen commited on
Commit
6871c2a
·
verified ·
1 Parent(s): 22c3ac0

Training in progress, step 2576, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ff07db1019ff730ef79b24091064f8a49cdefb9061c2509b0cec41d0929caa8e
3
  size 147770496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc2136cfbd00690f3783dc17e8800db2dc8ad727fd6fe2cc8c6c953b2599e207
3
  size 147770496
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1ef2da130652bf6a0a5337026ef09e6a2a500d20f2fc74f97ed1de83989cd348
3
  size 75455810
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ea2b85354be8b703c4c2920b8b2fb21f3e0e76e72838162be0d20e1b9317205
3
  size 75455810
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:10a3f2d483e1271610cf1160deebfe4b7cd7e846596d1240bcea8a0fb41c92d6
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3fd8571639a832ac3a6bdbe8b83a9d5174ee7552fcb8cdc3d2effb6c405a5866
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3bf3c143616698704049e834c6a36949b23673acd7aeab20c33c0b99d558ad3f
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83a970aca6d8492ea417337203ea2e40d5dcc9a674de898354c4c9743b9beef6
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.9965838692637534,
5
  "eval_steps": 500,
6
- "global_step": 2572,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -18011,6 +18011,34 @@
18011
  "learning_rate": 2.470938929571842e-10,
18012
  "loss": 1.4405,
18013
  "step": 2572
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
18014
  }
18015
  ],
18016
  "logging_steps": 1.0,
@@ -18030,7 +18058,7 @@
18030
  "attributes": {}
18031
  }
18032
  },
18033
- "total_flos": 2.5435260736329185e+18,
18034
  "train_batch_size": 1,
18035
  "trial_name": null,
18036
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.9981337664165743,
5
  "eval_steps": 500,
6
+ "global_step": 2576,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
18011
  "learning_rate": 2.470938929571842e-10,
18012
  "loss": 1.4405,
18013
  "step": 2572
18014
+ },
18015
+ {
18016
+ "epoch": 0.9969713435519586,
18017
+ "grad_norm": 0.18423818051815033,
18018
+ "learning_rate": 1.8918162699887997e-10,
18019
+ "loss": 1.3955,
18020
+ "step": 2573
18021
+ },
18022
+ {
18023
+ "epoch": 0.9973588178401638,
18024
+ "grad_norm": 0.17565150558948517,
18025
+ "learning_rate": 1.3899081563906182e-10,
18026
+ "loss": 1.3862,
18027
+ "step": 2574
18028
+ },
18029
+ {
18030
+ "epoch": 0.9977462921283691,
18031
+ "grad_norm": 0.18880826234817505,
18032
+ "learning_rate": 9.65215363907257e-11,
18033
+ "loss": 1.3653,
18034
+ "step": 2575
18035
+ },
18036
+ {
18037
+ "epoch": 0.9981337664165743,
18038
+ "grad_norm": 0.18108192086219788,
18039
+ "learning_rate": 6.177385484029685e-11,
18040
+ "loss": 1.3969,
18041
+ "step": 2576
18042
  }
18043
  ],
18044
  "logging_steps": 1.0,
 
18058
  "attributes": {}
18059
  }
18060
  },
18061
+ "total_flos": 2.547514766592221e+18,
18062
  "train_batch_size": 1,
18063
  "trial_name": null,
18064
  "trial_params": null