iamnguyen commited on
Commit
8c5b49f
·
verified ·
1 Parent(s): 1a55793

Training in progress, step 2572, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b50a8963fd836f2d4a75e5d6e5108f8838240dead3cf06ed2b83dbc79100fd4d
3
  size 147770496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff07db1019ff730ef79b24091064f8a49cdefb9061c2509b0cec41d0929caa8e
3
  size 147770496
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a9a9e65bd995e2f2ee324878ea2d0c3f2b57d027cac2eb0f3d9cb367e065083f
3
  size 75455810
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ef2da130652bf6a0a5337026ef09e6a2a500d20f2fc74f97ed1de83989cd348
3
  size 75455810
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:947d372c2ac3d768de08720e6fe18650bc49df377d135d0776536e2ae0cb6658
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10a3f2d483e1271610cf1160deebfe4b7cd7e846596d1240bcea8a0fb41c92d6
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9314e03f72b7117ed44ca7e4c2d8792153e93080b1ee70d313c7ca3f40ea0cdc
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3bf3c143616698704049e834c6a36949b23673acd7aeab20c33c0b99d558ad3f
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.9950339721109327,
5
  "eval_steps": 500,
6
- "global_step": 2568,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -17983,6 +17983,34 @@
17983
  "learning_rate": 5.559555351780655e-10,
17984
  "loss": 1.4324,
17985
  "step": 2568
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
17986
  }
17987
  ],
17988
  "logging_steps": 1.0,
@@ -18002,7 +18030,7 @@
18002
  "attributes": {}
18003
  }
18004
  },
18005
- "total_flos": 2.5393388307779543e+18,
18006
  "train_batch_size": 1,
18007
  "trial_name": null,
18008
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.9965838692637534,
5
  "eval_steps": 500,
6
+ "global_step": 2572,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
17983
  "learning_rate": 5.559555351780655e-10,
17984
  "loss": 1.4324,
17985
  "step": 2568
17986
+ },
17987
+ {
17988
+ "epoch": 0.9954214463991379,
17989
+ "grad_norm": 0.19253648817539215,
17990
+ "learning_rate": 4.671584644355687e-10,
17991
+ "loss": 1.426,
17992
+ "step": 2569
17993
+ },
17994
+ {
17995
+ "epoch": 0.9958089206873431,
17996
+ "grad_norm": 0.18221616744995117,
17997
+ "learning_rate": 3.8608241899940056e-10,
17998
+ "loss": 1.3659,
17999
+ "step": 2570
18000
+ },
18001
+ {
18002
+ "epoch": 0.9961963949755482,
18003
+ "grad_norm": 0.1748841404914856,
18004
+ "learning_rate": 3.1272752407773834e-10,
18005
+ "loss": 1.4263,
18006
+ "step": 2571
18007
+ },
18008
+ {
18009
+ "epoch": 0.9965838692637534,
18010
+ "grad_norm": 0.18521229922771454,
18011
+ "learning_rate": 2.470938929571842e-10,
18012
+ "loss": 1.4405,
18013
+ "step": 2572
18014
  }
18015
  ],
18016
  "logging_steps": 1.0,
 
18030
  "attributes": {}
18031
  }
18032
  },
18033
+ "total_flos": 2.5435260736329185e+18,
18034
  "train_batch_size": 1,
18035
  "trial_name": null,
18036
  "trial_params": null