ben81828 commited on
Commit
d3258a0
·
verified ·
1 Parent(s): 6dc6326

Training in progress, step 11950, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3e8de84e70949c404703b483e8ebea4616aa76e5550ebb518d001960ca8ef412
3
  size 29034840
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a57664fd8d118ea748563fac31ce8b78efe07e2901fabe33b4a76bc7345922c9
3
  size 29034840
last-checkpoint/global_step11950/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6733855f172a5e17147ab58fec551e4ea68d38f7c8b043a81415f2f81de978ae
3
+ size 43429616
last-checkpoint/global_step11950/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a14fd26ce8b1c99df35716828da1813b8c9f7e648038e92de7f170e8f1e8df78
3
+ size 43429616
last-checkpoint/global_step11950/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b11be5688b759767dfe0b26e8b368c3ec9a825be029288bdd45daf1f8dc9e542
3
+ size 43429616
last-checkpoint/global_step11950/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7293d5975b919a36400df7829d4e0f36d37efa4d5c7e4240a2f5634fd26f338a
3
+ size 43429616
last-checkpoint/global_step11950/zero_pp_rank_0_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a5f271fa2f28b74adf81f2429845e7461d0523805e900a34230860e44d7b755
3
+ size 637299
last-checkpoint/global_step11950/zero_pp_rank_1_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc694465bd460fe216d014082735cf94361930876706b4b89d8a5eb4a3e00493
3
+ size 637171
last-checkpoint/global_step11950/zero_pp_rank_2_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:093cb3139bae9ab33a671e4a148753a1bab1b301804a6384f38f53ee9d737d84
3
+ size 637171
last-checkpoint/global_step11950/zero_pp_rank_3_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e862425fb635b75ddaac1025f544ee91d4d7524b13c28cc38af2a83613c7a45
3
+ size 637171
last-checkpoint/latest CHANGED
@@ -1 +1 @@
1
- global_step11900
 
1
+ global_step11950
last-checkpoint/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fb2a68c98af16e47cb51ea6c0d0f4754f4a9407b30762d8f82e852b8e76a4608
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:08e63a1507ec8e6eb80eb4db3a04627f60726c8dfe128fa7ae48233814d3c7f7
3
  size 15024
last-checkpoint/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:509f6201054bbd37b518f6dab3ae8c057b4576b4a4b8e4e1585626d0721f975c
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f71e62374e9d96336f152733a8b55ad85a31bf315e29c8dae76917a6d5e9f50
3
  size 15024
last-checkpoint/rng_state_2.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ec94ec63816216d52664e59485f0d6307134f2239d5479100054cf3405e88af1
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:397b8002672366d97746debf09460a776c5721f87ae8f9aeb54c1b99133afdf5
3
  size 15024
last-checkpoint/rng_state_3.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:77e323bec972f8af2e04175c687b7e671d36cc07f3e45d61c276284be2f0eccc
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f1601c6a8a6f02b3ac80fc66c7243adaee34f52863db1b04674d7ba56e1a4a4
3
  size 15024
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3f3e1129e4ae45d39f9b6a7f47b9533d694c97a78bedd9094c78473d467ee729
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85333fd159e56c8beb00c8b6eec934d4a08a9b94a3fe141d6435a974bade77a4
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.04092838987708092,
3
  "best_model_checkpoint": "saves/CADICA_qwenvl_detect_classify_augmented/lora/sft/checkpoint-9050",
4
- "epoch": 1.4579339070778843,
5
  "eval_steps": 50,
6
- "global_step": 11900,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -21189,11 +21189,100 @@
21189
  "eval_steps_per_second": 0.778,
21190
  "num_input_tokens_seen": 79560856,
21191
  "step": 11900
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
21192
  }
21193
  ],
21194
  "logging_steps": 5,
21195
  "max_steps": 16324,
21196
- "num_input_tokens_seen": 79560856,
21197
  "num_train_epochs": 2,
21198
  "save_steps": 50,
21199
  "stateful_callbacks": {
@@ -21208,7 +21297,7 @@
21208
  "attributes": {}
21209
  }
21210
  },
21211
- "total_flos": 4990354859491328.0,
21212
  "train_batch_size": 1,
21213
  "trial_name": null,
21214
  "trial_params": null
 
1
  {
2
  "best_metric": 0.04092838987708092,
3
  "best_model_checkpoint": "saves/CADICA_qwenvl_detect_classify_augmented/lora/sft/checkpoint-9050",
4
+ "epoch": 1.4640592937429175,
5
  "eval_steps": 50,
6
+ "global_step": 11950,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
21189
  "eval_steps_per_second": 0.778,
21190
  "num_input_tokens_seen": 79560856,
21191
  "step": 11900
21192
+ },
21193
+ {
21194
+ "epoch": 1.4585464457443877,
21195
+ "grad_norm": 1.3679759784014216,
21196
+ "learning_rate": 1.8733954145586564e-05,
21197
+ "loss": 0.1984,
21198
+ "num_input_tokens_seen": 79594856,
21199
+ "step": 11905
21200
+ },
21201
+ {
21202
+ "epoch": 1.459158984410891,
21203
+ "grad_norm": 1.5475246361886241,
21204
+ "learning_rate": 1.869444614855496e-05,
21205
+ "loss": 0.2249,
21206
+ "num_input_tokens_seen": 79628416,
21207
+ "step": 11910
21208
+ },
21209
+ {
21210
+ "epoch": 1.4597715230773942,
21211
+ "grad_norm": 1.2083117859520573,
21212
+ "learning_rate": 1.865497027374209e-05,
21213
+ "loss": 0.2391,
21214
+ "num_input_tokens_seen": 79661656,
21215
+ "step": 11915
21216
+ },
21217
+ {
21218
+ "epoch": 1.4603840617438975,
21219
+ "grad_norm": 0.8092871912401626,
21220
+ "learning_rate": 1.861552656165359e-05,
21221
+ "loss": 0.1946,
21222
+ "num_input_tokens_seen": 79695424,
21223
+ "step": 11920
21224
+ },
21225
+ {
21226
+ "epoch": 1.460996600410401,
21227
+ "grad_norm": 1.8712024643764127,
21228
+ "learning_rate": 1.857611505276216e-05,
21229
+ "loss": 0.2245,
21230
+ "num_input_tokens_seen": 79728696,
21231
+ "step": 11925
21232
+ },
21233
+ {
21234
+ "epoch": 1.4616091390769042,
21235
+ "grad_norm": 1.708438260077005,
21236
+ "learning_rate": 1.8536735787507447e-05,
21237
+ "loss": 0.2071,
21238
+ "num_input_tokens_seen": 79762584,
21239
+ "step": 11930
21240
+ },
21241
+ {
21242
+ "epoch": 1.4622216777434076,
21243
+ "grad_norm": 1.3106131028698333,
21244
+ "learning_rate": 1.849738880629597e-05,
21245
+ "loss": 0.1768,
21246
+ "num_input_tokens_seen": 79796928,
21247
+ "step": 11935
21248
+ },
21249
+ {
21250
+ "epoch": 1.4628342164099108,
21251
+ "grad_norm": 1.4141721431066059,
21252
+ "learning_rate": 1.8458074149501153e-05,
21253
+ "loss": 0.223,
21254
+ "num_input_tokens_seen": 79830296,
21255
+ "step": 11940
21256
+ },
21257
+ {
21258
+ "epoch": 1.463446755076414,
21259
+ "grad_norm": 1.2084116693938611,
21260
+ "learning_rate": 1.8418791857463287e-05,
21261
+ "loss": 0.1694,
21262
+ "num_input_tokens_seen": 79864528,
21263
+ "step": 11945
21264
+ },
21265
+ {
21266
+ "epoch": 1.4640592937429175,
21267
+ "grad_norm": 1.2782206880022664,
21268
+ "learning_rate": 1.8379541970489366e-05,
21269
+ "loss": 0.1818,
21270
+ "num_input_tokens_seen": 79898064,
21271
+ "step": 11950
21272
+ },
21273
+ {
21274
+ "epoch": 1.4640592937429175,
21275
+ "eval_loss": 0.1088213175535202,
21276
+ "eval_runtime": 19.4183,
21277
+ "eval_samples_per_second": 3.09,
21278
+ "eval_steps_per_second": 0.772,
21279
+ "num_input_tokens_seen": 79898064,
21280
+ "step": 11950
21281
  }
21282
  ],
21283
  "logging_steps": 5,
21284
  "max_steps": 16324,
21285
+ "num_input_tokens_seen": 79898064,
21286
  "num_train_epochs": 2,
21287
  "save_steps": 50,
21288
  "stateful_callbacks": {
 
21297
  "attributes": {}
21298
  }
21299
  },
21300
+ "total_flos": 5011542192029696.0,
21301
  "train_batch_size": 1,
21302
  "trial_name": null,
21303
  "trial_params": null