dada22231 commited on
Commit
35357d1
1 Parent(s): f2286e2

Training in progress, step 42, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0a5f4d69ef98f96d85ef62e959c511b789ac80b6926f5b5207250d410778f8a8
3
  size 97728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06514da0eb15c610c3cec1974eed5f06de40ac94b4921072f7329a44b94d3748
3
  size 97728
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b95496f96d4ce56660e74c67f8fe76d34c21c492bdd40e8f6f81138472f68ef5
3
  size 212298
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4babe9608fe860fdf6c9eb5ba2bb866151491b6b28e9d881cf6f1b7daef98057
3
  size 212298
last-checkpoint/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f08773135f6f3b1ccb473bddedf248bbe0f93f221d52b2f5e8c66e4f8a675da6
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06e5990d5c191137f2e0cbe87b7c0f53d6455adf4bfee54f77b0f672ce6a5e0a
3
  size 15024
last-checkpoint/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c902b69959077f1b25492c58692753fcb90792100df2fb944f56d7ba1999df6f
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:16a2d6dea43355ffaa27153c0e8a009b7ff825795264f8492b7c7ca81413571a
3
  size 15024
last-checkpoint/rng_state_2.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d61497b68f95bb126b871bfdeefa66a68dbe8089dd6ba113e5c6c600a44cdd71
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c5d98186efceaf694923ca756096afd2d6023dfa26b7c3196ffe233b03efc8a7
3
  size 15024
last-checkpoint/rng_state_3.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1fb32e6e72470bcb40351f78e8ee0fd0cb32f4d50d8b2bfa5132f137b22b1ba8
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:661a43aed25b761f274dcb9e08ee625bb7c4a14e8d9afe10811aa54f72a95d7e
3
  size 15024
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e8f5af44da266da277e7531f687942d737458dea336c0a84bf03af6d775c87ac
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dde78f0152770b6d9ff9d100a4e49fd7ae2ec975980251062a46af49a9f251c7
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 1.8310810810810811,
5
  "eval_steps": 25,
6
- "global_step": 25,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -198,6 +198,125 @@
198
  "eval_samples_per_second": 495.946,
199
  "eval_steps_per_second": 128.946,
200
  "step": 25
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
201
  }
202
  ],
203
  "logging_steps": 1,
@@ -212,12 +331,12 @@
212
  "should_evaluate": false,
213
  "should_log": false,
214
  "should_save": true,
215
- "should_training_stop": false
216
  },
217
  "attributes": {}
218
  }
219
  },
220
- "total_flos": 21384029798400.0,
221
  "train_batch_size": 1,
222
  "trial_name": null,
223
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 3.114864864864865,
5
  "eval_steps": 25,
6
+ "global_step": 42,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
198
  "eval_samples_per_second": 495.946,
199
  "eval_steps_per_second": 128.946,
200
  "step": 25
201
+ },
202
+ {
203
+ "epoch": 1.9031531531531531,
204
+ "grad_norm": 0.6299550533294678,
205
+ "learning_rate": 4.109423525312738e-05,
206
+ "loss": 11.6034,
207
+ "step": 26
208
+ },
209
+ {
210
+ "epoch": 2.0045045045045047,
211
+ "grad_norm": 0.6582178473472595,
212
+ "learning_rate": 3.777924554357096e-05,
213
+ "loss": 14.3429,
214
+ "step": 27
215
+ },
216
+ {
217
+ "epoch": 2.0765765765765765,
218
+ "grad_norm": 0.4390489459037781,
219
+ "learning_rate": 3.45704275117204e-05,
220
+ "loss": 10.5127,
221
+ "step": 28
222
+ },
223
+ {
224
+ "epoch": 2.1486486486486487,
225
+ "grad_norm": 0.5411266088485718,
226
+ "learning_rate": 3.1487564587782306e-05,
227
+ "loss": 10.7874,
228
+ "step": 29
229
+ },
230
+ {
231
+ "epoch": 2.220720720720721,
232
+ "grad_norm": 0.5138953328132629,
233
+ "learning_rate": 2.854966364683872e-05,
234
+ "loss": 9.6413,
235
+ "step": 30
236
+ },
237
+ {
238
+ "epoch": 2.2927927927927927,
239
+ "grad_norm": 0.45892640948295593,
240
+ "learning_rate": 2.577483782514174e-05,
241
+ "loss": 10.7246,
242
+ "step": 31
243
+ },
244
+ {
245
+ "epoch": 2.364864864864865,
246
+ "grad_norm": 0.4862615466117859,
247
+ "learning_rate": 2.3180194846605367e-05,
248
+ "loss": 10.6531,
249
+ "step": 32
250
+ },
251
+ {
252
+ "epoch": 2.436936936936937,
253
+ "grad_norm": 0.5631672143936157,
254
+ "learning_rate": 2.0781731547998614e-05,
255
+ "loss": 10.1699,
256
+ "step": 33
257
+ },
258
+ {
259
+ "epoch": 2.509009009009009,
260
+ "grad_norm": 0.4757976233959198,
261
+ "learning_rate": 1.8594235253127375e-05,
262
+ "loss": 9.9347,
263
+ "step": 34
264
+ },
265
+ {
266
+ "epoch": 2.581081081081081,
267
+ "grad_norm": 0.5032361745834351,
268
+ "learning_rate": 1.6631192604065855e-05,
269
+ "loss": 10.5629,
270
+ "step": 35
271
+ },
272
+ {
273
+ "epoch": 2.653153153153153,
274
+ "grad_norm": 0.514340877532959,
275
+ "learning_rate": 1.490470641152345e-05,
276
+ "loss": 10.7382,
277
+ "step": 36
278
+ },
279
+ {
280
+ "epoch": 2.725225225225225,
281
+ "grad_norm": 0.501396656036377,
282
+ "learning_rate": 1.3425421036992098e-05,
283
+ "loss": 10.0469,
284
+ "step": 37
285
+ },
286
+ {
287
+ "epoch": 2.7972972972972974,
288
+ "grad_norm": 0.4644583463668823,
289
+ "learning_rate": 1.2202456766718093e-05,
290
+ "loss": 10.503,
291
+ "step": 38
292
+ },
293
+ {
294
+ "epoch": 2.8693693693693696,
295
+ "grad_norm": 0.5667948126792908,
296
+ "learning_rate": 1.1243353582104556e-05,
297
+ "loss": 10.8346,
298
+ "step": 39
299
+ },
300
+ {
301
+ "epoch": 2.9414414414414414,
302
+ "grad_norm": 0.5867165923118591,
303
+ "learning_rate": 1.0554024673218807e-05,
304
+ "loss": 10.6693,
305
+ "step": 40
306
+ },
307
+ {
308
+ "epoch": 3.0427927927927927,
309
+ "grad_norm": 0.7453739047050476,
310
+ "learning_rate": 1.0138719982009242e-05,
311
+ "loss": 15.0657,
312
+ "step": 41
313
+ },
314
+ {
315
+ "epoch": 3.114864864864865,
316
+ "grad_norm": 0.5105739235877991,
317
+ "learning_rate": 1e-05,
318
+ "loss": 10.1017,
319
+ "step": 42
320
  }
321
  ],
322
  "logging_steps": 1,
 
331
  "should_evaluate": false,
332
  "should_log": false,
333
  "should_save": true,
334
+ "should_training_stop": true
335
  },
336
  "attributes": {}
337
  }
338
  },
339
+ "total_flos": 35925170061312.0,
340
  "train_batch_size": 1,
341
  "trial_name": null,
342
  "trial_params": null