plip commited on
Commit
c08fa7c
1 Parent(s): bc6cb0f

Training in progress, step 30000

Browse files
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:65debe850847edf2f62bba0deb29483af297ffa34c6cdcd66a55832044fcc8ec
3
  size 202193937
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5028e680b90d0d5deb58c782fcd2654d9815241c8789819c0ef4b4a5d0227dd1
3
  size 202193937
last-checkpoint/pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e6f55f7fb7116bd42203f7b9313d6b9f08f6a023278c3f67be7ce3fba873b5dc
3
  size 102501541
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:48ec0fa394e10e50451293d1035a84eba2d3d01f6d72f6c0c1bd4739fe3ee776
3
  size 102501541
last-checkpoint/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:006a29e4d6e5dacaa8f0d3ba56263cd7f1775a1129713c79ccc62a2c04246894
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d43991d7af84e9540d8f27d2f0d1e420a0cf71dfe1da000898814055b8f62473
3
  size 14503
last-checkpoint/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:006a29e4d6e5dacaa8f0d3ba56263cd7f1775a1129713c79ccc62a2c04246894
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d43991d7af84e9540d8f27d2f0d1e420a0cf71dfe1da000898814055b8f62473
3
  size 14503
last-checkpoint/rng_state_2.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:006a29e4d6e5dacaa8f0d3ba56263cd7f1775a1129713c79ccc62a2c04246894
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d43991d7af84e9540d8f27d2f0d1e420a0cf71dfe1da000898814055b8f62473
3
  size 14503
last-checkpoint/rng_state_3.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:006a29e4d6e5dacaa8f0d3ba56263cd7f1775a1129713c79ccc62a2c04246894
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d43991d7af84e9540d8f27d2f0d1e420a0cf71dfe1da000898814055b8f62473
3
  size 14503
last-checkpoint/rng_state_4.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:006a29e4d6e5dacaa8f0d3ba56263cd7f1775a1129713c79ccc62a2c04246894
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d43991d7af84e9540d8f27d2f0d1e420a0cf71dfe1da000898814055b8f62473
3
  size 14503
last-checkpoint/rng_state_5.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:006a29e4d6e5dacaa8f0d3ba56263cd7f1775a1129713c79ccc62a2c04246894
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d43991d7af84e9540d8f27d2f0d1e420a0cf71dfe1da000898814055b8f62473
3
  size 14503
last-checkpoint/rng_state_6.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:006a29e4d6e5dacaa8f0d3ba56263cd7f1775a1129713c79ccc62a2c04246894
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d43991d7af84e9540d8f27d2f0d1e420a0cf71dfe1da000898814055b8f62473
3
  size 14503
last-checkpoint/rng_state_7.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:006a29e4d6e5dacaa8f0d3ba56263cd7f1775a1129713c79ccc62a2c04246894
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d43991d7af84e9540d8f27d2f0d1e420a0cf71dfe1da000898814055b8f62473
3
  size 14503
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c84272d9e8a15cd78c706d16b77833d53dbfbd182e8ad79e3cd658ef6c3eaaf6
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff9b79eba2b3c3487d57b5f9509b60a9305753a31bdaa8757ad442412114850e
3
  size 623
last-checkpoint/trainer_state.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.46599408187516017,
5
- "global_step": 20000,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -406,11 +406,211 @@
406
  "eval_samples_per_second": 1670.425,
407
  "eval_steps_per_second": 26.599,
408
  "step": 20000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
409
  }
410
  ],
411
  "max_steps": 500000,
412
  "num_train_epochs": 12,
413
- "total_flos": 6.38974277549e+20,
414
  "trial_name": null,
415
  "trial_params": null
416
  }
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.6989911228127402,
5
+ "global_step": 30000,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
406
  "eval_samples_per_second": 1670.425,
407
  "eval_steps_per_second": 26.599,
408
  "step": 20000
409
+ },
410
+ {
411
+ "epoch": 0.48,
412
+ "learning_rate": 0.00024599999999999996,
413
+ "loss": 0.4278,
414
+ "step": 20500
415
+ },
416
+ {
417
+ "epoch": 0.49,
418
+ "learning_rate": 0.00025199999999999995,
419
+ "loss": 0.4233,
420
+ "step": 21000
421
+ },
422
+ {
423
+ "epoch": 0.49,
424
+ "eval_loss": 0.4032529294490814,
425
+ "eval_runtime": 1.2842,
426
+ "eval_samples_per_second": 1711.577,
427
+ "eval_steps_per_second": 27.254,
428
+ "step": 21000
429
+ },
430
+ {
431
+ "epoch": 0.5,
432
+ "learning_rate": 0.000258,
433
+ "loss": 0.4189,
434
+ "step": 21500
435
+ },
436
+ {
437
+ "epoch": 0.51,
438
+ "learning_rate": 0.00026399999999999997,
439
+ "loss": 0.4146,
440
+ "step": 22000
441
+ },
442
+ {
443
+ "epoch": 0.51,
444
+ "eval_loss": 0.39436739683151245,
445
+ "eval_runtime": 1.2633,
446
+ "eval_samples_per_second": 1739.901,
447
+ "eval_steps_per_second": 27.705,
448
+ "step": 22000
449
+ },
450
+ {
451
+ "epoch": 0.52,
452
+ "learning_rate": 0.00027,
453
+ "loss": 0.4109,
454
+ "step": 22500
455
+ },
456
+ {
457
+ "epoch": 0.54,
458
+ "learning_rate": 0.000276,
459
+ "loss": 0.4077,
460
+ "step": 23000
461
+ },
462
+ {
463
+ "epoch": 0.54,
464
+ "eval_loss": 0.38647642731666565,
465
+ "eval_runtime": 1.2503,
466
+ "eval_samples_per_second": 1758.021,
467
+ "eval_steps_per_second": 27.994,
468
+ "step": 23000
469
+ },
470
+ {
471
+ "epoch": 0.55,
472
+ "learning_rate": 0.00028199999999999997,
473
+ "loss": 0.4047,
474
+ "step": 23500
475
+ },
476
+ {
477
+ "epoch": 0.56,
478
+ "learning_rate": 0.00028799999999999995,
479
+ "loss": 0.4019,
480
+ "step": 24000
481
+ },
482
+ {
483
+ "epoch": 0.56,
484
+ "eval_loss": 0.381785124540329,
485
+ "eval_runtime": 1.2485,
486
+ "eval_samples_per_second": 1760.518,
487
+ "eval_steps_per_second": 28.034,
488
+ "step": 24000
489
+ },
490
+ {
491
+ "epoch": 0.57,
492
+ "learning_rate": 0.000294,
493
+ "loss": 0.3994,
494
+ "step": 24500
495
+ },
496
+ {
497
+ "epoch": 0.58,
498
+ "learning_rate": 0.0003,
499
+ "loss": 0.3968,
500
+ "step": 25000
501
+ },
502
+ {
503
+ "epoch": 0.58,
504
+ "eval_loss": 0.3785216510295868,
505
+ "eval_runtime": 1.2913,
506
+ "eval_samples_per_second": 1702.153,
507
+ "eval_steps_per_second": 27.104,
508
+ "step": 25000
509
+ },
510
+ {
511
+ "epoch": 0.59,
512
+ "learning_rate": 0.00029999920715161553,
513
+ "loss": 0.3945,
514
+ "step": 25500
515
+ },
516
+ {
517
+ "epoch": 0.61,
518
+ "learning_rate": 0.0002999968286151326,
519
+ "loss": 0.392,
520
+ "step": 26000
521
+ },
522
+ {
523
+ "epoch": 0.61,
524
+ "eval_loss": 0.37318962812423706,
525
+ "eval_runtime": 1.2758,
526
+ "eval_samples_per_second": 1722.782,
527
+ "eval_steps_per_second": 27.433,
528
+ "step": 26000
529
+ },
530
+ {
531
+ "epoch": 0.62,
532
+ "learning_rate": 0.0002999928644165624,
533
+ "loss": 0.3897,
534
+ "step": 26500
535
+ },
536
+ {
537
+ "epoch": 0.63,
538
+ "learning_rate": 0.0002999873145992569,
539
+ "loss": 0.3869,
540
+ "step": 27000
541
+ },
542
+ {
543
+ "epoch": 0.63,
544
+ "eval_loss": 0.3669659197330475,
545
+ "eval_runtime": 1.2869,
546
+ "eval_samples_per_second": 1707.967,
547
+ "eval_steps_per_second": 27.197,
548
+ "step": 27000
549
+ },
550
+ {
551
+ "epoch": 0.64,
552
+ "learning_rate": 0.000299980179223908,
553
+ "loss": 0.385,
554
+ "step": 27500
555
+ },
556
+ {
557
+ "epoch": 0.65,
558
+ "learning_rate": 0.0002999714583685469,
559
+ "loss": 0.3828,
560
+ "step": 28000
561
+ },
562
+ {
563
+ "epoch": 0.65,
564
+ "eval_loss": 0.3615981340408325,
565
+ "eval_runtime": 1.2961,
566
+ "eval_samples_per_second": 1695.914,
567
+ "eval_steps_per_second": 27.005,
568
+ "step": 28000
569
+ },
570
+ {
571
+ "epoch": 0.66,
572
+ "learning_rate": 0.00029996115212854366,
573
+ "loss": 0.3808,
574
+ "step": 28500
575
+ },
576
+ {
577
+ "epoch": 0.68,
578
+ "learning_rate": 0.00029994926061660554,
579
+ "loss": 0.3786,
580
+ "step": 29000
581
+ },
582
+ {
583
+ "epoch": 0.68,
584
+ "eval_loss": 0.3613170087337494,
585
+ "eval_runtime": 1.3013,
586
+ "eval_samples_per_second": 1689.114,
587
+ "eval_steps_per_second": 26.897,
588
+ "step": 29000
589
+ },
590
+ {
591
+ "epoch": 0.69,
592
+ "learning_rate": 0.0002999357839627762,
593
+ "loss": 0.377,
594
+ "step": 29500
595
+ },
596
+ {
597
+ "epoch": 0.7,
598
+ "learning_rate": 0.00029992072231443425,
599
+ "loss": 0.3751,
600
+ "step": 30000
601
+ },
602
+ {
603
+ "epoch": 0.7,
604
+ "eval_loss": 0.35551005601882935,
605
+ "eval_runtime": 1.2661,
606
+ "eval_samples_per_second": 1736.053,
607
+ "eval_steps_per_second": 27.644,
608
+ "step": 30000
609
  }
610
  ],
611
  "max_steps": 500000,
612
  "num_train_epochs": 12,
613
+ "total_flos": 9.584614163235e+20,
614
  "trial_name": null,
615
  "trial_params": null
616
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e6f55f7fb7116bd42203f7b9313d6b9f08f6a023278c3f67be7ce3fba873b5dc
3
  size 102501541
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:48ec0fa394e10e50451293d1035a84eba2d3d01f6d72f6c0c1bd4739fe3ee776
3
  size 102501541