|
{ |
|
"best_metric": 0.44340217113494873, |
|
"best_model_checkpoint": "saved_model/lop_jun2024/checkpoint-10982", |
|
"epoch": 2.999931707983337, |
|
"eval_steps": 500, |
|
"global_step": 10982, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0, |
|
"loss": 74.2157, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"grad_norm": 13.99621295928955, |
|
"learning_rate": 2.5e-06, |
|
"loss": 74.5237, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 14.391114234924316, |
|
"learning_rate": 6.5000000000000004e-06, |
|
"loss": 74.2338, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 15.539421081542969, |
|
"learning_rate": 1.1500000000000002e-05, |
|
"loss": 72.8847, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 16.487953186035156, |
|
"learning_rate": 1.65e-05, |
|
"loss": 70.7699, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 21.840412139892578, |
|
"learning_rate": 2.15e-05, |
|
"loss": 66.4777, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 23.74176025390625, |
|
"learning_rate": 2.6000000000000002e-05, |
|
"loss": 59.438, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 26.7894287109375, |
|
"learning_rate": 3.1e-05, |
|
"loss": 45.7214, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 21.767683029174805, |
|
"learning_rate": 3.55e-05, |
|
"loss": 27.7672, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 15.309523582458496, |
|
"learning_rate": 4.05e-05, |
|
"loss": 12.9366, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 7.069633960723877, |
|
"learning_rate": 4.55e-05, |
|
"loss": 4.342, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 10.602043151855469, |
|
"learning_rate": 5.05e-05, |
|
"loss": 1.7191, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 8.00312614440918, |
|
"learning_rate": 5.550000000000001e-05, |
|
"loss": 1.3934, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 7.026750087738037, |
|
"learning_rate": 6.05e-05, |
|
"loss": 1.3728, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 10.686494827270508, |
|
"learning_rate": 6.55e-05, |
|
"loss": 1.2595, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 7.127823352813721, |
|
"learning_rate": 7.05e-05, |
|
"loss": 1.2226, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 6.253384590148926, |
|
"learning_rate": 7.55e-05, |
|
"loss": 1.2934, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 7.56700325012207, |
|
"learning_rate": 8.05e-05, |
|
"loss": 1.2571, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 9.671475410461426, |
|
"learning_rate": 8.55e-05, |
|
"loss": 1.2005, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 7.772515296936035, |
|
"learning_rate": 9.05e-05, |
|
"loss": 1.253, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 6.593009948730469, |
|
"learning_rate": 9.55e-05, |
|
"loss": 1.2269, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 9.796707153320312, |
|
"learning_rate": 9.999725274725276e-05, |
|
"loss": 1.2425, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 5.557741641998291, |
|
"learning_rate": 9.996978021978023e-05, |
|
"loss": 1.2383, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 7.321559906005859, |
|
"learning_rate": 9.994230769230771e-05, |
|
"loss": 1.2026, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 3.337256908416748, |
|
"learning_rate": 9.991483516483518e-05, |
|
"loss": 1.1289, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 4.283127307891846, |
|
"learning_rate": 9.988736263736264e-05, |
|
"loss": 1.0357, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 4.2405242919921875, |
|
"learning_rate": 9.985989010989013e-05, |
|
"loss": 1.0067, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 6.622239112854004, |
|
"learning_rate": 9.98324175824176e-05, |
|
"loss": 0.9816, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 5.521809101104736, |
|
"learning_rate": 9.980494505494506e-05, |
|
"loss": 0.9578, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 3.7347898483276367, |
|
"learning_rate": 9.977747252747254e-05, |
|
"loss": 0.8762, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 6.452709674835205, |
|
"learning_rate": 9.975000000000001e-05, |
|
"loss": 0.8188, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 3.2089035511016846, |
|
"learning_rate": 9.972252747252748e-05, |
|
"loss": 0.8247, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 4.455322742462158, |
|
"learning_rate": 9.969505494505496e-05, |
|
"loss": 0.7599, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 4.537048816680908, |
|
"learning_rate": 9.966758241758242e-05, |
|
"loss": 0.7968, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 3.754340648651123, |
|
"learning_rate": 9.964010989010988e-05, |
|
"loss": 0.72, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 4.538580894470215, |
|
"learning_rate": 9.961263736263737e-05, |
|
"loss": 0.7207, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 4.3613457679748535, |
|
"learning_rate": 9.958516483516483e-05, |
|
"loss": 0.7235, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 2.2669098377227783, |
|
"learning_rate": 9.95576923076923e-05, |
|
"loss": 0.7018, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 5.329736232757568, |
|
"learning_rate": 9.953021978021978e-05, |
|
"loss": 0.7193, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 2.5049781799316406, |
|
"learning_rate": 9.950274725274725e-05, |
|
"loss": 0.6948, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 3.5907931327819824, |
|
"learning_rate": 9.947527472527472e-05, |
|
"loss": 0.6637, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 3.264726161956787, |
|
"learning_rate": 9.94478021978022e-05, |
|
"loss": 0.6588, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 3.763380765914917, |
|
"learning_rate": 9.942032967032967e-05, |
|
"loss": 0.6479, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 1.9857630729675293, |
|
"learning_rate": 9.939285714285714e-05, |
|
"loss": 0.6481, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 2.649977922439575, |
|
"learning_rate": 9.936538461538462e-05, |
|
"loss": 0.6284, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 1.4232739210128784, |
|
"learning_rate": 9.933791208791209e-05, |
|
"loss": 0.6214, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 5.327674865722656, |
|
"learning_rate": 9.931043956043956e-05, |
|
"loss": 0.633, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 3.673241376876831, |
|
"learning_rate": 9.928296703296704e-05, |
|
"loss": 0.6792, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 3.6698083877563477, |
|
"learning_rate": 9.92554945054945e-05, |
|
"loss": 0.6297, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 2.056175470352173, |
|
"learning_rate": 9.922802197802197e-05, |
|
"loss": 0.6172, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 1.345395803451538, |
|
"learning_rate": 9.920054945054946e-05, |
|
"loss": 0.605, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 1.7540231943130493, |
|
"learning_rate": 9.917307692307692e-05, |
|
"loss": 0.6288, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 4.050562858581543, |
|
"learning_rate": 9.914560439560439e-05, |
|
"loss": 0.6281, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 5.210700511932373, |
|
"learning_rate": 9.911813186813187e-05, |
|
"loss": 0.6138, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 3.6927013397216797, |
|
"learning_rate": 9.909065934065934e-05, |
|
"loss": 0.6033, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 3.4765613079071045, |
|
"learning_rate": 9.906318681318681e-05, |
|
"loss": 0.5991, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 1.8127527236938477, |
|
"learning_rate": 9.903571428571429e-05, |
|
"loss": 0.5994, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 1.6926054954528809, |
|
"learning_rate": 9.900824175824176e-05, |
|
"loss": 0.5988, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 2.3790557384490967, |
|
"learning_rate": 9.898076923076923e-05, |
|
"loss": 0.6147, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 3.558159589767456, |
|
"learning_rate": 9.895329670329671e-05, |
|
"loss": 0.5858, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 3.5203194618225098, |
|
"learning_rate": 9.892582417582418e-05, |
|
"loss": 0.6036, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 2.4314634799957275, |
|
"learning_rate": 9.889835164835165e-05, |
|
"loss": 0.5789, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 2.956555128097534, |
|
"learning_rate": 9.887087912087913e-05, |
|
"loss": 0.6021, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 2.231612205505371, |
|
"learning_rate": 9.88434065934066e-05, |
|
"loss": 0.5825, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 2.1588950157165527, |
|
"learning_rate": 9.881593406593406e-05, |
|
"loss": 0.5924, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 2.8479809761047363, |
|
"learning_rate": 9.878846153846155e-05, |
|
"loss": 0.584, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 2.2029404640197754, |
|
"learning_rate": 9.876098901098901e-05, |
|
"loss": 0.5855, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 2.243802070617676, |
|
"learning_rate": 9.873351648351648e-05, |
|
"loss": 0.584, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 2.8837687969207764, |
|
"learning_rate": 9.870604395604396e-05, |
|
"loss": 0.5867, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 2.0768215656280518, |
|
"learning_rate": 9.867857142857143e-05, |
|
"loss": 0.5823, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 2.950639247894287, |
|
"learning_rate": 9.86510989010989e-05, |
|
"loss": 0.5832, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 1.0848056077957153, |
|
"learning_rate": 9.862362637362638e-05, |
|
"loss": 0.5595, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 1.293261170387268, |
|
"learning_rate": 9.859615384615385e-05, |
|
"loss": 0.5926, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 2.286271095275879, |
|
"learning_rate": 9.856868131868132e-05, |
|
"loss": 0.5658, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 2.9983866214752197, |
|
"learning_rate": 9.85412087912088e-05, |
|
"loss": 0.5749, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 2.712332248687744, |
|
"learning_rate": 9.851373626373627e-05, |
|
"loss": 0.5757, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 3.5591819286346436, |
|
"learning_rate": 9.848626373626374e-05, |
|
"loss": 0.5706, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 3.00175142288208, |
|
"learning_rate": 9.845879120879122e-05, |
|
"loss": 0.5654, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 1.5290532112121582, |
|
"learning_rate": 9.843131868131869e-05, |
|
"loss": 0.578, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 1.8737398386001587, |
|
"learning_rate": 9.840384615384615e-05, |
|
"loss": 0.5598, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 1.190427541732788, |
|
"learning_rate": 9.837637362637364e-05, |
|
"loss": 0.5507, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 2.8329577445983887, |
|
"learning_rate": 9.83489010989011e-05, |
|
"loss": 0.5394, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 2.6253738403320312, |
|
"learning_rate": 9.832142857142857e-05, |
|
"loss": 0.5404, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 2.1394851207733154, |
|
"learning_rate": 9.829395604395605e-05, |
|
"loss": 0.5616, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 1.5708709955215454, |
|
"learning_rate": 9.826648351648352e-05, |
|
"loss": 0.5606, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 1.521190881729126, |
|
"learning_rate": 9.823901098901099e-05, |
|
"loss": 0.5605, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 1.7570327520370483, |
|
"learning_rate": 9.821153846153847e-05, |
|
"loss": 0.5475, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 2.039400577545166, |
|
"learning_rate": 9.818406593406594e-05, |
|
"loss": 0.5451, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 1.9855557680130005, |
|
"learning_rate": 9.815659340659341e-05, |
|
"loss": 0.5378, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 1.3518325090408325, |
|
"learning_rate": 9.812912087912089e-05, |
|
"loss": 0.5447, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 2.36753511428833, |
|
"learning_rate": 9.810164835164836e-05, |
|
"loss": 0.5593, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 1.611568570137024, |
|
"learning_rate": 9.807417582417583e-05, |
|
"loss": 0.536, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 1.3011342287063599, |
|
"learning_rate": 9.804670329670331e-05, |
|
"loss": 0.5508, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 1.6673119068145752, |
|
"learning_rate": 9.801923076923078e-05, |
|
"loss": 0.5427, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 1.4966050386428833, |
|
"learning_rate": 9.799175824175824e-05, |
|
"loss": 0.5364, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 1.2000831365585327, |
|
"learning_rate": 9.796428571428573e-05, |
|
"loss": 0.5452, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 1.9247843027114868, |
|
"learning_rate": 9.79368131868132e-05, |
|
"loss": 0.529, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 1.6240330934524536, |
|
"learning_rate": 9.790934065934066e-05, |
|
"loss": 0.5523, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 1.4320908784866333, |
|
"learning_rate": 9.788186813186814e-05, |
|
"loss": 0.5567, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 1.7884244918823242, |
|
"learning_rate": 9.785439560439561e-05, |
|
"loss": 0.5244, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 1.9070675373077393, |
|
"learning_rate": 9.782692307692308e-05, |
|
"loss": 0.5354, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 1.0304174423217773, |
|
"learning_rate": 9.779945054945056e-05, |
|
"loss": 0.5443, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 2.184297800064087, |
|
"learning_rate": 9.777197802197803e-05, |
|
"loss": 0.5201, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 1.2110681533813477, |
|
"learning_rate": 9.77445054945055e-05, |
|
"loss": 0.5363, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 2.25685977935791, |
|
"learning_rate": 9.771703296703298e-05, |
|
"loss": 0.5392, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 2.558457851409912, |
|
"learning_rate": 9.768956043956045e-05, |
|
"loss": 0.5466, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 1.5898990631103516, |
|
"learning_rate": 9.766208791208792e-05, |
|
"loss": 0.5386, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 2.6933228969573975, |
|
"learning_rate": 9.76346153846154e-05, |
|
"loss": 0.5565, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 1.440042495727539, |
|
"learning_rate": 9.760714285714287e-05, |
|
"loss": 0.5507, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 1.9476529359817505, |
|
"learning_rate": 9.757967032967033e-05, |
|
"loss": 0.5276, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 3.142998218536377, |
|
"learning_rate": 9.755219780219782e-05, |
|
"loss": 0.5238, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 2.13360857963562, |
|
"learning_rate": 9.752472527472528e-05, |
|
"loss": 0.5326, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 1.2551095485687256, |
|
"learning_rate": 9.749725274725275e-05, |
|
"loss": 0.5569, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 1.0858782529830933, |
|
"learning_rate": 9.746978021978023e-05, |
|
"loss": 0.5425, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 1.3724364042282104, |
|
"learning_rate": 9.74423076923077e-05, |
|
"loss": 0.5381, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 1.2211335897445679, |
|
"learning_rate": 9.741483516483517e-05, |
|
"loss": 0.5147, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 1.806498646736145, |
|
"learning_rate": 9.738736263736264e-05, |
|
"loss": 0.5218, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.7817394733428955, |
|
"learning_rate": 9.735989010989012e-05, |
|
"loss": 0.539, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 2.054344415664673, |
|
"learning_rate": 9.733241758241759e-05, |
|
"loss": 0.5305, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.8021188974380493, |
|
"learning_rate": 9.730494505494506e-05, |
|
"loss": 0.519, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 1.833899736404419, |
|
"learning_rate": 9.727747252747254e-05, |
|
"loss": 0.5589, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 2.6109519004821777, |
|
"learning_rate": 9.725e-05, |
|
"loss": 0.5333, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 2.0866036415100098, |
|
"learning_rate": 9.722252747252747e-05, |
|
"loss": 0.5436, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 2.150935649871826, |
|
"learning_rate": 9.719505494505496e-05, |
|
"loss": 0.5202, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 1.280685544013977, |
|
"learning_rate": 9.716758241758242e-05, |
|
"loss": 0.5146, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.7753365635871887, |
|
"learning_rate": 9.714010989010989e-05, |
|
"loss": 0.5106, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.8706464767456055, |
|
"learning_rate": 9.711263736263737e-05, |
|
"loss": 0.4952, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 1.3467265367507935, |
|
"learning_rate": 9.708516483516484e-05, |
|
"loss": 0.5105, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 2.5618174076080322, |
|
"learning_rate": 9.705769230769231e-05, |
|
"loss": 0.5065, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 1.3450369834899902, |
|
"learning_rate": 9.703021978021979e-05, |
|
"loss": 0.5201, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 1.3272931575775146, |
|
"learning_rate": 9.700274725274726e-05, |
|
"loss": 0.5189, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 2.324739694595337, |
|
"learning_rate": 9.697527472527473e-05, |
|
"loss": 0.5298, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 1.375444769859314, |
|
"learning_rate": 9.694780219780221e-05, |
|
"loss": 0.527, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 1.1685923337936401, |
|
"learning_rate": 9.692032967032968e-05, |
|
"loss": 0.5303, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 1.6014115810394287, |
|
"learning_rate": 9.689285714285715e-05, |
|
"loss": 0.5399, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 2.3533401489257812, |
|
"learning_rate": 9.686538461538463e-05, |
|
"loss": 0.5185, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 1.6813533306121826, |
|
"learning_rate": 9.68379120879121e-05, |
|
"loss": 0.4986, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 1.1102067232131958, |
|
"learning_rate": 9.681043956043956e-05, |
|
"loss": 0.5313, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 1.576412320137024, |
|
"learning_rate": 9.678296703296705e-05, |
|
"loss": 0.5087, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 1.5507354736328125, |
|
"learning_rate": 9.675549450549451e-05, |
|
"loss": 0.532, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 1.4840365648269653, |
|
"learning_rate": 9.672802197802198e-05, |
|
"loss": 0.5191, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 1.8767573833465576, |
|
"learning_rate": 9.670054945054946e-05, |
|
"loss": 0.5252, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 1.0486218929290771, |
|
"learning_rate": 9.667307692307693e-05, |
|
"loss": 0.5152, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 2.066174030303955, |
|
"learning_rate": 9.66456043956044e-05, |
|
"loss": 0.5163, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 1.7480829954147339, |
|
"learning_rate": 9.661813186813188e-05, |
|
"loss": 0.5292, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 1.6554828882217407, |
|
"learning_rate": 9.659065934065935e-05, |
|
"loss": 0.5301, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 1.0723321437835693, |
|
"learning_rate": 9.656318681318682e-05, |
|
"loss": 0.5118, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 1.7398673295974731, |
|
"learning_rate": 9.653571428571429e-05, |
|
"loss": 0.5268, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 1.2732940912246704, |
|
"learning_rate": 9.650824175824175e-05, |
|
"loss": 0.5085, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 1.0260119438171387, |
|
"learning_rate": 9.648076923076924e-05, |
|
"loss": 0.5243, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.8816061019897461, |
|
"learning_rate": 9.64532967032967e-05, |
|
"loss": 0.5158, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 2.7940518856048584, |
|
"learning_rate": 9.642582417582417e-05, |
|
"loss": 0.5051, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 1.1889257431030273, |
|
"learning_rate": 9.639835164835165e-05, |
|
"loss": 0.5111, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 1.8643608093261719, |
|
"learning_rate": 9.637087912087912e-05, |
|
"loss": 0.5136, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 1.658339023590088, |
|
"learning_rate": 9.634340659340659e-05, |
|
"loss": 0.5132, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.6262873411178589, |
|
"learning_rate": 9.631593406593407e-05, |
|
"loss": 0.5047, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 1.2276610136032104, |
|
"learning_rate": 9.628846153846154e-05, |
|
"loss": 0.4997, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 1.3661067485809326, |
|
"learning_rate": 9.626098901098901e-05, |
|
"loss": 0.5113, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 1.423560380935669, |
|
"learning_rate": 9.623351648351649e-05, |
|
"loss": 0.5167, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 1.391201376914978, |
|
"learning_rate": 9.620604395604396e-05, |
|
"loss": 0.5057, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 1.5331265926361084, |
|
"learning_rate": 9.617857142857143e-05, |
|
"loss": 0.5105, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 2.0285804271698, |
|
"learning_rate": 9.615109890109891e-05, |
|
"loss": 0.5009, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 1.470106840133667, |
|
"learning_rate": 9.612362637362638e-05, |
|
"loss": 0.5082, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.8231136798858643, |
|
"learning_rate": 9.609615384615384e-05, |
|
"loss": 0.4893, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 1.4825439453125, |
|
"learning_rate": 9.606868131868133e-05, |
|
"loss": 0.5132, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 1.8054287433624268, |
|
"learning_rate": 9.60412087912088e-05, |
|
"loss": 0.4878, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.9411340355873108, |
|
"learning_rate": 9.601373626373626e-05, |
|
"loss": 0.4908, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 1.754782795906067, |
|
"learning_rate": 9.598626373626374e-05, |
|
"loss": 0.5206, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 1.1740814447402954, |
|
"learning_rate": 9.595879120879121e-05, |
|
"loss": 0.514, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 1.0887936353683472, |
|
"learning_rate": 9.593131868131868e-05, |
|
"loss": 0.5095, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 2.0167980194091797, |
|
"learning_rate": 9.590384615384616e-05, |
|
"loss": 0.5066, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 1.6176742315292358, |
|
"learning_rate": 9.587637362637363e-05, |
|
"loss": 0.5016, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 1.0261584520339966, |
|
"learning_rate": 9.58489010989011e-05, |
|
"loss": 0.5032, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 1.0735365152359009, |
|
"learning_rate": 9.582142857142858e-05, |
|
"loss": 0.4894, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 1.4293698072433472, |
|
"learning_rate": 9.579395604395605e-05, |
|
"loss": 0.5111, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 1.296971321105957, |
|
"learning_rate": 9.576648351648352e-05, |
|
"loss": 0.5078, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.7727321982383728, |
|
"learning_rate": 9.5739010989011e-05, |
|
"loss": 0.4923, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 2.6028566360473633, |
|
"learning_rate": 9.571153846153847e-05, |
|
"loss": 0.4994, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 1.2606744766235352, |
|
"learning_rate": 9.568406593406593e-05, |
|
"loss": 0.4885, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 1.812680721282959, |
|
"learning_rate": 9.565659340659342e-05, |
|
"loss": 0.5101, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 1.2432329654693604, |
|
"learning_rate": 9.562912087912088e-05, |
|
"loss": 0.5012, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 1.365795373916626, |
|
"learning_rate": 9.560164835164835e-05, |
|
"loss": 0.5093, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 1.4509589672088623, |
|
"learning_rate": 9.557417582417583e-05, |
|
"loss": 0.4921, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 1.0986443758010864, |
|
"learning_rate": 9.55467032967033e-05, |
|
"loss": 0.5003, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 1.46748685836792, |
|
"learning_rate": 9.551923076923077e-05, |
|
"loss": 0.4925, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 1.3228458166122437, |
|
"learning_rate": 9.549175824175825e-05, |
|
"loss": 0.5032, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 0.8542389273643494, |
|
"learning_rate": 9.546428571428572e-05, |
|
"loss": 0.4896, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 1.7039434909820557, |
|
"learning_rate": 9.543681318681319e-05, |
|
"loss": 0.4852, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 1.1597633361816406, |
|
"learning_rate": 9.540934065934067e-05, |
|
"loss": 0.5015, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 2.178933620452881, |
|
"learning_rate": 9.538186813186814e-05, |
|
"loss": 0.5298, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 1.1454273462295532, |
|
"learning_rate": 9.53543956043956e-05, |
|
"loss": 0.5046, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.7407316565513611, |
|
"learning_rate": 9.532692307692309e-05, |
|
"loss": 0.486, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 1.7394108772277832, |
|
"learning_rate": 9.529945054945056e-05, |
|
"loss": 0.4977, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.9852740168571472, |
|
"learning_rate": 9.527197802197802e-05, |
|
"loss": 0.4871, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 1.4632673263549805, |
|
"learning_rate": 9.52445054945055e-05, |
|
"loss": 0.5082, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 1.4848504066467285, |
|
"learning_rate": 9.521703296703297e-05, |
|
"loss": 0.5067, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 1.5140033960342407, |
|
"learning_rate": 9.518956043956044e-05, |
|
"loss": 0.5063, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 2.4344804286956787, |
|
"learning_rate": 9.516208791208791e-05, |
|
"loss": 0.5099, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.8636808395385742, |
|
"learning_rate": 9.513461538461539e-05, |
|
"loss": 0.4902, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 1.4363267421722412, |
|
"learning_rate": 9.510714285714286e-05, |
|
"loss": 0.498, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.7588983774185181, |
|
"learning_rate": 9.507967032967033e-05, |
|
"loss": 0.4884, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 1.4090018272399902, |
|
"learning_rate": 9.505219780219781e-05, |
|
"loss": 0.5082, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 2.2110331058502197, |
|
"learning_rate": 9.502472527472528e-05, |
|
"loss": 0.5064, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 1.2035874128341675, |
|
"learning_rate": 9.499725274725275e-05, |
|
"loss": 0.4952, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 1.597834825515747, |
|
"learning_rate": 9.496978021978023e-05, |
|
"loss": 0.4956, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 1.4878921508789062, |
|
"learning_rate": 9.49423076923077e-05, |
|
"loss": 0.4865, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 1.6747124195098877, |
|
"learning_rate": 9.491483516483516e-05, |
|
"loss": 0.4935, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 1.3186380863189697, |
|
"learning_rate": 9.488736263736264e-05, |
|
"loss": 0.4938, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 2.2865145206451416, |
|
"learning_rate": 9.485989010989011e-05, |
|
"loss": 0.4985, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 2.515855550765991, |
|
"learning_rate": 9.483241758241758e-05, |
|
"loss": 0.5049, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 1.2123080492019653, |
|
"learning_rate": 9.480494505494506e-05, |
|
"loss": 0.5004, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 1.0919526815414429, |
|
"learning_rate": 9.477747252747253e-05, |
|
"loss": 0.4854, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 1.3105583190917969, |
|
"learning_rate": 9.475e-05, |
|
"loss": 0.4866, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 1.745622158050537, |
|
"learning_rate": 9.472252747252748e-05, |
|
"loss": 0.4892, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.6633902192115784, |
|
"learning_rate": 9.469505494505495e-05, |
|
"loss": 0.507, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 1.0530184507369995, |
|
"learning_rate": 9.466758241758242e-05, |
|
"loss": 0.5035, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 1.2395071983337402, |
|
"learning_rate": 9.46401098901099e-05, |
|
"loss": 0.4926, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 1.139224886894226, |
|
"learning_rate": 9.461263736263737e-05, |
|
"loss": 0.4861, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.844968318939209, |
|
"learning_rate": 9.458516483516484e-05, |
|
"loss": 0.4855, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 1.2388585805892944, |
|
"learning_rate": 9.455769230769232e-05, |
|
"loss": 0.4887, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 1.4267388582229614, |
|
"learning_rate": 9.453021978021978e-05, |
|
"loss": 0.5031, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.7797860503196716, |
|
"learning_rate": 9.450274725274725e-05, |
|
"loss": 0.5014, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 1.3723804950714111, |
|
"learning_rate": 9.447527472527473e-05, |
|
"loss": 0.4887, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 1.396022081375122, |
|
"learning_rate": 9.44478021978022e-05, |
|
"loss": 0.4869, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 1.3555471897125244, |
|
"learning_rate": 9.442032967032967e-05, |
|
"loss": 0.4879, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 0.9702646136283875, |
|
"learning_rate": 9.439285714285715e-05, |
|
"loss": 0.4894, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.7753974199295044, |
|
"learning_rate": 9.436538461538462e-05, |
|
"loss": 0.5092, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 1.020336389541626, |
|
"learning_rate": 9.433791208791209e-05, |
|
"loss": 0.4961, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 1.2176239490509033, |
|
"learning_rate": 9.431043956043957e-05, |
|
"loss": 0.4827, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.8337396383285522, |
|
"learning_rate": 9.428296703296704e-05, |
|
"loss": 0.4931, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.8999798893928528, |
|
"learning_rate": 9.425549450549451e-05, |
|
"loss": 0.4912, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 1.6992242336273193, |
|
"learning_rate": 9.422802197802199e-05, |
|
"loss": 0.4835, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.7739867568016052, |
|
"learning_rate": 9.420054945054946e-05, |
|
"loss": 0.4783, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 1.2350486516952515, |
|
"learning_rate": 9.417307692307692e-05, |
|
"loss": 0.4904, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 1.1423838138580322, |
|
"learning_rate": 9.41456043956044e-05, |
|
"loss": 0.4884, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 1.0562171936035156, |
|
"learning_rate": 9.411813186813187e-05, |
|
"loss": 0.5082, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.7696706056594849, |
|
"learning_rate": 9.409065934065934e-05, |
|
"loss": 0.4889, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.5980774760246277, |
|
"learning_rate": 9.406318681318682e-05, |
|
"loss": 0.48, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 1.413926124572754, |
|
"learning_rate": 9.403571428571429e-05, |
|
"loss": 0.4971, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.8020157217979431, |
|
"learning_rate": 9.400824175824176e-05, |
|
"loss": 0.4926, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.6620104312896729, |
|
"learning_rate": 9.398076923076924e-05, |
|
"loss": 0.4941, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 1.9766381978988647, |
|
"learning_rate": 9.395329670329671e-05, |
|
"loss": 0.4867, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 1.3235608339309692, |
|
"learning_rate": 9.392582417582418e-05, |
|
"loss": 0.5061, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.8266246914863586, |
|
"learning_rate": 9.389835164835166e-05, |
|
"loss": 0.4842, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 1.1611332893371582, |
|
"learning_rate": 9.387087912087913e-05, |
|
"loss": 0.4795, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 1.750346064567566, |
|
"learning_rate": 9.38434065934066e-05, |
|
"loss": 0.4859, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 0.7080681324005127, |
|
"learning_rate": 9.381593406593408e-05, |
|
"loss": 0.4762, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 1.3531988859176636, |
|
"learning_rate": 9.378846153846155e-05, |
|
"loss": 0.4921, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.8100531697273254, |
|
"learning_rate": 9.376098901098901e-05, |
|
"loss": 0.4912, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 1.6417909860610962, |
|
"learning_rate": 9.37335164835165e-05, |
|
"loss": 0.4868, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 2.7357239723205566, |
|
"learning_rate": 9.370604395604396e-05, |
|
"loss": 0.4837, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.7493315935134888, |
|
"learning_rate": 9.367857142857143e-05, |
|
"loss": 0.494, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 1.4594992399215698, |
|
"learning_rate": 9.365109890109891e-05, |
|
"loss": 0.4967, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 1.2568315267562866, |
|
"learning_rate": 9.362362637362638e-05, |
|
"loss": 0.4865, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.8711192607879639, |
|
"learning_rate": 9.359615384615385e-05, |
|
"loss": 0.4915, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 1.6470551490783691, |
|
"learning_rate": 9.356868131868133e-05, |
|
"loss": 0.4736, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 2.410200357437134, |
|
"learning_rate": 9.35412087912088e-05, |
|
"loss": 0.5078, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 1.0611746311187744, |
|
"learning_rate": 9.351373626373627e-05, |
|
"loss": 0.4986, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 0.820321798324585, |
|
"learning_rate": 9.348626373626375e-05, |
|
"loss": 0.4788, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.7887302041053772, |
|
"learning_rate": 9.345879120879122e-05, |
|
"loss": 0.4878, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 1.2939894199371338, |
|
"learning_rate": 9.343131868131869e-05, |
|
"loss": 0.4896, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.7468405961990356, |
|
"learning_rate": 9.340384615384615e-05, |
|
"loss": 0.4932, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 1.63861882686615, |
|
"learning_rate": 9.337637362637362e-05, |
|
"loss": 0.4739, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 1.1842321157455444, |
|
"learning_rate": 9.33489010989011e-05, |
|
"loss": 0.4845, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 1.0613666772842407, |
|
"learning_rate": 9.332142857142857e-05, |
|
"loss": 0.4806, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.9317752718925476, |
|
"learning_rate": 9.329395604395604e-05, |
|
"loss": 0.4803, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 1.378394603729248, |
|
"learning_rate": 9.326648351648352e-05, |
|
"loss": 0.4942, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.748068630695343, |
|
"learning_rate": 9.323901098901099e-05, |
|
"loss": 0.4763, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 1.4142990112304688, |
|
"learning_rate": 9.321153846153846e-05, |
|
"loss": 0.4913, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.6746355891227722, |
|
"learning_rate": 9.318406593406594e-05, |
|
"loss": 0.4781, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 1.146001935005188, |
|
"learning_rate": 9.315659340659341e-05, |
|
"loss": 0.4789, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 1.0291472673416138, |
|
"learning_rate": 9.312912087912088e-05, |
|
"loss": 0.4813, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 1.7905545234680176, |
|
"learning_rate": 9.310164835164836e-05, |
|
"loss": 0.475, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.8916068077087402, |
|
"learning_rate": 9.307417582417583e-05, |
|
"loss": 0.4891, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 1.0133769512176514, |
|
"learning_rate": 9.30467032967033e-05, |
|
"loss": 0.4857, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 1.6718567609786987, |
|
"learning_rate": 9.301923076923078e-05, |
|
"loss": 0.4915, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 1.2999911308288574, |
|
"learning_rate": 9.299175824175824e-05, |
|
"loss": 0.4872, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 1.4812780618667603, |
|
"learning_rate": 9.296428571428571e-05, |
|
"loss": 0.4886, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 1.3728208541870117, |
|
"learning_rate": 9.293681318681318e-05, |
|
"loss": 0.5005, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 1.2773633003234863, |
|
"learning_rate": 9.290934065934066e-05, |
|
"loss": 0.4796, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.8197434544563293, |
|
"learning_rate": 9.288186813186813e-05, |
|
"loss": 0.4817, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 0.830235481262207, |
|
"learning_rate": 9.28543956043956e-05, |
|
"loss": 0.4956, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 3.1286449432373047, |
|
"learning_rate": 9.282692307692308e-05, |
|
"loss": 0.4822, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 0.9349818825721741, |
|
"learning_rate": 9.279945054945055e-05, |
|
"loss": 0.4924, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 1.8471746444702148, |
|
"learning_rate": 9.277197802197802e-05, |
|
"loss": 0.4771, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.9180948734283447, |
|
"learning_rate": 9.27445054945055e-05, |
|
"loss": 0.4855, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.7180835604667664, |
|
"learning_rate": 9.271703296703297e-05, |
|
"loss": 0.4801, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 1.5345978736877441, |
|
"learning_rate": 9.268956043956043e-05, |
|
"loss": 0.4791, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 1.521747350692749, |
|
"learning_rate": 9.266208791208792e-05, |
|
"loss": 0.4793, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 1.2364057302474976, |
|
"learning_rate": 9.263461538461538e-05, |
|
"loss": 0.4756, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 0.5741211771965027, |
|
"learning_rate": 9.260714285714285e-05, |
|
"loss": 0.4849, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 1.7251524925231934, |
|
"learning_rate": 9.257967032967033e-05, |
|
"loss": 0.4762, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 2.704491138458252, |
|
"learning_rate": 9.25521978021978e-05, |
|
"loss": 0.4777, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.9178886413574219, |
|
"learning_rate": 9.252472527472527e-05, |
|
"loss": 0.5077, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 1.2007173299789429, |
|
"learning_rate": 9.249725274725275e-05, |
|
"loss": 0.4916, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.8556911945343018, |
|
"learning_rate": 9.246978021978022e-05, |
|
"loss": 0.477, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.7754554152488708, |
|
"learning_rate": 9.244230769230769e-05, |
|
"loss": 0.4859, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 2.333521842956543, |
|
"learning_rate": 9.241483516483517e-05, |
|
"loss": 0.4811, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.9892757534980774, |
|
"learning_rate": 9.238736263736264e-05, |
|
"loss": 0.4777, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 1.0326083898544312, |
|
"learning_rate": 9.23598901098901e-05, |
|
"loss": 0.4891, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 1.4555460214614868, |
|
"learning_rate": 9.233241758241759e-05, |
|
"loss": 0.4802, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 0.8390610814094543, |
|
"learning_rate": 9.230494505494506e-05, |
|
"loss": 0.4835, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 1.0490790605545044, |
|
"learning_rate": 9.227747252747252e-05, |
|
"loss": 0.4806, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.8793150186538696, |
|
"learning_rate": 9.225e-05, |
|
"loss": 0.4846, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 1.2842731475830078, |
|
"learning_rate": 9.222252747252747e-05, |
|
"loss": 0.4715, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.7668898701667786, |
|
"learning_rate": 9.219505494505494e-05, |
|
"loss": 0.4749, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.8144903182983398, |
|
"learning_rate": 9.216758241758242e-05, |
|
"loss": 0.4705, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.6727710962295532, |
|
"learning_rate": 9.214010989010989e-05, |
|
"loss": 0.4826, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.8702958822250366, |
|
"learning_rate": 9.211263736263736e-05, |
|
"loss": 0.4679, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.9223732352256775, |
|
"learning_rate": 9.208516483516484e-05, |
|
"loss": 0.4875, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.7086817622184753, |
|
"learning_rate": 9.205769230769231e-05, |
|
"loss": 0.4771, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 1.000447154045105, |
|
"learning_rate": 9.203021978021978e-05, |
|
"loss": 0.4674, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 1.6982853412628174, |
|
"learning_rate": 9.200274725274726e-05, |
|
"loss": 0.4853, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 0.8975854516029358, |
|
"learning_rate": 9.197527472527473e-05, |
|
"loss": 0.4757, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 1.638816237449646, |
|
"learning_rate": 9.19478021978022e-05, |
|
"loss": 0.4862, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 1.0290194749832153, |
|
"learning_rate": 9.192032967032968e-05, |
|
"loss": 0.4758, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 1.0205775499343872, |
|
"learning_rate": 9.189285714285715e-05, |
|
"loss": 0.4723, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 1.2193140983581543, |
|
"learning_rate": 9.186538461538461e-05, |
|
"loss": 0.4869, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 1.0402013063430786, |
|
"learning_rate": 9.18379120879121e-05, |
|
"loss": 0.4911, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 1.1915889978408813, |
|
"learning_rate": 9.181043956043956e-05, |
|
"loss": 0.4757, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 1.2371594905853271, |
|
"learning_rate": 9.178296703296703e-05, |
|
"loss": 0.4811, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.9872990846633911, |
|
"learning_rate": 9.175549450549451e-05, |
|
"loss": 0.4784, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.9502705335617065, |
|
"learning_rate": 9.172802197802198e-05, |
|
"loss": 0.4716, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 1.9108206033706665, |
|
"learning_rate": 9.170054945054945e-05, |
|
"loss": 0.4828, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.5662224888801575, |
|
"learning_rate": 9.167307692307693e-05, |
|
"loss": 0.485, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.9142247438430786, |
|
"learning_rate": 9.16456043956044e-05, |
|
"loss": 0.4851, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 1.817579984664917, |
|
"learning_rate": 9.161813186813187e-05, |
|
"loss": 0.4807, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 1.3611973524093628, |
|
"learning_rate": 9.159065934065935e-05, |
|
"loss": 0.4815, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 1.4995872974395752, |
|
"learning_rate": 9.156318681318682e-05, |
|
"loss": 0.4779, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.8315975069999695, |
|
"learning_rate": 9.153571428571429e-05, |
|
"loss": 0.4732, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 1.285447359085083, |
|
"learning_rate": 9.150824175824177e-05, |
|
"loss": 0.4769, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 1.265435814857483, |
|
"learning_rate": 9.148076923076924e-05, |
|
"loss": 0.4884, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.7437517046928406, |
|
"learning_rate": 9.14532967032967e-05, |
|
"loss": 0.4807, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.8307008743286133, |
|
"learning_rate": 9.142582417582419e-05, |
|
"loss": 0.4723, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 1.2148619890213013, |
|
"learning_rate": 9.139835164835165e-05, |
|
"loss": 0.461, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.5682019591331482, |
|
"learning_rate": 9.137087912087912e-05, |
|
"loss": 0.4673, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.582708477973938, |
|
"learning_rate": 9.13434065934066e-05, |
|
"loss": 0.4657, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 1.0724083185195923, |
|
"learning_rate": 9.131593406593407e-05, |
|
"loss": 0.4634, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 1.511168360710144, |
|
"learning_rate": 9.128846153846154e-05, |
|
"loss": 0.4888, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 1.5445438623428345, |
|
"learning_rate": 9.126098901098902e-05, |
|
"loss": 0.4757, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 0.9609503746032715, |
|
"learning_rate": 9.123351648351649e-05, |
|
"loss": 0.4813, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 2.094918966293335, |
|
"learning_rate": 9.120604395604396e-05, |
|
"loss": 0.4963, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 1.6844788789749146, |
|
"learning_rate": 9.117857142857144e-05, |
|
"loss": 0.4902, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 1.2747464179992676, |
|
"learning_rate": 9.115109890109891e-05, |
|
"loss": 0.4884, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 1.4035594463348389, |
|
"learning_rate": 9.112362637362638e-05, |
|
"loss": 0.482, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 0.6047748923301697, |
|
"learning_rate": 9.109615384615386e-05, |
|
"loss": 0.4846, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 1.6141401529312134, |
|
"learning_rate": 9.106868131868133e-05, |
|
"loss": 0.4744, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.8329553008079529, |
|
"learning_rate": 9.10412087912088e-05, |
|
"loss": 0.4738, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.7500654458999634, |
|
"learning_rate": 9.101373626373628e-05, |
|
"loss": 0.4763, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.8617835640907288, |
|
"learning_rate": 9.098626373626374e-05, |
|
"loss": 0.4738, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 1.0339410305023193, |
|
"learning_rate": 9.095879120879121e-05, |
|
"loss": 0.4726, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 1.13742995262146, |
|
"learning_rate": 9.09313186813187e-05, |
|
"loss": 0.4827, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 1.611344575881958, |
|
"learning_rate": 9.090384615384616e-05, |
|
"loss": 0.4844, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 1.056773066520691, |
|
"learning_rate": 9.087637362637363e-05, |
|
"loss": 0.4799, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 1.2290446758270264, |
|
"learning_rate": 9.084890109890111e-05, |
|
"loss": 0.4764, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.7461050748825073, |
|
"learning_rate": 9.082142857142858e-05, |
|
"loss": 0.4664, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.5372506380081177, |
|
"learning_rate": 9.079395604395605e-05, |
|
"loss": 0.4795, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 1.658233642578125, |
|
"learning_rate": 9.076648351648353e-05, |
|
"loss": 0.4655, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 1.7918599843978882, |
|
"learning_rate": 9.0739010989011e-05, |
|
"loss": 0.4777, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.9504883289337158, |
|
"learning_rate": 9.071153846153847e-05, |
|
"loss": 0.4791, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 1.5631986856460571, |
|
"learning_rate": 9.068406593406595e-05, |
|
"loss": 0.4807, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.9991289377212524, |
|
"learning_rate": 9.065659340659342e-05, |
|
"loss": 0.4679, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 1.130434274673462, |
|
"learning_rate": 9.062912087912088e-05, |
|
"loss": 0.4781, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.639451265335083, |
|
"learning_rate": 9.060164835164837e-05, |
|
"loss": 0.4815, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 1.1776291131973267, |
|
"learning_rate": 9.057417582417583e-05, |
|
"loss": 0.4983, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 1.4073742628097534, |
|
"learning_rate": 9.05467032967033e-05, |
|
"loss": 0.4711, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 1.308344841003418, |
|
"learning_rate": 9.051923076923078e-05, |
|
"loss": 0.4847, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 0.44810301065444946, |
|
"eval_runtime": 1256.2985, |
|
"eval_samples_per_second": 232.516, |
|
"eval_steps_per_second": 3.634, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.6774665117263794, |
|
"learning_rate": 9.049175824175825e-05, |
|
"loss": 0.4791, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 0.6826792359352112, |
|
"learning_rate": 9.046428571428572e-05, |
|
"loss": 0.4775, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 0.7845907211303711, |
|
"learning_rate": 9.04368131868132e-05, |
|
"loss": 0.4765, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 0.9099870324134827, |
|
"learning_rate": 9.040934065934067e-05, |
|
"loss": 0.4689, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 1.374972939491272, |
|
"learning_rate": 9.038186813186814e-05, |
|
"loss": 0.469, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 0.8698770403862, |
|
"learning_rate": 9.035439560439562e-05, |
|
"loss": 0.4717, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 0.8832442164421082, |
|
"learning_rate": 9.032692307692309e-05, |
|
"loss": 0.4653, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 0.722259521484375, |
|
"learning_rate": 9.029945054945054e-05, |
|
"loss": 0.4818, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 1.1362375020980835, |
|
"learning_rate": 9.027197802197802e-05, |
|
"loss": 0.4674, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"grad_norm": 1.0420037508010864, |
|
"learning_rate": 9.024450549450549e-05, |
|
"loss": 0.4806, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"grad_norm": 0.9857313632965088, |
|
"learning_rate": 9.021703296703296e-05, |
|
"loss": 0.478, |
|
"step": 3770 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"grad_norm": 1.1428401470184326, |
|
"learning_rate": 9.018956043956044e-05, |
|
"loss": 0.4742, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 1.5837888717651367, |
|
"learning_rate": 9.016208791208791e-05, |
|
"loss": 0.4841, |
|
"step": 3790 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 1.9123620986938477, |
|
"learning_rate": 9.013461538461538e-05, |
|
"loss": 0.4707, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 0.9019024968147278, |
|
"learning_rate": 9.010714285714286e-05, |
|
"loss": 0.4731, |
|
"step": 3810 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 0.9601519107818604, |
|
"learning_rate": 9.007967032967033e-05, |
|
"loss": 0.4644, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"grad_norm": 0.4769773483276367, |
|
"learning_rate": 9.00521978021978e-05, |
|
"loss": 0.4755, |
|
"step": 3830 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"grad_norm": 1.2578529119491577, |
|
"learning_rate": 9.002472527472528e-05, |
|
"loss": 0.4703, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"grad_norm": 1.1451194286346436, |
|
"learning_rate": 8.999725274725275e-05, |
|
"loss": 0.4748, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"grad_norm": 0.4718180000782013, |
|
"learning_rate": 8.996978021978021e-05, |
|
"loss": 0.4673, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"grad_norm": 1.0323463678359985, |
|
"learning_rate": 8.99423076923077e-05, |
|
"loss": 0.4611, |
|
"step": 3870 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"grad_norm": 0.905623197555542, |
|
"learning_rate": 8.991483516483516e-05, |
|
"loss": 0.4651, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"grad_norm": 0.6803371906280518, |
|
"learning_rate": 8.988736263736263e-05, |
|
"loss": 0.4693, |
|
"step": 3890 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 0.7653062343597412, |
|
"learning_rate": 8.985989010989011e-05, |
|
"loss": 0.4676, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 0.6493901610374451, |
|
"learning_rate": 8.983241758241758e-05, |
|
"loss": 0.4796, |
|
"step": 3910 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 1.3194290399551392, |
|
"learning_rate": 8.980494505494505e-05, |
|
"loss": 0.4709, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 1.3907560110092163, |
|
"learning_rate": 8.977747252747253e-05, |
|
"loss": 0.4791, |
|
"step": 3930 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"grad_norm": 0.9326291084289551, |
|
"learning_rate": 8.975e-05, |
|
"loss": 0.4733, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"grad_norm": 0.6711992025375366, |
|
"learning_rate": 8.972252747252747e-05, |
|
"loss": 0.4607, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"grad_norm": 1.0052268505096436, |
|
"learning_rate": 8.969505494505495e-05, |
|
"loss": 0.4565, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"grad_norm": 1.893130898475647, |
|
"learning_rate": 8.966758241758242e-05, |
|
"loss": 0.4678, |
|
"step": 3970 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"grad_norm": 1.019831657409668, |
|
"learning_rate": 8.964010989010989e-05, |
|
"loss": 0.4929, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"grad_norm": 1.096341848373413, |
|
"learning_rate": 8.961263736263737e-05, |
|
"loss": 0.4812, |
|
"step": 3990 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"grad_norm": 1.3617221117019653, |
|
"learning_rate": 8.958516483516484e-05, |
|
"loss": 0.479, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"grad_norm": 0.6594845056533813, |
|
"learning_rate": 8.95576923076923e-05, |
|
"loss": 0.4784, |
|
"step": 4010 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"grad_norm": 0.7108995914459229, |
|
"learning_rate": 8.953021978021979e-05, |
|
"loss": 0.4624, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"grad_norm": 0.4914230704307556, |
|
"learning_rate": 8.950274725274725e-05, |
|
"loss": 0.4573, |
|
"step": 4030 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"grad_norm": 0.738325834274292, |
|
"learning_rate": 8.947527472527472e-05, |
|
"loss": 0.4598, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"grad_norm": 1.5622313022613525, |
|
"learning_rate": 8.94478021978022e-05, |
|
"loss": 0.4769, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"grad_norm": 0.84259033203125, |
|
"learning_rate": 8.942032967032967e-05, |
|
"loss": 0.4587, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"grad_norm": 1.1515220403671265, |
|
"learning_rate": 8.939285714285714e-05, |
|
"loss": 0.4698, |
|
"step": 4070 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"grad_norm": 1.2757478952407837, |
|
"learning_rate": 8.936538461538462e-05, |
|
"loss": 0.4662, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"grad_norm": 1.413041591644287, |
|
"learning_rate": 8.933791208791209e-05, |
|
"loss": 0.4739, |
|
"step": 4090 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"grad_norm": 1.3950985670089722, |
|
"learning_rate": 8.931043956043956e-05, |
|
"loss": 0.4652, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"grad_norm": 1.1247773170471191, |
|
"learning_rate": 8.928296703296704e-05, |
|
"loss": 0.4704, |
|
"step": 4110 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"grad_norm": 0.7197868824005127, |
|
"learning_rate": 8.925549450549451e-05, |
|
"loss": 0.466, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"grad_norm": 1.606062889099121, |
|
"learning_rate": 8.922802197802198e-05, |
|
"loss": 0.4697, |
|
"step": 4130 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"grad_norm": 0.8834720849990845, |
|
"learning_rate": 8.920054945054946e-05, |
|
"loss": 0.4787, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"grad_norm": 0.6989302635192871, |
|
"learning_rate": 8.917307692307693e-05, |
|
"loss": 0.4553, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"grad_norm": 1.1178115606307983, |
|
"learning_rate": 8.91456043956044e-05, |
|
"loss": 0.4679, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"grad_norm": 0.8493313789367676, |
|
"learning_rate": 8.911813186813188e-05, |
|
"loss": 0.4915, |
|
"step": 4170 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"grad_norm": 0.6862457990646362, |
|
"learning_rate": 8.909065934065934e-05, |
|
"loss": 0.4667, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"grad_norm": 0.5265287160873413, |
|
"learning_rate": 8.906318681318681e-05, |
|
"loss": 0.4675, |
|
"step": 4190 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"grad_norm": 1.6895486116409302, |
|
"learning_rate": 8.903571428571429e-05, |
|
"loss": 0.4694, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"grad_norm": 0.9672292470932007, |
|
"learning_rate": 8.900824175824176e-05, |
|
"loss": 0.4788, |
|
"step": 4210 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"grad_norm": 0.9348291158676147, |
|
"learning_rate": 8.898076923076923e-05, |
|
"loss": 0.461, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 0.5047799348831177, |
|
"learning_rate": 8.895329670329671e-05, |
|
"loss": 0.467, |
|
"step": 4230 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 1.6985957622528076, |
|
"learning_rate": 8.892582417582418e-05, |
|
"loss": 0.4581, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 0.8573766946792603, |
|
"learning_rate": 8.889835164835165e-05, |
|
"loss": 0.4624, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 0.6403746008872986, |
|
"learning_rate": 8.887087912087913e-05, |
|
"loss": 0.4574, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"grad_norm": 0.9546878933906555, |
|
"learning_rate": 8.88434065934066e-05, |
|
"loss": 0.4803, |
|
"step": 4270 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"grad_norm": 0.6992652416229248, |
|
"learning_rate": 8.881593406593407e-05, |
|
"loss": 0.4696, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"grad_norm": 1.035543441772461, |
|
"learning_rate": 8.878846153846155e-05, |
|
"loss": 0.4787, |
|
"step": 4290 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"grad_norm": 1.5563474893569946, |
|
"learning_rate": 8.876098901098902e-05, |
|
"loss": 0.4708, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"grad_norm": 1.7203930616378784, |
|
"learning_rate": 8.873351648351648e-05, |
|
"loss": 0.4725, |
|
"step": 4310 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"grad_norm": 1.4295490980148315, |
|
"learning_rate": 8.870604395604396e-05, |
|
"loss": 0.464, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"grad_norm": 1.1832698583602905, |
|
"learning_rate": 8.867857142857143e-05, |
|
"loss": 0.475, |
|
"step": 4330 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"grad_norm": 0.9679028391838074, |
|
"learning_rate": 8.86510989010989e-05, |
|
"loss": 0.4656, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"grad_norm": 0.9646287560462952, |
|
"learning_rate": 8.862362637362638e-05, |
|
"loss": 0.4707, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"grad_norm": 1.2284992933273315, |
|
"learning_rate": 8.859615384615385e-05, |
|
"loss": 0.4812, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"grad_norm": 0.7886459231376648, |
|
"learning_rate": 8.856868131868132e-05, |
|
"loss": 0.4641, |
|
"step": 4370 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 0.8464663028717041, |
|
"learning_rate": 8.85412087912088e-05, |
|
"loss": 0.4685, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 1.0549769401550293, |
|
"learning_rate": 8.851373626373627e-05, |
|
"loss": 0.4573, |
|
"step": 4390 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 1.3452584743499756, |
|
"learning_rate": 8.848626373626374e-05, |
|
"loss": 0.4823, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 1.5250259637832642, |
|
"learning_rate": 8.845879120879122e-05, |
|
"loss": 0.4762, |
|
"step": 4410 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"grad_norm": 0.7843578457832336, |
|
"learning_rate": 8.843131868131869e-05, |
|
"loss": 0.4714, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"grad_norm": 1.0666967630386353, |
|
"learning_rate": 8.840384615384616e-05, |
|
"loss": 0.4743, |
|
"step": 4430 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"grad_norm": 1.4348440170288086, |
|
"learning_rate": 8.837637362637364e-05, |
|
"loss": 0.4879, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"grad_norm": 1.036281704902649, |
|
"learning_rate": 8.83489010989011e-05, |
|
"loss": 0.469, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"grad_norm": 0.7014042139053345, |
|
"learning_rate": 8.832142857142857e-05, |
|
"loss": 0.4716, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"grad_norm": 1.593533992767334, |
|
"learning_rate": 8.829395604395605e-05, |
|
"loss": 0.4745, |
|
"step": 4470 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"grad_norm": 0.4499157965183258, |
|
"learning_rate": 8.826648351648352e-05, |
|
"loss": 0.4764, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"grad_norm": 0.7330986857414246, |
|
"learning_rate": 8.823901098901099e-05, |
|
"loss": 0.4733, |
|
"step": 4490 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"grad_norm": 0.7915944457054138, |
|
"learning_rate": 8.821153846153847e-05, |
|
"loss": 0.4605, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"grad_norm": 0.9147601127624512, |
|
"learning_rate": 8.818406593406594e-05, |
|
"loss": 0.4729, |
|
"step": 4510 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"grad_norm": 1.1145246028900146, |
|
"learning_rate": 8.815659340659341e-05, |
|
"loss": 0.4757, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 1.0525473356246948, |
|
"learning_rate": 8.812912087912089e-05, |
|
"loss": 0.4585, |
|
"step": 4530 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 0.7515787482261658, |
|
"learning_rate": 8.810164835164836e-05, |
|
"loss": 0.4557, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 0.9737600088119507, |
|
"learning_rate": 8.807417582417583e-05, |
|
"loss": 0.4647, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 1.1443486213684082, |
|
"learning_rate": 8.804670329670331e-05, |
|
"loss": 0.4685, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 1.4745419025421143, |
|
"learning_rate": 8.801923076923078e-05, |
|
"loss": 0.4736, |
|
"step": 4570 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 1.1491049528121948, |
|
"learning_rate": 8.799175824175824e-05, |
|
"loss": 0.4604, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 0.742324948310852, |
|
"learning_rate": 8.796428571428573e-05, |
|
"loss": 0.457, |
|
"step": 4590 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"grad_norm": 1.1914622783660889, |
|
"learning_rate": 8.79368131868132e-05, |
|
"loss": 0.4638, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"grad_norm": 0.6555135846138, |
|
"learning_rate": 8.790934065934066e-05, |
|
"loss": 0.4806, |
|
"step": 4610 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"grad_norm": 0.8627590537071228, |
|
"learning_rate": 8.788186813186814e-05, |
|
"loss": 0.4803, |
|
"step": 4620 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"grad_norm": 1.0868836641311646, |
|
"learning_rate": 8.785439560439561e-05, |
|
"loss": 0.4598, |
|
"step": 4630 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"grad_norm": 1.291483998298645, |
|
"learning_rate": 8.782692307692308e-05, |
|
"loss": 0.4707, |
|
"step": 4640 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"grad_norm": 0.9766128659248352, |
|
"learning_rate": 8.779945054945056e-05, |
|
"loss": 0.4707, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"grad_norm": 1.0301162004470825, |
|
"learning_rate": 8.777197802197803e-05, |
|
"loss": 0.4821, |
|
"step": 4660 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 1.0474865436553955, |
|
"learning_rate": 8.77445054945055e-05, |
|
"loss": 0.4603, |
|
"step": 4670 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 1.1218842267990112, |
|
"learning_rate": 8.771703296703298e-05, |
|
"loss": 0.4827, |
|
"step": 4680 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 1.0587611198425293, |
|
"learning_rate": 8.768956043956045e-05, |
|
"loss": 0.4746, |
|
"step": 4690 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 0.8249474763870239, |
|
"learning_rate": 8.766208791208792e-05, |
|
"loss": 0.4648, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"grad_norm": 0.6225512027740479, |
|
"learning_rate": 8.76346153846154e-05, |
|
"loss": 0.4582, |
|
"step": 4710 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"grad_norm": 1.2126622200012207, |
|
"learning_rate": 8.760714285714287e-05, |
|
"loss": 0.4617, |
|
"step": 4720 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"grad_norm": 1.6850072145462036, |
|
"learning_rate": 8.757967032967033e-05, |
|
"loss": 0.4793, |
|
"step": 4730 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"grad_norm": 0.5053130984306335, |
|
"learning_rate": 8.755219780219782e-05, |
|
"loss": 0.4648, |
|
"step": 4740 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"grad_norm": 0.4924371838569641, |
|
"learning_rate": 8.752472527472528e-05, |
|
"loss": 0.4694, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"grad_norm": 0.7182402014732361, |
|
"learning_rate": 8.749725274725275e-05, |
|
"loss": 0.4611, |
|
"step": 4760 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"grad_norm": 0.5569953322410583, |
|
"learning_rate": 8.746978021978023e-05, |
|
"loss": 0.4632, |
|
"step": 4770 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"grad_norm": 1.0777251720428467, |
|
"learning_rate": 8.74423076923077e-05, |
|
"loss": 0.4735, |
|
"step": 4780 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"grad_norm": 0.6310242414474487, |
|
"learning_rate": 8.741483516483517e-05, |
|
"loss": 0.4567, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"grad_norm": 0.94218510389328, |
|
"learning_rate": 8.738736263736265e-05, |
|
"loss": 0.4626, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"grad_norm": 0.4554530084133148, |
|
"learning_rate": 8.735989010989012e-05, |
|
"loss": 0.4681, |
|
"step": 4810 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 1.2381364107131958, |
|
"learning_rate": 8.733241758241759e-05, |
|
"loss": 0.467, |
|
"step": 4820 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 1.1469919681549072, |
|
"learning_rate": 8.730494505494507e-05, |
|
"loss": 0.4643, |
|
"step": 4830 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 0.7159097194671631, |
|
"learning_rate": 8.727747252747254e-05, |
|
"loss": 0.4731, |
|
"step": 4840 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 0.8695803880691528, |
|
"learning_rate": 8.725e-05, |
|
"loss": 0.4646, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"grad_norm": 0.8535722494125366, |
|
"learning_rate": 8.722252747252749e-05, |
|
"loss": 0.4612, |
|
"step": 4860 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"grad_norm": 1.231641173362732, |
|
"learning_rate": 8.719505494505496e-05, |
|
"loss": 0.48, |
|
"step": 4870 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"grad_norm": 1.0003236532211304, |
|
"learning_rate": 8.716758241758241e-05, |
|
"loss": 0.4741, |
|
"step": 4880 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"grad_norm": 1.8520704507827759, |
|
"learning_rate": 8.714010989010989e-05, |
|
"loss": 0.4693, |
|
"step": 4890 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"grad_norm": 1.0661410093307495, |
|
"learning_rate": 8.711263736263736e-05, |
|
"loss": 0.4848, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"grad_norm": 0.8280912637710571, |
|
"learning_rate": 8.708516483516483e-05, |
|
"loss": 0.4703, |
|
"step": 4910 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"grad_norm": 1.169047474861145, |
|
"learning_rate": 8.705769230769231e-05, |
|
"loss": 0.4777, |
|
"step": 4920 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"grad_norm": 0.7568297982215881, |
|
"learning_rate": 8.703021978021978e-05, |
|
"loss": 0.4752, |
|
"step": 4930 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"grad_norm": 1.5538333654403687, |
|
"learning_rate": 8.700274725274725e-05, |
|
"loss": 0.4569, |
|
"step": 4940 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"grad_norm": 0.6673538684844971, |
|
"learning_rate": 8.697527472527473e-05, |
|
"loss": 0.4659, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"grad_norm": 0.9370101690292358, |
|
"learning_rate": 8.69478021978022e-05, |
|
"loss": 0.4644, |
|
"step": 4960 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"grad_norm": 1.748659610748291, |
|
"learning_rate": 8.692032967032966e-05, |
|
"loss": 0.458, |
|
"step": 4970 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"grad_norm": 0.8026989102363586, |
|
"learning_rate": 8.689285714285715e-05, |
|
"loss": 0.4696, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"grad_norm": 1.7652183771133423, |
|
"learning_rate": 8.686538461538461e-05, |
|
"loss": 0.4676, |
|
"step": 4990 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"grad_norm": 0.6286242008209229, |
|
"learning_rate": 8.683791208791208e-05, |
|
"loss": 0.4638, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"grad_norm": 2.5077078342437744, |
|
"learning_rate": 8.681043956043956e-05, |
|
"loss": 0.4774, |
|
"step": 5010 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"grad_norm": 0.6874992251396179, |
|
"learning_rate": 8.678296703296703e-05, |
|
"loss": 0.4621, |
|
"step": 5020 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"grad_norm": 1.2562603950500488, |
|
"learning_rate": 8.67554945054945e-05, |
|
"loss": 0.4745, |
|
"step": 5030 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"grad_norm": 1.0717673301696777, |
|
"learning_rate": 8.672802197802198e-05, |
|
"loss": 0.4726, |
|
"step": 5040 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"grad_norm": 0.744522750377655, |
|
"learning_rate": 8.670054945054945e-05, |
|
"loss": 0.4823, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"grad_norm": 0.879859447479248, |
|
"learning_rate": 8.667307692307692e-05, |
|
"loss": 0.4651, |
|
"step": 5060 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"grad_norm": 0.6319553852081299, |
|
"learning_rate": 8.66456043956044e-05, |
|
"loss": 0.4714, |
|
"step": 5070 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"grad_norm": 0.8513883948326111, |
|
"learning_rate": 8.661813186813187e-05, |
|
"loss": 0.4644, |
|
"step": 5080 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"grad_norm": 1.7870076894760132, |
|
"learning_rate": 8.659065934065934e-05, |
|
"loss": 0.4714, |
|
"step": 5090 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"grad_norm": 0.8447625041007996, |
|
"learning_rate": 8.656318681318682e-05, |
|
"loss": 0.4717, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 0.7471101880073547, |
|
"learning_rate": 8.653571428571429e-05, |
|
"loss": 0.4594, |
|
"step": 5110 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 0.7438103556632996, |
|
"learning_rate": 8.650824175824175e-05, |
|
"loss": 0.4622, |
|
"step": 5120 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 0.8779565095901489, |
|
"learning_rate": 8.648076923076924e-05, |
|
"loss": 0.4639, |
|
"step": 5130 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 1.4103574752807617, |
|
"learning_rate": 8.64532967032967e-05, |
|
"loss": 0.4568, |
|
"step": 5140 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"grad_norm": 1.2644096612930298, |
|
"learning_rate": 8.642582417582417e-05, |
|
"loss": 0.4696, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"grad_norm": 0.7385484576225281, |
|
"learning_rate": 8.639835164835165e-05, |
|
"loss": 0.4638, |
|
"step": 5160 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"grad_norm": 0.7698835730552673, |
|
"learning_rate": 8.637087912087912e-05, |
|
"loss": 0.4762, |
|
"step": 5170 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 1.049194574356079, |
|
"learning_rate": 8.634340659340659e-05, |
|
"loss": 0.4588, |
|
"step": 5180 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 0.7376847267150879, |
|
"learning_rate": 8.631593406593407e-05, |
|
"loss": 0.46, |
|
"step": 5190 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 1.1413201093673706, |
|
"learning_rate": 8.628846153846154e-05, |
|
"loss": 0.4662, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 0.9843841195106506, |
|
"learning_rate": 8.626098901098901e-05, |
|
"loss": 0.4663, |
|
"step": 5210 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"grad_norm": 0.6186099052429199, |
|
"learning_rate": 8.623351648351649e-05, |
|
"loss": 0.4683, |
|
"step": 5220 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"grad_norm": 1.4082993268966675, |
|
"learning_rate": 8.620604395604396e-05, |
|
"loss": 0.4639, |
|
"step": 5230 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"grad_norm": 0.9625537395477295, |
|
"learning_rate": 8.617857142857143e-05, |
|
"loss": 0.4646, |
|
"step": 5240 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"grad_norm": 1.0915727615356445, |
|
"learning_rate": 8.615109890109891e-05, |
|
"loss": 0.4669, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"grad_norm": 0.9645349383354187, |
|
"learning_rate": 8.612362637362638e-05, |
|
"loss": 0.4572, |
|
"step": 5260 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"grad_norm": 0.7993157505989075, |
|
"learning_rate": 8.609615384615384e-05, |
|
"loss": 0.4623, |
|
"step": 5270 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"grad_norm": 1.3415577411651611, |
|
"learning_rate": 8.606868131868133e-05, |
|
"loss": 0.4569, |
|
"step": 5280 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"grad_norm": 1.1217197179794312, |
|
"learning_rate": 8.60412087912088e-05, |
|
"loss": 0.4632, |
|
"step": 5290 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"grad_norm": 0.6021980047225952, |
|
"learning_rate": 8.601373626373626e-05, |
|
"loss": 0.4688, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"grad_norm": 0.9672883749008179, |
|
"learning_rate": 8.598626373626374e-05, |
|
"loss": 0.4594, |
|
"step": 5310 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"grad_norm": 0.8534507155418396, |
|
"learning_rate": 8.595879120879121e-05, |
|
"loss": 0.4496, |
|
"step": 5320 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"grad_norm": 1.6172966957092285, |
|
"learning_rate": 8.593131868131868e-05, |
|
"loss": 0.4751, |
|
"step": 5330 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"grad_norm": 1.378061056137085, |
|
"learning_rate": 8.590384615384616e-05, |
|
"loss": 0.4597, |
|
"step": 5340 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"grad_norm": 0.8239535093307495, |
|
"learning_rate": 8.587637362637363e-05, |
|
"loss": 0.4549, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"grad_norm": 0.6672975420951843, |
|
"learning_rate": 8.58489010989011e-05, |
|
"loss": 0.4694, |
|
"step": 5360 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"grad_norm": 1.042051076889038, |
|
"learning_rate": 8.582142857142858e-05, |
|
"loss": 0.4667, |
|
"step": 5370 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"grad_norm": 1.6398990154266357, |
|
"learning_rate": 8.579395604395605e-05, |
|
"loss": 0.4657, |
|
"step": 5380 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"grad_norm": 0.5243840217590332, |
|
"learning_rate": 8.576648351648352e-05, |
|
"loss": 0.4733, |
|
"step": 5390 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"grad_norm": 1.3996922969818115, |
|
"learning_rate": 8.5739010989011e-05, |
|
"loss": 0.4747, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"grad_norm": 0.7381965517997742, |
|
"learning_rate": 8.571153846153847e-05, |
|
"loss": 0.4694, |
|
"step": 5410 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"grad_norm": 1.1151881217956543, |
|
"learning_rate": 8.568406593406593e-05, |
|
"loss": 0.4658, |
|
"step": 5420 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"grad_norm": 1.0418914556503296, |
|
"learning_rate": 8.565659340659342e-05, |
|
"loss": 0.4568, |
|
"step": 5430 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"grad_norm": 0.7290151119232178, |
|
"learning_rate": 8.562912087912088e-05, |
|
"loss": 0.4642, |
|
"step": 5440 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"grad_norm": 0.8483603000640869, |
|
"learning_rate": 8.560164835164835e-05, |
|
"loss": 0.4912, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"grad_norm": 1.1545953750610352, |
|
"learning_rate": 8.557417582417583e-05, |
|
"loss": 0.4738, |
|
"step": 5460 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"grad_norm": 1.0104776620864868, |
|
"learning_rate": 8.55467032967033e-05, |
|
"loss": 0.4751, |
|
"step": 5470 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"grad_norm": 0.9386792778968811, |
|
"learning_rate": 8.551923076923077e-05, |
|
"loss": 0.4549, |
|
"step": 5480 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"grad_norm": 0.6779844760894775, |
|
"learning_rate": 8.549175824175825e-05, |
|
"loss": 0.4682, |
|
"step": 5490 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"grad_norm": 0.7957237958908081, |
|
"learning_rate": 8.546428571428572e-05, |
|
"loss": 0.4561, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"grad_norm": 0.8393499255180359, |
|
"learning_rate": 8.543681318681319e-05, |
|
"loss": 0.4659, |
|
"step": 5510 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"grad_norm": 0.9744943380355835, |
|
"learning_rate": 8.540934065934067e-05, |
|
"loss": 0.4545, |
|
"step": 5520 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"grad_norm": 1.1456732749938965, |
|
"learning_rate": 8.538186813186814e-05, |
|
"loss": 0.4614, |
|
"step": 5530 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"grad_norm": 0.8066684603691101, |
|
"learning_rate": 8.53543956043956e-05, |
|
"loss": 0.475, |
|
"step": 5540 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"grad_norm": 1.4929931163787842, |
|
"learning_rate": 8.532692307692309e-05, |
|
"loss": 0.4695, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"grad_norm": 0.7223564386367798, |
|
"learning_rate": 8.529945054945056e-05, |
|
"loss": 0.4692, |
|
"step": 5560 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"grad_norm": 1.3222299814224243, |
|
"learning_rate": 8.527197802197802e-05, |
|
"loss": 0.4533, |
|
"step": 5570 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"grad_norm": 1.1262595653533936, |
|
"learning_rate": 8.52445054945055e-05, |
|
"loss": 0.4637, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"grad_norm": 0.9097437858581543, |
|
"learning_rate": 8.521703296703297e-05, |
|
"loss": 0.4692, |
|
"step": 5590 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"grad_norm": 1.5121077299118042, |
|
"learning_rate": 8.518956043956044e-05, |
|
"loss": 0.4591, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"grad_norm": 1.1432225704193115, |
|
"learning_rate": 8.516208791208792e-05, |
|
"loss": 0.4723, |
|
"step": 5610 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"grad_norm": 1.3745416402816772, |
|
"learning_rate": 8.513461538461539e-05, |
|
"loss": 0.4725, |
|
"step": 5620 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"grad_norm": 0.8508789539337158, |
|
"learning_rate": 8.510714285714286e-05, |
|
"loss": 0.4607, |
|
"step": 5630 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"grad_norm": 1.3097485303878784, |
|
"learning_rate": 8.507967032967034e-05, |
|
"loss": 0.4635, |
|
"step": 5640 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"grad_norm": 1.29444420337677, |
|
"learning_rate": 8.505219780219781e-05, |
|
"loss": 0.4628, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"grad_norm": 0.5336117744445801, |
|
"learning_rate": 8.502472527472528e-05, |
|
"loss": 0.4568, |
|
"step": 5660 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"grad_norm": 0.7139732837677002, |
|
"learning_rate": 8.499725274725276e-05, |
|
"loss": 0.4574, |
|
"step": 5670 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"grad_norm": 0.7436287999153137, |
|
"learning_rate": 8.496978021978023e-05, |
|
"loss": 0.455, |
|
"step": 5680 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"grad_norm": 0.9436907172203064, |
|
"learning_rate": 8.49423076923077e-05, |
|
"loss": 0.4678, |
|
"step": 5690 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"grad_norm": 0.9827747344970703, |
|
"learning_rate": 8.491483516483518e-05, |
|
"loss": 0.4614, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"grad_norm": 0.7332632541656494, |
|
"learning_rate": 8.488736263736265e-05, |
|
"loss": 0.4675, |
|
"step": 5710 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"grad_norm": 1.1933000087738037, |
|
"learning_rate": 8.485989010989011e-05, |
|
"loss": 0.4528, |
|
"step": 5720 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"grad_norm": 1.1396616697311401, |
|
"learning_rate": 8.48324175824176e-05, |
|
"loss": 0.4663, |
|
"step": 5730 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"grad_norm": 1.4335391521453857, |
|
"learning_rate": 8.480494505494506e-05, |
|
"loss": 0.4658, |
|
"step": 5740 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"grad_norm": 0.85484778881073, |
|
"learning_rate": 8.477747252747253e-05, |
|
"loss": 0.4536, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"grad_norm": 1.206037163734436, |
|
"learning_rate": 8.475000000000001e-05, |
|
"loss": 0.4665, |
|
"step": 5760 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"grad_norm": 1.0052751302719116, |
|
"learning_rate": 8.472252747252748e-05, |
|
"loss": 0.4657, |
|
"step": 5770 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"grad_norm": 0.6977716088294983, |
|
"learning_rate": 8.469505494505495e-05, |
|
"loss": 0.4594, |
|
"step": 5780 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"grad_norm": 0.975903332233429, |
|
"learning_rate": 8.466758241758243e-05, |
|
"loss": 0.4739, |
|
"step": 5790 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"grad_norm": 1.0114597082138062, |
|
"learning_rate": 8.46401098901099e-05, |
|
"loss": 0.4732, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"grad_norm": 0.5334266424179077, |
|
"learning_rate": 8.461263736263737e-05, |
|
"loss": 0.4585, |
|
"step": 5810 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"grad_norm": 1.6815012693405151, |
|
"learning_rate": 8.458516483516485e-05, |
|
"loss": 0.4737, |
|
"step": 5820 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"grad_norm": 1.1055243015289307, |
|
"learning_rate": 8.455769230769232e-05, |
|
"loss": 0.4608, |
|
"step": 5830 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"grad_norm": 1.3854832649230957, |
|
"learning_rate": 8.453021978021979e-05, |
|
"loss": 0.4659, |
|
"step": 5840 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"grad_norm": 1.0487163066864014, |
|
"learning_rate": 8.450274725274727e-05, |
|
"loss": 0.4519, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"grad_norm": 1.0301164388656616, |
|
"learning_rate": 8.447527472527474e-05, |
|
"loss": 0.4598, |
|
"step": 5860 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"grad_norm": 0.7901486754417419, |
|
"learning_rate": 8.44478021978022e-05, |
|
"loss": 0.4727, |
|
"step": 5870 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"grad_norm": 0.6557927131652832, |
|
"learning_rate": 8.442032967032969e-05, |
|
"loss": 0.4523, |
|
"step": 5880 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"grad_norm": 0.8954731225967407, |
|
"learning_rate": 8.439285714285715e-05, |
|
"loss": 0.4672, |
|
"step": 5890 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"grad_norm": 0.6038153171539307, |
|
"learning_rate": 8.436538461538462e-05, |
|
"loss": 0.4604, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"grad_norm": 0.9118047952651978, |
|
"learning_rate": 8.433791208791209e-05, |
|
"loss": 0.4592, |
|
"step": 5910 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"grad_norm": 1.125707745552063, |
|
"learning_rate": 8.431043956043957e-05, |
|
"loss": 0.4713, |
|
"step": 5920 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"grad_norm": 0.7669263482093811, |
|
"learning_rate": 8.428296703296704e-05, |
|
"loss": 0.465, |
|
"step": 5930 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"grad_norm": 1.0012739896774292, |
|
"learning_rate": 8.425549450549451e-05, |
|
"loss": 0.4661, |
|
"step": 5940 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"grad_norm": 0.650155782699585, |
|
"learning_rate": 8.422802197802199e-05, |
|
"loss": 0.4649, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"grad_norm": 1.224027156829834, |
|
"learning_rate": 8.420054945054946e-05, |
|
"loss": 0.4486, |
|
"step": 5960 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"grad_norm": 1.350738763809204, |
|
"learning_rate": 8.417307692307693e-05, |
|
"loss": 0.4714, |
|
"step": 5970 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"grad_norm": 1.238620400428772, |
|
"learning_rate": 8.414560439560441e-05, |
|
"loss": 0.4644, |
|
"step": 5980 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"grad_norm": 1.4456909894943237, |
|
"learning_rate": 8.411813186813188e-05, |
|
"loss": 0.4674, |
|
"step": 5990 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"grad_norm": 0.9819515943527222, |
|
"learning_rate": 8.409065934065934e-05, |
|
"loss": 0.4681, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"grad_norm": 0.7437455058097839, |
|
"learning_rate": 8.406318681318683e-05, |
|
"loss": 0.4701, |
|
"step": 6010 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"grad_norm": 0.7628225088119507, |
|
"learning_rate": 8.403571428571428e-05, |
|
"loss": 0.4608, |
|
"step": 6020 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"grad_norm": 1.0726768970489502, |
|
"learning_rate": 8.400824175824176e-05, |
|
"loss": 0.461, |
|
"step": 6030 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"grad_norm": 0.799857497215271, |
|
"learning_rate": 8.398076923076923e-05, |
|
"loss": 0.4587, |
|
"step": 6040 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"grad_norm": 1.242092251777649, |
|
"learning_rate": 8.39532967032967e-05, |
|
"loss": 0.4569, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"grad_norm": 0.971057116985321, |
|
"learning_rate": 8.392582417582418e-05, |
|
"loss": 0.4537, |
|
"step": 6060 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"grad_norm": 1.1566013097763062, |
|
"learning_rate": 8.389835164835165e-05, |
|
"loss": 0.4638, |
|
"step": 6070 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"grad_norm": 1.2802190780639648, |
|
"learning_rate": 8.387087912087912e-05, |
|
"loss": 0.4663, |
|
"step": 6080 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"grad_norm": 1.0269036293029785, |
|
"learning_rate": 8.38434065934066e-05, |
|
"loss": 0.4628, |
|
"step": 6090 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"grad_norm": 0.7012946605682373, |
|
"learning_rate": 8.381593406593407e-05, |
|
"loss": 0.4513, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"grad_norm": 0.5895891189575195, |
|
"learning_rate": 8.378846153846153e-05, |
|
"loss": 0.4584, |
|
"step": 6110 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"grad_norm": 0.9521509408950806, |
|
"learning_rate": 8.376098901098902e-05, |
|
"loss": 0.4587, |
|
"step": 6120 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"grad_norm": 1.409681797027588, |
|
"learning_rate": 8.373351648351648e-05, |
|
"loss": 0.4592, |
|
"step": 6130 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"grad_norm": 1.0833555459976196, |
|
"learning_rate": 8.370604395604395e-05, |
|
"loss": 0.4562, |
|
"step": 6140 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"grad_norm": 0.8994373679161072, |
|
"learning_rate": 8.367857142857143e-05, |
|
"loss": 0.4593, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"grad_norm": 1.2591134309768677, |
|
"learning_rate": 8.36510989010989e-05, |
|
"loss": 0.4632, |
|
"step": 6160 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"grad_norm": 0.8163917064666748, |
|
"learning_rate": 8.362362637362637e-05, |
|
"loss": 0.4767, |
|
"step": 6170 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"grad_norm": 0.6897978186607361, |
|
"learning_rate": 8.359615384615385e-05, |
|
"loss": 0.4569, |
|
"step": 6180 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"grad_norm": 1.0527364015579224, |
|
"learning_rate": 8.356868131868132e-05, |
|
"loss": 0.4618, |
|
"step": 6190 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"grad_norm": 1.4118075370788574, |
|
"learning_rate": 8.354120879120879e-05, |
|
"loss": 0.4542, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"grad_norm": 0.9567866325378418, |
|
"learning_rate": 8.351373626373627e-05, |
|
"loss": 0.4673, |
|
"step": 6210 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"grad_norm": 1.0207067728042603, |
|
"learning_rate": 8.348626373626374e-05, |
|
"loss": 0.4538, |
|
"step": 6220 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"grad_norm": 0.5897815227508545, |
|
"learning_rate": 8.34587912087912e-05, |
|
"loss": 0.4577, |
|
"step": 6230 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"grad_norm": 0.5948817133903503, |
|
"learning_rate": 8.343131868131869e-05, |
|
"loss": 0.4629, |
|
"step": 6240 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"grad_norm": 1.6685130596160889, |
|
"learning_rate": 8.340384615384616e-05, |
|
"loss": 0.461, |
|
"step": 6250 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"grad_norm": 1.502945899963379, |
|
"learning_rate": 8.337637362637362e-05, |
|
"loss": 0.4649, |
|
"step": 6260 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"grad_norm": 1.5422593355178833, |
|
"learning_rate": 8.33489010989011e-05, |
|
"loss": 0.4494, |
|
"step": 6270 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"grad_norm": 0.999934732913971, |
|
"learning_rate": 8.332142857142857e-05, |
|
"loss": 0.464, |
|
"step": 6280 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"grad_norm": 0.6580365896224976, |
|
"learning_rate": 8.329395604395604e-05, |
|
"loss": 0.4382, |
|
"step": 6290 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"grad_norm": 0.9826914668083191, |
|
"learning_rate": 8.326648351648352e-05, |
|
"loss": 0.4646, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"grad_norm": 1.6006591320037842, |
|
"learning_rate": 8.323901098901099e-05, |
|
"loss": 0.4709, |
|
"step": 6310 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"grad_norm": 0.46454253792762756, |
|
"learning_rate": 8.321153846153846e-05, |
|
"loss": 0.4572, |
|
"step": 6320 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"grad_norm": 1.1660053730010986, |
|
"learning_rate": 8.318406593406594e-05, |
|
"loss": 0.4537, |
|
"step": 6330 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"grad_norm": 0.7951260805130005, |
|
"learning_rate": 8.315659340659341e-05, |
|
"loss": 0.4491, |
|
"step": 6340 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"grad_norm": 0.9338926672935486, |
|
"learning_rate": 8.312912087912088e-05, |
|
"loss": 0.4608, |
|
"step": 6350 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"grad_norm": 1.2365031242370605, |
|
"learning_rate": 8.310164835164836e-05, |
|
"loss": 0.4582, |
|
"step": 6360 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"grad_norm": 0.7614923119544983, |
|
"learning_rate": 8.307417582417583e-05, |
|
"loss": 0.4665, |
|
"step": 6370 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"grad_norm": 0.9391283392906189, |
|
"learning_rate": 8.30467032967033e-05, |
|
"loss": 0.4593, |
|
"step": 6380 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"grad_norm": 1.4633766412734985, |
|
"learning_rate": 8.301923076923078e-05, |
|
"loss": 0.4611, |
|
"step": 6390 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"grad_norm": 1.159623146057129, |
|
"learning_rate": 8.299175824175825e-05, |
|
"loss": 0.4622, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"grad_norm": 1.0221785306930542, |
|
"learning_rate": 8.296428571428571e-05, |
|
"loss": 0.4483, |
|
"step": 6410 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"grad_norm": 1.5070372819900513, |
|
"learning_rate": 8.29368131868132e-05, |
|
"loss": 0.4677, |
|
"step": 6420 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"grad_norm": 0.9225922226905823, |
|
"learning_rate": 8.290934065934066e-05, |
|
"loss": 0.4711, |
|
"step": 6430 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"grad_norm": 0.9344289302825928, |
|
"learning_rate": 8.288186813186813e-05, |
|
"loss": 0.4591, |
|
"step": 6440 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"grad_norm": 0.5448782444000244, |
|
"learning_rate": 8.285439560439561e-05, |
|
"loss": 0.4635, |
|
"step": 6450 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"grad_norm": 0.465524822473526, |
|
"learning_rate": 8.282692307692308e-05, |
|
"loss": 0.4581, |
|
"step": 6460 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"grad_norm": 1.3583608865737915, |
|
"learning_rate": 8.279945054945055e-05, |
|
"loss": 0.4718, |
|
"step": 6470 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"grad_norm": 1.3173762559890747, |
|
"learning_rate": 8.277197802197803e-05, |
|
"loss": 0.4577, |
|
"step": 6480 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"grad_norm": 1.3845728635787964, |
|
"learning_rate": 8.27445054945055e-05, |
|
"loss": 0.4623, |
|
"step": 6490 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"grad_norm": 1.0992422103881836, |
|
"learning_rate": 8.271703296703297e-05, |
|
"loss": 0.4543, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"grad_norm": 0.5556319355964661, |
|
"learning_rate": 8.268956043956045e-05, |
|
"loss": 0.4613, |
|
"step": 6510 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"grad_norm": 0.7095728516578674, |
|
"learning_rate": 8.266208791208792e-05, |
|
"loss": 0.455, |
|
"step": 6520 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"grad_norm": 1.1960381269454956, |
|
"learning_rate": 8.263461538461539e-05, |
|
"loss": 0.4545, |
|
"step": 6530 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"grad_norm": 1.0696423053741455, |
|
"learning_rate": 8.260714285714287e-05, |
|
"loss": 0.4666, |
|
"step": 6540 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"grad_norm": 0.5211924314498901, |
|
"learning_rate": 8.257967032967034e-05, |
|
"loss": 0.4535, |
|
"step": 6550 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"grad_norm": 1.1284767389297485, |
|
"learning_rate": 8.25521978021978e-05, |
|
"loss": 0.4633, |
|
"step": 6560 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"grad_norm": 1.3033208847045898, |
|
"learning_rate": 8.252472527472529e-05, |
|
"loss": 0.4812, |
|
"step": 6570 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"grad_norm": 1.2627853155136108, |
|
"learning_rate": 8.249725274725275e-05, |
|
"loss": 0.4676, |
|
"step": 6580 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"grad_norm": 0.5073243379592896, |
|
"learning_rate": 8.246978021978022e-05, |
|
"loss": 0.4638, |
|
"step": 6590 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"grad_norm": 0.6492868661880493, |
|
"learning_rate": 8.24423076923077e-05, |
|
"loss": 0.46, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"grad_norm": 0.8350780010223389, |
|
"learning_rate": 8.241483516483517e-05, |
|
"loss": 0.4549, |
|
"step": 6610 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"grad_norm": 1.3048648834228516, |
|
"learning_rate": 8.238736263736264e-05, |
|
"loss": 0.47, |
|
"step": 6620 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"grad_norm": 1.2212404012680054, |
|
"learning_rate": 8.235989010989012e-05, |
|
"loss": 0.4623, |
|
"step": 6630 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"grad_norm": 0.9617910981178284, |
|
"learning_rate": 8.233241758241759e-05, |
|
"loss": 0.4566, |
|
"step": 6640 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"grad_norm": 0.5445905923843384, |
|
"learning_rate": 8.230494505494506e-05, |
|
"loss": 0.458, |
|
"step": 6650 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"grad_norm": 0.6512438654899597, |
|
"learning_rate": 8.227747252747254e-05, |
|
"loss": 0.4516, |
|
"step": 6660 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"grad_norm": 1.2203713655471802, |
|
"learning_rate": 8.225000000000001e-05, |
|
"loss": 0.4566, |
|
"step": 6670 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"grad_norm": 0.5803031325340271, |
|
"learning_rate": 8.222252747252748e-05, |
|
"loss": 0.4622, |
|
"step": 6680 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"grad_norm": 0.5457717180252075, |
|
"learning_rate": 8.219505494505496e-05, |
|
"loss": 0.4509, |
|
"step": 6690 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"grad_norm": 0.4855049252510071, |
|
"learning_rate": 8.216758241758242e-05, |
|
"loss": 0.4517, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"grad_norm": 0.49980971217155457, |
|
"learning_rate": 8.214010989010989e-05, |
|
"loss": 0.4592, |
|
"step": 6710 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"grad_norm": 0.8323667049407959, |
|
"learning_rate": 8.211263736263736e-05, |
|
"loss": 0.4515, |
|
"step": 6720 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"grad_norm": 0.6058827638626099, |
|
"learning_rate": 8.208516483516484e-05, |
|
"loss": 0.4437, |
|
"step": 6730 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"grad_norm": 1.3227874040603638, |
|
"learning_rate": 8.205769230769231e-05, |
|
"loss": 0.4497, |
|
"step": 6740 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"grad_norm": 1.14075767993927, |
|
"learning_rate": 8.203021978021978e-05, |
|
"loss": 0.455, |
|
"step": 6750 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"grad_norm": 1.0633773803710938, |
|
"learning_rate": 8.200274725274726e-05, |
|
"loss": 0.4582, |
|
"step": 6760 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"grad_norm": 0.671051025390625, |
|
"learning_rate": 8.197527472527473e-05, |
|
"loss": 0.4584, |
|
"step": 6770 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"grad_norm": 0.8227337002754211, |
|
"learning_rate": 8.19478021978022e-05, |
|
"loss": 0.4564, |
|
"step": 6780 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"grad_norm": 1.3290925025939941, |
|
"learning_rate": 8.192032967032968e-05, |
|
"loss": 0.4527, |
|
"step": 6790 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"grad_norm": 1.160757303237915, |
|
"learning_rate": 8.189285714285715e-05, |
|
"loss": 0.4417, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"grad_norm": 0.8233458399772644, |
|
"learning_rate": 8.186538461538462e-05, |
|
"loss": 0.4459, |
|
"step": 6810 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"grad_norm": 0.7575914859771729, |
|
"learning_rate": 8.18379120879121e-05, |
|
"loss": 0.4636, |
|
"step": 6820 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"grad_norm": 0.35609036684036255, |
|
"learning_rate": 8.181043956043956e-05, |
|
"loss": 0.4479, |
|
"step": 6830 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"grad_norm": 1.0687757730484009, |
|
"learning_rate": 8.178296703296703e-05, |
|
"loss": 0.4494, |
|
"step": 6840 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"grad_norm": 0.936469316482544, |
|
"learning_rate": 8.175549450549451e-05, |
|
"loss": 0.4605, |
|
"step": 6850 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"grad_norm": 0.6257254481315613, |
|
"learning_rate": 8.172802197802198e-05, |
|
"loss": 0.4559, |
|
"step": 6860 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"grad_norm": 1.0161064863204956, |
|
"learning_rate": 8.170054945054945e-05, |
|
"loss": 0.4677, |
|
"step": 6870 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"grad_norm": 1.0407649278640747, |
|
"learning_rate": 8.167307692307693e-05, |
|
"loss": 0.4542, |
|
"step": 6880 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"grad_norm": 0.7439064383506775, |
|
"learning_rate": 8.16456043956044e-05, |
|
"loss": 0.4747, |
|
"step": 6890 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"grad_norm": 0.37432214617729187, |
|
"learning_rate": 8.161813186813187e-05, |
|
"loss": 0.4595, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"grad_norm": 0.9039118885993958, |
|
"learning_rate": 8.159065934065935e-05, |
|
"loss": 0.4701, |
|
"step": 6910 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"grad_norm": 0.5294177532196045, |
|
"learning_rate": 8.156318681318682e-05, |
|
"loss": 0.4533, |
|
"step": 6920 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"grad_norm": 0.8023412227630615, |
|
"learning_rate": 8.153571428571429e-05, |
|
"loss": 0.4632, |
|
"step": 6930 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"grad_norm": 1.1089890003204346, |
|
"learning_rate": 8.150824175824177e-05, |
|
"loss": 0.4627, |
|
"step": 6940 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"grad_norm": 0.5587044954299927, |
|
"learning_rate": 8.148076923076924e-05, |
|
"loss": 0.4613, |
|
"step": 6950 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"grad_norm": 0.7494495511054993, |
|
"learning_rate": 8.14532967032967e-05, |
|
"loss": 0.4478, |
|
"step": 6960 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"grad_norm": 0.7425660490989685, |
|
"learning_rate": 8.142582417582419e-05, |
|
"loss": 0.4585, |
|
"step": 6970 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"grad_norm": 0.9916179180145264, |
|
"learning_rate": 8.139835164835165e-05, |
|
"loss": 0.464, |
|
"step": 6980 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"grad_norm": 0.9130895137786865, |
|
"learning_rate": 8.137087912087912e-05, |
|
"loss": 0.4628, |
|
"step": 6990 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"grad_norm": 0.7386452555656433, |
|
"learning_rate": 8.13434065934066e-05, |
|
"loss": 0.457, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"grad_norm": 1.5726503133773804, |
|
"learning_rate": 8.131593406593407e-05, |
|
"loss": 0.461, |
|
"step": 7010 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"grad_norm": 0.6793839335441589, |
|
"learning_rate": 8.128846153846154e-05, |
|
"loss": 0.4521, |
|
"step": 7020 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"grad_norm": 1.5302358865737915, |
|
"learning_rate": 8.126098901098902e-05, |
|
"loss": 0.4598, |
|
"step": 7030 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"grad_norm": 0.9924741387367249, |
|
"learning_rate": 8.123351648351649e-05, |
|
"loss": 0.4599, |
|
"step": 7040 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"grad_norm": 0.7319537997245789, |
|
"learning_rate": 8.120604395604396e-05, |
|
"loss": 0.4585, |
|
"step": 7050 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"grad_norm": 1.067428469657898, |
|
"learning_rate": 8.117857142857144e-05, |
|
"loss": 0.4509, |
|
"step": 7060 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"grad_norm": 1.2225853204727173, |
|
"learning_rate": 8.115109890109891e-05, |
|
"loss": 0.4591, |
|
"step": 7070 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"grad_norm": 0.7233381867408752, |
|
"learning_rate": 8.112362637362638e-05, |
|
"loss": 0.4568, |
|
"step": 7080 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"grad_norm": 1.3318099975585938, |
|
"learning_rate": 8.109615384615386e-05, |
|
"loss": 0.4715, |
|
"step": 7090 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"grad_norm": 1.4362515211105347, |
|
"learning_rate": 8.106868131868133e-05, |
|
"loss": 0.4539, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"grad_norm": 0.8498636484146118, |
|
"learning_rate": 8.10412087912088e-05, |
|
"loss": 0.4531, |
|
"step": 7110 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"grad_norm": 0.5961604118347168, |
|
"learning_rate": 8.101373626373628e-05, |
|
"loss": 0.4561, |
|
"step": 7120 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"grad_norm": 1.013180136680603, |
|
"learning_rate": 8.098626373626374e-05, |
|
"loss": 0.4581, |
|
"step": 7130 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"grad_norm": 0.9159004092216492, |
|
"learning_rate": 8.095879120879121e-05, |
|
"loss": 0.4498, |
|
"step": 7140 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"grad_norm": 0.9847846031188965, |
|
"learning_rate": 8.093131868131868e-05, |
|
"loss": 0.4584, |
|
"step": 7150 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"grad_norm": 1.5007630586624146, |
|
"learning_rate": 8.090384615384615e-05, |
|
"loss": 0.4524, |
|
"step": 7160 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"grad_norm": 0.8937559723854065, |
|
"learning_rate": 8.087637362637363e-05, |
|
"loss": 0.4701, |
|
"step": 7170 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"grad_norm": 0.9551745653152466, |
|
"learning_rate": 8.08489010989011e-05, |
|
"loss": 0.4412, |
|
"step": 7180 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"grad_norm": 0.7207891345024109, |
|
"learning_rate": 8.082142857142857e-05, |
|
"loss": 0.4577, |
|
"step": 7190 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"grad_norm": 0.5330765843391418, |
|
"learning_rate": 8.079395604395605e-05, |
|
"loss": 0.4531, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"grad_norm": 1.0517854690551758, |
|
"learning_rate": 8.076648351648352e-05, |
|
"loss": 0.4339, |
|
"step": 7210 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"grad_norm": 1.2942813634872437, |
|
"learning_rate": 8.073901098901098e-05, |
|
"loss": 0.4575, |
|
"step": 7220 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"grad_norm": 0.9348652362823486, |
|
"learning_rate": 8.071153846153847e-05, |
|
"loss": 0.4472, |
|
"step": 7230 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"grad_norm": 1.095439076423645, |
|
"learning_rate": 8.068406593406593e-05, |
|
"loss": 0.4503, |
|
"step": 7240 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"grad_norm": 0.5965198278427124, |
|
"learning_rate": 8.06565934065934e-05, |
|
"loss": 0.4499, |
|
"step": 7250 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"grad_norm": 0.5533403158187866, |
|
"learning_rate": 8.062912087912088e-05, |
|
"loss": 0.4676, |
|
"step": 7260 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"grad_norm": 1.1052013635635376, |
|
"learning_rate": 8.060164835164835e-05, |
|
"loss": 0.4635, |
|
"step": 7270 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"grad_norm": 0.9924907088279724, |
|
"learning_rate": 8.057417582417582e-05, |
|
"loss": 0.4606, |
|
"step": 7280 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"grad_norm": 0.5882344245910645, |
|
"learning_rate": 8.05467032967033e-05, |
|
"loss": 0.4456, |
|
"step": 7290 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"grad_norm": 1.0022166967391968, |
|
"learning_rate": 8.051923076923077e-05, |
|
"loss": 0.4505, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 0.8352431058883667, |
|
"learning_rate": 8.049175824175824e-05, |
|
"loss": 0.4567, |
|
"step": 7310 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 1.0245552062988281, |
|
"learning_rate": 8.046428571428572e-05, |
|
"loss": 0.4509, |
|
"step": 7320 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 0.44436022639274597, |
|
"eval_runtime": 1246.3805, |
|
"eval_samples_per_second": 234.366, |
|
"eval_steps_per_second": 3.663, |
|
"step": 7321 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 0.6890380382537842, |
|
"learning_rate": 8.043681318681319e-05, |
|
"loss": 0.4559, |
|
"step": 7330 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"grad_norm": 0.8071466088294983, |
|
"learning_rate": 8.040934065934066e-05, |
|
"loss": 0.4494, |
|
"step": 7340 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"grad_norm": 0.5615882873535156, |
|
"learning_rate": 8.038186813186814e-05, |
|
"loss": 0.444, |
|
"step": 7350 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"grad_norm": 1.312966227531433, |
|
"learning_rate": 8.03543956043956e-05, |
|
"loss": 0.4549, |
|
"step": 7360 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"grad_norm": 0.7778192162513733, |
|
"learning_rate": 8.032692307692307e-05, |
|
"loss": 0.4561, |
|
"step": 7370 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"grad_norm": 0.9343681335449219, |
|
"learning_rate": 8.029945054945056e-05, |
|
"loss": 0.4565, |
|
"step": 7380 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"grad_norm": 0.585110068321228, |
|
"learning_rate": 8.027197802197802e-05, |
|
"loss": 0.4502, |
|
"step": 7390 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"grad_norm": 1.0197051763534546, |
|
"learning_rate": 8.024450549450549e-05, |
|
"loss": 0.4595, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"grad_norm": 1.0265450477600098, |
|
"learning_rate": 8.021703296703297e-05, |
|
"loss": 0.4425, |
|
"step": 7410 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"grad_norm": 1.1836851835250854, |
|
"learning_rate": 8.018956043956044e-05, |
|
"loss": 0.4606, |
|
"step": 7420 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"grad_norm": 0.9758961796760559, |
|
"learning_rate": 8.016208791208791e-05, |
|
"loss": 0.4494, |
|
"step": 7430 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"grad_norm": 1.0412979125976562, |
|
"learning_rate": 8.013461538461539e-05, |
|
"loss": 0.4593, |
|
"step": 7440 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"grad_norm": 1.13507878780365, |
|
"learning_rate": 8.010714285714286e-05, |
|
"loss": 0.4598, |
|
"step": 7450 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"grad_norm": 1.118143081665039, |
|
"learning_rate": 8.007967032967033e-05, |
|
"loss": 0.4597, |
|
"step": 7460 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"grad_norm": 0.6100110411643982, |
|
"learning_rate": 8.005219780219781e-05, |
|
"loss": 0.46, |
|
"step": 7470 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"grad_norm": 1.4135487079620361, |
|
"learning_rate": 8.002472527472528e-05, |
|
"loss": 0.455, |
|
"step": 7480 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"grad_norm": 0.7120094299316406, |
|
"learning_rate": 7.999725274725275e-05, |
|
"loss": 0.4459, |
|
"step": 7490 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"grad_norm": 0.7432197332382202, |
|
"learning_rate": 7.996978021978023e-05, |
|
"loss": 0.4509, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"grad_norm": 1.1492975950241089, |
|
"learning_rate": 7.99423076923077e-05, |
|
"loss": 0.4638, |
|
"step": 7510 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"grad_norm": 1.0090208053588867, |
|
"learning_rate": 7.991483516483516e-05, |
|
"loss": 0.4702, |
|
"step": 7520 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"grad_norm": 1.09323251247406, |
|
"learning_rate": 7.988736263736263e-05, |
|
"loss": 0.4518, |
|
"step": 7530 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"grad_norm": 1.0353306531906128, |
|
"learning_rate": 7.985989010989011e-05, |
|
"loss": 0.4522, |
|
"step": 7540 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"grad_norm": 0.8499657511711121, |
|
"learning_rate": 7.983241758241758e-05, |
|
"loss": 0.4637, |
|
"step": 7550 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"grad_norm": 1.0327955484390259, |
|
"learning_rate": 7.980494505494505e-05, |
|
"loss": 0.4444, |
|
"step": 7560 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"grad_norm": 1.239608883857727, |
|
"learning_rate": 7.977747252747253e-05, |
|
"loss": 0.4536, |
|
"step": 7570 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"grad_norm": 0.6841428279876709, |
|
"learning_rate": 7.975e-05, |
|
"loss": 0.4519, |
|
"step": 7580 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"grad_norm": 1.1058799028396606, |
|
"learning_rate": 7.972252747252747e-05, |
|
"loss": 0.4511, |
|
"step": 7590 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"grad_norm": 1.1671322584152222, |
|
"learning_rate": 7.969505494505495e-05, |
|
"loss": 0.4512, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"grad_norm": 0.6706550121307373, |
|
"learning_rate": 7.966758241758242e-05, |
|
"loss": 0.4493, |
|
"step": 7610 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"grad_norm": 0.752795398235321, |
|
"learning_rate": 7.964010989010989e-05, |
|
"loss": 0.4494, |
|
"step": 7620 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"grad_norm": 0.6255919933319092, |
|
"learning_rate": 7.961263736263737e-05, |
|
"loss": 0.4525, |
|
"step": 7630 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"grad_norm": 0.4543440341949463, |
|
"learning_rate": 7.958516483516484e-05, |
|
"loss": 0.4527, |
|
"step": 7640 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"grad_norm": 0.6038147807121277, |
|
"learning_rate": 7.95576923076923e-05, |
|
"loss": 0.4431, |
|
"step": 7650 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"grad_norm": 1.010972261428833, |
|
"learning_rate": 7.953021978021979e-05, |
|
"loss": 0.451, |
|
"step": 7660 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"grad_norm": 1.0627601146697998, |
|
"learning_rate": 7.950274725274725e-05, |
|
"loss": 0.4441, |
|
"step": 7670 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"grad_norm": 0.6361072063446045, |
|
"learning_rate": 7.947527472527472e-05, |
|
"loss": 0.4592, |
|
"step": 7680 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"grad_norm": 0.897415280342102, |
|
"learning_rate": 7.94478021978022e-05, |
|
"loss": 0.4535, |
|
"step": 7690 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"grad_norm": 0.6545829772949219, |
|
"learning_rate": 7.942032967032967e-05, |
|
"loss": 0.4589, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"grad_norm": 0.9311442375183105, |
|
"learning_rate": 7.939285714285714e-05, |
|
"loss": 0.4417, |
|
"step": 7710 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"grad_norm": 1.5107172727584839, |
|
"learning_rate": 7.936538461538462e-05, |
|
"loss": 0.4503, |
|
"step": 7720 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"grad_norm": 0.9508922100067139, |
|
"learning_rate": 7.933791208791209e-05, |
|
"loss": 0.4532, |
|
"step": 7730 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"grad_norm": 0.8631314039230347, |
|
"learning_rate": 7.931043956043956e-05, |
|
"loss": 0.4551, |
|
"step": 7740 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"grad_norm": 0.639757513999939, |
|
"learning_rate": 7.928296703296704e-05, |
|
"loss": 0.4531, |
|
"step": 7750 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"grad_norm": 0.8901143074035645, |
|
"learning_rate": 7.925549450549451e-05, |
|
"loss": 0.4563, |
|
"step": 7760 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"grad_norm": 1.7677491903305054, |
|
"learning_rate": 7.922802197802198e-05, |
|
"loss": 0.4467, |
|
"step": 7770 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"grad_norm": 0.9918643236160278, |
|
"learning_rate": 7.920054945054946e-05, |
|
"loss": 0.4604, |
|
"step": 7780 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"grad_norm": 0.7385686635971069, |
|
"learning_rate": 7.917307692307693e-05, |
|
"loss": 0.4484, |
|
"step": 7790 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"grad_norm": 0.4740804433822632, |
|
"learning_rate": 7.91456043956044e-05, |
|
"loss": 0.4581, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"grad_norm": 0.3722800016403198, |
|
"learning_rate": 7.911813186813188e-05, |
|
"loss": 0.4437, |
|
"step": 7810 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"grad_norm": 0.5825007557868958, |
|
"learning_rate": 7.909065934065934e-05, |
|
"loss": 0.4495, |
|
"step": 7820 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"grad_norm": 0.7577399015426636, |
|
"learning_rate": 7.906318681318681e-05, |
|
"loss": 0.4507, |
|
"step": 7830 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"grad_norm": 0.5723931193351746, |
|
"learning_rate": 7.90357142857143e-05, |
|
"loss": 0.4512, |
|
"step": 7840 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"grad_norm": 1.0552576780319214, |
|
"learning_rate": 7.900824175824176e-05, |
|
"loss": 0.4493, |
|
"step": 7850 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"grad_norm": 1.1965206861495972, |
|
"learning_rate": 7.898076923076923e-05, |
|
"loss": 0.4502, |
|
"step": 7860 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"grad_norm": 0.7538771629333496, |
|
"learning_rate": 7.895329670329671e-05, |
|
"loss": 0.4576, |
|
"step": 7870 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"grad_norm": 1.2679303884506226, |
|
"learning_rate": 7.892582417582418e-05, |
|
"loss": 0.4464, |
|
"step": 7880 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"grad_norm": 0.8973550200462341, |
|
"learning_rate": 7.889835164835165e-05, |
|
"loss": 0.4596, |
|
"step": 7890 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"grad_norm": 0.7081739902496338, |
|
"learning_rate": 7.887087912087913e-05, |
|
"loss": 0.4368, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"grad_norm": 0.7710354328155518, |
|
"learning_rate": 7.88434065934066e-05, |
|
"loss": 0.4463, |
|
"step": 7910 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"grad_norm": 0.6749390363693237, |
|
"learning_rate": 7.881593406593407e-05, |
|
"loss": 0.4515, |
|
"step": 7920 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"grad_norm": 0.6680644154548645, |
|
"learning_rate": 7.878846153846155e-05, |
|
"loss": 0.4433, |
|
"step": 7930 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"grad_norm": 1.5634280443191528, |
|
"learning_rate": 7.876098901098902e-05, |
|
"loss": 0.4702, |
|
"step": 7940 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"grad_norm": 0.6959145665168762, |
|
"learning_rate": 7.873351648351648e-05, |
|
"loss": 0.4504, |
|
"step": 7950 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"grad_norm": 0.7739858627319336, |
|
"learning_rate": 7.870604395604397e-05, |
|
"loss": 0.4606, |
|
"step": 7960 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"grad_norm": 0.5519704818725586, |
|
"learning_rate": 7.867857142857143e-05, |
|
"loss": 0.4658, |
|
"step": 7970 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"grad_norm": 0.5425352454185486, |
|
"learning_rate": 7.86510989010989e-05, |
|
"loss": 0.4445, |
|
"step": 7980 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"grad_norm": 0.8816283941268921, |
|
"learning_rate": 7.862362637362638e-05, |
|
"loss": 0.4621, |
|
"step": 7990 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"grad_norm": 1.0816186666488647, |
|
"learning_rate": 7.859615384615385e-05, |
|
"loss": 0.4622, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"grad_norm": 0.8483924865722656, |
|
"learning_rate": 7.856868131868132e-05, |
|
"loss": 0.4533, |
|
"step": 8010 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"grad_norm": 0.45229572057724, |
|
"learning_rate": 7.85412087912088e-05, |
|
"loss": 0.4626, |
|
"step": 8020 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"grad_norm": 0.6593546867370605, |
|
"learning_rate": 7.851373626373627e-05, |
|
"loss": 0.4498, |
|
"step": 8030 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"grad_norm": 1.214553952217102, |
|
"learning_rate": 7.848626373626374e-05, |
|
"loss": 0.4497, |
|
"step": 8040 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"grad_norm": 0.7151647210121155, |
|
"learning_rate": 7.845879120879122e-05, |
|
"loss": 0.4447, |
|
"step": 8050 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"grad_norm": 0.7126444578170776, |
|
"learning_rate": 7.843131868131869e-05, |
|
"loss": 0.4622, |
|
"step": 8060 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"grad_norm": 0.8902627229690552, |
|
"learning_rate": 7.840384615384616e-05, |
|
"loss": 0.4439, |
|
"step": 8070 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"grad_norm": 0.7599775195121765, |
|
"learning_rate": 7.837637362637364e-05, |
|
"loss": 0.4537, |
|
"step": 8080 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"grad_norm": 0.9020054340362549, |
|
"learning_rate": 7.83489010989011e-05, |
|
"loss": 0.4658, |
|
"step": 8090 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"grad_norm": 0.8659080266952515, |
|
"learning_rate": 7.832142857142857e-05, |
|
"loss": 0.4553, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"grad_norm": 0.6685529351234436, |
|
"learning_rate": 7.829395604395606e-05, |
|
"loss": 0.4604, |
|
"step": 8110 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"grad_norm": 0.6612065434455872, |
|
"learning_rate": 7.826648351648352e-05, |
|
"loss": 0.4522, |
|
"step": 8120 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"grad_norm": 0.7674763202667236, |
|
"learning_rate": 7.823901098901099e-05, |
|
"loss": 0.4523, |
|
"step": 8130 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"grad_norm": 0.5680274963378906, |
|
"learning_rate": 7.821153846153847e-05, |
|
"loss": 0.4649, |
|
"step": 8140 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"grad_norm": 1.0545960664749146, |
|
"learning_rate": 7.818406593406594e-05, |
|
"loss": 0.4507, |
|
"step": 8150 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"grad_norm": 0.5723569393157959, |
|
"learning_rate": 7.815659340659341e-05, |
|
"loss": 0.4418, |
|
"step": 8160 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"grad_norm": 0.9119620323181152, |
|
"learning_rate": 7.812912087912089e-05, |
|
"loss": 0.4535, |
|
"step": 8170 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"grad_norm": 1.0251758098602295, |
|
"learning_rate": 7.810164835164836e-05, |
|
"loss": 0.4503, |
|
"step": 8180 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"grad_norm": 0.6839396357536316, |
|
"learning_rate": 7.807417582417583e-05, |
|
"loss": 0.452, |
|
"step": 8190 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"grad_norm": 1.3468687534332275, |
|
"learning_rate": 7.804670329670331e-05, |
|
"loss": 0.4442, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"grad_norm": 0.9562901854515076, |
|
"learning_rate": 7.801923076923078e-05, |
|
"loss": 0.4596, |
|
"step": 8210 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"grad_norm": 1.092677354812622, |
|
"learning_rate": 7.799175824175825e-05, |
|
"loss": 0.4454, |
|
"step": 8220 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"grad_norm": 0.7029426097869873, |
|
"learning_rate": 7.796428571428573e-05, |
|
"loss": 0.4463, |
|
"step": 8230 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"grad_norm": 0.7543437480926514, |
|
"learning_rate": 7.79368131868132e-05, |
|
"loss": 0.4588, |
|
"step": 8240 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"grad_norm": 1.5089532136917114, |
|
"learning_rate": 7.790934065934066e-05, |
|
"loss": 0.4531, |
|
"step": 8250 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"grad_norm": 0.6691949963569641, |
|
"learning_rate": 7.788186813186815e-05, |
|
"loss": 0.4494, |
|
"step": 8260 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"grad_norm": 1.0210888385772705, |
|
"learning_rate": 7.785439560439561e-05, |
|
"loss": 0.4548, |
|
"step": 8270 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"grad_norm": 0.9251496195793152, |
|
"learning_rate": 7.782692307692308e-05, |
|
"loss": 0.4576, |
|
"step": 8280 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"grad_norm": 1.2618598937988281, |
|
"learning_rate": 7.779945054945055e-05, |
|
"loss": 0.4478, |
|
"step": 8290 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"grad_norm": 1.2252585887908936, |
|
"learning_rate": 7.777197802197802e-05, |
|
"loss": 0.4573, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"grad_norm": 0.6124866604804993, |
|
"learning_rate": 7.77445054945055e-05, |
|
"loss": 0.4506, |
|
"step": 8310 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"grad_norm": 1.0547674894332886, |
|
"learning_rate": 7.771703296703297e-05, |
|
"loss": 0.4418, |
|
"step": 8320 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"grad_norm": 1.1824958324432373, |
|
"learning_rate": 7.768956043956044e-05, |
|
"loss": 0.4529, |
|
"step": 8330 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"grad_norm": 0.9790034890174866, |
|
"learning_rate": 7.76620879120879e-05, |
|
"loss": 0.455, |
|
"step": 8340 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"grad_norm": 0.9925460815429688, |
|
"learning_rate": 7.763461538461539e-05, |
|
"loss": 0.4607, |
|
"step": 8350 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"grad_norm": 0.8586925864219666, |
|
"learning_rate": 7.760714285714285e-05, |
|
"loss": 0.4644, |
|
"step": 8360 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"grad_norm": 1.4445006847381592, |
|
"learning_rate": 7.757967032967032e-05, |
|
"loss": 0.4513, |
|
"step": 8370 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"grad_norm": 0.9150937795639038, |
|
"learning_rate": 7.75521978021978e-05, |
|
"loss": 0.4534, |
|
"step": 8380 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"grad_norm": 0.9148111939430237, |
|
"learning_rate": 7.752472527472527e-05, |
|
"loss": 0.4424, |
|
"step": 8390 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"grad_norm": 0.9973462224006653, |
|
"learning_rate": 7.749725274725274e-05, |
|
"loss": 0.4523, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"grad_norm": 1.022926926612854, |
|
"learning_rate": 7.746978021978022e-05, |
|
"loss": 0.444, |
|
"step": 8410 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"grad_norm": 0.7249711155891418, |
|
"learning_rate": 7.744230769230769e-05, |
|
"loss": 0.4473, |
|
"step": 8420 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"grad_norm": 0.44306471943855286, |
|
"learning_rate": 7.741483516483516e-05, |
|
"loss": 0.4452, |
|
"step": 8430 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"grad_norm": 0.9625614881515503, |
|
"learning_rate": 7.738736263736264e-05, |
|
"loss": 0.4373, |
|
"step": 8440 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"grad_norm": 0.5116167664527893, |
|
"learning_rate": 7.735989010989011e-05, |
|
"loss": 0.4505, |
|
"step": 8450 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"grad_norm": 0.7496402859687805, |
|
"learning_rate": 7.733241758241758e-05, |
|
"loss": 0.4451, |
|
"step": 8460 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"grad_norm": 1.3204615116119385, |
|
"learning_rate": 7.730494505494506e-05, |
|
"loss": 0.4576, |
|
"step": 8470 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"grad_norm": 0.7599005699157715, |
|
"learning_rate": 7.727747252747253e-05, |
|
"loss": 0.4493, |
|
"step": 8480 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"grad_norm": 0.5699270367622375, |
|
"learning_rate": 7.725e-05, |
|
"loss": 0.4568, |
|
"step": 8490 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"grad_norm": 0.6626598238945007, |
|
"learning_rate": 7.722252747252748e-05, |
|
"loss": 0.4504, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"grad_norm": 0.6977862119674683, |
|
"learning_rate": 7.719505494505494e-05, |
|
"loss": 0.4529, |
|
"step": 8510 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"grad_norm": 0.883980393409729, |
|
"learning_rate": 7.716758241758241e-05, |
|
"loss": 0.4527, |
|
"step": 8520 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"grad_norm": 0.9930477738380432, |
|
"learning_rate": 7.71401098901099e-05, |
|
"loss": 0.4524, |
|
"step": 8530 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"grad_norm": 0.5075715780258179, |
|
"learning_rate": 7.711263736263736e-05, |
|
"loss": 0.4436, |
|
"step": 8540 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"grad_norm": 0.7683893442153931, |
|
"learning_rate": 7.708516483516483e-05, |
|
"loss": 0.4608, |
|
"step": 8550 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"grad_norm": 1.6546239852905273, |
|
"learning_rate": 7.705769230769231e-05, |
|
"loss": 0.4521, |
|
"step": 8560 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"grad_norm": 0.8132153153419495, |
|
"learning_rate": 7.703021978021978e-05, |
|
"loss": 0.4542, |
|
"step": 8570 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"grad_norm": 0.5142685770988464, |
|
"learning_rate": 7.700274725274725e-05, |
|
"loss": 0.4548, |
|
"step": 8580 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"grad_norm": 1.100368618965149, |
|
"learning_rate": 7.697527472527473e-05, |
|
"loss": 0.4602, |
|
"step": 8590 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"grad_norm": 1.018351674079895, |
|
"learning_rate": 7.69478021978022e-05, |
|
"loss": 0.4424, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"grad_norm": 0.557134211063385, |
|
"learning_rate": 7.692032967032967e-05, |
|
"loss": 0.4486, |
|
"step": 8610 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"grad_norm": 0.701932966709137, |
|
"learning_rate": 7.689285714285715e-05, |
|
"loss": 0.4543, |
|
"step": 8620 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"grad_norm": 1.0351616144180298, |
|
"learning_rate": 7.686538461538462e-05, |
|
"loss": 0.4628, |
|
"step": 8630 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"grad_norm": 0.5028539896011353, |
|
"learning_rate": 7.683791208791208e-05, |
|
"loss": 0.4566, |
|
"step": 8640 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"grad_norm": 0.7769728302955627, |
|
"learning_rate": 7.681043956043957e-05, |
|
"loss": 0.4429, |
|
"step": 8650 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"grad_norm": 0.7966272830963135, |
|
"learning_rate": 7.678296703296703e-05, |
|
"loss": 0.4377, |
|
"step": 8660 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"grad_norm": 1.1576814651489258, |
|
"learning_rate": 7.67554945054945e-05, |
|
"loss": 0.4507, |
|
"step": 8670 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"grad_norm": 0.8654575943946838, |
|
"learning_rate": 7.672802197802198e-05, |
|
"loss": 0.4598, |
|
"step": 8680 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"grad_norm": 0.7794680595397949, |
|
"learning_rate": 7.670054945054945e-05, |
|
"loss": 0.4424, |
|
"step": 8690 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"grad_norm": 0.8271937370300293, |
|
"learning_rate": 7.667307692307692e-05, |
|
"loss": 0.4586, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"grad_norm": 0.8499754071235657, |
|
"learning_rate": 7.66456043956044e-05, |
|
"loss": 0.4476, |
|
"step": 8710 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"grad_norm": 0.7553250789642334, |
|
"learning_rate": 7.661813186813187e-05, |
|
"loss": 0.4435, |
|
"step": 8720 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"grad_norm": 0.5660779476165771, |
|
"learning_rate": 7.659065934065934e-05, |
|
"loss": 0.4518, |
|
"step": 8730 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"grad_norm": 1.4875767230987549, |
|
"learning_rate": 7.656318681318682e-05, |
|
"loss": 0.4542, |
|
"step": 8740 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"grad_norm": 1.1859349012374878, |
|
"learning_rate": 7.653846153846153e-05, |
|
"loss": 0.4528, |
|
"step": 8750 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"grad_norm": 0.6259100437164307, |
|
"learning_rate": 7.651098901098902e-05, |
|
"loss": 0.4629, |
|
"step": 8760 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"grad_norm": 0.4763346314430237, |
|
"learning_rate": 7.648351648351648e-05, |
|
"loss": 0.4479, |
|
"step": 8770 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"grad_norm": 0.6478772759437561, |
|
"learning_rate": 7.645604395604395e-05, |
|
"loss": 0.4313, |
|
"step": 8780 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"grad_norm": 1.1578466892242432, |
|
"learning_rate": 7.642857142857143e-05, |
|
"loss": 0.4499, |
|
"step": 8790 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"grad_norm": 0.5163312554359436, |
|
"learning_rate": 7.64010989010989e-05, |
|
"loss": 0.4479, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"grad_norm": 1.183361530303955, |
|
"learning_rate": 7.637362637362637e-05, |
|
"loss": 0.4524, |
|
"step": 8810 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"grad_norm": 0.6872965097427368, |
|
"learning_rate": 7.634615384615385e-05, |
|
"loss": 0.4445, |
|
"step": 8820 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"grad_norm": 0.7339246273040771, |
|
"learning_rate": 7.631868131868132e-05, |
|
"loss": 0.4446, |
|
"step": 8830 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"grad_norm": 0.7101340889930725, |
|
"learning_rate": 7.629120879120879e-05, |
|
"loss": 0.45, |
|
"step": 8840 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"grad_norm": 0.5717099905014038, |
|
"learning_rate": 7.626373626373627e-05, |
|
"loss": 0.4503, |
|
"step": 8850 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"grad_norm": 0.6713473796844482, |
|
"learning_rate": 7.623626373626374e-05, |
|
"loss": 0.4528, |
|
"step": 8860 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"grad_norm": 0.9480137228965759, |
|
"learning_rate": 7.62087912087912e-05, |
|
"loss": 0.4569, |
|
"step": 8870 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"grad_norm": 0.594673752784729, |
|
"learning_rate": 7.618131868131869e-05, |
|
"loss": 0.4627, |
|
"step": 8880 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"grad_norm": 0.8691803216934204, |
|
"learning_rate": 7.615384615384616e-05, |
|
"loss": 0.4459, |
|
"step": 8890 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"grad_norm": 0.8882876634597778, |
|
"learning_rate": 7.612637362637362e-05, |
|
"loss": 0.4478, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"grad_norm": 0.48794516921043396, |
|
"learning_rate": 7.60989010989011e-05, |
|
"loss": 0.4479, |
|
"step": 8910 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"grad_norm": 0.4966289699077606, |
|
"learning_rate": 7.607142857142857e-05, |
|
"loss": 0.4395, |
|
"step": 8920 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"grad_norm": 0.792293131351471, |
|
"learning_rate": 7.604395604395604e-05, |
|
"loss": 0.4353, |
|
"step": 8930 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"grad_norm": 0.6615070104598999, |
|
"learning_rate": 7.601648351648351e-05, |
|
"loss": 0.4478, |
|
"step": 8940 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"grad_norm": 0.8686419725418091, |
|
"learning_rate": 7.598901098901099e-05, |
|
"loss": 0.4516, |
|
"step": 8950 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"grad_norm": 1.2833001613616943, |
|
"learning_rate": 7.596153846153846e-05, |
|
"loss": 0.4526, |
|
"step": 8960 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"grad_norm": 0.7414630055427551, |
|
"learning_rate": 7.593406593406593e-05, |
|
"loss": 0.4549, |
|
"step": 8970 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"grad_norm": 0.9745138883590698, |
|
"learning_rate": 7.590659340659341e-05, |
|
"loss": 0.4412, |
|
"step": 8980 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"grad_norm": 0.8407401442527771, |
|
"learning_rate": 7.587912087912088e-05, |
|
"loss": 0.4511, |
|
"step": 8990 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"grad_norm": 1.1383010149002075, |
|
"learning_rate": 7.585164835164835e-05, |
|
"loss": 0.4556, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"grad_norm": 1.0419915914535522, |
|
"learning_rate": 7.582417582417583e-05, |
|
"loss": 0.4504, |
|
"step": 9010 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"grad_norm": 0.931141197681427, |
|
"learning_rate": 7.57967032967033e-05, |
|
"loss": 0.4464, |
|
"step": 9020 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"grad_norm": 0.6264098286628723, |
|
"learning_rate": 7.576923076923076e-05, |
|
"loss": 0.4627, |
|
"step": 9030 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"grad_norm": 0.8912045359611511, |
|
"learning_rate": 7.574175824175825e-05, |
|
"loss": 0.4452, |
|
"step": 9040 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"grad_norm": 1.3010141849517822, |
|
"learning_rate": 7.571428571428571e-05, |
|
"loss": 0.4595, |
|
"step": 9050 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"grad_norm": 0.8275282979011536, |
|
"learning_rate": 7.568681318681318e-05, |
|
"loss": 0.4486, |
|
"step": 9060 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"grad_norm": 0.8537049293518066, |
|
"learning_rate": 7.565934065934066e-05, |
|
"loss": 0.4506, |
|
"step": 9070 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"grad_norm": 0.5747140645980835, |
|
"learning_rate": 7.563186813186813e-05, |
|
"loss": 0.4488, |
|
"step": 9080 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"grad_norm": 0.8508340716362, |
|
"learning_rate": 7.56043956043956e-05, |
|
"loss": 0.4436, |
|
"step": 9090 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"grad_norm": 1.1689362525939941, |
|
"learning_rate": 7.557692307692308e-05, |
|
"loss": 0.4586, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"grad_norm": 1.1048357486724854, |
|
"learning_rate": 7.554945054945055e-05, |
|
"loss": 0.4568, |
|
"step": 9110 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"grad_norm": 1.1694691181182861, |
|
"learning_rate": 7.552197802197802e-05, |
|
"loss": 0.4571, |
|
"step": 9120 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"grad_norm": 0.9257465600967407, |
|
"learning_rate": 7.54945054945055e-05, |
|
"loss": 0.4544, |
|
"step": 9130 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"grad_norm": 1.3826684951782227, |
|
"learning_rate": 7.546703296703297e-05, |
|
"loss": 0.44, |
|
"step": 9140 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"grad_norm": 0.8815349340438843, |
|
"learning_rate": 7.543956043956044e-05, |
|
"loss": 0.4438, |
|
"step": 9150 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"grad_norm": 0.661222517490387, |
|
"learning_rate": 7.541208791208792e-05, |
|
"loss": 0.4479, |
|
"step": 9160 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"grad_norm": 1.222886085510254, |
|
"learning_rate": 7.538461538461539e-05, |
|
"loss": 0.4487, |
|
"step": 9170 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"grad_norm": 1.1454687118530273, |
|
"learning_rate": 7.535714285714285e-05, |
|
"loss": 0.4625, |
|
"step": 9180 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"grad_norm": 0.49560263752937317, |
|
"learning_rate": 7.532967032967034e-05, |
|
"loss": 0.4518, |
|
"step": 9190 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"grad_norm": 0.8997675776481628, |
|
"learning_rate": 7.53021978021978e-05, |
|
"loss": 0.4426, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"grad_norm": 0.9074673056602478, |
|
"learning_rate": 7.527472527472527e-05, |
|
"loss": 0.4524, |
|
"step": 9210 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"grad_norm": 0.7328269481658936, |
|
"learning_rate": 7.524725274725275e-05, |
|
"loss": 0.4512, |
|
"step": 9220 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"grad_norm": 0.7607899904251099, |
|
"learning_rate": 7.521978021978022e-05, |
|
"loss": 0.4491, |
|
"step": 9230 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"grad_norm": 0.5627162456512451, |
|
"learning_rate": 7.519230769230769e-05, |
|
"loss": 0.4511, |
|
"step": 9240 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"grad_norm": 0.9358956813812256, |
|
"learning_rate": 7.516483516483517e-05, |
|
"loss": 0.453, |
|
"step": 9250 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"grad_norm": 1.2619479894638062, |
|
"learning_rate": 7.513736263736264e-05, |
|
"loss": 0.4432, |
|
"step": 9260 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"grad_norm": 0.9375318884849548, |
|
"learning_rate": 7.510989010989011e-05, |
|
"loss": 0.4539, |
|
"step": 9270 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"grad_norm": 0.8454050421714783, |
|
"learning_rate": 7.508241758241759e-05, |
|
"loss": 0.4515, |
|
"step": 9280 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"grad_norm": 0.9856541156768799, |
|
"learning_rate": 7.505494505494506e-05, |
|
"loss": 0.445, |
|
"step": 9290 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"grad_norm": 0.6597672700881958, |
|
"learning_rate": 7.502747252747253e-05, |
|
"loss": 0.4505, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"grad_norm": 1.0249258279800415, |
|
"learning_rate": 7.500000000000001e-05, |
|
"loss": 0.4516, |
|
"step": 9310 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"grad_norm": 1.003083348274231, |
|
"learning_rate": 7.497252747252748e-05, |
|
"loss": 0.4457, |
|
"step": 9320 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"grad_norm": 1.236377477645874, |
|
"learning_rate": 7.494505494505494e-05, |
|
"loss": 0.4573, |
|
"step": 9330 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"grad_norm": 1.5360130071640015, |
|
"learning_rate": 7.491758241758242e-05, |
|
"loss": 0.4478, |
|
"step": 9340 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"grad_norm": 0.9379085898399353, |
|
"learning_rate": 7.489010989010989e-05, |
|
"loss": 0.4575, |
|
"step": 9350 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"grad_norm": 0.7734600901603699, |
|
"learning_rate": 7.486263736263736e-05, |
|
"loss": 0.4571, |
|
"step": 9360 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"grad_norm": 0.6792673468589783, |
|
"learning_rate": 7.483516483516484e-05, |
|
"loss": 0.4522, |
|
"step": 9370 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"grad_norm": 0.8936319351196289, |
|
"learning_rate": 7.480769230769231e-05, |
|
"loss": 0.4576, |
|
"step": 9380 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"grad_norm": 0.7541890144348145, |
|
"learning_rate": 7.478021978021978e-05, |
|
"loss": 0.4456, |
|
"step": 9390 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"grad_norm": 0.8661054968833923, |
|
"learning_rate": 7.475274725274726e-05, |
|
"loss": 0.4428, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"grad_norm": 0.919201672077179, |
|
"learning_rate": 7.472527472527473e-05, |
|
"loss": 0.4507, |
|
"step": 9410 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"grad_norm": 1.0688129663467407, |
|
"learning_rate": 7.46978021978022e-05, |
|
"loss": 0.4469, |
|
"step": 9420 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"grad_norm": 7.527632236480713, |
|
"learning_rate": 7.467032967032968e-05, |
|
"loss": 0.4512, |
|
"step": 9430 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"grad_norm": 0.5668841600418091, |
|
"learning_rate": 7.464285714285715e-05, |
|
"loss": 0.4484, |
|
"step": 9440 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"grad_norm": 0.7174908518791199, |
|
"learning_rate": 7.461538461538462e-05, |
|
"loss": 0.4361, |
|
"step": 9450 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"grad_norm": 0.693144679069519, |
|
"learning_rate": 7.45879120879121e-05, |
|
"loss": 0.4483, |
|
"step": 9460 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"grad_norm": 1.2530773878097534, |
|
"learning_rate": 7.456043956043956e-05, |
|
"loss": 0.4495, |
|
"step": 9470 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"grad_norm": 0.7528218626976013, |
|
"learning_rate": 7.453296703296703e-05, |
|
"loss": 0.4429, |
|
"step": 9480 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"grad_norm": 0.596477746963501, |
|
"learning_rate": 7.450549450549451e-05, |
|
"loss": 0.4526, |
|
"step": 9490 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"grad_norm": 0.5093569755554199, |
|
"learning_rate": 7.447802197802198e-05, |
|
"loss": 0.4419, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"grad_norm": 1.0681297779083252, |
|
"learning_rate": 7.445054945054945e-05, |
|
"loss": 0.4394, |
|
"step": 9510 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"grad_norm": 1.280721664428711, |
|
"learning_rate": 7.442307692307693e-05, |
|
"loss": 0.4425, |
|
"step": 9520 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"grad_norm": 0.8306493759155273, |
|
"learning_rate": 7.43956043956044e-05, |
|
"loss": 0.4464, |
|
"step": 9530 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"grad_norm": 0.7929884195327759, |
|
"learning_rate": 7.436813186813187e-05, |
|
"loss": 0.4532, |
|
"step": 9540 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"grad_norm": 0.9622000455856323, |
|
"learning_rate": 7.434065934065935e-05, |
|
"loss": 0.4516, |
|
"step": 9550 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"grad_norm": 0.8442659378051758, |
|
"learning_rate": 7.431318681318682e-05, |
|
"loss": 0.4417, |
|
"step": 9560 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"grad_norm": 0.6593378782272339, |
|
"learning_rate": 7.428571428571429e-05, |
|
"loss": 0.4417, |
|
"step": 9570 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"grad_norm": 0.8290765881538391, |
|
"learning_rate": 7.425824175824177e-05, |
|
"loss": 0.4471, |
|
"step": 9580 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"grad_norm": 0.9711737632751465, |
|
"learning_rate": 7.423076923076924e-05, |
|
"loss": 0.4403, |
|
"step": 9590 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"grad_norm": 0.9451545476913452, |
|
"learning_rate": 7.42032967032967e-05, |
|
"loss": 0.4522, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"grad_norm": 1.1259006261825562, |
|
"learning_rate": 7.417582417582419e-05, |
|
"loss": 0.4557, |
|
"step": 9610 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"grad_norm": 0.6328985095024109, |
|
"learning_rate": 7.414835164835165e-05, |
|
"loss": 0.4454, |
|
"step": 9620 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"grad_norm": 0.9557839035987854, |
|
"learning_rate": 7.412087912087912e-05, |
|
"loss": 0.4627, |
|
"step": 9630 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"grad_norm": 0.6751061081886292, |
|
"learning_rate": 7.40934065934066e-05, |
|
"loss": 0.4521, |
|
"step": 9640 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"grad_norm": 0.49840301275253296, |
|
"learning_rate": 7.406593406593407e-05, |
|
"loss": 0.4541, |
|
"step": 9650 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"grad_norm": 1.2455097436904907, |
|
"learning_rate": 7.403846153846154e-05, |
|
"loss": 0.4466, |
|
"step": 9660 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"grad_norm": 1.1257638931274414, |
|
"learning_rate": 7.401098901098902e-05, |
|
"loss": 0.4492, |
|
"step": 9670 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"grad_norm": 0.623653769493103, |
|
"learning_rate": 7.398351648351649e-05, |
|
"loss": 0.4475, |
|
"step": 9680 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"grad_norm": 0.6021344661712646, |
|
"learning_rate": 7.395604395604396e-05, |
|
"loss": 0.4393, |
|
"step": 9690 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"grad_norm": 0.7180885672569275, |
|
"learning_rate": 7.392857142857144e-05, |
|
"loss": 0.4571, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"grad_norm": 0.5711954832077026, |
|
"learning_rate": 7.390109890109891e-05, |
|
"loss": 0.4534, |
|
"step": 9710 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"grad_norm": 0.8202961683273315, |
|
"learning_rate": 7.387362637362638e-05, |
|
"loss": 0.4568, |
|
"step": 9720 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"grad_norm": 0.5622140765190125, |
|
"learning_rate": 7.384615384615386e-05, |
|
"loss": 0.4415, |
|
"step": 9730 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"grad_norm": 0.9798756837844849, |
|
"learning_rate": 7.381868131868133e-05, |
|
"loss": 0.4371, |
|
"step": 9740 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"grad_norm": 0.6842076778411865, |
|
"learning_rate": 7.37912087912088e-05, |
|
"loss": 0.4495, |
|
"step": 9750 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"grad_norm": 0.6150854229927063, |
|
"learning_rate": 7.376373626373628e-05, |
|
"loss": 0.4566, |
|
"step": 9760 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"grad_norm": 0.5068964958190918, |
|
"learning_rate": 7.373626373626374e-05, |
|
"loss": 0.4445, |
|
"step": 9770 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"grad_norm": 0.8547071218490601, |
|
"learning_rate": 7.370879120879121e-05, |
|
"loss": 0.4398, |
|
"step": 9780 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"grad_norm": 0.9119787216186523, |
|
"learning_rate": 7.36813186813187e-05, |
|
"loss": 0.4622, |
|
"step": 9790 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"grad_norm": 0.6831972002983093, |
|
"learning_rate": 7.365384615384616e-05, |
|
"loss": 0.4493, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"grad_norm": 1.850950002670288, |
|
"learning_rate": 7.362637362637363e-05, |
|
"loss": 0.4418, |
|
"step": 9810 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"grad_norm": 1.3177289962768555, |
|
"learning_rate": 7.359890109890111e-05, |
|
"loss": 0.4514, |
|
"step": 9820 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"grad_norm": 0.8253613114356995, |
|
"learning_rate": 7.357142857142858e-05, |
|
"loss": 0.4536, |
|
"step": 9830 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"grad_norm": 0.9882771968841553, |
|
"learning_rate": 7.354395604395605e-05, |
|
"loss": 0.4388, |
|
"step": 9840 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"grad_norm": 1.1343753337860107, |
|
"learning_rate": 7.351648351648353e-05, |
|
"loss": 0.4495, |
|
"step": 9850 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"grad_norm": 0.9710454344749451, |
|
"learning_rate": 7.3489010989011e-05, |
|
"loss": 0.4546, |
|
"step": 9860 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"grad_norm": 1.0125503540039062, |
|
"learning_rate": 7.346153846153847e-05, |
|
"loss": 0.4436, |
|
"step": 9870 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"grad_norm": 0.972205400466919, |
|
"learning_rate": 7.343406593406593e-05, |
|
"loss": 0.4603, |
|
"step": 9880 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"grad_norm": 0.8217471837997437, |
|
"learning_rate": 7.34065934065934e-05, |
|
"loss": 0.4551, |
|
"step": 9890 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"grad_norm": 0.9497684836387634, |
|
"learning_rate": 7.337912087912087e-05, |
|
"loss": 0.4398, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"grad_norm": 1.2341935634613037, |
|
"learning_rate": 7.335164835164835e-05, |
|
"loss": 0.4412, |
|
"step": 9910 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"grad_norm": 1.0097248554229736, |
|
"learning_rate": 7.332417582417582e-05, |
|
"loss": 0.4536, |
|
"step": 9920 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"grad_norm": 1.1135015487670898, |
|
"learning_rate": 7.329670329670329e-05, |
|
"loss": 0.4482, |
|
"step": 9930 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"grad_norm": 1.0130164623260498, |
|
"learning_rate": 7.326923076923077e-05, |
|
"loss": 0.4499, |
|
"step": 9940 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"grad_norm": 0.802757740020752, |
|
"learning_rate": 7.324175824175824e-05, |
|
"loss": 0.4544, |
|
"step": 9950 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"grad_norm": 0.5492908954620361, |
|
"learning_rate": 7.321428571428571e-05, |
|
"loss": 0.4457, |
|
"step": 9960 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"grad_norm": 1.1192032098770142, |
|
"learning_rate": 7.318681318681319e-05, |
|
"loss": 0.4464, |
|
"step": 9970 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"grad_norm": 0.5543531179428101, |
|
"learning_rate": 7.315934065934066e-05, |
|
"loss": 0.456, |
|
"step": 9980 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"grad_norm": 0.741788387298584, |
|
"learning_rate": 7.313186813186812e-05, |
|
"loss": 0.4384, |
|
"step": 9990 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"grad_norm": 0.6588950157165527, |
|
"learning_rate": 7.31043956043956e-05, |
|
"loss": 0.4606, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"grad_norm": 0.5929410457611084, |
|
"learning_rate": 7.307692307692307e-05, |
|
"loss": 0.4474, |
|
"step": 10010 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"grad_norm": 0.624821662902832, |
|
"learning_rate": 7.304945054945054e-05, |
|
"loss": 0.4486, |
|
"step": 10020 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"grad_norm": 0.7971046566963196, |
|
"learning_rate": 7.302197802197802e-05, |
|
"loss": 0.4433, |
|
"step": 10030 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"grad_norm": 0.81767737865448, |
|
"learning_rate": 7.299450549450549e-05, |
|
"loss": 0.4583, |
|
"step": 10040 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"grad_norm": 0.7773674726486206, |
|
"learning_rate": 7.296703296703296e-05, |
|
"loss": 0.4488, |
|
"step": 10050 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"grad_norm": 0.7860414981842041, |
|
"learning_rate": 7.293956043956044e-05, |
|
"loss": 0.4548, |
|
"step": 10060 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"grad_norm": 0.627571702003479, |
|
"learning_rate": 7.291208791208791e-05, |
|
"loss": 0.4477, |
|
"step": 10070 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"grad_norm": 0.7049760818481445, |
|
"learning_rate": 7.288461538461538e-05, |
|
"loss": 0.4376, |
|
"step": 10080 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"grad_norm": 0.5372808575630188, |
|
"learning_rate": 7.285714285714286e-05, |
|
"loss": 0.4552, |
|
"step": 10090 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"grad_norm": 0.6712595224380493, |
|
"learning_rate": 7.282967032967033e-05, |
|
"loss": 0.4555, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"grad_norm": 0.4722815752029419, |
|
"learning_rate": 7.28021978021978e-05, |
|
"loss": 0.4536, |
|
"step": 10110 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"grad_norm": 0.48171865940093994, |
|
"learning_rate": 7.277472527472528e-05, |
|
"loss": 0.4542, |
|
"step": 10120 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"grad_norm": 0.8346227407455444, |
|
"learning_rate": 7.274725274725275e-05, |
|
"loss": 0.454, |
|
"step": 10130 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"grad_norm": 0.42196372151374817, |
|
"learning_rate": 7.271978021978021e-05, |
|
"loss": 0.4514, |
|
"step": 10140 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"grad_norm": 0.655756950378418, |
|
"learning_rate": 7.26923076923077e-05, |
|
"loss": 0.458, |
|
"step": 10150 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"grad_norm": 1.0113352537155151, |
|
"learning_rate": 7.266483516483516e-05, |
|
"loss": 0.4537, |
|
"step": 10160 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"grad_norm": 0.5985142588615417, |
|
"learning_rate": 7.263736263736263e-05, |
|
"loss": 0.4408, |
|
"step": 10170 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"grad_norm": 1.50113844871521, |
|
"learning_rate": 7.260989010989011e-05, |
|
"loss": 0.4516, |
|
"step": 10180 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"grad_norm": 1.536289095878601, |
|
"learning_rate": 7.258241758241758e-05, |
|
"loss": 0.4465, |
|
"step": 10190 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"grad_norm": 0.6726506948471069, |
|
"learning_rate": 7.255494505494505e-05, |
|
"loss": 0.4365, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"grad_norm": 0.6390563249588013, |
|
"learning_rate": 7.252747252747253e-05, |
|
"loss": 0.4401, |
|
"step": 10210 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"grad_norm": 0.7105517983436584, |
|
"learning_rate": 7.25e-05, |
|
"loss": 0.445, |
|
"step": 10220 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"grad_norm": 0.9596685767173767, |
|
"learning_rate": 7.247252747252747e-05, |
|
"loss": 0.4367, |
|
"step": 10230 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"grad_norm": 1.161080002784729, |
|
"learning_rate": 7.244505494505495e-05, |
|
"loss": 0.4372, |
|
"step": 10240 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"grad_norm": 0.9476664662361145, |
|
"learning_rate": 7.241758241758242e-05, |
|
"loss": 0.4416, |
|
"step": 10250 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"grad_norm": 0.7320071458816528, |
|
"learning_rate": 7.239010989010989e-05, |
|
"loss": 0.4549, |
|
"step": 10260 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"grad_norm": 1.2132612466812134, |
|
"learning_rate": 7.236263736263737e-05, |
|
"loss": 0.4573, |
|
"step": 10270 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"grad_norm": 0.5772107839584351, |
|
"learning_rate": 7.233516483516484e-05, |
|
"loss": 0.4502, |
|
"step": 10280 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"grad_norm": 1.0478837490081787, |
|
"learning_rate": 7.23076923076923e-05, |
|
"loss": 0.452, |
|
"step": 10290 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"grad_norm": 0.6411312222480774, |
|
"learning_rate": 7.228021978021979e-05, |
|
"loss": 0.4424, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"grad_norm": 0.7831696271896362, |
|
"learning_rate": 7.225274725274725e-05, |
|
"loss": 0.4525, |
|
"step": 10310 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"grad_norm": 0.45390376448631287, |
|
"learning_rate": 7.222527472527472e-05, |
|
"loss": 0.4401, |
|
"step": 10320 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"grad_norm": 0.8081623911857605, |
|
"learning_rate": 7.21978021978022e-05, |
|
"loss": 0.4396, |
|
"step": 10330 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"grad_norm": 1.091840147972107, |
|
"learning_rate": 7.217032967032967e-05, |
|
"loss": 0.4481, |
|
"step": 10340 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"grad_norm": 1.3704334497451782, |
|
"learning_rate": 7.214285714285714e-05, |
|
"loss": 0.4563, |
|
"step": 10350 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"grad_norm": 0.8897873163223267, |
|
"learning_rate": 7.211538461538462e-05, |
|
"loss": 0.4537, |
|
"step": 10360 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"grad_norm": 1.2123398780822754, |
|
"learning_rate": 7.208791208791209e-05, |
|
"loss": 0.4421, |
|
"step": 10370 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"grad_norm": 0.7756679058074951, |
|
"learning_rate": 7.206043956043956e-05, |
|
"loss": 0.4455, |
|
"step": 10380 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"grad_norm": 0.5030421614646912, |
|
"learning_rate": 7.203296703296704e-05, |
|
"loss": 0.4482, |
|
"step": 10390 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"grad_norm": 0.5804952383041382, |
|
"learning_rate": 7.200549450549451e-05, |
|
"loss": 0.4468, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"grad_norm": 0.5755320191383362, |
|
"learning_rate": 7.197802197802198e-05, |
|
"loss": 0.4473, |
|
"step": 10410 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"grad_norm": 0.7598744034767151, |
|
"learning_rate": 7.195054945054946e-05, |
|
"loss": 0.4609, |
|
"step": 10420 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"grad_norm": 0.5830199122428894, |
|
"learning_rate": 7.192307692307693e-05, |
|
"loss": 0.4574, |
|
"step": 10430 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"grad_norm": 0.6908131837844849, |
|
"learning_rate": 7.18956043956044e-05, |
|
"loss": 0.4419, |
|
"step": 10440 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"grad_norm": 1.064139485359192, |
|
"learning_rate": 7.186813186813188e-05, |
|
"loss": 0.4467, |
|
"step": 10450 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"grad_norm": 1.5694611072540283, |
|
"learning_rate": 7.184065934065934e-05, |
|
"loss": 0.4638, |
|
"step": 10460 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"grad_norm": 0.6071065664291382, |
|
"learning_rate": 7.181318681318681e-05, |
|
"loss": 0.4449, |
|
"step": 10470 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"grad_norm": 0.7417973279953003, |
|
"learning_rate": 7.17857142857143e-05, |
|
"loss": 0.4452, |
|
"step": 10480 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"grad_norm": 1.0433728694915771, |
|
"learning_rate": 7.175824175824176e-05, |
|
"loss": 0.4413, |
|
"step": 10490 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"grad_norm": 0.5444348454475403, |
|
"learning_rate": 7.173076923076923e-05, |
|
"loss": 0.453, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"grad_norm": 1.0602647066116333, |
|
"learning_rate": 7.170329670329671e-05, |
|
"loss": 0.4489, |
|
"step": 10510 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"grad_norm": 0.7549176216125488, |
|
"learning_rate": 7.167582417582418e-05, |
|
"loss": 0.4322, |
|
"step": 10520 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"grad_norm": 0.7759855389595032, |
|
"learning_rate": 7.164835164835165e-05, |
|
"loss": 0.4459, |
|
"step": 10530 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"grad_norm": 0.6525812149047852, |
|
"learning_rate": 7.162087912087913e-05, |
|
"loss": 0.4533, |
|
"step": 10540 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"grad_norm": 0.7948151230812073, |
|
"learning_rate": 7.15934065934066e-05, |
|
"loss": 0.4412, |
|
"step": 10550 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"grad_norm": 0.8408129215240479, |
|
"learning_rate": 7.156593406593407e-05, |
|
"loss": 0.4441, |
|
"step": 10560 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"grad_norm": 0.8035346865653992, |
|
"learning_rate": 7.153846153846155e-05, |
|
"loss": 0.463, |
|
"step": 10570 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"grad_norm": 0.9992803335189819, |
|
"learning_rate": 7.151098901098902e-05, |
|
"loss": 0.4426, |
|
"step": 10580 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"grad_norm": 0.7550234794616699, |
|
"learning_rate": 7.148351648351648e-05, |
|
"loss": 0.4455, |
|
"step": 10590 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"grad_norm": 0.47757139801979065, |
|
"learning_rate": 7.145604395604397e-05, |
|
"loss": 0.4487, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"grad_norm": 0.6475192308425903, |
|
"learning_rate": 7.142857142857143e-05, |
|
"loss": 0.4493, |
|
"step": 10610 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"grad_norm": 1.0151840448379517, |
|
"learning_rate": 7.14010989010989e-05, |
|
"loss": 0.432, |
|
"step": 10620 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"grad_norm": 0.586617648601532, |
|
"learning_rate": 7.137362637362638e-05, |
|
"loss": 0.44, |
|
"step": 10630 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"grad_norm": 0.49713996052742004, |
|
"learning_rate": 7.134615384615385e-05, |
|
"loss": 0.4411, |
|
"step": 10640 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"grad_norm": 0.8117252588272095, |
|
"learning_rate": 7.131868131868132e-05, |
|
"loss": 0.4505, |
|
"step": 10650 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"grad_norm": 0.9269289374351501, |
|
"learning_rate": 7.12912087912088e-05, |
|
"loss": 0.4481, |
|
"step": 10660 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"grad_norm": 1.0942814350128174, |
|
"learning_rate": 7.126373626373627e-05, |
|
"loss": 0.4407, |
|
"step": 10670 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"grad_norm": 0.6903828382492065, |
|
"learning_rate": 7.123626373626374e-05, |
|
"loss": 0.439, |
|
"step": 10680 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"grad_norm": 0.8993271589279175, |
|
"learning_rate": 7.120879120879122e-05, |
|
"loss": 0.4421, |
|
"step": 10690 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"grad_norm": 0.8367154002189636, |
|
"learning_rate": 7.118131868131869e-05, |
|
"loss": 0.4492, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"grad_norm": 0.7720317244529724, |
|
"learning_rate": 7.115384615384616e-05, |
|
"loss": 0.4521, |
|
"step": 10710 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"grad_norm": 1.0028263330459595, |
|
"learning_rate": 7.112637362637364e-05, |
|
"loss": 0.4537, |
|
"step": 10720 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"grad_norm": 0.663088321685791, |
|
"learning_rate": 7.10989010989011e-05, |
|
"loss": 0.4484, |
|
"step": 10730 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"grad_norm": 1.360903263092041, |
|
"learning_rate": 7.107142857142857e-05, |
|
"loss": 0.4663, |
|
"step": 10740 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"grad_norm": 1.5924625396728516, |
|
"learning_rate": 7.104395604395606e-05, |
|
"loss": 0.4599, |
|
"step": 10750 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"grad_norm": 0.7223337888717651, |
|
"learning_rate": 7.101648351648352e-05, |
|
"loss": 0.4528, |
|
"step": 10760 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"grad_norm": 0.8362312316894531, |
|
"learning_rate": 7.098901098901099e-05, |
|
"loss": 0.4469, |
|
"step": 10770 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"grad_norm": 0.5594087839126587, |
|
"learning_rate": 7.096153846153847e-05, |
|
"loss": 0.4484, |
|
"step": 10780 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"grad_norm": 0.8571014404296875, |
|
"learning_rate": 7.093406593406594e-05, |
|
"loss": 0.4414, |
|
"step": 10790 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"grad_norm": 0.9150867462158203, |
|
"learning_rate": 7.090659340659341e-05, |
|
"loss": 0.453, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"grad_norm": 0.6607587337493896, |
|
"learning_rate": 7.087912087912089e-05, |
|
"loss": 0.4507, |
|
"step": 10810 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"grad_norm": 0.7428970336914062, |
|
"learning_rate": 7.085164835164836e-05, |
|
"loss": 0.4498, |
|
"step": 10820 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"grad_norm": 0.9267810583114624, |
|
"learning_rate": 7.082417582417583e-05, |
|
"loss": 0.4536, |
|
"step": 10830 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"grad_norm": 0.4540146291255951, |
|
"learning_rate": 7.079670329670331e-05, |
|
"loss": 0.4396, |
|
"step": 10840 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"grad_norm": 1.3146427869796753, |
|
"learning_rate": 7.076923076923078e-05, |
|
"loss": 0.4439, |
|
"step": 10850 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"grad_norm": 0.7954951524734497, |
|
"learning_rate": 7.074175824175825e-05, |
|
"loss": 0.4452, |
|
"step": 10860 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"grad_norm": 0.9023503661155701, |
|
"learning_rate": 7.071428571428573e-05, |
|
"loss": 0.4248, |
|
"step": 10870 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"grad_norm": 0.7970243692398071, |
|
"learning_rate": 7.06868131868132e-05, |
|
"loss": 0.4546, |
|
"step": 10880 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"grad_norm": 0.6860815286636353, |
|
"learning_rate": 7.065934065934066e-05, |
|
"loss": 0.4393, |
|
"step": 10890 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"grad_norm": 0.6077113747596741, |
|
"learning_rate": 7.063186813186815e-05, |
|
"loss": 0.4438, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"grad_norm": 0.5479728579521179, |
|
"learning_rate": 7.060439560439561e-05, |
|
"loss": 0.4474, |
|
"step": 10910 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"grad_norm": 0.8500062823295593, |
|
"learning_rate": 7.057692307692308e-05, |
|
"loss": 0.4403, |
|
"step": 10920 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"grad_norm": 0.8599127531051636, |
|
"learning_rate": 7.054945054945056e-05, |
|
"loss": 0.4367, |
|
"step": 10930 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"grad_norm": 1.3294146060943604, |
|
"learning_rate": 7.052197802197803e-05, |
|
"loss": 0.4352, |
|
"step": 10940 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"grad_norm": 0.7617499828338623, |
|
"learning_rate": 7.04945054945055e-05, |
|
"loss": 0.4454, |
|
"step": 10950 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"grad_norm": 1.2405683994293213, |
|
"learning_rate": 7.046703296703298e-05, |
|
"loss": 0.4513, |
|
"step": 10960 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"grad_norm": 0.768068253993988, |
|
"learning_rate": 7.043956043956045e-05, |
|
"loss": 0.4448, |
|
"step": 10970 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"grad_norm": 1.0528051853179932, |
|
"learning_rate": 7.041208791208792e-05, |
|
"loss": 0.4436, |
|
"step": 10980 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 0.44340217113494873, |
|
"eval_runtime": 1239.9169, |
|
"eval_samples_per_second": 235.588, |
|
"eval_steps_per_second": 3.682, |
|
"step": 10982 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 36600, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 10, |
|
"save_steps": 500, |
|
"total_flos": 4.683801644117262e+18, |
|
"train_batch_size": 6, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|