|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.994059405940594, |
|
"eval_steps": 500, |
|
"global_step": 378, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.007920792079207921, |
|
"grad_norm": 5.2691683098946545, |
|
"learning_rate": 5.263157894736843e-07, |
|
"loss": 1.568, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.015841584158415842, |
|
"grad_norm": 5.00511067394956, |
|
"learning_rate": 1.0526315789473685e-06, |
|
"loss": 1.5221, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.023762376237623763, |
|
"grad_norm": 4.54506426631759, |
|
"learning_rate": 1.5789473684210526e-06, |
|
"loss": 1.5171, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.031683168316831684, |
|
"grad_norm": 4.36896006607012, |
|
"learning_rate": 2.105263157894737e-06, |
|
"loss": 1.5391, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.039603960396039604, |
|
"grad_norm": 3.353266992703984, |
|
"learning_rate": 2.631578947368421e-06, |
|
"loss": 1.4111, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.047524752475247525, |
|
"grad_norm": 2.920541207093266, |
|
"learning_rate": 3.157894736842105e-06, |
|
"loss": 1.3994, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.055445544554455446, |
|
"grad_norm": 1.916780394273478, |
|
"learning_rate": 3.6842105263157896e-06, |
|
"loss": 1.3978, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.06336633663366337, |
|
"grad_norm": 1.7850272367832638, |
|
"learning_rate": 4.210526315789474e-06, |
|
"loss": 1.3282, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.07128712871287128, |
|
"grad_norm": 1.7482580453609655, |
|
"learning_rate": 4.736842105263158e-06, |
|
"loss": 1.2986, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.07920792079207921, |
|
"grad_norm": 1.449543281626886, |
|
"learning_rate": 5.263157894736842e-06, |
|
"loss": 1.2898, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.08712871287128712, |
|
"grad_norm": 1.8720398819253754, |
|
"learning_rate": 5.789473684210527e-06, |
|
"loss": 1.2781, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.09504950495049505, |
|
"grad_norm": 2.2015426049634503, |
|
"learning_rate": 6.31578947368421e-06, |
|
"loss": 1.2935, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.10297029702970296, |
|
"grad_norm": 2.615874858377271, |
|
"learning_rate": 6.842105263157896e-06, |
|
"loss": 1.2973, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.11089108910891089, |
|
"grad_norm": 3.8682995800654187, |
|
"learning_rate": 7.368421052631579e-06, |
|
"loss": 1.2402, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.1188118811881188, |
|
"grad_norm": 2.746699538417491, |
|
"learning_rate": 7.894736842105265e-06, |
|
"loss": 1.2562, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.12673267326732673, |
|
"grad_norm": 2.682031995657898, |
|
"learning_rate": 8.421052631578948e-06, |
|
"loss": 1.2414, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.13465346534653466, |
|
"grad_norm": 2.9242715554029663, |
|
"learning_rate": 8.947368421052632e-06, |
|
"loss": 1.173, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.14257425742574256, |
|
"grad_norm": 2.7715021870832386, |
|
"learning_rate": 9.473684210526315e-06, |
|
"loss": 1.218, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.1504950495049505, |
|
"grad_norm": 2.3528632221189607, |
|
"learning_rate": 1e-05, |
|
"loss": 1.1863, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.15841584158415842, |
|
"grad_norm": 2.034765686956333, |
|
"learning_rate": 1.0526315789473684e-05, |
|
"loss": 1.1888, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.16633663366336635, |
|
"grad_norm": 3.1288870944867666, |
|
"learning_rate": 1.105263157894737e-05, |
|
"loss": 1.1971, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.17425742574257425, |
|
"grad_norm": 2.195732986258774, |
|
"learning_rate": 1.1578947368421053e-05, |
|
"loss": 1.1866, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.18217821782178217, |
|
"grad_norm": 4.112336771867653, |
|
"learning_rate": 1.2105263157894737e-05, |
|
"loss": 1.1563, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.1900990099009901, |
|
"grad_norm": 3.6328533281903526, |
|
"learning_rate": 1.263157894736842e-05, |
|
"loss": 1.175, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.19801980198019803, |
|
"grad_norm": 1.6325540508979153, |
|
"learning_rate": 1.3157894736842108e-05, |
|
"loss": 1.1646, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.20594059405940593, |
|
"grad_norm": 3.9494616695340077, |
|
"learning_rate": 1.3684210526315791e-05, |
|
"loss": 1.1323, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.21386138613861386, |
|
"grad_norm": 3.27189931951213, |
|
"learning_rate": 1.4210526315789475e-05, |
|
"loss": 1.1408, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.22178217821782178, |
|
"grad_norm": 2.6804790210906573, |
|
"learning_rate": 1.4736842105263159e-05, |
|
"loss": 1.1319, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.2297029702970297, |
|
"grad_norm": 1.1471627340033614, |
|
"learning_rate": 1.5263157894736846e-05, |
|
"loss": 1.1482, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.2376237623762376, |
|
"grad_norm": 2.002020630505531, |
|
"learning_rate": 1.578947368421053e-05, |
|
"loss": 1.1494, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.24554455445544554, |
|
"grad_norm": 2.4723778287869616, |
|
"learning_rate": 1.6315789473684213e-05, |
|
"loss": 1.1441, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.25346534653465347, |
|
"grad_norm": 2.229781993318989, |
|
"learning_rate": 1.6842105263157896e-05, |
|
"loss": 1.1235, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.2613861386138614, |
|
"grad_norm": 1.7701112722120012, |
|
"learning_rate": 1.736842105263158e-05, |
|
"loss": 1.1126, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.2693069306930693, |
|
"grad_norm": 1.064748424759578, |
|
"learning_rate": 1.7894736842105264e-05, |
|
"loss": 1.1595, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.27722772277227725, |
|
"grad_norm": 1.2312931929901163, |
|
"learning_rate": 1.8421052631578947e-05, |
|
"loss": 1.1668, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.2851485148514851, |
|
"grad_norm": 1.7551020997420819, |
|
"learning_rate": 1.894736842105263e-05, |
|
"loss": 1.0913, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.29306930693069305, |
|
"grad_norm": 1.7687041331840054, |
|
"learning_rate": 1.9473684210526318e-05, |
|
"loss": 1.1152, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.300990099009901, |
|
"grad_norm": 0.5975775886980378, |
|
"learning_rate": 2e-05, |
|
"loss": 1.1279, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.3089108910891089, |
|
"grad_norm": 1.7305083714637768, |
|
"learning_rate": 1.999957311703368e-05, |
|
"loss": 1.0947, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.31683168316831684, |
|
"grad_norm": 1.844146398722995, |
|
"learning_rate": 1.9998292504580528e-05, |
|
"loss": 1.0906, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.32475247524752476, |
|
"grad_norm": 0.7796143095441499, |
|
"learning_rate": 1.9996158271974875e-05, |
|
"loss": 1.1236, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.3326732673267327, |
|
"grad_norm": 1.5056201561723581, |
|
"learning_rate": 1.9993170601430233e-05, |
|
"loss": 1.0662, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.3405940594059406, |
|
"grad_norm": 1.8181218416149059, |
|
"learning_rate": 1.9989329748023728e-05, |
|
"loss": 1.1041, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.3485148514851485, |
|
"grad_norm": 1.0925270287101072, |
|
"learning_rate": 1.9984636039674342e-05, |
|
"loss": 1.1226, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.3564356435643564, |
|
"grad_norm": 0.9921305736361253, |
|
"learning_rate": 1.9979089877114905e-05, |
|
"loss": 1.0756, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.36435643564356435, |
|
"grad_norm": 1.2754799989749876, |
|
"learning_rate": 1.997269173385788e-05, |
|
"loss": 1.1173, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.3722772277227723, |
|
"grad_norm": 0.8010641417555526, |
|
"learning_rate": 1.9965442156154947e-05, |
|
"loss": 1.1113, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.3801980198019802, |
|
"grad_norm": 0.7444563010730426, |
|
"learning_rate": 1.9957341762950346e-05, |
|
"loss": 1.1396, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.38811881188118813, |
|
"grad_norm": 1.1638152394745782, |
|
"learning_rate": 1.994839124582806e-05, |
|
"loss": 1.1117, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.39603960396039606, |
|
"grad_norm": 0.7261236266090463, |
|
"learning_rate": 1.993859136895274e-05, |
|
"loss": 1.0745, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.403960396039604, |
|
"grad_norm": 0.6556537970861287, |
|
"learning_rate": 1.9927942969004493e-05, |
|
"loss": 1.0783, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.41188118811881186, |
|
"grad_norm": 1.0338775189848177, |
|
"learning_rate": 1.991644695510743e-05, |
|
"loss": 1.1053, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.4198019801980198, |
|
"grad_norm": 0.5045982803365276, |
|
"learning_rate": 1.9904104308752053e-05, |
|
"loss": 1.1187, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.4277227722772277, |
|
"grad_norm": 0.8811507158721841, |
|
"learning_rate": 1.9890916083711463e-05, |
|
"loss": 1.0666, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.43564356435643564, |
|
"grad_norm": 0.7323789997357792, |
|
"learning_rate": 1.9876883405951378e-05, |
|
"loss": 1.0613, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.44356435643564357, |
|
"grad_norm": 0.43382558782064967, |
|
"learning_rate": 1.9862007473534026e-05, |
|
"loss": 1.0946, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.4514851485148515, |
|
"grad_norm": 0.7855920569177011, |
|
"learning_rate": 1.9846289556515835e-05, |
|
"loss": 1.0801, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.4594059405940594, |
|
"grad_norm": 0.5060051358313292, |
|
"learning_rate": 1.982973099683902e-05, |
|
"loss": 1.0559, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.46732673267326735, |
|
"grad_norm": 0.6612232058281015, |
|
"learning_rate": 1.981233320821699e-05, |
|
"loss": 1.0684, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.4752475247524752, |
|
"grad_norm": 0.6407758050553994, |
|
"learning_rate": 1.979409767601366e-05, |
|
"loss": 1.0392, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.48316831683168315, |
|
"grad_norm": 0.4238055278476331, |
|
"learning_rate": 1.9775025957116657e-05, |
|
"loss": 1.0852, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.4910891089108911, |
|
"grad_norm": 0.6366553572873023, |
|
"learning_rate": 1.975511967980437e-05, |
|
"loss": 1.0718, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.499009900990099, |
|
"grad_norm": 0.4560763634964661, |
|
"learning_rate": 1.9734380543606932e-05, |
|
"loss": 1.0491, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.5069306930693069, |
|
"grad_norm": 0.47744475456040364, |
|
"learning_rate": 1.971281031916114e-05, |
|
"loss": 1.0682, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.5148514851485149, |
|
"grad_norm": 0.6139591331770152, |
|
"learning_rate": 1.9690410848059278e-05, |
|
"loss": 1.0859, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.5227722772277228, |
|
"grad_norm": 0.4633776390020583, |
|
"learning_rate": 1.9667184042691877e-05, |
|
"loss": 1.0637, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.5306930693069307, |
|
"grad_norm": 0.48132534481018163, |
|
"learning_rate": 1.964313188608445e-05, |
|
"loss": 1.1127, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.5386138613861386, |
|
"grad_norm": 0.5078593046327664, |
|
"learning_rate": 1.961825643172819e-05, |
|
"loss": 1.0666, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.5465346534653466, |
|
"grad_norm": 0.4641018705804529, |
|
"learning_rate": 1.9592559803404652e-05, |
|
"loss": 1.0412, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.5544554455445545, |
|
"grad_norm": 0.4477782377674348, |
|
"learning_rate": 1.956604419500441e-05, |
|
"loss": 1.0319, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.5623762376237624, |
|
"grad_norm": 0.4399022823054154, |
|
"learning_rate": 1.953871187033978e-05, |
|
"loss": 1.0433, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.5702970297029702, |
|
"grad_norm": 0.49392160981903666, |
|
"learning_rate": 1.9510565162951538e-05, |
|
"loss": 1.063, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.5782178217821782, |
|
"grad_norm": 0.47294899191298795, |
|
"learning_rate": 1.948160647590966e-05, |
|
"loss": 1.0749, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.5861386138613861, |
|
"grad_norm": 0.41344043915965834, |
|
"learning_rate": 1.94518382816082e-05, |
|
"loss": 1.0979, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.594059405940594, |
|
"grad_norm": 0.5011686560663466, |
|
"learning_rate": 1.9421263121554163e-05, |
|
"loss": 1.015, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.601980198019802, |
|
"grad_norm": 0.39582414491214485, |
|
"learning_rate": 1.938988360615057e-05, |
|
"loss": 1.0383, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.6099009900990099, |
|
"grad_norm": 0.48768604319919506, |
|
"learning_rate": 1.9357702414473528e-05, |
|
"loss": 1.0526, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.6178217821782178, |
|
"grad_norm": 0.45230653840089824, |
|
"learning_rate": 1.932472229404356e-05, |
|
"loss": 1.0308, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.6257425742574257, |
|
"grad_norm": 0.4138418438484353, |
|
"learning_rate": 1.9290946060590992e-05, |
|
"loss": 1.0532, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.6336633663366337, |
|
"grad_norm": 0.4370072818318615, |
|
"learning_rate": 1.9256376597815565e-05, |
|
"loss": 1.0673, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.6415841584158416, |
|
"grad_norm": 0.39792387068550145, |
|
"learning_rate": 1.9221016857140244e-05, |
|
"loss": 1.0101, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.6495049504950495, |
|
"grad_norm": 0.5007989711284164, |
|
"learning_rate": 1.9184869857459233e-05, |
|
"loss": 1.012, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.6574257425742575, |
|
"grad_norm": 0.4103210495096538, |
|
"learning_rate": 1.9147938684880213e-05, |
|
"loss": 1.0374, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.6653465346534654, |
|
"grad_norm": 0.4872855085217254, |
|
"learning_rate": 1.9110226492460886e-05, |
|
"loss": 1.0859, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.6732673267326733, |
|
"grad_norm": 0.422627898101724, |
|
"learning_rate": 1.9071736499939765e-05, |
|
"loss": 1.0518, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.6811881188118812, |
|
"grad_norm": 0.41564147530018053, |
|
"learning_rate": 1.903247199346129e-05, |
|
"loss": 1.0178, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.689108910891089, |
|
"grad_norm": 0.4358424246570489, |
|
"learning_rate": 1.8992436325295258e-05, |
|
"loss": 1.0677, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.697029702970297, |
|
"grad_norm": 0.41355898391860707, |
|
"learning_rate": 1.8951632913550625e-05, |
|
"loss": 1.0237, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.7049504950495049, |
|
"grad_norm": 0.3823609847824327, |
|
"learning_rate": 1.891006524188368e-05, |
|
"loss": 0.9833, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.7128712871287128, |
|
"grad_norm": 0.39209473964132163, |
|
"learning_rate": 1.886773685920062e-05, |
|
"loss": 1.0451, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.7207920792079208, |
|
"grad_norm": 0.4140041300883924, |
|
"learning_rate": 1.882465137935456e-05, |
|
"loss": 1.0222, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.7287128712871287, |
|
"grad_norm": 0.40375021194700533, |
|
"learning_rate": 1.878081248083698e-05, |
|
"loss": 1.0952, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.7366336633663366, |
|
"grad_norm": 0.4414196189759205, |
|
"learning_rate": 1.8736223906463698e-05, |
|
"loss": 1.0156, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.7445544554455445, |
|
"grad_norm": 0.4386744025847356, |
|
"learning_rate": 1.8690889463055285e-05, |
|
"loss": 1.0279, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.7524752475247525, |
|
"grad_norm": 0.7947840810413794, |
|
"learning_rate": 1.864481302111208e-05, |
|
"loss": 1.0593, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.7603960396039604, |
|
"grad_norm": 0.42527281294211106, |
|
"learning_rate": 1.8597998514483724e-05, |
|
"loss": 1.0055, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.7683168316831683, |
|
"grad_norm": 0.44600952063933513, |
|
"learning_rate": 1.855044994003331e-05, |
|
"loss": 1.0189, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.7762376237623763, |
|
"grad_norm": 0.41381659543669813, |
|
"learning_rate": 1.8502171357296144e-05, |
|
"loss": 1.0385, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.7841584158415842, |
|
"grad_norm": 0.4356737265816363, |
|
"learning_rate": 1.845316688813314e-05, |
|
"loss": 1.0188, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.7920792079207921, |
|
"grad_norm": 0.3910640622588359, |
|
"learning_rate": 1.840344071637893e-05, |
|
"loss": 1.0, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 2.5929595435725368, |
|
"learning_rate": 1.8352997087484657e-05, |
|
"loss": 1.0768, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.807920792079208, |
|
"grad_norm": 0.39724371010118187, |
|
"learning_rate": 1.8301840308155507e-05, |
|
"loss": 1.0546, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.8158415841584158, |
|
"grad_norm": 0.3924179902332133, |
|
"learning_rate": 1.8249974745983023e-05, |
|
"loss": 1.0163, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.8237623762376237, |
|
"grad_norm": 0.4531863562551441, |
|
"learning_rate": 1.8197404829072214e-05, |
|
"loss": 1.0125, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.8316831683168316, |
|
"grad_norm": 0.44574269385841797, |
|
"learning_rate": 1.8144135045663486e-05, |
|
"loss": 1.0399, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.8396039603960396, |
|
"grad_norm": 0.3994623879554383, |
|
"learning_rate": 1.8090169943749477e-05, |
|
"loss": 1.0103, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.8475247524752475, |
|
"grad_norm": 0.41462016834859783, |
|
"learning_rate": 1.8035514130686737e-05, |
|
"loss": 1.0474, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.8554455445544554, |
|
"grad_norm": 0.44011181806180777, |
|
"learning_rate": 1.7980172272802398e-05, |
|
"loss": 1.0274, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.8633663366336634, |
|
"grad_norm": 0.4581477822740845, |
|
"learning_rate": 1.792414909499574e-05, |
|
"loss": 1.0291, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.8712871287128713, |
|
"grad_norm": 0.3880729293901145, |
|
"learning_rate": 1.7867449380334834e-05, |
|
"loss": 1.0304, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.8792079207920792, |
|
"grad_norm": 0.3756477226180992, |
|
"learning_rate": 1.7810077969648157e-05, |
|
"loss": 0.9737, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.8871287128712871, |
|
"grad_norm": 0.38501494156072325, |
|
"learning_rate": 1.77520397611113e-05, |
|
"loss": 1.0222, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.8950495049504951, |
|
"grad_norm": 0.3812791473176217, |
|
"learning_rate": 1.769333970982879e-05, |
|
"loss": 1.0568, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.902970297029703, |
|
"grad_norm": 0.38636764366182014, |
|
"learning_rate": 1.763398282741103e-05, |
|
"loss": 1.0556, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.9108910891089109, |
|
"grad_norm": 0.3689101295336194, |
|
"learning_rate": 1.757397418154643e-05, |
|
"loss": 1.0143, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.9188118811881189, |
|
"grad_norm": 0.36527609469057676, |
|
"learning_rate": 1.7513318895568734e-05, |
|
"loss": 0.9986, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.9267326732673268, |
|
"grad_norm": 0.3986163851527506, |
|
"learning_rate": 1.7452022148019626e-05, |
|
"loss": 1.0462, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.9346534653465347, |
|
"grad_norm": 0.37735673327905117, |
|
"learning_rate": 1.7390089172206594e-05, |
|
"loss": 1.0072, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.9425742574257425, |
|
"grad_norm": 0.37706765852085494, |
|
"learning_rate": 1.7327525255756118e-05, |
|
"loss": 1.0323, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.9504950495049505, |
|
"grad_norm": 0.39377310916431657, |
|
"learning_rate": 1.7264335740162244e-05, |
|
"loss": 1.0422, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.9584158415841584, |
|
"grad_norm": 0.41314537071793267, |
|
"learning_rate": 1.720052602033055e-05, |
|
"loss": 0.9771, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.9663366336633663, |
|
"grad_norm": 0.37875715597090986, |
|
"learning_rate": 1.7136101544117526e-05, |
|
"loss": 0.9968, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.9742574257425742, |
|
"grad_norm": 0.37159716453417896, |
|
"learning_rate": 1.7071067811865477e-05, |
|
"loss": 1.0158, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.9821782178217822, |
|
"grad_norm": 0.36633287049364816, |
|
"learning_rate": 1.700543037593291e-05, |
|
"loss": 1.0113, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.9900990099009901, |
|
"grad_norm": 0.3639299308468012, |
|
"learning_rate": 1.6939194840220497e-05, |
|
"loss": 1.0108, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.998019801980198, |
|
"grad_norm": 0.3687515525421076, |
|
"learning_rate": 1.687236685969263e-05, |
|
"loss": 1.0183, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.998019801980198, |
|
"eval_loss": 1.0126549005508423, |
|
"eval_runtime": 87.7583, |
|
"eval_samples_per_second": 3.031, |
|
"eval_steps_per_second": 0.194, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 1.005940594059406, |
|
"grad_norm": 0.6033515169314877, |
|
"learning_rate": 1.6804952139894618e-05, |
|
"loss": 0.9244, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 1.0138613861386139, |
|
"grad_norm": 0.5458375892374859, |
|
"learning_rate": 1.6736956436465573e-05, |
|
"loss": 0.8684, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 1.0217821782178218, |
|
"grad_norm": 0.6652946196593673, |
|
"learning_rate": 1.6668385554647017e-05, |
|
"loss": 0.8744, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 1.0297029702970297, |
|
"grad_norm": 0.47420107863397504, |
|
"learning_rate": 1.659924534878723e-05, |
|
"loss": 0.8562, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 1.0376237623762377, |
|
"grad_norm": 0.6395196023840517, |
|
"learning_rate": 1.6529541721841444e-05, |
|
"loss": 0.8653, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 1.0455445544554456, |
|
"grad_norm": 0.49490432639088666, |
|
"learning_rate": 1.6459280624867876e-05, |
|
"loss": 0.8939, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 1.0534653465346535, |
|
"grad_norm": 0.662156431335598, |
|
"learning_rate": 1.638846805651961e-05, |
|
"loss": 0.8772, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 1.0613861386138614, |
|
"grad_norm": 0.4315607187688008, |
|
"learning_rate": 1.631711006253251e-05, |
|
"loss": 0.8719, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 1.0693069306930694, |
|
"grad_norm": 0.5428981275030659, |
|
"learning_rate": 1.6245212735208994e-05, |
|
"loss": 0.8902, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 1.0772277227722773, |
|
"grad_norm": 0.4212431283579326, |
|
"learning_rate": 1.617278221289793e-05, |
|
"loss": 0.8958, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 1.0851485148514852, |
|
"grad_norm": 0.45956779312336216, |
|
"learning_rate": 1.609982467947057e-05, |
|
"loss": 0.8798, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 1.0930693069306932, |
|
"grad_norm": 0.5318677463313782, |
|
"learning_rate": 1.6026346363792565e-05, |
|
"loss": 0.8691, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 1.100990099009901, |
|
"grad_norm": 0.420504194389614, |
|
"learning_rate": 1.595235353919219e-05, |
|
"loss": 0.8711, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 1.108910891089109, |
|
"grad_norm": 0.4826547764308898, |
|
"learning_rate": 1.5877852522924733e-05, |
|
"loss": 0.8908, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 1.116831683168317, |
|
"grad_norm": 0.3992971335438903, |
|
"learning_rate": 1.580284967563316e-05, |
|
"loss": 0.8591, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 1.1247524752475249, |
|
"grad_norm": 0.4402406328520412, |
|
"learning_rate": 1.5727351400805054e-05, |
|
"loss": 0.839, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 1.1326732673267328, |
|
"grad_norm": 0.41454275441137073, |
|
"learning_rate": 1.565136414422592e-05, |
|
"loss": 0.8566, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 1.1405940594059407, |
|
"grad_norm": 0.4296709034986113, |
|
"learning_rate": 1.5574894393428856e-05, |
|
"loss": 0.8881, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 1.1485148514851484, |
|
"grad_norm": 0.45117629451958274, |
|
"learning_rate": 1.5497948677140673e-05, |
|
"loss": 0.8457, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 1.1564356435643564, |
|
"grad_norm": 0.3989063232783904, |
|
"learning_rate": 1.5420533564724495e-05, |
|
"loss": 0.8805, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 1.1643564356435643, |
|
"grad_norm": 0.44568328942669366, |
|
"learning_rate": 1.5342655665618885e-05, |
|
"loss": 0.853, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 1.1722772277227722, |
|
"grad_norm": 0.39828075995328915, |
|
"learning_rate": 1.526432162877356e-05, |
|
"loss": 0.8831, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 1.1801980198019801, |
|
"grad_norm": 0.41294672710780717, |
|
"learning_rate": 1.5185538142081721e-05, |
|
"loss": 0.8786, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 1.188118811881188, |
|
"grad_norm": 0.40873959614869637, |
|
"learning_rate": 1.510631193180907e-05, |
|
"loss": 0.8684, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 1.196039603960396, |
|
"grad_norm": 0.3978636394954677, |
|
"learning_rate": 1.5026649762019539e-05, |
|
"loss": 0.8756, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 1.203960396039604, |
|
"grad_norm": 0.5036601026033098, |
|
"learning_rate": 1.4946558433997792e-05, |
|
"loss": 0.844, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 1.2118811881188118, |
|
"grad_norm": 0.3932657455702579, |
|
"learning_rate": 1.4866044785668563e-05, |
|
"loss": 0.801, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 1.2198019801980198, |
|
"grad_norm": 0.41396973144650107, |
|
"learning_rate": 1.4785115691012866e-05, |
|
"loss": 0.8854, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 1.2277227722772277, |
|
"grad_norm": 0.4095963357893977, |
|
"learning_rate": 1.4703778059481096e-05, |
|
"loss": 0.8629, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 1.2356435643564356, |
|
"grad_norm": 0.3844996926085625, |
|
"learning_rate": 1.4622038835403135e-05, |
|
"loss": 0.8802, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 1.2435643564356436, |
|
"grad_norm": 0.38098631899426383, |
|
"learning_rate": 1.4539904997395468e-05, |
|
"loss": 0.8546, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 1.2514851485148515, |
|
"grad_norm": 0.38645301150023, |
|
"learning_rate": 1.4457383557765385e-05, |
|
"loss": 0.8745, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 1.2594059405940594, |
|
"grad_norm": 0.38284619702545075, |
|
"learning_rate": 1.4374481561912266e-05, |
|
"loss": 0.8618, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 1.2673267326732673, |
|
"grad_norm": 0.4209251000095392, |
|
"learning_rate": 1.429120608772609e-05, |
|
"loss": 0.8839, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 1.2752475247524753, |
|
"grad_norm": 0.402410646316122, |
|
"learning_rate": 1.4207564244983154e-05, |
|
"loss": 0.8741, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 1.2831683168316832, |
|
"grad_norm": 0.39477095820040203, |
|
"learning_rate": 1.4123563174739036e-05, |
|
"loss": 0.8883, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 1.2910891089108911, |
|
"grad_norm": 0.3924015414958881, |
|
"learning_rate": 1.403921004871895e-05, |
|
"loss": 0.8576, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 1.299009900990099, |
|
"grad_norm": 0.4161025068008364, |
|
"learning_rate": 1.3954512068705425e-05, |
|
"loss": 0.8305, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 1.306930693069307, |
|
"grad_norm": 0.4263635509257888, |
|
"learning_rate": 1.3869476465923455e-05, |
|
"loss": 0.8647, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 1.314851485148515, |
|
"grad_norm": 0.39404329089929485, |
|
"learning_rate": 1.3784110500423104e-05, |
|
"loss": 0.8622, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 1.3227722772277228, |
|
"grad_norm": 0.3805109485657045, |
|
"learning_rate": 1.3698421460459692e-05, |
|
"loss": 0.862, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 1.3306930693069308, |
|
"grad_norm": 0.40690294201503635, |
|
"learning_rate": 1.3612416661871532e-05, |
|
"loss": 0.8647, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 1.3386138613861387, |
|
"grad_norm": 0.4259562065252122, |
|
"learning_rate": 1.3526103447455326e-05, |
|
"loss": 0.8563, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 1.3465346534653464, |
|
"grad_norm": 0.3810965967304589, |
|
"learning_rate": 1.3439489186339283e-05, |
|
"loss": 0.8311, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 1.3544554455445543, |
|
"grad_norm": 0.39136590416953965, |
|
"learning_rate": 1.335258127335394e-05, |
|
"loss": 0.8471, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 1.3623762376237623, |
|
"grad_norm": 0.41483855969298006, |
|
"learning_rate": 1.3265387128400833e-05, |
|
"loss": 0.8609, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 1.3702970297029702, |
|
"grad_norm": 0.40516699447727095, |
|
"learning_rate": 1.3177914195819018e-05, |
|
"loss": 0.8275, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 1.378217821782178, |
|
"grad_norm": 0.4793820142436531, |
|
"learning_rate": 1.3090169943749475e-05, |
|
"loss": 0.8749, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 1.386138613861386, |
|
"grad_norm": 0.3840703344721545, |
|
"learning_rate": 1.3002161863497529e-05, |
|
"loss": 0.8966, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 1.394059405940594, |
|
"grad_norm": 0.41308469917146595, |
|
"learning_rate": 1.2913897468893249e-05, |
|
"loss": 0.8914, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 1.401980198019802, |
|
"grad_norm": 0.3827637794217713, |
|
"learning_rate": 1.2825384295649952e-05, |
|
"loss": 0.8299, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 1.4099009900990098, |
|
"grad_norm": 0.4492756954997741, |
|
"learning_rate": 1.2736629900720832e-05, |
|
"loss": 0.8736, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 1.4178217821782177, |
|
"grad_norm": 0.3770575870144539, |
|
"learning_rate": 1.2647641861653759e-05, |
|
"loss": 0.8602, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 1.4257425742574257, |
|
"grad_norm": 0.3900899921685637, |
|
"learning_rate": 1.2558427775944357e-05, |
|
"loss": 0.8365, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 1.4336633663366336, |
|
"grad_norm": 0.41840094530944894, |
|
"learning_rate": 1.2468995260387332e-05, |
|
"loss": 0.8993, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 1.4415841584158415, |
|
"grad_norm": 0.40583158631957733, |
|
"learning_rate": 1.2379351950426188e-05, |
|
"loss": 0.8691, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 1.4495049504950495, |
|
"grad_norm": 0.3966569139417299, |
|
"learning_rate": 1.2289505499501341e-05, |
|
"loss": 0.8711, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 1.4574257425742574, |
|
"grad_norm": 0.41194466483513786, |
|
"learning_rate": 1.2199463578396688e-05, |
|
"loss": 0.8338, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 1.4653465346534653, |
|
"grad_norm": 0.3818733353460163, |
|
"learning_rate": 1.21092338745847e-05, |
|
"loss": 0.8403, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 1.4732673267326732, |
|
"grad_norm": 0.48110222651862666, |
|
"learning_rate": 1.2018824091570103e-05, |
|
"loss": 0.8359, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 1.4811881188118812, |
|
"grad_norm": 0.39846082622633, |
|
"learning_rate": 1.192824194823217e-05, |
|
"loss": 0.8862, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 1.489108910891089, |
|
"grad_norm": 0.4794826007648338, |
|
"learning_rate": 1.1837495178165706e-05, |
|
"loss": 0.8611, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 1.497029702970297, |
|
"grad_norm": 0.41925943616423056, |
|
"learning_rate": 1.1746591529020789e-05, |
|
"loss": 0.8734, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 1.504950495049505, |
|
"grad_norm": 0.40682041747226527, |
|
"learning_rate": 1.16555387618413e-05, |
|
"loss": 0.85, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 1.5128712871287129, |
|
"grad_norm": 0.3925259196133537, |
|
"learning_rate": 1.156434465040231e-05, |
|
"loss": 0.8685, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 1.5207920792079208, |
|
"grad_norm": 0.3917596296811381, |
|
"learning_rate": 1.1473016980546377e-05, |
|
"loss": 0.8613, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 1.5287128712871287, |
|
"grad_norm": 0.3916872938793011, |
|
"learning_rate": 1.1381563549518823e-05, |
|
"loss": 0.8708, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 1.5366336633663367, |
|
"grad_norm": 0.4079178768846547, |
|
"learning_rate": 1.1289992165302036e-05, |
|
"loss": 0.8444, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 1.5445544554455446, |
|
"grad_norm": 0.4239468808643011, |
|
"learning_rate": 1.1198310645948833e-05, |
|
"loss": 0.8475, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 1.5524752475247525, |
|
"grad_norm": 0.4143090508458373, |
|
"learning_rate": 1.1106526818915008e-05, |
|
"loss": 0.8584, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 1.5603960396039604, |
|
"grad_norm": 0.3749410819598408, |
|
"learning_rate": 1.1014648520391031e-05, |
|
"loss": 0.8694, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 1.5683168316831684, |
|
"grad_norm": 0.38698513255681644, |
|
"learning_rate": 1.092268359463302e-05, |
|
"loss": 0.8895, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 1.5762376237623763, |
|
"grad_norm": 0.3945257947528202, |
|
"learning_rate": 1.083063989329304e-05, |
|
"loss": 0.8437, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 1.5841584158415842, |
|
"grad_norm": 0.38726772534551934, |
|
"learning_rate": 1.073852527474874e-05, |
|
"loss": 0.9063, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 1.5920792079207922, |
|
"grad_norm": 0.3806123804826243, |
|
"learning_rate": 1.0646347603432443e-05, |
|
"loss": 0.8432, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"grad_norm": 0.39146660949081974, |
|
"learning_rate": 1.05541147491597e-05, |
|
"loss": 0.8676, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 1.607920792079208, |
|
"grad_norm": 0.3845422288308008, |
|
"learning_rate": 1.0461834586457398e-05, |
|
"loss": 0.852, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 1.615841584158416, |
|
"grad_norm": 0.3854898725988794, |
|
"learning_rate": 1.0369514993891451e-05, |
|
"loss": 0.8376, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 1.6237623762376239, |
|
"grad_norm": 0.40353321866721364, |
|
"learning_rate": 1.0277163853394166e-05, |
|
"loss": 0.852, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 1.6316831683168318, |
|
"grad_norm": 0.38444130174939833, |
|
"learning_rate": 1.01847890495913e-05, |
|
"loss": 0.889, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 1.6396039603960397, |
|
"grad_norm": 0.38108640413144107, |
|
"learning_rate": 1.009239846912891e-05, |
|
"loss": 0.8359, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 1.6475247524752477, |
|
"grad_norm": 0.4747347151817832, |
|
"learning_rate": 1e-05, |
|
"loss": 0.8865, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 1.6554455445544556, |
|
"grad_norm": 0.40452870424857584, |
|
"learning_rate": 9.907601530871094e-06, |
|
"loss": 0.8253, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 1.6633663366336635, |
|
"grad_norm": 0.38987399344119467, |
|
"learning_rate": 9.815210950408703e-06, |
|
"loss": 0.849, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 1.6712871287128714, |
|
"grad_norm": 0.39493082174252364, |
|
"learning_rate": 9.722836146605838e-06, |
|
"loss": 0.8483, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 1.6792079207920794, |
|
"grad_norm": 0.39165994965141043, |
|
"learning_rate": 9.630485006108554e-06, |
|
"loss": 0.8871, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 1.6871287128712873, |
|
"grad_norm": 0.38092018061735106, |
|
"learning_rate": 9.538165413542607e-06, |
|
"loss": 0.8518, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 1.695049504950495, |
|
"grad_norm": 0.3855485261645927, |
|
"learning_rate": 9.445885250840301e-06, |
|
"loss": 0.8356, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 1.702970297029703, |
|
"grad_norm": 0.39414553727691437, |
|
"learning_rate": 9.353652396567558e-06, |
|
"loss": 0.8689, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 1.7108910891089109, |
|
"grad_norm": 0.383608920878984, |
|
"learning_rate": 9.261474725251261e-06, |
|
"loss": 0.8462, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 1.7188118811881188, |
|
"grad_norm": 0.37956984957800455, |
|
"learning_rate": 9.169360106706962e-06, |
|
"loss": 0.8636, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 1.7267326732673267, |
|
"grad_norm": 0.3952521158053759, |
|
"learning_rate": 9.07731640536698e-06, |
|
"loss": 0.8371, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 1.7346534653465346, |
|
"grad_norm": 0.39432968092671433, |
|
"learning_rate": 8.985351479608972e-06, |
|
"loss": 0.8686, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 1.7425742574257426, |
|
"grad_norm": 0.4091695075116408, |
|
"learning_rate": 8.893473181084993e-06, |
|
"loss": 0.8854, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 1.7504950495049505, |
|
"grad_norm": 0.38879198281346644, |
|
"learning_rate": 8.80168935405117e-06, |
|
"loss": 0.8258, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 1.7584158415841584, |
|
"grad_norm": 0.3787674450951966, |
|
"learning_rate": 8.71000783469797e-06, |
|
"loss": 0.8577, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 1.7663366336633664, |
|
"grad_norm": 0.4140574481556825, |
|
"learning_rate": 8.618436450481182e-06, |
|
"loss": 0.8717, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 1.7742574257425743, |
|
"grad_norm": 0.3816315351422027, |
|
"learning_rate": 8.526983019453624e-06, |
|
"loss": 0.8328, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 1.7821782178217822, |
|
"grad_norm": 0.3778129363378664, |
|
"learning_rate": 8.43565534959769e-06, |
|
"loss": 0.8625, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 1.7900990099009901, |
|
"grad_norm": 0.39307248999316513, |
|
"learning_rate": 8.3444612381587e-06, |
|
"loss": 0.8472, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 1.798019801980198, |
|
"grad_norm": 0.567116321231214, |
|
"learning_rate": 8.253408470979212e-06, |
|
"loss": 0.8435, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 1.805940594059406, |
|
"grad_norm": 0.3826967696218752, |
|
"learning_rate": 8.162504821834296e-06, |
|
"loss": 0.8287, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 1.8138613861386137, |
|
"grad_norm": 0.4031524752294553, |
|
"learning_rate": 8.071758051767833e-06, |
|
"loss": 0.8372, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 1.8217821782178216, |
|
"grad_norm": 0.38474330971078624, |
|
"learning_rate": 7.9811759084299e-06, |
|
"loss": 0.8722, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 1.8297029702970296, |
|
"grad_norm": 0.3835735801490187, |
|
"learning_rate": 7.890766125415304e-06, |
|
"loss": 0.9066, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 1.8376237623762375, |
|
"grad_norm": 0.4060540049015593, |
|
"learning_rate": 7.800536421603317e-06, |
|
"loss": 0.8252, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 1.8455445544554454, |
|
"grad_norm": 0.386875601730567, |
|
"learning_rate": 7.710494500498662e-06, |
|
"loss": 0.8643, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 1.8534653465346533, |
|
"grad_norm": 0.38667621959177934, |
|
"learning_rate": 7.620648049573815e-06, |
|
"loss": 0.857, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 1.8613861386138613, |
|
"grad_norm": 0.38808086535439623, |
|
"learning_rate": 7.531004739612668e-06, |
|
"loss": 0.846, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 1.8693069306930692, |
|
"grad_norm": 0.3980360267748509, |
|
"learning_rate": 7.441572224055644e-06, |
|
"loss": 0.8394, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 1.8772277227722771, |
|
"grad_norm": 0.38040890589313575, |
|
"learning_rate": 7.352358138346241e-06, |
|
"loss": 0.8321, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 1.885148514851485, |
|
"grad_norm": 0.36846374610523613, |
|
"learning_rate": 7.263370099279173e-06, |
|
"loss": 0.8497, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 1.893069306930693, |
|
"grad_norm": 0.3864744771519634, |
|
"learning_rate": 7.17461570435005e-06, |
|
"loss": 0.8568, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 1.900990099009901, |
|
"grad_norm": 0.4061916523025884, |
|
"learning_rate": 7.086102531106755e-06, |
|
"loss": 0.9014, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 1.9089108910891088, |
|
"grad_norm": 0.4041879159841448, |
|
"learning_rate": 6.997838136502474e-06, |
|
"loss": 0.8792, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 1.9168316831683168, |
|
"grad_norm": 0.42210348978875756, |
|
"learning_rate": 6.909830056250527e-06, |
|
"loss": 0.8597, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 1.9247524752475247, |
|
"grad_norm": 0.3878637842247315, |
|
"learning_rate": 6.822085804180985e-06, |
|
"loss": 0.889, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 1.9326732673267326, |
|
"grad_norm": 0.3903220656513824, |
|
"learning_rate": 6.734612871599169e-06, |
|
"loss": 0.8518, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 1.9405940594059405, |
|
"grad_norm": 0.38140916668736696, |
|
"learning_rate": 6.647418726646065e-06, |
|
"loss": 0.8664, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 1.9485148514851485, |
|
"grad_norm": 0.37076093981349967, |
|
"learning_rate": 6.560510813660719e-06, |
|
"loss": 0.8115, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 1.9564356435643564, |
|
"grad_norm": 0.38782683714399413, |
|
"learning_rate": 6.473896552544674e-06, |
|
"loss": 0.8616, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 1.9643564356435643, |
|
"grad_norm": 0.3806774040873917, |
|
"learning_rate": 6.387583338128471e-06, |
|
"loss": 0.8547, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 1.9722772277227723, |
|
"grad_norm": 0.39269059714326215, |
|
"learning_rate": 6.30157853954031e-06, |
|
"loss": 0.8466, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 1.9801980198019802, |
|
"grad_norm": 0.3887611447725159, |
|
"learning_rate": 6.215889499576898e-06, |
|
"loss": 0.8347, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 1.988118811881188, |
|
"grad_norm": 0.3776184815702374, |
|
"learning_rate": 6.130523534076549e-06, |
|
"loss": 0.8313, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 1.996039603960396, |
|
"grad_norm": 0.39378215320119175, |
|
"learning_rate": 6.0454879312945755e-06, |
|
"loss": 0.869, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 1.996039603960396, |
|
"eval_loss": 1.002571702003479, |
|
"eval_runtime": 88.5175, |
|
"eval_samples_per_second": 3.005, |
|
"eval_steps_per_second": 0.192, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 2.003960396039604, |
|
"grad_norm": 0.576420587207012, |
|
"learning_rate": 5.960789951281052e-06, |
|
"loss": 0.7911, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 2.011881188118812, |
|
"grad_norm": 0.781752019713, |
|
"learning_rate": 5.876436825260967e-06, |
|
"loss": 0.7259, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 2.01980198019802, |
|
"grad_norm": 0.6274336393038759, |
|
"learning_rate": 5.7924357550168534e-06, |
|
"loss": 0.7553, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 2.0277227722772277, |
|
"grad_norm": 0.638695579314472, |
|
"learning_rate": 5.708793912273911e-06, |
|
"loss": 0.7508, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 2.0356435643564357, |
|
"grad_norm": 0.708507864502287, |
|
"learning_rate": 5.625518438087738e-06, |
|
"loss": 0.7477, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 2.0435643564356436, |
|
"grad_norm": 0.5492723143519291, |
|
"learning_rate": 5.542616442234618e-06, |
|
"loss": 0.7429, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 2.0514851485148515, |
|
"grad_norm": 0.6026775840172064, |
|
"learning_rate": 5.460095002604533e-06, |
|
"loss": 0.7175, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 2.0594059405940595, |
|
"grad_norm": 0.6699241593183407, |
|
"learning_rate": 5.3779611645968696e-06, |
|
"loss": 0.7093, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 2.0673267326732674, |
|
"grad_norm": 0.5343067117420846, |
|
"learning_rate": 5.296221940518908e-06, |
|
"loss": 0.7232, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 2.0752475247524753, |
|
"grad_norm": 0.4911057371956158, |
|
"learning_rate": 5.214884308987136e-06, |
|
"loss": 0.7802, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 2.0831683168316832, |
|
"grad_norm": 0.5433218831841199, |
|
"learning_rate": 5.133955214331439e-06, |
|
"loss": 0.7053, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 2.091089108910891, |
|
"grad_norm": 0.5400653655431835, |
|
"learning_rate": 5.053441566002214e-06, |
|
"loss": 0.7098, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 2.099009900990099, |
|
"grad_norm": 0.4898471275769769, |
|
"learning_rate": 4.973350237980466e-06, |
|
"loss": 0.739, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 2.106930693069307, |
|
"grad_norm": 0.4939480668131728, |
|
"learning_rate": 4.893688068190933e-06, |
|
"loss": 0.7255, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 2.114851485148515, |
|
"grad_norm": 0.5066963493135274, |
|
"learning_rate": 4.814461857918279e-06, |
|
"loss": 0.7502, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 2.122772277227723, |
|
"grad_norm": 0.5064134093537243, |
|
"learning_rate": 4.7356783712264405e-06, |
|
"loss": 0.7113, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 2.130693069306931, |
|
"grad_norm": 0.4224734666377468, |
|
"learning_rate": 4.657344334381116e-06, |
|
"loss": 0.7215, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 2.1386138613861387, |
|
"grad_norm": 0.459279750043329, |
|
"learning_rate": 4.579466435275506e-06, |
|
"loss": 0.7407, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 2.1465346534653467, |
|
"grad_norm": 0.46742630810280705, |
|
"learning_rate": 4.5020513228593275e-06, |
|
"loss": 0.6973, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 2.1544554455445546, |
|
"grad_norm": 0.4943363620760823, |
|
"learning_rate": 4.425105606571145e-06, |
|
"loss": 0.7338, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 2.1623762376237625, |
|
"grad_norm": 0.433433891249229, |
|
"learning_rate": 4.348635855774082e-06, |
|
"loss": 0.721, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 2.1702970297029704, |
|
"grad_norm": 0.44511761437854985, |
|
"learning_rate": 4.272648599194948e-06, |
|
"loss": 0.7183, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 2.1782178217821784, |
|
"grad_norm": 0.4758876220892502, |
|
"learning_rate": 4.197150324366844e-06, |
|
"loss": 0.6928, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 2.1861386138613863, |
|
"grad_norm": 0.49467814256335324, |
|
"learning_rate": 4.12214747707527e-06, |
|
"loss": 0.7151, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 2.1940594059405942, |
|
"grad_norm": 0.4332527533958727, |
|
"learning_rate": 4.047646460807814e-06, |
|
"loss": 0.7231, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 2.201980198019802, |
|
"grad_norm": 0.4341696748234564, |
|
"learning_rate": 3.973653636207437e-06, |
|
"loss": 0.7161, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 2.20990099009901, |
|
"grad_norm": 0.4892958618243702, |
|
"learning_rate": 3.9001753205294335e-06, |
|
"loss": 0.7385, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 2.217821782178218, |
|
"grad_norm": 0.42779412099186476, |
|
"learning_rate": 3.827217787102072e-06, |
|
"loss": 0.6949, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 2.225742574257426, |
|
"grad_norm": 0.43010675128621945, |
|
"learning_rate": 3.754787264791011e-06, |
|
"loss": 0.7279, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 2.233663366336634, |
|
"grad_norm": 0.4189623315492079, |
|
"learning_rate": 3.6828899374674933e-06, |
|
"loss": 0.7151, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 2.241584158415842, |
|
"grad_norm": 0.4283610207819412, |
|
"learning_rate": 3.6115319434803897e-06, |
|
"loss": 0.767, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 2.2495049504950497, |
|
"grad_norm": 0.4344389534943353, |
|
"learning_rate": 3.540719375132129e-06, |
|
"loss": 0.7271, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 2.2574257425742577, |
|
"grad_norm": 0.4361317858889359, |
|
"learning_rate": 3.4704582781585596e-06, |
|
"loss": 0.7248, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 2.2653465346534656, |
|
"grad_norm": 0.43728115762338654, |
|
"learning_rate": 3.4007546512127764e-06, |
|
"loss": 0.7231, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 2.2732673267326735, |
|
"grad_norm": 0.4264346314477906, |
|
"learning_rate": 3.3316144453529897e-06, |
|
"loss": 0.7265, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 2.2811881188118814, |
|
"grad_norm": 0.4328763606745891, |
|
"learning_rate": 3.2630435635344283e-06, |
|
"loss": 0.726, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 2.289108910891089, |
|
"grad_norm": 0.40543132963978806, |
|
"learning_rate": 3.1950478601053847e-06, |
|
"loss": 0.7226, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 2.297029702970297, |
|
"grad_norm": 0.40808543584405227, |
|
"learning_rate": 3.1276331403073733e-06, |
|
"loss": 0.6789, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 2.3049504950495048, |
|
"grad_norm": 0.40839769173881413, |
|
"learning_rate": 3.0608051597795043e-06, |
|
"loss": 0.7187, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 2.3128712871287127, |
|
"grad_norm": 0.4369573292307265, |
|
"learning_rate": 2.9945696240670905e-06, |
|
"loss": 0.7301, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 2.3207920792079206, |
|
"grad_norm": 0.40752917080947654, |
|
"learning_rate": 2.9289321881345257e-06, |
|
"loss": 0.7371, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 2.3287128712871286, |
|
"grad_norm": 0.40681720921863107, |
|
"learning_rate": 2.8638984558824777e-06, |
|
"loss": 0.7053, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 2.3366336633663365, |
|
"grad_norm": 0.39704090696075117, |
|
"learning_rate": 2.799473979669456e-06, |
|
"loss": 0.7222, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 2.3445544554455444, |
|
"grad_norm": 0.4127972906314945, |
|
"learning_rate": 2.7356642598377604e-06, |
|
"loss": 0.7227, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 2.3524752475247523, |
|
"grad_norm": 0.40247162704682216, |
|
"learning_rate": 2.672474744243888e-06, |
|
"loss": 0.7178, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 2.3603960396039603, |
|
"grad_norm": 0.4070193008615235, |
|
"learning_rate": 2.6099108277934105e-06, |
|
"loss": 0.7164, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 2.368316831683168, |
|
"grad_norm": 0.3958026290470292, |
|
"learning_rate": 2.547977851980373e-06, |
|
"loss": 0.7226, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 2.376237623762376, |
|
"grad_norm": 0.414324012823234, |
|
"learning_rate": 2.4866811044312667e-06, |
|
"loss": 0.722, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 2.384158415841584, |
|
"grad_norm": 0.3924655139400483, |
|
"learning_rate": 2.426025818453572e-06, |
|
"loss": 0.6922, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 2.392079207920792, |
|
"grad_norm": 0.39324797222146385, |
|
"learning_rate": 2.3660171725889703e-06, |
|
"loss": 0.7168, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"grad_norm": 0.4018290113845056, |
|
"learning_rate": 2.306660290171211e-06, |
|
"loss": 0.7058, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 2.407920792079208, |
|
"grad_norm": 0.4186975177176807, |
|
"learning_rate": 2.2479602388887013e-06, |
|
"loss": 0.7308, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 2.4158415841584158, |
|
"grad_norm": 0.41697995471077964, |
|
"learning_rate": 2.1899220303518465e-06, |
|
"loss": 0.7428, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 2.4237623762376237, |
|
"grad_norm": 0.4024605920230618, |
|
"learning_rate": 2.132550619665168e-06, |
|
"loss": 0.6937, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 2.4316831683168316, |
|
"grad_norm": 0.415257008654532, |
|
"learning_rate": 2.075850905004262e-06, |
|
"loss": 0.7351, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 2.4396039603960396, |
|
"grad_norm": 0.40407977073992135, |
|
"learning_rate": 2.019827727197605e-06, |
|
"loss": 0.7028, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 2.4475247524752475, |
|
"grad_norm": 0.3935541096805544, |
|
"learning_rate": 1.9644858693132627e-06, |
|
"loss": 0.7073, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 2.4554455445544554, |
|
"grad_norm": 0.40998948849676764, |
|
"learning_rate": 1.9098300562505266e-06, |
|
"loss": 0.7084, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 2.4633663366336633, |
|
"grad_norm": 0.4064763838316956, |
|
"learning_rate": 1.855864954336517e-06, |
|
"loss": 0.7541, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 2.4712871287128713, |
|
"grad_norm": 0.40702963715268836, |
|
"learning_rate": 1.80259517092779e-06, |
|
"loss": 0.7118, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 2.479207920792079, |
|
"grad_norm": 0.40962380739362325, |
|
"learning_rate": 1.7500252540169782e-06, |
|
"loss": 0.7149, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 2.487128712871287, |
|
"grad_norm": 0.4225405507530112, |
|
"learning_rate": 1.6981596918444953e-06, |
|
"loss": 0.7097, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 2.495049504950495, |
|
"grad_norm": 0.3974259872963068, |
|
"learning_rate": 1.6470029125153463e-06, |
|
"loss": 0.7468, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 2.502970297029703, |
|
"grad_norm": 0.4075694075831413, |
|
"learning_rate": 1.596559283621074e-06, |
|
"loss": 0.7223, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 2.510891089108911, |
|
"grad_norm": 0.4032028821038701, |
|
"learning_rate": 1.5468331118668655e-06, |
|
"loss": 0.7233, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 2.518811881188119, |
|
"grad_norm": 0.3984358554986335, |
|
"learning_rate": 1.4978286427038602e-06, |
|
"loss": 0.7163, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 2.5267326732673268, |
|
"grad_norm": 0.39826123904027727, |
|
"learning_rate": 1.44955005996669e-06, |
|
"loss": 0.7174, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 2.5346534653465347, |
|
"grad_norm": 0.40725326569026266, |
|
"learning_rate": 1.4020014855162755e-06, |
|
"loss": 0.7112, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 2.5425742574257426, |
|
"grad_norm": 0.39814818108116296, |
|
"learning_rate": 1.3551869788879213e-06, |
|
"loss": 0.7301, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 2.5504950495049505, |
|
"grad_norm": 0.39471994779707076, |
|
"learning_rate": 1.3091105369447166e-06, |
|
"loss": 0.7092, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 2.5584158415841585, |
|
"grad_norm": 0.4165929129276557, |
|
"learning_rate": 1.2637760935363053e-06, |
|
"loss": 0.7131, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 2.5663366336633664, |
|
"grad_norm": 0.40067988266060006, |
|
"learning_rate": 1.2191875191630209e-06, |
|
"loss": 0.7242, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 2.5742574257425743, |
|
"grad_norm": 0.4150381954280917, |
|
"learning_rate": 1.1753486206454433e-06, |
|
"loss": 0.7402, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 2.5821782178217823, |
|
"grad_norm": 0.3920384804019307, |
|
"learning_rate": 1.132263140799381e-06, |
|
"loss": 0.7105, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 2.59009900990099, |
|
"grad_norm": 0.3987176952287612, |
|
"learning_rate": 1.0899347581163222e-06, |
|
"loss": 0.7144, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 2.598019801980198, |
|
"grad_norm": 0.41078279785040733, |
|
"learning_rate": 1.0483670864493777e-06, |
|
"loss": 0.7249, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 2.605940594059406, |
|
"grad_norm": 0.4040238292412966, |
|
"learning_rate": 1.0075636747047446e-06, |
|
"loss": 0.7272, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 2.613861386138614, |
|
"grad_norm": 0.43047282777392387, |
|
"learning_rate": 9.675280065387117e-07, |
|
"loss": 0.7433, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 2.621782178217822, |
|
"grad_norm": 0.3922293398745348, |
|
"learning_rate": 9.282635000602346e-07, |
|
"loss": 0.7071, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 2.62970297029703, |
|
"grad_norm": 0.41685890998877895, |
|
"learning_rate": 8.897735075391156e-07, |
|
"loss": 0.7435, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 2.6376237623762377, |
|
"grad_norm": 0.4060427125681242, |
|
"learning_rate": 8.520613151197899e-07, |
|
"loss": 0.7065, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 2.6455445544554457, |
|
"grad_norm": 0.3975045499192803, |
|
"learning_rate": 8.151301425407699e-07, |
|
"loss": 0.7324, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 2.6534653465346536, |
|
"grad_norm": 0.39196801398755104, |
|
"learning_rate": 7.78983142859755e-07, |
|
"loss": 0.714, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 2.6613861386138615, |
|
"grad_norm": 0.3956193577383561, |
|
"learning_rate": 7.43623402184438e-07, |
|
"loss": 0.7276, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 2.6693069306930695, |
|
"grad_norm": 0.3990830893440769, |
|
"learning_rate": 7.090539394090135e-07, |
|
"loss": 0.7247, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 2.6772277227722774, |
|
"grad_norm": 0.3930307996524001, |
|
"learning_rate": 6.752777059564431e-07, |
|
"loss": 0.6925, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 2.6851485148514853, |
|
"grad_norm": 0.41056613538755815, |
|
"learning_rate": 6.422975855264757e-07, |
|
"loss": 0.7119, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 2.693069306930693, |
|
"grad_norm": 0.5308864436453282, |
|
"learning_rate": 6.101163938494359e-07, |
|
"loss": 0.6739, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 2.7009900990099007, |
|
"grad_norm": 0.3970396905428287, |
|
"learning_rate": 5.78736878445837e-07, |
|
"loss": 0.6944, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 2.7089108910891087, |
|
"grad_norm": 0.40116648564130203, |
|
"learning_rate": 5.481617183918053e-07, |
|
"loss": 0.6752, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 2.7168316831683166, |
|
"grad_norm": 0.4288212628472651, |
|
"learning_rate": 5.183935240903415e-07, |
|
"loss": 0.7268, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 2.7247524752475245, |
|
"grad_norm": 0.4018223127502873, |
|
"learning_rate": 4.894348370484648e-07, |
|
"loss": 0.7069, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 2.7326732673267324, |
|
"grad_norm": 0.45512141312553184, |
|
"learning_rate": 4.6128812966021894e-07, |
|
"loss": 0.7236, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 2.7405940594059404, |
|
"grad_norm": 0.40647740230337975, |
|
"learning_rate": 4.3395580499559276e-07, |
|
"loss": 0.7056, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 2.7485148514851483, |
|
"grad_norm": 0.7851062640103578, |
|
"learning_rate": 4.0744019659535116e-07, |
|
"loss": 0.7191, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 2.756435643564356, |
|
"grad_norm": 0.3918292935567612, |
|
"learning_rate": 3.817435682718096e-07, |
|
"loss": 0.7331, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 2.764356435643564, |
|
"grad_norm": 0.40947897586494714, |
|
"learning_rate": 3.5686811391555164e-07, |
|
"loss": 0.6874, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 2.772277227722772, |
|
"grad_norm": 0.3980289846234701, |
|
"learning_rate": 3.328159573081258e-07, |
|
"loss": 0.7299, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 2.78019801980198, |
|
"grad_norm": 0.3883876533346389, |
|
"learning_rate": 3.095891519407246e-07, |
|
"loss": 0.705, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 2.788118811881188, |
|
"grad_norm": 0.4021695381629565, |
|
"learning_rate": 2.871896808388608e-07, |
|
"loss": 0.7105, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 2.796039603960396, |
|
"grad_norm": 0.40524014008780573, |
|
"learning_rate": 2.656194563930714e-07, |
|
"loss": 0.6829, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 2.803960396039604, |
|
"grad_norm": 0.400098347999415, |
|
"learning_rate": 2.44880320195634e-07, |
|
"loss": 0.7134, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 2.8118811881188117, |
|
"grad_norm": 0.39959879213796157, |
|
"learning_rate": 2.2497404288334245e-07, |
|
"loss": 0.7349, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 2.8198019801980196, |
|
"grad_norm": 0.404829435915328, |
|
"learning_rate": 2.0590232398634114e-07, |
|
"loss": 0.7224, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 2.8277227722772276, |
|
"grad_norm": 0.3988593166231128, |
|
"learning_rate": 1.876667917830155e-07, |
|
"loss": 0.7137, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 2.8356435643564355, |
|
"grad_norm": 0.39999015184518205, |
|
"learning_rate": 1.7026900316098217e-07, |
|
"loss": 0.7214, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 2.8435643564356434, |
|
"grad_norm": 0.40532384840797636, |
|
"learning_rate": 1.537104434841641e-07, |
|
"loss": 0.7403, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 2.8514851485148514, |
|
"grad_norm": 0.4072129955656424, |
|
"learning_rate": 1.3799252646597428e-07, |
|
"loss": 0.7525, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 2.8594059405940593, |
|
"grad_norm": 0.4081067652201019, |
|
"learning_rate": 1.231165940486234e-07, |
|
"loss": 0.7119, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 2.867326732673267, |
|
"grad_norm": 0.40127768445127304, |
|
"learning_rate": 1.0908391628854042e-07, |
|
"loss": 0.7068, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 2.875247524752475, |
|
"grad_norm": 0.4056569585309908, |
|
"learning_rate": 9.589569124794918e-08, |
|
"loss": 0.7327, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 2.883168316831683, |
|
"grad_norm": 0.39394071807420317, |
|
"learning_rate": 8.355304489257254e-08, |
|
"loss": 0.7217, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 2.891089108910891, |
|
"grad_norm": 0.3987850556623756, |
|
"learning_rate": 7.205703099551042e-08, |
|
"loss": 0.696, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 2.899009900990099, |
|
"grad_norm": 0.40217224086214826, |
|
"learning_rate": 6.140863104726391e-08, |
|
"loss": 0.7303, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 2.906930693069307, |
|
"grad_norm": 0.40009663993330546, |
|
"learning_rate": 5.1608754171944555e-08, |
|
"loss": 0.7105, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 2.9148514851485148, |
|
"grad_norm": 0.3967938165718884, |
|
"learning_rate": 4.2658237049655325e-08, |
|
"loss": 0.7334, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 2.9227722772277227, |
|
"grad_norm": 0.4086603962701755, |
|
"learning_rate": 3.455784384505445e-08, |
|
"loss": 0.6938, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 2.9306930693069306, |
|
"grad_norm": 0.3899469279832273, |
|
"learning_rate": 2.7308266142119788e-08, |
|
"loss": 0.7172, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 2.9386138613861386, |
|
"grad_norm": 0.4126769814687196, |
|
"learning_rate": 2.0910122885097194e-08, |
|
"loss": 0.7112, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 2.9465346534653465, |
|
"grad_norm": 1.0101608008566858, |
|
"learning_rate": 1.5363960325660565e-08, |
|
"loss": 0.7177, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 2.9544554455445544, |
|
"grad_norm": 0.422375267138987, |
|
"learning_rate": 1.0670251976275803e-08, |
|
"loss": 0.7275, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 2.9623762376237623, |
|
"grad_norm": 0.3978598838942699, |
|
"learning_rate": 6.82939856977094e-09, |
|
"loss": 0.7221, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 2.9702970297029703, |
|
"grad_norm": 0.4100316902452377, |
|
"learning_rate": 3.8417280251257235e-09, |
|
"loss": 0.7309, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 2.978217821782178, |
|
"grad_norm": 0.39841317020419215, |
|
"learning_rate": 1.7074954194729044e-09, |
|
"loss": 0.7361, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 2.986138613861386, |
|
"grad_norm": 0.39205883398967273, |
|
"learning_rate": 4.2688296632120705e-10, |
|
"loss": 0.7139, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 2.994059405940594, |
|
"grad_norm": 0.3914037221232023, |
|
"learning_rate": 0.0, |
|
"loss": 0.714, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 2.994059405940594, |
|
"eval_loss": 1.0379831790924072, |
|
"eval_runtime": 89.0288, |
|
"eval_samples_per_second": 2.988, |
|
"eval_steps_per_second": 0.191, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 2.994059405940594, |
|
"step": 378, |
|
"total_flos": 158238674780160.0, |
|
"train_loss": 0.8970171990533354, |
|
"train_runtime": 67109.8008, |
|
"train_samples_per_second": 0.722, |
|
"train_steps_per_second": 0.006 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 378, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 158238674780160.0, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|