{ "best_metric": null, "best_model_checkpoint": null, "epoch": 9.997593261131167, "eval_steps": 500, "global_step": 20770, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.002888086642599278, "grad_norm": 1.6207919120788574, "learning_rate": 9.997111218103033e-05, "loss": 2.3195, "step": 6 }, { "epoch": 0.005776173285198556, "grad_norm": 1.1273815631866455, "learning_rate": 9.994222436206067e-05, "loss": 2.2944, "step": 12 }, { "epoch": 0.008664259927797834, "grad_norm": 1.0821865797042847, "learning_rate": 9.991333654309099e-05, "loss": 2.1311, "step": 18 }, { "epoch": 0.011552346570397111, "grad_norm": 1.6518632173538208, "learning_rate": 9.988444872412133e-05, "loss": 2.1438, "step": 24 }, { "epoch": 0.01444043321299639, "grad_norm": 1.0577433109283447, "learning_rate": 9.985556090515167e-05, "loss": 2.0555, "step": 30 }, { "epoch": 0.017328519855595668, "grad_norm": 0.9200044274330139, "learning_rate": 9.9826673086182e-05, "loss": 2.0851, "step": 36 }, { "epoch": 0.020216606498194945, "grad_norm": 0.8418582677841187, "learning_rate": 9.979778526721233e-05, "loss": 2.1405, "step": 42 }, { "epoch": 0.023104693140794223, "grad_norm": 0.8012077808380127, "learning_rate": 9.976889744824266e-05, "loss": 2.0837, "step": 48 }, { "epoch": 0.025992779783393503, "grad_norm": 1.0480042695999146, "learning_rate": 9.9740009629273e-05, "loss": 2.2047, "step": 54 }, { "epoch": 0.02888086642599278, "grad_norm": 0.7061931490898132, "learning_rate": 9.971112181030333e-05, "loss": 2.1014, "step": 60 }, { "epoch": 0.03176895306859206, "grad_norm": 1.2098735570907593, "learning_rate": 9.968223399133366e-05, "loss": 1.9964, "step": 66 }, { "epoch": 0.034657039711191336, "grad_norm": 0.793174684047699, "learning_rate": 9.9653346172364e-05, "loss": 2.1728, "step": 72 }, { "epoch": 0.03754512635379061, "grad_norm": 0.7847986221313477, "learning_rate": 9.962445835339432e-05, "loss": 2.0764, "step": 78 }, { "epoch": 0.04043321299638989, "grad_norm": 0.8436242938041687, "learning_rate": 9.959557053442465e-05, "loss": 2.0054, "step": 84 }, { "epoch": 0.04332129963898917, "grad_norm": 0.8447696566581726, "learning_rate": 9.956668271545499e-05, "loss": 2.0105, "step": 90 }, { "epoch": 0.046209386281588445, "grad_norm": 0.9139009714126587, "learning_rate": 9.953779489648531e-05, "loss": 1.9058, "step": 96 }, { "epoch": 0.04909747292418772, "grad_norm": 0.8787450790405273, "learning_rate": 9.950890707751566e-05, "loss": 2.0143, "step": 102 }, { "epoch": 0.05198555956678701, "grad_norm": 0.7764399647712708, "learning_rate": 9.948001925854599e-05, "loss": 2.0461, "step": 108 }, { "epoch": 0.054873646209386284, "grad_norm": 0.9490439295768738, "learning_rate": 9.945113143957631e-05, "loss": 2.0075, "step": 114 }, { "epoch": 0.05776173285198556, "grad_norm": 0.9537297487258911, "learning_rate": 9.942224362060665e-05, "loss": 1.9721, "step": 120 }, { "epoch": 0.06064981949458484, "grad_norm": 1.0197209119796753, "learning_rate": 9.939335580163698e-05, "loss": 1.9598, "step": 126 }, { "epoch": 0.06353790613718412, "grad_norm": 0.8615917563438416, "learning_rate": 9.936446798266731e-05, "loss": 1.9034, "step": 132 }, { "epoch": 0.0664259927797834, "grad_norm": 0.8871897459030151, "learning_rate": 9.933558016369765e-05, "loss": 1.8403, "step": 138 }, { "epoch": 0.06931407942238267, "grad_norm": 0.8798001408576965, "learning_rate": 9.930669234472798e-05, "loss": 2.138, "step": 144 }, { "epoch": 0.07220216606498195, "grad_norm": 1.137473702430725, "learning_rate": 9.927780452575832e-05, "loss": 2.1137, "step": 150 }, { "epoch": 0.07509025270758123, "grad_norm": 0.9626398682594299, "learning_rate": 9.924891670678864e-05, "loss": 2.02, "step": 156 }, { "epoch": 0.0779783393501805, "grad_norm": 0.9745214581489563, "learning_rate": 9.922002888781898e-05, "loss": 2.0703, "step": 162 }, { "epoch": 0.08086642599277978, "grad_norm": 1.4741218090057373, "learning_rate": 9.91911410688493e-05, "loss": 1.9356, "step": 168 }, { "epoch": 0.08375451263537906, "grad_norm": 0.9096198081970215, "learning_rate": 9.916225324987964e-05, "loss": 2.0368, "step": 174 }, { "epoch": 0.08664259927797834, "grad_norm": 0.8218715786933899, "learning_rate": 9.913336543090998e-05, "loss": 1.9897, "step": 180 }, { "epoch": 0.08953068592057761, "grad_norm": 0.829740047454834, "learning_rate": 9.91044776119403e-05, "loss": 1.995, "step": 186 }, { "epoch": 0.09241877256317689, "grad_norm": 0.9626950025558472, "learning_rate": 9.907558979297063e-05, "loss": 2.067, "step": 192 }, { "epoch": 0.09530685920577617, "grad_norm": 0.8067222833633423, "learning_rate": 9.904670197400097e-05, "loss": 2.0922, "step": 198 }, { "epoch": 0.09819494584837545, "grad_norm": 0.9139990210533142, "learning_rate": 9.901781415503129e-05, "loss": 2.0073, "step": 204 }, { "epoch": 0.10108303249097472, "grad_norm": 0.9457979798316956, "learning_rate": 9.898892633606163e-05, "loss": 2.0179, "step": 210 }, { "epoch": 0.10397111913357401, "grad_norm": 1.036190152168274, "learning_rate": 9.896003851709197e-05, "loss": 2.0313, "step": 216 }, { "epoch": 0.10685920577617329, "grad_norm": 0.8998299241065979, "learning_rate": 9.89311506981223e-05, "loss": 2.1345, "step": 222 }, { "epoch": 0.10974729241877257, "grad_norm": 0.8423928022384644, "learning_rate": 9.890226287915263e-05, "loss": 2.0542, "step": 228 }, { "epoch": 0.11263537906137185, "grad_norm": 1.0024595260620117, "learning_rate": 9.887337506018296e-05, "loss": 1.8846, "step": 234 }, { "epoch": 0.11552346570397112, "grad_norm": 0.8751068115234375, "learning_rate": 9.88444872412133e-05, "loss": 2.0634, "step": 240 }, { "epoch": 0.1184115523465704, "grad_norm": 0.9097185730934143, "learning_rate": 9.881559942224362e-05, "loss": 2.0407, "step": 246 }, { "epoch": 0.12129963898916968, "grad_norm": 0.8266995549201965, "learning_rate": 9.878671160327396e-05, "loss": 1.8854, "step": 252 }, { "epoch": 0.12418772563176896, "grad_norm": 0.8577411770820618, "learning_rate": 9.87578237843043e-05, "loss": 2.0752, "step": 258 }, { "epoch": 0.12707581227436823, "grad_norm": 0.8443561792373657, "learning_rate": 9.872893596533462e-05, "loss": 1.9415, "step": 264 }, { "epoch": 0.1299638989169675, "grad_norm": 0.8937883377075195, "learning_rate": 9.870004814636496e-05, "loss": 1.9728, "step": 270 }, { "epoch": 0.1328519855595668, "grad_norm": 0.8667436838150024, "learning_rate": 9.867116032739529e-05, "loss": 1.8851, "step": 276 }, { "epoch": 0.13574007220216608, "grad_norm": 0.9499288201332092, "learning_rate": 9.864227250842561e-05, "loss": 1.9715, "step": 282 }, { "epoch": 0.13862815884476534, "grad_norm": 0.8656870126724243, "learning_rate": 9.861338468945595e-05, "loss": 1.9229, "step": 288 }, { "epoch": 0.14151624548736463, "grad_norm": 1.013107419013977, "learning_rate": 9.858449687048629e-05, "loss": 1.9925, "step": 294 }, { "epoch": 0.1444043321299639, "grad_norm": 0.8389509916305542, "learning_rate": 9.855560905151661e-05, "loss": 2.1104, "step": 300 }, { "epoch": 0.1472924187725632, "grad_norm": 0.9546544551849365, "learning_rate": 9.852672123254695e-05, "loss": 1.9642, "step": 306 }, { "epoch": 0.15018050541516245, "grad_norm": 0.8441974520683289, "learning_rate": 9.849783341357727e-05, "loss": 1.7983, "step": 312 }, { "epoch": 0.15306859205776174, "grad_norm": 0.9391993284225464, "learning_rate": 9.846894559460761e-05, "loss": 1.8326, "step": 318 }, { "epoch": 0.155956678700361, "grad_norm": 0.9327995777130127, "learning_rate": 9.844005777563794e-05, "loss": 1.8797, "step": 324 }, { "epoch": 0.1588447653429603, "grad_norm": 0.8677220344543457, "learning_rate": 9.841116995666828e-05, "loss": 1.9258, "step": 330 }, { "epoch": 0.16173285198555956, "grad_norm": 0.9613078236579895, "learning_rate": 9.838228213769861e-05, "loss": 2.102, "step": 336 }, { "epoch": 0.16462093862815885, "grad_norm": 0.7909504771232605, "learning_rate": 9.835339431872894e-05, "loss": 2.0768, "step": 342 }, { "epoch": 0.16750902527075812, "grad_norm": 0.9365721344947815, "learning_rate": 9.832450649975928e-05, "loss": 1.9474, "step": 348 }, { "epoch": 0.1703971119133574, "grad_norm": 0.7811998724937439, "learning_rate": 9.82956186807896e-05, "loss": 2.0919, "step": 354 }, { "epoch": 0.17328519855595667, "grad_norm": 0.8282377123832703, "learning_rate": 9.826673086181993e-05, "loss": 1.9231, "step": 360 }, { "epoch": 0.17617328519855596, "grad_norm": 0.9420356750488281, "learning_rate": 9.823784304285028e-05, "loss": 2.111, "step": 366 }, { "epoch": 0.17906137184115523, "grad_norm": 0.9016260504722595, "learning_rate": 9.82089552238806e-05, "loss": 1.8339, "step": 372 }, { "epoch": 0.18194945848375452, "grad_norm": 0.7841916084289551, "learning_rate": 9.818006740491093e-05, "loss": 1.7858, "step": 378 }, { "epoch": 0.18483754512635378, "grad_norm": 0.89506995677948, "learning_rate": 9.815117958594127e-05, "loss": 1.8919, "step": 384 }, { "epoch": 0.18772563176895307, "grad_norm": 0.7988371253013611, "learning_rate": 9.812229176697159e-05, "loss": 1.9171, "step": 390 }, { "epoch": 0.19061371841155234, "grad_norm": 0.9598416686058044, "learning_rate": 9.809340394800193e-05, "loss": 1.9174, "step": 396 }, { "epoch": 0.19350180505415163, "grad_norm": 0.7619362473487854, "learning_rate": 9.806451612903226e-05, "loss": 2.1082, "step": 402 }, { "epoch": 0.1963898916967509, "grad_norm": 0.8834238052368164, "learning_rate": 9.80356283100626e-05, "loss": 1.9402, "step": 408 }, { "epoch": 0.19927797833935018, "grad_norm": 0.874588131904602, "learning_rate": 9.800674049109293e-05, "loss": 1.9408, "step": 414 }, { "epoch": 0.20216606498194944, "grad_norm": 0.847490668296814, "learning_rate": 9.797785267212326e-05, "loss": 1.9609, "step": 420 }, { "epoch": 0.20505415162454874, "grad_norm": 0.8169350624084473, "learning_rate": 9.79489648531536e-05, "loss": 1.8515, "step": 426 }, { "epoch": 0.20794223826714803, "grad_norm": 0.7894175052642822, "learning_rate": 9.792007703418392e-05, "loss": 1.8431, "step": 432 }, { "epoch": 0.2108303249097473, "grad_norm": 0.9915683269500732, "learning_rate": 9.789118921521425e-05, "loss": 1.8726, "step": 438 }, { "epoch": 0.21371841155234658, "grad_norm": 0.9293184876441956, "learning_rate": 9.78623013962446e-05, "loss": 2.0217, "step": 444 }, { "epoch": 0.21660649819494585, "grad_norm": 0.8520947694778442, "learning_rate": 9.783341357727492e-05, "loss": 1.9865, "step": 450 }, { "epoch": 0.21949458483754514, "grad_norm": 0.8678386211395264, "learning_rate": 9.780452575830526e-05, "loss": 2.0288, "step": 456 }, { "epoch": 0.2223826714801444, "grad_norm": 0.8332152366638184, "learning_rate": 9.777563793933559e-05, "loss": 1.9502, "step": 462 }, { "epoch": 0.2252707581227437, "grad_norm": 0.8556843400001526, "learning_rate": 9.774675012036591e-05, "loss": 1.9655, "step": 468 }, { "epoch": 0.22815884476534296, "grad_norm": 0.7289130091667175, "learning_rate": 9.771786230139625e-05, "loss": 1.8524, "step": 474 }, { "epoch": 0.23104693140794225, "grad_norm": 0.8300471901893616, "learning_rate": 9.768897448242657e-05, "loss": 1.8902, "step": 480 }, { "epoch": 0.2339350180505415, "grad_norm": 0.8754612803459167, "learning_rate": 9.766008666345691e-05, "loss": 1.8918, "step": 486 }, { "epoch": 0.2368231046931408, "grad_norm": 0.9495061635971069, "learning_rate": 9.763119884448725e-05, "loss": 1.94, "step": 492 }, { "epoch": 0.23971119133574006, "grad_norm": 0.865957498550415, "learning_rate": 9.760231102551757e-05, "loss": 1.8889, "step": 498 }, { "epoch": 0.24259927797833936, "grad_norm": 0.9952989816665649, "learning_rate": 9.757342320654791e-05, "loss": 1.932, "step": 504 }, { "epoch": 0.24548736462093862, "grad_norm": 0.8675450682640076, "learning_rate": 9.754453538757824e-05, "loss": 2.0051, "step": 510 }, { "epoch": 0.2483754512635379, "grad_norm": 0.7922016978263855, "learning_rate": 9.751564756860858e-05, "loss": 1.9325, "step": 516 }, { "epoch": 0.2512635379061372, "grad_norm": 0.8986628651618958, "learning_rate": 9.748675974963891e-05, "loss": 1.7155, "step": 522 }, { "epoch": 0.25415162454873647, "grad_norm": 0.9812318086624146, "learning_rate": 9.745787193066924e-05, "loss": 1.9009, "step": 528 }, { "epoch": 0.25703971119133573, "grad_norm": 0.7845908403396606, "learning_rate": 9.742898411169958e-05, "loss": 1.8757, "step": 534 }, { "epoch": 0.259927797833935, "grad_norm": 0.8557993769645691, "learning_rate": 9.74000962927299e-05, "loss": 1.8207, "step": 540 }, { "epoch": 0.2628158844765343, "grad_norm": 0.8217154741287231, "learning_rate": 9.737120847376023e-05, "loss": 1.8544, "step": 546 }, { "epoch": 0.2657039711191336, "grad_norm": 0.9092362523078918, "learning_rate": 9.734232065479057e-05, "loss": 2.0436, "step": 552 }, { "epoch": 0.26859205776173284, "grad_norm": 0.784461259841919, "learning_rate": 9.731343283582089e-05, "loss": 1.7741, "step": 558 }, { "epoch": 0.27148014440433216, "grad_norm": 0.9133092761039734, "learning_rate": 9.728454501685124e-05, "loss": 1.8878, "step": 564 }, { "epoch": 0.2743682310469314, "grad_norm": 1.314786672592163, "learning_rate": 9.725565719788157e-05, "loss": 1.94, "step": 570 }, { "epoch": 0.2772563176895307, "grad_norm": 0.7933446168899536, "learning_rate": 9.722676937891189e-05, "loss": 2.1515, "step": 576 }, { "epoch": 0.28014440433212995, "grad_norm": 0.848456859588623, "learning_rate": 9.719788155994223e-05, "loss": 1.8808, "step": 582 }, { "epoch": 0.28303249097472927, "grad_norm": 0.8681577444076538, "learning_rate": 9.716899374097256e-05, "loss": 1.9584, "step": 588 }, { "epoch": 0.28592057761732853, "grad_norm": 0.9784762263298035, "learning_rate": 9.71401059220029e-05, "loss": 1.9836, "step": 594 }, { "epoch": 0.2888086642599278, "grad_norm": 0.7724260091781616, "learning_rate": 9.711121810303323e-05, "loss": 2.0378, "step": 600 }, { "epoch": 0.29169675090252706, "grad_norm": 0.8997105956077576, "learning_rate": 9.708233028406356e-05, "loss": 2.0086, "step": 606 }, { "epoch": 0.2945848375451264, "grad_norm": 0.8949249982833862, "learning_rate": 9.70534424650939e-05, "loss": 1.9181, "step": 612 }, { "epoch": 0.29747292418772564, "grad_norm": 0.8346735835075378, "learning_rate": 9.702455464612422e-05, "loss": 1.8219, "step": 618 }, { "epoch": 0.3003610108303249, "grad_norm": 0.9193317294120789, "learning_rate": 9.699566682715456e-05, "loss": 2.0467, "step": 624 }, { "epoch": 0.30324909747292417, "grad_norm": 0.9070912003517151, "learning_rate": 9.696677900818488e-05, "loss": 1.7912, "step": 630 }, { "epoch": 0.3061371841155235, "grad_norm": 0.850788950920105, "learning_rate": 9.693789118921522e-05, "loss": 1.9246, "step": 636 }, { "epoch": 0.30902527075812275, "grad_norm": 0.8708522319793701, "learning_rate": 9.690900337024556e-05, "loss": 1.8161, "step": 642 }, { "epoch": 0.311913357400722, "grad_norm": 0.9008936285972595, "learning_rate": 9.688011555127588e-05, "loss": 1.9789, "step": 648 }, { "epoch": 0.3148014440433213, "grad_norm": 0.9246857762336731, "learning_rate": 9.685122773230621e-05, "loss": 2.0147, "step": 654 }, { "epoch": 0.3176895306859206, "grad_norm": 0.854885458946228, "learning_rate": 9.682233991333655e-05, "loss": 1.993, "step": 660 }, { "epoch": 0.32057761732851986, "grad_norm": 0.82230144739151, "learning_rate": 9.679345209436687e-05, "loss": 1.9439, "step": 666 }, { "epoch": 0.3234657039711191, "grad_norm": 0.8475018739700317, "learning_rate": 9.676456427539721e-05, "loss": 1.8485, "step": 672 }, { "epoch": 0.3263537906137184, "grad_norm": 1.3083347082138062, "learning_rate": 9.673567645642755e-05, "loss": 1.8313, "step": 678 }, { "epoch": 0.3292418772563177, "grad_norm": 0.9547069668769836, "learning_rate": 9.670678863745787e-05, "loss": 2.0282, "step": 684 }, { "epoch": 0.33212996389891697, "grad_norm": 0.9360696077346802, "learning_rate": 9.667790081848821e-05, "loss": 1.9996, "step": 690 }, { "epoch": 0.33501805054151623, "grad_norm": 0.9126375913619995, "learning_rate": 9.664901299951854e-05, "loss": 1.9998, "step": 696 }, { "epoch": 0.3379061371841155, "grad_norm": 0.8213159441947937, "learning_rate": 9.662012518054888e-05, "loss": 1.8704, "step": 702 }, { "epoch": 0.3407942238267148, "grad_norm": 0.7915050983428955, "learning_rate": 9.65912373615792e-05, "loss": 1.8156, "step": 708 }, { "epoch": 0.3436823104693141, "grad_norm": 0.8599154353141785, "learning_rate": 9.656234954260954e-05, "loss": 1.9933, "step": 714 }, { "epoch": 0.34657039711191334, "grad_norm": 0.8983302116394043, "learning_rate": 9.653346172363988e-05, "loss": 2.0004, "step": 720 }, { "epoch": 0.34945848375451266, "grad_norm": 0.8538333773612976, "learning_rate": 9.65045739046702e-05, "loss": 1.857, "step": 726 }, { "epoch": 0.3523465703971119, "grad_norm": 1.0243035554885864, "learning_rate": 9.647568608570053e-05, "loss": 2.0338, "step": 732 }, { "epoch": 0.3552346570397112, "grad_norm": 0.8577588200569153, "learning_rate": 9.644679826673087e-05, "loss": 1.8522, "step": 738 }, { "epoch": 0.35812274368231045, "grad_norm": 0.8270348906517029, "learning_rate": 9.641791044776119e-05, "loss": 1.8739, "step": 744 }, { "epoch": 0.36101083032490977, "grad_norm": 0.8861861228942871, "learning_rate": 9.638902262879153e-05, "loss": 1.7514, "step": 750 }, { "epoch": 0.36389891696750903, "grad_norm": 0.937230110168457, "learning_rate": 9.636013480982187e-05, "loss": 2.0671, "step": 756 }, { "epoch": 0.3667870036101083, "grad_norm": 0.8627128601074219, "learning_rate": 9.633124699085219e-05, "loss": 1.8855, "step": 762 }, { "epoch": 0.36967509025270756, "grad_norm": 0.8378495573997498, "learning_rate": 9.630235917188253e-05, "loss": 1.8708, "step": 768 }, { "epoch": 0.3725631768953069, "grad_norm": 0.848907470703125, "learning_rate": 9.627347135291286e-05, "loss": 1.9985, "step": 774 }, { "epoch": 0.37545126353790614, "grad_norm": 0.8259303569793701, "learning_rate": 9.62445835339432e-05, "loss": 1.8486, "step": 780 }, { "epoch": 0.3783393501805054, "grad_norm": 0.9302844405174255, "learning_rate": 9.621569571497352e-05, "loss": 1.8938, "step": 786 }, { "epoch": 0.38122743682310467, "grad_norm": 0.8346471190452576, "learning_rate": 9.618680789600386e-05, "loss": 1.8976, "step": 792 }, { "epoch": 0.384115523465704, "grad_norm": 0.9030067920684814, "learning_rate": 9.61579200770342e-05, "loss": 1.9017, "step": 798 }, { "epoch": 0.38700361010830325, "grad_norm": 0.8201918005943298, "learning_rate": 9.612903225806452e-05, "loss": 1.9068, "step": 804 }, { "epoch": 0.3898916967509025, "grad_norm": 0.9122233986854553, "learning_rate": 9.610014443909486e-05, "loss": 1.8501, "step": 810 }, { "epoch": 0.3927797833935018, "grad_norm": 0.8368625640869141, "learning_rate": 9.607125662012518e-05, "loss": 1.7682, "step": 816 }, { "epoch": 0.3956678700361011, "grad_norm": 0.783745527267456, "learning_rate": 9.604236880115551e-05, "loss": 1.8868, "step": 822 }, { "epoch": 0.39855595667870036, "grad_norm": 0.8005663156509399, "learning_rate": 9.601348098218586e-05, "loss": 1.8014, "step": 828 }, { "epoch": 0.4014440433212996, "grad_norm": 0.9079298973083496, "learning_rate": 9.598459316321618e-05, "loss": 2.0508, "step": 834 }, { "epoch": 0.4043321299638989, "grad_norm": 0.9480348229408264, "learning_rate": 9.595570534424651e-05, "loss": 1.9297, "step": 840 }, { "epoch": 0.4072202166064982, "grad_norm": 0.906222939491272, "learning_rate": 9.592681752527685e-05, "loss": 1.9085, "step": 846 }, { "epoch": 0.41010830324909747, "grad_norm": 0.8765448927879333, "learning_rate": 9.589792970630717e-05, "loss": 1.9402, "step": 852 }, { "epoch": 0.41299638989169674, "grad_norm": 0.92059326171875, "learning_rate": 9.586904188733751e-05, "loss": 1.7604, "step": 858 }, { "epoch": 0.41588447653429605, "grad_norm": 0.9235778450965881, "learning_rate": 9.584015406836784e-05, "loss": 1.8196, "step": 864 }, { "epoch": 0.4187725631768953, "grad_norm": 0.7899588942527771, "learning_rate": 9.581126624939817e-05, "loss": 1.8106, "step": 870 }, { "epoch": 0.4216606498194946, "grad_norm": 0.885440468788147, "learning_rate": 9.578237843042851e-05, "loss": 1.6675, "step": 876 }, { "epoch": 0.42454873646209385, "grad_norm": 0.8909257650375366, "learning_rate": 9.575349061145884e-05, "loss": 2.0604, "step": 882 }, { "epoch": 0.42743682310469316, "grad_norm": 0.986153244972229, "learning_rate": 9.572460279248918e-05, "loss": 1.7694, "step": 888 }, { "epoch": 0.43032490974729243, "grad_norm": 0.8067103028297424, "learning_rate": 9.56957149735195e-05, "loss": 1.9669, "step": 894 }, { "epoch": 0.4332129963898917, "grad_norm": 0.8143800497055054, "learning_rate": 9.566682715454983e-05, "loss": 1.8005, "step": 900 }, { "epoch": 0.43610108303249095, "grad_norm": 0.874967098236084, "learning_rate": 9.563793933558018e-05, "loss": 1.8154, "step": 906 }, { "epoch": 0.4389891696750903, "grad_norm": 0.7974651455879211, "learning_rate": 9.56090515166105e-05, "loss": 1.8649, "step": 912 }, { "epoch": 0.44187725631768954, "grad_norm": 1.027692437171936, "learning_rate": 9.558016369764084e-05, "loss": 1.9057, "step": 918 }, { "epoch": 0.4447653429602888, "grad_norm": 0.8275982737541199, "learning_rate": 9.555127587867117e-05, "loss": 1.9914, "step": 924 }, { "epoch": 0.44765342960288806, "grad_norm": 0.9577564597129822, "learning_rate": 9.552238805970149e-05, "loss": 1.9274, "step": 930 }, { "epoch": 0.4505415162454874, "grad_norm": 0.9420509934425354, "learning_rate": 9.549350024073183e-05, "loss": 1.8138, "step": 936 }, { "epoch": 0.45342960288808665, "grad_norm": 0.8478227257728577, "learning_rate": 9.546461242176215e-05, "loss": 2.0249, "step": 942 }, { "epoch": 0.4563176895306859, "grad_norm": 0.9064255952835083, "learning_rate": 9.543572460279249e-05, "loss": 1.8886, "step": 948 }, { "epoch": 0.4592057761732852, "grad_norm": 0.9979930520057678, "learning_rate": 9.540683678382283e-05, "loss": 1.9625, "step": 954 }, { "epoch": 0.4620938628158845, "grad_norm": 0.7970936894416809, "learning_rate": 9.537794896485315e-05, "loss": 1.8813, "step": 960 }, { "epoch": 0.46498194945848376, "grad_norm": 0.8390293717384338, "learning_rate": 9.53490611458835e-05, "loss": 1.9366, "step": 966 }, { "epoch": 0.467870036101083, "grad_norm": 0.9146532416343689, "learning_rate": 9.532017332691382e-05, "loss": 1.919, "step": 972 }, { "epoch": 0.4707581227436823, "grad_norm": 0.9212022423744202, "learning_rate": 9.529128550794416e-05, "loss": 1.7952, "step": 978 }, { "epoch": 0.4736462093862816, "grad_norm": 0.8925318121910095, "learning_rate": 9.52623976889745e-05, "loss": 1.9377, "step": 984 }, { "epoch": 0.47653429602888087, "grad_norm": 1.0243972539901733, "learning_rate": 9.523350987000482e-05, "loss": 1.8776, "step": 990 }, { "epoch": 0.47942238267148013, "grad_norm": 0.8607951402664185, "learning_rate": 9.520462205103516e-05, "loss": 1.9668, "step": 996 }, { "epoch": 0.48231046931407945, "grad_norm": 0.8962234258651733, "learning_rate": 9.517573423206548e-05, "loss": 1.7967, "step": 1002 }, { "epoch": 0.4851985559566787, "grad_norm": 0.8797476291656494, "learning_rate": 9.514684641309581e-05, "loss": 1.9335, "step": 1008 }, { "epoch": 0.488086642599278, "grad_norm": 0.990982711315155, "learning_rate": 9.511795859412615e-05, "loss": 1.8913, "step": 1014 }, { "epoch": 0.49097472924187724, "grad_norm": 0.9242013096809387, "learning_rate": 9.508907077515648e-05, "loss": 1.9286, "step": 1020 }, { "epoch": 0.49386281588447656, "grad_norm": 0.88871830701828, "learning_rate": 9.506018295618682e-05, "loss": 2.0455, "step": 1026 }, { "epoch": 0.4967509025270758, "grad_norm": 0.8407121896743774, "learning_rate": 9.503129513721715e-05, "loss": 1.8595, "step": 1032 }, { "epoch": 0.4996389891696751, "grad_norm": 1.0398621559143066, "learning_rate": 9.500240731824747e-05, "loss": 1.915, "step": 1038 }, { "epoch": 0.5025270758122744, "grad_norm": 0.9600794315338135, "learning_rate": 9.497351949927781e-05, "loss": 1.9156, "step": 1044 }, { "epoch": 0.5054151624548736, "grad_norm": 0.8989099860191345, "learning_rate": 9.494463168030814e-05, "loss": 2.0227, "step": 1050 }, { "epoch": 0.5083032490974729, "grad_norm": 0.9669433832168579, "learning_rate": 9.491574386133847e-05, "loss": 1.9437, "step": 1056 }, { "epoch": 0.5111913357400723, "grad_norm": 0.9267261624336243, "learning_rate": 9.488685604236881e-05, "loss": 1.9161, "step": 1062 }, { "epoch": 0.5140794223826715, "grad_norm": 0.8845149278640747, "learning_rate": 9.485796822339914e-05, "loss": 1.945, "step": 1068 }, { "epoch": 0.5169675090252708, "grad_norm": 0.9711998105049133, "learning_rate": 9.482908040442948e-05, "loss": 1.8136, "step": 1074 }, { "epoch": 0.51985559566787, "grad_norm": 0.9887317419052124, "learning_rate": 9.48001925854598e-05, "loss": 2.0559, "step": 1080 }, { "epoch": 0.5227436823104693, "grad_norm": 0.9018107652664185, "learning_rate": 9.477130476649014e-05, "loss": 2.0157, "step": 1086 }, { "epoch": 0.5256317689530686, "grad_norm": 0.8001646399497986, "learning_rate": 9.474241694752046e-05, "loss": 1.9274, "step": 1092 }, { "epoch": 0.5285198555956678, "grad_norm": 0.9303708076477051, "learning_rate": 9.47135291285508e-05, "loss": 1.8443, "step": 1098 }, { "epoch": 0.5314079422382672, "grad_norm": 0.8899023532867432, "learning_rate": 9.468464130958114e-05, "loss": 1.87, "step": 1104 }, { "epoch": 0.5342960288808665, "grad_norm": 0.9111254811286926, "learning_rate": 9.465575349061147e-05, "loss": 1.8897, "step": 1110 }, { "epoch": 0.5371841155234657, "grad_norm": 0.9461797475814819, "learning_rate": 9.462686567164179e-05, "loss": 1.8811, "step": 1116 }, { "epoch": 0.540072202166065, "grad_norm": 0.8209288716316223, "learning_rate": 9.459797785267213e-05, "loss": 2.0091, "step": 1122 }, { "epoch": 0.5429602888086643, "grad_norm": 0.9299931526184082, "learning_rate": 9.456909003370245e-05, "loss": 1.7928, "step": 1128 }, { "epoch": 0.5458483754512635, "grad_norm": 0.901559591293335, "learning_rate": 9.454020221473279e-05, "loss": 2.0195, "step": 1134 }, { "epoch": 0.5487364620938628, "grad_norm": 0.7455955743789673, "learning_rate": 9.451131439576313e-05, "loss": 1.7071, "step": 1140 }, { "epoch": 0.551624548736462, "grad_norm": 0.8501541018486023, "learning_rate": 9.448242657679345e-05, "loss": 2.0273, "step": 1146 }, { "epoch": 0.5545126353790614, "grad_norm": 0.8086917996406555, "learning_rate": 9.445353875782379e-05, "loss": 1.8067, "step": 1152 }, { "epoch": 0.5574007220216607, "grad_norm": 0.8288106918334961, "learning_rate": 9.442465093885412e-05, "loss": 1.8646, "step": 1158 }, { "epoch": 0.5602888086642599, "grad_norm": 0.9436626434326172, "learning_rate": 9.439576311988446e-05, "loss": 1.8146, "step": 1164 }, { "epoch": 0.5631768953068592, "grad_norm": 0.9421128034591675, "learning_rate": 9.436687530091478e-05, "loss": 1.8354, "step": 1170 }, { "epoch": 0.5660649819494585, "grad_norm": 1.0243662595748901, "learning_rate": 9.433798748194512e-05, "loss": 1.9703, "step": 1176 }, { "epoch": 0.5689530685920577, "grad_norm": 1.0636711120605469, "learning_rate": 9.430909966297546e-05, "loss": 1.8092, "step": 1182 }, { "epoch": 0.5718411552346571, "grad_norm": 0.8733149766921997, "learning_rate": 9.428021184400578e-05, "loss": 1.876, "step": 1188 }, { "epoch": 0.5747292418772563, "grad_norm": 0.9452032446861267, "learning_rate": 9.425132402503611e-05, "loss": 1.9518, "step": 1194 }, { "epoch": 0.5776173285198556, "grad_norm": 0.8435307145118713, "learning_rate": 9.422243620606645e-05, "loss": 1.9722, "step": 1200 }, { "epoch": 0.5805054151624549, "grad_norm": 1.0450546741485596, "learning_rate": 9.419354838709677e-05, "loss": 1.827, "step": 1206 }, { "epoch": 0.5833935018050541, "grad_norm": 0.7851176261901855, "learning_rate": 9.416466056812711e-05, "loss": 1.7594, "step": 1212 }, { "epoch": 0.5862815884476534, "grad_norm": 0.9468355774879456, "learning_rate": 9.413577274915745e-05, "loss": 1.8362, "step": 1218 }, { "epoch": 0.5891696750902528, "grad_norm": 0.9826107621192932, "learning_rate": 9.410688493018777e-05, "loss": 1.7532, "step": 1224 }, { "epoch": 0.592057761732852, "grad_norm": 0.8601197600364685, "learning_rate": 9.407799711121811e-05, "loss": 1.9296, "step": 1230 }, { "epoch": 0.5949458483754513, "grad_norm": 0.9739350080490112, "learning_rate": 9.404910929224844e-05, "loss": 1.8009, "step": 1236 }, { "epoch": 0.5978339350180505, "grad_norm": 0.8854568600654602, "learning_rate": 9.402022147327877e-05, "loss": 1.7029, "step": 1242 }, { "epoch": 0.6007220216606498, "grad_norm": 0.8913396000862122, "learning_rate": 9.39913336543091e-05, "loss": 1.8105, "step": 1248 }, { "epoch": 0.6036101083032491, "grad_norm": 0.8291392922401428, "learning_rate": 9.396244583533944e-05, "loss": 1.8937, "step": 1254 }, { "epoch": 0.6064981949458483, "grad_norm": 0.8857879638671875, "learning_rate": 9.393355801636978e-05, "loss": 1.8274, "step": 1260 }, { "epoch": 0.6093862815884477, "grad_norm": 0.8503677845001221, "learning_rate": 9.39046701974001e-05, "loss": 1.7653, "step": 1266 }, { "epoch": 0.612274368231047, "grad_norm": 0.8525097370147705, "learning_rate": 9.387578237843044e-05, "loss": 1.9174, "step": 1272 }, { "epoch": 0.6151624548736462, "grad_norm": 0.9329962134361267, "learning_rate": 9.384689455946076e-05, "loss": 1.9305, "step": 1278 }, { "epoch": 0.6180505415162455, "grad_norm": 0.868125319480896, "learning_rate": 9.381800674049109e-05, "loss": 1.8828, "step": 1284 }, { "epoch": 0.6209386281588448, "grad_norm": 1.0041636228561401, "learning_rate": 9.378911892152144e-05, "loss": 1.8758, "step": 1290 }, { "epoch": 0.623826714801444, "grad_norm": 0.8816074728965759, "learning_rate": 9.376023110255177e-05, "loss": 1.768, "step": 1296 }, { "epoch": 0.6267148014440433, "grad_norm": 0.8880153298377991, "learning_rate": 9.373134328358209e-05, "loss": 1.6417, "step": 1302 }, { "epoch": 0.6296028880866426, "grad_norm": 0.8256460428237915, "learning_rate": 9.370245546461243e-05, "loss": 1.9324, "step": 1308 }, { "epoch": 0.6324909747292419, "grad_norm": 0.9612888097763062, "learning_rate": 9.367356764564275e-05, "loss": 1.9422, "step": 1314 }, { "epoch": 0.6353790613718412, "grad_norm": 0.8787070512771606, "learning_rate": 9.364467982667309e-05, "loss": 1.7575, "step": 1320 }, { "epoch": 0.6382671480144404, "grad_norm": 0.8224942088127136, "learning_rate": 9.361579200770342e-05, "loss": 1.6799, "step": 1326 }, { "epoch": 0.6411552346570397, "grad_norm": 0.9040127396583557, "learning_rate": 9.358690418873375e-05, "loss": 1.9106, "step": 1332 }, { "epoch": 0.644043321299639, "grad_norm": 0.9278034567832947, "learning_rate": 9.355801636976409e-05, "loss": 1.8981, "step": 1338 }, { "epoch": 0.6469314079422382, "grad_norm": 0.9739242792129517, "learning_rate": 9.352912855079442e-05, "loss": 1.8388, "step": 1344 }, { "epoch": 0.6498194945848376, "grad_norm": 0.847473680973053, "learning_rate": 9.350024073182476e-05, "loss": 1.8612, "step": 1350 }, { "epoch": 0.6527075812274368, "grad_norm": 0.9112971425056458, "learning_rate": 9.347135291285508e-05, "loss": 1.7993, "step": 1356 }, { "epoch": 0.6555956678700361, "grad_norm": 0.9686988592147827, "learning_rate": 9.34424650938854e-05, "loss": 1.9013, "step": 1362 }, { "epoch": 0.6584837545126354, "grad_norm": 0.8313731551170349, "learning_rate": 9.341357727491576e-05, "loss": 1.9114, "step": 1368 }, { "epoch": 0.6613718411552346, "grad_norm": 0.8467274904251099, "learning_rate": 9.338468945594608e-05, "loss": 1.8708, "step": 1374 }, { "epoch": 0.6642599277978339, "grad_norm": 0.9020529389381409, "learning_rate": 9.335580163697642e-05, "loss": 1.9062, "step": 1380 }, { "epoch": 0.6671480144404333, "grad_norm": 0.8216947913169861, "learning_rate": 9.332691381800675e-05, "loss": 1.823, "step": 1386 }, { "epoch": 0.6700361010830325, "grad_norm": 0.9465901851654053, "learning_rate": 9.329802599903707e-05, "loss": 1.9349, "step": 1392 }, { "epoch": 0.6729241877256318, "grad_norm": 0.8203220367431641, "learning_rate": 9.326913818006741e-05, "loss": 2.0348, "step": 1398 }, { "epoch": 0.675812274368231, "grad_norm": 0.951603889465332, "learning_rate": 9.324025036109773e-05, "loss": 1.8948, "step": 1404 }, { "epoch": 0.6787003610108303, "grad_norm": 0.9840219616889954, "learning_rate": 9.321136254212807e-05, "loss": 1.9281, "step": 1410 }, { "epoch": 0.6815884476534296, "grad_norm": 0.8682270646095276, "learning_rate": 9.318247472315841e-05, "loss": 1.7359, "step": 1416 }, { "epoch": 0.6844765342960288, "grad_norm": 1.0090640783309937, "learning_rate": 9.315358690418874e-05, "loss": 1.8792, "step": 1422 }, { "epoch": 0.6873646209386282, "grad_norm": 0.9532263278961182, "learning_rate": 9.312469908521907e-05, "loss": 1.829, "step": 1428 }, { "epoch": 0.6902527075812275, "grad_norm": 0.8232157826423645, "learning_rate": 9.30958112662494e-05, "loss": 1.8748, "step": 1434 }, { "epoch": 0.6931407942238267, "grad_norm": 1.0355809926986694, "learning_rate": 9.306692344727974e-05, "loss": 1.9308, "step": 1440 }, { "epoch": 0.696028880866426, "grad_norm": 0.9785148501396179, "learning_rate": 9.303803562831008e-05, "loss": 1.9289, "step": 1446 }, { "epoch": 0.6989169675090253, "grad_norm": 0.8625774383544922, "learning_rate": 9.30091478093404e-05, "loss": 1.6764, "step": 1452 }, { "epoch": 0.7018050541516245, "grad_norm": 0.9302477836608887, "learning_rate": 9.298025999037074e-05, "loss": 1.9143, "step": 1458 }, { "epoch": 0.7046931407942238, "grad_norm": 0.9322661757469177, "learning_rate": 9.295137217140106e-05, "loss": 1.9104, "step": 1464 }, { "epoch": 0.7075812274368231, "grad_norm": 0.9997643232345581, "learning_rate": 9.292248435243139e-05, "loss": 1.8039, "step": 1470 }, { "epoch": 0.7104693140794224, "grad_norm": 0.8950901627540588, "learning_rate": 9.289359653346173e-05, "loss": 1.7502, "step": 1476 }, { "epoch": 0.7133574007220217, "grad_norm": 0.8296883702278137, "learning_rate": 9.286470871449206e-05, "loss": 1.8705, "step": 1482 }, { "epoch": 0.7162454873646209, "grad_norm": 1.0351941585540771, "learning_rate": 9.283582089552239e-05, "loss": 1.7832, "step": 1488 }, { "epoch": 0.7191335740072202, "grad_norm": 0.8638412356376648, "learning_rate": 9.280693307655273e-05, "loss": 1.9876, "step": 1494 }, { "epoch": 0.7220216606498195, "grad_norm": 0.8901738524436951, "learning_rate": 9.277804525758305e-05, "loss": 1.9947, "step": 1500 }, { "epoch": 0.7249097472924187, "grad_norm": 0.8631407618522644, "learning_rate": 9.274915743861339e-05, "loss": 1.9187, "step": 1506 }, { "epoch": 0.7277978339350181, "grad_norm": 0.7957249283790588, "learning_rate": 9.272026961964372e-05, "loss": 1.7462, "step": 1512 }, { "epoch": 0.7306859205776173, "grad_norm": 0.934535562992096, "learning_rate": 9.269138180067405e-05, "loss": 1.8903, "step": 1518 }, { "epoch": 0.7335740072202166, "grad_norm": 0.9951024055480957, "learning_rate": 9.266249398170439e-05, "loss": 1.8042, "step": 1524 }, { "epoch": 0.7364620938628159, "grad_norm": 0.9872926473617554, "learning_rate": 9.263360616273472e-05, "loss": 1.7489, "step": 1530 }, { "epoch": 0.7393501805054151, "grad_norm": 0.9420842528343201, "learning_rate": 9.260471834376506e-05, "loss": 1.8499, "step": 1536 }, { "epoch": 0.7422382671480144, "grad_norm": 0.8420530557632446, "learning_rate": 9.257583052479538e-05, "loss": 1.8963, "step": 1542 }, { "epoch": 0.7451263537906138, "grad_norm": 1.0515209436416626, "learning_rate": 9.25469427058257e-05, "loss": 1.8133, "step": 1548 }, { "epoch": 0.748014440433213, "grad_norm": 0.8648954033851624, "learning_rate": 9.251805488685604e-05, "loss": 1.6847, "step": 1554 }, { "epoch": 0.7509025270758123, "grad_norm": 0.8986181616783142, "learning_rate": 9.248916706788638e-05, "loss": 1.8927, "step": 1560 }, { "epoch": 0.7537906137184116, "grad_norm": 0.9706621170043945, "learning_rate": 9.246027924891672e-05, "loss": 1.9018, "step": 1566 }, { "epoch": 0.7566787003610108, "grad_norm": 0.9827377200126648, "learning_rate": 9.243139142994705e-05, "loss": 1.8079, "step": 1572 }, { "epoch": 0.7595667870036101, "grad_norm": 0.9754669666290283, "learning_rate": 9.240250361097737e-05, "loss": 1.7695, "step": 1578 }, { "epoch": 0.7624548736462093, "grad_norm": 0.8769993185997009, "learning_rate": 9.237361579200771e-05, "loss": 1.8216, "step": 1584 }, { "epoch": 0.7653429602888087, "grad_norm": 0.944936215877533, "learning_rate": 9.234472797303803e-05, "loss": 1.859, "step": 1590 }, { "epoch": 0.768231046931408, "grad_norm": 0.9633384943008423, "learning_rate": 9.231584015406837e-05, "loss": 2.0226, "step": 1596 }, { "epoch": 0.7711191335740072, "grad_norm": 0.7893475890159607, "learning_rate": 9.228695233509871e-05, "loss": 1.8286, "step": 1602 }, { "epoch": 0.7740072202166065, "grad_norm": 0.8420344591140747, "learning_rate": 9.225806451612904e-05, "loss": 1.8814, "step": 1608 }, { "epoch": 0.7768953068592058, "grad_norm": 0.8740891218185425, "learning_rate": 9.222917669715937e-05, "loss": 1.892, "step": 1614 }, { "epoch": 0.779783393501805, "grad_norm": 0.8923601508140564, "learning_rate": 9.22002888781897e-05, "loss": 1.8693, "step": 1620 }, { "epoch": 0.7826714801444044, "grad_norm": 0.8936988711357117, "learning_rate": 9.217140105922004e-05, "loss": 1.83, "step": 1626 }, { "epoch": 0.7855595667870036, "grad_norm": 0.8722962141036987, "learning_rate": 9.214251324025036e-05, "loss": 1.8782, "step": 1632 }, { "epoch": 0.7884476534296029, "grad_norm": 0.8476417660713196, "learning_rate": 9.21136254212807e-05, "loss": 1.9952, "step": 1638 }, { "epoch": 0.7913357400722022, "grad_norm": 0.8244451284408569, "learning_rate": 9.208473760231104e-05, "loss": 1.9606, "step": 1644 }, { "epoch": 0.7942238267148014, "grad_norm": 0.8247030377388, "learning_rate": 9.205584978334136e-05, "loss": 1.8087, "step": 1650 }, { "epoch": 0.7971119133574007, "grad_norm": 0.8857212066650391, "learning_rate": 9.202696196437169e-05, "loss": 1.7191, "step": 1656 }, { "epoch": 0.8, "grad_norm": 0.8654845952987671, "learning_rate": 9.199807414540203e-05, "loss": 1.905, "step": 1662 }, { "epoch": 0.8028880866425993, "grad_norm": 0.9047902226448059, "learning_rate": 9.196918632643235e-05, "loss": 1.8211, "step": 1668 }, { "epoch": 0.8057761732851986, "grad_norm": 1.3159352540969849, "learning_rate": 9.194029850746269e-05, "loss": 1.7811, "step": 1674 }, { "epoch": 0.8086642599277978, "grad_norm": 0.7594133615493774, "learning_rate": 9.191141068849303e-05, "loss": 1.8047, "step": 1680 }, { "epoch": 0.8115523465703971, "grad_norm": 0.9146220684051514, "learning_rate": 9.188252286952335e-05, "loss": 1.7584, "step": 1686 }, { "epoch": 0.8144404332129964, "grad_norm": 0.9372876882553101, "learning_rate": 9.185363505055369e-05, "loss": 1.7256, "step": 1692 }, { "epoch": 0.8173285198555956, "grad_norm": 0.8805015683174133, "learning_rate": 9.182474723158402e-05, "loss": 1.6554, "step": 1698 }, { "epoch": 0.8202166064981949, "grad_norm": 0.932763397693634, "learning_rate": 9.179585941261435e-05, "loss": 1.8781, "step": 1704 }, { "epoch": 0.8231046931407943, "grad_norm": 0.8992255926132202, "learning_rate": 9.176697159364468e-05, "loss": 1.9043, "step": 1710 }, { "epoch": 0.8259927797833935, "grad_norm": 0.93336021900177, "learning_rate": 9.173808377467502e-05, "loss": 1.7806, "step": 1716 }, { "epoch": 0.8288808664259928, "grad_norm": 1.3005733489990234, "learning_rate": 9.170919595570536e-05, "loss": 1.8429, "step": 1722 }, { "epoch": 0.8317689530685921, "grad_norm": 0.943427562713623, "learning_rate": 9.168030813673568e-05, "loss": 1.9096, "step": 1728 }, { "epoch": 0.8346570397111913, "grad_norm": 0.9351149201393127, "learning_rate": 9.165142031776602e-05, "loss": 2.0987, "step": 1734 }, { "epoch": 0.8375451263537906, "grad_norm": 0.8708287477493286, "learning_rate": 9.162253249879634e-05, "loss": 1.7912, "step": 1740 }, { "epoch": 0.8404332129963898, "grad_norm": 0.9620802402496338, "learning_rate": 9.159364467982667e-05, "loss": 1.825, "step": 1746 }, { "epoch": 0.8433212996389892, "grad_norm": 0.9509764313697815, "learning_rate": 9.156475686085702e-05, "loss": 1.8606, "step": 1752 }, { "epoch": 0.8462093862815885, "grad_norm": 0.8942235112190247, "learning_rate": 9.153586904188735e-05, "loss": 1.9101, "step": 1758 }, { "epoch": 0.8490974729241877, "grad_norm": 0.8848027586936951, "learning_rate": 9.150698122291767e-05, "loss": 1.7782, "step": 1764 }, { "epoch": 0.851985559566787, "grad_norm": 0.9881941676139832, "learning_rate": 9.147809340394801e-05, "loss": 1.774, "step": 1770 }, { "epoch": 0.8548736462093863, "grad_norm": 0.9735055565834045, "learning_rate": 9.144920558497833e-05, "loss": 1.8205, "step": 1776 }, { "epoch": 0.8577617328519855, "grad_norm": 0.9619297981262207, "learning_rate": 9.142031776600867e-05, "loss": 1.7643, "step": 1782 }, { "epoch": 0.8606498194945849, "grad_norm": 0.8577748537063599, "learning_rate": 9.1391429947039e-05, "loss": 1.8891, "step": 1788 }, { "epoch": 0.8635379061371841, "grad_norm": 0.844664990901947, "learning_rate": 9.136254212806933e-05, "loss": 1.8615, "step": 1794 }, { "epoch": 0.8664259927797834, "grad_norm": 1.0318166017532349, "learning_rate": 9.133365430909967e-05, "loss": 1.7101, "step": 1800 }, { "epoch": 0.8693140794223827, "grad_norm": 0.894458532333374, "learning_rate": 9.130476649013e-05, "loss": 1.7533, "step": 1806 }, { "epoch": 0.8722021660649819, "grad_norm": 0.9777185916900635, "learning_rate": 9.127587867116034e-05, "loss": 1.8554, "step": 1812 }, { "epoch": 0.8750902527075812, "grad_norm": 0.9938101768493652, "learning_rate": 9.124699085219066e-05, "loss": 1.6738, "step": 1818 }, { "epoch": 0.8779783393501805, "grad_norm": 0.8971199989318848, "learning_rate": 9.121810303322099e-05, "loss": 1.7175, "step": 1824 }, { "epoch": 0.8808664259927798, "grad_norm": 1.0048614740371704, "learning_rate": 9.118921521425134e-05, "loss": 1.8605, "step": 1830 }, { "epoch": 0.8837545126353791, "grad_norm": 1.1836994886398315, "learning_rate": 9.116032739528166e-05, "loss": 2.0019, "step": 1836 }, { "epoch": 0.8866425992779784, "grad_norm": 0.9762217402458191, "learning_rate": 9.113143957631199e-05, "loss": 2.1297, "step": 1842 }, { "epoch": 0.8895306859205776, "grad_norm": 0.8256924748420715, "learning_rate": 9.110255175734233e-05, "loss": 1.867, "step": 1848 }, { "epoch": 0.8924187725631769, "grad_norm": 1.045291543006897, "learning_rate": 9.107366393837265e-05, "loss": 1.6801, "step": 1854 }, { "epoch": 0.8953068592057761, "grad_norm": 0.815434455871582, "learning_rate": 9.104477611940299e-05, "loss": 1.7546, "step": 1860 }, { "epoch": 0.8981949458483754, "grad_norm": 0.8940648436546326, "learning_rate": 9.101588830043331e-05, "loss": 1.8755, "step": 1866 }, { "epoch": 0.9010830324909748, "grad_norm": 0.9004398584365845, "learning_rate": 9.098700048146365e-05, "loss": 1.8167, "step": 1872 }, { "epoch": 0.903971119133574, "grad_norm": 0.9392117261886597, "learning_rate": 9.095811266249399e-05, "loss": 1.7528, "step": 1878 }, { "epoch": 0.9068592057761733, "grad_norm": 0.8730528950691223, "learning_rate": 9.092922484352432e-05, "loss": 1.7747, "step": 1884 }, { "epoch": 0.9097472924187726, "grad_norm": 0.8894970417022705, "learning_rate": 9.090033702455465e-05, "loss": 1.8911, "step": 1890 }, { "epoch": 0.9126353790613718, "grad_norm": 0.8260467648506165, "learning_rate": 9.087144920558498e-05, "loss": 1.819, "step": 1896 }, { "epoch": 0.9155234657039711, "grad_norm": 0.9672936201095581, "learning_rate": 9.08425613866153e-05, "loss": 1.769, "step": 1902 }, { "epoch": 0.9184115523465703, "grad_norm": 0.9402258992195129, "learning_rate": 9.081367356764566e-05, "loss": 1.7315, "step": 1908 }, { "epoch": 0.9212996389891697, "grad_norm": 0.8740605115890503, "learning_rate": 9.078478574867598e-05, "loss": 1.7074, "step": 1914 }, { "epoch": 0.924187725631769, "grad_norm": 0.982002854347229, "learning_rate": 9.075589792970632e-05, "loss": 1.5706, "step": 1920 }, { "epoch": 0.9270758122743682, "grad_norm": 0.8364567160606384, "learning_rate": 9.072701011073664e-05, "loss": 1.7983, "step": 1926 }, { "epoch": 0.9299638989169675, "grad_norm": 0.934249222278595, "learning_rate": 9.069812229176697e-05, "loss": 1.7702, "step": 1932 }, { "epoch": 0.9328519855595668, "grad_norm": 0.902988612651825, "learning_rate": 9.06692344727973e-05, "loss": 1.796, "step": 1938 }, { "epoch": 0.935740072202166, "grad_norm": 0.9595827460289001, "learning_rate": 9.064034665382765e-05, "loss": 1.7816, "step": 1944 }, { "epoch": 0.9386281588447654, "grad_norm": 0.9334012269973755, "learning_rate": 9.061145883485797e-05, "loss": 1.8743, "step": 1950 }, { "epoch": 0.9415162454873646, "grad_norm": 0.9705057144165039, "learning_rate": 9.058257101588831e-05, "loss": 1.8199, "step": 1956 }, { "epoch": 0.9444043321299639, "grad_norm": 0.7907630205154419, "learning_rate": 9.055368319691863e-05, "loss": 1.676, "step": 1962 }, { "epoch": 0.9472924187725632, "grad_norm": 0.9483066201210022, "learning_rate": 9.052479537794897e-05, "loss": 1.7532, "step": 1968 }, { "epoch": 0.9501805054151624, "grad_norm": 0.9475246071815491, "learning_rate": 9.04959075589793e-05, "loss": 1.8601, "step": 1974 }, { "epoch": 0.9530685920577617, "grad_norm": 0.9407988786697388, "learning_rate": 9.046701974000963e-05, "loss": 1.6404, "step": 1980 }, { "epoch": 0.955956678700361, "grad_norm": 1.0300847291946411, "learning_rate": 9.043813192103997e-05, "loss": 1.808, "step": 1986 }, { "epoch": 0.9588447653429603, "grad_norm": 0.8636395335197449, "learning_rate": 9.04092441020703e-05, "loss": 1.8415, "step": 1992 }, { "epoch": 0.9617328519855596, "grad_norm": 0.9217292070388794, "learning_rate": 9.038035628310064e-05, "loss": 1.7774, "step": 1998 }, { "epoch": 0.9646209386281589, "grad_norm": 0.8736981749534607, "learning_rate": 9.035146846413096e-05, "loss": 1.8354, "step": 2004 }, { "epoch": 0.9675090252707581, "grad_norm": 0.8969728946685791, "learning_rate": 9.032258064516129e-05, "loss": 1.8406, "step": 2010 }, { "epoch": 0.9703971119133574, "grad_norm": 1.0683832168579102, "learning_rate": 9.029369282619162e-05, "loss": 1.9605, "step": 2016 }, { "epoch": 0.9732851985559566, "grad_norm": 0.8833884000778198, "learning_rate": 9.026480500722196e-05, "loss": 1.8209, "step": 2022 }, { "epoch": 0.976173285198556, "grad_norm": 0.8806989789009094, "learning_rate": 9.02359171882523e-05, "loss": 1.9671, "step": 2028 }, { "epoch": 0.9790613718411553, "grad_norm": 1.344590663909912, "learning_rate": 9.020702936928263e-05, "loss": 1.8679, "step": 2034 }, { "epoch": 0.9819494584837545, "grad_norm": 0.9840295910835266, "learning_rate": 9.017814155031295e-05, "loss": 1.934, "step": 2040 }, { "epoch": 0.9848375451263538, "grad_norm": 0.9418866634368896, "learning_rate": 9.014925373134329e-05, "loss": 1.8794, "step": 2046 }, { "epoch": 0.9877256317689531, "grad_norm": 0.9919633269309998, "learning_rate": 9.012036591237361e-05, "loss": 1.8323, "step": 2052 }, { "epoch": 0.9906137184115523, "grad_norm": 0.9582290053367615, "learning_rate": 9.009147809340395e-05, "loss": 1.8054, "step": 2058 }, { "epoch": 0.9935018050541516, "grad_norm": 1.2520999908447266, "learning_rate": 9.006259027443429e-05, "loss": 1.9384, "step": 2064 }, { "epoch": 0.9963898916967509, "grad_norm": 0.9360920190811157, "learning_rate": 9.003370245546462e-05, "loss": 1.7446, "step": 2070 }, { "epoch": 0.9992779783393502, "grad_norm": 1.016358494758606, "learning_rate": 9.000481463649495e-05, "loss": 1.8447, "step": 2076 }, { "epoch": 1.0021660649819495, "grad_norm": 0.9083847403526306, "learning_rate": 8.997592681752528e-05, "loss": 1.7043, "step": 2082 }, { "epoch": 1.0050541516245488, "grad_norm": 0.9961937069892883, "learning_rate": 8.994703899855562e-05, "loss": 1.738, "step": 2088 }, { "epoch": 1.007942238267148, "grad_norm": 0.8415824770927429, "learning_rate": 8.991815117958594e-05, "loss": 1.6261, "step": 2094 }, { "epoch": 1.0108303249097472, "grad_norm": 0.9961141347885132, "learning_rate": 8.988926336061628e-05, "loss": 1.8498, "step": 2100 }, { "epoch": 1.0137184115523465, "grad_norm": 0.8841677308082581, "learning_rate": 8.986037554164662e-05, "loss": 1.8537, "step": 2106 }, { "epoch": 1.0166064981949459, "grad_norm": 0.9442580342292786, "learning_rate": 8.983148772267694e-05, "loss": 1.6557, "step": 2112 }, { "epoch": 1.0194945848375452, "grad_norm": 0.908871054649353, "learning_rate": 8.980259990370727e-05, "loss": 1.7399, "step": 2118 }, { "epoch": 1.0223826714801445, "grad_norm": 0.9388368129730225, "learning_rate": 8.97737120847376e-05, "loss": 1.8534, "step": 2124 }, { "epoch": 1.0252707581227436, "grad_norm": 0.994142472743988, "learning_rate": 8.974482426576793e-05, "loss": 1.8387, "step": 2130 }, { "epoch": 1.028158844765343, "grad_norm": 1.1008354425430298, "learning_rate": 8.971593644679827e-05, "loss": 1.925, "step": 2136 }, { "epoch": 1.0310469314079422, "grad_norm": 0.9457741379737854, "learning_rate": 8.968704862782861e-05, "loss": 1.6821, "step": 2142 }, { "epoch": 1.0339350180505416, "grad_norm": 0.9406821131706238, "learning_rate": 8.965816080885893e-05, "loss": 1.8124, "step": 2148 }, { "epoch": 1.0368231046931409, "grad_norm": 0.971881628036499, "learning_rate": 8.962927298988927e-05, "loss": 1.8048, "step": 2154 }, { "epoch": 1.03971119133574, "grad_norm": 1.0449204444885254, "learning_rate": 8.96003851709196e-05, "loss": 1.7271, "step": 2160 }, { "epoch": 1.0425992779783393, "grad_norm": 1.07316255569458, "learning_rate": 8.957149735194993e-05, "loss": 1.7932, "step": 2166 }, { "epoch": 1.0454873646209386, "grad_norm": 0.9846547245979309, "learning_rate": 8.954260953298026e-05, "loss": 1.6516, "step": 2172 }, { "epoch": 1.048375451263538, "grad_norm": 0.9487558603286743, "learning_rate": 8.95137217140106e-05, "loss": 1.7006, "step": 2178 }, { "epoch": 1.0512635379061372, "grad_norm": 0.9893501400947571, "learning_rate": 8.948483389504094e-05, "loss": 1.8402, "step": 2184 }, { "epoch": 1.0541516245487366, "grad_norm": 0.9936769604682922, "learning_rate": 8.945594607607126e-05, "loss": 1.6854, "step": 2190 }, { "epoch": 1.0570397111913357, "grad_norm": 0.965240478515625, "learning_rate": 8.942705825710159e-05, "loss": 1.6902, "step": 2196 }, { "epoch": 1.059927797833935, "grad_norm": 0.9786770343780518, "learning_rate": 8.939817043813192e-05, "loss": 1.7748, "step": 2202 }, { "epoch": 1.0628158844765343, "grad_norm": 0.9702706336975098, "learning_rate": 8.936928261916225e-05, "loss": 1.6175, "step": 2208 }, { "epoch": 1.0657039711191336, "grad_norm": 1.0654871463775635, "learning_rate": 8.93403948001926e-05, "loss": 1.8488, "step": 2214 }, { "epoch": 1.068592057761733, "grad_norm": 1.0505008697509766, "learning_rate": 8.931150698122293e-05, "loss": 1.834, "step": 2220 }, { "epoch": 1.071480144404332, "grad_norm": 1.0047574043273926, "learning_rate": 8.928261916225325e-05, "loss": 1.8519, "step": 2226 }, { "epoch": 1.0743682310469314, "grad_norm": 0.996280312538147, "learning_rate": 8.925373134328359e-05, "loss": 1.7325, "step": 2232 }, { "epoch": 1.0772563176895307, "grad_norm": 0.9753589034080505, "learning_rate": 8.922484352431391e-05, "loss": 1.8097, "step": 2238 }, { "epoch": 1.08014440433213, "grad_norm": 1.0616793632507324, "learning_rate": 8.919595570534425e-05, "loss": 1.8369, "step": 2244 }, { "epoch": 1.0830324909747293, "grad_norm": 1.1079219579696655, "learning_rate": 8.916706788637458e-05, "loss": 1.6744, "step": 2250 }, { "epoch": 1.0859205776173284, "grad_norm": 1.1477363109588623, "learning_rate": 8.913818006740492e-05, "loss": 1.6888, "step": 2256 }, { "epoch": 1.0888086642599277, "grad_norm": 1.048020601272583, "learning_rate": 8.910929224843525e-05, "loss": 1.8622, "step": 2262 }, { "epoch": 1.091696750902527, "grad_norm": 1.2666881084442139, "learning_rate": 8.908040442946558e-05, "loss": 1.7299, "step": 2268 }, { "epoch": 1.0945848375451264, "grad_norm": 0.9427263736724854, "learning_rate": 8.905151661049592e-05, "loss": 1.6548, "step": 2274 }, { "epoch": 1.0974729241877257, "grad_norm": 1.073176622390747, "learning_rate": 8.902262879152624e-05, "loss": 1.8006, "step": 2280 }, { "epoch": 1.100361010830325, "grad_norm": 0.9325835704803467, "learning_rate": 8.899374097255657e-05, "loss": 1.7079, "step": 2286 }, { "epoch": 1.103249097472924, "grad_norm": 0.8748703002929688, "learning_rate": 8.896485315358692e-05, "loss": 1.7671, "step": 2292 }, { "epoch": 1.1061371841155234, "grad_norm": 0.997786283493042, "learning_rate": 8.893596533461724e-05, "loss": 1.7821, "step": 2298 }, { "epoch": 1.1090252707581227, "grad_norm": 1.0615136623382568, "learning_rate": 8.890707751564757e-05, "loss": 1.8833, "step": 2304 }, { "epoch": 1.111913357400722, "grad_norm": 1.0575331449508667, "learning_rate": 8.88781896966779e-05, "loss": 1.5439, "step": 2310 }, { "epoch": 1.1148014440433214, "grad_norm": 1.117900013923645, "learning_rate": 8.884930187770823e-05, "loss": 1.736, "step": 2316 }, { "epoch": 1.1176895306859205, "grad_norm": 1.0114376544952393, "learning_rate": 8.882041405873857e-05, "loss": 1.4941, "step": 2322 }, { "epoch": 1.1205776173285198, "grad_norm": 1.1415714025497437, "learning_rate": 8.87915262397689e-05, "loss": 1.8393, "step": 2328 }, { "epoch": 1.1234657039711191, "grad_norm": 1.101880431175232, "learning_rate": 8.876263842079923e-05, "loss": 1.6394, "step": 2334 }, { "epoch": 1.1263537906137184, "grad_norm": 0.9148219227790833, "learning_rate": 8.873375060182957e-05, "loss": 1.5605, "step": 2340 }, { "epoch": 1.1292418772563177, "grad_norm": 0.9689401984214783, "learning_rate": 8.87048627828599e-05, "loss": 1.9746, "step": 2346 }, { "epoch": 1.132129963898917, "grad_norm": 0.9616325497627258, "learning_rate": 8.867597496389023e-05, "loss": 1.616, "step": 2352 }, { "epoch": 1.1350180505415162, "grad_norm": 1.0750771760940552, "learning_rate": 8.864708714492056e-05, "loss": 1.7647, "step": 2358 }, { "epoch": 1.1379061371841155, "grad_norm": 1.1549386978149414, "learning_rate": 8.861819932595088e-05, "loss": 1.7737, "step": 2364 }, { "epoch": 1.1407942238267148, "grad_norm": 0.9160277247428894, "learning_rate": 8.858931150698124e-05, "loss": 1.6646, "step": 2370 }, { "epoch": 1.1436823104693141, "grad_norm": 1.0932331085205078, "learning_rate": 8.856042368801156e-05, "loss": 1.6903, "step": 2376 }, { "epoch": 1.1465703971119134, "grad_norm": 1.0549309253692627, "learning_rate": 8.85315358690419e-05, "loss": 1.8913, "step": 2382 }, { "epoch": 1.1494584837545125, "grad_norm": 1.1628798246383667, "learning_rate": 8.850264805007222e-05, "loss": 1.751, "step": 2388 }, { "epoch": 1.1523465703971119, "grad_norm": 0.9804364442825317, "learning_rate": 8.847376023110255e-05, "loss": 1.8117, "step": 2394 }, { "epoch": 1.1552346570397112, "grad_norm": 1.0072076320648193, "learning_rate": 8.844487241213289e-05, "loss": 1.76, "step": 2400 }, { "epoch": 1.1581227436823105, "grad_norm": 1.0015157461166382, "learning_rate": 8.841598459316323e-05, "loss": 1.8551, "step": 2406 }, { "epoch": 1.1610108303249098, "grad_norm": 1.0217664241790771, "learning_rate": 8.838709677419355e-05, "loss": 1.7495, "step": 2412 }, { "epoch": 1.1638989169675091, "grad_norm": 1.1116472482681274, "learning_rate": 8.835820895522389e-05, "loss": 1.7464, "step": 2418 }, { "epoch": 1.1667870036101082, "grad_norm": 1.1815686225891113, "learning_rate": 8.832932113625421e-05, "loss": 1.681, "step": 2424 }, { "epoch": 1.1696750902527075, "grad_norm": 1.0796726942062378, "learning_rate": 8.830043331728455e-05, "loss": 1.7095, "step": 2430 }, { "epoch": 1.1725631768953069, "grad_norm": 0.9273179173469543, "learning_rate": 8.827154549831488e-05, "loss": 1.6938, "step": 2436 }, { "epoch": 1.1754512635379062, "grad_norm": 0.9109070897102356, "learning_rate": 8.824265767934522e-05, "loss": 1.6893, "step": 2442 }, { "epoch": 1.1783393501805055, "grad_norm": 0.9104015827178955, "learning_rate": 8.821376986037555e-05, "loss": 1.7906, "step": 2448 }, { "epoch": 1.1812274368231046, "grad_norm": 0.9482967853546143, "learning_rate": 8.818488204140588e-05, "loss": 1.746, "step": 2454 }, { "epoch": 1.184115523465704, "grad_norm": 1.0127427577972412, "learning_rate": 8.815599422243622e-05, "loss": 1.6273, "step": 2460 }, { "epoch": 1.1870036101083032, "grad_norm": 0.9931307435035706, "learning_rate": 8.812710640346654e-05, "loss": 1.7379, "step": 2466 }, { "epoch": 1.1898916967509026, "grad_norm": 0.9653785824775696, "learning_rate": 8.809821858449687e-05, "loss": 1.6758, "step": 2472 }, { "epoch": 1.1927797833935019, "grad_norm": 1.1514537334442139, "learning_rate": 8.80693307655272e-05, "loss": 1.6432, "step": 2478 }, { "epoch": 1.1956678700361012, "grad_norm": 1.0019876956939697, "learning_rate": 8.804044294655754e-05, "loss": 1.9578, "step": 2484 }, { "epoch": 1.1985559566787003, "grad_norm": 0.9717727303504944, "learning_rate": 8.801155512758788e-05, "loss": 1.7929, "step": 2490 }, { "epoch": 1.2014440433212996, "grad_norm": 0.9696713089942932, "learning_rate": 8.79826673086182e-05, "loss": 1.634, "step": 2496 }, { "epoch": 1.204332129963899, "grad_norm": 0.9992627501487732, "learning_rate": 8.795377948964853e-05, "loss": 1.6882, "step": 2502 }, { "epoch": 1.2072202166064983, "grad_norm": 1.0690054893493652, "learning_rate": 8.792489167067887e-05, "loss": 1.6777, "step": 2508 }, { "epoch": 1.2101083032490974, "grad_norm": 1.05001699924469, "learning_rate": 8.78960038517092e-05, "loss": 1.7851, "step": 2514 }, { "epoch": 1.2129963898916967, "grad_norm": 0.9634299874305725, "learning_rate": 8.786711603273953e-05, "loss": 1.8262, "step": 2520 }, { "epoch": 1.215884476534296, "grad_norm": 1.1295297145843506, "learning_rate": 8.783822821376987e-05, "loss": 1.8273, "step": 2526 }, { "epoch": 1.2187725631768953, "grad_norm": 1.2978733777999878, "learning_rate": 8.78093403948002e-05, "loss": 1.8286, "step": 2532 }, { "epoch": 1.2216606498194946, "grad_norm": 0.9385501742362976, "learning_rate": 8.778045257583053e-05, "loss": 1.6883, "step": 2538 }, { "epoch": 1.224548736462094, "grad_norm": 0.9171926975250244, "learning_rate": 8.775156475686086e-05, "loss": 1.662, "step": 2544 }, { "epoch": 1.2274368231046933, "grad_norm": 0.9665663242340088, "learning_rate": 8.77226769378912e-05, "loss": 1.5996, "step": 2550 }, { "epoch": 1.2303249097472924, "grad_norm": 1.0602624416351318, "learning_rate": 8.769378911892152e-05, "loss": 1.7716, "step": 2556 }, { "epoch": 1.2332129963898917, "grad_norm": 1.0323963165283203, "learning_rate": 8.766490129995186e-05, "loss": 1.7883, "step": 2562 }, { "epoch": 1.236101083032491, "grad_norm": 0.9718726277351379, "learning_rate": 8.76360134809822e-05, "loss": 1.8158, "step": 2568 }, { "epoch": 1.2389891696750903, "grad_norm": 1.142627477645874, "learning_rate": 8.760712566201252e-05, "loss": 1.7251, "step": 2574 }, { "epoch": 1.2418772563176894, "grad_norm": 1.004875659942627, "learning_rate": 8.757823784304285e-05, "loss": 1.6949, "step": 2580 }, { "epoch": 1.2447653429602887, "grad_norm": 1.0258288383483887, "learning_rate": 8.754935002407319e-05, "loss": 1.7968, "step": 2586 }, { "epoch": 1.247653429602888, "grad_norm": 1.0146340131759644, "learning_rate": 8.752046220510351e-05, "loss": 1.7542, "step": 2592 }, { "epoch": 1.2505415162454874, "grad_norm": 1.0843137502670288, "learning_rate": 8.749157438613385e-05, "loss": 1.6311, "step": 2598 }, { "epoch": 1.2534296028880867, "grad_norm": 1.0937719345092773, "learning_rate": 8.746268656716419e-05, "loss": 1.8367, "step": 2604 }, { "epoch": 1.256317689530686, "grad_norm": 1.0456905364990234, "learning_rate": 8.743379874819451e-05, "loss": 1.9052, "step": 2610 }, { "epoch": 1.2592057761732853, "grad_norm": 1.0397287607192993, "learning_rate": 8.740491092922485e-05, "loss": 1.7185, "step": 2616 }, { "epoch": 1.2620938628158844, "grad_norm": 1.0215904712677002, "learning_rate": 8.737602311025518e-05, "loss": 1.7426, "step": 2622 }, { "epoch": 1.2649819494584837, "grad_norm": 0.9434764385223389, "learning_rate": 8.734713529128551e-05, "loss": 1.6707, "step": 2628 }, { "epoch": 1.267870036101083, "grad_norm": 1.181675910949707, "learning_rate": 8.731824747231584e-05, "loss": 1.7161, "step": 2634 }, { "epoch": 1.2707581227436824, "grad_norm": 1.0122061967849731, "learning_rate": 8.728935965334618e-05, "loss": 1.6888, "step": 2640 }, { "epoch": 1.2736462093862815, "grad_norm": 1.0125913619995117, "learning_rate": 8.726047183437652e-05, "loss": 1.7665, "step": 2646 }, { "epoch": 1.2765342960288808, "grad_norm": 0.9477354288101196, "learning_rate": 8.723158401540684e-05, "loss": 1.7568, "step": 2652 }, { "epoch": 1.2794223826714801, "grad_norm": 0.8869631886482239, "learning_rate": 8.720269619643717e-05, "loss": 1.7527, "step": 2658 }, { "epoch": 1.2823104693140794, "grad_norm": 1.2832481861114502, "learning_rate": 8.71738083774675e-05, "loss": 1.7923, "step": 2664 }, { "epoch": 1.2851985559566788, "grad_norm": 0.9761554598808289, "learning_rate": 8.714492055849783e-05, "loss": 1.7485, "step": 2670 }, { "epoch": 1.288086642599278, "grad_norm": 1.0984383821487427, "learning_rate": 8.711603273952818e-05, "loss": 1.7928, "step": 2676 }, { "epoch": 1.2909747292418774, "grad_norm": 0.9601598978042603, "learning_rate": 8.70871449205585e-05, "loss": 1.6855, "step": 2682 }, { "epoch": 1.2938628158844765, "grad_norm": 1.125381350517273, "learning_rate": 8.705825710158883e-05, "loss": 1.7978, "step": 2688 }, { "epoch": 1.2967509025270758, "grad_norm": 1.049444317817688, "learning_rate": 8.702936928261917e-05, "loss": 1.7855, "step": 2694 }, { "epoch": 1.2996389891696751, "grad_norm": 1.0342787504196167, "learning_rate": 8.70004814636495e-05, "loss": 1.8188, "step": 2700 }, { "epoch": 1.3025270758122744, "grad_norm": 1.0101394653320312, "learning_rate": 8.697159364467983e-05, "loss": 1.6311, "step": 2706 }, { "epoch": 1.3054151624548735, "grad_norm": 0.9851915836334229, "learning_rate": 8.694270582571016e-05, "loss": 1.6549, "step": 2712 }, { "epoch": 1.3083032490974729, "grad_norm": 1.1822905540466309, "learning_rate": 8.69138180067405e-05, "loss": 1.838, "step": 2718 }, { "epoch": 1.3111913357400722, "grad_norm": 1.008961796760559, "learning_rate": 8.688493018777083e-05, "loss": 1.683, "step": 2724 }, { "epoch": 1.3140794223826715, "grad_norm": 1.1166210174560547, "learning_rate": 8.685604236880116e-05, "loss": 1.6471, "step": 2730 }, { "epoch": 1.3169675090252708, "grad_norm": 0.9904022216796875, "learning_rate": 8.68271545498315e-05, "loss": 1.8926, "step": 2736 }, { "epoch": 1.3198555956678701, "grad_norm": 1.0302913188934326, "learning_rate": 8.679826673086182e-05, "loss": 1.6918, "step": 2742 }, { "epoch": 1.3227436823104692, "grad_norm": 0.9691470861434937, "learning_rate": 8.676937891189215e-05, "loss": 1.7238, "step": 2748 }, { "epoch": 1.3256317689530686, "grad_norm": 1.0059096813201904, "learning_rate": 8.67404910929225e-05, "loss": 1.7318, "step": 2754 }, { "epoch": 1.3285198555956679, "grad_norm": 1.0634851455688477, "learning_rate": 8.671160327395282e-05, "loss": 1.8358, "step": 2760 }, { "epoch": 1.3314079422382672, "grad_norm": 0.9774668216705322, "learning_rate": 8.668271545498315e-05, "loss": 1.7196, "step": 2766 }, { "epoch": 1.3342960288808663, "grad_norm": 1.03284752368927, "learning_rate": 8.665382763601349e-05, "loss": 1.6497, "step": 2772 }, { "epoch": 1.3371841155234656, "grad_norm": 0.9450129270553589, "learning_rate": 8.662493981704381e-05, "loss": 1.6532, "step": 2778 }, { "epoch": 1.340072202166065, "grad_norm": 0.953349232673645, "learning_rate": 8.659605199807415e-05, "loss": 1.6214, "step": 2784 }, { "epoch": 1.3429602888086642, "grad_norm": 1.1071418523788452, "learning_rate": 8.656716417910447e-05, "loss": 1.7346, "step": 2790 }, { "epoch": 1.3458483754512636, "grad_norm": 1.0929051637649536, "learning_rate": 8.653827636013481e-05, "loss": 1.7152, "step": 2796 }, { "epoch": 1.3487364620938629, "grad_norm": 1.0930378437042236, "learning_rate": 8.650938854116515e-05, "loss": 1.8128, "step": 2802 }, { "epoch": 1.3516245487364622, "grad_norm": 1.1772379875183105, "learning_rate": 8.648050072219548e-05, "loss": 1.7629, "step": 2808 }, { "epoch": 1.3545126353790613, "grad_norm": 1.0636848211288452, "learning_rate": 8.645161290322581e-05, "loss": 1.8735, "step": 2814 }, { "epoch": 1.3574007220216606, "grad_norm": 0.9733750820159912, "learning_rate": 8.642272508425614e-05, "loss": 1.694, "step": 2820 }, { "epoch": 1.36028880866426, "grad_norm": 1.1212573051452637, "learning_rate": 8.639383726528646e-05, "loss": 1.9241, "step": 2826 }, { "epoch": 1.3631768953068593, "grad_norm": 0.9903728365898132, "learning_rate": 8.636494944631682e-05, "loss": 1.6626, "step": 2832 }, { "epoch": 1.3660649819494584, "grad_norm": 0.9836233854293823, "learning_rate": 8.633606162734714e-05, "loss": 1.7094, "step": 2838 }, { "epoch": 1.3689530685920577, "grad_norm": 1.0178714990615845, "learning_rate": 8.630717380837748e-05, "loss": 1.6921, "step": 2844 }, { "epoch": 1.371841155234657, "grad_norm": 1.15469491481781, "learning_rate": 8.62782859894078e-05, "loss": 1.7139, "step": 2850 }, { "epoch": 1.3747292418772563, "grad_norm": 0.9500898718833923, "learning_rate": 8.624939817043813e-05, "loss": 1.6894, "step": 2856 }, { "epoch": 1.3776173285198556, "grad_norm": 1.0546015501022339, "learning_rate": 8.622051035146847e-05, "loss": 1.6646, "step": 2862 }, { "epoch": 1.380505415162455, "grad_norm": 0.9874374866485596, "learning_rate": 8.61916225324988e-05, "loss": 1.8495, "step": 2868 }, { "epoch": 1.3833935018050543, "grad_norm": 0.9840148091316223, "learning_rate": 8.616273471352913e-05, "loss": 1.8157, "step": 2874 }, { "epoch": 1.3862815884476534, "grad_norm": 1.1555536985397339, "learning_rate": 8.613384689455947e-05, "loss": 1.7951, "step": 2880 }, { "epoch": 1.3891696750902527, "grad_norm": 1.0564396381378174, "learning_rate": 8.61049590755898e-05, "loss": 1.67, "step": 2886 }, { "epoch": 1.392057761732852, "grad_norm": 1.0742223262786865, "learning_rate": 8.607607125662013e-05, "loss": 1.691, "step": 2892 }, { "epoch": 1.3949458483754513, "grad_norm": 1.0695408582687378, "learning_rate": 8.604718343765046e-05, "loss": 1.6352, "step": 2898 }, { "epoch": 1.3978339350180504, "grad_norm": 1.2662606239318848, "learning_rate": 8.60182956186808e-05, "loss": 1.7069, "step": 2904 }, { "epoch": 1.4007220216606497, "grad_norm": 1.053790807723999, "learning_rate": 8.598940779971113e-05, "loss": 1.7029, "step": 2910 }, { "epoch": 1.403610108303249, "grad_norm": 1.0109453201293945, "learning_rate": 8.596051998074146e-05, "loss": 1.6386, "step": 2916 }, { "epoch": 1.4064981949458484, "grad_norm": 1.2060072422027588, "learning_rate": 8.59316321617718e-05, "loss": 1.9432, "step": 2922 }, { "epoch": 1.4093862815884477, "grad_norm": 1.0275509357452393, "learning_rate": 8.590274434280212e-05, "loss": 1.6342, "step": 2928 }, { "epoch": 1.412274368231047, "grad_norm": 0.8910785913467407, "learning_rate": 8.587385652383245e-05, "loss": 1.7677, "step": 2934 }, { "epoch": 1.4151624548736463, "grad_norm": 1.1656451225280762, "learning_rate": 8.584496870486278e-05, "loss": 1.651, "step": 2940 }, { "epoch": 1.4180505415162454, "grad_norm": 1.296594500541687, "learning_rate": 8.581608088589312e-05, "loss": 1.6522, "step": 2946 }, { "epoch": 1.4209386281588448, "grad_norm": 1.0698925256729126, "learning_rate": 8.578719306692345e-05, "loss": 1.8092, "step": 2952 }, { "epoch": 1.423826714801444, "grad_norm": 1.7982864379882812, "learning_rate": 8.575830524795379e-05, "loss": 1.8233, "step": 2958 }, { "epoch": 1.4267148014440434, "grad_norm": 1.1080901622772217, "learning_rate": 8.572941742898411e-05, "loss": 1.7939, "step": 2964 }, { "epoch": 1.4296028880866425, "grad_norm": 1.0394750833511353, "learning_rate": 8.570052961001445e-05, "loss": 1.7214, "step": 2970 }, { "epoch": 1.4324909747292418, "grad_norm": 1.1438641548156738, "learning_rate": 8.567164179104477e-05, "loss": 1.6488, "step": 2976 }, { "epoch": 1.4353790613718411, "grad_norm": 1.1957541704177856, "learning_rate": 8.564275397207511e-05, "loss": 1.7588, "step": 2982 }, { "epoch": 1.4382671480144404, "grad_norm": 1.1108427047729492, "learning_rate": 8.561386615310545e-05, "loss": 1.7508, "step": 2988 }, { "epoch": 1.4411552346570398, "grad_norm": 0.9800836443901062, "learning_rate": 8.558497833413578e-05, "loss": 1.7038, "step": 2994 }, { "epoch": 1.444043321299639, "grad_norm": 1.004544734954834, "learning_rate": 8.555609051516611e-05, "loss": 1.7543, "step": 3000 }, { "epoch": 1.4469314079422384, "grad_norm": 1.067256212234497, "learning_rate": 8.552720269619644e-05, "loss": 1.5596, "step": 3006 }, { "epoch": 1.4498194945848375, "grad_norm": 1.2239118814468384, "learning_rate": 8.549831487722676e-05, "loss": 1.7551, "step": 3012 }, { "epoch": 1.4527075812274368, "grad_norm": 1.0731794834136963, "learning_rate": 8.54694270582571e-05, "loss": 1.7561, "step": 3018 }, { "epoch": 1.4555956678700361, "grad_norm": 1.232521653175354, "learning_rate": 8.544053923928744e-05, "loss": 1.7201, "step": 3024 }, { "epoch": 1.4584837545126355, "grad_norm": 1.1533269882202148, "learning_rate": 8.541165142031778e-05, "loss": 1.6789, "step": 3030 }, { "epoch": 1.4613718411552346, "grad_norm": 1.063510537147522, "learning_rate": 8.53827636013481e-05, "loss": 1.7389, "step": 3036 }, { "epoch": 1.4642599277978339, "grad_norm": 1.2335575819015503, "learning_rate": 8.535387578237843e-05, "loss": 1.8308, "step": 3042 }, { "epoch": 1.4671480144404332, "grad_norm": 1.0851231813430786, "learning_rate": 8.532498796340877e-05, "loss": 1.7899, "step": 3048 }, { "epoch": 1.4700361010830325, "grad_norm": 0.9476112127304077, "learning_rate": 8.529610014443909e-05, "loss": 1.7315, "step": 3054 }, { "epoch": 1.4729241877256318, "grad_norm": 1.1753737926483154, "learning_rate": 8.526721232546943e-05, "loss": 1.7615, "step": 3060 }, { "epoch": 1.4758122743682311, "grad_norm": 1.1020028591156006, "learning_rate": 8.523832450649977e-05, "loss": 1.8995, "step": 3066 }, { "epoch": 1.4787003610108302, "grad_norm": 1.0446275472640991, "learning_rate": 8.52094366875301e-05, "loss": 1.6421, "step": 3072 }, { "epoch": 1.4815884476534296, "grad_norm": 1.0213720798492432, "learning_rate": 8.518054886856043e-05, "loss": 1.5305, "step": 3078 }, { "epoch": 1.4844765342960289, "grad_norm": 1.2649009227752686, "learning_rate": 8.515166104959076e-05, "loss": 1.725, "step": 3084 }, { "epoch": 1.4873646209386282, "grad_norm": 1.0405287742614746, "learning_rate": 8.51227732306211e-05, "loss": 1.7605, "step": 3090 }, { "epoch": 1.4902527075812273, "grad_norm": 1.008787989616394, "learning_rate": 8.509388541165142e-05, "loss": 1.7242, "step": 3096 }, { "epoch": 1.4931407942238266, "grad_norm": 1.0666959285736084, "learning_rate": 8.506499759268176e-05, "loss": 1.7276, "step": 3102 }, { "epoch": 1.496028880866426, "grad_norm": 1.2428715229034424, "learning_rate": 8.50361097737121e-05, "loss": 1.8922, "step": 3108 }, { "epoch": 1.4989169675090253, "grad_norm": 1.181990385055542, "learning_rate": 8.500722195474242e-05, "loss": 1.6425, "step": 3114 }, { "epoch": 1.5018050541516246, "grad_norm": 1.168373703956604, "learning_rate": 8.497833413577275e-05, "loss": 1.766, "step": 3120 }, { "epoch": 1.504693140794224, "grad_norm": 1.2343201637268066, "learning_rate": 8.494944631680308e-05, "loss": 1.7044, "step": 3126 }, { "epoch": 1.5075812274368232, "grad_norm": 1.1096030473709106, "learning_rate": 8.492055849783341e-05, "loss": 1.7391, "step": 3132 }, { "epoch": 1.5104693140794225, "grad_norm": 1.1505628824234009, "learning_rate": 8.489167067886376e-05, "loss": 1.9354, "step": 3138 }, { "epoch": 1.5133574007220216, "grad_norm": 1.1855266094207764, "learning_rate": 8.486278285989409e-05, "loss": 1.7407, "step": 3144 }, { "epoch": 1.516245487364621, "grad_norm": 1.023950457572937, "learning_rate": 8.483389504092441e-05, "loss": 1.8246, "step": 3150 }, { "epoch": 1.5191335740072203, "grad_norm": 1.0361186265945435, "learning_rate": 8.480500722195475e-05, "loss": 1.792, "step": 3156 }, { "epoch": 1.5220216606498194, "grad_norm": 0.877399206161499, "learning_rate": 8.477611940298507e-05, "loss": 1.6681, "step": 3162 }, { "epoch": 1.5249097472924187, "grad_norm": 1.049506425857544, "learning_rate": 8.474723158401541e-05, "loss": 1.8189, "step": 3168 }, { "epoch": 1.527797833935018, "grad_norm": 1.033416748046875, "learning_rate": 8.471834376504574e-05, "loss": 1.6792, "step": 3174 }, { "epoch": 1.5306859205776173, "grad_norm": 1.1268037557601929, "learning_rate": 8.468945594607608e-05, "loss": 1.7762, "step": 3180 }, { "epoch": 1.5335740072202166, "grad_norm": 1.0766969919204712, "learning_rate": 8.466056812710641e-05, "loss": 1.7018, "step": 3186 }, { "epoch": 1.536462093862816, "grad_norm": 1.1467041969299316, "learning_rate": 8.463168030813674e-05, "loss": 1.6004, "step": 3192 }, { "epoch": 1.5393501805054153, "grad_norm": 0.9107251167297363, "learning_rate": 8.460279248916708e-05, "loss": 1.7406, "step": 3198 }, { "epoch": 1.5422382671480146, "grad_norm": 1.2722817659378052, "learning_rate": 8.45739046701974e-05, "loss": 1.8712, "step": 3204 }, { "epoch": 1.5451263537906137, "grad_norm": 1.139040231704712, "learning_rate": 8.454501685122773e-05, "loss": 1.6943, "step": 3210 }, { "epoch": 1.548014440433213, "grad_norm": 1.1248966455459595, "learning_rate": 8.451612903225808e-05, "loss": 1.7745, "step": 3216 }, { "epoch": 1.550902527075812, "grad_norm": 1.2232081890106201, "learning_rate": 8.44872412132884e-05, "loss": 1.585, "step": 3222 }, { "epoch": 1.5537906137184114, "grad_norm": 1.0943208932876587, "learning_rate": 8.445835339431873e-05, "loss": 1.7541, "step": 3228 }, { "epoch": 1.5566787003610107, "grad_norm": 1.0940830707550049, "learning_rate": 8.442946557534907e-05, "loss": 1.6583, "step": 3234 }, { "epoch": 1.55956678700361, "grad_norm": 0.9460192322731018, "learning_rate": 8.440057775637939e-05, "loss": 1.718, "step": 3240 }, { "epoch": 1.5624548736462094, "grad_norm": 1.0804071426391602, "learning_rate": 8.437168993740973e-05, "loss": 1.8761, "step": 3246 }, { "epoch": 1.5653429602888087, "grad_norm": 1.0630260705947876, "learning_rate": 8.434280211844005e-05, "loss": 1.6661, "step": 3252 }, { "epoch": 1.568231046931408, "grad_norm": 1.1862587928771973, "learning_rate": 8.43139142994704e-05, "loss": 1.8148, "step": 3258 }, { "epoch": 1.5711191335740073, "grad_norm": 1.0556730031967163, "learning_rate": 8.428502648050073e-05, "loss": 1.55, "step": 3264 }, { "epoch": 1.5740072202166067, "grad_norm": 1.0431818962097168, "learning_rate": 8.425613866153106e-05, "loss": 1.6068, "step": 3270 }, { "epoch": 1.5768953068592058, "grad_norm": 1.0159975290298462, "learning_rate": 8.42272508425614e-05, "loss": 1.6361, "step": 3276 }, { "epoch": 1.579783393501805, "grad_norm": 1.1946203708648682, "learning_rate": 8.419836302359172e-05, "loss": 1.772, "step": 3282 }, { "epoch": 1.5826714801444042, "grad_norm": 1.0768563747406006, "learning_rate": 8.416947520462204e-05, "loss": 1.646, "step": 3288 }, { "epoch": 1.5855595667870035, "grad_norm": 1.034365177154541, "learning_rate": 8.41405873856524e-05, "loss": 1.6392, "step": 3294 }, { "epoch": 1.5884476534296028, "grad_norm": 0.9765658378601074, "learning_rate": 8.411169956668272e-05, "loss": 1.5322, "step": 3300 }, { "epoch": 1.5913357400722021, "grad_norm": 1.0680170059204102, "learning_rate": 8.408281174771305e-05, "loss": 1.6113, "step": 3306 }, { "epoch": 1.5942238267148015, "grad_norm": 1.2642759084701538, "learning_rate": 8.405392392874338e-05, "loss": 1.9051, "step": 3312 }, { "epoch": 1.5971119133574008, "grad_norm": 1.1315819025039673, "learning_rate": 8.402503610977371e-05, "loss": 1.7804, "step": 3318 }, { "epoch": 1.6, "grad_norm": 1.0824874639511108, "learning_rate": 8.399614829080405e-05, "loss": 1.7895, "step": 3324 }, { "epoch": 1.6028880866425994, "grad_norm": 1.066275954246521, "learning_rate": 8.396726047183439e-05, "loss": 1.7041, "step": 3330 }, { "epoch": 1.6057761732851985, "grad_norm": 1.1938848495483398, "learning_rate": 8.393837265286471e-05, "loss": 1.6845, "step": 3336 }, { "epoch": 1.6086642599277978, "grad_norm": 1.0698251724243164, "learning_rate": 8.390948483389505e-05, "loss": 1.4932, "step": 3342 }, { "epoch": 1.6115523465703971, "grad_norm": 0.9851581454277039, "learning_rate": 8.388059701492537e-05, "loss": 1.7086, "step": 3348 }, { "epoch": 1.6144404332129962, "grad_norm": 1.108106255531311, "learning_rate": 8.385170919595571e-05, "loss": 1.6334, "step": 3354 }, { "epoch": 1.6173285198555956, "grad_norm": 1.020650863647461, "learning_rate": 8.382282137698604e-05, "loss": 1.6268, "step": 3360 }, { "epoch": 1.6202166064981949, "grad_norm": 1.2234477996826172, "learning_rate": 8.379393355801636e-05, "loss": 1.7606, "step": 3366 }, { "epoch": 1.6231046931407942, "grad_norm": 1.1748366355895996, "learning_rate": 8.376504573904671e-05, "loss": 1.7543, "step": 3372 }, { "epoch": 1.6259927797833935, "grad_norm": 0.9767166972160339, "learning_rate": 8.373615792007704e-05, "loss": 1.6512, "step": 3378 }, { "epoch": 1.6288808664259928, "grad_norm": 1.1096307039260864, "learning_rate": 8.370727010110738e-05, "loss": 1.7136, "step": 3384 }, { "epoch": 1.6317689530685922, "grad_norm": 1.1528247594833374, "learning_rate": 8.36783822821377e-05, "loss": 1.7028, "step": 3390 }, { "epoch": 1.6346570397111915, "grad_norm": 1.0301446914672852, "learning_rate": 8.364949446316803e-05, "loss": 1.7274, "step": 3396 }, { "epoch": 1.6375451263537906, "grad_norm": 1.2318360805511475, "learning_rate": 8.362060664419837e-05, "loss": 1.6507, "step": 3402 }, { "epoch": 1.6404332129963899, "grad_norm": 1.1347384452819824, "learning_rate": 8.35917188252287e-05, "loss": 1.7352, "step": 3408 }, { "epoch": 1.6433212996389892, "grad_norm": 1.00263249874115, "learning_rate": 8.356283100625903e-05, "loss": 1.807, "step": 3414 }, { "epoch": 1.6462093862815883, "grad_norm": 1.035452127456665, "learning_rate": 8.353394318728937e-05, "loss": 1.6491, "step": 3420 }, { "epoch": 1.6490974729241876, "grad_norm": 1.1256824731826782, "learning_rate": 8.350505536831969e-05, "loss": 1.7076, "step": 3426 }, { "epoch": 1.651985559566787, "grad_norm": 1.0680978298187256, "learning_rate": 8.347616754935003e-05, "loss": 1.8565, "step": 3432 }, { "epoch": 1.6548736462093863, "grad_norm": 1.0038206577301025, "learning_rate": 8.344727973038035e-05, "loss": 1.7132, "step": 3438 }, { "epoch": 1.6577617328519856, "grad_norm": 1.055567741394043, "learning_rate": 8.341839191141069e-05, "loss": 1.7351, "step": 3444 }, { "epoch": 1.660649819494585, "grad_norm": 1.1556047201156616, "learning_rate": 8.338950409244103e-05, "loss": 1.7888, "step": 3450 }, { "epoch": 1.6635379061371842, "grad_norm": 0.9870200753211975, "learning_rate": 8.336061627347136e-05, "loss": 1.5747, "step": 3456 }, { "epoch": 1.6664259927797835, "grad_norm": 1.1878291368484497, "learning_rate": 8.33317284545017e-05, "loss": 1.7293, "step": 3462 }, { "epoch": 1.6693140794223826, "grad_norm": 0.9205313324928284, "learning_rate": 8.330284063553202e-05, "loss": 1.6705, "step": 3468 }, { "epoch": 1.672202166064982, "grad_norm": 1.0989576578140259, "learning_rate": 8.327395281656234e-05, "loss": 1.6474, "step": 3474 }, { "epoch": 1.6750902527075813, "grad_norm": 1.1472797393798828, "learning_rate": 8.324506499759268e-05, "loss": 1.7974, "step": 3480 }, { "epoch": 1.6779783393501804, "grad_norm": 0.9694781303405762, "learning_rate": 8.321617717862302e-05, "loss": 1.6223, "step": 3486 }, { "epoch": 1.6808664259927797, "grad_norm": 1.0390557050704956, "learning_rate": 8.318728935965336e-05, "loss": 1.8201, "step": 3492 }, { "epoch": 1.683754512635379, "grad_norm": 1.0823266506195068, "learning_rate": 8.315840154068368e-05, "loss": 1.732, "step": 3498 }, { "epoch": 1.6866425992779783, "grad_norm": 1.2140618562698364, "learning_rate": 8.312951372171401e-05, "loss": 1.7687, "step": 3504 }, { "epoch": 1.6895306859205776, "grad_norm": 1.2177654504776, "learning_rate": 8.310062590274435e-05, "loss": 1.5734, "step": 3510 }, { "epoch": 1.692418772563177, "grad_norm": 1.0990577936172485, "learning_rate": 8.307173808377467e-05, "loss": 1.5655, "step": 3516 }, { "epoch": 1.6953068592057763, "grad_norm": 1.1155422925949097, "learning_rate": 8.304285026480501e-05, "loss": 1.8938, "step": 3522 }, { "epoch": 1.6981949458483756, "grad_norm": 0.9653927683830261, "learning_rate": 8.301396244583535e-05, "loss": 1.6425, "step": 3528 }, { "epoch": 1.7010830324909747, "grad_norm": 0.9776820540428162, "learning_rate": 8.298507462686567e-05, "loss": 1.6066, "step": 3534 }, { "epoch": 1.703971119133574, "grad_norm": 1.2487496137619019, "learning_rate": 8.295618680789601e-05, "loss": 1.7163, "step": 3540 }, { "epoch": 1.7068592057761733, "grad_norm": 1.2020738124847412, "learning_rate": 8.292729898892634e-05, "loss": 1.7985, "step": 3546 }, { "epoch": 1.7097472924187724, "grad_norm": 1.0812251567840576, "learning_rate": 8.289841116995668e-05, "loss": 1.708, "step": 3552 }, { "epoch": 1.7126353790613718, "grad_norm": 1.088507056236267, "learning_rate": 8.2869523350987e-05, "loss": 1.5565, "step": 3558 }, { "epoch": 1.715523465703971, "grad_norm": 1.2732123136520386, "learning_rate": 8.284063553201734e-05, "loss": 1.7935, "step": 3564 }, { "epoch": 1.7184115523465704, "grad_norm": 0.9990867376327515, "learning_rate": 8.281174771304768e-05, "loss": 1.5686, "step": 3570 }, { "epoch": 1.7212996389891697, "grad_norm": 1.0261949300765991, "learning_rate": 8.2782859894078e-05, "loss": 1.7717, "step": 3576 }, { "epoch": 1.724187725631769, "grad_norm": 1.0716211795806885, "learning_rate": 8.275397207510833e-05, "loss": 1.5839, "step": 3582 }, { "epoch": 1.7270758122743683, "grad_norm": 1.0850791931152344, "learning_rate": 8.272508425613867e-05, "loss": 1.754, "step": 3588 }, { "epoch": 1.7299638989169677, "grad_norm": 1.084754228591919, "learning_rate": 8.269619643716899e-05, "loss": 1.7777, "step": 3594 }, { "epoch": 1.7328519855595668, "grad_norm": 1.1335595846176147, "learning_rate": 8.266730861819934e-05, "loss": 1.5596, "step": 3600 }, { "epoch": 1.735740072202166, "grad_norm": 1.1492549180984497, "learning_rate": 8.263842079922967e-05, "loss": 1.7441, "step": 3606 }, { "epoch": 1.7386281588447652, "grad_norm": 1.1537525653839111, "learning_rate": 8.260953298025999e-05, "loss": 1.649, "step": 3612 }, { "epoch": 1.7415162454873645, "grad_norm": 1.0606080293655396, "learning_rate": 8.258064516129033e-05, "loss": 1.6707, "step": 3618 }, { "epoch": 1.7444043321299638, "grad_norm": 1.1513628959655762, "learning_rate": 8.255175734232065e-05, "loss": 1.6525, "step": 3624 }, { "epoch": 1.7472924187725631, "grad_norm": 1.0390511751174927, "learning_rate": 8.252286952335099e-05, "loss": 1.7454, "step": 3630 }, { "epoch": 1.7501805054151625, "grad_norm": 0.9583371877670288, "learning_rate": 8.249398170438132e-05, "loss": 1.7848, "step": 3636 }, { "epoch": 1.7530685920577618, "grad_norm": 1.0754055976867676, "learning_rate": 8.246509388541166e-05, "loss": 1.8745, "step": 3642 }, { "epoch": 1.755956678700361, "grad_norm": 1.07400381565094, "learning_rate": 8.2436206066442e-05, "loss": 1.6418, "step": 3648 }, { "epoch": 1.7588447653429604, "grad_norm": 1.07863450050354, "learning_rate": 8.240731824747232e-05, "loss": 1.8029, "step": 3654 }, { "epoch": 1.7617328519855595, "grad_norm": 1.0401897430419922, "learning_rate": 8.237843042850264e-05, "loss": 1.6719, "step": 3660 }, { "epoch": 1.7646209386281588, "grad_norm": 1.112797498703003, "learning_rate": 8.234954260953298e-05, "loss": 1.7945, "step": 3666 }, { "epoch": 1.7675090252707581, "grad_norm": 1.3708149194717407, "learning_rate": 8.232065479056331e-05, "loss": 1.8079, "step": 3672 }, { "epoch": 1.7703971119133572, "grad_norm": 1.3755298852920532, "learning_rate": 8.229176697159366e-05, "loss": 1.8289, "step": 3678 }, { "epoch": 1.7732851985559566, "grad_norm": 1.259181022644043, "learning_rate": 8.226287915262398e-05, "loss": 1.6958, "step": 3684 }, { "epoch": 1.7761732851985559, "grad_norm": 1.0492124557495117, "learning_rate": 8.223399133365431e-05, "loss": 1.6361, "step": 3690 }, { "epoch": 1.7790613718411552, "grad_norm": 1.1529468297958374, "learning_rate": 8.220510351468465e-05, "loss": 1.7051, "step": 3696 }, { "epoch": 1.7819494584837545, "grad_norm": 1.2096043825149536, "learning_rate": 8.217621569571497e-05, "loss": 1.676, "step": 3702 }, { "epoch": 1.7848375451263538, "grad_norm": 1.0143238306045532, "learning_rate": 8.214732787674531e-05, "loss": 1.7523, "step": 3708 }, { "epoch": 1.7877256317689532, "grad_norm": 1.068037986755371, "learning_rate": 8.211844005777565e-05, "loss": 1.7135, "step": 3714 }, { "epoch": 1.7906137184115525, "grad_norm": 1.2734490633010864, "learning_rate": 8.208955223880597e-05, "loss": 1.7715, "step": 3720 }, { "epoch": 1.7935018050541516, "grad_norm": 1.2356928586959839, "learning_rate": 8.206066441983631e-05, "loss": 1.6567, "step": 3726 }, { "epoch": 1.796389891696751, "grad_norm": 1.1506340503692627, "learning_rate": 8.203177660086664e-05, "loss": 1.7113, "step": 3732 }, { "epoch": 1.7992779783393502, "grad_norm": 1.1448200941085815, "learning_rate": 8.200288878189698e-05, "loss": 1.7475, "step": 3738 }, { "epoch": 1.8021660649819493, "grad_norm": 1.0730384588241577, "learning_rate": 8.19740009629273e-05, "loss": 1.7723, "step": 3744 }, { "epoch": 1.8050541516245486, "grad_norm": 1.154062032699585, "learning_rate": 8.194511314395762e-05, "loss": 1.6937, "step": 3750 }, { "epoch": 1.807942238267148, "grad_norm": 0.9958991408348083, "learning_rate": 8.191622532498798e-05, "loss": 1.8599, "step": 3756 }, { "epoch": 1.8108303249097473, "grad_norm": 1.2138396501541138, "learning_rate": 8.18873375060183e-05, "loss": 1.7354, "step": 3762 }, { "epoch": 1.8137184115523466, "grad_norm": 1.1917569637298584, "learning_rate": 8.185844968704863e-05, "loss": 1.7503, "step": 3768 }, { "epoch": 1.816606498194946, "grad_norm": 1.1662646532058716, "learning_rate": 8.182956186807896e-05, "loss": 1.636, "step": 3774 }, { "epoch": 1.8194945848375452, "grad_norm": 1.1966273784637451, "learning_rate": 8.180067404910929e-05, "loss": 1.6651, "step": 3780 }, { "epoch": 1.8223826714801445, "grad_norm": 1.1080706119537354, "learning_rate": 8.177178623013963e-05, "loss": 1.5983, "step": 3786 }, { "epoch": 1.8252707581227436, "grad_norm": 1.3055298328399658, "learning_rate": 8.174289841116997e-05, "loss": 1.7489, "step": 3792 }, { "epoch": 1.828158844765343, "grad_norm": 1.1526329517364502, "learning_rate": 8.171401059220029e-05, "loss": 1.7338, "step": 3798 }, { "epoch": 1.8310469314079423, "grad_norm": 1.1349650621414185, "learning_rate": 8.168512277323063e-05, "loss": 1.6961, "step": 3804 }, { "epoch": 1.8339350180505414, "grad_norm": 0.9573349356651306, "learning_rate": 8.165623495426095e-05, "loss": 1.6846, "step": 3810 }, { "epoch": 1.8368231046931407, "grad_norm": 1.1585997343063354, "learning_rate": 8.162734713529129e-05, "loss": 1.843, "step": 3816 }, { "epoch": 1.83971119133574, "grad_norm": 1.0836002826690674, "learning_rate": 8.159845931632162e-05, "loss": 1.6475, "step": 3822 }, { "epoch": 1.8425992779783393, "grad_norm": 1.2193304300308228, "learning_rate": 8.156957149735194e-05, "loss": 1.669, "step": 3828 }, { "epoch": 1.8454873646209387, "grad_norm": 1.1085386276245117, "learning_rate": 8.15406836783823e-05, "loss": 1.75, "step": 3834 }, { "epoch": 1.848375451263538, "grad_norm": 1.1439108848571777, "learning_rate": 8.151179585941262e-05, "loss": 1.7212, "step": 3840 }, { "epoch": 1.8512635379061373, "grad_norm": 1.1636065244674683, "learning_rate": 8.148290804044296e-05, "loss": 1.5981, "step": 3846 }, { "epoch": 1.8541516245487366, "grad_norm": 1.1237648725509644, "learning_rate": 8.145402022147328e-05, "loss": 1.8418, "step": 3852 }, { "epoch": 1.8570397111913357, "grad_norm": 1.091884970664978, "learning_rate": 8.142513240250361e-05, "loss": 1.6305, "step": 3858 }, { "epoch": 1.859927797833935, "grad_norm": 1.1279999017715454, "learning_rate": 8.139624458353395e-05, "loss": 1.4908, "step": 3864 }, { "epoch": 1.8628158844765343, "grad_norm": 1.1477504968643188, "learning_rate": 8.136735676456428e-05, "loss": 1.7972, "step": 3870 }, { "epoch": 1.8657039711191334, "grad_norm": 1.090254306793213, "learning_rate": 8.133846894559461e-05, "loss": 1.523, "step": 3876 }, { "epoch": 1.8685920577617328, "grad_norm": 1.1562796831130981, "learning_rate": 8.130958112662495e-05, "loss": 1.8617, "step": 3882 }, { "epoch": 1.871480144404332, "grad_norm": 1.1004424095153809, "learning_rate": 8.128069330765527e-05, "loss": 1.6834, "step": 3888 }, { "epoch": 1.8743682310469314, "grad_norm": 1.1712908744812012, "learning_rate": 8.125180548868561e-05, "loss": 1.8313, "step": 3894 }, { "epoch": 1.8772563176895307, "grad_norm": 1.1347545385360718, "learning_rate": 8.122291766971594e-05, "loss": 1.7099, "step": 3900 }, { "epoch": 1.88014440433213, "grad_norm": 1.0436241626739502, "learning_rate": 8.119402985074627e-05, "loss": 1.6774, "step": 3906 }, { "epoch": 1.8830324909747294, "grad_norm": 1.176244854927063, "learning_rate": 8.116514203177661e-05, "loss": 1.74, "step": 3912 }, { "epoch": 1.8859205776173287, "grad_norm": 1.2282298803329468, "learning_rate": 8.113625421280694e-05, "loss": 1.7166, "step": 3918 }, { "epoch": 1.8888086642599278, "grad_norm": 1.0349675416946411, "learning_rate": 8.110736639383728e-05, "loss": 1.6607, "step": 3924 }, { "epoch": 1.891696750902527, "grad_norm": 1.0478503704071045, "learning_rate": 8.10784785748676e-05, "loss": 1.7279, "step": 3930 }, { "epoch": 1.8945848375451262, "grad_norm": 1.0904722213745117, "learning_rate": 8.104959075589792e-05, "loss": 1.6563, "step": 3936 }, { "epoch": 1.8974729241877255, "grad_norm": 1.0156768560409546, "learning_rate": 8.102070293692826e-05, "loss": 1.7356, "step": 3942 }, { "epoch": 1.9003610108303248, "grad_norm": 1.0968583822250366, "learning_rate": 8.09918151179586e-05, "loss": 1.7761, "step": 3948 }, { "epoch": 1.9032490974729241, "grad_norm": 1.078016996383667, "learning_rate": 8.096292729898894e-05, "loss": 1.7503, "step": 3954 }, { "epoch": 1.9061371841155235, "grad_norm": 1.0822505950927734, "learning_rate": 8.093403948001926e-05, "loss": 1.674, "step": 3960 }, { "epoch": 1.9090252707581228, "grad_norm": 1.0806671380996704, "learning_rate": 8.090515166104959e-05, "loss": 1.7122, "step": 3966 }, { "epoch": 1.911913357400722, "grad_norm": 1.3015724420547485, "learning_rate": 8.087626384207993e-05, "loss": 1.7696, "step": 3972 }, { "epoch": 1.9148014440433214, "grad_norm": 1.287876844406128, "learning_rate": 8.084737602311025e-05, "loss": 1.7275, "step": 3978 }, { "epoch": 1.9176895306859207, "grad_norm": 0.9350766539573669, "learning_rate": 8.081848820414059e-05, "loss": 1.5426, "step": 3984 }, { "epoch": 1.9205776173285198, "grad_norm": 1.0319976806640625, "learning_rate": 8.078960038517093e-05, "loss": 1.6149, "step": 3990 }, { "epoch": 1.9234657039711192, "grad_norm": 0.9956598877906799, "learning_rate": 8.076071256620125e-05, "loss": 1.625, "step": 3996 }, { "epoch": 1.9263537906137183, "grad_norm": 1.1281410455703735, "learning_rate": 8.073182474723159e-05, "loss": 1.6262, "step": 4002 }, { "epoch": 1.9292418772563176, "grad_norm": 1.24624764919281, "learning_rate": 8.070293692826192e-05, "loss": 1.5947, "step": 4008 }, { "epoch": 1.932129963898917, "grad_norm": 1.185560703277588, "learning_rate": 8.067404910929226e-05, "loss": 1.6313, "step": 4014 }, { "epoch": 1.9350180505415162, "grad_norm": 0.9983828067779541, "learning_rate": 8.064516129032258e-05, "loss": 1.6558, "step": 4020 }, { "epoch": 1.9379061371841155, "grad_norm": 1.1473082304000854, "learning_rate": 8.061627347135292e-05, "loss": 1.8027, "step": 4026 }, { "epoch": 1.9407942238267148, "grad_norm": 0.9994457960128784, "learning_rate": 8.058738565238326e-05, "loss": 1.69, "step": 4032 }, { "epoch": 1.9436823104693142, "grad_norm": 1.0906177759170532, "learning_rate": 8.055849783341358e-05, "loss": 1.8217, "step": 4038 }, { "epoch": 1.9465703971119135, "grad_norm": 1.0700476169586182, "learning_rate": 8.052961001444391e-05, "loss": 1.6365, "step": 4044 }, { "epoch": 1.9494584837545126, "grad_norm": 1.1224262714385986, "learning_rate": 8.050072219547425e-05, "loss": 1.7322, "step": 4050 }, { "epoch": 1.952346570397112, "grad_norm": 1.0347113609313965, "learning_rate": 8.047183437650457e-05, "loss": 1.6879, "step": 4056 }, { "epoch": 1.9552346570397112, "grad_norm": 1.154697060585022, "learning_rate": 8.044294655753491e-05, "loss": 1.4966, "step": 4062 }, { "epoch": 1.9581227436823103, "grad_norm": 1.11407470703125, "learning_rate": 8.041405873856525e-05, "loss": 1.67, "step": 4068 }, { "epoch": 1.9610108303249096, "grad_norm": 1.093591570854187, "learning_rate": 8.038517091959557e-05, "loss": 1.6351, "step": 4074 }, { "epoch": 1.963898916967509, "grad_norm": 0.9935932159423828, "learning_rate": 8.035628310062591e-05, "loss": 1.7314, "step": 4080 }, { "epoch": 1.9667870036101083, "grad_norm": 1.095932126045227, "learning_rate": 8.032739528165623e-05, "loss": 1.7439, "step": 4086 }, { "epoch": 1.9696750902527076, "grad_norm": 1.0556317567825317, "learning_rate": 8.029850746268657e-05, "loss": 1.7958, "step": 4092 }, { "epoch": 1.972563176895307, "grad_norm": 1.1100959777832031, "learning_rate": 8.02696196437169e-05, "loss": 1.7097, "step": 4098 }, { "epoch": 1.9754512635379062, "grad_norm": 1.2633126974105835, "learning_rate": 8.024073182474724e-05, "loss": 1.7605, "step": 4104 }, { "epoch": 1.9783393501805056, "grad_norm": 1.178573727607727, "learning_rate": 8.021184400577757e-05, "loss": 1.8145, "step": 4110 }, { "epoch": 1.9812274368231046, "grad_norm": 1.2489348649978638, "learning_rate": 8.01829561868079e-05, "loss": 1.8325, "step": 4116 }, { "epoch": 1.984115523465704, "grad_norm": 0.9644096493721008, "learning_rate": 8.015406836783822e-05, "loss": 1.733, "step": 4122 }, { "epoch": 1.9870036101083033, "grad_norm": 1.1063843965530396, "learning_rate": 8.012518054886856e-05, "loss": 1.6017, "step": 4128 }, { "epoch": 1.9898916967509024, "grad_norm": 1.2632142305374146, "learning_rate": 8.009629272989889e-05, "loss": 1.8401, "step": 4134 }, { "epoch": 1.9927797833935017, "grad_norm": 1.0902996063232422, "learning_rate": 8.006740491092924e-05, "loss": 1.7511, "step": 4140 }, { "epoch": 1.995667870036101, "grad_norm": 1.1063785552978516, "learning_rate": 8.003851709195956e-05, "loss": 1.6363, "step": 4146 }, { "epoch": 1.9985559566787003, "grad_norm": 0.9343324303627014, "learning_rate": 8.000962927298989e-05, "loss": 1.7942, "step": 4152 }, { "epoch": 2.0014440433212997, "grad_norm": 1.1438519954681396, "learning_rate": 7.998074145402023e-05, "loss": 1.75, "step": 4158 }, { "epoch": 2.004332129963899, "grad_norm": 1.2648952007293701, "learning_rate": 7.995185363505055e-05, "loss": 1.6695, "step": 4164 }, { "epoch": 2.0072202166064983, "grad_norm": 1.157775640487671, "learning_rate": 7.992296581608089e-05, "loss": 1.6517, "step": 4170 }, { "epoch": 2.0101083032490976, "grad_norm": 1.195163369178772, "learning_rate": 7.989407799711123e-05, "loss": 1.6822, "step": 4176 }, { "epoch": 2.012996389891697, "grad_norm": 1.1555876731872559, "learning_rate": 7.986519017814155e-05, "loss": 1.4847, "step": 4182 }, { "epoch": 2.015884476534296, "grad_norm": 1.1627624034881592, "learning_rate": 7.983630235917189e-05, "loss": 1.6968, "step": 4188 }, { "epoch": 2.018772563176895, "grad_norm": 1.1089667081832886, "learning_rate": 7.980741454020222e-05, "loss": 1.7204, "step": 4194 }, { "epoch": 2.0216606498194944, "grad_norm": 1.3893928527832031, "learning_rate": 7.977852672123256e-05, "loss": 1.6803, "step": 4200 }, { "epoch": 2.0245487364620938, "grad_norm": 1.0562649965286255, "learning_rate": 7.974963890226288e-05, "loss": 1.6126, "step": 4206 }, { "epoch": 2.027436823104693, "grad_norm": 1.1291148662567139, "learning_rate": 7.97207510832932e-05, "loss": 1.7663, "step": 4212 }, { "epoch": 2.0303249097472924, "grad_norm": 1.1951361894607544, "learning_rate": 7.969186326432356e-05, "loss": 1.6993, "step": 4218 }, { "epoch": 2.0332129963898917, "grad_norm": 1.2842034101486206, "learning_rate": 7.966297544535388e-05, "loss": 1.6134, "step": 4224 }, { "epoch": 2.036101083032491, "grad_norm": 1.1778056621551514, "learning_rate": 7.96340876263842e-05, "loss": 1.5604, "step": 4230 }, { "epoch": 2.0389891696750904, "grad_norm": 1.306934118270874, "learning_rate": 7.960519980741455e-05, "loss": 1.7132, "step": 4236 }, { "epoch": 2.0418772563176897, "grad_norm": 1.1847249269485474, "learning_rate": 7.957631198844487e-05, "loss": 1.4934, "step": 4242 }, { "epoch": 2.044765342960289, "grad_norm": 1.2349610328674316, "learning_rate": 7.954742416947521e-05, "loss": 1.5731, "step": 4248 }, { "epoch": 2.047653429602888, "grad_norm": 1.2912768125534058, "learning_rate": 7.951853635050555e-05, "loss": 1.597, "step": 4254 }, { "epoch": 2.050541516245487, "grad_norm": 1.2363039255142212, "learning_rate": 7.948964853153587e-05, "loss": 1.6562, "step": 4260 }, { "epoch": 2.0534296028880865, "grad_norm": 1.2619967460632324, "learning_rate": 7.946076071256621e-05, "loss": 1.6379, "step": 4266 }, { "epoch": 2.056317689530686, "grad_norm": 1.4596139192581177, "learning_rate": 7.943187289359653e-05, "loss": 1.5272, "step": 4272 }, { "epoch": 2.059205776173285, "grad_norm": 1.1500120162963867, "learning_rate": 7.940298507462687e-05, "loss": 1.4515, "step": 4278 }, { "epoch": 2.0620938628158845, "grad_norm": 1.223166823387146, "learning_rate": 7.93740972556572e-05, "loss": 1.5532, "step": 4284 }, { "epoch": 2.064981949458484, "grad_norm": 1.3942859172821045, "learning_rate": 7.934520943668752e-05, "loss": 1.529, "step": 4290 }, { "epoch": 2.067870036101083, "grad_norm": 1.349802017211914, "learning_rate": 7.931632161771787e-05, "loss": 1.6175, "step": 4296 }, { "epoch": 2.0707581227436824, "grad_norm": 1.394556999206543, "learning_rate": 7.92874337987482e-05, "loss": 1.7616, "step": 4302 }, { "epoch": 2.0736462093862817, "grad_norm": 1.2381508350372314, "learning_rate": 7.925854597977854e-05, "loss": 1.73, "step": 4308 }, { "epoch": 2.076534296028881, "grad_norm": 1.226277470588684, "learning_rate": 7.922965816080886e-05, "loss": 1.6947, "step": 4314 }, { "epoch": 2.07942238267148, "grad_norm": 1.110539436340332, "learning_rate": 7.920077034183919e-05, "loss": 1.4744, "step": 4320 }, { "epoch": 2.0823104693140793, "grad_norm": 1.1478859186172485, "learning_rate": 7.917188252286953e-05, "loss": 1.5649, "step": 4326 }, { "epoch": 2.0851985559566786, "grad_norm": 1.3074700832366943, "learning_rate": 7.914299470389986e-05, "loss": 1.63, "step": 4332 }, { "epoch": 2.088086642599278, "grad_norm": 1.345037817955017, "learning_rate": 7.911410688493019e-05, "loss": 1.6094, "step": 4338 }, { "epoch": 2.090974729241877, "grad_norm": 1.1591620445251465, "learning_rate": 7.908521906596053e-05, "loss": 1.5836, "step": 4344 }, { "epoch": 2.0938628158844765, "grad_norm": 1.1762259006500244, "learning_rate": 7.905633124699085e-05, "loss": 1.5235, "step": 4350 }, { "epoch": 2.096750902527076, "grad_norm": 1.1976076364517212, "learning_rate": 7.902744342802119e-05, "loss": 1.5648, "step": 4356 }, { "epoch": 2.099638989169675, "grad_norm": 1.2209484577178955, "learning_rate": 7.899855560905152e-05, "loss": 1.5787, "step": 4362 }, { "epoch": 2.1025270758122745, "grad_norm": 1.134056568145752, "learning_rate": 7.896966779008185e-05, "loss": 1.4751, "step": 4368 }, { "epoch": 2.105415162454874, "grad_norm": 1.2926255464553833, "learning_rate": 7.894077997111219e-05, "loss": 1.5271, "step": 4374 }, { "epoch": 2.108303249097473, "grad_norm": 1.137678861618042, "learning_rate": 7.891189215214252e-05, "loss": 1.7616, "step": 4380 }, { "epoch": 2.111191335740072, "grad_norm": 1.2014248371124268, "learning_rate": 7.888300433317286e-05, "loss": 1.6895, "step": 4386 }, { "epoch": 2.1140794223826713, "grad_norm": 1.2278836965560913, "learning_rate": 7.885411651420318e-05, "loss": 1.7106, "step": 4392 }, { "epoch": 2.1169675090252706, "grad_norm": 1.2308661937713623, "learning_rate": 7.88252286952335e-05, "loss": 1.7689, "step": 4398 }, { "epoch": 2.11985559566787, "grad_norm": 1.4774951934814453, "learning_rate": 7.879634087626384e-05, "loss": 1.4952, "step": 4404 }, { "epoch": 2.1227436823104693, "grad_norm": 1.3602070808410645, "learning_rate": 7.876745305729418e-05, "loss": 1.6509, "step": 4410 }, { "epoch": 2.1256317689530686, "grad_norm": 1.3164474964141846, "learning_rate": 7.87385652383245e-05, "loss": 1.8139, "step": 4416 }, { "epoch": 2.128519855595668, "grad_norm": 1.2632802724838257, "learning_rate": 7.870967741935484e-05, "loss": 1.6474, "step": 4422 }, { "epoch": 2.1314079422382672, "grad_norm": 1.1633061170578003, "learning_rate": 7.868078960038517e-05, "loss": 1.8168, "step": 4428 }, { "epoch": 2.1342960288808666, "grad_norm": 1.203873634338379, "learning_rate": 7.865190178141551e-05, "loss": 1.5914, "step": 4434 }, { "epoch": 2.137184115523466, "grad_norm": 1.043006181716919, "learning_rate": 7.862301396244583e-05, "loss": 1.5726, "step": 4440 }, { "epoch": 2.140072202166065, "grad_norm": 1.362917423248291, "learning_rate": 7.859412614347617e-05, "loss": 1.5982, "step": 4446 }, { "epoch": 2.142960288808664, "grad_norm": 1.2041537761688232, "learning_rate": 7.856523832450651e-05, "loss": 1.4713, "step": 4452 }, { "epoch": 2.1458483754512634, "grad_norm": 1.0579909086227417, "learning_rate": 7.853635050553683e-05, "loss": 1.5698, "step": 4458 }, { "epoch": 2.1487364620938627, "grad_norm": 1.3498187065124512, "learning_rate": 7.850746268656717e-05, "loss": 1.717, "step": 4464 }, { "epoch": 2.151624548736462, "grad_norm": 1.4977130889892578, "learning_rate": 7.84785748675975e-05, "loss": 1.5898, "step": 4470 }, { "epoch": 2.1545126353790613, "grad_norm": 1.1892290115356445, "learning_rate": 7.844968704862782e-05, "loss": 1.7988, "step": 4476 }, { "epoch": 2.1574007220216607, "grad_norm": 1.1984620094299316, "learning_rate": 7.842079922965816e-05, "loss": 1.7374, "step": 4482 }, { "epoch": 2.16028880866426, "grad_norm": 1.3232879638671875, "learning_rate": 7.83919114106885e-05, "loss": 1.554, "step": 4488 }, { "epoch": 2.1631768953068593, "grad_norm": 1.0719797611236572, "learning_rate": 7.836302359171884e-05, "loss": 1.696, "step": 4494 }, { "epoch": 2.1660649819494586, "grad_norm": 1.0943427085876465, "learning_rate": 7.833413577274916e-05, "loss": 1.5985, "step": 4500 }, { "epoch": 2.168953068592058, "grad_norm": 1.1994340419769287, "learning_rate": 7.830524795377949e-05, "loss": 1.596, "step": 4506 }, { "epoch": 2.171841155234657, "grad_norm": 1.276389718055725, "learning_rate": 7.827636013480983e-05, "loss": 1.6607, "step": 4512 }, { "epoch": 2.174729241877256, "grad_norm": 1.394392490386963, "learning_rate": 7.824747231584015e-05, "loss": 1.5467, "step": 4518 }, { "epoch": 2.1776173285198555, "grad_norm": 1.184099793434143, "learning_rate": 7.821858449687049e-05, "loss": 1.6001, "step": 4524 }, { "epoch": 2.1805054151624548, "grad_norm": 1.1102205514907837, "learning_rate": 7.818969667790083e-05, "loss": 1.6193, "step": 4530 }, { "epoch": 2.183393501805054, "grad_norm": 1.205987572669983, "learning_rate": 7.816080885893115e-05, "loss": 1.5394, "step": 4536 }, { "epoch": 2.1862815884476534, "grad_norm": 1.1822541952133179, "learning_rate": 7.813192103996149e-05, "loss": 1.4661, "step": 4542 }, { "epoch": 2.1891696750902527, "grad_norm": 1.1073379516601562, "learning_rate": 7.810303322099182e-05, "loss": 1.6141, "step": 4548 }, { "epoch": 2.192057761732852, "grad_norm": 1.101296305656433, "learning_rate": 7.807414540202215e-05, "loss": 1.4932, "step": 4554 }, { "epoch": 2.1949458483754514, "grad_norm": 1.3230780363082886, "learning_rate": 7.804525758305248e-05, "loss": 1.5654, "step": 4560 }, { "epoch": 2.1978339350180507, "grad_norm": 1.2974190711975098, "learning_rate": 7.801636976408282e-05, "loss": 1.6838, "step": 4566 }, { "epoch": 2.20072202166065, "grad_norm": 1.3460841178894043, "learning_rate": 7.798748194511316e-05, "loss": 1.6584, "step": 4572 }, { "epoch": 2.2036101083032493, "grad_norm": 1.2820940017700195, "learning_rate": 7.795859412614348e-05, "loss": 1.5599, "step": 4578 }, { "epoch": 2.206498194945848, "grad_norm": 1.424225926399231, "learning_rate": 7.79297063071738e-05, "loss": 1.6186, "step": 4584 }, { "epoch": 2.2093862815884475, "grad_norm": 1.4198392629623413, "learning_rate": 7.790081848820414e-05, "loss": 1.6396, "step": 4590 }, { "epoch": 2.212274368231047, "grad_norm": 1.1684870719909668, "learning_rate": 7.787193066923447e-05, "loss": 1.6655, "step": 4596 }, { "epoch": 2.215162454873646, "grad_norm": 1.0981298685073853, "learning_rate": 7.784304285026482e-05, "loss": 1.6641, "step": 4602 }, { "epoch": 2.2180505415162455, "grad_norm": 1.3571112155914307, "learning_rate": 7.781415503129514e-05, "loss": 1.5771, "step": 4608 }, { "epoch": 2.220938628158845, "grad_norm": 1.2179030179977417, "learning_rate": 7.778526721232547e-05, "loss": 1.6925, "step": 4614 }, { "epoch": 2.223826714801444, "grad_norm": 1.3599469661712646, "learning_rate": 7.775637939335581e-05, "loss": 1.5381, "step": 4620 }, { "epoch": 2.2267148014440434, "grad_norm": 1.2987323999404907, "learning_rate": 7.772749157438613e-05, "loss": 1.6514, "step": 4626 }, { "epoch": 2.2296028880866428, "grad_norm": 1.2273507118225098, "learning_rate": 7.769860375541647e-05, "loss": 1.6317, "step": 4632 }, { "epoch": 2.232490974729242, "grad_norm": 1.2115448713302612, "learning_rate": 7.766971593644681e-05, "loss": 1.6433, "step": 4638 }, { "epoch": 2.235379061371841, "grad_norm": 1.1618566513061523, "learning_rate": 7.764082811747713e-05, "loss": 1.6355, "step": 4644 }, { "epoch": 2.2382671480144403, "grad_norm": 1.1803876161575317, "learning_rate": 7.761194029850747e-05, "loss": 1.6649, "step": 4650 }, { "epoch": 2.2411552346570396, "grad_norm": 1.2069780826568604, "learning_rate": 7.75830524795378e-05, "loss": 1.6584, "step": 4656 }, { "epoch": 2.244043321299639, "grad_norm": 1.25930655002594, "learning_rate": 7.755416466056814e-05, "loss": 1.6516, "step": 4662 }, { "epoch": 2.2469314079422382, "grad_norm": 1.2072900533676147, "learning_rate": 7.752527684159846e-05, "loss": 1.5028, "step": 4668 }, { "epoch": 2.2498194945848375, "grad_norm": 1.15311861038208, "learning_rate": 7.749638902262879e-05, "loss": 1.6321, "step": 4674 }, { "epoch": 2.252707581227437, "grad_norm": 1.5071076154708862, "learning_rate": 7.746750120365914e-05, "loss": 1.7516, "step": 4680 }, { "epoch": 2.255595667870036, "grad_norm": 1.3569972515106201, "learning_rate": 7.743861338468946e-05, "loss": 1.4522, "step": 4686 }, { "epoch": 2.2584837545126355, "grad_norm": 1.1228210926055908, "learning_rate": 7.740972556571979e-05, "loss": 1.5638, "step": 4692 }, { "epoch": 2.261371841155235, "grad_norm": 1.3490697145462036, "learning_rate": 7.738083774675013e-05, "loss": 1.6415, "step": 4698 }, { "epoch": 2.264259927797834, "grad_norm": 1.2531455755233765, "learning_rate": 7.735194992778045e-05, "loss": 1.5478, "step": 4704 }, { "epoch": 2.2671480144404335, "grad_norm": 1.3723077774047852, "learning_rate": 7.732306210881079e-05, "loss": 1.6667, "step": 4710 }, { "epoch": 2.2700361010830323, "grad_norm": 1.2446554899215698, "learning_rate": 7.729417428984113e-05, "loss": 1.6147, "step": 4716 }, { "epoch": 2.2729241877256317, "grad_norm": 1.336427092552185, "learning_rate": 7.726528647087145e-05, "loss": 1.5869, "step": 4722 }, { "epoch": 2.275812274368231, "grad_norm": 1.3884146213531494, "learning_rate": 7.723639865190179e-05, "loss": 1.5991, "step": 4728 }, { "epoch": 2.2787003610108303, "grad_norm": 1.3205794095993042, "learning_rate": 7.720751083293212e-05, "loss": 1.7311, "step": 4734 }, { "epoch": 2.2815884476534296, "grad_norm": 1.229244589805603, "learning_rate": 7.717862301396245e-05, "loss": 1.6598, "step": 4740 }, { "epoch": 2.284476534296029, "grad_norm": 1.1891183853149414, "learning_rate": 7.714973519499278e-05, "loss": 1.6262, "step": 4746 }, { "epoch": 2.2873646209386282, "grad_norm": 1.3675397634506226, "learning_rate": 7.71208473760231e-05, "loss": 1.818, "step": 4752 }, { "epoch": 2.2902527075812276, "grad_norm": 1.458994746208191, "learning_rate": 7.709195955705345e-05, "loss": 1.7083, "step": 4758 }, { "epoch": 2.293140794223827, "grad_norm": 1.2246507406234741, "learning_rate": 7.706307173808378e-05, "loss": 1.5738, "step": 4764 }, { "epoch": 2.2960288808664258, "grad_norm": 1.2233506441116333, "learning_rate": 7.70341839191141e-05, "loss": 1.6291, "step": 4770 }, { "epoch": 2.298916967509025, "grad_norm": 1.0909838676452637, "learning_rate": 7.700529610014444e-05, "loss": 1.5458, "step": 4776 }, { "epoch": 2.3018050541516244, "grad_norm": 1.3471211194992065, "learning_rate": 7.697640828117477e-05, "loss": 1.7282, "step": 4782 }, { "epoch": 2.3046931407942237, "grad_norm": 1.2401286363601685, "learning_rate": 7.69475204622051e-05, "loss": 1.5071, "step": 4788 }, { "epoch": 2.307581227436823, "grad_norm": 1.1695021390914917, "learning_rate": 7.691863264323544e-05, "loss": 1.5697, "step": 4794 }, { "epoch": 2.3104693140794224, "grad_norm": 1.1990652084350586, "learning_rate": 7.688974482426577e-05, "loss": 1.52, "step": 4800 }, { "epoch": 2.3133574007220217, "grad_norm": 1.1849806308746338, "learning_rate": 7.686085700529611e-05, "loss": 1.6041, "step": 4806 }, { "epoch": 2.316245487364621, "grad_norm": 1.389315128326416, "learning_rate": 7.683196918632643e-05, "loss": 1.6196, "step": 4812 }, { "epoch": 2.3191335740072203, "grad_norm": 1.4406752586364746, "learning_rate": 7.680308136735677e-05, "loss": 1.7943, "step": 4818 }, { "epoch": 2.3220216606498196, "grad_norm": 1.5369993448257446, "learning_rate": 7.67741935483871e-05, "loss": 1.4947, "step": 4824 }, { "epoch": 2.324909747292419, "grad_norm": 1.3118836879730225, "learning_rate": 7.674530572941742e-05, "loss": 1.5228, "step": 4830 }, { "epoch": 2.3277978339350183, "grad_norm": 1.1526684761047363, "learning_rate": 7.671641791044777e-05, "loss": 1.5266, "step": 4836 }, { "epoch": 2.330685920577617, "grad_norm": 1.265730857849121, "learning_rate": 7.66875300914781e-05, "loss": 1.5851, "step": 4842 }, { "epoch": 2.3335740072202165, "grad_norm": 1.272842526435852, "learning_rate": 7.665864227250844e-05, "loss": 1.76, "step": 4848 }, { "epoch": 2.336462093862816, "grad_norm": 1.2320207357406616, "learning_rate": 7.662975445353876e-05, "loss": 1.6304, "step": 4854 }, { "epoch": 2.339350180505415, "grad_norm": 1.3562240600585938, "learning_rate": 7.660086663456909e-05, "loss": 1.63, "step": 4860 }, { "epoch": 2.3422382671480144, "grad_norm": 1.1847113370895386, "learning_rate": 7.657197881559942e-05, "loss": 1.6142, "step": 4866 }, { "epoch": 2.3451263537906137, "grad_norm": 1.1257132291793823, "learning_rate": 7.654309099662976e-05, "loss": 1.5736, "step": 4872 }, { "epoch": 2.348014440433213, "grad_norm": 1.3064936399459839, "learning_rate": 7.651420317766009e-05, "loss": 1.603, "step": 4878 }, { "epoch": 2.3509025270758124, "grad_norm": 1.1850717067718506, "learning_rate": 7.648531535869043e-05, "loss": 1.5756, "step": 4884 }, { "epoch": 2.3537906137184117, "grad_norm": 1.2616254091262817, "learning_rate": 7.645642753972075e-05, "loss": 1.5929, "step": 4890 }, { "epoch": 2.356678700361011, "grad_norm": 1.2740598917007446, "learning_rate": 7.642753972075109e-05, "loss": 1.5728, "step": 4896 }, { "epoch": 2.35956678700361, "grad_norm": 1.1042790412902832, "learning_rate": 7.639865190178141e-05, "loss": 1.4185, "step": 4902 }, { "epoch": 2.362454873646209, "grad_norm": 1.224420428276062, "learning_rate": 7.636976408281175e-05, "loss": 1.6455, "step": 4908 }, { "epoch": 2.3653429602888085, "grad_norm": 1.2900187969207764, "learning_rate": 7.634087626384209e-05, "loss": 1.678, "step": 4914 }, { "epoch": 2.368231046931408, "grad_norm": 1.372710943222046, "learning_rate": 7.631198844487241e-05, "loss": 1.7301, "step": 4920 }, { "epoch": 2.371119133574007, "grad_norm": 1.4258843660354614, "learning_rate": 7.628310062590275e-05, "loss": 1.7446, "step": 4926 }, { "epoch": 2.3740072202166065, "grad_norm": 1.386684536933899, "learning_rate": 7.625421280693308e-05, "loss": 1.6287, "step": 4932 }, { "epoch": 2.376895306859206, "grad_norm": 1.326601266860962, "learning_rate": 7.62253249879634e-05, "loss": 1.674, "step": 4938 }, { "epoch": 2.379783393501805, "grad_norm": 1.24738347530365, "learning_rate": 7.619643716899374e-05, "loss": 1.5161, "step": 4944 }, { "epoch": 2.3826714801444044, "grad_norm": 1.15766441822052, "learning_rate": 7.616754935002408e-05, "loss": 1.6868, "step": 4950 }, { "epoch": 2.3855595667870038, "grad_norm": 1.292614221572876, "learning_rate": 7.613866153105442e-05, "loss": 1.6901, "step": 4956 }, { "epoch": 2.388447653429603, "grad_norm": 1.143007755279541, "learning_rate": 7.610977371208474e-05, "loss": 1.538, "step": 4962 }, { "epoch": 2.3913357400722024, "grad_norm": 1.186502456665039, "learning_rate": 7.608088589311507e-05, "loss": 1.5079, "step": 4968 }, { "epoch": 2.3942238267148013, "grad_norm": 1.1861786842346191, "learning_rate": 7.60519980741454e-05, "loss": 1.6263, "step": 4974 }, { "epoch": 2.3971119133574006, "grad_norm": 1.1094098091125488, "learning_rate": 7.602311025517573e-05, "loss": 1.5522, "step": 4980 }, { "epoch": 2.4, "grad_norm": 1.4490925073623657, "learning_rate": 7.599422243620607e-05, "loss": 1.5524, "step": 4986 }, { "epoch": 2.4028880866425992, "grad_norm": 1.348564624786377, "learning_rate": 7.596533461723641e-05, "loss": 1.6943, "step": 4992 }, { "epoch": 2.4057761732851985, "grad_norm": 1.4004294872283936, "learning_rate": 7.593644679826673e-05, "loss": 1.591, "step": 4998 }, { "epoch": 2.408664259927798, "grad_norm": 1.3113765716552734, "learning_rate": 7.590755897929707e-05, "loss": 1.6834, "step": 5004 }, { "epoch": 2.411552346570397, "grad_norm": 1.082801103591919, "learning_rate": 7.58786711603274e-05, "loss": 1.4812, "step": 5010 }, { "epoch": 2.4144404332129965, "grad_norm": 1.180828332901001, "learning_rate": 7.584978334135773e-05, "loss": 1.6264, "step": 5016 }, { "epoch": 2.417328519855596, "grad_norm": 1.5142802000045776, "learning_rate": 7.582089552238806e-05, "loss": 1.5092, "step": 5022 }, { "epoch": 2.4202166064981947, "grad_norm": 1.3384352922439575, "learning_rate": 7.57920077034184e-05, "loss": 1.6612, "step": 5028 }, { "epoch": 2.423104693140794, "grad_norm": 1.5020110607147217, "learning_rate": 7.576311988444874e-05, "loss": 1.489, "step": 5034 }, { "epoch": 2.4259927797833933, "grad_norm": 1.3337640762329102, "learning_rate": 7.573423206547906e-05, "loss": 1.5282, "step": 5040 }, { "epoch": 2.4288808664259927, "grad_norm": 1.255208134651184, "learning_rate": 7.570534424650939e-05, "loss": 1.4907, "step": 5046 }, { "epoch": 2.431768953068592, "grad_norm": 1.3977291584014893, "learning_rate": 7.567645642753972e-05, "loss": 1.698, "step": 5052 }, { "epoch": 2.4346570397111913, "grad_norm": 1.219262719154358, "learning_rate": 7.564756860857005e-05, "loss": 1.5961, "step": 5058 }, { "epoch": 2.4375451263537906, "grad_norm": 1.2772104740142822, "learning_rate": 7.56186807896004e-05, "loss": 1.6224, "step": 5064 }, { "epoch": 2.44043321299639, "grad_norm": 1.1492831707000732, "learning_rate": 7.558979297063073e-05, "loss": 1.63, "step": 5070 }, { "epoch": 2.4433212996389893, "grad_norm": 1.2771663665771484, "learning_rate": 7.556090515166105e-05, "loss": 1.6917, "step": 5076 }, { "epoch": 2.4462093862815886, "grad_norm": 1.322731614112854, "learning_rate": 7.553201733269139e-05, "loss": 1.8313, "step": 5082 }, { "epoch": 2.449097472924188, "grad_norm": 1.3147132396697998, "learning_rate": 7.550312951372171e-05, "loss": 1.6955, "step": 5088 }, { "epoch": 2.451985559566787, "grad_norm": 1.3510420322418213, "learning_rate": 7.547424169475205e-05, "loss": 1.4538, "step": 5094 }, { "epoch": 2.4548736462093865, "grad_norm": 1.2206655740737915, "learning_rate": 7.544535387578239e-05, "loss": 1.5752, "step": 5100 }, { "epoch": 2.4577617328519854, "grad_norm": 1.3175653219223022, "learning_rate": 7.541646605681271e-05, "loss": 1.7799, "step": 5106 }, { "epoch": 2.4606498194945847, "grad_norm": 1.2858401536941528, "learning_rate": 7.538757823784305e-05, "loss": 1.6604, "step": 5112 }, { "epoch": 2.463537906137184, "grad_norm": 1.3517030477523804, "learning_rate": 7.535869041887338e-05, "loss": 1.5863, "step": 5118 }, { "epoch": 2.4664259927797834, "grad_norm": 1.0484105348587036, "learning_rate": 7.53298025999037e-05, "loss": 1.4965, "step": 5124 }, { "epoch": 2.4693140794223827, "grad_norm": 1.1144850254058838, "learning_rate": 7.530091478093404e-05, "loss": 1.8233, "step": 5130 }, { "epoch": 2.472202166064982, "grad_norm": 1.3140604496002197, "learning_rate": 7.527202696196437e-05, "loss": 1.5329, "step": 5136 }, { "epoch": 2.4750902527075813, "grad_norm": 1.4496768712997437, "learning_rate": 7.524313914299472e-05, "loss": 1.7751, "step": 5142 }, { "epoch": 2.4779783393501806, "grad_norm": 1.388692855834961, "learning_rate": 7.521425132402504e-05, "loss": 1.5578, "step": 5148 }, { "epoch": 2.48086642599278, "grad_norm": 1.5865966081619263, "learning_rate": 7.518536350505537e-05, "loss": 1.5705, "step": 5154 }, { "epoch": 2.483754512635379, "grad_norm": 1.154073715209961, "learning_rate": 7.51564756860857e-05, "loss": 1.5505, "step": 5160 }, { "epoch": 2.486642599277978, "grad_norm": 1.1993553638458252, "learning_rate": 7.512758786711603e-05, "loss": 1.6222, "step": 5166 }, { "epoch": 2.4895306859205775, "grad_norm": 1.4118672609329224, "learning_rate": 7.509870004814637e-05, "loss": 1.7492, "step": 5172 }, { "epoch": 2.492418772563177, "grad_norm": 1.1973785161972046, "learning_rate": 7.506981222917671e-05, "loss": 1.5083, "step": 5178 }, { "epoch": 2.495306859205776, "grad_norm": 1.3421173095703125, "learning_rate": 7.504092441020703e-05, "loss": 1.6671, "step": 5184 }, { "epoch": 2.4981949458483754, "grad_norm": 1.30620276927948, "learning_rate": 7.501203659123737e-05, "loss": 1.5303, "step": 5190 }, { "epoch": 2.5010830324909747, "grad_norm": 1.1590051651000977, "learning_rate": 7.49831487722677e-05, "loss": 1.7932, "step": 5196 }, { "epoch": 2.503971119133574, "grad_norm": 1.36478853225708, "learning_rate": 7.495426095329803e-05, "loss": 1.7876, "step": 5202 }, { "epoch": 2.5068592057761734, "grad_norm": 1.346425175666809, "learning_rate": 7.492537313432836e-05, "loss": 1.6745, "step": 5208 }, { "epoch": 2.5097472924187727, "grad_norm": 1.2543796300888062, "learning_rate": 7.489648531535868e-05, "loss": 1.6298, "step": 5214 }, { "epoch": 2.512635379061372, "grad_norm": 1.3284118175506592, "learning_rate": 7.486759749638904e-05, "loss": 1.6211, "step": 5220 }, { "epoch": 2.5155234657039713, "grad_norm": 1.2074249982833862, "learning_rate": 7.483870967741936e-05, "loss": 1.5981, "step": 5226 }, { "epoch": 2.5184115523465707, "grad_norm": 1.3378170728683472, "learning_rate": 7.480982185844968e-05, "loss": 1.7754, "step": 5232 }, { "epoch": 2.5212996389891695, "grad_norm": 1.3198719024658203, "learning_rate": 7.478093403948002e-05, "loss": 1.5556, "step": 5238 }, { "epoch": 2.524187725631769, "grad_norm": 1.230972409248352, "learning_rate": 7.475204622051035e-05, "loss": 1.5516, "step": 5244 }, { "epoch": 2.527075812274368, "grad_norm": 1.551426649093628, "learning_rate": 7.472315840154069e-05, "loss": 1.6522, "step": 5250 }, { "epoch": 2.5299638989169675, "grad_norm": 1.327309012413025, "learning_rate": 7.469427058257102e-05, "loss": 1.4459, "step": 5256 }, { "epoch": 2.532851985559567, "grad_norm": 1.2913554906845093, "learning_rate": 7.466538276360135e-05, "loss": 1.5076, "step": 5262 }, { "epoch": 2.535740072202166, "grad_norm": 1.1913667917251587, "learning_rate": 7.463649494463169e-05, "loss": 1.6733, "step": 5268 }, { "epoch": 2.5386281588447654, "grad_norm": 1.493749976158142, "learning_rate": 7.460760712566201e-05, "loss": 1.5806, "step": 5274 }, { "epoch": 2.5415162454873648, "grad_norm": 1.4761329889297485, "learning_rate": 7.457871930669235e-05, "loss": 1.7495, "step": 5280 }, { "epoch": 2.5444043321299636, "grad_norm": 1.2897906303405762, "learning_rate": 7.454983148772268e-05, "loss": 1.5485, "step": 5286 }, { "epoch": 2.547292418772563, "grad_norm": 1.251991868019104, "learning_rate": 7.452094366875301e-05, "loss": 1.6951, "step": 5292 }, { "epoch": 2.5501805054151623, "grad_norm": 1.2617682218551636, "learning_rate": 7.449205584978335e-05, "loss": 1.5836, "step": 5298 }, { "epoch": 2.5530685920577616, "grad_norm": 1.217786431312561, "learning_rate": 7.446316803081368e-05, "loss": 1.6093, "step": 5304 }, { "epoch": 2.555956678700361, "grad_norm": 1.31692636013031, "learning_rate": 7.443428021184402e-05, "loss": 1.5695, "step": 5310 }, { "epoch": 2.5588447653429602, "grad_norm": 1.4154953956604004, "learning_rate": 7.440539239287434e-05, "loss": 1.7063, "step": 5316 }, { "epoch": 2.5617328519855596, "grad_norm": 1.3738391399383545, "learning_rate": 7.437650457390467e-05, "loss": 1.5702, "step": 5322 }, { "epoch": 2.564620938628159, "grad_norm": 1.416218638420105, "learning_rate": 7.4347616754935e-05, "loss": 1.6785, "step": 5328 }, { "epoch": 2.567509025270758, "grad_norm": 1.4179434776306152, "learning_rate": 7.431872893596534e-05, "loss": 1.6065, "step": 5334 }, { "epoch": 2.5703971119133575, "grad_norm": 1.0944335460662842, "learning_rate": 7.428984111699567e-05, "loss": 1.5835, "step": 5340 }, { "epoch": 2.573285198555957, "grad_norm": 1.5338119268417358, "learning_rate": 7.4260953298026e-05, "loss": 1.5217, "step": 5346 }, { "epoch": 2.576173285198556, "grad_norm": 1.2250339984893799, "learning_rate": 7.423206547905633e-05, "loss": 1.7028, "step": 5352 }, { "epoch": 2.5790613718411555, "grad_norm": 1.5273693799972534, "learning_rate": 7.420317766008667e-05, "loss": 1.5176, "step": 5358 }, { "epoch": 2.581949458483755, "grad_norm": 1.4212782382965088, "learning_rate": 7.4174289841117e-05, "loss": 1.7383, "step": 5364 }, { "epoch": 2.5848375451263537, "grad_norm": 1.3202974796295166, "learning_rate": 7.414540202214733e-05, "loss": 1.4809, "step": 5370 }, { "epoch": 2.587725631768953, "grad_norm": 1.2213678359985352, "learning_rate": 7.411651420317767e-05, "loss": 1.5184, "step": 5376 }, { "epoch": 2.5906137184115523, "grad_norm": 1.4411945343017578, "learning_rate": 7.4087626384208e-05, "loss": 1.6339, "step": 5382 }, { "epoch": 2.5935018050541516, "grad_norm": 1.2993313074111938, "learning_rate": 7.405873856523833e-05, "loss": 1.6329, "step": 5388 }, { "epoch": 2.596389891696751, "grad_norm": 1.4830660820007324, "learning_rate": 7.402985074626866e-05, "loss": 1.6706, "step": 5394 }, { "epoch": 2.5992779783393503, "grad_norm": 1.219641089439392, "learning_rate": 7.400096292729898e-05, "loss": 1.5362, "step": 5400 }, { "epoch": 2.6021660649819496, "grad_norm": 1.321448564529419, "learning_rate": 7.397207510832932e-05, "loss": 1.5515, "step": 5406 }, { "epoch": 2.605054151624549, "grad_norm": 1.5268959999084473, "learning_rate": 7.394318728935966e-05, "loss": 1.6884, "step": 5412 }, { "epoch": 2.6079422382671478, "grad_norm": 1.310424566268921, "learning_rate": 7.391429947039e-05, "loss": 1.7589, "step": 5418 }, { "epoch": 2.610830324909747, "grad_norm": 1.2116687297821045, "learning_rate": 7.388541165142032e-05, "loss": 1.5911, "step": 5424 }, { "epoch": 2.6137184115523464, "grad_norm": 1.214961290359497, "learning_rate": 7.385652383245065e-05, "loss": 1.6176, "step": 5430 }, { "epoch": 2.6166064981949457, "grad_norm": 1.1844496726989746, "learning_rate": 7.382763601348099e-05, "loss": 1.5616, "step": 5436 }, { "epoch": 2.619494584837545, "grad_norm": 1.3806785345077515, "learning_rate": 7.379874819451131e-05, "loss": 1.5609, "step": 5442 }, { "epoch": 2.6223826714801444, "grad_norm": 1.1580435037612915, "learning_rate": 7.376986037554165e-05, "loss": 1.6087, "step": 5448 }, { "epoch": 2.6252707581227437, "grad_norm": 1.1602493524551392, "learning_rate": 7.374097255657199e-05, "loss": 1.7464, "step": 5454 }, { "epoch": 2.628158844765343, "grad_norm": 1.3840891122817993, "learning_rate": 7.371208473760231e-05, "loss": 1.5239, "step": 5460 }, { "epoch": 2.6310469314079423, "grad_norm": 1.3290904760360718, "learning_rate": 7.368319691863265e-05, "loss": 1.5744, "step": 5466 }, { "epoch": 2.6339350180505416, "grad_norm": 1.3490746021270752, "learning_rate": 7.365430909966298e-05, "loss": 1.5793, "step": 5472 }, { "epoch": 2.636823104693141, "grad_norm": 1.405853509902954, "learning_rate": 7.362542128069331e-05, "loss": 1.5953, "step": 5478 }, { "epoch": 2.6397111913357403, "grad_norm": 1.3156434297561646, "learning_rate": 7.359653346172364e-05, "loss": 1.6069, "step": 5484 }, { "epoch": 2.6425992779783396, "grad_norm": 1.1551438570022583, "learning_rate": 7.356764564275398e-05, "loss": 1.5401, "step": 5490 }, { "epoch": 2.6454873646209385, "grad_norm": 1.4044383764266968, "learning_rate": 7.353875782378432e-05, "loss": 1.6629, "step": 5496 }, { "epoch": 2.648375451263538, "grad_norm": 1.2237492799758911, "learning_rate": 7.350987000481464e-05, "loss": 1.6188, "step": 5502 }, { "epoch": 2.651263537906137, "grad_norm": 1.298865795135498, "learning_rate": 7.348098218584497e-05, "loss": 1.6026, "step": 5508 }, { "epoch": 2.6541516245487364, "grad_norm": 1.4190306663513184, "learning_rate": 7.34520943668753e-05, "loss": 1.5575, "step": 5514 }, { "epoch": 2.6570397111913358, "grad_norm": 1.4638572931289673, "learning_rate": 7.342320654790563e-05, "loss": 1.4777, "step": 5520 }, { "epoch": 2.659927797833935, "grad_norm": 1.3402222394943237, "learning_rate": 7.339431872893597e-05, "loss": 1.7415, "step": 5526 }, { "epoch": 2.6628158844765344, "grad_norm": 1.12592613697052, "learning_rate": 7.33654309099663e-05, "loss": 1.5952, "step": 5532 }, { "epoch": 2.6657039711191337, "grad_norm": 1.2370685338974, "learning_rate": 7.333654309099663e-05, "loss": 1.7116, "step": 5538 }, { "epoch": 2.6685920577617326, "grad_norm": 1.281273365020752, "learning_rate": 7.330765527202697e-05, "loss": 1.5687, "step": 5544 }, { "epoch": 2.671480144404332, "grad_norm": 1.3616689443588257, "learning_rate": 7.32787674530573e-05, "loss": 1.5347, "step": 5550 }, { "epoch": 2.674368231046931, "grad_norm": 1.3972671031951904, "learning_rate": 7.324987963408763e-05, "loss": 1.5767, "step": 5556 }, { "epoch": 2.6772563176895305, "grad_norm": 1.358963966369629, "learning_rate": 7.322099181511797e-05, "loss": 1.5608, "step": 5562 }, { "epoch": 2.68014440433213, "grad_norm": 1.3625091314315796, "learning_rate": 7.31921039961483e-05, "loss": 1.6922, "step": 5568 }, { "epoch": 2.683032490974729, "grad_norm": 1.3537369966506958, "learning_rate": 7.316321617717863e-05, "loss": 1.5137, "step": 5574 }, { "epoch": 2.6859205776173285, "grad_norm": 1.3839062452316284, "learning_rate": 7.313432835820896e-05, "loss": 1.5753, "step": 5580 }, { "epoch": 2.688808664259928, "grad_norm": 1.3717292547225952, "learning_rate": 7.310544053923928e-05, "loss": 1.6706, "step": 5586 }, { "epoch": 2.691696750902527, "grad_norm": 1.3445582389831543, "learning_rate": 7.307655272026962e-05, "loss": 1.8085, "step": 5592 }, { "epoch": 2.6945848375451265, "grad_norm": 1.2958651781082153, "learning_rate": 7.304766490129995e-05, "loss": 1.6474, "step": 5598 }, { "epoch": 2.6974729241877258, "grad_norm": 1.3864260911941528, "learning_rate": 7.30187770823303e-05, "loss": 1.6932, "step": 5604 }, { "epoch": 2.700361010830325, "grad_norm": 1.2720165252685547, "learning_rate": 7.298988926336062e-05, "loss": 1.6862, "step": 5610 }, { "epoch": 2.7032490974729244, "grad_norm": 1.2167384624481201, "learning_rate": 7.296100144439095e-05, "loss": 1.6075, "step": 5616 }, { "epoch": 2.7061371841155237, "grad_norm": 1.2950185537338257, "learning_rate": 7.293211362542129e-05, "loss": 1.7357, "step": 5622 }, { "epoch": 2.7090252707581226, "grad_norm": 1.424984335899353, "learning_rate": 7.290322580645161e-05, "loss": 1.5795, "step": 5628 }, { "epoch": 2.711913357400722, "grad_norm": 1.364794373512268, "learning_rate": 7.287433798748195e-05, "loss": 1.6716, "step": 5634 }, { "epoch": 2.7148014440433212, "grad_norm": 1.2985285520553589, "learning_rate": 7.284545016851229e-05, "loss": 1.614, "step": 5640 }, { "epoch": 2.7176895306859206, "grad_norm": 1.2109993696212769, "learning_rate": 7.281656234954261e-05, "loss": 1.5145, "step": 5646 }, { "epoch": 2.72057761732852, "grad_norm": 1.2661532163619995, "learning_rate": 7.278767453057295e-05, "loss": 1.7077, "step": 5652 }, { "epoch": 2.723465703971119, "grad_norm": 1.266586184501648, "learning_rate": 7.275878671160328e-05, "loss": 1.6358, "step": 5658 }, { "epoch": 2.7263537906137185, "grad_norm": 1.286040186882019, "learning_rate": 7.272989889263361e-05, "loss": 1.6334, "step": 5664 }, { "epoch": 2.729241877256318, "grad_norm": 1.4455010890960693, "learning_rate": 7.270101107366394e-05, "loss": 1.656, "step": 5670 }, { "epoch": 2.7321299638989167, "grad_norm": 1.1945879459381104, "learning_rate": 7.267212325469426e-05, "loss": 1.5453, "step": 5676 }, { "epoch": 2.735018050541516, "grad_norm": 1.2462996244430542, "learning_rate": 7.264323543572462e-05, "loss": 1.659, "step": 5682 }, { "epoch": 2.7379061371841154, "grad_norm": 1.2458769083023071, "learning_rate": 7.261434761675494e-05, "loss": 1.665, "step": 5688 }, { "epoch": 2.7407942238267147, "grad_norm": 1.2947547435760498, "learning_rate": 7.258545979778527e-05, "loss": 1.4766, "step": 5694 }, { "epoch": 2.743682310469314, "grad_norm": 1.3986401557922363, "learning_rate": 7.25565719788156e-05, "loss": 1.5921, "step": 5700 }, { "epoch": 2.7465703971119133, "grad_norm": 1.3267658948898315, "learning_rate": 7.252768415984593e-05, "loss": 1.5958, "step": 5706 }, { "epoch": 2.7494584837545126, "grad_norm": 1.2047417163848877, "learning_rate": 7.249879634087627e-05, "loss": 1.5399, "step": 5712 }, { "epoch": 2.752346570397112, "grad_norm": 1.3697569370269775, "learning_rate": 7.24699085219066e-05, "loss": 1.6239, "step": 5718 }, { "epoch": 2.7552346570397113, "grad_norm": 1.355702519416809, "learning_rate": 7.244102070293693e-05, "loss": 1.6487, "step": 5724 }, { "epoch": 2.7581227436823106, "grad_norm": 1.4432755708694458, "learning_rate": 7.241213288396727e-05, "loss": 1.6086, "step": 5730 }, { "epoch": 2.76101083032491, "grad_norm": 1.4248381853103638, "learning_rate": 7.238324506499759e-05, "loss": 1.677, "step": 5736 }, { "epoch": 2.7638989169675092, "grad_norm": 1.3859144449234009, "learning_rate": 7.235435724602793e-05, "loss": 1.7666, "step": 5742 }, { "epoch": 2.7667870036101085, "grad_norm": 1.3488315343856812, "learning_rate": 7.232546942705826e-05, "loss": 1.5927, "step": 5748 }, { "epoch": 2.7696750902527074, "grad_norm": 1.596721887588501, "learning_rate": 7.22965816080886e-05, "loss": 1.8206, "step": 5754 }, { "epoch": 2.7725631768953067, "grad_norm": 1.315971851348877, "learning_rate": 7.226769378911893e-05, "loss": 1.6721, "step": 5760 }, { "epoch": 2.775451263537906, "grad_norm": 1.0635358095169067, "learning_rate": 7.223880597014926e-05, "loss": 1.6032, "step": 5766 }, { "epoch": 2.7783393501805054, "grad_norm": 1.2257412672042847, "learning_rate": 7.22099181511796e-05, "loss": 1.5735, "step": 5772 }, { "epoch": 2.7812274368231047, "grad_norm": 1.2224736213684082, "learning_rate": 7.218103033220992e-05, "loss": 1.6036, "step": 5778 }, { "epoch": 2.784115523465704, "grad_norm": 1.3514684438705444, "learning_rate": 7.215214251324025e-05, "loss": 1.4362, "step": 5784 }, { "epoch": 2.7870036101083033, "grad_norm": 1.1783185005187988, "learning_rate": 7.212325469427058e-05, "loss": 1.5864, "step": 5790 }, { "epoch": 2.7898916967509026, "grad_norm": 1.3172597885131836, "learning_rate": 7.209436687530092e-05, "loss": 1.4758, "step": 5796 }, { "epoch": 2.792779783393502, "grad_norm": 1.2571547031402588, "learning_rate": 7.206547905633125e-05, "loss": 1.6945, "step": 5802 }, { "epoch": 2.795667870036101, "grad_norm": 1.1622651815414429, "learning_rate": 7.203659123736159e-05, "loss": 1.6206, "step": 5808 }, { "epoch": 2.7985559566787, "grad_norm": 1.49088454246521, "learning_rate": 7.200770341839191e-05, "loss": 1.6189, "step": 5814 }, { "epoch": 2.8014440433212995, "grad_norm": 1.616474986076355, "learning_rate": 7.197881559942225e-05, "loss": 1.5603, "step": 5820 }, { "epoch": 2.804332129963899, "grad_norm": 1.1622787714004517, "learning_rate": 7.194992778045257e-05, "loss": 1.6995, "step": 5826 }, { "epoch": 2.807220216606498, "grad_norm": 1.2091171741485596, "learning_rate": 7.192103996148291e-05, "loss": 1.7526, "step": 5832 }, { "epoch": 2.8101083032490974, "grad_norm": 1.3821436166763306, "learning_rate": 7.189215214251325e-05, "loss": 1.5683, "step": 5838 }, { "epoch": 2.8129963898916968, "grad_norm": 1.3422421216964722, "learning_rate": 7.186326432354358e-05, "loss": 1.6817, "step": 5844 }, { "epoch": 2.815884476534296, "grad_norm": 1.3110400438308716, "learning_rate": 7.183437650457391e-05, "loss": 1.7442, "step": 5850 }, { "epoch": 2.8187725631768954, "grad_norm": 1.5378167629241943, "learning_rate": 7.180548868560424e-05, "loss": 1.6124, "step": 5856 }, { "epoch": 2.8216606498194947, "grad_norm": 1.289829134941101, "learning_rate": 7.177660086663456e-05, "loss": 1.5071, "step": 5862 }, { "epoch": 2.824548736462094, "grad_norm": 1.1729573011398315, "learning_rate": 7.17477130476649e-05, "loss": 1.7035, "step": 5868 }, { "epoch": 2.8274368231046934, "grad_norm": 1.5683456659317017, "learning_rate": 7.171882522869524e-05, "loss": 1.6186, "step": 5874 }, { "epoch": 2.8303249097472927, "grad_norm": 1.2638441324234009, "learning_rate": 7.168993740972557e-05, "loss": 1.5563, "step": 5880 }, { "epoch": 2.8332129963898915, "grad_norm": 1.2925350666046143, "learning_rate": 7.16610495907559e-05, "loss": 1.6974, "step": 5886 }, { "epoch": 2.836101083032491, "grad_norm": 1.3925011157989502, "learning_rate": 7.163216177178623e-05, "loss": 1.7928, "step": 5892 }, { "epoch": 2.83898916967509, "grad_norm": 1.4501144886016846, "learning_rate": 7.160327395281657e-05, "loss": 1.4659, "step": 5898 }, { "epoch": 2.8418772563176895, "grad_norm": 1.219335913658142, "learning_rate": 7.157438613384689e-05, "loss": 1.4762, "step": 5904 }, { "epoch": 2.844765342960289, "grad_norm": 1.1340190172195435, "learning_rate": 7.154549831487723e-05, "loss": 1.4428, "step": 5910 }, { "epoch": 2.847653429602888, "grad_norm": 1.3847441673278809, "learning_rate": 7.151661049590757e-05, "loss": 1.6348, "step": 5916 }, { "epoch": 2.8505415162454875, "grad_norm": 1.447758674621582, "learning_rate": 7.148772267693789e-05, "loss": 1.5288, "step": 5922 }, { "epoch": 2.853429602888087, "grad_norm": 1.2214792966842651, "learning_rate": 7.145883485796823e-05, "loss": 1.5558, "step": 5928 }, { "epoch": 2.8563176895306857, "grad_norm": 1.4466886520385742, "learning_rate": 7.142994703899856e-05, "loss": 1.6775, "step": 5934 }, { "epoch": 2.859205776173285, "grad_norm": 1.2865900993347168, "learning_rate": 7.140105922002888e-05, "loss": 1.5486, "step": 5940 }, { "epoch": 2.8620938628158843, "grad_norm": 1.5181447267532349, "learning_rate": 7.137217140105922e-05, "loss": 1.5823, "step": 5946 }, { "epoch": 2.8649819494584836, "grad_norm": 1.541406273841858, "learning_rate": 7.134328358208956e-05, "loss": 1.5834, "step": 5952 }, { "epoch": 2.867870036101083, "grad_norm": 1.297784447669983, "learning_rate": 7.13143957631199e-05, "loss": 1.6832, "step": 5958 }, { "epoch": 2.8707581227436823, "grad_norm": 1.313137412071228, "learning_rate": 7.128550794415022e-05, "loss": 1.5613, "step": 5964 }, { "epoch": 2.8736462093862816, "grad_norm": 1.3346741199493408, "learning_rate": 7.125662012518055e-05, "loss": 1.6823, "step": 5970 }, { "epoch": 2.876534296028881, "grad_norm": 1.3883317708969116, "learning_rate": 7.122773230621088e-05, "loss": 1.6169, "step": 5976 }, { "epoch": 2.87942238267148, "grad_norm": 1.1874194145202637, "learning_rate": 7.119884448724121e-05, "loss": 1.4746, "step": 5982 }, { "epoch": 2.8823104693140795, "grad_norm": 1.185779333114624, "learning_rate": 7.116995666827155e-05, "loss": 1.5456, "step": 5988 }, { "epoch": 2.885198555956679, "grad_norm": 1.3507730960845947, "learning_rate": 7.114106884930189e-05, "loss": 1.645, "step": 5994 }, { "epoch": 2.888086642599278, "grad_norm": 1.2408970594406128, "learning_rate": 7.111218103033221e-05, "loss": 1.6393, "step": 6000 }, { "epoch": 2.8909747292418775, "grad_norm": 1.2361149787902832, "learning_rate": 7.108329321136255e-05, "loss": 1.5963, "step": 6006 }, { "epoch": 2.893862815884477, "grad_norm": 1.188944935798645, "learning_rate": 7.105440539239287e-05, "loss": 1.6087, "step": 6012 }, { "epoch": 2.8967509025270757, "grad_norm": 1.2834843397140503, "learning_rate": 7.102551757342321e-05, "loss": 1.5252, "step": 6018 }, { "epoch": 2.899638989169675, "grad_norm": 1.0705162286758423, "learning_rate": 7.099662975445355e-05, "loss": 1.5861, "step": 6024 }, { "epoch": 2.9025270758122743, "grad_norm": 1.2680517435073853, "learning_rate": 7.096774193548388e-05, "loss": 1.5593, "step": 6030 }, { "epoch": 2.9054151624548736, "grad_norm": 1.3138911724090576, "learning_rate": 7.093885411651421e-05, "loss": 1.663, "step": 6036 }, { "epoch": 2.908303249097473, "grad_norm": 1.159525752067566, "learning_rate": 7.090996629754454e-05, "loss": 1.6696, "step": 6042 }, { "epoch": 2.9111913357400723, "grad_norm": 1.2861100435256958, "learning_rate": 7.088107847857486e-05, "loss": 1.6043, "step": 6048 }, { "epoch": 2.9140794223826716, "grad_norm": 1.3583109378814697, "learning_rate": 7.08521906596052e-05, "loss": 1.5967, "step": 6054 }, { "epoch": 2.916967509025271, "grad_norm": 1.1275051832199097, "learning_rate": 7.082330284063553e-05, "loss": 1.4659, "step": 6060 }, { "epoch": 2.91985559566787, "grad_norm": 1.322127103805542, "learning_rate": 7.079441502166588e-05, "loss": 1.5323, "step": 6066 }, { "epoch": 2.922743682310469, "grad_norm": 1.6170438528060913, "learning_rate": 7.07655272026962e-05, "loss": 1.5573, "step": 6072 }, { "epoch": 2.9256317689530684, "grad_norm": 1.362372636795044, "learning_rate": 7.073663938372653e-05, "loss": 1.5177, "step": 6078 }, { "epoch": 2.9285198555956677, "grad_norm": 1.4996514320373535, "learning_rate": 7.070775156475687e-05, "loss": 1.693, "step": 6084 }, { "epoch": 2.931407942238267, "grad_norm": 1.6922024488449097, "learning_rate": 7.067886374578719e-05, "loss": 1.6693, "step": 6090 }, { "epoch": 2.9342960288808664, "grad_norm": 1.5298017263412476, "learning_rate": 7.064997592681753e-05, "loss": 1.7013, "step": 6096 }, { "epoch": 2.9371841155234657, "grad_norm": 1.326957106590271, "learning_rate": 7.062108810784787e-05, "loss": 1.5126, "step": 6102 }, { "epoch": 2.940072202166065, "grad_norm": 1.3751074075698853, "learning_rate": 7.059220028887819e-05, "loss": 1.5895, "step": 6108 }, { "epoch": 2.9429602888086643, "grad_norm": 1.3946486711502075, "learning_rate": 7.056331246990853e-05, "loss": 1.4357, "step": 6114 }, { "epoch": 2.9458483754512637, "grad_norm": 1.3499393463134766, "learning_rate": 7.053442465093886e-05, "loss": 1.4927, "step": 6120 }, { "epoch": 2.948736462093863, "grad_norm": 1.2250151634216309, "learning_rate": 7.05055368319692e-05, "loss": 1.3964, "step": 6126 }, { "epoch": 2.9516245487364623, "grad_norm": 1.2756565809249878, "learning_rate": 7.047664901299952e-05, "loss": 1.7204, "step": 6132 }, { "epoch": 2.9545126353790616, "grad_norm": 1.3110105991363525, "learning_rate": 7.044776119402984e-05, "loss": 1.5094, "step": 6138 }, { "epoch": 2.9574007220216605, "grad_norm": 1.2724515199661255, "learning_rate": 7.04188733750602e-05, "loss": 1.7711, "step": 6144 }, { "epoch": 2.96028880866426, "grad_norm": 1.350330114364624, "learning_rate": 7.038998555609052e-05, "loss": 1.5933, "step": 6150 }, { "epoch": 2.963176895306859, "grad_norm": 1.2350198030471802, "learning_rate": 7.036109773712085e-05, "loss": 1.6471, "step": 6156 }, { "epoch": 2.9660649819494584, "grad_norm": 1.2377517223358154, "learning_rate": 7.033220991815118e-05, "loss": 1.6777, "step": 6162 }, { "epoch": 2.9689530685920578, "grad_norm": 1.2064777612686157, "learning_rate": 7.030332209918151e-05, "loss": 1.6778, "step": 6168 }, { "epoch": 2.971841155234657, "grad_norm": 1.2704119682312012, "learning_rate": 7.027443428021185e-05, "loss": 1.6394, "step": 6174 }, { "epoch": 2.9747292418772564, "grad_norm": 1.2109112739562988, "learning_rate": 7.024554646124219e-05, "loss": 1.5224, "step": 6180 }, { "epoch": 2.9776173285198557, "grad_norm": 1.492051362991333, "learning_rate": 7.021665864227251e-05, "loss": 1.4729, "step": 6186 }, { "epoch": 2.9805054151624546, "grad_norm": 1.5079621076583862, "learning_rate": 7.018777082330285e-05, "loss": 1.5897, "step": 6192 }, { "epoch": 2.983393501805054, "grad_norm": 1.1070969104766846, "learning_rate": 7.015888300433317e-05, "loss": 1.5476, "step": 6198 }, { "epoch": 2.9862815884476532, "grad_norm": 1.4847484827041626, "learning_rate": 7.012999518536351e-05, "loss": 1.6039, "step": 6204 }, { "epoch": 2.9891696750902526, "grad_norm": 1.1884548664093018, "learning_rate": 7.010110736639384e-05, "loss": 1.7351, "step": 6210 }, { "epoch": 2.992057761732852, "grad_norm": 1.3621864318847656, "learning_rate": 7.007221954742418e-05, "loss": 1.6624, "step": 6216 }, { "epoch": 2.994945848375451, "grad_norm": 1.4305689334869385, "learning_rate": 7.004333172845451e-05, "loss": 1.6281, "step": 6222 }, { "epoch": 2.9978339350180505, "grad_norm": 1.2927610874176025, "learning_rate": 7.001444390948484e-05, "loss": 1.6954, "step": 6228 }, { "epoch": 3.00072202166065, "grad_norm": 1.282721996307373, "learning_rate": 6.998555609051516e-05, "loss": 1.4308, "step": 6234 }, { "epoch": 3.003610108303249, "grad_norm": 1.273368000984192, "learning_rate": 6.99566682715455e-05, "loss": 1.4508, "step": 6240 }, { "epoch": 3.0064981949458485, "grad_norm": 1.4955154657363892, "learning_rate": 6.992778045257583e-05, "loss": 1.6074, "step": 6246 }, { "epoch": 3.009386281588448, "grad_norm": 1.42247474193573, "learning_rate": 6.989889263360616e-05, "loss": 1.534, "step": 6252 }, { "epoch": 3.012274368231047, "grad_norm": 1.3951181173324585, "learning_rate": 6.98700048146365e-05, "loss": 1.6169, "step": 6258 }, { "epoch": 3.0151624548736464, "grad_norm": 1.2139549255371094, "learning_rate": 6.984111699566683e-05, "loss": 1.4012, "step": 6264 }, { "epoch": 3.0180505415162453, "grad_norm": 1.2227681875228882, "learning_rate": 6.981222917669717e-05, "loss": 1.5491, "step": 6270 }, { "epoch": 3.0209386281588446, "grad_norm": 1.4078668355941772, "learning_rate": 6.978334135772749e-05, "loss": 1.5155, "step": 6276 }, { "epoch": 3.023826714801444, "grad_norm": 1.220913290977478, "learning_rate": 6.975445353875783e-05, "loss": 1.5441, "step": 6282 }, { "epoch": 3.0267148014440433, "grad_norm": 1.6128947734832764, "learning_rate": 6.972556571978815e-05, "loss": 1.5433, "step": 6288 }, { "epoch": 3.0296028880866426, "grad_norm": 1.4628114700317383, "learning_rate": 6.969667790081849e-05, "loss": 1.5636, "step": 6294 }, { "epoch": 3.032490974729242, "grad_norm": 1.2599527835845947, "learning_rate": 6.966779008184883e-05, "loss": 1.4994, "step": 6300 }, { "epoch": 3.035379061371841, "grad_norm": 1.3851128816604614, "learning_rate": 6.963890226287916e-05, "loss": 1.5261, "step": 6306 }, { "epoch": 3.0382671480144405, "grad_norm": 1.6416099071502686, "learning_rate": 6.96100144439095e-05, "loss": 1.519, "step": 6312 }, { "epoch": 3.04115523465704, "grad_norm": 1.536676287651062, "learning_rate": 6.958112662493982e-05, "loss": 1.6529, "step": 6318 }, { "epoch": 3.044043321299639, "grad_norm": 1.2807570695877075, "learning_rate": 6.955223880597014e-05, "loss": 1.4145, "step": 6324 }, { "epoch": 3.046931407942238, "grad_norm": 1.264595866203308, "learning_rate": 6.952335098700048e-05, "loss": 1.5285, "step": 6330 }, { "epoch": 3.0498194945848374, "grad_norm": 1.3549964427947998, "learning_rate": 6.949446316803082e-05, "loss": 1.4394, "step": 6336 }, { "epoch": 3.0527075812274367, "grad_norm": 1.537376880645752, "learning_rate": 6.946557534906115e-05, "loss": 1.584, "step": 6342 }, { "epoch": 3.055595667870036, "grad_norm": 1.3610512018203735, "learning_rate": 6.943668753009148e-05, "loss": 1.5712, "step": 6348 }, { "epoch": 3.0584837545126353, "grad_norm": 1.3872708082199097, "learning_rate": 6.940779971112181e-05, "loss": 1.6135, "step": 6354 }, { "epoch": 3.0613718411552346, "grad_norm": 1.401494026184082, "learning_rate": 6.937891189215215e-05, "loss": 1.398, "step": 6360 }, { "epoch": 3.064259927797834, "grad_norm": 1.371489405632019, "learning_rate": 6.935002407318247e-05, "loss": 1.5593, "step": 6366 }, { "epoch": 3.0671480144404333, "grad_norm": 1.2828617095947266, "learning_rate": 6.932113625421281e-05, "loss": 1.4622, "step": 6372 }, { "epoch": 3.0700361010830326, "grad_norm": 1.3589963912963867, "learning_rate": 6.929224843524315e-05, "loss": 1.6289, "step": 6378 }, { "epoch": 3.072924187725632, "grad_norm": 1.4497342109680176, "learning_rate": 6.926336061627347e-05, "loss": 1.5249, "step": 6384 }, { "epoch": 3.0758122743682312, "grad_norm": 1.512769103050232, "learning_rate": 6.923447279730381e-05, "loss": 1.461, "step": 6390 }, { "epoch": 3.07870036101083, "grad_norm": 1.3837759494781494, "learning_rate": 6.920558497833414e-05, "loss": 1.3899, "step": 6396 }, { "epoch": 3.0815884476534294, "grad_norm": 1.8521764278411865, "learning_rate": 6.917669715936446e-05, "loss": 1.6093, "step": 6402 }, { "epoch": 3.0844765342960287, "grad_norm": 1.350327491760254, "learning_rate": 6.91478093403948e-05, "loss": 1.5044, "step": 6408 }, { "epoch": 3.087364620938628, "grad_norm": 1.4750343561172485, "learning_rate": 6.911892152142514e-05, "loss": 1.4215, "step": 6414 }, { "epoch": 3.0902527075812274, "grad_norm": 1.394968867301941, "learning_rate": 6.909003370245548e-05, "loss": 1.578, "step": 6420 }, { "epoch": 3.0931407942238267, "grad_norm": 1.3584859371185303, "learning_rate": 6.90611458834858e-05, "loss": 1.6288, "step": 6426 }, { "epoch": 3.096028880866426, "grad_norm": 1.435828685760498, "learning_rate": 6.903225806451613e-05, "loss": 1.4614, "step": 6432 }, { "epoch": 3.0989169675090253, "grad_norm": 1.6963458061218262, "learning_rate": 6.900337024554646e-05, "loss": 1.6315, "step": 6438 }, { "epoch": 3.1018050541516247, "grad_norm": 1.4279038906097412, "learning_rate": 6.897448242657679e-05, "loss": 1.5825, "step": 6444 }, { "epoch": 3.104693140794224, "grad_norm": 1.2581316232681274, "learning_rate": 6.894559460760713e-05, "loss": 1.4823, "step": 6450 }, { "epoch": 3.1075812274368233, "grad_norm": 1.436333417892456, "learning_rate": 6.891670678863747e-05, "loss": 1.3954, "step": 6456 }, { "epoch": 3.110469314079422, "grad_norm": 1.5141692161560059, "learning_rate": 6.888781896966779e-05, "loss": 1.5268, "step": 6462 }, { "epoch": 3.1133574007220215, "grad_norm": 1.4098795652389526, "learning_rate": 6.885893115069813e-05, "loss": 1.5579, "step": 6468 }, { "epoch": 3.116245487364621, "grad_norm": 1.2756410837173462, "learning_rate": 6.883004333172845e-05, "loss": 1.5175, "step": 6474 }, { "epoch": 3.11913357400722, "grad_norm": 1.5627847909927368, "learning_rate": 6.880115551275879e-05, "loss": 1.4705, "step": 6480 }, { "epoch": 3.1220216606498195, "grad_norm": 1.3875834941864014, "learning_rate": 6.877226769378913e-05, "loss": 1.5465, "step": 6486 }, { "epoch": 3.1249097472924188, "grad_norm": 1.395652174949646, "learning_rate": 6.874337987481946e-05, "loss": 1.5263, "step": 6492 }, { "epoch": 3.127797833935018, "grad_norm": 1.372431755065918, "learning_rate": 6.87144920558498e-05, "loss": 1.6131, "step": 6498 }, { "epoch": 3.1306859205776174, "grad_norm": 1.5990840196609497, "learning_rate": 6.868560423688012e-05, "loss": 1.6711, "step": 6504 }, { "epoch": 3.1335740072202167, "grad_norm": 1.4993387460708618, "learning_rate": 6.865671641791044e-05, "loss": 1.6545, "step": 6510 }, { "epoch": 3.136462093862816, "grad_norm": 1.3550641536712646, "learning_rate": 6.862782859894078e-05, "loss": 1.4062, "step": 6516 }, { "epoch": 3.1393501805054154, "grad_norm": 1.390467643737793, "learning_rate": 6.85989407799711e-05, "loss": 1.4803, "step": 6522 }, { "epoch": 3.1422382671480142, "grad_norm": 1.4581964015960693, "learning_rate": 6.857005296100146e-05, "loss": 1.4739, "step": 6528 }, { "epoch": 3.1451263537906136, "grad_norm": 1.3887474536895752, "learning_rate": 6.854116514203178e-05, "loss": 1.4054, "step": 6534 }, { "epoch": 3.148014440433213, "grad_norm": 1.4484093189239502, "learning_rate": 6.851227732306211e-05, "loss": 1.4714, "step": 6540 }, { "epoch": 3.150902527075812, "grad_norm": 1.5083246231079102, "learning_rate": 6.848338950409245e-05, "loss": 1.6308, "step": 6546 }, { "epoch": 3.1537906137184115, "grad_norm": 1.328951120376587, "learning_rate": 6.845450168512277e-05, "loss": 1.5347, "step": 6552 }, { "epoch": 3.156678700361011, "grad_norm": 1.4870861768722534, "learning_rate": 6.842561386615311e-05, "loss": 1.656, "step": 6558 }, { "epoch": 3.15956678700361, "grad_norm": 1.4287126064300537, "learning_rate": 6.839672604718345e-05, "loss": 1.4072, "step": 6564 }, { "epoch": 3.1624548736462095, "grad_norm": 1.5289545059204102, "learning_rate": 6.836783822821377e-05, "loss": 1.44, "step": 6570 }, { "epoch": 3.165342960288809, "grad_norm": 1.4592528343200684, "learning_rate": 6.833895040924411e-05, "loss": 1.3888, "step": 6576 }, { "epoch": 3.168231046931408, "grad_norm": 1.6642377376556396, "learning_rate": 6.831006259027444e-05, "loss": 1.5433, "step": 6582 }, { "epoch": 3.171119133574007, "grad_norm": 1.2856703996658325, "learning_rate": 6.828117477130476e-05, "loss": 1.5278, "step": 6588 }, { "epoch": 3.1740072202166063, "grad_norm": 1.4379680156707764, "learning_rate": 6.82522869523351e-05, "loss": 1.6534, "step": 6594 }, { "epoch": 3.1768953068592056, "grad_norm": 1.4422087669372559, "learning_rate": 6.822339913336542e-05, "loss": 1.5405, "step": 6600 }, { "epoch": 3.179783393501805, "grad_norm": 1.5906853675842285, "learning_rate": 6.819451131439578e-05, "loss": 1.5102, "step": 6606 }, { "epoch": 3.1826714801444043, "grad_norm": 1.231186032295227, "learning_rate": 6.81656234954261e-05, "loss": 1.5602, "step": 6612 }, { "epoch": 3.1855595667870036, "grad_norm": 1.3421540260314941, "learning_rate": 6.813673567645643e-05, "loss": 1.4416, "step": 6618 }, { "epoch": 3.188447653429603, "grad_norm": 1.606502652168274, "learning_rate": 6.810784785748676e-05, "loss": 1.4712, "step": 6624 }, { "epoch": 3.191335740072202, "grad_norm": 1.6454626321792603, "learning_rate": 6.807896003851709e-05, "loss": 1.5368, "step": 6630 }, { "epoch": 3.1942238267148015, "grad_norm": 1.5166757106781006, "learning_rate": 6.805007221954743e-05, "loss": 1.5122, "step": 6636 }, { "epoch": 3.197111913357401, "grad_norm": 1.3923163414001465, "learning_rate": 6.802118440057777e-05, "loss": 1.5438, "step": 6642 }, { "epoch": 3.2, "grad_norm": 1.4296586513519287, "learning_rate": 6.799229658160809e-05, "loss": 1.5347, "step": 6648 }, { "epoch": 3.2028880866425995, "grad_norm": 1.3495914936065674, "learning_rate": 6.796340876263843e-05, "loss": 1.4089, "step": 6654 }, { "epoch": 3.2057761732851984, "grad_norm": 1.422104835510254, "learning_rate": 6.793452094366875e-05, "loss": 1.5055, "step": 6660 }, { "epoch": 3.2086642599277977, "grad_norm": 1.559430480003357, "learning_rate": 6.790563312469909e-05, "loss": 1.5793, "step": 6666 }, { "epoch": 3.211552346570397, "grad_norm": 1.4511831998825073, "learning_rate": 6.787674530572942e-05, "loss": 1.5523, "step": 6672 }, { "epoch": 3.2144404332129963, "grad_norm": 1.6502745151519775, "learning_rate": 6.784785748675976e-05, "loss": 1.5111, "step": 6678 }, { "epoch": 3.2173285198555956, "grad_norm": 1.5004009008407593, "learning_rate": 6.78189696677901e-05, "loss": 1.5731, "step": 6684 }, { "epoch": 3.220216606498195, "grad_norm": 1.4580602645874023, "learning_rate": 6.779008184882042e-05, "loss": 1.4664, "step": 6690 }, { "epoch": 3.2231046931407943, "grad_norm": 1.2027837038040161, "learning_rate": 6.776119402985074e-05, "loss": 1.5317, "step": 6696 }, { "epoch": 3.2259927797833936, "grad_norm": 1.5642435550689697, "learning_rate": 6.773230621088108e-05, "loss": 1.5712, "step": 6702 }, { "epoch": 3.228880866425993, "grad_norm": 1.517662525177002, "learning_rate": 6.77034183919114e-05, "loss": 1.4691, "step": 6708 }, { "epoch": 3.2317689530685922, "grad_norm": 1.5193718671798706, "learning_rate": 6.767453057294174e-05, "loss": 1.5351, "step": 6714 }, { "epoch": 3.234657039711191, "grad_norm": 1.2706571817398071, "learning_rate": 6.764564275397208e-05, "loss": 1.5433, "step": 6720 }, { "epoch": 3.2375451263537904, "grad_norm": 1.4653698205947876, "learning_rate": 6.761675493500241e-05, "loss": 1.5458, "step": 6726 }, { "epoch": 3.2404332129963898, "grad_norm": 1.2928682565689087, "learning_rate": 6.758786711603275e-05, "loss": 1.5147, "step": 6732 }, { "epoch": 3.243321299638989, "grad_norm": 1.398807406425476, "learning_rate": 6.755897929706307e-05, "loss": 1.4942, "step": 6738 }, { "epoch": 3.2462093862815884, "grad_norm": 1.617550015449524, "learning_rate": 6.753009147809341e-05, "loss": 1.6212, "step": 6744 }, { "epoch": 3.2490974729241877, "grad_norm": 1.5323724746704102, "learning_rate": 6.750120365912373e-05, "loss": 1.5397, "step": 6750 }, { "epoch": 3.251985559566787, "grad_norm": 1.6292804479599, "learning_rate": 6.747231584015407e-05, "loss": 1.4601, "step": 6756 }, { "epoch": 3.2548736462093864, "grad_norm": 1.5685395002365112, "learning_rate": 6.744342802118441e-05, "loss": 1.4972, "step": 6762 }, { "epoch": 3.2577617328519857, "grad_norm": 1.6193629503250122, "learning_rate": 6.741454020221474e-05, "loss": 1.4892, "step": 6768 }, { "epoch": 3.260649819494585, "grad_norm": 1.3756766319274902, "learning_rate": 6.738565238324507e-05, "loss": 1.6205, "step": 6774 }, { "epoch": 3.2635379061371843, "grad_norm": 1.3481699228286743, "learning_rate": 6.73567645642754e-05, "loss": 1.4917, "step": 6780 }, { "epoch": 3.2664259927797836, "grad_norm": 1.5346568822860718, "learning_rate": 6.732787674530572e-05, "loss": 1.6203, "step": 6786 }, { "epoch": 3.2693140794223825, "grad_norm": 1.7245608568191528, "learning_rate": 6.729898892633606e-05, "loss": 1.6326, "step": 6792 }, { "epoch": 3.272202166064982, "grad_norm": 1.554915428161621, "learning_rate": 6.72701011073664e-05, "loss": 1.3797, "step": 6798 }, { "epoch": 3.275090252707581, "grad_norm": 1.5676765441894531, "learning_rate": 6.724121328839673e-05, "loss": 1.5011, "step": 6804 }, { "epoch": 3.2779783393501805, "grad_norm": 1.4126967191696167, "learning_rate": 6.721232546942706e-05, "loss": 1.5041, "step": 6810 }, { "epoch": 3.2808664259927798, "grad_norm": 1.8876826763153076, "learning_rate": 6.718343765045739e-05, "loss": 1.5403, "step": 6816 }, { "epoch": 3.283754512635379, "grad_norm": 1.2787379026412964, "learning_rate": 6.715454983148773e-05, "loss": 1.4815, "step": 6822 }, { "epoch": 3.2866425992779784, "grad_norm": 1.3978190422058105, "learning_rate": 6.712566201251805e-05, "loss": 1.5324, "step": 6828 }, { "epoch": 3.2895306859205777, "grad_norm": 1.5534523725509644, "learning_rate": 6.709677419354839e-05, "loss": 1.5045, "step": 6834 }, { "epoch": 3.292418772563177, "grad_norm": 1.1673310995101929, "learning_rate": 6.706788637457873e-05, "loss": 1.4586, "step": 6840 }, { "epoch": 3.295306859205776, "grad_norm": 1.436120867729187, "learning_rate": 6.703899855560905e-05, "loss": 1.4821, "step": 6846 }, { "epoch": 3.2981949458483752, "grad_norm": 1.386696219444275, "learning_rate": 6.701011073663939e-05, "loss": 1.5751, "step": 6852 }, { "epoch": 3.3010830324909746, "grad_norm": 1.4546778202056885, "learning_rate": 6.698122291766972e-05, "loss": 1.6171, "step": 6858 }, { "epoch": 3.303971119133574, "grad_norm": 1.287412166595459, "learning_rate": 6.695233509870004e-05, "loss": 1.5816, "step": 6864 }, { "epoch": 3.306859205776173, "grad_norm": 1.2950351238250732, "learning_rate": 6.69234472797304e-05, "loss": 1.4545, "step": 6870 }, { "epoch": 3.3097472924187725, "grad_norm": 1.4955419301986694, "learning_rate": 6.689455946076072e-05, "loss": 1.4505, "step": 6876 }, { "epoch": 3.312635379061372, "grad_norm": 1.2548589706420898, "learning_rate": 6.686567164179106e-05, "loss": 1.4659, "step": 6882 }, { "epoch": 3.315523465703971, "grad_norm": 1.8329297304153442, "learning_rate": 6.683678382282138e-05, "loss": 1.4436, "step": 6888 }, { "epoch": 3.3184115523465705, "grad_norm": 1.476171851158142, "learning_rate": 6.68078960038517e-05, "loss": 1.57, "step": 6894 }, { "epoch": 3.32129963898917, "grad_norm": 1.6188886165618896, "learning_rate": 6.677900818488204e-05, "loss": 1.6481, "step": 6900 }, { "epoch": 3.324187725631769, "grad_norm": 1.5443732738494873, "learning_rate": 6.675012036591237e-05, "loss": 1.368, "step": 6906 }, { "epoch": 3.3270758122743684, "grad_norm": 1.4383240938186646, "learning_rate": 6.672123254694271e-05, "loss": 1.5098, "step": 6912 }, { "epoch": 3.3299638989169678, "grad_norm": 1.5492448806762695, "learning_rate": 6.669234472797305e-05, "loss": 1.5829, "step": 6918 }, { "epoch": 3.3328519855595666, "grad_norm": 1.4505982398986816, "learning_rate": 6.666345690900337e-05, "loss": 1.5352, "step": 6924 }, { "epoch": 3.335740072202166, "grad_norm": 1.4156904220581055, "learning_rate": 6.663456909003371e-05, "loss": 1.6596, "step": 6930 }, { "epoch": 3.3386281588447653, "grad_norm": 1.4835001230239868, "learning_rate": 6.660568127106403e-05, "loss": 1.5499, "step": 6936 }, { "epoch": 3.3415162454873646, "grad_norm": 1.4488086700439453, "learning_rate": 6.657679345209437e-05, "loss": 1.4835, "step": 6942 }, { "epoch": 3.344404332129964, "grad_norm": 1.6607515811920166, "learning_rate": 6.654790563312471e-05, "loss": 1.5113, "step": 6948 }, { "epoch": 3.3472924187725632, "grad_norm": 1.690685749053955, "learning_rate": 6.651901781415504e-05, "loss": 1.6099, "step": 6954 }, { "epoch": 3.3501805054151625, "grad_norm": 1.4607224464416504, "learning_rate": 6.649012999518537e-05, "loss": 1.4669, "step": 6960 }, { "epoch": 3.353068592057762, "grad_norm": 1.6601324081420898, "learning_rate": 6.64612421762157e-05, "loss": 1.4777, "step": 6966 }, { "epoch": 3.355956678700361, "grad_norm": 1.3485480546951294, "learning_rate": 6.643235435724602e-05, "loss": 1.5004, "step": 6972 }, { "epoch": 3.35884476534296, "grad_norm": 1.571258306503296, "learning_rate": 6.640346653827636e-05, "loss": 1.4316, "step": 6978 }, { "epoch": 3.3617328519855594, "grad_norm": 1.436010718345642, "learning_rate": 6.637457871930669e-05, "loss": 1.5176, "step": 6984 }, { "epoch": 3.3646209386281587, "grad_norm": 1.6512060165405273, "learning_rate": 6.634569090033703e-05, "loss": 1.5103, "step": 6990 }, { "epoch": 3.367509025270758, "grad_norm": 1.4819453954696655, "learning_rate": 6.631680308136736e-05, "loss": 1.5654, "step": 6996 }, { "epoch": 3.3703971119133573, "grad_norm": 1.4926148653030396, "learning_rate": 6.628791526239769e-05, "loss": 1.474, "step": 7002 }, { "epoch": 3.3732851985559567, "grad_norm": 1.4325014352798462, "learning_rate": 6.625902744342803e-05, "loss": 1.4935, "step": 7008 }, { "epoch": 3.376173285198556, "grad_norm": 1.5342340469360352, "learning_rate": 6.623013962445835e-05, "loss": 1.5243, "step": 7014 }, { "epoch": 3.3790613718411553, "grad_norm": 1.8693536520004272, "learning_rate": 6.620125180548869e-05, "loss": 1.5669, "step": 7020 }, { "epoch": 3.3819494584837546, "grad_norm": 1.433168649673462, "learning_rate": 6.617236398651903e-05, "loss": 1.4978, "step": 7026 }, { "epoch": 3.384837545126354, "grad_norm": 1.5898305177688599, "learning_rate": 6.614347616754935e-05, "loss": 1.6241, "step": 7032 }, { "epoch": 3.3877256317689532, "grad_norm": 1.5472948551177979, "learning_rate": 6.611458834857969e-05, "loss": 1.44, "step": 7038 }, { "epoch": 3.3906137184115526, "grad_norm": 1.4879659414291382, "learning_rate": 6.608570052961002e-05, "loss": 1.5091, "step": 7044 }, { "epoch": 3.3935018050541514, "grad_norm": 1.6339657306671143, "learning_rate": 6.605681271064034e-05, "loss": 1.6722, "step": 7050 }, { "epoch": 3.3963898916967508, "grad_norm": 1.4274871349334717, "learning_rate": 6.602792489167068e-05, "loss": 1.5392, "step": 7056 }, { "epoch": 3.39927797833935, "grad_norm": 1.4243069887161255, "learning_rate": 6.5999037072701e-05, "loss": 1.3447, "step": 7062 }, { "epoch": 3.4021660649819494, "grad_norm": 1.4953285455703735, "learning_rate": 6.597014925373136e-05, "loss": 1.4886, "step": 7068 }, { "epoch": 3.4050541516245487, "grad_norm": 1.6193463802337646, "learning_rate": 6.594126143476168e-05, "loss": 1.6539, "step": 7074 }, { "epoch": 3.407942238267148, "grad_norm": 1.5607357025146484, "learning_rate": 6.5912373615792e-05, "loss": 1.496, "step": 7080 }, { "epoch": 3.4108303249097474, "grad_norm": 1.4780343770980835, "learning_rate": 6.588348579682234e-05, "loss": 1.6454, "step": 7086 }, { "epoch": 3.4137184115523467, "grad_norm": 1.4360201358795166, "learning_rate": 6.585459797785267e-05, "loss": 1.5863, "step": 7092 }, { "epoch": 3.416606498194946, "grad_norm": 1.3535892963409424, "learning_rate": 6.582571015888301e-05, "loss": 1.5526, "step": 7098 }, { "epoch": 3.4194945848375453, "grad_norm": 1.3327372074127197, "learning_rate": 6.579682233991335e-05, "loss": 1.6812, "step": 7104 }, { "epoch": 3.422382671480144, "grad_norm": 1.7549904584884644, "learning_rate": 6.576793452094367e-05, "loss": 1.7746, "step": 7110 }, { "epoch": 3.4252707581227435, "grad_norm": 1.4379364252090454, "learning_rate": 6.573904670197401e-05, "loss": 1.5785, "step": 7116 }, { "epoch": 3.428158844765343, "grad_norm": 1.5202213525772095, "learning_rate": 6.571015888300433e-05, "loss": 1.6704, "step": 7122 }, { "epoch": 3.431046931407942, "grad_norm": 1.3667750358581543, "learning_rate": 6.568127106403467e-05, "loss": 1.5697, "step": 7128 }, { "epoch": 3.4339350180505415, "grad_norm": 1.7388198375701904, "learning_rate": 6.5652383245065e-05, "loss": 1.5042, "step": 7134 }, { "epoch": 3.436823104693141, "grad_norm": 1.4183061122894287, "learning_rate": 6.562349542609534e-05, "loss": 1.6761, "step": 7140 }, { "epoch": 3.43971119133574, "grad_norm": 1.6383647918701172, "learning_rate": 6.559460760712567e-05, "loss": 1.5824, "step": 7146 }, { "epoch": 3.4425992779783394, "grad_norm": 1.4391705989837646, "learning_rate": 6.5565719788156e-05, "loss": 1.6134, "step": 7152 }, { "epoch": 3.4454873646209387, "grad_norm": 1.4116913080215454, "learning_rate": 6.553683196918632e-05, "loss": 1.5476, "step": 7158 }, { "epoch": 3.448375451263538, "grad_norm": 1.3838624954223633, "learning_rate": 6.550794415021666e-05, "loss": 1.5168, "step": 7164 }, { "epoch": 3.4512635379061374, "grad_norm": 1.6727395057678223, "learning_rate": 6.547905633124699e-05, "loss": 1.5165, "step": 7170 }, { "epoch": 3.4541516245487367, "grad_norm": 1.5257728099822998, "learning_rate": 6.545016851227733e-05, "loss": 1.5048, "step": 7176 }, { "epoch": 3.4570397111913356, "grad_norm": 1.5113712549209595, "learning_rate": 6.542128069330766e-05, "loss": 1.4783, "step": 7182 }, { "epoch": 3.459927797833935, "grad_norm": 1.4760724306106567, "learning_rate": 6.539239287433799e-05, "loss": 1.4726, "step": 7188 }, { "epoch": 3.462815884476534, "grad_norm": 1.4856163263320923, "learning_rate": 6.536350505536833e-05, "loss": 1.5903, "step": 7194 }, { "epoch": 3.4657039711191335, "grad_norm": 1.7575163841247559, "learning_rate": 6.533461723639865e-05, "loss": 1.6314, "step": 7200 }, { "epoch": 3.468592057761733, "grad_norm": 1.4918440580368042, "learning_rate": 6.530572941742899e-05, "loss": 1.5808, "step": 7206 }, { "epoch": 3.471480144404332, "grad_norm": 1.3790260553359985, "learning_rate": 6.527684159845931e-05, "loss": 1.6271, "step": 7212 }, { "epoch": 3.4743682310469315, "grad_norm": 1.574593424797058, "learning_rate": 6.524795377948965e-05, "loss": 1.4264, "step": 7218 }, { "epoch": 3.477256317689531, "grad_norm": 1.7312476634979248, "learning_rate": 6.521906596051999e-05, "loss": 1.4523, "step": 7224 }, { "epoch": 3.48014440433213, "grad_norm": 1.541854977607727, "learning_rate": 6.519017814155032e-05, "loss": 1.5271, "step": 7230 }, { "epoch": 3.483032490974729, "grad_norm": 1.322721004486084, "learning_rate": 6.516129032258065e-05, "loss": 1.4823, "step": 7236 }, { "epoch": 3.4859205776173283, "grad_norm": 1.7561603784561157, "learning_rate": 6.513240250361098e-05, "loss": 1.5794, "step": 7242 }, { "epoch": 3.4888086642599276, "grad_norm": 1.424901008605957, "learning_rate": 6.51035146846413e-05, "loss": 1.5423, "step": 7248 }, { "epoch": 3.491696750902527, "grad_norm": 1.2866737842559814, "learning_rate": 6.507462686567164e-05, "loss": 1.6048, "step": 7254 }, { "epoch": 3.4945848375451263, "grad_norm": 1.5464754104614258, "learning_rate": 6.504573904670198e-05, "loss": 1.3892, "step": 7260 }, { "epoch": 3.4974729241877256, "grad_norm": 1.4413031339645386, "learning_rate": 6.50168512277323e-05, "loss": 1.46, "step": 7266 }, { "epoch": 3.500361010830325, "grad_norm": 1.6262658834457397, "learning_rate": 6.498796340876264e-05, "loss": 1.4199, "step": 7272 }, { "epoch": 3.5032490974729242, "grad_norm": 1.4883025884628296, "learning_rate": 6.495907558979297e-05, "loss": 1.63, "step": 7278 }, { "epoch": 3.5061371841155236, "grad_norm": 1.5281471014022827, "learning_rate": 6.493018777082331e-05, "loss": 1.4694, "step": 7284 }, { "epoch": 3.509025270758123, "grad_norm": 1.6533699035644531, "learning_rate": 6.490129995185363e-05, "loss": 1.6792, "step": 7290 }, { "epoch": 3.511913357400722, "grad_norm": 1.645508885383606, "learning_rate": 6.487241213288397e-05, "loss": 1.7064, "step": 7296 }, { "epoch": 3.5148014440433215, "grad_norm": 1.1872341632843018, "learning_rate": 6.484352431391431e-05, "loss": 1.5519, "step": 7302 }, { "epoch": 3.517689530685921, "grad_norm": 1.2986491918563843, "learning_rate": 6.481463649494463e-05, "loss": 1.4083, "step": 7308 }, { "epoch": 3.5205776173285197, "grad_norm": 1.4779576063156128, "learning_rate": 6.478574867597497e-05, "loss": 1.256, "step": 7314 }, { "epoch": 3.523465703971119, "grad_norm": 1.3267526626586914, "learning_rate": 6.47568608570053e-05, "loss": 1.4446, "step": 7320 }, { "epoch": 3.5263537906137183, "grad_norm": 1.3313223123550415, "learning_rate": 6.472797303803562e-05, "loss": 1.5269, "step": 7326 }, { "epoch": 3.5292418772563177, "grad_norm": 1.308917760848999, "learning_rate": 6.469908521906597e-05, "loss": 1.3902, "step": 7332 }, { "epoch": 3.532129963898917, "grad_norm": 1.3230177164077759, "learning_rate": 6.46701974000963e-05, "loss": 1.52, "step": 7338 }, { "epoch": 3.5350180505415163, "grad_norm": 1.4954818487167358, "learning_rate": 6.464130958112662e-05, "loss": 1.4969, "step": 7344 }, { "epoch": 3.5379061371841156, "grad_norm": 1.678186058998108, "learning_rate": 6.461242176215696e-05, "loss": 1.5565, "step": 7350 }, { "epoch": 3.540794223826715, "grad_norm": 1.5359582901000977, "learning_rate": 6.458353394318729e-05, "loss": 1.5765, "step": 7356 }, { "epoch": 3.543682310469314, "grad_norm": 1.4628596305847168, "learning_rate": 6.455464612421763e-05, "loss": 1.4644, "step": 7362 }, { "epoch": 3.546570397111913, "grad_norm": 1.4721364974975586, "learning_rate": 6.452575830524795e-05, "loss": 1.4648, "step": 7368 }, { "epoch": 3.5494584837545125, "grad_norm": 1.507857084274292, "learning_rate": 6.449687048627829e-05, "loss": 1.5581, "step": 7374 }, { "epoch": 3.5523465703971118, "grad_norm": 1.5103305578231812, "learning_rate": 6.446798266730863e-05, "loss": 1.52, "step": 7380 }, { "epoch": 3.555234657039711, "grad_norm": 1.5159964561462402, "learning_rate": 6.443909484833895e-05, "loss": 1.6344, "step": 7386 }, { "epoch": 3.5581227436823104, "grad_norm": 1.5743902921676636, "learning_rate": 6.441020702936929e-05, "loss": 1.5543, "step": 7392 }, { "epoch": 3.5610108303249097, "grad_norm": 1.4101184606552124, "learning_rate": 6.438131921039961e-05, "loss": 1.335, "step": 7398 }, { "epoch": 3.563898916967509, "grad_norm": 1.6320714950561523, "learning_rate": 6.435243139142994e-05, "loss": 1.4472, "step": 7404 }, { "epoch": 3.5667870036101084, "grad_norm": 1.3284647464752197, "learning_rate": 6.432354357246029e-05, "loss": 1.583, "step": 7410 }, { "epoch": 3.5696750902527077, "grad_norm": 1.445041298866272, "learning_rate": 6.429465575349062e-05, "loss": 1.5604, "step": 7416 }, { "epoch": 3.572563176895307, "grad_norm": 1.4162532091140747, "learning_rate": 6.426576793452095e-05, "loss": 1.4675, "step": 7422 }, { "epoch": 3.5754512635379063, "grad_norm": 1.3426916599273682, "learning_rate": 6.423688011555128e-05, "loss": 1.5954, "step": 7428 }, { "epoch": 3.5783393501805056, "grad_norm": 1.5268527269363403, "learning_rate": 6.42079922965816e-05, "loss": 1.4583, "step": 7434 }, { "epoch": 3.581227436823105, "grad_norm": 1.4427614212036133, "learning_rate": 6.417910447761194e-05, "loss": 1.6517, "step": 7440 }, { "epoch": 3.584115523465704, "grad_norm": 1.420042872428894, "learning_rate": 6.415021665864227e-05, "loss": 1.5428, "step": 7446 }, { "epoch": 3.587003610108303, "grad_norm": 1.4562898874282837, "learning_rate": 6.41213288396726e-05, "loss": 1.7234, "step": 7452 }, { "epoch": 3.5898916967509025, "grad_norm": 1.3193168640136719, "learning_rate": 6.409244102070294e-05, "loss": 1.6164, "step": 7458 }, { "epoch": 3.592779783393502, "grad_norm": 1.7231054306030273, "learning_rate": 6.406355320173327e-05, "loss": 1.5238, "step": 7464 }, { "epoch": 3.595667870036101, "grad_norm": 1.7380363941192627, "learning_rate": 6.403466538276361e-05, "loss": 1.6431, "step": 7470 }, { "epoch": 3.5985559566787004, "grad_norm": 1.3448389768600464, "learning_rate": 6.400577756379393e-05, "loss": 1.3804, "step": 7476 }, { "epoch": 3.6014440433212997, "grad_norm": 1.618169903755188, "learning_rate": 6.397688974482427e-05, "loss": 1.5849, "step": 7482 }, { "epoch": 3.604332129963899, "grad_norm": 1.4316123723983765, "learning_rate": 6.394800192585461e-05, "loss": 1.4229, "step": 7488 }, { "epoch": 3.607220216606498, "grad_norm": 1.4174507856369019, "learning_rate": 6.391911410688493e-05, "loss": 1.5361, "step": 7494 }, { "epoch": 3.6101083032490973, "grad_norm": 1.4997692108154297, "learning_rate": 6.389022628791527e-05, "loss": 1.4183, "step": 7500 }, { "epoch": 3.6129963898916966, "grad_norm": 1.6012548208236694, "learning_rate": 6.38613384689456e-05, "loss": 1.4574, "step": 7506 }, { "epoch": 3.615884476534296, "grad_norm": 1.5662246942520142, "learning_rate": 6.383245064997592e-05, "loss": 1.491, "step": 7512 }, { "epoch": 3.618772563176895, "grad_norm": 1.4726227521896362, "learning_rate": 6.380356283100626e-05, "loss": 1.6551, "step": 7518 }, { "epoch": 3.6216606498194945, "grad_norm": 1.7899826765060425, "learning_rate": 6.377467501203658e-05, "loss": 1.575, "step": 7524 }, { "epoch": 3.624548736462094, "grad_norm": 1.3145309686660767, "learning_rate": 6.374578719306694e-05, "loss": 1.4415, "step": 7530 }, { "epoch": 3.627436823104693, "grad_norm": 1.2957236766815186, "learning_rate": 6.371689937409726e-05, "loss": 1.5145, "step": 7536 }, { "epoch": 3.6303249097472925, "grad_norm": 1.4462108612060547, "learning_rate": 6.368801155512759e-05, "loss": 1.4974, "step": 7542 }, { "epoch": 3.633212996389892, "grad_norm": 1.3840019702911377, "learning_rate": 6.365912373615792e-05, "loss": 1.4705, "step": 7548 }, { "epoch": 3.636101083032491, "grad_norm": 1.8745399713516235, "learning_rate": 6.363023591718825e-05, "loss": 1.4415, "step": 7554 }, { "epoch": 3.6389891696750905, "grad_norm": 1.6287518739700317, "learning_rate": 6.360134809821859e-05, "loss": 1.621, "step": 7560 }, { "epoch": 3.6418772563176898, "grad_norm": 1.4659773111343384, "learning_rate": 6.357246027924893e-05, "loss": 1.5174, "step": 7566 }, { "epoch": 3.6447653429602886, "grad_norm": 1.3289144039154053, "learning_rate": 6.354357246027925e-05, "loss": 1.5571, "step": 7572 }, { "epoch": 3.647653429602888, "grad_norm": 1.6725703477859497, "learning_rate": 6.351468464130959e-05, "loss": 1.5611, "step": 7578 }, { "epoch": 3.6505415162454873, "grad_norm": 1.5543458461761475, "learning_rate": 6.348579682233991e-05, "loss": 1.4958, "step": 7584 }, { "epoch": 3.6534296028880866, "grad_norm": 1.556384563446045, "learning_rate": 6.345690900337025e-05, "loss": 1.4885, "step": 7590 }, { "epoch": 3.656317689530686, "grad_norm": 1.4587044715881348, "learning_rate": 6.342802118440058e-05, "loss": 1.37, "step": 7596 }, { "epoch": 3.6592057761732852, "grad_norm": 1.4932681322097778, "learning_rate": 6.339913336543092e-05, "loss": 1.5462, "step": 7602 }, { "epoch": 3.6620938628158846, "grad_norm": 1.5371379852294922, "learning_rate": 6.337024554646125e-05, "loss": 1.6041, "step": 7608 }, { "epoch": 3.664981949458484, "grad_norm": 1.4784220457077026, "learning_rate": 6.334135772749158e-05, "loss": 1.4612, "step": 7614 }, { "epoch": 3.667870036101083, "grad_norm": 1.6144012212753296, "learning_rate": 6.33124699085219e-05, "loss": 1.5867, "step": 7620 }, { "epoch": 3.670758122743682, "grad_norm": 1.5069010257720947, "learning_rate": 6.328358208955224e-05, "loss": 1.5127, "step": 7626 }, { "epoch": 3.6736462093862814, "grad_norm": 1.629775047302246, "learning_rate": 6.325469427058257e-05, "loss": 1.6359, "step": 7632 }, { "epoch": 3.6765342960288807, "grad_norm": 1.8102917671203613, "learning_rate": 6.32258064516129e-05, "loss": 1.5538, "step": 7638 }, { "epoch": 3.67942238267148, "grad_norm": 1.4897123575210571, "learning_rate": 6.319691863264324e-05, "loss": 1.5643, "step": 7644 }, { "epoch": 3.6823104693140793, "grad_norm": 1.4273055791854858, "learning_rate": 6.316803081367357e-05, "loss": 1.5429, "step": 7650 }, { "epoch": 3.6851985559566787, "grad_norm": 1.451066017150879, "learning_rate": 6.313914299470391e-05, "loss": 1.5614, "step": 7656 }, { "epoch": 3.688086642599278, "grad_norm": 1.972150206565857, "learning_rate": 6.311025517573423e-05, "loss": 1.4888, "step": 7662 }, { "epoch": 3.6909747292418773, "grad_norm": 1.441041350364685, "learning_rate": 6.308136735676457e-05, "loss": 1.6074, "step": 7668 }, { "epoch": 3.6938628158844766, "grad_norm": 1.5164940357208252, "learning_rate": 6.30524795377949e-05, "loss": 1.4738, "step": 7674 }, { "epoch": 3.696750902527076, "grad_norm": 1.4767781496047974, "learning_rate": 6.302359171882523e-05, "loss": 1.6453, "step": 7680 }, { "epoch": 3.6996389891696753, "grad_norm": 1.625365972518921, "learning_rate": 6.299470389985557e-05, "loss": 1.563, "step": 7686 }, { "epoch": 3.7025270758122746, "grad_norm": 1.53213369846344, "learning_rate": 6.29658160808859e-05, "loss": 1.5084, "step": 7692 }, { "epoch": 3.705415162454874, "grad_norm": 1.6262837648391724, "learning_rate": 6.293692826191622e-05, "loss": 1.3751, "step": 7698 }, { "epoch": 3.7083032490974728, "grad_norm": 1.5903650522232056, "learning_rate": 6.290804044294656e-05, "loss": 1.3668, "step": 7704 }, { "epoch": 3.711191335740072, "grad_norm": 1.6617803573608398, "learning_rate": 6.287915262397688e-05, "loss": 1.5651, "step": 7710 }, { "epoch": 3.7140794223826714, "grad_norm": 1.388130784034729, "learning_rate": 6.285026480500722e-05, "loss": 1.4406, "step": 7716 }, { "epoch": 3.7169675090252707, "grad_norm": 1.403623342514038, "learning_rate": 6.282137698603756e-05, "loss": 1.4963, "step": 7722 }, { "epoch": 3.71985559566787, "grad_norm": 1.2051951885223389, "learning_rate": 6.279248916706789e-05, "loss": 1.4565, "step": 7728 }, { "epoch": 3.7227436823104694, "grad_norm": 1.4764797687530518, "learning_rate": 6.276360134809822e-05, "loss": 1.502, "step": 7734 }, { "epoch": 3.7256317689530687, "grad_norm": 1.4363739490509033, "learning_rate": 6.273471352912855e-05, "loss": 1.5337, "step": 7740 }, { "epoch": 3.728519855595668, "grad_norm": 1.173463225364685, "learning_rate": 6.270582571015889e-05, "loss": 1.4593, "step": 7746 }, { "epoch": 3.731407942238267, "grad_norm": 1.469289779663086, "learning_rate": 6.267693789118921e-05, "loss": 1.5526, "step": 7752 }, { "epoch": 3.734296028880866, "grad_norm": 1.530397891998291, "learning_rate": 6.264805007221955e-05, "loss": 1.5535, "step": 7758 }, { "epoch": 3.7371841155234655, "grad_norm": 1.5095781087875366, "learning_rate": 6.261916225324989e-05, "loss": 1.6442, "step": 7764 }, { "epoch": 3.740072202166065, "grad_norm": 1.4158457517623901, "learning_rate": 6.259027443428021e-05, "loss": 1.4654, "step": 7770 }, { "epoch": 3.742960288808664, "grad_norm": 1.5007539987564087, "learning_rate": 6.256138661531055e-05, "loss": 1.5055, "step": 7776 }, { "epoch": 3.7458483754512635, "grad_norm": 1.7604862451553345, "learning_rate": 6.253249879634088e-05, "loss": 1.6218, "step": 7782 }, { "epoch": 3.748736462093863, "grad_norm": 1.394851803779602, "learning_rate": 6.25036109773712e-05, "loss": 1.6479, "step": 7788 }, { "epoch": 3.751624548736462, "grad_norm": 1.5454732179641724, "learning_rate": 6.247472315840155e-05, "loss": 1.6242, "step": 7794 }, { "epoch": 3.7545126353790614, "grad_norm": 1.8919693231582642, "learning_rate": 6.244583533943188e-05, "loss": 1.6416, "step": 7800 }, { "epoch": 3.7574007220216608, "grad_norm": 1.6387457847595215, "learning_rate": 6.24169475204622e-05, "loss": 1.4174, "step": 7806 }, { "epoch": 3.76028880866426, "grad_norm": 1.399137258529663, "learning_rate": 6.238805970149254e-05, "loss": 1.5456, "step": 7812 }, { "epoch": 3.7631768953068594, "grad_norm": 1.443626880645752, "learning_rate": 6.235917188252287e-05, "loss": 1.5683, "step": 7818 }, { "epoch": 3.7660649819494587, "grad_norm": 1.6055556535720825, "learning_rate": 6.23302840635532e-05, "loss": 1.6042, "step": 7824 }, { "epoch": 3.768953068592058, "grad_norm": 1.5026847124099731, "learning_rate": 6.230139624458353e-05, "loss": 1.497, "step": 7830 }, { "epoch": 3.771841155234657, "grad_norm": 1.5990947484970093, "learning_rate": 6.227250842561387e-05, "loss": 1.496, "step": 7836 }, { "epoch": 3.7747292418772562, "grad_norm": 1.368952751159668, "learning_rate": 6.224362060664421e-05, "loss": 1.4256, "step": 7842 }, { "epoch": 3.7776173285198555, "grad_norm": 1.292877435684204, "learning_rate": 6.221473278767453e-05, "loss": 1.4361, "step": 7848 }, { "epoch": 3.780505415162455, "grad_norm": 1.2545397281646729, "learning_rate": 6.218584496870487e-05, "loss": 1.5185, "step": 7854 }, { "epoch": 3.783393501805054, "grad_norm": 1.4091216325759888, "learning_rate": 6.21569571497352e-05, "loss": 1.4895, "step": 7860 }, { "epoch": 3.7862815884476535, "grad_norm": 1.6132863759994507, "learning_rate": 6.212806933076552e-05, "loss": 1.5802, "step": 7866 }, { "epoch": 3.789169675090253, "grad_norm": 1.5012065172195435, "learning_rate": 6.209918151179587e-05, "loss": 1.5185, "step": 7872 }, { "epoch": 3.792057761732852, "grad_norm": 1.6006042957305908, "learning_rate": 6.20702936928262e-05, "loss": 1.6084, "step": 7878 }, { "epoch": 3.794945848375451, "grad_norm": 1.474250078201294, "learning_rate": 6.204140587385653e-05, "loss": 1.4051, "step": 7884 }, { "epoch": 3.7978339350180503, "grad_norm": 1.7603445053100586, "learning_rate": 6.201251805488686e-05, "loss": 1.4651, "step": 7890 }, { "epoch": 3.8007220216606497, "grad_norm": 1.5761584043502808, "learning_rate": 6.198363023591718e-05, "loss": 1.5937, "step": 7896 }, { "epoch": 3.803610108303249, "grad_norm": 1.6909846067428589, "learning_rate": 6.195474241694752e-05, "loss": 1.4438, "step": 7902 }, { "epoch": 3.8064981949458483, "grad_norm": 1.4626883268356323, "learning_rate": 6.192585459797785e-05, "loss": 1.4724, "step": 7908 }, { "epoch": 3.8093862815884476, "grad_norm": 1.4689635038375854, "learning_rate": 6.189696677900819e-05, "loss": 1.5172, "step": 7914 }, { "epoch": 3.812274368231047, "grad_norm": 1.8288270235061646, "learning_rate": 6.186807896003852e-05, "loss": 1.6703, "step": 7920 }, { "epoch": 3.8151624548736462, "grad_norm": 1.2827084064483643, "learning_rate": 6.183919114106885e-05, "loss": 1.4645, "step": 7926 }, { "epoch": 3.8180505415162456, "grad_norm": 1.618679165840149, "learning_rate": 6.181030332209919e-05, "loss": 1.6419, "step": 7932 }, { "epoch": 3.820938628158845, "grad_norm": 1.5698423385620117, "learning_rate": 6.178141550312951e-05, "loss": 1.5364, "step": 7938 }, { "epoch": 3.823826714801444, "grad_norm": 1.4559266567230225, "learning_rate": 6.175252768415985e-05, "loss": 1.7049, "step": 7944 }, { "epoch": 3.8267148014440435, "grad_norm": 1.3951289653778076, "learning_rate": 6.172363986519019e-05, "loss": 1.4616, "step": 7950 }, { "epoch": 3.829602888086643, "grad_norm": 1.2115695476531982, "learning_rate": 6.169475204622051e-05, "loss": 1.312, "step": 7956 }, { "epoch": 3.8324909747292417, "grad_norm": 1.58408784866333, "learning_rate": 6.166586422725085e-05, "loss": 1.6526, "step": 7962 }, { "epoch": 3.835379061371841, "grad_norm": 1.5165067911148071, "learning_rate": 6.163697640828118e-05, "loss": 1.6011, "step": 7968 }, { "epoch": 3.8382671480144404, "grad_norm": 1.629097819328308, "learning_rate": 6.16080885893115e-05, "loss": 1.6349, "step": 7974 }, { "epoch": 3.8411552346570397, "grad_norm": 1.4361348152160645, "learning_rate": 6.157920077034184e-05, "loss": 1.4666, "step": 7980 }, { "epoch": 3.844043321299639, "grad_norm": 1.3918006420135498, "learning_rate": 6.155031295137218e-05, "loss": 1.5626, "step": 7986 }, { "epoch": 3.8469314079422383, "grad_norm": 1.6155132055282593, "learning_rate": 6.152142513240252e-05, "loss": 1.458, "step": 7992 }, { "epoch": 3.8498194945848376, "grad_norm": 1.6548750400543213, "learning_rate": 6.149253731343284e-05, "loss": 1.5995, "step": 7998 }, { "epoch": 3.852707581227437, "grad_norm": 1.725041389465332, "learning_rate": 6.146364949446317e-05, "loss": 1.4828, "step": 8004 }, { "epoch": 3.855595667870036, "grad_norm": 1.688523292541504, "learning_rate": 6.14347616754935e-05, "loss": 1.5678, "step": 8010 }, { "epoch": 3.858483754512635, "grad_norm": 1.3089724779129028, "learning_rate": 6.140587385652383e-05, "loss": 1.4564, "step": 8016 }, { "epoch": 3.8613718411552345, "grad_norm": 1.7244492769241333, "learning_rate": 6.137698603755417e-05, "loss": 1.4731, "step": 8022 }, { "epoch": 3.864259927797834, "grad_norm": 1.654013991355896, "learning_rate": 6.13480982185845e-05, "loss": 1.4557, "step": 8028 }, { "epoch": 3.867148014440433, "grad_norm": 1.608169674873352, "learning_rate": 6.131921039961483e-05, "loss": 1.6386, "step": 8034 }, { "epoch": 3.8700361010830324, "grad_norm": 1.621324062347412, "learning_rate": 6.129032258064517e-05, "loss": 1.6564, "step": 8040 }, { "epoch": 3.8729241877256317, "grad_norm": 1.643832802772522, "learning_rate": 6.12614347616755e-05, "loss": 1.5032, "step": 8046 }, { "epoch": 3.875812274368231, "grad_norm": 1.272944688796997, "learning_rate": 6.123254694270582e-05, "loss": 1.4866, "step": 8052 }, { "epoch": 3.8787003610108304, "grad_norm": 1.2037947177886963, "learning_rate": 6.120365912373616e-05, "loss": 1.379, "step": 8058 }, { "epoch": 3.8815884476534297, "grad_norm": 1.4692788124084473, "learning_rate": 6.11747713047665e-05, "loss": 1.5345, "step": 8064 }, { "epoch": 3.884476534296029, "grad_norm": 1.6058166027069092, "learning_rate": 6.114588348579683e-05, "loss": 1.5306, "step": 8070 }, { "epoch": 3.8873646209386283, "grad_norm": 1.6075516939163208, "learning_rate": 6.111699566682716e-05, "loss": 1.6014, "step": 8076 }, { "epoch": 3.8902527075812277, "grad_norm": 1.639082431793213, "learning_rate": 6.108810784785748e-05, "loss": 1.5178, "step": 8082 }, { "epoch": 3.893140794223827, "grad_norm": 1.494981288909912, "learning_rate": 6.105922002888782e-05, "loss": 1.4526, "step": 8088 }, { "epoch": 3.896028880866426, "grad_norm": 1.6544820070266724, "learning_rate": 6.103033220991815e-05, "loss": 1.4136, "step": 8094 }, { "epoch": 3.898916967509025, "grad_norm": 1.713814616203308, "learning_rate": 6.100144439094848e-05, "loss": 1.5209, "step": 8100 }, { "epoch": 3.9018050541516245, "grad_norm": 1.5945087671279907, "learning_rate": 6.0972556571978824e-05, "loss": 1.5685, "step": 8106 }, { "epoch": 3.904693140794224, "grad_norm": 1.448152780532837, "learning_rate": 6.0943668753009156e-05, "loss": 1.4336, "step": 8112 }, { "epoch": 3.907581227436823, "grad_norm": 1.4396202564239502, "learning_rate": 6.091478093403948e-05, "loss": 1.5036, "step": 8118 }, { "epoch": 3.9104693140794224, "grad_norm": 1.5016582012176514, "learning_rate": 6.088589311506981e-05, "loss": 1.5235, "step": 8124 }, { "epoch": 3.9133574007220218, "grad_norm": 1.4177602529525757, "learning_rate": 6.0857005296100144e-05, "loss": 1.4779, "step": 8130 }, { "epoch": 3.916245487364621, "grad_norm": 1.6093776226043701, "learning_rate": 6.0828117477130475e-05, "loss": 1.511, "step": 8136 }, { "epoch": 3.91913357400722, "grad_norm": 1.4724836349487305, "learning_rate": 6.0799229658160814e-05, "loss": 1.4643, "step": 8142 }, { "epoch": 3.9220216606498193, "grad_norm": 1.5472164154052734, "learning_rate": 6.0770341839191145e-05, "loss": 1.4958, "step": 8148 }, { "epoch": 3.9249097472924186, "grad_norm": 1.290418267250061, "learning_rate": 6.074145402022148e-05, "loss": 1.5664, "step": 8154 }, { "epoch": 3.927797833935018, "grad_norm": 1.6291229724884033, "learning_rate": 6.071256620125181e-05, "loss": 1.6923, "step": 8160 }, { "epoch": 3.9306859205776172, "grad_norm": 1.438417911529541, "learning_rate": 6.068367838228214e-05, "loss": 1.6395, "step": 8166 }, { "epoch": 3.9335740072202166, "grad_norm": 1.5091540813446045, "learning_rate": 6.065479056331247e-05, "loss": 1.5377, "step": 8172 }, { "epoch": 3.936462093862816, "grad_norm": 1.3645180463790894, "learning_rate": 6.06259027443428e-05, "loss": 1.6578, "step": 8178 }, { "epoch": 3.939350180505415, "grad_norm": 1.888291597366333, "learning_rate": 6.059701492537314e-05, "loss": 1.5703, "step": 8184 }, { "epoch": 3.9422382671480145, "grad_norm": 1.4239270687103271, "learning_rate": 6.056812710640347e-05, "loss": 1.5591, "step": 8190 }, { "epoch": 3.945126353790614, "grad_norm": 1.5898613929748535, "learning_rate": 6.0539239287433805e-05, "loss": 1.5799, "step": 8196 }, { "epoch": 3.948014440433213, "grad_norm": 1.8115546703338623, "learning_rate": 6.051035146846413e-05, "loss": 1.4434, "step": 8202 }, { "epoch": 3.9509025270758125, "grad_norm": 1.4307985305786133, "learning_rate": 6.048146364949446e-05, "loss": 1.4422, "step": 8208 }, { "epoch": 3.953790613718412, "grad_norm": 1.6043577194213867, "learning_rate": 6.045257583052479e-05, "loss": 1.5482, "step": 8214 }, { "epoch": 3.956678700361011, "grad_norm": 1.5922268629074097, "learning_rate": 6.042368801155514e-05, "loss": 1.5734, "step": 8220 }, { "epoch": 3.95956678700361, "grad_norm": 1.5308772325515747, "learning_rate": 6.039480019258546e-05, "loss": 1.5369, "step": 8226 }, { "epoch": 3.9624548736462093, "grad_norm": 1.6243884563446045, "learning_rate": 6.0365912373615795e-05, "loss": 1.5979, "step": 8232 }, { "epoch": 3.9653429602888086, "grad_norm": 1.6606591939926147, "learning_rate": 6.0337024554646126e-05, "loss": 1.6306, "step": 8238 }, { "epoch": 3.968231046931408, "grad_norm": 1.545500636100769, "learning_rate": 6.030813673567646e-05, "loss": 1.4779, "step": 8244 }, { "epoch": 3.9711191335740073, "grad_norm": 1.5263632535934448, "learning_rate": 6.027924891670679e-05, "loss": 1.5485, "step": 8250 }, { "epoch": 3.9740072202166066, "grad_norm": 1.7132220268249512, "learning_rate": 6.025036109773713e-05, "loss": 1.4637, "step": 8256 }, { "epoch": 3.976895306859206, "grad_norm": 1.455425500869751, "learning_rate": 6.022147327876746e-05, "loss": 1.4088, "step": 8262 }, { "epoch": 3.979783393501805, "grad_norm": 1.7424649000167847, "learning_rate": 6.019258545979779e-05, "loss": 1.577, "step": 8268 }, { "epoch": 3.982671480144404, "grad_norm": 1.442482352256775, "learning_rate": 6.016369764082812e-05, "loss": 1.4951, "step": 8274 }, { "epoch": 3.9855595667870034, "grad_norm": 1.6424578428268433, "learning_rate": 6.0134809821858454e-05, "loss": 1.4175, "step": 8280 }, { "epoch": 3.9884476534296027, "grad_norm": 1.3829751014709473, "learning_rate": 6.010592200288878e-05, "loss": 1.6003, "step": 8286 }, { "epoch": 3.991335740072202, "grad_norm": 1.3388317823410034, "learning_rate": 6.007703418391911e-05, "loss": 1.5075, "step": 8292 }, { "epoch": 3.9942238267148014, "grad_norm": 1.4209929704666138, "learning_rate": 6.0048146364949456e-05, "loss": 1.5223, "step": 8298 }, { "epoch": 3.9971119133574007, "grad_norm": 1.5496662855148315, "learning_rate": 6.001925854597979e-05, "loss": 1.4865, "step": 8304 }, { "epoch": 4.0, "grad_norm": 1.6789218187332153, "learning_rate": 5.999037072701011e-05, "loss": 1.6393, "step": 8310 }, { "epoch": 4.002888086642599, "grad_norm": 1.4550832509994507, "learning_rate": 5.9961482908040444e-05, "loss": 1.3915, "step": 8316 }, { "epoch": 4.005776173285199, "grad_norm": 1.614750862121582, "learning_rate": 5.9932595089070775e-05, "loss": 1.5198, "step": 8322 }, { "epoch": 4.008664259927798, "grad_norm": 1.721107840538025, "learning_rate": 5.990370727010111e-05, "loss": 1.2546, "step": 8328 }, { "epoch": 4.011552346570397, "grad_norm": 1.33194899559021, "learning_rate": 5.9874819451131445e-05, "loss": 1.3978, "step": 8334 }, { "epoch": 4.014440433212997, "grad_norm": 1.782246470451355, "learning_rate": 5.984593163216178e-05, "loss": 1.4508, "step": 8340 }, { "epoch": 4.017328519855596, "grad_norm": 1.2263399362564087, "learning_rate": 5.981704381319211e-05, "loss": 1.3852, "step": 8346 }, { "epoch": 4.020216606498195, "grad_norm": 1.714895248413086, "learning_rate": 5.978815599422244e-05, "loss": 1.3465, "step": 8352 }, { "epoch": 4.0231046931407946, "grad_norm": 1.4772042036056519, "learning_rate": 5.975926817525277e-05, "loss": 1.344, "step": 8358 }, { "epoch": 4.025992779783394, "grad_norm": 1.587766170501709, "learning_rate": 5.97303803562831e-05, "loss": 1.4376, "step": 8364 }, { "epoch": 4.028880866425993, "grad_norm": 1.5323612689971924, "learning_rate": 5.970149253731343e-05, "loss": 1.3739, "step": 8370 }, { "epoch": 4.031768953068592, "grad_norm": 1.409971833229065, "learning_rate": 5.967260471834377e-05, "loss": 1.402, "step": 8376 }, { "epoch": 4.034657039711191, "grad_norm": 1.6884794235229492, "learning_rate": 5.9643716899374105e-05, "loss": 1.4966, "step": 8382 }, { "epoch": 4.03754512635379, "grad_norm": 1.635025978088379, "learning_rate": 5.961482908040443e-05, "loss": 1.3885, "step": 8388 }, { "epoch": 4.04043321299639, "grad_norm": 1.669643759727478, "learning_rate": 5.958594126143476e-05, "loss": 1.4058, "step": 8394 }, { "epoch": 4.043321299638989, "grad_norm": 1.804304599761963, "learning_rate": 5.955705344246509e-05, "loss": 1.3034, "step": 8400 }, { "epoch": 4.046209386281588, "grad_norm": 1.658978819847107, "learning_rate": 5.9528165623495424e-05, "loss": 1.5572, "step": 8406 }, { "epoch": 4.0490974729241875, "grad_norm": 1.4568359851837158, "learning_rate": 5.949927780452576e-05, "loss": 1.4984, "step": 8412 }, { "epoch": 4.051985559566787, "grad_norm": 1.5212243795394897, "learning_rate": 5.9470389985556094e-05, "loss": 1.3974, "step": 8418 }, { "epoch": 4.054873646209386, "grad_norm": 1.70310640335083, "learning_rate": 5.9441502166586426e-05, "loss": 1.4399, "step": 8424 }, { "epoch": 4.0577617328519855, "grad_norm": 2.1572511196136475, "learning_rate": 5.941261434761676e-05, "loss": 1.4426, "step": 8430 }, { "epoch": 4.060649819494585, "grad_norm": 1.5291939973831177, "learning_rate": 5.938372652864709e-05, "loss": 1.4652, "step": 8436 }, { "epoch": 4.063537906137184, "grad_norm": 1.7144441604614258, "learning_rate": 5.935483870967742e-05, "loss": 1.3935, "step": 8442 }, { "epoch": 4.0664259927797834, "grad_norm": 1.729713797569275, "learning_rate": 5.932595089070776e-05, "loss": 1.5098, "step": 8448 }, { "epoch": 4.069314079422383, "grad_norm": 1.5398902893066406, "learning_rate": 5.929706307173809e-05, "loss": 1.5447, "step": 8454 }, { "epoch": 4.072202166064982, "grad_norm": 1.5530447959899902, "learning_rate": 5.926817525276842e-05, "loss": 1.6461, "step": 8460 }, { "epoch": 4.075090252707581, "grad_norm": 1.6429003477096558, "learning_rate": 5.9239287433798754e-05, "loss": 1.5015, "step": 8466 }, { "epoch": 4.077978339350181, "grad_norm": 1.656868815422058, "learning_rate": 5.921039961482908e-05, "loss": 1.4334, "step": 8472 }, { "epoch": 4.08086642599278, "grad_norm": 1.3499794006347656, "learning_rate": 5.918151179585941e-05, "loss": 1.4527, "step": 8478 }, { "epoch": 4.083754512635379, "grad_norm": 1.504385232925415, "learning_rate": 5.915262397688974e-05, "loss": 1.4374, "step": 8484 }, { "epoch": 4.086642599277979, "grad_norm": 1.5548646450042725, "learning_rate": 5.912373615792009e-05, "loss": 1.4621, "step": 8490 }, { "epoch": 4.089530685920578, "grad_norm": 1.591664433479309, "learning_rate": 5.909484833895041e-05, "loss": 1.3524, "step": 8496 }, { "epoch": 4.092418772563177, "grad_norm": 1.6754616498947144, "learning_rate": 5.9065960519980743e-05, "loss": 1.5838, "step": 8502 }, { "epoch": 4.095306859205776, "grad_norm": 1.681044340133667, "learning_rate": 5.9037072701011075e-05, "loss": 1.5389, "step": 8508 }, { "epoch": 4.098194945848375, "grad_norm": 1.4432076215744019, "learning_rate": 5.900818488204141e-05, "loss": 1.4055, "step": 8514 }, { "epoch": 4.101083032490974, "grad_norm": 1.4049861431121826, "learning_rate": 5.897929706307174e-05, "loss": 1.3669, "step": 8520 }, { "epoch": 4.103971119133574, "grad_norm": 1.8418153524398804, "learning_rate": 5.8950409244102077e-05, "loss": 1.3375, "step": 8526 }, { "epoch": 4.106859205776173, "grad_norm": 1.7224644422531128, "learning_rate": 5.892152142513241e-05, "loss": 1.4919, "step": 8532 }, { "epoch": 4.109747292418772, "grad_norm": 1.5333194732666016, "learning_rate": 5.889263360616274e-05, "loss": 1.5308, "step": 8538 }, { "epoch": 4.112635379061372, "grad_norm": 1.6399935483932495, "learning_rate": 5.886374578719307e-05, "loss": 1.4282, "step": 8544 }, { "epoch": 4.115523465703971, "grad_norm": 1.6538561582565308, "learning_rate": 5.88348579682234e-05, "loss": 1.3706, "step": 8550 }, { "epoch": 4.11841155234657, "grad_norm": 1.8446022272109985, "learning_rate": 5.880597014925373e-05, "loss": 1.4265, "step": 8556 }, { "epoch": 4.12129963898917, "grad_norm": 1.59902822971344, "learning_rate": 5.877708233028406e-05, "loss": 1.5447, "step": 8562 }, { "epoch": 4.124187725631769, "grad_norm": 1.3995146751403809, "learning_rate": 5.8748194511314405e-05, "loss": 1.4068, "step": 8568 }, { "epoch": 4.127075812274368, "grad_norm": 1.9567158222198486, "learning_rate": 5.8719306692344736e-05, "loss": 1.4449, "step": 8574 }, { "epoch": 4.129963898916968, "grad_norm": 1.66590416431427, "learning_rate": 5.869041887337506e-05, "loss": 1.6368, "step": 8580 }, { "epoch": 4.132851985559567, "grad_norm": 1.5213643312454224, "learning_rate": 5.866153105440539e-05, "loss": 1.3459, "step": 8586 }, { "epoch": 4.135740072202166, "grad_norm": 1.710214376449585, "learning_rate": 5.8632643235435724e-05, "loss": 1.5188, "step": 8592 }, { "epoch": 4.1386281588447655, "grad_norm": 1.5988775491714478, "learning_rate": 5.8603755416466056e-05, "loss": 1.5097, "step": 8598 }, { "epoch": 4.141516245487365, "grad_norm": 1.4921354055404663, "learning_rate": 5.8574867597496394e-05, "loss": 1.4448, "step": 8604 }, { "epoch": 4.144404332129964, "grad_norm": 1.5055233240127563, "learning_rate": 5.8545979778526726e-05, "loss": 1.3636, "step": 8610 }, { "epoch": 4.1472924187725635, "grad_norm": 1.6548657417297363, "learning_rate": 5.851709195955706e-05, "loss": 1.6389, "step": 8616 }, { "epoch": 4.150180505415163, "grad_norm": 1.7043691873550415, "learning_rate": 5.848820414058739e-05, "loss": 1.5006, "step": 8622 }, { "epoch": 4.153068592057762, "grad_norm": 1.5166929960250854, "learning_rate": 5.845931632161772e-05, "loss": 1.3257, "step": 8628 }, { "epoch": 4.1559566787003615, "grad_norm": 1.477417230606079, "learning_rate": 5.843042850264805e-05, "loss": 1.4928, "step": 8634 }, { "epoch": 4.15884476534296, "grad_norm": 1.579978585243225, "learning_rate": 5.840154068367838e-05, "loss": 1.3719, "step": 8640 }, { "epoch": 4.161732851985559, "grad_norm": 1.4910101890563965, "learning_rate": 5.837265286470872e-05, "loss": 1.3811, "step": 8646 }, { "epoch": 4.1646209386281585, "grad_norm": 1.7950223684310913, "learning_rate": 5.8343765045739054e-05, "loss": 1.5086, "step": 8652 }, { "epoch": 4.167509025270758, "grad_norm": 1.7879935503005981, "learning_rate": 5.8314877226769385e-05, "loss": 1.3616, "step": 8658 }, { "epoch": 4.170397111913357, "grad_norm": 1.643619418144226, "learning_rate": 5.828598940779971e-05, "loss": 1.4296, "step": 8664 }, { "epoch": 4.1732851985559565, "grad_norm": 1.7300169467926025, "learning_rate": 5.825710158883004e-05, "loss": 1.4791, "step": 8670 }, { "epoch": 4.176173285198556, "grad_norm": 1.6170413494110107, "learning_rate": 5.822821376986037e-05, "loss": 1.4922, "step": 8676 }, { "epoch": 4.179061371841155, "grad_norm": 1.7448548078536987, "learning_rate": 5.819932595089072e-05, "loss": 1.3818, "step": 8682 }, { "epoch": 4.181949458483754, "grad_norm": 1.7214919328689575, "learning_rate": 5.817043813192104e-05, "loss": 1.2601, "step": 8688 }, { "epoch": 4.184837545126354, "grad_norm": 1.7466048002243042, "learning_rate": 5.8141550312951375e-05, "loss": 1.4668, "step": 8694 }, { "epoch": 4.187725631768953, "grad_norm": 1.852961540222168, "learning_rate": 5.8112662493981706e-05, "loss": 1.5136, "step": 8700 }, { "epoch": 4.190613718411552, "grad_norm": 1.7246805429458618, "learning_rate": 5.808377467501204e-05, "loss": 1.4941, "step": 8706 }, { "epoch": 4.193501805054152, "grad_norm": 1.7242628335952759, "learning_rate": 5.805488685604237e-05, "loss": 1.3926, "step": 8712 }, { "epoch": 4.196389891696751, "grad_norm": 1.6819355487823486, "learning_rate": 5.802599903707271e-05, "loss": 1.4201, "step": 8718 }, { "epoch": 4.19927797833935, "grad_norm": 1.6086903810501099, "learning_rate": 5.799711121810304e-05, "loss": 1.507, "step": 8724 }, { "epoch": 4.20216606498195, "grad_norm": 1.4008262157440186, "learning_rate": 5.796822339913337e-05, "loss": 1.4915, "step": 8730 }, { "epoch": 4.205054151624549, "grad_norm": 1.459985613822937, "learning_rate": 5.79393355801637e-05, "loss": 1.398, "step": 8736 }, { "epoch": 4.207942238267148, "grad_norm": 1.5876463651657104, "learning_rate": 5.7910447761194034e-05, "loss": 1.4019, "step": 8742 }, { "epoch": 4.210830324909748, "grad_norm": 1.475845217704773, "learning_rate": 5.788155994222436e-05, "loss": 1.4685, "step": 8748 }, { "epoch": 4.213718411552347, "grad_norm": 1.6649187803268433, "learning_rate": 5.785267212325469e-05, "loss": 1.4118, "step": 8754 }, { "epoch": 4.216606498194946, "grad_norm": 1.4283579587936401, "learning_rate": 5.7823784304285036e-05, "loss": 1.4122, "step": 8760 }, { "epoch": 4.219494584837545, "grad_norm": 1.4697599411010742, "learning_rate": 5.779489648531536e-05, "loss": 1.5015, "step": 8766 }, { "epoch": 4.222382671480144, "grad_norm": 1.707942008972168, "learning_rate": 5.776600866634569e-05, "loss": 1.4145, "step": 8772 }, { "epoch": 4.225270758122743, "grad_norm": 1.509848713874817, "learning_rate": 5.7737120847376024e-05, "loss": 1.4207, "step": 8778 }, { "epoch": 4.228158844765343, "grad_norm": 1.780031442642212, "learning_rate": 5.7708233028406356e-05, "loss": 1.5637, "step": 8784 }, { "epoch": 4.231046931407942, "grad_norm": 1.8217121362686157, "learning_rate": 5.767934520943669e-05, "loss": 1.5933, "step": 8790 }, { "epoch": 4.233935018050541, "grad_norm": 1.7729730606079102, "learning_rate": 5.7650457390467026e-05, "loss": 1.4196, "step": 8796 }, { "epoch": 4.236823104693141, "grad_norm": 1.7698053121566772, "learning_rate": 5.762156957149736e-05, "loss": 1.4205, "step": 8802 }, { "epoch": 4.23971119133574, "grad_norm": 1.5443717241287231, "learning_rate": 5.759268175252769e-05, "loss": 1.3968, "step": 8808 }, { "epoch": 4.242599277978339, "grad_norm": 1.4556883573532104, "learning_rate": 5.756379393355802e-05, "loss": 1.438, "step": 8814 }, { "epoch": 4.245487364620939, "grad_norm": 1.509716510772705, "learning_rate": 5.753490611458835e-05, "loss": 1.5343, "step": 8820 }, { "epoch": 4.248375451263538, "grad_norm": 1.5431232452392578, "learning_rate": 5.750601829561868e-05, "loss": 1.4181, "step": 8826 }, { "epoch": 4.251263537906137, "grad_norm": 1.5298374891281128, "learning_rate": 5.747713047664901e-05, "loss": 1.45, "step": 8832 }, { "epoch": 4.2541516245487365, "grad_norm": 1.5995231866836548, "learning_rate": 5.7448242657679353e-05, "loss": 1.3873, "step": 8838 }, { "epoch": 4.257039711191336, "grad_norm": 1.6229076385498047, "learning_rate": 5.7419354838709685e-05, "loss": 1.4067, "step": 8844 }, { "epoch": 4.259927797833935, "grad_norm": 1.5539122819900513, "learning_rate": 5.739046701974001e-05, "loss": 1.3197, "step": 8850 }, { "epoch": 4.2628158844765345, "grad_norm": 1.8184165954589844, "learning_rate": 5.736157920077034e-05, "loss": 1.4066, "step": 8856 }, { "epoch": 4.265703971119134, "grad_norm": 1.7610480785369873, "learning_rate": 5.733269138180067e-05, "loss": 1.5089, "step": 8862 }, { "epoch": 4.268592057761733, "grad_norm": 1.8405461311340332, "learning_rate": 5.7303803562831005e-05, "loss": 1.5206, "step": 8868 }, { "epoch": 4.271480144404332, "grad_norm": 1.6631218194961548, "learning_rate": 5.727491574386134e-05, "loss": 1.3852, "step": 8874 }, { "epoch": 4.274368231046932, "grad_norm": 1.7204480171203613, "learning_rate": 5.7246027924891675e-05, "loss": 1.4988, "step": 8880 }, { "epoch": 4.277256317689531, "grad_norm": 1.4067143201828003, "learning_rate": 5.7217140105922006e-05, "loss": 1.3906, "step": 8886 }, { "epoch": 4.28014440433213, "grad_norm": 1.669914722442627, "learning_rate": 5.718825228695234e-05, "loss": 1.4517, "step": 8892 }, { "epoch": 4.28303249097473, "grad_norm": 1.4268051385879517, "learning_rate": 5.715936446798267e-05, "loss": 1.4833, "step": 8898 }, { "epoch": 4.285920577617328, "grad_norm": 1.7413396835327148, "learning_rate": 5.7130476649013e-05, "loss": 1.3277, "step": 8904 }, { "epoch": 4.2888086642599275, "grad_norm": 1.869745135307312, "learning_rate": 5.710158883004334e-05, "loss": 1.4482, "step": 8910 }, { "epoch": 4.291696750902527, "grad_norm": 1.8538944721221924, "learning_rate": 5.707270101107367e-05, "loss": 1.4419, "step": 8916 }, { "epoch": 4.294584837545126, "grad_norm": 1.546785831451416, "learning_rate": 5.7043813192104e-05, "loss": 1.3637, "step": 8922 }, { "epoch": 4.297472924187725, "grad_norm": 1.6526366472244263, "learning_rate": 5.7014925373134334e-05, "loss": 1.5787, "step": 8928 }, { "epoch": 4.300361010830325, "grad_norm": 1.8506531715393066, "learning_rate": 5.698603755416466e-05, "loss": 1.5486, "step": 8934 }, { "epoch": 4.303249097472924, "grad_norm": 2.049315929412842, "learning_rate": 5.695714973519499e-05, "loss": 1.4305, "step": 8940 }, { "epoch": 4.306137184115523, "grad_norm": 1.675057291984558, "learning_rate": 5.692826191622532e-05, "loss": 1.4907, "step": 8946 }, { "epoch": 4.309025270758123, "grad_norm": 1.872334361076355, "learning_rate": 5.689937409725567e-05, "loss": 1.4617, "step": 8952 }, { "epoch": 4.311913357400722, "grad_norm": 1.6291019916534424, "learning_rate": 5.687048627828599e-05, "loss": 1.4988, "step": 8958 }, { "epoch": 4.314801444043321, "grad_norm": 1.7041306495666504, "learning_rate": 5.6841598459316324e-05, "loss": 1.292, "step": 8964 }, { "epoch": 4.317689530685921, "grad_norm": 1.5324870347976685, "learning_rate": 5.6812710640346655e-05, "loss": 1.4945, "step": 8970 }, { "epoch": 4.32057761732852, "grad_norm": 1.5529515743255615, "learning_rate": 5.678382282137699e-05, "loss": 1.4235, "step": 8976 }, { "epoch": 4.323465703971119, "grad_norm": 1.7138432264328003, "learning_rate": 5.675493500240732e-05, "loss": 1.5024, "step": 8982 }, { "epoch": 4.326353790613719, "grad_norm": 1.7132987976074219, "learning_rate": 5.672604718343766e-05, "loss": 1.5104, "step": 8988 }, { "epoch": 4.329241877256318, "grad_norm": 1.5370500087738037, "learning_rate": 5.669715936446799e-05, "loss": 1.5101, "step": 8994 }, { "epoch": 4.332129963898917, "grad_norm": 1.563990592956543, "learning_rate": 5.666827154549832e-05, "loss": 1.4713, "step": 9000 }, { "epoch": 4.335018050541517, "grad_norm": 1.632399320602417, "learning_rate": 5.663938372652865e-05, "loss": 1.3429, "step": 9006 }, { "epoch": 4.337906137184116, "grad_norm": 1.7472485303878784, "learning_rate": 5.661049590755898e-05, "loss": 1.5486, "step": 9012 }, { "epoch": 4.340794223826715, "grad_norm": 1.7046183347702026, "learning_rate": 5.658160808858931e-05, "loss": 1.4564, "step": 9018 }, { "epoch": 4.343682310469314, "grad_norm": 1.6135555505752563, "learning_rate": 5.655272026961964e-05, "loss": 1.5037, "step": 9024 }, { "epoch": 4.346570397111913, "grad_norm": 1.5906718969345093, "learning_rate": 5.6523832450649985e-05, "loss": 1.5493, "step": 9030 }, { "epoch": 4.349458483754512, "grad_norm": 1.6016156673431396, "learning_rate": 5.6494944631680316e-05, "loss": 1.5504, "step": 9036 }, { "epoch": 4.352346570397112, "grad_norm": 1.591470718383789, "learning_rate": 5.646605681271064e-05, "loss": 1.5179, "step": 9042 }, { "epoch": 4.355234657039711, "grad_norm": 1.8730591535568237, "learning_rate": 5.643716899374097e-05, "loss": 1.5929, "step": 9048 }, { "epoch": 4.35812274368231, "grad_norm": 1.90233314037323, "learning_rate": 5.6408281174771304e-05, "loss": 1.3865, "step": 9054 }, { "epoch": 4.3610108303249095, "grad_norm": 1.4737846851348877, "learning_rate": 5.6379393355801636e-05, "loss": 1.4932, "step": 9060 }, { "epoch": 4.363898916967509, "grad_norm": 1.6005144119262695, "learning_rate": 5.6350505536831974e-05, "loss": 1.4599, "step": 9066 }, { "epoch": 4.366787003610108, "grad_norm": 1.5976680517196655, "learning_rate": 5.6321617717862306e-05, "loss": 1.4245, "step": 9072 }, { "epoch": 4.3696750902527075, "grad_norm": 1.507001519203186, "learning_rate": 5.629272989889264e-05, "loss": 1.2839, "step": 9078 }, { "epoch": 4.372563176895307, "grad_norm": 1.8390591144561768, "learning_rate": 5.626384207992297e-05, "loss": 1.41, "step": 9084 }, { "epoch": 4.375451263537906, "grad_norm": 1.7298332452774048, "learning_rate": 5.62349542609533e-05, "loss": 1.5863, "step": 9090 }, { "epoch": 4.3783393501805055, "grad_norm": 1.573898196220398, "learning_rate": 5.620606644198363e-05, "loss": 1.4118, "step": 9096 }, { "epoch": 4.381227436823105, "grad_norm": 1.811855435371399, "learning_rate": 5.617717862301396e-05, "loss": 1.487, "step": 9102 }, { "epoch": 4.384115523465704, "grad_norm": 1.55027174949646, "learning_rate": 5.61482908040443e-05, "loss": 1.5078, "step": 9108 }, { "epoch": 4.387003610108303, "grad_norm": 1.7810839414596558, "learning_rate": 5.6119402985074634e-05, "loss": 1.5177, "step": 9114 }, { "epoch": 4.389891696750903, "grad_norm": 1.6995246410369873, "learning_rate": 5.609051516610496e-05, "loss": 1.3558, "step": 9120 }, { "epoch": 4.392779783393502, "grad_norm": 1.6509652137756348, "learning_rate": 5.606162734713529e-05, "loss": 1.4263, "step": 9126 }, { "epoch": 4.395667870036101, "grad_norm": 1.7169989347457886, "learning_rate": 5.603273952816562e-05, "loss": 1.6184, "step": 9132 }, { "epoch": 4.398555956678701, "grad_norm": 1.903829574584961, "learning_rate": 5.6003851709195954e-05, "loss": 1.5542, "step": 9138 }, { "epoch": 4.4014440433213, "grad_norm": 1.6400285959243774, "learning_rate": 5.597496389022629e-05, "loss": 1.4157, "step": 9144 }, { "epoch": 4.404332129963899, "grad_norm": 1.5079020261764526, "learning_rate": 5.5946076071256624e-05, "loss": 1.5349, "step": 9150 }, { "epoch": 4.407220216606499, "grad_norm": 1.9373408555984497, "learning_rate": 5.5917188252286955e-05, "loss": 1.6024, "step": 9156 }, { "epoch": 4.410108303249097, "grad_norm": 1.7451376914978027, "learning_rate": 5.588830043331729e-05, "loss": 1.5143, "step": 9162 }, { "epoch": 4.412996389891696, "grad_norm": 1.9356104135513306, "learning_rate": 5.585941261434762e-05, "loss": 1.5062, "step": 9168 }, { "epoch": 4.415884476534296, "grad_norm": 1.4982830286026, "learning_rate": 5.583052479537795e-05, "loss": 1.4954, "step": 9174 }, { "epoch": 4.418772563176895, "grad_norm": 1.6297886371612549, "learning_rate": 5.580163697640829e-05, "loss": 1.3941, "step": 9180 }, { "epoch": 4.421660649819494, "grad_norm": 1.8183754682540894, "learning_rate": 5.577274915743862e-05, "loss": 1.4009, "step": 9186 }, { "epoch": 4.424548736462094, "grad_norm": 1.594931721687317, "learning_rate": 5.574386133846895e-05, "loss": 1.4866, "step": 9192 }, { "epoch": 4.427436823104693, "grad_norm": 1.69497549533844, "learning_rate": 5.571497351949928e-05, "loss": 1.4225, "step": 9198 }, { "epoch": 4.430324909747292, "grad_norm": 1.618679165840149, "learning_rate": 5.568608570052961e-05, "loss": 1.552, "step": 9204 }, { "epoch": 4.433212996389892, "grad_norm": 1.8175188302993774, "learning_rate": 5.565719788155994e-05, "loss": 1.4269, "step": 9210 }, { "epoch": 4.436101083032491, "grad_norm": 1.5205835103988647, "learning_rate": 5.562831006259027e-05, "loss": 1.4365, "step": 9216 }, { "epoch": 4.43898916967509, "grad_norm": 1.6575679779052734, "learning_rate": 5.5599422243620616e-05, "loss": 1.446, "step": 9222 }, { "epoch": 4.44187725631769, "grad_norm": 1.6318070888519287, "learning_rate": 5.557053442465094e-05, "loss": 1.3628, "step": 9228 }, { "epoch": 4.444765342960289, "grad_norm": 1.6535283327102661, "learning_rate": 5.554164660568127e-05, "loss": 1.3825, "step": 9234 }, { "epoch": 4.447653429602888, "grad_norm": 1.847111463546753, "learning_rate": 5.5512758786711604e-05, "loss": 1.425, "step": 9240 }, { "epoch": 4.4505415162454875, "grad_norm": 1.736524224281311, "learning_rate": 5.5483870967741936e-05, "loss": 1.4276, "step": 9246 }, { "epoch": 4.453429602888087, "grad_norm": 1.6557468175888062, "learning_rate": 5.545498314877227e-05, "loss": 1.403, "step": 9252 }, { "epoch": 4.456317689530686, "grad_norm": 1.5843019485473633, "learning_rate": 5.5426095329802606e-05, "loss": 1.5629, "step": 9258 }, { "epoch": 4.4592057761732855, "grad_norm": 1.6323384046554565, "learning_rate": 5.539720751083294e-05, "loss": 1.5658, "step": 9264 }, { "epoch": 4.462093862815885, "grad_norm": 1.6166220903396606, "learning_rate": 5.536831969186327e-05, "loss": 1.5843, "step": 9270 }, { "epoch": 4.464981949458484, "grad_norm": 1.837344765663147, "learning_rate": 5.53394318728936e-05, "loss": 1.5197, "step": 9276 }, { "epoch": 4.467870036101083, "grad_norm": 1.6901663541793823, "learning_rate": 5.531054405392393e-05, "loss": 1.4853, "step": 9282 }, { "epoch": 4.470758122743682, "grad_norm": 1.9212669134140015, "learning_rate": 5.528165623495426e-05, "loss": 1.334, "step": 9288 }, { "epoch": 4.473646209386281, "grad_norm": 1.685388207435608, "learning_rate": 5.525276841598459e-05, "loss": 1.5215, "step": 9294 }, { "epoch": 4.4765342960288805, "grad_norm": 1.6588802337646484, "learning_rate": 5.5223880597014934e-05, "loss": 1.4299, "step": 9300 }, { "epoch": 4.47942238267148, "grad_norm": 1.871759057044983, "learning_rate": 5.5194992778045265e-05, "loss": 1.4823, "step": 9306 }, { "epoch": 4.482310469314079, "grad_norm": 1.6634881496429443, "learning_rate": 5.516610495907559e-05, "loss": 1.2605, "step": 9312 }, { "epoch": 4.4851985559566785, "grad_norm": 1.7368061542510986, "learning_rate": 5.513721714010592e-05, "loss": 1.4792, "step": 9318 }, { "epoch": 4.488086642599278, "grad_norm": 2.0436112880706787, "learning_rate": 5.5108329321136253e-05, "loss": 1.4499, "step": 9324 }, { "epoch": 4.490974729241877, "grad_norm": 1.9877756834030151, "learning_rate": 5.5079441502166585e-05, "loss": 1.51, "step": 9330 }, { "epoch": 4.4938628158844764, "grad_norm": 1.7253186702728271, "learning_rate": 5.505055368319692e-05, "loss": 1.3984, "step": 9336 }, { "epoch": 4.496750902527076, "grad_norm": 1.4645874500274658, "learning_rate": 5.5021665864227255e-05, "loss": 1.5101, "step": 9342 }, { "epoch": 4.499638989169675, "grad_norm": 1.8681585788726807, "learning_rate": 5.4992778045257587e-05, "loss": 1.4815, "step": 9348 }, { "epoch": 4.502527075812274, "grad_norm": 1.3689117431640625, "learning_rate": 5.496389022628792e-05, "loss": 1.524, "step": 9354 }, { "epoch": 4.505415162454874, "grad_norm": 1.8421396017074585, "learning_rate": 5.493500240731825e-05, "loss": 1.4463, "step": 9360 }, { "epoch": 4.508303249097473, "grad_norm": 1.8004907369613647, "learning_rate": 5.490611458834858e-05, "loss": 1.3948, "step": 9366 }, { "epoch": 4.511191335740072, "grad_norm": 1.8060587644577026, "learning_rate": 5.487722676937892e-05, "loss": 1.4625, "step": 9372 }, { "epoch": 4.514079422382672, "grad_norm": 1.668420433998108, "learning_rate": 5.484833895040925e-05, "loss": 1.4216, "step": 9378 }, { "epoch": 4.516967509025271, "grad_norm": 1.8279979228973389, "learning_rate": 5.481945113143958e-05, "loss": 1.5199, "step": 9384 }, { "epoch": 4.51985559566787, "grad_norm": 1.526413083076477, "learning_rate": 5.4790563312469914e-05, "loss": 1.3356, "step": 9390 }, { "epoch": 4.52274368231047, "grad_norm": 1.6100770235061646, "learning_rate": 5.476167549350024e-05, "loss": 1.4125, "step": 9396 }, { "epoch": 4.525631768953069, "grad_norm": 1.6549354791641235, "learning_rate": 5.473278767453057e-05, "loss": 1.5174, "step": 9402 }, { "epoch": 4.528519855595668, "grad_norm": 1.7690056562423706, "learning_rate": 5.47038998555609e-05, "loss": 1.5016, "step": 9408 }, { "epoch": 4.531407942238268, "grad_norm": 1.5413516759872437, "learning_rate": 5.467501203659125e-05, "loss": 1.4267, "step": 9414 }, { "epoch": 4.534296028880867, "grad_norm": 1.622559666633606, "learning_rate": 5.464612421762157e-05, "loss": 1.4433, "step": 9420 }, { "epoch": 4.537184115523465, "grad_norm": 1.778174877166748, "learning_rate": 5.4617236398651904e-05, "loss": 1.4509, "step": 9426 }, { "epoch": 4.540072202166065, "grad_norm": 1.3961842060089111, "learning_rate": 5.4588348579682236e-05, "loss": 1.3933, "step": 9432 }, { "epoch": 4.542960288808664, "grad_norm": 1.6277543306350708, "learning_rate": 5.455946076071257e-05, "loss": 1.4046, "step": 9438 }, { "epoch": 4.545848375451263, "grad_norm": 1.5735429525375366, "learning_rate": 5.45305729417429e-05, "loss": 1.5233, "step": 9444 }, { "epoch": 4.548736462093863, "grad_norm": 1.7940515279769897, "learning_rate": 5.450168512277324e-05, "loss": 1.4729, "step": 9450 }, { "epoch": 4.551624548736462, "grad_norm": 1.885544776916504, "learning_rate": 5.447279730380357e-05, "loss": 1.4493, "step": 9456 }, { "epoch": 4.554512635379061, "grad_norm": 2.1958835124969482, "learning_rate": 5.44439094848339e-05, "loss": 1.5431, "step": 9462 }, { "epoch": 4.557400722021661, "grad_norm": 1.7695215940475464, "learning_rate": 5.441502166586423e-05, "loss": 1.5892, "step": 9468 }, { "epoch": 4.56028880866426, "grad_norm": 1.927868366241455, "learning_rate": 5.4386133846894564e-05, "loss": 1.4764, "step": 9474 }, { "epoch": 4.563176895306859, "grad_norm": 1.688446044921875, "learning_rate": 5.435724602792489e-05, "loss": 1.5092, "step": 9480 }, { "epoch": 4.5660649819494585, "grad_norm": 2.230762481689453, "learning_rate": 5.432835820895522e-05, "loss": 1.5608, "step": 9486 }, { "epoch": 4.568953068592058, "grad_norm": 1.5881056785583496, "learning_rate": 5.4299470389985565e-05, "loss": 1.4303, "step": 9492 }, { "epoch": 4.571841155234657, "grad_norm": 1.8297187089920044, "learning_rate": 5.427058257101589e-05, "loss": 1.6535, "step": 9498 }, { "epoch": 4.5747292418772565, "grad_norm": 1.9879727363586426, "learning_rate": 5.424169475204622e-05, "loss": 1.5715, "step": 9504 }, { "epoch": 4.577617328519856, "grad_norm": 1.6415268182754517, "learning_rate": 5.421280693307655e-05, "loss": 1.5046, "step": 9510 }, { "epoch": 4.580505415162455, "grad_norm": 2.1128270626068115, "learning_rate": 5.4183919114106885e-05, "loss": 1.4155, "step": 9516 }, { "epoch": 4.5833935018050544, "grad_norm": 1.764696478843689, "learning_rate": 5.4155031295137216e-05, "loss": 1.4763, "step": 9522 }, { "epoch": 4.586281588447654, "grad_norm": 1.6416302919387817, "learning_rate": 5.4126143476167555e-05, "loss": 1.4315, "step": 9528 }, { "epoch": 4.589169675090253, "grad_norm": 1.7634718418121338, "learning_rate": 5.4097255657197886e-05, "loss": 1.5054, "step": 9534 }, { "epoch": 4.5920577617328515, "grad_norm": 1.4948071241378784, "learning_rate": 5.406836783822822e-05, "loss": 1.5497, "step": 9540 }, { "epoch": 4.594945848375451, "grad_norm": 1.7105985879898071, "learning_rate": 5.403948001925855e-05, "loss": 1.3874, "step": 9546 }, { "epoch": 4.59783393501805, "grad_norm": 1.6949584484100342, "learning_rate": 5.401059220028888e-05, "loss": 1.3458, "step": 9552 }, { "epoch": 4.6007220216606495, "grad_norm": 1.6944713592529297, "learning_rate": 5.3981704381319206e-05, "loss": 1.4146, "step": 9558 }, { "epoch": 4.603610108303249, "grad_norm": 1.967438817024231, "learning_rate": 5.395281656234955e-05, "loss": 1.413, "step": 9564 }, { "epoch": 4.606498194945848, "grad_norm": 1.5702255964279175, "learning_rate": 5.392392874337988e-05, "loss": 1.4504, "step": 9570 }, { "epoch": 4.609386281588447, "grad_norm": 1.758432388305664, "learning_rate": 5.3895040924410214e-05, "loss": 1.6061, "step": 9576 }, { "epoch": 4.612274368231047, "grad_norm": 1.8751901388168335, "learning_rate": 5.386615310544054e-05, "loss": 1.4571, "step": 9582 }, { "epoch": 4.615162454873646, "grad_norm": 1.5731037855148315, "learning_rate": 5.383726528647087e-05, "loss": 1.5664, "step": 9588 }, { "epoch": 4.618050541516245, "grad_norm": 1.5141605138778687, "learning_rate": 5.38083774675012e-05, "loss": 1.5261, "step": 9594 }, { "epoch": 4.620938628158845, "grad_norm": 1.7852624654769897, "learning_rate": 5.3779489648531534e-05, "loss": 1.4412, "step": 9600 }, { "epoch": 4.623826714801444, "grad_norm": 1.6854467391967773, "learning_rate": 5.375060182956187e-05, "loss": 1.3936, "step": 9606 }, { "epoch": 4.626714801444043, "grad_norm": 1.8110824823379517, "learning_rate": 5.3721714010592204e-05, "loss": 1.4346, "step": 9612 }, { "epoch": 4.629602888086643, "grad_norm": 1.6684165000915527, "learning_rate": 5.3692826191622535e-05, "loss": 1.4315, "step": 9618 }, { "epoch": 4.632490974729242, "grad_norm": 1.730546474456787, "learning_rate": 5.366393837265287e-05, "loss": 1.447, "step": 9624 }, { "epoch": 4.635379061371841, "grad_norm": 1.6766997575759888, "learning_rate": 5.36350505536832e-05, "loss": 1.5172, "step": 9630 }, { "epoch": 4.638267148014441, "grad_norm": 1.752596378326416, "learning_rate": 5.360616273471353e-05, "loss": 1.5703, "step": 9636 }, { "epoch": 4.64115523465704, "grad_norm": 1.516764521598816, "learning_rate": 5.357727491574387e-05, "loss": 1.4207, "step": 9642 }, { "epoch": 4.644043321299639, "grad_norm": 1.6425399780273438, "learning_rate": 5.35483870967742e-05, "loss": 1.5561, "step": 9648 }, { "epoch": 4.646931407942239, "grad_norm": 1.6749035120010376, "learning_rate": 5.351949927780453e-05, "loss": 1.4242, "step": 9654 }, { "epoch": 4.649819494584838, "grad_norm": 1.4819715023040771, "learning_rate": 5.3490611458834863e-05, "loss": 1.5135, "step": 9660 }, { "epoch": 4.652707581227437, "grad_norm": 1.80519437789917, "learning_rate": 5.346172363986519e-05, "loss": 1.3649, "step": 9666 }, { "epoch": 4.6555956678700365, "grad_norm": 1.6599200963974, "learning_rate": 5.343283582089552e-05, "loss": 1.4734, "step": 9672 }, { "epoch": 4.658483754512636, "grad_norm": 1.5368373394012451, "learning_rate": 5.340394800192585e-05, "loss": 1.4531, "step": 9678 }, { "epoch": 4.661371841155234, "grad_norm": 2.1384847164154053, "learning_rate": 5.3375060182956197e-05, "loss": 1.45, "step": 9684 }, { "epoch": 4.664259927797834, "grad_norm": 1.7415045499801636, "learning_rate": 5.334617236398652e-05, "loss": 1.3706, "step": 9690 }, { "epoch": 4.667148014440433, "grad_norm": 1.7436102628707886, "learning_rate": 5.331728454501685e-05, "loss": 1.4278, "step": 9696 }, { "epoch": 4.670036101083032, "grad_norm": 1.6706188917160034, "learning_rate": 5.3288396726047185e-05, "loss": 1.5001, "step": 9702 }, { "epoch": 4.672924187725632, "grad_norm": 1.6849429607391357, "learning_rate": 5.3259508907077516e-05, "loss": 1.4343, "step": 9708 }, { "epoch": 4.675812274368231, "grad_norm": 1.7769701480865479, "learning_rate": 5.323062108810785e-05, "loss": 1.4805, "step": 9714 }, { "epoch": 4.67870036101083, "grad_norm": 1.4511619806289673, "learning_rate": 5.3201733269138186e-05, "loss": 1.4884, "step": 9720 }, { "epoch": 4.6815884476534295, "grad_norm": 1.520058274269104, "learning_rate": 5.317284545016852e-05, "loss": 1.4967, "step": 9726 }, { "epoch": 4.684476534296029, "grad_norm": 1.7556337118148804, "learning_rate": 5.314395763119885e-05, "loss": 1.3493, "step": 9732 }, { "epoch": 4.687364620938628, "grad_norm": 1.6768261194229126, "learning_rate": 5.311506981222918e-05, "loss": 1.4091, "step": 9738 }, { "epoch": 4.6902527075812275, "grad_norm": 1.4439074993133545, "learning_rate": 5.308618199325951e-05, "loss": 1.4622, "step": 9744 }, { "epoch": 4.693140794223827, "grad_norm": 1.722645878791809, "learning_rate": 5.305729417428984e-05, "loss": 1.3739, "step": 9750 }, { "epoch": 4.696028880866426, "grad_norm": 1.792325735092163, "learning_rate": 5.302840635532017e-05, "loss": 1.3588, "step": 9756 }, { "epoch": 4.698916967509025, "grad_norm": 1.6889091730117798, "learning_rate": 5.2999518536350514e-05, "loss": 1.4995, "step": 9762 }, { "epoch": 4.701805054151625, "grad_norm": 1.7436286211013794, "learning_rate": 5.2970630717380846e-05, "loss": 1.5922, "step": 9768 }, { "epoch": 4.704693140794224, "grad_norm": 1.7228326797485352, "learning_rate": 5.294174289841117e-05, "loss": 1.3615, "step": 9774 }, { "epoch": 4.707581227436823, "grad_norm": 2.043823480606079, "learning_rate": 5.29128550794415e-05, "loss": 1.4896, "step": 9780 }, { "epoch": 4.710469314079423, "grad_norm": 1.7681926488876343, "learning_rate": 5.2883967260471834e-05, "loss": 1.3888, "step": 9786 }, { "epoch": 4.713357400722022, "grad_norm": 1.6977531909942627, "learning_rate": 5.2855079441502165e-05, "loss": 1.4375, "step": 9792 }, { "epoch": 4.7162454873646205, "grad_norm": 1.9429376125335693, "learning_rate": 5.2826191622532504e-05, "loss": 1.5145, "step": 9798 }, { "epoch": 4.71913357400722, "grad_norm": 1.6318238973617554, "learning_rate": 5.2797303803562835e-05, "loss": 1.3205, "step": 9804 }, { "epoch": 4.722021660649819, "grad_norm": 1.835802435874939, "learning_rate": 5.276841598459317e-05, "loss": 1.4839, "step": 9810 }, { "epoch": 4.724909747292418, "grad_norm": 1.6703810691833496, "learning_rate": 5.27395281656235e-05, "loss": 1.4048, "step": 9816 }, { "epoch": 4.727797833935018, "grad_norm": 1.7998586893081665, "learning_rate": 5.271064034665383e-05, "loss": 1.4814, "step": 9822 }, { "epoch": 4.730685920577617, "grad_norm": 1.9564024209976196, "learning_rate": 5.268175252768416e-05, "loss": 1.382, "step": 9828 }, { "epoch": 4.733574007220216, "grad_norm": 1.9755386114120483, "learning_rate": 5.26528647087145e-05, "loss": 1.3448, "step": 9834 }, { "epoch": 4.736462093862816, "grad_norm": 1.7144132852554321, "learning_rate": 5.262397688974483e-05, "loss": 1.5541, "step": 9840 }, { "epoch": 4.739350180505415, "grad_norm": 1.9930962324142456, "learning_rate": 5.259508907077516e-05, "loss": 1.3482, "step": 9846 }, { "epoch": 4.742238267148014, "grad_norm": 1.810982584953308, "learning_rate": 5.256620125180549e-05, "loss": 1.576, "step": 9852 }, { "epoch": 4.745126353790614, "grad_norm": 1.8291996717453003, "learning_rate": 5.253731343283582e-05, "loss": 1.3761, "step": 9858 }, { "epoch": 4.748014440433213, "grad_norm": 1.922634482383728, "learning_rate": 5.250842561386615e-05, "loss": 1.4655, "step": 9864 }, { "epoch": 4.750902527075812, "grad_norm": 1.7774097919464111, "learning_rate": 5.247953779489648e-05, "loss": 1.4643, "step": 9870 }, { "epoch": 4.753790613718412, "grad_norm": 1.6503722667694092, "learning_rate": 5.245064997592682e-05, "loss": 1.3222, "step": 9876 }, { "epoch": 4.756678700361011, "grad_norm": 1.9438482522964478, "learning_rate": 5.242176215695715e-05, "loss": 1.5228, "step": 9882 }, { "epoch": 4.75956678700361, "grad_norm": 1.7219537496566772, "learning_rate": 5.2392874337987484e-05, "loss": 1.4703, "step": 9888 }, { "epoch": 4.76245487364621, "grad_norm": 1.8766907453536987, "learning_rate": 5.2363986519017816e-05, "loss": 1.4265, "step": 9894 }, { "epoch": 4.765342960288809, "grad_norm": 1.75528883934021, "learning_rate": 5.233509870004815e-05, "loss": 1.5039, "step": 9900 }, { "epoch": 4.768231046931408, "grad_norm": 1.657808542251587, "learning_rate": 5.230621088107848e-05, "loss": 1.5231, "step": 9906 }, { "epoch": 4.7711191335740075, "grad_norm": 1.679816722869873, "learning_rate": 5.227732306210882e-05, "loss": 1.3617, "step": 9912 }, { "epoch": 4.774007220216607, "grad_norm": 1.6203405857086182, "learning_rate": 5.224843524313915e-05, "loss": 1.5331, "step": 9918 }, { "epoch": 4.776895306859206, "grad_norm": 1.549669623374939, "learning_rate": 5.221954742416948e-05, "loss": 1.483, "step": 9924 }, { "epoch": 4.7797833935018055, "grad_norm": 1.7331022024154663, "learning_rate": 5.219065960519981e-05, "loss": 1.4742, "step": 9930 }, { "epoch": 4.782671480144405, "grad_norm": 1.8189924955368042, "learning_rate": 5.216177178623014e-05, "loss": 1.3581, "step": 9936 }, { "epoch": 4.785559566787003, "grad_norm": 1.6937721967697144, "learning_rate": 5.213288396726047e-05, "loss": 1.4103, "step": 9942 }, { "epoch": 4.7884476534296025, "grad_norm": 1.8959146738052368, "learning_rate": 5.21039961482908e-05, "loss": 1.4093, "step": 9948 }, { "epoch": 4.791335740072202, "grad_norm": 1.5573374032974243, "learning_rate": 5.2075108329321145e-05, "loss": 1.3581, "step": 9954 }, { "epoch": 4.794223826714801, "grad_norm": 1.790752649307251, "learning_rate": 5.204622051035147e-05, "loss": 1.4013, "step": 9960 }, { "epoch": 4.7971119133574005, "grad_norm": 2.056004762649536, "learning_rate": 5.20173326913818e-05, "loss": 1.5989, "step": 9966 }, { "epoch": 4.8, "grad_norm": 1.7498806715011597, "learning_rate": 5.1988444872412134e-05, "loss": 1.3658, "step": 9972 }, { "epoch": 4.802888086642599, "grad_norm": 1.7703509330749512, "learning_rate": 5.1959557053442465e-05, "loss": 1.4478, "step": 9978 }, { "epoch": 4.8057761732851985, "grad_norm": 1.846375584602356, "learning_rate": 5.19306692344728e-05, "loss": 1.4157, "step": 9984 }, { "epoch": 4.808664259927798, "grad_norm": 1.5509352684020996, "learning_rate": 5.1901781415503135e-05, "loss": 1.5404, "step": 9990 }, { "epoch": 4.811552346570397, "grad_norm": 1.6880102157592773, "learning_rate": 5.187289359653347e-05, "loss": 1.5267, "step": 9996 }, { "epoch": 4.814440433212996, "grad_norm": 1.7758433818817139, "learning_rate": 5.18440057775638e-05, "loss": 1.4681, "step": 10002 }, { "epoch": 4.817328519855596, "grad_norm": 1.7554033994674683, "learning_rate": 5.181511795859413e-05, "loss": 1.6025, "step": 10008 }, { "epoch": 4.820216606498195, "grad_norm": 1.6940950155258179, "learning_rate": 5.178623013962446e-05, "loss": 1.4669, "step": 10014 }, { "epoch": 4.823104693140794, "grad_norm": 1.516883373260498, "learning_rate": 5.1757342320654786e-05, "loss": 1.4401, "step": 10020 }, { "epoch": 4.825992779783394, "grad_norm": 1.7707339525222778, "learning_rate": 5.172845450168513e-05, "loss": 1.4268, "step": 10026 }, { "epoch": 4.828880866425993, "grad_norm": 1.9699839353561401, "learning_rate": 5.169956668271546e-05, "loss": 1.5255, "step": 10032 }, { "epoch": 4.831768953068592, "grad_norm": 1.8428404331207275, "learning_rate": 5.1670678863745795e-05, "loss": 1.4885, "step": 10038 }, { "epoch": 4.834657039711192, "grad_norm": 1.6280450820922852, "learning_rate": 5.164179104477612e-05, "loss": 1.4275, "step": 10044 }, { "epoch": 4.837545126353791, "grad_norm": 1.8140079975128174, "learning_rate": 5.161290322580645e-05, "loss": 1.3047, "step": 10050 }, { "epoch": 4.840433212996389, "grad_norm": 1.8450980186462402, "learning_rate": 5.158401540683678e-05, "loss": 1.426, "step": 10056 }, { "epoch": 4.843321299638989, "grad_norm": 1.3705717325210571, "learning_rate": 5.1555127587867114e-05, "loss": 1.3611, "step": 10062 }, { "epoch": 4.846209386281588, "grad_norm": 1.5663145780563354, "learning_rate": 5.152623976889745e-05, "loss": 1.4092, "step": 10068 }, { "epoch": 4.849097472924187, "grad_norm": 1.74745512008667, "learning_rate": 5.1497351949927784e-05, "loss": 1.4206, "step": 10074 }, { "epoch": 4.851985559566787, "grad_norm": 1.5562224388122559, "learning_rate": 5.1468464130958116e-05, "loss": 1.3529, "step": 10080 }, { "epoch": 4.854873646209386, "grad_norm": 1.7399548292160034, "learning_rate": 5.143957631198845e-05, "loss": 1.385, "step": 10086 }, { "epoch": 4.857761732851985, "grad_norm": 1.86776602268219, "learning_rate": 5.141068849301878e-05, "loss": 1.4186, "step": 10092 }, { "epoch": 4.860649819494585, "grad_norm": 1.6189292669296265, "learning_rate": 5.138180067404911e-05, "loss": 1.2879, "step": 10098 }, { "epoch": 4.863537906137184, "grad_norm": 1.8929914236068726, "learning_rate": 5.135291285507945e-05, "loss": 1.4662, "step": 10104 }, { "epoch": 4.866425992779783, "grad_norm": 1.8193767070770264, "learning_rate": 5.132402503610978e-05, "loss": 1.3527, "step": 10110 }, { "epoch": 4.869314079422383, "grad_norm": 1.6652731895446777, "learning_rate": 5.129513721714011e-05, "loss": 1.3338, "step": 10116 }, { "epoch": 4.872202166064982, "grad_norm": 1.8011685609817505, "learning_rate": 5.1266249398170444e-05, "loss": 1.5244, "step": 10122 }, { "epoch": 4.875090252707581, "grad_norm": 1.6646233797073364, "learning_rate": 5.123736157920077e-05, "loss": 1.4337, "step": 10128 }, { "epoch": 4.8779783393501805, "grad_norm": 2.006948947906494, "learning_rate": 5.12084737602311e-05, "loss": 1.4993, "step": 10134 }, { "epoch": 4.88086642599278, "grad_norm": 1.7132530212402344, "learning_rate": 5.117958594126143e-05, "loss": 1.497, "step": 10140 }, { "epoch": 4.883754512635379, "grad_norm": 1.7648813724517822, "learning_rate": 5.115069812229178e-05, "loss": 1.4389, "step": 10146 }, { "epoch": 4.8866425992779785, "grad_norm": 1.7599231004714966, "learning_rate": 5.11218103033221e-05, "loss": 1.4478, "step": 10152 }, { "epoch": 4.889530685920578, "grad_norm": 1.8418164253234863, "learning_rate": 5.109292248435243e-05, "loss": 1.5462, "step": 10158 }, { "epoch": 4.892418772563177, "grad_norm": 1.6681691408157349, "learning_rate": 5.1064034665382765e-05, "loss": 1.5786, "step": 10164 }, { "epoch": 4.8953068592057765, "grad_norm": 1.5516400337219238, "learning_rate": 5.1035146846413096e-05, "loss": 1.4357, "step": 10170 }, { "epoch": 4.898194945848376, "grad_norm": 1.6380077600479126, "learning_rate": 5.100625902744343e-05, "loss": 1.4772, "step": 10176 }, { "epoch": 4.901083032490975, "grad_norm": 1.7370389699935913, "learning_rate": 5.0977371208473766e-05, "loss": 1.5053, "step": 10182 }, { "epoch": 4.903971119133574, "grad_norm": 1.6210397481918335, "learning_rate": 5.09484833895041e-05, "loss": 1.533, "step": 10188 }, { "epoch": 4.906859205776174, "grad_norm": 1.888321876525879, "learning_rate": 5.091959557053443e-05, "loss": 1.5205, "step": 10194 }, { "epoch": 4.909747292418773, "grad_norm": 1.8798712491989136, "learning_rate": 5.089070775156476e-05, "loss": 1.502, "step": 10200 }, { "epoch": 4.9126353790613715, "grad_norm": 1.8763666152954102, "learning_rate": 5.086181993259509e-05, "loss": 1.5679, "step": 10206 }, { "epoch": 4.915523465703971, "grad_norm": 1.5806293487548828, "learning_rate": 5.083293211362542e-05, "loss": 1.4718, "step": 10212 }, { "epoch": 4.91841155234657, "grad_norm": 1.6880789995193481, "learning_rate": 5.080404429465575e-05, "loss": 1.4112, "step": 10218 }, { "epoch": 4.9212996389891694, "grad_norm": 1.7200547456741333, "learning_rate": 5.0775156475686094e-05, "loss": 1.4805, "step": 10224 }, { "epoch": 4.924187725631769, "grad_norm": 1.59951651096344, "learning_rate": 5.074626865671642e-05, "loss": 1.3402, "step": 10230 }, { "epoch": 4.927075812274368, "grad_norm": 1.6900615692138672, "learning_rate": 5.071738083774675e-05, "loss": 1.4021, "step": 10236 }, { "epoch": 4.929963898916967, "grad_norm": 1.716254472732544, "learning_rate": 5.068849301877708e-05, "loss": 1.3708, "step": 10242 }, { "epoch": 4.932851985559567, "grad_norm": 1.8260782957077026, "learning_rate": 5.0659605199807414e-05, "loss": 1.404, "step": 10248 }, { "epoch": 4.935740072202166, "grad_norm": 2.0131442546844482, "learning_rate": 5.0630717380837746e-05, "loss": 1.4794, "step": 10254 }, { "epoch": 4.938628158844765, "grad_norm": 1.735527753829956, "learning_rate": 5.0601829561868084e-05, "loss": 1.5615, "step": 10260 }, { "epoch": 4.941516245487365, "grad_norm": 1.5134721994400024, "learning_rate": 5.0572941742898416e-05, "loss": 1.3088, "step": 10266 }, { "epoch": 4.944404332129964, "grad_norm": 1.7269394397735596, "learning_rate": 5.054405392392875e-05, "loss": 1.5247, "step": 10272 }, { "epoch": 4.947292418772563, "grad_norm": 1.5956223011016846, "learning_rate": 5.051516610495908e-05, "loss": 1.4059, "step": 10278 }, { "epoch": 4.950180505415163, "grad_norm": 1.6663627624511719, "learning_rate": 5.048627828598941e-05, "loss": 1.5972, "step": 10284 }, { "epoch": 4.953068592057762, "grad_norm": 1.6441720724105835, "learning_rate": 5.0457390467019735e-05, "loss": 1.5017, "step": 10290 }, { "epoch": 4.955956678700361, "grad_norm": 1.6909894943237305, "learning_rate": 5.042850264805008e-05, "loss": 1.4872, "step": 10296 }, { "epoch": 4.958844765342961, "grad_norm": 1.69085693359375, "learning_rate": 5.039961482908041e-05, "loss": 1.4999, "step": 10302 }, { "epoch": 4.96173285198556, "grad_norm": 1.5892562866210938, "learning_rate": 5.0370727010110744e-05, "loss": 1.5178, "step": 10308 }, { "epoch": 4.964620938628159, "grad_norm": 1.8232535123825073, "learning_rate": 5.034183919114107e-05, "loss": 1.6112, "step": 10314 }, { "epoch": 4.967509025270758, "grad_norm": 1.6040140390396118, "learning_rate": 5.03129513721714e-05, "loss": 1.3749, "step": 10320 }, { "epoch": 4.970397111913357, "grad_norm": 1.9546236991882324, "learning_rate": 5.028406355320173e-05, "loss": 1.5562, "step": 10326 }, { "epoch": 4.973285198555956, "grad_norm": 1.8370152711868286, "learning_rate": 5.025517573423206e-05, "loss": 1.6396, "step": 10332 }, { "epoch": 4.976173285198556, "grad_norm": 2.182056427001953, "learning_rate": 5.02262879152624e-05, "loss": 1.5834, "step": 10338 }, { "epoch": 4.979061371841155, "grad_norm": 1.7879499197006226, "learning_rate": 5.019740009629273e-05, "loss": 1.4824, "step": 10344 }, { "epoch": 4.981949458483754, "grad_norm": 1.386134386062622, "learning_rate": 5.0168512277323065e-05, "loss": 1.377, "step": 10350 }, { "epoch": 4.984837545126354, "grad_norm": 1.8046499490737915, "learning_rate": 5.0139624458353396e-05, "loss": 1.4398, "step": 10356 }, { "epoch": 4.987725631768953, "grad_norm": 1.970961332321167, "learning_rate": 5.011073663938373e-05, "loss": 1.4907, "step": 10362 }, { "epoch": 4.990613718411552, "grad_norm": 1.6951099634170532, "learning_rate": 5.008184882041406e-05, "loss": 1.4999, "step": 10368 }, { "epoch": 4.9935018050541515, "grad_norm": 1.6805800199508667, "learning_rate": 5.00529610014444e-05, "loss": 1.3735, "step": 10374 }, { "epoch": 4.996389891696751, "grad_norm": 1.982224464416504, "learning_rate": 5.002407318247473e-05, "loss": 1.3677, "step": 10380 }, { "epoch": 4.99927797833935, "grad_norm": 2.183117389678955, "learning_rate": 4.999518536350506e-05, "loss": 1.3067, "step": 10386 }, { "epoch": 5.0021660649819495, "grad_norm": 1.384118676185608, "learning_rate": 4.996629754453539e-05, "loss": 1.3873, "step": 10392 }, { "epoch": 5.005054151624549, "grad_norm": 1.6283146142959595, "learning_rate": 4.993740972556572e-05, "loss": 1.2813, "step": 10398 }, { "epoch": 5.007942238267148, "grad_norm": 1.6414759159088135, "learning_rate": 4.9908521906596056e-05, "loss": 1.3912, "step": 10404 }, { "epoch": 5.0108303249097474, "grad_norm": 1.9007381200790405, "learning_rate": 4.987963408762639e-05, "loss": 1.4488, "step": 10410 }, { "epoch": 5.013718411552347, "grad_norm": 1.7227401733398438, "learning_rate": 4.985074626865672e-05, "loss": 1.3655, "step": 10416 }, { "epoch": 5.016606498194946, "grad_norm": 1.7134157419204712, "learning_rate": 4.982185844968705e-05, "loss": 1.4936, "step": 10422 }, { "epoch": 5.019494584837545, "grad_norm": 1.9248756170272827, "learning_rate": 4.979297063071738e-05, "loss": 1.4322, "step": 10428 }, { "epoch": 5.022382671480145, "grad_norm": 1.8274937868118286, "learning_rate": 4.9764082811747714e-05, "loss": 1.2389, "step": 10434 }, { "epoch": 5.025270758122744, "grad_norm": 2.0861730575561523, "learning_rate": 4.973519499277805e-05, "loss": 1.4014, "step": 10440 }, { "epoch": 5.028158844765343, "grad_norm": 1.7571805715560913, "learning_rate": 4.9706307173808384e-05, "loss": 1.4832, "step": 10446 }, { "epoch": 5.031046931407943, "grad_norm": 2.3004283905029297, "learning_rate": 4.967741935483871e-05, "loss": 1.4241, "step": 10452 }, { "epoch": 5.033935018050541, "grad_norm": 1.5680732727050781, "learning_rate": 4.964853153586904e-05, "loss": 1.332, "step": 10458 }, { "epoch": 5.03682310469314, "grad_norm": 1.9814610481262207, "learning_rate": 4.961964371689938e-05, "loss": 1.4567, "step": 10464 }, { "epoch": 5.03971119133574, "grad_norm": 1.6604129076004028, "learning_rate": 4.959075589792971e-05, "loss": 1.4299, "step": 10470 }, { "epoch": 5.042599277978339, "grad_norm": 1.5325173139572144, "learning_rate": 4.956186807896004e-05, "loss": 1.2592, "step": 10476 }, { "epoch": 5.045487364620938, "grad_norm": 1.6775814294815063, "learning_rate": 4.953298025999037e-05, "loss": 1.309, "step": 10482 }, { "epoch": 5.048375451263538, "grad_norm": 2.0211949348449707, "learning_rate": 4.9504092441020705e-05, "loss": 1.3468, "step": 10488 }, { "epoch": 5.051263537906137, "grad_norm": 1.9157582521438599, "learning_rate": 4.9475204622051037e-05, "loss": 1.4286, "step": 10494 }, { "epoch": 5.054151624548736, "grad_norm": 2.0210371017456055, "learning_rate": 4.9446316803081375e-05, "loss": 1.305, "step": 10500 }, { "epoch": 5.057039711191336, "grad_norm": 1.9095145463943481, "learning_rate": 4.94174289841117e-05, "loss": 1.4491, "step": 10506 }, { "epoch": 5.059927797833935, "grad_norm": 1.4906312227249146, "learning_rate": 4.938854116514203e-05, "loss": 1.4041, "step": 10512 }, { "epoch": 5.062815884476534, "grad_norm": 1.779810905456543, "learning_rate": 4.935965334617237e-05, "loss": 1.4542, "step": 10518 }, { "epoch": 5.065703971119134, "grad_norm": 1.5957283973693848, "learning_rate": 4.93307655272027e-05, "loss": 1.3471, "step": 10524 }, { "epoch": 5.068592057761733, "grad_norm": 1.9749889373779297, "learning_rate": 4.930187770823303e-05, "loss": 1.3474, "step": 10530 }, { "epoch": 5.071480144404332, "grad_norm": 1.9881882667541504, "learning_rate": 4.9272989889263364e-05, "loss": 1.4045, "step": 10536 }, { "epoch": 5.074368231046932, "grad_norm": 1.9918802976608276, "learning_rate": 4.9244102070293696e-05, "loss": 1.4391, "step": 10542 }, { "epoch": 5.077256317689531, "grad_norm": 1.6849573850631714, "learning_rate": 4.921521425132403e-05, "loss": 1.3363, "step": 10548 }, { "epoch": 5.08014440433213, "grad_norm": 1.6692942380905151, "learning_rate": 4.918632643235436e-05, "loss": 1.4115, "step": 10554 }, { "epoch": 5.0830324909747295, "grad_norm": 1.8065398931503296, "learning_rate": 4.915743861338469e-05, "loss": 1.3753, "step": 10560 }, { "epoch": 5.085920577617329, "grad_norm": 2.1003754138946533, "learning_rate": 4.912855079441502e-05, "loss": 1.4076, "step": 10566 }, { "epoch": 5.088808664259928, "grad_norm": 2.0016491413116455, "learning_rate": 4.9099662975445354e-05, "loss": 1.3546, "step": 10572 }, { "epoch": 5.0916967509025275, "grad_norm": 1.8994795083999634, "learning_rate": 4.907077515647569e-05, "loss": 1.459, "step": 10578 }, { "epoch": 5.094584837545126, "grad_norm": 1.8071502447128296, "learning_rate": 4.904188733750602e-05, "loss": 1.4814, "step": 10584 }, { "epoch": 5.097472924187725, "grad_norm": 1.936405062675476, "learning_rate": 4.901299951853635e-05, "loss": 1.4163, "step": 10590 }, { "epoch": 5.100361010830325, "grad_norm": 1.9142526388168335, "learning_rate": 4.898411169956669e-05, "loss": 1.4571, "step": 10596 }, { "epoch": 5.103249097472924, "grad_norm": 1.7836267948150635, "learning_rate": 4.895522388059702e-05, "loss": 1.4234, "step": 10602 }, { "epoch": 5.106137184115523, "grad_norm": 1.8793550729751587, "learning_rate": 4.892633606162735e-05, "loss": 1.4718, "step": 10608 }, { "epoch": 5.1090252707581225, "grad_norm": 1.8568651676177979, "learning_rate": 4.889744824265768e-05, "loss": 1.3384, "step": 10614 }, { "epoch": 5.111913357400722, "grad_norm": 1.4710673093795776, "learning_rate": 4.8868560423688014e-05, "loss": 1.3012, "step": 10620 }, { "epoch": 5.114801444043321, "grad_norm": 1.819098949432373, "learning_rate": 4.8839672604718345e-05, "loss": 1.3454, "step": 10626 }, { "epoch": 5.1176895306859205, "grad_norm": 1.6424278020858765, "learning_rate": 4.8810784785748684e-05, "loss": 1.4711, "step": 10632 }, { "epoch": 5.12057761732852, "grad_norm": 1.8864530324935913, "learning_rate": 4.878189696677901e-05, "loss": 1.3401, "step": 10638 }, { "epoch": 5.123465703971119, "grad_norm": 1.7134032249450684, "learning_rate": 4.875300914780934e-05, "loss": 1.2558, "step": 10644 }, { "epoch": 5.126353790613718, "grad_norm": 1.6353322267532349, "learning_rate": 4.872412132883967e-05, "loss": 1.3548, "step": 10650 }, { "epoch": 5.129241877256318, "grad_norm": 2.1162588596343994, "learning_rate": 4.869523350987001e-05, "loss": 1.4099, "step": 10656 }, { "epoch": 5.132129963898917, "grad_norm": 1.846817135810852, "learning_rate": 4.866634569090034e-05, "loss": 1.4067, "step": 10662 }, { "epoch": 5.135018050541516, "grad_norm": 1.5485557317733765, "learning_rate": 4.8637457871930666e-05, "loss": 1.3518, "step": 10668 }, { "epoch": 5.137906137184116, "grad_norm": 1.8367717266082764, "learning_rate": 4.8608570052961005e-05, "loss": 1.4269, "step": 10674 }, { "epoch": 5.140794223826715, "grad_norm": 1.6887692213058472, "learning_rate": 4.8579682233991336e-05, "loss": 1.2343, "step": 10680 }, { "epoch": 5.143682310469314, "grad_norm": 2.087960720062256, "learning_rate": 4.855079441502167e-05, "loss": 1.4093, "step": 10686 }, { "epoch": 5.146570397111914, "grad_norm": 1.9119988679885864, "learning_rate": 4.8521906596052e-05, "loss": 1.3467, "step": 10692 }, { "epoch": 5.149458483754513, "grad_norm": 1.5580964088439941, "learning_rate": 4.849301877708233e-05, "loss": 1.3639, "step": 10698 }, { "epoch": 5.152346570397112, "grad_norm": 1.5769884586334229, "learning_rate": 4.846413095811266e-05, "loss": 1.4331, "step": 10704 }, { "epoch": 5.155234657039712, "grad_norm": 1.8357503414154053, "learning_rate": 4.8435243139143e-05, "loss": 1.3628, "step": 10710 }, { "epoch": 5.15812274368231, "grad_norm": 2.0516934394836426, "learning_rate": 4.840635532017333e-05, "loss": 1.5278, "step": 10716 }, { "epoch": 5.161010830324909, "grad_norm": 1.849205732345581, "learning_rate": 4.837746750120366e-05, "loss": 1.3952, "step": 10722 }, { "epoch": 5.163898916967509, "grad_norm": 1.7021790742874146, "learning_rate": 4.834857968223399e-05, "loss": 1.4927, "step": 10728 }, { "epoch": 5.166787003610108, "grad_norm": 1.7733180522918701, "learning_rate": 4.831969186326433e-05, "loss": 1.5298, "step": 10734 }, { "epoch": 5.169675090252707, "grad_norm": 1.9406448602676392, "learning_rate": 4.829080404429466e-05, "loss": 1.5383, "step": 10740 }, { "epoch": 5.172563176895307, "grad_norm": 1.9060615301132202, "learning_rate": 4.826191622532499e-05, "loss": 1.2133, "step": 10746 }, { "epoch": 5.175451263537906, "grad_norm": 2.1986987590789795, "learning_rate": 4.823302840635532e-05, "loss": 1.3893, "step": 10752 }, { "epoch": 5.178339350180505, "grad_norm": 2.101058006286621, "learning_rate": 4.8204140587385654e-05, "loss": 1.3781, "step": 10758 }, { "epoch": 5.181227436823105, "grad_norm": 1.965269684791565, "learning_rate": 4.8175252768415985e-05, "loss": 1.1875, "step": 10764 }, { "epoch": 5.184115523465704, "grad_norm": 1.8783873319625854, "learning_rate": 4.8146364949446324e-05, "loss": 1.5101, "step": 10770 }, { "epoch": 5.187003610108303, "grad_norm": 1.776719331741333, "learning_rate": 4.811747713047665e-05, "loss": 1.5005, "step": 10776 }, { "epoch": 5.189891696750903, "grad_norm": 1.7807551622390747, "learning_rate": 4.808858931150698e-05, "loss": 1.3807, "step": 10782 }, { "epoch": 5.192779783393502, "grad_norm": 2.1187584400177, "learning_rate": 4.805970149253732e-05, "loss": 1.4385, "step": 10788 }, { "epoch": 5.195667870036101, "grad_norm": 1.9886709451675415, "learning_rate": 4.803081367356765e-05, "loss": 1.4442, "step": 10794 }, { "epoch": 5.1985559566787005, "grad_norm": 2.038381338119507, "learning_rate": 4.800192585459798e-05, "loss": 1.4191, "step": 10800 }, { "epoch": 5.2014440433213, "grad_norm": 1.9075435400009155, "learning_rate": 4.7973038035628313e-05, "loss": 1.377, "step": 10806 }, { "epoch": 5.204332129963899, "grad_norm": 1.8431333303451538, "learning_rate": 4.7944150216658645e-05, "loss": 1.3622, "step": 10812 }, { "epoch": 5.2072202166064985, "grad_norm": 1.5886873006820679, "learning_rate": 4.7915262397688977e-05, "loss": 1.4011, "step": 10818 }, { "epoch": 5.210108303249098, "grad_norm": 2.1114776134490967, "learning_rate": 4.788637457871931e-05, "loss": 1.6096, "step": 10824 }, { "epoch": 5.212996389891697, "grad_norm": 1.8112022876739502, "learning_rate": 4.785748675974964e-05, "loss": 1.3961, "step": 10830 }, { "epoch": 5.215884476534296, "grad_norm": 1.9229594469070435, "learning_rate": 4.782859894077997e-05, "loss": 1.2632, "step": 10836 }, { "epoch": 5.218772563176895, "grad_norm": 1.8223298788070679, "learning_rate": 4.77997111218103e-05, "loss": 1.279, "step": 10842 }, { "epoch": 5.221660649819494, "grad_norm": 1.8965673446655273, "learning_rate": 4.777082330284064e-05, "loss": 1.3592, "step": 10848 }, { "epoch": 5.2245487364620935, "grad_norm": 2.0837786197662354, "learning_rate": 4.774193548387097e-05, "loss": 1.4595, "step": 10854 }, { "epoch": 5.227436823104693, "grad_norm": 1.8152995109558105, "learning_rate": 4.77130476649013e-05, "loss": 1.2914, "step": 10860 }, { "epoch": 5.230324909747292, "grad_norm": 1.8160494565963745, "learning_rate": 4.7684159845931636e-05, "loss": 1.3206, "step": 10866 }, { "epoch": 5.2332129963898915, "grad_norm": 1.8240723609924316, "learning_rate": 4.765527202696197e-05, "loss": 1.4125, "step": 10872 }, { "epoch": 5.236101083032491, "grad_norm": 1.9265865087509155, "learning_rate": 4.76263842079923e-05, "loss": 1.368, "step": 10878 }, { "epoch": 5.23898916967509, "grad_norm": 1.8223111629486084, "learning_rate": 4.759749638902263e-05, "loss": 1.3415, "step": 10884 }, { "epoch": 5.241877256317689, "grad_norm": 1.8942539691925049, "learning_rate": 4.756860857005296e-05, "loss": 1.4183, "step": 10890 }, { "epoch": 5.244765342960289, "grad_norm": 1.979552149772644, "learning_rate": 4.7539720751083294e-05, "loss": 1.3745, "step": 10896 }, { "epoch": 5.247653429602888, "grad_norm": 1.865634560585022, "learning_rate": 4.751083293211363e-05, "loss": 1.4074, "step": 10902 }, { "epoch": 5.250541516245487, "grad_norm": 2.0252764225006104, "learning_rate": 4.7481945113143964e-05, "loss": 1.4073, "step": 10908 }, { "epoch": 5.253429602888087, "grad_norm": 1.8616702556610107, "learning_rate": 4.745305729417429e-05, "loss": 1.4077, "step": 10914 }, { "epoch": 5.256317689530686, "grad_norm": 1.682409644126892, "learning_rate": 4.742416947520462e-05, "loss": 1.4566, "step": 10920 }, { "epoch": 5.259205776173285, "grad_norm": 1.925624132156372, "learning_rate": 4.739528165623496e-05, "loss": 1.375, "step": 10926 }, { "epoch": 5.262093862815885, "grad_norm": 1.9925528764724731, "learning_rate": 4.736639383726529e-05, "loss": 1.3658, "step": 10932 }, { "epoch": 5.264981949458484, "grad_norm": 1.7103776931762695, "learning_rate": 4.733750601829562e-05, "loss": 1.4078, "step": 10938 }, { "epoch": 5.267870036101083, "grad_norm": 1.7259843349456787, "learning_rate": 4.7308618199325954e-05, "loss": 1.4094, "step": 10944 }, { "epoch": 5.270758122743683, "grad_norm": 1.6239246129989624, "learning_rate": 4.7279730380356285e-05, "loss": 1.3552, "step": 10950 }, { "epoch": 5.273646209386282, "grad_norm": 1.7632781267166138, "learning_rate": 4.725084256138662e-05, "loss": 1.3738, "step": 10956 }, { "epoch": 5.276534296028881, "grad_norm": 1.8627060651779175, "learning_rate": 4.722195474241695e-05, "loss": 1.3842, "step": 10962 }, { "epoch": 5.279422382671481, "grad_norm": 1.7716244459152222, "learning_rate": 4.719306692344728e-05, "loss": 1.1686, "step": 10968 }, { "epoch": 5.28231046931408, "grad_norm": 1.9946244955062866, "learning_rate": 4.716417910447761e-05, "loss": 1.3803, "step": 10974 }, { "epoch": 5.285198555956678, "grad_norm": 1.9587725400924683, "learning_rate": 4.713529128550795e-05, "loss": 1.5221, "step": 10980 }, { "epoch": 5.288086642599278, "grad_norm": 1.7351337671279907, "learning_rate": 4.710640346653828e-05, "loss": 1.385, "step": 10986 }, { "epoch": 5.290974729241877, "grad_norm": 1.7889926433563232, "learning_rate": 4.7077515647568606e-05, "loss": 1.3701, "step": 10992 }, { "epoch": 5.293862815884476, "grad_norm": 1.9065725803375244, "learning_rate": 4.7048627828598945e-05, "loss": 1.4602, "step": 10998 }, { "epoch": 5.296750902527076, "grad_norm": 1.7426867485046387, "learning_rate": 4.7019740009629276e-05, "loss": 1.423, "step": 11004 }, { "epoch": 5.299638989169675, "grad_norm": 1.6339125633239746, "learning_rate": 4.699085219065961e-05, "loss": 1.372, "step": 11010 }, { "epoch": 5.302527075812274, "grad_norm": 1.5886833667755127, "learning_rate": 4.696196437168994e-05, "loss": 1.2921, "step": 11016 }, { "epoch": 5.3054151624548735, "grad_norm": 1.9326857328414917, "learning_rate": 4.693307655272027e-05, "loss": 1.3793, "step": 11022 }, { "epoch": 5.308303249097473, "grad_norm": 1.7117496728897095, "learning_rate": 4.69041887337506e-05, "loss": 1.3847, "step": 11028 }, { "epoch": 5.311191335740072, "grad_norm": 1.6420360803604126, "learning_rate": 4.6875300914780934e-05, "loss": 1.3076, "step": 11034 }, { "epoch": 5.3140794223826715, "grad_norm": 1.6558014154434204, "learning_rate": 4.684641309581127e-05, "loss": 1.224, "step": 11040 }, { "epoch": 5.316967509025271, "grad_norm": 1.8222453594207764, "learning_rate": 4.68175252768416e-05, "loss": 1.2835, "step": 11046 }, { "epoch": 5.31985559566787, "grad_norm": 1.741629719734192, "learning_rate": 4.678863745787193e-05, "loss": 1.402, "step": 11052 }, { "epoch": 5.3227436823104695, "grad_norm": 1.9874197244644165, "learning_rate": 4.675974963890227e-05, "loss": 1.4474, "step": 11058 }, { "epoch": 5.325631768953069, "grad_norm": 1.8994001150131226, "learning_rate": 4.67308618199326e-05, "loss": 1.3897, "step": 11064 }, { "epoch": 5.328519855595668, "grad_norm": 1.8804703950881958, "learning_rate": 4.670197400096293e-05, "loss": 1.4019, "step": 11070 }, { "epoch": 5.331407942238267, "grad_norm": 1.9062293767929077, "learning_rate": 4.667308618199326e-05, "loss": 1.4114, "step": 11076 }, { "epoch": 5.334296028880867, "grad_norm": 1.8367713689804077, "learning_rate": 4.6644198363023594e-05, "loss": 1.2645, "step": 11082 }, { "epoch": 5.337184115523466, "grad_norm": 1.8398098945617676, "learning_rate": 4.6615310544053926e-05, "loss": 1.4482, "step": 11088 }, { "epoch": 5.340072202166065, "grad_norm": 1.9933046102523804, "learning_rate": 4.6586422725084264e-05, "loss": 1.5075, "step": 11094 }, { "epoch": 5.342960288808664, "grad_norm": 1.787484049797058, "learning_rate": 4.655753490611459e-05, "loss": 1.292, "step": 11100 }, { "epoch": 5.345848375451263, "grad_norm": 1.7917931079864502, "learning_rate": 4.652864708714492e-05, "loss": 1.2276, "step": 11106 }, { "epoch": 5.348736462093862, "grad_norm": 1.9087529182434082, "learning_rate": 4.649975926817525e-05, "loss": 1.2995, "step": 11112 }, { "epoch": 5.351624548736462, "grad_norm": 1.7566546201705933, "learning_rate": 4.647087144920559e-05, "loss": 1.3288, "step": 11118 }, { "epoch": 5.354512635379061, "grad_norm": 2.1683685779571533, "learning_rate": 4.644198363023592e-05, "loss": 1.3751, "step": 11124 }, { "epoch": 5.35740072202166, "grad_norm": 1.5682117938995361, "learning_rate": 4.641309581126625e-05, "loss": 1.4448, "step": 11130 }, { "epoch": 5.36028880866426, "grad_norm": 1.962820053100586, "learning_rate": 4.6384207992296585e-05, "loss": 1.5982, "step": 11136 }, { "epoch": 5.363176895306859, "grad_norm": 1.957924723625183, "learning_rate": 4.635532017332692e-05, "loss": 1.3756, "step": 11142 }, { "epoch": 5.366064981949458, "grad_norm": 1.878746747970581, "learning_rate": 4.632643235435725e-05, "loss": 1.4451, "step": 11148 }, { "epoch": 5.368953068592058, "grad_norm": 1.7947072982788086, "learning_rate": 4.629754453538758e-05, "loss": 1.3482, "step": 11154 }, { "epoch": 5.371841155234657, "grad_norm": 1.9547878503799438, "learning_rate": 4.626865671641791e-05, "loss": 1.3312, "step": 11160 }, { "epoch": 5.374729241877256, "grad_norm": 2.0302963256835938, "learning_rate": 4.623976889744824e-05, "loss": 1.5307, "step": 11166 }, { "epoch": 5.377617328519856, "grad_norm": 1.9372237920761108, "learning_rate": 4.621088107847858e-05, "loss": 1.3882, "step": 11172 }, { "epoch": 5.380505415162455, "grad_norm": 2.100524425506592, "learning_rate": 4.618199325950891e-05, "loss": 1.4805, "step": 11178 }, { "epoch": 5.383393501805054, "grad_norm": 1.7742443084716797, "learning_rate": 4.615310544053924e-05, "loss": 1.5495, "step": 11184 }, { "epoch": 5.386281588447654, "grad_norm": 2.0399422645568848, "learning_rate": 4.612421762156957e-05, "loss": 1.4574, "step": 11190 }, { "epoch": 5.389169675090253, "grad_norm": 2.0389344692230225, "learning_rate": 4.609532980259991e-05, "loss": 1.4501, "step": 11196 }, { "epoch": 5.392057761732852, "grad_norm": 1.9308079481124878, "learning_rate": 4.606644198363024e-05, "loss": 1.4544, "step": 11202 }, { "epoch": 5.3949458483754515, "grad_norm": 1.829406976699829, "learning_rate": 4.603755416466057e-05, "loss": 1.307, "step": 11208 }, { "epoch": 5.397833935018051, "grad_norm": 1.9424519538879395, "learning_rate": 4.60086663456909e-05, "loss": 1.4305, "step": 11214 }, { "epoch": 5.40072202166065, "grad_norm": 1.7839441299438477, "learning_rate": 4.5979778526721234e-05, "loss": 1.3377, "step": 11220 }, { "epoch": 5.4036101083032495, "grad_norm": 1.7180062532424927, "learning_rate": 4.5950890707751566e-05, "loss": 1.467, "step": 11226 }, { "epoch": 5.406498194945849, "grad_norm": 2.0764577388763428, "learning_rate": 4.5922002888781904e-05, "loss": 1.4136, "step": 11232 }, { "epoch": 5.409386281588447, "grad_norm": 1.75456702709198, "learning_rate": 4.589311506981223e-05, "loss": 1.4363, "step": 11238 }, { "epoch": 5.412274368231047, "grad_norm": 1.8039367198944092, "learning_rate": 4.586422725084256e-05, "loss": 1.4926, "step": 11244 }, { "epoch": 5.415162454873646, "grad_norm": 2.1852633953094482, "learning_rate": 4.58353394318729e-05, "loss": 1.412, "step": 11250 }, { "epoch": 5.418050541516245, "grad_norm": 1.8858367204666138, "learning_rate": 4.580645161290323e-05, "loss": 1.3036, "step": 11256 }, { "epoch": 5.4209386281588445, "grad_norm": 1.864302158355713, "learning_rate": 4.577756379393356e-05, "loss": 1.3319, "step": 11262 }, { "epoch": 5.423826714801444, "grad_norm": 1.948822021484375, "learning_rate": 4.5748675974963894e-05, "loss": 1.3825, "step": 11268 }, { "epoch": 5.426714801444043, "grad_norm": 1.9043256044387817, "learning_rate": 4.5719788155994225e-05, "loss": 1.399, "step": 11274 }, { "epoch": 5.4296028880866425, "grad_norm": 2.0592966079711914, "learning_rate": 4.569090033702456e-05, "loss": 1.4343, "step": 11280 }, { "epoch": 5.432490974729242, "grad_norm": 1.8048512935638428, "learning_rate": 4.566201251805489e-05, "loss": 1.4055, "step": 11286 }, { "epoch": 5.435379061371841, "grad_norm": 2.163090944290161, "learning_rate": 4.563312469908522e-05, "loss": 1.3877, "step": 11292 }, { "epoch": 5.4382671480144404, "grad_norm": 1.9854857921600342, "learning_rate": 4.560423688011555e-05, "loss": 1.3265, "step": 11298 }, { "epoch": 5.44115523465704, "grad_norm": 1.6429569721221924, "learning_rate": 4.557534906114588e-05, "loss": 1.3253, "step": 11304 }, { "epoch": 5.444043321299639, "grad_norm": 1.7292470932006836, "learning_rate": 4.554646124217622e-05, "loss": 1.3812, "step": 11310 }, { "epoch": 5.446931407942238, "grad_norm": 1.913120150566101, "learning_rate": 4.5517573423206546e-05, "loss": 1.4822, "step": 11316 }, { "epoch": 5.449819494584838, "grad_norm": 1.9358288049697876, "learning_rate": 4.548868560423688e-05, "loss": 1.3817, "step": 11322 }, { "epoch": 5.452707581227437, "grad_norm": 1.8093414306640625, "learning_rate": 4.5459797785267216e-05, "loss": 1.331, "step": 11328 }, { "epoch": 5.455595667870036, "grad_norm": 1.7647932767868042, "learning_rate": 4.543090996629755e-05, "loss": 1.4005, "step": 11334 }, { "epoch": 5.458483754512636, "grad_norm": 1.7140213251113892, "learning_rate": 4.540202214732788e-05, "loss": 1.5008, "step": 11340 }, { "epoch": 5.461371841155235, "grad_norm": 2.1872005462646484, "learning_rate": 4.537313432835821e-05, "loss": 1.3719, "step": 11346 }, { "epoch": 5.464259927797834, "grad_norm": 1.745508074760437, "learning_rate": 4.534424650938854e-05, "loss": 1.3898, "step": 11352 }, { "epoch": 5.467148014440433, "grad_norm": 1.695009708404541, "learning_rate": 4.5315358690418874e-05, "loss": 1.4349, "step": 11358 }, { "epoch": 5.470036101083032, "grad_norm": 1.6629071235656738, "learning_rate": 4.528647087144921e-05, "loss": 1.4291, "step": 11364 }, { "epoch": 5.472924187725631, "grad_norm": 1.9073585271835327, "learning_rate": 4.525758305247954e-05, "loss": 1.4247, "step": 11370 }, { "epoch": 5.475812274368231, "grad_norm": 1.589372158050537, "learning_rate": 4.522869523350987e-05, "loss": 1.3748, "step": 11376 }, { "epoch": 5.47870036101083, "grad_norm": 1.7802690267562866, "learning_rate": 4.51998074145402e-05, "loss": 1.4203, "step": 11382 }, { "epoch": 5.481588447653429, "grad_norm": 1.7933183908462524, "learning_rate": 4.517091959557054e-05, "loss": 1.4391, "step": 11388 }, { "epoch": 5.484476534296029, "grad_norm": 1.7182660102844238, "learning_rate": 4.514203177660087e-05, "loss": 1.2877, "step": 11394 }, { "epoch": 5.487364620938628, "grad_norm": 1.9912599325180054, "learning_rate": 4.5113143957631196e-05, "loss": 1.5119, "step": 11400 }, { "epoch": 5.490252707581227, "grad_norm": 1.6722325086593628, "learning_rate": 4.5084256138661534e-05, "loss": 1.4839, "step": 11406 }, { "epoch": 5.493140794223827, "grad_norm": 2.12286376953125, "learning_rate": 4.5055368319691866e-05, "loss": 1.2948, "step": 11412 }, { "epoch": 5.496028880866426, "grad_norm": 1.656951904296875, "learning_rate": 4.50264805007222e-05, "loss": 1.2704, "step": 11418 }, { "epoch": 5.498916967509025, "grad_norm": 1.8386222124099731, "learning_rate": 4.499759268175253e-05, "loss": 1.5161, "step": 11424 }, { "epoch": 5.501805054151625, "grad_norm": 1.9824771881103516, "learning_rate": 4.496870486278286e-05, "loss": 1.3773, "step": 11430 }, { "epoch": 5.504693140794224, "grad_norm": 2.0316433906555176, "learning_rate": 4.493981704381319e-05, "loss": 1.4863, "step": 11436 }, { "epoch": 5.507581227436823, "grad_norm": 2.0281896591186523, "learning_rate": 4.491092922484353e-05, "loss": 1.4575, "step": 11442 }, { "epoch": 5.5104693140794225, "grad_norm": 1.7858606576919556, "learning_rate": 4.488204140587386e-05, "loss": 1.3199, "step": 11448 }, { "epoch": 5.513357400722022, "grad_norm": 1.764918327331543, "learning_rate": 4.485315358690419e-05, "loss": 1.4856, "step": 11454 }, { "epoch": 5.516245487364621, "grad_norm": 2.0273139476776123, "learning_rate": 4.4824265767934525e-05, "loss": 1.4077, "step": 11460 }, { "epoch": 5.5191335740072205, "grad_norm": 2.00801157951355, "learning_rate": 4.479537794896486e-05, "loss": 1.3524, "step": 11466 }, { "epoch": 5.52202166064982, "grad_norm": 2.129699230194092, "learning_rate": 4.476649012999519e-05, "loss": 1.3521, "step": 11472 }, { "epoch": 5.524909747292419, "grad_norm": 2.2441952228546143, "learning_rate": 4.473760231102552e-05, "loss": 1.3439, "step": 11478 }, { "epoch": 5.5277978339350184, "grad_norm": 2.0186970233917236, "learning_rate": 4.470871449205585e-05, "loss": 1.3312, "step": 11484 }, { "epoch": 5.530685920577618, "grad_norm": 1.6965993642807007, "learning_rate": 4.467982667308618e-05, "loss": 1.3982, "step": 11490 }, { "epoch": 5.533574007220217, "grad_norm": 1.9102481603622437, "learning_rate": 4.4650938854116515e-05, "loss": 1.404, "step": 11496 }, { "epoch": 5.5364620938628155, "grad_norm": 1.7596397399902344, "learning_rate": 4.462205103514685e-05, "loss": 1.4844, "step": 11502 }, { "epoch": 5.539350180505415, "grad_norm": 2.059607744216919, "learning_rate": 4.459316321617718e-05, "loss": 1.384, "step": 11508 }, { "epoch": 5.542238267148014, "grad_norm": 1.9029583930969238, "learning_rate": 4.456427539720751e-05, "loss": 1.3698, "step": 11514 }, { "epoch": 5.5451263537906135, "grad_norm": 1.7408952713012695, "learning_rate": 4.453538757823785e-05, "loss": 1.2599, "step": 11520 }, { "epoch": 5.548014440433213, "grad_norm": 1.862351417541504, "learning_rate": 4.450649975926818e-05, "loss": 1.3424, "step": 11526 }, { "epoch": 5.550902527075812, "grad_norm": 1.9627162218093872, "learning_rate": 4.447761194029851e-05, "loss": 1.5071, "step": 11532 }, { "epoch": 5.553790613718411, "grad_norm": 1.933674931526184, "learning_rate": 4.444872412132884e-05, "loss": 1.2882, "step": 11538 }, { "epoch": 5.556678700361011, "grad_norm": 2.037869691848755, "learning_rate": 4.4419836302359174e-05, "loss": 1.489, "step": 11544 }, { "epoch": 5.55956678700361, "grad_norm": 2.019050121307373, "learning_rate": 4.4390948483389506e-05, "loss": 1.4629, "step": 11550 }, { "epoch": 5.562454873646209, "grad_norm": 1.7724860906600952, "learning_rate": 4.4362060664419844e-05, "loss": 1.3927, "step": 11556 }, { "epoch": 5.565342960288809, "grad_norm": 1.7266159057617188, "learning_rate": 4.433317284545017e-05, "loss": 1.2914, "step": 11562 }, { "epoch": 5.568231046931408, "grad_norm": 1.5895179510116577, "learning_rate": 4.43042850264805e-05, "loss": 1.3507, "step": 11568 }, { "epoch": 5.571119133574007, "grad_norm": 1.7363311052322388, "learning_rate": 4.427539720751083e-05, "loss": 1.3289, "step": 11574 }, { "epoch": 5.574007220216607, "grad_norm": 1.9808260202407837, "learning_rate": 4.424650938854117e-05, "loss": 1.3685, "step": 11580 }, { "epoch": 5.576895306859206, "grad_norm": 1.8071736097335815, "learning_rate": 4.42176215695715e-05, "loss": 1.4803, "step": 11586 }, { "epoch": 5.579783393501805, "grad_norm": 1.9605588912963867, "learning_rate": 4.418873375060183e-05, "loss": 1.4801, "step": 11592 }, { "epoch": 5.582671480144405, "grad_norm": 1.611019492149353, "learning_rate": 4.4159845931632165e-05, "loss": 1.4711, "step": 11598 }, { "epoch": 5.585559566787004, "grad_norm": 1.9906893968582153, "learning_rate": 4.41309581126625e-05, "loss": 1.4382, "step": 11604 }, { "epoch": 5.588447653429603, "grad_norm": 1.9580456018447876, "learning_rate": 4.410207029369283e-05, "loss": 1.3805, "step": 11610 }, { "epoch": 5.591335740072202, "grad_norm": 1.885762095451355, "learning_rate": 4.407318247472316e-05, "loss": 1.4118, "step": 11616 }, { "epoch": 5.594223826714801, "grad_norm": 1.8096966743469238, "learning_rate": 4.404429465575349e-05, "loss": 1.3606, "step": 11622 }, { "epoch": 5.5971119133574, "grad_norm": 1.8512705564498901, "learning_rate": 4.401540683678382e-05, "loss": 1.4005, "step": 11628 }, { "epoch": 5.6, "grad_norm": 1.7062269449234009, "learning_rate": 4.398651901781416e-05, "loss": 1.4909, "step": 11634 }, { "epoch": 5.602888086642599, "grad_norm": 1.894586205482483, "learning_rate": 4.395763119884449e-05, "loss": 1.4788, "step": 11640 }, { "epoch": 5.605776173285198, "grad_norm": 1.649200439453125, "learning_rate": 4.392874337987482e-05, "loss": 1.3886, "step": 11646 }, { "epoch": 5.608664259927798, "grad_norm": 1.546499490737915, "learning_rate": 4.3899855560905157e-05, "loss": 1.3351, "step": 11652 }, { "epoch": 5.611552346570397, "grad_norm": 2.165597677230835, "learning_rate": 4.387096774193549e-05, "loss": 1.4347, "step": 11658 }, { "epoch": 5.614440433212996, "grad_norm": 1.9112956523895264, "learning_rate": 4.384207992296582e-05, "loss": 1.4344, "step": 11664 }, { "epoch": 5.617328519855596, "grad_norm": 2.0621337890625, "learning_rate": 4.381319210399615e-05, "loss": 1.327, "step": 11670 }, { "epoch": 5.620216606498195, "grad_norm": 2.179347038269043, "learning_rate": 4.378430428502648e-05, "loss": 1.4222, "step": 11676 }, { "epoch": 5.623104693140794, "grad_norm": 1.7475500106811523, "learning_rate": 4.3755416466056814e-05, "loss": 1.4572, "step": 11682 }, { "epoch": 5.6259927797833935, "grad_norm": 1.9678518772125244, "learning_rate": 4.3726528647087146e-05, "loss": 1.43, "step": 11688 }, { "epoch": 5.628880866425993, "grad_norm": 2.0170772075653076, "learning_rate": 4.369764082811748e-05, "loss": 1.4801, "step": 11694 }, { "epoch": 5.631768953068592, "grad_norm": 1.8827463388442993, "learning_rate": 4.366875300914781e-05, "loss": 1.5281, "step": 11700 }, { "epoch": 5.6346570397111915, "grad_norm": 1.859522819519043, "learning_rate": 4.363986519017814e-05, "loss": 1.2258, "step": 11706 }, { "epoch": 5.637545126353791, "grad_norm": 1.8803082704544067, "learning_rate": 4.361097737120848e-05, "loss": 1.3057, "step": 11712 }, { "epoch": 5.64043321299639, "grad_norm": 1.6335639953613281, "learning_rate": 4.358208955223881e-05, "loss": 1.3201, "step": 11718 }, { "epoch": 5.643321299638989, "grad_norm": 1.5964938402175903, "learning_rate": 4.3553201733269136e-05, "loss": 1.3912, "step": 11724 }, { "epoch": 5.646209386281589, "grad_norm": 1.800778865814209, "learning_rate": 4.3524313914299474e-05, "loss": 1.4468, "step": 11730 }, { "epoch": 5.649097472924188, "grad_norm": 2.1224846839904785, "learning_rate": 4.3495426095329806e-05, "loss": 1.3376, "step": 11736 }, { "epoch": 5.651985559566787, "grad_norm": 2.138456344604492, "learning_rate": 4.346653827636014e-05, "loss": 1.3204, "step": 11742 }, { "epoch": 5.654873646209387, "grad_norm": 1.773726463317871, "learning_rate": 4.343765045739047e-05, "loss": 1.2861, "step": 11748 }, { "epoch": 5.657761732851986, "grad_norm": 1.9812120199203491, "learning_rate": 4.34087626384208e-05, "loss": 1.4671, "step": 11754 }, { "epoch": 5.6606498194945845, "grad_norm": 1.9356069564819336, "learning_rate": 4.337987481945113e-05, "loss": 1.4953, "step": 11760 }, { "epoch": 5.663537906137184, "grad_norm": 1.928642988204956, "learning_rate": 4.3350987000481464e-05, "loss": 1.4253, "step": 11766 }, { "epoch": 5.666425992779783, "grad_norm": 1.6639690399169922, "learning_rate": 4.33220991815118e-05, "loss": 1.3495, "step": 11772 }, { "epoch": 5.669314079422382, "grad_norm": 2.0531044006347656, "learning_rate": 4.329321136254213e-05, "loss": 1.4934, "step": 11778 }, { "epoch": 5.672202166064982, "grad_norm": 1.7838382720947266, "learning_rate": 4.326432354357246e-05, "loss": 1.3705, "step": 11784 }, { "epoch": 5.675090252707581, "grad_norm": 1.8908065557479858, "learning_rate": 4.32354357246028e-05, "loss": 1.3966, "step": 11790 }, { "epoch": 5.67797833935018, "grad_norm": 1.8372167348861694, "learning_rate": 4.320654790563313e-05, "loss": 1.4891, "step": 11796 }, { "epoch": 5.68086642599278, "grad_norm": 1.9004453420639038, "learning_rate": 4.317766008666346e-05, "loss": 1.3588, "step": 11802 }, { "epoch": 5.683754512635379, "grad_norm": 1.8517347574234009, "learning_rate": 4.314877226769379e-05, "loss": 1.2687, "step": 11808 }, { "epoch": 5.686642599277978, "grad_norm": 1.7763683795928955, "learning_rate": 4.311988444872412e-05, "loss": 1.4178, "step": 11814 }, { "epoch": 5.689530685920578, "grad_norm": 1.7509212493896484, "learning_rate": 4.3090996629754455e-05, "loss": 1.3796, "step": 11820 }, { "epoch": 5.692418772563177, "grad_norm": 2.1566200256347656, "learning_rate": 4.306210881078479e-05, "loss": 1.4543, "step": 11826 }, { "epoch": 5.695306859205776, "grad_norm": 1.7460936307907104, "learning_rate": 4.303322099181512e-05, "loss": 1.2682, "step": 11832 }, { "epoch": 5.698194945848376, "grad_norm": 1.8554408550262451, "learning_rate": 4.300433317284545e-05, "loss": 1.4274, "step": 11838 }, { "epoch": 5.701083032490975, "grad_norm": 2.0894148349761963, "learning_rate": 4.297544535387578e-05, "loss": 1.4654, "step": 11844 }, { "epoch": 5.703971119133574, "grad_norm": 2.163367986679077, "learning_rate": 4.294655753490612e-05, "loss": 1.4042, "step": 11850 }, { "epoch": 5.706859205776174, "grad_norm": 2.003779411315918, "learning_rate": 4.291766971593645e-05, "loss": 1.4397, "step": 11856 }, { "epoch": 5.709747292418773, "grad_norm": 2.179804563522339, "learning_rate": 4.2888781896966776e-05, "loss": 1.4992, "step": 11862 }, { "epoch": 5.712635379061372, "grad_norm": 2.2084672451019287, "learning_rate": 4.2859894077997114e-05, "loss": 1.426, "step": 11868 }, { "epoch": 5.715523465703971, "grad_norm": 1.7292015552520752, "learning_rate": 4.2831006259027446e-05, "loss": 1.4317, "step": 11874 }, { "epoch": 5.71841155234657, "grad_norm": 2.035912036895752, "learning_rate": 4.280211844005778e-05, "loss": 1.4239, "step": 11880 }, { "epoch": 5.721299638989169, "grad_norm": 1.6785438060760498, "learning_rate": 4.277323062108811e-05, "loss": 1.3988, "step": 11886 }, { "epoch": 5.724187725631769, "grad_norm": 1.8216668367385864, "learning_rate": 4.274434280211844e-05, "loss": 1.304, "step": 11892 }, { "epoch": 5.727075812274368, "grad_norm": 1.792708158493042, "learning_rate": 4.271545498314877e-05, "loss": 1.3306, "step": 11898 }, { "epoch": 5.729963898916967, "grad_norm": 1.9805876016616821, "learning_rate": 4.268656716417911e-05, "loss": 1.2536, "step": 11904 }, { "epoch": 5.7328519855595665, "grad_norm": 1.6256234645843506, "learning_rate": 4.265767934520944e-05, "loss": 1.3431, "step": 11910 }, { "epoch": 5.735740072202166, "grad_norm": 2.1218743324279785, "learning_rate": 4.262879152623977e-05, "loss": 1.4304, "step": 11916 }, { "epoch": 5.738628158844765, "grad_norm": 1.910498857498169, "learning_rate": 4.2599903707270105e-05, "loss": 1.4401, "step": 11922 }, { "epoch": 5.7415162454873645, "grad_norm": 2.004836320877075, "learning_rate": 4.257101588830044e-05, "loss": 1.4144, "step": 11928 }, { "epoch": 5.744404332129964, "grad_norm": 1.9637408256530762, "learning_rate": 4.254212806933077e-05, "loss": 1.3516, "step": 11934 }, { "epoch": 5.747292418772563, "grad_norm": 2.0260169506073, "learning_rate": 4.25132402503611e-05, "loss": 1.3499, "step": 11940 }, { "epoch": 5.7501805054151625, "grad_norm": 1.7120596170425415, "learning_rate": 4.248435243139143e-05, "loss": 1.2908, "step": 11946 }, { "epoch": 5.753068592057762, "grad_norm": 2.0504469871520996, "learning_rate": 4.2455464612421763e-05, "loss": 1.4879, "step": 11952 }, { "epoch": 5.755956678700361, "grad_norm": 1.7862015962600708, "learning_rate": 4.2426576793452095e-05, "loss": 1.5042, "step": 11958 }, { "epoch": 5.75884476534296, "grad_norm": 1.7038929462432861, "learning_rate": 4.239768897448243e-05, "loss": 1.4017, "step": 11964 }, { "epoch": 5.76173285198556, "grad_norm": 1.7859505414962769, "learning_rate": 4.236880115551276e-05, "loss": 1.5061, "step": 11970 }, { "epoch": 5.764620938628159, "grad_norm": 1.9234548807144165, "learning_rate": 4.233991333654309e-05, "loss": 1.4679, "step": 11976 }, { "epoch": 5.767509025270758, "grad_norm": 2.1028268337249756, "learning_rate": 4.231102551757343e-05, "loss": 1.3843, "step": 11982 }, { "epoch": 5.770397111913358, "grad_norm": 1.603543758392334, "learning_rate": 4.228213769860376e-05, "loss": 1.3642, "step": 11988 }, { "epoch": 5.773285198555957, "grad_norm": 1.8545292615890503, "learning_rate": 4.225324987963409e-05, "loss": 1.4179, "step": 11994 }, { "epoch": 5.776173285198556, "grad_norm": 1.8924285173416138, "learning_rate": 4.222436206066442e-05, "loss": 1.5207, "step": 12000 }, { "epoch": 5.779061371841156, "grad_norm": 2.0622525215148926, "learning_rate": 4.2195474241694755e-05, "loss": 1.347, "step": 12006 }, { "epoch": 5.781949458483755, "grad_norm": 2.026139497756958, "learning_rate": 4.2166586422725086e-05, "loss": 1.485, "step": 12012 }, { "epoch": 5.784837545126354, "grad_norm": 1.7109187841415405, "learning_rate": 4.213769860375542e-05, "loss": 1.4288, "step": 12018 }, { "epoch": 5.787725631768953, "grad_norm": 1.9268308877944946, "learning_rate": 4.210881078478575e-05, "loss": 1.5425, "step": 12024 }, { "epoch": 5.790613718411552, "grad_norm": 1.6217255592346191, "learning_rate": 4.207992296581608e-05, "loss": 1.2631, "step": 12030 }, { "epoch": 5.793501805054151, "grad_norm": 1.9938968420028687, "learning_rate": 4.205103514684641e-05, "loss": 1.4996, "step": 12036 }, { "epoch": 5.796389891696751, "grad_norm": 1.8144477605819702, "learning_rate": 4.202214732787675e-05, "loss": 1.3683, "step": 12042 }, { "epoch": 5.79927797833935, "grad_norm": 1.8964409828186035, "learning_rate": 4.1993259508907076e-05, "loss": 1.4834, "step": 12048 }, { "epoch": 5.802166064981949, "grad_norm": 1.7576426267623901, "learning_rate": 4.196437168993741e-05, "loss": 1.4151, "step": 12054 }, { "epoch": 5.805054151624549, "grad_norm": 1.8021008968353271, "learning_rate": 4.1935483870967746e-05, "loss": 1.4186, "step": 12060 }, { "epoch": 5.807942238267148, "grad_norm": 1.885892391204834, "learning_rate": 4.190659605199808e-05, "loss": 1.5866, "step": 12066 }, { "epoch": 5.810830324909747, "grad_norm": 2.0988616943359375, "learning_rate": 4.187770823302841e-05, "loss": 1.2874, "step": 12072 }, { "epoch": 5.813718411552347, "grad_norm": 1.953670620918274, "learning_rate": 4.184882041405874e-05, "loss": 1.3713, "step": 12078 }, { "epoch": 5.816606498194946, "grad_norm": 1.6839842796325684, "learning_rate": 4.181993259508907e-05, "loss": 1.3297, "step": 12084 }, { "epoch": 5.819494584837545, "grad_norm": 1.9384211301803589, "learning_rate": 4.1791044776119404e-05, "loss": 1.5461, "step": 12090 }, { "epoch": 5.8223826714801445, "grad_norm": 1.9669500589370728, "learning_rate": 4.176215695714974e-05, "loss": 1.2696, "step": 12096 }, { "epoch": 5.825270758122744, "grad_norm": 1.9413832426071167, "learning_rate": 4.173326913818007e-05, "loss": 1.4677, "step": 12102 }, { "epoch": 5.828158844765343, "grad_norm": 2.022766590118408, "learning_rate": 4.17043813192104e-05, "loss": 1.307, "step": 12108 }, { "epoch": 5.8310469314079425, "grad_norm": 1.9298158884048462, "learning_rate": 4.167549350024074e-05, "loss": 1.415, "step": 12114 }, { "epoch": 5.833935018050542, "grad_norm": 1.8130465745925903, "learning_rate": 4.164660568127107e-05, "loss": 1.4829, "step": 12120 }, { "epoch": 5.836823104693141, "grad_norm": 1.8297646045684814, "learning_rate": 4.16177178623014e-05, "loss": 1.3587, "step": 12126 }, { "epoch": 5.8397111913357405, "grad_norm": 1.8857874870300293, "learning_rate": 4.1588830043331725e-05, "loss": 1.3789, "step": 12132 }, { "epoch": 5.842599277978339, "grad_norm": 1.9266512393951416, "learning_rate": 4.155994222436206e-05, "loss": 1.4016, "step": 12138 }, { "epoch": 5.845487364620938, "grad_norm": 1.751286268234253, "learning_rate": 4.1531054405392395e-05, "loss": 1.4379, "step": 12144 }, { "epoch": 5.8483754512635375, "grad_norm": 1.9500097036361694, "learning_rate": 4.1502166586422726e-05, "loss": 1.243, "step": 12150 }, { "epoch": 5.851263537906137, "grad_norm": 1.7020072937011719, "learning_rate": 4.147327876745306e-05, "loss": 1.5905, "step": 12156 }, { "epoch": 5.854151624548736, "grad_norm": 1.7563939094543457, "learning_rate": 4.144439094848339e-05, "loss": 1.3604, "step": 12162 }, { "epoch": 5.8570397111913355, "grad_norm": 1.6418819427490234, "learning_rate": 4.141550312951372e-05, "loss": 1.3549, "step": 12168 }, { "epoch": 5.859927797833935, "grad_norm": 2.0404348373413086, "learning_rate": 4.138661531054406e-05, "loss": 1.3621, "step": 12174 }, { "epoch": 5.862815884476534, "grad_norm": 1.8847910165786743, "learning_rate": 4.135772749157439e-05, "loss": 1.4716, "step": 12180 }, { "epoch": 5.865703971119133, "grad_norm": 1.6954399347305298, "learning_rate": 4.1328839672604716e-05, "loss": 1.429, "step": 12186 }, { "epoch": 5.868592057761733, "grad_norm": 2.1019704341888428, "learning_rate": 4.1299951853635054e-05, "loss": 1.4496, "step": 12192 }, { "epoch": 5.871480144404332, "grad_norm": 1.893194556236267, "learning_rate": 4.1271064034665386e-05, "loss": 1.4882, "step": 12198 }, { "epoch": 5.874368231046931, "grad_norm": 1.6725478172302246, "learning_rate": 4.124217621569572e-05, "loss": 1.4139, "step": 12204 }, { "epoch": 5.877256317689531, "grad_norm": 1.995352029800415, "learning_rate": 4.121328839672605e-05, "loss": 1.4551, "step": 12210 }, { "epoch": 5.88014440433213, "grad_norm": 1.6621922254562378, "learning_rate": 4.118440057775638e-05, "loss": 1.303, "step": 12216 }, { "epoch": 5.883032490974729, "grad_norm": 1.9246941804885864, "learning_rate": 4.115551275878671e-05, "loss": 1.359, "step": 12222 }, { "epoch": 5.885920577617329, "grad_norm": 2.0179104804992676, "learning_rate": 4.1126624939817044e-05, "loss": 1.4741, "step": 12228 }, { "epoch": 5.888808664259928, "grad_norm": 1.6908698081970215, "learning_rate": 4.109773712084738e-05, "loss": 1.3707, "step": 12234 }, { "epoch": 5.891696750902527, "grad_norm": 1.841015100479126, "learning_rate": 4.106884930187771e-05, "loss": 1.2912, "step": 12240 }, { "epoch": 5.894584837545127, "grad_norm": 1.7567088603973389, "learning_rate": 4.103996148290804e-05, "loss": 1.426, "step": 12246 }, { "epoch": 5.897472924187726, "grad_norm": 1.5711897611618042, "learning_rate": 4.101107366393838e-05, "loss": 1.2977, "step": 12252 }, { "epoch": 5.900361010830325, "grad_norm": 2.0007364749908447, "learning_rate": 4.098218584496871e-05, "loss": 1.4897, "step": 12258 }, { "epoch": 5.903249097472925, "grad_norm": 1.6352111101150513, "learning_rate": 4.095329802599904e-05, "loss": 1.4219, "step": 12264 }, { "epoch": 5.906137184115524, "grad_norm": 1.771353006362915, "learning_rate": 4.092441020702937e-05, "loss": 1.2484, "step": 12270 }, { "epoch": 5.909025270758123, "grad_norm": 1.7339236736297607, "learning_rate": 4.0895522388059703e-05, "loss": 1.3603, "step": 12276 }, { "epoch": 5.911913357400722, "grad_norm": 1.9230111837387085, "learning_rate": 4.0866634569090035e-05, "loss": 1.4506, "step": 12282 }, { "epoch": 5.914801444043321, "grad_norm": 1.6120918989181519, "learning_rate": 4.0837746750120373e-05, "loss": 1.355, "step": 12288 }, { "epoch": 5.91768953068592, "grad_norm": 1.7437494993209839, "learning_rate": 4.08088589311507e-05, "loss": 1.2657, "step": 12294 }, { "epoch": 5.92057761732852, "grad_norm": 1.81756591796875, "learning_rate": 4.077997111218103e-05, "loss": 1.4559, "step": 12300 }, { "epoch": 5.923465703971119, "grad_norm": 1.6596312522888184, "learning_rate": 4.075108329321136e-05, "loss": 1.3035, "step": 12306 }, { "epoch": 5.926353790613718, "grad_norm": 1.717152714729309, "learning_rate": 4.07221954742417e-05, "loss": 1.362, "step": 12312 }, { "epoch": 5.929241877256318, "grad_norm": 1.8736882209777832, "learning_rate": 4.069330765527203e-05, "loss": 1.2596, "step": 12318 }, { "epoch": 5.932129963898917, "grad_norm": 2.1841235160827637, "learning_rate": 4.0664419836302356e-05, "loss": 1.507, "step": 12324 }, { "epoch": 5.935018050541516, "grad_norm": 1.8083436489105225, "learning_rate": 4.0635532017332695e-05, "loss": 1.4993, "step": 12330 }, { "epoch": 5.9379061371841155, "grad_norm": 1.800974726676941, "learning_rate": 4.0606644198363026e-05, "loss": 1.5012, "step": 12336 }, { "epoch": 5.940794223826715, "grad_norm": 1.8853440284729004, "learning_rate": 4.057775637939336e-05, "loss": 1.3611, "step": 12342 }, { "epoch": 5.943682310469314, "grad_norm": 2.1409683227539062, "learning_rate": 4.054886856042369e-05, "loss": 1.2798, "step": 12348 }, { "epoch": 5.9465703971119135, "grad_norm": 1.922129511833191, "learning_rate": 4.051998074145402e-05, "loss": 1.4087, "step": 12354 }, { "epoch": 5.949458483754513, "grad_norm": 1.9033502340316772, "learning_rate": 4.049109292248435e-05, "loss": 1.2344, "step": 12360 }, { "epoch": 5.952346570397112, "grad_norm": 1.6752101182937622, "learning_rate": 4.046220510351469e-05, "loss": 1.2973, "step": 12366 }, { "epoch": 5.955234657039711, "grad_norm": 1.8972097635269165, "learning_rate": 4.043331728454502e-05, "loss": 1.2313, "step": 12372 }, { "epoch": 5.958122743682311, "grad_norm": 1.8748421669006348, "learning_rate": 4.040442946557535e-05, "loss": 1.3155, "step": 12378 }, { "epoch": 5.96101083032491, "grad_norm": 1.7344088554382324, "learning_rate": 4.0375541646605686e-05, "loss": 1.4494, "step": 12384 }, { "epoch": 5.963898916967509, "grad_norm": 2.3799328804016113, "learning_rate": 4.034665382763602e-05, "loss": 1.5491, "step": 12390 }, { "epoch": 5.966787003610108, "grad_norm": 1.8558484315872192, "learning_rate": 4.031776600866635e-05, "loss": 1.4472, "step": 12396 }, { "epoch": 5.969675090252707, "grad_norm": 1.8774383068084717, "learning_rate": 4.028887818969668e-05, "loss": 1.3833, "step": 12402 }, { "epoch": 5.9725631768953065, "grad_norm": 2.15177845954895, "learning_rate": 4.025999037072701e-05, "loss": 1.5606, "step": 12408 }, { "epoch": 5.975451263537906, "grad_norm": 1.8685375452041626, "learning_rate": 4.0231102551757344e-05, "loss": 1.3502, "step": 12414 }, { "epoch": 5.978339350180505, "grad_norm": 1.6772881746292114, "learning_rate": 4.0202214732787675e-05, "loss": 1.3623, "step": 12420 }, { "epoch": 5.981227436823104, "grad_norm": 2.2881078720092773, "learning_rate": 4.017332691381801e-05, "loss": 1.3219, "step": 12426 }, { "epoch": 5.984115523465704, "grad_norm": 1.5685398578643799, "learning_rate": 4.014443909484834e-05, "loss": 1.3142, "step": 12432 }, { "epoch": 5.987003610108303, "grad_norm": 1.7786245346069336, "learning_rate": 4.011555127587867e-05, "loss": 1.3855, "step": 12438 }, { "epoch": 5.989891696750902, "grad_norm": 1.9545096158981323, "learning_rate": 4.008666345690901e-05, "loss": 1.3725, "step": 12444 }, { "epoch": 5.992779783393502, "grad_norm": 2.0671491622924805, "learning_rate": 4.005777563793934e-05, "loss": 1.3956, "step": 12450 }, { "epoch": 5.995667870036101, "grad_norm": 2.138389825820923, "learning_rate": 4.0028887818969665e-05, "loss": 1.3893, "step": 12456 }, { "epoch": 5.9985559566787, "grad_norm": 1.794791579246521, "learning_rate": 4e-05, "loss": 1.3896, "step": 12462 }, { "epoch": 6.0014440433213, "grad_norm": 2.1021018028259277, "learning_rate": 3.9971112181030335e-05, "loss": 1.3539, "step": 12468 }, { "epoch": 6.004332129963899, "grad_norm": 1.4785107374191284, "learning_rate": 3.9942224362060666e-05, "loss": 1.3274, "step": 12474 }, { "epoch": 6.007220216606498, "grad_norm": 2.1230878829956055, "learning_rate": 3.9913336543091e-05, "loss": 1.3729, "step": 12480 }, { "epoch": 6.010108303249098, "grad_norm": 1.7914018630981445, "learning_rate": 3.988444872412133e-05, "loss": 1.3188, "step": 12486 }, { "epoch": 6.012996389891697, "grad_norm": 1.6376656293869019, "learning_rate": 3.985556090515166e-05, "loss": 1.2418, "step": 12492 }, { "epoch": 6.015884476534296, "grad_norm": 2.012544870376587, "learning_rate": 3.982667308618199e-05, "loss": 1.2535, "step": 12498 }, { "epoch": 6.018772563176896, "grad_norm": 2.1302671432495117, "learning_rate": 3.979778526721233e-05, "loss": 1.3524, "step": 12504 }, { "epoch": 6.021660649819495, "grad_norm": 1.9295530319213867, "learning_rate": 3.9768897448242656e-05, "loss": 1.1391, "step": 12510 }, { "epoch": 6.024548736462094, "grad_norm": 1.8256555795669556, "learning_rate": 3.974000962927299e-05, "loss": 1.2022, "step": 12516 }, { "epoch": 6.0274368231046935, "grad_norm": 2.0539205074310303, "learning_rate": 3.9711121810303326e-05, "loss": 1.3815, "step": 12522 }, { "epoch": 6.030324909747293, "grad_norm": 1.771124005317688, "learning_rate": 3.968223399133366e-05, "loss": 1.2214, "step": 12528 }, { "epoch": 6.033212996389891, "grad_norm": 1.78920578956604, "learning_rate": 3.965334617236399e-05, "loss": 1.2692, "step": 12534 }, { "epoch": 6.036101083032491, "grad_norm": 2.1181342601776123, "learning_rate": 3.962445835339432e-05, "loss": 1.4804, "step": 12540 }, { "epoch": 6.03898916967509, "grad_norm": 1.9736062288284302, "learning_rate": 3.959557053442465e-05, "loss": 1.271, "step": 12546 }, { "epoch": 6.041877256317689, "grad_norm": 2.2139220237731934, "learning_rate": 3.9566682715454984e-05, "loss": 1.2811, "step": 12552 }, { "epoch": 6.044765342960289, "grad_norm": 1.9621381759643555, "learning_rate": 3.953779489648532e-05, "loss": 1.2532, "step": 12558 }, { "epoch": 6.047653429602888, "grad_norm": 2.1034348011016846, "learning_rate": 3.950890707751565e-05, "loss": 1.3667, "step": 12564 }, { "epoch": 6.050541516245487, "grad_norm": 1.9443104267120361, "learning_rate": 3.948001925854598e-05, "loss": 1.3451, "step": 12570 }, { "epoch": 6.0534296028880865, "grad_norm": 2.2242839336395264, "learning_rate": 3.945113143957632e-05, "loss": 1.3838, "step": 12576 }, { "epoch": 6.056317689530686, "grad_norm": 2.152243137359619, "learning_rate": 3.942224362060665e-05, "loss": 1.2769, "step": 12582 }, { "epoch": 6.059205776173285, "grad_norm": 2.035865306854248, "learning_rate": 3.939335580163698e-05, "loss": 1.4616, "step": 12588 }, { "epoch": 6.0620938628158845, "grad_norm": 2.161909580230713, "learning_rate": 3.9364467982667305e-05, "loss": 1.2669, "step": 12594 }, { "epoch": 6.064981949458484, "grad_norm": 1.8491957187652588, "learning_rate": 3.9335580163697644e-05, "loss": 1.4949, "step": 12600 }, { "epoch": 6.067870036101083, "grad_norm": 2.2088770866394043, "learning_rate": 3.9306692344727975e-05, "loss": 1.281, "step": 12606 }, { "epoch": 6.070758122743682, "grad_norm": 2.0980498790740967, "learning_rate": 3.927780452575831e-05, "loss": 1.3526, "step": 12612 }, { "epoch": 6.073646209386282, "grad_norm": 2.0058248043060303, "learning_rate": 3.924891670678864e-05, "loss": 1.3189, "step": 12618 }, { "epoch": 6.076534296028881, "grad_norm": 2.131687879562378, "learning_rate": 3.922002888781897e-05, "loss": 1.3599, "step": 12624 }, { "epoch": 6.07942238267148, "grad_norm": 2.022653341293335, "learning_rate": 3.91911410688493e-05, "loss": 1.3949, "step": 12630 }, { "epoch": 6.08231046931408, "grad_norm": 1.9371111392974854, "learning_rate": 3.916225324987964e-05, "loss": 1.2097, "step": 12636 }, { "epoch": 6.085198555956679, "grad_norm": 2.0980279445648193, "learning_rate": 3.913336543090997e-05, "loss": 1.4379, "step": 12642 }, { "epoch": 6.088086642599278, "grad_norm": 1.9599767923355103, "learning_rate": 3.9104477611940296e-05, "loss": 1.2288, "step": 12648 }, { "epoch": 6.090974729241878, "grad_norm": 2.140772581100464, "learning_rate": 3.9075589792970635e-05, "loss": 1.2956, "step": 12654 }, { "epoch": 6.093862815884476, "grad_norm": 1.8847256898880005, "learning_rate": 3.9046701974000966e-05, "loss": 1.2393, "step": 12660 }, { "epoch": 6.096750902527075, "grad_norm": 2.1283531188964844, "learning_rate": 3.90178141550313e-05, "loss": 1.355, "step": 12666 }, { "epoch": 6.099638989169675, "grad_norm": 2.0285041332244873, "learning_rate": 3.898892633606163e-05, "loss": 1.3532, "step": 12672 }, { "epoch": 6.102527075812274, "grad_norm": 2.1453232765197754, "learning_rate": 3.896003851709196e-05, "loss": 1.2172, "step": 12678 }, { "epoch": 6.105415162454873, "grad_norm": 1.7310751676559448, "learning_rate": 3.893115069812229e-05, "loss": 1.3127, "step": 12684 }, { "epoch": 6.108303249097473, "grad_norm": 2.142002820968628, "learning_rate": 3.8902262879152624e-05, "loss": 1.4237, "step": 12690 }, { "epoch": 6.111191335740072, "grad_norm": 1.7624256610870361, "learning_rate": 3.887337506018296e-05, "loss": 1.2464, "step": 12696 }, { "epoch": 6.114079422382671, "grad_norm": 1.9412416219711304, "learning_rate": 3.884448724121329e-05, "loss": 1.3444, "step": 12702 }, { "epoch": 6.116967509025271, "grad_norm": 1.8871822357177734, "learning_rate": 3.881559942224362e-05, "loss": 1.3087, "step": 12708 }, { "epoch": 6.11985559566787, "grad_norm": 2.3506288528442383, "learning_rate": 3.878671160327396e-05, "loss": 1.4325, "step": 12714 }, { "epoch": 6.122743682310469, "grad_norm": 2.187023878097534, "learning_rate": 3.875782378430429e-05, "loss": 1.3793, "step": 12720 }, { "epoch": 6.125631768953069, "grad_norm": 2.065196990966797, "learning_rate": 3.872893596533462e-05, "loss": 1.2981, "step": 12726 }, { "epoch": 6.128519855595668, "grad_norm": 2.088632822036743, "learning_rate": 3.870004814636495e-05, "loss": 1.3564, "step": 12732 }, { "epoch": 6.131407942238267, "grad_norm": 1.8097286224365234, "learning_rate": 3.8671160327395284e-05, "loss": 1.3854, "step": 12738 }, { "epoch": 6.134296028880867, "grad_norm": 1.868072509765625, "learning_rate": 3.8642272508425615e-05, "loss": 1.2408, "step": 12744 }, { "epoch": 6.137184115523466, "grad_norm": 1.9194953441619873, "learning_rate": 3.861338468945595e-05, "loss": 1.3297, "step": 12750 }, { "epoch": 6.140072202166065, "grad_norm": 2.0026934146881104, "learning_rate": 3.858449687048628e-05, "loss": 1.247, "step": 12756 }, { "epoch": 6.1429602888086645, "grad_norm": 2.0791313648223877, "learning_rate": 3.855560905151661e-05, "loss": 1.2765, "step": 12762 }, { "epoch": 6.145848375451264, "grad_norm": 1.9057096242904663, "learning_rate": 3.852672123254695e-05, "loss": 1.3153, "step": 12768 }, { "epoch": 6.148736462093863, "grad_norm": 1.8137800693511963, "learning_rate": 3.849783341357728e-05, "loss": 1.3655, "step": 12774 }, { "epoch": 6.1516245487364625, "grad_norm": 2.344698905944824, "learning_rate": 3.8468945594607605e-05, "loss": 1.2772, "step": 12780 }, { "epoch": 6.154512635379062, "grad_norm": 2.158886671066284, "learning_rate": 3.8440057775637937e-05, "loss": 1.2718, "step": 12786 }, { "epoch": 6.15740072202166, "grad_norm": 1.7736636400222778, "learning_rate": 3.8411169956668275e-05, "loss": 1.1999, "step": 12792 }, { "epoch": 6.1602888086642595, "grad_norm": 1.7308954000473022, "learning_rate": 3.8382282137698607e-05, "loss": 1.3106, "step": 12798 }, { "epoch": 6.163176895306859, "grad_norm": 1.8739991188049316, "learning_rate": 3.835339431872894e-05, "loss": 1.3106, "step": 12804 }, { "epoch": 6.166064981949458, "grad_norm": 1.8555270433425903, "learning_rate": 3.832450649975927e-05, "loss": 1.2647, "step": 12810 }, { "epoch": 6.1689530685920575, "grad_norm": 1.8527487516403198, "learning_rate": 3.82956186807896e-05, "loss": 1.3396, "step": 12816 }, { "epoch": 6.171841155234657, "grad_norm": 2.089982748031616, "learning_rate": 3.826673086181993e-05, "loss": 1.3519, "step": 12822 }, { "epoch": 6.174729241877256, "grad_norm": 2.250741481781006, "learning_rate": 3.823784304285027e-05, "loss": 1.3304, "step": 12828 }, { "epoch": 6.1776173285198555, "grad_norm": 2.0107860565185547, "learning_rate": 3.8208955223880596e-05, "loss": 1.3853, "step": 12834 }, { "epoch": 6.180505415162455, "grad_norm": 1.8693338632583618, "learning_rate": 3.818006740491093e-05, "loss": 1.2968, "step": 12840 }, { "epoch": 6.183393501805054, "grad_norm": 1.8955572843551636, "learning_rate": 3.8151179585941266e-05, "loss": 1.2442, "step": 12846 }, { "epoch": 6.186281588447653, "grad_norm": 2.148103952407837, "learning_rate": 3.81222917669716e-05, "loss": 1.3565, "step": 12852 }, { "epoch": 6.189169675090253, "grad_norm": 1.9617493152618408, "learning_rate": 3.809340394800193e-05, "loss": 1.4681, "step": 12858 }, { "epoch": 6.192057761732852, "grad_norm": 2.200397253036499, "learning_rate": 3.8064516129032254e-05, "loss": 1.3018, "step": 12864 }, { "epoch": 6.194945848375451, "grad_norm": 2.2233638763427734, "learning_rate": 3.803562831006259e-05, "loss": 1.3091, "step": 12870 }, { "epoch": 6.197833935018051, "grad_norm": 1.996849536895752, "learning_rate": 3.8006740491092924e-05, "loss": 1.1425, "step": 12876 }, { "epoch": 6.20072202166065, "grad_norm": 2.093247413635254, "learning_rate": 3.7977852672123256e-05, "loss": 1.4268, "step": 12882 }, { "epoch": 6.203610108303249, "grad_norm": 2.1991209983825684, "learning_rate": 3.794896485315359e-05, "loss": 1.2938, "step": 12888 }, { "epoch": 6.206498194945849, "grad_norm": 1.8842002153396606, "learning_rate": 3.792007703418392e-05, "loss": 1.2709, "step": 12894 }, { "epoch": 6.209386281588448, "grad_norm": 1.7757829427719116, "learning_rate": 3.789118921521425e-05, "loss": 1.3033, "step": 12900 }, { "epoch": 6.212274368231047, "grad_norm": 2.1156203746795654, "learning_rate": 3.786230139624459e-05, "loss": 1.4486, "step": 12906 }, { "epoch": 6.215162454873647, "grad_norm": 1.8334699869155884, "learning_rate": 3.783341357727492e-05, "loss": 1.1874, "step": 12912 }, { "epoch": 6.218050541516245, "grad_norm": 1.947995662689209, "learning_rate": 3.7804525758305245e-05, "loss": 1.4739, "step": 12918 }, { "epoch": 6.220938628158844, "grad_norm": 2.290067195892334, "learning_rate": 3.7775637939335584e-05, "loss": 1.3384, "step": 12924 }, { "epoch": 6.223826714801444, "grad_norm": 1.830267310142517, "learning_rate": 3.7746750120365915e-05, "loss": 1.2771, "step": 12930 }, { "epoch": 6.226714801444043, "grad_norm": 2.1026790142059326, "learning_rate": 3.771786230139625e-05, "loss": 1.3727, "step": 12936 }, { "epoch": 6.229602888086642, "grad_norm": 1.8266420364379883, "learning_rate": 3.768897448242658e-05, "loss": 1.3723, "step": 12942 }, { "epoch": 6.232490974729242, "grad_norm": 2.2374696731567383, "learning_rate": 3.766008666345691e-05, "loss": 1.4521, "step": 12948 }, { "epoch": 6.235379061371841, "grad_norm": 1.578538179397583, "learning_rate": 3.763119884448724e-05, "loss": 1.2017, "step": 12954 }, { "epoch": 6.23826714801444, "grad_norm": 2.0472538471221924, "learning_rate": 3.760231102551757e-05, "loss": 1.2753, "step": 12960 }, { "epoch": 6.24115523465704, "grad_norm": 1.9548298120498657, "learning_rate": 3.757342320654791e-05, "loss": 1.2513, "step": 12966 }, { "epoch": 6.244043321299639, "grad_norm": 1.9903643131256104, "learning_rate": 3.7544535387578236e-05, "loss": 1.3641, "step": 12972 }, { "epoch": 6.246931407942238, "grad_norm": 1.9193955659866333, "learning_rate": 3.751564756860857e-05, "loss": 1.3596, "step": 12978 }, { "epoch": 6.2498194945848375, "grad_norm": 2.2139837741851807, "learning_rate": 3.7486759749638906e-05, "loss": 1.3646, "step": 12984 }, { "epoch": 6.252707581227437, "grad_norm": 2.080420970916748, "learning_rate": 3.745787193066924e-05, "loss": 1.2801, "step": 12990 }, { "epoch": 6.255595667870036, "grad_norm": 1.8530288934707642, "learning_rate": 3.742898411169957e-05, "loss": 1.2188, "step": 12996 }, { "epoch": 6.2584837545126355, "grad_norm": 2.3050553798675537, "learning_rate": 3.74000962927299e-05, "loss": 1.383, "step": 13002 }, { "epoch": 6.261371841155235, "grad_norm": 2.2345874309539795, "learning_rate": 3.737120847376023e-05, "loss": 1.3457, "step": 13008 }, { "epoch": 6.264259927797834, "grad_norm": 1.7052682638168335, "learning_rate": 3.7342320654790564e-05, "loss": 1.2608, "step": 13014 }, { "epoch": 6.2671480144404335, "grad_norm": 2.027405023574829, "learning_rate": 3.73134328358209e-05, "loss": 1.4392, "step": 13020 }, { "epoch": 6.270036101083033, "grad_norm": 2.044365406036377, "learning_rate": 3.728454501685123e-05, "loss": 1.3284, "step": 13026 }, { "epoch": 6.272924187725632, "grad_norm": 1.931955099105835, "learning_rate": 3.725565719788156e-05, "loss": 1.4159, "step": 13032 }, { "epoch": 6.275812274368231, "grad_norm": 1.8093472719192505, "learning_rate": 3.72267693789119e-05, "loss": 1.2494, "step": 13038 }, { "epoch": 6.278700361010831, "grad_norm": 1.576155662536621, "learning_rate": 3.719788155994223e-05, "loss": 1.2074, "step": 13044 }, { "epoch": 6.28158844765343, "grad_norm": 1.9408760070800781, "learning_rate": 3.716899374097256e-05, "loss": 1.4148, "step": 13050 }, { "epoch": 6.2844765342960285, "grad_norm": 2.0714027881622314, "learning_rate": 3.7140105922002885e-05, "loss": 1.2891, "step": 13056 }, { "epoch": 6.287364620938628, "grad_norm": 1.9418405294418335, "learning_rate": 3.7111218103033224e-05, "loss": 1.4627, "step": 13062 }, { "epoch": 6.290252707581227, "grad_norm": 2.369185447692871, "learning_rate": 3.7082330284063555e-05, "loss": 1.3421, "step": 13068 }, { "epoch": 6.293140794223826, "grad_norm": 2.00561785697937, "learning_rate": 3.705344246509389e-05, "loss": 1.3632, "step": 13074 }, { "epoch": 6.296028880866426, "grad_norm": 2.30251145362854, "learning_rate": 3.702455464612422e-05, "loss": 1.2571, "step": 13080 }, { "epoch": 6.298916967509025, "grad_norm": 2.082336664199829, "learning_rate": 3.699566682715455e-05, "loss": 1.3731, "step": 13086 }, { "epoch": 6.301805054151624, "grad_norm": 2.041266918182373, "learning_rate": 3.696677900818488e-05, "loss": 1.323, "step": 13092 }, { "epoch": 6.304693140794224, "grad_norm": 1.982439637184143, "learning_rate": 3.693789118921522e-05, "loss": 1.3554, "step": 13098 }, { "epoch": 6.307581227436823, "grad_norm": 2.0283870697021484, "learning_rate": 3.690900337024555e-05, "loss": 1.3792, "step": 13104 }, { "epoch": 6.310469314079422, "grad_norm": 1.9673742055892944, "learning_rate": 3.6880115551275877e-05, "loss": 1.3812, "step": 13110 }, { "epoch": 6.313357400722022, "grad_norm": 1.7935924530029297, "learning_rate": 3.6851227732306215e-05, "loss": 1.3314, "step": 13116 }, { "epoch": 6.316245487364621, "grad_norm": 1.8954070806503296, "learning_rate": 3.6822339913336547e-05, "loss": 1.3668, "step": 13122 }, { "epoch": 6.31913357400722, "grad_norm": 1.998505711555481, "learning_rate": 3.679345209436688e-05, "loss": 1.2209, "step": 13128 }, { "epoch": 6.32202166064982, "grad_norm": 2.058203935623169, "learning_rate": 3.676456427539721e-05, "loss": 1.412, "step": 13134 }, { "epoch": 6.324909747292419, "grad_norm": 2.163821220397949, "learning_rate": 3.673567645642754e-05, "loss": 1.3243, "step": 13140 }, { "epoch": 6.327797833935018, "grad_norm": 2.0029637813568115, "learning_rate": 3.670678863745787e-05, "loss": 1.4012, "step": 13146 }, { "epoch": 6.330685920577618, "grad_norm": 1.9769597053527832, "learning_rate": 3.6677900818488205e-05, "loss": 1.2926, "step": 13152 }, { "epoch": 6.333574007220217, "grad_norm": 2.097107410430908, "learning_rate": 3.6649012999518536e-05, "loss": 1.41, "step": 13158 }, { "epoch": 6.336462093862816, "grad_norm": 2.1978600025177, "learning_rate": 3.662012518054887e-05, "loss": 1.3383, "step": 13164 }, { "epoch": 6.3393501805054155, "grad_norm": 2.2370049953460693, "learning_rate": 3.65912373615792e-05, "loss": 1.3727, "step": 13170 }, { "epoch": 6.342238267148014, "grad_norm": 1.9544514417648315, "learning_rate": 3.656234954260954e-05, "loss": 1.2355, "step": 13176 }, { "epoch": 6.345126353790613, "grad_norm": 1.92686927318573, "learning_rate": 3.653346172363987e-05, "loss": 1.3302, "step": 13182 }, { "epoch": 6.348014440433213, "grad_norm": 2.5676190853118896, "learning_rate": 3.6504573904670194e-05, "loss": 1.3914, "step": 13188 }, { "epoch": 6.350902527075812, "grad_norm": 2.0156211853027344, "learning_rate": 3.647568608570053e-05, "loss": 1.3592, "step": 13194 }, { "epoch": 6.353790613718411, "grad_norm": 2.064894437789917, "learning_rate": 3.6446798266730864e-05, "loss": 1.404, "step": 13200 }, { "epoch": 6.356678700361011, "grad_norm": 2.2358882427215576, "learning_rate": 3.6417910447761196e-05, "loss": 1.3712, "step": 13206 }, { "epoch": 6.35956678700361, "grad_norm": 1.95382559299469, "learning_rate": 3.638902262879153e-05, "loss": 1.4407, "step": 13212 }, { "epoch": 6.362454873646209, "grad_norm": 2.282245635986328, "learning_rate": 3.636013480982186e-05, "loss": 1.317, "step": 13218 }, { "epoch": 6.3653429602888085, "grad_norm": 2.4631175994873047, "learning_rate": 3.633124699085219e-05, "loss": 1.4832, "step": 13224 }, { "epoch": 6.368231046931408, "grad_norm": 1.8183037042617798, "learning_rate": 3.630235917188253e-05, "loss": 1.2859, "step": 13230 }, { "epoch": 6.371119133574007, "grad_norm": 1.6793725490570068, "learning_rate": 3.627347135291286e-05, "loss": 1.2912, "step": 13236 }, { "epoch": 6.3740072202166065, "grad_norm": 2.045598030090332, "learning_rate": 3.6244583533943185e-05, "loss": 1.3943, "step": 13242 }, { "epoch": 6.376895306859206, "grad_norm": 2.3434243202209473, "learning_rate": 3.621569571497352e-05, "loss": 1.3872, "step": 13248 }, { "epoch": 6.379783393501805, "grad_norm": 2.1125779151916504, "learning_rate": 3.6186807896003855e-05, "loss": 1.3097, "step": 13254 }, { "epoch": 6.382671480144404, "grad_norm": 2.4115211963653564, "learning_rate": 3.615792007703419e-05, "loss": 1.3275, "step": 13260 }, { "epoch": 6.385559566787004, "grad_norm": 1.8257559537887573, "learning_rate": 3.612903225806452e-05, "loss": 1.3454, "step": 13266 }, { "epoch": 6.388447653429603, "grad_norm": 2.0949103832244873, "learning_rate": 3.610014443909485e-05, "loss": 1.3901, "step": 13272 }, { "epoch": 6.391335740072202, "grad_norm": 2.2928035259246826, "learning_rate": 3.607125662012518e-05, "loss": 1.2302, "step": 13278 }, { "epoch": 6.394223826714802, "grad_norm": 2.2742626667022705, "learning_rate": 3.604236880115551e-05, "loss": 1.3243, "step": 13284 }, { "epoch": 6.397111913357401, "grad_norm": 2.023577928543091, "learning_rate": 3.601348098218585e-05, "loss": 1.3695, "step": 13290 }, { "epoch": 6.4, "grad_norm": 1.7191404104232788, "learning_rate": 3.5984593163216176e-05, "loss": 1.4046, "step": 13296 }, { "epoch": 6.4028880866426, "grad_norm": 1.8727620840072632, "learning_rate": 3.595570534424651e-05, "loss": 1.2754, "step": 13302 }, { "epoch": 6.405776173285199, "grad_norm": 1.6530207395553589, "learning_rate": 3.5926817525276846e-05, "loss": 1.1796, "step": 13308 }, { "epoch": 6.408664259927797, "grad_norm": 2.0881009101867676, "learning_rate": 3.589792970630718e-05, "loss": 1.2178, "step": 13314 }, { "epoch": 6.411552346570397, "grad_norm": 1.7391626834869385, "learning_rate": 3.586904188733751e-05, "loss": 1.3469, "step": 13320 }, { "epoch": 6.414440433212996, "grad_norm": 2.2128288745880127, "learning_rate": 3.5840154068367834e-05, "loss": 1.3513, "step": 13326 }, { "epoch": 6.417328519855595, "grad_norm": 1.7997967004776, "learning_rate": 3.581126624939817e-05, "loss": 1.4823, "step": 13332 }, { "epoch": 6.420216606498195, "grad_norm": 2.0978808403015137, "learning_rate": 3.5782378430428504e-05, "loss": 1.3952, "step": 13338 }, { "epoch": 6.423104693140794, "grad_norm": 2.0215840339660645, "learning_rate": 3.5753490611458836e-05, "loss": 1.4898, "step": 13344 }, { "epoch": 6.425992779783393, "grad_norm": 2.0582714080810547, "learning_rate": 3.572460279248917e-05, "loss": 1.3305, "step": 13350 }, { "epoch": 6.428880866425993, "grad_norm": 1.978116512298584, "learning_rate": 3.56957149735195e-05, "loss": 1.3496, "step": 13356 }, { "epoch": 6.431768953068592, "grad_norm": 1.9430569410324097, "learning_rate": 3.566682715454983e-05, "loss": 1.1726, "step": 13362 }, { "epoch": 6.434657039711191, "grad_norm": 1.9519298076629639, "learning_rate": 3.563793933558017e-05, "loss": 1.2401, "step": 13368 }, { "epoch": 6.437545126353791, "grad_norm": 2.2902326583862305, "learning_rate": 3.56090515166105e-05, "loss": 1.4338, "step": 13374 }, { "epoch": 6.44043321299639, "grad_norm": 2.048083543777466, "learning_rate": 3.5580163697640826e-05, "loss": 1.444, "step": 13380 }, { "epoch": 6.443321299638989, "grad_norm": 1.8564822673797607, "learning_rate": 3.5551275878671164e-05, "loss": 1.3981, "step": 13386 }, { "epoch": 6.446209386281589, "grad_norm": 2.2005515098571777, "learning_rate": 3.5522388059701495e-05, "loss": 1.3701, "step": 13392 }, { "epoch": 6.449097472924188, "grad_norm": 2.1352627277374268, "learning_rate": 3.549350024073183e-05, "loss": 1.3862, "step": 13398 }, { "epoch": 6.451985559566787, "grad_norm": 2.0659098625183105, "learning_rate": 3.546461242176216e-05, "loss": 1.4609, "step": 13404 }, { "epoch": 6.4548736462093865, "grad_norm": 1.885061264038086, "learning_rate": 3.543572460279249e-05, "loss": 1.261, "step": 13410 }, { "epoch": 6.457761732851986, "grad_norm": 2.1201045513153076, "learning_rate": 3.540683678382282e-05, "loss": 1.403, "step": 13416 }, { "epoch": 6.460649819494585, "grad_norm": 2.0786256790161133, "learning_rate": 3.5377948964853153e-05, "loss": 1.3455, "step": 13422 }, { "epoch": 6.4635379061371845, "grad_norm": 2.0821759700775146, "learning_rate": 3.534906114588349e-05, "loss": 1.3853, "step": 13428 }, { "epoch": 6.466425992779784, "grad_norm": 1.978011965751648, "learning_rate": 3.532017332691382e-05, "loss": 1.3304, "step": 13434 }, { "epoch": 6.469314079422382, "grad_norm": 2.0738377571105957, "learning_rate": 3.529128550794415e-05, "loss": 1.2215, "step": 13440 }, { "epoch": 6.4722021660649816, "grad_norm": 2.0587310791015625, "learning_rate": 3.526239768897449e-05, "loss": 1.3263, "step": 13446 }, { "epoch": 6.475090252707581, "grad_norm": 1.9080551862716675, "learning_rate": 3.523350987000482e-05, "loss": 1.3467, "step": 13452 }, { "epoch": 6.47797833935018, "grad_norm": 2.13201904296875, "learning_rate": 3.520462205103515e-05, "loss": 1.4372, "step": 13458 }, { "epoch": 6.4808664259927795, "grad_norm": 2.2397775650024414, "learning_rate": 3.517573423206548e-05, "loss": 1.3803, "step": 13464 }, { "epoch": 6.483754512635379, "grad_norm": 1.9216378927230835, "learning_rate": 3.514684641309581e-05, "loss": 1.2832, "step": 13470 }, { "epoch": 6.486642599277978, "grad_norm": 1.981704592704773, "learning_rate": 3.5117958594126145e-05, "loss": 1.2586, "step": 13476 }, { "epoch": 6.4895306859205775, "grad_norm": 1.913421869277954, "learning_rate": 3.5089070775156476e-05, "loss": 1.1993, "step": 13482 }, { "epoch": 6.492418772563177, "grad_norm": 1.9839037656784058, "learning_rate": 3.506018295618681e-05, "loss": 1.3218, "step": 13488 }, { "epoch": 6.495306859205776, "grad_norm": 2.0137104988098145, "learning_rate": 3.503129513721714e-05, "loss": 1.357, "step": 13494 }, { "epoch": 6.498194945848375, "grad_norm": 2.1069047451019287, "learning_rate": 3.500240731824748e-05, "loss": 1.394, "step": 13500 }, { "epoch": 6.501083032490975, "grad_norm": 2.3328614234924316, "learning_rate": 3.497351949927781e-05, "loss": 1.4354, "step": 13506 }, { "epoch": 6.503971119133574, "grad_norm": 1.9804739952087402, "learning_rate": 3.4944631680308134e-05, "loss": 1.3258, "step": 13512 }, { "epoch": 6.506859205776173, "grad_norm": 1.8180625438690186, "learning_rate": 3.4915743861338466e-05, "loss": 1.2977, "step": 13518 }, { "epoch": 6.509747292418773, "grad_norm": 1.8896945714950562, "learning_rate": 3.4886856042368804e-05, "loss": 1.2187, "step": 13524 }, { "epoch": 6.512635379061372, "grad_norm": 2.192333459854126, "learning_rate": 3.4857968223399136e-05, "loss": 1.3298, "step": 13530 }, { "epoch": 6.515523465703971, "grad_norm": 2.363006353378296, "learning_rate": 3.482908040442947e-05, "loss": 1.2285, "step": 13536 }, { "epoch": 6.518411552346571, "grad_norm": 1.9818804264068604, "learning_rate": 3.48001925854598e-05, "loss": 1.3959, "step": 13542 }, { "epoch": 6.52129963898917, "grad_norm": 2.2765305042266846, "learning_rate": 3.477130476649013e-05, "loss": 1.3849, "step": 13548 }, { "epoch": 6.524187725631769, "grad_norm": 2.075432777404785, "learning_rate": 3.474241694752046e-05, "loss": 1.3629, "step": 13554 }, { "epoch": 6.527075812274369, "grad_norm": 2.052759885787964, "learning_rate": 3.47135291285508e-05, "loss": 1.4023, "step": 13560 }, { "epoch": 6.529963898916968, "grad_norm": 2.1682379245758057, "learning_rate": 3.4684641309581125e-05, "loss": 1.413, "step": 13566 }, { "epoch": 6.532851985559567, "grad_norm": 2.1048147678375244, "learning_rate": 3.465575349061146e-05, "loss": 1.3033, "step": 13572 }, { "epoch": 6.535740072202166, "grad_norm": 2.1120145320892334, "learning_rate": 3.4626865671641795e-05, "loss": 1.2475, "step": 13578 }, { "epoch": 6.538628158844765, "grad_norm": 2.0168371200561523, "learning_rate": 3.459797785267213e-05, "loss": 1.476, "step": 13584 }, { "epoch": 6.541516245487364, "grad_norm": 2.1569178104400635, "learning_rate": 3.456909003370246e-05, "loss": 1.2847, "step": 13590 }, { "epoch": 6.544404332129964, "grad_norm": 2.1205708980560303, "learning_rate": 3.454020221473279e-05, "loss": 1.463, "step": 13596 }, { "epoch": 6.547292418772563, "grad_norm": 2.044015407562256, "learning_rate": 3.451131439576312e-05, "loss": 1.3442, "step": 13602 }, { "epoch": 6.550180505415162, "grad_norm": 1.8246663808822632, "learning_rate": 3.448242657679345e-05, "loss": 1.3143, "step": 13608 }, { "epoch": 6.553068592057762, "grad_norm": 2.079723596572876, "learning_rate": 3.4453538757823785e-05, "loss": 1.3255, "step": 13614 }, { "epoch": 6.555956678700361, "grad_norm": 2.0468392372131348, "learning_rate": 3.4424650938854116e-05, "loss": 1.4191, "step": 13620 }, { "epoch": 6.55884476534296, "grad_norm": 2.2494707107543945, "learning_rate": 3.439576311988445e-05, "loss": 1.2621, "step": 13626 }, { "epoch": 6.5617328519855596, "grad_norm": 2.35994815826416, "learning_rate": 3.436687530091478e-05, "loss": 1.3047, "step": 13632 }, { "epoch": 6.564620938628159, "grad_norm": 2.431471109390259, "learning_rate": 3.433798748194512e-05, "loss": 1.4564, "step": 13638 }, { "epoch": 6.567509025270758, "grad_norm": 2.2207846641540527, "learning_rate": 3.430909966297545e-05, "loss": 1.3681, "step": 13644 }, { "epoch": 6.5703971119133575, "grad_norm": 1.9036294221878052, "learning_rate": 3.4280211844005774e-05, "loss": 1.4099, "step": 13650 }, { "epoch": 6.573285198555957, "grad_norm": 2.3382749557495117, "learning_rate": 3.425132402503611e-05, "loss": 1.4212, "step": 13656 }, { "epoch": 6.576173285198556, "grad_norm": 2.3738365173339844, "learning_rate": 3.4222436206066444e-05, "loss": 1.3965, "step": 13662 }, { "epoch": 6.5790613718411555, "grad_norm": 2.221010208129883, "learning_rate": 3.4193548387096776e-05, "loss": 1.4896, "step": 13668 }, { "epoch": 6.581949458483755, "grad_norm": 2.301729917526245, "learning_rate": 3.416466056812711e-05, "loss": 1.5285, "step": 13674 }, { "epoch": 6.584837545126354, "grad_norm": 2.1413276195526123, "learning_rate": 3.413577274915744e-05, "loss": 1.3464, "step": 13680 }, { "epoch": 6.587725631768953, "grad_norm": 2.1112895011901855, "learning_rate": 3.410688493018777e-05, "loss": 1.3382, "step": 13686 }, { "epoch": 6.590613718411552, "grad_norm": 1.9682643413543701, "learning_rate": 3.407799711121811e-05, "loss": 1.209, "step": 13692 }, { "epoch": 6.593501805054151, "grad_norm": 2.167492151260376, "learning_rate": 3.404910929224844e-05, "loss": 1.4147, "step": 13698 }, { "epoch": 6.5963898916967505, "grad_norm": 1.9941587448120117, "learning_rate": 3.4020221473278766e-05, "loss": 1.4182, "step": 13704 }, { "epoch": 6.59927797833935, "grad_norm": 2.2123749256134033, "learning_rate": 3.39913336543091e-05, "loss": 1.3698, "step": 13710 }, { "epoch": 6.602166064981949, "grad_norm": 2.3372397422790527, "learning_rate": 3.3962445835339436e-05, "loss": 1.2855, "step": 13716 }, { "epoch": 6.6050541516245485, "grad_norm": 1.7581604719161987, "learning_rate": 3.393355801636977e-05, "loss": 1.2992, "step": 13722 }, { "epoch": 6.607942238267148, "grad_norm": 2.133925676345825, "learning_rate": 3.39046701974001e-05, "loss": 1.2404, "step": 13728 }, { "epoch": 6.610830324909747, "grad_norm": 1.8793286085128784, "learning_rate": 3.387578237843043e-05, "loss": 1.337, "step": 13734 }, { "epoch": 6.613718411552346, "grad_norm": 2.182652235031128, "learning_rate": 3.384689455946076e-05, "loss": 1.4463, "step": 13740 }, { "epoch": 6.616606498194946, "grad_norm": 1.9356266260147095, "learning_rate": 3.3818006740491094e-05, "loss": 1.4431, "step": 13746 }, { "epoch": 6.619494584837545, "grad_norm": 2.241422176361084, "learning_rate": 3.378911892152143e-05, "loss": 1.4396, "step": 13752 }, { "epoch": 6.622382671480144, "grad_norm": 2.35361647605896, "learning_rate": 3.376023110255176e-05, "loss": 1.4235, "step": 13758 }, { "epoch": 6.625270758122744, "grad_norm": 1.8256207704544067, "learning_rate": 3.373134328358209e-05, "loss": 1.3318, "step": 13764 }, { "epoch": 6.628158844765343, "grad_norm": 1.9854552745819092, "learning_rate": 3.370245546461243e-05, "loss": 1.3332, "step": 13770 }, { "epoch": 6.631046931407942, "grad_norm": 1.5280933380126953, "learning_rate": 3.367356764564276e-05, "loss": 1.2768, "step": 13776 }, { "epoch": 6.633935018050542, "grad_norm": 2.0747978687286377, "learning_rate": 3.364467982667309e-05, "loss": 1.373, "step": 13782 }, { "epoch": 6.636823104693141, "grad_norm": 2.1441943645477295, "learning_rate": 3.361579200770342e-05, "loss": 1.3539, "step": 13788 }, { "epoch": 6.63971119133574, "grad_norm": 2.275808572769165, "learning_rate": 3.358690418873375e-05, "loss": 1.3835, "step": 13794 }, { "epoch": 6.64259927797834, "grad_norm": 1.8950169086456299, "learning_rate": 3.3558016369764085e-05, "loss": 1.3601, "step": 13800 }, { "epoch": 6.645487364620939, "grad_norm": 2.5102648735046387, "learning_rate": 3.3529128550794416e-05, "loss": 1.3797, "step": 13806 }, { "epoch": 6.648375451263538, "grad_norm": 1.9168519973754883, "learning_rate": 3.350024073182475e-05, "loss": 1.3502, "step": 13812 }, { "epoch": 6.651263537906138, "grad_norm": 2.168057441711426, "learning_rate": 3.347135291285508e-05, "loss": 1.4576, "step": 13818 }, { "epoch": 6.654151624548737, "grad_norm": 2.0834097862243652, "learning_rate": 3.344246509388541e-05, "loss": 1.2719, "step": 13824 }, { "epoch": 6.657039711191336, "grad_norm": 1.9164800643920898, "learning_rate": 3.341357727491575e-05, "loss": 1.3443, "step": 13830 }, { "epoch": 6.6599277978339355, "grad_norm": 2.057058095932007, "learning_rate": 3.338468945594608e-05, "loss": 1.4268, "step": 13836 }, { "epoch": 6.662815884476534, "grad_norm": 2.0304322242736816, "learning_rate": 3.3355801636976406e-05, "loss": 1.3626, "step": 13842 }, { "epoch": 6.665703971119133, "grad_norm": 2.240356206893921, "learning_rate": 3.3326913818006744e-05, "loss": 1.432, "step": 13848 }, { "epoch": 6.668592057761733, "grad_norm": 2.2326059341430664, "learning_rate": 3.3298025999037076e-05, "loss": 1.3062, "step": 13854 }, { "epoch": 6.671480144404332, "grad_norm": 2.1657516956329346, "learning_rate": 3.326913818006741e-05, "loss": 1.3565, "step": 13860 }, { "epoch": 6.674368231046931, "grad_norm": 1.7401317358016968, "learning_rate": 3.324025036109774e-05, "loss": 1.3971, "step": 13866 }, { "epoch": 6.6772563176895305, "grad_norm": 2.1301217079162598, "learning_rate": 3.321136254212807e-05, "loss": 1.2814, "step": 13872 }, { "epoch": 6.68014440433213, "grad_norm": 1.9313137531280518, "learning_rate": 3.31824747231584e-05, "loss": 1.2632, "step": 13878 }, { "epoch": 6.683032490974729, "grad_norm": 2.148108720779419, "learning_rate": 3.3153586904188734e-05, "loss": 1.4191, "step": 13884 }, { "epoch": 6.6859205776173285, "grad_norm": 2.225947141647339, "learning_rate": 3.3124699085219065e-05, "loss": 1.3946, "step": 13890 }, { "epoch": 6.688808664259928, "grad_norm": 1.983741283416748, "learning_rate": 3.30958112662494e-05, "loss": 1.443, "step": 13896 }, { "epoch": 6.691696750902527, "grad_norm": 2.1427829265594482, "learning_rate": 3.306692344727973e-05, "loss": 1.2984, "step": 13902 }, { "epoch": 6.6945848375451265, "grad_norm": 2.236668348312378, "learning_rate": 3.303803562831007e-05, "loss": 1.3872, "step": 13908 }, { "epoch": 6.697472924187726, "grad_norm": 2.0448758602142334, "learning_rate": 3.30091478093404e-05, "loss": 1.4, "step": 13914 }, { "epoch": 6.700361010830325, "grad_norm": 2.7093663215637207, "learning_rate": 3.298025999037072e-05, "loss": 1.356, "step": 13920 }, { "epoch": 6.703249097472924, "grad_norm": 2.056847333908081, "learning_rate": 3.295137217140106e-05, "loss": 1.3289, "step": 13926 }, { "epoch": 6.706137184115524, "grad_norm": 1.8913872241973877, "learning_rate": 3.292248435243139e-05, "loss": 1.1991, "step": 13932 }, { "epoch": 6.709025270758123, "grad_norm": 2.1027042865753174, "learning_rate": 3.2893596533461725e-05, "loss": 1.3705, "step": 13938 }, { "epoch": 6.711913357400722, "grad_norm": 2.358003616333008, "learning_rate": 3.2864708714492057e-05, "loss": 1.1891, "step": 13944 }, { "epoch": 6.714801444043322, "grad_norm": 1.9723337888717651, "learning_rate": 3.283582089552239e-05, "loss": 1.4677, "step": 13950 }, { "epoch": 6.71768953068592, "grad_norm": 2.266407012939453, "learning_rate": 3.280693307655272e-05, "loss": 1.3154, "step": 13956 }, { "epoch": 6.720577617328519, "grad_norm": 2.0430335998535156, "learning_rate": 3.277804525758306e-05, "loss": 1.3088, "step": 13962 }, { "epoch": 6.723465703971119, "grad_norm": 2.1435258388519287, "learning_rate": 3.274915743861339e-05, "loss": 1.3142, "step": 13968 }, { "epoch": 6.726353790613718, "grad_norm": 1.9704300165176392, "learning_rate": 3.2720269619643714e-05, "loss": 1.3091, "step": 13974 }, { "epoch": 6.729241877256317, "grad_norm": 2.029926061630249, "learning_rate": 3.2691381800674046e-05, "loss": 1.3992, "step": 13980 }, { "epoch": 6.732129963898917, "grad_norm": 2.1187069416046143, "learning_rate": 3.2662493981704384e-05, "loss": 1.4783, "step": 13986 }, { "epoch": 6.735018050541516, "grad_norm": 1.9642643928527832, "learning_rate": 3.2633606162734716e-05, "loss": 1.3502, "step": 13992 }, { "epoch": 6.737906137184115, "grad_norm": 2.056483268737793, "learning_rate": 3.260471834376505e-05, "loss": 1.4096, "step": 13998 }, { "epoch": 6.740794223826715, "grad_norm": 2.340914487838745, "learning_rate": 3.257583052479538e-05, "loss": 1.3773, "step": 14004 }, { "epoch": 6.743682310469314, "grad_norm": 2.0396006107330322, "learning_rate": 3.254694270582571e-05, "loss": 1.3172, "step": 14010 }, { "epoch": 6.746570397111913, "grad_norm": 1.9927539825439453, "learning_rate": 3.251805488685604e-05, "loss": 1.2938, "step": 14016 }, { "epoch": 6.749458483754513, "grad_norm": 2.0610477924346924, "learning_rate": 3.248916706788638e-05, "loss": 1.2913, "step": 14022 }, { "epoch": 6.752346570397112, "grad_norm": 2.142198324203491, "learning_rate": 3.2460279248916706e-05, "loss": 1.309, "step": 14028 }, { "epoch": 6.755234657039711, "grad_norm": 2.022963047027588, "learning_rate": 3.243139142994704e-05, "loss": 1.2602, "step": 14034 }, { "epoch": 6.758122743682311, "grad_norm": 1.996519923210144, "learning_rate": 3.2402503610977376e-05, "loss": 1.367, "step": 14040 }, { "epoch": 6.76101083032491, "grad_norm": 2.1520936489105225, "learning_rate": 3.237361579200771e-05, "loss": 1.3815, "step": 14046 }, { "epoch": 6.763898916967509, "grad_norm": 2.2342798709869385, "learning_rate": 3.234472797303804e-05, "loss": 1.5146, "step": 14052 }, { "epoch": 6.7667870036101085, "grad_norm": 2.027791738510132, "learning_rate": 3.231584015406837e-05, "loss": 1.4273, "step": 14058 }, { "epoch": 6.769675090252708, "grad_norm": 1.990384578704834, "learning_rate": 3.22869523350987e-05, "loss": 1.1993, "step": 14064 }, { "epoch": 6.772563176895307, "grad_norm": 1.694976806640625, "learning_rate": 3.2258064516129034e-05, "loss": 1.3092, "step": 14070 }, { "epoch": 6.7754512635379065, "grad_norm": 2.4921963214874268, "learning_rate": 3.2229176697159365e-05, "loss": 1.3887, "step": 14076 }, { "epoch": 6.778339350180506, "grad_norm": 1.9910262823104858, "learning_rate": 3.22002888781897e-05, "loss": 1.323, "step": 14082 }, { "epoch": 6.781227436823105, "grad_norm": 2.1973018646240234, "learning_rate": 3.217140105922003e-05, "loss": 1.338, "step": 14088 }, { "epoch": 6.7841155234657045, "grad_norm": 2.0339014530181885, "learning_rate": 3.214251324025036e-05, "loss": 1.3863, "step": 14094 }, { "epoch": 6.787003610108303, "grad_norm": 2.1608047485351562, "learning_rate": 3.21136254212807e-05, "loss": 1.3811, "step": 14100 }, { "epoch": 6.789891696750902, "grad_norm": 2.3583872318267822, "learning_rate": 3.208473760231103e-05, "loss": 1.4725, "step": 14106 }, { "epoch": 6.7927797833935015, "grad_norm": 1.8191965818405151, "learning_rate": 3.2055849783341355e-05, "loss": 1.364, "step": 14112 }, { "epoch": 6.795667870036101, "grad_norm": 2.0503878593444824, "learning_rate": 3.202696196437169e-05, "loss": 1.4024, "step": 14118 }, { "epoch": 6.7985559566787, "grad_norm": 1.83818781375885, "learning_rate": 3.1998074145402025e-05, "loss": 1.2563, "step": 14124 }, { "epoch": 6.8014440433212995, "grad_norm": 2.3008313179016113, "learning_rate": 3.1969186326432356e-05, "loss": 1.4091, "step": 14130 }, { "epoch": 6.804332129963899, "grad_norm": 1.8955360651016235, "learning_rate": 3.194029850746269e-05, "loss": 1.395, "step": 14136 }, { "epoch": 6.807220216606498, "grad_norm": 2.0710670948028564, "learning_rate": 3.191141068849302e-05, "loss": 1.3762, "step": 14142 }, { "epoch": 6.810108303249097, "grad_norm": 2.323211908340454, "learning_rate": 3.188252286952335e-05, "loss": 1.3479, "step": 14148 }, { "epoch": 6.812996389891697, "grad_norm": 2.0192136764526367, "learning_rate": 3.185363505055369e-05, "loss": 1.3498, "step": 14154 }, { "epoch": 6.815884476534296, "grad_norm": 1.9817941188812256, "learning_rate": 3.182474723158402e-05, "loss": 1.318, "step": 14160 }, { "epoch": 6.818772563176895, "grad_norm": 2.1749179363250732, "learning_rate": 3.1795859412614346e-05, "loss": 1.3641, "step": 14166 }, { "epoch": 6.821660649819495, "grad_norm": 1.8101519346237183, "learning_rate": 3.176697159364468e-05, "loss": 1.3643, "step": 14172 }, { "epoch": 6.824548736462094, "grad_norm": 2.3149006366729736, "learning_rate": 3.1738083774675016e-05, "loss": 1.3616, "step": 14178 }, { "epoch": 6.827436823104693, "grad_norm": 1.8692431449890137, "learning_rate": 3.170919595570535e-05, "loss": 1.2194, "step": 14184 }, { "epoch": 6.830324909747293, "grad_norm": 1.9948549270629883, "learning_rate": 3.168030813673568e-05, "loss": 1.2977, "step": 14190 }, { "epoch": 6.833212996389892, "grad_norm": 1.9564955234527588, "learning_rate": 3.165142031776601e-05, "loss": 1.5387, "step": 14196 }, { "epoch": 6.836101083032491, "grad_norm": 2.008981704711914, "learning_rate": 3.162253249879634e-05, "loss": 1.2663, "step": 14202 }, { "epoch": 6.838989169675091, "grad_norm": 2.125422477722168, "learning_rate": 3.1593644679826674e-05, "loss": 1.3545, "step": 14208 }, { "epoch": 6.841877256317689, "grad_norm": 2.1856260299682617, "learning_rate": 3.1564756860857005e-05, "loss": 1.3446, "step": 14214 }, { "epoch": 6.844765342960288, "grad_norm": 2.2684152126312256, "learning_rate": 3.153586904188734e-05, "loss": 1.3561, "step": 14220 }, { "epoch": 6.847653429602888, "grad_norm": 2.106973171234131, "learning_rate": 3.150698122291767e-05, "loss": 1.4036, "step": 14226 }, { "epoch": 6.850541516245487, "grad_norm": 1.9567675590515137, "learning_rate": 3.147809340394801e-05, "loss": 1.3276, "step": 14232 }, { "epoch": 6.853429602888086, "grad_norm": 1.8021240234375, "learning_rate": 3.144920558497834e-05, "loss": 1.4046, "step": 14238 }, { "epoch": 6.856317689530686, "grad_norm": 2.179922580718994, "learning_rate": 3.1420317766008663e-05, "loss": 1.467, "step": 14244 }, { "epoch": 6.859205776173285, "grad_norm": 1.7912051677703857, "learning_rate": 3.1391429947039e-05, "loss": 1.2755, "step": 14250 }, { "epoch": 6.862093862815884, "grad_norm": 2.4015791416168213, "learning_rate": 3.136254212806933e-05, "loss": 1.4817, "step": 14256 }, { "epoch": 6.864981949458484, "grad_norm": 1.8893228769302368, "learning_rate": 3.1333654309099665e-05, "loss": 1.2788, "step": 14262 }, { "epoch": 6.867870036101083, "grad_norm": 1.8940479755401611, "learning_rate": 3.1304766490129997e-05, "loss": 1.3268, "step": 14268 }, { "epoch": 6.870758122743682, "grad_norm": 2.376673460006714, "learning_rate": 3.127587867116033e-05, "loss": 1.4446, "step": 14274 }, { "epoch": 6.873646209386282, "grad_norm": 2.0661487579345703, "learning_rate": 3.124699085219066e-05, "loss": 1.3979, "step": 14280 }, { "epoch": 6.876534296028881, "grad_norm": 2.1720292568206787, "learning_rate": 3.121810303322099e-05, "loss": 1.4312, "step": 14286 }, { "epoch": 6.87942238267148, "grad_norm": 2.253237247467041, "learning_rate": 3.118921521425133e-05, "loss": 1.2651, "step": 14292 }, { "epoch": 6.8823104693140795, "grad_norm": 2.205434799194336, "learning_rate": 3.1160327395281655e-05, "loss": 1.3593, "step": 14298 }, { "epoch": 6.885198555956679, "grad_norm": 1.892343521118164, "learning_rate": 3.1131439576311986e-05, "loss": 1.5479, "step": 14304 }, { "epoch": 6.888086642599278, "grad_norm": 1.9750456809997559, "learning_rate": 3.1102551757342325e-05, "loss": 1.3699, "step": 14310 }, { "epoch": 6.8909747292418775, "grad_norm": 2.2541091442108154, "learning_rate": 3.1073663938372656e-05, "loss": 1.3554, "step": 14316 }, { "epoch": 6.893862815884477, "grad_norm": 1.857726812362671, "learning_rate": 3.104477611940299e-05, "loss": 1.4027, "step": 14322 }, { "epoch": 6.896750902527076, "grad_norm": 1.8090591430664062, "learning_rate": 3.101588830043332e-05, "loss": 1.2621, "step": 14328 }, { "epoch": 6.899638989169675, "grad_norm": 2.0047237873077393, "learning_rate": 3.098700048146365e-05, "loss": 1.3457, "step": 14334 }, { "epoch": 6.902527075812275, "grad_norm": 2.0962107181549072, "learning_rate": 3.095811266249398e-05, "loss": 1.2916, "step": 14340 }, { "epoch": 6.905415162454874, "grad_norm": 2.1170454025268555, "learning_rate": 3.092922484352432e-05, "loss": 1.3188, "step": 14346 }, { "epoch": 6.908303249097473, "grad_norm": 2.256338596343994, "learning_rate": 3.0900337024554646e-05, "loss": 1.3448, "step": 14352 }, { "epoch": 6.911191335740072, "grad_norm": 2.071233034133911, "learning_rate": 3.087144920558498e-05, "loss": 1.565, "step": 14358 }, { "epoch": 6.914079422382671, "grad_norm": 1.736685872077942, "learning_rate": 3.084256138661531e-05, "loss": 1.1786, "step": 14364 }, { "epoch": 6.9169675090252705, "grad_norm": 2.0045108795166016, "learning_rate": 3.081367356764565e-05, "loss": 1.2745, "step": 14370 }, { "epoch": 6.91985559566787, "grad_norm": 2.081310272216797, "learning_rate": 3.078478574867598e-05, "loss": 1.3371, "step": 14376 }, { "epoch": 6.922743682310469, "grad_norm": 2.074223518371582, "learning_rate": 3.0755897929706304e-05, "loss": 1.315, "step": 14382 }, { "epoch": 6.925631768953068, "grad_norm": 2.059849262237549, "learning_rate": 3.072701011073664e-05, "loss": 1.3754, "step": 14388 }, { "epoch": 6.928519855595668, "grad_norm": 1.9962570667266846, "learning_rate": 3.0698122291766974e-05, "loss": 1.1744, "step": 14394 }, { "epoch": 6.931407942238267, "grad_norm": 2.0568816661834717, "learning_rate": 3.0669234472797305e-05, "loss": 1.3736, "step": 14400 }, { "epoch": 6.934296028880866, "grad_norm": 1.9728388786315918, "learning_rate": 3.064034665382764e-05, "loss": 1.4668, "step": 14406 }, { "epoch": 6.937184115523466, "grad_norm": 2.0445895195007324, "learning_rate": 3.061145883485797e-05, "loss": 1.1722, "step": 14412 }, { "epoch": 6.940072202166065, "grad_norm": 2.015259265899658, "learning_rate": 3.05825710158883e-05, "loss": 1.4915, "step": 14418 }, { "epoch": 6.942960288808664, "grad_norm": 2.0763933658599854, "learning_rate": 3.055368319691864e-05, "loss": 1.428, "step": 14424 }, { "epoch": 6.945848375451264, "grad_norm": 1.7801214456558228, "learning_rate": 3.052479537794897e-05, "loss": 1.3822, "step": 14430 }, { "epoch": 6.948736462093863, "grad_norm": 2.1252527236938477, "learning_rate": 3.0495907558979298e-05, "loss": 1.2063, "step": 14436 }, { "epoch": 6.951624548736462, "grad_norm": 2.281691312789917, "learning_rate": 3.0467019740009626e-05, "loss": 1.3636, "step": 14442 }, { "epoch": 6.954512635379062, "grad_norm": 2.1052117347717285, "learning_rate": 3.0438131921039965e-05, "loss": 1.2888, "step": 14448 }, { "epoch": 6.957400722021661, "grad_norm": 2.346125841140747, "learning_rate": 3.0409244102070293e-05, "loss": 1.3503, "step": 14454 }, { "epoch": 6.96028880866426, "grad_norm": 2.145902633666992, "learning_rate": 3.0380356283100625e-05, "loss": 1.3705, "step": 14460 }, { "epoch": 6.96317689530686, "grad_norm": 2.02055287361145, "learning_rate": 3.035146846413096e-05, "loss": 1.3689, "step": 14466 }, { "epoch": 6.966064981949458, "grad_norm": 2.3224618434906006, "learning_rate": 3.032258064516129e-05, "loss": 1.2915, "step": 14472 }, { "epoch": 6.968953068592057, "grad_norm": 2.037391185760498, "learning_rate": 3.0293692826191623e-05, "loss": 1.292, "step": 14478 }, { "epoch": 6.971841155234657, "grad_norm": 2.049356698989868, "learning_rate": 3.0264805007221958e-05, "loss": 1.3921, "step": 14484 }, { "epoch": 6.974729241877256, "grad_norm": 2.0224785804748535, "learning_rate": 3.023591718825229e-05, "loss": 1.4085, "step": 14490 }, { "epoch": 6.977617328519855, "grad_norm": 2.0940916538238525, "learning_rate": 3.0207029369282618e-05, "loss": 1.3331, "step": 14496 }, { "epoch": 6.980505415162455, "grad_norm": 2.0449419021606445, "learning_rate": 3.0178141550312956e-05, "loss": 1.4808, "step": 14502 }, { "epoch": 6.983393501805054, "grad_norm": 1.9780464172363281, "learning_rate": 3.0149253731343284e-05, "loss": 1.2405, "step": 14508 }, { "epoch": 6.986281588447653, "grad_norm": 2.1067726612091064, "learning_rate": 3.0120365912373616e-05, "loss": 1.395, "step": 14514 }, { "epoch": 6.9891696750902526, "grad_norm": 2.156315326690674, "learning_rate": 3.009147809340395e-05, "loss": 1.2828, "step": 14520 }, { "epoch": 6.992057761732852, "grad_norm": 2.262876033782959, "learning_rate": 3.0062590274434282e-05, "loss": 1.2427, "step": 14526 }, { "epoch": 6.994945848375451, "grad_norm": 2.247741937637329, "learning_rate": 3.0033702455464614e-05, "loss": 1.4121, "step": 14532 }, { "epoch": 6.9978339350180505, "grad_norm": 1.9669195413589478, "learning_rate": 3.0004814636494942e-05, "loss": 1.2883, "step": 14538 }, { "epoch": 7.00072202166065, "grad_norm": 1.9752209186553955, "learning_rate": 2.997592681752528e-05, "loss": 1.4106, "step": 14544 }, { "epoch": 7.003610108303249, "grad_norm": 2.2150895595550537, "learning_rate": 2.994703899855561e-05, "loss": 1.3538, "step": 14550 }, { "epoch": 7.0064981949458485, "grad_norm": 2.0981974601745605, "learning_rate": 2.991815117958594e-05, "loss": 1.2394, "step": 14556 }, { "epoch": 7.009386281588448, "grad_norm": 2.1260874271392822, "learning_rate": 2.9889263360616275e-05, "loss": 1.3165, "step": 14562 }, { "epoch": 7.012274368231047, "grad_norm": 2.4060280323028564, "learning_rate": 2.9860375541646607e-05, "loss": 1.3916, "step": 14568 }, { "epoch": 7.015162454873646, "grad_norm": 1.848456859588623, "learning_rate": 2.983148772267694e-05, "loss": 1.1962, "step": 14574 }, { "epoch": 7.018050541516246, "grad_norm": 1.816550850868225, "learning_rate": 2.9802599903707273e-05, "loss": 1.2513, "step": 14580 }, { "epoch": 7.020938628158845, "grad_norm": 2.3506507873535156, "learning_rate": 2.9773712084737605e-05, "loss": 1.4183, "step": 14586 }, { "epoch": 7.023826714801444, "grad_norm": 2.238711357116699, "learning_rate": 2.9744824265767933e-05, "loss": 1.3133, "step": 14592 }, { "epoch": 7.026714801444044, "grad_norm": 2.173229217529297, "learning_rate": 2.971593644679827e-05, "loss": 1.2464, "step": 14598 }, { "epoch": 7.029602888086643, "grad_norm": 2.2146549224853516, "learning_rate": 2.96870486278286e-05, "loss": 1.3215, "step": 14604 }, { "epoch": 7.0324909747292415, "grad_norm": 1.9929778575897217, "learning_rate": 2.965816080885893e-05, "loss": 1.1125, "step": 14610 }, { "epoch": 7.035379061371841, "grad_norm": 2.01400089263916, "learning_rate": 2.9629272989889266e-05, "loss": 1.3311, "step": 14616 }, { "epoch": 7.03826714801444, "grad_norm": 1.9139404296875, "learning_rate": 2.9600385170919598e-05, "loss": 1.2441, "step": 14622 }, { "epoch": 7.041155234657039, "grad_norm": 1.9874911308288574, "learning_rate": 2.957149735194993e-05, "loss": 1.2744, "step": 14628 }, { "epoch": 7.044043321299639, "grad_norm": 2.335536479949951, "learning_rate": 2.9542609532980258e-05, "loss": 1.4193, "step": 14634 }, { "epoch": 7.046931407942238, "grad_norm": 2.133687973022461, "learning_rate": 2.9513721714010596e-05, "loss": 1.4011, "step": 14640 }, { "epoch": 7.049819494584837, "grad_norm": 2.414933919906616, "learning_rate": 2.9484833895040924e-05, "loss": 1.3516, "step": 14646 }, { "epoch": 7.052707581227437, "grad_norm": 1.9889665842056274, "learning_rate": 2.9455946076071256e-05, "loss": 1.3184, "step": 14652 }, { "epoch": 7.055595667870036, "grad_norm": 2.0444176197052, "learning_rate": 2.942705825710159e-05, "loss": 1.3442, "step": 14658 }, { "epoch": 7.058483754512635, "grad_norm": 2.2121970653533936, "learning_rate": 2.9398170438131923e-05, "loss": 1.3738, "step": 14664 }, { "epoch": 7.061371841155235, "grad_norm": 2.242079257965088, "learning_rate": 2.9369282619162254e-05, "loss": 1.0973, "step": 14670 }, { "epoch": 7.064259927797834, "grad_norm": 2.384089946746826, "learning_rate": 2.934039480019259e-05, "loss": 1.3272, "step": 14676 }, { "epoch": 7.067148014440433, "grad_norm": 2.2968733310699463, "learning_rate": 2.931150698122292e-05, "loss": 1.3051, "step": 14682 }, { "epoch": 7.070036101083033, "grad_norm": 1.9639430046081543, "learning_rate": 2.928261916225325e-05, "loss": 1.3448, "step": 14688 }, { "epoch": 7.072924187725632, "grad_norm": 2.189775228500366, "learning_rate": 2.9253731343283587e-05, "loss": 1.3671, "step": 14694 }, { "epoch": 7.075812274368231, "grad_norm": 2.142028570175171, "learning_rate": 2.9224843524313916e-05, "loss": 1.3145, "step": 14700 }, { "epoch": 7.0787003610108306, "grad_norm": 2.326184034347534, "learning_rate": 2.9195955705344247e-05, "loss": 1.3027, "step": 14706 }, { "epoch": 7.08158844765343, "grad_norm": 2.1447956562042236, "learning_rate": 2.9167067886374582e-05, "loss": 1.2744, "step": 14712 }, { "epoch": 7.084476534296029, "grad_norm": 2.285935878753662, "learning_rate": 2.9138180067404914e-05, "loss": 1.2937, "step": 14718 }, { "epoch": 7.0873646209386285, "grad_norm": 2.1844773292541504, "learning_rate": 2.9109292248435245e-05, "loss": 1.2143, "step": 14724 }, { "epoch": 7.090252707581228, "grad_norm": 2.2296836376190186, "learning_rate": 2.9080404429465574e-05, "loss": 1.2352, "step": 14730 }, { "epoch": 7.093140794223827, "grad_norm": 2.0895187854766846, "learning_rate": 2.9051516610495912e-05, "loss": 1.3759, "step": 14736 }, { "epoch": 7.096028880866426, "grad_norm": 1.8765606880187988, "learning_rate": 2.902262879152624e-05, "loss": 1.1983, "step": 14742 }, { "epoch": 7.098916967509025, "grad_norm": 2.3796579837799072, "learning_rate": 2.899374097255657e-05, "loss": 1.246, "step": 14748 }, { "epoch": 7.101805054151624, "grad_norm": 2.4822275638580322, "learning_rate": 2.8964853153586907e-05, "loss": 1.2945, "step": 14754 }, { "epoch": 7.1046931407942235, "grad_norm": 2.052121877670288, "learning_rate": 2.8935965334617238e-05, "loss": 1.2594, "step": 14760 }, { "epoch": 7.107581227436823, "grad_norm": 1.9186431169509888, "learning_rate": 2.890707751564757e-05, "loss": 1.2334, "step": 14766 }, { "epoch": 7.110469314079422, "grad_norm": 1.9390826225280762, "learning_rate": 2.8878189696677905e-05, "loss": 1.3567, "step": 14772 }, { "epoch": 7.1133574007220215, "grad_norm": 2.223958730697632, "learning_rate": 2.8849301877708233e-05, "loss": 1.4377, "step": 14778 }, { "epoch": 7.116245487364621, "grad_norm": 2.67101788520813, "learning_rate": 2.8820414058738565e-05, "loss": 1.2773, "step": 14784 }, { "epoch": 7.11913357400722, "grad_norm": 2.0902130603790283, "learning_rate": 2.87915262397689e-05, "loss": 1.3599, "step": 14790 }, { "epoch": 7.1220216606498195, "grad_norm": 2.046217203140259, "learning_rate": 2.876263842079923e-05, "loss": 1.4173, "step": 14796 }, { "epoch": 7.124909747292419, "grad_norm": 2.2035062313079834, "learning_rate": 2.8733750601829563e-05, "loss": 1.2139, "step": 14802 }, { "epoch": 7.127797833935018, "grad_norm": 2.053544521331787, "learning_rate": 2.8704862782859898e-05, "loss": 1.2916, "step": 14808 }, { "epoch": 7.130685920577617, "grad_norm": 2.3134024143218994, "learning_rate": 2.867597496389023e-05, "loss": 1.4039, "step": 14814 }, { "epoch": 7.133574007220217, "grad_norm": 2.3305368423461914, "learning_rate": 2.8647087144920558e-05, "loss": 1.3655, "step": 14820 }, { "epoch": 7.136462093862816, "grad_norm": 2.1926629543304443, "learning_rate": 2.861819932595089e-05, "loss": 1.3164, "step": 14826 }, { "epoch": 7.139350180505415, "grad_norm": 2.133531093597412, "learning_rate": 2.8589311506981224e-05, "loss": 1.365, "step": 14832 }, { "epoch": 7.142238267148015, "grad_norm": 2.121251344680786, "learning_rate": 2.8560423688011556e-05, "loss": 1.3233, "step": 14838 }, { "epoch": 7.145126353790614, "grad_norm": 2.065376043319702, "learning_rate": 2.8531535869041887e-05, "loss": 1.3902, "step": 14844 }, { "epoch": 7.148014440433213, "grad_norm": 2.0350940227508545, "learning_rate": 2.8502648050072222e-05, "loss": 1.2904, "step": 14850 }, { "epoch": 7.150902527075813, "grad_norm": 2.3373630046844482, "learning_rate": 2.8473760231102554e-05, "loss": 1.1022, "step": 14856 }, { "epoch": 7.153790613718412, "grad_norm": 1.8651994466781616, "learning_rate": 2.8444872412132882e-05, "loss": 1.2238, "step": 14862 }, { "epoch": 7.156678700361011, "grad_norm": 2.035008668899536, "learning_rate": 2.841598459316322e-05, "loss": 1.1198, "step": 14868 }, { "epoch": 7.15956678700361, "grad_norm": 2.3811092376708984, "learning_rate": 2.838709677419355e-05, "loss": 1.27, "step": 14874 }, { "epoch": 7.162454873646209, "grad_norm": 2.1494102478027344, "learning_rate": 2.835820895522388e-05, "loss": 1.2916, "step": 14880 }, { "epoch": 7.165342960288808, "grad_norm": 2.3356525897979736, "learning_rate": 2.8329321136254215e-05, "loss": 1.2394, "step": 14886 }, { "epoch": 7.168231046931408, "grad_norm": 2.2005722522735596, "learning_rate": 2.8300433317284547e-05, "loss": 1.1998, "step": 14892 }, { "epoch": 7.171119133574007, "grad_norm": 2.1327602863311768, "learning_rate": 2.827154549831488e-05, "loss": 1.2998, "step": 14898 }, { "epoch": 7.174007220216606, "grad_norm": 2.115985870361328, "learning_rate": 2.8242657679345213e-05, "loss": 1.2974, "step": 14904 }, { "epoch": 7.176895306859206, "grad_norm": 2.350025177001953, "learning_rate": 2.8213769860375545e-05, "loss": 1.219, "step": 14910 }, { "epoch": 7.179783393501805, "grad_norm": 1.9179502725601196, "learning_rate": 2.8184882041405873e-05, "loss": 1.2573, "step": 14916 }, { "epoch": 7.182671480144404, "grad_norm": 1.8451985120773315, "learning_rate": 2.8155994222436205e-05, "loss": 1.2692, "step": 14922 }, { "epoch": 7.185559566787004, "grad_norm": 2.443845272064209, "learning_rate": 2.812710640346654e-05, "loss": 1.3154, "step": 14928 }, { "epoch": 7.188447653429603, "grad_norm": 2.3334531784057617, "learning_rate": 2.809821858449687e-05, "loss": 1.3123, "step": 14934 }, { "epoch": 7.191335740072202, "grad_norm": 1.9302979707717896, "learning_rate": 2.8069330765527203e-05, "loss": 1.2007, "step": 14940 }, { "epoch": 7.1942238267148015, "grad_norm": 2.392970323562622, "learning_rate": 2.8040442946557538e-05, "loss": 1.2354, "step": 14946 }, { "epoch": 7.197111913357401, "grad_norm": 1.94162118434906, "learning_rate": 2.801155512758787e-05, "loss": 1.2999, "step": 14952 }, { "epoch": 7.2, "grad_norm": 2.1090657711029053, "learning_rate": 2.7982667308618198e-05, "loss": 1.3802, "step": 14958 }, { "epoch": 7.2028880866425995, "grad_norm": 2.19401478767395, "learning_rate": 2.7953779489648536e-05, "loss": 1.3477, "step": 14964 }, { "epoch": 7.205776173285199, "grad_norm": 2.3425323963165283, "learning_rate": 2.7924891670678864e-05, "loss": 1.3398, "step": 14970 }, { "epoch": 7.208664259927798, "grad_norm": 2.421382427215576, "learning_rate": 2.7896003851709196e-05, "loss": 1.338, "step": 14976 }, { "epoch": 7.2115523465703975, "grad_norm": 1.950007677078247, "learning_rate": 2.786711603273953e-05, "loss": 1.1895, "step": 14982 }, { "epoch": 7.214440433212997, "grad_norm": 1.9278783798217773, "learning_rate": 2.7838228213769863e-05, "loss": 1.3792, "step": 14988 }, { "epoch": 7.217328519855596, "grad_norm": 2.1478066444396973, "learning_rate": 2.7809340394800194e-05, "loss": 1.3093, "step": 14994 }, { "epoch": 7.2202166064981945, "grad_norm": 2.0234525203704834, "learning_rate": 2.7780452575830522e-05, "loss": 1.2938, "step": 15000 }, { "epoch": 7.223104693140794, "grad_norm": 2.3196847438812256, "learning_rate": 2.775156475686086e-05, "loss": 1.4355, "step": 15006 }, { "epoch": 7.225992779783393, "grad_norm": 2.343885660171509, "learning_rate": 2.772267693789119e-05, "loss": 1.2604, "step": 15012 }, { "epoch": 7.2288808664259925, "grad_norm": 2.5910885334014893, "learning_rate": 2.769378911892152e-05, "loss": 1.3775, "step": 15018 }, { "epoch": 7.231768953068592, "grad_norm": 2.301701307296753, "learning_rate": 2.7664901299951856e-05, "loss": 1.3175, "step": 15024 }, { "epoch": 7.234657039711191, "grad_norm": 2.0991172790527344, "learning_rate": 2.7636013480982187e-05, "loss": 1.4067, "step": 15030 }, { "epoch": 7.23754512635379, "grad_norm": 2.3040804862976074, "learning_rate": 2.760712566201252e-05, "loss": 1.382, "step": 15036 }, { "epoch": 7.24043321299639, "grad_norm": 2.273277521133423, "learning_rate": 2.7578237843042854e-05, "loss": 1.3132, "step": 15042 }, { "epoch": 7.243321299638989, "grad_norm": 2.07668137550354, "learning_rate": 2.7549350024073185e-05, "loss": 1.2136, "step": 15048 }, { "epoch": 7.246209386281588, "grad_norm": 2.1075737476348877, "learning_rate": 2.7520462205103514e-05, "loss": 1.257, "step": 15054 }, { "epoch": 7.249097472924188, "grad_norm": 2.0710859298706055, "learning_rate": 2.7491574386133852e-05, "loss": 1.3443, "step": 15060 }, { "epoch": 7.251985559566787, "grad_norm": 2.1196136474609375, "learning_rate": 2.746268656716418e-05, "loss": 1.275, "step": 15066 }, { "epoch": 7.254873646209386, "grad_norm": 1.972159504890442, "learning_rate": 2.7433798748194512e-05, "loss": 1.2677, "step": 15072 }, { "epoch": 7.257761732851986, "grad_norm": 2.2575905323028564, "learning_rate": 2.7404910929224847e-05, "loss": 1.331, "step": 15078 }, { "epoch": 7.260649819494585, "grad_norm": 2.2448132038116455, "learning_rate": 2.737602311025518e-05, "loss": 1.3121, "step": 15084 }, { "epoch": 7.263537906137184, "grad_norm": 2.3498339653015137, "learning_rate": 2.734713529128551e-05, "loss": 1.375, "step": 15090 }, { "epoch": 7.266425992779784, "grad_norm": 2.009660005569458, "learning_rate": 2.7318247472315838e-05, "loss": 1.3038, "step": 15096 }, { "epoch": 7.269314079422383, "grad_norm": 1.905214548110962, "learning_rate": 2.7289359653346176e-05, "loss": 1.2524, "step": 15102 }, { "epoch": 7.272202166064982, "grad_norm": 2.166912317276001, "learning_rate": 2.7260471834376505e-05, "loss": 1.2201, "step": 15108 }, { "epoch": 7.275090252707582, "grad_norm": 2.3297297954559326, "learning_rate": 2.7231584015406836e-05, "loss": 1.2938, "step": 15114 }, { "epoch": 7.277978339350181, "grad_norm": 2.211768865585327, "learning_rate": 2.720269619643717e-05, "loss": 1.3224, "step": 15120 }, { "epoch": 7.28086642599278, "grad_norm": 2.2896621227264404, "learning_rate": 2.7173808377467503e-05, "loss": 1.3836, "step": 15126 }, { "epoch": 7.283754512635379, "grad_norm": 2.1588120460510254, "learning_rate": 2.7144920558497834e-05, "loss": 1.288, "step": 15132 }, { "epoch": 7.286642599277978, "grad_norm": 1.9184274673461914, "learning_rate": 2.711603273952817e-05, "loss": 1.2347, "step": 15138 }, { "epoch": 7.289530685920577, "grad_norm": 2.086366891860962, "learning_rate": 2.7087144920558498e-05, "loss": 1.1043, "step": 15144 }, { "epoch": 7.292418772563177, "grad_norm": 2.0788636207580566, "learning_rate": 2.705825710158883e-05, "loss": 1.2034, "step": 15150 }, { "epoch": 7.295306859205776, "grad_norm": 2.1878442764282227, "learning_rate": 2.7029369282619164e-05, "loss": 1.2915, "step": 15156 }, { "epoch": 7.298194945848375, "grad_norm": 2.0171077251434326, "learning_rate": 2.7000481463649496e-05, "loss": 1.2652, "step": 15162 }, { "epoch": 7.301083032490975, "grad_norm": 2.3236310482025146, "learning_rate": 2.6971593644679827e-05, "loss": 1.3666, "step": 15168 }, { "epoch": 7.303971119133574, "grad_norm": 2.1256818771362305, "learning_rate": 2.6942705825710162e-05, "loss": 1.3746, "step": 15174 }, { "epoch": 7.306859205776173, "grad_norm": 1.715587854385376, "learning_rate": 2.6913818006740494e-05, "loss": 1.2231, "step": 15180 }, { "epoch": 7.3097472924187725, "grad_norm": 2.300208330154419, "learning_rate": 2.6884930187770822e-05, "loss": 1.3619, "step": 15186 }, { "epoch": 7.312635379061372, "grad_norm": 1.9693948030471802, "learning_rate": 2.6856042368801154e-05, "loss": 1.4643, "step": 15192 }, { "epoch": 7.315523465703971, "grad_norm": 2.2002835273742676, "learning_rate": 2.682715454983149e-05, "loss": 1.2849, "step": 15198 }, { "epoch": 7.3184115523465705, "grad_norm": 2.3935768604278564, "learning_rate": 2.679826673086182e-05, "loss": 1.3346, "step": 15204 }, { "epoch": 7.32129963898917, "grad_norm": 2.001189947128296, "learning_rate": 2.6769378911892152e-05, "loss": 1.3506, "step": 15210 }, { "epoch": 7.324187725631769, "grad_norm": 1.9659454822540283, "learning_rate": 2.6740491092922487e-05, "loss": 1.2355, "step": 15216 }, { "epoch": 7.327075812274368, "grad_norm": 2.500028610229492, "learning_rate": 2.671160327395282e-05, "loss": 1.4146, "step": 15222 }, { "epoch": 7.329963898916968, "grad_norm": 2.1833832263946533, "learning_rate": 2.6682715454983147e-05, "loss": 1.1947, "step": 15228 }, { "epoch": 7.332851985559567, "grad_norm": 2.3576362133026123, "learning_rate": 2.6653827636013485e-05, "loss": 1.2563, "step": 15234 }, { "epoch": 7.335740072202166, "grad_norm": 2.56644868850708, "learning_rate": 2.6624939817043813e-05, "loss": 1.3279, "step": 15240 }, { "epoch": 7.338628158844766, "grad_norm": 2.2322018146514893, "learning_rate": 2.6596051998074145e-05, "loss": 1.1986, "step": 15246 }, { "epoch": 7.341516245487365, "grad_norm": 2.3351545333862305, "learning_rate": 2.656716417910448e-05, "loss": 1.3528, "step": 15252 }, { "epoch": 7.3444043321299635, "grad_norm": 2.35760235786438, "learning_rate": 2.653827636013481e-05, "loss": 1.2846, "step": 15258 }, { "epoch": 7.347292418772563, "grad_norm": 2.394383192062378, "learning_rate": 2.6509388541165143e-05, "loss": 1.3756, "step": 15264 }, { "epoch": 7.350180505415162, "grad_norm": 2.3711929321289062, "learning_rate": 2.6480500722195478e-05, "loss": 1.3535, "step": 15270 }, { "epoch": 7.353068592057761, "grad_norm": 2.146742105484009, "learning_rate": 2.645161290322581e-05, "loss": 1.2427, "step": 15276 }, { "epoch": 7.355956678700361, "grad_norm": 2.2010059356689453, "learning_rate": 2.6422725084256138e-05, "loss": 1.2125, "step": 15282 }, { "epoch": 7.35884476534296, "grad_norm": 2.119807004928589, "learning_rate": 2.639383726528647e-05, "loss": 1.2248, "step": 15288 }, { "epoch": 7.361732851985559, "grad_norm": 2.2936084270477295, "learning_rate": 2.6364949446316804e-05, "loss": 1.3009, "step": 15294 }, { "epoch": 7.364620938628159, "grad_norm": 2.394885301589966, "learning_rate": 2.6336061627347136e-05, "loss": 1.2633, "step": 15300 }, { "epoch": 7.367509025270758, "grad_norm": 2.3763296604156494, "learning_rate": 2.6307173808377468e-05, "loss": 1.2518, "step": 15306 }, { "epoch": 7.370397111913357, "grad_norm": 2.0544917583465576, "learning_rate": 2.6278285989407803e-05, "loss": 1.1122, "step": 15312 }, { "epoch": 7.373285198555957, "grad_norm": 2.1504461765289307, "learning_rate": 2.6249398170438134e-05, "loss": 1.2244, "step": 15318 }, { "epoch": 7.376173285198556, "grad_norm": 2.255826711654663, "learning_rate": 2.6220510351468462e-05, "loss": 1.2144, "step": 15324 }, { "epoch": 7.379061371841155, "grad_norm": 2.161203145980835, "learning_rate": 2.61916225324988e-05, "loss": 1.3007, "step": 15330 }, { "epoch": 7.381949458483755, "grad_norm": 2.2940750122070312, "learning_rate": 2.616273471352913e-05, "loss": 1.2294, "step": 15336 }, { "epoch": 7.384837545126354, "grad_norm": 2.398928165435791, "learning_rate": 2.613384689455946e-05, "loss": 1.3316, "step": 15342 }, { "epoch": 7.387725631768953, "grad_norm": 2.2840466499328613, "learning_rate": 2.6104959075589796e-05, "loss": 1.3146, "step": 15348 }, { "epoch": 7.390613718411553, "grad_norm": 2.188737154006958, "learning_rate": 2.6076071256620127e-05, "loss": 1.1755, "step": 15354 }, { "epoch": 7.393501805054152, "grad_norm": 1.9792438745498657, "learning_rate": 2.604718343765046e-05, "loss": 1.2856, "step": 15360 }, { "epoch": 7.396389891696751, "grad_norm": 1.8341894149780273, "learning_rate": 2.6018295618680794e-05, "loss": 1.2336, "step": 15366 }, { "epoch": 7.3992779783393505, "grad_norm": 1.9442938566207886, "learning_rate": 2.5989407799711125e-05, "loss": 1.2694, "step": 15372 }, { "epoch": 7.40216606498195, "grad_norm": 2.2578675746917725, "learning_rate": 2.5960519980741454e-05, "loss": 1.2984, "step": 15378 }, { "epoch": 7.405054151624549, "grad_norm": 2.1602561473846436, "learning_rate": 2.5931632161771785e-05, "loss": 1.2081, "step": 15384 }, { "epoch": 7.4079422382671485, "grad_norm": 2.152763843536377, "learning_rate": 2.590274434280212e-05, "loss": 1.2767, "step": 15390 }, { "epoch": 7.410830324909747, "grad_norm": 2.323044538497925, "learning_rate": 2.5873856523832452e-05, "loss": 1.381, "step": 15396 }, { "epoch": 7.413718411552346, "grad_norm": 2.1355791091918945, "learning_rate": 2.5844968704862783e-05, "loss": 1.325, "step": 15402 }, { "epoch": 7.4166064981949456, "grad_norm": 2.0928754806518555, "learning_rate": 2.581608088589312e-05, "loss": 1.3601, "step": 15408 }, { "epoch": 7.419494584837545, "grad_norm": 2.303727149963379, "learning_rate": 2.578719306692345e-05, "loss": 1.2936, "step": 15414 }, { "epoch": 7.422382671480144, "grad_norm": 2.4098784923553467, "learning_rate": 2.5758305247953778e-05, "loss": 1.3833, "step": 15420 }, { "epoch": 7.4252707581227435, "grad_norm": 2.2777061462402344, "learning_rate": 2.5729417428984117e-05, "loss": 1.4038, "step": 15426 }, { "epoch": 7.428158844765343, "grad_norm": 2.4891586303710938, "learning_rate": 2.5700529610014445e-05, "loss": 1.2615, "step": 15432 }, { "epoch": 7.431046931407942, "grad_norm": 2.1244149208068848, "learning_rate": 2.5671641791044776e-05, "loss": 1.2295, "step": 15438 }, { "epoch": 7.4339350180505415, "grad_norm": 1.9814139604568481, "learning_rate": 2.564275397207511e-05, "loss": 1.293, "step": 15444 }, { "epoch": 7.436823104693141, "grad_norm": 2.2788689136505127, "learning_rate": 2.5613866153105443e-05, "loss": 1.342, "step": 15450 }, { "epoch": 7.43971119133574, "grad_norm": 2.245830774307251, "learning_rate": 2.5584978334135775e-05, "loss": 1.3638, "step": 15456 }, { "epoch": 7.442599277978339, "grad_norm": 2.276754379272461, "learning_rate": 2.5556090515166103e-05, "loss": 1.2818, "step": 15462 }, { "epoch": 7.445487364620939, "grad_norm": 2.163547992706299, "learning_rate": 2.552720269619644e-05, "loss": 1.2507, "step": 15468 }, { "epoch": 7.448375451263538, "grad_norm": 2.194054126739502, "learning_rate": 2.549831487722677e-05, "loss": 1.3224, "step": 15474 }, { "epoch": 7.451263537906137, "grad_norm": 1.9761089086532593, "learning_rate": 2.54694270582571e-05, "loss": 1.3325, "step": 15480 }, { "epoch": 7.454151624548737, "grad_norm": 1.8696566820144653, "learning_rate": 2.5440539239287436e-05, "loss": 1.3439, "step": 15486 }, { "epoch": 7.457039711191336, "grad_norm": 2.3281335830688477, "learning_rate": 2.5411651420317767e-05, "loss": 1.3318, "step": 15492 }, { "epoch": 7.459927797833935, "grad_norm": 2.3628556728363037, "learning_rate": 2.5382763601348096e-05, "loss": 1.2404, "step": 15498 }, { "epoch": 7.462815884476535, "grad_norm": 2.215751886367798, "learning_rate": 2.5353875782378434e-05, "loss": 1.1838, "step": 15504 }, { "epoch": 7.465703971119134, "grad_norm": 2.060454845428467, "learning_rate": 2.5324987963408762e-05, "loss": 1.3695, "step": 15510 }, { "epoch": 7.468592057761732, "grad_norm": 2.123988628387451, "learning_rate": 2.5296100144439094e-05, "loss": 1.3107, "step": 15516 }, { "epoch": 7.471480144404332, "grad_norm": 2.330810546875, "learning_rate": 2.526721232546943e-05, "loss": 1.3545, "step": 15522 }, { "epoch": 7.474368231046931, "grad_norm": 2.525790214538574, "learning_rate": 2.523832450649976e-05, "loss": 1.2855, "step": 15528 }, { "epoch": 7.47725631768953, "grad_norm": 2.218888521194458, "learning_rate": 2.5209436687530092e-05, "loss": 1.2346, "step": 15534 }, { "epoch": 7.48014440433213, "grad_norm": 2.0760385990142822, "learning_rate": 2.5180548868560427e-05, "loss": 1.3186, "step": 15540 }, { "epoch": 7.483032490974729, "grad_norm": 2.115798234939575, "learning_rate": 2.515166104959076e-05, "loss": 1.1811, "step": 15546 }, { "epoch": 7.485920577617328, "grad_norm": 2.450348377227783, "learning_rate": 2.5122773230621087e-05, "loss": 1.2819, "step": 15552 }, { "epoch": 7.488808664259928, "grad_norm": 2.43626070022583, "learning_rate": 2.509388541165142e-05, "loss": 1.2627, "step": 15558 }, { "epoch": 7.491696750902527, "grad_norm": 2.107074499130249, "learning_rate": 2.5064997592681753e-05, "loss": 1.3235, "step": 15564 }, { "epoch": 7.494584837545126, "grad_norm": 2.344689130783081, "learning_rate": 2.5036109773712085e-05, "loss": 1.2635, "step": 15570 }, { "epoch": 7.497472924187726, "grad_norm": 2.2031452655792236, "learning_rate": 2.5007221954742417e-05, "loss": 1.2766, "step": 15576 }, { "epoch": 7.500361010830325, "grad_norm": 2.1732189655303955, "learning_rate": 2.4978334135772748e-05, "loss": 1.3086, "step": 15582 }, { "epoch": 7.503249097472924, "grad_norm": 2.0421383380889893, "learning_rate": 2.4949446316803083e-05, "loss": 1.3997, "step": 15588 }, { "epoch": 7.5061371841155236, "grad_norm": 2.419647693634033, "learning_rate": 2.4920558497833415e-05, "loss": 1.3815, "step": 15594 }, { "epoch": 7.509025270758123, "grad_norm": 2.365783929824829, "learning_rate": 2.4891670678863746e-05, "loss": 1.1775, "step": 15600 }, { "epoch": 7.511913357400722, "grad_norm": 2.1844465732574463, "learning_rate": 2.4862782859894078e-05, "loss": 1.2811, "step": 15606 }, { "epoch": 7.5148014440433215, "grad_norm": 2.2472074031829834, "learning_rate": 2.4833895040924413e-05, "loss": 1.2786, "step": 15612 }, { "epoch": 7.517689530685921, "grad_norm": 1.9421433210372925, "learning_rate": 2.4805007221954745e-05, "loss": 1.3926, "step": 15618 }, { "epoch": 7.52057761732852, "grad_norm": 1.8379909992218018, "learning_rate": 2.4776119402985076e-05, "loss": 1.2464, "step": 15624 }, { "epoch": 7.5234657039711195, "grad_norm": 2.046856164932251, "learning_rate": 2.4747231584015408e-05, "loss": 1.3559, "step": 15630 }, { "epoch": 7.526353790613719, "grad_norm": 2.136666774749756, "learning_rate": 2.471834376504574e-05, "loss": 1.1779, "step": 15636 }, { "epoch": 7.529241877256318, "grad_norm": 2.172560930252075, "learning_rate": 2.4689455946076074e-05, "loss": 1.3734, "step": 15642 }, { "epoch": 7.532129963898917, "grad_norm": 2.22676944732666, "learning_rate": 2.4660568127106403e-05, "loss": 1.3636, "step": 15648 }, { "epoch": 7.535018050541517, "grad_norm": 2.2356252670288086, "learning_rate": 2.4631680308136738e-05, "loss": 1.3715, "step": 15654 }, { "epoch": 7.537906137184115, "grad_norm": 2.270413637161255, "learning_rate": 2.460279248916707e-05, "loss": 1.2597, "step": 15660 }, { "epoch": 7.5407942238267145, "grad_norm": 2.3443989753723145, "learning_rate": 2.45739046701974e-05, "loss": 1.2092, "step": 15666 }, { "epoch": 7.543682310469314, "grad_norm": 2.1875009536743164, "learning_rate": 2.4545016851227732e-05, "loss": 1.4733, "step": 15672 }, { "epoch": 7.546570397111913, "grad_norm": 2.1493067741394043, "learning_rate": 2.4516129032258064e-05, "loss": 1.297, "step": 15678 }, { "epoch": 7.5494584837545125, "grad_norm": 2.1837875843048096, "learning_rate": 2.44872412132884e-05, "loss": 1.2341, "step": 15684 }, { "epoch": 7.552346570397112, "grad_norm": 2.251175880432129, "learning_rate": 2.445835339431873e-05, "loss": 1.3765, "step": 15690 }, { "epoch": 7.555234657039711, "grad_norm": 2.441986560821533, "learning_rate": 2.4429465575349062e-05, "loss": 1.1737, "step": 15696 }, { "epoch": 7.55812274368231, "grad_norm": 2.1945173740386963, "learning_rate": 2.4400577756379394e-05, "loss": 1.2024, "step": 15702 }, { "epoch": 7.56101083032491, "grad_norm": 2.1951329708099365, "learning_rate": 2.437168993740973e-05, "loss": 1.3401, "step": 15708 }, { "epoch": 7.563898916967509, "grad_norm": 2.546407699584961, "learning_rate": 2.4342802118440057e-05, "loss": 1.3875, "step": 15714 }, { "epoch": 7.566787003610108, "grad_norm": 2.2147834300994873, "learning_rate": 2.4313914299470392e-05, "loss": 1.2098, "step": 15720 }, { "epoch": 7.569675090252708, "grad_norm": 2.4005002975463867, "learning_rate": 2.4285026480500723e-05, "loss": 1.1926, "step": 15726 }, { "epoch": 7.572563176895307, "grad_norm": 2.0462913513183594, "learning_rate": 2.4256138661531055e-05, "loss": 1.239, "step": 15732 }, { "epoch": 7.575451263537906, "grad_norm": 1.9525500535964966, "learning_rate": 2.422725084256139e-05, "loss": 1.3496, "step": 15738 }, { "epoch": 7.578339350180506, "grad_norm": 2.169572353363037, "learning_rate": 2.4198363023591718e-05, "loss": 1.1806, "step": 15744 }, { "epoch": 7.581227436823105, "grad_norm": 1.8244518041610718, "learning_rate": 2.4169475204622053e-05, "loss": 1.2999, "step": 15750 }, { "epoch": 7.584115523465704, "grad_norm": 2.338318347930908, "learning_rate": 2.4140587385652385e-05, "loss": 1.3395, "step": 15756 }, { "epoch": 7.587003610108304, "grad_norm": 1.9967395067214966, "learning_rate": 2.4111699566682716e-05, "loss": 1.3191, "step": 15762 }, { "epoch": 7.589891696750902, "grad_norm": 2.0763633251190186, "learning_rate": 2.4082811747713048e-05, "loss": 1.3108, "step": 15768 }, { "epoch": 7.592779783393501, "grad_norm": 1.897753357887268, "learning_rate": 2.405392392874338e-05, "loss": 1.4415, "step": 15774 }, { "epoch": 7.595667870036101, "grad_norm": 2.3038172721862793, "learning_rate": 2.4025036109773715e-05, "loss": 1.2775, "step": 15780 }, { "epoch": 7.5985559566787, "grad_norm": 2.3705053329467773, "learning_rate": 2.3996148290804046e-05, "loss": 1.3493, "step": 15786 }, { "epoch": 7.601444043321299, "grad_norm": 2.4415407180786133, "learning_rate": 2.3967260471834378e-05, "loss": 1.2544, "step": 15792 }, { "epoch": 7.604332129963899, "grad_norm": 2.0733916759490967, "learning_rate": 2.393837265286471e-05, "loss": 1.2509, "step": 15798 }, { "epoch": 7.607220216606498, "grad_norm": 2.2098817825317383, "learning_rate": 2.3909484833895044e-05, "loss": 1.2673, "step": 15804 }, { "epoch": 7.610108303249097, "grad_norm": 2.112529754638672, "learning_rate": 2.3880597014925373e-05, "loss": 1.2983, "step": 15810 }, { "epoch": 7.612996389891697, "grad_norm": 2.100017786026001, "learning_rate": 2.3851709195955708e-05, "loss": 1.2783, "step": 15816 }, { "epoch": 7.615884476534296, "grad_norm": 2.44110107421875, "learning_rate": 2.382282137698604e-05, "loss": 1.4142, "step": 15822 }, { "epoch": 7.618772563176895, "grad_norm": 2.0864884853363037, "learning_rate": 2.379393355801637e-05, "loss": 1.3575, "step": 15828 }, { "epoch": 7.6216606498194945, "grad_norm": 1.9816993474960327, "learning_rate": 2.3765045739046706e-05, "loss": 1.3836, "step": 15834 }, { "epoch": 7.624548736462094, "grad_norm": 2.47248911857605, "learning_rate": 2.3736157920077034e-05, "loss": 1.2608, "step": 15840 }, { "epoch": 7.627436823104693, "grad_norm": 1.9025063514709473, "learning_rate": 2.370727010110737e-05, "loss": 1.1618, "step": 15846 }, { "epoch": 7.6303249097472925, "grad_norm": 2.3164050579071045, "learning_rate": 2.36783822821377e-05, "loss": 1.1918, "step": 15852 }, { "epoch": 7.633212996389892, "grad_norm": 2.2230441570281982, "learning_rate": 2.3649494463168032e-05, "loss": 1.2957, "step": 15858 }, { "epoch": 7.636101083032491, "grad_norm": 2.2598307132720947, "learning_rate": 2.3620606644198364e-05, "loss": 1.131, "step": 15864 }, { "epoch": 7.6389891696750905, "grad_norm": 2.1465461254119873, "learning_rate": 2.3591718825228695e-05, "loss": 1.2635, "step": 15870 }, { "epoch": 7.64187725631769, "grad_norm": 1.8952943086624146, "learning_rate": 2.3562831006259027e-05, "loss": 1.2564, "step": 15876 }, { "epoch": 7.644765342960289, "grad_norm": 2.119030475616455, "learning_rate": 2.3533943187289362e-05, "loss": 1.2215, "step": 15882 }, { "epoch": 7.647653429602888, "grad_norm": 2.214770555496216, "learning_rate": 2.3505055368319693e-05, "loss": 1.3819, "step": 15888 }, { "epoch": 7.650541516245488, "grad_norm": 2.0744829177856445, "learning_rate": 2.3476167549350025e-05, "loss": 1.3014, "step": 15894 }, { "epoch": 7.653429602888087, "grad_norm": 2.2595269680023193, "learning_rate": 2.344727973038036e-05, "loss": 1.2156, "step": 15900 }, { "epoch": 7.656317689530686, "grad_norm": 2.287863254547119, "learning_rate": 2.3418391911410688e-05, "loss": 1.3724, "step": 15906 }, { "epoch": 7.659205776173286, "grad_norm": 2.2104246616363525, "learning_rate": 2.3389504092441023e-05, "loss": 1.2346, "step": 15912 }, { "epoch": 7.662093862815884, "grad_norm": 2.4807827472686768, "learning_rate": 2.336061627347135e-05, "loss": 1.3007, "step": 15918 }, { "epoch": 7.664981949458483, "grad_norm": 2.278895378112793, "learning_rate": 2.3331728454501686e-05, "loss": 1.3355, "step": 15924 }, { "epoch": 7.667870036101083, "grad_norm": 2.234372615814209, "learning_rate": 2.3302840635532018e-05, "loss": 1.2544, "step": 15930 }, { "epoch": 7.670758122743682, "grad_norm": 2.052826404571533, "learning_rate": 2.327395281656235e-05, "loss": 1.3949, "step": 15936 }, { "epoch": 7.673646209386281, "grad_norm": 1.9531551599502563, "learning_rate": 2.3245064997592685e-05, "loss": 1.2471, "step": 15942 }, { "epoch": 7.676534296028881, "grad_norm": 2.0562710762023926, "learning_rate": 2.3216177178623016e-05, "loss": 1.2146, "step": 15948 }, { "epoch": 7.67942238267148, "grad_norm": 2.192960500717163, "learning_rate": 2.3187289359653348e-05, "loss": 1.3115, "step": 15954 }, { "epoch": 7.682310469314079, "grad_norm": 2.1669225692749023, "learning_rate": 2.315840154068368e-05, "loss": 1.3585, "step": 15960 }, { "epoch": 7.685198555956679, "grad_norm": 1.9748342037200928, "learning_rate": 2.312951372171401e-05, "loss": 1.2605, "step": 15966 }, { "epoch": 7.688086642599278, "grad_norm": 1.8606023788452148, "learning_rate": 2.3100625902744343e-05, "loss": 1.3438, "step": 15972 }, { "epoch": 7.690974729241877, "grad_norm": 2.2940800189971924, "learning_rate": 2.3071738083774678e-05, "loss": 1.3865, "step": 15978 }, { "epoch": 7.693862815884477, "grad_norm": 2.0730743408203125, "learning_rate": 2.304285026480501e-05, "loss": 1.2666, "step": 15984 }, { "epoch": 7.696750902527076, "grad_norm": 2.118253469467163, "learning_rate": 2.301396244583534e-05, "loss": 1.1733, "step": 15990 }, { "epoch": 7.699638989169675, "grad_norm": 2.2693004608154297, "learning_rate": 2.2985074626865672e-05, "loss": 1.3097, "step": 15996 }, { "epoch": 7.702527075812275, "grad_norm": 2.030654191970825, "learning_rate": 2.2956186807896004e-05, "loss": 1.1816, "step": 16002 }, { "epoch": 7.705415162454874, "grad_norm": 2.3118362426757812, "learning_rate": 2.292729898892634e-05, "loss": 1.2739, "step": 16008 }, { "epoch": 7.708303249097473, "grad_norm": 1.9208492040634155, "learning_rate": 2.2898411169956667e-05, "loss": 1.1821, "step": 16014 }, { "epoch": 7.7111913357400725, "grad_norm": 2.0396127700805664, "learning_rate": 2.2869523350987002e-05, "loss": 1.2552, "step": 16020 }, { "epoch": 7.714079422382672, "grad_norm": 2.1712639331817627, "learning_rate": 2.2840635532017334e-05, "loss": 1.3156, "step": 16026 }, { "epoch": 7.71696750902527, "grad_norm": 2.2475435733795166, "learning_rate": 2.2811747713047665e-05, "loss": 1.4739, "step": 16032 }, { "epoch": 7.71985559566787, "grad_norm": 2.171123743057251, "learning_rate": 2.2782859894077997e-05, "loss": 1.3501, "step": 16038 }, { "epoch": 7.722743682310469, "grad_norm": 2.1378371715545654, "learning_rate": 2.275397207510833e-05, "loss": 1.3207, "step": 16044 }, { "epoch": 7.725631768953068, "grad_norm": 2.3567566871643066, "learning_rate": 2.2725084256138663e-05, "loss": 1.3739, "step": 16050 }, { "epoch": 7.728519855595668, "grad_norm": 2.44980525970459, "learning_rate": 2.2696196437168995e-05, "loss": 1.3297, "step": 16056 }, { "epoch": 7.731407942238267, "grad_norm": 2.353928804397583, "learning_rate": 2.2667308618199327e-05, "loss": 1.3842, "step": 16062 }, { "epoch": 7.734296028880866, "grad_norm": 2.527003765106201, "learning_rate": 2.2638420799229658e-05, "loss": 1.2404, "step": 16068 }, { "epoch": 7.7371841155234655, "grad_norm": 2.4323041439056396, "learning_rate": 2.2609532980259993e-05, "loss": 1.3556, "step": 16074 }, { "epoch": 7.740072202166065, "grad_norm": 1.9534416198730469, "learning_rate": 2.258064516129032e-05, "loss": 1.2974, "step": 16080 }, { "epoch": 7.742960288808664, "grad_norm": 2.319335460662842, "learning_rate": 2.2551757342320656e-05, "loss": 1.3315, "step": 16086 }, { "epoch": 7.7458483754512635, "grad_norm": 1.9822558164596558, "learning_rate": 2.2522869523350988e-05, "loss": 1.2499, "step": 16092 }, { "epoch": 7.748736462093863, "grad_norm": 2.3497676849365234, "learning_rate": 2.249398170438132e-05, "loss": 1.3806, "step": 16098 }, { "epoch": 7.751624548736462, "grad_norm": 1.9077353477478027, "learning_rate": 2.2465093885411655e-05, "loss": 1.2693, "step": 16104 }, { "epoch": 7.754512635379061, "grad_norm": 2.0622823238372803, "learning_rate": 2.2436206066441983e-05, "loss": 1.3246, "step": 16110 }, { "epoch": 7.757400722021661, "grad_norm": 1.926512598991394, "learning_rate": 2.2407318247472318e-05, "loss": 1.1968, "step": 16116 }, { "epoch": 7.76028880866426, "grad_norm": 2.166417360305786, "learning_rate": 2.237843042850265e-05, "loss": 1.379, "step": 16122 }, { "epoch": 7.763176895306859, "grad_norm": 2.130384683609009, "learning_rate": 2.234954260953298e-05, "loss": 1.3433, "step": 16128 }, { "epoch": 7.766064981949459, "grad_norm": 2.3311727046966553, "learning_rate": 2.2320654790563313e-05, "loss": 1.2313, "step": 16134 }, { "epoch": 7.768953068592058, "grad_norm": 2.3167972564697266, "learning_rate": 2.2291766971593644e-05, "loss": 1.236, "step": 16140 }, { "epoch": 7.771841155234657, "grad_norm": 2.083965301513672, "learning_rate": 2.226287915262398e-05, "loss": 1.2455, "step": 16146 }, { "epoch": 7.774729241877257, "grad_norm": 2.0586471557617188, "learning_rate": 2.223399133365431e-05, "loss": 1.2408, "step": 16152 }, { "epoch": 7.777617328519856, "grad_norm": 2.188206195831299, "learning_rate": 2.2205103514684642e-05, "loss": 1.4237, "step": 16158 }, { "epoch": 7.780505415162455, "grad_norm": 2.235292911529541, "learning_rate": 2.2176215695714974e-05, "loss": 1.4806, "step": 16164 }, { "epoch": 7.783393501805055, "grad_norm": 1.9259231090545654, "learning_rate": 2.214732787674531e-05, "loss": 1.3406, "step": 16170 }, { "epoch": 7.786281588447653, "grad_norm": 2.202165365219116, "learning_rate": 2.2118440057775637e-05, "loss": 1.342, "step": 16176 }, { "epoch": 7.789169675090252, "grad_norm": 2.375990629196167, "learning_rate": 2.2089552238805972e-05, "loss": 1.3303, "step": 16182 }, { "epoch": 7.792057761732852, "grad_norm": 2.4719150066375732, "learning_rate": 2.2060664419836304e-05, "loss": 1.492, "step": 16188 }, { "epoch": 7.794945848375451, "grad_norm": 2.0408966541290283, "learning_rate": 2.2031776600866635e-05, "loss": 1.3023, "step": 16194 }, { "epoch": 7.79783393501805, "grad_norm": 2.4042892456054688, "learning_rate": 2.200288878189697e-05, "loss": 1.3463, "step": 16200 }, { "epoch": 7.80072202166065, "grad_norm": 2.2976527214050293, "learning_rate": 2.19740009629273e-05, "loss": 1.2498, "step": 16206 }, { "epoch": 7.803610108303249, "grad_norm": 2.295166492462158, "learning_rate": 2.1945113143957634e-05, "loss": 1.2767, "step": 16212 }, { "epoch": 7.806498194945848, "grad_norm": 1.8925294876098633, "learning_rate": 2.1916225324987965e-05, "loss": 1.247, "step": 16218 }, { "epoch": 7.809386281588448, "grad_norm": 2.0340781211853027, "learning_rate": 2.1887337506018297e-05, "loss": 1.3864, "step": 16224 }, { "epoch": 7.812274368231047, "grad_norm": 2.1586222648620605, "learning_rate": 2.185844968704863e-05, "loss": 1.3745, "step": 16230 }, { "epoch": 7.815162454873646, "grad_norm": 2.243001699447632, "learning_rate": 2.182956186807896e-05, "loss": 1.386, "step": 16236 }, { "epoch": 7.818050541516246, "grad_norm": 2.397334337234497, "learning_rate": 2.180067404910929e-05, "loss": 1.4374, "step": 16242 }, { "epoch": 7.820938628158845, "grad_norm": 2.4068939685821533, "learning_rate": 2.1771786230139626e-05, "loss": 1.2975, "step": 16248 }, { "epoch": 7.823826714801444, "grad_norm": 2.103208541870117, "learning_rate": 2.1742898411169958e-05, "loss": 1.4149, "step": 16254 }, { "epoch": 7.8267148014440435, "grad_norm": 2.4357903003692627, "learning_rate": 2.171401059220029e-05, "loss": 1.3357, "step": 16260 }, { "epoch": 7.829602888086643, "grad_norm": 1.9648973941802979, "learning_rate": 2.1685122773230625e-05, "loss": 1.272, "step": 16266 }, { "epoch": 7.832490974729242, "grad_norm": 2.2688581943511963, "learning_rate": 2.1656234954260953e-05, "loss": 1.3403, "step": 16272 }, { "epoch": 7.8353790613718415, "grad_norm": 2.1761794090270996, "learning_rate": 2.1627347135291288e-05, "loss": 1.2463, "step": 16278 }, { "epoch": 7.838267148014441, "grad_norm": 2.2070188522338867, "learning_rate": 2.1598459316321616e-05, "loss": 1.3232, "step": 16284 }, { "epoch": 7.841155234657039, "grad_norm": 2.521026849746704, "learning_rate": 2.156957149735195e-05, "loss": 1.3464, "step": 16290 }, { "epoch": 7.8440433212996385, "grad_norm": 2.164336681365967, "learning_rate": 2.1540683678382283e-05, "loss": 1.4074, "step": 16296 }, { "epoch": 7.846931407942238, "grad_norm": 2.223737955093384, "learning_rate": 2.1511795859412614e-05, "loss": 1.3089, "step": 16302 }, { "epoch": 7.849819494584837, "grad_norm": 2.2604787349700928, "learning_rate": 2.148290804044295e-05, "loss": 1.3591, "step": 16308 }, { "epoch": 7.8527075812274365, "grad_norm": 2.3126556873321533, "learning_rate": 2.145402022147328e-05, "loss": 1.2578, "step": 16314 }, { "epoch": 7.855595667870036, "grad_norm": 2.5499024391174316, "learning_rate": 2.1425132402503612e-05, "loss": 1.3405, "step": 16320 }, { "epoch": 7.858483754512635, "grad_norm": 2.315791606903076, "learning_rate": 2.1396244583533944e-05, "loss": 1.3278, "step": 16326 }, { "epoch": 7.8613718411552345, "grad_norm": 2.1160778999328613, "learning_rate": 2.1367356764564276e-05, "loss": 1.249, "step": 16332 }, { "epoch": 7.864259927797834, "grad_norm": 2.5632481575012207, "learning_rate": 2.1338468945594607e-05, "loss": 1.3378, "step": 16338 }, { "epoch": 7.867148014440433, "grad_norm": 1.9587711095809937, "learning_rate": 2.1309581126624942e-05, "loss": 1.2298, "step": 16344 }, { "epoch": 7.870036101083032, "grad_norm": 2.283735990524292, "learning_rate": 2.1280693307655274e-05, "loss": 1.2515, "step": 16350 }, { "epoch": 7.872924187725632, "grad_norm": 2.4281904697418213, "learning_rate": 2.1251805488685605e-05, "loss": 1.3551, "step": 16356 }, { "epoch": 7.875812274368231, "grad_norm": 2.2724380493164062, "learning_rate": 2.122291766971594e-05, "loss": 1.2991, "step": 16362 }, { "epoch": 7.87870036101083, "grad_norm": 2.122723340988159, "learning_rate": 2.119402985074627e-05, "loss": 1.3254, "step": 16368 }, { "epoch": 7.88158844765343, "grad_norm": 2.341156482696533, "learning_rate": 2.1165142031776604e-05, "loss": 1.3489, "step": 16374 }, { "epoch": 7.884476534296029, "grad_norm": 2.3297278881073, "learning_rate": 2.1136254212806932e-05, "loss": 1.3176, "step": 16380 }, { "epoch": 7.887364620938628, "grad_norm": 2.1843454837799072, "learning_rate": 2.1107366393837267e-05, "loss": 1.2248, "step": 16386 }, { "epoch": 7.890252707581228, "grad_norm": 2.317122220993042, "learning_rate": 2.10784785748676e-05, "loss": 1.3177, "step": 16392 }, { "epoch": 7.893140794223827, "grad_norm": 2.240217447280884, "learning_rate": 2.104959075589793e-05, "loss": 1.3877, "step": 16398 }, { "epoch": 7.896028880866426, "grad_norm": 1.932973027229309, "learning_rate": 2.102070293692826e-05, "loss": 1.2084, "step": 16404 }, { "epoch": 7.898916967509026, "grad_norm": 2.116739273071289, "learning_rate": 2.0991815117958597e-05, "loss": 1.4071, "step": 16410 }, { "epoch": 7.901805054151625, "grad_norm": 2.2456881999969482, "learning_rate": 2.0962927298988928e-05, "loss": 1.4123, "step": 16416 }, { "epoch": 7.904693140794224, "grad_norm": 2.4015941619873047, "learning_rate": 2.093403948001926e-05, "loss": 1.2512, "step": 16422 }, { "epoch": 7.907581227436824, "grad_norm": 1.8345890045166016, "learning_rate": 2.090515166104959e-05, "loss": 1.2217, "step": 16428 }, { "epoch": 7.910469314079422, "grad_norm": 2.0231847763061523, "learning_rate": 2.0876263842079923e-05, "loss": 1.234, "step": 16434 }, { "epoch": 7.913357400722021, "grad_norm": 2.3847780227661133, "learning_rate": 2.0847376023110258e-05, "loss": 1.2205, "step": 16440 }, { "epoch": 7.916245487364621, "grad_norm": 2.4380481243133545, "learning_rate": 2.0818488204140586e-05, "loss": 1.2812, "step": 16446 }, { "epoch": 7.91913357400722, "grad_norm": 2.1847198009490967, "learning_rate": 2.078960038517092e-05, "loss": 1.3718, "step": 16452 }, { "epoch": 7.922021660649819, "grad_norm": 1.8653346300125122, "learning_rate": 2.0760712566201253e-05, "loss": 1.3163, "step": 16458 }, { "epoch": 7.924909747292419, "grad_norm": 2.2642664909362793, "learning_rate": 2.0731824747231584e-05, "loss": 1.3803, "step": 16464 }, { "epoch": 7.927797833935018, "grad_norm": 2.1146976947784424, "learning_rate": 2.070293692826192e-05, "loss": 1.2607, "step": 16470 }, { "epoch": 7.930685920577617, "grad_norm": 2.4533612728118896, "learning_rate": 2.0674049109292247e-05, "loss": 1.3034, "step": 16476 }, { "epoch": 7.9335740072202166, "grad_norm": 2.4898979663848877, "learning_rate": 2.0645161290322582e-05, "loss": 1.2945, "step": 16482 }, { "epoch": 7.936462093862816, "grad_norm": 2.1610875129699707, "learning_rate": 2.0616273471352914e-05, "loss": 1.2708, "step": 16488 }, { "epoch": 7.939350180505415, "grad_norm": 2.1210849285125732, "learning_rate": 2.0587385652383246e-05, "loss": 1.2912, "step": 16494 }, { "epoch": 7.9422382671480145, "grad_norm": 1.8365246057510376, "learning_rate": 2.0558497833413577e-05, "loss": 1.2479, "step": 16500 }, { "epoch": 7.945126353790614, "grad_norm": 2.2168519496917725, "learning_rate": 2.0529610014443912e-05, "loss": 1.4159, "step": 16506 }, { "epoch": 7.948014440433213, "grad_norm": 2.4118411540985107, "learning_rate": 2.0500722195474244e-05, "loss": 1.2747, "step": 16512 }, { "epoch": 7.9509025270758125, "grad_norm": 2.0903749465942383, "learning_rate": 2.0471834376504575e-05, "loss": 1.2367, "step": 16518 }, { "epoch": 7.953790613718412, "grad_norm": 2.249903678894043, "learning_rate": 2.0442946557534907e-05, "loss": 1.2555, "step": 16524 }, { "epoch": 7.956678700361011, "grad_norm": 2.186875104904175, "learning_rate": 2.041405873856524e-05, "loss": 1.4082, "step": 16530 }, { "epoch": 7.95956678700361, "grad_norm": 1.9269988536834717, "learning_rate": 2.0385170919595574e-05, "loss": 1.2618, "step": 16536 }, { "epoch": 7.96245487364621, "grad_norm": 2.3233482837677, "learning_rate": 2.0356283100625902e-05, "loss": 1.3687, "step": 16542 }, { "epoch": 7.965342960288808, "grad_norm": 1.9730572700500488, "learning_rate": 2.0327395281656237e-05, "loss": 1.1027, "step": 16548 }, { "epoch": 7.9682310469314075, "grad_norm": 2.241650342941284, "learning_rate": 2.029850746268657e-05, "loss": 1.3069, "step": 16554 }, { "epoch": 7.971119133574007, "grad_norm": 2.293813705444336, "learning_rate": 2.02696196437169e-05, "loss": 1.2866, "step": 16560 }, { "epoch": 7.974007220216606, "grad_norm": 2.4939346313476562, "learning_rate": 2.0240731824747235e-05, "loss": 1.297, "step": 16566 }, { "epoch": 7.9768953068592054, "grad_norm": 2.3601834774017334, "learning_rate": 2.0211844005777563e-05, "loss": 1.2761, "step": 16572 }, { "epoch": 7.979783393501805, "grad_norm": 2.249720811843872, "learning_rate": 2.0182956186807898e-05, "loss": 1.2052, "step": 16578 }, { "epoch": 7.982671480144404, "grad_norm": 2.2592575550079346, "learning_rate": 2.015406836783823e-05, "loss": 1.3232, "step": 16584 }, { "epoch": 7.985559566787003, "grad_norm": 1.9863622188568115, "learning_rate": 2.012518054886856e-05, "loss": 1.2443, "step": 16590 }, { "epoch": 7.988447653429603, "grad_norm": 2.1748459339141846, "learning_rate": 2.0096292729898893e-05, "loss": 1.231, "step": 16596 }, { "epoch": 7.991335740072202, "grad_norm": 2.2211921215057373, "learning_rate": 2.0067404910929225e-05, "loss": 1.3949, "step": 16602 }, { "epoch": 7.994223826714801, "grad_norm": 1.9200705289840698, "learning_rate": 2.0038517091959556e-05, "loss": 1.2257, "step": 16608 }, { "epoch": 7.997111913357401, "grad_norm": 2.959439992904663, "learning_rate": 2.000962927298989e-05, "loss": 1.4959, "step": 16614 }, { "epoch": 8.0, "grad_norm": 2.120746612548828, "learning_rate": 1.9980741454020223e-05, "loss": 1.1812, "step": 16620 }, { "epoch": 8.0028880866426, "grad_norm": 1.988132119178772, "learning_rate": 1.9951853635050554e-05, "loss": 1.1609, "step": 16626 }, { "epoch": 8.005776173285199, "grad_norm": 2.171072483062744, "learning_rate": 1.992296581608089e-05, "loss": 1.3039, "step": 16632 }, { "epoch": 8.008664259927798, "grad_norm": 2.1484792232513428, "learning_rate": 1.9894077997111217e-05, "loss": 1.148, "step": 16638 }, { "epoch": 8.011552346570397, "grad_norm": 2.3488759994506836, "learning_rate": 1.9865190178141552e-05, "loss": 1.3898, "step": 16644 }, { "epoch": 8.014440433212997, "grad_norm": 2.2173473834991455, "learning_rate": 1.983630235917188e-05, "loss": 1.1605, "step": 16650 }, { "epoch": 8.017328519855596, "grad_norm": 2.2034502029418945, "learning_rate": 1.9807414540202216e-05, "loss": 1.2757, "step": 16656 }, { "epoch": 8.020216606498195, "grad_norm": 2.248469352722168, "learning_rate": 1.9778526721232547e-05, "loss": 1.1598, "step": 16662 }, { "epoch": 8.023104693140795, "grad_norm": 2.5009453296661377, "learning_rate": 1.974963890226288e-05, "loss": 1.328, "step": 16668 }, { "epoch": 8.025992779783394, "grad_norm": 2.2033584117889404, "learning_rate": 1.9720751083293214e-05, "loss": 1.2769, "step": 16674 }, { "epoch": 8.028880866425993, "grad_norm": 2.0796737670898438, "learning_rate": 1.9691863264323545e-05, "loss": 1.1288, "step": 16680 }, { "epoch": 8.031768953068593, "grad_norm": 2.2955288887023926, "learning_rate": 1.9662975445353877e-05, "loss": 1.256, "step": 16686 }, { "epoch": 8.034657039711192, "grad_norm": 2.1415369510650635, "learning_rate": 1.963408762638421e-05, "loss": 1.2708, "step": 16692 }, { "epoch": 8.037545126353791, "grad_norm": 2.1770899295806885, "learning_rate": 1.960519980741454e-05, "loss": 1.3123, "step": 16698 }, { "epoch": 8.04043321299639, "grad_norm": 2.363538980484009, "learning_rate": 1.9576311988444872e-05, "loss": 1.1373, "step": 16704 }, { "epoch": 8.04332129963899, "grad_norm": 2.270289182662964, "learning_rate": 1.9547424169475207e-05, "loss": 1.3011, "step": 16710 }, { "epoch": 8.046209386281589, "grad_norm": 2.316013813018799, "learning_rate": 1.951853635050554e-05, "loss": 1.2992, "step": 16716 }, { "epoch": 8.049097472924188, "grad_norm": 2.366546869277954, "learning_rate": 1.948964853153587e-05, "loss": 1.4213, "step": 16722 }, { "epoch": 8.051985559566788, "grad_norm": 2.0727810859680176, "learning_rate": 1.9460760712566205e-05, "loss": 1.2366, "step": 16728 }, { "epoch": 8.054873646209387, "grad_norm": 2.18540096282959, "learning_rate": 1.9431872893596533e-05, "loss": 1.2684, "step": 16734 }, { "epoch": 8.057761732851986, "grad_norm": 2.131272792816162, "learning_rate": 1.9402985074626868e-05, "loss": 1.1976, "step": 16740 }, { "epoch": 8.060649819494586, "grad_norm": 2.4119958877563477, "learning_rate": 1.9374097255657196e-05, "loss": 1.2853, "step": 16746 }, { "epoch": 8.063537906137183, "grad_norm": 2.211803436279297, "learning_rate": 1.934520943668753e-05, "loss": 1.2122, "step": 16752 }, { "epoch": 8.066425992779783, "grad_norm": 2.6786718368530273, "learning_rate": 1.9316321617717863e-05, "loss": 1.4132, "step": 16758 }, { "epoch": 8.069314079422382, "grad_norm": 2.1564385890960693, "learning_rate": 1.9287433798748195e-05, "loss": 1.2724, "step": 16764 }, { "epoch": 8.072202166064981, "grad_norm": 2.204904794692993, "learning_rate": 1.9258545979778526e-05, "loss": 1.1021, "step": 16770 }, { "epoch": 8.07509025270758, "grad_norm": 2.45331072807312, "learning_rate": 1.922965816080886e-05, "loss": 1.3473, "step": 16776 }, { "epoch": 8.07797833935018, "grad_norm": 2.11112642288208, "learning_rate": 1.9200770341839193e-05, "loss": 1.2628, "step": 16782 }, { "epoch": 8.08086642599278, "grad_norm": 2.207327365875244, "learning_rate": 1.9171882522869524e-05, "loss": 1.2196, "step": 16788 }, { "epoch": 8.083754512635378, "grad_norm": 2.6712491512298584, "learning_rate": 1.9142994703899856e-05, "loss": 1.3813, "step": 16794 }, { "epoch": 8.086642599277978, "grad_norm": 2.066424608230591, "learning_rate": 1.9114106884930188e-05, "loss": 1.3301, "step": 16800 }, { "epoch": 8.089530685920577, "grad_norm": 2.372260332107544, "learning_rate": 1.9085219065960522e-05, "loss": 1.2512, "step": 16806 }, { "epoch": 8.092418772563176, "grad_norm": 2.711108922958374, "learning_rate": 1.905633124699085e-05, "loss": 1.3282, "step": 16812 }, { "epoch": 8.095306859205776, "grad_norm": 2.478761672973633, "learning_rate": 1.9027443428021186e-05, "loss": 1.403, "step": 16818 }, { "epoch": 8.098194945848375, "grad_norm": 2.3473591804504395, "learning_rate": 1.8998555609051517e-05, "loss": 1.2231, "step": 16824 }, { "epoch": 8.101083032490974, "grad_norm": 2.4410910606384277, "learning_rate": 1.896966779008185e-05, "loss": 1.2006, "step": 16830 }, { "epoch": 8.103971119133574, "grad_norm": 2.4634766578674316, "learning_rate": 1.8940779971112184e-05, "loss": 1.2607, "step": 16836 }, { "epoch": 8.106859205776173, "grad_norm": 2.0777995586395264, "learning_rate": 1.8911892152142512e-05, "loss": 1.3691, "step": 16842 }, { "epoch": 8.109747292418772, "grad_norm": 2.0884387493133545, "learning_rate": 1.8883004333172847e-05, "loss": 1.2213, "step": 16848 }, { "epoch": 8.112635379061372, "grad_norm": 2.0958762168884277, "learning_rate": 1.885411651420318e-05, "loss": 1.2888, "step": 16854 }, { "epoch": 8.115523465703971, "grad_norm": 2.4269843101501465, "learning_rate": 1.882522869523351e-05, "loss": 1.3085, "step": 16860 }, { "epoch": 8.11841155234657, "grad_norm": 2.5322747230529785, "learning_rate": 1.8796340876263842e-05, "loss": 1.2346, "step": 16866 }, { "epoch": 8.12129963898917, "grad_norm": 2.2674336433410645, "learning_rate": 1.8767453057294177e-05, "loss": 1.1784, "step": 16872 }, { "epoch": 8.124187725631769, "grad_norm": 2.4527480602264404, "learning_rate": 1.873856523832451e-05, "loss": 1.29, "step": 16878 }, { "epoch": 8.127075812274368, "grad_norm": 2.3671751022338867, "learning_rate": 1.870967741935484e-05, "loss": 1.2726, "step": 16884 }, { "epoch": 8.129963898916968, "grad_norm": 2.130805492401123, "learning_rate": 1.868078960038517e-05, "loss": 1.2646, "step": 16890 }, { "epoch": 8.132851985559567, "grad_norm": 2.6094465255737305, "learning_rate": 1.8651901781415503e-05, "loss": 1.1325, "step": 16896 }, { "epoch": 8.135740072202166, "grad_norm": 2.297860622406006, "learning_rate": 1.8623013962445838e-05, "loss": 1.2653, "step": 16902 }, { "epoch": 8.138628158844766, "grad_norm": 2.2653355598449707, "learning_rate": 1.8594126143476166e-05, "loss": 1.2186, "step": 16908 }, { "epoch": 8.141516245487365, "grad_norm": 2.123230457305908, "learning_rate": 1.85652383245065e-05, "loss": 1.242, "step": 16914 }, { "epoch": 8.144404332129964, "grad_norm": 2.172114372253418, "learning_rate": 1.8536350505536833e-05, "loss": 1.2729, "step": 16920 }, { "epoch": 8.147292418772563, "grad_norm": 2.377122402191162, "learning_rate": 1.8507462686567165e-05, "loss": 1.2464, "step": 16926 }, { "epoch": 8.150180505415163, "grad_norm": 2.143918752670288, "learning_rate": 1.84785748675975e-05, "loss": 1.1236, "step": 16932 }, { "epoch": 8.153068592057762, "grad_norm": 2.2139859199523926, "learning_rate": 1.8449687048627828e-05, "loss": 1.3103, "step": 16938 }, { "epoch": 8.155956678700361, "grad_norm": 2.4742610454559326, "learning_rate": 1.8420799229658163e-05, "loss": 1.2898, "step": 16944 }, { "epoch": 8.15884476534296, "grad_norm": 2.4500892162323, "learning_rate": 1.8391911410688494e-05, "loss": 1.2318, "step": 16950 }, { "epoch": 8.16173285198556, "grad_norm": 2.1122260093688965, "learning_rate": 1.8363023591718826e-05, "loss": 1.255, "step": 16956 }, { "epoch": 8.16462093862816, "grad_norm": 2.236144781112671, "learning_rate": 1.8334135772749158e-05, "loss": 1.2963, "step": 16962 }, { "epoch": 8.167509025270759, "grad_norm": 2.1741039752960205, "learning_rate": 1.8305247953779493e-05, "loss": 1.241, "step": 16968 }, { "epoch": 8.170397111913358, "grad_norm": 2.303325891494751, "learning_rate": 1.827636013480982e-05, "loss": 1.1864, "step": 16974 }, { "epoch": 8.173285198555957, "grad_norm": 2.6378703117370605, "learning_rate": 1.8247472315840156e-05, "loss": 1.3628, "step": 16980 }, { "epoch": 8.176173285198557, "grad_norm": 2.388031482696533, "learning_rate": 1.8218584496870487e-05, "loss": 1.2503, "step": 16986 }, { "epoch": 8.179061371841156, "grad_norm": 2.207179307937622, "learning_rate": 1.818969667790082e-05, "loss": 1.1969, "step": 16992 }, { "epoch": 8.181949458483755, "grad_norm": 2.154195547103882, "learning_rate": 1.8160808858931154e-05, "loss": 1.3397, "step": 16998 }, { "epoch": 8.184837545126355, "grad_norm": 2.130171060562134, "learning_rate": 1.8131921039961482e-05, "loss": 1.2894, "step": 17004 }, { "epoch": 8.187725631768952, "grad_norm": 2.609776735305786, "learning_rate": 1.8103033220991817e-05, "loss": 1.224, "step": 17010 }, { "epoch": 8.190613718411552, "grad_norm": 2.525320291519165, "learning_rate": 1.807414540202215e-05, "loss": 1.2534, "step": 17016 }, { "epoch": 8.19350180505415, "grad_norm": 2.2384321689605713, "learning_rate": 1.804525758305248e-05, "loss": 1.2748, "step": 17022 }, { "epoch": 8.19638989169675, "grad_norm": 2.0817294120788574, "learning_rate": 1.8016369764082812e-05, "loss": 1.2498, "step": 17028 }, { "epoch": 8.19927797833935, "grad_norm": 2.1168110370635986, "learning_rate": 1.7987481945113143e-05, "loss": 1.4173, "step": 17034 }, { "epoch": 8.202166064981949, "grad_norm": 2.5823731422424316, "learning_rate": 1.795859412614348e-05, "loss": 1.2147, "step": 17040 }, { "epoch": 8.205054151624548, "grad_norm": 2.2747976779937744, "learning_rate": 1.792970630717381e-05, "loss": 1.3457, "step": 17046 }, { "epoch": 8.207942238267147, "grad_norm": 2.5621488094329834, "learning_rate": 1.790081848820414e-05, "loss": 1.3016, "step": 17052 }, { "epoch": 8.210830324909747, "grad_norm": 2.567328453063965, "learning_rate": 1.7871930669234473e-05, "loss": 1.127, "step": 17058 }, { "epoch": 8.213718411552346, "grad_norm": 1.903593897819519, "learning_rate": 1.7843042850264805e-05, "loss": 1.1042, "step": 17064 }, { "epoch": 8.216606498194945, "grad_norm": 2.408254623413086, "learning_rate": 1.7814155031295136e-05, "loss": 1.2246, "step": 17070 }, { "epoch": 8.219494584837545, "grad_norm": 2.1230661869049072, "learning_rate": 1.778526721232547e-05, "loss": 1.281, "step": 17076 }, { "epoch": 8.222382671480144, "grad_norm": 2.2255263328552246, "learning_rate": 1.7756379393355803e-05, "loss": 1.3212, "step": 17082 }, { "epoch": 8.225270758122743, "grad_norm": 2.534367322921753, "learning_rate": 1.7727491574386135e-05, "loss": 1.3432, "step": 17088 }, { "epoch": 8.228158844765343, "grad_norm": 2.166191339492798, "learning_rate": 1.769860375541647e-05, "loss": 1.3121, "step": 17094 }, { "epoch": 8.231046931407942, "grad_norm": 2.0163776874542236, "learning_rate": 1.7669715936446798e-05, "loss": 1.2918, "step": 17100 }, { "epoch": 8.233935018050541, "grad_norm": 2.253397226333618, "learning_rate": 1.7640828117477133e-05, "loss": 1.422, "step": 17106 }, { "epoch": 8.23682310469314, "grad_norm": 2.276719093322754, "learning_rate": 1.761194029850746e-05, "loss": 1.3579, "step": 17112 }, { "epoch": 8.23971119133574, "grad_norm": 2.245455026626587, "learning_rate": 1.7583052479537796e-05, "loss": 1.2866, "step": 17118 }, { "epoch": 8.24259927797834, "grad_norm": 2.421773910522461, "learning_rate": 1.7554164660568128e-05, "loss": 1.3998, "step": 17124 }, { "epoch": 8.245487364620939, "grad_norm": 2.3611483573913574, "learning_rate": 1.752527684159846e-05, "loss": 1.2705, "step": 17130 }, { "epoch": 8.248375451263538, "grad_norm": 2.3981902599334717, "learning_rate": 1.749638902262879e-05, "loss": 1.1715, "step": 17136 }, { "epoch": 8.251263537906137, "grad_norm": 2.3771636486053467, "learning_rate": 1.7467501203659126e-05, "loss": 1.2222, "step": 17142 }, { "epoch": 8.254151624548737, "grad_norm": 2.2969510555267334, "learning_rate": 1.7438613384689457e-05, "loss": 1.2845, "step": 17148 }, { "epoch": 8.257039711191336, "grad_norm": 2.4261081218719482, "learning_rate": 1.740972556571979e-05, "loss": 1.276, "step": 17154 }, { "epoch": 8.259927797833935, "grad_norm": 2.1763529777526855, "learning_rate": 1.738083774675012e-05, "loss": 1.1341, "step": 17160 }, { "epoch": 8.262815884476534, "grad_norm": 2.4904167652130127, "learning_rate": 1.7351949927780452e-05, "loss": 1.2704, "step": 17166 }, { "epoch": 8.265703971119134, "grad_norm": 2.148839235305786, "learning_rate": 1.7323062108810787e-05, "loss": 1.3151, "step": 17172 }, { "epoch": 8.268592057761733, "grad_norm": 2.406240463256836, "learning_rate": 1.7294174289841115e-05, "loss": 1.2702, "step": 17178 }, { "epoch": 8.271480144404332, "grad_norm": 2.2829699516296387, "learning_rate": 1.726528647087145e-05, "loss": 1.2528, "step": 17184 }, { "epoch": 8.274368231046932, "grad_norm": 2.0860342979431152, "learning_rate": 1.7236398651901782e-05, "loss": 1.2648, "step": 17190 }, { "epoch": 8.277256317689531, "grad_norm": 2.3210463523864746, "learning_rate": 1.7207510832932113e-05, "loss": 1.2914, "step": 17196 }, { "epoch": 8.28014440433213, "grad_norm": 2.1169650554656982, "learning_rate": 1.717862301396245e-05, "loss": 1.113, "step": 17202 }, { "epoch": 8.28303249097473, "grad_norm": 2.2125871181488037, "learning_rate": 1.7149735194992777e-05, "loss": 1.2936, "step": 17208 }, { "epoch": 8.285920577617329, "grad_norm": 2.2111034393310547, "learning_rate": 1.712084737602311e-05, "loss": 1.2662, "step": 17214 }, { "epoch": 8.288808664259928, "grad_norm": 2.231214761734009, "learning_rate": 1.7091959557053443e-05, "loss": 1.3312, "step": 17220 }, { "epoch": 8.291696750902528, "grad_norm": 2.1983861923217773, "learning_rate": 1.7063071738083775e-05, "loss": 1.2868, "step": 17226 }, { "epoch": 8.294584837545127, "grad_norm": 2.284274101257324, "learning_rate": 1.7034183919114106e-05, "loss": 1.2405, "step": 17232 }, { "epoch": 8.297472924187726, "grad_norm": 2.414581537246704, "learning_rate": 1.700529610014444e-05, "loss": 1.2527, "step": 17238 }, { "epoch": 8.300361010830326, "grad_norm": 1.95023512840271, "learning_rate": 1.6976408281174773e-05, "loss": 1.2732, "step": 17244 }, { "epoch": 8.303249097472925, "grad_norm": 2.495514154434204, "learning_rate": 1.6947520462205105e-05, "loss": 1.4328, "step": 17250 }, { "epoch": 8.306137184115524, "grad_norm": 2.2045085430145264, "learning_rate": 1.6918632643235436e-05, "loss": 1.2695, "step": 17256 }, { "epoch": 8.309025270758124, "grad_norm": 2.746220588684082, "learning_rate": 1.6889744824265768e-05, "loss": 1.3488, "step": 17262 }, { "epoch": 8.311913357400723, "grad_norm": 2.3221893310546875, "learning_rate": 1.6860857005296103e-05, "loss": 1.3656, "step": 17268 }, { "epoch": 8.31480144404332, "grad_norm": 2.0858867168426514, "learning_rate": 1.683196918632643e-05, "loss": 1.3104, "step": 17274 }, { "epoch": 8.31768953068592, "grad_norm": 2.353863477706909, "learning_rate": 1.6803081367356766e-05, "loss": 1.2053, "step": 17280 }, { "epoch": 8.320577617328519, "grad_norm": 2.11897349357605, "learning_rate": 1.6774193548387098e-05, "loss": 1.3375, "step": 17286 }, { "epoch": 8.323465703971118, "grad_norm": 2.036802291870117, "learning_rate": 1.674530572941743e-05, "loss": 1.1849, "step": 17292 }, { "epoch": 8.326353790613718, "grad_norm": 2.544525384902954, "learning_rate": 1.6716417910447764e-05, "loss": 1.3988, "step": 17298 }, { "epoch": 8.329241877256317, "grad_norm": 2.266232490539551, "learning_rate": 1.6687530091478092e-05, "loss": 1.2461, "step": 17304 }, { "epoch": 8.332129963898916, "grad_norm": 2.111287832260132, "learning_rate": 1.6658642272508427e-05, "loss": 1.0941, "step": 17310 }, { "epoch": 8.335018050541516, "grad_norm": 2.4048843383789062, "learning_rate": 1.662975445353876e-05, "loss": 1.1809, "step": 17316 }, { "epoch": 8.337906137184115, "grad_norm": 1.9914991855621338, "learning_rate": 1.660086663456909e-05, "loss": 1.2197, "step": 17322 }, { "epoch": 8.340794223826714, "grad_norm": 2.3968265056610107, "learning_rate": 1.6571978815599422e-05, "loss": 1.2374, "step": 17328 }, { "epoch": 8.343682310469314, "grad_norm": 2.3766844272613525, "learning_rate": 1.6543090996629757e-05, "loss": 1.0829, "step": 17334 }, { "epoch": 8.346570397111913, "grad_norm": 2.2232887744903564, "learning_rate": 1.6514203177660085e-05, "loss": 1.223, "step": 17340 }, { "epoch": 8.349458483754512, "grad_norm": 2.310328245162964, "learning_rate": 1.648531535869042e-05, "loss": 1.2608, "step": 17346 }, { "epoch": 8.352346570397112, "grad_norm": 2.3563790321350098, "learning_rate": 1.6456427539720752e-05, "loss": 1.1935, "step": 17352 }, { "epoch": 8.355234657039711, "grad_norm": 2.3894340991973877, "learning_rate": 1.6427539720751084e-05, "loss": 1.1878, "step": 17358 }, { "epoch": 8.35812274368231, "grad_norm": 2.2936031818389893, "learning_rate": 1.639865190178142e-05, "loss": 1.1293, "step": 17364 }, { "epoch": 8.36101083032491, "grad_norm": 2.3730051517486572, "learning_rate": 1.6369764082811747e-05, "loss": 1.2688, "step": 17370 }, { "epoch": 8.363898916967509, "grad_norm": 2.157003164291382, "learning_rate": 1.6340876263842082e-05, "loss": 1.1687, "step": 17376 }, { "epoch": 8.366787003610108, "grad_norm": 2.203092098236084, "learning_rate": 1.6311988444872413e-05, "loss": 1.2793, "step": 17382 }, { "epoch": 8.369675090252708, "grad_norm": 2.4802985191345215, "learning_rate": 1.6283100625902745e-05, "loss": 1.2731, "step": 17388 }, { "epoch": 8.372563176895307, "grad_norm": 2.398092269897461, "learning_rate": 1.6254212806933076e-05, "loss": 1.2905, "step": 17394 }, { "epoch": 8.375451263537906, "grad_norm": 2.3800745010375977, "learning_rate": 1.6225324987963408e-05, "loss": 1.3073, "step": 17400 }, { "epoch": 8.378339350180505, "grad_norm": 2.025825262069702, "learning_rate": 1.6196437168993743e-05, "loss": 1.2893, "step": 17406 }, { "epoch": 8.381227436823105, "grad_norm": 2.394500494003296, "learning_rate": 1.6167549350024075e-05, "loss": 1.2533, "step": 17412 }, { "epoch": 8.384115523465704, "grad_norm": 2.4356296062469482, "learning_rate": 1.6138661531054406e-05, "loss": 1.371, "step": 17418 }, { "epoch": 8.387003610108303, "grad_norm": 2.3099329471588135, "learning_rate": 1.6109773712084738e-05, "loss": 1.2613, "step": 17424 }, { "epoch": 8.389891696750903, "grad_norm": 2.2050700187683105, "learning_rate": 1.6080885893115073e-05, "loss": 1.2681, "step": 17430 }, { "epoch": 8.392779783393502, "grad_norm": 2.278611660003662, "learning_rate": 1.60519980741454e-05, "loss": 1.2909, "step": 17436 }, { "epoch": 8.395667870036101, "grad_norm": 2.3865113258361816, "learning_rate": 1.6023110255175736e-05, "loss": 1.2619, "step": 17442 }, { "epoch": 8.3985559566787, "grad_norm": 2.5550620555877686, "learning_rate": 1.5994222436206068e-05, "loss": 1.2956, "step": 17448 }, { "epoch": 8.4014440433213, "grad_norm": 2.35084867477417, "learning_rate": 1.59653346172364e-05, "loss": 1.1275, "step": 17454 }, { "epoch": 8.4043321299639, "grad_norm": 2.3950724601745605, "learning_rate": 1.5936446798266734e-05, "loss": 1.324, "step": 17460 }, { "epoch": 8.407220216606499, "grad_norm": 2.308840751647949, "learning_rate": 1.5907558979297062e-05, "loss": 1.269, "step": 17466 }, { "epoch": 8.410108303249098, "grad_norm": 2.020993232727051, "learning_rate": 1.5878671160327397e-05, "loss": 1.2197, "step": 17472 }, { "epoch": 8.412996389891697, "grad_norm": 2.2119052410125732, "learning_rate": 1.584978334135773e-05, "loss": 1.312, "step": 17478 }, { "epoch": 8.415884476534297, "grad_norm": 2.2268126010894775, "learning_rate": 1.582089552238806e-05, "loss": 1.2613, "step": 17484 }, { "epoch": 8.418772563176896, "grad_norm": 2.221973180770874, "learning_rate": 1.5792007703418392e-05, "loss": 1.1341, "step": 17490 }, { "epoch": 8.421660649819495, "grad_norm": 2.129371166229248, "learning_rate": 1.5763119884448724e-05, "loss": 1.3255, "step": 17496 }, { "epoch": 8.424548736462095, "grad_norm": 2.120737314224243, "learning_rate": 1.5734232065479055e-05, "loss": 1.256, "step": 17502 }, { "epoch": 8.427436823104694, "grad_norm": 2.3367464542388916, "learning_rate": 1.570534424650939e-05, "loss": 1.3222, "step": 17508 }, { "epoch": 8.430324909747293, "grad_norm": 2.4470677375793457, "learning_rate": 1.5676456427539722e-05, "loss": 1.2591, "step": 17514 }, { "epoch": 8.433212996389893, "grad_norm": 2.701282501220703, "learning_rate": 1.5647568608570054e-05, "loss": 1.3445, "step": 17520 }, { "epoch": 8.43610108303249, "grad_norm": 2.3028478622436523, "learning_rate": 1.561868078960039e-05, "loss": 1.4258, "step": 17526 }, { "epoch": 8.43898916967509, "grad_norm": 2.6721808910369873, "learning_rate": 1.5589792970630717e-05, "loss": 1.2543, "step": 17532 }, { "epoch": 8.441877256317689, "grad_norm": 2.5805273056030273, "learning_rate": 1.5560905151661052e-05, "loss": 1.2922, "step": 17538 }, { "epoch": 8.444765342960288, "grad_norm": 2.4856879711151123, "learning_rate": 1.553201733269138e-05, "loss": 1.209, "step": 17544 }, { "epoch": 8.447653429602887, "grad_norm": 2.628204584121704, "learning_rate": 1.5503129513721715e-05, "loss": 1.36, "step": 17550 }, { "epoch": 8.450541516245487, "grad_norm": 2.131279706954956, "learning_rate": 1.5474241694752047e-05, "loss": 1.3696, "step": 17556 }, { "epoch": 8.453429602888086, "grad_norm": 2.3374860286712646, "learning_rate": 1.5445353875782378e-05, "loss": 1.3339, "step": 17562 }, { "epoch": 8.456317689530685, "grad_norm": 2.1739320755004883, "learning_rate": 1.5416466056812713e-05, "loss": 1.3101, "step": 17568 }, { "epoch": 8.459205776173285, "grad_norm": 2.405245780944824, "learning_rate": 1.5387578237843045e-05, "loss": 1.3154, "step": 17574 }, { "epoch": 8.462093862815884, "grad_norm": 2.2778072357177734, "learning_rate": 1.5358690418873376e-05, "loss": 1.2817, "step": 17580 }, { "epoch": 8.464981949458483, "grad_norm": 2.1930065155029297, "learning_rate": 1.5329802599903708e-05, "loss": 1.2649, "step": 17586 }, { "epoch": 8.467870036101083, "grad_norm": 2.13317608833313, "learning_rate": 1.530091478093404e-05, "loss": 1.1289, "step": 17592 }, { "epoch": 8.470758122743682, "grad_norm": 2.2151834964752197, "learning_rate": 1.527202696196437e-05, "loss": 1.2932, "step": 17598 }, { "epoch": 8.473646209386281, "grad_norm": 2.348048448562622, "learning_rate": 1.5243139142994706e-05, "loss": 1.1951, "step": 17604 }, { "epoch": 8.47653429602888, "grad_norm": 2.026750087738037, "learning_rate": 1.5214251324025036e-05, "loss": 1.2442, "step": 17610 }, { "epoch": 8.47942238267148, "grad_norm": 2.391514301300049, "learning_rate": 1.518536350505537e-05, "loss": 1.2756, "step": 17616 }, { "epoch": 8.48231046931408, "grad_norm": 2.2480287551879883, "learning_rate": 1.51564756860857e-05, "loss": 1.2273, "step": 17622 }, { "epoch": 8.485198555956678, "grad_norm": 2.3557629585266113, "learning_rate": 1.5127587867116032e-05, "loss": 1.1096, "step": 17628 }, { "epoch": 8.488086642599278, "grad_norm": 2.4257090091705322, "learning_rate": 1.5098700048146366e-05, "loss": 1.2636, "step": 17634 }, { "epoch": 8.490974729241877, "grad_norm": 2.2749366760253906, "learning_rate": 1.5069812229176697e-05, "loss": 1.1466, "step": 17640 }, { "epoch": 8.493862815884476, "grad_norm": 2.304753065109253, "learning_rate": 1.504092441020703e-05, "loss": 1.3301, "step": 17646 }, { "epoch": 8.496750902527076, "grad_norm": 2.760322332382202, "learning_rate": 1.5012036591237364e-05, "loss": 1.2303, "step": 17652 }, { "epoch": 8.499638989169675, "grad_norm": 2.03890323638916, "learning_rate": 1.4983148772267694e-05, "loss": 1.2691, "step": 17658 }, { "epoch": 8.502527075812274, "grad_norm": 2.330463171005249, "learning_rate": 1.4954260953298027e-05, "loss": 1.1421, "step": 17664 }, { "epoch": 8.505415162454874, "grad_norm": 2.1985552310943604, "learning_rate": 1.4925373134328357e-05, "loss": 1.3091, "step": 17670 }, { "epoch": 8.508303249097473, "grad_norm": 2.1054491996765137, "learning_rate": 1.489648531535869e-05, "loss": 1.2742, "step": 17676 }, { "epoch": 8.511191335740072, "grad_norm": 2.192638874053955, "learning_rate": 1.4867597496389024e-05, "loss": 1.4513, "step": 17682 }, { "epoch": 8.514079422382672, "grad_norm": 2.386117458343506, "learning_rate": 1.4838709677419355e-05, "loss": 1.2242, "step": 17688 }, { "epoch": 8.516967509025271, "grad_norm": 2.301132917404175, "learning_rate": 1.4809821858449688e-05, "loss": 1.2362, "step": 17694 }, { "epoch": 8.51985559566787, "grad_norm": 2.5709292888641357, "learning_rate": 1.4780934039480022e-05, "loss": 1.4222, "step": 17700 }, { "epoch": 8.52274368231047, "grad_norm": 2.29669451713562, "learning_rate": 1.4752046220510352e-05, "loss": 1.1632, "step": 17706 }, { "epoch": 8.525631768953069, "grad_norm": 2.3466362953186035, "learning_rate": 1.4723158401540685e-05, "loss": 1.2778, "step": 17712 }, { "epoch": 8.528519855595668, "grad_norm": 2.1661696434020996, "learning_rate": 1.4694270582571015e-05, "loss": 1.2564, "step": 17718 }, { "epoch": 8.531407942238268, "grad_norm": 2.1351566314697266, "learning_rate": 1.4665382763601348e-05, "loss": 1.2306, "step": 17724 }, { "epoch": 8.534296028880867, "grad_norm": 1.991471767425537, "learning_rate": 1.4636494944631681e-05, "loss": 1.2461, "step": 17730 }, { "epoch": 8.537184115523466, "grad_norm": 2.018778085708618, "learning_rate": 1.4607607125662013e-05, "loss": 1.2116, "step": 17736 }, { "epoch": 8.540072202166066, "grad_norm": 2.251307725906372, "learning_rate": 1.4578719306692346e-05, "loss": 1.2558, "step": 17742 }, { "epoch": 8.542960288808665, "grad_norm": 2.222395896911621, "learning_rate": 1.454983148772268e-05, "loss": 1.2548, "step": 17748 }, { "epoch": 8.545848375451264, "grad_norm": 2.24300479888916, "learning_rate": 1.452094366875301e-05, "loss": 1.1839, "step": 17754 }, { "epoch": 8.548736462093864, "grad_norm": 2.212580919265747, "learning_rate": 1.4492055849783343e-05, "loss": 1.1896, "step": 17760 }, { "epoch": 8.551624548736463, "grad_norm": 2.499786853790283, "learning_rate": 1.4463168030813673e-05, "loss": 1.3114, "step": 17766 }, { "epoch": 8.554512635379062, "grad_norm": 2.1249430179595947, "learning_rate": 1.4434280211844006e-05, "loss": 1.2706, "step": 17772 }, { "epoch": 8.557400722021661, "grad_norm": 2.2962262630462646, "learning_rate": 1.440539239287434e-05, "loss": 1.2703, "step": 17778 }, { "epoch": 8.56028880866426, "grad_norm": 2.2279868125915527, "learning_rate": 1.437650457390467e-05, "loss": 1.2306, "step": 17784 }, { "epoch": 8.56317689530686, "grad_norm": 2.161259889602661, "learning_rate": 1.4347616754935002e-05, "loss": 1.3125, "step": 17790 }, { "epoch": 8.56606498194946, "grad_norm": 2.3654041290283203, "learning_rate": 1.4318728935965336e-05, "loss": 1.3942, "step": 17796 }, { "epoch": 8.568953068592057, "grad_norm": 2.3930516242980957, "learning_rate": 1.4289841116995667e-05, "loss": 1.2896, "step": 17802 }, { "epoch": 8.571841155234656, "grad_norm": 1.858183741569519, "learning_rate": 1.4260953298026e-05, "loss": 1.0676, "step": 17808 }, { "epoch": 8.574729241877256, "grad_norm": 2.210345983505249, "learning_rate": 1.423206547905633e-05, "loss": 1.1811, "step": 17814 }, { "epoch": 8.577617328519855, "grad_norm": 2.233337879180908, "learning_rate": 1.4203177660086664e-05, "loss": 1.2054, "step": 17820 }, { "epoch": 8.580505415162454, "grad_norm": 2.1268064975738525, "learning_rate": 1.4174289841116997e-05, "loss": 1.2619, "step": 17826 }, { "epoch": 8.583393501805054, "grad_norm": 2.21042799949646, "learning_rate": 1.4145402022147327e-05, "loss": 1.2871, "step": 17832 }, { "epoch": 8.586281588447653, "grad_norm": 2.4033517837524414, "learning_rate": 1.411651420317766e-05, "loss": 1.2691, "step": 17838 }, { "epoch": 8.589169675090252, "grad_norm": 2.493499517440796, "learning_rate": 1.4087626384207994e-05, "loss": 1.3392, "step": 17844 }, { "epoch": 8.592057761732852, "grad_norm": 2.213696002960205, "learning_rate": 1.4058738565238325e-05, "loss": 1.2165, "step": 17850 }, { "epoch": 8.59494584837545, "grad_norm": 2.264890432357788, "learning_rate": 1.4029850746268658e-05, "loss": 1.2817, "step": 17856 }, { "epoch": 8.59783393501805, "grad_norm": 2.7451865673065186, "learning_rate": 1.4000962927298988e-05, "loss": 1.2082, "step": 17862 }, { "epoch": 8.60072202166065, "grad_norm": 2.2848448753356934, "learning_rate": 1.3972075108329322e-05, "loss": 1.2534, "step": 17868 }, { "epoch": 8.603610108303249, "grad_norm": 2.466820478439331, "learning_rate": 1.3943187289359655e-05, "loss": 1.2637, "step": 17874 }, { "epoch": 8.606498194945848, "grad_norm": 2.500640392303467, "learning_rate": 1.3914299470389985e-05, "loss": 1.2286, "step": 17880 }, { "epoch": 8.609386281588447, "grad_norm": 2.3085384368896484, "learning_rate": 1.3885411651420318e-05, "loss": 1.2998, "step": 17886 }, { "epoch": 8.612274368231047, "grad_norm": 2.303452491760254, "learning_rate": 1.3856523832450651e-05, "loss": 1.2753, "step": 17892 }, { "epoch": 8.615162454873646, "grad_norm": 2.467526912689209, "learning_rate": 1.3827636013480983e-05, "loss": 1.2452, "step": 17898 }, { "epoch": 8.618050541516245, "grad_norm": 2.1237597465515137, "learning_rate": 1.3798748194511316e-05, "loss": 1.2552, "step": 17904 }, { "epoch": 8.620938628158845, "grad_norm": 2.326815605163574, "learning_rate": 1.3769860375541646e-05, "loss": 1.193, "step": 17910 }, { "epoch": 8.623826714801444, "grad_norm": 2.6419262886047363, "learning_rate": 1.374097255657198e-05, "loss": 1.4011, "step": 17916 }, { "epoch": 8.626714801444043, "grad_norm": 2.4191136360168457, "learning_rate": 1.3712084737602313e-05, "loss": 1.2028, "step": 17922 }, { "epoch": 8.629602888086643, "grad_norm": 2.124406337738037, "learning_rate": 1.3683196918632643e-05, "loss": 1.2453, "step": 17928 }, { "epoch": 8.632490974729242, "grad_norm": 2.7630136013031006, "learning_rate": 1.3654309099662976e-05, "loss": 1.3387, "step": 17934 }, { "epoch": 8.635379061371841, "grad_norm": 2.578667163848877, "learning_rate": 1.362542128069331e-05, "loss": 1.2398, "step": 17940 }, { "epoch": 8.63826714801444, "grad_norm": 2.016751289367676, "learning_rate": 1.3596533461723641e-05, "loss": 1.2243, "step": 17946 }, { "epoch": 8.64115523465704, "grad_norm": 2.352961540222168, "learning_rate": 1.3567645642753972e-05, "loss": 1.3052, "step": 17952 }, { "epoch": 8.64404332129964, "grad_norm": 2.430095672607422, "learning_rate": 1.3538757823784304e-05, "loss": 1.2241, "step": 17958 }, { "epoch": 8.646931407942239, "grad_norm": 2.5071909427642822, "learning_rate": 1.3509870004814637e-05, "loss": 1.2961, "step": 17964 }, { "epoch": 8.649819494584838, "grad_norm": 2.2662789821624756, "learning_rate": 1.348098218584497e-05, "loss": 1.3365, "step": 17970 }, { "epoch": 8.652707581227437, "grad_norm": 2.04911470413208, "learning_rate": 1.34520943668753e-05, "loss": 1.2096, "step": 17976 }, { "epoch": 8.655595667870037, "grad_norm": 2.543955087661743, "learning_rate": 1.3423206547905634e-05, "loss": 1.2682, "step": 17982 }, { "epoch": 8.658483754512636, "grad_norm": 2.413886547088623, "learning_rate": 1.3394318728935967e-05, "loss": 1.2559, "step": 17988 }, { "epoch": 8.661371841155235, "grad_norm": 2.550696849822998, "learning_rate": 1.3365430909966297e-05, "loss": 1.3128, "step": 17994 }, { "epoch": 8.664259927797834, "grad_norm": 2.3017125129699707, "learning_rate": 1.333654309099663e-05, "loss": 1.366, "step": 18000 }, { "epoch": 8.667148014440434, "grad_norm": 2.19319486618042, "learning_rate": 1.3307655272026962e-05, "loss": 1.3143, "step": 18006 }, { "epoch": 8.670036101083033, "grad_norm": 2.4145078659057617, "learning_rate": 1.3278767453057295e-05, "loss": 1.2275, "step": 18012 }, { "epoch": 8.672924187725632, "grad_norm": 2.5456371307373047, "learning_rate": 1.3249879634087629e-05, "loss": 1.4188, "step": 18018 }, { "epoch": 8.675812274368232, "grad_norm": 2.3706424236297607, "learning_rate": 1.3220991815117958e-05, "loss": 1.1982, "step": 18024 }, { "epoch": 8.678700361010831, "grad_norm": 2.320688009262085, "learning_rate": 1.3192103996148292e-05, "loss": 1.2633, "step": 18030 }, { "epoch": 8.68158844765343, "grad_norm": 2.41300368309021, "learning_rate": 1.3163216177178625e-05, "loss": 1.228, "step": 18036 }, { "epoch": 8.684476534296028, "grad_norm": 2.73764967918396, "learning_rate": 1.3134328358208955e-05, "loss": 1.1853, "step": 18042 }, { "epoch": 8.687364620938627, "grad_norm": 2.4584436416625977, "learning_rate": 1.3105440539239288e-05, "loss": 1.3114, "step": 18048 }, { "epoch": 8.690252707581227, "grad_norm": 2.294552803039551, "learning_rate": 1.307655272026962e-05, "loss": 1.2345, "step": 18054 }, { "epoch": 8.693140794223826, "grad_norm": 2.203077793121338, "learning_rate": 1.3047664901299953e-05, "loss": 1.2168, "step": 18060 }, { "epoch": 8.696028880866425, "grad_norm": 3.0376579761505127, "learning_rate": 1.3018777082330286e-05, "loss": 1.3537, "step": 18066 }, { "epoch": 8.698916967509025, "grad_norm": 2.1634063720703125, "learning_rate": 1.2989889263360616e-05, "loss": 1.163, "step": 18072 }, { "epoch": 8.701805054151624, "grad_norm": 2.4643237590789795, "learning_rate": 1.296100144439095e-05, "loss": 1.2528, "step": 18078 }, { "epoch": 8.704693140794223, "grad_norm": 2.3502261638641357, "learning_rate": 1.2932113625421283e-05, "loss": 1.3329, "step": 18084 }, { "epoch": 8.707581227436823, "grad_norm": 2.4813005924224854, "learning_rate": 1.2903225806451613e-05, "loss": 1.24, "step": 18090 }, { "epoch": 8.710469314079422, "grad_norm": 2.2660582065582275, "learning_rate": 1.2874337987481946e-05, "loss": 1.258, "step": 18096 }, { "epoch": 8.713357400722021, "grad_norm": 2.5988929271698, "learning_rate": 1.2845450168512278e-05, "loss": 1.2589, "step": 18102 }, { "epoch": 8.71624548736462, "grad_norm": 2.450751781463623, "learning_rate": 1.2816562349542611e-05, "loss": 1.3144, "step": 18108 }, { "epoch": 8.71913357400722, "grad_norm": 1.9000325202941895, "learning_rate": 1.2787674530572944e-05, "loss": 1.2914, "step": 18114 }, { "epoch": 8.722021660649819, "grad_norm": 2.470663547515869, "learning_rate": 1.2758786711603274e-05, "loss": 1.2587, "step": 18120 }, { "epoch": 8.724909747292418, "grad_norm": 2.261643409729004, "learning_rate": 1.2729898892633607e-05, "loss": 1.1797, "step": 18126 }, { "epoch": 8.727797833935018, "grad_norm": 2.584998846054077, "learning_rate": 1.2701011073663937e-05, "loss": 1.2501, "step": 18132 }, { "epoch": 8.730685920577617, "grad_norm": 2.45704984664917, "learning_rate": 1.267212325469427e-05, "loss": 1.2108, "step": 18138 }, { "epoch": 8.733574007220216, "grad_norm": 2.7792510986328125, "learning_rate": 1.2643235435724604e-05, "loss": 1.2287, "step": 18144 }, { "epoch": 8.736462093862816, "grad_norm": 1.9823635816574097, "learning_rate": 1.2614347616754934e-05, "loss": 1.3288, "step": 18150 }, { "epoch": 8.739350180505415, "grad_norm": 2.5700464248657227, "learning_rate": 1.2585459797785267e-05, "loss": 1.2915, "step": 18156 }, { "epoch": 8.742238267148014, "grad_norm": 2.414431571960449, "learning_rate": 1.25565719788156e-05, "loss": 1.217, "step": 18162 }, { "epoch": 8.745126353790614, "grad_norm": 2.1502439975738525, "learning_rate": 1.2527684159845932e-05, "loss": 1.2745, "step": 18168 }, { "epoch": 8.748014440433213, "grad_norm": 2.067995309829712, "learning_rate": 1.2498796340876265e-05, "loss": 1.1061, "step": 18174 }, { "epoch": 8.750902527075812, "grad_norm": 2.3790807723999023, "learning_rate": 1.2469908521906597e-05, "loss": 1.2044, "step": 18180 }, { "epoch": 8.753790613718412, "grad_norm": 2.572530746459961, "learning_rate": 1.2441020702936928e-05, "loss": 1.2665, "step": 18186 }, { "epoch": 8.756678700361011, "grad_norm": 2.5196290016174316, "learning_rate": 1.241213288396726e-05, "loss": 1.2, "step": 18192 }, { "epoch": 8.75956678700361, "grad_norm": 2.1105809211730957, "learning_rate": 1.2383245064997593e-05, "loss": 1.2445, "step": 18198 }, { "epoch": 8.76245487364621, "grad_norm": 2.057403802871704, "learning_rate": 1.2354357246027925e-05, "loss": 1.216, "step": 18204 }, { "epoch": 8.765342960288809, "grad_norm": 2.366905450820923, "learning_rate": 1.2325469427058258e-05, "loss": 1.2085, "step": 18210 }, { "epoch": 8.768231046931408, "grad_norm": 2.42722225189209, "learning_rate": 1.229658160808859e-05, "loss": 1.4205, "step": 18216 }, { "epoch": 8.771119133574008, "grad_norm": 2.1595468521118164, "learning_rate": 1.2267693789118923e-05, "loss": 1.2274, "step": 18222 }, { "epoch": 8.774007220216607, "grad_norm": 2.4783761501312256, "learning_rate": 1.2238805970149255e-05, "loss": 1.387, "step": 18228 }, { "epoch": 8.776895306859206, "grad_norm": 2.4428951740264893, "learning_rate": 1.2209918151179586e-05, "loss": 1.2652, "step": 18234 }, { "epoch": 8.779783393501805, "grad_norm": 2.3700435161590576, "learning_rate": 1.2181030332209918e-05, "loss": 1.4046, "step": 18240 }, { "epoch": 8.782671480144405, "grad_norm": 2.0197060108184814, "learning_rate": 1.2152142513240251e-05, "loss": 1.1329, "step": 18246 }, { "epoch": 8.785559566787004, "grad_norm": 2.057305097579956, "learning_rate": 1.2123254694270583e-05, "loss": 1.2474, "step": 18252 }, { "epoch": 8.788447653429603, "grad_norm": 2.625067949295044, "learning_rate": 1.2094366875300914e-05, "loss": 1.2005, "step": 18258 }, { "epoch": 8.791335740072203, "grad_norm": 2.0948245525360107, "learning_rate": 1.2065479056331248e-05, "loss": 1.2556, "step": 18264 }, { "epoch": 8.794223826714802, "grad_norm": 2.6602585315704346, "learning_rate": 1.2036591237361581e-05, "loss": 1.2975, "step": 18270 }, { "epoch": 8.797111913357401, "grad_norm": 2.17158579826355, "learning_rate": 1.2007703418391913e-05, "loss": 1.2918, "step": 18276 }, { "epoch": 8.8, "grad_norm": 2.0830094814300537, "learning_rate": 1.1978815599422244e-05, "loss": 1.2504, "step": 18282 }, { "epoch": 8.8028880866426, "grad_norm": 2.2799277305603027, "learning_rate": 1.1949927780452576e-05, "loss": 1.1772, "step": 18288 }, { "epoch": 8.8057761732852, "grad_norm": 2.0804333686828613, "learning_rate": 1.1921039961482909e-05, "loss": 1.2008, "step": 18294 }, { "epoch": 8.808664259927799, "grad_norm": 2.667393445968628, "learning_rate": 1.189215214251324e-05, "loss": 1.3473, "step": 18300 }, { "epoch": 8.811552346570398, "grad_norm": 2.790308713912964, "learning_rate": 1.1863264323543572e-05, "loss": 1.2445, "step": 18306 }, { "epoch": 8.814440433212997, "grad_norm": 2.193434000015259, "learning_rate": 1.1834376504573906e-05, "loss": 1.2088, "step": 18312 }, { "epoch": 8.817328519855595, "grad_norm": 2.1138932704925537, "learning_rate": 1.1805488685604237e-05, "loss": 1.3489, "step": 18318 }, { "epoch": 8.820216606498194, "grad_norm": 2.269848108291626, "learning_rate": 1.177660086663457e-05, "loss": 1.3868, "step": 18324 }, { "epoch": 8.823104693140793, "grad_norm": 2.1187031269073486, "learning_rate": 1.1747713047664902e-05, "loss": 1.2639, "step": 18330 }, { "epoch": 8.825992779783393, "grad_norm": 2.3019652366638184, "learning_rate": 1.1718825228695234e-05, "loss": 1.2975, "step": 18336 }, { "epoch": 8.828880866425992, "grad_norm": 2.399311065673828, "learning_rate": 1.1689937409725567e-05, "loss": 1.3889, "step": 18342 }, { "epoch": 8.831768953068591, "grad_norm": 2.372480869293213, "learning_rate": 1.1661049590755898e-05, "loss": 1.2904, "step": 18348 }, { "epoch": 8.83465703971119, "grad_norm": 2.18190336227417, "learning_rate": 1.163216177178623e-05, "loss": 1.2442, "step": 18354 }, { "epoch": 8.83754512635379, "grad_norm": 1.957169771194458, "learning_rate": 1.1603273952816562e-05, "loss": 1.2804, "step": 18360 }, { "epoch": 8.84043321299639, "grad_norm": 2.157301187515259, "learning_rate": 1.1574386133846895e-05, "loss": 1.3628, "step": 18366 }, { "epoch": 8.843321299638989, "grad_norm": 2.4432151317596436, "learning_rate": 1.1545498314877228e-05, "loss": 1.3248, "step": 18372 }, { "epoch": 8.846209386281588, "grad_norm": 2.5118889808654785, "learning_rate": 1.151661049590756e-05, "loss": 1.2824, "step": 18378 }, { "epoch": 8.849097472924187, "grad_norm": 2.317892074584961, "learning_rate": 1.1487722676937891e-05, "loss": 1.2653, "step": 18384 }, { "epoch": 8.851985559566787, "grad_norm": 2.2736213207244873, "learning_rate": 1.1458834857968225e-05, "loss": 1.2226, "step": 18390 }, { "epoch": 8.854873646209386, "grad_norm": 2.4439244270324707, "learning_rate": 1.1429947038998556e-05, "loss": 1.3379, "step": 18396 }, { "epoch": 8.857761732851985, "grad_norm": 2.3487303256988525, "learning_rate": 1.1401059220028888e-05, "loss": 1.1986, "step": 18402 }, { "epoch": 8.860649819494585, "grad_norm": 2.3188843727111816, "learning_rate": 1.137217140105922e-05, "loss": 1.3077, "step": 18408 }, { "epoch": 8.863537906137184, "grad_norm": 2.075413227081299, "learning_rate": 1.1343283582089553e-05, "loss": 1.1356, "step": 18414 }, { "epoch": 8.866425992779783, "grad_norm": 2.3944144248962402, "learning_rate": 1.1314395763119884e-05, "loss": 1.3585, "step": 18420 }, { "epoch": 8.869314079422383, "grad_norm": 2.1392505168914795, "learning_rate": 1.1285507944150218e-05, "loss": 1.1838, "step": 18426 }, { "epoch": 8.872202166064982, "grad_norm": 2.530421257019043, "learning_rate": 1.125662012518055e-05, "loss": 1.1748, "step": 18432 }, { "epoch": 8.875090252707581, "grad_norm": 2.2624752521514893, "learning_rate": 1.1227732306210883e-05, "loss": 1.2439, "step": 18438 }, { "epoch": 8.87797833935018, "grad_norm": 2.3706676959991455, "learning_rate": 1.1198844487241214e-05, "loss": 1.2369, "step": 18444 }, { "epoch": 8.88086642599278, "grad_norm": 2.620652675628662, "learning_rate": 1.1169956668271546e-05, "loss": 1.3092, "step": 18450 }, { "epoch": 8.88375451263538, "grad_norm": 2.5827722549438477, "learning_rate": 1.1141068849301877e-05, "loss": 1.3212, "step": 18456 }, { "epoch": 8.886642599277979, "grad_norm": 2.1854054927825928, "learning_rate": 1.111218103033221e-05, "loss": 1.2563, "step": 18462 }, { "epoch": 8.889530685920578, "grad_norm": 2.2618489265441895, "learning_rate": 1.1083293211362542e-05, "loss": 1.2422, "step": 18468 }, { "epoch": 8.892418772563177, "grad_norm": 2.1838231086730957, "learning_rate": 1.1054405392392876e-05, "loss": 1.3097, "step": 18474 }, { "epoch": 8.895306859205776, "grad_norm": 2.340770721435547, "learning_rate": 1.1025517573423207e-05, "loss": 1.2988, "step": 18480 }, { "epoch": 8.898194945848376, "grad_norm": 2.3193206787109375, "learning_rate": 1.099662975445354e-05, "loss": 1.3336, "step": 18486 }, { "epoch": 8.901083032490975, "grad_norm": 2.505052328109741, "learning_rate": 1.0967741935483872e-05, "loss": 1.1517, "step": 18492 }, { "epoch": 8.903971119133574, "grad_norm": 2.2726972103118896, "learning_rate": 1.0938854116514204e-05, "loss": 1.2276, "step": 18498 }, { "epoch": 8.906859205776174, "grad_norm": 1.866655945777893, "learning_rate": 1.0909966297544535e-05, "loss": 1.1217, "step": 18504 }, { "epoch": 8.909747292418773, "grad_norm": 2.383863687515259, "learning_rate": 1.0881078478574869e-05, "loss": 1.2635, "step": 18510 }, { "epoch": 8.912635379061372, "grad_norm": 2.292257070541382, "learning_rate": 1.08521906596052e-05, "loss": 1.3634, "step": 18516 }, { "epoch": 8.915523465703972, "grad_norm": 2.5861167907714844, "learning_rate": 1.0823302840635532e-05, "loss": 1.3605, "step": 18522 }, { "epoch": 8.918411552346571, "grad_norm": 2.3552374839782715, "learning_rate": 1.0794415021665865e-05, "loss": 1.1637, "step": 18528 }, { "epoch": 8.92129963898917, "grad_norm": 2.5096309185028076, "learning_rate": 1.0765527202696198e-05, "loss": 1.2542, "step": 18534 }, { "epoch": 8.92418772563177, "grad_norm": 2.089756965637207, "learning_rate": 1.073663938372653e-05, "loss": 1.1837, "step": 18540 }, { "epoch": 8.927075812274369, "grad_norm": 1.9912588596343994, "learning_rate": 1.0707751564756861e-05, "loss": 1.354, "step": 18546 }, { "epoch": 8.929963898916968, "grad_norm": 2.392772912979126, "learning_rate": 1.0678863745787193e-05, "loss": 1.2032, "step": 18552 }, { "epoch": 8.932851985559568, "grad_norm": 2.478771686553955, "learning_rate": 1.0649975926817526e-05, "loss": 1.1603, "step": 18558 }, { "epoch": 8.935740072202165, "grad_norm": 2.6634716987609863, "learning_rate": 1.0621088107847858e-05, "loss": 1.3318, "step": 18564 }, { "epoch": 8.938628158844764, "grad_norm": 2.05354642868042, "learning_rate": 1.059220028887819e-05, "loss": 1.2958, "step": 18570 }, { "epoch": 8.941516245487364, "grad_norm": 2.3808186054229736, "learning_rate": 1.0563312469908523e-05, "loss": 1.3212, "step": 18576 }, { "epoch": 8.944404332129963, "grad_norm": 2.39864444732666, "learning_rate": 1.0534424650938854e-05, "loss": 1.3567, "step": 18582 }, { "epoch": 8.947292418772562, "grad_norm": 2.1512749195098877, "learning_rate": 1.0505536831969188e-05, "loss": 1.3226, "step": 18588 }, { "epoch": 8.950180505415162, "grad_norm": 2.3179657459259033, "learning_rate": 1.047664901299952e-05, "loss": 1.292, "step": 18594 }, { "epoch": 8.953068592057761, "grad_norm": 2.172163963317871, "learning_rate": 1.0447761194029851e-05, "loss": 1.3006, "step": 18600 }, { "epoch": 8.95595667870036, "grad_norm": 2.064714193344116, "learning_rate": 1.0418873375060184e-05, "loss": 1.3457, "step": 18606 }, { "epoch": 8.95884476534296, "grad_norm": 2.309373617172241, "learning_rate": 1.0389985556090516e-05, "loss": 1.1547, "step": 18612 }, { "epoch": 8.961732851985559, "grad_norm": 2.2805683612823486, "learning_rate": 1.0361097737120847e-05, "loss": 1.2597, "step": 18618 }, { "epoch": 8.964620938628158, "grad_norm": 2.391455888748169, "learning_rate": 1.0332209918151179e-05, "loss": 1.1654, "step": 18624 }, { "epoch": 8.967509025270758, "grad_norm": 2.1680045127868652, "learning_rate": 1.0303322099181512e-05, "loss": 1.0595, "step": 18630 }, { "epoch": 8.970397111913357, "grad_norm": 2.623579978942871, "learning_rate": 1.0274434280211846e-05, "loss": 1.289, "step": 18636 }, { "epoch": 8.973285198555956, "grad_norm": 2.032925605773926, "learning_rate": 1.0245546461242177e-05, "loss": 1.2721, "step": 18642 }, { "epoch": 8.976173285198556, "grad_norm": 2.2639307975769043, "learning_rate": 1.0216658642272509e-05, "loss": 1.2487, "step": 18648 }, { "epoch": 8.979061371841155, "grad_norm": 2.4102444648742676, "learning_rate": 1.018777082330284e-05, "loss": 1.2144, "step": 18654 }, { "epoch": 8.981949458483754, "grad_norm": 2.1024041175842285, "learning_rate": 1.0158883004333174e-05, "loss": 1.2042, "step": 18660 }, { "epoch": 8.984837545126354, "grad_norm": 2.392308473587036, "learning_rate": 1.0129995185363505e-05, "loss": 1.3555, "step": 18666 }, { "epoch": 8.987725631768953, "grad_norm": 2.2143936157226562, "learning_rate": 1.0101107366393837e-05, "loss": 1.3028, "step": 18672 }, { "epoch": 8.990613718411552, "grad_norm": 2.146360397338867, "learning_rate": 1.007221954742417e-05, "loss": 1.083, "step": 18678 }, { "epoch": 8.993501805054152, "grad_norm": 2.4372308254241943, "learning_rate": 1.0043331728454502e-05, "loss": 1.3489, "step": 18684 }, { "epoch": 8.99638989169675, "grad_norm": 2.30122447013855, "learning_rate": 1.0014443909484835e-05, "loss": 1.2091, "step": 18690 }, { "epoch": 8.99927797833935, "grad_norm": 2.203500747680664, "learning_rate": 9.985556090515167e-06, "loss": 1.2261, "step": 18696 }, { "epoch": 9.00216606498195, "grad_norm": 2.0938429832458496, "learning_rate": 9.956668271545498e-06, "loss": 1.1729, "step": 18702 }, { "epoch": 9.005054151624549, "grad_norm": 2.4375829696655273, "learning_rate": 9.927780452575832e-06, "loss": 1.2984, "step": 18708 }, { "epoch": 9.007942238267148, "grad_norm": 2.4390084743499756, "learning_rate": 9.898892633606163e-06, "loss": 1.2798, "step": 18714 }, { "epoch": 9.010830324909747, "grad_norm": 2.5862436294555664, "learning_rate": 9.870004814636495e-06, "loss": 1.21, "step": 18720 }, { "epoch": 9.013718411552347, "grad_norm": 2.2012178897857666, "learning_rate": 9.841116995666826e-06, "loss": 1.1105, "step": 18726 }, { "epoch": 9.016606498194946, "grad_norm": 2.0476772785186768, "learning_rate": 9.81222917669716e-06, "loss": 1.2424, "step": 18732 }, { "epoch": 9.019494584837545, "grad_norm": 2.598489761352539, "learning_rate": 9.783341357727493e-06, "loss": 1.1982, "step": 18738 }, { "epoch": 9.022382671480145, "grad_norm": 2.651494026184082, "learning_rate": 9.754453538757824e-06, "loss": 1.2385, "step": 18744 }, { "epoch": 9.025270758122744, "grad_norm": 2.3831934928894043, "learning_rate": 9.725565719788156e-06, "loss": 1.2102, "step": 18750 }, { "epoch": 9.028158844765343, "grad_norm": 2.5753979682922363, "learning_rate": 9.69667790081849e-06, "loss": 1.2983, "step": 18756 }, { "epoch": 9.031046931407943, "grad_norm": 2.4834184646606445, "learning_rate": 9.667790081848821e-06, "loss": 1.203, "step": 18762 }, { "epoch": 9.033935018050542, "grad_norm": 2.3223676681518555, "learning_rate": 9.638902262879153e-06, "loss": 1.1585, "step": 18768 }, { "epoch": 9.036823104693141, "grad_norm": 2.319796323776245, "learning_rate": 9.610014443909484e-06, "loss": 1.2328, "step": 18774 }, { "epoch": 9.03971119133574, "grad_norm": 2.118867874145508, "learning_rate": 9.581126624939817e-06, "loss": 1.2439, "step": 18780 }, { "epoch": 9.04259927797834, "grad_norm": 2.129068613052368, "learning_rate": 9.552238805970149e-06, "loss": 1.2185, "step": 18786 }, { "epoch": 9.04548736462094, "grad_norm": 2.3292746543884277, "learning_rate": 9.523350987000482e-06, "loss": 1.2153, "step": 18792 }, { "epoch": 9.048375451263539, "grad_norm": 2.2010996341705322, "learning_rate": 9.494463168030814e-06, "loss": 1.2633, "step": 18798 }, { "epoch": 9.051263537906138, "grad_norm": 2.1080682277679443, "learning_rate": 9.465575349061147e-06, "loss": 1.1362, "step": 18804 }, { "epoch": 9.054151624548737, "grad_norm": 2.2506561279296875, "learning_rate": 9.436687530091479e-06, "loss": 1.2364, "step": 18810 }, { "epoch": 9.057039711191337, "grad_norm": 2.3619260787963867, "learning_rate": 9.40779971112181e-06, "loss": 1.2053, "step": 18816 }, { "epoch": 9.059927797833936, "grad_norm": 2.383169412612915, "learning_rate": 9.378911892152142e-06, "loss": 1.2208, "step": 18822 }, { "epoch": 9.062815884476535, "grad_norm": 2.522207021713257, "learning_rate": 9.350024073182475e-06, "loss": 1.199, "step": 18828 }, { "epoch": 9.065703971119133, "grad_norm": 1.974516749382019, "learning_rate": 9.321136254212807e-06, "loss": 1.2266, "step": 18834 }, { "epoch": 9.068592057761732, "grad_norm": 2.480916976928711, "learning_rate": 9.29224843524314e-06, "loss": 1.383, "step": 18840 }, { "epoch": 9.071480144404331, "grad_norm": 2.1840980052948, "learning_rate": 9.263360616273472e-06, "loss": 1.2099, "step": 18846 }, { "epoch": 9.07436823104693, "grad_norm": 2.32034969329834, "learning_rate": 9.234472797303805e-06, "loss": 1.2653, "step": 18852 }, { "epoch": 9.07725631768953, "grad_norm": 2.301502227783203, "learning_rate": 9.205584978334137e-06, "loss": 1.0421, "step": 18858 }, { "epoch": 9.08014440433213, "grad_norm": 2.1938793659210205, "learning_rate": 9.176697159364468e-06, "loss": 1.0918, "step": 18864 }, { "epoch": 9.083032490974729, "grad_norm": 2.462716817855835, "learning_rate": 9.1478093403948e-06, "loss": 1.2317, "step": 18870 }, { "epoch": 9.085920577617328, "grad_norm": 2.3070485591888428, "learning_rate": 9.118921521425133e-06, "loss": 1.2478, "step": 18876 }, { "epoch": 9.088808664259927, "grad_norm": 2.242806911468506, "learning_rate": 9.090033702455465e-06, "loss": 1.1931, "step": 18882 }, { "epoch": 9.091696750902527, "grad_norm": 2.1841952800750732, "learning_rate": 9.061145883485796e-06, "loss": 1.1011, "step": 18888 }, { "epoch": 9.094584837545126, "grad_norm": 2.3345274925231934, "learning_rate": 9.03225806451613e-06, "loss": 1.2616, "step": 18894 }, { "epoch": 9.097472924187725, "grad_norm": 2.521969795227051, "learning_rate": 9.003370245546463e-06, "loss": 1.3138, "step": 18900 }, { "epoch": 9.100361010830325, "grad_norm": 2.1150646209716797, "learning_rate": 8.974482426576794e-06, "loss": 1.1819, "step": 18906 }, { "epoch": 9.103249097472924, "grad_norm": 2.7031123638153076, "learning_rate": 8.945594607607126e-06, "loss": 1.3979, "step": 18912 }, { "epoch": 9.106137184115523, "grad_norm": 2.1842849254608154, "learning_rate": 8.916706788637458e-06, "loss": 1.1825, "step": 18918 }, { "epoch": 9.109025270758123, "grad_norm": 2.470853805541992, "learning_rate": 8.887818969667791e-06, "loss": 1.3493, "step": 18924 }, { "epoch": 9.111913357400722, "grad_norm": 1.817830204963684, "learning_rate": 8.858931150698123e-06, "loss": 1.0576, "step": 18930 }, { "epoch": 9.114801444043321, "grad_norm": 2.2351763248443604, "learning_rate": 8.830043331728454e-06, "loss": 1.1753, "step": 18936 }, { "epoch": 9.11768953068592, "grad_norm": 2.032766819000244, "learning_rate": 8.801155512758787e-06, "loss": 1.2103, "step": 18942 }, { "epoch": 9.12057761732852, "grad_norm": 2.360675573348999, "learning_rate": 8.772267693789119e-06, "loss": 1.1919, "step": 18948 }, { "epoch": 9.123465703971119, "grad_norm": 2.2888801097869873, "learning_rate": 8.743379874819452e-06, "loss": 1.2562, "step": 18954 }, { "epoch": 9.126353790613718, "grad_norm": 2.5365374088287354, "learning_rate": 8.714492055849784e-06, "loss": 1.1366, "step": 18960 }, { "epoch": 9.129241877256318, "grad_norm": 2.241122245788574, "learning_rate": 8.685604236880116e-06, "loss": 1.2506, "step": 18966 }, { "epoch": 9.132129963898917, "grad_norm": 2.3539435863494873, "learning_rate": 8.656716417910449e-06, "loss": 1.348, "step": 18972 }, { "epoch": 9.135018050541516, "grad_norm": 2.176654815673828, "learning_rate": 8.62782859894078e-06, "loss": 1.0906, "step": 18978 }, { "epoch": 9.137906137184116, "grad_norm": 2.243553876876831, "learning_rate": 8.598940779971112e-06, "loss": 1.2037, "step": 18984 }, { "epoch": 9.140794223826715, "grad_norm": 2.432035207748413, "learning_rate": 8.570052961001444e-06, "loss": 1.2252, "step": 18990 }, { "epoch": 9.143682310469314, "grad_norm": 2.093798875808716, "learning_rate": 8.541165142031777e-06, "loss": 1.1197, "step": 18996 }, { "epoch": 9.146570397111914, "grad_norm": 2.336071729660034, "learning_rate": 8.51227732306211e-06, "loss": 1.1986, "step": 19002 }, { "epoch": 9.149458483754513, "grad_norm": 1.9607352018356323, "learning_rate": 8.483389504092442e-06, "loss": 1.0825, "step": 19008 }, { "epoch": 9.152346570397112, "grad_norm": 2.1687610149383545, "learning_rate": 8.454501685122773e-06, "loss": 1.2321, "step": 19014 }, { "epoch": 9.155234657039712, "grad_norm": 2.3614039421081543, "learning_rate": 8.425613866153107e-06, "loss": 1.1779, "step": 19020 }, { "epoch": 9.158122743682311, "grad_norm": 2.327784299850464, "learning_rate": 8.396726047183438e-06, "loss": 1.1449, "step": 19026 }, { "epoch": 9.16101083032491, "grad_norm": 2.3062336444854736, "learning_rate": 8.36783822821377e-06, "loss": 1.2111, "step": 19032 }, { "epoch": 9.16389891696751, "grad_norm": 2.3239548206329346, "learning_rate": 8.338950409244101e-06, "loss": 1.2746, "step": 19038 }, { "epoch": 9.166787003610109, "grad_norm": 2.522998332977295, "learning_rate": 8.310062590274435e-06, "loss": 1.3188, "step": 19044 }, { "epoch": 9.169675090252708, "grad_norm": 2.138909101486206, "learning_rate": 8.281174771304766e-06, "loss": 1.2402, "step": 19050 }, { "epoch": 9.172563176895308, "grad_norm": 2.192629814147949, "learning_rate": 8.2522869523351e-06, "loss": 1.2996, "step": 19056 }, { "epoch": 9.175451263537907, "grad_norm": 2.5591354370117188, "learning_rate": 8.223399133365431e-06, "loss": 1.2934, "step": 19062 }, { "epoch": 9.178339350180506, "grad_norm": 2.4939780235290527, "learning_rate": 8.194511314395765e-06, "loss": 1.2114, "step": 19068 }, { "epoch": 9.181227436823105, "grad_norm": 2.7413809299468994, "learning_rate": 8.165623495426096e-06, "loss": 1.2314, "step": 19074 }, { "epoch": 9.184115523465705, "grad_norm": 2.552177906036377, "learning_rate": 8.136735676456428e-06, "loss": 1.2025, "step": 19080 }, { "epoch": 9.187003610108302, "grad_norm": 2.4776370525360107, "learning_rate": 8.10784785748676e-06, "loss": 1.2825, "step": 19086 }, { "epoch": 9.189891696750902, "grad_norm": 2.753856658935547, "learning_rate": 8.078960038517093e-06, "loss": 1.2069, "step": 19092 }, { "epoch": 9.192779783393501, "grad_norm": 2.3003361225128174, "learning_rate": 8.050072219547424e-06, "loss": 1.1785, "step": 19098 }, { "epoch": 9.1956678700361, "grad_norm": 2.434595823287964, "learning_rate": 8.021184400577757e-06, "loss": 1.1571, "step": 19104 }, { "epoch": 9.1985559566787, "grad_norm": 1.9893741607666016, "learning_rate": 7.992296581608089e-06, "loss": 1.1891, "step": 19110 }, { "epoch": 9.201444043321299, "grad_norm": 2.3171331882476807, "learning_rate": 7.963408762638422e-06, "loss": 1.2647, "step": 19116 }, { "epoch": 9.204332129963898, "grad_norm": 2.598099708557129, "learning_rate": 7.934520943668754e-06, "loss": 1.1773, "step": 19122 }, { "epoch": 9.207220216606498, "grad_norm": 2.2868220806121826, "learning_rate": 7.905633124699086e-06, "loss": 1.1873, "step": 19128 }, { "epoch": 9.210108303249097, "grad_norm": 2.527674674987793, "learning_rate": 7.876745305729417e-06, "loss": 1.2308, "step": 19134 }, { "epoch": 9.212996389891696, "grad_norm": 2.488891839981079, "learning_rate": 7.84785748675975e-06, "loss": 1.2123, "step": 19140 }, { "epoch": 9.215884476534296, "grad_norm": 2.1088197231292725, "learning_rate": 7.818969667790082e-06, "loss": 1.2055, "step": 19146 }, { "epoch": 9.218772563176895, "grad_norm": 2.5725884437561035, "learning_rate": 7.790081848820414e-06, "loss": 1.2431, "step": 19152 }, { "epoch": 9.221660649819494, "grad_norm": 2.211540937423706, "learning_rate": 7.761194029850747e-06, "loss": 1.2861, "step": 19158 }, { "epoch": 9.224548736462094, "grad_norm": 2.1979074478149414, "learning_rate": 7.73230621088108e-06, "loss": 1.1315, "step": 19164 }, { "epoch": 9.227436823104693, "grad_norm": 2.3574771881103516, "learning_rate": 7.703418391911412e-06, "loss": 1.2926, "step": 19170 }, { "epoch": 9.230324909747292, "grad_norm": 2.263415575027466, "learning_rate": 7.674530572941743e-06, "loss": 1.2263, "step": 19176 }, { "epoch": 9.233212996389891, "grad_norm": 2.4815309047698975, "learning_rate": 7.645642753972075e-06, "loss": 1.2512, "step": 19182 }, { "epoch": 9.23610108303249, "grad_norm": 2.6005685329437256, "learning_rate": 7.616754935002407e-06, "loss": 1.3248, "step": 19188 }, { "epoch": 9.23898916967509, "grad_norm": 1.9450905323028564, "learning_rate": 7.58786711603274e-06, "loss": 1.1772, "step": 19194 }, { "epoch": 9.24187725631769, "grad_norm": 2.4414079189300537, "learning_rate": 7.558979297063072e-06, "loss": 1.2072, "step": 19200 }, { "epoch": 9.244765342960289, "grad_norm": 2.335949659347534, "learning_rate": 7.530091478093404e-06, "loss": 1.1541, "step": 19206 }, { "epoch": 9.247653429602888, "grad_norm": 2.511228084564209, "learning_rate": 7.5012036591237355e-06, "loss": 1.2218, "step": 19212 }, { "epoch": 9.250541516245487, "grad_norm": 2.8051741123199463, "learning_rate": 7.472315840154069e-06, "loss": 1.1682, "step": 19218 }, { "epoch": 9.253429602888087, "grad_norm": 2.5831925868988037, "learning_rate": 7.443428021184401e-06, "loss": 1.2585, "step": 19224 }, { "epoch": 9.256317689530686, "grad_norm": 2.599363088607788, "learning_rate": 7.414540202214733e-06, "loss": 1.3139, "step": 19230 }, { "epoch": 9.259205776173285, "grad_norm": 2.2100281715393066, "learning_rate": 7.3856523832450645e-06, "loss": 1.1415, "step": 19236 }, { "epoch": 9.262093862815885, "grad_norm": 2.539811849594116, "learning_rate": 7.356764564275398e-06, "loss": 1.1547, "step": 19242 }, { "epoch": 9.264981949458484, "grad_norm": 2.559650182723999, "learning_rate": 7.32787674530573e-06, "loss": 1.4179, "step": 19248 }, { "epoch": 9.267870036101083, "grad_norm": 2.257063627243042, "learning_rate": 7.298988926336062e-06, "loss": 1.3225, "step": 19254 }, { "epoch": 9.270758122743683, "grad_norm": 2.3699631690979004, "learning_rate": 7.270101107366393e-06, "loss": 1.1899, "step": 19260 }, { "epoch": 9.273646209386282, "grad_norm": 2.1957170963287354, "learning_rate": 7.241213288396727e-06, "loss": 1.316, "step": 19266 }, { "epoch": 9.276534296028881, "grad_norm": 2.2408201694488525, "learning_rate": 7.212325469427058e-06, "loss": 1.2037, "step": 19272 }, { "epoch": 9.27942238267148, "grad_norm": 2.225553512573242, "learning_rate": 7.183437650457391e-06, "loss": 1.2787, "step": 19278 }, { "epoch": 9.28231046931408, "grad_norm": 2.5694825649261475, "learning_rate": 7.154549831487722e-06, "loss": 1.1188, "step": 19284 }, { "epoch": 9.28519855595668, "grad_norm": 2.226738452911377, "learning_rate": 7.125662012518056e-06, "loss": 1.2788, "step": 19290 }, { "epoch": 9.288086642599279, "grad_norm": 2.3538405895233154, "learning_rate": 7.096774193548387e-06, "loss": 1.2383, "step": 19296 }, { "epoch": 9.290974729241878, "grad_norm": 2.3546793460845947, "learning_rate": 7.06788637457872e-06, "loss": 1.1774, "step": 19302 }, { "epoch": 9.293862815884477, "grad_norm": 2.406064033508301, "learning_rate": 7.038998555609051e-06, "loss": 1.2815, "step": 19308 }, { "epoch": 9.296750902527076, "grad_norm": 2.1413636207580566, "learning_rate": 7.0101107366393845e-06, "loss": 1.1875, "step": 19314 }, { "epoch": 9.299638989169676, "grad_norm": 2.19785213470459, "learning_rate": 6.981222917669716e-06, "loss": 1.1489, "step": 19320 }, { "epoch": 9.302527075812275, "grad_norm": 2.236283540725708, "learning_rate": 6.9523350987000486e-06, "loss": 1.3078, "step": 19326 }, { "epoch": 9.305415162454874, "grad_norm": 2.316331624984741, "learning_rate": 6.92344727973038e-06, "loss": 1.2069, "step": 19332 }, { "epoch": 9.308303249097474, "grad_norm": 2.3903753757476807, "learning_rate": 6.8945594607607134e-06, "loss": 1.1561, "step": 19338 }, { "epoch": 9.311191335740073, "grad_norm": 2.7022266387939453, "learning_rate": 6.865671641791045e-06, "loss": 1.1806, "step": 19344 }, { "epoch": 9.314079422382672, "grad_norm": 2.459681749343872, "learning_rate": 6.8367838228213775e-06, "loss": 1.2518, "step": 19350 }, { "epoch": 9.31696750902527, "grad_norm": 2.4115102291107178, "learning_rate": 6.807896003851709e-06, "loss": 1.2335, "step": 19356 }, { "epoch": 9.31985559566787, "grad_norm": 2.0440561771392822, "learning_rate": 6.779008184882042e-06, "loss": 1.135, "step": 19362 }, { "epoch": 9.322743682310469, "grad_norm": 2.5518269538879395, "learning_rate": 6.750120365912374e-06, "loss": 1.3179, "step": 19368 }, { "epoch": 9.325631768953068, "grad_norm": 2.353300094604492, "learning_rate": 6.7212325469427056e-06, "loss": 1.2528, "step": 19374 }, { "epoch": 9.328519855595667, "grad_norm": 2.3808093070983887, "learning_rate": 6.692344727973038e-06, "loss": 1.1128, "step": 19380 }, { "epoch": 9.331407942238267, "grad_norm": 2.2478206157684326, "learning_rate": 6.663456909003371e-06, "loss": 1.2679, "step": 19386 }, { "epoch": 9.334296028880866, "grad_norm": 2.514036178588867, "learning_rate": 6.634569090033703e-06, "loss": 1.4101, "step": 19392 }, { "epoch": 9.337184115523465, "grad_norm": 2.4391374588012695, "learning_rate": 6.6056812710640345e-06, "loss": 1.2781, "step": 19398 }, { "epoch": 9.340072202166064, "grad_norm": 2.2296199798583984, "learning_rate": 6.576793452094367e-06, "loss": 1.2095, "step": 19404 }, { "epoch": 9.342960288808664, "grad_norm": 2.545651912689209, "learning_rate": 6.5479056331247e-06, "loss": 1.2342, "step": 19410 }, { "epoch": 9.345848375451263, "grad_norm": 2.2817232608795166, "learning_rate": 6.519017814155032e-06, "loss": 1.2203, "step": 19416 }, { "epoch": 9.348736462093862, "grad_norm": 2.2757492065429688, "learning_rate": 6.490129995185363e-06, "loss": 1.161, "step": 19422 }, { "epoch": 9.351624548736462, "grad_norm": 2.294544219970703, "learning_rate": 6.461242176215696e-06, "loss": 1.2576, "step": 19428 }, { "epoch": 9.354512635379061, "grad_norm": 2.206552267074585, "learning_rate": 6.432354357246029e-06, "loss": 1.1727, "step": 19434 }, { "epoch": 9.35740072202166, "grad_norm": 2.4329841136932373, "learning_rate": 6.403466538276361e-06, "loss": 1.1816, "step": 19440 }, { "epoch": 9.36028880866426, "grad_norm": 2.1045424938201904, "learning_rate": 6.374578719306692e-06, "loss": 1.2103, "step": 19446 }, { "epoch": 9.363176895306859, "grad_norm": 2.241314649581909, "learning_rate": 6.345690900337024e-06, "loss": 1.2671, "step": 19452 }, { "epoch": 9.366064981949458, "grad_norm": 2.329519748687744, "learning_rate": 6.316803081367357e-06, "loss": 1.2101, "step": 19458 }, { "epoch": 9.368953068592058, "grad_norm": 2.289459705352783, "learning_rate": 6.28791526239769e-06, "loss": 1.3074, "step": 19464 }, { "epoch": 9.371841155234657, "grad_norm": 2.4800870418548584, "learning_rate": 6.259027443428021e-06, "loss": 1.221, "step": 19470 }, { "epoch": 9.374729241877256, "grad_norm": 2.063279151916504, "learning_rate": 6.230139624458354e-06, "loss": 1.1588, "step": 19476 }, { "epoch": 9.377617328519856, "grad_norm": 2.440152883529663, "learning_rate": 6.201251805488686e-06, "loss": 1.1956, "step": 19482 }, { "epoch": 9.380505415162455, "grad_norm": 2.5867509841918945, "learning_rate": 6.172363986519019e-06, "loss": 1.2076, "step": 19488 }, { "epoch": 9.383393501805054, "grad_norm": 2.4591619968414307, "learning_rate": 6.14347616754935e-06, "loss": 1.1986, "step": 19494 }, { "epoch": 9.386281588447654, "grad_norm": 2.0710132122039795, "learning_rate": 6.114588348579683e-06, "loss": 1.3345, "step": 19500 }, { "epoch": 9.389169675090253, "grad_norm": 2.1490485668182373, "learning_rate": 6.085700529610014e-06, "loss": 1.2387, "step": 19506 }, { "epoch": 9.392057761732852, "grad_norm": 2.1383213996887207, "learning_rate": 6.0568127106403475e-06, "loss": 1.1552, "step": 19512 }, { "epoch": 9.394945848375452, "grad_norm": 2.439613103866577, "learning_rate": 6.027924891670679e-06, "loss": 1.2303, "step": 19518 }, { "epoch": 9.39783393501805, "grad_norm": 2.1893656253814697, "learning_rate": 5.9990370727010115e-06, "loss": 1.2211, "step": 19524 }, { "epoch": 9.40072202166065, "grad_norm": 2.3151159286499023, "learning_rate": 5.970149253731343e-06, "loss": 1.2428, "step": 19530 }, { "epoch": 9.40361010830325, "grad_norm": 2.4251465797424316, "learning_rate": 5.9412614347616764e-06, "loss": 1.3268, "step": 19536 }, { "epoch": 9.406498194945849, "grad_norm": 2.222810983657837, "learning_rate": 5.912373615792008e-06, "loss": 1.298, "step": 19542 }, { "epoch": 9.409386281588448, "grad_norm": 2.1088998317718506, "learning_rate": 5.8834857968223405e-06, "loss": 1.2684, "step": 19548 }, { "epoch": 9.412274368231047, "grad_norm": 2.0928173065185547, "learning_rate": 5.854597977852672e-06, "loss": 1.2001, "step": 19554 }, { "epoch": 9.415162454873647, "grad_norm": 2.520779609680176, "learning_rate": 5.8257101588830045e-06, "loss": 1.1817, "step": 19560 }, { "epoch": 9.418050541516246, "grad_norm": 2.101513385772705, "learning_rate": 5.796822339913337e-06, "loss": 1.2678, "step": 19566 }, { "epoch": 9.420938628158845, "grad_norm": 2.093722105026245, "learning_rate": 5.767934520943669e-06, "loss": 1.2303, "step": 19572 }, { "epoch": 9.423826714801445, "grad_norm": 2.366917371749878, "learning_rate": 5.739046701974001e-06, "loss": 1.313, "step": 19578 }, { "epoch": 9.426714801444044, "grad_norm": 2.31256103515625, "learning_rate": 5.7101588830043334e-06, "loss": 1.2776, "step": 19584 }, { "epoch": 9.429602888086643, "grad_norm": 2.154003620147705, "learning_rate": 5.681271064034666e-06, "loss": 1.2525, "step": 19590 }, { "epoch": 9.432490974729243, "grad_norm": 2.5784707069396973, "learning_rate": 5.652383245064998e-06, "loss": 1.2446, "step": 19596 }, { "epoch": 9.435379061371842, "grad_norm": 2.6449742317199707, "learning_rate": 5.62349542609533e-06, "loss": 1.3151, "step": 19602 }, { "epoch": 9.43826714801444, "grad_norm": 2.592254161834717, "learning_rate": 5.594607607125662e-06, "loss": 1.2733, "step": 19608 }, { "epoch": 9.441155234657039, "grad_norm": 2.5075817108154297, "learning_rate": 5.565719788155995e-06, "loss": 1.2551, "step": 19614 }, { "epoch": 9.444043321299638, "grad_norm": 2.5247716903686523, "learning_rate": 5.536831969186327e-06, "loss": 1.2028, "step": 19620 }, { "epoch": 9.446931407942238, "grad_norm": 2.1357192993164062, "learning_rate": 5.507944150216659e-06, "loss": 1.1714, "step": 19626 }, { "epoch": 9.449819494584837, "grad_norm": 2.196584463119507, "learning_rate": 5.479056331246991e-06, "loss": 1.0848, "step": 19632 }, { "epoch": 9.452707581227436, "grad_norm": 2.4527227878570557, "learning_rate": 5.450168512277323e-06, "loss": 1.3413, "step": 19638 }, { "epoch": 9.455595667870035, "grad_norm": 2.255542278289795, "learning_rate": 5.421280693307656e-06, "loss": 1.1527, "step": 19644 }, { "epoch": 9.458483754512635, "grad_norm": 2.334282636642456, "learning_rate": 5.392392874337988e-06, "loss": 1.4042, "step": 19650 }, { "epoch": 9.461371841155234, "grad_norm": 2.1921286582946777, "learning_rate": 5.36350505536832e-06, "loss": 1.0973, "step": 19656 }, { "epoch": 9.464259927797833, "grad_norm": 2.344468593597412, "learning_rate": 5.334617236398652e-06, "loss": 1.2175, "step": 19662 }, { "epoch": 9.467148014440433, "grad_norm": 2.628777503967285, "learning_rate": 5.305729417428985e-06, "loss": 1.4088, "step": 19668 }, { "epoch": 9.470036101083032, "grad_norm": 2.4402968883514404, "learning_rate": 5.276841598459317e-06, "loss": 1.2727, "step": 19674 }, { "epoch": 9.472924187725631, "grad_norm": 2.2141613960266113, "learning_rate": 5.247953779489649e-06, "loss": 1.2681, "step": 19680 }, { "epoch": 9.47581227436823, "grad_norm": 2.2613868713378906, "learning_rate": 5.219065960519981e-06, "loss": 1.3074, "step": 19686 }, { "epoch": 9.47870036101083, "grad_norm": 2.443639039993286, "learning_rate": 5.190178141550313e-06, "loss": 1.2533, "step": 19692 }, { "epoch": 9.48158844765343, "grad_norm": 2.311923027038574, "learning_rate": 5.161290322580646e-06, "loss": 1.1957, "step": 19698 }, { "epoch": 9.484476534296029, "grad_norm": 2.622516632080078, "learning_rate": 5.132402503610978e-06, "loss": 1.331, "step": 19704 }, { "epoch": 9.487364620938628, "grad_norm": 2.3957629203796387, "learning_rate": 5.10351468464131e-06, "loss": 1.2969, "step": 19710 }, { "epoch": 9.490252707581227, "grad_norm": 2.3772497177124023, "learning_rate": 5.074626865671642e-06, "loss": 1.3025, "step": 19716 }, { "epoch": 9.493140794223827, "grad_norm": 2.3489632606506348, "learning_rate": 5.0457390467019745e-06, "loss": 1.4028, "step": 19722 }, { "epoch": 9.496028880866426, "grad_norm": 2.45080304145813, "learning_rate": 5.016851227732306e-06, "loss": 1.2148, "step": 19728 }, { "epoch": 9.498916967509025, "grad_norm": 2.2912306785583496, "learning_rate": 4.9879634087626386e-06, "loss": 1.1665, "step": 19734 }, { "epoch": 9.501805054151625, "grad_norm": 2.171609878540039, "learning_rate": 4.95907558979297e-06, "loss": 1.2231, "step": 19740 }, { "epoch": 9.504693140794224, "grad_norm": 2.311638832092285, "learning_rate": 4.9301877708233035e-06, "loss": 1.2469, "step": 19746 }, { "epoch": 9.507581227436823, "grad_norm": 2.70458722114563, "learning_rate": 4.901299951853635e-06, "loss": 1.2975, "step": 19752 }, { "epoch": 9.510469314079423, "grad_norm": 1.958707571029663, "learning_rate": 4.8724121328839675e-06, "loss": 1.2312, "step": 19758 }, { "epoch": 9.513357400722022, "grad_norm": 2.60730242729187, "learning_rate": 4.843524313914299e-06, "loss": 1.1709, "step": 19764 }, { "epoch": 9.516245487364621, "grad_norm": 2.1283106803894043, "learning_rate": 4.8146364949446315e-06, "loss": 1.2252, "step": 19770 }, { "epoch": 9.51913357400722, "grad_norm": 2.304222345352173, "learning_rate": 4.785748675974964e-06, "loss": 1.2442, "step": 19776 }, { "epoch": 9.52202166064982, "grad_norm": 2.251542806625366, "learning_rate": 4.756860857005296e-06, "loss": 1.2059, "step": 19782 }, { "epoch": 9.52490974729242, "grad_norm": 2.4506092071533203, "learning_rate": 4.727973038035628e-06, "loss": 1.2424, "step": 19788 }, { "epoch": 9.527797833935018, "grad_norm": 2.4053351879119873, "learning_rate": 4.6990852190659605e-06, "loss": 1.2832, "step": 19794 }, { "epoch": 9.530685920577618, "grad_norm": 2.1441526412963867, "learning_rate": 4.670197400096293e-06, "loss": 1.1581, "step": 19800 }, { "epoch": 9.533574007220217, "grad_norm": 2.0671143531799316, "learning_rate": 4.641309581126625e-06, "loss": 1.152, "step": 19806 }, { "epoch": 9.536462093862816, "grad_norm": 2.3222920894622803, "learning_rate": 4.612421762156957e-06, "loss": 1.2355, "step": 19812 }, { "epoch": 9.539350180505416, "grad_norm": 2.150838613510132, "learning_rate": 4.583533943187289e-06, "loss": 1.1765, "step": 19818 }, { "epoch": 9.542238267148015, "grad_norm": 2.379537582397461, "learning_rate": 4.554646124217622e-06, "loss": 1.3396, "step": 19824 }, { "epoch": 9.545126353790614, "grad_norm": 2.533644676208496, "learning_rate": 4.525758305247954e-06, "loss": 1.3229, "step": 19830 }, { "epoch": 9.548014440433214, "grad_norm": 2.58528208732605, "learning_rate": 4.496870486278286e-06, "loss": 1.347, "step": 19836 }, { "epoch": 9.550902527075813, "grad_norm": 2.374589443206787, "learning_rate": 4.467982667308618e-06, "loss": 1.1991, "step": 19842 }, { "epoch": 9.553790613718412, "grad_norm": 2.2971291542053223, "learning_rate": 4.439094848338951e-06, "loss": 1.2671, "step": 19848 }, { "epoch": 9.556678700361012, "grad_norm": 2.4392921924591064, "learning_rate": 4.410207029369283e-06, "loss": 1.3072, "step": 19854 }, { "epoch": 9.559566787003611, "grad_norm": 2.393974542617798, "learning_rate": 4.381319210399615e-06, "loss": 1.3455, "step": 19860 }, { "epoch": 9.56245487364621, "grad_norm": 1.8708361387252808, "learning_rate": 4.352431391429947e-06, "loss": 1.1551, "step": 19866 }, { "epoch": 9.56534296028881, "grad_norm": 2.3878774642944336, "learning_rate": 4.323543572460279e-06, "loss": 1.3259, "step": 19872 }, { "epoch": 9.568231046931407, "grad_norm": 2.1509549617767334, "learning_rate": 4.294655753490612e-06, "loss": 1.1399, "step": 19878 }, { "epoch": 9.571119133574006, "grad_norm": 2.306751012802124, "learning_rate": 4.265767934520944e-06, "loss": 1.2453, "step": 19884 }, { "epoch": 9.574007220216606, "grad_norm": 2.291106939315796, "learning_rate": 4.236880115551276e-06, "loss": 1.2484, "step": 19890 }, { "epoch": 9.576895306859205, "grad_norm": 2.401792049407959, "learning_rate": 4.207992296581608e-06, "loss": 1.2544, "step": 19896 }, { "epoch": 9.579783393501804, "grad_norm": 2.256636142730713, "learning_rate": 4.179104477611941e-06, "loss": 1.3103, "step": 19902 }, { "epoch": 9.582671480144404, "grad_norm": 2.2943310737609863, "learning_rate": 4.150216658642273e-06, "loss": 1.2298, "step": 19908 }, { "epoch": 9.585559566787003, "grad_norm": 2.4245386123657227, "learning_rate": 4.121328839672605e-06, "loss": 1.1794, "step": 19914 }, { "epoch": 9.588447653429602, "grad_norm": 2.251491069793701, "learning_rate": 4.092441020702937e-06, "loss": 1.2366, "step": 19920 }, { "epoch": 9.591335740072202, "grad_norm": 2.245015859603882, "learning_rate": 4.063553201733269e-06, "loss": 1.3019, "step": 19926 }, { "epoch": 9.594223826714801, "grad_norm": 2.509939670562744, "learning_rate": 4.0346653827636016e-06, "loss": 1.2061, "step": 19932 }, { "epoch": 9.5971119133574, "grad_norm": 2.4469003677368164, "learning_rate": 4.005777563793934e-06, "loss": 1.2124, "step": 19938 }, { "epoch": 9.6, "grad_norm": 2.3611714839935303, "learning_rate": 3.976889744824266e-06, "loss": 1.2976, "step": 19944 }, { "epoch": 9.602888086642599, "grad_norm": 2.4009909629821777, "learning_rate": 3.948001925854598e-06, "loss": 1.2515, "step": 19950 }, { "epoch": 9.605776173285198, "grad_norm": 2.2847542762756348, "learning_rate": 3.9191141068849305e-06, "loss": 1.2346, "step": 19956 }, { "epoch": 9.608664259927798, "grad_norm": 2.497036933898926, "learning_rate": 3.890226287915263e-06, "loss": 1.2457, "step": 19962 }, { "epoch": 9.611552346570397, "grad_norm": 2.2243359088897705, "learning_rate": 3.8613384689455945e-06, "loss": 1.2477, "step": 19968 }, { "epoch": 9.614440433212996, "grad_norm": 2.5689961910247803, "learning_rate": 3.832450649975927e-06, "loss": 1.2615, "step": 19974 }, { "epoch": 9.617328519855596, "grad_norm": 2.196061611175537, "learning_rate": 3.803562831006259e-06, "loss": 1.2868, "step": 19980 }, { "epoch": 9.620216606498195, "grad_norm": 2.2471742630004883, "learning_rate": 3.7746750120365914e-06, "loss": 1.2319, "step": 19986 }, { "epoch": 9.623104693140794, "grad_norm": 1.8405648469924927, "learning_rate": 3.7457871930669235e-06, "loss": 1.1748, "step": 19992 }, { "epoch": 9.625992779783394, "grad_norm": 2.7079999446868896, "learning_rate": 3.716899374097256e-06, "loss": 1.2952, "step": 19998 }, { "epoch": 9.628880866425993, "grad_norm": 2.3641443252563477, "learning_rate": 3.688011555127588e-06, "loss": 1.2141, "step": 20004 }, { "epoch": 9.631768953068592, "grad_norm": 2.3246045112609863, "learning_rate": 3.6591237361579204e-06, "loss": 1.3115, "step": 20010 }, { "epoch": 9.634657039711191, "grad_norm": 1.9954503774642944, "learning_rate": 3.6302359171882524e-06, "loss": 1.3166, "step": 20016 }, { "epoch": 9.63754512635379, "grad_norm": 2.402571201324463, "learning_rate": 3.601348098218585e-06, "loss": 1.2189, "step": 20022 }, { "epoch": 9.64043321299639, "grad_norm": 2.270003318786621, "learning_rate": 3.572460279248917e-06, "loss": 1.1893, "step": 20028 }, { "epoch": 9.64332129963899, "grad_norm": 2.128403425216675, "learning_rate": 3.5435724602792493e-06, "loss": 1.2742, "step": 20034 }, { "epoch": 9.646209386281589, "grad_norm": 2.1447091102600098, "learning_rate": 3.5146846413095813e-06, "loss": 1.1451, "step": 20040 }, { "epoch": 9.649097472924188, "grad_norm": 2.3995559215545654, "learning_rate": 3.4857968223399137e-06, "loss": 1.164, "step": 20046 }, { "epoch": 9.651985559566787, "grad_norm": 2.4563231468200684, "learning_rate": 3.4569090033702458e-06, "loss": 1.3851, "step": 20052 }, { "epoch": 9.654873646209387, "grad_norm": 2.525294542312622, "learning_rate": 3.428021184400578e-06, "loss": 1.3327, "step": 20058 }, { "epoch": 9.657761732851986, "grad_norm": 2.028841018676758, "learning_rate": 3.3991333654309102e-06, "loss": 1.1771, "step": 20064 }, { "epoch": 9.660649819494585, "grad_norm": 2.365696907043457, "learning_rate": 3.3702455464612427e-06, "loss": 1.278, "step": 20070 }, { "epoch": 9.663537906137185, "grad_norm": 2.2780346870422363, "learning_rate": 3.3413577274915743e-06, "loss": 1.302, "step": 20076 }, { "epoch": 9.666425992779784, "grad_norm": 2.292342185974121, "learning_rate": 3.312469908521907e-06, "loss": 1.2553, "step": 20082 }, { "epoch": 9.669314079422383, "grad_norm": 2.7302663326263428, "learning_rate": 3.2835820895522387e-06, "loss": 1.1942, "step": 20088 }, { "epoch": 9.672202166064983, "grad_norm": 2.676071882247925, "learning_rate": 3.2546942705825716e-06, "loss": 1.161, "step": 20094 }, { "epoch": 9.675090252707582, "grad_norm": 2.5632307529449463, "learning_rate": 3.225806451612903e-06, "loss": 1.206, "step": 20100 }, { "epoch": 9.677978339350181, "grad_norm": 2.478809118270874, "learning_rate": 3.196918632643236e-06, "loss": 1.2128, "step": 20106 }, { "epoch": 9.68086642599278, "grad_norm": 2.3033599853515625, "learning_rate": 3.1680308136735677e-06, "loss": 1.2519, "step": 20112 }, { "epoch": 9.683754512635378, "grad_norm": 2.238093614578247, "learning_rate": 3.1391429947039e-06, "loss": 1.2498, "step": 20118 }, { "epoch": 9.686642599277977, "grad_norm": 2.7841029167175293, "learning_rate": 3.110255175734232e-06, "loss": 1.1223, "step": 20124 }, { "epoch": 9.689530685920577, "grad_norm": 2.403424024581909, "learning_rate": 3.0813673567645646e-06, "loss": 1.1357, "step": 20130 }, { "epoch": 9.692418772563176, "grad_norm": 2.3034446239471436, "learning_rate": 3.0524795377948966e-06, "loss": 1.1616, "step": 20136 }, { "epoch": 9.695306859205775, "grad_norm": 2.740278720855713, "learning_rate": 3.0235917188252286e-06, "loss": 1.2645, "step": 20142 }, { "epoch": 9.698194945848375, "grad_norm": 2.888977289199829, "learning_rate": 2.994703899855561e-06, "loss": 1.3376, "step": 20148 }, { "epoch": 9.701083032490974, "grad_norm": 2.268575668334961, "learning_rate": 2.965816080885893e-06, "loss": 1.173, "step": 20154 }, { "epoch": 9.703971119133573, "grad_norm": 2.339740753173828, "learning_rate": 2.9369282619162255e-06, "loss": 1.2167, "step": 20160 }, { "epoch": 9.706859205776173, "grad_norm": 2.295725107192993, "learning_rate": 2.9080404429465575e-06, "loss": 1.3193, "step": 20166 }, { "epoch": 9.709747292418772, "grad_norm": 2.6091272830963135, "learning_rate": 2.87915262397689e-06, "loss": 1.3258, "step": 20172 }, { "epoch": 9.712635379061371, "grad_norm": 2.257789134979248, "learning_rate": 2.850264805007222e-06, "loss": 1.2064, "step": 20178 }, { "epoch": 9.71552346570397, "grad_norm": 2.2011852264404297, "learning_rate": 2.8213769860375544e-06, "loss": 1.0562, "step": 20184 }, { "epoch": 9.71841155234657, "grad_norm": 2.018612861633301, "learning_rate": 2.7924891670678864e-06, "loss": 1.1794, "step": 20190 }, { "epoch": 9.72129963898917, "grad_norm": 2.580288887023926, "learning_rate": 2.763601348098219e-06, "loss": 1.2002, "step": 20196 }, { "epoch": 9.724187725631769, "grad_norm": 2.3544936180114746, "learning_rate": 2.734713529128551e-06, "loss": 1.254, "step": 20202 }, { "epoch": 9.727075812274368, "grad_norm": 2.0447824001312256, "learning_rate": 2.705825710158883e-06, "loss": 1.2016, "step": 20208 }, { "epoch": 9.729963898916967, "grad_norm": 2.4457359313964844, "learning_rate": 2.6769378911892154e-06, "loss": 1.3403, "step": 20214 }, { "epoch": 9.732851985559567, "grad_norm": 2.1593923568725586, "learning_rate": 2.6480500722195474e-06, "loss": 1.2905, "step": 20220 }, { "epoch": 9.735740072202166, "grad_norm": 2.201643228530884, "learning_rate": 2.61916225324988e-06, "loss": 1.0454, "step": 20226 }, { "epoch": 9.738628158844765, "grad_norm": 2.8895528316497803, "learning_rate": 2.590274434280212e-06, "loss": 1.2768, "step": 20232 }, { "epoch": 9.741516245487365, "grad_norm": 2.2888689041137695, "learning_rate": 2.5613866153105443e-06, "loss": 1.106, "step": 20238 }, { "epoch": 9.744404332129964, "grad_norm": 2.139033317565918, "learning_rate": 2.5324987963408763e-06, "loss": 1.1288, "step": 20244 }, { "epoch": 9.747292418772563, "grad_norm": 2.434142827987671, "learning_rate": 2.5036109773712088e-06, "loss": 1.3009, "step": 20250 }, { "epoch": 9.750180505415162, "grad_norm": 2.657888174057007, "learning_rate": 2.4747231584015408e-06, "loss": 1.1796, "step": 20256 }, { "epoch": 9.753068592057762, "grad_norm": 1.8752620220184326, "learning_rate": 2.4458353394318732e-06, "loss": 1.1962, "step": 20262 }, { "epoch": 9.755956678700361, "grad_norm": 2.256107807159424, "learning_rate": 2.4169475204622052e-06, "loss": 1.2385, "step": 20268 }, { "epoch": 9.75884476534296, "grad_norm": 2.506643056869507, "learning_rate": 2.3880597014925373e-06, "loss": 1.3532, "step": 20274 }, { "epoch": 9.76173285198556, "grad_norm": 2.461879014968872, "learning_rate": 2.3591718825228697e-06, "loss": 1.1955, "step": 20280 }, { "epoch": 9.764620938628159, "grad_norm": 2.3208186626434326, "learning_rate": 2.3302840635532017e-06, "loss": 1.282, "step": 20286 }, { "epoch": 9.767509025270758, "grad_norm": 1.9891459941864014, "learning_rate": 2.301396244583534e-06, "loss": 1.23, "step": 20292 }, { "epoch": 9.770397111913358, "grad_norm": 2.2118470668792725, "learning_rate": 2.272508425613866e-06, "loss": 1.3848, "step": 20298 }, { "epoch": 9.773285198555957, "grad_norm": 2.091510534286499, "learning_rate": 2.2436206066441986e-06, "loss": 1.2262, "step": 20304 }, { "epoch": 9.776173285198556, "grad_norm": 2.447319507598877, "learning_rate": 2.2147327876745306e-06, "loss": 1.1995, "step": 20310 }, { "epoch": 9.779061371841156, "grad_norm": 2.52054762840271, "learning_rate": 2.185844968704863e-06, "loss": 1.3131, "step": 20316 }, { "epoch": 9.781949458483755, "grad_norm": 2.1203324794769287, "learning_rate": 2.156957149735195e-06, "loss": 1.1343, "step": 20322 }, { "epoch": 9.784837545126354, "grad_norm": 2.2780799865722656, "learning_rate": 2.1280693307655275e-06, "loss": 1.1423, "step": 20328 }, { "epoch": 9.787725631768954, "grad_norm": 2.0878875255584717, "learning_rate": 2.0991815117958596e-06, "loss": 1.2441, "step": 20334 }, { "epoch": 9.790613718411553, "grad_norm": 1.9169219732284546, "learning_rate": 2.0702936928261916e-06, "loss": 1.2983, "step": 20340 }, { "epoch": 9.793501805054152, "grad_norm": 2.319692611694336, "learning_rate": 2.041405873856524e-06, "loss": 1.2886, "step": 20346 }, { "epoch": 9.796389891696752, "grad_norm": 2.6197574138641357, "learning_rate": 2.012518054886856e-06, "loss": 1.1908, "step": 20352 }, { "epoch": 9.79927797833935, "grad_norm": 2.584380626678467, "learning_rate": 1.9836302359171885e-06, "loss": 1.2787, "step": 20358 }, { "epoch": 9.80216606498195, "grad_norm": 2.341965675354004, "learning_rate": 1.9547424169475205e-06, "loss": 1.2356, "step": 20364 }, { "epoch": 9.80505415162455, "grad_norm": 2.1206750869750977, "learning_rate": 1.925854597977853e-06, "loss": 1.2322, "step": 20370 }, { "epoch": 9.807942238267149, "grad_norm": 2.2975590229034424, "learning_rate": 1.896966779008185e-06, "loss": 1.3277, "step": 20376 }, { "epoch": 9.810830324909748, "grad_norm": 2.240583658218384, "learning_rate": 1.8680789600385172e-06, "loss": 1.1679, "step": 20382 }, { "epoch": 9.813718411552347, "grad_norm": 2.3908703327178955, "learning_rate": 1.8391911410688494e-06, "loss": 1.2889, "step": 20388 }, { "epoch": 9.816606498194945, "grad_norm": 2.1877002716064453, "learning_rate": 1.8103033220991817e-06, "loss": 1.3055, "step": 20394 }, { "epoch": 9.819494584837544, "grad_norm": 2.1622610092163086, "learning_rate": 1.781415503129514e-06, "loss": 1.1946, "step": 20400 }, { "epoch": 9.822382671480144, "grad_norm": 2.543584108352661, "learning_rate": 1.7525276841598461e-06, "loss": 1.2881, "step": 20406 }, { "epoch": 9.825270758122743, "grad_norm": 2.404539108276367, "learning_rate": 1.7236398651901784e-06, "loss": 1.1853, "step": 20412 }, { "epoch": 9.828158844765342, "grad_norm": 2.565101146697998, "learning_rate": 1.6947520462205106e-06, "loss": 1.2663, "step": 20418 }, { "epoch": 9.831046931407942, "grad_norm": 2.497796058654785, "learning_rate": 1.6658642272508428e-06, "loss": 1.3226, "step": 20424 }, { "epoch": 9.833935018050541, "grad_norm": 2.155773162841797, "learning_rate": 1.636976408281175e-06, "loss": 1.3225, "step": 20430 }, { "epoch": 9.83682310469314, "grad_norm": 2.215646743774414, "learning_rate": 1.6080885893115073e-06, "loss": 1.2775, "step": 20436 }, { "epoch": 9.83971119133574, "grad_norm": 2.4456937313079834, "learning_rate": 1.5792007703418393e-06, "loss": 1.1621, "step": 20442 }, { "epoch": 9.842599277978339, "grad_norm": 2.8060905933380127, "learning_rate": 1.5503129513721715e-06, "loss": 1.2299, "step": 20448 }, { "epoch": 9.845487364620938, "grad_norm": 2.5546038150787354, "learning_rate": 1.5214251324025036e-06, "loss": 1.2299, "step": 20454 }, { "epoch": 9.848375451263538, "grad_norm": 2.2446370124816895, "learning_rate": 1.4925373134328358e-06, "loss": 1.3042, "step": 20460 }, { "epoch": 9.851263537906137, "grad_norm": 2.345034122467041, "learning_rate": 1.463649494463168e-06, "loss": 1.238, "step": 20466 }, { "epoch": 9.854151624548736, "grad_norm": 2.3552098274230957, "learning_rate": 1.4347616754935002e-06, "loss": 1.3188, "step": 20472 }, { "epoch": 9.857039711191335, "grad_norm": 2.3168089389801025, "learning_rate": 1.4058738565238325e-06, "loss": 1.1818, "step": 20478 }, { "epoch": 9.859927797833935, "grad_norm": 2.3557307720184326, "learning_rate": 1.3769860375541647e-06, "loss": 1.2258, "step": 20484 }, { "epoch": 9.862815884476534, "grad_norm": 2.6589162349700928, "learning_rate": 1.348098218584497e-06, "loss": 1.3465, "step": 20490 }, { "epoch": 9.865703971119133, "grad_norm": 2.3183438777923584, "learning_rate": 1.3192103996148292e-06, "loss": 1.4191, "step": 20496 }, { "epoch": 9.868592057761733, "grad_norm": 2.5920135974884033, "learning_rate": 1.2903225806451614e-06, "loss": 1.1622, "step": 20502 }, { "epoch": 9.871480144404332, "grad_norm": 2.04130220413208, "learning_rate": 1.2614347616754936e-06, "loss": 1.3239, "step": 20508 }, { "epoch": 9.874368231046931, "grad_norm": 2.4695188999176025, "learning_rate": 1.2325469427058259e-06, "loss": 1.3346, "step": 20514 }, { "epoch": 9.87725631768953, "grad_norm": 2.267341136932373, "learning_rate": 1.2036591237361579e-06, "loss": 1.28, "step": 20520 }, { "epoch": 9.88014440433213, "grad_norm": 2.379171848297119, "learning_rate": 1.1747713047664901e-06, "loss": 1.2921, "step": 20526 }, { "epoch": 9.88303249097473, "grad_norm": 2.4448933601379395, "learning_rate": 1.1458834857968223e-06, "loss": 1.2723, "step": 20532 }, { "epoch": 9.885920577617329, "grad_norm": 2.8071084022521973, "learning_rate": 1.1169956668271546e-06, "loss": 1.1818, "step": 20538 }, { "epoch": 9.888808664259928, "grad_norm": 2.5098917484283447, "learning_rate": 1.0881078478574868e-06, "loss": 1.3448, "step": 20544 }, { "epoch": 9.891696750902527, "grad_norm": 2.482367753982544, "learning_rate": 1.059220028887819e-06, "loss": 1.2131, "step": 20550 }, { "epoch": 9.894584837545127, "grad_norm": 2.487971067428589, "learning_rate": 1.0303322099181513e-06, "loss": 1.3766, "step": 20556 }, { "epoch": 9.897472924187726, "grad_norm": 2.509375810623169, "learning_rate": 1.0014443909484835e-06, "loss": 1.2973, "step": 20562 }, { "epoch": 9.900361010830325, "grad_norm": 2.0165855884552, "learning_rate": 9.725565719788157e-07, "loss": 1.0617, "step": 20568 }, { "epoch": 9.903249097472925, "grad_norm": 2.2034783363342285, "learning_rate": 9.436687530091479e-07, "loss": 1.1605, "step": 20574 }, { "epoch": 9.906137184115524, "grad_norm": 2.674926519393921, "learning_rate": 9.147809340394801e-07, "loss": 1.2985, "step": 20580 }, { "epoch": 9.909025270758123, "grad_norm": 2.781643867492676, "learning_rate": 8.858931150698123e-07, "loss": 1.2798, "step": 20586 }, { "epoch": 9.911913357400723, "grad_norm": 2.177703619003296, "learning_rate": 8.570052961001446e-07, "loss": 1.101, "step": 20592 }, { "epoch": 9.914801444043322, "grad_norm": 2.28141713142395, "learning_rate": 8.281174771304768e-07, "loss": 1.2065, "step": 20598 }, { "epoch": 9.917689530685921, "grad_norm": 2.3427016735076904, "learning_rate": 7.99229658160809e-07, "loss": 1.209, "step": 20604 }, { "epoch": 9.92057761732852, "grad_norm": 1.973497748374939, "learning_rate": 7.703418391911411e-07, "loss": 1.0934, "step": 20610 }, { "epoch": 9.92346570397112, "grad_norm": 2.3422048091888428, "learning_rate": 7.414540202214733e-07, "loss": 1.1634, "step": 20616 }, { "epoch": 9.92635379061372, "grad_norm": 2.1770737171173096, "learning_rate": 7.125662012518055e-07, "loss": 1.1271, "step": 20622 }, { "epoch": 9.929241877256318, "grad_norm": 2.296694278717041, "learning_rate": 6.836783822821377e-07, "loss": 1.3701, "step": 20628 }, { "epoch": 9.932129963898918, "grad_norm": 2.438347101211548, "learning_rate": 6.5479056331247e-07, "loss": 1.1755, "step": 20634 }, { "epoch": 9.935018050541515, "grad_norm": 2.4034717082977295, "learning_rate": 6.259027443428022e-07, "loss": 1.1493, "step": 20640 }, { "epoch": 9.937906137184115, "grad_norm": 2.531728506088257, "learning_rate": 5.970149253731343e-07, "loss": 1.1656, "step": 20646 }, { "epoch": 9.940794223826714, "grad_norm": 2.172436237335205, "learning_rate": 5.681271064034665e-07, "loss": 1.2747, "step": 20652 }, { "epoch": 9.943682310469313, "grad_norm": 2.661789655685425, "learning_rate": 5.392392874337988e-07, "loss": 1.2295, "step": 20658 }, { "epoch": 9.946570397111913, "grad_norm": 2.3129892349243164, "learning_rate": 5.10351468464131e-07, "loss": 1.2107, "step": 20664 }, { "epoch": 9.949458483754512, "grad_norm": 2.8129472732543945, "learning_rate": 4.814636494944632e-07, "loss": 1.4251, "step": 20670 }, { "epoch": 9.952346570397111, "grad_norm": 2.4403560161590576, "learning_rate": 4.525758305247954e-07, "loss": 1.3063, "step": 20676 }, { "epoch": 9.95523465703971, "grad_norm": 2.532114267349243, "learning_rate": 4.2368801155512765e-07, "loss": 1.2486, "step": 20682 }, { "epoch": 9.95812274368231, "grad_norm": 2.604034662246704, "learning_rate": 3.948001925854598e-07, "loss": 1.2903, "step": 20688 }, { "epoch": 9.96101083032491, "grad_norm": 2.4501640796661377, "learning_rate": 3.65912373615792e-07, "loss": 1.3946, "step": 20694 }, { "epoch": 9.963898916967509, "grad_norm": 2.498872995376587, "learning_rate": 3.3702455464612424e-07, "loss": 1.2135, "step": 20700 }, { "epoch": 9.966787003610108, "grad_norm": 2.491468667984009, "learning_rate": 3.0813673567645647e-07, "loss": 1.2328, "step": 20706 }, { "epoch": 9.969675090252707, "grad_norm": 2.552875518798828, "learning_rate": 2.7924891670678864e-07, "loss": 1.2454, "step": 20712 }, { "epoch": 9.972563176895306, "grad_norm": 2.2347304821014404, "learning_rate": 2.503610977371209e-07, "loss": 1.2647, "step": 20718 }, { "epoch": 9.975451263537906, "grad_norm": 2.3349459171295166, "learning_rate": 2.2147327876745308e-07, "loss": 1.1757, "step": 20724 }, { "epoch": 9.978339350180505, "grad_norm": 2.179368734359741, "learning_rate": 1.9258545979778528e-07, "loss": 1.1906, "step": 20730 }, { "epoch": 9.981227436823104, "grad_norm": 2.6869025230407715, "learning_rate": 1.636976408281175e-07, "loss": 1.2625, "step": 20736 }, { "epoch": 9.984115523465704, "grad_norm": 2.1900277137756348, "learning_rate": 1.348098218584497e-07, "loss": 1.2562, "step": 20742 }, { "epoch": 9.987003610108303, "grad_norm": 2.6487183570861816, "learning_rate": 1.0592200288878191e-07, "loss": 1.2621, "step": 20748 }, { "epoch": 9.989891696750902, "grad_norm": 2.5393707752227783, "learning_rate": 7.703418391911412e-08, "loss": 1.3284, "step": 20754 }, { "epoch": 9.992779783393502, "grad_norm": 2.517340660095215, "learning_rate": 4.814636494944632e-08, "loss": 1.2272, "step": 20760 }, { "epoch": 9.995667870036101, "grad_norm": 2.670698642730713, "learning_rate": 1.925854597977853e-08, "loss": 1.3195, "step": 20766 } ], "logging_steps": 6, "max_steps": 20770, "num_input_tokens_seen": 0, "num_train_epochs": 10, "save_steps": 500, "total_flos": 1.6650231241009213e+18, "train_batch_size": 16, "trial_name": null, "trial_params": null }